[lvm-devel] LVM2/libdm libdevmapper.h libdm-deptree.c
zkabelac at sourceware.org
zkabelac at sourceware.org
Mon Oct 17 14:16:26 UTC 2011
CVSROOT: /cvs/lvm2
Module name: LVM2
Changes by: zkabelac at sourceware.org 2011-10-17 14:16:25
Modified files:
libdm : libdevmapper.h libdm-deptree.c
Log message:
Add thin_pool dm message support
Experimental support for kernel message via resume sequence.
Patches:
http://sourceware.org/cgi-bin/cvsweb.cgi/LVM2/libdm/libdevmapper.h.diff?cvsroot=lvm2&r1=1.160&r2=1.161
http://sourceware.org/cgi-bin/cvsweb.cgi/LVM2/libdm/libdm-deptree.c.diff?cvsroot=lvm2&r1=1.127&r2=1.128
--- LVM2/libdm/libdevmapper.h 2011/10/17 14:15:01 1.160
+++ LVM2/libdm/libdevmapper.h 2011/10/17 14:16:25 1.161
@@ -555,6 +555,42 @@
uint64_t low_water_mark_size,
unsigned skip_block_zeroing);
+/* Supported messages for thin provision target */
+typedef enum {
+ DM_THIN_MESSAGE_CREATE_SNAP,
+ DM_THIN_MESSAGE_CREATE_THIN,
+ DM_THIN_MESSAGE_DELETE,
+ DM_THIN_MESSAGE_SET_TRANSACTION_ID,
+ DM_THIN_MESSAGE_TRIM
+} dm_thin_message_t;
+
+struct dm_thin_message {
+ dm_thin_message_t type;
+ union {
+ struct {
+ uint32_t device_id;
+ uint32_t origin_id;
+ } m_create_snap;
+ struct {
+ uint32_t device_id;
+ } m_create_thin;
+ struct {
+ uint32_t device_id;
+ } m_delete;
+ struct {
+ uint64_t current_id;
+ uint64_t new_id;
+ } m_set_transaction_id;
+ struct {
+ uint32_t device_id;
+ uint64_t new_size;
+ } m_trim;
+ } u;
+};
+
+int dm_tree_node_add_thin_pool_message(struct dm_tree_node *node,
+ struct dm_thin_message *message);
+
int dm_tree_node_add_thin_target(struct dm_tree_node *node,
uint64_t size,
const char *thin_pool_uuid,
--- LVM2/libdm/libdm-deptree.c 2011/10/17 14:15:26 1.127
+++ LVM2/libdm/libdm-deptree.c 2011/10/17 14:16:25 1.128
@@ -108,6 +108,11 @@
uint32_t flags; /* Replicator sync log flags */
};
+struct thin_message {
+ struct dm_list list;
+ struct dm_thin_message message;
+};
+
/* Replicator-log has a list of sites */
/* FIXME: maybe move to seg_area too? */
struct replicator_site {
@@ -163,6 +168,7 @@
struct dm_tree_node *metadata; /* Thin_pool */
struct dm_tree_node *pool; /* Thin_pool, Thin */
+ struct dm_list thin_messages; /* Thin_pool */
uint64_t low_water_mark_size; /* Thin_pool */
uint32_t data_block_size; /* Thin_pool */
unsigned skip_block_zeroing; /* Thin_pool */
@@ -1211,50 +1217,159 @@
return r;
}
-static int _check_thin_pool_transaction_id(const char *name, uint32_t major, uint32_t minor,
- uint64_t transaction_id)
+static int _thin_pool_status_transaction_id(struct dm_tree_node *dnode, uint64_t *transaction_id)
{
struct dm_task *dmt;
int r = 0;
uint64_t start, length;
char *type = NULL;
char *params = NULL;
- uint64_t t_id = transaction_id; // FIXME: fake
- log_verbose("Checking transaction id %s (%" PRIu32 ":%" PRIu32 ")", name, major, minor);
+ if (!(dmt = dm_task_create(DM_DEVICE_STATUS)))
+ return_0;
- if (!(dmt = dm_task_create(DM_DEVICE_STATUS))) {
- log_debug("Device status dm_task creation failed for %s.", name);
- return 0;
+ if (!dm_task_set_major(dmt, dnode->info.major) ||
+ !dm_task_set_minor(dmt, dnode->info.minor)) {
+ log_error("Failed to set major minor.");
+ goto out;
}
- if (!dm_task_set_name(dmt, name)) {
- log_debug("Failed to set device name for %s status.", name);
+ if (!dm_task_run(dmt))
+ goto_out;
+
+ dm_get_next_target(dmt, NULL, &start, &length, &type, ¶ms);
+
+ if (type && (strcmp(type, "thin-pool") != 0)) {
+ log_error(INTERNAL_ERROR
+ "Expected thin-pool target for %d:%d and got %s.",
+ dnode->info.major, dnode->info.minor, type);
goto out;
}
- if (!dm_task_set_major_minor(dmt, major, minor, 1)) {
- log_error("Failed to set device number for %s status.", name);
+ if (!params || (sscanf(params, "%" PRIu64, transaction_id) != 1)) {
+ log_error(INTERNAL_ERROR
+ "Failed to parse transaction_id from %s.", params);
goto out;
}
- if (!dm_task_no_open_count(dmt))
- log_error("Failed to disable open_count");
+ log_debug("Thin pool transaction id: %" PRIu64 " status: %s.", *transaction_id, params);
- if (!(r = dm_task_run(dmt)))
- goto_out;
+ r = 1;
+out:
+ dm_task_destroy(dmt);
- dm_get_next_target(dmt, NULL, &start, &length, &type, ¶ms);
- log_verbose("PARSE params %s", params); // FIXME: parse status
+ return r;
+}
- r = (transaction_id == t_id);
+static int _thin_pool_node_message(struct dm_tree_node *dnode, struct thin_message *tm)
+{
+ struct dm_task *dmt;
+ struct dm_thin_message *m = &tm->message;
+ char buf[64];
+ int r;
+ switch (m->type) {
+ case DM_THIN_MESSAGE_CREATE_SNAP:
+ r = dm_snprintf(buf, sizeof(buf), "create_snap %u %u",
+ m->u.m_create_snap.device_id,
+ m->u.m_create_snap.origin_id);
+ break;
+ case DM_THIN_MESSAGE_CREATE_THIN:
+ r = dm_snprintf(buf, sizeof(buf), "create_thin %u",
+ m->u.m_create_thin.device_id);
+ break;
+ case DM_THIN_MESSAGE_DELETE:
+ r = dm_snprintf(buf, sizeof(buf), "delete %u",
+ m->u.m_delete.device_id);
+ break;
+ case DM_THIN_MESSAGE_TRIM:
+ r = dm_snprintf(buf, sizeof(buf), "trim %u %" PRIu64,
+ m->u.m_trim.device_id,
+ m->u.m_trim.new_size);
+ break;
+ case DM_THIN_MESSAGE_SET_TRANSACTION_ID:
+ r = dm_snprintf(buf, sizeof(buf),
+ "set_transaction_id %" PRIu64 " %" PRIu64,
+ m->u.m_set_transaction_id.current_id,
+ m->u.m_set_transaction_id.new_id);
+ break;
+ }
+
+ if (!r) {
+ log_error("Failed to prepare message.");
+ return 0;
+ }
+
+ r = 0;
+
+ if (!(dmt = dm_task_create(DM_DEVICE_TARGET_MSG)))
+ return_0;
+
+ if (!dm_task_set_major(dmt, dnode->info.major) ||
+ !dm_task_set_minor(dmt, dnode->info.minor)) {
+ log_error("Failed to set message major minor.");
+ goto out;
+ }
+
+ if (!dm_task_set_message(dmt, buf))
+ goto_out;
+
+ if (!dm_task_run(dmt))
+ goto_out;
+
+ r = 1;
out:
dm_task_destroy(dmt);
return r;
}
+static int _thin_pool_node_send_messages(struct dm_tree_node *dnode,
+ const char *uuid_prefix,
+ size_t uuid_prefix_len)
+{
+ struct load_segment *seg;
+ struct thin_message *tmsg;
+ uint64_t current_id;
+ const char *uuid;
+
+ if ((dnode == &dnode->dtree->root) || /* root has rops.segs uninitialized */
+ (dm_list_size(&dnode->props.segs) != 1))
+ return 1;
+
+ seg = dm_list_item(dm_list_last(&dnode->props.segs), struct load_segment);
+
+ if (seg->type != SEG_THIN_POOL)
+ return 1;
+
+ if (!(uuid = dm_tree_node_get_uuid(dnode)))
+ return_0;
+
+ if (!_uuid_prefix_matches(uuid, uuid_prefix, uuid_prefix_len)) {
+ log_debug("UUID \"%s\" does not match.", uuid);
+ return 1;
+ }
+
+ if (!_thin_pool_status_transaction_id(dnode, ¤t_id))
+ return_0;
+
+ log_debug("Expecting transaction_id: %" PRIu64, dnode->props.thin_pool_transaction_id);
+ if (current_id == dnode->props.thin_pool_transaction_id)
+ return 1; /* In sync - skip messages */
+
+ if (current_id != (dnode->props.thin_pool_transaction_id - 1)) {
+ log_error("Thin pool transaction_id=%" PRIu64 ", while expected: %" PRIu64 ".",
+ current_id, dnode->props.thin_pool_transaction_id - 1);
+ return 0; /* Nothing to send */
+ }
+
+ dm_list_iterate_items(tmsg, &seg->thin_messages)
+ if (!(_thin_pool_node_message(dnode, tmsg)))
+ return_0;
+
+ return 1;
+}
+
/*
* FIXME Don't attempt to deactivate known internal dependencies.
*/
@@ -2205,6 +2320,13 @@
dm_tree_set_cookie(dnode, 0);
}
+ if (r && !_thin_pool_node_send_messages(dnode, uuid_prefix, uuid_prefix_len)) {
+ stack;
+ if (!(dm_tree_deactivate_children(dnode, uuid_prefix, uuid_prefix_len)))
+ log_error("Failed to deactivate %s", dnode->name);
+ r = 0;
+ }
+
return r;
}
@@ -2744,6 +2866,82 @@
seg->low_water_mark_size = low_water_mark_size;
seg->data_block_size = data_block_size;
seg->skip_block_zeroing = skip_block_zeroing;
+ dm_list_init(&seg->thin_messages);
+
+ return 1;
+}
+
+int dm_tree_node_add_thin_pool_message(struct dm_tree_node *node,
+ struct dm_thin_message *message)
+{
+ struct load_segment *seg;
+ struct thin_message *tm;
+
+ if (node->props.segment_count != 1) {
+ log_error(INTERNAL_ERROR "Attempt to use non thin pool segment.");
+ return 0;
+ }
+
+ seg = dm_list_item(dm_list_last(&node->props.segs), struct load_segment);
+
+ if (seg->type != SEG_THIN_POOL) {
+ log_error(INTERNAL_ERROR "Attempt to use non thin pool segment %s.",
+ dm_segtypes[seg->type].target);
+ return 0;
+ }
+
+ if (!(tm = dm_pool_zalloc(node->dtree->mem, sizeof (*tm)))) {
+ log_error("Failed to allocate thin message.");
+ return 0;
+ }
+
+ switch (message->type) {
+ case DM_THIN_MESSAGE_CREATE_SNAP:
+ /* Origin MUST be suspend! */
+ if (message->u.m_create_snap.device_id == message->u.m_create_snap.origin_id) {
+ log_error("Same origin used for thin snapshot.");
+ return 0;
+ }
+ if (!_thin_validate_device_id(message->u.m_create_snap.device_id) ||
+ !_thin_validate_device_id(message->u.m_create_snap.origin_id))
+ return_0;
+ tm->message.u.m_create_snap.device_id = message->u.m_create_snap.device_id;
+ tm->message.u.m_create_snap.origin_id = message->u.m_create_snap.origin_id;
+ break;
+ case DM_THIN_MESSAGE_CREATE_THIN:
+ if (!_thin_validate_device_id(message->u.m_create_thin.device_id))
+ return_0;
+ tm->message.u.m_create_thin.device_id = message->u.m_create_thin.device_id;
+ break;
+ case DM_THIN_MESSAGE_DELETE:
+ if (!_thin_validate_device_id(message->u.m_delete.device_id))
+ return_0;
+ tm->message.u.m_delete.device_id = message->u.m_delete.device_id;
+ break;
+ case DM_THIN_MESSAGE_TRIM:
+ if (!_thin_validate_device_id(message->u.m_trim.device_id))
+ return_0;
+ tm->message.u.m_trim.device_id = message->u.m_trim.device_id;
+ tm->message.u.m_trim.new_size = message->u.m_trim.new_size;
+ break;
+ case DM_THIN_MESSAGE_SET_TRANSACTION_ID:
+ if (message->u.m_set_transaction_id.current_id !=
+ (message->u.m_set_transaction_id.new_id - 1)) {
+ log_error("New transaction_id must be sequential.");
+ return 0; /* FIXME: Maybe too strict here? */
+ }
+ tm->message.u.m_set_transaction_id.current_id =
+ message->u.m_set_transaction_id.current_id;
+ tm->message.u.m_set_transaction_id.new_id =
+ message->u.m_set_transaction_id.new_id;
+ break;
+ default:
+ log_error("Unsupported message type %d.", (int) message->type);
+ return 0;
+ }
+
+ tm->message.type = message->type;
+ dm_list_add(&seg->thin_messages, &tm->list);
return 1;
}
More information about the lvm-devel
mailing list