From: jbrassow@sourceware.org <jbrassow@sourceware.org>
To: cluster-devel.redhat.com
Subject: [Cluster-devel] cluster/cmirror-kernel src/dm-clog-tfr.c src/d ...
Date: 23 Aug 2007 19:54:59 -0000 [thread overview]
Message-ID: <20070823195459.29033.qmail@sourceware.org> (raw)
CVSROOT: /cvs/cluster
Module name: cluster
Branch: RHEL5
Changes by: jbrassow at sourceware.org 2007-08-23 19:54:58
Modified files:
cmirror-kernel/src: dm-clog-tfr.c dm-clog-tfr.h dm-clog.c
Added files:
cmirror-kernel : Makefile configure
cmirror-kernel/make: defines.mk.input release.mk.input
cmirror-kernel/scripts: uninstall.pl
cmirror-kernel/src: Makefile
Log message:
- RHEL5 cluster mirror module
Patches:
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/cmirror-kernel/Makefile.diff?cvsroot=cluster&only_with_tag=RHEL5&r1=NONE&r2=1.1.6.1
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/cmirror-kernel/configure.diff?cvsroot=cluster&only_with_tag=RHEL5&r1=NONE&r2=1.1.6.1
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/cmirror-kernel/make/defines.mk.input.diff?cvsroot=cluster&only_with_tag=RHEL5&r1=NONE&r2=1.1.6.1
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/cmirror-kernel/make/release.mk.input.diff?cvsroot=cluster&only_with_tag=RHEL5&r1=NONE&r2=1.1.6.1
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/cmirror-kernel/scripts/uninstall.pl.diff?cvsroot=cluster&only_with_tag=RHEL5&r1=NONE&r2=1.1.6.1
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/cmirror-kernel/src/Makefile.diff?cvsroot=cluster&only_with_tag=RHEL5&r1=NONE&r2=1.1.6.1
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/cmirror-kernel/src/dm-clog-tfr.c.diff?cvsroot=cluster&only_with_tag=RHEL5&r1=1.1&r2=1.1.2.1
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/cmirror-kernel/src/dm-clog-tfr.h.diff?cvsroot=cluster&only_with_tag=RHEL5&r1=1.1&r2=1.1.2.1
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/cmirror-kernel/src/dm-clog.c.diff?cvsroot=cluster&only_with_tag=RHEL5&r1=1.2&r2=1.2.2.1
--- cluster/cmirror-kernel/src/dm-clog-tfr.c 2006/06/21 01:41:43 1.1
+++ cluster/cmirror-kernel/src/dm-clog-tfr.c 2007/08/23 19:54:57 1.1.2.1
@@ -4,19 +4,106 @@
* This file is released under the LGPL.
*/
-#include "dm-clog-tfr.h"
+#include <linux/kernel.h>
+#include <linux/module.h>
+#include <net/sock.h>
+#include <linux/workqueue.h>
+#include <linux/dm-clog-tfr.h>
+#include <linux/connector.h>
+#include "dm.h"
+
+static uint64_t seq = 0;
/*
- * Pre-allocated nominal request area for speed
+ * Pre-allocated space for speed
*/
-#define DM_CLOG_NOMINAL_REQUEST_SIZE 512
-static char nominal_request[DM_CLOG_NOMINAL_REQUEST_SIZE];
+#define DM_CLOG_PREALLOCED_SIZE 512
+static struct cn_msg *prealloced_cn_msg = NULL;
+static struct clog_tfr *prealloced_clog_tfr = NULL;
+
+static struct cb_id cn_clog_id = { 0x4, 0x1 };
+static DEFINE_MUTEX(_lock);
+
+struct recieving_pkg {
+ struct list_head list;
+ struct completion complete;
+
+ uint64_t seq;
+
+ int error;
+ int *data_size;
+ char *data;
+};
+
+static spinlock_t recieving_list_lock = SPIN_LOCK_UNLOCKED;
+static struct list_head recieving_list;
-static DECLARE_MUTEX(consult_server_lock);
+static int dm_clog_sendto_server(struct clog_tfr *tfr)
+{
+ int r;
+ int size;
+ struct cn_msg *msg = prealloced_cn_msg;
+
+ if (tfr != prealloced_clog_tfr) {
+ size = sizeof(*msg) + sizeof(struct clog_tfr) + tfr->data_size;
+ msg = kmalloc(size, GFP_NOIO);
+ if (!msg)
+ return -ENOMEM;
+ memcpy((msg + 1), tfr, sizeof(struct clog_tfr) + tfr->data_size);
+ }
+
+ memset(msg, 0, sizeof(*msg));
+
+ msg->id.idx = cn_clog_id.idx;
+ msg->id.val = cn_clog_id.val;
+ msg->ack = 0;
+ msg->seq = 0;
+ msg->len = sizeof(struct clog_tfr) + tfr->data_size;
+
+ r = cn_netlink_send(msg, 0, gfp_any());
+ if (msg != prealloced_cn_msg)
+ kfree(msg);
+
+ return r;
+}
+
+static void cn_clog_callback(void *data)
+{
+ struct cn_msg *msg = (struct cn_msg *)data;
+ struct clog_tfr *tfr = (struct clog_tfr *)(msg + 1);
+ struct recieving_pkg *pkg;
+
+ spin_lock(&recieving_list_lock);
+ if (msg->len < sizeof(*tfr)) {
+ DMERR("Incomplete message recieved from cluster log server");
+ spin_unlock(&recieving_list_lock);
+ return;
+ }
+
+ list_for_each_entry(pkg, &recieving_list, list) {
+ if (tfr->seq == pkg->seq) {
+ if (tfr->data_size > *(pkg->data_size)) {
+ DMERR("Insufficient space to recieve package [%s]",
+ RQ_TYPE(tfr->request_type));
+ *(pkg->data_size) = 0;
+ pkg->error = -ENOSPC;
+ } else {
+ pkg->error = tfr->error;
+ memcpy(pkg->data, tfr->data, tfr->data_size);
+ *(pkg->data_size) = tfr->data_size;
+ }
+
+ complete(&pkg->complete);
+ break;
+ }
+ }
+
+ spin_unlock(&recieving_list_lock);
+}
/*
* dm_clog_consult_server
- * @uuid: log's uuid (must be MAX_NAME_LEN in size)
+ * @uuid: log's uuid (must be DM_UUID_LEN in size)
* @request_type:
* @data: data to tx to the server
* @data_size: size of data in bytes
@@ -24,60 +111,117 @@
* @rdata_size: value-result (amount of space given/amount of space used)
*
* Only one process at a time can communicate with the server.
- * Possible error return values:
- * +XXX: Server-side error
- * -XXX: Client-side error
- * -ENOSPC: Not enough space in rdata
- * -ENOMEM: Unable to allocate memory to complete request
- * -ESRCH: Unable to contact server
- * EIO: Server unable to commit request
+ * rdata_size is undefined on failure.
*
- * Returns: 0 on success, otherwise failure
+ * Returns: 0 on success, -EXXX on failure
*/
int dm_clog_consult_server(const char *uuid, int request_type,
char *data, int data_size,
char *rdata, int *rdata_size)
{
int r = 0;
- struct clog_tfr *tfr = (struct clog_tfr *)nominal_request;
+ int dummy = 0;
+ int overhead_size = sizeof(struct clog_tfr *) + sizeof(struct cn_msg);
+ struct clog_tfr *tfr = prealloced_clog_tfr;
+ struct recieving_pkg pkg;
- mutex_lock(&consult_server_lock);
- if (data_size > (DM_CLOG_NOMINAL_REQUEST_SIZE - sizeof(*tfr)))
+ if (data_size > (DM_CLOG_PREALLOCED_SIZE - overhead_size)) {
+ DMINFO("Size of tfr exceeds preallocated size");
/* FIXME: is kmalloc sufficient if we need this much space? */
- tfr = kmalloc(data_size + sizeof(*tfr), GFP_KERNEL);
+ tfr = kmalloc(data_size + sizeof(*tfr), GFP_NOIO);
+ }
if (!tfr)
return -ENOMEM;
- memcpy(tfr->uuid, uuid, MAX_NAME_LEN);
+ if (!rdata_size)
+ rdata_size = &dummy;
+resend:
+ /*
+ * We serialize the sending of requests so we can
+ * use the preallocated space.
+ */
+ mutex_lock(&_lock);
+
+ memset(tfr, 0, data_size + sizeof(*tfr));
+ memcpy(tfr->uuid, uuid, DM_UUID_LEN);
+ tfr->seq = seq++;
tfr->request_type = request_type;
tfr->data_size = data_size;
+ if (data && data_size)
+ memcpy(tfr->data, data, data_size);
- /*
- * FIXME: Send to server
- */
+ memset(&pkg, 0, sizeof(pkg));
+ init_completion(&pkg.complete);
+ pkg.seq = tfr->seq;
+ pkg.data_size = rdata_size;
+ pkg.data = rdata;
+ spin_lock(&recieving_list_lock);
+ list_add(&(pkg.list), &recieving_list);
+ spin_unlock(&recieving_list_lock);
+
+ r = dm_clog_sendto_server(tfr);
+
+ mutex_unlock(&_lock);
+
+ if (r) {
+ DMERR("Unable to send cluster log request [%s] to server: %d",
+ RQ_TYPE(request_type), r);
+ goto out;
+ }
- if (rdata) {
- /* FIXME: receive from server */
- if (tfr->error) {
- r = tfr->error;
- } else if (tfr->data_size > *rdata_size) {
- r = -ENOSPC;
- } else {
- *rdata_size = tfr->data_size;
- memcpy(rdata, tft->data, tfr->data_size);
- }
- /* FIXME: If using netlink, we may wish to ack back */
- } else {
- /*
- * FIXME: If we are using netlink, we may want an
- * ack from the server to know that it got the
- * request. (Ack is implicit if we are receiving
- * data.)
- */
+ r = wait_for_completion_timeout(&(pkg.complete), 5 * HZ);
+ if (!r) {
+ DMWARN("Timed out waiting for cluster log server [%s]",
+ RQ_TYPE(request_type));
+ DMWARN("Retrying request [%s]", RQ_TYPE(request_type));
+ spin_lock(&recieving_list_lock);
+ list_del_init(&(pkg.list));
+ spin_unlock(&recieving_list_lock);
+
+ goto resend;
}
- r = ENOSYS;
+ r = pkg.error;
+ if (r)
+ DMERR("Server error while processing request [%s]: %d",
+ RQ_TYPE(request_type), r);
+
+out:
+ spin_lock(&recieving_list_lock);
+ list_del_init(&(pkg.list));
+ spin_unlock(&recieving_list_lock);
+
+ if (tfr != (struct clog_tfr *)prealloced_clog_tfr)
+ kfree(tfr);
- mutex_unlock(&consult_server_lock);
return r;
}
+
+int dm_clog_tfr_init(void)
+{
+ int r;
+ void *prealloced;
+
+ INIT_LIST_HEAD(&recieving_list);
+
+ prealloced = kmalloc(DM_CLOG_PREALLOCED_SIZE, GFP_KERNEL);
+ if (!prealloced)
+ return -ENOMEM;
+
+ prealloced_cn_msg = prealloced;
+ prealloced_clog_tfr = prealloced + sizeof(struct cn_msg);
+
+ r = cn_add_callback(&cn_clog_id, "clulog", cn_clog_callback);
+ if (r) {
+ cn_del_callback(&cn_clog_id);
+ return r;
+ }
+
+ return 0;
+}
+
+void dm_clog_tfr_exit(void)
+{
+ cn_del_callback(&cn_clog_id);
+ kfree(prealloced_cn_msg);
+}
--- cluster/cmirror-kernel/src/dm-clog-tfr.h 2006/06/21 01:41:43 1.1
+++ cluster/cmirror-kernel/src/dm-clog-tfr.h 2007/08/23 19:54:57 1.1.2.1
@@ -5,6 +5,11 @@
*/
#ifndef __DM_CLOG_TFR_H__
+#define __DM_CLOG_TFR_H__
+
+#include <linux/dm-ioctl.h> /* For DM_UUID_LEN */
+
+#define DM_CLOG_TFR_SIZE 1024
#define DM_CLOG_CTR 1
#define DM_CLOG_DTR 2
@@ -13,28 +18,56 @@
#define DM_CLOG_RESUME 5
#define DM_CLOG_GET_REGION_SIZE 6
#define DM_CLOG_IS_CLEAN 7
-#define DM_CLOG_IS_REMOTE_RECOVERING 8
-#define DM_CLOG_IN_SYNC 9
-#define DM_CLOG_FLUSH 10
-#define DM_CLOG_MARK_REGION 11
-#define DM_CLOG_CLEAR_REGION 12
-#define DM_CLOG_GET_RESYNC_WORK 13
-#define DM_CLOG_SET_REGION_SYNC 14
-#define DM_CLOG_GET_SYNC_COUNT 15
-#define DM_CLOG_STATUS 16
-#define DM_CLOG_GET_FAILURE_RESPONSE 17
+#define DM_CLOG_IN_SYNC 8
+#define DM_CLOG_FLUSH 9
+#define DM_CLOG_MARK_REGION 10
+#define DM_CLOG_CLEAR_REGION 11
+#define DM_CLOG_GET_RESYNC_WORK 12
+#define DM_CLOG_SET_REGION_SYNC 13
+#define DM_CLOG_GET_SYNC_COUNT 14
+#define DM_CLOG_STATUS_INFO 15
+#define DM_CLOG_STATUS_TABLE 16
+
+#define RQ_TYPE(x) \
+ ((x) == DM_CLOG_CTR) ? "DM_CLOG_CTR" : \
+ ((x) == DM_CLOG_DTR) ? "DM_CLOG_DTR" : \
+ ((x) == DM_CLOG_PRESUSPEND) ? "DM_CLOG_PRESUSPEND" : \
+ ((x) == DM_CLOG_POSTSUSPEND) ? "DM_CLOG_POSTSUSPEND" : \
+ ((x) == DM_CLOG_RESUME) ? "DM_CLOG_RESUME" : \
+ ((x) == DM_CLOG_GET_REGION_SIZE) ? "DM_CLOG_GET_REGION_SIZE" : \
+ ((x) == DM_CLOG_IS_CLEAN) ? "DM_CLOG_IS_CLEAN" : \
+ ((x) == DM_CLOG_IN_SYNC) ? "DM_CLOG_IN_SYNC" : \
+ ((x) == DM_CLOG_FLUSH) ? "DM_CLOG_FLUSH" : \
+ ((x) == DM_CLOG_MARK_REGION) ? "DM_CLOG_MARK_REGION" : \
+ ((x) == DM_CLOG_CLEAR_REGION) ? "DM_CLOG_CLEAR_REGION" : \
+ ((x) == DM_CLOG_GET_RESYNC_WORK) ? "DM_CLOG_GET_RESYNC_WORK" : \
+ ((x) == DM_CLOG_SET_REGION_SYNC) ? "DM_CLOG_SET_REGION_SYNC" : \
+ ((x) == DM_CLOG_GET_SYNC_COUNT) ? "DM_CLOG_GET_SYNC_COUNT" : \
+ ((x) == DM_CLOG_STATUS_INFO) ? "DM_CLOG_STATUS_INFO" : \
+ ((x) == DM_CLOG_STATUS_TABLE) ? "DM_CLOG_STATUS_TABLE" : \
+ NULL
struct clog_tfr {
- char uuid[MAX_NAME_LEN];
- int error; /* Used by server to inform of errors */
- int request_type;
- int data_size;
+ uint64_t private[2];
+ char uuid[DM_UUID_LEN]; /* Ties a request to a specific mirror log */
+
+ int error; /* Used by server to inform of errors */
+ uint32_t originator; /* Cluster ID of this machine */
+
+ uint64_t seq; /* Sequence number for request */
+ uint32_t request_type; /* DM_CLOG_* */
+ uint32_t data_size; /* How much data (not including this struct) */
char data[0];
};
+#ifdef __KERNEL__
+#define DM_MSG_PREFIX "clulog"
+int dm_clog_tfr_init(void);
+void dm_clog_tfr_exit(void);
int dm_clog_consult_server(const char *uuid, int request_type,
char *data, int data_size,
char *rdata, int *rdata_size);
+#endif
#endif /* __DM_CLOG_TFR_H__ */
--- cluster/cmirror-kernel/src/dm-clog.c 2006/06/26 20:10:51 1.2
+++ cluster/cmirror-kernel/src/dm-clog.c 2007/08/23 19:54:57 1.2.2.1
@@ -4,7 +4,9 @@
* This file is released under the LGPL.
*/
-#include "dm-clog-tfr.h"
+#include "dm.h"
+#include "dm-log.h"
+#include <linux/dm-clog-tfr.h>
struct flush_entry {
int type;
@@ -16,8 +18,7 @@
struct dm_target *ti;
uint32_t region_size;
region_t region_count;
- int failure_response;
- char uuid[MAX_NAME_LEN];
+ char uuid[DM_UUID_LEN];
spinlock_t flush_lock;
struct list_head flush_list; /* only for clear and mark requests */
@@ -25,7 +26,7 @@
static mempool_t *flush_entry_pool = NULL;
-static void *flush_entry_alloc(int gfp_mask, void *pool_data)
+static void *flush_entry_alloc(gfp_t gfp_mask, void *pool_data)
{
return kmalloc(sizeof(struct flush_entry), gfp_mask);
}
@@ -40,24 +41,17 @@
{
int i;
int r = 0;
- int failure_response = FR_NONBLOCK;
+ int str_size;
+ char *ctr_str = NULL;
struct log_c *lc = NULL;
uint32_t region_size;
region_t region_count;
/* Already checked argument count */
-
- /* Check for block_on_error. It must be present. */
- for (i = 1; i < argc; i++) {
- if (!strcmp(argv[i], "block_on_error"))
- failure_response = FR_BLOCK;
- }
- if (failure_response != FR_BLOCK) {
- DMWARN("Required \"block_on_error\" argument not supplied.");
+ if (disk_log != 0 && disk_log != 1)
return -EINVAL;
- }
- if (sscanf(argv[0], SECTOR_FORMAT, ®ion_size) != 1) {
+ if (sscanf(argv[disk_log], "%u", ®ion_size) != 1) {
DMWARN("Invalid region size string");
return -EINVAL;
}
@@ -69,17 +63,47 @@
DMWARN("Unable to allocate cluster log context.");
return -ENOMEM;
}
+
lc->ti = ti;
lc->region_size = region_size;
lc->region_count = region_count;
- /* FIXME: Send table string to server */
+ /* FIXME: Need to check size of uuid arg */
+ memcpy(lc->uuid, argv[1 + disk_log], DM_UUID_LEN);
+ spin_lock_init(&lc->flush_lock);
+ INIT_LIST_HEAD(&lc->flush_list);
+
+ for (i = 0, str_size = 0; i < argc; i++)
+ str_size += strlen(argv[i]) + 1; /* +1 for space between args */
+
+ str_size += 20; /* Max number of chars in a printed u64 number */
+
+ ctr_str = kmalloc(str_size, GFP_KERNEL);
+ if (!ctr_str) {
+ DMWARN("Unable to allocate memory for constructor string");
+ kfree(lc);
+ return -ENOMEM;
+ }
-fail:
- if (lc)
+ for (i = 0, str_size = 0; i < argc; i++)
+ str_size += sprintf(ctr_str + str_size, "%s ", argv[i]);
+ str_size += sprintf(ctr_str + str_size, "%llu", ti->len);
+
+ /* Send table string */
+ r = dm_clog_consult_server(lc->uuid, DM_CLOG_CTR,
+ ctr_str, str_size, NULL, NULL);
+
+ if (r == -ESRCH)
+ DMERR(" Userspace cluster log server not found");
+
+ log->context = lc;
+
+ if (r && lc)
kfree(lc);
-
- return -ENOSYS;
+ if (ctr_str)
+ kfree(ctr_str);
+
+ return r;
}
/*
@@ -90,24 +114,26 @@
* @argv
*
* argv contains:
- * <region_size> <uuid> [[no]sync] "block_on_error"
+ * <region_size> <uuid> [[no]sync] <block_on_error>
*
* Returns: 0 on success, -XXX on failure
*/
static int cluster_core_ctr(struct dirty_log *log, struct dm_target *ti,
unsigned int argc, char **argv)
{
- int i;
+ int i, r;
if ((argc < 3) || (argc > 4)) {
DMERR("Too %s arguments to clustered_core mirror log type.",
- (argc < 3) ? "few" : "many");
+ (argc < 2) ? "few" : "many");
DMERR(" %d arguments supplied:", argc);
for (i = 0; i < argc; i++)
DMERR(" %s", argv[i]);
return -EINVAL;
}
- return cluster_ctr(log, ti, argc, argv, 0);
+ r = cluster_ctr(log, ti, argc, argv, 0);
+
+ return r;
}
@@ -119,7 +145,7 @@
* @argv
*
* argv contains:
- * <disk> <region_size> <uuid> [[no]sync] "block_on_error"
+ * <disk> <region_size> <uuid> [[no]sync] <block_on_error>
*
* Returns: 0 on success, -XXX on failure
*/
@@ -129,7 +155,7 @@
int i;
if ((argc < 4) || (argc > 5)) {
DMERR("Too %s arguments to clustered_disk mirror log type.",
- (argc < 4) ? "few" : "many");
+ (argc < 3) ? "few" : "many");
DMERR(" %d arguments supplied:", argc);
for (i = 0; i < argc; i++)
DMERR(" %s", argv[i]);
@@ -153,6 +179,7 @@
NULL, NULL);
/* FIXME: What do we do on failure? */
+
kfree(lc);
return;
@@ -171,7 +198,7 @@
NULL, 0,
NULL, NULL);
- return (r > 0) ? -r : r;
+ return r;
}
/*
@@ -187,7 +214,7 @@
NULL, 0,
NULL, NULL);
- return (r > 0) ? -r : r;
+ return r;
}
/*
@@ -203,7 +230,7 @@
NULL, 0,
NULL, NULL);
- return (r > 0) ? -r : r;
+ return r;
}
/*
@@ -247,32 +274,6 @@
}
/*
- * cluster_is_remote_recovering
- * @log
- * @region
- *
- * Check whether a region is being resync'ed on a remote node.
- * If there is any sort of failure when consulting the server,
- * we assume that the region is being remotely recovered.
- *
- * Returns: 1 if remote recovering, 0 otherwise
- */
-static int cluster_is_remote_recovering(struct dirty_log *log, region_t region)
-{
- int r;
- int is_recovering;
- int rdata_size;
- struct log_c *lc = (struct log_c *)log->context;
-
- rdata_size = sizeof(is_recovering);
- r = dm_clog_consult_server(lc->uuid, DM_CLOG_IS_REMOTE_RECOVERING,
- (char *)®ion, sizeof(region),
- (char *)&is_recovering, &rdata_size);
-
- return (r) ? 1 : is_recovering;
-}
-
-/*
* cluster_in_sync
* @log
* @region
@@ -298,7 +299,6 @@
r = dm_clog_consult_server(lc->uuid, DM_CLOG_IN_SYNC,
(char *)®ion, sizeof(region),
(char *)&in_sync, &rdata_size);
-
return (r) ? 0 : in_sync;
}
@@ -324,15 +324,17 @@
{
int r = 0;
int flags;
- region_t region;
struct log_c *lc = (struct log_c *)log->context;
- struct list_head flush_list;
+ LIST_HEAD(flush_list);
struct flush_entry *fe, *tmp_fe;
spin_lock_irqsave(&lc->flush_lock, flags);
- flush_list = lc->flush_list;
+ list_splice_init(&lc->flush_list, &flush_list);
spin_unlock_irqrestore(&lc->flush_lock, flags);
+ if (list_empty(&flush_list))
+ return 0;
+
/*
* FIXME: Count up requests, group request types,
* allocate memory to stick all requests in and
@@ -345,25 +347,32 @@
(char *)&fe->region,
sizeof(fe->region),
NULL, NULL);
- if (r) {
- r = (r > 0) ? -r : r;
+ if (r)
goto fail;
- }
}
- r = dm_clog_consult_server(lc->uuid, DM_CLOG_FLUSH,
- NULL, 0, NULL, NULL);
- if (r)
- r = (r > 0) ? -r : r;
+ do {
+ r = dm_clog_consult_server(lc->uuid, DM_CLOG_FLUSH,
+ NULL, 0, NULL, NULL);
+ if (r != -EAGAIN)
+ break;
+
+ DMINFO("Flush conflicts with recovery [delaying]");
+ set_current_state(TASK_INTERRUPTIBLE);
+ schedule_timeout(HZ/4);
+ } while (1);
fail:
+ /*
+ * We can safely remove these entries, even if failure.
+ * Calling code will recieve an error and will know that
+ * the log facility has failed.
+ */
list_for_each_entry_safe(fe, tmp_fe, &flush_list, list) {
list_del(&fe->list);
mempool_free(fe, flush_entry_pool);
}
- r = -EIO;
-
return r;
}
@@ -382,7 +391,7 @@
struct flush_entry *fe;
/* Wait for an allocation, but _never_ fail */
- fe = mempool_alloc(flush_enrty_pool, GFP_KERNEL);
+ fe = mempool_alloc(flush_entry_pool, GFP_NOIO);
BUG_ON(!fe);
spin_lock_irqsave(&lc->flush_lock, flags);
@@ -390,7 +399,7 @@
fe->region = region;
list_add(&fe->list, &lc->flush_list);
spin_unlock_irqrestore(&lc->flush_lock, flags);
-
+
return;
}
@@ -412,17 +421,24 @@
struct log_c *lc = (struct log_c *)log->context;
struct flush_entry *fe;
- fe = mempool_alloc(flush_enrty_pool, GFP_ATOMIC);
+ /*
+ * If we fail to allocate, we skip the clearing of
+ * the region. This doesn't hurt us in any way, except
+ * to cause the region to be resync'ed when the
+ * device is activated next time.
+ */
+ fe = mempool_alloc(flush_entry_pool, GFP_ATOMIC);
if (!fe) {
DMERR("Failed to allocate memory to clear region.");
return;
}
+
spin_lock_irqsave(&lc->flush_lock, flags);
fe->type = DM_CLOG_CLEAR_REGION;
fe->region = region;
list_add(&fe->list, &lc->flush_list);
spin_unlock_irqrestore(&lc->flush_lock, flags);
-
+
return;
}
@@ -446,12 +462,9 @@
rdata_size = sizeof(pkg);
r = dm_clog_consult_server(lc->uuid, DM_CLOG_GET_RESYNC_WORK,
NULL, 0,
- &pkg, &rdata_size);
-
- r = (r > 0) ? -r : r;
+ (char *)&pkg, &rdata_size);
*region = pkg.r;
-
return (r) ? r : pkg.i;
}
@@ -475,7 +488,7 @@
pkg.i = in_sync;
r = dm_clog_consult_server(lc->uuid, DM_CLOG_SET_REGION_SYNC,
- &pkg, sizeof(pkg),
+ (char *)&pkg, sizeof(pkg),
NULL, NULL);
/* FIXME: It would be nice to be able to report failures */
@@ -518,18 +531,22 @@
static int cluster_status(struct dirty_log *log, status_type_t status_type,
char *result, unsigned int maxlen)
{
- int r;
+ int r = 0;
unsigned int sz = maxlen;
struct log_c *lc = (struct log_c *)log->context;
- switch(status) {
+ switch(status_type) {
case STATUSTYPE_INFO:
r = dm_clog_consult_server(lc->uuid, DM_CLOG_STATUS_INFO,
NULL, 0,
result, &sz);
+ /*
+ * FIXME: If we fail to contact server, we should still
+ * populate this with parsible results
+ */
break;
case STATUSTYPE_TABLE:
- r = dm_clog_consult_server(lc->uuid, DM_CLOG_STATUS_INFO,
+ r = dm_clog_consult_server(lc->uuid, DM_CLOG_STATUS_TABLE,
NULL, 0,
result, &sz);
break;
@@ -537,11 +554,10 @@
return (r) ? 0: sz;
}
-status int cluster_get_failure_response(struct dirty_log *log)
+static int cluster_get_failure_response(struct dirty_log *log)
{
- struct log_c *lc = (struct log_c *)log->context;
-
- return lc->failure_response;
+ /* This type of return is not an option */
+ return DMLOG_IOERR_BLOCK;
}
static struct dirty_log_type _clustered_core_type = {
@@ -554,7 +570,6 @@
.resume = cluster_resume,
.get_region_size = cluster_get_region_size,
.is_clean = cluster_is_clean,
- .is_remote_recovering = cluster_is_remote_recovering,
.in_sync = cluster_in_sync,
.flush = cluster_flush,
.mark_region = cluster_mark_region,
@@ -576,7 +591,6 @@
.resume = cluster_resume,
.get_region_size = cluster_get_region_size,
.is_clean = cluster_is_clean,
- .is_remote_recovering = cluster_is_remote_recovering,
.in_sync = cluster_in_sync,
.flush = cluster_flush,
.mark_region = cluster_mark_region,
@@ -596,13 +610,22 @@
flush_entry_free, NULL);
if (!flush_entry_pool) {
- DMERR("Unable to create flush_entry_pool: No memory.");
+ DMWARN("Unable to create flush_entry_pool: No memory.");
return -ENOMEM;
}
+ r = dm_clog_tfr_init();
+ if (r) {
+ DMWARN("Unable to initialize cluster log communications");
+ mempool_destroy(flush_entry_pool);
+ return r;
+ }
+
r = dm_register_dirty_log_type(&_clustered_core_type);
if (r) {
DMWARN("Couldn't register clustered_core dirty log type");
+ dm_clog_tfr_exit();
+ mempool_destroy(flush_entry_pool);
return r;
}
@@ -610,15 +633,28 @@
if (r) {
DMWARN("Couldn't register clustered_disk dirty log type");
dm_unregister_dirty_log_type(&_clustered_core_type);
+ dm_clog_tfr_exit();
+ mempool_destroy(flush_entry_pool);
return r;
}
- return r;
+ DMINFO("dm-clulog (built %s %s) installed", __DATE__, __TIME__);
+ return 0;
}
static void __exit cluster_dirty_log_exit(void)
{
dm_unregister_dirty_log_type(&_clustered_disk_type);
dm_unregister_dirty_log_type(&_clustered_core_type);
+ dm_clog_tfr_exit();
+ mempool_destroy(flush_entry_pool);
+ DMINFO("dm-clulog (built %s %s) removed", __DATE__, __TIME__);
return;
}
+
+module_init(cluster_dirty_log_init);
+module_exit(cluster_dirty_log_exit);
+
+MODULE_DESCRIPTION(DM_NAME " mirror cluster-aware log");
+MODULE_AUTHOR("Jonathan Brassow");
+MODULE_LICENSE("GPL");
reply other threads:[~2007-08-23 19:54 UTC|newest]
Thread overview: [no followups] expand[flat|nested] mbox.gz Atom feed
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20070823195459.29033.qmail@sourceware.org \
--to=jbrassow@sourceware.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).