From: pcaulfield@sourceware.org <pcaulfield@sourceware.org>
To: cluster-devel.redhat.com
Subject: [Cluster-devel] cluster/dlm/man dlm_create_lockspace.3 dlm_loc ...
Date: 9 Jul 2007 09:00:32 -0000 [thread overview]
Message-ID: <20070709090032.2930.qmail@sourceware.org> (raw)
CVSROOT: /cvs/cluster
Module name: cluster
Branch: RHEL4
Changes by: pcaulfield at sourceware.org 2007-07-09 09:00:29
Modified files:
dlm/man : dlm_create_lockspace.3 dlm_lock.3 dlm_unlock.3
libdlm.3
Log message:
Fix some typos and cut-n-pasteos
Patches:
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/dlm/man/dlm_create_lockspace.3.diff?cvsroot=cluster&only_with_tag=RHEL4&r1=1.1.2.1&r2=1.1.2.2
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/dlm/man/dlm_lock.3.diff?cvsroot=cluster&only_with_tag=RHEL4&r1=1.1.2.1&r2=1.1.2.2
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/dlm/man/dlm_unlock.3.diff?cvsroot=cluster&only_with_tag=RHEL4&r1=1.1.2.1&r2=1.1.2.2
http://sourceware.org/cgi-bin/cvsweb.cgi/cluster/dlm/man/libdlm.3.diff?cvsroot=cluster&only_with_tag=RHEL4&r1=1.1.2.1&r2=1.1.2.2
--- cluster/dlm/man/Attic/dlm_create_lockspace.3 2007/07/06 12:32:30 1.1.2.1
+++ cluster/dlm/man/Attic/dlm_create_lockspace.3 2007/07/09 09:00:29 1.1.2.2
@@ -14,7 +14,7 @@
.SH DESCRIPTION
The DLM allows locks to be partitioned into "lockspaces", and these can be manipulated by userspace calls. It is possible (though not recommended) for an application to have multiple lockspaces open at one time.
-Many of the DLM calls work on the "default" lockspace, which should be fine for most users. The calls with _ls_ in them allow you to isolate your application from all others running in the cluster. Remember, lockspaces are a cluster-wide resource, so if you create a lockspace called "myls" it will share locks with a lockspace called "myls" on all nodes. These calls allow users to create & remove lockspaces, and users to connecto to existing lockspace to store their locks there.
+Many of the DLM calls work on the "default" lockspace, which should be fine for most users. The calls with _ls_ in them allow you to isolate your application from all others running in the cluster. Remember, lockspaces are a cluster-wide resource, so if you create a lockspace called "myls" it will share locks with a lockspace called "myls" on all nodes. These calls allow users to create & remove lockspaces, and users to connect to existing lockspace to store their locks there.
.PP
.SS
dlm_lshandle_t dlm_create_lockspace(const char *name, mode_t mode);
--- cluster/dlm/man/Attic/dlm_lock.3 2007/07/06 12:32:30 1.1.2.1
+++ cluster/dlm/man/Attic/dlm_lock.3 2007/07/09 09:00:29 1.1.2.2
@@ -139,10 +139,10 @@
.B range
.br
an optional structure of two uint64_t that indicate the range
-of the lock. Locks with overlaPPing ranges will be granted only
-if the lock modes are compatible. locks with non-overlaPPing
+of the lock. Locks with overlapping ranges will be granted only
+if the lock modes are compatible. locks with non-overlapping
ranges (on the same resource) do not conflict. A lock with no
-range is assumed to have a range emcompassing the largest
+range is assumed to have a range encompassing the largest
possible range. ie. 0-0xFFFFFFFFFFFFFFFF. Note that is is more
efficient to specify no range than to specify the full range
above.
--- cluster/dlm/man/Attic/dlm_unlock.3 2007/07/06 12:32:30 1.1.2.1
+++ cluster/dlm/man/Attic/dlm_unlock.3 2007/07/09 09:00:29 1.1.2.2
@@ -3,13 +3,13 @@
dlm_unlock \- unlock a DLM lock
.SH SYNOPSIS
.nf
- #include <libdlm.h>
+#include <libdlm.h>
-extern int dlm_unlock(uint32_t lkid,
- uint32_t flags, struct dlm_lksb *lksb, void *astarg);
+int dlm_unlock(uint32_t lkid,
+ uint32_t flags, struct dlm_lksb *lksb, void *astarg);
-extern int dlm_unlock_wait(uint32_t lkid,
- uint32_t flags, struct dlm_lksb *lksb);
+int dlm_unlock_wait(uint32_t lkid,
+ uint32_t flags, struct dlm_lksb *lksb);
.fi
.SH DESCRIPTION
--- cluster/dlm/man/Attic/libdlm.3 2007/07/06 12:32:30 1.1.2.1
+++ cluster/dlm/man/Attic/libdlm.3 2007/07/09 09:00:29 1.1.2.2
@@ -3,13 +3,13 @@
libdlm, dlm_get_fd, dlm_dispatch, dlm_pthread_init, dlm_ls_pthread_init, dlm_cleanup
.SH SYNOPSIS
.nf
- #include <libdlm.h>
+#include <libdlm.h>
.nf
-extern int dlm_pthread_init();
-extern int dlm_ls_pthread_init(dlm_lshandle_t lockspace);
-extern int dlm_pthread_cleanup();
-extern int dlm_get_fd(void);
-extern int dlm_dispatch(int fd);
+int dlm_pthread_init();
+int dlm_ls_pthread_init(dlm_lshandle_t lockspace);
+int dlm_pthread_cleanup();
+int dlm_get_fd(void);
+int dlm_dispatch(int fd);
link with -ldlm
.fi
next reply other threads:[~2007-07-09 9:00 UTC|newest]
Thread overview: 2+ messages / expand[flat|nested] mbox.gz Atom feed top
2007-07-09 9:00 pcaulfield [this message]
-- strict thread matches above, loose matches on Subject: below --
2007-07-09 16:53 [Cluster-devel] cluster/dlm/man dlm_create_lockspace.3 dlm_loc teigland
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20070709090032.2930.qmail@sourceware.org \
--to=pcaulfield@sourceware.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).