linux-nfs.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Trond Myklebust <Trond.Myklebust@netapp.com>
To: Chuck Lever <chuck.lever@oracle.com>
Cc: linux-nfs@vger.kernel.org
Subject: Re: [PATCH 2/4] NFS: Move cl_state_owners and related fields to the nfs_server struct
Date: Thu, 23 Dec 2010 17:04:34 -0500	[thread overview]
Message-ID: <1293141874.9718.1.camel@heimdal.trondhjem.org> (raw)
In-Reply-To: <20101223215425.2344.36322.stgit@matisse.1015granger.net>

On Thu, 2010-12-23 at 16:54 -0500, Chuck Lever wrote:
> NFSv4 migration needs to reassociate state owners from the source to
> the destination nfs_server data structures.  To make that easier, move
> the cl_state_owners field to the nfs_server struct.  cl_openowner_id
> and cl_lockowner_id accompany this move, as they are used in
> conjunction with cl_state_owners.
> 
> The cl_lock field in the parent nfs_client continues to protect all
> three of these fields.
> 
> Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
> ---
> 
>  fs/nfs/nfs4_fs.h          |    2 
>  fs/nfs/nfs4state.c        |  239 ++++++++++++++++++++++++++++++++-------------
>  include/linux/nfs_fs_sb.h |    9 +-
>  3 files changed, 177 insertions(+), 73 deletions(-)
> 
> diff --git a/fs/nfs/nfs4_fs.h b/fs/nfs/nfs4_fs.h
> index 7a6eecf..fa5c5d3 100644
> --- a/fs/nfs/nfs4_fs.h
> +++ b/fs/nfs/nfs4_fs.h
> @@ -109,7 +109,7 @@ struct nfs_unique_id {
>  struct nfs4_state_owner {
>  	struct nfs_unique_id so_owner_id;
>  	struct nfs_server    *so_server;
> -	struct rb_node	     so_client_node;
> +	struct rb_node	     so_server_node;
>  
>  	struct rpc_cred	     *so_cred;	 /* Associated cred */
>  
> diff --git a/fs/nfs/nfs4state.c b/fs/nfs/nfs4state.c
> index f575a31..fbcfe72 100644
> --- a/fs/nfs/nfs4state.c
> +++ b/fs/nfs/nfs4state.c
> @@ -105,14 +105,17 @@ static void nfs4_clear_machine_cred(struct nfs_client *clp)
>  		put_rpccred(cred);
>  }
>  
> -struct rpc_cred *nfs4_get_renew_cred_locked(struct nfs_client *clp)
> +static struct rpc_cred *
> +nfs4_get_renew_cred_server_locked(struct nfs_server *server)
>  {
> +	struct rpc_cred *cred = NULL;
>  	struct nfs4_state_owner *sp;
>  	struct rb_node *pos;
> -	struct rpc_cred *cred = NULL;
>  
> -	for (pos = rb_first(&clp->cl_state_owners); pos != NULL; pos = rb_next(pos)) {
> -		sp = rb_entry(pos, struct nfs4_state_owner, so_client_node);
> +	for (pos = rb_first(&server->state_owners);
> +	     pos != NULL;
> +	     pos = rb_next(pos)) {
> +		sp = rb_entry(pos, struct nfs4_state_owner, so_server_node);
>  		if (list_empty(&sp->so_states))
>  			continue;
>  		cred = get_rpccred(sp->so_cred);
> @@ -121,6 +124,28 @@ struct rpc_cred *nfs4_get_renew_cred_locked(struct nfs_client *clp)
>  	return cred;
>  }
>  
> +/**
> + * nfs4_get_renew_cred_locked - Acquire credential for a renew operation
> + * @clp: client state handle
> + *
> + * Returns an rpc_cred with reference count bumped, or NULL.
> + * Caller must hold clp->cl_lock.
> + */
> +struct rpc_cred *nfs4_get_renew_cred_locked(struct nfs_client *clp)
> +{
> +	struct rpc_cred *cred = NULL;
> +	struct nfs_server *server;
> +
> +	rcu_read_lock();
> +	list_for_each_entry_rcu(server, &clp->cl_superblocks, client_link) {
> +		cred = nfs4_get_renew_cred_server_locked(server);
> +		if (cred != NULL)
> +			break;
> +	}
> +	rcu_read_unlock();
> +	return cred;
> +}
> +
>  #if defined(CONFIG_NFS_V4_1)
>  
>  static int nfs41_setup_state_renewal(struct nfs_client *clp)
> @@ -210,21 +235,45 @@ struct rpc_cred *nfs4_get_exchange_id_cred(struct nfs_client *clp)
>  
>  #endif /* CONFIG_NFS_V4_1 */
>  
> -struct rpc_cred *nfs4_get_setclientid_cred(struct nfs_client *clp)
> +static struct rpc_cred *
> +nfs4_get_setclientid_cred_server_locked(struct nfs_server *server)
>  {
> +	struct rpc_cred *cred = NULL;
>  	struct nfs4_state_owner *sp;
>  	struct rb_node *pos;
> +
> +	pos = rb_first(&server->state_owners);
> +	if (pos != NULL) {
> +		sp = rb_entry(pos, struct nfs4_state_owner, so_server_node);
> +		cred = get_rpccred(sp->so_cred);
> +	}
> +	return cred;
> +}
> +
> +/**
> + * nfs4_get_setclientid_cred - Acquire credential for a setclientid operation
> + * @clp: client state handle
> + *
> + * Returns an rpc_cred with reference count bumped, or NULL.
> + */
> +struct rpc_cred *nfs4_get_setclientid_cred(struct nfs_client *clp)
> +{
> +	struct nfs_server *server;
>  	struct rpc_cred *cred;
>  
>  	spin_lock(&clp->cl_lock);
>  	cred = nfs4_get_machine_cred_locked(clp);
>  	if (cred != NULL)
>  		goto out;
> -	pos = rb_first(&clp->cl_state_owners);
> -	if (pos != NULL) {
> -		sp = rb_entry(pos, struct nfs4_state_owner, so_client_node);
> -		cred = get_rpccred(sp->so_cred);
> +
> +	rcu_read_lock();

Here's another rcu_read_lock() nested inside a spin locked section.

> +	list_for_each_entry_rcu(server, &clp->cl_superblocks, client_link) {
> +		cred = nfs4_get_setclientid_cred_server_locked(server);
> +		if (cred != NULL)
> +			break;
>  	}
> +	rcu_read_unlock();
> +
>  out:
>  	spin_unlock(&clp->cl_lock);
>  	return cred;
> @@ -286,16 +335,15 @@ static void nfs_free_unique_id(struct rb_root *root, struct nfs_unique_id *id)
>  }
>  
>  static struct nfs4_state_owner *
> -nfs4_find_state_owner(struct nfs_server *server, struct rpc_cred *cred)
> +nfs4_find_state_owner_locked(struct nfs_server *server, struct rpc_cred *cred)
>  {
> -	struct nfs_client *clp = server->nfs_client;
> -	struct rb_node **p = &clp->cl_state_owners.rb_node,
> +	struct rb_node **p = &server->state_owners.rb_node,
>  		       *parent = NULL;
>  	struct nfs4_state_owner *sp, *res = NULL;
>  
>  	while (*p != NULL) {
>  		parent = *p;
> -		sp = rb_entry(parent, struct nfs4_state_owner, so_client_node);
> +		sp = rb_entry(parent, struct nfs4_state_owner, so_server_node);
>  
>  		if (server < sp->so_server) {
>  			p = &parent->rb_left;
> @@ -319,24 +367,17 @@ nfs4_find_state_owner(struct nfs_server *server, struct rpc_cred *cred)
>  }
>  
>  static struct nfs4_state_owner *
> -nfs4_insert_state_owner(struct nfs_client *clp, struct nfs4_state_owner *new)
> +nfs4_insert_state_owner_locked(struct nfs4_state_owner *new)
>  {
> -	struct rb_node **p = &clp->cl_state_owners.rb_node,
> +	struct nfs_server *server = new->so_server;
> +	struct rb_node **p = &server->state_owners.rb_node,
>  		       *parent = NULL;
>  	struct nfs4_state_owner *sp;
>  
>  	while (*p != NULL) {
>  		parent = *p;
> -		sp = rb_entry(parent, struct nfs4_state_owner, so_client_node);
> +		sp = rb_entry(parent, struct nfs4_state_owner, so_server_node);
>  
> -		if (new->so_server < sp->so_server) {
> -			p = &parent->rb_left;
> -			continue;
> -		}
> -		if (new->so_server > sp->so_server) {
> -			p = &parent->rb_right;
> -			continue;
> -		}
>  		if (new->so_cred < sp->so_cred)
>  			p = &parent->rb_left;
>  		else if (new->so_cred > sp->so_cred)
> @@ -346,18 +387,20 @@ nfs4_insert_state_owner(struct nfs_client *clp, struct nfs4_state_owner *new)
>  			return sp;
>  		}
>  	}
> -	nfs_alloc_unique_id(&clp->cl_openowner_id, &new->so_owner_id, 1, 64);
> -	rb_link_node(&new->so_client_node, parent, p);
> -	rb_insert_color(&new->so_client_node, &clp->cl_state_owners);
> +	nfs_alloc_unique_id(&server->openowner_id, &new->so_owner_id, 1, 64);
> +	rb_link_node(&new->so_server_node, parent, p);
> +	rb_insert_color(&new->so_server_node, &server->state_owners);
>  	return new;
>  }
>  
>  static void
> -nfs4_remove_state_owner(struct nfs_client *clp, struct nfs4_state_owner *sp)
> +nfs4_remove_state_owner_locked(struct nfs4_state_owner *sp)
>  {
> -	if (!RB_EMPTY_NODE(&sp->so_client_node))
> -		rb_erase(&sp->so_client_node, &clp->cl_state_owners);
> -	nfs_free_unique_id(&clp->cl_openowner_id, &sp->so_owner_id);
> +	struct nfs_server *server = sp->so_server;
> +
> +	if (!RB_EMPTY_NODE(&sp->so_server_node))
> +		rb_erase(&sp->so_server_node, &server->state_owners);
> +	nfs_free_unique_id(&server->openowner_id, &sp->so_owner_id);
>  }
>  
>  /*
> @@ -386,23 +429,32 @@ nfs4_alloc_state_owner(void)
>  static void
>  nfs4_drop_state_owner(struct nfs4_state_owner *sp)
>  {
> -	if (!RB_EMPTY_NODE(&sp->so_client_node)) {
> -		struct nfs_client *clp = sp->so_server->nfs_client;
> +	if (!RB_EMPTY_NODE(&sp->so_server_node)) {
> +		struct nfs_server *server = sp->so_server;
> +		struct nfs_client *clp = server->nfs_client;
>  
>  		spin_lock(&clp->cl_lock);
> -		rb_erase(&sp->so_client_node, &clp->cl_state_owners);
> -		RB_CLEAR_NODE(&sp->so_client_node);
> +		rb_erase(&sp->so_server_node, &server->state_owners);
> +		RB_CLEAR_NODE(&sp->so_server_node);
>  		spin_unlock(&clp->cl_lock);
>  	}
>  }
>  
> -struct nfs4_state_owner *nfs4_get_state_owner(struct nfs_server *server, struct rpc_cred *cred)
> +/**
> + * nfs4_get_state_owner - Look up a state owner given a credential
> + * @server: nfs_server to search
> + * @cred: RPC credential to match
> + *
> + * Returns a pointer to an instantiated nfs4_state_owner struct, or NULL.
> + */
> +struct nfs4_state_owner *nfs4_get_state_owner(struct nfs_server *server,
> +					      struct rpc_cred *cred)
>  {
>  	struct nfs_client *clp = server->nfs_client;
>  	struct nfs4_state_owner *sp, *new;
>  
>  	spin_lock(&clp->cl_lock);
> -	sp = nfs4_find_state_owner(server, cred);
> +	sp = nfs4_find_state_owner_locked(server, cred);
>  	spin_unlock(&clp->cl_lock);
>  	if (sp != NULL)
>  		return sp;
> @@ -412,7 +464,7 @@ struct nfs4_state_owner *nfs4_get_state_owner(struct nfs_server *server, struct
>  	new->so_server = server;
>  	new->so_cred = cred;
>  	spin_lock(&clp->cl_lock);
> -	sp = nfs4_insert_state_owner(clp, new);
> +	sp = nfs4_insert_state_owner_locked(new);
>  	spin_unlock(&clp->cl_lock);
>  	if (sp == new)
>  		get_rpccred(cred);
> @@ -423,6 +475,11 @@ struct nfs4_state_owner *nfs4_get_state_owner(struct nfs_server *server, struct
>  	return sp;
>  }
>  
> +/**
> + * nfs4_put_state_owner - Release a nfs4_state_owner
> + * @sp: state owner data to release
> + *
> + */
>  void nfs4_put_state_owner(struct nfs4_state_owner *sp)
>  {
>  	struct nfs_client *clp = sp->so_server->nfs_client;
> @@ -430,7 +487,7 @@ void nfs4_put_state_owner(struct nfs4_state_owner *sp)
>  
>  	if (!atomic_dec_and_lock(&sp->so_count, &clp->cl_lock))
>  		return;
> -	nfs4_remove_state_owner(clp, sp);
> +	nfs4_remove_state_owner_locked(sp);
>  	spin_unlock(&clp->cl_lock);
>  	rpc_destroy_wait_queue(&sp->so_sequence.wait);
>  	put_rpccred(cred);
> @@ -633,7 +690,8 @@ __nfs4_find_lock_state(struct nfs4_state *state, fl_owner_t fl_owner, pid_t fl_p
>  static struct nfs4_lock_state *nfs4_alloc_lock_state(struct nfs4_state *state, fl_owner_t fl_owner, pid_t fl_pid, unsigned int type)
>  {
>  	struct nfs4_lock_state *lsp;
> -	struct nfs_client *clp = state->owner->so_server->nfs_client;
> +	struct nfs_server *server = state->owner->so_server;
> +	struct nfs_client *clp = server->nfs_client;
>  
>  	lsp = kzalloc(sizeof(*lsp), GFP_NOFS);
>  	if (lsp == NULL)
> @@ -657,7 +715,7 @@ static struct nfs4_lock_state *nfs4_alloc_lock_state(struct nfs4_state *state, f
>  		return NULL;
>  	}
>  	spin_lock(&clp->cl_lock);
> -	nfs_alloc_unique_id(&clp->cl_lockowner_id, &lsp->ls_id, 1, 64);
> +	nfs_alloc_unique_id(&server->lockowner_id, &lsp->ls_id, 1, 64);
>  	spin_unlock(&clp->cl_lock);
>  	INIT_LIST_HEAD(&lsp->ls_locks);
>  	return lsp;
> @@ -665,10 +723,11 @@ static struct nfs4_lock_state *nfs4_alloc_lock_state(struct nfs4_state *state, f
>  
>  static void nfs4_free_lock_state(struct nfs4_lock_state *lsp)
>  {
> -	struct nfs_client *clp = lsp->ls_state->owner->so_server->nfs_client;
> +	struct nfs_server *server = lsp->ls_state->owner->so_server;
> +	struct nfs_client *clp = server->nfs_client;
>  
>  	spin_lock(&clp->cl_lock);
> -	nfs_free_unique_id(&clp->cl_lockowner_id, &lsp->ls_id);
> +	nfs_free_unique_id(&server->lockowner_id, &lsp->ls_id);
>  	spin_unlock(&clp->cl_lock);
>  	rpc_destroy_wait_queue(&lsp->ls_sequence.wait);
>  	kfree(lsp);
> @@ -1114,25 +1173,40 @@ static void nfs4_clear_open_state(struct nfs4_state *state)
>  	}
>  }
>  
> -static void nfs4_state_mark_reclaim_helper(struct nfs_client *clp, int (*mark_reclaim)(struct nfs_client *clp, struct nfs4_state *state))
> +static void nfs4_reset_seqids_locked(struct nfs_server *server,
> +	int (*mark_reclaim)(struct nfs_client *clp, struct nfs4_state *state))
>  {
>  	struct nfs4_state_owner *sp;
>  	struct rb_node *pos;
>  	struct nfs4_state *state;
>  
> -	/* Reset all sequence ids to zero */
> -	for (pos = rb_first(&clp->cl_state_owners); pos != NULL; pos = rb_next(pos)) {
> -		sp = rb_entry(pos, struct nfs4_state_owner, so_client_node);
> +	for (pos = rb_first(&server->state_owners);
> +	     pos != NULL;
> +	     pos = rb_next(pos)) {
> +		sp = rb_entry(pos, struct nfs4_state_owner, so_server_node);
>  		sp->so_seqid.flags = 0;
>  		spin_lock(&sp->so_lock);
>  		list_for_each_entry(state, &sp->so_states, open_states) {
> -			if (mark_reclaim(clp, state))
> +			if (mark_reclaim(server->nfs_client, state))
>  				nfs4_clear_open_state(state);
>  		}
>  		spin_unlock(&sp->so_lock);
>  	}
>  }
>  
> +static void nfs4_state_mark_reclaim_helper(struct nfs_client *clp,
> +	int (*mark_reclaim)(struct nfs_client *clp, struct nfs4_state *state))
> +{
> +	struct nfs_server *server;
> +
> +	spin_lock(&clp->cl_lock);
> +	rcu_read_lock();

Ditto.

> +	list_for_each_entry_rcu(server, &clp->cl_superblocks, client_link)
> +		nfs4_reset_seqids_locked(server, mark_reclaim);
> +	rcu_read_unlock();
> +	spin_unlock(&clp->cl_lock);
> +}
> +
>  static void nfs4_state_start_reclaim_reboot(struct nfs_client *clp)
>  {
>  	/* Mark all delegations for reclaim */
> @@ -1148,26 +1222,42 @@ static void nfs4_reclaim_complete(struct nfs_client *clp,
>  		(void)ops->reclaim_complete(clp);
>  }
>  
> -static int nfs4_state_clear_reclaim_reboot(struct nfs_client *clp)
> +static void nfs4_clear_reclaim_server_locked(struct nfs_server *server)
>  {
>  	struct nfs4_state_owner *sp;
>  	struct rb_node *pos;
>  	struct nfs4_state *state;
>  
> -	if (!test_and_clear_bit(NFS4CLNT_RECLAIM_REBOOT, &clp->cl_state))
> -		return 0;
> -
> -	for (pos = rb_first(&clp->cl_state_owners); pos != NULL; pos = rb_next(pos)) {
> -		sp = rb_entry(pos, struct nfs4_state_owner, so_client_node);
> +	for (pos = rb_first(&server->state_owners);
> +	     pos != NULL;
> +	     pos = rb_next(pos)) {
> +		sp = rb_entry(pos, struct nfs4_state_owner, so_server_node);
>  		spin_lock(&sp->so_lock);
>  		list_for_each_entry(state, &sp->so_states, open_states) {
> -			if (!test_and_clear_bit(NFS_STATE_RECLAIM_REBOOT, &state->flags))
> +			if (!test_and_clear_bit(NFS_STATE_RECLAIM_REBOOT,
> +						&state->flags))
>  				continue;
> -			nfs4_state_mark_reclaim_nograce(clp, state);
> +			nfs4_state_mark_reclaim_nograce(server->nfs_client,
> +							state);
>  		}
>  		spin_unlock(&sp->so_lock);
>  	}
>  
> +}
> +
> +static int nfs4_state_clear_reclaim_reboot(struct nfs_client *clp)
> +{
> +	struct nfs_server *server;
> +
> +	if (!test_and_clear_bit(NFS4CLNT_RECLAIM_REBOOT, &clp->cl_state))
> +		return 0;
> +
> +	spin_lock(&clp->cl_lock);
> +	rcu_read_lock();
> +	list_for_each_entry_rcu(server, &clp->cl_superblocks, client_link)
> +		nfs4_clear_reclaim_server_locked(server);
> +	rcu_read_unlock();

Ditto

> +	spin_unlock(&clp->cl_lock);
>  	nfs_delegation_reap_unclaimed(clp);
>  	return 1;
>  }
> @@ -1238,26 +1328,39 @@ static int nfs4_recovery_handle_error(struct nfs_client *clp, int error)
>  
>  static int nfs4_do_reclaim(struct nfs_client *clp, const struct nfs4_state_recovery_ops *ops)
>  {
> +	struct nfs4_state_owner *sp;
> +	struct nfs_server *server;
>  	struct rb_node *pos;
>  	int status = 0;
>  
>  restart:
>  	spin_lock(&clp->cl_lock);
> -	for (pos = rb_first(&clp->cl_state_owners); pos != NULL; pos = rb_next(pos)) {
> -		struct nfs4_state_owner *sp = rb_entry(pos, struct nfs4_state_owner, so_client_node);
> -		if (!test_and_clear_bit(ops->owner_flag_bit, &sp->so_flags))
> -			continue;
> -		atomic_inc(&sp->so_count);
> -		spin_unlock(&clp->cl_lock);
> -		status = nfs4_reclaim_open_state(sp, ops);
> -		if (status < 0) {
> -			set_bit(ops->owner_flag_bit, &sp->so_flags);
> +	rcu_read_lock();

Ditto.
> +	list_for_each_entry_rcu(server, &clp->cl_superblocks, client_link) {
> +		for (pos = rb_first(&server->state_owners);
> +		     pos != NULL;
> +		     pos = rb_next(pos)) {
> +			sp = rb_entry(pos,
> +				struct nfs4_state_owner, so_server_node);
> +			if (!test_and_clear_bit(ops->owner_flag_bit,
> +							&sp->so_flags))
> +				continue;
> +			atomic_inc(&sp->so_count);
> +			rcu_read_unlock();
> +			spin_unlock(&clp->cl_lock);
> +
> +			status = nfs4_reclaim_open_state(sp, ops);
> +			if (status < 0) {
> +				set_bit(ops->owner_flag_bit, &sp->so_flags);
> +				nfs4_put_state_owner(sp);
> +				return nfs4_recovery_handle_error(clp, status);
> +			}
> +
>  			nfs4_put_state_owner(sp);
> -			return nfs4_recovery_handle_error(clp, status);
> +			goto restart;
>  		}
> -		nfs4_put_state_owner(sp);
> -		goto restart;
>  	}
> +	rcu_read_unlock();
>  	spin_unlock(&clp->cl_lock);
>  	return status;
>  }
> diff --git a/include/linux/nfs_fs_sb.h b/include/linux/nfs_fs_sb.h
> index 452d964..e96ec55 100644
> --- a/include/linux/nfs_fs_sb.h
> +++ b/include/linux/nfs_fs_sb.h
> @@ -47,11 +47,7 @@ struct nfs_client {
>  	u64			cl_clientid;	/* constant */
>  	unsigned long		cl_state;
>  
> -	struct rb_root		cl_openowner_id;
> -	struct rb_root		cl_lockowner_id;
> -
>  	struct list_head	cl_delegations;
> -	struct rb_root		cl_state_owners;
>  	spinlock_t		cl_lock;
>  
>  	unsigned long		cl_lease_time;
> @@ -148,6 +144,11 @@ struct nfs_server {
>  						   that are supported on this
>  						   filesystem */
>  	struct pnfs_layoutdriver_type  *pnfs_curr_ld; /* Active layout driver */
> +
> +	/* the following fields are protected by nfs_client->cl_lock */
> +	struct rb_root		state_owners;
> +	struct rb_root		openowner_id;
> +	struct rb_root		lockowner_id;
>  #endif
>  	void (*destroy)(struct nfs_server *);
>  
> 

-- 
Trond Myklebust
Linux NFS client maintainer

NetApp
Trond.Myklebust@netapp.com
www.netapp.com


  reply	other threads:[~2010-12-23 22:04 UTC|newest]

Thread overview: 11+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2010-12-23 21:54 [PATCH 0/4] Refactor data structures to support NFSv4 migration Chuck Lever
2010-12-23 21:54 ` [PATCH 1/4] NFS: Allow walking nfs_client.cl_superblocks list outside client.c Chuck Lever
2010-12-23 21:54 ` [PATCH 2/4] NFS: Move cl_state_owners and related fields to the nfs_server struct Chuck Lever
2010-12-23 22:04   ` Trond Myklebust [this message]
2010-12-23 22:49     ` Chuck Lever
2010-12-23 23:27       ` Trond Myklebust
2010-12-23 21:54 ` [PATCH 3/4] NFS: Introduce nfs_detach_delegations() Chuck Lever
2010-12-23 21:54 ` [PATCH 4/4] NFS: Move cl_delegations to the nfs_server struct Chuck Lever
2010-12-23 22:09   ` Trond Myklebust
     [not found]     ` <1293142182.9718.3.camel-rJ7iovZKK19ZJLDQqaL3InhyD016LWXt@public.gmane.org>
2010-12-23 22:14       ` Chuck Lever
  -- strict thread matches above, loose matches on Subject: below --
2010-12-24  1:32 [PATCH 0/4] Refactor data structures to support NFSv4 migration (2) Chuck Lever
2010-12-24  1:32 ` [PATCH 2/4] NFS: Move cl_state_owners and related fields to the nfs_server struct Chuck Lever

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1293141874.9718.1.camel@heimdal.trondhjem.org \
    --to=trond.myklebust@netapp.com \
    --cc=chuck.lever@oracle.com \
    --cc=linux-nfs@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).