Linux NFS development
 help / color / mirror / Atom feed
* [PATCH] SUNRPC: Fix backchannel reply, again
@ 2024-06-14 14:18 cel
  2024-06-18 20:31 ` Chuck Lever
  2024-06-18 21:17 ` Jeff Layton
  0 siblings, 2 replies; 4+ messages in thread
From: cel @ 2024-06-14 14:18 UTC (permalink / raw)
  To: linux-nfs; +Cc: Ben Coddington, Chuck Lever

From: Chuck Lever <chuck.lever@oracle.com>

I still see "RPC: Could not send backchannel reply error: -110"
quite often, along with slow-running tests. Debugging shows that the
backchannel is still stumbling when it has to queue a callback reply
on a busy transport.

Note that every one of these timeouts causes a connection loss by
virtue of the xprt_conditional_disconnect() call in that arm of
call_cb_transmit_status().

I found that setting to_maxval is necessary to get the RPC timeout
logic to behave whenever to_exponential is not set.

Fixes: 57331a59ac0d ("NFSv4.1: Use the nfs_client's rpc timeouts for backchannel")
Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
---
 net/sunrpc/svc.c | 1 +
 1 file changed, 1 insertion(+)

diff --git a/net/sunrpc/svc.c b/net/sunrpc/svc.c
index 965a27806bfd..f4ddb2961042 100644
--- a/net/sunrpc/svc.c
+++ b/net/sunrpc/svc.c
@@ -1643,6 +1643,7 @@ void svc_process_bc(struct rpc_rqst *req, struct svc_rqst *rqstp)
 		timeout.to_initval = req->rq_xprt->timeout->to_initval;
 		timeout.to_retries = req->rq_xprt->timeout->to_retries;
 	}
+	timeout.to_maxval = timeout.to_initval;
 	memcpy(&req->rq_snd_buf, &rqstp->rq_res, sizeof(req->rq_snd_buf));
 	task = rpc_run_bc_task(req, &timeout);
 
-- 
2.45.1


^ permalink raw reply related	[flat|nested] 4+ messages in thread

* Re: [PATCH] SUNRPC: Fix backchannel reply, again
  2024-06-14 14:18 [PATCH] SUNRPC: Fix backchannel reply, again cel
@ 2024-06-18 20:31 ` Chuck Lever
  2024-06-18 22:45   ` Trond Myklebust
  2024-06-18 21:17 ` Jeff Layton
  1 sibling, 1 reply; 4+ messages in thread
From: Chuck Lever @ 2024-06-18 20:31 UTC (permalink / raw)
  To: Trond Myklebust, Anna Schumaker; +Cc: linux-nfs, Ben Coddington

On Fri, Jun 14, 2024 at 10:18:52AM -0400, cel@kernel.org wrote:
> From: Chuck Lever <chuck.lever@oracle.com>
> 
> I still see "RPC: Could not send backchannel reply error: -110"
> quite often, along with slow-running tests. Debugging shows that the
> backchannel is still stumbling when it has to queue a callback reply
> on a busy transport.
> 
> Note that every one of these timeouts causes a connection loss by
> virtue of the xprt_conditional_disconnect() call in that arm of
> call_cb_transmit_status().
> 
> I found that setting to_maxval is necessary to get the RPC timeout
> logic to behave whenever to_exponential is not set.
> 
> Fixes: 57331a59ac0d ("NFSv4.1: Use the nfs_client's rpc timeouts for backchannel")
> Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
> ---
>  net/sunrpc/svc.c | 1 +
>  1 file changed, 1 insertion(+)
> 
> diff --git a/net/sunrpc/svc.c b/net/sunrpc/svc.c
> index 965a27806bfd..f4ddb2961042 100644
> --- a/net/sunrpc/svc.c
> +++ b/net/sunrpc/svc.c
> @@ -1643,6 +1643,7 @@ void svc_process_bc(struct rpc_rqst *req, struct svc_rqst *rqstp)
>  		timeout.to_initval = req->rq_xprt->timeout->to_initval;
>  		timeout.to_retries = req->rq_xprt->timeout->to_retries;
>  	}
> +	timeout.to_maxval = timeout.to_initval;
>  	memcpy(&req->rq_snd_buf, &rqstp->rq_res, sizeof(req->rq_snd_buf));
>  	task = rpc_run_bc_task(req, &timeout);
>  
> -- 
> 2.45.1
> 

Hi - would love to see this in 6.10-rc. Is there a chance that
could happen?

-- 
Chuck Lever

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH] SUNRPC: Fix backchannel reply, again
  2024-06-14 14:18 [PATCH] SUNRPC: Fix backchannel reply, again cel
  2024-06-18 20:31 ` Chuck Lever
@ 2024-06-18 21:17 ` Jeff Layton
  1 sibling, 0 replies; 4+ messages in thread
From: Jeff Layton @ 2024-06-18 21:17 UTC (permalink / raw)
  To: cel, linux-nfs; +Cc: Ben Coddington, Chuck Lever

On Fri, 2024-06-14 at 10:18 -0400, cel@kernel.org wrote:
> From: Chuck Lever <chuck.lever@oracle.com>
> 
> I still see "RPC: Could not send backchannel reply error: -110"
> quite often, along with slow-running tests. Debugging shows that the
> backchannel is still stumbling when it has to queue a callback reply
> on a busy transport.
> 
> Note that every one of these timeouts causes a connection loss by
> virtue of the xprt_conditional_disconnect() call in that arm of
> call_cb_transmit_status().
> 
> I found that setting to_maxval is necessary to get the RPC timeout
> logic to behave whenever to_exponential is not set.
> 
> Fixes: 57331a59ac0d ("NFSv4.1: Use the nfs_client's rpc timeouts for
> backchannel")
> Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
> ---
>  net/sunrpc/svc.c | 1 +
>  1 file changed, 1 insertion(+)
> 
> diff --git a/net/sunrpc/svc.c b/net/sunrpc/svc.c
> index 965a27806bfd..f4ddb2961042 100644
> --- a/net/sunrpc/svc.c
> +++ b/net/sunrpc/svc.c
> @@ -1643,6 +1643,7 @@ void svc_process_bc(struct rpc_rqst *req,
> struct svc_rqst *rqstp)
>  		timeout.to_initval = req->rq_xprt->timeout-
> >to_initval;
>  		timeout.to_retries = req->rq_xprt->timeout-
> >to_retries;
>  	}
> +	timeout.to_maxval = timeout.to_initval;
>  	memcpy(&req->rq_snd_buf, &rqstp->rq_res, sizeof(req-
> >rq_snd_buf));
>  	task = rpc_run_bc_task(req, &timeout);
>  

Reviewed-by: Jeff Layton <jlayton@kernel.org>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH] SUNRPC: Fix backchannel reply, again
  2024-06-18 20:31 ` Chuck Lever
@ 2024-06-18 22:45   ` Trond Myklebust
  0 siblings, 0 replies; 4+ messages in thread
From: Trond Myklebust @ 2024-06-18 22:45 UTC (permalink / raw)
  To: anna@kernel.org, chuck.lever@oracle.com
  Cc: linux-nfs@vger.kernel.org, bcodding@redhat.com

On Tue, 2024-06-18 at 16:31 -0400, Chuck Lever wrote:
> On Fri, Jun 14, 2024 at 10:18:52AM -0400, cel@kernel.org wrote:
> > From: Chuck Lever <chuck.lever@oracle.com>
> > 
> > I still see "RPC: Could not send backchannel reply error: -110"
> > quite often, along with slow-running tests. Debugging shows that
> > the
> > backchannel is still stumbling when it has to queue a callback
> > reply
> > on a busy transport.
> > 
> > Note that every one of these timeouts causes a connection loss by
> > virtue of the xprt_conditional_disconnect() call in that arm of
> > call_cb_transmit_status().
> > 
> > I found that setting to_maxval is necessary to get the RPC timeout
> > logic to behave whenever to_exponential is not set.
> > 
> > Fixes: 57331a59ac0d ("NFSv4.1: Use the nfs_client's rpc timeouts
> > for backchannel")
> > Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
> > ---
> >  net/sunrpc/svc.c | 1 +
> >  1 file changed, 1 insertion(+)
> > 
> > diff --git a/net/sunrpc/svc.c b/net/sunrpc/svc.c
> > index 965a27806bfd..f4ddb2961042 100644
> > --- a/net/sunrpc/svc.c
> > +++ b/net/sunrpc/svc.c
> > @@ -1643,6 +1643,7 @@ void svc_process_bc(struct rpc_rqst *req,
> > struct svc_rqst *rqstp)
> >  		timeout.to_initval = req->rq_xprt->timeout-
> > >to_initval;
> >  		timeout.to_retries = req->rq_xprt->timeout-
> > >to_retries;
> >  	}
> > +	timeout.to_maxval = timeout.to_initval;
> >  	memcpy(&req->rq_snd_buf, &rqstp->rq_res, sizeof(req-
> > >rq_snd_buf));
> >  	task = rpc_run_bc_task(req, &timeout);
> >  
> > -- 
> > 2.45.1
> > 
> 
> Hi - would love to see this in 6.10-rc. Is there a chance that
> could happen?

Hmm... Can we please also set the remaining fields in timeout to 0?
Otherwise, we're still playing roulette with what actually ends up
happening in xprt_calc_majortimeo(). If to_increment happens to be
large enough, we could overflow and end up with a silly small timeout
value on a retry.


-- 
Trond Myklebust
Linux NFS client maintainer, Hammerspace
trond.myklebust@hammerspace.com



^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2024-06-18 22:45 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2024-06-14 14:18 [PATCH] SUNRPC: Fix backchannel reply, again cel
2024-06-18 20:31 ` Chuck Lever
2024-06-18 22:45   ` Trond Myklebust
2024-06-18 21:17 ` Jeff Layton

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox