From: Trond Myklebust <trondmy-7I+n7zu2hftEKMMhf/gKZA@public.gmane.org>
To: Lever Chuck <chuck.lever-QHcLZuEGTsvQT0dZR+AlfA@public.gmane.org>
Cc: List Linux RDMA Mailing
<linux-rdma-u79uwXL29TY76Z2rM5mHXA@public.gmane.org>,
List Linux NFS Mailing
<linux-nfs-u79uwXL29TY76Z2rM5mHXA@public.gmane.org>
Subject: Re: [PATCH v1 03/22] SUNRPC: Generalize the RPC buffer allocation API
Date: Mon, 15 Aug 2016 21:59:41 +0000 [thread overview]
Message-ID: <CD81109C-4EB7-4925-9D30-852D3440ACF3@primarydata.com> (raw)
In-Reply-To: <20160815205038.11652.47749.stgit-FYjufvaPoItvLzlybtyyYzGyq/o6K9yX@public.gmane.org>
> On Aug 15, 2016, at 16:50, Chuck Lever <chuck.lever-QHcLZuEGTsvQT0dZR+AlfA@public.gmane.org> wrote:
>
> xprtrdma needs to allocate the Call and Reply buffers separately.
> TBH, the reliance on using a single buffer for the pair of XDR
> buffers is transport implementation-specific.
>
> Transports that want to allocate separate Call and Reply buffers
> will ignore the "size" argument anyway. Don't bother passing it.
>
> The buf_alloc method can't return two pointers. Instead, make the
> method's return value an error code, and set the rq_buffer pointer
> in the method itself.
>
> This gives call_allocate an opportunity to terminate an RPC instead
> of looping forever when a permanent problem occurs. If a request is
> just bogus, or the transport is in a state where it can't allocate
> resources for any request, there needs to be a way to kill the RPC
> right there and not loop.
>
> This immediately fixes a rare problem in the backchannel send path,
> which loops if the server happens to send a CB request whose
> call+reply size is larger than a page (which it shouldn't do yet).
>
> One more issue: looks like xprt_inject_disconnect was incorrectly
> placed in the failure path in call_allocate. It needs to be in the
> success path, as it is for other call-sites.
>
> Signed-off-by: Chuck Lever <chuck.lever-QHcLZuEGTsvQT0dZR+AlfA@public.gmane.org>
> ---
> include/linux/sunrpc/sched.h | 2 +-
> include/linux/sunrpc/xprt.h | 2 +-
> net/sunrpc/clnt.c | 12 ++++++++----
> net/sunrpc/sched.c | 24 +++++++++++++++---------
> net/sunrpc/sunrpc.h | 2 +-
> net/sunrpc/xprtrdma/svc_rdma_backchannel.c | 17 +++++++++--------
> net/sunrpc/xprtrdma/transport.c | 26 ++++++++++++++++++--------
> net/sunrpc/xprtsock.c | 17 +++++++++++------
> 8 files changed, 64 insertions(+), 38 deletions(-)
>
> diff --git a/include/linux/sunrpc/sched.h b/include/linux/sunrpc/sched.h
> index 817af0b..38d4c1b 100644
> --- a/include/linux/sunrpc/sched.h
> +++ b/include/linux/sunrpc/sched.h
> @@ -239,7 +239,7 @@ struct rpc_task *rpc_wake_up_first(struct rpc_wait_queue *,
> void *);
> void rpc_wake_up_status(struct rpc_wait_queue *, int);
> void rpc_delay(struct rpc_task *, unsigned long);
> -void * rpc_malloc(struct rpc_task *, size_t);
> +int rpc_malloc(struct rpc_task *);
> void rpc_free(void *);
> int rpciod_up(void);
> void rpciod_down(void);
> diff --git a/include/linux/sunrpc/xprt.h b/include/linux/sunrpc/xprt.h
> index 559dad3..6f0f796 100644
> --- a/include/linux/sunrpc/xprt.h
> +++ b/include/linux/sunrpc/xprt.h
> @@ -127,7 +127,7 @@ struct rpc_xprt_ops {
> void (*rpcbind)(struct rpc_task *task);
> void (*set_port)(struct rpc_xprt *xprt, unsigned short port);
> void (*connect)(struct rpc_xprt *xprt, struct rpc_task *task);
> - void * (*buf_alloc)(struct rpc_task *task, size_t size);
> + int (*buf_alloc)(struct rpc_task *task);
> void (*buf_free)(void *buffer);
> int (*send_request)(struct rpc_task *task);
> void (*set_retrans_timeout)(struct rpc_task *task);
> diff --git a/net/sunrpc/clnt.c b/net/sunrpc/clnt.c
> index 0e75369..601b3ca 100644
> --- a/net/sunrpc/clnt.c
> +++ b/net/sunrpc/clnt.c
> @@ -1693,6 +1693,7 @@ call_allocate(struct rpc_task *task)
> struct rpc_rqst *req = task->tk_rqstp;
> struct rpc_xprt *xprt = req->rq_xprt;
> struct rpc_procinfo *proc = task->tk_msg.rpc_proc;
> + int status;
>
> dprint_status(task);
>
> @@ -1718,11 +1719,14 @@ call_allocate(struct rpc_task *task)
> req->rq_rcvsize = RPC_REPHDRSIZE + slack + proc->p_replen;
> req->rq_rcvsize <<= 2;
>
> - req->rq_buffer = xprt->ops->buf_alloc(task,
> - req->rq_callsize + req->rq_rcvsize);
> - if (req->rq_buffer != NULL)
> - return;
> + status = xprt->ops->buf_alloc(task);
> xprt_inject_disconnect(xprt);
> + if (status == 0)
> + return;
> + if (status == -EIO) {
> + rpc_exit(task, -EIO);
What does EIO mean in this context, and why should it be a fatal error? A buffer allocation is not normally considered to be I/O.
> + return;
> + }
>
> dprintk("RPC: %5u rpc_buffer allocation failed\n", task->tk_pid);
>
> diff --git a/net/sunrpc/sched.c b/net/sunrpc/sched.c
> index 9ae5885..b964d40 100644
> --- a/net/sunrpc/sched.c
> +++ b/net/sunrpc/sched.c
> @@ -849,14 +849,17 @@ static void rpc_async_schedule(struct work_struct *work)
> }
>
> /**
> - * rpc_malloc - allocate an RPC buffer
> - * @task: RPC task that will use this buffer
> - * @size: requested byte size
> + * rpc_malloc - allocate RPC buffer resources
> + * @task: RPC task
> + *
> + * A single memory region is allocated, which is split between the
> + * RPC call and RPC reply that this task is being used for. When
> + * this RPC is retired, the memory is released by calling rpc_free.
> *
> * To prevent rpciod from hanging, this allocator never sleeps,
> - * returning NULL and suppressing warning if the request cannot be serviced
> - * immediately.
> - * The caller can arrange to sleep in a way that is safe for rpciod.
> + * returning -ENOMEM and suppressing warning if the request cannot
> + * be serviced immediately. The caller can arrange to sleep in a
> + * way that is safe for rpciod.
> *
> * Most requests are 'small' (under 2KiB) and can be serviced from a
> * mempool, ensuring that NFS reads and writes can always proceed,
> @@ -865,8 +868,10 @@ static void rpc_async_schedule(struct work_struct *work)
> * In order to avoid memory starvation triggering more writebacks of
> * NFS requests, we avoid using GFP_KERNEL.
> */
> -void *rpc_malloc(struct rpc_task *task, size_t size)
> +int rpc_malloc(struct rpc_task *task)
> {
> + struct rpc_rqst *rqst = task->tk_rqstp;
> + size_t size = rqst->rq_callsize + rqst->rq_rcvsize;
> struct rpc_buffer *buf;
> gfp_t gfp = GFP_NOIO | __GFP_NOWARN;
>
> @@ -880,12 +885,13 @@ void *rpc_malloc(struct rpc_task *task, size_t size)
> buf = kmalloc(size, gfp);
>
> if (!buf)
> - return NULL;
> + return -ENOMEM;
>
> buf->len = size;
> dprintk("RPC: %5u allocated buffer of size %zu at %p\n",
> task->tk_pid, size, buf);
> - return &buf->data;
> + rqst->rq_buffer = buf->data;
> + return 0;
> }
> EXPORT_SYMBOL_GPL(rpc_malloc);
>
> diff --git a/net/sunrpc/sunrpc.h b/net/sunrpc/sunrpc.h
> index f2b7cb5..a4f44ca 100644
> --- a/net/sunrpc/sunrpc.h
> +++ b/net/sunrpc/sunrpc.h
> @@ -34,7 +34,7 @@ SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
> */
> struct rpc_buffer {
> size_t len;
> - char data[];
> + __be32 data[];
Again, why? Nothing should be touching the data contents directly from this pointer.
> };
>
> static inline int rpc_reply_expected(struct rpc_task *task)
> diff --git a/net/sunrpc/xprtrdma/svc_rdma_backchannel.c b/net/sunrpc/xprtrdma/svc_rdma_backchannel.c
> index a2a7519..b7c698f7f 100644
> --- a/net/sunrpc/xprtrdma/svc_rdma_backchannel.c
> +++ b/net/sunrpc/xprtrdma/svc_rdma_backchannel.c
> @@ -159,29 +159,30 @@ out_unmap:
> /* Server-side transport endpoint wants a whole page for its send
> * buffer. The client RPC code constructs the RPC header in this
> * buffer before it invokes ->send_request.
> - *
> - * Returns NULL if there was a temporary allocation failure.
> */
> -static void *
> -xprt_rdma_bc_allocate(struct rpc_task *task, size_t size)
> +static int
> +xprt_rdma_bc_allocate(struct rpc_task *task)
> {
> struct rpc_rqst *rqst = task->tk_rqstp;
> struct svc_xprt *sxprt = rqst->rq_xprt->bc_xprt;
> + size_t size = rqst->rq_callsize;
> struct svcxprt_rdma *rdma;
> struct page *page;
>
> rdma = container_of(sxprt, struct svcxprt_rdma, sc_xprt);
>
> - /* Prevent an infinite loop: try to make this case work */
> - if (size > PAGE_SIZE)
> + if (size > PAGE_SIZE) {
> WARN_ONCE(1, "svcrdma: large bc buffer request (size %zu)\n",
> size);
> + return -EIO;
> + }
>
> page = alloc_page(RPCRDMA_DEF_GFP);
> if (!page)
> - return NULL;
> + return -ENOMEM;
>
> - return page_address(page);
> + rqst->rq_buffer = page_address(page);
> + return 0;
> }
>
> static void
> diff --git a/net/sunrpc/xprtrdma/transport.c b/net/sunrpc/xprtrdma/transport.c
> index be95ece..daa7d4d 100644
> --- a/net/sunrpc/xprtrdma/transport.c
> +++ b/net/sunrpc/xprtrdma/transport.c
> @@ -477,7 +477,15 @@ xprt_rdma_connect(struct rpc_xprt *xprt, struct rpc_task *task)
> }
> }
>
> -/*
> +/**
> + * xprt_rdma_allocate - allocate transport resources for an RPC
> + * @task: RPC task
> + *
> + * Return values:
> + * 0: Success; rq_buffer points to RPC buffer to use
> + * ENOMEM: Out of memory, call again later
> + * EIO: A permanent error occurred, do not retry
> + *
> * The RDMA allocate/free functions need the task structure as a place
> * to hide the struct rpcrdma_req, which is necessary for the actual send/recv
> * sequence.
> @@ -486,11 +494,12 @@ xprt_rdma_connect(struct rpc_xprt *xprt, struct rpc_task *task)
> * (rq_send_buf and rq_rcv_buf are both part of a single contiguous buffer).
> * We may register rq_rcv_buf when using reply chunks.
> */
> -static void *
> -xprt_rdma_allocate(struct rpc_task *task, size_t size)
> +static int
> +xprt_rdma_allocate(struct rpc_task *task)
> {
> - struct rpc_xprt *xprt = task->tk_rqstp->rq_xprt;
> - struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(xprt);
> + struct rpc_rqst *rqst = task->tk_rqstp;
> + size_t size = rqst->rq_callsize + rqst->rq_rcvsize;
> + struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(rqst->rq_xprt);
> struct rpcrdma_regbuf *rb;
> struct rpcrdma_req *req;
> size_t min_size;
> @@ -498,7 +507,7 @@ xprt_rdma_allocate(struct rpc_task *task, size_t size)
>
> req = rpcrdma_buffer_get(&r_xprt->rx_buf);
> if (req == NULL)
> - return NULL;
> + return -ENOMEM;
>
> flags = RPCRDMA_DEF_GFP;
> if (RPC_IS_SWAPPER(task))
> @@ -515,7 +524,8 @@ out:
> dprintk("RPC: %s: size %zd, request 0x%p\n", __func__, size, req);
> req->rl_connect_cookie = 0; /* our reserved value */
> req->rl_task = task;
> - return req->rl_sendbuf->rg_base;
> + rqst->rq_buffer = req->rl_sendbuf->rg_base;
> + return 0;
>
> out_rdmabuf:
> min_size = r_xprt->rx_data.inline_wsize;
> @@ -558,7 +568,7 @@ out_sendbuf:
>
> out_fail:
> rpcrdma_buffer_put(req);
> - return NULL;
> + return -ENOMEM;
> }
>
> /*
> diff --git a/net/sunrpc/xprtsock.c b/net/sunrpc/xprtsock.c
> index 8ede3bc..d6db5cf 100644
> --- a/net/sunrpc/xprtsock.c
> +++ b/net/sunrpc/xprtsock.c
> @@ -2533,23 +2533,28 @@ static void xs_tcp_print_stats(struct rpc_xprt *xprt, struct seq_file *seq)
> * we allocate pages instead doing a kmalloc like rpc_malloc is because we want
> * to use the server side send routines.
> */
> -static void *bc_malloc(struct rpc_task *task, size_t size)
> +static int bc_malloc(struct rpc_task *task)
> {
> + struct rpc_rqst *rqst = task->tk_rqstp;
> + size_t size = rqst->rq_callsize;
> struct page *page;
> struct rpc_buffer *buf;
>
> - WARN_ON_ONCE(size > PAGE_SIZE - sizeof(struct rpc_buffer));
> - if (size > PAGE_SIZE - sizeof(struct rpc_buffer))
> - return NULL;
> + if (size > PAGE_SIZE - sizeof(struct rpc_buffer)) {
> + WARN_ONCE(1, "xprtsock: large bc buffer request (size %zu)\n",
> + size);
> + return -EIO;
> + }
>
> page = alloc_page(GFP_KERNEL);
> if (!page)
> - return NULL;
> + return -ENOMEM;
>
> buf = page_address(page);
> buf->len = PAGE_SIZE;
>
> - return buf->data;
> + rqst->rq_buffer = buf->data;
> + return 0;
> }
>
> /*
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-nfs" in
> the body of a message to majordomo-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
>
--
To unsubscribe from this list: send the line "unsubscribe linux-nfs" in
the body of a message to majordomo-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
next prev parent reply other threads:[~2016-08-15 21:59 UTC|newest]
Thread overview: 28+ messages / expand[flat|nested] mbox.gz Atom feed top
2016-08-15 20:50 [PATCH v1 00/22] client-side NFS/RDMA patches proposed for v4.9 Chuck Lever
[not found] ` <20160815195649.11652.32252.stgit-FYjufvaPoItvLzlybtyyYzGyq/o6K9yX@public.gmane.org>
2016-08-15 20:50 ` [PATCH v1 01/22] xprtrdma: Eliminate INLINE_THRESHOLD macros Chuck Lever
2016-08-15 20:50 ` [PATCH v1 02/22] SUNRPC: Refactor rpc_xdr_buf_init() Chuck Lever
[not found] ` <20160815205030.11652.6620.stgit-FYjufvaPoItvLzlybtyyYzGyq/o6K9yX@public.gmane.org>
2016-08-15 21:51 ` Trond Myklebust
[not found] ` <0750E5E3-335B-4D49-8B17-45DC0A616225-7I+n7zu2hftEKMMhf/gKZA@public.gmane.org>
2016-08-16 1:24 ` Chuck Lever
2016-08-15 20:50 ` [PATCH v1 03/22] SUNRPC: Generalize the RPC buffer allocation API Chuck Lever
[not found] ` <20160815205038.11652.47749.stgit-FYjufvaPoItvLzlybtyyYzGyq/o6K9yX@public.gmane.org>
2016-08-15 21:59 ` Trond Myklebust [this message]
[not found] ` <CD81109C-4EB7-4925-9D30-852D3440ACF3-7I+n7zu2hftEKMMhf/gKZA@public.gmane.org>
2016-08-16 1:20 ` Chuck Lever
2016-08-15 20:50 ` [PATCH v1 04/22] SUNRPC: Generalize the RPC buffer release API Chuck Lever
2016-08-15 20:50 ` [PATCH v1 05/22] SUNRPC: Separate buffer pointers for RPC Call and Reply messages Chuck Lever
2016-08-15 20:51 ` [PATCH v1 06/22] SUNRPC: Add a transport-specific private field in rpc_rqst Chuck Lever
2016-08-15 20:51 ` [PATCH v1 07/22] xprtrdma: Initialize separate RPC call and reply buffers Chuck Lever
2016-08-15 20:51 ` [PATCH v1 08/22] xprtrdma: Use smaller buffers for RPC-over-RDMA headers Chuck Lever
2016-08-15 20:51 ` [PATCH v1 09/22] xprtrdma: Replace DMA_BIDIRECTIONAL Chuck Lever
2016-08-15 20:51 ` [PATCH v1 10/22] xprtrdma: Delay DMA mapping Send and Receive buffers Chuck Lever
2016-08-15 20:51 ` [PATCH v1 11/22] xprtrdma: Eliminate "ia" argument in rpcrdma_{alloc, free}_regbuf Chuck Lever
2016-08-15 20:51 ` [PATCH v1 12/22] xprtrdma: Simplify rpcrdma_ep_post_recv() Chuck Lever
2016-08-15 20:52 ` [PATCH v1 13/22] xprtrdma: Move send_wr to struct rpcrdma_req Chuck Lever
2016-08-15 20:52 ` [PATCH v1 14/22] xprtrdma: Move recv_wr to struct rpcrdma_rep Chuck Lever
2016-08-15 20:52 ` [PATCH v1 15/22] xprtrmda: Report address of frmr, not mw Chuck Lever
2016-08-15 20:52 ` [PATCH v1 16/22] rpcrdma: RDMA/CM private message data structure Chuck Lever
2016-08-15 20:52 ` [PATCH v1 17/22] xprtrdma: Client-side support for rpcrdma_connect_private Chuck Lever
2016-08-15 20:52 ` [PATCH v1 18/22] xprtrdma: Basic support for Remote Invalidation Chuck Lever
2016-08-15 20:52 ` [PATCH v1 19/22] xprtrdma: Use gathered Send for large inline messages Chuck Lever
[not found] ` <20160815205249.11652.23617.stgit-FYjufvaPoItvLzlybtyyYzGyq/o6K9yX@public.gmane.org>
2016-08-15 21:46 ` kbuild test robot
2016-08-15 20:52 ` [PATCH v1 20/22] xprtrdma: Support larger inline thresholds Chuck Lever
2016-08-15 20:53 ` [PATCH v1 21/22] xprtrdma: Rename rpcrdma_receive_wc() Chuck Lever
2016-08-15 20:53 ` [PATCH v1 22/22] xprtrdma: Eliminate rpcrdma_receive_worker() Chuck Lever
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=CD81109C-4EB7-4925-9D30-852D3440ACF3@primarydata.com \
--to=trondmy-7i+n7zu2hftekmmhf/gkza@public.gmane.org \
--cc=chuck.lever-QHcLZuEGTsvQT0dZR+AlfA@public.gmane.org \
--cc=linux-nfs-u79uwXL29TY76Z2rM5mHXA@public.gmane.org \
--cc=linux-rdma-u79uwXL29TY76Z2rM5mHXA@public.gmane.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox