From: swise@opengridcomputing.com (Steve Wise)
Subject: [PATCH v3 3/3] nvmet-rdma: support 16K inline data
Date: Sun, 3 Jun 2018 13:25:47 -0500 [thread overview]
Message-ID: <00fe01d3fb68$4b5e2cf0$e21a86d0$@opengridcomputing.com> (raw)
In-Reply-To: <ff362b8e-c223-2d81-b355-80397745bd2a@mellanox.com>
> -----Original Message-----
> From: Max Gurtovoy <maxg at mellanox.com>
> Sent: Sunday, June 3, 2018 3:40 AM
> To: Steve Wise <swise at opengridcomputing.com>; axboe at kernel.dk;
> hch at lst.de; keith.busch at intel.com; sagi at grimberg.me; linux-
> nvme at lists.infradead.org
> Cc: parav at mellanox.com; linux-rdma at vger.kernel.org
> Subject: Re: [PATCH v3 3/3] nvmet-rdma: support 16K inline data
>
>
>
> On 5/29/2018 9:25 PM, Steve Wise wrote:
> > Add a new configfs port attribute, called inline_data_size,
> > to allow configuring the size of inline data for a given port.
> > The maximum size allowed is still enforced by nvmet-rdma with
> > NVMET_RDMA_MAX_INLINE_DATA_SIZE, which is increased to max(16KB,
> > PAGE_SIZE). And the default size, if not specified via configfs,
> > is still PAGE_SIZE. This preserves the existing behavior, but allows
> > larger inline sizes.
> >
> > Also support a configuration where inline_data_size is 0, which disables
> > using inline data.
> >
> > Signed-off-by: Steve Wise <swise at opengridcomputing.com>
> > ---
> > drivers/nvme/target/admin-cmd.c | 4 ++--
> > drivers/nvme/target/configfs.c | 31 ++++++++++++++++++++++++++++
> > drivers/nvme/target/core.c | 4 ++++
> > drivers/nvme/target/discovery.c | 2 +-
> > drivers/nvme/target/nvmet.h | 2 +-
> > drivers/nvme/target/rdma.c | 45 ++++++++++++++++++++++++++++-----
> --------
> > 6 files changed, 70 insertions(+), 18 deletions(-)
>
> snip..
>
> >
> > diff --git a/drivers/nvme/target/rdma.c b/drivers/nvme/target/rdma.c
> > index 52e0c5d..2f0b08e 100644
> > --- a/drivers/nvme/target/rdma.c
> > +++ b/drivers/nvme/target/rdma.c
> > @@ -33,9 +33,10 @@
> > #include "nvmet.h"
> >
> > /*
> > - * We allow up to a page of inline data to go with the SQE
> > + * We allow at least 1 page, and up to 16KB of inline data to go with
the
> SQE
> > */
> > -#define NVMET_RDMA_INLINE_DATA_SIZE PAGE_SIZE
> > +#define NVMET_RDMA_DEFAULT_INLINE_DATA_SIZE PAGE_SIZE
> > +#define NVMET_RDMA_MAX_INLINE_DATA_SIZE max_t(int,
> SZ_16K, PAGE_SIZE)
>
> why not use SZ_16K ? why we need to mention the PAGE_SIZE ?
>
The idea is to allow at least 1 page. So for, say, a 64K page system, we'll
allow 64K since we're allocating a page minimum for the buffer.
> >
> > struct nvmet_rdma_cmd {
> > struct ib_sge sge[2];
> > @@ -116,6 +117,7 @@ struct nvmet_rdma_device {
> > size_t srq_size;
> > struct kref ref;
> > struct list_head entry;
> > + int inline_data_size;
> > };
> >
> > static bool nvmet_rdma_use_srq;
> > @@ -187,6 +189,8 @@ static inline bool
> nvmet_rdma_need_data_out(struct nvmet_rdma_rsp *rsp)
> > static int nvmet_rdma_alloc_cmd(struct nvmet_rdma_device *ndev,
> > struct nvmet_rdma_cmd *c, bool admin)
> > {
> > + int inline_data_size = ndev->inline_data_size;
> > +
> > /* NVMe command / RDMA RECV */
> > c->nvme_cmd = kmalloc(sizeof(*c->nvme_cmd), GFP_KERNEL);
> > if (!c->nvme_cmd)
> > @@ -200,17 +204,17 @@ static int nvmet_rdma_alloc_cmd(struct
> nvmet_rdma_device *ndev,
> > c->sge[0].length = sizeof(*c->nvme_cmd);
> > c->sge[0].lkey = ndev->pd->local_dma_lkey;
> >
> > - if (!admin) {
> > + if (!admin && inline_data_size) {
> > c->inline_page = alloc_pages(GFP_KERNEL,
> > -
> get_order(NVMET_RDMA_INLINE_DATA_SIZE));
> > + get_order(inline_data_size));
> > if (!c->inline_page)
> > goto out_unmap_cmd;
> > c->sge[1].addr = ib_dma_map_page(ndev->device,
> > - c->inline_page, 0,
> NVMET_RDMA_INLINE_DATA_SIZE,
> > + c->inline_page, 0, inline_data_size,
> > DMA_FROM_DEVICE);
> > if (ib_dma_mapping_error(ndev->device, c->sge[1].addr))
> > goto out_free_inline_page;
> > - c->sge[1].length = NVMET_RDMA_INLINE_DATA_SIZE;
> > + c->sge[1].length = inline_data_size;
> > c->sge[1].lkey = ndev->pd->local_dma_lkey;
> > }
> >
> > @@ -225,7 +229,7 @@ static int nvmet_rdma_alloc_cmd(struct
> nvmet_rdma_device *ndev,
> > out_free_inline_page:
> > if (!admin) {
> > __free_pages(c->inline_page,
> > -
> get_order(NVMET_RDMA_INLINE_DATA_SIZE));
> > + get_order(inline_data_size));
> > }
> > out_unmap_cmd:
> > ib_dma_unmap_single(ndev->device, c->sge[0].addr,
> > @@ -240,11 +244,13 @@ static int nvmet_rdma_alloc_cmd(struct
> nvmet_rdma_device *ndev,
> > static void nvmet_rdma_free_cmd(struct nvmet_rdma_device *ndev,
> > struct nvmet_rdma_cmd *c, bool admin)
> > {
> > - if (!admin) {
> > + int inline_data_size = ndev->inline_data_size;
> > +
> > + if (!admin && inline_data_size) {
> > ib_dma_unmap_page(ndev->device, c->sge[1].addr,
> > - NVMET_RDMA_INLINE_DATA_SIZE,
> DMA_FROM_DEVICE);
> > + inline_data_size, DMA_FROM_DEVICE);
> > __free_pages(c->inline_page,
> > -
> get_order(NVMET_RDMA_INLINE_DATA_SIZE));
> > + get_order(inline_data_size));
> > }
> > ib_dma_unmap_single(ndev->device, c->sge[0].addr,
> > sizeof(*c->nvme_cmd),
> DMA_FROM_DEVICE);
> > @@ -544,7 +550,7 @@ static u16 nvmet_rdma_map_sgl_inline(struct
> nvmet_rdma_rsp *rsp)
> > if (!nvme_is_write(rsp->req.cmd))
> > return NVME_SC_INVALID_FIELD | NVME_SC_DNR;
> >
> > - if (off + len > NVMET_RDMA_INLINE_DATA_SIZE) {
> > + if (off + len > rsp->queue->dev->inline_data_size) {
> > pr_err("invalid inline data offset!\n");
> > return NVME_SC_SGL_INVALID_OFFSET | NVME_SC_DNR;
> > }
> > @@ -793,6 +799,7 @@ static void nvmet_rdma_free_dev(struct kref *ref)
> > static struct nvmet_rdma_device *
> > nvmet_rdma_find_get_device(struct rdma_cm_id *cm_id)
> > {
> > + struct nvmet_port *port = cm_id->context;
> > struct nvmet_rdma_device *ndev;
> > int ret;
> >
> > @@ -807,6 +814,7 @@ static void nvmet_rdma_free_dev(struct kref *ref)
> > if (!ndev)
> > goto out_err;
> >
> > + ndev->inline_data_size = port->inline_data_size;
> > ndev->device = cm_id->device;
> > kref_init(&ndev->ref);
> >
> > @@ -1379,6 +1387,15 @@ static int nvmet_rdma_add_port(struct
> nvmet_port *port)
> > return -EINVAL;
> > }
> >
> > + if (port->inline_data_size < 0) {
> > + port->inline_data_size =
> NVMET_RDMA_DEFAULT_INLINE_DATA_SIZE;
> > + } else if (port->inline_data_size >
> NVMET_RDMA_MAX_INLINE_DATA_SIZE) {
> > + pr_err("invalid inline_data_size %d (max supported is
> %u)\n",
> > + port->inline_data_size,
> > + NVMET_RDMA_MAX_INLINE_DATA_SIZE);
> > + return -EINVAL;
> > + }
> > +
> > ret = inet_pton_with_scope(&init_net, af, port->disc_addr.traddr,
> > port->disc_addr.trsvcid, &addr);
> > if (ret) {
> > @@ -1418,8 +1435,9 @@ static int nvmet_rdma_add_port(struct
> nvmet_port *port)
> > goto out_destroy_id;
> > }
> >
> > - pr_info("enabling port %d (%pISpcs)\n",
> > - le16_to_cpu(port->disc_addr.portid), (struct sockaddr
> *)&addr);
> > + pr_info("enabling port %d (%pISpcs) inline_data_size %d\n",
> > + le16_to_cpu(port->disc_addr.portid), (struct sockaddr
> *)&addr,
> > + port->inline_data_size);
> > port->priv = cm_id;
> > return 0;
> >
> > @@ -1456,7 +1474,6 @@ static void nvmet_rdma_disc_port_addr(struct
> nvmet_req *req,
> > static const struct nvmet_fabrics_ops nvmet_rdma_ops = {
> > .owner = THIS_MODULE,
> > .type = NVMF_TRTYPE_RDMA,
> > - .sqe_inline_size = NVMET_RDMA_INLINE_DATA_SIZE,
> > .msdbd = 1,
> > .has_keyed_sgls = 1,
> > .add_port = nvmet_rdma_add_port,
> >
next prev parent reply other threads:[~2018-06-03 18:25 UTC|newest]
Thread overview: 39+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-05-29 18:26 [PATCH v3 0/3] NVMF/RDMA 16K Inline Support Steve Wise
2018-05-29 18:25 ` [PATCH v3 1/3] nvme-rdma: correctly check for target keyed sgl support Steve Wise
2018-05-29 20:23 ` Ruhl, Michael J
2018-05-30 14:39 ` Steve Wise
2018-05-30 15:11 ` Steve Wise
2018-05-30 21:37 ` Sagi Grimberg
2018-05-31 17:02 ` hch
2018-05-31 17:17 ` Steve Wise
2018-05-31 17:25 ` hch
2018-06-01 13:08 ` Steve Wise
2018-06-03 11:57 ` Sagi Grimberg
2018-06-03 18:27 ` Steve Wise
2018-06-04 12:01 ` Sagi Grimberg
2018-06-04 12:11 ` Christoph Hellwig
2018-06-04 12:17 ` Steve Wise
2018-06-04 13:52 ` Max Gurtovoy
2018-06-04 14:21 ` Steve Wise
2018-06-04 14:29 ` Max Gurtovoy
2018-06-04 14:31 ` Steve Wise
2018-06-04 14:37 ` Max Gurtovoy
2018-06-04 14:45 ` Steve Wise
2018-05-31 17:00 ` hch
2018-05-29 18:25 ` [PATCH v3 2/3] nvme-rdma: support up to 4 segments of inline data Steve Wise
2018-05-30 21:42 ` Sagi Grimberg
2018-05-30 21:46 ` Steve Wise
2018-05-29 18:25 ` [PATCH v3 3/3] nvmet-rdma: support 16K " Steve Wise
2018-05-30 15:49 ` Christopher Lameter
2018-05-30 16:46 ` Steve Wise
2018-05-30 17:02 ` Christopher Lameter
2018-05-30 21:45 ` Sagi Grimberg
2018-05-30 21:52 ` Steve Wise
2018-05-30 22:13 ` Sagi Grimberg
2018-05-30 22:26 ` Steve Wise
2018-06-03 8:39 ` Max Gurtovoy
2018-06-03 18:25 ` Steve Wise [this message]
2018-06-04 13:58 ` Max Gurtovoy
2018-06-04 14:18 ` Steve Wise
2018-06-05 8:52 ` Max Gurtovoy
2018-06-05 14:28 ` Steve Wise
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to='00fe01d3fb68$4b5e2cf0$e21a86d0$@opengridcomputing.com' \
--to=swise@opengridcomputing.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).