From: "Steve Wise" <swise-7bPotxP6k4+P2YhJcF5u+vpXobYPEAuW@public.gmane.org>
To: 'Doug Ledford' <dledford-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>,
linux-rdma-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
Cc: bart.vanassche-XdAiOPVOjttBDgjK7y7TUQ@public.gmane.org
Subject: RE: [PATCH 1/3] IB: new common API for draining queues
Date: Mon, 15 Feb 2016 15:20:36 -0600 [thread overview]
Message-ID: <01aa01d16836$b6510220$22f30660$@opengridcomputing.com> (raw)
In-Reply-To: <56C23DA8.40905-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>
> -----Original Message-----
> From: Doug Ledford [mailto:dledford-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org]
> Sent: Monday, February 15, 2016 3:06 PM
> To: Steve Wise; linux-rdma-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
> Cc: bart.vanassche-XdAiOPVOjttBDgjK7y7TUQ@public.gmane.org
> Subject: Re: [PATCH 1/3] IB: new common API for draining queues
>
> On 02/05/2016 04:13 PM, Steve Wise wrote:
> > From: Steve Wise <swise-7bPotxP6k4+P2YhJcF5u+p6NamaJ0bNTAL8bYrjMMd8@public.gmane.org>
> >
> > Add provider-specific drain_sq/drain_rq functions for providers needing
> > special drain logic.
> >
> > Add static functions __ib_drain_sq() and __ib_drain_rq() which post noop
> > WRs to the SQ or RQ and block until their completions are processed.
> > This ensures the applications completions have all been processed.
>
> Except it doesn't, comments inline below...and applications is
> possessive, so needs a '
>
> > Add API functions ib_drain_sq(), ib_drain_rq(), and ib_drain_qp().
> >
> > Reviewed-by: Chuck Lever <chuck.lever-QHcLZuEGTsvQT0dZR+AlfA@public.gmane.org>
> > Signed-off-by: Steve Wise <swise-7bPotxP6k4+P2YhJcF5u+vpXobYPEAuW@public.gmane.org>
> > ---
> > drivers/infiniband/core/verbs.c | 144 ++++++++++++++++++++++++++++++++++++++++
> > include/rdma/ib_verbs.h | 5 ++
> > 2 files changed, 149 insertions(+)
> >
> > diff --git a/drivers/infiniband/core/verbs.c b/drivers/infiniband/core/verbs.c
> > index 5af6d02..aed521e 100644
> > --- a/drivers/infiniband/core/verbs.c
> > +++ b/drivers/infiniband/core/verbs.c
> > @@ -1657,3 +1657,147 @@ next_page:
> > return i;
> > }
> > EXPORT_SYMBOL(ib_sg_to_pages);
> > +
> > +struct ib_drain_cqe {
> > + struct ib_cqe cqe;
> > + struct completion done;
> > +};
> > +
> > +static void ib_drain_qp_done(struct ib_cq *cq, struct ib_wc *wc)
> > +{
> > + struct ib_drain_cqe *cqe = container_of(wc->wr_cqe, struct ib_drain_cqe,
> > + cqe);
> > +
> > + complete(&cqe->done);
> > +}
> > +
> > +static void wait_for_drain(struct ib_cq *cq, struct completion *c)
> > +{
> > + if (cq->poll_ctx == IB_POLL_DIRECT)
> > + do
> > + ib_process_cq_direct(cq, 1024);
> > + while (!wait_for_completion_timeout(c, msecs_to_jiffies(100)));
> > + else
> > + wait_for_completion(c);
> > +}
> > +
> > +/*
> > + * Post a WR and block until its completion is reaped for the SQ.
> > + */
> > +static void __ib_drain_sq(struct ib_qp *qp)
> > +{
> > + struct ib_qp_attr attr = { .qp_state = IB_QPS_ERR };
> > + struct ib_drain_cqe sdrain;
> > + struct ib_send_wr swr = {}, *bad_swr;
> > + int ret;
> > +
> > + swr.wr_cqe = &sdrain.cqe;
> > + sdrain.cqe.done = ib_drain_qp_done;
> > + init_completion(&sdrain.done);
> > +
> > + ret = ib_modify_qp(qp, &attr, IB_QP_STATE);
> > + if (ret) {
> > + WARN_ONCE(ret, "failed to drain QP: %d\n", ret);
> > + return;
> > + }
>
> Set QP to ERR state...OK...
>
> > + ret = ib_post_send(qp, &swr, &bad_swr);
> > + if (ret) {
> > + WARN_ONCE(ret, "failed to drain send queue: %d\n", ret);
> > + return;
> > + }
>
> Post command to QP in ERR state (admittedly, I never do this and hadn't
> even thought about whether or not it is allowed...obviously it is, but
> my initial reaction would have been "won't ib_post_send return an error
> when you try to post to a QP in ERR state?")....OK...
>
IBTA spec sez a post to a QP in ERR will result in a CQE in the CQ with status FLUSHED. IE it mandates this behavior.
> > + wait_for_drain(qp->send_cq, &sdrain.done);
>
> Wait for your posted WR to complete...OK...
>
> As I mentioned in my comment above, I would have thought that the
> attempt to post a send to a QP in ERR state would have returned an
> error. It must not or else this patch is worthless because of the order
> of actions. What that highlights though, is that this code will drain a
> QP, but only if the caller has taken the time to stop all possible
> contexts that might run on other cores and post commands to the QP.
> Those commands will error out, but the caller must, none the less, take
> steps to block other contexts from sending or else this drain is
> useless. That might be fine for the API, but it should be clearly
> documented, and currently it isn't.
>
I'll clarify this in the function header comments.
> > +}
> > +
> > +/*
> > + * Post a WR and block until its completion is reaped for the RQ.
> > + */
> > +static void __ib_drain_rq(struct ib_qp *qp)
> > +{
> > + struct ib_qp_attr attr = { .qp_state = IB_QPS_ERR };
> > + struct ib_drain_cqe rdrain;
> > + struct ib_recv_wr rwr = {}, *bad_rwr;
> > + int ret;
> > +
> > + rwr.wr_cqe = &rdrain.cqe;
> > + rdrain.cqe.done = ib_drain_qp_done;
> > + init_completion(&rdrain.done);
> > +
> > + ret = ib_modify_qp(qp, &attr, IB_QP_STATE);
> > + if (ret) {
> > + WARN_ONCE(ret, "failed to drain QP: %d\n", ret);
> > + return;
> > + }
> > +
> > + ret = ib_post_recv(qp, &rwr, &bad_rwr);
> > + if (ret) {
> > + WARN_ONCE(ret, "failed to drain send queue: %d\n", ret);
> > + return;
> > + }
> > +
> > + wait_for_drain(qp->recv_cq, &rdrain.done);
> > +}
> > +
> > +/**
> > + * ib_drain_sq() - Block until all SQ CQEs have been consumed by the
> > + * application.
> > + * @qp: queue pair to drain
> > + *
> > + * If the device has a provider-specific drain function, then
> > + * call that. Otherwise call the generic drain function
> > + * __ib_drain_sq().
> > + *
> > + * The consumer must ensure there is room in the CQ and SQ
> > + * for a drain work requests. Also the consumer must allocate the
>
> Either requests should be singular, or you should remove the article 'a'.
>
agreed (and agree to all the dittos)
> > + * CQ using ib_alloc_cq(). It is up to the consumer to handle concurrency
> > + * issues if the CQs are using the IB_POLL_DIRECT polling context.
> > + */
> > +void ib_drain_sq(struct ib_qp *qp)
> > +{
> > + if (qp->device->drain_sq)
> > + qp->device->drain_sq(qp);
> > + else
> > + __ib_drain_sq(qp);
> > +}
> > +EXPORT_SYMBOL(ib_drain_sq);
> > +
> > +/**
> > + * ib_drain_rq() - Block until all RQ CQEs have been consumed by the
> > + * application.
> > + * @qp: queue pair to drain
> > + *
> > + * If the device has a provider-specific drain function, then
> > + * call that. Otherwise call the generic drain function
> > + * __ib_drain_rq().
> > + *
> > + * The consumer must ensure there is room in the CQ and RQ
> > + * for a drain work requests. Also the consumer must allocate the
>
> Ditto...
>
> > + * CQ using ib_alloc_cq(). It is up to the consumer to handle concurrency
> > + * issues if the CQs are using the IB_POLL_DIRECT polling context.
> > + */
> > +void ib_drain_rq(struct ib_qp *qp)
> > +{
> > + if (qp->device->drain_rq)
> > + qp->device->drain_rq(qp);
> > + else
> > + __ib_drain_rq(qp);
> > +}
> > +EXPORT_SYMBOL(ib_drain_rq);
> > +
> > +/**
> > + * ib_drain_qp() - Block until all CQEs have been consumed by the
> > + * application on both the RQ and SQ.
> > + * @qp: queue pair to drain
> > + *
> > + * The consumer must ensure there is room in the CQ(s), SQ, and RQ
> > + * for a drain work requests. Also the consumer must allocate the
>
> Ditto...
>
> > + * CQ using ib_alloc_cq(). It is up to the consumer to handle concurrency
> > + * issues if the CQs are using the IB_POLL_DIRECT polling context.
> > + */
> > +void ib_drain_qp(struct ib_qp *qp)
> > +{
> > + ib_drain_sq(qp);
> > + ib_drain_rq(qp);
> > +}
> > +EXPORT_SYMBOL(ib_drain_qp);
> > diff --git a/include/rdma/ib_verbs.h b/include/rdma/ib_verbs.h
> > index 284b00c..68b7e97 100644
> > --- a/include/rdma/ib_verbs.h
> > +++ b/include/rdma/ib_verbs.h
> > @@ -1846,6 +1846,8 @@ struct ib_device {
> > int (*check_mr_status)(struct ib_mr *mr, u32 check_mask,
> > struct ib_mr_status *mr_status);
> > void (*disassociate_ucontext)(struct ib_ucontext *ibcontext);
> > + void (*drain_rq)(struct ib_qp *qp);
> > + void (*drain_sq)(struct ib_qp *qp);
> >
> > struct ib_dma_mapping_ops *dma_ops;
> >
> > @@ -3094,4 +3096,7 @@ int ib_sg_to_pages(struct ib_mr *mr,
> > int sg_nents,
> > int (*set_page)(struct ib_mr *, u64));
> >
> > +void ib_drain_rq(struct ib_qp *qp);
> > +void ib_drain_sq(struct ib_qp *qp);
> > +void ib_drain_qp(struct ib_qp *qp);
> > #endif /* IB_VERBS_H */
> >
>
>
> --
> Doug Ledford <dledford-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>
> GPG KeyID: 0E572FDD
>
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
next prev parent reply other threads:[~2016-02-15 21:20 UTC|newest]
Thread overview: 22+ messages / expand[flat|nested] mbox.gz Atom feed top
2016-02-11 22:44 [PATCH v3 0/3] new ib_drain_qp() API Steve Wise
[not found] ` <cover.1455230646.git.swise-ut6Up61K2wZBDgjK7y7TUQ@public.gmane.org>
2016-01-14 18:24 ` [PATCH 2/3] iw_cxgb4: add queue drain functions Steve Wise
2016-01-27 20:09 ` [PATCH 3/3] IB/srp: use ib_drain_rq() Steve Wise
[not found] ` <c11baa726a6440549ab46b9525116d9fe74eb5a0.1455230646.git.swise-ut6Up61K2wZBDgjK7y7TUQ@public.gmane.org>
2016-02-11 23:01 ` Bart Van Assche
2016-02-05 21:13 ` [PATCH 1/3] IB: new common API for draining queues Steve Wise
[not found] ` <3e7261d1436d33320223d365974ff38945f0d558.1455230646.git.swise-ut6Up61K2wZBDgjK7y7TUQ@public.gmane.org>
2016-02-11 22:59 ` Bart Van Assche
[not found] ` <56BD1248.80805-XdAiOPVOjttBDgjK7y7TUQ@public.gmane.org>
2016-02-11 23:15 ` Bart Van Assche
[not found] ` <56BD1624.4090309-XdAiOPVOjttBDgjK7y7TUQ@public.gmane.org>
2016-02-11 23:20 ` Steve Wise
2016-02-11 23:23 ` Bart Van Assche
[not found] ` <56BD1800.9050508-XdAiOPVOjttBDgjK7y7TUQ@public.gmane.org>
2016-02-12 0:32 ` Steve Wise
2016-02-11 23:18 ` Steve Wise
2016-02-12 5:19 ` Devesh Sharma
[not found] ` <CANjDDBg=B33kRDTZ=NnZ-cZhNwXnpJ950dLy6qY0QZBjDaNisQ-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
2016-02-12 15:17 ` Steve Wise
2016-02-13 15:41 ` Devesh Sharma
2016-02-15 21:05 ` Doug Ledford
[not found] ` <56C23DA8.40905-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>
2016-02-15 21:20 ` Steve Wise [this message]
2016-02-16 11:00 ` Sagi Grimberg
2016-02-13 16:10 ` [PATCH v3 0/3] new ib_drain_qp() API Leon Romanovsky
[not found] ` <20160213161049.GB14741-2ukJVAZIZ/Y@public.gmane.org>
2016-02-13 16:32 ` Christoph Hellwig
[not found] ` <20160213163253.GA8843-wEGCiKHe2LqWVfeAwA7xHQ@public.gmane.org>
2016-02-13 16:50 ` Leon Romanovsky
[not found] ` <20160213165001.GC14741-2ukJVAZIZ/Y@public.gmane.org>
2016-02-14 14:51 ` Steve Wise
[not found] ` <56C0946E.50100-7bPotxP6k4+P2YhJcF5u+vpXobYPEAuW@public.gmane.org>
2016-02-14 14:58 ` Christoph Hellwig
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to='01aa01d16836$b6510220$22f30660$@opengridcomputing.com' \
--to=swise-7bpotxp6k4+p2yhjcf5u+vpxobypeauw@public.gmane.org \
--cc=bart.vanassche-XdAiOPVOjttBDgjK7y7TUQ@public.gmane.org \
--cc=dledford-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org \
--cc=linux-rdma-u79uwXL29TY76Z2rM5mHXA@public.gmane.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).