From: Maxim Levitsky <mlevitsk@redhat.com>
To: Fam Zheng <fam@euphon.net>
Cc: Kevin Wolf <kwolf@redhat.com>,
qemu-devel@nongnu.org, qemu-block@nongnu.org,
Max Reitz <mreitz@redhat.com>
Subject: Re: [Qemu-devel] [PATCH v2 5/5] block/nvme: add support for discard
Date: Thu, 06 Jun 2019 10:31:29 +0300 [thread overview]
Message-ID: <d2ed59512f533566cbe1750fc7e52c9c5396e49c.camel@redhat.com> (raw)
In-Reply-To: <20190606031932.3nowbnt67o4omxjv@debian>
On Thu, 2019-06-06 at 11:19 +0800, Fam Zheng wrote:
> On Wed, 04/17 22:53, Maxim Levitsky wrote:
> > Signed-off-by: Maxim Levitsky <mlevitsk@redhat.com>
> > ---
> > block/nvme.c | 80 ++++++++++++++++++++++++++++++++++++++++++++++
> > block/trace-events | 2 ++
> > 2 files changed, 82 insertions(+)
> >
> > diff --git a/block/nvme.c b/block/nvme.c
> > index 35b925899f..b83912c627 100644
> > --- a/block/nvme.c
> > +++ b/block/nvme.c
> > @@ -110,6 +110,7 @@ typedef struct {
> > bool plugged;
> >
> > bool supports_write_zeros;
> > + bool supports_discard;
> >
> > CoMutex dma_map_lock;
> > CoQueue dma_flush_queue;
> > @@ -462,6 +463,7 @@ static void nvme_identify(BlockDriverState *bs, int namespace, Error **errp)
> >
> >
> > s->supports_write_zeros = (idctrl->oncs & NVME_ONCS_WRITE_ZEROS) != 0;
> > + s->supports_discard = (idctrl->oncs & NVME_ONCS_DSM) != 0;
> >
> > memset(resp, 0, 4096);
> >
> > @@ -1144,6 +1146,83 @@ static coroutine_fn int nvme_co_pwrite_zeroes(BlockDriverState *bs,
> > }
> >
> >
> > +static int coroutine_fn nvme_co_pdiscard(BlockDriverState *bs,
> > + int64_t offset, int bytes)
>
> While you respin, you can align the parameters.
Hi Fam!!
I didn't knew that this is also required by qemu codeing style (it kind of suggested in the kernel)
I'll be more that glad to do so!
>
> > +{
> > + BDRVNVMeState *s = bs->opaque;
> > + NVMeQueuePair *ioq = s->queues[1];
> > + NVMeRequest *req;
> > + NvmeDsmRange *buf;
> > + QEMUIOVector local_qiov;
> > + int r;
> > +
> > + NvmeCmd cmd = {
> > + .opcode = NVME_CMD_DSM,
> > + .nsid = cpu_to_le32(s->nsid),
> > + .cdw10 = 0, /*number of ranges - 0 based*/
> > + .cdw11 = cpu_to_le32(1 << 2), /*deallocate bit*/
> > + };
> > +
> > + NVMeCoData data = {
> > + .ctx = bdrv_get_aio_context(bs),
> > + .ret = -EINPROGRESS,
> > + };
> > +
> > + if (!s->supports_discard) {
> > + return -ENOTSUP;
> > + }
> > +
> > + assert(s->nr_queues > 1);
> > +
> > + buf = qemu_try_blockalign0(bs, 4096);
> > + if (!buf) {
> > + return -ENOMEM;
> > + }
> > +
> > + buf->nlb = bytes >> s->blkshift;
> > + buf->slba = offset >> s->blkshift;
>
> This buffer is for the device, do we need to do anything about the endianness?
Thank you very very much, this is indeed an endianess bug.
Thanks a lot for the review,
Best regards,
Maxim Levitsky
>
> > + buf->cattr = 0;
> > +
> > + qemu_iovec_init(&local_qiov, 1);
> > + qemu_iovec_add(&local_qiov, buf, 4096);
> > +
> > + req = nvme_get_free_req(ioq);
> > + assert(req);
> > +
> > + qemu_co_mutex_lock(&s->dma_map_lock);
> > + r = nvme_cmd_map_qiov(bs, &cmd, req, &local_qiov);
> > + qemu_co_mutex_unlock(&s->dma_map_lock);
> > +
> > + if (r) {
> > + req->busy = false;
> > + return r;
> > + }
> > +
> > + trace_nvme_dsm(s, offset, bytes);
> > +
> > + nvme_submit_command(s, ioq, req, &cmd, nvme_rw_cb, &data);
> > +
> > + data.co = qemu_coroutine_self();
> > + while (data.ret == -EINPROGRESS) {
> > + qemu_coroutine_yield();
> > + }
> > +
> > + qemu_co_mutex_lock(&s->dma_map_lock);
> > + r = nvme_cmd_unmap_qiov(bs, &local_qiov);
> > + qemu_co_mutex_unlock(&s->dma_map_lock);
> > + if (r) {
> > + return r;
> > + }
> > +
> > + trace_nvme_dsm_done(s, offset, bytes, data.ret);
> > +
> > + qemu_iovec_destroy(&local_qiov);
> > + qemu_vfree(buf);
> > + return data.ret;
> > +
> > +}
> > +
> > +
> > static int nvme_reopen_prepare(BDRVReopenState *reopen_state,
> > BlockReopenQueue *queue, Error **errp)
> > {
> > @@ -1250,6 +1329,7 @@ static BlockDriver bdrv_nvme = {
> > .bdrv_co_pwritev = nvme_co_pwritev,
> >
> > .bdrv_co_pwrite_zeroes = nvme_co_pwrite_zeroes,
> > + .bdrv_co_pdiscard = nvme_co_pdiscard,
> >
> > .bdrv_co_flush_to_disk = nvme_co_flush,
> > .bdrv_reopen_prepare = nvme_reopen_prepare,
> > diff --git a/block/trace-events b/block/trace-events
> > index 943a58569f..e55ac5c40b 100644
> > --- a/block/trace-events
> > +++ b/block/trace-events
> > @@ -148,6 +148,8 @@ nvme_write_zeros(void *s, uint64_t offset, uint64_t bytes, int flags) "s %p offs
> > nvme_qiov_unaligned(const void *qiov, int n, void *base, size_t size, int align) "qiov %p n %d base %p size 0x%zx align 0x%x"
> > nvme_prw_buffered(void *s, uint64_t offset, uint64_t bytes, int niov, int is_write) "s %p offset %"PRId64" bytes %"PRId64" niov %d is_write %d"
> > nvme_rw_done(void *s, int is_write, uint64_t offset, uint64_t bytes, int ret) "s %p is_write %d offset %"PRId64" bytes %"PRId64" ret %d"
> > +nvme_dsm(void *s, uint64_t offset, uint64_t bytes) "s %p offset %"PRId64" bytes %"PRId64""
> > +nvme_dsm_done(void *s, uint64_t offset, uint64_t bytes, int ret) "s %p offset %"PRId64" bytes %"PRId64" ret %d"
> > nvme_dma_map_flush(void *s) "s %p"
> > nvme_free_req_queue_wait(void *q) "q %p"
> > nvme_cmd_map_qiov(void *s, void *cmd, void *req, void *qiov, int entries) "s %p cmd %p req %p qiov %p entries %d"
> > --
> > 2.17.2
> >
>
>
next prev parent reply other threads:[~2019-06-06 7:32 UTC|newest]
Thread overview: 22+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-04-17 19:53 [Qemu-devel] [PATCH v2 0/5] Few fixes for userspace NVME driver Maxim Levitsky
2019-04-17 19:53 ` Maxim Levitsky
2019-04-17 19:53 ` [Qemu-devel] [PATCH v2 1/5] block/nvme: don't flip CQ phase bits Maxim Levitsky
2019-04-17 19:53 ` Maxim Levitsky
2019-06-03 22:25 ` [Qemu-devel] [Qemu-block] " John Snow
2019-06-05 7:47 ` Maxim Levitsky
2019-06-06 21:23 ` John Snow
2019-06-07 11:08 ` Paolo Bonzini
2019-06-07 19:28 ` John Snow
2019-06-11 8:50 ` Maxim Levitsky
2019-04-17 19:53 ` [Qemu-devel] [PATCH v2 2/5] block/nvme: fix doorbell stride Maxim Levitsky
2019-04-17 19:53 ` Maxim Levitsky
2019-04-17 19:53 ` [Qemu-devel] [PATCH v2 3/5] block/nvme: support larger that 512 bytes sector devices Maxim Levitsky
2019-04-17 19:53 ` Maxim Levitsky
2019-04-17 19:53 ` [Qemu-devel] [PATCH v2 4/5] block/nvme: add support for write zeros Maxim Levitsky
2019-04-17 19:53 ` Maxim Levitsky
2019-06-06 2:56 ` Fam Zheng
2019-04-17 19:53 ` [Qemu-devel] [PATCH v2 5/5] block/nvme: add support for discard Maxim Levitsky
2019-04-17 19:53 ` Maxim Levitsky
2019-06-06 3:19 ` Fam Zheng
2019-06-06 7:31 ` Maxim Levitsky [this message]
2019-06-03 12:26 ` [Qemu-devel] [PATCH v2 0/5] Few fixes for userspace NVME driver Maxim Levitsky
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=d2ed59512f533566cbe1750fc7e52c9c5396e49c.camel@redhat.com \
--to=mlevitsk@redhat.com \
--cc=fam@euphon.net \
--cc=kwolf@redhat.com \
--cc=mreitz@redhat.com \
--cc=qemu-block@nongnu.org \
--cc=qemu-devel@nongnu.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).