linux-scsi.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Thomas Hellstrom <thellstrom@vmware.com>
To: Josh Boyer <jwboyer@fedoraproject.org>,
	Arvind Kumar <arvindkumar@vmware.com>,
	James Bottomley <JBottomley@parallels.com>
Cc: VMware PV-Drivers <pv-drivers@vmware.com>,
	Linus Torvalds <torvalds@linux-foundation.org>,
	"linux-scsi@vger.kernel.org" <linux-scsi@vger.kernel.org>
Subject: Re: [Pv-drivers] [PATCH] VMW_PVSCSI: Fix the issue of DMA-API related warnings.
Date: Thu, 11 Jun 2015 20:11:26 +0200	[thread overview]
Message-ID: <5579CF4E.8040006@vmware.com> (raw)
In-Reply-To: <CA+5PVA5tghb=ZhUHwt7A6+yK1hmcFzPbwYL1BN2wkd76p=o54Q@mail.gmail.com>

On 06/11/2015 02:32 PM, Josh Boyer wrote:
> On Fri, Mar 21, 2014 at 2:08 PM, Arvind Kumar <arvindkumar@vmware.com> wrote:
>> The driver is missing calls to pci_dma_mapping_error() after
>> performing the DMA mapping, which caused DMA-API warning to
>> show up in dmesg's output. Though that happens only when
>> DMA_API_DEBUG option is enabled. This change fixes the issue
>> and makes pvscsi_map_buffers() function more robust.
>>
>> Signed-off-by: Arvind Kumar <arvindkumar@vmware.com>
>> Cc: Josh Boyer <jwboyer@fedoraproject.org>
> This patch has been sent and pinged for 3 months now.  It's gotten no
> comments at all.  Should we send it to Linus so it actually gets
> picked up?
>
> josh

Reviewed-by: Thomas Hellstrom <thellstrom@vmware.com>

>
>> ---
>>  drivers/scsi/vmw_pvscsi.c |   45 +++++++++++++++++++++++++++++++++++++++------
>>  drivers/scsi/vmw_pvscsi.h |    2 +-
>>  2 files changed, 40 insertions(+), 7 deletions(-)
>>
>> diff --git a/drivers/scsi/vmw_pvscsi.c b/drivers/scsi/vmw_pvscsi.c
>> index c88e146..9478a00 100644
>> --- a/drivers/scsi/vmw_pvscsi.c
>> +++ b/drivers/scsi/vmw_pvscsi.c
>> @@ -349,9 +349,9 @@ static void pvscsi_create_sg(struct pvscsi_ctx *ctx,
>>   * Map all data buffers for a command into PCI space and
>>   * setup the scatter/gather list if needed.
>>   */
>> -static void pvscsi_map_buffers(struct pvscsi_adapter *adapter,
>> -                              struct pvscsi_ctx *ctx, struct scsi_cmnd *cmd,
>> -                              struct PVSCSIRingReqDesc *e)
>> +static int pvscsi_map_buffers(struct pvscsi_adapter *adapter,
>> +                             struct pvscsi_ctx *ctx, struct scsi_cmnd *cmd,
>> +                             struct PVSCSIRingReqDesc *e)
>>  {
>>         unsigned count;
>>         unsigned bufflen = scsi_bufflen(cmd);
>> @@ -360,18 +360,30 @@ static void pvscsi_map_buffers(struct pvscsi_adapter *adapter,
>>         e->dataLen = bufflen;
>>         e->dataAddr = 0;
>>         if (bufflen == 0)
>> -               return;
>> +               return 0;
>>
>>         sg = scsi_sglist(cmd);
>>         count = scsi_sg_count(cmd);
>>         if (count != 0) {
>>                 int segs = scsi_dma_map(cmd);
>> -               if (segs > 1) {
>> +
>> +               if (segs == -ENOMEM) {
>> +                       scmd_printk(KERN_ERR, cmd,
>> +                                   "vmw_pvscsi: Failed to map cmd sglist for DMA.\n");
>> +                       return -1;
>> +               } else if (segs > 1) {
>>                         pvscsi_create_sg(ctx, sg, segs);
>>
>>                         e->flags |= PVSCSI_FLAG_CMD_WITH_SG_LIST;
>>                         ctx->sglPA = pci_map_single(adapter->dev, ctx->sgl,
>>                                                     SGL_SIZE, PCI_DMA_TODEVICE);
>> +                       if (pci_dma_mapping_error(adapter->dev, ctx->sglPA)) {
>> +                               scmd_printk(KERN_ERR, cmd,
>> +                                           "vmw_pvscsi: Failed to map ctx sglist for DMA.\n");
>> +                               scsi_dma_unmap(cmd);
>> +                               ctx->sglPA = 0;
>> +                               return -1;
>> +                       }
>>                         e->dataAddr = ctx->sglPA;
>>                 } else
>>                         e->dataAddr = sg_dma_address(sg);
>> @@ -382,8 +394,15 @@ static void pvscsi_map_buffers(struct pvscsi_adapter *adapter,
>>                  */
>>                 ctx->dataPA = pci_map_single(adapter->dev, sg, bufflen,
>>                                              cmd->sc_data_direction);
>> +               if (pci_dma_mapping_error(adapter->dev, ctx->dataPA)) {
>> +                       scmd_printk(KERN_ERR, cmd,
>> +                                   "vmw_pvscsi: Failed to map direct data buffer for DMA.\n");
>> +                       return -1;
>> +               }
>>                 e->dataAddr = ctx->dataPA;
>>         }
>> +
>> +       return 0;
>>  }
>>
>>  static void pvscsi_unmap_buffers(const struct pvscsi_adapter *adapter,
>> @@ -712,6 +731,12 @@ static int pvscsi_queue_ring(struct pvscsi_adapter *adapter,
>>                 ctx->sensePA = pci_map_single(adapter->dev, cmd->sense_buffer,
>>                                               SCSI_SENSE_BUFFERSIZE,
>>                                               PCI_DMA_FROMDEVICE);
>> +               if (pci_dma_mapping_error(adapter->dev, ctx->sensePA)) {
>> +                       scmd_printk(KERN_ERR, cmd,
>> +                                   "vmw_pvscsi: Failed to map sense buffer for DMA.\n");
>> +                       ctx->sensePA = 0;
>> +                       return -1;
>> +               }
>>                 e->senseAddr = ctx->sensePA;
>>                 e->senseLen = SCSI_SENSE_BUFFERSIZE;
>>         } else {
>> @@ -737,7 +762,15 @@ static int pvscsi_queue_ring(struct pvscsi_adapter *adapter,
>>         else
>>                 e->flags = 0;
>>
>> -       pvscsi_map_buffers(adapter, ctx, cmd, e);
>> +       if (pvscsi_map_buffers(adapter, ctx, cmd, e) != 0) {
>> +               if (cmd->sense_buffer) {
>> +                       pci_unmap_single(adapter->dev, ctx->sensePA,
>> +                                        SCSI_SENSE_BUFFERSIZE,
>> +                                        PCI_DMA_FROMDEVICE);
>> +                       ctx->sensePA = 0;
>> +               }
>> +               return -1;
>> +       }
>>
>>         e->context = pvscsi_map_context(adapter, ctx);
>>
>> diff --git a/drivers/scsi/vmw_pvscsi.h b/drivers/scsi/vmw_pvscsi.h
>> index ce45888..fe3b759 100644
>> --- a/drivers/scsi/vmw_pvscsi.h
>> +++ b/drivers/scsi/vmw_pvscsi.h
>> @@ -26,7 +26,7 @@
>>
>>  #include <linux/types.h>
>>
>> -#define PVSCSI_DRIVER_VERSION_STRING   "1.0.5.0-k"
>> +#define PVSCSI_DRIVER_VERSION_STRING   "1.0.6.0-k"
>>
>>  #define PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT 128
>>
>> --
>> 1.7.9.5
>>
>>
>>
> _______________________________________________
> Pv-drivers mailing list
> Pv-drivers@vmware.com
> http://mailman2.vmware.com/mailman/listinfo/pv-drivers


  parent reply	other threads:[~2015-06-11 18:11 UTC|newest]

Thread overview: 7+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2014-03-21 18:08 [PATCH] VMW_PVSCSI: Fix the issue of DMA-API related warnings Arvind Kumar
2014-04-14 18:17 ` [Pv-drivers] " Arvind Kumar
2014-05-15 17:59   ` Arvind Kumar
2015-06-11 12:32 ` Josh Boyer
2015-06-11 13:23   ` James Bottomley
2015-06-11 18:11   ` Thomas Hellstrom [this message]
2015-07-07 21:00     ` [Pv-drivers] " Arvind Kumar

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=5579CF4E.8040006@vmware.com \
    --to=thellstrom@vmware.com \
    --cc=JBottomley@parallels.com \
    --cc=arvindkumar@vmware.com \
    --cc=jwboyer@fedoraproject.org \
    --cc=linux-scsi@vger.kernel.org \
    --cc=pv-drivers@vmware.com \
    --cc=torvalds@linux-foundation.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).