* [PATCH] swiotlb: Max mapping size takes min align mask into account @ 2022-05-10 14:21 Tianyu Lan 2022-05-10 16:33 ` Robin Murphy 2022-05-17 9:22 ` Christoph Hellwig 0 siblings, 2 replies; 5+ messages in thread From: Tianyu Lan @ 2022-05-10 14:21 UTC (permalink / raw) To: kys, haiyangz, sthemmin, wei.liu, decui, jejb, martin.petersen, hch, m.szyprowski, robin.murphy, michael.h.kelley Cc: Tianyu Lan, iommu, linux-hyperv, linux-kernel, linux-scsi, vkuznets, konrad.wilk, hch, parri.andrea, thomas.lendacky From: Tianyu Lan <Tianyu.Lan@microsoft.com> swiotlb_find_slots() skips slots according to io tlb aligned mask calculated from min aligned mask and original physical address offset. This affects max mapping size. The mapping size can't achieve the IO_TLB_SEGSIZE * IO_TLB_SIZE when original offset is non-zero. This will cause system boot up failure in Hyper-V Isolation VM where swiotlb force is enabled. Scsi layer use return value of dma_max_mapping_size() to set max segment size and it finally calls swiotlb_max_mapping_size(). Hyper-V storage driver sets min align mask to 4k - 1. Scsi layer may pass 256k length of request buffer with 0~4k offset and Hyper-V storage driver can't get swiotlb bounce buffer via DMA API. Swiotlb_find_slots() can't find 256k length bounce buffer with offset. Make swiotlb_max_mapping _size() take min align mask into account. Signed-off-by: Tianyu Lan <Tianyu.Lan@microsoft.com> --- kernel/dma/swiotlb.c | 13 ++++++++++++- 1 file changed, 12 insertions(+), 1 deletion(-) diff --git a/kernel/dma/swiotlb.c b/kernel/dma/swiotlb.c index 73a41cec9e38..0d6684ca7eab 100644 --- a/kernel/dma/swiotlb.c +++ b/kernel/dma/swiotlb.c @@ -743,7 +743,18 @@ dma_addr_t swiotlb_map(struct device *dev, phys_addr_t paddr, size_t size, size_t swiotlb_max_mapping_size(struct device *dev) { - return ((size_t)IO_TLB_SIZE) * IO_TLB_SEGSIZE; + int min_align_mask = dma_get_min_align_mask(dev); + int min_align = 0; + + /* + * swiotlb_find_slots() skips slots according to + * min align mask. This affects max mapping size. + * Take it into acount here. + */ + if (min_align_mask) + min_align = roundup(min_align_mask, IO_TLB_SIZE); + + return ((size_t)IO_TLB_SIZE) * IO_TLB_SEGSIZE - min_align; } bool is_swiotlb_active(struct device *dev) -- 2.25.1 ^ permalink raw reply related [flat|nested] 5+ messages in thread
* Re: [PATCH] swiotlb: Max mapping size takes min align mask into account 2022-05-10 14:21 [PATCH] swiotlb: Max mapping size takes min align mask into account Tianyu Lan @ 2022-05-10 16:33 ` Robin Murphy 2022-05-10 18:26 ` Michael Kelley (LINUX) 2022-05-17 9:22 ` Christoph Hellwig 1 sibling, 1 reply; 5+ messages in thread From: Robin Murphy @ 2022-05-10 16:33 UTC (permalink / raw) To: Tianyu Lan, kys, haiyangz, sthemmin, wei.liu, decui, jejb, martin.petersen, hch, m.szyprowski, michael.h.kelley Cc: Tianyu Lan, iommu, linux-hyperv, linux-kernel, linux-scsi, vkuznets, konrad.wilk, hch, parri.andrea, thomas.lendacky On 2022-05-10 15:21, Tianyu Lan wrote: > From: Tianyu Lan <Tianyu.Lan@microsoft.com> > > swiotlb_find_slots() skips slots according to io tlb aligned mask > calculated from min aligned mask and original physical address > offset. This affects max mapping size. The mapping size can't > achieve the IO_TLB_SEGSIZE * IO_TLB_SIZE when original offset is > non-zero. This will cause system boot up failure in Hyper-V > Isolation VM where swiotlb force is enabled. Scsi layer use return > value of dma_max_mapping_size() to set max segment size and it > finally calls swiotlb_max_mapping_size(). Hyper-V storage driver > sets min align mask to 4k - 1. Scsi layer may pass 256k length of > request buffer with 0~4k offset and Hyper-V storage driver can't > get swiotlb bounce buffer via DMA API. Swiotlb_find_slots() can't > find 256k length bounce buffer with offset. Make swiotlb_max_mapping > _size() take min align mask into account. Hmm, this seems a bit pessimistic - the offset can vary per mapping, so it feels to me like it should really be the caller's responsibility to account for it if they're already involved enough to care about both constraints. But I'm not sure how practical that would be. Robin. > Signed-off-by: Tianyu Lan <Tianyu.Lan@microsoft.com> > --- > kernel/dma/swiotlb.c | 13 ++++++++++++- > 1 file changed, 12 insertions(+), 1 deletion(-) > > diff --git a/kernel/dma/swiotlb.c b/kernel/dma/swiotlb.c > index 73a41cec9e38..0d6684ca7eab 100644 > --- a/kernel/dma/swiotlb.c > +++ b/kernel/dma/swiotlb.c > @@ -743,7 +743,18 @@ dma_addr_t swiotlb_map(struct device *dev, phys_addr_t paddr, size_t size, > > size_t swiotlb_max_mapping_size(struct device *dev) > { > - return ((size_t)IO_TLB_SIZE) * IO_TLB_SEGSIZE; > + int min_align_mask = dma_get_min_align_mask(dev); > + int min_align = 0; > + > + /* > + * swiotlb_find_slots() skips slots according to > + * min align mask. This affects max mapping size. > + * Take it into acount here. > + */ > + if (min_align_mask) > + min_align = roundup(min_align_mask, IO_TLB_SIZE); > + > + return ((size_t)IO_TLB_SIZE) * IO_TLB_SEGSIZE - min_align; > } > > bool is_swiotlb_active(struct device *dev) ^ permalink raw reply [flat|nested] 5+ messages in thread
* RE: [PATCH] swiotlb: Max mapping size takes min align mask into account 2022-05-10 16:33 ` Robin Murphy @ 2022-05-10 18:26 ` Michael Kelley (LINUX) 2022-05-11 6:02 ` hch 0 siblings, 1 reply; 5+ messages in thread From: Michael Kelley (LINUX) @ 2022-05-10 18:26 UTC (permalink / raw) To: Robin Murphy, Tianyu Lan, KY Srinivasan, Haiyang Zhang, Stephen Hemminger, wei.liu@kernel.org, Dexuan Cui, jejb@linux.ibm.com, martin.petersen@oracle.com, hch@infradead.org, m.szyprowski@samsung.com Cc: Tianyu Lan, iommu@lists.linux-foundation.org, linux-hyperv@vger.kernel.org, linux-kernel@vger.kernel.org, linux-scsi@vger.kernel.org, vkuznets, konrad.wilk@oracle.com, hch@lst.de, parri.andrea@gmail.com, thomas.lendacky@amd.com From: Robin Murphy <robin.murphy@arm.com> Sent: Tuesday, May 10, 2022 9:34 AM > > On 2022-05-10 15:21, Tianyu Lan wrote: > > From: Tianyu Lan <Tianyu.Lan@microsoft.com> > > > > swiotlb_find_slots() skips slots according to io tlb aligned mask > > calculated from min aligned mask and original physical address > > offset. This affects max mapping size. The mapping size can't > > achieve the IO_TLB_SEGSIZE * IO_TLB_SIZE when original offset is > > non-zero. This will cause system boot up failure in Hyper-V > > Isolation VM where swiotlb force is enabled. Scsi layer use return > > value of dma_max_mapping_size() to set max segment size and it > > finally calls swiotlb_max_mapping_size(). Hyper-V storage driver > > sets min align mask to 4k - 1. Scsi layer may pass 256k length of > > request buffer with 0~4k offset and Hyper-V storage driver can't > > get swiotlb bounce buffer via DMA API. Swiotlb_find_slots() can't > > find 256k length bounce buffer with offset. Make swiotlb_max_mapping > > _size() take min align mask into account. > > Hmm, this seems a bit pessimistic - the offset can vary per mapping, so > it feels to me like it should really be the caller's responsibility to > account for it if they're already involved enough to care about both > constraints. But I'm not sure how practical that would be. Tianyu and I discussed this prior to his submitting the patch. Presumably dma_max_mapping_size() exists so that the higher level blk-mq code can limit the size of I/O requests to something that will "fit" in the swiotlb when bounce buffering is enabled. Unfortunately, the current code is just giving the wrong answer when the offset is non-zero. The offset would be less than PAGE_SIZE, so the impact would be dma_max_mapping_size() returning 252 Kbytes instead of 256 Kbytes, but only for devices where dma min align mask is set. And any I/O sizes less than 252 Kbytes are unaffected even when dma min align mask is set. Net, the impact would be only in a fairly rare edge case. Even on ARM64 with a 64K page size, the Hyper-V storage driver is setting the dma min align mask to only 4K (which is correct because the Hyper-V host uses a 4K page size even if the guest is using something larger), so again the limit becomes 252 Kbytes instead of 256 Kbytes, and any impact is rare. As you mentioned, how else would a caller handle this situation? Michael > > Robin. > > > Signed-off-by: Tianyu Lan <Tianyu.Lan@microsoft.com> > > --- > > kernel/dma/swiotlb.c | 13 ++++++++++++- > > 1 file changed, 12 insertions(+), 1 deletion(-) > > > > diff --git a/kernel/dma/swiotlb.c b/kernel/dma/swiotlb.c > > index 73a41cec9e38..0d6684ca7eab 100644 > > --- a/kernel/dma/swiotlb.c > > +++ b/kernel/dma/swiotlb.c > > @@ -743,7 +743,18 @@ dma_addr_t swiotlb_map(struct device *dev, phys_addr_t > paddr, size_t size, > > > > size_t swiotlb_max_mapping_size(struct device *dev) > > { > > - return ((size_t)IO_TLB_SIZE) * IO_TLB_SEGSIZE; > > + int min_align_mask = dma_get_min_align_mask(dev); > > + int min_align = 0; > > + > > + /* > > + * swiotlb_find_slots() skips slots according to > > + * min align mask. This affects max mapping size. > > + * Take it into acount here. > > + */ > > + if (min_align_mask) > > + min_align = roundup(min_align_mask, IO_TLB_SIZE); > > + > > + return ((size_t)IO_TLB_SIZE) * IO_TLB_SEGSIZE - min_align; > > } > > > > bool is_swiotlb_active(struct device *dev) ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH] swiotlb: Max mapping size takes min align mask into account 2022-05-10 18:26 ` Michael Kelley (LINUX) @ 2022-05-11 6:02 ` hch 0 siblings, 0 replies; 5+ messages in thread From: hch @ 2022-05-11 6:02 UTC (permalink / raw) To: Michael Kelley (LINUX) Cc: Robin Murphy, Tianyu Lan, KY Srinivasan, Haiyang Zhang, Stephen Hemminger, wei.liu@kernel.org, Dexuan Cui, jejb@linux.ibm.com, martin.petersen@oracle.com, hch@infradead.org, m.szyprowski@samsung.com, Tianyu Lan, iommu@lists.linux-foundation.org, linux-hyperv@vger.kernel.org, linux-kernel@vger.kernel.org, linux-scsi@vger.kernel.org, vkuznets, konrad.wilk@oracle.com, hch@lst.de, parri.andrea@gmail.com, thomas.lendacky@amd.com On Tue, May 10, 2022 at 06:26:55PM +0000, Michael Kelley (LINUX) wrote: > > Hmm, this seems a bit pessimistic - the offset can vary per mapping, so > > it feels to me like it should really be the caller's responsibility to > > account for it if they're already involved enough to care about both > > constraints. But I'm not sure how practical that would be. > > Tianyu and I discussed this prior to his submitting the patch. > Presumably dma_max_mapping_size() exists so that the higher > level blk-mq code can limit the size of I/O requests to something > that will "fit" in the swiotlb when bounce buffering is enabled. Yes, the idea that upper level code doesn't need to care was very much the idea behind dma_max_mapping_size(). > As you mentioned, how else would a caller handle this situation? Well, we could look at dma_get_min_align_mask in the caller and do the calculation there, but I really don't think that is a good idea. So this patch looks sensible to me. ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH] swiotlb: Max mapping size takes min align mask into account 2022-05-10 14:21 [PATCH] swiotlb: Max mapping size takes min align mask into account Tianyu Lan 2022-05-10 16:33 ` Robin Murphy @ 2022-05-17 9:22 ` Christoph Hellwig 1 sibling, 0 replies; 5+ messages in thread From: Christoph Hellwig @ 2022-05-17 9:22 UTC (permalink / raw) To: Tianyu Lan Cc: kys, haiyangz, sthemmin, wei.liu, decui, jejb, martin.petersen, hch, m.szyprowski, robin.murphy, michael.h.kelley, Tianyu Lan, iommu, linux-hyperv, linux-kernel, linux-scsi, vkuznets, konrad.wilk, hch, parri.andrea, thomas.lendacky Thanks, applied to the dma-mapping for-next tree. ^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2022-05-17 9:23 UTC | newest] Thread overview: 5+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2022-05-10 14:21 [PATCH] swiotlb: Max mapping size takes min align mask into account Tianyu Lan 2022-05-10 16:33 ` Robin Murphy 2022-05-10 18:26 ` Michael Kelley (LINUX) 2022-05-11 6:02 ` hch 2022-05-17 9:22 ` Christoph Hellwig
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).