* [PATCH for-rc] RDMA: Fix ib block iterator counter overflow
@ 2023-01-02 16:03 Yonatan Nachum
2023-01-03 9:43 ` Gal Pressman
` (2 more replies)
0 siblings, 3 replies; 4+ messages in thread
From: Yonatan Nachum @ 2023-01-02 16:03 UTC (permalink / raw)
To: jgg, leon, linux-rdma; +Cc: mrgolin
When registering a new DMA MR after selecting the best aligned page size
for it, we iterate over the given sglist to split each entry to smaller,
aligned to the selected page size, DMA blocks.
In given circumstances where the sg entry and page size fit certain sizes
and the sg entry is not aligned to the selected page size, the total size
of the aligned pages we need to cover the sg entry is >= 4GB. Under this
circumstances, while iterating page aligned blocks, the counter responsible
for counting how much we advanced from the start of the sg entry is
overflowed because its type is u32 and we pass 4GB in size. This can
lead to an infinite loop inside the iterator function because in some
cases the overflow prevents the counter to be larger than the size of
the sg entry.
Fix the presented problem with changing the counter type to u64.
Backtrace:
[ 192.374329] efa_reg_user_mr_dmabuf
[ 192.376783] efa_register_mr
[ 192.382579] pgsz_bitmap 0xfffff000 rounddown 0x80000000
[ 192.386423] pg_sz [0x80000000] umem_length[0xc0000000]
[ 192.392657] start 0x0 length 0xc0000000 params.page_shift 31 params.page_num 3
[ 192.399559] hp_cnt[3], pages_in_hp[524288]
[ 192.403690] umem->sgt_append.sgt.nents[1]
[ 192.407905] number entries: [1], pg_bit: [31]
[ 192.411397] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8]
[ 192.415601] biter->__sg_advance [665837568] sg_dma_len[3221225472]
[ 192.419823] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8]
[ 192.423976] biter->__sg_advance [2813321216] sg_dma_len[3221225472]
[ 192.428243] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8]
[ 192.432397] biter->__sg_advance [665837568] sg_dma_len[3221225472]
Fixes: a808273a495c
Signed-off-by: Yonatan Nachum <ynachum@amazon.com>
---
include/rdma/ib_verbs.h | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/include/rdma/ib_verbs.h b/include/rdma/ib_verbs.h
index 975d6e9efbcb..6821c7951363 100644
--- a/include/rdma/ib_verbs.h
+++ b/include/rdma/ib_verbs.h
@@ -2836,7 +2836,7 @@ struct ib_block_iter {
struct scatterlist *__sg; /* sg holding the current aligned block */
dma_addr_t __dma_addr; /* unaligned DMA address of this block */
unsigned int __sg_nents; /* number of SG entries */
- unsigned int __sg_advance; /* number of bytes to advance in sg in next step */
+ u64 __sg_advance; /* number of bytes to advance in sg in next step */
unsigned int __pg_bit; /* alignment of current block */
};
--
2.38.1
^ permalink raw reply related [flat|nested] 4+ messages in thread
* Re: [PATCH for-rc] RDMA: Fix ib block iterator counter overflow
2023-01-02 16:03 [PATCH for-rc] RDMA: Fix ib block iterator counter overflow Yonatan Nachum
@ 2023-01-03 9:43 ` Gal Pressman
2023-01-03 9:46 ` Leon Romanovsky
2023-01-03 13:29 ` Jason Gunthorpe
2 siblings, 0 replies; 4+ messages in thread
From: Gal Pressman @ 2023-01-03 9:43 UTC (permalink / raw)
To: Yonatan Nachum, jgg, leon, linux-rdma; +Cc: mrgolin
On 02/01/2023 18:03, Yonatan Nachum wrote:
> When registering a new DMA MR after selecting the best aligned page size
> for it, we iterate over the given sglist to split each entry to smaller,
> aligned to the selected page size, DMA blocks.
>
> In given circumstances where the sg entry and page size fit certain sizes
> and the sg entry is not aligned to the selected page size, the total size
> of the aligned pages we need to cover the sg entry is >= 4GB. Under this
> circumstances, while iterating page aligned blocks, the counter responsible
> for counting how much we advanced from the start of the sg entry is
> overflowed because its type is u32 and we pass 4GB in size. This can
> lead to an infinite loop inside the iterator function because in some
> cases the overflow prevents the counter to be larger than the size of
> the sg entry.
>
> Fix the presented problem with changing the counter type to u64.
>
> Backtrace:
> [ 192.374329] efa_reg_user_mr_dmabuf
> [ 192.376783] efa_register_mr
> [ 192.382579] pgsz_bitmap 0xfffff000 rounddown 0x80000000
> [ 192.386423] pg_sz [0x80000000] umem_length[0xc0000000]
> [ 192.392657] start 0x0 length 0xc0000000 params.page_shift 31 params.page_num 3
> [ 192.399559] hp_cnt[3], pages_in_hp[524288]
> [ 192.403690] umem->sgt_append.sgt.nents[1]
> [ 192.407905] number entries: [1], pg_bit: [31]
> [ 192.411397] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8]
> [ 192.415601] biter->__sg_advance [665837568] sg_dma_len[3221225472]
> [ 192.419823] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8]
> [ 192.423976] biter->__sg_advance [2813321216] sg_dma_len[3221225472]
> [ 192.428243] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8]
> [ 192.432397] biter->__sg_advance [665837568] sg_dma_len[3221225472]
>
> Fixes: a808273a495c
Missing the patch subject line, please see:
https://www.kernel.org/doc/html/latest/process/submitting-patches.html#describe-your-changes
Also, there shouldn't be a blank line here.
>
> Signed-off-by: Yonatan Nachum <ynachum@amazon.com>
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [PATCH for-rc] RDMA: Fix ib block iterator counter overflow
2023-01-02 16:03 [PATCH for-rc] RDMA: Fix ib block iterator counter overflow Yonatan Nachum
2023-01-03 9:43 ` Gal Pressman
@ 2023-01-03 9:46 ` Leon Romanovsky
2023-01-03 13:29 ` Jason Gunthorpe
2 siblings, 0 replies; 4+ messages in thread
From: Leon Romanovsky @ 2023-01-03 9:46 UTC (permalink / raw)
To: Yonatan Nachum; +Cc: jgg, linux-rdma, mrgolin
On Mon, Jan 02, 2023 at 06:03:17PM +0200, Yonatan Nachum wrote:
> When registering a new DMA MR after selecting the best aligned page size
> for it, we iterate over the given sglist to split each entry to smaller,
> aligned to the selected page size, DMA blocks.
>
> In given circumstances where the sg entry and page size fit certain sizes
> and the sg entry is not aligned to the selected page size, the total size
> of the aligned pages we need to cover the sg entry is >= 4GB. Under this
> circumstances, while iterating page aligned blocks, the counter responsible
> for counting how much we advanced from the start of the sg entry is
> overflowed because its type is u32 and we pass 4GB in size. This can
> lead to an infinite loop inside the iterator function because in some
> cases the overflow prevents the counter to be larger than the size of
> the sg entry.
>
> Fix the presented problem with changing the counter type to u64.
>
> Backtrace:
> [ 192.374329] efa_reg_user_mr_dmabuf
> [ 192.376783] efa_register_mr
> [ 192.382579] pgsz_bitmap 0xfffff000 rounddown 0x80000000
> [ 192.386423] pg_sz [0x80000000] umem_length[0xc0000000]
> [ 192.392657] start 0x0 length 0xc0000000 params.page_shift 31 params.page_num 3
> [ 192.399559] hp_cnt[3], pages_in_hp[524288]
> [ 192.403690] umem->sgt_append.sgt.nents[1]
> [ 192.407905] number entries: [1], pg_bit: [31]
> [ 192.411397] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8]
> [ 192.415601] biter->__sg_advance [665837568] sg_dma_len[3221225472]
> [ 192.419823] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8]
> [ 192.423976] biter->__sg_advance [2813321216] sg_dma_len[3221225472]
> [ 192.428243] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8]
> [ 192.432397] biter->__sg_advance [665837568] sg_dma_len[3221225472]
>
> Fixes: a808273a495c
The fixes line is truncated. Please fix and resend.
Thanks
>
> Signed-off-by: Yonatan Nachum <ynachum@amazon.com>
> ---
> include/rdma/ib_verbs.h | 2 +-
> 1 file changed, 1 insertion(+), 1 deletion(-)
>
> diff --git a/include/rdma/ib_verbs.h b/include/rdma/ib_verbs.h
> index 975d6e9efbcb..6821c7951363 100644
> --- a/include/rdma/ib_verbs.h
> +++ b/include/rdma/ib_verbs.h
> @@ -2836,7 +2836,7 @@ struct ib_block_iter {
> struct scatterlist *__sg; /* sg holding the current aligned block */
> dma_addr_t __dma_addr; /* unaligned DMA address of this block */
> unsigned int __sg_nents; /* number of SG entries */
> - unsigned int __sg_advance; /* number of bytes to advance in sg in next step */
> + u64 __sg_advance; /* number of bytes to advance in sg in next step */
> unsigned int __pg_bit; /* alignment of current block */
> };
>
> --
> 2.38.1
>
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [PATCH for-rc] RDMA: Fix ib block iterator counter overflow
2023-01-02 16:03 [PATCH for-rc] RDMA: Fix ib block iterator counter overflow Yonatan Nachum
2023-01-03 9:43 ` Gal Pressman
2023-01-03 9:46 ` Leon Romanovsky
@ 2023-01-03 13:29 ` Jason Gunthorpe
2 siblings, 0 replies; 4+ messages in thread
From: Jason Gunthorpe @ 2023-01-03 13:29 UTC (permalink / raw)
To: Yonatan Nachum; +Cc: leon, linux-rdma, mrgolin
On Mon, Jan 02, 2023 at 06:03:17PM +0200, Yonatan Nachum wrote:
> When registering a new DMA MR after selecting the best aligned page size
> for it, we iterate over the given sglist to split each entry to smaller,
> aligned to the selected page size, DMA blocks.
>
> In given circumstances where the sg entry and page size fit certain sizes
> and the sg entry is not aligned to the selected page size, the total size
> of the aligned pages we need to cover the sg entry is >= 4GB.
Huh? The max length of a scatterlist element is unsigned int:
struct scatterlist {
unsigned long page_link;
unsigned int offset;
unsigned int length;
The problem is the construction of the end of iteration test:
block_offset = biter->__dma_addr & (BIT_ULL(biter->__pg_bit) - 1);
biter->__sg_advance += BIT_ULL(biter->__pg_bit) - block_offset;
if (biter->__sg_advance >= sg_dma_len(biter->__sg)) {
This should be reworked so it doesn't increment until it knows it is
OK and leave the types alone.
Jason
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2023-01-03 13:29 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2023-01-02 16:03 [PATCH for-rc] RDMA: Fix ib block iterator counter overflow Yonatan Nachum
2023-01-03 9:43 ` Gal Pressman
2023-01-03 9:46 ` Leon Romanovsky
2023-01-03 13:29 ` Jason Gunthorpe
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).