* [PATCH rdma-next 0/4] Small set of mlx5 related fixes
@ 2019-03-19 9:24 Leon Romanovsky
2019-03-19 9:24 ` [PATCH rdma-next 1/4] IB/mlx5: Reset access mask when looping inside page fault handler Leon Romanovsky
` (4 more replies)
0 siblings, 5 replies; 11+ messages in thread
From: Leon Romanovsky @ 2019-03-19 9:24 UTC (permalink / raw)
To: Doug Ledford, Jason Gunthorpe
Cc: Leon Romanovsky, RDMA mailing list, Artemy Kovalyov, Guy Levi,
Haggai Eran, Jerome Glisse, Moni Shoua, Saeed Mahameed,
linux-netdev
From: Leon Romanovsky <leonro@mellanox.com>
Hi,
Despite one change in mlx5_core, all those changes are for mlx5_ib. The
change in mlx5_core is related to MR cache initialization that is not
used in mlx5_core at all.
Thanks
Artemy Kovalyov (3):
IB/mlx5: WQE dump jumps over first 16 bytes
net/mlx5: Decrease default mr cache size
IB/mlx5: Compare only index part of a memory window rkey
Moni Shoua (1):
IB/mlx5: Reset access mask when looping inside page fault handler
drivers/infiniband/hw/mlx5/odp.c | 14 +++++++++++--
drivers/infiniband/hw/mlx5/qp.c | 5 ++---
.../net/ethernet/mellanox/mlx5/core/main.c | 20 -------------------
3 files changed, 14 insertions(+), 25 deletions(-)
--
2.20.1
^ permalink raw reply [flat|nested] 11+ messages in thread
* [PATCH rdma-next 1/4] IB/mlx5: Reset access mask when looping inside page fault handler
2019-03-19 9:24 [PATCH rdma-next 0/4] Small set of mlx5 related fixes Leon Romanovsky
@ 2019-03-19 9:24 ` Leon Romanovsky
2019-03-19 9:24 ` [PATCH rdma-next 2/4] IB/mlx5: WQE dump jumps over first 16 bytes Leon Romanovsky
` (3 subsequent siblings)
4 siblings, 0 replies; 11+ messages in thread
From: Leon Romanovsky @ 2019-03-19 9:24 UTC (permalink / raw)
To: Doug Ledford, Jason Gunthorpe
Cc: Leon Romanovsky, RDMA mailing list, Artemy Kovalyov, Guy Levi,
Haggai Eran, Jerome Glisse, Moni Shoua, Saeed Mahameed,
linux-netdev
From: Moni Shoua <monis@mellanox.com>
If page-fault handler spans multiple MRs then the access mask needs to
be reset before each MR handling or otherwise write access will be
granted to mapped pages instead of read-only.
Cc: <stable@vger.kernel.org> # 3.19
Fixes: 7bdf65d411c1 ("IB/mlx5: Handle page faults")
Reported-by: Jerome Glisse <jglisse@redhat.com>
Signed-off-by: Moni Shoua <monis@mellanox.com>
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
---
drivers/infiniband/hw/mlx5/odp.c | 3 ++-
1 file changed, 2 insertions(+), 1 deletion(-)
diff --git a/drivers/infiniband/hw/mlx5/odp.c b/drivers/infiniband/hw/mlx5/odp.c
index ec8381d1e0f2..95dccac1f844 100644
--- a/drivers/infiniband/hw/mlx5/odp.c
+++ b/drivers/infiniband/hw/mlx5/odp.c
@@ -585,7 +585,7 @@ static int pagefault_mr(struct mlx5_ib_dev *dev, struct mlx5_ib_mr *mr,
struct ib_umem_odp *odp_mr = to_ib_umem_odp(mr->umem);
bool downgrade = flags & MLX5_PF_FLAGS_DOWNGRADE;
bool prefetch = flags & MLX5_PF_FLAGS_PREFETCH;
- u64 access_mask = ODP_READ_ALLOWED_BIT;
+ u64 access_mask;
u64 start_idx, page_mask;
struct ib_umem_odp *odp;
size_t size;
@@ -607,6 +607,7 @@ static int pagefault_mr(struct mlx5_ib_dev *dev, struct mlx5_ib_mr *mr,
page_shift = mr->umem->page_shift;
page_mask = ~(BIT(page_shift) - 1);
start_idx = (io_virt - (mr->mmkey.iova & page_mask)) >> page_shift;
+ access_mask = ODP_READ_ALLOWED_BIT;
if (prefetch && !downgrade && !mr->umem->writable) {
/* prefetch with write-access must
--
2.20.1
^ permalink raw reply related [flat|nested] 11+ messages in thread
* [PATCH rdma-next 2/4] IB/mlx5: WQE dump jumps over first 16 bytes
2019-03-19 9:24 [PATCH rdma-next 0/4] Small set of mlx5 related fixes Leon Romanovsky
2019-03-19 9:24 ` [PATCH rdma-next 1/4] IB/mlx5: Reset access mask when looping inside page fault handler Leon Romanovsky
@ 2019-03-19 9:24 ` Leon Romanovsky
2019-03-19 9:24 ` [PATCH mlx5-next 3/4] net/mlx5: Decrease default mr cache size Leon Romanovsky
` (2 subsequent siblings)
4 siblings, 0 replies; 11+ messages in thread
From: Leon Romanovsky @ 2019-03-19 9:24 UTC (permalink / raw)
To: Doug Ledford, Jason Gunthorpe
Cc: Leon Romanovsky, RDMA mailing list, Artemy Kovalyov, Guy Levi,
Haggai Eran, Jerome Glisse, Moni Shoua, Saeed Mahameed,
linux-netdev
From: Artemy Kovalyov <artemyko@mellanox.com>
Move index increment after its is used or otherwise it will start the
dump of the WQE from second WQE BB.
Fixes: 34f4c9554d8b ("IB/mlx5: Use fragmented QP's buffer for in-kernel users")
Signed-off-by: Artemy Kovalyov <artemyko@mellanox.com>
Signed-off-by: Moni Shoua <monis@mellanox.com>
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
---
drivers/infiniband/hw/mlx5/qp.c | 5 ++---
1 file changed, 2 insertions(+), 3 deletions(-)
diff --git a/drivers/infiniband/hw/mlx5/qp.c b/drivers/infiniband/hw/mlx5/qp.c
index 6b1f0e76900b..2014fd0fddc7 100644
--- a/drivers/infiniband/hw/mlx5/qp.c
+++ b/drivers/infiniband/hw/mlx5/qp.c
@@ -4724,16 +4724,15 @@ static void set_linv_wr(struct mlx5_ib_qp *qp, void **seg, int *size,
static void dump_wqe(struct mlx5_ib_qp *qp, u32 idx, int size_16)
{
__be32 *p = NULL;
- u32 tidx = idx;
int i, j;
pr_debug("dump WQE index %u:\n", idx);
for (i = 0, j = 0; i < size_16 * 4; i += 4, j += 4) {
if ((i & 0xf) == 0) {
- tidx = (tidx + 1) & (qp->sq.wqe_cnt - 1);
- p = mlx5_frag_buf_get_wqe(&qp->sq.fbc, tidx);
+ p = mlx5_frag_buf_get_wqe(&qp->sq.fbc, idx);
pr_debug("WQBB at %p:\n", (void *)p);
j = 0;
+ idx = (idx + 1) & (qp->sq.wqe_cnt - 1);
}
pr_debug("%08x %08x %08x %08x\n", be32_to_cpu(p[j]),
be32_to_cpu(p[j + 1]), be32_to_cpu(p[j + 2]),
--
2.20.1
^ permalink raw reply related [flat|nested] 11+ messages in thread
* [PATCH mlx5-next 3/4] net/mlx5: Decrease default mr cache size
2019-03-19 9:24 [PATCH rdma-next 0/4] Small set of mlx5 related fixes Leon Romanovsky
2019-03-19 9:24 ` [PATCH rdma-next 1/4] IB/mlx5: Reset access mask when looping inside page fault handler Leon Romanovsky
2019-03-19 9:24 ` [PATCH rdma-next 2/4] IB/mlx5: WQE dump jumps over first 16 bytes Leon Romanovsky
@ 2019-03-19 9:24 ` Leon Romanovsky
2019-03-27 10:07 ` Or Gerlitz
2019-03-19 9:24 ` [PATCH rdma-next 4/4] IB/mlx5: Compare only index part of a memory window rkey Leon Romanovsky
2019-03-27 18:29 ` [PATCH rdma-next 0/4] Small set of mlx5 related fixes Jason Gunthorpe
4 siblings, 1 reply; 11+ messages in thread
From: Leon Romanovsky @ 2019-03-19 9:24 UTC (permalink / raw)
To: Doug Ledford, Jason Gunthorpe
Cc: Leon Romanovsky, RDMA mailing list, Artemy Kovalyov, Guy Levi,
Haggai Eran, Jerome Glisse, Moni Shoua, Saeed Mahameed,
linux-netdev
From: Artemy Kovalyov <artemyko@mellanox.com>
Delete initialization of high order entries in mr cache to decrease initial
memory footprint. When required, the administrator can populate the
entries with memory keys via the /sys interface.
Signed-off-by: Artemy Kovalyov <artemyko@mellanox.com>
Signed-off-by: Moni Shoua <monis@mellanox.com>
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
---
.../net/ethernet/mellanox/mlx5/core/main.c | 20 -------------------
1 file changed, 20 deletions(-)
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/main.c b/drivers/net/ethernet/mellanox/mlx5/core/main.c
index 70cc906a102b..76716419370d 100644
--- a/drivers/net/ethernet/mellanox/mlx5/core/main.c
+++ b/drivers/net/ethernet/mellanox/mlx5/core/main.c
@@ -164,26 +164,6 @@ static struct mlx5_profile profile[] = {
.size = 8,
.limit = 4
},
- .mr_cache[16] = {
- .size = 8,
- .limit = 4
- },
- .mr_cache[17] = {
- .size = 8,
- .limit = 4
- },
- .mr_cache[18] = {
- .size = 8,
- .limit = 4
- },
- .mr_cache[19] = {
- .size = 4,
- .limit = 2
- },
- .mr_cache[20] = {
- .size = 4,
- .limit = 2
- },
},
};
--
2.20.1
^ permalink raw reply related [flat|nested] 11+ messages in thread
* [PATCH rdma-next 4/4] IB/mlx5: Compare only index part of a memory window rkey
2019-03-19 9:24 [PATCH rdma-next 0/4] Small set of mlx5 related fixes Leon Romanovsky
` (2 preceding siblings ...)
2019-03-19 9:24 ` [PATCH mlx5-next 3/4] net/mlx5: Decrease default mr cache size Leon Romanovsky
@ 2019-03-19 9:24 ` Leon Romanovsky
2019-03-27 18:29 ` [PATCH rdma-next 0/4] Small set of mlx5 related fixes Jason Gunthorpe
4 siblings, 0 replies; 11+ messages in thread
From: Leon Romanovsky @ 2019-03-19 9:24 UTC (permalink / raw)
To: Doug Ledford, Jason Gunthorpe
Cc: Leon Romanovsky, RDMA mailing list, Artemy Kovalyov, Guy Levi,
Haggai Eran, Jerome Glisse, Moni Shoua, Saeed Mahameed,
linux-netdev
From: Artemy Kovalyov <artemyko@mellanox.com>
The InfiniBand Architecture Specification section 10.6.7.2.4 TYPE
2 MEMORY WINDOWS says that if the CI supports the Base Memory Management
Extensions defined in this specification, the R_Key format for a Type
2 Memory Window must consist of:
* 24 bit index in the most significant bits of the R_Key, which is
owned by the CI, and
* 8 bit key in the least significant bits of the R_Key, which is
owned by the Consumer.
This means that the kernel should compare only the index part of a R_Key
to determine equality with another R_Key.
Fixes: db570d7deafb ("IB/mlx5: Add ODP support to MW")
Signed-off-by: Artemy Kovalyov <artemyko@mellanox.com>
Signed-off-by: Moni Shoua <monis@mellanox.com>
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
---
drivers/infiniband/hw/mlx5/odp.c | 11 ++++++++++-
1 file changed, 10 insertions(+), 1 deletion(-)
diff --git a/drivers/infiniband/hw/mlx5/odp.c b/drivers/infiniband/hw/mlx5/odp.c
index 95dccac1f844..2b197da1cc37 100644
--- a/drivers/infiniband/hw/mlx5/odp.c
+++ b/drivers/infiniband/hw/mlx5/odp.c
@@ -711,6 +711,15 @@ struct pf_frame {
int depth;
};
+static bool mkey_is_eq(struct mlx5_core_mkey *mmkey, u32 key)
+{
+ if (!mmkey)
+ return false;
+ if (mmkey->type == MLX5_MKEY_MW)
+ return mlx5_base_mkey(mmkey->key) == mlx5_base_mkey(key);
+ return mmkey->key == key;
+}
+
static int get_indirect_num_descs(struct mlx5_core_mkey *mmkey)
{
struct mlx5_ib_mw *mw;
@@ -760,7 +769,7 @@ static int pagefault_single_data_segment(struct mlx5_ib_dev *dev,
next_mr:
mmkey = __mlx5_mr_lookup(dev->mdev, mlx5_base_mkey(key));
- if (!mmkey || mmkey->key != key) {
+ if (!mkey_is_eq(mmkey, key)) {
mlx5_ib_dbg(dev, "failed to find mkey %x\n", key);
ret = -EFAULT;
goto srcu_unlock;
--
2.20.1
^ permalink raw reply related [flat|nested] 11+ messages in thread
* Re: [PATCH mlx5-next 3/4] net/mlx5: Decrease default mr cache size
2019-03-19 9:24 ` [PATCH mlx5-next 3/4] net/mlx5: Decrease default mr cache size Leon Romanovsky
@ 2019-03-27 10:07 ` Or Gerlitz
2019-03-27 11:41 ` Leon Romanovsky
0 siblings, 1 reply; 11+ messages in thread
From: Or Gerlitz @ 2019-03-27 10:07 UTC (permalink / raw)
To: Leon Romanovsky
Cc: Doug Ledford, Jason Gunthorpe, Leon Romanovsky, RDMA mailing list,
Artemy Kovalyov, Guy Levi, Haggai Eran, Jerome Glisse, Moni Shoua,
Saeed Mahameed, linux-netdev
On Tue, Mar 19, 2019 at 11:25 AM Leon Romanovsky <leon@kernel.org> wrote:
> From: Artemy Kovalyov <artemyko@mellanox.com>
>
> Delete initialization of high order entries in mr cache to decrease initial
> memory footprint. When required, the administrator can populate the
> entries with memory keys via the /sys interface.
Please add here:
This approach is very helpful to reduce the per HW function memory
footprint in environments such as VMs. Before the patch we see
consumption of 0.9GB per function and after the patch about 0.1GB
> Signed-off-by: Artemy Kovalyov <artemyko@mellanox.com>
> Signed-off-by: Moni Shoua <monis@mellanox.com>
> Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
Reported-by: Shalom Toledo <shalomt@mellanox.com>
Acked-by: Or Gerlitz <ogerlitz@mellanox.com>
Lets push it into stable kernels, a Fixes tag here will cause that
to happen more easily, so please consider that.
> ---
> .../net/ethernet/mellanox/mlx5/core/main.c | 20 -------------------
> 1 file changed, 20 deletions(-)
>
> diff --git a/drivers/net/ethernet/mellanox/mlx5/core/main.c b/drivers/net/ethernet/mellanox/mlx5/core/main.c
> index 70cc906a102b..76716419370d 100644
> --- a/drivers/net/ethernet/mellanox/mlx5/core/main.c
> +++ b/drivers/net/ethernet/mellanox/mlx5/core/main.c
> @@ -164,26 +164,6 @@ static struct mlx5_profile profile[] = {
> .size = 8,
> .limit = 4
> },
> - .mr_cache[16] = {
> - .size = 8,
> - .limit = 4
> - },
> - .mr_cache[17] = {
> - .size = 8,
> - .limit = 4
> - },
> - .mr_cache[18] = {
> - .size = 8,
> - .limit = 4
> - },
> - .mr_cache[19] = {
> - .size = 4,
> - .limit = 2
> - },
> - .mr_cache[20] = {
> - .size = 4,
> - .limit = 2
> - },
> },
> };
>
> --
> 2.20.1
>
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PATCH mlx5-next 3/4] net/mlx5: Decrease default mr cache size
2019-03-27 10:07 ` Or Gerlitz
@ 2019-03-27 11:41 ` Leon Romanovsky
2019-03-27 11:58 ` Or Gerlitz
0 siblings, 1 reply; 11+ messages in thread
From: Leon Romanovsky @ 2019-03-27 11:41 UTC (permalink / raw)
To: Or Gerlitz
Cc: Doug Ledford, Jason Gunthorpe, RDMA mailing list, Artemy Kovalyov,
Guy Levi, Haggai Eran, Jerome Glisse, Moni Shoua, Saeed Mahameed,
linux-netdev
[-- Attachment #1: Type: text/plain, Size: 1056 bytes --]
On Wed, Mar 27, 2019 at 12:07:54PM +0200, Or Gerlitz wrote:
> On Tue, Mar 19, 2019 at 11:25 AM Leon Romanovsky <leon@kernel.org> wrote:
> > From: Artemy Kovalyov <artemyko@mellanox.com>
> >
> > Delete initialization of high order entries in mr cache to decrease initial
> > memory footprint. When required, the administrator can populate the
> > entries with memory keys via the /sys interface.
>
> Please add here:
>
> This approach is very helpful to reduce the per HW function memory
> footprint in environments such as VMs. Before the patch we see
> consumption of 0.9GB per function and after the patch about 0.1GB
>
> > Signed-off-by: Artemy Kovalyov <artemyko@mellanox.com>
> > Signed-off-by: Moni Shoua <monis@mellanox.com>
> > Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
>
> Reported-by: Shalom Toledo <shalomt@mellanox.com>
> Acked-by: Or Gerlitz <ogerlitz@mellanox.com>
>
> Lets push it into stable kernels, a Fixes tag here will cause that
> to happen more easily, so please consider that.
Thanks,
I'll add it at the "apply" stage.
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 801 bytes --]
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PATCH mlx5-next 3/4] net/mlx5: Decrease default mr cache size
2019-03-27 11:41 ` Leon Romanovsky
@ 2019-03-27 11:58 ` Or Gerlitz
2019-03-27 13:36 ` Leon Romanovsky
0 siblings, 1 reply; 11+ messages in thread
From: Or Gerlitz @ 2019-03-27 11:58 UTC (permalink / raw)
To: Leon Romanovsky
Cc: Doug Ledford, Jason Gunthorpe, RDMA mailing list, Artemy Kovalyov,
Guy Levi, Haggai Eran, Jerome Glisse, Moni Shoua, Saeed Mahameed,
linux-netdev
On Wed, Mar 27, 2019 at 1:41 PM Leon Romanovsky <leon@kernel.org> wrote:
> On Wed, Mar 27, 2019 at 12:07:54PM +0200, Or Gerlitz wrote:
> > On Tue, Mar 19, 2019 at 11:25 AM Leon Romanovsky <leon@kernel.org> wrote:
> > > From: Artemy Kovalyov <artemyko@mellanox.com>
> > >
> > > Delete initialization of high order entries in mr cache to decrease initial
> > > memory footprint. When required, the administrator can populate the
> > > entries with memory keys via the /sys interface.
> >
> > Please add here:
> >
> > This approach is very helpful to reduce the per HW function memory
> > footprint in environments such as VMs. Before the patch we see
> > consumption of 0.9GB per function and after the patch about 0.1GB
> >
> > > Signed-off-by: Artemy Kovalyov <artemyko@mellanox.com>
> > > Signed-off-by: Moni Shoua <monis@mellanox.com>
> > > Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
> >
> > Reported-by: Shalom Toledo <shalomt@mellanox.com>
> > Acked-by: Or Gerlitz <ogerlitz@mellanox.com>
> >
> > Lets push it into stable kernels, a Fixes tag here will cause that
> > to happen more easily, so please consider that.
> I'll add it at the "apply" stage.
Just to make sure, by "add it" you mean the signatures, the text and
the Fixes that?
Also, is there any reason not to eliminate the mr cache pre-population
all together?
Currently we consume 100-200MB per function after the patch which is
also problematic for some
environments.
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PATCH mlx5-next 3/4] net/mlx5: Decrease default mr cache size
2019-03-27 11:58 ` Or Gerlitz
@ 2019-03-27 13:36 ` Leon Romanovsky
2019-03-27 14:24 ` Or Gerlitz
0 siblings, 1 reply; 11+ messages in thread
From: Leon Romanovsky @ 2019-03-27 13:36 UTC (permalink / raw)
To: Or Gerlitz
Cc: Doug Ledford, Jason Gunthorpe, RDMA mailing list, Artemy Kovalyov,
Guy Levi, Haggai Eran, Jerome Glisse, Moni Shoua, Saeed Mahameed,
linux-netdev
[-- Attachment #1: Type: text/plain, Size: 1762 bytes --]
On Wed, Mar 27, 2019 at 01:58:17PM +0200, Or Gerlitz wrote:
> On Wed, Mar 27, 2019 at 1:41 PM Leon Romanovsky <leon@kernel.org> wrote:
> > On Wed, Mar 27, 2019 at 12:07:54PM +0200, Or Gerlitz wrote:
> > > On Tue, Mar 19, 2019 at 11:25 AM Leon Romanovsky <leon@kernel.org> wrote:
> > > > From: Artemy Kovalyov <artemyko@mellanox.com>
> > > >
> > > > Delete initialization of high order entries in mr cache to decrease initial
> > > > memory footprint. When required, the administrator can populate the
> > > > entries with memory keys via the /sys interface.
> > >
> > > Please add here:
> > >
> > > This approach is very helpful to reduce the per HW function memory
> > > footprint in environments such as VMs. Before the patch we see
> > > consumption of 0.9GB per function and after the patch about 0.1GB
> > >
> > > > Signed-off-by: Artemy Kovalyov <artemyko@mellanox.com>
> > > > Signed-off-by: Moni Shoua <monis@mellanox.com>
> > > > Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
> > >
> > > Reported-by: Shalom Toledo <shalomt@mellanox.com>
> > > Acked-by: Or Gerlitz <ogerlitz@mellanox.com>
> > >
> > > Lets push it into stable kernels, a Fixes tag here will cause that
> > > to happen more easily, so please consider that.
>
> > I'll add it at the "apply" stage.
>
> Just to make sure, by "add it" you mean the signatures, the text and
> the Fixes that?
Yes, of course.
>
> Also, is there any reason not to eliminate the mr cache pre-population
> all together?
AFAIK, pre-populated cache helps MPI application to start faster and
with out-of-box experience. Nobody seems to care enough to challenge
this internal assumption.
>
> Currently we consume 100-200MB per function after the patch which is
> also problematic for some
> environments.
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 801 bytes --]
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PATCH mlx5-next 3/4] net/mlx5: Decrease default mr cache size
2019-03-27 13:36 ` Leon Romanovsky
@ 2019-03-27 14:24 ` Or Gerlitz
0 siblings, 0 replies; 11+ messages in thread
From: Or Gerlitz @ 2019-03-27 14:24 UTC (permalink / raw)
To: Leon Romanovsky
Cc: Doug Ledford, Jason Gunthorpe, RDMA mailing list, Artemy Kovalyov,
Guy Levi, Haggai Eran, Jerome Glisse, Moni Shoua, Saeed Mahameed,
linux-netdev
On Wed, Mar 27, 2019 at 3:36 PM Leon Romanovsky <leon@kernel.org> wrote:
> On Wed, Mar 27, 2019 at 01:58:17PM +0200, Or Gerlitz wrote:
> > On Wed, Mar 27, 2019 at 1:41 PM Leon Romanovsky <leon@kernel.org> wrote:
> > > On Wed, Mar 27, 2019 at 12:07:54PM +0200, Or Gerlitz wrote:
> > > > On Tue, Mar 19, 2019 at 11:25 AM Leon Romanovsky <leon@kernel.org> wrote:
> > > > > From: Artemy Kovalyov <artemyko@mellanox.com>
> > > > >
> > > > > Delete initialization of high order entries in mr cache to decrease initial
> > > > > memory footprint. When required, the administrator can populate the
> > > > > entries with memory keys via the /sys interface.
> > > >
> > > > Please add here:
> > > >
> > > > This approach is very helpful to reduce the per HW function memory
> > > > footprint in environments such as VMs. Before the patch we see
> > > > consumption of 0.9GB per function and after the patch about 0.1GB
here it needs to be "per physical function" and not "per function"
> > > > Lets push it into stable kernels, a Fixes tag here will cause that
> > > > to happen more easily, so please consider that.
>>> I'll add it at the "apply" stage.
>> Just to make sure, by "add it" you mean the signatures, the text and
>> the Fixes that?
> Yes, of course.
thanks
>> Also, is there any reason not to eliminate the mr cache pre-population
>> all together?
> AFAIK, pre-populated cache helps MPI application to start faster and
> with out-of-box experience. Nobody seems to care enough to challenge
> this internal assumption.
I see, we are cutting here ~0.8GB so lets just do it and wonder on the
remaining 0.1GB later..
It would be good to push it into 5.1-rc so the fallout into stables
will start right away and
not only by 5.2 time - can we do that?
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [PATCH rdma-next 0/4] Small set of mlx5 related fixes
2019-03-19 9:24 [PATCH rdma-next 0/4] Small set of mlx5 related fixes Leon Romanovsky
` (3 preceding siblings ...)
2019-03-19 9:24 ` [PATCH rdma-next 4/4] IB/mlx5: Compare only index part of a memory window rkey Leon Romanovsky
@ 2019-03-27 18:29 ` Jason Gunthorpe
4 siblings, 0 replies; 11+ messages in thread
From: Jason Gunthorpe @ 2019-03-27 18:29 UTC (permalink / raw)
To: Leon Romanovsky
Cc: Doug Ledford, Leon Romanovsky, RDMA mailing list, Artemy Kovalyov,
Guy Levi, Haggai Eran, Jerome Glisse, Moni Shoua, Saeed Mahameed,
linux-netdev
On Tue, Mar 19, 2019 at 11:24:35AM +0200, Leon Romanovsky wrote:
> From: Leon Romanovsky <leonro@mellanox.com>
>
> Hi,
>
> Despite one change in mlx5_core, all those changes are for mlx5_ib. The
> change in mlx5_core is related to MR cache initialization that is not
> used in mlx5_core at all.
>
> Thanks
>
> Artemy Kovalyov (3):
> IB/mlx5: WQE dump jumps over first 16 bytes
> IB/mlx5: Compare only index part of a memory window rkey
Applied to for-next
> net/mlx5: Decrease default mr cache size
For you and Saeed
> Moni Shoua (1):
> IB/mlx5: Reset access mask when looping inside page fault handler
I sent this one to for-rc since it is @stable
Thanks,
Jason
^ permalink raw reply [flat|nested] 11+ messages in thread
end of thread, other threads:[~2019-03-27 18:30 UTC | newest]
Thread overview: 11+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2019-03-19 9:24 [PATCH rdma-next 0/4] Small set of mlx5 related fixes Leon Romanovsky
2019-03-19 9:24 ` [PATCH rdma-next 1/4] IB/mlx5: Reset access mask when looping inside page fault handler Leon Romanovsky
2019-03-19 9:24 ` [PATCH rdma-next 2/4] IB/mlx5: WQE dump jumps over first 16 bytes Leon Romanovsky
2019-03-19 9:24 ` [PATCH mlx5-next 3/4] net/mlx5: Decrease default mr cache size Leon Romanovsky
2019-03-27 10:07 ` Or Gerlitz
2019-03-27 11:41 ` Leon Romanovsky
2019-03-27 11:58 ` Or Gerlitz
2019-03-27 13:36 ` Leon Romanovsky
2019-03-27 14:24 ` Or Gerlitz
2019-03-19 9:24 ` [PATCH rdma-next 4/4] IB/mlx5: Compare only index part of a memory window rkey Leon Romanovsky
2019-03-27 18:29 ` [PATCH rdma-next 0/4] Small set of mlx5 related fixes Jason Gunthorpe
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).