From: Ilias Apalodimas <ilias.apalodimas@linaro.org>
To: "Li,Rongqing" <lirongqing@baidu.com>
Cc: Yunsheng Lin <linyunsheng@huawei.com>,
Saeed Mahameed <saeedm@mellanox.com>,
"jonathan.lemon@gmail.com" <jonathan.lemon@gmail.com>,
"netdev@vger.kernel.org" <netdev@vger.kernel.org>,
"brouer@redhat.com" <brouer@redhat.com>,
"ivan.khoronzhuk@linaro.org" <ivan.khoronzhuk@linaro.org>,
"grygorii.strashko@ti.com" <grygorii.strashko@ti.com>
Subject: Re: 答复: [PATCH][v2] page_pool: handle page recycle for NUMA_NO_NODE condition
Date: Mon, 9 Dec 2019 11:30:14 +0200 [thread overview]
Message-ID: <20191209093014.GA25360@apalos.home> (raw)
In-Reply-To: <96bc5e8351a54adc8f00c18a61e2555d@baidu.com>
On Mon, Dec 09, 2019 at 03:47:50AM +0000, Li,Rongqing wrote:
> > >
[...]
> > > Cc'ed Jesper, Ilias & Jonathan.
> > >
> > > I don't think it is correct to check that the page nid is same as
> > > numa_mem_id() if pool is NUMA_NO_NODE. In such case we should allow
> > > all pages to recycle, because you can't assume where pages are
> > > allocated from and where they are being handled.
> > >
> > > I suggest the following:
> > >
> > > return !page_pfmemalloc() &&
> > > ( page_to_nid(page) == pool->p.nid || pool->p.nid == NUMA_NO_NODE );
> > >
> > > 1) never recycle emergency pages, regardless of pool nid.
> > > 2) always recycle if pool is NUMA_NO_NODE.
> >
> > As I can see, below are the cases that the pool->p.nid could be
> > NUMA_NO_NODE:
> >
> > 1. kernel built with the CONFIG_NUMA being off.
> >
> > 2. kernel built with the CONFIG_NUMA being on, but FW/BIOS dose not provide
> > a valid node id through ACPI/DT, and it has the below cases:
> >
> > a). the hardware itself is single numa node system, so maybe it is valid
> > to not provide a valid node for the device.
> > b). the hardware itself is multi numa nodes system, and the FW/BIOS is
> > broken that it does not provide a valid one.
> >
> > 3. kernel built with the CONFIG_NUMA being on, and FW/BIOS dose provide a
> > valid node id through ACPI/DT, but the driver does not pass the valid
> > node id when calling page_pool_init().
> >
> > I am not sure which case this patch is trying to fix, maybe Rongqing can help to
> > clarify.
> >
> > For case 1 and case 2 (a), I suppose checking pool->p.nid being
> > NUMA_NO_NODE is enough.
> >
> > For case 2 (b), There may be argument that it should be fixed in the BIOS/FW,
> > But it also can be argued that the numa_mem_id() checking has been done in
> > the driver that has not using page pool yet when deciding whether to do
> > recycling, see [1]. If I understanding correctly, recycling is normally done at the
> > NAPI pooling, which has the same affinity as the rx interrupt, and rx interrupt is
> > not changed very often. If it does change to different memory node, maybe it
> > makes sense not to recycle the old page belongs to old node?
> >
> >
> > For case 3, I am not sure if any driver is doing that, and if the page pool API
> > even allow that?
> >
>
> I think pool_page_reusable should support NUMA_NO_NODE no matter which cases
>
Yes
>
> And recycling is normally done at the NAPI pooling, NUMA_NO_NODE hint to use the
> local node, except memory usage is unbalance, so I add the check that the page nid is
> same as numa_mem_id() if pool is NUMA_NO_NODE
I am not sure i am following here.
Recycling done at NAPI or not should have nothing to do with NUMA.
What do you mean by 'memory usage is unbalance'?
Thanks
/Ilias
next prev parent reply other threads:[~2019-12-09 9:30 UTC|newest]
Thread overview: 44+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-12-06 9:32 [PATCH][v2] page_pool: handle page recycle for NUMA_NO_NODE condition Li RongQing
2019-12-07 3:52 ` Saeed Mahameed
2019-12-09 1:31 ` Yunsheng Lin
2019-12-09 3:47 ` 答复: " Li,Rongqing
2019-12-09 9:30 ` Ilias Apalodimas [this message]
2019-12-09 10:37 ` 答复: " Li,Rongqing
2019-12-09 12:14 ` Jesper Dangaard Brouer
2019-12-09 23:34 ` Saeed Mahameed
2019-12-10 1:31 ` Yunsheng Lin
2019-12-10 9:39 ` 答复: " Li,Rongqing
2019-12-10 14:52 ` Ilias Apalodimas
2019-12-10 19:56 ` Saeed Mahameed
2019-12-10 19:45 ` Saeed Mahameed
2019-12-11 3:01 ` Yunsheng Lin
2019-12-11 3:06 ` Yunsheng Lin
2019-12-11 20:57 ` Saeed Mahameed
2019-12-12 1:04 ` Yunsheng Lin
2019-12-10 15:02 ` Ilias Apalodimas
2019-12-10 20:02 ` Saeed Mahameed
2019-12-10 20:10 ` Ilias Apalodimas
2019-12-11 18:49 ` Jesper Dangaard Brouer
2019-12-11 21:24 ` Saeed Mahameed
2019-12-12 1:34 ` Yunsheng Lin
2019-12-12 10:18 ` Jesper Dangaard Brouer
2019-12-13 3:40 ` Yunsheng Lin
2019-12-13 6:27 ` 答复: " Li,Rongqing
2019-12-13 6:53 ` Yunsheng Lin
2019-12-13 8:48 ` Jesper Dangaard Brouer
2019-12-16 1:51 ` Yunsheng Lin
2019-12-16 4:02 ` 答复: " Li,Rongqing
2019-12-16 10:13 ` Ilias Apalodimas
2019-12-16 10:16 ` Ilias Apalodimas
2019-12-16 10:57 ` 答复: " Li,Rongqing
2019-12-17 19:38 ` Saeed Mahameed
2019-12-17 19:35 ` Saeed Mahameed
2019-12-17 19:27 ` Saeed Mahameed
2019-12-16 12:15 ` Michal Hocko
2019-12-16 12:34 ` Ilias Apalodimas
2019-12-16 13:08 ` Michal Hocko
2019-12-16 13:21 ` Ilias Apalodimas
2019-12-17 2:11 ` Yunsheng Lin
2019-12-17 9:11 ` Michal Hocko
2019-12-19 2:09 ` Yunsheng Lin
2019-12-19 11:53 ` Michal Hocko
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20191209093014.GA25360@apalos.home \
--to=ilias.apalodimas@linaro.org \
--cc=brouer@redhat.com \
--cc=grygorii.strashko@ti.com \
--cc=ivan.khoronzhuk@linaro.org \
--cc=jonathan.lemon@gmail.com \
--cc=linyunsheng@huawei.com \
--cc=lirongqing@baidu.com \
--cc=netdev@vger.kernel.org \
--cc=saeedm@mellanox.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).