From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id C9942C4707C for ; Fri, 12 Jan 2024 07:09:53 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 420796B0098; Fri, 12 Jan 2024 02:09:53 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 3D0486B0099; Fri, 12 Jan 2024 02:09:53 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 299816B009A; Fri, 12 Jan 2024 02:09:53 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0017.hostedemail.com [216.40.44.17]) by kanga.kvack.org (Postfix) with ESMTP id 190726B0098 for ; Fri, 12 Jan 2024 02:09:53 -0500 (EST) Received: from smtpin08.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay02.hostedemail.com (Postfix) with ESMTP id E5D85120CE5 for ; Fri, 12 Jan 2024 07:09:52 +0000 (UTC) X-FDA: 81669784224.08.4386769 Received: from out-183.mta0.migadu.com (out-183.mta0.migadu.com [91.218.175.183]) by imf27.hostedemail.com (Postfix) with ESMTP id E59E04000A for ; Fri, 12 Jan 2024 07:09:49 +0000 (UTC) Authentication-Results: imf27.hostedemail.com; dkim=pass header.d=linux.dev header.s=key1 header.b="rl6/mhHP"; spf=pass (imf27.hostedemail.com: domain of gang.li@linux.dev designates 91.218.175.183 as permitted sender) smtp.mailfrom=gang.li@linux.dev; dmarc=pass (policy=none) header.from=linux.dev ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1705043390; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=x1/E36ja6hSjXcFMnhf7nMX3AJMPWbXzIlOegBOBTgw=; b=CYBCjDQtl0EryfQKJslun9lQXkKZepy7/hrhTZDLn8vGZbLilDYq4dYduY1srJlNTkFPBr p2sQNvQMCL7viqggZzw3pGuLXU7ZB8yYTcbrZR7beSuSo3SWEeneZ5+M01Rb5kqr9fri9y 5f4V3kPg4KAt63pzfVF4FED1RJpGov8= ARC-Authentication-Results: i=1; imf27.hostedemail.com; dkim=pass header.d=linux.dev header.s=key1 header.b="rl6/mhHP"; spf=pass (imf27.hostedemail.com: domain of gang.li@linux.dev designates 91.218.175.183 as permitted sender) smtp.mailfrom=gang.li@linux.dev; dmarc=pass (policy=none) header.from=linux.dev ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1705043390; a=rsa-sha256; cv=none; b=Ua6a3Et+O6yMk1YAal/iV+hGwCvXV/MXENHMvGdjvuZfyM/5N4XrX2f+P6L9acjpa9QuOx Zx9I8FEqruDTqABTtD4gK9QWbMCKbloEAG120AZX+RRoyjoDFm6/UTs1sekhPwTDc2PFoo TqOs6B3t2/XMGpGsoyHtxOMZ/PYifrQ= Message-ID: DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1705043387; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=x1/E36ja6hSjXcFMnhf7nMX3AJMPWbXzIlOegBOBTgw=; b=rl6/mhHP8qss1zJkt8tAb+O2UQRGEreguRM5ECbwehucdUCmiM3gdzMM1AvxJ6bX7sZ25K tKZs7xxGHKIplmLH6QebIJt7P/KRkt3ckrFNsgOC+P31/Re+W92p+KVm7w3R9j9bpYCWaL Zx12flbfZLdvnRGUe7PLpX4I8UkNgSs= Date: Fri, 12 Jan 2024 15:09:18 +0800 MIME-Version: 1.0 Subject: Re: [PATCH v3 3/7] padata: dispatch works on different nodes Content-Language: en-US To: Tim Chen Cc: linux-mm@kvack.org, Andrew Morton , Mike Kravetz , David Rientjes , linux-kernel@vger.kernel.org, ligang.bdlg@bytedance.com, David Hildenbrand , Muchun Song References: <20240102131249.76622-1-gang.li@linux.dev> <20240102131249.76622-4-gang.li@linux.dev> <1d9074955618ea0b4b155701f7c1b8b18a43fa8d.camel@linux.intel.com> X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. From: Gang Li In-Reply-To: <1d9074955618ea0b4b155701f7c1b8b18a43fa8d.camel@linux.intel.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-Migadu-Flow: FLOW_OUT X-Rspamd-Queue-Id: E59E04000A X-Rspam-User: X-Stat-Signature: b7yg4ctn7f8c11e9zs6um4hqk49dkayi X-Rspamd-Server: rspam01 X-HE-Tag: 1705043389-282907 X-HE-Meta: U2FsdGVkX189heWDp6fLBTy2M0Wsz/7KiHbU0vtXruHHFQFTD95/vTYRQbK/OP3OpGC0YSXHjWHIbUuznjmx8o2TMDA9ueTRJLYKGHc60yH8H5rZeJwswJweHe7rmr4EUZNYY9AqRrb6M5PL1yh78wqwgNA2ZMTvryHnHIV1I5K4r9MLqQoLrF8TY6hU1sSDk8fRpTeREwIzLa5xgTmP7st94xKKog/Qpu6MgMpPnUCFYvTfWYWmVQLsA/Zvw66gxUC2JMQdAzc1RhpHcIIDfSSduu1ce7egEnrtboy86xshHNNFhS0EEuJLen5zp4i17BUcjqyJFwBSOaxS5rrttNBbNkBeup3z3Lz7NZtz9Uc7fyQNxRIEJq/+c2GmfqGj8BzIyiOLC1rqh5MHZItXZOGksu7jt7VpbGzsEhZDZ/u7xVCcc1uP7UEI+yRy9eG2vplYEzVPrgrQfAgO0h1ptO7GsalN0dfeo5n9eByV3LUzVqyP8NzjiX1cVlSshp6szBKI3t+m1uvdUL39Kj+Ggf1a5VJcoUoW6TTOdeg9freb7fXictSNv0NjNrTyaNHt3NpQLoW9o4li+zHuOg45p+hHK6ivXJfhv/ZZC4ucY/1v+/3LvEPbkyd0dOUHa6EueUeD8yWzkLE68egy7lx+SFn+eVEFK5O6dxLoBcyzf5d3r9+LFc5Dy+0FWEwQy4JrdmLTs+Wl7CObiMerV2jpKQF1yOnnyBfStOWZ5fOe03dh5YbU4ipnY/m0MBpSG0RnSGjFDajjXdfdPNCG1tfWvf+5doCbLt1WEWoG+KL2M1S6AcrBWYle2bNIyscCBv+qzjHXplbv6KWzZQFJcTAQt2pyeGt9RHCYRO3TRvwXpB9+3uyQBs2GKVJq6uHUHGGLwnVjezpGqE0ydpFYk01/WCPixz+FtHumaUSJ6GV3SFdgn6X9R0ZvadSExG0b93VkzY48NsN2wRr8PQhty/V Y1UhBLdR GObOSI++LUJy8HbXsNC/JSlAY7QAtM9DKAAUb+QUV5wJx6fh7yzs2w6+0FbrvZBblv75zgphCHRk+CgzUetcPsQIVfg== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: On 2024/1/12 01:50, Tim Chen wrote: > On Tue, 2024-01-02 at 21:12 +0800, Gang Li wrote: >> When a group of tasks that access different nodes are scheduled on the >> same node, they may encounter bandwidth bottlenecks and access latency. >> >> Thus, numa_aware flag is introduced here, allowing tasks to be >> distributed across different nodes to fully utilize the advantage of >> multi-node systems. >> >> Signed-off-by: Gang Li >> --- >> include/linux/padata.h | 3 +++ >> kernel/padata.c | 8 ++++++-- >> mm/mm_init.c | 1 + >> 3 files changed, 10 insertions(+), 2 deletions(-) >> >> diff --git a/include/linux/padata.h b/include/linux/padata.h >> index 495b16b6b4d72..f79ccd50e7f40 100644 >> --- a/include/linux/padata.h >> +++ b/include/linux/padata.h >> @@ -137,6 +137,8 @@ struct padata_shell { >> * appropriate for one worker thread to do at once. >> * @max_threads: Max threads to use for the job, actual number may be less >> * depending on task size and minimum chunk size. >> + * @numa_aware: Dispatch jobs to different nodes. If a node only has memory but >> + * no CPU, dispatch its jobs to a random CPU. >> */ >> struct padata_mt_job { >> void (*thread_fn)(unsigned long start, unsigned long end, void *arg); >> @@ -146,6 +148,7 @@ struct padata_mt_job { >> unsigned long align; >> unsigned long min_chunk; >> int max_threads; >> + bool numa_aware; >> }; >> >> /** >> diff --git a/kernel/padata.c b/kernel/padata.c >> index 179fb1518070c..1c2b3a337479e 100644 >> --- a/kernel/padata.c >> +++ b/kernel/padata.c >> @@ -485,7 +485,7 @@ void __init padata_do_multithreaded(struct padata_mt_job *job) >> struct padata_work my_work, *pw; >> struct padata_mt_job_state ps; >> LIST_HEAD(works); >> - int nworks; >> + int nworks, nid = 0; > > If we always start from 0, we may be biased towards the low numbered node, > and not use high numbered nodes at all. Suggest you do > static nid = 0; > When we use `static`, if there are multiple parallel calls to `padata_do_multithreaded`, it may result in an uneven distribution of tasks for each padata_do_multithreaded. We can make the following modifications to address this issue. ``` diff --git a/kernel/padata.c b/kernel/padata.c index 1c2b3a337479e..925e48df6dd8d 100644 --- a/kernel/padata.c +++ b/kernel/padata.c @@ -485,7 +485,8 @@ void __init padata_do_multithreaded(struct padata_mt_job *job) struct padata_work my_work, *pw; struct padata_mt_job_state ps; LIST_HEAD(works); - int nworks, nid = 0; + int nworks, nid; + static volatile int global_nid = 0; if (job->size == 0) return; @@ -516,12 +517,15 @@ void __init padata_do_multithreaded(struct padata_mt_job *job) ps.chunk_size = max(ps.chunk_size, job->min_chunk); ps.chunk_size = roundup(ps.chunk_size, job->align); + nid = global_nid; list_for_each_entry(pw, &works, pw_list) - if (job->numa_aware) - queue_work_node((++nid % num_node_state(N_MEMORY)), - system_unbound_wq, &pw->pw_work); - else + if (job->numa_aware) { + queue_work_node(nid, system_unbound_wq, &pw->pw_work); + nid = next_node(nid, node_states[N_CPU]); + } else queue_work(system_unbound_wq, &pw->pw_work); + if (job->numa_aware) + global_nid = nid; /* Use the current thread, which saves starting a workqueue worker. */ padata_work_init(&my_work, padata_mt_helper, &ps, PADATA_WORK_ONSTACK); ``` >> >> if (job->size == 0) >> return; >> @@ -517,7 +517,11 @@ void __init padata_do_multithreaded(struct padata_mt_job *job) >> ps.chunk_size = roundup(ps.chunk_size, job->align); >> >> list_for_each_entry(pw, &works, pw_list) >> - queue_work(system_unbound_wq, &pw->pw_work); >> + if (job->numa_aware) >> + queue_work_node((++nid % num_node_state(N_MEMORY)), >> + system_unbound_wq, &pw->pw_work); > > I think we should use nid = next_node(nid, node_states[N_CPU]) instead of > ++nid % num_node_state(N_MEMORY). You are picking the next node with CPU > to handle the job. > > Tim > I agree.