linuxppc-dev.lists.ozlabs.org archive mirror
 help / color / mirror / Atom feed
From: Sergey Senozhatsky <senozhatsky@chromium.org>
To: Nhat Pham <nphamcs@gmail.com>
Cc: Erhard Furtner <erhard_f@mailbox.org>,
	Yu Zhao <yuzhao@google.com>,
	linux-mm@kvack.org, Minchan Kim <minchan@kernel.org>,
	linux-kernel@vger.kernel.org, Yosry Ahmed <yosryahmed@google.com>,
	Sergey Senozhatsky <senozhatsky@chromium.org>,
	Johannes Weiner <hannes@cmpxchg.org>,
	Chengming Zhou <chengming.zhou@linux.dev>,
	linuxppc-dev@lists.ozlabs.org,
	"Vlastimil Babka \(SUSE\)" <vbabka@kernel.org>
Subject: Re: kswapd0: page allocation failure: order:0, mode:0x820(GFP_ATOMIC), nodemask=(null),cpuset=/,mems_allowed=0 (Kernel v6.5.9, 32bit ppc)
Date: Fri, 7 Jun 2024 20:20:52 +0900	[thread overview]
Message-ID: <20240607112052.GA479513@google.com> (raw)
In-Reply-To: <CAKEwX=OAEzKW9zO9pn41V71A8Ua2dJjD3DY6RsDMNdSd-4ts5w@mail.gmail.com>

On (24/06/07 10:40), Nhat Pham wrote:
> Personally, I'm not super convinced about class locks. We're
> essentially relying on the post-compression size of the data to
> load-balance the queries - I can imagine a scenario where a workload
> has a concentrated distribution of post-compression data (i.e its
> pages are compressed to similar-ish sizes), and we're once again
> contending for a (few) lock(s) again.
> 
> That said, I'll let the data tell the story :) We don't need a perfect
> solution, just a good enough solution for now.

Speaking of size class locks:

One thing to mention is that zsmalloc merges size classes, we never have
documented/claimed 256 size classe, the actual number is always much
much lower.  Each such "cluster" (merged size classes) holds a range of
objects' sizes (e.g. 3504-3584 bytes).  The wider the cluster's size range
the more likely the (size class) lock contention is.

Setting CONFIG_ZSMALLOC_CHAIN_SIZE to 10 or higher makes zsmalloc pool
to be configured with more size class clusters (which means that clusters
hold narrower size intervals).

  reply	other threads:[~2024-06-07 11:21 UTC|newest]

Thread overview: 46+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-05-08 18:21 kswapd0: page allocation failure: order:0, mode:0x820(GFP_ATOMIC), nodemask=(null),cpuset=/,mems_allowed=0 (Kernel v6.5.9, 32bit ppc) Erhard Furtner
2024-05-15 20:45 ` Erhard Furtner
2024-05-15 22:06   ` Yu Zhao
2024-06-01  6:01     ` Yu Zhao
2024-06-01 15:37       ` David Hildenbrand
2024-06-06  3:11         ` Michael Ellerman
2024-06-06  3:38           ` Yu Zhao
2024-06-06 12:08             ` Michael Ellerman
2024-06-06 16:05               ` Erhard Furtner
2024-06-02 18:03       ` Erhard Furtner
2024-06-02 20:38         ` Yu Zhao
2024-06-02 21:36           ` Erhard Furtner
2024-06-03 22:13         ` Erhard Furtner
2024-06-03 23:24           ` Yosry Ahmed
2024-06-04 11:44             ` Erhard Furtner
2024-06-04 16:11               ` Yosry Ahmed
2024-06-04 17:18                 ` Yu Zhao
2024-06-04 17:34                   ` Yosry Ahmed
2024-06-04 17:53                     ` Yu Zhao
2024-06-04 18:01                       ` Yosry Ahmed
2024-06-04 21:00                         ` Vlastimil Babka (SUSE)
2024-06-04 21:10                         ` Erhard Furtner
2024-06-05  3:03                           ` Yosry Ahmed
2024-06-05 23:04                             ` Erhard Furtner
2024-06-05 23:41                               ` Yosry Ahmed
2024-06-05 23:52                                 ` Yu Zhao
2024-06-05 23:58                                   ` Yosry Ahmed
2024-06-06 13:28                                     ` Erhard Furtner
2024-06-06 16:42                                       ` Yosry Ahmed
2024-06-06  2:49                                 ` Chengming Zhou
2024-06-06  4:31                                   ` Sergey Senozhatsky
2024-06-06  4:46                                     ` Chengming Zhou
2024-06-06  5:43                                       ` Sergey Senozhatsky
2024-06-06  5:55                                         ` Chengming Zhou
2024-06-07  9:40                                         ` Nhat Pham
2024-06-07 11:20                                           ` Sergey Senozhatsky [this message]
2024-06-06  7:24                                 ` Vlastimil Babka (SUSE)
2024-06-06 13:32                                   ` Erhard Furtner
2024-06-06 16:53                                     ` Vlastimil Babka (SUSE)
2024-06-06 17:14                                 ` Takero Funaki
2024-06-06 17:41                                   ` Yosry Ahmed
2024-06-06 17:55                                     ` Yu Zhao
2024-06-06 18:03                                       ` Yosry Ahmed
2024-06-04 22:17                   ` Erhard Furtner
2024-06-04 20:52             ` Vlastimil Babka (SUSE)
2024-06-04 20:55               ` Yosry Ahmed

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20240607112052.GA479513@google.com \
    --to=senozhatsky@chromium.org \
    --cc=chengming.zhou@linux.dev \
    --cc=erhard_f@mailbox.org \
    --cc=hannes@cmpxchg.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linuxppc-dev@lists.ozlabs.org \
    --cc=minchan@kernel.org \
    --cc=nphamcs@gmail.com \
    --cc=vbabka@kernel.org \
    --cc=yosryahmed@google.com \
    --cc=yuzhao@google.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).