From: Larry Bassel <larry.bassel@oracle.com>
To: mike.kravetz@oracle.com, willy@infradead.org,
dan.j.williams@intel.com, linux-mm@kvack.org,
linux-kernel@vger.kernel.org, linux-nvdimm@lists.01.org
Cc: Larry Bassel <larry.bassel@oracle.com>
Subject: [RFC PATCH v2 0/2] Share PMDs for FS/DAX on x86
Date: Fri, 7 Jun 2019 12:51:01 -0700 [thread overview]
Message-ID: <1559937063-8323-1-git-send-email-larry.bassel@oracle.com> (raw)
Changes from v1 to v2:
* Rebased on v5.2-rc3
* An incorrect reference to "page table entries" was fixed (pointed
out by Kirill Shutemov)
* Renamed CONFIG_ARCH_WANT_HUGE_PMD_SHARE
to CONFIG_ARCH_HAS_HUGE_PMD_SHARE instead of introducing
a new config option (suggested by Dan Williams)
* Removed some unnecessary #ifdef stubs (suggested by Matt Wilcox)
* A previously overlooked case involving mprotect() is now handled
properly (pointed out by Mike Kravetz)
---
This patchset implements sharing of page tables pointing
to 2MiB pages (PMDs) for FS/DAX on x86.
Only shared mmapings of files (i.e. neither private mmapings nor
anonymous pages) are eligible for PMD sharing.
Due to the characteristics of DAX, this code is simpler and
less intrusive than the general case would be.
In our use case (high end Oracle database using DAX/XFS/PMEM/2MiB
pages) there would be significant memory savings.
A future system might have 6 TiB of PMEM on it and
there might be 10000 processes each mapping all of this 6 TiB.
Here the savings would be approximately
(6 TiB / 2 MiB) * 8 bytes (page table size) * 10000 = 240 GiB
(and these page tables themselves would probably be in
non-PMEM (ordinary RAM)).
There would also be a reduction in page faults because in
some cases the page fault has already been satisfied and
the page table entry has been filled in (and so the processes
after the first would not take a fault).
The code for detecting whether PMDs can be shared and
the implementation of sharing and unsharing is based
on, but somewhat different than that in mm/hugetlb.c,
though some of the code from this file could be reused and
thus was made non-static.
Larry Bassel (2):
Rename CONFIG_ARCH_WANT_HUGE_PMD_SHARE to
CONFIG_ARCH_HAS_HUGE_PMD_SHARE
Implement sharing/unsharing of PMDs for FS/DAX
arch/arm64/Kconfig | 2 +-
arch/arm64/mm/hugetlbpage.c | 2 +-
arch/x86/Kconfig | 2 +-
include/linux/hugetlb.h | 4 ++
mm/huge_memory.c | 37 +++++++++++++++
mm/hugetlb.c | 14 +++---
mm/memory.c | 108 +++++++++++++++++++++++++++++++++++++++++++-
7 files changed, 158 insertions(+), 11 deletions(-)
--
1.8.3.1
next reply other threads:[~2019-06-07 19:52 UTC|newest]
Thread overview: 4+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-06-07 19:51 Larry Bassel [this message]
2019-06-07 19:51 ` [RFC PATCH v2 1/2] Rename CONFIG_ARCH_WANT_HUGE_PMD_SHARE to CONFIG_ARCH_HAS_HUGE_PMD_SHARE Larry Bassel
2019-06-07 19:51 ` [RFC PATCH v2 2/2] Implement sharing/unsharing of PMDs for FS/DAX Larry Bassel
2019-06-12 2:33 ` Kirill A. Shutemov
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1559937063-8323-1-git-send-email-larry.bassel@oracle.com \
--to=larry.bassel@oracle.com \
--cc=dan.j.williams@intel.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=linux-nvdimm@lists.01.org \
--cc=mike.kravetz@oracle.com \
--cc=willy@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).