From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-pg1-f199.google.com (mail-pg1-f199.google.com [209.85.215.199]) by kanga.kvack.org (Postfix) with ESMTP id CD35B6B0276 for ; Wed, 25 Jul 2018 11:52:56 -0400 (EDT) Received: by mail-pg1-f199.google.com with SMTP id e19-v6so5014374pgv.11 for ; Wed, 25 Jul 2018 08:52:56 -0700 (PDT) Received: from mail-sor-f41.google.com (mail-sor-f41.google.com. [209.85.220.41]) by mx.google.com with SMTPS id l22-v6sor3998003pgo.200.2018.07.25.08.52.55 for (Google Transport Security); Wed, 25 Jul 2018 08:52:55 -0700 (PDT) From: Nicholas Piggin Subject: [RFC PATCH 0/4] possibilities for improving invalidations Date: Thu, 26 Jul 2018 01:52:42 +1000 Message-Id: <20180725155246.1085-1-npiggin@gmail.com> Sender: owner-linux-mm@kvack.org List-ID: To: linux-mm@kvack.org Cc: Nicholas Piggin , linux-arch@vger.kernel.org I wonder if we could make some improvements to zapping pages to reduce TLB flushes under PTL, and to single threaded pte updates to reduce atomic operations. This might require some changes to arch code, particularly the last patch. I'd just like to see if I've missed something fundamental with the mm or with pte/tlb behaviour. Thanks, Nick Nicholas Piggin (4): mm: munmap optimise single threaded page freeing mm: zap_pte_range only flush under ptl if a dirty shared page was unmapped mm: zap_pte_range optimise fullmm handling for dirty shared pages mm: optimise flushing and pte manipulation for single threaded access include/asm-generic/tlb.h | 3 +++ mm/huge_memory.c | 4 ++-- mm/madvise.c | 4 ++-- mm/memory.c | 40 ++++++++++++++++++++++++++++++++------- 4 files changed, 40 insertions(+), 11 deletions(-) -- 2.17.0