* [PATCH] mm: thp: Set the accessed flag for old pages on access fault.
@ 2012-10-01 13:51 Will Deacon
2012-10-01 13:51 ` Will Deacon
` (2 more replies)
0 siblings, 3 replies; 9+ messages in thread
From: Will Deacon @ 2012-10-01 13:51 UTC (permalink / raw)
To: linux-mm
Cc: linux-kernel, linux-arch, akpm, mhocko, Steve Capper,
Andrea Arcangeli, Chris Metcalf, Will Deacon
From: Steve Capper <steve.capper@arm.com>
On x86 memory accesses to pages without the ACCESSED flag set result in the
ACCESSED flag being set automatically. With the ARM architecture a page access
fault is raised instead (and it will continue to be raised until the ACCESSED
flag is set for the appropriate PTE/PMD).
For normal memory pages, handle_pte_fault will call pte_mkyoung (effectively
setting the ACCESSED flag). For transparent huge pages, pmd_mkyoung will only
be called for a write fault.
This patch ensures that faults on transparent hugepages which do not result
in a CoW update the access flags for the faulting pmd.
Cc: Andrea Arcangeli <aarcange@redhat.com>
Cc: Chris Metcalf <cmetcalf@tilera.com>
Signed-off-by: Steve Capper <steve.capper@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
---
Hello again,
This is another fix for an issue that we discovered when porting THP to
ARM but it somehow managed to slip through the cracks.
Will
include/linux/huge_mm.h | 3 +++
mm/huge_memory.c | 12 ++++++++++++
mm/memory.c | 4 ++++
3 files changed, 19 insertions(+), 0 deletions(-)
diff --git a/include/linux/huge_mm.h b/include/linux/huge_mm.h
index 4c59b11..bbc62ad 100644
--- a/include/linux/huge_mm.h
+++ b/include/linux/huge_mm.h
@@ -8,6 +8,9 @@ extern int do_huge_pmd_anonymous_page(struct mm_struct *mm,
extern int copy_huge_pmd(struct mm_struct *dst_mm, struct mm_struct *src_mm,
pmd_t *dst_pmd, pmd_t *src_pmd, unsigned long addr,
struct vm_area_struct *vma);
+extern void huge_pmd_set_accessed(struct mm_struct *mm, struct vm_area_struct *vma,
+ unsigned long address, pmd_t *pmd,
+ pmd_t orig_pmd);
extern int do_huge_pmd_wp_page(struct mm_struct *mm, struct vm_area_struct *vma,
unsigned long address, pmd_t *pmd,
pmd_t orig_pmd);
diff --git a/mm/huge_memory.c b/mm/huge_memory.c
index d684934..ee9cc3b 100644
--- a/mm/huge_memory.c
+++ b/mm/huge_memory.c
@@ -914,6 +914,18 @@ out_free_pages:
goto out;
}
+void huge_pmd_set_accessed(struct mm_struct *mm, struct vm_area_struct *vma,
+ unsigned long address, pmd_t *pmd, pmd_t orig_pmd)
+{
+ pmd_t entry;
+
+ spin_lock(&mm->page_table_lock);
+ entry = pmd_mkyoung(orig_pmd);
+ if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, 0))
+ update_mmu_cache(vma, address, pmd);
+ spin_unlock(&mm->page_table_lock);
+}
+
int do_huge_pmd_wp_page(struct mm_struct *mm, struct vm_area_struct *vma,
unsigned long address, pmd_t *pmd, pmd_t orig_pmd)
{
diff --git a/mm/memory.c b/mm/memory.c
index 5736170..d5c007d 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -3537,7 +3537,11 @@ retry:
if (unlikely(ret & VM_FAULT_OOM))
goto retry;
return ret;
+ } else {
+ huge_pmd_set_accessed(mm, vma, address, pmd,
+ orig_pmd);
}
+
return 0;
}
}
--
1.7.4.1
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply related [flat|nested] 9+ messages in thread* [PATCH] mm: thp: Set the accessed flag for old pages on access fault. 2012-10-01 13:51 [PATCH] mm: thp: Set the accessed flag for old pages on access fault Will Deacon @ 2012-10-01 13:51 ` Will Deacon 2012-10-01 14:03 ` Kirill A. Shutemov 2012-10-01 14:59 ` Andrea Arcangeli 2 siblings, 0 replies; 9+ messages in thread From: Will Deacon @ 2012-10-01 13:51 UTC (permalink / raw) To: linux-mm Cc: linux-kernel, linux-arch, akpm, mhocko, Steve Capper, Andrea Arcangeli, Chris Metcalf, Will Deacon From: Steve Capper <steve.capper@arm.com> On x86 memory accesses to pages without the ACCESSED flag set result in the ACCESSED flag being set automatically. With the ARM architecture a page access fault is raised instead (and it will continue to be raised until the ACCESSED flag is set for the appropriate PTE/PMD). For normal memory pages, handle_pte_fault will call pte_mkyoung (effectively setting the ACCESSED flag). For transparent huge pages, pmd_mkyoung will only be called for a write fault. This patch ensures that faults on transparent hugepages which do not result in a CoW update the access flags for the faulting pmd. Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Chris Metcalf <cmetcalf@tilera.com> Signed-off-by: Steve Capper <steve.capper@arm.com> Signed-off-by: Will Deacon <will.deacon@arm.com> --- Hello again, This is another fix for an issue that we discovered when porting THP to ARM but it somehow managed to slip through the cracks. Will include/linux/huge_mm.h | 3 +++ mm/huge_memory.c | 12 ++++++++++++ mm/memory.c | 4 ++++ 3 files changed, 19 insertions(+), 0 deletions(-) diff --git a/include/linux/huge_mm.h b/include/linux/huge_mm.h index 4c59b11..bbc62ad 100644 --- a/include/linux/huge_mm.h +++ b/include/linux/huge_mm.h @@ -8,6 +8,9 @@ extern int do_huge_pmd_anonymous_page(struct mm_struct *mm, extern int copy_huge_pmd(struct mm_struct *dst_mm, struct mm_struct *src_mm, pmd_t *dst_pmd, pmd_t *src_pmd, unsigned long addr, struct vm_area_struct *vma); +extern void huge_pmd_set_accessed(struct mm_struct *mm, struct vm_area_struct *vma, + unsigned long address, pmd_t *pmd, + pmd_t orig_pmd); extern int do_huge_pmd_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pmd_t *pmd, pmd_t orig_pmd); diff --git a/mm/huge_memory.c b/mm/huge_memory.c index d684934..ee9cc3b 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -914,6 +914,18 @@ out_free_pages: goto out; } +void huge_pmd_set_accessed(struct mm_struct *mm, struct vm_area_struct *vma, + unsigned long address, pmd_t *pmd, pmd_t orig_pmd) +{ + pmd_t entry; + + spin_lock(&mm->page_table_lock); + entry = pmd_mkyoung(orig_pmd); + if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, 0)) + update_mmu_cache(vma, address, pmd); + spin_unlock(&mm->page_table_lock); +} + int do_huge_pmd_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pmd_t *pmd, pmd_t orig_pmd) { diff --git a/mm/memory.c b/mm/memory.c index 5736170..d5c007d 100644 --- a/mm/memory.c +++ b/mm/memory.c @@ -3537,7 +3537,11 @@ retry: if (unlikely(ret & VM_FAULT_OOM)) goto retry; return ret; + } else { + huge_pmd_set_accessed(mm, vma, address, pmd, + orig_pmd); } + return 0; } } -- 1.7.4.1 ^ permalink raw reply related [flat|nested] 9+ messages in thread
* Re: [PATCH] mm: thp: Set the accessed flag for old pages on access fault. 2012-10-01 13:51 [PATCH] mm: thp: Set the accessed flag for old pages on access fault Will Deacon 2012-10-01 13:51 ` Will Deacon @ 2012-10-01 14:03 ` Kirill A. Shutemov 2012-10-01 14:03 ` Kirill A. Shutemov 2012-10-01 14:59 ` Andrea Arcangeli 2 siblings, 1 reply; 9+ messages in thread From: Kirill A. Shutemov @ 2012-10-01 14:03 UTC (permalink / raw) To: Will Deacon Cc: linux-mm, linux-kernel, linux-arch, akpm, mhocko, Steve Capper, Andrea Arcangeli, Chris Metcalf On Mon, Oct 01, 2012 at 02:51:45PM +0100, Will Deacon wrote: > diff --git a/mm/memory.c b/mm/memory.c > index 5736170..d5c007d 100644 > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -3537,7 +3537,11 @@ retry: > if (unlikely(ret & VM_FAULT_OOM)) > goto retry; > return ret; > + } else { > + huge_pmd_set_accessed(mm, vma, address, pmd, > + orig_pmd); I think putting it to 'else' is wrong. You should not touch pmd, if it's under splitting. > } > + > return 0; > } > } -- Kirill A. Shutemov -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a> ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] mm: thp: Set the accessed flag for old pages on access fault. 2012-10-01 14:03 ` Kirill A. Shutemov @ 2012-10-01 14:03 ` Kirill A. Shutemov 0 siblings, 0 replies; 9+ messages in thread From: Kirill A. Shutemov @ 2012-10-01 14:03 UTC (permalink / raw) To: Will Deacon Cc: linux-mm, linux-kernel, linux-arch, akpm, mhocko, Steve Capper, Andrea Arcangeli, Chris Metcalf On Mon, Oct 01, 2012 at 02:51:45PM +0100, Will Deacon wrote: > diff --git a/mm/memory.c b/mm/memory.c > index 5736170..d5c007d 100644 > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -3537,7 +3537,11 @@ retry: > if (unlikely(ret & VM_FAULT_OOM)) > goto retry; > return ret; > + } else { > + huge_pmd_set_accessed(mm, vma, address, pmd, > + orig_pmd); I think putting it to 'else' is wrong. You should not touch pmd, if it's under splitting. > } > + > return 0; > } > } -- Kirill A. Shutemov ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] mm: thp: Set the accessed flag for old pages on access fault. 2012-10-01 13:51 [PATCH] mm: thp: Set the accessed flag for old pages on access fault Will Deacon 2012-10-01 13:51 ` Will Deacon 2012-10-01 14:03 ` Kirill A. Shutemov @ 2012-10-01 14:59 ` Andrea Arcangeli 2012-10-01 14:59 ` Andrea Arcangeli ` (2 more replies) 2 siblings, 3 replies; 9+ messages in thread From: Andrea Arcangeli @ 2012-10-01 14:59 UTC (permalink / raw) To: Will Deacon Cc: linux-mm, linux-kernel, linux-arch, akpm, mhocko, Steve Capper, Chris Metcalf Hi Will, On Mon, Oct 01, 2012 at 02:51:45PM +0100, Will Deacon wrote: > +void huge_pmd_set_accessed(struct mm_struct *mm, struct vm_area_struct *vma, > + unsigned long address, pmd_t *pmd, pmd_t orig_pmd) > +{ > + pmd_t entry; > + > + spin_lock(&mm->page_table_lock); > + entry = pmd_mkyoung(orig_pmd); > + if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, 0)) > + update_mmu_cache(vma, address, pmd); If the pmd is being splitted, this may not be a trasnhuge pmd anymore by the time you obtained the lock. (orig_pmd could be stale, and it wasn't verified with pmd_same either) The lock should be obtained through pmd_trans_huge_lock. if (pmd_trans_huge_lock(orig_pmd, vma) == 1) { set young bit spin_unlock(&mm->page_table_lock); } On x86: int pmdp_set_access_flags(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp, pmd_t entry, int dirty) { int changed = !pmd_same(*pmdp, entry); VM_BUG_ON(address & ~HPAGE_PMD_MASK); if (changed && dirty) { *pmdp = entry; with dirty == 0 it looks like it won't make any difference, but I guess your arm pmdp_set_access_flag is different. However it seems "dirty" means write access and so the invocation would better match the pte case: if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, flags & FAULT_FLAG_WRITE)) But note, you still have to update it even when "dirty" == 0, or it'll still infinite loop for read accesses. > + spin_unlock(&mm->page_table_lock); > +} > + > int do_huge_pmd_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, > unsigned long address, pmd_t *pmd, pmd_t orig_pmd) > { > diff --git a/mm/memory.c b/mm/memory.c > index 5736170..d5c007d 100644 > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -3537,7 +3537,11 @@ retry: > if (unlikely(ret & VM_FAULT_OOM)) > goto retry; > return ret; > + } else { > + huge_pmd_set_accessed(mm, vma, address, pmd, > + orig_pmd); > } > + > return 0; Thanks, Andrea -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a> ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] mm: thp: Set the accessed flag for old pages on access fault. 2012-10-01 14:59 ` Andrea Arcangeli @ 2012-10-01 14:59 ` Andrea Arcangeli 2012-10-01 16:32 ` Will Deacon 2012-10-16 9:37 ` Ni zhan Chen 2 siblings, 0 replies; 9+ messages in thread From: Andrea Arcangeli @ 2012-10-01 14:59 UTC (permalink / raw) To: Will Deacon Cc: linux-mm, linux-kernel, linux-arch, akpm, mhocko, Steve Capper, Chris Metcalf Hi Will, On Mon, Oct 01, 2012 at 02:51:45PM +0100, Will Deacon wrote: > +void huge_pmd_set_accessed(struct mm_struct *mm, struct vm_area_struct *vma, > + unsigned long address, pmd_t *pmd, pmd_t orig_pmd) > +{ > + pmd_t entry; > + > + spin_lock(&mm->page_table_lock); > + entry = pmd_mkyoung(orig_pmd); > + if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, 0)) > + update_mmu_cache(vma, address, pmd); If the pmd is being splitted, this may not be a trasnhuge pmd anymore by the time you obtained the lock. (orig_pmd could be stale, and it wasn't verified with pmd_same either) The lock should be obtained through pmd_trans_huge_lock. if (pmd_trans_huge_lock(orig_pmd, vma) == 1) { set young bit spin_unlock(&mm->page_table_lock); } On x86: int pmdp_set_access_flags(struct vm_area_struct *vma, unsigned long address, pmd_t *pmdp, pmd_t entry, int dirty) { int changed = !pmd_same(*pmdp, entry); VM_BUG_ON(address & ~HPAGE_PMD_MASK); if (changed && dirty) { *pmdp = entry; with dirty == 0 it looks like it won't make any difference, but I guess your arm pmdp_set_access_flag is different. However it seems "dirty" means write access and so the invocation would better match the pte case: if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, flags & FAULT_FLAG_WRITE)) But note, you still have to update it even when "dirty" == 0, or it'll still infinite loop for read accesses. > + spin_unlock(&mm->page_table_lock); > +} > + > int do_huge_pmd_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, > unsigned long address, pmd_t *pmd, pmd_t orig_pmd) > { > diff --git a/mm/memory.c b/mm/memory.c > index 5736170..d5c007d 100644 > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -3537,7 +3537,11 @@ retry: > if (unlikely(ret & VM_FAULT_OOM)) > goto retry; > return ret; > + } else { > + huge_pmd_set_accessed(mm, vma, address, pmd, > + orig_pmd); > } > + > return 0; Thanks, Andrea ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] mm: thp: Set the accessed flag for old pages on access fault. 2012-10-01 14:59 ` Andrea Arcangeli 2012-10-01 14:59 ` Andrea Arcangeli @ 2012-10-01 16:32 ` Will Deacon 2012-10-01 16:32 ` Will Deacon 2012-10-16 9:37 ` Ni zhan Chen 2 siblings, 1 reply; 9+ messages in thread From: Will Deacon @ 2012-10-01 16:32 UTC (permalink / raw) To: Andrea Arcangeli Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, linux-arch@vger.kernel.org, akpm@linux-foundation.org, mhocko@suse.cz, Steve Capper, Chris Metcalf On Mon, Oct 01, 2012 at 03:59:44PM +0100, Andrea Arcangeli wrote: > Hi Will, Hi Andrea, Kirill, Thanks for the comments. > On Mon, Oct 01, 2012 at 02:51:45PM +0100, Will Deacon wrote: > > +void huge_pmd_set_accessed(struct mm_struct *mm, struct vm_area_struct *vma, > > + unsigned long address, pmd_t *pmd, pmd_t orig_pmd) > > +{ > > + pmd_t entry; > > + > > + spin_lock(&mm->page_table_lock); > > + entry = pmd_mkyoung(orig_pmd); > > + if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, 0)) > > + update_mmu_cache(vma, address, pmd); > > If the pmd is being splitted, this may not be a trasnhuge pmd anymore > by the time you obtained the lock. (orig_pmd could be stale, and it > wasn't verified with pmd_same either) > > The lock should be obtained through pmd_trans_huge_lock. > > if (pmd_trans_huge_lock(orig_pmd, vma) == 1) > { > set young bit > spin_unlock(&mm->page_table_lock); > } I didn't notice that -- thanks. I'll move the locking outside of the _set_accessed function and direct it via that function instead. > On x86: > > int pmdp_set_access_flags(struct vm_area_struct *vma, > unsigned long address, pmd_t *pmdp, > pmd_t entry, int dirty) > { > int changed = !pmd_same(*pmdp, entry); > > VM_BUG_ON(address & ~HPAGE_PMD_MASK); > > if (changed && dirty) { > *pmdp = entry; > > with dirty == 0 it looks like it won't make any difference, but I > guess your arm pmdp_set_access_flag is different. We use the generic code, which ignores the dirty argument. Still, we should pass the correct value through anyway, so I'll fix that too. > However it seems "dirty" means write access and so the invocation > would better match the pte case: > > if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, > flags & FAULT_FLAG_WRITE)) > > > But note, you still have to update it even when "dirty" == 0, or it'll > still infinite loop for read accesses. Yup. v2 to follow once we've re-run our testing. Cheers, Will -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a> ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] mm: thp: Set the accessed flag for old pages on access fault. 2012-10-01 16:32 ` Will Deacon @ 2012-10-01 16:32 ` Will Deacon 0 siblings, 0 replies; 9+ messages in thread From: Will Deacon @ 2012-10-01 16:32 UTC (permalink / raw) To: Andrea Arcangeli Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, linux-arch@vger.kernel.org, akpm@linux-foundation.org, mhocko@suse.cz, Steve Capper, Chris Metcalf On Mon, Oct 01, 2012 at 03:59:44PM +0100, Andrea Arcangeli wrote: > Hi Will, Hi Andrea, Kirill, Thanks for the comments. > On Mon, Oct 01, 2012 at 02:51:45PM +0100, Will Deacon wrote: > > +void huge_pmd_set_accessed(struct mm_struct *mm, struct vm_area_struct *vma, > > + unsigned long address, pmd_t *pmd, pmd_t orig_pmd) > > +{ > > + pmd_t entry; > > + > > + spin_lock(&mm->page_table_lock); > > + entry = pmd_mkyoung(orig_pmd); > > + if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, 0)) > > + update_mmu_cache(vma, address, pmd); > > If the pmd is being splitted, this may not be a trasnhuge pmd anymore > by the time you obtained the lock. (orig_pmd could be stale, and it > wasn't verified with pmd_same either) > > The lock should be obtained through pmd_trans_huge_lock. > > if (pmd_trans_huge_lock(orig_pmd, vma) == 1) > { > set young bit > spin_unlock(&mm->page_table_lock); > } I didn't notice that -- thanks. I'll move the locking outside of the _set_accessed function and direct it via that function instead. > On x86: > > int pmdp_set_access_flags(struct vm_area_struct *vma, > unsigned long address, pmd_t *pmdp, > pmd_t entry, int dirty) > { > int changed = !pmd_same(*pmdp, entry); > > VM_BUG_ON(address & ~HPAGE_PMD_MASK); > > if (changed && dirty) { > *pmdp = entry; > > with dirty == 0 it looks like it won't make any difference, but I > guess your arm pmdp_set_access_flag is different. We use the generic code, which ignores the dirty argument. Still, we should pass the correct value through anyway, so I'll fix that too. > However it seems "dirty" means write access and so the invocation > would better match the pte case: > > if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, > flags & FAULT_FLAG_WRITE)) > > > But note, you still have to update it even when "dirty" == 0, or it'll > still infinite loop for read accesses. Yup. v2 to follow once we've re-run our testing. Cheers, Will ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] mm: thp: Set the accessed flag for old pages on access fault. 2012-10-01 14:59 ` Andrea Arcangeli 2012-10-01 14:59 ` Andrea Arcangeli 2012-10-01 16:32 ` Will Deacon @ 2012-10-16 9:37 ` Ni zhan Chen 2 siblings, 0 replies; 9+ messages in thread From: Ni zhan Chen @ 2012-10-16 9:37 UTC (permalink / raw) To: Andrea Arcangeli Cc: Will Deacon, linux-mm, linux-kernel, linux-arch, akpm, mhocko, Steve Capper, Chris Metcalf On 10/01/2012 10:59 PM, Andrea Arcangeli wrote: > Hi Will, > > On Mon, Oct 01, 2012 at 02:51:45PM +0100, Will Deacon wrote: >> +void huge_pmd_set_accessed(struct mm_struct *mm, struct vm_area_struct *vma, >> + unsigned long address, pmd_t *pmd, pmd_t orig_pmd) >> +{ >> + pmd_t entry; >> + >> + spin_lock(&mm->page_table_lock); >> + entry = pmd_mkyoung(orig_pmd); >> + if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, 0)) >> + update_mmu_cache(vma, address, pmd); > If the pmd is being splitted, this may not be a trasnhuge pmd anymore > by the time you obtained the lock. (orig_pmd could be stale, and it > wasn't verified with pmd_same either) Could you tell me when should call pmd_same in general? > > The lock should be obtained through pmd_trans_huge_lock. > > if (pmd_trans_huge_lock(orig_pmd, vma) == 1) > { > set young bit > spin_unlock(&mm->page_table_lock); > } > > > On x86: > > int pmdp_set_access_flags(struct vm_area_struct *vma, > unsigned long address, pmd_t *pmdp, > pmd_t entry, int dirty) > { > int changed = !pmd_same(*pmdp, entry); > > VM_BUG_ON(address & ~HPAGE_PMD_MASK); > > if (changed && dirty) { > *pmdp = entry; > > with dirty == 0 it looks like it won't make any difference, but I > guess your arm pmdp_set_access_flag is different. > > However it seems "dirty" means write access and so the invocation > would better match the pte case: > > if (pmdp_set_access_flags(vma, address & HPAGE_PMD_MASK, pmd, entry, > flags & FAULT_FLAG_WRITE)) > > > But note, you still have to update it even when "dirty" == 0, or it'll > still infinite loop for read accesses. > >> + spin_unlock(&mm->page_table_lock); >> +} >> + >> int do_huge_pmd_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, >> unsigned long address, pmd_t *pmd, pmd_t orig_pmd) >> { >> diff --git a/mm/memory.c b/mm/memory.c >> index 5736170..d5c007d 100644 >> --- a/mm/memory.c >> +++ b/mm/memory.c >> @@ -3537,7 +3537,11 @@ retry: >> if (unlikely(ret & VM_FAULT_OOM)) >> goto retry; >> return ret; >> + } else { >> + huge_pmd_set_accessed(mm, vma, address, pmd, >> + orig_pmd); >> } >> + >> return 0; > Thanks, > Andrea > > -- > To unsubscribe, send a message with 'unsubscribe linux-mm' in > the body to majordomo@kvack.org. For more info on Linux MM, > see: http://www.linux-mm.org/ . > Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a> > ^ permalink raw reply [flat|nested] 9+ messages in thread
end of thread, other threads:[~2012-10-16 9:37 UTC | newest] Thread overview: 9+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2012-10-01 13:51 [PATCH] mm: thp: Set the accessed flag for old pages on access fault Will Deacon 2012-10-01 13:51 ` Will Deacon 2012-10-01 14:03 ` Kirill A. Shutemov 2012-10-01 14:03 ` Kirill A. Shutemov 2012-10-01 14:59 ` Andrea Arcangeli 2012-10-01 14:59 ` Andrea Arcangeli 2012-10-01 16:32 ` Will Deacon 2012-10-01 16:32 ` Will Deacon 2012-10-16 9:37 ` Ni zhan Chen
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).