* Why does unmap_area_sections() depend on !CONFIG_SMP?
@ 2013-10-01 9:59 Joonsoo Kim
2013-10-01 18:23 ` Nicolas Pitre
0 siblings, 1 reply; 3+ messages in thread
From: Joonsoo Kim @ 2013-10-01 9:59 UTC (permalink / raw)
To: linux-arm-kernel
Hello, Russell.
I looked at ioremap code in arm tree and found that unmap_area_sections()
is enabled only if !CONFIG_SMP. I can't understand the comments
above this function and it comes from you. Could you elaborate more
on this?
I guess that flush_cache_vunmap() before clearing page table and
flush_tlb_kernel_range() after clearing page table is safe enough to cache
consistency regardless CONFIG_SMP configuration. I think that 4K vunmap()
also depends on this flushing logic.
Please let me know what I am missing here.
Thanks.
^ permalink raw reply [flat|nested] 3+ messages in thread
* Why does unmap_area_sections() depend on !CONFIG_SMP?
2013-10-01 9:59 Why does unmap_area_sections() depend on !CONFIG_SMP? Joonsoo Kim
@ 2013-10-01 18:23 ` Nicolas Pitre
2013-10-02 5:05 ` 김준수
0 siblings, 1 reply; 3+ messages in thread
From: Nicolas Pitre @ 2013-10-01 18:23 UTC (permalink / raw)
To: linux-arm-kernel
On Tue, 1 Oct 2013, Joonsoo Kim wrote:
> Hello, Russell.
>
> I looked at ioremap code in arm tree and found that unmap_area_sections()
> is enabled only if !CONFIG_SMP. I can't understand the comments
> above this function and it comes from you. Could you elaborate more
> on this?
>
> I guess that flush_cache_vunmap() before clearing page table and
> flush_tlb_kernel_range() after clearing page table is safe enough to cache
> consistency regardless CONFIG_SMP configuration. I think that 4K vunmap()
> also depends on this flushing logic.
>
> Please let me know what I am missing here.
This is all related to the page table level involved.
Each entry in the first level page table may refer to a second level
page table covering 1MB worth of virtual space, or it may be a direct
mapping corresponding to 1MB of contiguous physical memory.
In Linux, all tasks, including kernel threads, have their own first
level page table. On process creation, the top entries covering
TASK_SIZE and above in the first level page table is copied from init_mm
into the new page table as the kernel address space is meant to be
identical across all tasks.
This is however not always the case though. Consider one call to
ioremap() which does create a new entry in the kernel virtual space.
In order to ensure that the kernel virtual space is indeed the same
across all tasks, the ioremap code would have to walk the entire task
list just to update their own copy of the kernel virtual mapping. So
what we do instead is to create the new page table entry in init_mm
only, and lazily update the other task's page table when they fault on
access due to their own page table being incomplete.
What about iounmap() then. When a mapping is removed, we don't want it
to be accessible through some random task's page table. Well, in the
normal ioremap() case, the actual mapping is created into a second level
page table which happens to be common to all tasks. Hence the first
level page table entry being created is actually a pointer to that
second level page table, and when a mapping is removed it is only
removed from that second level page table. The second level table
itself remains in memory forever, ready to be reused for any other call
to ioremap(). Therefore there is no need to update each task's first
level table again.
So far so good.
Now comes the section mapping for ioremap(). Since this is handled into
the first level page table only with no common second level table, we
needed a mechanism to ensure that any mapping removal gets propagated to
all first level tables in the system. This is accomplished with a
sequence counter namely vmalloc_seq which is incremented whenever such a
change occurs. Upon every task switch, this counter is checked against
the master copy to detect when the next task to be scheduled has its
first level page table out of date, and if so it is updated before the
new memory context is instated.
But... this works only if not SMP. On SMP, different tasks might be
running on the other CPUs and incrementing vmalloc_seq won't have any
effect on them. This is why section mappings for ioremap() is not
available if SMP.
This could probably be fixed by sending an IPI to the other processors,
forcing them to resync their page table right after clearing the mapping
from the master table. But no one implemented it so far.
Nicolas
^ permalink raw reply [flat|nested] 3+ messages in thread
* Why does unmap_area_sections() depend on !CONFIG_SMP?
2013-10-01 18:23 ` Nicolas Pitre
@ 2013-10-02 5:05 ` 김준수
0 siblings, 0 replies; 3+ messages in thread
From: 김준수 @ 2013-10-02 5:05 UTC (permalink / raw)
To: linux-arm-kernel
> -----Original Message-----
> From: Nicolas Pitre [mailto:nicolas.pitre at linaro.org]
> Sent: Wednesday, October 02, 2013 3:24 AM
> To: Joonsoo Kim
> Cc: Russell King; linux-arm-kernel at lists.infradead.org
> Subject: Re: Why does unmap_area_sections() depend on !CONFIG_SMP?
>
> On Tue, 1 Oct 2013, Joonsoo Kim wrote:
>
> > Hello, Russell.
> >
> > I looked at ioremap code in arm tree and found that
> > unmap_area_sections() is enabled only if !CONFIG_SMP. I can't
> > understand the comments above this function and it comes from you.
> > Could you elaborate more on this?
> >
> > I guess that flush_cache_vunmap() before clearing page table and
> > flush_tlb_kernel_range() after clearing page table is safe enough to
> > cache consistency regardless CONFIG_SMP configuration. I think that 4K
> > vunmap() also depends on this flushing logic.
> >
> > Please let me know what I am missing here.
>
> This is all related to the page table level involved.
>
> Each entry in the first level page table may refer to a second level page
> table covering 1MB worth of virtual space, or it may be a direct mapping
> corresponding to 1MB of contiguous physical memory.
>
> In Linux, all tasks, including kernel threads, have their own first level
> page table. On process creation, the top entries covering TASK_SIZE and
> above in the first level page table is copied from init_mm into the new
> page table as the kernel address space is meant to be identical across all
> tasks.
>
> This is however not always the case though. Consider one call to
> ioremap() which does create a new entry in the kernel virtual space.
> In order to ensure that the kernel virtual space is indeed the same across
> all tasks, the ioremap code would have to walk the entire task list just
> to update their own copy of the kernel virtual mapping. So what we do
> instead is to create the new page table entry in init_mm only, and lazily
> update the other task's page table when they fault on access due to their
> own page table being incomplete.
>
> What about iounmap() then. When a mapping is removed, we don't want it to
> be accessible through some random task's page table. Well, in the normal
> ioremap() case, the actual mapping is created into a second level page
> table which happens to be common to all tasks. Hence the first level page
> table entry being created is actually a pointer to that second level page
> table, and when a mapping is removed it is only removed from that second
> level page table. The second level table itself remains in memory
forever,
> ready to be reused for any other call to ioremap(). Therefore there is no
> need to update each task's first level table again.
>
> So far so good.
>
> Now comes the section mapping for ioremap(). Since this is handled into
> the first level page table only with no common second level table, we
> needed a mechanism to ensure that any mapping removal gets propagated to
> all first level tables in the system. This is accomplished with a
> sequence counter namely vmalloc_seq which is incremented whenever such a
> change occurs. Upon every task switch, this counter is checked against
> the master copy to detect when the next task to be scheduled has its first
> level page table out of date, and if so it is updated before the new
> memory context is instated.
>
> But... this works only if not SMP. On SMP, different tasks might be
> running on the other CPUs and incrementing vmalloc_seq won't have any
> effect on them. This is why section mappings for ioremap() is not
> available if SMP.
>
> This could probably be fixed by sending an IPI to the other processors,
> forcing them to resync their page table right after clearing the mapping
> from the master table. But no one implemented it so far.
Hello, Nicolas.
Really thanks for kind explanation.
Now, I totally understand why it is not available if SMP.
I will investigate more on this and try to implement that section mapping
for ioremap() works on SMP.
Thanks.
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2013-10-02 5:05 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2013-10-01 9:59 Why does unmap_area_sections() depend on !CONFIG_SMP? Joonsoo Kim
2013-10-01 18:23 ` Nicolas Pitre
2013-10-02 5:05 ` 김준수
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).