* [RFC] A question about memcg/kmem @ 2015-01-13 9:24 Vladimir Davydov 2015-01-13 14:25 ` Johannes Weiner 0 siblings, 1 reply; 4+ messages in thread From: Vladimir Davydov @ 2015-01-13 9:24 UTC (permalink / raw) To: Johannes Weiner, Michal Hocko; +Cc: Andrew Morton, linux-mm, linux-kernel Hi, There's one thing about kmemcg implementation that's bothering me. It's about arrays holding per-memcg data (e.g. kmem_cache->memcg_params-> memcg_caches). On kmalloc or list_lru_{add,del} we want to quickly lookup the copy of kmem_cache or list_lru corresponding to the current cgroup. Currently, we hold all per-memcg caches/lists in an array indexed by mem_cgroup->kmemcg_id. This allows us to lookup quickly, and that's nice, but the arrays can grow indefinitely, because we reserve slots for all cgroups, including offlined, and this is disastrous and must be fixed. I see several ways how to sort this out, but none of them looks perfect to me, so I can't decide which one to choose. I would appreciate if you could share your thoughts on them. Here they are: 1. When we are about to grow arrays (new kmem-active memcg is created and there's no slot for it), try to reclaim memory from all offline kmem-active cgroups in the hope one of them will pass away and release its slot. This is not very reliable obviously, because we can fail to reclaim and have to grow arrays anyway. 2. On css offline, empty all list_lru's corresponding to the dying cgroup by moving items to the parent. Then, we could free kmemcg_id immediately on offline, and the arrays would store entries for online cgroups only, which is fine. This looks as a kind of reparenting, but it doesn't move charges, only list_lru elements, which is much easier to do. This does not conform to how we treat other charges though. 3. Use some reclaimable data structure instead of a raw array. E.g. radix tree, or idr. The structure would grow then, but it would also shrink when css's are reclaimed on memory pressure. This will probably affect performance, because we do lookups on each kmalloc, so it must be as fast as possible. It could be probably optimized by caching the result of the last lookup (hint), but hints must be per cpu then, which will make list_lru bulky. Currently, I incline to #1 or (most preferably) #2. I implemented per-memcg list_lru with this in mind, and I have patches bringing in list_lru "reparenting". #3 popped up in my mind just a few days ago. If we decide to give it a try, I'll have to drop the previous per-memcg list_lru implementation, and do a heavy rework of per-memcg kmem_cache handling as well, but I'm fine with it. I would be happy if we could opt out some of those design decisions above. E.g. "I really hate #X, it's a no-go, because..." :-) Otherwise, I'll most probably go with #2, which may become a nasty surprise to some of you. Thanks, Vladimir -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a> ^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [RFC] A question about memcg/kmem 2015-01-13 9:24 [RFC] A question about memcg/kmem Vladimir Davydov @ 2015-01-13 14:25 ` Johannes Weiner 2015-01-13 15:20 ` Vladimir Davydov 0 siblings, 1 reply; 4+ messages in thread From: Johannes Weiner @ 2015-01-13 14:25 UTC (permalink / raw) To: Vladimir Davydov; +Cc: Michal Hocko, Andrew Morton, linux-mm, linux-kernel On Tue, Jan 13, 2015 at 12:24:24PM +0300, Vladimir Davydov wrote: > Hi, > > There's one thing about kmemcg implementation that's bothering me. It's > about arrays holding per-memcg data (e.g. kmem_cache->memcg_params-> > memcg_caches). On kmalloc or list_lru_{add,del} we want to quickly > lookup the copy of kmem_cache or list_lru corresponding to the current > cgroup. Currently, we hold all per-memcg caches/lists in an array > indexed by mem_cgroup->kmemcg_id. This allows us to lookup quickly, and > that's nice, but the arrays can grow indefinitely, because we reserve > slots for all cgroups, including offlined, and this is disastrous and > must be fixed. > > I see several ways how to sort this out, but none of them looks perfect > to me, so I can't decide which one to choose. I would appreciate if you > could share your thoughts on them. Here they are: > > 1. When we are about to grow arrays (new kmem-active memcg is created > and there's no slot for it), try to reclaim memory from all offline > kmem-active cgroups in the hope one of them will pass away and > release its slot. > > This is not very reliable obviously, because we can fail to reclaim > and have to grow arrays anyway. I don't like this option because the user doesn't expect large swathes of page cache to be reclaimed simply because they created a new memcg. > 2. On css offline, empty all list_lru's corresponding to the dying > cgroup by moving items to the parent. Then, we could free kmemcg_id > immediately on offline, and the arrays would store entries for online > cgroups only, which is fine. This looks as a kind of reparenting, but > it doesn't move charges, only list_lru elements, which is much easier > to do. > > This does not conform to how we treat other charges though. This seems like the best way to do it to me. It shouldn't result in a user-visible difference in behavior and we get to keep the O(1) lookup during the allocation hotpath. Could even the reparenting be constant by using list_splice()? > 3. Use some reclaimable data structure instead of a raw array. E.g. > radix tree, or idr. The structure would grow then, but it would also > shrink when css's are reclaimed on memory pressure. > > This will probably affect performance, because we do lookups on each > kmalloc, so it must be as fast as possible. It could be probably > optimized by caching the result of the last lookup (hint), but hints > must be per cpu then, which will make list_lru bulky. I think the tree lookup in the slab allocation hotpath is prohibitive. > Currently, I incline to #1 or (most preferably) #2. I implemented > per-memcg list_lru with this in mind, and I have patches bringing in > list_lru "reparenting". #3 popped up in my mind just a few days ago. If > we decide to give it a try, I'll have to drop the previous per-memcg > list_lru implementation, and do a heavy rework of per-memcg kmem_cache > handling as well, but I'm fine with it. > > I would be happy if we could opt out some of those design decisions > above. E.g. "I really hate #X, it's a no-go, because..." :-) Otherwise, > I'll most probably go with #2, which may become a nasty surprise to some > of you. What aspects of #2 do you think are nasty? -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a> ^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [RFC] A question about memcg/kmem 2015-01-13 14:25 ` Johannes Weiner @ 2015-01-13 15:20 ` Vladimir Davydov 2015-01-13 16:28 ` Johannes Weiner 0 siblings, 1 reply; 4+ messages in thread From: Vladimir Davydov @ 2015-01-13 15:20 UTC (permalink / raw) To: Johannes Weiner; +Cc: Michal Hocko, Andrew Morton, linux-mm, linux-kernel On Tue, Jan 13, 2015 at 09:25:44AM -0500, Johannes Weiner wrote: > On Tue, Jan 13, 2015 at 12:24:24PM +0300, Vladimir Davydov wrote: > > Hi, > > > > There's one thing about kmemcg implementation that's bothering me. It's > > about arrays holding per-memcg data (e.g. kmem_cache->memcg_params-> > > memcg_caches). On kmalloc or list_lru_{add,del} we want to quickly > > lookup the copy of kmem_cache or list_lru corresponding to the current > > cgroup. Currently, we hold all per-memcg caches/lists in an array > > indexed by mem_cgroup->kmemcg_id. This allows us to lookup quickly, and > > that's nice, but the arrays can grow indefinitely, because we reserve > > slots for all cgroups, including offlined, and this is disastrous and > > must be fixed. > > > > I see several ways how to sort this out, but none of them looks perfect > > to me, so I can't decide which one to choose. I would appreciate if you > > could share your thoughts on them. Here they are: > > > > 1. When we are about to grow arrays (new kmem-active memcg is created > > and there's no slot for it), try to reclaim memory from all offline > > kmem-active cgroups in the hope one of them will pass away and > > release its slot. > > > > This is not very reliable obviously, because we can fail to reclaim > > and have to grow arrays anyway. > > I don't like this option because the user doesn't expect large swathes > of page cache to be reclaimed simply because they created a new memcg. > > > 2. On css offline, empty all list_lru's corresponding to the dying > > cgroup by moving items to the parent. Then, we could free kmemcg_id > > immediately on offline, and the arrays would store entries for online > > cgroups only, which is fine. This looks as a kind of reparenting, but > > it doesn't move charges, only list_lru elements, which is much easier > > to do. > > > > This does not conform to how we treat other charges though. > > This seems like the best way to do it to me. It shouldn't result in a > user-visible difference in behavior and we get to keep the O(1) lookup > during the allocation hotpath. Could even the reparenting be constant > by using list_splice()? Unfortunately, list_splice() doesn't seem to be an option with the list_lru API we have right now, because there's LRU_REMOVED_RETRY. It indicates that list_lru_walk callback removed an element, then dropped and reacquired the list_lru lock. In this case we first decrement nr_items to reflect an item removal, and then restart the loop. If we do list_splice() between the item removal and nr_items fix-up (when the lock was released) we'll end up with screwed nr_items. So we have to move elements one by one. Come to think of it, I believe we could change the list_lru API so that callbacks would fix nr_items by themselves. May be, we could add a special helper for walkers to remove items, say list_lru_isolate, that would fix nr_items? Anyway, I'll take a closer look in this direction. > > > 3. Use some reclaimable data structure instead of a raw array. E.g. > > radix tree, or idr. The structure would grow then, but it would also > > shrink when css's are reclaimed on memory pressure. > > > > This will probably affect performance, because we do lookups on each > > kmalloc, so it must be as fast as possible. It could be probably > > optimized by caching the result of the last lookup (hint), but hints > > must be per cpu then, which will make list_lru bulky. > > I think the tree lookup in the slab allocation hotpath is prohibitive. > > > Currently, I incline to #1 or (most preferably) #2. I implemented > > per-memcg list_lru with this in mind, and I have patches bringing in > > list_lru "reparenting". #3 popped up in my mind just a few days ago. If > > we decide to give it a try, I'll have to drop the previous per-memcg > > list_lru implementation, and do a heavy rework of per-memcg kmem_cache > > handling as well, but I'm fine with it. > > > > I would be happy if we could opt out some of those design decisions > > above. E.g. "I really hate #X, it's a no-go, because..." :-) Otherwise, > > I'll most probably go with #2, which may become a nasty surprise to some > > of you. > > What aspects of #2 do you think are nasty? We wouldn't be able to reclaim dentries/inodes accounted to an offline css w/o reclaiming objects accounted to its online ancestor. I'm not sure if we will ever want to do it though, so it isn't necessarily bad. That said, I don't see anything nasty in #2 now, but I may be short-sighted. I just want to make sure anyone interested is fine with the concept. Thank you! -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a> ^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [RFC] A question about memcg/kmem 2015-01-13 15:20 ` Vladimir Davydov @ 2015-01-13 16:28 ` Johannes Weiner 0 siblings, 0 replies; 4+ messages in thread From: Johannes Weiner @ 2015-01-13 16:28 UTC (permalink / raw) To: Vladimir Davydov; +Cc: Michal Hocko, Andrew Morton, linux-mm, linux-kernel On Tue, Jan 13, 2015 at 06:20:09PM +0300, Vladimir Davydov wrote: > On Tue, Jan 13, 2015 at 09:25:44AM -0500, Johannes Weiner wrote: > > On Tue, Jan 13, 2015 at 12:24:24PM +0300, Vladimir Davydov wrote: > > > 2. On css offline, empty all list_lru's corresponding to the dying > > > cgroup by moving items to the parent. Then, we could free kmemcg_id > > > immediately on offline, and the arrays would store entries for online > > > cgroups only, which is fine. This looks as a kind of reparenting, but > > > it doesn't move charges, only list_lru elements, which is much easier > > > to do. > > > > > > This does not conform to how we treat other charges though. > > > > This seems like the best way to do it to me. It shouldn't result in a > > user-visible difference in behavior and we get to keep the O(1) lookup > > during the allocation hotpath. Could even the reparenting be constant > > by using list_splice()? > > Unfortunately, list_splice() doesn't seem to be an option with the > list_lru API we have right now, because there's LRU_REMOVED_RETRY. It > indicates that list_lru_walk callback removed an element, then dropped > and reacquired the list_lru lock. In this case we first decrement > nr_items to reflect an item removal, and then restart the loop. If we do > list_splice() between the item removal and nr_items fix-up (when the > lock was released) we'll end up with screwed nr_items. So we have to > move elements one by one. > > Come to think of it, I believe we could change the list_lru API so that > callbacks would fix nr_items by themselves. May be, we could add a > special helper for walkers to remove items, say list_lru_isolate, that > would fix nr_items? Anyway, I'll take a closer look in this direction. The API is not set in stone. We should be able to add a function that can move pages in bulk, no? > > What aspects of #2 do you think are nasty? > > We wouldn't be able to reclaim dentries/inodes accounted to an offline > css w/o reclaiming objects accounted to its online ancestor. I'm not > sure if we will ever want to do it though, so it isn't necessarily bad. I don't think it is bad. Conceptually, the pages in any given cgroup belong to all its ancestors as well. Whether we reparent them or not, they get reclaimed during memory pressure on the hierarchy. Purging them from any other avenue besides parent pressure is unexpected, so I would like to avoid that. -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a> ^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2015-01-13 16:28 UTC | newest] Thread overview: 4+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2015-01-13 9:24 [RFC] A question about memcg/kmem Vladimir Davydov 2015-01-13 14:25 ` Johannes Weiner 2015-01-13 15:20 ` Vladimir Davydov 2015-01-13 16:28 ` Johannes Weiner
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).