From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751685Ab1IUTAm (ORCPT ); Wed, 21 Sep 2011 15:00:42 -0400 Received: from mx2.parallels.com ([64.131.90.16]:47809 "EHLO mx2.parallels.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1750932Ab1IUTAi (ORCPT ); Wed, 21 Sep 2011 15:00:38 -0400 Message-ID: <4E7A342B.5040608@parallels.com> Date: Wed, 21 Sep 2011 15:59:55 -0300 From: Glauber Costa User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:6.0.2) Gecko/20110906 Thunderbird/6.0.2 MIME-Version: 1.0 To: Greg Thelen CC: , , , , , , , , Subject: Re: [PATCH v3 2/7] socket: initial cgroup code. References: <1316393805-3005-1-git-send-email-glommer@parallels.com> <1316393805-3005-3-git-send-email-glommer@parallels.com> In-Reply-To: Content-Type: text/plain; charset="ISO-8859-1"; format=flowed Content-Transfer-Encoding: 7bit X-Originating-IP: [201.82.134.226] Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 09/21/2011 03:47 PM, Greg Thelen wrote: > On Sun, Sep 18, 2011 at 5:56 PM, Glauber Costa wrote: >> We aim to control the amount of kernel memory pinned at any >> time by tcp sockets. To lay the foundations for this work, >> this patch adds a pointer to the kmem_cgroup to the socket >> structure. >> >> Signed-off-by: Glauber Costa >> CC: David S. Miller >> CC: Hiroyouki Kamezawa >> CC: Eric W. Biederman > ... >> +void sock_update_memcg(struct sock *sk) >> +{ >> + /* right now a socket spends its whole life in the same cgroup */ >> + BUG_ON(sk->sk_cgrp); >> + >> + rcu_read_lock(); >> + sk->sk_cgrp = mem_cgroup_from_task(current); >> + >> + /* >> + * We don't need to protect against anything task-related, because >> + * we are basically stuck with the sock pointer that won't change, >> + * even if the task that originated the socket changes cgroups. >> + * >> + * What we do have to guarantee, is that the chain leading us to >> + * the top level won't change under our noses. Incrementing the >> + * reference count via cgroup_exclude_rmdir guarantees that. >> + */ >> + cgroup_exclude_rmdir(mem_cgroup_css(sk->sk_cgrp)); > > This grabs a css_get() reference, which prevents rmdir (will return > -EBUSY). Yes. How long is this reference held? For the socket lifetime. > I wonder about the case > where a process creates a socket in memcg M1 and later is moved into > memcg M2. At that point an admin would expect to be able to 'rmdir > M1'. I think this rmdir would return -EBUSY and I suspect it would be > difficult for the admin to understand why the rmdir of M1 failed. It > seems that to rmdir a memcg, an admin would have to kill all processes > that allocated sockets while in M1. Such processes may not still be > in M1. > >> + rcu_read_unlock(); >> +} I agree. But also, don't see too much ways around it without implementing full task migration. Right now I am working under the assumption that tasks are long lived inside the cgroup. Migration potentially introduces some nasty locking problems in the mem_schedule path. Also, unless I am missing something, the memcg already has the policy of not carrying charges around, probably because of this very same complexity. True that at least it won't EBUSY you... But I think this is at least a way to guarantee that the cgroup under our nose won't disappear in the middle of our allocations.