stable.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Doug Ledford <dledford@redhat.com>
To: Davidlohr Bueso <davidlohr@hp.com>,
	m@silodev.com, stable@vger.kernel.org, akpm@linux-foundation.org
Cc: linux-kernel@vger.kernel.org, Manfred Spraul <manfred@colorfullife.com>
Subject: Re: [PATCH] ipc,mqueue: remove limits for the amount of system-wide queues
Date: Tue, 11 Feb 2014 13:14:07 -0500	[thread overview]
Message-ID: <52FA686F.6070701@redhat.com> (raw)
In-Reply-To: <1391979963.1099.34.camel@buesod1.americas.hpqcorp.net>

[-- Attachment #1: Type: text/plain, Size: 1807 bytes --]

On 2/9/2014 4:06 PM, Davidlohr Bueso wrote:
> From: Davidlohr Bueso <davidlohr@hp.com>
> 
> Commit 93e6f119 (ipc/mqueue: cleanup definition names and locations) added
> global hardcoded limits to the amount of message queues that can be created.
> While these limits are per-namespace, reality is that it ends up breaking
> userspace applications. Historically users have, at least in theory, been able
> to create up to INT_MAX queues, and limiting it to just 1024 is way too low
> and dramatic for some workloads and use cases. For instance, Madars reports:
> 
> "This update imposes bad limits on our multi-process application. As our
> app uses approaches that each process opens its own set of queues (usually
> something about 3-5 queues per process). In some scenarios we might run up
> to 3000 processes or more (which of-course for linux is not a problem).
> Thus we might need up to 9000 queues or more. All processes run under one
> user."
> 
> Other affected users can be found in launchpad bug #1155695:
> https://bugs.launchpad.net/ubuntu/+source/manpages/+bug/1155695
> 
> Instead of increasing this limit, revert it entirely and fallback to the
> original way of dealing queue limits -- where once a user's resource limit
> is reached, and all memory is used, new queues cannot be created.
> 
> Reported-by: m@silodev.com
> Cc: Doug Ledford <dledford@redhat.com>

Acked-by: Doug Ledford <dledford@redhat.com>

> Cc: Manfred Spraul <manfred@colorfullife.com>
> Cc: stable@vger.kernel.org # v3.5+
> Signed-off-by: Davidlohr Bueso <davidlohr@hp.com>
> ---
>  include/linux/ipc_namespace.h |  2 --
>  ipc/mq_sysctl.c               | 18 ++++++++++++------
>  ipc/mqueue.c                  |  6 +++---
>  3 files changed, 15 insertions(+), 11 deletions(-)



[-- Attachment #2: OpenPGP digital signature --]
[-- Type: application/pgp-signature, Size: 899 bytes --]

  reply	other threads:[~2014-02-11 18:14 UTC|newest]

Thread overview: 3+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <11414.87.110.183.114.1391682066.squirrel@www.silodev.com>
     [not found] ` <1391803868.1099.23.camel@buesod1.americas.hpqcorp.net>
2014-02-09 21:06   ` [PATCH] ipc,mqueue: remove limits for the amount of system-wide queues Davidlohr Bueso
2014-02-11 18:14     ` Doug Ledford [this message]
2014-02-11 22:16     ` Andrew Morton

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=52FA686F.6070701@redhat.com \
    --to=dledford@redhat.com \
    --cc=akpm@linux-foundation.org \
    --cc=davidlohr@hp.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=m@silodev.com \
    --cc=manfred@colorfullife.com \
    --cc=stable@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).