From: Andrea Righi <righi.andrea@gmail.com>
To: Vivek Goyal <vgoyal@redhat.com>
Cc: Paul Menage <menage@google.com>,
Balbir Singh <balbir@linux.vnet.ibm.com>,
Gui Jianfeng <guijianfeng@cn.fujitsu.com>,
KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>,
agk@sourceware.org, akpm@linux-foundation.org, axboe@kernel.dk,
baramsori72@gmail.com, Carl Henrik Lunde <chlunde@ping.uio.no>,
dave@linux.vnet.ibm.com, Divyesh Shah <dpshah@google.com>,
eric.rannaud@gmail.com, fernando@oss.ntt.co.jp,
Hirokazu Takahashi <taka@valinux.co.jp>,
Li Zefan <lizf@cn.fujitsu.com>,
matt@bluehost.com, dradford@bluehost.com, ngupta@google.com,
randy.dunlap@oracle.com, roberto@unbit.it,
Ryo Tsuruta <ryov@valinux.co.jp>,
Satoshi UCHIDA <s-uchida@ap.jp.nec.com>,
subrata@linux.vnet.ibm.com, yoshikawa.takuya@oss.ntt.co.jp,
containers@lists.linux-foundation.org,
linux-kernel@vger.kernel.org
Subject: Re: [PATCH 1/9] io-throttle documentation
Date: Tue, 21 Apr 2009 00:05:12 +0200 [thread overview]
Message-ID: <20090420220511.GA8740@linux> (raw)
In-Reply-To: <20090420212827.GA9080@redhat.com>
On Mon, Apr 20, 2009 at 05:28:27PM -0400, Vivek Goyal wrote:
> On Sun, Apr 19, 2009 at 05:47:18PM +0200, Andrea Righi wrote:
> > On Sun, Apr 19, 2009 at 09:42:01AM -0400, Vivek Goyal wrote:
> > > > The difference between synchronous IO and writeback IO is that in the
> > > > first case the task itself is throttled via schedule_timeout_killable();
> > > > in the second case pdflush is never throttled, the IO requests instead
> > > > are simply added into a rbtree and dispatched asynchronously by another
> > > > kernel thread (kiothrottled) using a EDF-like scheduling. More exactly,
> > > > a deadline is evaluated for each writeback IO request looking at the
> > > > cgroup BW and iops/sec limits, then kiothrottled periodically selects
> > > > and dispatches the requests with an elapsed deadline.
> > > >
> > >
> > > Ok, i will look into the logic of translating cgroup BW limits into
> > > deadline. But as Nauman pointed out that we probably will run into
> > > issues of tasks with in cgroup as we loose that notion of class and prio.
> >
> > Correct. I've not addressed the IO class and priority inside cgroup, and
> > there is a lot of space for optimizations and tunings for this in the
> > io-throttle controller. In the current implementation the delay is only
> > imposed to the first task that hits the BW limit. This is not fair at
> > all.
> >
> > Ideally the throttling should be distributed equally among the tasks
> > within the same cgroup that exhaust the available BW. With equally I
> > mean depending of a function of the previous generated IO, class and IO
> > priority.
> >
> > The same concept of fairness (for ioprio and class) will be reflected to
> > the underlying IO scheduler (only CFQ at the moment) for the requests
> > that passed the BW limits.
> >
> > This doesn't seem a bad idea, well.. at least in theory... :) Do you see
> > evident weak points? or motivations to move to another direction?
> >
> > >
> > > > >
> > > > > If that's the case, will a process not see an increased rate of writes
> > > > > till we are not hitting dirty_background_ratio?
> > > >
> > > > Correct. And this is a good behaviour IMHO. At the same time we have a
> > > > smooth BW usage (according to the cgroup limits I mean) even in presence
> > > > of writeback IO only.
> > > >
> > >
> > > Hmm.., I am not able to understand this. The very fact that you will see
> > > a high rate of async writes (more than specified by cgroup max BW), till
> > > you hit dirty_background_ratio, isn't it against the goals of max bw
> > > controller? You wanted to see a consistent view of rate even if spare BW
> > > is available, and this scenario goes against that?
> >
> > The goal of the io-throttle controller is to guarantee a constant BW for
> > the IO to the block devices. If you write data in cache, buffers, etc.
> > you shouldn't be affected by any IO limitation, but you will be when the
> > data is be written out to the disk.
> >
> > OTOH if an application needs a predictable IO BW, we can always set a
> > max limit and use direct IO.
> >
> > >
> > > Think of an hypothetical configuration of 10G RAM with dirty ratio say
> > > set to 20%. Assume not much of write out is taking place in the system.
> > > So for first 2G of writes, application will be able to write it at cpu
> > > speed and no throttling will kick in and a cgroup will easily cross it
> > > max BW?
> >
> > Yes.
> >
> > >
> > > > >
> > > > > Secondly, if above is giving acceptable performance resutls, then we
> > > > > should be able to provide max bw control at IO scheduler level (along
> > > > > with proportional bw control)?
> > > > >
> > > > > So instead of doing max bw and proportional bw implementation in two
> > > > > places with the help of different controllers, I think we can do it
> > > > > with the help of one controller at one place.
> > > > >
> > > > > Please do have a look at my patches also to figure out if that's possible
> > > > > or not. I think it should be possible.
> > > > >
> > > > > Keeping both at single place should simplify the things.
> > > >
> > > > Absolutely agree to do both proportional and max BW limiting in a single
> > > > place. I still need to figure which is the best place, if the IO
> > > > scheduler in the elevator, when the IO requests are submitted. A natural
> > > > way IMHO is to control the submission of requests, also Andrew seemed to
> > > > be convinced about this approach. Anyway, I've already scheduled to test
> > > > your patchset and I'd like to see if it's possible to merge our works,
> > > > or select the best from ours patchsets.
> > > >
> > >
> > > Are we not already controlling submission of request (at crude level).
> > > If application is doing writeout at high rate, then it hits vm_dirty_ratio
> > > hits and this application is forced to do write out and hence it is slowed
> > > down and is not allowed to submit writes at high rate.
> > >
> > > Just that it is not a very fair scheme right now as during right out
> > > a high prio/high weight cgroup application can start writing out some
> > > other cgroups' pages.
> > >
> > > For this we probably need to have some combination of solutions like
> > > per cgroup upper limit on dirty pages. Secondly probably if an application
> > > is slowed down because of hitting vm_drity_ratio, it should try to
> > > write out the inode it is dirtying first instead of picking any random
> > > inode and associated pages. This will ensure that a high weight
> > > application can quickly get through the write outs and see higher
> > > throughput from the disk.
> >
> > For the first, I submitted a patchset some months ago to provide this
> > feature in the memory controller:
> >
> > https://lists.linux-foundation.org/pipermail/containers/2008-September/013140.html
> >
> > We focused on the best interface to use for setting the dirty pages
> > limit, but we didn't finalize it. I can rework on that and repost an
> > updated version. Now that we have the dirty_ratio/dirty_bytes to set the
> > global limit I think we can use the same interface and the same semantic
> > within the cgroup fs, something like:
> >
> > memory.dirty_ratio
> > memory.dirty_bytes
> >
> > For the second point something like this should be enough to force tasks
> > to write out only the inode they're actually dirtying when they hit the
> > vm_dirty_ratio limit. But it should be tested carefully and may cause
> > heavy performance regressions.
> >
> > Signed-off-by: Andrea Righi <righi.andrea@gmail.com>
> > ---
> > mm/page-writeback.c | 2 +-
> > 1 files changed, 1 insertions(+), 1 deletions(-)
> >
> > diff --git a/mm/page-writeback.c b/mm/page-writeback.c
> > index 2630937..1e07c9d 100644
> > --- a/mm/page-writeback.c
> > +++ b/mm/page-writeback.c
> > @@ -543,7 +543,7 @@ static void balance_dirty_pages(struct address_space *mapping)
> > * been flushed to permanent storage.
> > */
> > if (bdi_nr_reclaimable) {
> > - writeback_inodes(&wbc);
> > + sync_inode(mapping->host, &wbc);
> > pages_written += write_chunk - wbc.nr_to_write;
> > get_dirty_limits(&background_thresh, &dirty_thresh,
> > &bdi_thresh, bdi);
>
> This patch seems to be helping me a bit in getting more service
> differentiation between two writer dd of different weights. But strangely
> it is helping only for ext3 and not ext4. Debugging is on.
Are you explicitly mounting ext3 with data=ordered?
-Andrea
next prev parent reply other threads:[~2009-04-20 22:05 UTC|newest]
Thread overview: 104+ messages / expand[flat|nested] mbox.gz Atom feed top
2009-04-14 20:21 [PATCH 0/9] cgroup: io-throttle controller (v13) Andrea Righi
2009-04-14 20:21 ` [PATCH 1/9] io-throttle documentation Andrea Righi
2009-04-17 1:24 ` KAMEZAWA Hiroyuki
2009-04-17 1:56 ` Li Zefan
2009-04-17 10:25 ` Andrea Righi
2009-04-17 10:41 ` Andrea Righi
2009-04-17 11:35 ` Fernando Luis Vázquez Cao
2009-04-20 9:38 ` Ryo Tsuruta
2009-04-20 15:00 ` Andrea Righi
2009-04-27 10:45 ` Ryo Tsuruta
2009-04-27 12:15 ` Ryo Tsuruta
2009-04-27 21:56 ` Andrea Righi
2009-04-17 7:34 ` Gui Jianfeng
2009-04-17 7:43 ` KAMEZAWA Hiroyuki
2009-04-17 9:29 ` Gui Jianfeng
2009-04-17 9:55 ` Andrea Righi
2009-04-17 17:39 ` Vivek Goyal
2009-04-17 23:12 ` Andrea Righi
2009-04-19 13:42 ` Vivek Goyal
2009-04-19 15:47 ` Andrea Righi
2009-04-20 21:28 ` Vivek Goyal
2009-04-20 22:05 ` Andrea Righi [this message]
2009-04-21 1:08 ` Vivek Goyal
2009-04-21 8:37 ` Andrea Righi
2009-04-21 14:23 ` Vivek Goyal
2009-04-21 18:29 ` Vivek Goyal
2009-04-21 21:36 ` Andrea Righi
2009-04-21 21:28 ` Andrea Righi
2009-04-19 13:54 ` Vivek Goyal
2009-04-14 20:21 ` [PATCH 2/9] res_counter: introduce ratelimiting attributes Andrea Righi
2009-04-14 20:21 ` [PATCH 3/9] bio-cgroup controller Andrea Righi
2009-04-15 2:15 ` KAMEZAWA Hiroyuki
2009-04-15 9:37 ` Andrea Righi
2009-04-15 12:38 ` Ryo Tsuruta
2009-04-15 13:23 ` Andrea Righi
2009-04-15 23:58 ` KAMEZAWA Hiroyuki
2009-04-16 10:42 ` Andrea Righi
2009-04-16 12:00 ` Ryo Tsuruta
2009-04-17 0:04 ` KAMEZAWA Hiroyuki
2009-04-17 9:44 ` Andrea Righi
2009-04-15 13:07 ` Andrea Righi
2009-04-16 22:29 ` Andrew Morton
2009-04-17 0:20 ` KAMEZAWA Hiroyuki
2009-04-17 0:44 ` Andrew Morton
2009-04-17 1:44 ` Ryo Tsuruta
2009-04-17 4:15 ` Andrew Morton
2009-04-17 7:48 ` Ryo Tsuruta
2009-04-17 1:50 ` Balbir Singh
2009-04-17 9:40 ` Andrea Righi
2009-04-17 1:49 ` Takuya Yoshikawa
2009-04-17 2:24 ` KAMEZAWA Hiroyuki
2009-04-17 7:22 ` Ryo Tsuruta
2009-04-17 8:00 ` KAMEZAWA Hiroyuki
2009-04-17 8:48 ` KAMEZAWA Hiroyuki
2009-04-17 8:51 ` KAMEZAWA Hiroyuki
2009-04-17 11:27 ` Block I/O tracking (was Re: [PATCH 3/9] bio-cgroup controller) Fernando Luis Vázquez Cao
2009-04-17 22:09 ` Andrea Righi
2009-04-17 7:32 ` [PATCH 3/9] bio-cgroup controller Ryo Tsuruta
2009-04-17 10:22 ` Balbir Singh
2009-04-20 11:35 ` Ryo Tsuruta
2009-04-20 14:56 ` Andrea Righi
2009-04-21 11:39 ` Ryo Tsuruta
2009-04-21 15:31 ` Balbir Singh
2009-04-14 20:21 ` [PATCH 4/9] support checking of cgroup subsystem dependencies Andrea Righi
2009-04-14 20:21 ` [PATCH 5/9] io-throttle controller infrastructure Andrea Righi
2009-04-14 20:21 ` [PATCH 6/9] kiothrottled: throttle buffered (writeback) IO Andrea Righi
2009-04-14 20:21 ` [PATCH 7/9] io-throttle instrumentation Andrea Righi
2009-04-14 20:21 ` [PATCH 8/9] export per-task io-throttle statistics to userspace Andrea Righi
2009-04-14 20:21 ` [PATCH 9/9] ext3: do not throttle metadata and journal IO Andrea Righi
2009-04-17 12:38 ` Theodore Tso
2009-04-17 12:50 ` Jens Axboe
2009-04-17 14:39 ` Andrea Righi
2009-04-21 0:18 ` Theodore Tso
2009-04-21 8:30 ` Andrea Righi
2009-04-21 14:06 ` Theodore Tso
2009-04-21 14:31 ` Andrea Righi
2009-04-21 16:35 ` Theodore Tso
2009-04-21 17:23 ` Balbir Singh
2009-04-21 17:46 ` Theodore Tso
2009-04-21 18:14 ` Balbir Singh
2009-04-21 19:14 ` Theodore Tso
2009-04-21 20:49 ` Andrea Righi
2009-04-22 0:33 ` KAMEZAWA Hiroyuki
2009-04-22 1:21 ` KAMEZAWA Hiroyuki
2009-04-22 10:22 ` Andrea Righi
2009-04-23 0:05 ` KAMEZAWA Hiroyuki
2009-04-23 1:22 ` Theodore Tso
2009-04-23 2:54 ` KAMEZAWA Hiroyuki
2009-04-23 4:35 ` Theodore Tso
2009-04-23 4:58 ` Andrew Morton
2009-04-23 5:37 ` KAMEZAWA Hiroyuki
2009-04-23 9:44 ` Andrea Righi
2009-04-23 12:17 ` Theodore Tso
2009-04-23 12:27 ` Theodore Tso
2009-04-23 21:13 ` Andrea Righi
2009-04-24 0:26 ` KAMEZAWA Hiroyuki
2009-04-24 5:14 ` Balbir Singh
2009-04-23 10:03 ` Andrea Righi
2009-04-22 3:30 ` Balbir Singh
2009-04-24 15:10 ` Balbir Singh
2009-04-16 22:24 ` [PATCH 0/9] cgroup: io-throttle controller (v13) Andrew Morton
2009-04-17 9:37 ` Andrea Righi
2009-04-30 13:20 ` Alan D. Brunelle
2009-05-01 11:11 ` Andrea Righi
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20090420220511.GA8740@linux \
--to=righi.andrea@gmail.com \
--cc=agk@sourceware.org \
--cc=akpm@linux-foundation.org \
--cc=axboe@kernel.dk \
--cc=balbir@linux.vnet.ibm.com \
--cc=baramsori72@gmail.com \
--cc=chlunde@ping.uio.no \
--cc=containers@lists.linux-foundation.org \
--cc=dave@linux.vnet.ibm.com \
--cc=dpshah@google.com \
--cc=dradford@bluehost.com \
--cc=eric.rannaud@gmail.com \
--cc=fernando@oss.ntt.co.jp \
--cc=guijianfeng@cn.fujitsu.com \
--cc=kamezawa.hiroyu@jp.fujitsu.com \
--cc=linux-kernel@vger.kernel.org \
--cc=lizf@cn.fujitsu.com \
--cc=matt@bluehost.com \
--cc=menage@google.com \
--cc=ngupta@google.com \
--cc=randy.dunlap@oracle.com \
--cc=roberto@unbit.it \
--cc=ryov@valinux.co.jp \
--cc=s-uchida@ap.jp.nec.com \
--cc=subrata@linux.vnet.ibm.com \
--cc=taka@valinux.co.jp \
--cc=vgoyal@redhat.com \
--cc=yoshikawa.takuya@oss.ntt.co.jp \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).