From: Josef Bacik <jbacik@fusionio.com>
To: Alex Lyakas <alex.btrfs@zadarastorage.com>
Cc: linux-btrfs <linux-btrfs@vger.kernel.org>,
Josef Bacik <jbacik@fusionio.com>,
Lev Vainblat <lev@zadarastorage.com>,
Shyam Kaushik <shyam@zadarastorage.com>, <clmason@fusionio.com>
Subject: Re: btrfs:async-thread: atomic_start_pending=1 is set, but it's too late
Date: Thu, 29 Aug 2013 16:03:06 -0400 [thread overview]
Message-ID: <20130829200306.GF10591@localhost.localdomain> (raw)
In-Reply-To: <CAOcd+r0PfcX9Xs9ND8zXui3-YyJh+9THz1qrg7u9DsH1L-uO2g@mail.gmail.com>
On Mon, Aug 26, 2013 at 05:16:42PM +0300, Alex Lyakas wrote:
> Greetings all,
> I see a following issue with spawning new threads for btrfs_workers
> that have atomic_worker_start set:
>
> # I have BTRFS that has 24Gb of total metadata, out of which extent
> tree takes 11Gb; space_cache option is not used.
> # After mouting, cache_block_group() triggers ~250 work items to
> cache-in the needed block groups.
> # At this point, fs_info->caching_workers has one thread, which is
> considered "idle".
> # Work items start to add to this thread's "pending" list, until this
> thread becomes considered "busy".
> # Now workers->atomic_worker_start is set, but
> check_pending_worker_creates() has not run yet (it is called only from
> worker_loop), so the same single thread is picked as "fallback".
>
> The problem is that this thread is still running the "caching_thread"
> function, scanning for EXTENT_ITEMs of the first block-group. This
> takes 3-4seconds for 1Gb block group.
>
> # Once caching_thread() exits, check_pending_worker_creates() is
> called, and creates the second thread, but it's too late, because all
> the 250 work items are already sitting in the first thread's "pending"
> list. So the second thread doesn't help at all.
>
> As a result, all block-group caching is performed by the same thread,
> which, due to one-by-one scanning of EXTENT_ITEMs, takes forever for
> this BTRFS.
>
> How this can be fixed?
> - can perhaps check_pending_worker_creates() be called out of
> worker_loop, e.g., by find_worker()? Instead of just setting
> workers->atomic_start_pending?
> - maybe for fs_info->caching_workers we don't have to create new
> workers asynchronously, so we can pass NULL for async_helper in
> btrfs_init_workers()? (probably we have to, just checking)
So I looked at this, and I'm pretty sure we have an async_helper just because of
copy+paste. "Hey I want a new async group, let me copy this other one and
change the name!" So yes let's just pass NULL here. In fact the only cases
that we should be using an async helper is for super critical areas, so I'm
pretty sure _most_ of the cases that specify an async helper don't need to.
Chris is this correct, or am I missing something? Thanks,
Josef
next prev parent reply other threads:[~2013-08-29 20:03 UTC|newest]
Thread overview: 4+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-08-26 14:16 btrfs:async-thread: atomic_start_pending=1 is set, but it's too late Alex Lyakas
2013-08-29 20:03 ` Josef Bacik [this message]
2013-08-29 20:08 ` Chris Mason
2013-08-29 22:19 ` Alex Lyakas
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20130829200306.GF10591@localhost.localdomain \
--to=jbacik@fusionio.com \
--cc=alex.btrfs@zadarastorage.com \
--cc=clmason@fusionio.com \
--cc=lev@zadarastorage.com \
--cc=linux-btrfs@vger.kernel.org \
--cc=shyam@zadarastorage.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).