From: Zou Mingzhe <mingzhe.zou@easystack.cn>
To: linux-bcache@vger.kernel.org
Cc: zoumingzhe@qq.com
Subject: Re: [PATCH v3] bcache: dynamic incremental gc
Date: Fri, 20 May 2022 16:22:38 +0800 [thread overview]
Message-ID: <112eaaf7-05fd-3b4f-0190-958d0c85fa1f@easystack.cn> (raw)
In-Reply-To: <ecce38e7-8ba0-5fbf-61a6-2dfc21c7793d@suse.de>
在 2022/5/12 21:41, Coly Li 写道:
> On 5/11/22 3:39 PM, mingzhe.zou@easystack.cn wrote:
>> From: ZouMingzhe <mingzhe.zou@easystack.cn>
>>
>> Currently, GC wants no more than 100 times, with at least
>> 100 nodes each time, the maximum number of nodes each time
>> is not limited.
>>
>> ```
>> static size_t btree_gc_min_nodes(struct cache_set *c)
>> {
>> ......
>> min_nodes = c->gc_stats.nodes / MAX_GC_TIMES;
>> if (min_nodes < MIN_GC_NODES)
>> min_nodes = MIN_GC_NODES;
>>
>> return min_nodes;
>> }
>> ```
>>
>> According to our test data, when nvme is used as the cache,
>> it takes about 1ms for GC to handle each node (block 4k and
>> bucket 512k). This means that the latency during GC is at
>> least 100ms. During GC, IO performance would be reduced by
>> half or more.
>>
>> I want to optimize the IOPS and latency under high pressure.
>> This patch hold the inflight peak. When IO depth up to maximum,
>> GC only process very few(10) nodes, then sleep immediately and
>> handle these requests.
>>
>> bch_bucket_alloc() maybe wait for bch_allocator_thread() to
>> wake up, and and bch_allocator_thread() needs to wait for gc
>> to complete, in which case gc needs to end quickly. So, add
>> bucket_alloc_inflight to cache_set in v3.
>>
>> ```
>> long bch_bucket_alloc(struct cache *ca, unsigned int reserve, bool wait)
>> {
>> ......
>> do {
>> prepare_to_wait(&ca->set->bucket_wait, &w,
>> TASK_UNINTERRUPTIBLE);
>>
>> mutex_unlock(&ca->set->bucket_lock);
>> schedule();
>> mutex_lock(&ca->set->bucket_lock);
>> } while (!fifo_pop(&ca->free[RESERVE_NONE], r) &&
>> !fifo_pop(&ca->free[reserve], r));
>> ......
>> }
>>
>> static int bch_allocator_thread(void *arg)
>> {
>> ......
>> allocator_wait(ca, bch_allocator_push(ca, bucket));
>> wake_up(&ca->set->btree_cache_wait);
>> wake_up(&ca->set->bucket_wait);
>> ......
>> }
>>
>> static void bch_btree_gc(struct cache_set *c)
>> {
>> ......
>> bch_btree_gc_finish(c);
>> wake_up_allocators(c);
>> ......
>> }
>> ```
>>
>> Apply this patch, each GC maybe only process very few nodes,
>> GC would last a long time if sleep 100ms each time. So, the
>> sleep time should be calculated dynamically based on gc_cost.
>>
>> At the same time, I added some cost statistics in gc_stat,
>> hoping to provide useful information for future work.
>
>
> Hi Mingzhe,
>
> From the first glance, I feel this change may delay the small GC
> period, and finally result a large GC period, which is not expected.
>
> But it is possible that my feeling is incorrect. Do you have detailed
> performance number about both I/O latency and GC period, then I can
> have more understanding for this effort.
>
> BTW, I will add this patch to my testing set and experience myself.
>
>
> Thanks.
>
>
> Coly Li
>
>
Hi Coly,
First, your feeling is right. Then, I have some performance number abort
before and after this patch.
Since the mailing list does not accept attachments, I put them on the gist.
Please visit the page for details:
https://gist.github.com/zoumingzhe/69a353e7c6fffe43142c2f42b94a67b5
mingzhe
>
>
> [snipped]
>
>
>
next prev parent reply other threads:[~2022-05-20 8:22 UTC|newest]
Thread overview: 8+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-05-11 7:39 [PATCH v3] bcache: dynamic incremental gc mingzhe.zou
2022-05-12 13:41 ` Coly Li
2022-05-20 8:22 ` Zou Mingzhe [this message]
2022-05-20 18:24 ` Eric Wheeler
2022-05-23 2:52 ` Zou Mingzhe
2022-05-23 12:54 ` Zou Mingzhe
2022-05-23 17:55 ` Eric Wheeler
2022-05-24 2:47 ` Zou Mingzhe
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=112eaaf7-05fd-3b4f-0190-958d0c85fa1f@easystack.cn \
--to=mingzhe.zou@easystack.cn \
--cc=linux-bcache@vger.kernel.org \
--cc=zoumingzhe@qq.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox