linux-f2fs-devel.lists.sourceforge.net archive mirror
 help / color / mirror / Atom feed
From: Kinglong Mee <kinglongmee@gmail.com>
To: Chao Yu <yuchao0@huawei.com>
Cc: jaegeuk@kernel.org, chao@kernel.org,
	linux-kernel@vger.kernel.org,
	linux-f2fs-devel@lists.sourceforge.net
Subject: Re: [PATCH RFC] f2fs: combine nat_bits and free_nid_bitmap cache
Date: Thu, 2 Mar 2017 12:29:23 +0800	[thread overview]
Message-ID: <481d03f6-cbd4-aa80-3689-5736c686e11b@gmail.com> (raw)
In-Reply-To: <b6ad726f-acd4-fa46-0c75-568626a54a6d@huawei.com>

On 3/2/2017 09:35, Chao Yu wrote:
> On 2017/3/1 21:09, Kinglong Mee wrote:
>> On 3/1/2017 17:10, Chao Yu wrote:
>>> Both nat_bits cache and free_nid_bitmap cache provide same functionality
>>> as a intermediate cache between free nid cache and disk, but with
>>> different granularity of indicating free nid range, and different
>>> persistence policy. nat_bits cache provides better persistence ability,
>>> and free_nid_bitmap provides better granularity.
>>>
>>> In this patch we combine advantage of both caches, so finally policy of
>>> the intermediate cache would be:
>>> - init: load free nid status from nat_bits into free_nid_bitmap
>>> - lookup: scan free_nid_bitmap before load NAT blocks
>>
>> Why not scan the full_nat_bits/empty_nat_bits before load NAT blocks here?
>> If after an objects shrinker, the cached free nid will be empty quickly.
> 
> Since after this patch, all nids status (free or used) of
> full_nat_bits/empty_nat_bits will be loaded into free_nid_bitmap, so we can just
> check free_nid_bitmap instead of both cache before loading NAT blocks.

Yes, you are right.
I forgot f2fs also updates update_free_nid_bitmap in __flush_nat_entry_set.

thanks,
Kinglong Mee

>>> - update: update free_nid_bitmap in real-time
>>> - persistence: udpate and persist nat_bits in checkpoint
>>>
>>> Signed-off-by: Chao Yu <yuchao0@huawei.com>
>>> ---
>>>  fs/f2fs/node.c | 109 +++++++++++++++++++++------------------------------------
>>>  1 file changed, 39 insertions(+), 70 deletions(-)
>>>
>>> diff --git a/fs/f2fs/node.c b/fs/f2fs/node.c
>>> index 1a759d45b7e4..6c027b6833f4 100644
>>> --- a/fs/f2fs/node.c
>>> +++ b/fs/f2fs/node.c
>>> @@ -338,9 +338,6 @@ static void set_node_addr(struct f2fs_sb_info *sbi, struct node_info *ni,
>>>  		set_nat_flag(e, IS_CHECKPOINTED, false);
>>>  	__set_nat_cache_dirty(nm_i, e);
>>>  
>>> -	if (enabled_nat_bits(sbi, NULL) && new_blkaddr == NEW_ADDR)
>>> -		clear_bit_le(NAT_BLOCK_OFFSET(ni->nid), nm_i->empty_nat_bits);
>>> -
>>>  	/* update fsync_mark if its inode nat entry is still alive */
>>>  	if (ni->nid != ni->ino)
>>>  		e = __lookup_nat_cache(nm_i, ni->ino);
>>> @@ -1920,58 +1917,6 @@ static void scan_free_nid_bits(struct f2fs_sb_info *sbi)
>>>  	up_read(&nm_i->nat_tree_lock);
>>>  }
>>>  
>>> -static int scan_nat_bits(struct f2fs_sb_info *sbi)
>>> -{
>>> -	struct f2fs_nm_info *nm_i = NM_I(sbi);
>>> -	struct page *page;
>>> -	unsigned int i = 0;
>>> -	nid_t nid;
>>> -
>>> -	if (!enabled_nat_bits(sbi, NULL))
>>> -		return -EAGAIN;
>>> -
>>> -	down_read(&nm_i->nat_tree_lock);
>>> -check_empty:
>>> -	i = find_next_bit_le(nm_i->empty_nat_bits, nm_i->nat_blocks, i);
>>> -	if (i >= nm_i->nat_blocks) {
>>> -		i = 0;
>>> -		goto check_partial;
>>> -	}
>>> -
>>> -	for (nid = i * NAT_ENTRY_PER_BLOCK; nid < (i + 1) * NAT_ENTRY_PER_BLOCK;
>>> -									nid++) {
>>> -		if (unlikely(nid >= nm_i->max_nid))
>>> -			break;
>>> -		add_free_nid(sbi, nid, true);
>>> -	}
>>> -
>>> -	if (nm_i->nid_cnt[FREE_NID_LIST] >= MAX_FREE_NIDS)
>>> -		goto out;
>>> -	i++;
>>> -	goto check_empty;
>>> -
>>> -check_partial:
>>> -	i = find_next_zero_bit_le(nm_i->full_nat_bits, nm_i->nat_blocks, i);
>>> -	if (i >= nm_i->nat_blocks) {
>>> -		disable_nat_bits(sbi, true);
>>> -		up_read(&nm_i->nat_tree_lock);
>>> -		return -EINVAL;
>>> -	}
>>> -
>>> -	nid = i * NAT_ENTRY_PER_BLOCK;
>>> -	page = get_current_nat_page(sbi, nid);
>>> -	scan_nat_page(sbi, page, nid);
>>> -	f2fs_put_page(page, 1);
>>> -
>>> -	if (nm_i->nid_cnt[FREE_NID_LIST] < MAX_FREE_NIDS) {
>>> -		i++;
>>> -		goto check_partial;
>>> -	}
>>> -out:
>>> -	up_read(&nm_i->nat_tree_lock);
>>> -	return 0;
>>> -}
>>> -
>>>  static void __build_free_nids(struct f2fs_sb_info *sbi, bool sync, bool mount)
>>>  {
>>>  	struct f2fs_nm_info *nm_i = NM_I(sbi);
>>> @@ -1993,21 +1938,6 @@ static void __build_free_nids(struct f2fs_sb_info *sbi, bool sync, bool mount)
>>>  
>>>  		if (nm_i->nid_cnt[FREE_NID_LIST])
>>>  			return;
>>> -
>>> -		/* try to find free nids with nat_bits */
>>> -		if (!scan_nat_bits(sbi) && nm_i->nid_cnt[FREE_NID_LIST])
>>> -			return;
>>> -	}
>>> -
>>> -	/* find next valid candidate */
>>> -	if (enabled_nat_bits(sbi, NULL)) {
>>> -		int idx = find_next_zero_bit_le(nm_i->full_nat_bits,
>>> -					nm_i->nat_blocks, 0);
>>> -
>>> -		if (idx >= nm_i->nat_blocks)
>>> -			set_sbi_flag(sbi, SBI_NEED_FSCK);
>>> -		else
>>> -			nid = idx * NAT_ENTRY_PER_BLOCK;
>>>  	}
>>>  
>>>  	/* readahead nat pages to be scanned */
>>> @@ -2590,6 +2520,41 @@ static int __get_nat_bitmaps(struct f2fs_sb_info *sbi)
>>>  	return 0;
>>>  }
>>>  
>>> +inline void load_free_nid_bitmap(struct f2fs_sb_info *sbi)
>>> +{
>>> +	struct f2fs_nm_info *nm_i = NM_I(sbi);
>>> +	unsigned int i = 0;
>>> +	nid_t nid, last_nid;
>>> +
>>> +	for (i = 0; i < nm_i->nat_blocks; i++) {
>>> +		i = find_next_bit_le(nm_i->empty_nat_bits, nm_i->nat_blocks, i);
>>> +		if (i >= nm_i->nat_blocks)
>>> +			break;
>>> +
>>> +		set_bit_le(i, nm_i->nat_block_bitmap);
>>> +
>>> +		nid = i * NAT_ENTRY_PER_BLOCK;
>>> +		last_nid = (i + 1) * NAT_ENTRY_PER_BLOCK;
>>> +
>>> +		for (; nid < last_nid; nid++)
>>> +			update_free_nid_bitmap(sbi, nid, true, true);
>>> +	}
>>> +
>>> +	for (i = 0; i < nm_i->nat_blocks; i++) {
>>> +		i = find_next_bit_le(nm_i->full_nat_bits, nm_i->nat_blocks, i);
>>> +		if (i >= nm_i->nat_blocks)
>>> +			break;
>>> +
>>> +		set_bit_le(i, nm_i->nat_block_bitmap);
>>> +
>>> +		nid = i * NAT_ENTRY_PER_BLOCK;
>>> +		last_nid = (i + 1) * NAT_ENTRY_PER_BLOCK;
>>> +
>>> +		for (; nid < last_nid; nid++)
>>> +			update_free_nid_bitmap(sbi, nid, false, true);
>>> +	}
>>> +}
>>> +
>>>  static int init_node_manager(struct f2fs_sb_info *sbi)
>>>  {
>>>  	struct f2fs_super_block *sb_raw = F2FS_RAW_SUPER(sbi);
>>> @@ -2672,6 +2637,10 @@ int init_free_nid_cache(struct f2fs_sb_info *sbi)
>>>  
>>>  	spin_lock_init(&nm_i->free_nid_lock);
>>>  
>>> +	/* load free nid status from nat_bits table */
>>> +	if (enabled_nat_bits(sbi, NULL))
>>> +		load_free_nid_bitmap(sbi);
>>> +
>>>  	return 0;
>>>  }
>>>  
>>>
>>
>> .
>>
> 
> 

------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, SlashDot.org! http://sdm.link/slashdot

  reply	other threads:[~2017-03-02  4:29 UTC|newest]

Thread overview: 6+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2017-03-01  9:10 [PATCH RFC] f2fs: combine nat_bits and free_nid_bitmap cache Chao Yu
2017-03-01 13:09 ` Kinglong Mee
2017-03-02  1:35   ` Chao Yu
2017-03-02  4:29     ` Kinglong Mee [this message]
2017-03-02 19:10 ` Jaegeuk Kim
2017-03-06 10:31   ` Chao Yu

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=481d03f6-cbd4-aa80-3689-5736c686e11b@gmail.com \
    --to=kinglongmee@gmail.com \
    --cc=chao@kernel.org \
    --cc=jaegeuk@kernel.org \
    --cc=linux-f2fs-devel@lists.sourceforge.net \
    --cc=linux-kernel@vger.kernel.org \
    --cc=yuchao0@huawei.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).