From: Steve Dickson <SteveD@redhat.com>
To: Peter Staubach <staubach@redhat.com>
Cc: nfs@lists.sourceforge.net, linux-fsdevel@vger.kernel.org
Subject: Re: [NFS] Re: [PATCH][RFC] NFS: Improving the access cache
Date: Tue, 02 May 2006 10:38:28 -0400 [thread overview]
Message-ID: <44576EE4.4010704@RedHat.com> (raw)
In-Reply-To: <445763CF.5040506@redhat.com>
Peter Staubach wrote:
>> Basically we would maintain one global hlist (i.e. link list) that
>> would contain all of the cached entries; then each nfs_inode would
>> have its own LRU hlist that would contain entries that are associated
>> with that nfs_inode. So each entry would be on two lists, the
>> global hlist and hlist in the nfs_inode.
>>
>
> How are these lists used?
The inode hlist will be used to search and purge...
>
> I would suggest that a global set of hash queues would work better than
> a linked list and that these hash queues by used to find the cache entry
> for any particular user. Finding the entry for a particular (user,inode)
> needs to be fast and linearly searching a linked list is slow. Linear
> searching needs to be avoided. Comparing the fewest number of entries
> possible will result in the best performance because the comparisons
> need to take into account the entire user identification, including
> the groups list.
I guess we could have the VFS shrinker to purge a hash table just
as well as a link list... although a hash table will have an
small memory cost...
> The list in the inode seems useful, but only for purges. Searching via
> this list will be very slow once the list grows beyond a few entries.
> Purging needs to be fast because purging the access cache entries for a
> particular file will need to happen whenever the ctime on the file changes.
> This list can be used to make it easy to find the correct entries in the
> global access cache.
Seems reasonable assuming we use a hash table...
>
>> We would govern memory consumption by only allowing 30 entries
>> on any one hlist in the nfs_inode and by registering the globe
>> hlist with the VFS shrinker which will cause the list to be prune
>> when memory is needed. So this means, when the 31st entry was added
>> to the hlist in the nfs_inode, the least recently used entry would
>> be removed.
>>
>
> Why is there a limit at all and why is 30 the right number? This
> seems small and rather arbitrary. If there is some way to trigger
> memory reclaiming, then letting the list grow as appropriate seems
> like a good thing to do.
Well the vfs mechanism will be the trigger... so your saying we
should just let the purge hlist lists in the nfs_inode grow
untethered? How about read-only filesystems where the ctime
will not change... I would think we might want some type of
high water mark for that case, true?
>
> Making sure that you are one of the original 30 users accessing the
> file in order to get reasonable performance seems tricky to me. :-)
>
>> Locking might be a bit tricky, but do able... To make this scalable,
>> I would think we would need global read/write spin_lock. The read_lock()
>> would be taken when the hlist in the inode was searched and the
>> write_lock() would taken when the hlist in the inode was changed
>> and when the global list was prune.
>>
>
> Sorry, read/write spin lock? I thought that spin locks were exclusive,
> either the lock was held or the process spins waiting to acquire it.
See the rwlock_t lock type in asm/spinlock.h.. That's the one
I was planning on using...
steved.
next prev parent reply other threads:[~2006-05-02 14:38 UTC|newest]
Thread overview: 30+ messages / expand[flat|nested] mbox.gz Atom feed top
2006-04-26 1:14 [PATCH][RFC] NFS: Improving the access cache Steve Dickson
2006-04-26 1:31 ` Matthew Wilcox
2006-04-26 4:55 ` Neil Brown
2006-04-26 14:51 ` Steve Dickson
2006-04-26 22:32 ` Neil Brown
2006-05-02 9:49 ` Steve Dickson
2006-05-02 13:51 ` [NFS] " Peter Staubach
2006-05-02 14:38 ` Steve Dickson [this message]
2006-05-02 14:51 ` Peter Staubach
2006-05-02 15:26 ` [NFS] " Ian Kent
2006-05-03 4:42 ` Chuck Lever
2006-05-05 14:07 ` Steve Dickson
2006-05-05 14:53 ` Peter Staubach
2006-05-05 14:59 ` Peter Staubach
2006-05-06 14:35 ` [NFS] " Steve Dickson
2006-05-08 14:07 ` Peter Staubach
2006-05-08 17:09 ` Trond Myklebust
2006-05-08 17:20 ` Peter Staubach
2006-05-08 2:44 ` Neil Brown
2006-05-08 3:23 ` Chuck Lever
2006-05-08 3:28 ` Neil Brown
2006-04-26 13:03 ` Trond Myklebust
2006-04-26 13:14 ` Peter Staubach
2006-04-26 14:01 ` Trond Myklebust
2006-04-26 14:15 ` Peter Staubach
2006-04-26 15:44 ` Trond Myklebust
2006-04-26 17:01 ` Peter Staubach
2006-04-26 15:03 ` Steve Dickson
2006-04-26 13:17 ` [NFS] " Chuck Lever
2006-04-26 14:19 ` Steve Dickson
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=44576EE4.4010704@RedHat.com \
--to=steved@redhat.com \
--cc=linux-fsdevel@vger.kernel.org \
--cc=nfs@lists.sourceforge.net \
--cc=staubach@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).