linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH 01/15] userns: use new hashtable implementation
@ 2012-12-17 15:01 Sasha Levin
  2012-12-17 15:01 ` [PATCH 02/15] mm,ksm: " Sasha Levin
                   ` (13 more replies)
  0 siblings, 14 replies; 24+ messages in thread
From: Sasha Levin @ 2012-12-17 15:01 UTC (permalink / raw)
  To: Eric W. Biederman, Serge Hallyn, Sasha Levin, linux-kernel; +Cc: Sasha Levin

Switch to using the new hashtable implementation to store user structs.
This reduces the amount of generic unrelated code in kernel/user.c.

This patch depends on d9b482c ("hashtable: introduce a small and naive
hashtable") which was merged in v3.6.

Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
---
 kernel/user.c | 33 ++++++++++++---------------------
 1 file changed, 12 insertions(+), 21 deletions(-)

diff --git a/kernel/user.c b/kernel/user.c
index 33acb5e..5333a16 100644
--- a/kernel/user.c
+++ b/kernel/user.c
@@ -17,6 +17,7 @@
 #include <linux/export.h>
 #include <linux/user_namespace.h>
 #include <linux/proc_fs.h>
+#include <linux/hashtable.h>
 
 /*
  * userns count is 1 for root user, 1 for init_uts_ns,
@@ -62,13 +63,9 @@ EXPORT_SYMBOL_GPL(init_user_ns);
  */
 
 #define UIDHASH_BITS	(CONFIG_BASE_SMALL ? 3 : 7)
-#define UIDHASH_SZ	(1 << UIDHASH_BITS)
-#define UIDHASH_MASK		(UIDHASH_SZ - 1)
-#define __uidhashfn(uid)	(((uid >> UIDHASH_BITS) + uid) & UIDHASH_MASK)
-#define uidhashentry(uid)	(uidhash_table + __uidhashfn((__kuid_val(uid))))
 
 static struct kmem_cache *uid_cachep;
-struct hlist_head uidhash_table[UIDHASH_SZ];
+static DEFINE_HASHTABLE(uidhash_table, UIDHASH_BITS);
 
 /*
  * The uidhash_lock is mostly taken from process context, but it is
@@ -94,22 +91,22 @@ struct user_struct root_user = {
 /*
  * These routines must be called with the uidhash spinlock held!
  */
-static void uid_hash_insert(struct user_struct *up, struct hlist_head *hashent)
+static void uid_hash_insert(struct user_struct *up)
 {
-	hlist_add_head(&up->uidhash_node, hashent);
+	hash_add(uidhash_table, &up->uidhash_node, __kuid_val(up->uid));
 }
 
 static void uid_hash_remove(struct user_struct *up)
 {
-	hlist_del_init(&up->uidhash_node);
+	hash_del(&up->uidhash_node);
 }
 
-static struct user_struct *uid_hash_find(kuid_t uid, struct hlist_head *hashent)
+static struct user_struct *uid_hash_find(kuid_t uid)
 {
 	struct user_struct *user;
 	struct hlist_node *h;
 
-	hlist_for_each_entry(user, h, hashent, uidhash_node) {
+	hash_for_each_possible(uidhash_table, user, h, uidhash_node, __kuid_val(uid)) {
 		if (uid_eq(user->uid, uid)) {
 			atomic_inc(&user->__count);
 			return user;
@@ -145,7 +142,7 @@ struct user_struct *find_user(kuid_t uid)
 	unsigned long flags;
 
 	spin_lock_irqsave(&uidhash_lock, flags);
-	ret = uid_hash_find(uid, uidhashentry(uid));
+	ret = uid_hash_find(uid);
 	spin_unlock_irqrestore(&uidhash_lock, flags);
 	return ret;
 }
@@ -166,11 +163,10 @@ void free_uid(struct user_struct *up)
 
 struct user_struct *alloc_uid(kuid_t uid)
 {
-	struct hlist_head *hashent = uidhashentry(uid);
 	struct user_struct *up, *new;
 
 	spin_lock_irq(&uidhash_lock);
-	up = uid_hash_find(uid, hashent);
+	up = uid_hash_find(uid);
 	spin_unlock_irq(&uidhash_lock);
 
 	if (!up) {
@@ -186,13 +182,13 @@ struct user_struct *alloc_uid(kuid_t uid)
 		 * on adding the same user already..
 		 */
 		spin_lock_irq(&uidhash_lock);
-		up = uid_hash_find(uid, hashent);
+		up = uid_hash_find(uid);
 		if (up) {
 			key_put(new->uid_keyring);
 			key_put(new->session_keyring);
 			kmem_cache_free(uid_cachep, new);
 		} else {
-			uid_hash_insert(new, hashent);
+			uid_hash_insert(new);
 			up = new;
 		}
 		spin_unlock_irq(&uidhash_lock);
@@ -206,17 +202,12 @@ out_unlock:
 
 static int __init uid_cache_init(void)
 {
-	int n;
-
 	uid_cachep = kmem_cache_create("uid_cache", sizeof(struct user_struct),
 			0, SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL);
 
-	for(n = 0; n < UIDHASH_SZ; ++n)
-		INIT_HLIST_HEAD(uidhash_table + n);
-
 	/* Insert the root user immediately (init already runs as root) */
 	spin_lock_irq(&uidhash_lock);
-	uid_hash_insert(&root_user, uidhashentry(GLOBAL_ROOT_UID));
+	uid_hash_insert(&root_user);
 	spin_unlock_irq(&uidhash_lock);
 
 	return 0;
-- 
1.8.0


^ permalink raw reply related	[flat|nested] 24+ messages in thread

end of thread, other threads:[~2013-01-11 18:50 UTC | newest]

Thread overview: 24+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2012-12-17 15:01 [PATCH 01/15] userns: use new hashtable implementation Sasha Levin
2012-12-17 15:01 ` [PATCH 02/15] mm,ksm: " Sasha Levin
2012-12-18  3:25   ` Hugh Dickins
2012-12-17 15:01 ` [PATCH 03/15] workqueue: " Sasha Levin
2012-12-18 17:24   ` Tejun Heo
2012-12-17 15:01 ` [PATCH 04/15] mm/huge_memory: " Sasha Levin
2012-12-19 22:26   ` David Rientjes
2012-12-20  2:25     ` Sasha Levin
2012-12-20 20:28       ` David Rientjes
2012-12-20 20:30         ` Sasha Levin
2012-12-17 15:01 ` [PATCH 05/15] tracepoint: " Sasha Levin
2012-12-17 15:48   ` Steven Rostedt
2012-12-17 15:01 ` [PATCH 06/15] net,9p: " Sasha Levin
2012-12-17 15:01 ` [PATCH 07/15] block,elevator: " Sasha Levin
2012-12-17 15:18   ` Jens Axboe
2012-12-17 15:01 ` [PATCH 08/15] SUNRPC/cache: " Sasha Levin
2012-12-17 15:01 ` [PATCH 09/15] dlm: " Sasha Levin
2012-12-17 15:01 ` [PATCH 10/15] net,l2tp: " Sasha Levin
2012-12-17 15:01 ` [PATCH 11/15] dm: " Sasha Levin
2013-01-11 18:49   ` Sasha Levin
2012-12-17 15:01 ` [PATCH 12/15] lockd: " Sasha Levin
2012-12-17 15:01 ` [PATCH 13/15] net,rds: " Sasha Levin
2012-12-17 15:01 ` [PATCH 14/15] openvswitch: " Sasha Levin
2012-12-17 15:01 ` [PATCH 15/15] tracing output: " Sasha Levin

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).