Commit 6c7005f6 authored by David S. Miller's avatar David S. Miller

Merge branch 'rhashtable-cleanups'

Herbert Xu says:

====================
rhashtable hash cleanups

This is a rebase on top of the nested lock annotation fix.

Nothing to see here, just a bunch of simple clean-ups before
I move onto something more substantial (hopefully).
====================
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parents 6b9f53bc ec9f71c5
...@@ -63,34 +63,25 @@ static void *rht_obj(const struct rhashtable *ht, const struct rhash_head *he) ...@@ -63,34 +63,25 @@ static void *rht_obj(const struct rhashtable *ht, const struct rhash_head *he)
static u32 rht_bucket_index(const struct bucket_table *tbl, u32 hash) static u32 rht_bucket_index(const struct bucket_table *tbl, u32 hash)
{ {
return hash & (tbl->size - 1); return (hash >> HASH_RESERVED_SPACE) & (tbl->size - 1);
}
static u32 obj_raw_hashfn(struct rhashtable *ht,
const struct bucket_table *tbl, const void *ptr)
{
u32 hash;
if (unlikely(!ht->p.key_len))
hash = ht->p.obj_hashfn(ptr, tbl->hash_rnd);
else
hash = ht->p.hashfn(ptr + ht->p.key_offset, ht->p.key_len,
tbl->hash_rnd);
return hash >> HASH_RESERVED_SPACE;
} }
static u32 key_hashfn(struct rhashtable *ht, const struct bucket_table *tbl, static u32 key_hashfn(struct rhashtable *ht, const struct bucket_table *tbl,
const void *key, u32 len) const void *key)
{ {
return ht->p.hashfn(key, len, tbl->hash_rnd) >> HASH_RESERVED_SPACE; return rht_bucket_index(tbl, ht->p.hashfn(key, ht->p.key_len,
tbl->hash_rnd));
} }
static u32 head_hashfn(struct rhashtable *ht, static u32 head_hashfn(struct rhashtable *ht,
const struct bucket_table *tbl, const struct bucket_table *tbl,
const struct rhash_head *he) const struct rhash_head *he)
{ {
return rht_bucket_index(tbl, obj_raw_hashfn(ht, tbl, rht_obj(ht, he))); const char *ptr = rht_obj(ht, he);
return likely(ht->p.key_len) ?
key_hashfn(ht, tbl, ptr + ht->p.key_offset) :
rht_bucket_index(tbl, ht->p.obj_hashfn(ptr, tbl->hash_rnd));
} }
#ifdef CONFIG_PROVE_LOCKING #ifdef CONFIG_PROVE_LOCKING
...@@ -402,7 +393,7 @@ static bool __rhashtable_insert(struct rhashtable *ht, struct rhash_head *obj, ...@@ -402,7 +393,7 @@ static bool __rhashtable_insert(struct rhashtable *ht, struct rhash_head *obj,
rcu_read_lock(); rcu_read_lock();
old_tbl = rht_dereference_rcu(ht->tbl, ht); old_tbl = rht_dereference_rcu(ht->tbl, ht);
hash = obj_raw_hashfn(ht, old_tbl, rht_obj(ht, obj)); hash = head_hashfn(ht, old_tbl, obj);
spin_lock_bh(bucket_lock(old_tbl, hash)); spin_lock_bh(bucket_lock(old_tbl, hash));
...@@ -414,7 +405,7 @@ static bool __rhashtable_insert(struct rhashtable *ht, struct rhash_head *obj, ...@@ -414,7 +405,7 @@ static bool __rhashtable_insert(struct rhashtable *ht, struct rhash_head *obj,
*/ */
tbl = rht_dereference_rcu(ht->future_tbl, ht); tbl = rht_dereference_rcu(ht->future_tbl, ht);
if (tbl != old_tbl) { if (tbl != old_tbl) {
hash = obj_raw_hashfn(ht, tbl, rht_obj(ht, obj)); hash = head_hashfn(ht, tbl, obj);
spin_lock_nested(bucket_lock(tbl, hash), RHT_LOCK_NESTED); spin_lock_nested(bucket_lock(tbl, hash), RHT_LOCK_NESTED);
} }
...@@ -427,7 +418,6 @@ static bool __rhashtable_insert(struct rhashtable *ht, struct rhash_head *obj, ...@@ -427,7 +418,6 @@ static bool __rhashtable_insert(struct rhashtable *ht, struct rhash_head *obj,
no_resize_running = tbl == old_tbl; no_resize_running = tbl == old_tbl;
hash = rht_bucket_index(tbl, hash);
head = rht_dereference_bucket(tbl->buckets[hash], tbl, hash); head = rht_dereference_bucket(tbl->buckets[hash], tbl, hash);
if (rht_is_a_nulls(head)) if (rht_is_a_nulls(head))
...@@ -443,11 +433,11 @@ static bool __rhashtable_insert(struct rhashtable *ht, struct rhash_head *obj, ...@@ -443,11 +433,11 @@ static bool __rhashtable_insert(struct rhashtable *ht, struct rhash_head *obj,
exit: exit:
if (tbl != old_tbl) { if (tbl != old_tbl) {
hash = obj_raw_hashfn(ht, tbl, rht_obj(ht, obj)); hash = head_hashfn(ht, tbl, obj);
spin_unlock(bucket_lock(tbl, hash)); spin_unlock(bucket_lock(tbl, hash));
} }
hash = obj_raw_hashfn(ht, old_tbl, rht_obj(ht, obj)); hash = head_hashfn(ht, old_tbl, obj);
spin_unlock_bh(bucket_lock(old_tbl, hash)); spin_unlock_bh(bucket_lock(old_tbl, hash));
rcu_read_unlock(); rcu_read_unlock();
...@@ -486,9 +476,8 @@ static bool __rhashtable_remove(struct rhashtable *ht, ...@@ -486,9 +476,8 @@ static bool __rhashtable_remove(struct rhashtable *ht,
unsigned hash; unsigned hash;
bool ret = false; bool ret = false;
hash = obj_raw_hashfn(ht, tbl, rht_obj(ht, obj)); hash = head_hashfn(ht, tbl, obj);
lock = bucket_lock(tbl, hash); lock = bucket_lock(tbl, hash);
hash = rht_bucket_index(tbl, hash);
spin_lock_bh(lock); spin_lock_bh(lock);
...@@ -620,9 +609,9 @@ void *rhashtable_lookup_compare(struct rhashtable *ht, const void *key, ...@@ -620,9 +609,9 @@ void *rhashtable_lookup_compare(struct rhashtable *ht, const void *key,
rcu_read_lock(); rcu_read_lock();
tbl = rht_dereference_rcu(ht->tbl, ht); tbl = rht_dereference_rcu(ht->tbl, ht);
hash = key_hashfn(ht, tbl, key, ht->p.key_len); hash = key_hashfn(ht, tbl, key);
restart: restart:
rht_for_each_rcu(he, tbl, rht_bucket_index(tbl, hash)) { rht_for_each_rcu(he, tbl, hash) {
if (!compare(rht_obj(ht, he), arg)) if (!compare(rht_obj(ht, he), arg))
continue; continue;
rcu_read_unlock(); rcu_read_unlock();
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment