struct rhashtable *ht, const struct bucket_table *tbl,
const void *key, const struct rhashtable_params params)
{
- unsigned hash;
+ unsigned int hash;
/* params must be equal to ht->p if it isn't constant. */
if (!__builtin_constant_p(params.key_len))
hash = ht->p.hashfn(key, ht->key_len, tbl->hash_rnd);
else if (params.key_len) {
- unsigned key_len = params.key_len;
+ unsigned int key_len = params.key_len;
if (params.hashfn)
hash = params.hashfn(key, key_len, tbl->hash_rnd);
hash = jhash2(key, key_len / sizeof(u32),
tbl->hash_rnd);
} else {
- unsigned key_len = ht->p.key_len;
+ unsigned int key_len = ht->p.key_len;
if (params.hashfn)
hash = params.hashfn(key, key_len, tbl->hash_rnd);
};
const struct bucket_table *tbl;
struct rhash_head *he;
- unsigned hash;
+ unsigned int hash;
rcu_read_lock();
struct bucket_table *tbl, *new_tbl;
struct rhash_head *head;
spinlock_t *lock;
- unsigned elasticity;
- unsigned hash;
+ unsigned int elasticity;
+ unsigned int hash;
int err;
restart:
if (unlikely(rht_grow_above_100(ht, tbl))) {
slow_path:
spin_unlock_bh(lock);
- rcu_read_unlock();
err = rhashtable_insert_rehash(ht);
+ rcu_read_unlock();
if (err)
return err;
struct rhash_head __rcu **pprev;
struct rhash_head *he;
spinlock_t * lock;
- unsigned hash;
+ unsigned int hash;
int err = -ENOENT;
hash = rht_head_hashfn(ht, tbl, obj, params);