HASH_SIZE
for (i = 0; i < HASH_SIZE(src->fences); ++i)
jhash(_key, _key_len, 0) % HASH_SIZE(*__tbl2); \
return &table->hashtable[hash & (HASH_SIZE(table->hashtable) - 1)];
(HASH_SIZE(table->hashtable) - 1)];
&vclock_hash[vclock->clock->index % HASH_SIZE(vclock_hash)]);
unsigned int hash = vclock_index % HASH_SIZE(vclock_hash);
for (i = 0; i < HASH_SIZE(afs_permits_cache); i++)
for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
#define HASH_BITS(name) ilog2(HASH_SIZE(name))
#define hash_init(hashtable) __hash_init(hashtable, HASH_SIZE(hashtable))
#define hash_empty(hashtable) __hash_empty(hashtable, HASH_SIZE(hashtable))
for (i = 0; i < HASH_SIZE; i++)
static struct list_head chunk_hash_heads[HASH_SIZE];
return chunk_hash_heads + n % HASH_SIZE;
#define HASH_FN_MASK (HASH_SIZE - 1)
int limit = min(HASH_SIZE, (index.dev_addr >> HASH_FN_SHIFT) + 1);
for (idx = 0; idx < HASH_SIZE; idx++) {
for (idx = 0; idx < HASH_SIZE; idx++) {
for (i = 0; i < HASH_SIZE; ++i) {
for (i = 0; i < HASH_SIZE; ++i) {
static struct hash_bucket dma_entry_hash[HASH_SIZE];
u8 actual_hash[HASH_SIZE];
test, actual_hash, hash_testvecs[i].digest, HASH_SIZE,
u8 hash[HASH_SIZE];
HASH_UPDATE(&ctx, hash, HASH_SIZE);
KUNIT_ASSERT_MEMEQ(test, hash, hash_testvec_consolidated, HASH_SIZE);
u8 hash1[HASH_SIZE];
u8 hash2[HASH_SIZE];
test, hash1, hash2, HASH_SIZE,
u8 hash[HASH_SIZE];
HASH(test_buf, len, buf_end - HASH_SIZE);
HASH_FINAL(&ctx, buf_end - HASH_SIZE);
const size_t max_tested_len = TEST_BUF_LEN - HASH_SIZE;
u8 hash[HASH_SIZE];
size_t offset = HASH_SIZE + rand_offset(max_tested_len - len);
&test_buf[offset + len - HASH_SIZE];
test, hash, ovl_hash, HASH_SIZE,
test, hash, ovl_hash, HASH_SIZE,
test, hash, ovl_hash, HASH_SIZE,
u8 hash1[128 + HASH_SIZE];
u8 hash2[128 + HASH_SIZE];
test, &hash1[hash_offs1], &hash2[hash_offs2], HASH_SIZE,
u8 expected_hashes[IRQ_TEST_NUM_BUFFERS][HASH_SIZE];
u8 actual_hash[HASH_SIZE];
return memcmp(actual_hash, state->expected_hashes[i], HASH_SIZE) == 0;
u8 expected_hash[HASH_SIZE];
u8 actual_hash[HASH_SIZE];
if (memcmp(actual_hash, state->expected_hash, HASH_SIZE) != 0)
u8 mac[HASH_SIZE];
u8 mac2[HASH_SIZE];
HMAC_UPDATE(&ctx, mac, HASH_SIZE);
test, mac, mac2, HASH_SIZE,
KUNIT_EXPECT_MEMEQ_MSG(test, mac, hmac_testvec_consolidated, HASH_SIZE,
u8 hash[HASH_SIZE];
&napi_hash[napi->napi_id % HASH_SIZE(napi_hash)]);
unsigned int hash = napi_id % HASH_SIZE(napi_hash);
for (i = 0; i < HASH_SIZE(rxnet->peer_hash); i++) {
if (bucket >= HASH_SIZE(rxnet->peer_hash)) {
if (bucket >= HASH_SIZE(rxnet->peer_hash)) {
nsyms, HASH_SIZE(symbol_hashtable),
(double)nsyms / HASH_SIZE(symbol_hashtable));
#define hash_head(table, key) (&(table)[(key) % HASH_SIZE(table)])
#define hash_init(table) __hash_init(table, HASH_SIZE(table))
for (int _bkt = 0; _bkt < HASH_SIZE(table); _bkt++) \
for (int _bkt = 0; _bkt < HASH_SIZE(table); _bkt++) \
if (derived_buf_len < HASH_SIZE)
derived_buf_len = HASH_SIZE;
u8 derived_key[HASH_SIZE];
dump_hmac(NULL, digest, HASH_SIZE);
u8 derived_key[HASH_SIZE];
u8 digest[HASH_SIZE];
HASH_SIZE);
dump_hmac("calc", digest, HASH_SIZE);
+ HASH_SIZE + 1);
datablob_len + HASH_SIZE + 1, GFP_KERNEL);
u8 derived_key[HASH_SIZE];
asciilen = (ivsize + 1 + encrypted_datalen + HASH_SIZE) * 2;
HASH_SIZE);
char derived_key[HASH_SIZE];
+ (HASH_SIZE * 2);
for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
#define HASH_BITS(name) ilog2(HASH_SIZE(name))
#define hash_init(hashtable) __hash_init(hashtable, HASH_SIZE(hashtable))
#define hash_empty(hashtable) __hash_empty(hashtable, HASH_SIZE(hashtable))
#define HASH_MASK (HASH_SIZE - 1)
static unsigned long nr_pages[HASH_SIZE];
static uint64_t page_flags[HASH_SIZE];