Symbol: hash
arch/arm64/net/bpf_jit_comp.c
186
static inline void emit_kcfi(u32 hash, struct jit_ctx *ctx)
arch/arm64/net/bpf_jit_comp.c
189
emit_u32_data(hash, ctx);
arch/loongarch/kernel/relocate.c
101
if (size < diff + sizeof(hash))
arch/loongarch/kernel/relocate.c
102
return hash;
arch/loongarch/kernel/relocate.c
104
size = ALIGN_DOWN(size - diff, sizeof(hash));
arch/loongarch/kernel/relocate.c
106
for (i = 0; i < size / sizeof(hash); i++) {
arch/loongarch/kernel/relocate.c
108
hash = (hash << ((sizeof(hash) * 8) - 7)) | (hash >> 7);
arch/loongarch/kernel/relocate.c
109
hash ^= ptr[i];
arch/loongarch/kernel/relocate.c
112
return hash;
arch/loongarch/kernel/relocate.c
117
unsigned long hash = 0;
arch/loongarch/kernel/relocate.c
121
hash = rotate_xor(hash, linux_banner, strlen(linux_banner));
arch/loongarch/kernel/relocate.c
124
hash = rotate_xor(hash, &entropy, sizeof(entropy));
arch/loongarch/kernel/relocate.c
126
return hash;
arch/loongarch/kernel/relocate.c
94
static inline __init unsigned long rotate_xor(unsigned long hash,
arch/loongarch/kernel/relocate.c
98
const typeof(hash) *ptr = PTR_ALIGN(area, sizeof(hash));
arch/mips/kernel/relocate.c
194
static inline __init unsigned long rotate_xor(unsigned long hash,
arch/mips/kernel/relocate.c
197
const typeof(hash) *ptr = PTR_ALIGN(area, sizeof(hash));
arch/mips/kernel/relocate.c
201
if (unlikely(size < diff + sizeof(hash)))
arch/mips/kernel/relocate.c
202
return hash;
arch/mips/kernel/relocate.c
204
size = ALIGN_DOWN(size - diff, sizeof(hash));
arch/mips/kernel/relocate.c
206
for (i = 0; i < size / sizeof(hash); i++) {
arch/mips/kernel/relocate.c
208
hash = (hash << ((sizeof(hash) * 8) - 7)) | (hash >> 7);
arch/mips/kernel/relocate.c
209
hash ^= ptr[i];
arch/mips/kernel/relocate.c
212
return hash;
arch/mips/kernel/relocate.c
218
unsigned long hash = 0;
arch/mips/kernel/relocate.c
221
hash = rotate_xor(hash, linux_banner, strlen(linux_banner));
arch/mips/kernel/relocate.c
224
hash = rotate_xor(hash, &entropy, sizeof(entropy));
arch/mips/kernel/relocate.c
237
hash = rotate_xor(hash, prop, sizeof(*prop));
arch/mips/kernel/relocate.c
242
return hash;
arch/powerpc/crypto/aes-gcm-p10-glue.c
103
gcm_ghash_p10(nXi, hash->Htable+32, aad, i);
arch/powerpc/crypto/aes-gcm-p10-glue.c
112
gcm_ghash_p10(gctx->aad_hash, hash->Htable+32, nXi, 16);
arch/powerpc/crypto/aes-gcm-p10-glue.c
117
memcpy(hash->Htable, gctx->aad_hash, 16);
arch/powerpc/crypto/aes-gcm-p10-glue.c
121
struct Hash_ctx *hash, u8 *assoc, unsigned int assoclen)
arch/powerpc/crypto/aes-gcm-p10-glue.c
125
aes_p10_encrypt(hash->H, hash->H, rdkey);
arch/powerpc/crypto/aes-gcm-p10-glue.c
126
set_subkey(hash->H);
arch/powerpc/crypto/aes-gcm-p10-glue.c
127
gcm_init_htable(hash->Htable+32, hash->H);
arch/powerpc/crypto/aes-gcm-p10-glue.c
145
set_aad(gctx, hash, assoc, assoclen);
arch/powerpc/crypto/aes-gcm-p10-glue.c
148
static void finish_tag(struct gcm_ctx *gctx, struct Hash_ctx *hash, int len)
arch/powerpc/crypto/aes-gcm-p10-glue.c
157
memcpy(hash->Htable, gctx->ivtag, 16);
arch/powerpc/crypto/aes-gcm-p10-glue.c
170
gcm_ghash_p10(hash->Htable, hash->Htable+32, aclen, 16);
arch/powerpc/crypto/aes-gcm-p10-glue.c
173
hash->Htable[i] ^= gctx->ivtag[i];
arch/powerpc/crypto/aes-gcm-p10-glue.c
216
struct Hash_ctx *hash = PTR_ALIGN((void *)hashbuf, PPC_ALIGN);
arch/powerpc/crypto/aes-gcm-p10-glue.c
251
gcmp10_init(gctx, iv, (unsigned char *) &ctx->enc_key, hash, assoc, assoclen);
arch/powerpc/crypto/aes-gcm-p10-glue.c
274
&ctx->enc_key, gctx->iv, hash->Htable);
arch/powerpc/crypto/aes-gcm-p10-glue.c
277
&ctx->enc_key, gctx->iv, hash->Htable);
arch/powerpc/crypto/aes-gcm-p10-glue.c
293
gcm_update(gctx->iv, hash->Htable);
arch/powerpc/crypto/aes-gcm-p10-glue.c
294
finish_tag(gctx, hash, total_processed);
arch/powerpc/crypto/aes-gcm-p10-glue.c
299
scatterwalk_map_and_copy(hash->Htable, req->dst, req->assoclen + cryptlen,
arch/powerpc/crypto/aes-gcm-p10-glue.c
306
if (crypto_memneq(otag, hash->Htable, auth_tag_len)) {
arch/powerpc/crypto/aes-gcm-p10-glue.c
307
memzero_explicit(hash->Htable, 16);
arch/powerpc/crypto/aes-gcm-p10-glue.c
84
static void set_subkey(unsigned char *hash)
arch/powerpc/crypto/aes-gcm-p10-glue.c
86
*(u64 *)&hash[0] = be64_to_cpup((__be64 *)&hash[0]);
arch/powerpc/crypto/aes-gcm-p10-glue.c
87
*(u64 *)&hash[8] = be64_to_cpup((__be64 *)&hash[8]);
arch/powerpc/crypto/aes-gcm-p10-glue.c
94
static void set_aad(struct gcm_ctx *gctx, struct Hash_ctx *hash,
arch/powerpc/include/asm/book3s/64/mmu-hash.h
437
unsigned long hash, vsid;
arch/powerpc/include/asm/book3s/64/mmu-hash.h
442
hash = (vpn >> (SID_SHIFT - VPN_SHIFT)) ^
arch/powerpc/include/asm/book3s/64/mmu-hash.h
447
hash = vsid ^ (vsid << 25) ^
arch/powerpc/include/asm/book3s/64/mmu-hash.h
450
return hash & 0x7fffffffffUL;
arch/powerpc/include/asm/book3s/64/mmu-hash.h
457
long hpte_insert_repeating(unsigned long hash, unsigned long vpn, unsigned long pa,
arch/powerpc/kvm/book3s_32_mmu.c
104
hash = ((sr_vsid(sre) ^ page) << 6);
arch/powerpc/kvm/book3s_32_mmu.c
106
hash = ~hash;
arch/powerpc/kvm/book3s_32_mmu.c
107
hash &= htabmask;
arch/powerpc/kvm/book3s_32_mmu.c
109
pteg = (vcpu_book3s->sdr1 & 0xffff0000) | hash;
arch/powerpc/kvm/book3s_32_mmu.c
98
u32 page, hash, pteg, htabmask;
arch/powerpc/kvm/book3s_32_mmu_host.c
109
u32 page, hash;
arch/powerpc/kvm/book3s_32_mmu_host.c
114
hash = ((vsid ^ page) << 6);
arch/powerpc/kvm/book3s_32_mmu_host.c
116
hash = ~hash;
arch/powerpc/kvm/book3s_32_mmu_host.c
118
hash &= htabmask;
arch/powerpc/kvm/book3s_32_mmu_host.c
120
pteg |= hash;
arch/powerpc/kvm/book3s_32_mmu_host.c
123
htab, hash, htabmask, pteg);
arch/powerpc/kvm/book3s_64_mmu.c
121
u64 hash, pteg, htabsize;
arch/powerpc/kvm/book3s_64_mmu.c
130
hash = hpt_hash(vpn, kvmppc_mmu_book3s_64_get_pagesize(slbe), ssize);
arch/powerpc/kvm/book3s_64_mmu.c
132
hash = ~hash;
arch/powerpc/kvm/book3s_64_mmu.c
133
hash &= ((1ULL << 39ULL) - 1ULL);
arch/powerpc/kvm/book3s_64_mmu.c
134
hash &= htabsize;
arch/powerpc/kvm/book3s_64_mmu.c
135
hash <<= 7ULL;
arch/powerpc/kvm/book3s_64_mmu.c
138
pteg |= hash;
arch/powerpc/kvm/book3s_64_mmu_host.c
147
hash = hpt_hash(vpn, mmu_psize_defs[hpsize].shift, MMU_SEGSIZE_256M);
arch/powerpc/kvm/book3s_64_mmu_host.c
158
hpteg = ((hash & htab_hash_mask) * HPTES_PER_GROUP);
arch/powerpc/kvm/book3s_64_mmu_host.c
172
hash = ~hash;
arch/powerpc/kvm/book3s_64_mmu_host.c
188
hash = ~hash;
arch/powerpc/kvm/book3s_64_mmu_host.c
189
hpteg = ((hash & htab_hash_mask) * HPTES_PER_GROUP);
arch/powerpc/kvm/book3s_64_mmu_host.c
76
ulong hash, hpteg;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1209
unsigned long apsize, avpn, pteg, hash;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1296
hash = vsid ^ (offset >> pshift);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1306
hash = vsid ^ (vsid << 25) ^ (offset >> pshift);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1309
new_pteg = hash & new_hash_mask;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1311
new_pteg = ~hash & new_hash_mask;
arch/powerpc/kvm/book3s_64_mmu_hv.c
212
unsigned long addr, hash;
arch/powerpc/kvm/book3s_64_mmu_hv.c
237
hash = (i ^ (VRMA_VSID ^ (VRMA_VSID << 25)))
arch/powerpc/kvm/book3s_64_mmu_hv.c
245
hash = (hash << 3) + 7;
arch/powerpc/kvm/book3s_64_mmu_hv.c
248
ret = kvmppc_virtmode_do_h_enter(kvm, H_EXACT, hash, hp_v, hp_r,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1103
unsigned long vsid, hash;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1126
hash = (vsid ^ ((eaddr & somask) >> pshift)) & kvmppc_hpt_mask(&kvm->arch.hpt);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1137
hpte = (__be64 *)(kvm->arch.hpt.virt + (hash << 7));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1165
return (hash << 3) + (i >> 1);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1173
hash = hash ^ kvmppc_hpt_mask(&kvm->arch.hpt);
arch/powerpc/mm/book3s32/mmu.c
394
unsigned int hash = (unsigned int)Hash - PAGE_OFFSET;
arch/powerpc/mm/book3s32/mmu.c
409
modify_instruction_site(&patch__hash_page_A0, 0xffff, hash >> 16);
arch/powerpc/mm/book3s32/mmu.c
418
modify_instruction_site(&patch__flush_hash_A0, 0xffff, hash >> 16);
arch/powerpc/mm/book3s64/hash_4k.c
102
hpte_group = (hash & htab_hash_mask) *
arch/powerpc/mm/book3s64/hash_4k.c
29
unsigned long vpn, hash, slot;
arch/powerpc/mm/book3s64/hash_4k.c
82
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hash_4k.c
85
hpte_group = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_4k.c
94
hpte_group = (~hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_64k.c
157
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hash_64k.c
159
hpte_group = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_64k.c
170
hpte_group = (~hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_64k.c
198
hpte_group = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_64k.c
236
unsigned long vpn, hash, slot;
arch/powerpc/mm/book3s64/hash_64k.c
292
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hash_64k.c
295
hpte_group = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_64k.c
305
hpte_group = (~hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_64k.c
313
hpte_group = (hash & htab_hash_mask) *
arch/powerpc/mm/book3s64/hash_64k.c
47
unsigned long vpn, hash, slot, gslot;
arch/powerpc/mm/book3s64/hash_hugepage.c
102
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hash_hugepage.c
105
hash = ~hash;
arch/powerpc/mm/book3s64/hash_hugepage.c
106
slot = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_hugepage.c
128
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hash_hugepage.c
134
hpte_group = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_hugepage.c
143
hpte_group = (~hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_hugepage.c
150
hpte_group = (hash & htab_hash_mask) *
arch/powerpc/mm/book3s64/hash_hugepage.c
30
unsigned long vpn, hash, shift, slot;
arch/powerpc/mm/book3s64/hash_native.c
451
unsigned long hash;
arch/powerpc/mm/book3s64/hash_native.c
454
hash = hpt_hash(vpn, mmu_psize_defs[psize].shift, ssize);
arch/powerpc/mm/book3s64/hash_native.c
461
hpte_group = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_native.c
465
hpte_group = (~hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_native.c
602
unsigned long hidx, vpn = 0, hash, slot;
arch/powerpc/mm/book3s64/hash_native.c
617
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hash_native.c
619
hash = ~hash;
arch/powerpc/mm/book3s64/hash_native.c
621
slot = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_native.c
778
unsigned long hash, index, hidx, shift, slot;
arch/powerpc/mm/book3s64/hash_native.c
796
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hash_native.c
799
hash = ~hash;
arch/powerpc/mm/book3s64/hash_native.c
800
slot = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_utils.c
2189
unsigned long hash, gslot, hidx;
arch/powerpc/mm/book3s64/hash_utils.c
2191
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hash_utils.c
2194
hash = ~hash;
arch/powerpc/mm/book3s64/hash_utils.c
2195
gslot = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_utils.c
2229
unsigned long hidx, shift, vpn, hash, slot;
arch/powerpc/mm/book3s64/hash_utils.c
2265
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hash_utils.c
2267
hash = ~hash;
arch/powerpc/mm/book3s64/hash_utils.c
2269
slot = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_utils.c
2294
long hpte_insert_repeating(unsigned long hash, unsigned long vpn,
arch/powerpc/mm/book3s64/hash_utils.c
2302
hpte_group = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_utils.c
2310
hpte_group = (~hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_utils.c
2316
hpte_group = (hash & htab_hash_mask) *
arch/powerpc/mm/book3s64/hash_utils.c
281
unsigned long hash;
arch/powerpc/mm/book3s64/hash_utils.c
287
hash = hpt_hash(vpn, PAGE_SHIFT, mmu_kernel_ssize);
arch/powerpc/mm/book3s64/hash_utils.c
296
ret = hpte_insert_repeating(hash, vpn, __pa(vaddr), mode,
arch/powerpc/mm/book3s64/hash_utils.c
310
unsigned long hash, hslot, slot;
arch/powerpc/mm/book3s64/hash_utils.c
314
hash = hpt_hash(vpn, PAGE_SHIFT, mmu_kernel_ssize);
arch/powerpc/mm/book3s64/hash_utils.c
324
hash = ~hash;
arch/powerpc/mm/book3s64/hash_utils.c
325
slot = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/book3s64/hash_utils.c
643
unsigned long hash, hpteg;
arch/powerpc/mm/book3s64/hash_utils.c
672
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hash_utils.c
673
hpteg = ((hash & htab_hash_mask) * HPTES_PER_GROUP);
arch/powerpc/mm/book3s64/hash_utils.c
692
hpteg = ((~hash & htab_hash_mask) * HPTES_PER_GROUP);
arch/powerpc/mm/book3s64/hugetlbpage.c
106
unsigned long hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/book3s64/hugetlbpage.c
113
slot = hpte_insert_repeating(hash, vpn, pa, rflags, 0,
arch/powerpc/mm/nohash/kaslr_booke.c
45
static unsigned long __init rotate_xor(unsigned long hash, const void *area,
arch/powerpc/mm/nohash/kaslr_booke.c
51
for (i = 0; i < size / sizeof(hash); i++) {
arch/powerpc/mm/nohash/kaslr_booke.c
53
hash = (hash << ((sizeof(hash) * 8) - 7)) | (hash >> 7);
arch/powerpc/mm/nohash/kaslr_booke.c
54
hash ^= ptr[i];
arch/powerpc/mm/nohash/kaslr_booke.c
57
return hash;
arch/powerpc/mm/nohash/kaslr_booke.c
66
unsigned long hash = 0;
arch/powerpc/mm/nohash/kaslr_booke.c
69
hash = rotate_xor(hash, linux_banner, strlen(linux_banner));
arch/powerpc/mm/nohash/kaslr_booke.c
70
hash = rotate_xor(hash, fdt, fdt_totalsize(fdt));
arch/powerpc/mm/nohash/kaslr_booke.c
72
return hash;
arch/powerpc/mm/ptdump/hashpagetable.c
210
unsigned long hash, vsid, vpn, hpte_group, want_v, hpte_v;
arch/powerpc/mm/ptdump/hashpagetable.c
217
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/ptdump/hashpagetable.c
224
hash = ~hash;
arch/powerpc/mm/ptdump/hashpagetable.c
225
hpte_group = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/mm/ptdump/hashpagetable.c
251
unsigned long vsid, vpn, hash, hpte_group, want_v;
arch/powerpc/mm/ptdump/hashpagetable.c
259
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/mm/ptdump/hashpagetable.c
264
hash = ~hash;
arch/powerpc/mm/ptdump/hashpagetable.c
265
hpte_group = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/platforms/pseries/lpar.c
1001
hpte_group = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/platforms/pseries/lpar.c
1005
hpte_group = (~hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/platforms/pseries/lpar.c
1259
unsigned long shift, hidx, vpn = 0, hash, slot;
arch/powerpc/platforms/pseries/lpar.c
1273
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/platforms/pseries/lpar.c
1275
hash = ~hash;
arch/powerpc/platforms/pseries/lpar.c
1277
slot = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/platforms/pseries/lpar.c
1335
unsigned long slot, hash, hidx;
arch/powerpc/platforms/pseries/lpar.c
1337
hash = hpt_hash(vpn, shift, ssize);
arch/powerpc/platforms/pseries/lpar.c
1340
hash = ~hash;
arch/powerpc/platforms/pseries/lpar.c
1341
slot = (hash & htab_hash_mask) * HPTES_PER_GROUP;
arch/powerpc/platforms/pseries/lpar.c
990
unsigned long hash;
arch/powerpc/platforms/pseries/lpar.c
994
hash = hpt_hash(vpn, mmu_psize_defs[psize].shift, ssize);
arch/riscv/kernel/module.c
664
unsigned long hash;
arch/riscv/kernel/module.c
675
hash = hash_min((uintptr_t)location, hashtable_bits);
arch/riscv/kernel/module.c
677
current_head = &relocation_hashtable[hash];
arch/riscv/net/bpf_jit_comp64.c
470
static inline void emit_kcfi(u32 hash, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp64.c
473
emit(hash, ctx);
arch/s390/kernel/cert_store.c
262
u8 hash[SHA256_DIGEST_SIZE];
arch/s390/kernel/cert_store.c
268
sha256((u8 *)vce + vce->vce_hdr.vc_offset, vce->vce_hdr.vc_length, hash);
arch/s390/kernel/cert_store.c
269
if (memcmp(vce_hash, hash, vc_hash_length) == 0)
arch/s390/kernel/cert_store.c
276
debug_event(cert_store_hexdump, 3, hash, SHA256_DIGEST_SIZE);
arch/s390/kernel/machine_kexec_file.c
60
ms->hash != 0 ||
arch/s390/kernel/uv.c
609
static inline ssize_t emit_hash(struct uv_key_hash *hash, char *buf, int at)
arch/s390/kernel/uv.c
612
hash->dword[0], hash->dword[1], hash->dword[2], hash->dword[3]);
arch/sparc/mm/tsb.c
60
unsigned long hash = tsb_hash(v, PAGE_SHIFT,
arch/sparc/mm/tsb.c
62
struct tsb *ent = &swapper_tsb[hash];
arch/sparc/mm/tsb.c
73
unsigned long tag, ent, hash;
arch/sparc/mm/tsb.c
76
hash = tsb_hash(v, hash_shift, nentries);
arch/sparc/mm/tsb.c
77
ent = tsb + (hash * sizeof(struct tsb));
arch/sparc/net/bpf_jit_comp_32.c
552
emit_skb_load32(hash, r_A);
arch/x86/boot/compressed/kaslr.c
41
static unsigned long rotate_xor(unsigned long hash, const void *area,
arch/x86/boot/compressed/kaslr.c
47
for (i = 0; i < size / sizeof(hash); i++) {
arch/x86/boot/compressed/kaslr.c
49
hash = (hash << ((sizeof(hash) * 8) - 7)) | (hash >> 7);
arch/x86/boot/compressed/kaslr.c
50
hash ^= ptr[i];
arch/x86/boot/compressed/kaslr.c
53
return hash;
arch/x86/boot/compressed/kaslr.c
59
unsigned long hash = 0;
arch/x86/boot/compressed/kaslr.c
61
hash = rotate_xor(hash, build_str, sizeof(build_str));
arch/x86/boot/compressed/kaslr.c
62
hash = rotate_xor(hash, boot_params_ptr, sizeof(*boot_params_ptr));
arch/x86/boot/compressed/kaslr.c
64
return hash;
arch/x86/kernel/alternative.c
1213
u32 hash;
arch/x86/kernel/alternative.c
1227
if (get_kernel_nofault(hash, func))
arch/x86/kernel/alternative.c
1230
return hash;
arch/x86/kernel/alternative.c
1258
static u32 cfi_rehash(u32 hash)
arch/x86/kernel/alternative.c
1260
hash ^= cfi_seed;
arch/x86/kernel/alternative.c
1261
while (unlikely(__is_endbr(hash) || __is_endbr(-hash))) {
arch/x86/kernel/alternative.c
1262
bool lsb = hash & 1;
arch/x86/kernel/alternative.c
1263
hash >>= 1;
arch/x86/kernel/alternative.c
1265
hash ^= 0x80200003;
arch/x86/kernel/alternative.c
1267
return hash;
arch/x86/kernel/alternative.c
1505
u32 hash;
arch/x86/kernel/alternative.c
1508
hash = decode_caller_hash(addr);
arch/x86/kernel/alternative.c
1509
if (!hash) /* nocfi callers */
arch/x86/kernel/alternative.c
1528
u32 hash;
arch/x86/kernel/alternative.c
1531
hash = decode_caller_hash(addr);
arch/x86/kernel/alternative.c
1532
if (!hash) /* nocfi callers */
arch/x86/kernel/alternative.c
1548
u32 hash;
arch/x86/kernel/alternative.c
1550
hash = decode_preamble_hash(addr, NULL);
arch/x86/kernel/alternative.c
1551
if (WARN(!hash, "no CFI hash found at: %pS %px %*ph\n",
arch/x86/kernel/alternative.c
1555
hash = cfi_rehash(hash);
arch/x86/kernel/alternative.c
1556
text_poke_early(addr + 1, &hash, 4);
arch/x86/kernel/alternative.c
1632
u32 hash;
arch/x86/kernel/alternative.c
1642
hash = decode_preamble_hash(addr, &arity);
arch/x86/kernel/alternative.c
1643
if (WARN(!hash, "no CFI hash found at: %pS %px %*ph\n",
arch/x86/kernel/alternative.c
1658
text_poke_early(addr + fineibt_preamble_hash, &hash, 4);
arch/x86/kernel/alternative.c
1692
u32 hash;
arch/x86/kernel/alternative.c
1695
hash = decode_caller_hash(addr);
arch/x86/kernel/alternative.c
1696
if (hash) {
arch/x86/kernel/alternative.c
1697
hash = -cfi_rehash(hash);
arch/x86/kernel/alternative.c
1698
text_poke_early(addr + 2, &hash, 4);
arch/x86/kernel/alternative.c
1726
u32 hash;
arch/x86/kernel/alternative.c
1731
hash = decode_caller_hash(addr);
arch/x86/kernel/alternative.c
1732
if (!hash)
arch/x86/kernel/alternative.c
1738
text_poke_early(addr + fineibt_caller_hash, &hash, 4);
arch/x86/kernel/alternative.c
1755
memcpy(bytes + fineibt_caller_hash, &hash, 4);
arch/x86/kernel/alternative.c
1953
u32 hash;
arch/x86/kernel/alternative.c
1960
__get_kernel_nofault(&hash, addr + fineibt_preamble_hash, u32, Efault);
arch/x86/kernel/alternative.c
1961
*type = (u32)regs->ax + hash;
arch/x86/kernel/alternative.c
1981
u32 hash;
arch/x86/kernel/alternative.c
2003
__get_kernel_nofault(&hash, addr + fineibt_preamble_hash, u32, Efault);
arch/x86/kernel/alternative.c
2004
*type = (u32)regs->ax + hash;
arch/x86/kernel/tboot.c
453
u8 hash[SHA1_SIZE];
arch/x86/net/bpf_jit_comp.c
418
static void emit_fineibt(u8 **pprog, u8 *ip, u32 hash, int arity)
arch/x86/net/bpf_jit_comp.c
423
EMIT1_off32(0x2d, hash); /* subl $hash, %eax */
arch/x86/net/bpf_jit_comp.c
435
static void emit_kcfi(u8 **pprog, u32 hash)
arch/x86/net/bpf_jit_comp.c
439
EMIT1_off32(0xb8, hash); /* movl $hash, %eax */
arch/x86/net/bpf_jit_comp.c
449
static void emit_cfi(u8 **pprog, u8 *ip, u32 hash, int arity)
arch/x86/net/bpf_jit_comp.c
455
emit_fineibt(&prog, ip, hash, arity);
arch/x86/net/bpf_jit_comp.c
459
emit_kcfi(&prog, hash);
block/blk-mq.c
381
INIT_HLIST_NODE(&rq->hash);
block/blk-mq.c
453
INIT_HLIST_NODE(&rq->hash);
block/elevator.c
136
hash_init(eq->hash);
block/elevator.c
167
hash_del(&rq->hash);
block/elevator.c
183
hash_add(e->hash, &rq->hash, rq_hash_key(rq));
block/elevator.c
200
hash_for_each_possible_safe(e->hash, rq, next, hash, offset) {
block/elevator.h
154
DECLARE_HASHTABLE(hash, ELV_HASH_BITS);
certs/blacklist.c
147
static char *get_raw_hash(const u8 *hash, size_t hash_len,
certs/blacklist.c
173
bin2hex(p, hash, hash_len);
certs/blacklist.c
183
static int mark_raw_hash_blacklisted(const char *hash)
certs/blacklist.c
189
hash,
certs/blacklist.c
197
pr_warn("Duplicate blacklisted hash %s\n", hash);
certs/blacklist.c
199
pr_err("Problem blacklisting hash %s: %pe\n", hash, key);
certs/blacklist.c
205
int mark_hash_blacklisted(const u8 *hash, size_t hash_len,
certs/blacklist.c
211
buffer = get_raw_hash(hash, hash_len, hash_type);
certs/blacklist.c
225
int is_hash_blacklisted(const u8 *hash, size_t hash_len,
certs/blacklist.c
232
buffer = get_raw_hash(hash, hash_len, hash_type);
certs/blacklist.c
247
int is_binary_blacklisted(const u8 *hash, size_t hash_len)
certs/blacklist.c
249
if (is_hash_blacklisted(hash, hash_len, BLACKLIST_HASH_BINARY) ==
crypto/ahash.c
702
struct crypto_ahash *hash = __crypto_ahash_cast(tfm);
crypto/ahash.c
703
struct ahash_alg *alg = crypto_ahash_alg(hash);
crypto/ahash.c
706
alg->exit_tfm(hash);
crypto/ahash.c
710
if (crypto_ahash_need_fallback(hash))
crypto/ahash.c
711
crypto_free_ahash(crypto_ahash_fb(hash));
crypto/ahash.c
716
struct crypto_ahash *hash = __crypto_ahash_cast(tfm);
crypto/ahash.c
717
struct ahash_alg *alg = crypto_ahash_alg(hash);
crypto/ahash.c
721
crypto_ahash_set_statesize(hash, alg->halg.statesize);
crypto/ahash.c
722
crypto_ahash_set_reqsize(hash, crypto_tfm_alg_reqsize(tfm));
crypto/ahash.c
727
if (crypto_ahash_need_fallback(hash)) {
crypto/ahash.c
728
fb = crypto_alloc_ahash(crypto_ahash_alg_name(hash),
crypto/ahash.c
739
ahash_set_needkey(hash, alg);
crypto/ahash.c
744
err = alg->init_tfm(hash);
crypto/ahash.c
753
if (!ahash_is_async(hash) && crypto_ahash_reqsize(hash) >
crypto/ahash.c
758
if (crypto_ahash_reqsize(hash) < HASH_MAX_DESCSIZE)
crypto/ahash.c
759
crypto_ahash_set_reqsize(hash, HASH_MAX_DESCSIZE);
crypto/ahash.c
765
alg->exit_tfm(hash);
crypto/ahash.c
865
struct crypto_ahash *crypto_clone_ahash(struct crypto_ahash *hash)
crypto/ahash.c
867
struct hash_alg_common *halg = crypto_hash_alg_common(hash);
crypto/ahash.c
868
struct crypto_tfm *tfm = crypto_ahash_tfm(hash);
crypto/ahash.c
879
return hash;
crypto/ahash.c
887
nhash->reqsize = hash->reqsize;
crypto/ahash.c
888
nhash->statesize = hash->statesize;
crypto/ahash.c
890
if (likely(hash->using_shash)) {
crypto/ahash.c
894
shash = crypto_clone_shash(ahash_to_shash(hash));
crypto/ahash.c
905
if (crypto_ahash_need_fallback(hash)) {
crypto/ahash.c
906
fb = crypto_clone_ahash(crypto_ahash_fb(hash));
crypto/ahash.c
915
alg = crypto_ahash_alg(hash);
crypto/ahash.c
919
err = alg->clone_tfm(nhash, hash);
crypto/authenc.c
155
u8 *hash = areq_ctx->tail;
crypto/authenc.c
159
ahash_request_set_crypt(ahreq, req->dst, hash,
crypto/authenc.c
169
scatterwalk_map_and_copy(hash, req->dst, req->assoclen + req->cryptlen,
crypto/authenc.c
288
u8 *hash = areq_ctx->tail;
crypto/authenc.c
292
ahash_request_set_crypt(ahreq, req->src, hash,
crypto/authencesn.c
101
scatterwalk_map_and_copy(hash, dst, assoclen + cryptlen, authsize, 1);
crypto/authencesn.c
120
u8 *hash = areq_ctx->tail;
crypto/authencesn.c
140
ahash_request_set_crypt(ahreq, dst, hash, assoclen + cryptlen);
crypto/authencesn.c
89
u8 *hash = areq_ctx->tail;
crypto/crc32.c
36
static int crc32_setkey(struct crypto_shash *hash, const u8 *key,
crypto/crc32.c
39
u32 *mctx = crypto_shash_ctx(hash);
crypto/cryptd.c
442
struct crypto_shash *hash;
crypto/cryptd.c
444
hash = crypto_spawn_shash(spawn);
crypto/cryptd.c
445
if (IS_ERR(hash))
crypto/cryptd.c
446
return PTR_ERR(hash);
crypto/cryptd.c
448
ctx->child = hash;
crypto/cryptd.c
451
crypto_shash_descsize(hash));
crypto/cryptd.c
460
struct crypto_shash *hash;
crypto/cryptd.c
462
hash = crypto_clone_shash(ctx->child);
crypto/cryptd.c
463
if (IS_ERR(hash))
crypto/cryptd.c
464
return PTR_ERR(hash);
crypto/cryptd.c
466
nctx->child = hash;
crypto/ecdsa.c
23
static int _ecdsa_verify(struct ecc_ctx *ctx, const u64 *hash, const u64 *r, const u64 *s)
crypto/ecdsa.c
41
hash[ndigits - 1], hash[ndigits - 2], hash[0]);
crypto/ecdsa.c
46
vli_mod_mult_slow(u1, hash, s1, curve->n, ndigits);
crypto/ecdsa.c
73
u64 hash[ECC_MAX_DIGITS];
crypto/ecdsa.c
84
ecc_digits_from_bytes(digest, bufsize, hash, ctx->curve->g.ndigits);
crypto/ecdsa.c
86
return _ecdsa_verify(ctx, hash, sig->r, sig->s);
crypto/essiv.c
112
desc->tfm = tctx->hash;
crypto/essiv.c
123
crypto_shash_digestsize(tctx->hash));
crypto/essiv.c
272
struct crypto_shash *hash;
crypto/essiv.c
279
hash = crypto_alloc_shash(ictx->shash_driver_name, 0, 0);
crypto/essiv.c
280
if (IS_ERR(hash)) {
crypto/essiv.c
281
err = PTR_ERR(hash);
crypto/essiv.c
286
tctx->hash = hash;
crypto/essiv.c
359
crypto_free_shash(tctx->hash);
crypto/essiv.c
368
crypto_free_shash(tctx->hash);
crypto/essiv.c
56
struct crypto_shash *hash;
crypto/essiv.c
81
err = crypto_shash_tfm_digest(tctx->hash, key, keylen, salt);
crypto/essiv.c
90
crypto_shash_digestsize(tctx->hash));
crypto/essiv.c
97
SHASH_DESC_ON_STACK(desc, tctx->hash);
crypto/gcm.c
118
sg_init_one(data->sg, &data->hash, sizeof(data->hash));
crypto/gcm.c
125
sizeof(data->hash), data->iv);
crypto/gcm.c
136
err = crypto_ahash_setkey(ghash, (u8 *)&data->hash, sizeof(be128));
crypto/gcm.c
95
be128 hash;
crypto/hmac.c
111
desc->tfm = tctx->hash;
crypto/hmac.c
148
struct crypto_shash *hash;
crypto/hmac.c
153
hash = crypto_spawn_shash(spawn);
crypto/hmac.c
154
if (IS_ERR(hash))
crypto/hmac.c
155
return PTR_ERR(hash);
crypto/hmac.c
157
tctx->hash = hash;
crypto/hmac.c
165
struct crypto_shash *hash;
crypto/hmac.c
167
hash = crypto_clone_shash(sctx->hash);
crypto/hmac.c
168
if (IS_ERR(hash))
crypto/hmac.c
169
return PTR_ERR(hash);
crypto/hmac.c
171
dctx->hash = hash;
crypto/hmac.c
179
crypto_free_shash(tctx->hash);
crypto/hmac.c
24
struct crypto_shash *hash;
crypto/hmac.c
255
struct crypto_ahash *fb = crypto_ahash_fb(tctx->hash);
crypto/hmac.c
30
struct crypto_ahash *hash;
crypto/hmac.c
314
ahash_request_set_tfm(req, tctx->hash);
crypto/hmac.c
329
ahash_request_set_tfm(req, tctx->hash);
crypto/hmac.c
412
struct crypto_ahash *hash;
crypto/hmac.c
414
hash = crypto_spawn_ahash(ahash_instance_ctx(inst));
crypto/hmac.c
415
if (IS_ERR(hash))
crypto/hmac.c
416
return PTR_ERR(hash);
crypto/hmac.c
419
crypto_ahash_reqsize(hash))
crypto/hmac.c
42
struct crypto_shash *hash = tctx->hash;
crypto/hmac.c
422
tctx->hash = hash;
crypto/hmac.c
431
struct crypto_ahash *hash;
crypto/hmac.c
433
hash = crypto_clone_ahash(sctx->hash);
crypto/hmac.c
434
if (IS_ERR(hash))
crypto/hmac.c
435
return PTR_ERR(hash);
crypto/hmac.c
437
dctx->hash = hash;
crypto/hmac.c
445
crypto_free_ahash(tctx->hash);
crypto/hmac.c
45
SHASH_DESC_ON_STACK(shash, hash);
crypto/hmac.c
51
shash->tfm = hash;
crypto/hmac.c
94
desc->tfm = tctx->hash;
crypto/jitterentropy-kcapi.c
227
struct crypto_shash *hash;
crypto/jitterentropy-kcapi.c
234
hash = crypto_alloc_shash(JENT_CONDITIONING_HASH, 0, 0);
crypto/jitterentropy-kcapi.c
235
if (IS_ERR(hash)) {
crypto/jitterentropy-kcapi.c
237
return PTR_ERR(hash);
crypto/jitterentropy-kcapi.c
239
rng->tfm = hash;
crypto/jitterentropy-kcapi.c
241
size = sizeof(struct shash_desc) + crypto_shash_descsize(hash);
crypto/jitterentropy-kcapi.c
248
sdesc->tfm = hash;
crypto/krb5enc.c
168
static void krb5enc_insert_checksum(struct aead_request *req, u8 *hash)
crypto/krb5enc.c
172
scatterwalk_map_and_copy(hash, req->dst,
crypto/krb5enc.c
214
u8 *hash = areq_ctx->tail;
crypto/krb5enc.c
220
ahash_request_set_crypt(ahreq, req->src, hash, req->assoclen + req->cryptlen);
crypto/krb5enc.c
226
krb5enc_insert_checksum(req, hash);
crypto/krb5enc.c
287
u8 *hash = areq_ctx->tail;
crypto/krb5enc.c
291
ahash_request_set_crypt(ahreq, req->dst, hash,
crypto/md4.c
127
hash[0] += a;
crypto/md4.c
128
hash[1] += b;
crypto/md4.c
129
hash[2] += c;
crypto/md4.c
130
hash[3] += d;
crypto/md4.c
136
md4_transform(ctx->hash, ctx->block);
crypto/md4.c
143
mctx->hash[0] = 0x67452301;
crypto/md4.c
144
mctx->hash[1] = 0xefcdab89;
crypto/md4.c
145
mctx->hash[2] = 0x98badcfe;
crypto/md4.c
146
mctx->hash[3] = 0x10325476;
crypto/md4.c
204
md4_transform(mctx->hash, mctx->block);
crypto/md4.c
205
cpu_to_le32_array(mctx->hash, ARRAY_SIZE(mctx->hash));
crypto/md4.c
206
memcpy(out, mctx->hash, sizeof(mctx->hash));
crypto/md4.c
37
u32 hash[MD4_HASH_WORDS];
crypto/md4.c
67
static void md4_transform(u32 *hash, u32 const *in)
crypto/md4.c
71
a = hash[0];
crypto/md4.c
72
b = hash[1];
crypto/md4.c
73
c = hash[2];
crypto/md4.c
74
d = hash[3];
crypto/md5.c
21
static_assert(offsetof(struct md5_ctx, state) == offsetof(struct md5_state, hash));
crypto/shash.c
304
struct crypto_shash *hash = __crypto_shash_cast(tfm);
crypto/shash.c
305
struct shash_alg *alg = crypto_shash_alg(hash);
crypto/shash.c
307
alg->exit_tfm(hash);
crypto/shash.c
312
struct crypto_shash *hash = __crypto_shash_cast(tfm);
crypto/shash.c
313
struct shash_alg *alg = crypto_shash_alg(hash);
crypto/shash.c
315
shash_set_needkey(hash, alg);
crypto/shash.c
323
return alg->init_tfm(hash);
crypto/shash.c
398
struct crypto_shash *crypto_clone_shash(struct crypto_shash *hash)
crypto/shash.c
400
struct crypto_tfm *tfm = crypto_shash_tfm(hash);
crypto/shash.c
401
struct shash_alg *alg = crypto_shash_alg(hash);
crypto/shash.c
410
return hash;
crypto/shash.c
421
err = alg->clone_tfm(nhash, hash);
crypto/sm3_generic.c
32
unsigned int len, u8 *hash)
crypto/sm3_generic.c
35
return sm3_base_finish(desc, hash);
crypto/streebog_generic.c
1020
memcpy(digest, &ctx->hash.qword[4], STREEBOG256_DIGEST_SIZE);
crypto/streebog_generic.c
1022
memcpy(digest, &ctx->hash.qword[0], STREEBOG512_DIGEST_SIZE);
crypto/streebog_generic.c
996
memcpy(&ctx->hash, &ctx->h, sizeof(struct streebog_uint512));
crypto/testmgr.c
151
struct hash_test_suite hash;
crypto/testmgr.c
1992
const struct hash_testvec *template = desc->suite.hash.vecs;
crypto/testmgr.c
1993
unsigned int tcount = desc->suite.hash.count;
crypto/testmgr.c
4267
.hash = __VECS(blake2b_160_tv_template)
crypto/testmgr.c
4275
.hash = __VECS(blake2b_256_tv_template)
crypto/testmgr.c
4283
.hash = __VECS(blake2b_384_tv_template)
crypto/testmgr.c
4291
.hash = __VECS(blake2b_512_tv_template)
crypto/testmgr.c
4394
.hash = __VECS(aes_cbcmac_tv_template)
crypto/testmgr.c
4400
.hash = __VECS(sm4_cbcmac_tv_template)
crypto/testmgr.c
4436
.hash = __VECS(aes_cmac128_tv_template)
crypto/testmgr.c
4442
.hash = __VECS(camellia_cmac128_tv_template)
crypto/testmgr.c
4448
.hash = __VECS(des3_ede_cmac64_tv_template)
crypto/testmgr.c
4454
.hash = __VECS(sm4_cmac128_tv_template)
crypto/testmgr.c
4462
.hash = __VECS(crc32_tv_template)
crypto/testmgr.c
4470
.hash = __VECS(crc32c_tv_template)
crypto/testmgr.c
4992
.hash = __VECS(ghash_tv_template)
crypto/testmgr.c
5006
.hash = __VECS(hmac_md5_tv_template)
crypto/testmgr.c
5012
.hash = __VECS(hmac_rmd160_tv_template)
crypto/testmgr.c
5020
.hash = __VECS(hmac_sha1_tv_template)
crypto/testmgr.c
5028
.hash = __VECS(hmac_sha224_tv_template)
crypto/testmgr.c
5036
.hash = __VECS(hmac_sha256_tv_template)
crypto/testmgr.c
5044
.hash = __VECS(hmac_sha3_224_tv_template)
crypto/testmgr.c
5052
.hash = __VECS(hmac_sha3_256_tv_template)
crypto/testmgr.c
5060
.hash = __VECS(hmac_sha3_384_tv_template)
crypto/testmgr.c
5068
.hash = __VECS(hmac_sha3_512_tv_template)
crypto/testmgr.c
5076
.hash = __VECS(hmac_sha384_tv_template)
crypto/testmgr.c
5084
.hash = __VECS(hmac_sha512_tv_template)
crypto/testmgr.c
5090
.hash = __VECS(hmac_sm3_tv_template)
crypto/testmgr.c
5096
.hash = __VECS(hmac_streebog256_tv_template)
crypto/testmgr.c
5102
.hash = __VECS(hmac_streebog512_tv_template)
crypto/testmgr.c
5191
.hash = __VECS(md4_tv_template)
crypto/testmgr.c
5198
.hash = __VECS(md5_tv_template)
crypto/testmgr.c
5204
.hash = __VECS(michael_mic_tv_template)
crypto/testmgr.c
5236
.hash = __VECS(hmac_sha224_tv_template)
crypto/testmgr.c
5243
.hash = __VECS(hmac_sha256_tv_template)
crypto/testmgr.c
5250
.hash = __VECS(hmac_sha384_tv_template)
crypto/testmgr.c
5257
.hash = __VECS(hmac_sha512_tv_template)
crypto/testmgr.c
5372
.hash = __VECS(rmd160_tv_template)
crypto/testmgr.c
5387
.hash = __VECS(sha1_tv_template)
crypto/testmgr.c
5395
.hash = __VECS(sha224_tv_template)
crypto/testmgr.c
5403
.hash = __VECS(sha256_tv_template)
crypto/testmgr.c
5411
.hash = __VECS(sha3_224_tv_template)
crypto/testmgr.c
5419
.hash = __VECS(sha3_256_tv_template)
crypto/testmgr.c
5427
.hash = __VECS(sha3_384_tv_template)
crypto/testmgr.c
5435
.hash = __VECS(sha3_512_tv_template)
crypto/testmgr.c
5443
.hash = __VECS(sha384_tv_template)
crypto/testmgr.c
5451
.hash = __VECS(sha512_tv_template)
crypto/testmgr.c
5457
.hash = __VECS(sm3_tv_template)
crypto/testmgr.c
5463
.hash = __VECS(streebog256_tv_template)
crypto/testmgr.c
5469
.hash = __VECS(streebog512_tv_template)
crypto/testmgr.c
5475
.hash = __VECS(wp256_tv_template)
crypto/testmgr.c
5481
.hash = __VECS(wp384_tv_template)
crypto/testmgr.c
5487
.hash = __VECS(wp512_tv_template)
crypto/testmgr.c
5521
.hash = __VECS(aes_xcbc128_tv_template)
crypto/testmgr.c
5527
.hash = __VECS(sm4_xcbc128_tv_template)
crypto/testmgr.c
5614
.hash = __VECS(xxhash64_tv_template)
crypto/wp512.c
1057
digest[i] = cpu_to_be64(wctx->hash[i]);
crypto/wp512.c
40
u64 hash[WP512_DIGEST_SIZE/8];
crypto/wp512.c
790
state[0] = block[0] ^ (K[0] = wctx->hash[0]);
crypto/wp512.c
791
state[1] = block[1] ^ (K[1] = wctx->hash[1]);
crypto/wp512.c
792
state[2] = block[2] ^ (K[2] = wctx->hash[2]);
crypto/wp512.c
793
state[3] = block[3] ^ (K[3] = wctx->hash[3]);
crypto/wp512.c
794
state[4] = block[4] ^ (K[4] = wctx->hash[4]);
crypto/wp512.c
795
state[5] = block[5] ^ (K[5] = wctx->hash[5]);
crypto/wp512.c
796
state[6] = block[6] ^ (K[6] = wctx->hash[6]);
crypto/wp512.c
797
state[7] = block[7] ^ (K[7] = wctx->hash[7]);
crypto/wp512.c
975
wctx->hash[0] ^= state[0] ^ block[0];
crypto/wp512.c
976
wctx->hash[1] ^= state[1] ^ block[1];
crypto/wp512.c
977
wctx->hash[2] ^= state[2] ^ block[2];
crypto/wp512.c
978
wctx->hash[3] ^= state[3] ^ block[3];
crypto/wp512.c
979
wctx->hash[4] ^= state[4] ^ block[4];
crypto/wp512.c
980
wctx->hash[5] ^= state[5] ^ block[5];
crypto/wp512.c
981
wctx->hash[6] ^= state[6] ^ block[6];
crypto/wp512.c
982
wctx->hash[7] ^= state[7] ^ block[7];
crypto/wp512.c
992
wctx->hash[i] = 0L;
drivers/accessibility/speakup/utils.h
58
int hash = 0;
drivers/accessibility/speakup/utils.h
61
hash = (hash * 17) & 0xfffffff;
drivers/accessibility/speakup/utils.h
64
hash += (int)*pn;
drivers/accessibility/speakup/utils.h
67
hash %= HASHSIZE;
drivers/accessibility/speakup/utils.h
68
return &key_table[hash];
drivers/base/power/trace.c
193
unsigned int hash = hash_string(lineno, file, FILEHASH);
drivers/base/power/trace.c
194
if (hash != value)
drivers/base/power/trace.c
211
unsigned int hash = hash_string(DEVSEED, dev_name(dev), DEVHASH);
drivers/base/power/trace.c
212
if (hash == value) {
drivers/base/power/trace.c
238
unsigned int hash = hash_string(DEVSEED, dev_name(dev),
drivers/base/power/trace.c
240
if (hash == value) {
drivers/char/random.c
629
struct blake2s_ctx hash;
drivers/char/random.c
633
.hash.h = { BLAKE2S_IV0 ^ (0x01010000 | BLAKE2S_HASH_SIZE),
drivers/char/random.c
636
.hash.outlen = BLAKE2S_HASH_SIZE,
drivers/char/random.c
642
blake2s_update(&input_pool.hash, buf, len);
drivers/char/random.c
690
blake2s_final(&input_pool.hash, seed);
drivers/char/random.c
695
blake2s_init_key(&input_pool.hash, BLAKE2S_HASH_SIZE, next_key, sizeof(next_key));
drivers/char/tpm/tpm.h
250
int tpm1_pcr_extend(struct tpm_chip *chip, u32 pcr_idx, const u8 *hash,
drivers/char/tpm/tpm1-cmd.c
463
int tpm1_pcr_extend(struct tpm_chip *chip, u32 pcr_idx, const u8 *hash,
drivers/char/tpm/tpm1-cmd.c
474
tpm_buf_append(&buf, hash, TPM_DIGEST_SIZE);
drivers/clk/clk.c
408
u32 hash;
drivers/clk/clk.c
413
hash = full_name_hash(NULL, name, strlen(name));
drivers/clk/clk.c
416
hash_for_each_possible(clk_hashtable, core, hashtable_node, hash)
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-core.c
258
ss_algs[i].alg.hash.halg.base.cra_driver_name,
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-core.c
259
ss_algs[i].alg.hash.halg.base.cra_name,
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-core.c
37
.alg.hash = {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-core.c
467
err = crypto_register_ahash(&ss_algs[i].alg.hash);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-core.c
470
ss_algs[i].alg.hash.halg.base.cra_name);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-core.c
498
crypto_unregister_ahash(&ss_algs[i].alg.hash);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-core.c
521
crypto_unregister_ahash(&ss_algs[i].alg.hash);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-core.c
63
.alg.hash = {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
116
octx->state[i] = op->hash[i];
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
142
op->hash[i] = ictx->state[i];
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
207
op->len, op->hash[0]);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
235
writel(op->hash[i], ss->base + SS_IV0 + i * 4);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
27
algt = container_of(alg, struct sun4i_ss_alg_template, alg.hash);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
386
op->hash[i] = readl(ss->base + SS_MD0 + i * 4);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
404
algt = container_of(alg, struct sun4i_ss_alg_template, alg.hash);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
56
algt = container_of(alg, struct sun4i_ss_alg_template, alg.hash);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
74
octx->hash[i] = op->hash[i];
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
76
octx->hash[0] = SHA1_H0;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
77
octx->hash[1] = SHA1_H1;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
78
octx->hash[2] = SHA1_H2;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
79
octx->hash[3] = SHA1_H3;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
99
op->hash[i] = ictx->hash[i];
drivers/crypto/allwinner/sun4i-ss/sun4i-ss.h
166
struct ahash_alg hash;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss.h
193
u32 hash[5]; /* for storing SS_IVx register */
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
425
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
451
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
458
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
484
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
490
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
516
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
522
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
548
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
554
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
580
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
586
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
612
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
681
ce_algs[i].alg.hash.base.halg.base.cra_driver_name,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
682
ce_algs[i].alg.hash.base.halg.base.cra_name,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
914
ce_algs[i].alg.hash.base.halg.base.cra_name);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
919
ce_algs[i].alg.hash.base.halg.base.cra_name);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
920
err = crypto_engine_register_ahash(&ce_algs[i].alg.hash);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
923
ce_algs[i].alg.hash.base.halg.base.cra_name);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
968
ce_algs[i].alg.hash.base.halg.base.cra_name);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
969
crypto_engine_unregister_ahash(&ce_algs[i].alg.hash);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
205
algt = container_of(alg, struct sun8i_ce_alg_template, alg.hash.base);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
33
alg.hash.base);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
333
algt = container_of(alg, struct sun8i_ce_alg_template, alg.hash.base);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
45
algt = container_of(alg, struct sun8i_ce_alg_template, alg.hash.base);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce.h
359
struct ahash_engine_alg hash;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
308
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
334
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
340
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
366
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
372
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
398
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
404
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
430
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
436
.alg.hash.base = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
463
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
512
ss_algs[i].alg.hash.base.halg.base.cra_driver_name,
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
513
ss_algs[i].alg.hash.base.halg.base.cra_name,
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
724
ss_algs[i].alg.hash.base.halg.base.cra_name);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
729
ss_algs[i].alg.hash.base.halg.base.cra_name);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
730
err = crypto_engine_register_ahash(&ss_algs[i].alg.hash);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
733
ss_algs[i].alg.hash.base.halg.base.cra_name);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
766
ss_algs[i].alg.hash.base.halg.base.cra_name);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
767
crypto_engine_unregister_ahash(&ss_algs[i].alg.hash);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
102
algt = container_of(alg, struct sun8i_ss_alg_template, alg.hash.base);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
205
alg.hash.base);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
248
alg.hash.base);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
276
alg.hash.base);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
352
algt = container_of(alg, struct sun8i_ss_alg_template, alg.hash.base);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
407
algt = container_of(alg, struct sun8i_ss_alg_template, alg.hash.base);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
503
algt = container_of(alg, struct sun8i_ss_alg_template, alg.hash.base);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss.h
278
struct ahash_engine_alg hash;
drivers/crypto/atmel-sha.c
523
const u32 *hash = (const u32 *)ctx->digest;
drivers/crypto/atmel-sha.c
534
atmel_sha_write(dd, SHA_REG_DIN(i), hash[i]);
drivers/crypto/atmel-sha.c
925
u32 *hash = (u32 *)ctx->digest;
drivers/crypto/atmel-sha.c
949
hash[i] = atmel_sha_read(ctx->dd, SHA_REG_DIGEST(i));
drivers/crypto/bcm/cipher.c
2016
struct crypto_shash *hash;
drivers/crypto/bcm/cipher.c
2027
hash = crypto_alloc_shash(alg_name, 0, 0);
drivers/crypto/bcm/cipher.c
2028
if (IS_ERR(hash)) {
drivers/crypto/bcm/cipher.c
2029
ret = PTR_ERR(hash);
drivers/crypto/bcm/cipher.c
2036
crypto_shash_descsize(hash), gfp);
drivers/crypto/bcm/cipher.c
2041
ctx->shash->tfm = hash;
drivers/crypto/bcm/cipher.c
2045
ret = crypto_shash_setkey(hash, ctx->authkey,
drivers/crypto/bcm/cipher.c
2065
crypto_free_shash(hash);
drivers/crypto/bcm/cipher.c
3671
.alg.hash = {
drivers/crypto/bcm/cipher.c
3692
.alg.hash = {
drivers/crypto/bcm/cipher.c
3710
.alg.hash = {
drivers/crypto/bcm/cipher.c
3728
.alg.hash = {
drivers/crypto/bcm/cipher.c
3746
.alg.hash = {
drivers/crypto/bcm/cipher.c
3764
.alg.hash = {
drivers/crypto/bcm/cipher.c
3782
.alg.hash = {
drivers/crypto/bcm/cipher.c
3800
.alg.hash = {
drivers/crypto/bcm/cipher.c
3819
.alg.hash = {
drivers/crypto/bcm/cipher.c
3838
.alg.hash = {
drivers/crypto/bcm/cipher.c
3857
.alg.hash = {
drivers/crypto/bcm/cipher.c
3876
.alg.hash = {
drivers/crypto/bcm/cipher.c
3895
.alg.hash = {
drivers/crypto/bcm/cipher.c
3914
.alg.hash = {
drivers/crypto/bcm/cipher.c
3933
.alg.hash = {
drivers/crypto/bcm/cipher.c
3952
.alg.hash = {
drivers/crypto/bcm/cipher.c
3971
.alg.hash = {
drivers/crypto/bcm/cipher.c
3990
.alg.hash = {
drivers/crypto/bcm/cipher.c
4009
.alg.hash = {
drivers/crypto/bcm/cipher.c
4028
.alg.hash = {
drivers/crypto/bcm/cipher.c
4047
.alg.hash = {
drivers/crypto/bcm/cipher.c
4066
.alg.hash = {
drivers/crypto/bcm/cipher.c
4133
alg.hash);
drivers/crypto/bcm/cipher.c
4397
struct ahash_alg *hash = &driver_alg->alg.hash;
drivers/crypto/bcm/cipher.c
4411
hash->halg.base.cra_module = THIS_MODULE;
drivers/crypto/bcm/cipher.c
4412
hash->halg.base.cra_priority = hash_pri;
drivers/crypto/bcm/cipher.c
4413
hash->halg.base.cra_alignmask = 0;
drivers/crypto/bcm/cipher.c
4414
hash->halg.base.cra_ctxsize = sizeof(struct iproc_ctx_s);
drivers/crypto/bcm/cipher.c
4415
hash->halg.base.cra_init = ahash_cra_init;
drivers/crypto/bcm/cipher.c
4416
hash->halg.base.cra_exit = generic_cra_exit;
drivers/crypto/bcm/cipher.c
4417
hash->halg.base.cra_flags = CRYPTO_ALG_ASYNC |
drivers/crypto/bcm/cipher.c
4419
hash->halg.statesize = sizeof(struct spu_hash_export_s);
drivers/crypto/bcm/cipher.c
4422
hash->init = ahash_init;
drivers/crypto/bcm/cipher.c
4423
hash->update = ahash_update;
drivers/crypto/bcm/cipher.c
4424
hash->final = ahash_final;
drivers/crypto/bcm/cipher.c
4425
hash->finup = ahash_finup;
drivers/crypto/bcm/cipher.c
4426
hash->digest = ahash_digest;
drivers/crypto/bcm/cipher.c
4430
hash->setkey = ahash_setkey;
drivers/crypto/bcm/cipher.c
4433
hash->setkey = ahash_hmac_setkey;
drivers/crypto/bcm/cipher.c
4434
hash->init = ahash_hmac_init;
drivers/crypto/bcm/cipher.c
4435
hash->update = ahash_hmac_update;
drivers/crypto/bcm/cipher.c
4436
hash->final = ahash_hmac_final;
drivers/crypto/bcm/cipher.c
4437
hash->finup = ahash_hmac_finup;
drivers/crypto/bcm/cipher.c
4438
hash->digest = ahash_hmac_digest;
drivers/crypto/bcm/cipher.c
4440
hash->export = ahash_export;
drivers/crypto/bcm/cipher.c
4441
hash->import = ahash_import;
drivers/crypto/bcm/cipher.c
4443
err = crypto_register_ahash(hash);
drivers/crypto/bcm/cipher.c
4448
hash->halg.base.cra_driver_name);
drivers/crypto/bcm/cipher.c
4522
crypto_unregister_ahash(&driver_algs[j].alg.hash);
drivers/crypto/bcm/cipher.c
4680
crypto_unregister_ahash(&driver_algs[i].alg.hash);
drivers/crypto/bcm/cipher.c
4681
cdn = driver_algs[i].alg.hash.halg.base.cra_driver_name;
drivers/crypto/bcm/cipher.h
109
struct ahash_alg hash;
drivers/crypto/bcm/util.c
217
struct crypto_shash *hash;
drivers/crypto/bcm/util.c
220
hash = crypto_alloc_shash(name, 0, 0);
drivers/crypto/bcm/util.c
221
if (IS_ERR(hash)) {
drivers/crypto/bcm/util.c
222
rc = PTR_ERR(hash);
drivers/crypto/bcm/util.c
227
size = sizeof(struct shash_desc) + crypto_shash_descsize(hash);
drivers/crypto/bcm/util.c
233
sdesc->shash.tfm = hash;
drivers/crypto/bcm/util.c
236
rc = crypto_shash_setkey(hash, key, key_len);
drivers/crypto/bcm/util.c
265
crypto_free_shash(hash);
drivers/crypto/caam/key_gen.h
16
static inline u32 split_key_len(u32 hash)
drivers/crypto/caam/key_gen.h
22
idx = (hash & OP_ALG_ALGSEL_SUBMASK) >> OP_ALG_ALGSEL_SHIFT;
drivers/crypto/caam/key_gen.h
34
static inline u32 split_key_pad_len(u32 hash)
drivers/crypto/caam/key_gen.h
36
return ALIGN(split_key_len(hash), 16);
drivers/crypto/caam/regs.h
872
struct rtic_memhash hash[4]; /* Block hash values A-D */
drivers/crypto/chelsio/chcr_algo.c
340
alg.hash);
drivers/crypto/chelsio/chcr_algo.c
3800
.alg.hash = {
drivers/crypto/chelsio/chcr_algo.c
3812
.alg.hash = {
drivers/crypto/chelsio/chcr_algo.c
3824
.alg.hash = {
drivers/crypto/chelsio/chcr_algo.c
3836
.alg.hash = {
drivers/crypto/chelsio/chcr_algo.c
3848
.alg.hash = {
drivers/crypto/chelsio/chcr_algo.c
3861
.alg.hash = {
drivers/crypto/chelsio/chcr_algo.c
3873
.alg.hash = {
drivers/crypto/chelsio/chcr_algo.c
3885
.alg.hash = {
drivers/crypto/chelsio/chcr_algo.c
3897
.alg.hash = {
drivers/crypto/chelsio/chcr_algo.c
3909
.alg.hash = {
drivers/crypto/chelsio/chcr_algo.c
4279
&driver_algs[i].alg.hash.halg.base.cra_refcnt)
drivers/crypto/chelsio/chcr_algo.c
4282
&driver_algs[i].alg.hash);
drivers/crypto/chelsio/chcr_algo.c
4338
a_hash = &driver_algs[i].alg.hash;
drivers/crypto/chelsio/chcr_algo.c
4363
err = crypto_register_ahash(&driver_algs[i].alg.hash);
drivers/crypto/chelsio/chcr_algo.c
4364
ai = driver_algs[i].alg.hash.halg.base;
drivers/crypto/chelsio/chcr_crypto.h
319
struct ahash_alg hash;
drivers/crypto/img-hash.c
286
__be32 *hash = (__be32 *)ctx->digest;
drivers/crypto/img-hash.c
289
for (i = (ctx->digsize / sizeof(*hash)) - 1; i >= 0; i--)
drivers/crypto/img-hash.c
290
hash[i] = img_hash_read_result_queue(ctx->hdev);
drivers/crypto/inside-secure/eip93/eip93-hash.c
116
switch (hash) {
drivers/crypto/inside-secure/eip93/eip93-hash.c
98
static void eip93_hash_init_sa_state_digest(u32 hash, u8 *digest)
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
1418
.hash = &hash_alg_md5,
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
1431
.hash = &hash_alg_md5,
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
1443
.hash = &hash_alg_sha1,
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
1456
.hash = &hash_alg_sha1,
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
1468
.hash = &hash_alg_md5,
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
1480
.hash = &hash_alg_sha1,
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
195
const struct ix_hash_algo *hash;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
204
const struct ix_hash_algo *hash;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
259
return container_of(tfm->__crt_alg, struct ixp_alg, crypto.base)->hash;
drivers/crypto/intel/qat/qat_common/qat_algs.c
111
__be32 *istate = (__be32 *)hash->sha.state1;
drivers/crypto/intel/qat/qat_common/qat_algs.c
112
__be32 *ostate = (__be32 *)(hash->sha.state1 +
drivers/crypto/intel/qat/qat_common/qat_algs.c
125
__be64 *istate = (__be64 *)hash->sha.state1;
drivers/crypto/intel/qat/qat_common/qat_algs.c
126
__be64 *ostate = (__be64 *)(hash->sha.state1 +
drivers/crypto/intel/qat/qat_common/qat_algs.c
169
struct icp_qat_hw_auth_algo_blk *hash =
drivers/crypto/intel/qat/qat_common/qat_algs.c
182
hash->sha.inner_setup.auth_config.config =
drivers/crypto/intel/qat/qat_common/qat_algs.c
185
hash->sha.inner_setup.auth_counter.counter =
drivers/crypto/intel/qat/qat_common/qat_algs.c
188
if (qat_alg_do_precomputes(hash, ctx, keys->authkey, keys->authkeylen))
drivers/crypto/intel/qat/qat_common/qat_algs.c
210
hash_cd_ctrl->hash_cfg_offset = ((char *)hash - (char *)cipher) >> 3;
drivers/crypto/intel/qat/qat_common/qat_algs.c
249
struct icp_qat_hw_auth_algo_blk *hash = &dec_ctx->hash;
drivers/crypto/intel/qat/qat_common/qat_algs.c
268
hash->sha.inner_setup.auth_config.config =
drivers/crypto/intel/qat/qat_common/qat_algs.c
272
hash->sha.inner_setup.auth_counter.counter =
drivers/crypto/intel/qat/qat_common/qat_algs.c
275
if (qat_alg_do_precomputes(hash, ctx, keys->authkey, keys->authkeylen))
drivers/crypto/intel/qat/qat_common/qat_algs.c
53
struct icp_qat_hw_auth_algo_blk hash;
drivers/crypto/intel/qat/qat_common/qat_algs.c
56
struct icp_qat_hw_auth_algo_blk hash;
drivers/crypto/intel/qat/qat_common/qat_algs.c
89
static int qat_alg_do_precomputes(struct icp_qat_hw_auth_algo_blk *hash,
drivers/crypto/intel/qat/qat_common/qat_algs.c
97
__be32 *istate = (__be32 *)hash->sha.state1;
drivers/crypto/intel/qat/qat_common/qat_algs.c
98
__be32 *ostate = (__be32 *)(hash->sha.state1 +
drivers/crypto/marvell/cesa/cesa.h
253
__le32 hash[8];
drivers/crypto/marvell/cesa/cesa.h
267
struct mv_cesa_hash_op_ctx hash;
drivers/crypto/marvell/cesa/hash.c
1250
memcpy(tmpl.ctx.hash.iv, ctx->iv, sizeof(ctx->iv));
drivers/crypto/marvell/cesa/hash.c
1268
for (i = 0; i < ARRAY_SIZE(istate.hash); i++)
drivers/crypto/marvell/cesa/hash.c
1269
ctx->iv[i] = cpu_to_be32(istate.hash[i]);
drivers/crypto/marvell/cesa/hash.c
1271
for (i = 0; i < ARRAY_SIZE(ostate.hash); i++)
drivers/crypto/marvell/cesa/hash.c
1272
ctx->iv[i + 8] = cpu_to_be32(ostate.hash[i]);
drivers/crypto/marvell/cesa/hash.c
1321
memcpy(tmpl.ctx.hash.iv, ctx->iv, sizeof(ctx->iv));
drivers/crypto/marvell/cesa/hash.c
1412
memcpy(tmpl.ctx.hash.iv, ctx->iv, sizeof(ctx->iv));
drivers/crypto/marvell/cesa/hash.c
374
data = creq->base.chain.last->op->ctx.hash.hash;
drivers/crypto/marvell/cesa/hash.c
838
static int mv_cesa_ahash_export(struct ahash_request *req, void *hash,
drivers/crypto/marvell/cesa/hash.c
849
memcpy(hash, creq->state, digsize);
drivers/crypto/marvell/cesa/hash.c
856
static int mv_cesa_ahash_import(struct ahash_request *req, const void *hash,
drivers/crypto/marvell/cesa/hash.c
877
memcpy(creq->state, hash, digsize);
drivers/crypto/marvell/cesa/hash.c
911
return mv_cesa_ahash_export(req, out_state->hash,
drivers/crypto/marvell/cesa/hash.c
919
return mv_cesa_ahash_import(req, in_state->hash, in_state->byte_count,
drivers/crypto/omap-sham.c
287
u32 *hash = (u32 *)ctx->digest;
drivers/crypto/omap-sham.c
292
hash[i] = omap_sham_read(dd, SHA_REG_IDIGEST(dd, i));
drivers/crypto/omap-sham.c
294
omap_sham_write(dd, SHA_REG_IDIGEST(dd, i), hash[i]);
drivers/crypto/omap-sham.c
327
u32 *hash = (u32 *)req->result;
drivers/crypto/omap-sham.c
330
if (!hash)
drivers/crypto/omap-sham.c
361
put_unaligned(be32_to_cpup((__be32 *)in + i), &hash[i]);
drivers/crypto/omap-sham.c
364
put_unaligned(le32_to_cpup((__le32 *)in + i), &hash[i]);
drivers/crypto/padlock-sha.c
158
static int padlock_init_tfm(struct crypto_shash *hash)
drivers/crypto/padlock-sha.c
160
const char *fallback_driver_name = crypto_shash_alg_name(hash);
drivers/crypto/padlock-sha.c
161
struct padlock_sha_ctx *ctx = crypto_shash_ctx(hash);
drivers/crypto/padlock-sha.c
174
if (crypto_shash_statesize(hash) !=
drivers/crypto/padlock-sha.c
185
static void padlock_exit_tfm(struct crypto_shash *hash)
drivers/crypto/padlock-sha.c
187
struct padlock_sha_ctx *ctx = crypto_shash_ctx(hash);
drivers/crypto/rockchip/rk3288_crypto.c
225
rk_cipher_algs[i]->alg.hash.base.halg.base.cra_driver_name,
drivers/crypto/rockchip/rk3288_crypto.c
226
rk_cipher_algs[i]->alg.hash.base.halg.base.cra_name,
drivers/crypto/rockchip/rk3288_crypto.c
268
rk_cipher_algs[i]->alg.hash.base.halg.base.cra_name,
drivers/crypto/rockchip/rk3288_crypto.c
269
rk_cipher_algs[i]->alg.hash.base.halg.base.cra_driver_name);
drivers/crypto/rockchip/rk3288_crypto.c
270
err = crypto_engine_register_ahash(&rk_cipher_algs[i]->alg.hash);
drivers/crypto/rockchip/rk3288_crypto.c
285
crypto_engine_unregister_ahash(&rk_cipher_algs[i]->alg.hash);
drivers/crypto/rockchip/rk3288_crypto.c
298
crypto_engine_unregister_ahash(&rk_cipher_algs[i]->alg.hash);
drivers/crypto/rockchip/rk3288_crypto.h
263
struct ahash_engine_alg hash;
drivers/crypto/rockchip/rk3288_crypto_ahash.c
266
struct rk_crypto_tmp *algt = container_of(alg, struct rk_crypto_tmp, alg.hash.base);
drivers/crypto/rockchip/rk3288_crypto_ahash.c
351
struct rk_crypto_tmp *algt = container_of(alg, struct rk_crypto_tmp, alg.hash.base);
drivers/crypto/rockchip/rk3288_crypto_ahash.c
377
.alg.hash.base = {
drivers/crypto/rockchip/rk3288_crypto_ahash.c
402
.alg.hash.op = {
drivers/crypto/rockchip/rk3288_crypto_ahash.c
409
.alg.hash.base = {
drivers/crypto/rockchip/rk3288_crypto_ahash.c
434
.alg.hash.op = {
drivers/crypto/rockchip/rk3288_crypto_ahash.c
441
.alg.hash.base = {
drivers/crypto/rockchip/rk3288_crypto_ahash.c
466
.alg.hash.op = {
drivers/crypto/rockchip/rk3288_crypto_ahash.c
50
struct rk_crypto_tmp *algt = container_of(alg, struct rk_crypto_tmp, alg.hash.base);
drivers/crypto/s5p-sss.c
770
u32 *hash = (u32 *)ctx->digest;
drivers/crypto/s5p-sss.c
774
hash[i] = s5p_hash_read(dd, SSS_REG_HASH_OUT(i));
drivers/crypto/s5p-sss.c
785
const u32 *hash = (const u32 *)ctx->digest;
drivers/crypto/s5p-sss.c
789
s5p_hash_write(dd, SSS_REG_HASH_IV(i), hash[i]);
drivers/crypto/sa2ul.c
1699
static int sa_cra_init_aead(struct crypto_aead *tfm, const char *hash,
drivers/crypto/sa2ul.c
1709
ctx->shash = crypto_alloc_shash(hash, 0, CRYPTO_ALG_NEED_FALLBACK);
drivers/crypto/sa2ul.c
1711
dev_err(sa_k3_dev, "base driver %s couldn't be loaded\n", hash);
drivers/crypto/sa2ul.c
408
static void sa_export_shash(void *state, struct shash_desc *hash,
drivers/crypto/sa2ul.c
430
crypto_shash_export(hash, state);
drivers/crypto/starfive/jh7110-cryp.h
212
union starfive_hash_csr hash;
drivers/crypto/starfive/jh7110-hash.c
210
rctx->csr.hash.v = 0;
drivers/crypto/starfive/jh7110-hash.c
211
rctx->csr.hash.mode = ctx->hash_mode;
drivers/crypto/starfive/jh7110-hash.c
218
rctx->csr.hash.start = 1;
drivers/crypto/starfive/jh7110-hash.c
219
rctx->csr.hash.firstb = 1;
drivers/crypto/starfive/jh7110-hash.c
220
writel(rctx->csr.hash.v, cryp->base + STARFIVE_HASH_SHACSR);
drivers/crypto/starfive/jh7110-hash.c
374
static int starfive_hash_init_tfm(struct crypto_ahash *hash,
drivers/crypto/starfive/jh7110-hash.c
379
struct starfive_cryp_ctx *ctx = crypto_ahash_ctx(hash);
drivers/crypto/starfive/jh7110-hash.c
393
crypto_ahash_set_statesize(hash, crypto_ahash_statesize(ctx->ahash_fbk));
drivers/crypto/starfive/jh7110-hash.c
394
crypto_ahash_set_reqsize(hash, sizeof(struct starfive_cryp_request_ctx) +
drivers/crypto/starfive/jh7110-hash.c
403
static void starfive_hash_exit_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
405
struct starfive_cryp_ctx *ctx = crypto_ahash_ctx(hash);
drivers/crypto/starfive/jh7110-hash.c
456
static int starfive_hash_setkey(struct crypto_ahash *hash,
drivers/crypto/starfive/jh7110-hash.c
459
struct starfive_cryp_ctx *ctx = crypto_ahash_ctx(hash);
drivers/crypto/starfive/jh7110-hash.c
460
unsigned int digestsize = crypto_ahash_digestsize(hash);
drivers/crypto/starfive/jh7110-hash.c
461
unsigned int blocksize = crypto_ahash_blocksize(hash);
drivers/crypto/starfive/jh7110-hash.c
497
static int starfive_sha224_init_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
499
return starfive_hash_init_tfm(hash, "sha224-lib",
drivers/crypto/starfive/jh7110-hash.c
503
static int starfive_sha256_init_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
505
return starfive_hash_init_tfm(hash, "sha256-lib",
drivers/crypto/starfive/jh7110-hash.c
509
static int starfive_sha384_init_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
511
return starfive_hash_init_tfm(hash, "sha384-lib",
drivers/crypto/starfive/jh7110-hash.c
515
static int starfive_sha512_init_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
517
return starfive_hash_init_tfm(hash, "sha512-lib",
drivers/crypto/starfive/jh7110-hash.c
521
static int starfive_sm3_init_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
523
return starfive_hash_init_tfm(hash, "sm3-generic",
drivers/crypto/starfive/jh7110-hash.c
527
static int starfive_hmac_sha224_init_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
529
return starfive_hash_init_tfm(hash, "hmac-sha224-lib",
drivers/crypto/starfive/jh7110-hash.c
533
static int starfive_hmac_sha256_init_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
535
return starfive_hash_init_tfm(hash, "hmac-sha256-lib",
drivers/crypto/starfive/jh7110-hash.c
539
static int starfive_hmac_sha384_init_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
541
return starfive_hash_init_tfm(hash, "hmac-sha384-lib",
drivers/crypto/starfive/jh7110-hash.c
545
static int starfive_hmac_sha512_init_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
547
return starfive_hash_init_tfm(hash, "hmac-sha512-lib",
drivers/crypto/starfive/jh7110-hash.c
551
static int starfive_hmac_sm3_init_tfm(struct crypto_ahash *hash)
drivers/crypto/starfive/jh7110-hash.c
553
return starfive_hash_init_tfm(hash, "hmac(sm3-generic)",
drivers/crypto/starfive/jh7110-hash.c
74
rctx->csr.hash.hmac = 1;
drivers/crypto/starfive/jh7110-hash.c
75
rctx->csr.hash.key_flag = 1;
drivers/crypto/starfive/jh7110-hash.c
77
writel(rctx->csr.hash.v, cryp->base + STARFIVE_HASH_SHACSR);
drivers/crypto/stm32/stm32-hash.c
962
__be32 *hash = (void *)rctx->digest;
drivers/crypto/stm32/stm32-hash.c
972
hash[i] = cpu_to_be32(stm32_hash_read(hdev,
drivers/crypto/stm32/stm32-hash.c
975
hash[i] = cpu_to_be32(stm32_hash_read(hdev,
drivers/crypto/talitos.c
2188
u8 *hash)
drivers/crypto/talitos.c
2210
ahash_request_set_crypt(req, sg, hash, keylen);
drivers/crypto/talitos.c
2227
u8 hash[SHA512_DIGEST_SIZE];
drivers/crypto/talitos.c
2234
ret = keyhash(tfm, key, keylen, hash);
drivers/crypto/talitos.c
2240
memcpy(ctx->key, hash, digestsize);
drivers/crypto/talitos.c
2258
struct ahash_alg hash;
drivers/crypto/talitos.c
2849
.alg.hash = {
drivers/crypto/talitos.c
2865
.alg.hash = {
drivers/crypto/talitos.c
2881
.alg.hash = {
drivers/crypto/talitos.c
2897
.alg.hash = {
drivers/crypto/talitos.c
2913
.alg.hash = {
drivers/crypto/talitos.c
2929
.alg.hash = {
drivers/crypto/talitos.c
2945
.alg.hash = {
drivers/crypto/talitos.c
2961
.alg.hash = {
drivers/crypto/talitos.c
2977
.alg.hash = {
drivers/crypto/talitos.c
2993
.alg.hash = {
drivers/crypto/talitos.c
3009
.alg.hash = {
drivers/crypto/talitos.c
3025
.alg.hash = {
drivers/crypto/talitos.c
3102
algt.alg.hash);
drivers/crypto/talitos.c
3156
crypto_unregister_ahash(&t_alg->algt.alg.hash);
drivers/crypto/talitos.c
3222
alg = &t_alg->algt.alg.hash.halg.base;
drivers/crypto/talitos.c
3225
t_alg->algt.alg.hash.init = ahash_init;
drivers/crypto/talitos.c
3226
t_alg->algt.alg.hash.update = ahash_update;
drivers/crypto/talitos.c
3227
t_alg->algt.alg.hash.final = ahash_final;
drivers/crypto/talitos.c
3228
t_alg->algt.alg.hash.finup = ahash_finup;
drivers/crypto/talitos.c
3229
t_alg->algt.alg.hash.digest = ahash_digest;
drivers/crypto/talitos.c
3231
t_alg->algt.alg.hash.setkey = ahash_setkey;
drivers/crypto/talitos.c
3232
t_alg->algt.alg.hash.import = ahash_import;
drivers/crypto/talitos.c
3233
t_alg->algt.alg.hash.export = ahash_export;
drivers/crypto/talitos.c
3243
t_alg->algt.alg.hash.init = ahash_init_sha224_swinit;
drivers/crypto/talitos.c
3244
t_alg->algt.alg.hash.digest =
drivers/crypto/talitos.c
3518
&t_alg->algt.alg.hash);
drivers/crypto/talitos.c
3519
alg = &t_alg->algt.alg.hash.halg.base;
drivers/crypto/xilinx/zynqmp-sha.c
44
static int zynqmp_sha_init_tfm(struct crypto_shash *hash)
drivers/crypto/xilinx/zynqmp-sha.c
46
const char *fallback_driver_name = crypto_shash_alg_name(hash);
drivers/crypto/xilinx/zynqmp-sha.c
47
struct zynqmp_sha_tfm_ctx *tfm_ctx = crypto_shash_ctx(hash);
drivers/crypto/xilinx/zynqmp-sha.c
48
struct shash_alg *alg = crypto_shash_alg(hash);
drivers/crypto/xilinx/zynqmp-sha.c
61
if (crypto_shash_descsize(hash) <
drivers/crypto/xilinx/zynqmp-sha.c
72
static void zynqmp_sha_exit_tfm(struct crypto_shash *hash)
drivers/crypto/xilinx/zynqmp-sha.c
74
struct zynqmp_sha_tfm_ctx *tfm_ctx = crypto_shash_ctx(hash);
drivers/dpll/zl3073x/flash.c
464
u32 hash;
drivers/dpll/zl3073x/flash.c
467
rc = zl3073x_read_u32(zldev, ZL_REG_FLASH_HASH, &hash);
drivers/dpll/zl3073x/flash.c
481
hash, family, release);
drivers/edac/igen6_edac.c
428
u64 hash_addr = addr & mask, hash = hash_init;
drivers/edac/igen6_edac.c
433
hash ^= (hash_addr >> i) & 1;
drivers/edac/igen6_edac.c
435
return hash ^ intlv;
drivers/edac/igen6_edac.c
440
u64 maddr, hash, mask, ms_s_size;
drivers/edac/igen6_edac.c
456
hash = mem_slice_hash(maddr, mask, mc, intlv_bit);
drivers/edac/igen6_edac.c
458
return maddr | (hash << intlv_bit);
drivers/edac/igen6_edac.c
749
u64 hash_addr = addr & mask, hash = 0;
drivers/edac/igen6_edac.c
754
hash ^= (hash_addr >> i) & 1;
drivers/edac/igen6_edac.c
756
return (int)hash ^ intlv;
drivers/edac/igen6_edac.c
770
static void decode_addr(u64 addr, u32 hash, u64 s_size, int l_map,
drivers/edac/igen6_edac.c
773
int intlv_bit = CHANNEL_HASH_LSB_MASK_BIT(hash) + 6;
drivers/edac/igen6_edac.c
781
if (CHANNEL_HASH_MODE(hash)) {
drivers/edac/igen6_edac.c
783
*idx = decode_chan_idx(addr, CHANNEL_HASH_MASK(hash), intlv_bit);
drivers/edac/igen6_edac.c
795
u32 hash;
drivers/edac/igen6_edac.c
803
hash = readl(imc->window + CHANNEL_HASH_OFFSET);
drivers/edac/igen6_edac.c
806
decode_addr(addr, hash, s_size, l_map, &idx, &sub_addr);
drivers/edac/igen6_edac.c
811
hash = readl(imc->window + CHANNEL_EHASH_OFFSET);
drivers/edac/igen6_edac.c
814
decode_addr(res->channel_addr, hash, s_size, l_map, &idx, &sub_addr);
drivers/firmware/arm_scmi/driver.c
196
unsigned long hash = 0;
drivers/firmware/arm_scmi/driver.c
206
hash = partial_name_hash(tolower(*p++), hash);
drivers/firmware/arm_scmi/driver.c
207
hash = end_name_hash(hash);
drivers/firmware/arm_scmi/driver.c
211
return hash;
drivers/firmware/arm_scmi/notify.c
131
hash_for_each_possible((__ht), obj_, hash, k_) \
drivers/firmware/arm_scmi/notify.c
1594
hash_for_each_safe(ni->pending_events_handlers, bkt, tmp, hndl, hash) {
drivers/firmware/arm_scmi/notify.c
369
struct hlist_node hash;
drivers/firmware/arm_scmi/notify.c
907
hash_add(ni->pending_events_handlers, &hndl->hash, hndl->key);
drivers/firmware/arm_scmi/notify.c
921
hash_del(&hndl->hash);
drivers/firmware/arm_scmi/notify.c
951
hash_del(&hndl->hash);
drivers/firmware/arm_scmi/notify.c
969
&hndl->hash, hndl->key);
drivers/firmware/arm_scmi/perf.c
174
hash_for_each_possible((__htp), _opp, hash, f_) \
drivers/firmware/arm_scmi/perf.c
428
hash_add(dom->opps_by_freq, &opp->hash, opp->indicative_freq);
drivers/firmware/arm_scmi/perf.c
57
struct hlist_node hash;
drivers/firmware/arm_scmi/quirks.c
200
unsigned long hash = 0;
drivers/firmware/arm_scmi/quirks.c
211
hash = partial_name_hash(tolower(*p++), hash);
drivers/firmware/arm_scmi/quirks.c
212
hash32 = end_name_hash(hash);
drivers/firmware/arm_scmi/quirks.c
277
hash_add(scmi_quirks_ht, &quirk->hash, quirk->hkey);
drivers/firmware/arm_scmi/quirks.c
304
hash_for_each_possible(scmi_quirks_ht, quirk, hash, hkey) {
drivers/firmware/arm_scmi/quirks.c
99
struct hlist_node hash;
drivers/firmware/efi/embedded-firmware.c
41
u8 hash[32];
drivers/firmware/efi/embedded-firmware.c
56
sha256(map + i, desc->length, hash);
drivers/firmware/efi/embedded-firmware.c
57
if (memcmp(hash, desc->sha256, 32) == 0)
drivers/firmware/google/gsmi.c
713
u64 hash = val;
drivers/firmware/google/gsmi.c
716
u64 n = hash;
drivers/firmware/google/gsmi.c
718
hash -= n;
drivers/firmware/google/gsmi.c
720
hash -= n;
drivers/firmware/google/gsmi.c
722
hash += n;
drivers/firmware/google/gsmi.c
724
hash -= n;
drivers/firmware/google/gsmi.c
726
hash += n;
drivers/firmware/google/gsmi.c
728
hash += n;
drivers/firmware/google/gsmi.c
731
return hash >> (64 - bits);
drivers/firmware/google/gsmi.c
760
u32 hash;
drivers/firmware/google/gsmi.c
780
hash = hash_oem_table_id(acpi_gbl_FADT.header.oem_table_id);
drivers/firmware/google/gsmi.c
781
if (hash == QUIRKY_BOARD_HASH) {
drivers/fpga/intel-m10-bmc-sec-update.c
112
u8 hash[REH_SHA384_SIZE];
drivers/fpga/intel-m10-bmc-sec-update.c
130
ret = m10bmc_sec_read(sec, hash, reh_addr, sha_num_bytes);
drivers/fpga/intel-m10-bmc-sec-update.c
137
cnt += sprintf(buf + cnt, "%02x", hash[i]);
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
424
uint32_t hash;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
437
hash = hash_64(key, AMDGPU_GMC_FAULT_HASH_ORDER);
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
438
fault = &gmc->fault_ring[gmc->fault_hash[hash].idx];
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
472
fault->next = gmc->fault_hash[hash].idx;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
473
gmc->fault_hash[hash].idx = gmc->last_fault++;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
496
uint32_t hash;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
515
hash = hash_64(key, AMDGPU_GMC_FAULT_HASH_ORDER);
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
516
fault = &gmc->fault_ring[gmc->fault_hash[hash].idx];
drivers/gpu/drm/nouveau/include/nvkm/subdev/fsp.h
19
u64 img_addr, const u8 *hash, const u8 *pkey, const u8 *sig);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
129
u8 *hash;
drivers/gpu/drm/nouveau/nvkm/core/ramht.c
29
u32 hash = 0;
drivers/gpu/drm/nouveau/nvkm/core/ramht.c
32
hash ^= (handle & ((1 << ramht->bits) - 1));
drivers/gpu/drm/nouveau/nvkm/core/ramht.c
36
hash ^= chid << (ramht->bits - 4);
drivers/gpu/drm/nouveau/nvkm/core/ramht.c
37
return hash;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
100
if (object->hash < 0)
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
101
return object->hash;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
62
int hash;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
70
nvkm_ramht_remove(object->disp->ramht, object->hash);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
99
object->hash = chan->func->bind(chan, object->oproxy.object, oclass->handle);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv04.c
181
nv04_eobj_ramht_del(struct nvkm_chan *chan, int hash)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv04.c
187
nvkm_ramht_remove(imem->ramht, hash);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv04.c
197
int hash;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv04.c
200
hash = nvkm_ramht_insert(imem->ramht, eobj, chan->id, 4, eobj->handle, context);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv04.c
202
return hash;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv40.c
116
int hash;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv40.c
119
hash = nvkm_ramht_insert(imem->ramht, eobj, chan->id, 4, eobj->handle, context);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv40.c
121
return hash;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
36
nv50_eobj_ramht_del(struct nvkm_chan *chan, int hash)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
38
nvkm_ramht_remove(chan->ramht, hash);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
23
void (*ramht_del)(struct nvkm_chan *, int hash);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
141
engn->func->ramht_del(uobj->chan, uobj->hash);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
195
uobj->hash = engn->func->ramht_add(engn, uobj->oproxy.object, uobj->chan);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
196
if (uobj->hash < 0)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
197
return uobj->hash;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
71
int hash;
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/base.c
12
img_addr, hash, pkey, sig);
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/base.c
9
u64 img_addr, const u8 *hash, const u8 *pkey, const u8 *sig)
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/gh100.c
206
u64 img_addr, const u8 *hash, const u8 *pkey, const u8 *sig)
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/gh100.c
231
memcpy(msg.cot.hash384, hash, fsp->func->cot.size_hash);
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/priv.h
19
u64 img_addr, const u8 *hash, const u8 *pkey, const u8 *sig);
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/priv.h
28
u64 img_addr, const u8 *hash, const u8 *pkey, const u8 *sig);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
106
gsp->fmc.fw.addr, gsp->fmc.hash, gsp->fmc.pkey, gsp->fmc.sig);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
269
const void *hash, *sig, *pkey, *img;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
280
hash = elf_section(fw, "hash", &hash_len);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
285
if (!hash || !sig || !pkey || !img) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
300
gsp->fmc.hash = kmemdup(hash, hash_len, GFP_KERNEL);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
303
if (!gsp->fmc.hash || !gsp->fmc.pkey || !gsp->fmc.sig)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
2104
kfree(gsp->fmc.hash);
drivers/gpu/drm/qxl/qxl_image.c
107
int depth, unsigned int hash,
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
12
static u32 hgsmi_hash_process(u32 hash, const u8 *data, int size)
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
15
hash += *data++;
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
16
hash += (hash << 10);
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
17
hash ^= (hash >> 6);
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
20
return hash;
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
23
static u32 hgsmi_hash_end(u32 hash)
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
25
hash += (hash << 3);
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
26
hash ^= (hash >> 11);
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
27
hash += (hash << 15);
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
29
return hash;
drivers/gpu/drm/vmwgfx/ttm_object.c
123
struct vmwgfx_hash_item hash;
drivers/gpu/drm/vmwgfx/ttm_object.c
143
struct vmwgfx_hash_item *hash;
drivers/gpu/drm/vmwgfx/ttm_object.c
145
hash_for_each_possible_rcu(tfile->ref_hash, hash, head, key) {
drivers/gpu/drm/vmwgfx/ttm_object.c
146
if (hash->key == key) {
drivers/gpu/drm/vmwgfx/ttm_object.c
147
*p_hash = hash;
drivers/gpu/drm/vmwgfx/ttm_object.c
158
struct vmwgfx_hash_item *hash;
drivers/gpu/drm/vmwgfx/ttm_object.c
160
hash_for_each_possible(tfile->ref_hash, hash, head, key) {
drivers/gpu/drm/vmwgfx/ttm_object.c
161
if (hash->key == key) {
drivers/gpu/drm/vmwgfx/ttm_object.c
162
*p_hash = hash;
drivers/gpu/drm/vmwgfx/ttm_object.c
258
struct vmwgfx_hash_item *hash;
drivers/gpu/drm/vmwgfx/ttm_object.c
262
ret = ttm_tfile_find_ref(tfile, key, &hash);
drivers/gpu/drm/vmwgfx/ttm_object.c
265
base = hlist_entry(hash, struct ttm_ref_object, hash)->obj;
drivers/gpu/drm/vmwgfx/ttm_object.c
296
struct vmwgfx_hash_item *hash;
drivers/gpu/drm/vmwgfx/ttm_object.c
307
ret = ttm_tfile_find_ref_rcu(tfile, base->handle, &hash);
drivers/gpu/drm/vmwgfx/ttm_object.c
310
ref = hlist_entry(hash, struct ttm_ref_object, hash);
drivers/gpu/drm/vmwgfx/ttm_object.c
326
ref->hash.key = base->handle;
drivers/gpu/drm/vmwgfx/ttm_object.c
332
hash_add_rcu(tfile->ref_hash, &ref->hash.head, ref->hash.key);
drivers/gpu/drm/vmwgfx/ttm_object.c
352
hash_del_rcu(&ref->hash.head);
drivers/gpu/drm/vmwgfx/ttm_object.c
365
struct vmwgfx_hash_item *hash;
drivers/gpu/drm/vmwgfx/ttm_object.c
369
ret = ttm_tfile_find_ref(tfile, key, &hash);
drivers/gpu/drm/vmwgfx/ttm_object.c
374
ref = hlist_entry(hash, struct ttm_ref_object, hash);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
109
hash_del_rcu(&entry->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
170
hash_add_rcu(entry->man->resources, &entry->hash.head,
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
171
entry->hash.key);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
207
cres->hash.key = user_key | (res_type << 24);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
208
hash_add_rcu(man->resources, &cres->hash.head, cres->hash.key);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
241
struct vmwgfx_hash_item *hash;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
244
hash_for_each_possible_rcu(man->resources, hash, head, key) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
245
if (hash->key == key) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
246
entry = hlist_entry(hash, struct vmw_cmdbuf_res, hash);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
259
hash_del_rcu(&entry->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
47
struct vmwgfx_hash_item hash;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
86
struct vmwgfx_hash_item *hash;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
89
hash_for_each_possible_rcu(man->resources, hash, head, key) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
90
if (hash->key == key)
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
91
return hlist_entry(hash, struct vmw_cmdbuf_res, hash)->res;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
166
struct vmwgfx_hash_item *hash;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
169
hash_for_each_possible_rcu(ctx->sw_context->res_ht, hash, head, key) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
170
if (hash->key == key) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
171
bo_node = container_of(hash, typeof(*bo_node), hash);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
208
struct vmwgfx_hash_item *hash;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
211
hash_for_each_possible_rcu(ctx->sw_context->res_ht, hash, head, key) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
212
if (hash->key == key) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
213
res_node = container_of(hash, typeof(*res_node), hash);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
260
bo_node->hash.key = (unsigned long) vbo;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
261
hash_add_rcu(ctx->sw_context->res_ht, &bo_node->hash.head,
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
262
bo_node->hash.key);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
307
node->hash.key = (unsigned long) res;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
308
hash_add_rcu(ctx->sw_context->res_ht, &node->hash.head, node->hash.key);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
312
hash_del_rcu(&node->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
47
struct vmwgfx_hash_item hash;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
635
hash_del_rcu(&entry->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
638
hash_del_rcu(&val->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
641
hash_del_rcu(&val->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
74
struct vmwgfx_hash_item hash;
drivers/hwtracing/coresight/coresight-etm-perf.c
864
unsigned long hash;
drivers/hwtracing/coresight/coresight-etm-perf.c
881
hash = hashlen_hash(hashlen_string(NULL, name));
drivers/hwtracing/coresight/coresight-etm-perf.c
889
ea->var = (unsigned long *)hash;
drivers/hwtracing/coresight/coresight-sysfs.c
174
u32 hash;
drivers/hwtracing/coresight/coresight-sysfs.c
245
hash = hashlen_hash(hashlen_string(NULL, dev_name(&csdev->dev)));
drivers/hwtracing/coresight/coresight-sysfs.c
246
ret = idr_alloc_u32(&path_idr, path, &hash, hash, GFP_KERNEL);
drivers/hwtracing/coresight/coresight-sysfs.c
272
u32 hash;
drivers/hwtracing/coresight/coresight-sysfs.c
292
hash = hashlen_hash(hashlen_string(NULL, dev_name(&csdev->dev)));
drivers/hwtracing/coresight/coresight-sysfs.c
294
path = idr_find(&path_idr, hash);
drivers/hwtracing/coresight/coresight-sysfs.c
299
idr_remove(&path_idr, hash);
drivers/infiniband/core/cma.c
3285
u32 hash, fl;
drivers/infiniband/core/cma.c
3292
hash = (u32)sport * 31 + dport;
drivers/infiniband/core/cma.c
3293
fl = hash & IB_GRH_FLOWLABEL_MASK;
drivers/infiniband/core/iwpm_util.c
517
u32 hash = jhash_2words(ipv6_hash, (__force u32) ipv6_sockaddr->sin6_port, 0);
drivers/infiniband/core/iwpm_util.c
518
return hash;
drivers/infiniband/core/iwpm_util.c
524
u32 hash = jhash_2words(ipv4_hash, (__force u32) ipv4_sockaddr->sin_port, 0);
drivers/infiniband/core/iwpm_util.c
525
return hash;
drivers/infiniband/core/iwpm_util.c
529
struct sockaddr_storage *b_sockaddr, u32 *hash)
drivers/infiniband/core/iwpm_util.c
546
*hash = a_hash;
drivers/infiniband/core/iwpm_util.c
548
*hash = jhash_2words(a_hash, b_hash, 0);
drivers/infiniband/core/iwpm_util.c
556
u32 hash;
drivers/infiniband/core/iwpm_util.c
559
ret = get_hash_bucket(local_sockaddr, mapped_sockaddr, &hash);
drivers/infiniband/core/iwpm_util.c
562
return &iwpm_hash_bucket[hash & IWPM_MAPINFO_HASH_MASK];
drivers/infiniband/core/iwpm_util.c
569
u32 hash;
drivers/infiniband/core/iwpm_util.c
572
ret = get_hash_bucket(mapped_loc_sockaddr, mapped_rem_sockaddr, &hash);
drivers/infiniband/core/iwpm_util.c
575
return &iwpm_reminfo_bucket[hash & IWPM_REMINFO_HASH_MASK];
drivers/infiniband/hw/hfi1/user_sdma.c
252
int hash;
drivers/infiniband/hw/hfi1/user_sdma.c
259
hash = ((dlid >> 8) ^ dlid) & 0xFF;
drivers/infiniband/hw/hfi1/user_sdma.c
260
if (mapping[hash] == 0xFF) {
drivers/infiniband/hw/hfi1/user_sdma.c
261
mapping[hash] = next;
drivers/infiniband/hw/hfi1/user_sdma.c
265
return mapping[hash];
drivers/infiniband/hw/mthca/mthca_cmd.c
1955
u16 *hash)
drivers/infiniband/hw/mthca/mthca_cmd.c
1963
*hash = imm;
drivers/infiniband/hw/mthca/mthca_cmd.h
322
u16 *hash);
drivers/infiniband/hw/mthca/mthca_mcg.c
125
u16 hash;
drivers/infiniband/hw/mthca/mthca_mcg.c
138
err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index);
drivers/infiniband/hw/mthca/mthca_mcg.c
219
u16 hash;
drivers/infiniband/hw/mthca/mthca_mcg.c
231
err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index);
drivers/infiniband/hw/mthca/mthca_mcg.c
65
u16 *hash, int *prev, int *index)
drivers/infiniband/hw/mthca/mthca_mcg.c
79
err = mthca_MGID_HASH(dev, mailbox, hash);
drivers/infiniband/hw/mthca/mthca_mcg.c
86
mthca_dbg(dev, "Hash for %pI6 is %04x\n", gid, *hash);
drivers/infiniband/hw/mthca/mthca_mcg.c
88
*index = *hash;
drivers/infiniband/hw/mthca/mthca_mcg.c
99
if (*index != *hash) {
drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c
449
u32 hash = skb_get_hash(skb);
drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c
452
hash ^= hash >> 8;
drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c
453
hash ^= hash >> 16;
drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c
456
return (u8)(hash & 0xFF);
drivers/md/bcache/bcache.h
288
struct hlist_node hash;
drivers/md/bcache/btree.c
539
hlist_del_init_rcu(&b->hash);
drivers/md/bcache/btree.c
858
hlist_for_each_entry_rcu(b, mca_hash(c, k), hash)
drivers/md/bcache/btree.c
965
hlist_del_init_rcu(&b->hash);
drivers/md/bcache/btree.c
966
hlist_add_head_rcu(&b->hash, mca_hash(c, k));
drivers/md/bcache/btree.h
119
struct hlist_node hash;
drivers/md/bcache/btree.h
208
hlist_for_each_entry_rcu((b), (c)->bucket_hash + iter, hash)
drivers/md/bcache/request.c
430
hlist_for_each_entry(i, iohash(dc, bio->bi_iter.bi_sector), hash)
drivers/md/bcache/request.c
447
hlist_del(&i->hash);
drivers/md/bcache/request.c
448
hlist_add_head(&i->hash, iohash(dc, i->last));
drivers/md/bcache/super.c
1423
hlist_add_head(&io->hash, dc->io_hash + RECENT_IO);
drivers/md/dm-ima.c
338
if (table->md->ima.active_table.hash != table->md->ima.inactive_table.hash)
drivers/md/dm-ima.c
339
kfree(table->md->ima.inactive_table.hash);
drivers/md/dm-ima.c
341
table->md->ima.inactive_table.hash = digest_buf;
drivers/md/dm-ima.c
390
if (md->ima.active_table.hash != md->ima.inactive_table.hash)
drivers/md/dm-ima.c
391
kfree(md->ima.active_table.hash);
drivers/md/dm-ima.c
393
md->ima.active_table.hash = NULL;
drivers/md/dm-ima.c
404
if (md->ima.inactive_table.hash) {
drivers/md/dm-ima.c
405
md->ima.active_table.hash = md->ima.inactive_table.hash;
drivers/md/dm-ima.c
407
md->ima.inactive_table.hash = NULL;
drivers/md/dm-ima.c
431
if (md->ima.active_table.hash) {
drivers/md/dm-ima.c
435
memcpy(device_table_data + l, md->ima.active_table.hash,
drivers/md/dm-ima.c
522
if (md->ima.active_table.hash) {
drivers/md/dm-ima.c
526
memcpy(device_table_data + l, md->ima.active_table.hash,
drivers/md/dm-ima.c
536
if (md->ima.inactive_table.hash) {
drivers/md/dm-ima.c
540
memcpy(device_table_data + l, md->ima.inactive_table.hash,
drivers/md/dm-ima.c
583
kfree(md->ima.active_table.hash);
drivers/md/dm-ima.c
585
if (md->ima.active_table.hash != md->ima.inactive_table.hash)
drivers/md/dm-ima.c
586
kfree(md->ima.inactive_table.hash);
drivers/md/dm-ima.c
627
memcpy(device_table_data + l, md->ima.inactive_table.hash,
drivers/md/dm-ima.c
653
if (md->ima.inactive_table.hash &&
drivers/md/dm-ima.c
654
md->ima.inactive_table.hash != md->ima.active_table.hash)
drivers/md/dm-ima.c
655
kfree(md->ima.inactive_table.hash);
drivers/md/dm-ima.c
657
md->ima.inactive_table.hash = NULL;
drivers/md/dm-ima.c
668
if (md->ima.active_table.hash) {
drivers/md/dm-ima.c
669
md->ima.inactive_table.hash = md->ima.active_table.hash;
drivers/md/dm-ima.h
45
char *hash;
drivers/md/dm-thin-metadata.c
149
struct hlist_node hash;
drivers/md/dm-vdo/dedupe.c
220
struct uds_record_name hash;
drivers/md/dm-vdo/dedupe.c
2716
u32 hash = name->name[0];
drivers/md/dm-vdo/dedupe.c
2724
hash = (hash * zones->zone_count) >> 8;
drivers/md/dm-vdo/dedupe.c
2725
return &zones->zones[hash];
drivers/md/dm-vdo/dedupe.c
374
return get_unaligned_le64(&lock->hash.name);
drivers/md/dm-vdo/dedupe.c
864
const struct uds_record_name *hash,
drivers/md/dm-vdo/dedupe.c
887
new_lock->hash = *hash;
drivers/md/dm-vdo/int-map.c
115
u64 hash = (input1 ^ input2);
drivers/md/dm-vdo/int-map.c
117
hash *= CITY_MULTIPLIER;
drivers/md/dm-vdo/int-map.c
118
hash ^= (hash >> 47);
drivers/md/dm-vdo/int-map.c
119
hash ^= input2;
drivers/md/dm-vdo/int-map.c
120
hash *= CITY_MULTIPLIER;
drivers/md/dm-vdo/int-map.c
121
hash ^= (hash >> 47);
drivers/md/dm-vdo/int-map.c
122
hash *= CITY_MULTIPLIER;
drivers/md/dm-vdo/int-map.c
123
return hash;
drivers/md/dm-vdo/int-map.c
299
u64 hash = hash_key(key) & 0xFFFFFFFF;
drivers/md/dm-vdo/int-map.c
307
return &map->buckets[(hash * map->capacity) >> 32];
drivers/md/raid5.c
100
spin_unlock_irq(conf->hash_locks + hash);
drivers/md/raid5.c
2539
int hash, cnt;
drivers/md/raid5.c
2575
hash = 0;
drivers/md/raid5.c
2578
lock_device_hash_lock(conf, hash);
drivers/md/raid5.c
2580
!list_empty(conf->inactive_list + hash),
drivers/md/raid5.c
2581
unlock_device_hash_lock(conf, hash),
drivers/md/raid5.c
2582
lock_device_hash_lock(conf, hash));
drivers/md/raid5.c
2583
osh = get_free_stripe(conf, hash);
drivers/md/raid5.c
2584
unlock_device_hash_lock(conf, hash);
drivers/md/raid5.c
2597
nsh->hash_lock_index = hash;
drivers/md/raid5.c
2601
!!((conf->max_nr_stripes % NR_STRIPE_HASH_LOCKS) > hash)) {
drivers/md/raid5.c
2602
hash++;
drivers/md/raid5.c
2685
int hash = (conf->max_nr_stripes - 1) & STRIPE_HASH_LOCKS_MASK;
drivers/md/raid5.c
2687
spin_lock_irq(conf->hash_locks + hash);
drivers/md/raid5.c
2688
sh = get_free_stripe(conf, hash);
drivers/md/raid5.c
2689
spin_unlock_irq(conf->hash_locks + hash);
drivers/md/raid5.c
328
int hash)
drivers/md/raid5.c
334
if (hash == NR_STRIPE_HASH_LOCKS) {
drivers/md/raid5.c
336
hash = NR_STRIPE_HASH_LOCKS - 1;
drivers/md/raid5.c
347
spin_lock_irqsave(conf->hash_locks + hash, flags);
drivers/md/raid5.c
348
if (list_empty(conf->inactive_list + hash) &&
drivers/md/raid5.c
351
list_splice_tail_init(list, conf->inactive_list + hash);
drivers/md/raid5.c
353
spin_unlock_irqrestore(conf->hash_locks + hash, flags);
drivers/md/raid5.c
356
hash--;
drivers/md/raid5.c
379
int hash;
drivers/md/raid5.c
389
hash = sh->hash_lock_index;
drivers/md/raid5.c
390
__release_stripe(conf, sh, &temp_inactive_list[hash]);
drivers/md/raid5.c
402
int hash;
drivers/md/raid5.c
4044
int hash;
drivers/md/raid5.c
4059
hash = sh->hash_lock_index;
drivers/md/raid5.c
4060
spin_lock_irq(conf->hash_locks + hash);
drivers/md/raid5.c
4062
spin_unlock_irq(conf->hash_locks + hash);
drivers/md/raid5.c
421
hash = sh->hash_lock_index;
drivers/md/raid5.c
424
release_inactive_stripe_list(conf, &list, hash);
drivers/md/raid5.c
433
hlist_del_init(&sh->hash);
drivers/md/raid5.c
443
hlist_add_head(&sh->hash, hp);
drivers/md/raid5.c
447
static struct stripe_head *get_free_stripe(struct r5conf *conf, int hash)
drivers/md/raid5.c
452
if (list_empty(conf->inactive_list + hash))
drivers/md/raid5.c
454
first = (conf->inactive_list + hash)->next;
drivers/md/raid5.c
459
BUG_ON(hash != sh->hash_lock_index);
drivers/md/raid5.c
460
if (list_empty(conf->inactive_list + hash))
drivers/md/raid5.c
5328
int hash;
drivers/md/raid5.c
5331
hash = sh->hash_lock_index;
drivers/md/raid5.c
5332
__release_stripe(conf, sh, &temp_inactive_list[hash]);
drivers/md/raid5.c
5630
int hash;
drivers/md/raid5.c
5648
hash = sh->hash_lock_index;
drivers/md/raid5.c
5649
__release_stripe(conf, sh, &cb->temp_inactive_list[hash]);
drivers/md/raid5.c
628
hlist_for_each_entry(sh, stripe_hash(conf, sector), hash)
drivers/md/raid5.c
636
sector_t sector, short generation, int hash)
drivers/md/raid5.c
661
if (!list_empty(conf->inactive_list + hash))
drivers/md/raid5.c
664
if (list_empty(conf->inactive_list + hash) &&
drivers/md/raid5.c
6669
int i, batch_size = 0, hash;
drivers/md/raid5.c
6707
hash = batch[i]->hash_lock_index;
drivers/md/raid5.c
6708
__release_stripe(conf, batch[i], &temp_inactive_list[hash]);
drivers/md/raid5.c
795
static bool is_inactive_blocked(struct r5conf *conf, int hash)
drivers/md/raid5.c
797
if (list_empty(conf->inactive_list + hash))
drivers/md/raid5.c
80
int hash = (sect >> RAID5_STRIPE_SHIFT(conf)) & HASH_MASK;
drivers/md/raid5.c
81
return &conf->stripe_hashtbl[hash];
drivers/md/raid5.c
812
int hash = stripe_hash_locks_hash(conf, sector);
drivers/md/raid5.c
817
spin_lock_irq(conf->hash_locks + hash);
drivers/md/raid5.c
835
*(conf->hash_locks + hash));
drivers/md/raid5.c
839
hash);
drivers/md/raid5.c
844
sh = get_free_stripe(conf, hash);
drivers/md/raid5.c
869
is_inactive_blocked(conf, hash),
drivers/md/raid5.c
870
*(conf->hash_locks + hash));
drivers/md/raid5.c
874
spin_unlock_irq(conf->hash_locks + hash);
drivers/md/raid5.c
89
static inline void lock_device_hash_lock(struct r5conf *conf, int hash)
drivers/md/raid5.c
92
spin_lock_irq(conf->hash_locks + hash);
drivers/md/raid5.c
922
int hash;
drivers/md/raid5.c
935
hash = stripe_hash_locks_hash(conf, head_sector);
drivers/md/raid5.c
936
spin_lock_irq(conf->hash_locks + hash);
drivers/md/raid5.c
938
hash);
drivers/md/raid5.c
939
spin_unlock_irq(conf->hash_locks + hash);
drivers/md/raid5.c
96
static inline void unlock_device_hash_lock(struct r5conf *conf, int hash)
drivers/md/raid5.h
201
struct hlist_node hash;
drivers/media/usb/em28xx/em28xx-cards.c
3131
if (dev->hash == em28xx_eeprom_hash[i].hash) {
drivers/media/usb/em28xx/em28xx-cards.c
3163
if (dev->i2c_hash == em28xx_i2c_hash[i].hash) {
drivers/media/usb/em28xx/em28xx-cards.c
3186
dev->hash, dev->i2c_hash);
drivers/media/usb/em28xx/em28xx-cards.c
57
unsigned long hash;
drivers/media/usb/em28xx/em28xx-i2c.c
607
unsigned long hash = 0;
drivers/media/usb/em28xx/em28xx-i2c.c
622
hash = ((hash ^ l) * 0x9e370001UL);
drivers/media/usb/em28xx/em28xx-i2c.c
625
return (hash >> (32 - bits)) & 0xffffffffUL;
drivers/media/usb/em28xx/em28xx-i2c.c
727
dev->hash = em28xx_hash_mem(data, len, 32);
drivers/media/usb/em28xx/em28xx-i2c.c
732
data, dev->hash);
drivers/media/usb/em28xx/em28xx-i2c.c
798
dev->hash = em28xx_hash_mem(data, len, 32);
drivers/media/usb/em28xx/em28xx-i2c.c
801
data, dev->hash);
drivers/media/usb/em28xx/em28xx.h
695
unsigned long hash; // eeprom hash - for boards with generic ID
drivers/mtd/tests/nandbiterrs.c
162
if (rbuffer[i] != hash(i+seed)) {
drivers/mtd/tests/nandbiterrs.c
164
i, hash(i+seed), rbuffer[i]);
drivers/mtd/tests/nandbiterrs.c
209
wbuffer[i] = hash(i+seed);
drivers/mtd/tests/nandbiterrs.c
272
wbuffer[i] = hash(i+seed);
drivers/net/amt.c
1190
u32 hash;
drivers/net/amt.c
125
u32 hash = jhash(src, sizeof(*src), tunnel->amt->hash_seed);
drivers/net/amt.c
127
return reciprocal_scale(hash, tunnel->amt->hash_buckets);
drivers/net/amt.c
1277
hash = amt_group_hash(tunnel, &group);
drivers/net/amt.c
1278
hlist_for_each_entry_rcu(gnode, &tunnel->groups[hash],
drivers/net/amt.c
1404
u32 hash;
drivers/net/amt.c
1433
hash = amt_source_hash(tunnel, &snode->source_addr);
drivers/net/amt.c
1434
hlist_add_head_rcu(&snode->node, &gnode->sources[hash]);
drivers/net/amt.c
189
u32 hash = amt_source_hash(tunnel, src);
drivers/net/amt.c
192
hlist_for_each_entry_rcu(snode, &gnode->sources[hash], node)
drivers/net/amt.c
202
u32 hash = jhash(group, sizeof(*group), tunnel->amt->hash_seed);
drivers/net/amt.c
204
return reciprocal_scale(hash, tunnel->amt->hash_buckets);
drivers/net/amt.c
212
u32 hash = amt_group_hash(tunnel, group);
drivers/net/amt.c
215
hlist_for_each_entry_rcu(gnode, &tunnel->groups[hash], node) {
drivers/net/amt.c
472
u32 hash;
drivers/net/amt.c
495
hash = amt_group_hash(tunnel, group);
drivers/net/amt.c
496
hlist_add_head_rcu(&gnode->node, &tunnel->groups[hash]);
drivers/net/bonding/bond_alb.c
64
u8 hash = 0;
drivers/net/bonding/bond_alb.c
67
hash ^= hash_start[i];
drivers/net/bonding/bond_alb.c
69
return hash;
drivers/net/bonding/bond_main.c
4220
static u32 bond_ip_hash(u32 hash, struct flow_keys *flow, int xmit_policy)
drivers/net/bonding/bond_main.c
4222
hash ^= (__force u32)flow_get_u32_dst(flow) ^
drivers/net/bonding/bond_main.c
4224
hash ^= (hash >> 16);
drivers/net/bonding/bond_main.c
4225
hash ^= (hash >> 8);
drivers/net/bonding/bond_main.c
4230
return hash >> 1;
drivers/net/bonding/bond_main.c
4232
return hash;
drivers/net/bonding/bond_main.c
4243
u32 hash;
drivers/net/bonding/bond_main.c
4254
hash = bond_eth_hash(skb, data, mhoff, hlen);
drivers/net/bonding/bond_main.c
4257
memcpy(&hash, &flow.icmp, sizeof(hash));
drivers/net/bonding/bond_main.c
4259
memcpy(&hash, &flow.ports.ports, sizeof(hash));
drivers/net/bonding/bond_main.c
4262
return bond_ip_hash(hash, &flow, bond->params.xmit_policy);
drivers/net/bonding/bond_main.c
4277
return skb->hash;
drivers/net/bonding/bond_main.c
5226
u32 hash;
drivers/net/bonding/bond_main.c
5228
hash = bond_xmit_hash(bond, skb);
drivers/net/bonding/bond_main.c
5233
slave = slaves->arr[hash % count];
drivers/net/bonding/bond_main.c
5242
u32 hash;
drivers/net/bonding/bond_main.c
5244
hash = bond_xmit_hash_xdp(bond, xdp);
drivers/net/bonding/bond_main.c
5250
return slaves->arr[hash % count];
drivers/net/bonding/bond_main.c
5483
u32 hash;
drivers/net/bonding/bond_main.c
5488
memcpy(&hash, &flow.ports.ports, sizeof(hash));
drivers/net/bonding/bond_main.c
5490
return bond_ip_hash(hash, &flow, BOND_XMIT_POLICY_LAYER34);
drivers/net/bonding/bond_main.c
5499
u32 hash;
drivers/net/bonding/bond_main.c
5506
hash = bond_sk_hash_l34(sk);
drivers/net/bonding/bond_main.c
5507
slave = slaves->arr[hash % count];
drivers/net/dsa/mv88e6xxx/chip.h
646
int (*atu_get_hash)(struct mv88e6xxx_chip *chip, u8 *hash);
drivers/net/dsa/mv88e6xxx/chip.h
647
int (*atu_set_hash)(struct mv88e6xxx_chip *chip, u8 hash);
drivers/net/dsa/mv88e6xxx/devlink.c
10
static int mv88e6xxx_atu_get_hash(struct mv88e6xxx_chip *chip, u8 *hash)
drivers/net/dsa/mv88e6xxx/devlink.c
13
return chip->info->ops->atu_get_hash(chip, hash);
drivers/net/dsa/mv88e6xxx/devlink.c
18
static int mv88e6xxx_atu_set_hash(struct mv88e6xxx_chip *chip, u8 hash)
drivers/net/dsa/mv88e6xxx/devlink.c
21
return chip->info->ops->atu_set_hash(chip, hash);
drivers/net/dsa/mv88e6xxx/global1.h
336
int mv88e6165_g1_atu_get_hash(struct mv88e6xxx_chip *chip, u8 *hash);
drivers/net/dsa/mv88e6xxx/global1.h
337
int mv88e6165_g1_atu_set_hash(struct mv88e6xxx_chip *chip, u8 hash);
drivers/net/dsa/mv88e6xxx/global1_atu.c
105
val |= hash;
drivers/net/dsa/mv88e6xxx/global1_atu.c
78
int mv88e6165_g1_atu_get_hash(struct mv88e6xxx_chip *chip, u8 *hash)
drivers/net/dsa/mv88e6xxx/global1_atu.c
87
*hash = val & MV88E6161_G1_ATU_CTL_HASH_MASK;
drivers/net/dsa/mv88e6xxx/global1_atu.c
92
int mv88e6165_g1_atu_set_hash(struct mv88e6xxx_chip *chip, u8 hash)
drivers/net/dsa/mv88e6xxx/global1_atu.c
97
if (hash & ~MV88E6161_G1_ATU_CTL_HASH_MASK)
drivers/net/dsa/mv88e6xxx/global2.c
130
bool hash, u16 mask)
drivers/net/dsa/mv88e6xxx/global2.c
134
if (hash)
drivers/net/dsa/mv88e6xxx/global2.h
361
bool hash, u16 mask);
drivers/net/dsa/qca/qca8k-common.c
1099
u32 hash = 0;
drivers/net/dsa/qca/qca8k-common.c
1103
hash |= QCA8K_TRUNK_HASH_SIP_EN;
drivers/net/dsa/qca/qca8k-common.c
1104
hash |= QCA8K_TRUNK_HASH_DIP_EN;
drivers/net/dsa/qca/qca8k-common.c
1107
hash |= QCA8K_TRUNK_HASH_SA_EN;
drivers/net/dsa/qca/qca8k-common.c
1108
hash |= QCA8K_TRUNK_HASH_DA_EN;
drivers/net/dsa/qca/qca8k-common.c
1129
priv->lag_hash_mode = hash;
drivers/net/dsa/qca/qca8k-common.c
1130
} else if (priv->lag_hash_mode != hash) {
drivers/net/dsa/qca/qca8k-common.c
1136
QCA8K_TRUNK_HASH_MASK, hash);
drivers/net/dsa/vitesse-vsc73xx-core.c
1909
u16 hash = FIELD_PREP(VSC73XX_HASH0_VID_TO_MASK,
drivers/net/dsa/vitesse-vsc73xx-core.c
1914
hash ^= FIELD_PREP(VSC73XX_HASH1_MAC0_TO_MASK,
drivers/net/dsa/vitesse-vsc73xx-core.c
1919
hash ^= FIELD_PREP(VSC73XX_HASH2_MAC1_TO_MASK,
drivers/net/dsa/vitesse-vsc73xx-core.c
1926
hash ^= FIELD_PREP(VSC73XX_HASH3_MAC3_TO_MASK,
drivers/net/dsa/vitesse-vsc73xx-core.c
1931
hash ^= FIELD_PREP(VSC73XX_HASH4_MAC4_TO_MASK,
drivers/net/dsa/vitesse-vsc73xx-core.c
1935
return hash;
drivers/net/dsa/vitesse-vsc73xx-core.c
2027
u16 hash, u16 cmd_mask, u16 cmd_val)
drivers/net/dsa/vitesse-vsc73xx-core.c
2050
hash);
drivers/net/dsa/vitesse-vsc73xx-core.c
2070
u16 hash = vsc73xx_calc_hash(addr, vid);
drivers/net/dsa/vitesse-vsc73xx-core.c
2075
ret = vsc73xx_port_read_mac_table_row(vsc, hash, fdb);
drivers/net/dsa/vitesse-vsc73xx-core.c
2091
ret = vsc73xx_fdb_operation(vsc, addr, vid, hash,
drivers/net/dsa/vitesse-vsc73xx-core.c
2103
u16 hash = vsc73xx_calc_hash(addr, vid);
drivers/net/dsa/vitesse-vsc73xx-core.c
2109
ret = vsc73xx_port_read_mac_table_row(vsc, hash, fdb);
drivers/net/dsa/vitesse-vsc73xx-core.c
2127
ret = vsc73xx_fdb_operation(vsc, addr, vid, hash,
drivers/net/ethernet/airoha/airoha_eth.c
615
u32 hash, reason, msg1 = le32_to_cpu(desc->msg1);
drivers/net/ethernet/airoha/airoha_eth.c
682
hash = FIELD_GET(AIROHA_RXD4_FOE_ENTRY, msg1);
drivers/net/ethernet/airoha/airoha_eth.c
683
if (hash != AIROHA_RXD4_FOE_ENTRY)
drivers/net/ethernet/airoha/airoha_eth.c
684
skb_set_hash(q->skb, jhash_1word(hash, 0),
drivers/net/ethernet/airoha/airoha_eth.c
689
airoha_ppe_check_skb(&eth->ppe->dev, q->skb, hash,
drivers/net/ethernet/airoha/airoha_eth.h
482
u32 hash;
drivers/net/ethernet/airoha/airoha_eth.h
654
u16 hash, bool rx_wlan);
drivers/net/ethernet/airoha/airoha_eth.h
661
u32 hash);
drivers/net/ethernet/airoha/airoha_eth.h
662
void airoha_ppe_foe_entry_get_stats(struct airoha_ppe *ppe, u32 hash,
drivers/net/ethernet/airoha/airoha_npu.c
397
u32 entry_size, u32 hash, bool ppe2)
drivers/net/ethernet/airoha/airoha_npu.c
419
ppe_data->set_info.data = hash;
drivers/net/ethernet/airoha/airoha_ppe.c
1222
void airoha_ppe_foe_entry_get_stats(struct airoha_ppe *ppe, u32 hash,
drivers/net/ethernet/airoha/airoha_ppe.c
1234
if (airoha_ppe_foe_get_flow_stats_index(ppe, hash, &index))
drivers/net/ethernet/airoha/airoha_ppe.c
1271
if (e->hash != 0xffff) {
drivers/net/ethernet/airoha/airoha_ppe.c
1274
airoha_ppe_foe_entry_get_stats(eth->ppe, e->hash, &stats);
drivers/net/ethernet/airoha/airoha_ppe.c
1391
u16 hash, bool rx_wlan)
drivers/net/ethernet/airoha/airoha_ppe.c
1397
if (hash > ppe_hash_mask)
drivers/net/ethernet/airoha/airoha_ppe.c
1401
diff = now - ppe->foe_check_time[hash];
drivers/net/ethernet/airoha/airoha_ppe.c
1405
ppe->foe_check_time[hash] = now;
drivers/net/ethernet/airoha/airoha_ppe.c
1406
airoha_ppe_foe_insert_entry(ppe, skb, hash, rx_wlan);
drivers/net/ethernet/airoha/airoha_ppe.c
464
u32 hash, hv1, hv2, hv3;
drivers/net/ethernet/airoha/airoha_ppe.c
504
hash = (hv1 & hv2) | ((~hv1) & hv3);
drivers/net/ethernet/airoha/airoha_ppe.c
505
hash = (hash >> 24) | ((hash & 0xffffff) << 8);
drivers/net/ethernet/airoha/airoha_ppe.c
506
hash ^= hv1 ^ hv2 ^ hv3;
drivers/net/ethernet/airoha/airoha_ppe.c
507
hash ^= hash >> 16;
drivers/net/ethernet/airoha/airoha_ppe.c
508
hash &= ppe_hash_mask;
drivers/net/ethernet/airoha/airoha_ppe.c
510
return hash;
drivers/net/ethernet/airoha/airoha_ppe.c
514
u32 hash, u32 *index)
drivers/net/ethernet/airoha/airoha_ppe.c
522
*index = hash >= ppe_num_stats_entries ? hash - PPE_STATS_NUM_ENTRIES
drivers/net/ethernet/airoha/airoha_ppe.c
523
: hash;
drivers/net/ethernet/airoha/airoha_ppe.c
552
u32 hash)
drivers/net/ethernet/airoha/airoha_ppe.c
563
if (airoha_ppe_foe_get_flow_stats_index(ppe, hash, &index))
drivers/net/ethernet/airoha/airoha_ppe.c
605
airoha_ppe_foe_get_entry_locked(struct airoha_ppe *ppe, u32 hash)
drivers/net/ethernet/airoha/airoha_ppe.c
611
if (hash < sram_num_entries) {
drivers/net/ethernet/airoha/airoha_ppe.c
612
u32 *hwe = ppe->foe + hash * sizeof(struct airoha_foe_entry);
drivers/net/ethernet/airoha/airoha_ppe.c
613
bool ppe2 = hash >= PPE_SRAM_NUM_ENTRIES;
drivers/net/ethernet/airoha/airoha_ppe.c
619
FIELD_PREP(PPE_SRAM_CTRL_ENTRY_MASK, hash) |
drivers/net/ethernet/airoha/airoha_ppe.c
633
return ppe->foe + hash * sizeof(struct airoha_foe_entry);
drivers/net/ethernet/airoha/airoha_ppe.c
637
u32 hash)
drivers/net/ethernet/airoha/airoha_ppe.c
642
hwe = airoha_ppe_foe_get_entry_locked(ppe, hash);
drivers/net/ethernet/airoha/airoha_ppe.c
665
static int airoha_ppe_foe_commit_sram_entry(struct airoha_ppe *ppe, u32 hash)
drivers/net/ethernet/airoha/airoha_ppe.c
667
struct airoha_foe_entry *hwe = ppe->foe + hash * sizeof(*hwe);
drivers/net/ethernet/airoha/airoha_ppe.c
668
bool ppe2 = hash >= PPE_SRAM_NUM_ENTRIES;
drivers/net/ethernet/airoha/airoha_ppe.c
677
FIELD_PREP(PPE_SRAM_CTRL_ENTRY_MASK, hash) |
drivers/net/ethernet/airoha/airoha_ppe.c
688
u32 hash, bool rx_wlan)
drivers/net/ethernet/airoha/airoha_ppe.c
691
struct airoha_foe_entry *hwe = ppe->foe + hash * sizeof(*hwe);
drivers/net/ethernet/airoha/airoha_ppe.c
713
airoha_ppe_foe_flow_stats_update(ppe, npu, hwe, hash);
drivers/net/ethernet/airoha/airoha_ppe.c
715
if (hash < sram_num_entries)
drivers/net/ethernet/airoha/airoha_ppe.c
716
err = airoha_ppe_foe_commit_sram_entry(ppe, hash);
drivers/net/ethernet/airoha/airoha_ppe.c
729
if (e->hash != 0xffff) {
drivers/net/ethernet/airoha/airoha_ppe.c
733
airoha_ppe_foe_commit_entry(ppe, &e->data, e->hash, false);
drivers/net/ethernet/airoha/airoha_ppe.c
734
e->hash = 0xffff;
drivers/net/ethernet/airoha/airoha_ppe.c
772
u32 hash, bool rx_wlan)
drivers/net/ethernet/airoha/airoha_ppe.c
779
hwe_p = airoha_ppe_foe_get_entry_locked(ppe, hash);
drivers/net/ethernet/airoha/airoha_ppe.c
789
f->hash = hash;
drivers/net/ethernet/airoha/airoha_ppe.c
813
airoha_ppe_foe_commit_entry(ppe, &hwe, hash, rx_wlan);
drivers/net/ethernet/airoha/airoha_ppe.c
820
u32 hash, bool rx_wlan)
drivers/net/ethernet/airoha/airoha_ppe.c
831
hwe = airoha_ppe_foe_get_entry_locked(ppe, hash);
drivers/net/ethernet/airoha/airoha_ppe.c
844
e->hash = 0xffff;
drivers/net/ethernet/airoha/airoha_ppe.c
853
airoha_ppe_foe_commit_entry(ppe, &e->data, hash, rx_wlan);
drivers/net/ethernet/airoha/airoha_ppe.c
855
e->hash = hash;
drivers/net/ethernet/airoha/airoha_ppe.c
865
airoha_ppe_foe_commit_subflow_entry(ppe, e, hash, rx_wlan);
drivers/net/ethernet/airoha/airoha_ppe.c
894
u32 hash;
drivers/net/ethernet/airoha/airoha_ppe.c
899
hash = airoha_ppe_foe_get_entry_hash(ppe, &e->data);
drivers/net/ethernet/airoha/airoha_ppe.c
901
e->hash = 0xffff;
drivers/net/ethernet/airoha/airoha_ppe.c
904
hlist_add_head(&e->list, &ppe->foe_flow[hash]);
drivers/net/ethernet/airoha/airoha_ppe.c
944
hwe = airoha_ppe_foe_get_entry_locked(ppe, iter->hash);
drivers/net/ethernet/airoha/airoha_ppe.c
951
iter->hash = 0xffff;
drivers/net/ethernet/airoha/airoha_ppe.c
978
if (e->hash == 0xffff)
drivers/net/ethernet/airoha/airoha_ppe.c
981
hwe_p = airoha_ppe_foe_get_entry_locked(ppe, e->hash);
drivers/net/ethernet/airoha/airoha_ppe.c
987
e->hash = 0xffff;
drivers/net/ethernet/altera/altera_tse_main.c
806
unsigned int hash = 0;
drivers/net/ethernet/altera/altera_tse_main.c
817
hash = (hash << 1) | xor_bit;
drivers/net/ethernet/altera/altera_tse_main.c
819
csrwr32(1, priv->mac_dev, tse_csroffs(hash_table) + hash * 4);
drivers/net/ethernet/amazon/ena/ena_eth_com.c
373
ena_rx_ctx->hash = cdesc->hash;
drivers/net/ethernet/amazon/ena/ena_eth_com.c
381
ena_rx_ctx->l4_csum_err, ena_rx_ctx->hash, ena_rx_ctx->frag, cdesc->status);
drivers/net/ethernet/amazon/ena/ena_eth_com.h
48
u32 hash;
drivers/net/ethernet/amazon/ena/ena_eth_io_defs.h
236
u32 hash;
drivers/net/ethernet/amazon/ena/ena_netdev.c
1144
skb_set_hash(skb, ena_rx_ctx->hash, hash_type);
drivers/net/ethernet/amazon/ena/ena_netdev.c
1233
ena_rx_ctx.l4_proto, ena_rx_ctx.hash);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1707
hlist_del_rcu(&fltr->base.hash);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1722
hlist_add_head_rcu(&fltr->base.hash, head);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1734
hlist_for_each_entry_rcu(fltr, head, base.hash) {
drivers/net/ethernet/broadcom/bnge/bnge_netdev.h
525
struct hlist_node hash;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15574
hlist_add_head_rcu(&fltr->base.hash, head);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15628
hlist_for_each_entry_rcu(f, head, base.hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15730
hlist_del_rcu(&fltr->base.hash);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15751
hlist_for_each_entry_safe(fltr, tmp, head, base.hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5368
hlist_del(&fltr->hash);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5392
hlist_for_each_entry_safe(fltr, tmp, head, base.hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5437
hlist_for_each_entry_safe(fltr, tmp, head, base.hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5925
hlist_del_rcu(&fltr->base.hash);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5942
hlist_for_each_entry_rcu(fltr, head, base.hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6004
u64 prefix = bp->toeplitz_prefix, hash = 0;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6030
hash ^= prefix;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6036
return (hash >> 32) & BNXT_NTP_FLTR_HASH_MASK;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6072
hlist_add_head_rcu(&fltr->base.hash, head);
drivers/net/ethernet/broadcom/bnxt/bnxt.h
1423
struct hlist_node hash;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1046
hlist_for_each_entry_rcu(fltr, head, hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1069
hlist_for_each_entry_rcu(fltr, head, hash) {
drivers/net/ethernet/cavium/liquidio/lio_core.c
661
u32 hash = be32_to_cpu(*hash_be);
drivers/net/ethernet/cavium/liquidio/lio_core.c
663
skb_set_hash(skb, hash, PKT_HASH_TYPE_L4);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
732
u32 hash;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
741
hash = cqe_rx->rss_tag;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
745
hash = cqe_rx->rss_tag;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
749
hash = 0;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
752
skb_set_hash(skb, hash, hash_type);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
204
int hash = arp_hash(e->addr, e->ifindex, d);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
206
for (p = &d->l2tab[hash].first; *p; p = &(*p)->next)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
272
int hash;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
294
hash = arp_hash(addr, ifidx, d);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
297
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
310
e->next = d->l2tab[hash].first;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
311
d->l2tab[hash].first = e;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
369
int hash = arp_hash(addr, ifidx, d);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
372
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb3/xgmac.c
289
int hash = 0, octet, bit, i = 0, c;
drivers/net/ethernet/chelsio/cxgb3/xgmac.c
293
hash ^= (c & 1) << i;
drivers/net/ethernet/chelsio/cxgb3/xgmac.c
297
return hash;
drivers/net/ethernet/chelsio/cxgb3/xgmac.c
323
int hash = hash_hw_addr(ha->addr);
drivers/net/ethernet/chelsio/cxgb3/xgmac.c
325
if (hash < 32)
drivers/net/ethernet/chelsio/cxgb3/xgmac.c
326
hash_lo |= (1 << hash);
drivers/net/ethernet/chelsio/cxgb3/xgmac.c
328
hash_hi |= (1 << (hash - 32));
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
113
list_add_tail(&ce->list, &ctbl->hash_list[hash]);
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
152
int hash;
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
158
hash = clip_addr_hash(ctbl, addr, v6);
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
161
list_for_each_entry(cte, &ctbl->hash_list[hash], list) {
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
80
int hash;
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
86
hash = clip_addr_hash(ctbl, addr, v6);
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
89
list_for_each_entry(cte, &ctbl->hash_list[hash], list) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1378
u32 hash:1; /* 0 => wild-card, 1 => exact-match */
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
2013
const u8 **addr, u16 *idx, u64 *hash, bool sleep_ok);
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
2139
u64 *hash, bool sleep_ok);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1856
else if (!f->fs.hash)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1920
else if (!fs.hash)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1540
if (fs->hash) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1811
if (fs && fs->hash) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
256
fconf = fs->hash ? adapter->params.tp.filter_mask :
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
358
u64 *pkts, u64 *bytes, bool hash)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
367
if (is_hashfilter(adapter) && hash) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
438
u64 *hitcnt, u64 *bytecnt, bool hash)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
442
return get_filter_count(adapter, fidx, hitcnt, bytecnt, hash);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
996
if ((f->fs.hash || is_t6(adap->params.chip)) && f->fs.type)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c
80
u16 *idx, u64 *hash, bool sleep_ok)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c
85
naddr, addr, idx, hash, sleep_ok);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
1040
flower_entry->fs.hash);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
1085
ch_flower->fs.hash);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
814
if (fe->fs.hash &&
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
881
fs->hash = is_filter_exact_match(adap, fs);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
888
fidx = cxgb4_get_free_ftid(dev, inet_family, fs->hash,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
898
fs->hash = 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
904
if (fs->hash)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
927
if (fs->hash)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
967
if (fs->hash)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
981
u8 hash;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
984
hash = fs->hash;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
990
if (hash)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
491
ingress->fs[i].hash);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
297
u64 *hitcnt, u64 *bytecnt, bool hash);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
287
for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
329
for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
425
int hash = addr_hash(d, addr, addr_len, ifidx);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
440
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
458
e->hash = hash;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
464
e->next = d->l2tab[hash].first;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
465
d->l2tab[hash].first = e;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
513
int hash, ifidx = neigh->dev->ifindex;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
518
hash = addr_hash(d, addr, addr_len, ifidx);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
520
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.h
84
u16 hash; /* hash bucket the entry is on */
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7974
const u8 **addr, u16 *idx, u64 *hash, bool sleep_ok)
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8029
else if (hash)
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8030
*hash |= (1ULL <<
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8209
c.u.hash.hashvec = cpu_to_be64(vec);
drivers/net/ethernet/chelsio/cxgb4/t4fw_api.h
2297
} hash;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
1529
u64 *hash, bool sleep_ok)
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
1583
else if (hash)
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
1584
*hash |= (1ULL << hash_mac_addr(addr[offset+i]));
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
1752
cmd.u.hash.hashvec = cpu_to_be64(vec);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_main.c
200
tlsdev->hash = chtls_create_hash;
drivers/net/ethernet/emulex/benet/be_cmds.c
2212
memcpy(req->hash, rss_hkey, RSS_HASH_KEY_LEN);
drivers/net/ethernet/emulex/benet/be_cmds.c
2213
be_dws_cpu_to_le(req->hash, sizeof(req->hash));
drivers/net/ethernet/emulex/benet/be_cmds.h
1220
u32 hash[10];
drivers/net/ethernet/ethoc.c
829
u32 hash[2] = { 0, 0 };
drivers/net/ethernet/ethoc.c
853
hash[0] = 0xffffffff;
drivers/net/ethernet/ethoc.c
854
hash[1] = 0xffffffff;
drivers/net/ethernet/ethoc.c
859
hash[bit >> 5] |= 1 << (bit & 0x1f);
drivers/net/ethernet/ethoc.c
863
ethoc_write(priv, ETH_HASH0, hash[0]);
drivers/net/ethernet/ethoc.c
864
ethoc_write(priv, ETH_HASH1, hash[1]);
drivers/net/ethernet/faraday/ftmac100.c
158
u32 hash = ether_crc(ETH_ALEN, ha->addr) >> 26;
drivers/net/ethernet/faraday/ftmac100.c
160
maht |= BIT_ULL(hash);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2704
u32 hash;
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2774
hash = be32_to_cpu(*(__be32 *)(vaddr + hash_offset));
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2823
skb_set_hash(skb, hash, type);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
101
u64 hash)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
103
enetc_port_wr(hw, ENETC4_PSIUMHFR0(si), lower_32_bits(hash));
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
104
enetc_port_wr(hw, ENETC4_PSIUMHFR1(si), upper_32_bits(hash));
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
108
u64 hash)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
110
enetc_port_wr(hw, ENETC4_PSIMMHFR0(si), lower_32_bits(hash));
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
111
enetc_port_wr(hw, ENETC4_PSIMMHFR1(si), upper_32_bits(hash));
drivers/net/ethernet/freescale/enetc/enetc_pf.c
103
lower_32_bits(hash));
drivers/net/ethernet/freescale/enetc/enetc_pf.c
105
upper_32_bits(hash));
drivers/net/ethernet/freescale/enetc/enetc_pf.c
108
lower_32_bits(hash));
drivers/net/ethernet/freescale/enetc/enetc_pf.c
110
upper_32_bits(hash));
drivers/net/ethernet/freescale/enetc/enetc_pf.c
97
unsigned long hash)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
391
int si_id, u64 hash)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
404
enetc_port_wr(hw, low_reg_off, lower_32_bits(hash));
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
405
enetc_port_wr(hw, high_reg_off, upper_32_bits(hash));
drivers/net/ethernet/freescale/fec_main.c
4280
unsigned char hash;
drivers/net/ethernet/freescale/fec_main.c
4312
hash = (crc >> (32 - FEC_HASH_BITS)) & 0x3f;
drivers/net/ethernet/freescale/fec_main.c
4314
if (hash > 31)
drivers/net/ethernet/freescale/fec_main.c
4315
hash_high |= 1 << (hash - 32);
drivers/net/ethernet/freescale/fec_main.c
4317
hash_low |= 1 << hash;
drivers/net/ethernet/freescale/fman/fman_mac.h
196
static inline void free_hash_table(struct eth_hash_t *hash)
drivers/net/ethernet/freescale/fman/fman_mac.h
201
if (hash) {
drivers/net/ethernet/freescale/fman/fman_mac.h
202
if (hash->lsts) {
drivers/net/ethernet/freescale/fman/fman_mac.h
203
for (i = 0; i < hash->size; i++) {
drivers/net/ethernet/freescale/fman/fman_mac.h
205
dequeue_addr_from_hash_entry(&hash->lsts[i]);
drivers/net/ethernet/freescale/fman/fman_mac.h
209
dequeue_addr_from_hash_entry(&hash->
drivers/net/ethernet/freescale/fman/fman_mac.h
214
kfree(hash->lsts);
drivers/net/ethernet/freescale/fman/fman_mac.h
217
kfree(hash);
drivers/net/ethernet/freescale/fman/fman_mac.h
224
struct eth_hash_t *hash;
drivers/net/ethernet/freescale/fman/fman_mac.h
227
hash = kmalloc_obj(*hash);
drivers/net/ethernet/freescale/fman/fman_mac.h
228
if (!hash)
drivers/net/ethernet/freescale/fman/fman_mac.h
231
hash->size = size;
drivers/net/ethernet/freescale/fman/fman_mac.h
233
hash->lsts = kmalloc_objs(struct list_head, hash->size);
drivers/net/ethernet/freescale/fman/fman_mac.h
234
if (!hash->lsts) {
drivers/net/ethernet/freescale/fman/fman_mac.h
235
kfree(hash);
drivers/net/ethernet/freescale/fman/fman_mac.h
239
for (i = 0; i < hash->size; i++)
drivers/net/ethernet/freescale/fman/fman_mac.h
240
INIT_LIST_HEAD(&hash->lsts[i]);
drivers/net/ethernet/freescale/fman/fman_mac.h
242
return hash;
drivers/net/ethernet/freescale/fman/fman_memac.c
801
u32 hash;
drivers/net/ethernet/freescale/fman/fman_memac.c
811
hash = get_mac_addr_hash_code(addr) & HASH_CTRL_ADDR_MASK;
drivers/net/ethernet/freescale/fman/fman_memac.c
821
&memac->multicast_addr_hash->lsts[hash]);
drivers/net/ethernet/freescale/fman/fman_memac.c
822
iowrite32be(hash | HASH_CTRL_MCAST_EN, &regs->hashtable_ctrl);
drivers/net/ethernet/freescale/fman/fman_memac.c
858
u32 hash;
drivers/net/ethernet/freescale/fman/fman_memac.c
863
hash = get_mac_addr_hash_code(addr) & HASH_CTRL_ADDR_MASK;
drivers/net/ethernet/freescale/fman/fman_memac.c
865
list_for_each(pos, &memac->multicast_addr_hash->lsts[hash]) {
drivers/net/ethernet/freescale/fman/fman_memac.c
875
if (list_empty(&memac->multicast_addr_hash->lsts[hash]))
drivers/net/ethernet/freescale/fman/fman_memac.c
876
iowrite32be(hash & ~HASH_CTRL_MCAST_EN,
drivers/net/ethernet/freescale/fman/fman_tgec.c
491
u32 crc = 0xFFFFFFFF, hash;
drivers/net/ethernet/freescale/fman/fman_tgec.c
505
hash = (crc >> TGEC_HASH_MCAST_SHIFT) & TGEC_HASH_ADR_MSK;
drivers/net/ethernet/freescale/fman/fman_tgec.c
515
&tgec->multicast_addr_hash->lsts[hash]);
drivers/net/ethernet/freescale/fman/fman_tgec.c
516
iowrite32be((hash | TGEC_HASH_MCAST_EN), &regs->hashtable_ctrl);
drivers/net/ethernet/freescale/fman/fman_tgec.c
564
u32 crc = 0xFFFFFFFF, hash;
drivers/net/ethernet/freescale/fman/fman_tgec.c
573
hash = (crc >> TGEC_HASH_MCAST_SHIFT) & TGEC_HASH_ADR_MSK;
drivers/net/ethernet/freescale/fman/fman_tgec.c
575
list_for_each(pos, &tgec->multicast_addr_hash->lsts[hash]) {
drivers/net/ethernet/freescale/fman/fman_tgec.c
585
if (list_empty(&tgec->multicast_addr_hash->lsts[hash]))
drivers/net/ethernet/freescale/fman/fman_tgec.c
586
iowrite32be((hash & ~TGEC_HASH_MCAST_EN),
drivers/net/ethernet/fungible/funcore/fun_hci.h
1172
__be32 hash;
drivers/net/ethernet/fungible/funeth/funeth_rx.c
434
if (likely((q->netdev->features & NETIF_F_RXHASH) && rxreq->hash))
drivers/net/ethernet/fungible/funeth/funeth_rx.c
435
skb_set_hash(skb, be32_to_cpu(rxreq->hash),
drivers/net/ethernet/fungible/funeth/funeth_rx.c
445
trace_funeth_rx(q, rxreq->nsgl, pkt_len, skb->hash, cv);
drivers/net/ethernet/fungible/funeth/funeth_trace.h
105
__entry->num_rqes, __entry->len, __entry->hash,
drivers/net/ethernet/fungible/funeth/funeth_trace.h
78
u32 hash,
drivers/net/ethernet/fungible/funeth/funeth_trace.h
81
TP_ARGS(rxq, num_rqes, pkt_len, hash, cls_vec),
drivers/net/ethernet/fungible/funeth/funeth_trace.h
88
__field(u32, hash)
drivers/net/ethernet/fungible/funeth/funeth_trace.h
98
__entry->hash = hash;
drivers/net/ethernet/google/gve/gve_desc_dqo.h
249
__le32 hash;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
447
skb_set_hash(skb, le32_to_cpu(compl_desc->hash), hash_type);
drivers/net/ethernet/google/gve/gve_tx.c
529
mtd_desc->mtd.path_hash = cpu_to_be32(skb->hash);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
534
u16 path_hash = skb->hash ^ (skb->hash >> 16);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1820
u32 hash;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1830
hash = le32_to_cpu(rx_desc->wb.qword0.hi_dword.rss);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1831
libeth_rx_pt_set_hash(skb, hash, decoded);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3726
u32 hash;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3729
hash = skb->sk->sk_hash;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3731
hash = (__force u16)skb->protocol ^ skb->hash;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3733
hash = jhash_1word(hash, jhash_initval_salt);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3735
return (u16)(((u64)hash * num_tx_queues) >> 32);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3748
u16 hash;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3765
hash = i40e_swdcb_skb_tx_hash(netdev, skb, qcount);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3768
return qoffset + hash;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1042
u32 hash;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1048
hash = le64_get_bits(qw0, IAVF_RXD_LEGACY_RSS_M);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1049
libeth_rx_pt_set_hash(skb, hash, decoded_pt);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1068
u32 hash;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1075
hash = le64_get_bits(qw1, IAVF_RXD_FLEX_RSS_HASH_M);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1076
libeth_rx_pt_set_hash(skb, hash, decoded_pt);
drivers/net/ethernet/intel/ice/devlink/devlink.c
169
snprintf(ctx->buf, sizeof(ctx->buf), "0x%08x", netlist->hash);
drivers/net/ethernet/intel/ice/devlink/devlink.c
193
snprintf(ctx->buf, sizeof(ctx->buf), "0x%08x", netlist->hash);
drivers/net/ethernet/intel/ice/ice_flow.c
2205
return (r->hash.hdr_type == hdr_type &&
drivers/net/ethernet/intel/ice/ice_flow.c
2206
r->hash.hash_flds == prof->segs[prof->segs_cnt - 1].match &&
drivers/net/ethernet/intel/ice/ice_flow.c
2207
r->hash.addl_hdrs == prof->segs[prof->segs_cnt - 1].hdrs);
drivers/net/ethernet/intel/ice/ice_flow.c
2266
rss_cfg->hash.hash_flds = prof->segs[prof->segs_cnt - 1].match;
drivers/net/ethernet/intel/ice/ice_flow.c
2267
rss_cfg->hash.addl_hdrs = prof->segs[prof->segs_cnt - 1].hdrs;
drivers/net/ethernet/intel/ice/ice_flow.c
2268
rss_cfg->hash.hdr_type = hdr_type;
drivers/net/ethernet/intel/ice/ice_flow.c
2269
rss_cfg->hash.symm = prof->symm;
drivers/net/ethernet/intel/ice/ice_flow.c
2943
if (test_bit(vsi_handle, r->vsis) && r->hash.symm != symm) {
drivers/net/ethernet/intel/ice/ice_flow.c
2944
local = r->hash;
drivers/net/ethernet/intel/ice/ice_flow.c
2946
if (symm && !rss_cfg_symm_valid(r->hash.hash_flds))
drivers/net/ethernet/intel/ice/ice_flow.c
2975
status = ice_add_rss_cfg_sync(hw, vsi_handle, &r->hash);
drivers/net/ethernet/intel/ice/ice_flow.c
3007
r->hash.addl_hdrs == hdrs) {
drivers/net/ethernet/intel/ice/ice_flow.c
3008
rss_hash = r->hash.hash_flds;
drivers/net/ethernet/intel/ice/ice_flow.c
3009
*symm = r->hash.symm;
drivers/net/ethernet/intel/ice/ice_flow.h
497
struct ice_rss_hash_cfg hash;
drivers/net/ethernet/intel/ice/ice_nvm.c
825
netlist->hash = id_blk[ICE_NETLIST_ID_BLK_SHA_HASH_WORD(15)] << 16 |
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
556
static int ice_xdp_rx_hash(const struct xdp_md *ctx, u32 *hash,
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
561
*hash = ice_get_rx_hash(xdp_ext->desc);
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
563
if (!likely(*hash))
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
70
u32 hash;
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
76
hash = ice_get_rx_hash(rx_desc);
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
77
if (likely(hash))
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
78
libeth_rx_pt_set_hash(skb, hash, decoded);
drivers/net/ethernet/intel/ice/ice_type.h
489
u32 hash; /* SHA-1 hash word */
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
803
u32 hash = le32_to_cpu(rx_desc->base_wb.qword0.hi_dword.rss);
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
805
libeth_rx_pt_set_hash(skb, hash, decoded);
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
829
u32 hash = le32_to_cpu(rx_desc->flex_nic_wb.rss_hash);
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
831
libeth_rx_pt_set_hash(skb, hash, decoded);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3188
u32 hash;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3193
hash = le16_to_cpu(rx_desc->hash1) |
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3197
libeth_rx_pt_set_hash(skb, hash, decoded);
drivers/net/ethernet/intel/idpf/xdp.c
384
static int idpf_xdpmo_rx_hash(const struct xdp_md *ctx, u32 *hash,
drivers/net/ethernet/intel/idpf/xdp.c
402
return libeth_xdpmo_rx_hash(hash, rss_type, idpf_xdp_rx_hash(&desc),
drivers/net/ethernet/intel/igc/igc_main.c
7048
static int igc_xdp_rx_hash(const struct xdp_md *_ctx, u32 *hash,
drivers/net/ethernet/intel/igc/igc_main.c
7056
*hash = le32_to_cpu(ctx->rx_desc->wb.lower.hi_dword.rss);
drivers/net/ethernet/intel/ixgbe/devlink/devlink.c
181
snprintf(ctx->buf, sizeof(ctx->buf), "0x%08x", netlist->hash);
drivers/net/ethernet/intel/ixgbe/ixgbe_e610.c
3311
netlist->hash = id_blk[IXGBE_NETLIST_ID_BLK_SHA_HASH_WORD(15)] << 16 |
drivers/net/ethernet/intel/ixgbe/ixgbe_type_e610.h
995
u32 hash; /* SHA-1 hash word */
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
147
mkex_hash->hash[intf][ld]);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
187
mkex_hash->hash[intf][ld]);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
433
u32 hash;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
446
hash = rvu_npc_toeplitz_hash(&ldata, (u64 *)hash_key, 64, 95);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
448
hash &= table->mem_table.hash_mask;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
449
hash += table->mem_table.hash_offset;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
450
dev_dbg(rvu->dev, "%s: hash=%x\n", __func__, hash);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
452
return hash;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
466
u32 *index, unsigned int hash)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
477
if (test_bit(hash + i * depth, table->mem_table.bmap))
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
480
set_bit(hash + i * depth, table->mem_table.bmap);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
484
__func__, i, hash);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
487
*index = hash;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
977
unsigned int hash;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
983
hash = rvu_exact_calculate_hash(rvu, chan, ctype, mac, table->mem_table.mask,
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
985
err = rvu_npc_exact_alloc_mem_table_entry(rvu, ways, index, hash);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.h
48
u64 hash[NPC_MAX_INTF][NPC_MAX_HASH];
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.h
89
.hash = {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
390
u8 hash[CN10K_MAX_HASH_LEN];
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
394
err = cn10k_ecb_aes_encrypt(pfvf, sak, secy->key_len, hash);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
404
hash_rev[i] = hash[CN10K_MAX_HASH_LEN - 1 - i];
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
50
u16 sak_len, u8 *hash)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
80
sg_init_one(&sg_dst, hash, CN10K_MAX_HASH_LEN);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
229
u32 hash = 0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
241
hash = cqe->hdr.flow_tag;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
243
skb_set_hash(skb, hash, hash_type);
drivers/net/ethernet/marvell/skge.h
2536
const u8 *hash)
drivers/net/ethernet/marvell/skge.h
2538
xm_write16(hw, port, reg, (u16)hash[0] | ((u16)hash[1] << 8));
drivers/net/ethernet/marvell/skge.h
2539
xm_write16(hw, port, reg+2, (u16)hash[2] | ((u16)hash[3] << 8));
drivers/net/ethernet/marvell/skge.h
2540
xm_write16(hw, port, reg+4, (u16)hash[4] | ((u16)hash[5] << 8));
drivers/net/ethernet/marvell/skge.h
2541
xm_write16(hw, port, reg+6, (u16)hash[6] | ((u16)hash[7] << 8));
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2196
u32 hash, reason;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2329
hash = trxd.rxd5 & MTK_RXD5_FOE_ENTRY;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2330
if (hash != MTK_RXD5_FOE_ENTRY)
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2331
skb_set_hash(skb, jhash_1word(hash, 0),
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2336
hash = trxd.rxd4 & MTK_RXD4_FOE_ENTRY;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2337
if (hash != MTK_RXD4_FOE_ENTRY)
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2338
skb_set_hash(skb, jhash_1word(hash, 0),
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2362
mtk_ppe_check_skb(eth->ppe[ppe_idx], skb, hash);
drivers/net/ethernet/mediatek/mtk_eth_soc.h
1402
mtk_foe_get_entry(struct mtk_ppe *ppe, u16 hash)
drivers/net/ethernet/mediatek/mtk_eth_soc.h
1406
return ppe->foe_table + hash * soc->foe_entry_size;
drivers/net/ethernet/mediatek/mtk_ppe.c
147
u32 hash;
drivers/net/ethernet/mediatek/mtk_ppe.c
174
hash = (hv1 & hv2) | ((~hv1) & hv3);
drivers/net/ethernet/mediatek/mtk_ppe.c
175
hash = (hash >> 24) | ((hash & 0xffffff) << 8);
drivers/net/ethernet/mediatek/mtk_ppe.c
176
hash ^= hv1 ^ hv2 ^ hv3;
drivers/net/ethernet/mediatek/mtk_ppe.c
177
hash ^= hash >> 16;
drivers/net/ethernet/mediatek/mtk_ppe.c
178
hash <<= (ffs(eth->soc->hash_offset) - 1);
drivers/net/ethernet/mediatek/mtk_ppe.c
179
hash &= MTK_PPE_ENTRIES - 1;
drivers/net/ethernet/mediatek/mtk_ppe.c
181
return hash;
drivers/net/ethernet/mediatek/mtk_ppe.c
517
if (entry->hash != 0xffff) {
drivers/net/ethernet/mediatek/mtk_ppe.c
518
struct mtk_foe_entry *hwe = mtk_foe_get_entry(ppe, entry->hash);
drivers/net/ethernet/mediatek/mtk_ppe.c
528
acct = ppe->acct_table + entry->hash * sizeof(*acct);
drivers/net/ethernet/mediatek/mtk_ppe.c
533
entry->hash = 0xffff;
drivers/net/ethernet/mediatek/mtk_ppe.c
568
hwe = mtk_foe_get_entry(ppe, cur->hash);
drivers/net/ethernet/mediatek/mtk_ppe.c
572
cur->hash = 0xffff;
drivers/net/ethernet/mediatek/mtk_ppe.c
600
if (entry->hash == 0xffff)
drivers/net/ethernet/mediatek/mtk_ppe.c
603
hwe = mtk_foe_get_entry(ppe, entry->hash);
drivers/net/ethernet/mediatek/mtk_ppe.c
606
entry->hash = 0xffff;
drivers/net/ethernet/mediatek/mtk_ppe.c
618
u16 hash)
drivers/net/ethernet/mediatek/mtk_ppe.c
635
hwe = mtk_foe_get_entry(ppe, hash);
drivers/net/ethernet/mediatek/mtk_ppe.c
683
u32 hash;
drivers/net/ethernet/mediatek/mtk_ppe.c
688
hash = mtk_ppe_hash_entry(ppe->eth, &entry->data);
drivers/net/ethernet/mediatek/mtk_ppe.c
689
entry->hash = 0xffff;
drivers/net/ethernet/mediatek/mtk_ppe.c
691
hlist_add_head(&entry->list, &ppe->foe_flow[hash / soc->hash_offset]);
drivers/net/ethernet/mediatek/mtk_ppe.c
699
u16 hash)
drivers/net/ethernet/mediatek/mtk_ppe.c
714
flow_info->hash = hash;
drivers/net/ethernet/mediatek/mtk_ppe.c
716
&ppe->foe_flow[hash / soc->hash_offset]);
drivers/net/ethernet/mediatek/mtk_ppe.c
719
hwe = mtk_foe_get_entry(ppe, hash);
drivers/net/ethernet/mediatek/mtk_ppe.c
735
__mtk_foe_entry_commit(ppe, &foe, hash);
drivers/net/ethernet/mediatek/mtk_ppe.c
738
void __mtk_ppe_check_skb(struct mtk_ppe *ppe, struct sk_buff *skb, u16 hash)
drivers/net/ethernet/mediatek/mtk_ppe.c
741
struct hlist_head *head = &ppe->foe_flow[hash / soc->hash_offset];
drivers/net/ethernet/mediatek/mtk_ppe.c
742
struct mtk_foe_entry *hwe = mtk_foe_get_entry(ppe, hash);
drivers/net/ethernet/mediatek/mtk_ppe.c
761
entry->hash = 0xffff;
drivers/net/ethernet/mediatek/mtk_ppe.c
767
if (entry->hash != 0xffff)
drivers/net/ethernet/mediatek/mtk_ppe.c
768
entry->hash = 0xffff;
drivers/net/ethernet/mediatek/mtk_ppe.c
772
entry->hash = hash;
drivers/net/ethernet/mediatek/mtk_ppe.c
773
__mtk_foe_entry_commit(ppe, &entry->data, hash);
drivers/net/ethernet/mediatek/mtk_ppe.c
811
mtk_foe_entry_commit_subflow(ppe, entry, hash);
drivers/net/ethernet/mediatek/mtk_ppe.h
298
u16 hash;
drivers/net/ethernet/mediatek/mtk_ppe.h
353
void __mtk_ppe_check_skb(struct mtk_ppe *ppe, struct sk_buff *skb, u16 hash);
drivers/net/ethernet/mediatek/mtk_ppe.h
356
mtk_ppe_check_skb(struct mtk_ppe *ppe, struct sk_buff *skb, u16 hash)
drivers/net/ethernet/mediatek/mtk_ppe.h
363
if (hash > MTK_PPE_HASH_MASK)
drivers/net/ethernet/mediatek/mtk_ppe.h
367
diff = now - ppe->foe_check_time[hash];
drivers/net/ethernet/mediatek/mtk_ppe.h
371
ppe->foe_check_time[hash] = now;
drivers/net/ethernet/mediatek/mtk_ppe.h
372
__mtk_ppe_check_skb(ppe, skb, hash);
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
555
if (entry->hash != 0xFFFF &&
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
556
mtk_foe_entry_get_mib(eth->ppe[entry->ppe_index], entry->hash,
drivers/net/ethernet/mediatek/mtk_wed.c
1960
u32 reason, u32 hash)
drivers/net/ethernet/mediatek/mtk_wed.c
1974
mtk_ppe_check_skb(eth->ppe[dev->hw->index], skb, hash);
drivers/net/ethernet/mellanox/mlx4/en_rx.c
655
int mlx4_en_xdp_rx_hash(const struct xdp_md *ctx, u32 *hash,
drivers/net/ethernet/mellanox/mlx4/en_rx.c
666
*hash = be32_to_cpu(cqe->immed_rss_invalid);
drivers/net/ethernet/mellanox/mlx4/mcg.c
108
u16 *hash, u8 op_mod)
drivers/net/ethernet/mellanox/mlx4/mcg.c
118
*hash = imm;
drivers/net/ethernet/mellanox/mlx4/mcg.c
703
u16 hash;
drivers/net/ethernet/mellanox/mlx4/mcg.c
714
err = mlx4_GID_HASH(dev, mailbox, &hash, op_mod);
drivers/net/ethernet/mellanox/mlx4/mcg.c
720
mlx4_dbg(dev, "Hash for %pI6 is %04x\n", gid, hash);
drivers/net/ethernet/mellanox/mlx4/mcg.c
722
*index = hash;
drivers/net/ethernet/mellanox/mlx4/mcg.c
731
if (*index != hash) {
drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
794
int mlx4_en_xdp_rx_hash(const struct xdp_md *ctx, u32 *hash,
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
412
&tracer->hash[mlx5_tracer_message_hash(tracer_event->string_event.tmsn)];
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
507
&tracer->hash[mlx5_tracer_message_hash(tracer_event->string_event.tmsn)];
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
584
hlist_for_each_entry_safe(str_frmt, n, &tracer->hash[i], hlist)
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.h
110
struct hlist_head hash[MESSAGE_HASH_SIZE];
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
164
rss->hash.symmetric = true;
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
165
rss->hash.hfunc = ETH_RSS_HASH_TOP;
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
166
netdev_rss_key_fill(rss->hash.toeplitz_hash_key,
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
167
sizeof(rss->hash.toeplitz_hash_key));
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
236
mlx5e_tir_builder_build_rss(builder, &rss->hash, &rss_tt, inner);
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
319
mlx5e_tir_builder_build_rss(builder, &rss->hash, &rss_tt, inner);
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
504
err = mlx5e_rqt_redirect_indir(&rss->rqt, rqns, vhca_ids, num_rqns, rss->hash.hfunc,
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
587
memcpy(key, rss->hash.toeplitz_hash_key,
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
588
sizeof(rss->hash.toeplitz_hash_key));
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
591
*hfunc = rss->hash.hfunc;
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
594
*symmetric = rss->hash.symmetric;
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
610
if (hfunc && *hfunc != rss->hash.hfunc) {
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
621
rss->hash.hfunc = *hfunc;
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
625
if (rss->hash.hfunc == ETH_RSS_HASH_TOP)
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
627
memcpy(rss->hash.toeplitz_hash_key, key,
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
628
sizeof(rss->hash.toeplitz_hash_key));
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
639
rss->hash.symmetric = *symmetric;
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
663
return rss->hash;
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
71
struct mlx5e_rss_params_hash hash;
drivers/net/ethernet/mellanox/mlx5/core/en/tc/act_stats.c
122
rhashtable_remove_fast(ht, &act_stats->hash,
drivers/net/ethernet/mellanox/mlx5/core/en/tc/act_stats.c
22
struct rhash_head hash;
drivers/net/ethernet/mellanox/mlx5/core/en/tc/act_stats.c
27
.head_offset = offsetof(struct mlx5e_tc_act_stats, hash),
drivers/net/ethernet/mellanox/mlx5/core/en/tc/act_stats.c
83
&act_stats->hash,
drivers/net/ethernet/mellanox/mlx5/core/en/tir.c
120
MLX5_SET(modify_tir_in, builder->in, bitmask.hash, 1);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
237
static int mlx5e_xdp_rx_hash(const struct xdp_md *ctx, u32 *hash,
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
247
*hash = be32_to_cpu(cqe->rss_hash_result);
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
102
#define mlx5e_for_each_hash_arfs_rule(hn, tmp, hash, j) \
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
104
hlist_for_each_entry_safe(hn, tmp, &hash[j], hlist)
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
114
static void mlx5e_add_l2_to_hash(struct hlist_head *hash, const u8 *addr)
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
120
hlist_for_each_entry(hn, &hash[ix], hlist)
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
138
hlist_add_head(&hn->hlist, &hash[ix]);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
575
#define mlx5e_for_each_hash_node(hn, tmp, hash, i) \
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
577
hlist_for_each_entry_safe(hn, tmp, &hash[i], hlist)
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
417
struct hlist_head *hash = esw->mc_table;
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
425
esw_mc = l2addr_hash_find(hash, mac, struct esw_mc_addr);
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
429
esw_mc = l2addr_hash_add(hash, mac, struct esw_mc_addr, GFP_KERNEL);
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
457
struct hlist_head *hash = esw->mc_table;
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
465
esw_mc = l2addr_hash_find(hash, mac, struct esw_mc_addr);
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
506
struct hlist_head *hash;
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
515
hash = is_uc ? vport->uc_list : vport->mc_list;
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
516
for_each_l2hash_node(node, tmp, hash, hi) {
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
539
struct hlist_head *hash;
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
553
hash = is_uc ? vport->uc_list : vport->mc_list;
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
555
for_each_l2hash_node(node, tmp, hash, hi) {
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
577
addr = l2addr_hash_find(hash, mac_list[i], struct vport_addr);
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
601
addr = l2addr_hash_add(hash, mac_list[i], struct vport_addr,
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
624
struct hlist_head *hash;
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
628
hash = vport->mc_list;
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
633
addr = l2addr_hash_find(hash, mac, struct vport_addr);
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
639
addr = l2addr_hash_add(hash, mac, struct vport_addr,
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
24
u32 hash;
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
26
hash = jhash_1word(flow_act->action, 0);
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
27
hash = jhash((const void *)&flow_act->vlan,
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
28
sizeof(flow_act->vlan), hash);
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
29
hash = jhash((const void *)&dest->vport.num,
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
30
sizeof(dest->vport.num), hash);
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
31
hash = jhash((const void *)&dest->vport.vhca_id,
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
32
sizeof(dest->vport.num), hash);
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
34
hash = jhash(flow_act->pkt_reformat,
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
36
hash);
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads_termtbl.c
37
return hash;
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
2090
rhl_for_each_entry_rcu(g, tmp, list, hash) {
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
344
.head_offset = offsetof(struct fs_fte, hash),
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
352
.head_offset = offsetof(struct mlx5_flow_group, hash),
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
770
&fte->hash,
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
813
&fg->hash,
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
831
&fte->hash,
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
926
&fg->hash,
drivers/net/ethernet/mellanox/mlx5/core/fs_core.h
254
struct rhash_head hash;
drivers/net/ethernet/mellanox/mlx5/core/fs_core.h
291
struct rhlist_head hash;
drivers/net/ethernet/mellanox/mlx5/core/lag/debugfs.c
106
bool hash = false;
drivers/net/ethernet/mellanox/mlx5/core/lag/debugfs.c
118
hash = true;
drivers/net/ethernet/mellanox/mlx5/core/lag/debugfs.c
130
if (hash)
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
202
enum netdev_lag_hash hash)
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
278
if (hash == NETDEV_LAG_HASH_VLAN_SRCMAC) {
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
298
mlx5_lag_create_definer(struct mlx5_lag *ldev, enum netdev_lag_hash hash,
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
323
format_id = mlx5_lag_set_definer(match_definer_mask, tt, tunnel, hash);
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
425
enum netdev_lag_hash hash)
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
429
switch (hash) {
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
165
.head_offset = offsetof(struct mlx5_fs_id, hash),
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
173
.head_offset = offsetof(struct mlx5_fs_id, hash),
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
680
rhashtable_remove_fast(hash_table, &fs_id_found->hash, *rhash);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
748
err = rhashtable_insert_fast(hash_table, &fs_id_iter->hash, *rhash);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
96
struct rhash_head hash;
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.c
131
WARN_ON(!hlist_empty(mpfs->hash));
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.c
148
l2addr = l2addr_hash_find(mpfs->hash, mac, struct l2table_node);
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.c
154
l2addr = l2addr_hash_add(mpfs->hash, mac, struct l2table_node, GFP_KERNEL);
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.c
200
l2addr = l2addr_hash_find(mpfs->hash, mac, struct l2table_node);
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.c
73
struct hlist_head hash[MLX5_L2_ADDR_HASH_SIZE];
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h
50
hlist_for_each_entry_safe(hs, tmp, &(mpfs)->hash[i], node.hlist)
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h
52
#define for_each_l2hash_node(hn, tmp, hash, i) \
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h
54
hlist_for_each_entry_safe(hn, tmp, &(hash)[i], hlist)
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h
56
#define l2addr_hash_find(hash, mac, type) ({ \
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h
61
hlist_for_each_entry(ptr, &(hash)[ix], node.hlist) \
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h
71
#define l2addr_hash_add(hash, mac, type, gfp) ({ \
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h
78
hlist_add_head(&ptr->node.hlist, &(hash)[ix]);\
drivers/net/ethernet/microchip/lan865x/lan865x.c
121
u32 hash = 0;
drivers/net/ethernet/microchip/lan865x/lan865x.c
124
hash ^= get_address_bit(addr, (j * 6) + i);
drivers/net/ethernet/microchip/lan865x/lan865x.c
126
hash_index |= (hash << i);
drivers/net/ethernet/microsoft/mana/mana_en.c
634
u32 hash = skb_get_hash(skb);
drivers/net/ethernet/microsoft/mana/mana_en.c
638
txq = apc->indir_table[hash & (apc->indir_table_sz - 1)];
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
203
cfg->ctrl_word.hash = NFP_IPSEC_HASH_MD5_96;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
206
cfg->ctrl_word.hash = NFP_IPSEC_HASH_MD5_128;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
217
cfg->ctrl_word.hash = NFP_IPSEC_HASH_SHA1_96;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
220
cfg->ctrl_word.hash = NFP_IPSEC_HASH_SHA1_80;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
231
cfg->ctrl_word.hash = NFP_IPSEC_HASH_SHA256_96;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
234
cfg->ctrl_word.hash = NFP_IPSEC_HASH_SHA256_128;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
245
cfg->ctrl_word.hash = NFP_IPSEC_HASH_SHA384_96;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
248
cfg->ctrl_word.hash = NFP_IPSEC_HASH_SHA384_192;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
259
cfg->ctrl_word.hash = NFP_IPSEC_HASH_SHA512_96;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
262
cfg->ctrl_word.hash = NFP_IPSEC_HASH_SHA512_256;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
335
cfg->ctrl_word.hash = NFP_IPSEC_HASH_NONE;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
404
cfg->ctrl_word.hash = NFP_IPSEC_HASH_POLY1305_128;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
433
cfg->ctrl_word.hash = NFP_IPSEC_HASH_GF128_128;
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
98
uint32_t hash :4; /* From nfp_ipsec_sa_hash_type */
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1095
skb_set_hash(skb, meta.hash, meta.hash_type);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
710
unsigned int type, __be32 *hash)
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
726
meta->hash = get_unaligned_be32(hash);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
739
&rx_hash->hash);
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
90
skb_set_hash(skb, meta->hash, meta->hash_type);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1214
skb_set_hash(skb, meta.hash, meta.hash_type);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
723
unsigned int type, __be32 *hash)
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
739
meta->hash = get_unaligned_be32(hash);
drivers/net/ethernet/netronome/nfp/nfp_net.h
260
u32 hash;
drivers/net/ethernet/netronome/nfp/nfp_net.h
277
__be32 hash;
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
124
u32 hash, struct ionic_admin_ctx *ctx,
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.h
39
u32 hash, struct ionic_admin_ctx *ctx,
drivers/net/ethernet/qlogic/qede/qede_fp.c
613
u32 hash = 0;
drivers/net/ethernet/qlogic/qede/qede_fp.c
620
hash = le32_to_cpu(rss_hash);
drivers/net/ethernet/qlogic/qede/qede_fp.c
622
skb_set_hash(skb, hash, hash_type);
drivers/net/ethernet/sfc/falcon/farch.c
2346
unsigned int hash = ef4_farch_filter_hash(key);
drivers/net/ethernet/sfc/falcon/farch.c
2353
unsigned int i = hash & (table->size - 1);
drivers/net/ethernet/sfc/mcdi_filters.c
351
unsigned int match_pri, hash;
drivers/net/ethernet/sfc/mcdi_filters.c
377
hash = efx_filter_spec_hash(spec);
drivers/net/ethernet/sfc/mcdi_filters.c
403
i = (hash + depth) & (EFX_MCDI_FILTER_TBL_ROWS - 1);
drivers/net/ethernet/sfc/mcdi_filters.c
521
i = (hash + depth) & (EFX_MCDI_FILTER_TBL_ROWS - 1);
drivers/net/ethernet/sfc/rx_common.c
666
u32 hash = efx_filter_spec_hash(spec);
drivers/net/ethernet/sfc/rx_common.c
671
return &efx->rps_hash_table[hash % EFX_ARFS_HASH_TABLE_SIZE];
drivers/net/ethernet/sfc/siena/farch.c
2439
unsigned int hash = efx_farch_filter_hash(key);
drivers/net/ethernet/sfc/siena/farch.c
2446
unsigned int i = hash & (table->size - 1);
drivers/net/ethernet/sfc/siena/rx_common.c
653
u32 hash = efx_siena_filter_spec_hash(spec);
drivers/net/ethernet/sfc/siena/rx_common.c
658
return &efx->rps_hash_table[hash % EFX_ARFS_HASH_TABLE_SIZE];
drivers/net/ethernet/spacemit/k1_emac.c
879
u32 hash, reg, bit, val;
drivers/net/ethernet/spacemit/k1_emac.c
902
hash = emac_ether_addr_hash(ha->addr);
drivers/net/ethernet/spacemit/k1_emac.c
903
reg = hash / 16;
drivers/net/ethernet/spacemit/k1_emac.c
904
bit = hash % 16;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_descs.c
272
static int dwxgmac2_get_rx_hash(struct dma_desc *p, u32 *hash,
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_descs.c
293
*hash = le32_to_cpu(p->des1);
drivers/net/ethernet/stmicro/stmmac/hwif.h
642
void (*update_vlan_hash)(struct mac_device_info *hw, u32 hash,
drivers/net/ethernet/stmicro/stmmac/hwif.h
98
int (*get_rx_hash)(struct dma_desc *p, u32 *hash,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5298
u32 hash;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5320
if (!stmmac_get_rx_hash(priv, p, &hash, &hash_type))
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5321
skb_set_hash(skb, hash, hash_type);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5635
u32 hash;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5840
if (!stmmac_get_rx_hash(priv, p, &hash, &hash_type))
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5841
skb_set_hash(skb, hash, hash_type);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6752
u32 crc, hash = 0;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6760
hash |= (1 << crc);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6769
hash = 0;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6775
return stmmac_update_vlan_hash(priv, priv->hw, hash, pmatch, is_double);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
308
if (tpriv->packet->exp_hash && !skb->hash)
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
500
u32 hash, hash_nr;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
503
hash = bitrev32(~crc32_le(~0, addr, 6)) >> mc_offset;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
504
hash_nr = hash >> 5;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
505
hash = 1 << (hash & 0x1f);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
510
if (((nr >> 5) == hash_nr) && ((1 << (nr & 0x1f)) == hash))
drivers/net/ethernet/stmicro/stmmac/stmmac_vlan.c
164
static void vlan_update_hash(struct mac_device_info *hw, u32 hash,
drivers/net/ethernet/stmicro/stmmac/stmmac_vlan.c
170
writel(hash, ioaddr + VLAN_HASH_TABLE);
drivers/net/ethernet/stmicro/stmmac/stmmac_vlan.c
174
if (hash) {
drivers/net/ethernet/stmicro/stmmac/stmmac_vlan.c
253
static void dwxgmac2_update_vlan_hash(struct mac_device_info *hw, u32 hash,
drivers/net/ethernet/stmicro/stmmac/stmmac_vlan.c
258
writel(hash, ioaddr + VLAN_HASH_TABLE);
drivers/net/ethernet/stmicro/stmmac/stmmac_vlan.c
260
if (hash) {
drivers/net/ethernet/sun/niu.c
6304
static void niu_load_hash_xmac(struct niu *np, u16 *hash)
drivers/net/ethernet/sun/niu.c
6309
nw64_mac(XMAC_HASH_TBL(i), hash[i]);
drivers/net/ethernet/sun/niu.c
6312
static void niu_load_hash_bmac(struct niu *np, u16 *hash)
drivers/net/ethernet/sun/niu.c
6317
nw64_mac(BMAC_HASH_TBL(i), hash[i]);
drivers/net/ethernet/sun/niu.c
6320
static void niu_load_hash(struct niu *np, u16 *hash)
drivers/net/ethernet/sun/niu.c
6323
niu_load_hash_xmac(np, hash);
drivers/net/ethernet/sun/niu.c
6325
niu_load_hash_bmac(np, hash);
drivers/net/ethernet/sun/niu.c
6334
u16 hash[16] = { 0, };
drivers/net/ethernet/sun/niu.c
6381
hash[i] = 0xffff;
drivers/net/ethernet/sun/niu.c
6387
hash[crc >> 4] |= (1 << (15 - (crc & 0xf)));
drivers/net/ethernet/sun/niu.c
6392
niu_load_hash(np, hash);
drivers/net/ethernet/sun/sunvnet.c
193
unsigned int hash = vnet_hashfn(skb->data);
drivers/net/ethernet/sun/sunvnet.c
194
struct hlist_head *hp = &vp->port_hash[hash];
drivers/net/ethernet/sun/sunvnet.c
197
hlist_for_each_entry_rcu(port, hp, hash) {
drivers/net/ethernet/sun/sunvnet.c
461
INIT_HLIST_NODE(&port->hash);
drivers/net/ethernet/sun/sunvnet.c
476
hlist_add_head_rcu(&port->hash,
drivers/net/ethernet/sun/sunvnet.c
513
hlist_del_rcu(&port->hash);
drivers/net/ethernet/sun/sunvnet_common.h
63
struct hlist_node hash;
drivers/net/ethernet/tehuti/tehuti.c
787
u8 hash;
drivers/net/ethernet/tehuti/tehuti.c
806
hash = 0;
drivers/net/ethernet/tehuti/tehuti.c
808
hash ^= ha->addr[i];
drivers/net/ethernet/tehuti/tehuti.c
809
reg = regRX_MCST_HASH0 + ((hash >> 5) << 2);
drivers/net/ethernet/tehuti/tehuti.c
811
val |= (1 << (hash % 32));
drivers/net/ethernet/tehuti/tn40.c
1483
u8 hash;
drivers/net/ethernet/tehuti/tn40.c
1505
hash = 0;
drivers/net/ethernet/tehuti/tn40.c
1507
hash ^= mclist->addr[i];
drivers/net/ethernet/tehuti/tn40.c
1509
reg = TN40_REG_RX_MCST_HASH0 + ((hash >> 5) << 2);
drivers/net/ethernet/tehuti/tn40.c
1511
val |= (1 << (hash % 32));
drivers/net/ethernet/ti/davinci_emac.c
586
u32 hash;
drivers/net/ethernet/ti/davinci_emac.c
589
hash = 0;
drivers/net/ethernet/ti/davinci_emac.c
593
hash ^= (tmpval >> 2) ^ (tmpval << 4);
drivers/net/ethernet/ti/davinci_emac.c
595
hash ^= (tmpval >> 4) ^ (tmpval << 2);
drivers/net/ethernet/ti/davinci_emac.c
597
hash ^= (tmpval >> 6) ^ (tmpval);
drivers/net/ethernet/ti/davinci_emac.c
600
return hash & 0x3F;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1314
static void icssm_emac_mc_filter_bin_update(struct prueth_emac *emac, u8 hash,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1324
writeb(val, mc_filter_tbl + hash);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1327
void icssm_emac_mc_filter_bin_allow(struct prueth_emac *emac, u8 hash)
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1330
(emac, hash,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1334
void icssm_emac_mc_filter_bin_disallow(struct prueth_emac *emac, u8 hash)
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1337
(emac, hash,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1343
u8 hash;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1346
for (j = 0, hash = 0; j < ETH_ALEN; j++)
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1347
hash ^= (mac[j] & mask[j]);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1349
return hash;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1368
u8 hash;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1418
hash = icssm_emac_get_mc_hash(ha->addr, emac->mc_filter_mask);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1419
icssm_emac_mc_filter_bin_allow(emac, hash);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1425
hash = icssm_emac_get_mc_hash(ha->addr,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1427
icssm_emac_mc_filter_bin_allow(emac, hash);
drivers/net/ethernet/ti/icssm/icssm_prueth.h
277
void icssm_emac_mc_filter_bin_allow(struct prueth_emac *emac, u8 hash);
drivers/net/ethernet/ti/icssm/icssm_prueth.h
278
void icssm_emac_mc_filter_bin_disallow(struct prueth_emac *emac, u8 hash);
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
409
unsigned int hash, hash_prev;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
427
hash = icssm_prueth_sw_fdb_hash(mac);
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
430
if (hash != hash_prev)
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
431
writew(i, &FDB_IDX_TBL_ENTRY(hash)->bucket_idx);
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
433
hash_prev = hash;
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
213
u8 hash;
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
219
hash = icssm_emac_get_mc_hash(mdb->addr, emac->mc_filter_mask);
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
220
icssm_emac_mc_filter_bin_allow(emac, hash);
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
238
u8 hash, tmp_hash;
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
248
hash = icssm_emac_get_mc_hash(mdb->addr, mask);
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
255
if (tmp_hash == hash)
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
258
icssm_emac_mc_filter_bin_disallow(emac, hash);
drivers/net/ethernet/ti/tlan.h
533
u8 hash;
drivers/net/ethernet/ti/tlan.h
535
hash = (a[0]^a[3]); /* & 077 */
drivers/net/ethernet/ti/tlan.h
536
hash ^= ((a[0]^a[3])>>6); /* & 003 */
drivers/net/ethernet/ti/tlan.h
537
hash ^= ((a[1]^a[4])<<2); /* & 074 */
drivers/net/ethernet/ti/tlan.h
538
hash ^= ((a[1]^a[4])>>4); /* & 017 */
drivers/net/ethernet/ti/tlan.h
539
hash ^= ((a[2]^a[5])<<4); /* & 060 */
drivers/net/ethernet/ti/tlan.h
540
hash ^= ((a[2]^a[5])>>2); /* & 077 */
drivers/net/ethernet/ti/tlan.h
542
return hash & 077;
drivers/net/ethernet/tundra/tsi108_eth.c
1176
u32 hash, crc;
drivers/net/ethernet/tundra/tsi108_eth.c
1179
hash = crc >> 23;
drivers/net/ethernet/tundra/tsi108_eth.c
1180
__set_bit(hash, &data->mc_hash[0]);
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
51
u32 *hash)
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
90
*hash = sig_hash ^ bucket_hash;
drivers/net/geneve.c
1069
__u32 hash;
drivers/net/geneve.c
1099
hash = geneve_net_vni_hash(vni);
drivers/net/geneve.c
1100
hlist_add_head_rcu(&node->hlist, &gs->vni_list[hash]);
drivers/net/geneve.c
179
__u32 hash;
drivers/net/geneve.c
182
hash = geneve_net_vni_hash(vni);
drivers/net/geneve.c
183
vni_list_head = &gs->vni_list[hash];
drivers/net/geneve.c
198
__u32 hash;
drivers/net/geneve.c
201
hash = geneve_net_vni_hash(vni);
drivers/net/geneve.c
202
vni_list_head = &gs->vni_list[hash];
drivers/net/hyperv/hyperv_net.h
1088
u32 hash, pkt_proto = 0;
drivers/net/hyperv/hyperv_net.h
1118
hash = jhash2((u32 *)&flow.addrs.v4addrs, 2, hashrnd);
drivers/net/hyperv/hyperv_net.h
1120
hash = jhash2((u32 *)&flow.addrs.v6addrs, 8, hashrnd);
drivers/net/hyperv/hyperv_net.h
1124
__skb_set_sw_hash(skb, hash, false);
drivers/net/hyperv/hyperv_net.h
1127
return hash;
drivers/net/hyperv/netvsc_drv.c
451
u32 hash;
drivers/net/hyperv/netvsc_drv.c
519
hash = skb_get_hash_raw(skb);
drivers/net/hyperv/netvsc_drv.c
520
if (hash != 0 && net->real_num_tx_queues > 1) {
drivers/net/hyperv/netvsc_drv.c
526
*hash_info = hash;
drivers/net/ipvlan/ipvlan_core.c
105
u8 hash;
drivers/net/ipvlan/ipvlan_core.c
107
hash = (addr->atype == IPVL_IPV6) ?
drivers/net/ipvlan/ipvlan_core.c
111
hlist_add_head_rcu(&addr->hlnode, &port->hlhead[hash]);
drivers/net/ipvlan/ipvlan_core.c
232
u32 hash = jhash_1word(get_unaligned((u32 *)(addr + 2)),
drivers/net/ipvlan/ipvlan_core.c
235
return hash & IPVLAN_MAC_FILTER_MASK;
drivers/net/ipvlan/ipvlan_core.c
79
u8 hash;
drivers/net/ipvlan/ipvlan_core.c
81
hash = ipvlan_get_v6_hash(iaddr);
drivers/net/ipvlan/ipvlan_core.c
82
hlist_for_each_entry_rcu(addr, &port->hlhead[hash], hlnode)
drivers/net/ipvlan/ipvlan_core.c
93
u8 hash;
drivers/net/ipvlan/ipvlan_core.c
95
hash = ipvlan_get_v4_hash(addr4);
drivers/net/ipvlan/ipvlan_core.c
96
hlist_for_each_entry_rcu(addr, &port->hlhead[hash], hlnode)
drivers/net/macvlan.c
274
unsigned int hash;
drivers/net/macvlan.c
283
hash = mc_hash(vlan, eth->h_dest);
drivers/net/macvlan.c
284
if (!test_bit(hash, vlan->mc_filter))
drivers/net/macvlan.c
470
unsigned int hash;
drivers/net/macvlan.c
492
hash = mc_hash(NULL, eth->h_dest);
drivers/net/macvlan.c
493
if (test_bit(hash, port->bc_filter))
drivers/net/macvlan.c
495
else if (test_bit(hash, port->mc_filter))
drivers/net/phy/dp83640.c
100
u16 hash;
drivers/net/phy/dp83640.c
284
rxts->hash = p->msgtype & 0x0fff;
drivers/net/phy/dp83640.c
800
u16 hash;
drivers/net/phy/dp83640.c
817
hash = ether_crc(DP83640_PACKET_HASH_LEN,
drivers/net/phy/dp83640.c
819
if (rxts->hash != hash)
drivers/net/ppp/pppoe.c
1095
int hash = hash_item(po->pppoe_pa.sid, po->pppoe_pa.remote);
drivers/net/ppp/pppoe.c
1098
while (++hash < PPPOE_HASH_SIZE) {
drivers/net/ppp/pppoe.c
1099
po = rcu_dereference(pn->hash_table[hash]);
drivers/net/ppp/pppoe.c
141
unsigned char hash = 0;
drivers/net/ppp/pppoe.c
145
hash ^= addr[i];
drivers/net/ppp/pppoe.c
147
hash ^= (__force __u32)sid >> i;
drivers/net/ppp/pppoe.c
149
hash ^= hash >> i;
drivers/net/ppp/pppoe.c
151
return hash & PPPOE_HASH_MASK;
drivers/net/ppp/pppoe.c
162
int hash = hash_item(sid, addr);
drivers/net/ppp/pppoe.c
165
ret = rcu_dereference(pn->hash_table[hash]);
drivers/net/ppp/pppoe.c
179
int hash = hash_item(po->pppoe_pa.sid, po->pppoe_pa.remote);
drivers/net/ppp/pppoe.c
182
first = rcu_dereference_protected(pn->hash_table[hash], lockdep_is_held(&pn->hash_lock));
drivers/net/ppp/pppoe.c
193
rcu_assign_pointer(pn->hash_table[hash], po);
drivers/net/ppp/pppoe.c
201
int hash = hash_item(sid, addr);
drivers/net/ppp/pppoe.c
204
ret = rcu_dereference_protected(pn->hash_table[hash], lockdep_is_held(&pn->hash_lock));
drivers/net/ppp/pppoe.c
205
src = &pn->hash_table[hash];
drivers/net/team/team_mode_loadbalance.c
119
unsigned char hash)
drivers/net/team/team_mode_loadbalance.c
121
int port_index = team_num_to_port_index(team, hash);
drivers/net/team/team_mode_loadbalance.c
128
unsigned char hash)
drivers/net/team/team_mode_loadbalance.c
133
port = rcu_dereference_bh(LB_HTPM_PORT_BY_HASH(lb_priv, hash));
drivers/net/team/team_mode_loadbalance.c
137
return lb_hash_select_tx_port(team, hash);
drivers/net/team/team_mode_loadbalance.c
202
unsigned char hash)
drivers/net/team/team_mode_loadbalance.c
210
hash_stats = &pcpu_stats->hash_stats[hash];
drivers/net/team/team_mode_loadbalance.c
222
unsigned char hash;
drivers/net/team/team_mode_loadbalance.c
225
hash = lb_get_skb_hash(lb_priv, skb);
drivers/net/team/team_mode_loadbalance.c
227
port = select_tx_port_func(team, hash);
drivers/net/team/team_mode_loadbalance.c
232
lb_update_tx_stats(tx_bytes, lb_priv, get_lb_port_priv(port), hash);
drivers/net/team/team_mode_loadbalance.c
358
unsigned char hash = info->array_index;
drivers/net/team/team_mode_loadbalance.c
360
LB_HTPM_OPT_INST_INFO_BY_HASH(lb_priv, hash) = info;
drivers/net/team/team_mode_loadbalance.c
368
unsigned char hash = ctx->info->array_index;
drivers/net/team/team_mode_loadbalance.c
370
port = LB_HTPM_PORT_BY_HASH(lb_priv, hash);
drivers/net/team/team_mode_loadbalance.c
379
unsigned char hash = ctx->info->array_index;
drivers/net/team/team_mode_loadbalance.c
384
rcu_assign_pointer(LB_HTPM_PORT_BY_HASH(lb_priv, hash),
drivers/net/team/team_mode_loadbalance.c
396
unsigned char hash = info->array_index;
drivers/net/team/team_mode_loadbalance.c
398
lb_priv->ex->stats.info[hash].opt_inst_info = info;
drivers/net/team/team_mode_loadbalance.c
404
unsigned char hash = ctx->info->array_index;
drivers/net/team/team_mode_loadbalance.c
406
ctx->data.bin_val.ptr = &lb_priv->ex->stats.info[hash].stats;
drivers/net/team/team_mode_loadbalance.c
90
#define LB_HTPM_PORT_BY_HASH(lp_priv, hash) \
drivers/net/team/team_mode_loadbalance.c
91
(lb_priv)->ex->tx_hash_to_port_mapping[hash].port
drivers/net/team/team_mode_loadbalance.c
93
#define LB_HTPM_OPT_INST_INFO_BY_HASH(lp_priv, hash) \
drivers/net/team/team_mode_loadbalance.c
94
(lb_priv)->ex->tx_hash_to_port_mapping[hash].opt_inst_info
drivers/net/thunderbolt/main.c
1322
u32 hash;
drivers/net/thunderbolt/main.c
1328
hash = jhash2((u32 *)xd->local_uuid, 4, 0);
drivers/net/thunderbolt/main.c
1329
memcpy(addr + 1, &hash, sizeof(hash));
drivers/net/thunderbolt/main.c
1330
hash = jhash2((u32 *)xd->local_uuid, 4, hash);
drivers/net/thunderbolt/main.c
1331
addr[5] = hash & 0xff;
drivers/net/tun.c
449
static inline void tun_flow_save_rps_rxhash(struct tun_flow_entry *e, u32 hash)
drivers/net/tun.c
451
if (unlikely(e->rps_rxhash != hash))
drivers/net/tun.c
452
e->rps_rxhash = hash;
drivers/net/usb/catc.c
206
#define f5u011_mchash_async(catc, hash) catc_ctrl_async(catc, USB_DIR_OUT, SetRxMode, 0, 2, &hash, 8, NULL)
drivers/net/veth.c
1667
static int veth_xdp_rx_hash(const struct xdp_md *ctx, u32 *hash,
drivers/net/veth.c
1676
*hash = skb_get_hash(skb);
drivers/net/virtio_net.c
6675
static int virtnet_xdp_rx_hash(const struct xdp_md *_ctx, u32 *hash,
drivers/net/virtio_net.c
6694
*hash = virtio_net_hash_value(hdr_hash);
drivers/net/vxlan/vxlan_core.c
2656
u32 hash;
drivers/net/vxlan/vxlan_core.c
2659
hash = skb_get_hash(skb);
drivers/net/vxlan/vxlan_core.c
2664
do_xmit = vxlan_fdb_nh_path_select(nh, hash, &nh_rdst);
drivers/net/vxlan/vxlan_core.c
2687
u32 hash;
drivers/net/vxlan/vxlan_core.c
2690
hash = skb_get_hash(skb);
drivers/net/vxlan/vxlan_core.c
2698
do_xmit = vxlan_fdb_nh_path_select(nh, hash, &nh_rdst);
drivers/net/wireguard/noise.c
432
static void mix_hash(u8 hash[NOISE_HASH_LEN], const u8 *src, size_t src_len)
drivers/net/wireguard/noise.c
437
blake2s_update(&blake, hash, NOISE_HASH_LEN);
drivers/net/wireguard/noise.c
439
blake2s_final(&blake, hash);
drivers/net/wireguard/noise.c
442
static void mix_psk(u8 chaining_key[NOISE_HASH_LEN], u8 hash[NOISE_HASH_LEN],
drivers/net/wireguard/noise.c
450
mix_hash(hash, temp_hash, NOISE_HASH_LEN);
drivers/net/wireguard/noise.c
455
u8 hash[NOISE_HASH_LEN],
drivers/net/wireguard/noise.c
458
memcpy(hash, handshake_init_hash, NOISE_HASH_LEN);
drivers/net/wireguard/noise.c
460
mix_hash(hash, remote_static, NOISE_PUBLIC_KEY_LEN);
drivers/net/wireguard/noise.c
465
u8 hash[NOISE_HASH_LEN])
drivers/net/wireguard/noise.c
467
chacha20poly1305_encrypt(dst_ciphertext, src_plaintext, src_len, hash,
drivers/net/wireguard/noise.c
470
mix_hash(hash, dst_ciphertext, noise_encrypted_len(src_len));
drivers/net/wireguard/noise.c
475
u8 hash[NOISE_HASH_LEN])
drivers/net/wireguard/noise.c
478
hash, NOISE_HASH_LEN,
drivers/net/wireguard/noise.c
481
mix_hash(hash, src_ciphertext, src_len);
drivers/net/wireguard/noise.c
488
u8 hash[NOISE_HASH_LEN])
drivers/net/wireguard/noise.c
492
mix_hash(hash, ephemeral_src, NOISE_PUBLIC_KEY_LEN);
drivers/net/wireguard/noise.c
537
handshake_init(handshake->chaining_key, handshake->hash,
drivers/net/wireguard/noise.c
547
handshake->hash);
drivers/net/wireguard/noise.c
557
NOISE_PUBLIC_KEY_LEN, key, handshake->hash);
drivers/net/wireguard/noise.c
567
NOISE_TIMESTAMP_LEN, key, handshake->hash);
drivers/net/wireguard/noise.c
592
u8 hash[NOISE_HASH_LEN];
drivers/net/wireguard/noise.c
602
handshake_init(chaining_key, hash, wg->static_identity.static_public);
drivers/net/wireguard/noise.c
605
message_ephemeral(e, src->unencrypted_ephemeral, chaining_key, hash);
drivers/net/wireguard/noise.c
613
sizeof(src->encrypted_static), key, hash))
drivers/net/wireguard/noise.c
629
sizeof(src->encrypted_timestamp), key, hash))
drivers/net/wireguard/noise.c
647
memcpy(handshake->hash, hash, NOISE_HASH_LEN);
drivers/net/wireguard/noise.c
659
memzero_explicit(hash, NOISE_HASH_LEN);
drivers/net/wireguard/noise.c
694
handshake->hash);
drivers/net/wireguard/noise.c
707
mix_psk(handshake->chaining_key, handshake->hash, key,
drivers/net/wireguard/noise.c
711
message_encrypt(dst->encrypted_nothing, NULL, 0, key, handshake->hash);
drivers/net/wireguard/noise.c
735
u8 hash[NOISE_HASH_LEN];
drivers/net/wireguard/noise.c
755
memcpy(hash, handshake->hash, NOISE_HASH_LEN);
drivers/net/wireguard/noise.c
767
message_ephemeral(e, src->unencrypted_ephemeral, chaining_key, hash);
drivers/net/wireguard/noise.c
778
mix_psk(chaining_key, hash, key, preshared_key);
drivers/net/wireguard/noise.c
782
sizeof(src->encrypted_nothing), key, hash))
drivers/net/wireguard/noise.c
795
memcpy(handshake->hash, hash, NOISE_HASH_LEN);
drivers/net/wireguard/noise.c
807
memzero_explicit(hash, NOISE_HASH_LEN);
drivers/net/wireguard/noise.c
82
memset(&handshake->hash, 0, NOISE_HASH_LEN);
drivers/net/wireguard/noise.h
80
u8 hash[NOISE_HASH_LEN];
drivers/net/wireguard/peerlookup.c
17
const u64 hash = siphash(pubkey, NOISE_PUBLIC_KEY_LEN, &table->key);
drivers/net/wireguard/peerlookup.c
19
return &table->hashtable[hash & (HASH_SIZE(table->hashtable) - 1)];
drivers/net/wireguard/queueing.h
80
u32 hash = skb->hash;
drivers/net/wireguard/queueing.h
86
skb->hash = hash;
drivers/net/wireguard/ratelimiter.c
111
hlist_for_each_entry_rcu(entry, bucket, hash) {
drivers/net/wireguard/ratelimiter.c
144
INIT_HLIST_NODE(&entry->hash);
drivers/net/wireguard/ratelimiter.c
149
hlist_add_head_rcu(&entry->hash, bucket);
drivers/net/wireguard/ratelimiter.c
30
struct hlist_node hash;
drivers/net/wireguard/ratelimiter.c
50
hlist_del_rcu(&entry->hash);
drivers/net/wireguard/ratelimiter.c
64
hlist_for_each_entry_safe(entry, temp, &table_v4[i], hash) {
drivers/net/wireguard/ratelimiter.c
70
hlist_for_each_entry_safe(entry, temp, &table_v6[i], hash) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
107
struct brcmf_flowring_hash *hash;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
130
hash = flow->hash;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
132
if ((hash[hash_idx].ifidx == BRCMF_FLOWRING_INVALID_IFIDX) &&
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
133
(is_zero_ether_addr(hash[hash_idx].mac))) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
152
memcpy(hash[hash_idx].mac, mac, ETH_ALEN);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
153
hash[hash_idx].fifo = fifo;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
154
hash[hash_idx].ifidx = ifidx;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
155
hash[hash_idx].flowid = i;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
174
return flow->hash[ring->hash_id].fifo;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
245
flow->hash[hash_idx].ifidx = BRCMF_FLOWRING_INVALID_IFIDX;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
246
eth_zero_addr(flow->hash[hash_idx].mac);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
354
return flow->hash[hash_idx].ifidx;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
370
for (i = 0; i < ARRAY_SIZE(flow->hash); i++)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
371
flow->hash[i].ifidx = BRCMF_FLOWRING_INVALID_IFIDX;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
416
for (i = 0; i < ARRAY_SIZE(flow->hash); i++) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
417
if (flow->hash[i].ifidx == ifidx) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
418
flowid = flow->hash[i].flowid;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
434
struct brcmf_flowring_hash *hash;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
454
hash = flow->hash;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
456
if ((sta || (memcmp(hash[i].mac, peer, ETH_ALEN) == 0)) &&
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
457
(hash[i].ifidx == ifidx)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
458
flowid = flow->hash[i].flowid;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
62
struct brcmf_flowring_hash *hash;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
85
hash = flow->hash;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
87
if ((sta || (memcmp(hash[hash_idx].mac, mac, ETH_ALEN) == 0)) &&
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
88
(hash[hash_idx].fifo == fifo) &&
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
89
(hash[hash_idx].ifidx == ifidx)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
97
return hash[hash_idx].flowid;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.h
40
struct brcmf_flowring_hash hash[BRCMF_FLOWRING_HASHSIZE];
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
1477
struct brcmf_flowring_hash *hash;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
1505
hash = &msgbuf->flow->hash[ring->hash_id];
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
1510
hash->ifidx, hash->fifo, hash->mac);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c
2569
u16 hash;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c
2664
hash = (target_gains->gm_gain << 8) |
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c
2670
if (hash == tbl_iqcal_gainparams_lcnphy[0][j][0]) {
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1785
u32 hash = mt7996_msdu_page_hash_from_addr(dma_addr);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1790
list_for_each_entry_safe(p, tmp, &dev->wed_rro.page_map[hash],
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1852
u32 hash;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1863
hash = mt7996_msdu_page_hash_from_addr(dma_addr);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1866
list_add_tail(&p->list, &dev->wed_rro.page_map[hash]);
drivers/net/wireless/mediatek/mt76/npu.c
120
u16 reason, hash;
drivers/net/wireless/mediatek/mt76/npu.c
131
hash = FIELD_GET(NPU_RX_DMA_FOE_ID_MASK, info);
drivers/net/wireless/mediatek/mt76/npu.c
132
skb_set_hash(skb, hash, PKT_HASH_TYPE_L4);
drivers/net/wireless/mediatek/mt76/npu.c
137
airoha_ppe_dev_check_skb(ppe_dev, skb, hash, true);
drivers/net/wireless/purelifi/plfxlc/chip.h
64
static inline void plfxlc_mc_add_all(struct plfxlc_mc_hash *hash)
drivers/net/wireless/purelifi/plfxlc/chip.h
66
hash->low = 0xffffffff;
drivers/net/wireless/purelifi/plfxlc/chip.h
67
hash->high = 0xffffffff;
drivers/net/wireless/purelifi/plfxlc/mac.c
542
struct plfxlc_mc_hash hash = {
drivers/net/wireless/purelifi/plfxlc/mac.c
560
plfxlc_mc_add_all(&hash);
drivers/net/wireless/purelifi/plfxlc/mac.c
565
mac->multicast_hash = hash;
drivers/net/wireless/realtek/rtw89/cam.c
805
u8 hash = 0;
drivers/net/wireless/realtek/rtw89/cam.c
809
hash ^= addr[i];
drivers/net/wireless/realtek/rtw89/cam.c
811
return hash;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
1518
struct zd_mc_hash *hash)
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
1521
{ CR_GROUP_HASH_P1, hash->low },
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
1522
{ CR_GROUP_HASH_P2, hash->high },
drivers/net/wireless/zydas/zd1211rw/zd_chip.h
942
static inline void zd_mc_clear(struct zd_mc_hash *hash)
drivers/net/wireless/zydas/zd1211rw/zd_chip.h
944
hash->low = 0;
drivers/net/wireless/zydas/zd1211rw/zd_chip.h
948
hash->high = 0x80000000;
drivers/net/wireless/zydas/zd1211rw/zd_chip.h
951
static inline void zd_mc_add_all(struct zd_mc_hash *hash)
drivers/net/wireless/zydas/zd1211rw/zd_chip.h
953
hash->low = hash->high = 0xffffffff;
drivers/net/wireless/zydas/zd1211rw/zd_chip.h
956
static inline void zd_mc_add_addr(struct zd_mc_hash *hash, u8 *addr)
drivers/net/wireless/zydas/zd1211rw/zd_chip.h
960
hash->low |= 1 << i;
drivers/net/wireless/zydas/zd1211rw/zd_chip.h
962
hash->high |= 1 << (i-32);
drivers/net/wireless/zydas/zd1211rw/zd_chip.h
967
struct zd_mc_hash *hash);
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
1207
struct zd_mc_hash hash;
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
1210
zd_mc_clear(&hash);
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
1214
zd_mc_add_addr(&hash, ha->addr);
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
1217
return hash.low | ((u64)hash.high << 32);
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
1228
struct zd_mc_hash hash = {
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
1248
zd_mc_add_all(&hash);
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
1253
mac->multicast_hash = hash;
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
1256
zd_chip_set_multicast_hash(&mac->chip, &hash);
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
264
struct zd_mc_hash hash;
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
265
zd_mc_clear(&hash);
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
266
return zd_chip_set_multicast_hash(&mac->chip, &hash);
drivers/net/xen-netback/common.h
312
struct xenvif_hash hash;
drivers/net/xen-netback/hash.c
104
spin_lock_irqsave(&vif->hash.cache.lock, flags);
drivers/net/xen-netback/hash.c
106
list_for_each_entry_safe(entry, n, &vif->hash.cache.list, link) {
drivers/net/xen-netback/hash.c
108
vif->hash.cache.count--;
drivers/net/xen-netback/hash.c
112
spin_unlock_irqrestore(&vif->hash.cache.lock, flags);
drivers/net/xen-netback/hash.c
132
list_for_each_entry_rcu(entry, &vif->hash.cache.list, link) {
drivers/net/xen-netback/hash.c
136
entry->seq = atomic_inc_return(&vif->hash.cache.seq);
drivers/net/xen-netback/hash.c
153
u32 hash = 0;
drivers/net/xen-netback/hash.c
155
u32 flags = vif->hash.flags;
drivers/net/xen-netback/hash.c
199
hash = xenvif_find_hash(vif, data, sizeof(data));
drivers/net/xen-netback/hash.c
207
hash = xenvif_find_hash(vif, data, sizeof(data));
drivers/net/xen-netback/hash.c
223
hash = xenvif_find_hash(vif, data, sizeof(data));
drivers/net/xen-netback/hash.c
231
hash = xenvif_find_hash(vif, data, sizeof(data));
drivers/net/xen-netback/hash.c
242
__skb_set_sw_hash(skb, hash, type == PKT_HASH_TYPE_L4);
drivers/net/xen-netback/hash.c
256
vif->hash.alg = alg;
drivers/net/xen-netback/hash.c
263
if (vif->hash.alg == XEN_NETIF_CTRL_HASH_ALGORITHM_NONE)
drivers/net/xen-netback/hash.c
282
if (vif->hash.alg == XEN_NETIF_CTRL_HASH_ALGORITHM_NONE)
drivers/net/xen-netback/hash.c
285
vif->hash.flags = flags;
drivers/net/xen-netback/hash.c
292
u8 *key = vif->hash.key;
drivers/net/xen-netback/hash.c
327
vif->hash.size = size;
drivers/net/xen-netback/hash.c
328
memset(vif->hash.mapping[vif->hash.mapping_sel], 0,
drivers/net/xen-netback/hash.c
337
u32 *mapping = vif->hash.mapping[!vif->hash.mapping_sel];
drivers/net/xen-netback/hash.c
347
if ((off + len < off) || (off + len > vif->hash.size) ||
drivers/net/xen-netback/hash.c
363
memcpy(mapping, vif->hash.mapping[vif->hash.mapping_sel],
drivers/net/xen-netback/hash.c
364
vif->hash.size * sizeof(*mapping));
drivers/net/xen-netback/hash.c
378
vif->hash.mapping_sel = !vif->hash.mapping_sel;
drivers/net/xen-netback/hash.c
388
switch (vif->hash.alg) {
drivers/net/xen-netback/hash.c
400
if (vif->hash.flags) {
drivers/net/xen-netback/hash.c
403
if (vif->hash.flags & XEN_NETIF_CTRL_HASH_TYPE_IPV4)
drivers/net/xen-netback/hash.c
405
if (vif->hash.flags & XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP)
drivers/net/xen-netback/hash.c
407
if (vif->hash.flags & XEN_NETIF_CTRL_HASH_TYPE_IPV6)
drivers/net/xen-netback/hash.c
409
if (vif->hash.flags & XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP)
drivers/net/xen-netback/hash.c
425
seq_printf(m, "%02x ", vif->hash.key[i]);
drivers/net/xen-netback/hash.c
430
if (vif->hash.size != 0) {
drivers/net/xen-netback/hash.c
431
const u32 *mapping = vif->hash.mapping[vif->hash.mapping_sel];
drivers/net/xen-netback/hash.c
435
for (i = 0; i < vif->hash.size; ) {
drivers/net/xen-netback/hash.c
439
if (i + n >= vif->hash.size)
drivers/net/xen-netback/hash.c
440
n = vif->hash.size - i;
drivers/net/xen-netback/hash.c
458
BUG_ON(vif->hash.cache.count);
drivers/net/xen-netback/hash.c
460
spin_lock_init(&vif->hash.cache.lock);
drivers/net/xen-netback/hash.c
461
INIT_LIST_HEAD(&vif->hash.cache.list);
drivers/net/xen-netback/hash.c
50
spin_lock_irqsave(&vif->hash.cache.lock, flags);
drivers/net/xen-netback/hash.c
54
list_for_each_entry_rcu(entry, &vif->hash.cache.list, link,
drivers/net/xen-netback/hash.c
55
lockdep_is_held(&vif->hash.cache.lock)) {
drivers/net/xen-netback/hash.c
65
new->seq = atomic_inc_return(&vif->hash.cache.seq);
drivers/net/xen-netback/hash.c
66
list_add_rcu(&new->link, &vif->hash.cache.list);
drivers/net/xen-netback/hash.c
68
if (++vif->hash.cache.count > xenvif_hash_cache_size) {
drivers/net/xen-netback/hash.c
70
vif->hash.cache.count--;
drivers/net/xen-netback/hash.c
75
spin_unlock_irqrestore(&vif->hash.cache.lock, flags);
drivers/net/xen-netback/hash.c
86
val = xen_netif_toeplitz_hash(vif->hash.key,
drivers/net/xen-netback/hash.c
87
sizeof(vif->hash.key),
drivers/net/xen-netback/interface.c
182
unsigned int size = vif->hash.size;
drivers/net/xen-netback/interface.c
191
if (vif->hash.alg == XEN_NETIF_CTRL_HASH_ALGORITHM_NONE)
drivers/net/xen-netback/interface.c
200
return vif->hash.mapping[vif->hash.mapping_sel]
drivers/net/xen-netback/interface.c
252
if (vif->hash.alg == XEN_NETIF_CTRL_HASH_ALGORITHM_NONE)
drivers/net/xen-netback/netback.c
1076
switch (extra->u.hash.type) {
drivers/net/xen-netback/netback.c
1093
*(u32 *)extra->u.hash.value,
drivers/net/xen-netback/rx.c
305
extra->u.hash.algorithm =
drivers/net/xen-netback/rx.c
309
extra->u.hash.type =
drivers/net/xen-netback/rx.c
314
extra->u.hash.type =
drivers/net/xen-netback/rx.c
319
*(uint32_t *)extra->u.hash.value = skb_get_hash_raw(skb);
drivers/net/xen-netfront.c
590
u32 hash;
drivers/net/xen-netfront.c
597
hash = skb_get_hash(skb);
drivers/net/xen-netfront.c
598
queue_idx = hash % num_queues;
drivers/nvme/common/auth.c
211
key->hash = key_hash;
drivers/nvme/common/auth.c
219
struct nvme_dhchap_key *nvme_auth_alloc_key(u32 len, u8 hash)
drivers/nvme/common/auth.c
226
key->hash = hash;
drivers/nvme/common/auth.c
253
if (key->hash == 0) {
drivers/nvme/common/auth.c
260
hmac_name = nvme_auth_hmac_name(key->hash);
drivers/nvme/common/auth.c
262
pr_warn("Invalid key hash id %d\n", key->hash);
drivers/nvme/common/auth.c
271
transformed_key = nvme_auth_alloc_key(key_len, key->hash);
drivers/nvme/target/auth.c
203
ctrl->host_key->hash > 0 ?
drivers/nvme/target/auth.c
204
nvme_auth_hmac_name(ctrl->host_key->hash) : "none",
drivers/nvme/target/auth.c
221
ctrl->ctrl_key->hash > 0 ?
drivers/nvme/target/auth.c
222
nvme_auth_hmac_name(ctrl->ctrl_key->hash) : "none",
drivers/nvme/target/nvmet.h
898
int nvmet_auth_set_host_hash(struct nvmet_host *host, const char *hash);
drivers/pci/controller/dwc/pcie-qcom.c
1119
u8 hash;
drivers/pci/controller/dwc/pcie-qcom.c
1121
hash = crc8(qcom_pcie_crc8_table, (u8 *)&bdf_be, sizeof(bdf_be), 0);
drivers/pci/controller/dwc/pcie-qcom.c
1123
val = readl(bdf_to_sid_base + hash * sizeof(u32));
drivers/pci/controller/dwc/pcie-qcom.c
1127
u8 current_hash = hash++;
drivers/pci/controller/dwc/pcie-qcom.c
1132
val |= (u32)hash;
drivers/pci/controller/dwc/pcie-qcom.c
1136
val = readl(bdf_to_sid_base + hash * sizeof(u32));
drivers/pci/controller/dwc/pcie-qcom.c
1141
writel(val, bdf_to_sid_base + hash * sizeof(u32));
drivers/pcmcia/ds.c
1078
u32 hash[4] = { 0, 0, 0, 0};
drivers/pcmcia/ds.c
1084
hash[i] = crc32(0, p_dev->prod_id[i],
drivers/pcmcia/ds.c
1092
hash[0], hash[1], hash[2], hash[3]);
drivers/pcmcia/ds.c
47
u32 hash;
drivers/pcmcia/ds.c
58
hash = crc32(0, did->prod_id[i], strlen(did->prod_id[i]));
drivers/pcmcia/ds.c
59
if (hash == did->prod_id_hash[i])
drivers/pcmcia/ds.c
65
did->prod_id_hash[i], hash);
drivers/pcmcia/ds.c
938
u32 hash[4] = { 0, 0, 0, 0};
drivers/pcmcia/ds.c
949
hash[i] = crc32(0, p_dev->prod_id[i], strlen(p_dev->prod_id[i]));
drivers/pcmcia/ds.c
965
hash[0],
drivers/pcmcia/ds.c
966
hash[1],
drivers/pcmcia/ds.c
967
hash[2],
drivers/pcmcia/ds.c
968
hash[3]))
drivers/ptp/ptp_vclock.c
277
unsigned int hash = vclock_index % HASH_SIZE(vclock_hash);
drivers/ptp/ptp_vclock.c
286
hlist_for_each_entry_rcu(vclock, &vclock_hash[hash], vclock_hash_node) {
drivers/s390/block/dasd_devmap.c
101
hash = 0;
drivers/s390/block/dasd_devmap.c
103
hash += *bus_id;
drivers/s390/block/dasd_devmap.c
104
return hash & 0xff;
drivers/s390/block/dasd_devmap.c
413
int hash;
drivers/s390/block/dasd_devmap.c
420
hash = dasd_hash_busid(bus_id);
drivers/s390/block/dasd_devmap.c
421
list_for_each_entry(tmp, &dasd_hashlists[hash], list)
drivers/s390/block/dasd_devmap.c
432
list_add(&new->list, &dasd_hashlists[hash]);
drivers/s390/block/dasd_devmap.c
445
int hash;
drivers/s390/block/dasd_devmap.c
448
hash = dasd_hash_busid(bus_id);
drivers/s390/block/dasd_devmap.c
449
list_for_each_entry(tmp, &dasd_hashlists[hash], list) {
drivers/s390/block/dasd_devmap.c
99
int hash, i;
drivers/scsi/elx/efct/efct_hw.c
2151
efct_hw_queue_hash_find(struct efct_queue_hash *hash, u16 id)
drivers/scsi/elx/efct/efct_hw.c
2162
if (hash[i].in_use && hash[i].id == id)
drivers/scsi/elx/efct/efct_hw.c
2163
index = hash[i].index;
drivers/scsi/elx/efct/efct_hw.c
2166
} while (index == -1 && hash[i].in_use);
drivers/scsi/elx/efct/efct_hw.c
832
efct_hw_queue_hash_add(struct efct_queue_hash *hash,
drivers/scsi/elx/efct/efct_hw.c
841
while (hash[hash_index].in_use)
drivers/scsi/elx/efct/efct_hw.c
845
hash[hash_index].id = id;
drivers/scsi/elx/efct/efct_hw.c
846
hash[hash_index].in_use = true;
drivers/scsi/elx/efct/efct_hw.c
847
hash[hash_index].index = index;
drivers/scsi/elx/efct/efct_hw.h
692
efct_hw_queue_hash_find(struct efct_queue_hash *hash, u16 id);
drivers/scsi/lpfc/lpfc_crtn.h
683
uint32_t hash, uint8_t *buf);
drivers/scsi/lpfc/lpfc_ct.c
3591
u32 cmd, hash, bucket;
drivers/scsi/lpfc/lpfc_ct.c
3634
hash = lpfc_vmid_hash_fn(app->obj.entity_id,
drivers/scsi/lpfc/lpfc_ct.c
3636
vmp = lpfc_get_vmid_from_hashtable(vport, hash,
drivers/scsi/lpfc/lpfc_ct.c
3649
app->obj.entity_id, hash);
drivers/scsi/lpfc/lpfc_vmid.c
132
u32 hash;
drivers/scsi/lpfc/lpfc_vmid.c
138
hash = lpfc_vmid_hash_fn(vmid->host_vmid, vmid->vmid_len);
drivers/scsi/lpfc/lpfc_vmid.c
140
lpfc_get_vmid_from_hashtable(vport->phba->pport, hash,
drivers/scsi/lpfc/lpfc_vmid.c
163
int hash, len, rc = -EPERM, i;
drivers/scsi/lpfc/lpfc_vmid.c
175
hash = lpfc_vmid_hash_fn(uuid, len);
drivers/scsi/lpfc/lpfc_vmid.c
179
vmp = lpfc_get_vmid_from_hashtable(vport, hash, uuid);
drivers/scsi/lpfc/lpfc_vmid.c
199
vmp = lpfc_get_vmid_from_hashtable(vport, hash, uuid);
drivers/scsi/lpfc/lpfc_vmid.c
231
lpfc_put_vmid_in_hashtable(vport, hash, vmp);
drivers/scsi/lpfc/lpfc_vmid.c
48
u32 hash, u8 *buf)
drivers/scsi/lpfc/lpfc_vmid.c
52
hash_for_each_possible(vport->hash_table, vmp, hnode, hash) {
drivers/scsi/lpfc/lpfc_vmid.c
69
lpfc_put_vmid_in_hashtable(struct lpfc_vport *vport, u32 hash,
drivers/scsi/lpfc/lpfc_vmid.c
72
hash_add(vport->hash_table, &vmp->hnode, hash);
drivers/scsi/lpfc/lpfc_vmid.c
84
int hash = 0;
drivers/scsi/lpfc/lpfc_vmid.c
93
hash = (hash + (c << LPFC_VMID_HASH_SHIFT) +
drivers/scsi/lpfc/lpfc_vmid.c
97
return hash & LPFC_VMID_HASH_MASK;
drivers/scsi/snic/snic_scsi.c
100
u32 hash = snic_cmd_tag(sc) & (SNIC_IO_LOCKS - 1);
drivers/scsi/snic/snic_scsi.c
102
return &snic->io_req_lock[hash];
drivers/staging/media/ipu7/ipu7-cpd.c
71
u8 hash[48];
drivers/target/tcm_fc/tcm_fc.h
35
struct hlist_node hash; /* linkage in ft_sess_hash table */
drivers/target/tcm_fc/tcm_fc.h
59
struct hlist_head hash[FT_SESS_HASH_SIZE]; /* list of sessions */
drivers/target/tcm_fc/tfc_sess.c
172
head = &tport->hash[ft_sess_hash(port_id)];
drivers/target/tcm_fc/tfc_sess.c
173
hlist_for_each_entry_rcu(sess, head, hash) {
drivers/target/tcm_fc/tfc_sess.c
194
struct hlist_head *head = &tport->hash[ft_sess_hash(sess->port_id)];
drivers/target/tcm_fc/tfc_sess.c
197
hlist_add_head_rcu(&sess->hash, head);
drivers/target/tcm_fc/tfc_sess.c
217
head = &tport->hash[ft_sess_hash(port_id)];
drivers/target/tcm_fc/tfc_sess.c
218
hlist_for_each_entry_rcu(sess, head, hash)
drivers/target/tcm_fc/tfc_sess.c
250
hlist_del_rcu(&sess->hash);
drivers/target/tcm_fc/tfc_sess.c
266
head = &tport->hash[ft_sess_hash(port_id)];
drivers/target/tcm_fc/tfc_sess.c
267
hlist_for_each_entry_rcu(sess, head, hash) {
drivers/target/tcm_fc/tfc_sess.c
292
for (head = tport->hash;
drivers/target/tcm_fc/tfc_sess.c
293
head < &tport->hash[FT_SESS_HASH_SIZE]; head++) {
drivers/target/tcm_fc/tfc_sess.c
294
hlist_for_each_entry_rcu(sess, head, hash) {
drivers/target/tcm_fc/tfc_sess.c
70
INIT_HLIST_HEAD(&tport->hash[i]);
drivers/tee/tee_core.c
152
unsigned char hash[SHA1_DIGEST_SIZE];
drivers/tee/tee_core.c
158
sha1_final(&ctx, hash);
drivers/tee/tee_core.c
160
memcpy(uuid->b, hash, UUID_SIZE);
drivers/tee/tee_core.c
163
uuid->b[6] = (hash[6] & 0x0F) | 0x50;
drivers/tee/tee_core.c
164
uuid->b[8] = (hash[8] & 0x3F) | 0x80;
drivers/usb/host/ohci-q.c
523
int hash;
drivers/usb/host/ohci-q.c
575
hash = TD_HASH_FUNC (td->td_dma);
drivers/usb/host/ohci-q.c
576
td->td_hash = ohci->td_hash [hash];
drivers/usb/host/ohci-q.c
577
ohci->td_hash [hash] = td;
drivers/vhost/vsock.c
55
struct hlist_node hash;
drivers/vhost/vsock.c
759
hash_del_rcu(&vsock->hash);
drivers/vhost/vsock.c
78
hash_for_each_possible_rcu(vhost_vsock_hash, vsock, hash, guest_cid,
drivers/vhost/vsock.c
816
hash_del_rcu(&vsock->hash);
drivers/vhost/vsock.c
819
hash_add_rcu(vhost_vsock_hash, &vsock->hash, vsock->guest_cid);
fs/adfs/dir.c
399
unsigned long hash;
fs/adfs/dir.c
407
hash = init_name_hash(parent);
fs/adfs/dir.c
409
hash = partial_name_hash(adfs_tolower(*name++), hash);
fs/adfs/dir.c
410
qstr->hash = end_name_hash(hash);
fs/affs/namei.c
160
u32 hash;
fs/affs/namei.c
162
hash = len = min(len, AFFSNAMEMAX);
fs/affs/namei.c
164
hash = (hash * 13 + fn(*name++)) & 0x7ff;
fs/affs/namei.c
166
return hash % AFFS_SB(sb)->s_hashsize;
fs/affs/namei.c
49
unsigned long hash;
fs/affs/namei.c
57
hash = init_name_hash(dentry);
fs/affs/namei.c
60
hash = partial_name_hash(fn(*name), hash);
fs/affs/namei.c
61
qstr->hash = end_name_hash(hash);
fs/afs/dir_search.c
24
unsigned int hash = 0, i;
fs/afs/dir_search.c
28
hash = (hash * 173) + p[i];
fs/afs/dir_search.c
29
bucket = hash & (AFS_DIR_HASHTBL_SIZE - 1);
fs/afs/dir_search.c
30
if (hash > INT_MAX) {
fs/autofs/root.c
137
unsigned int hash = name->hash;
fs/autofs/root.c
161
if (active->d_name.hash != hash)
fs/autofs/root.c
192
unsigned int hash = name->hash;
fs/autofs/root.c
221
if (expiring->d_name.hash != hash)
fs/autofs/waitq.c
185
if (wq->name.hash == qstr->hash &&
fs/autofs/waitq.c
362
qstr.hash = full_name_hash(dentry, qstr.name, qstr.len);
fs/btrfs/disk-io.c
1870
unsigned long hash = btrfs_inode_hash(BTRFS_BTREE_INODE_OBJECTID,
fs/btrfs/disk-io.c
1895
__insert_inode_hash(inode, hash);
fs/btrfs/props.c
35
static const struct hlist_head *find_prop_handlers_by_hash(const u64 hash)
fs/btrfs/props.c
39
h = &prop_handlers_ht[hash_min(hash, BTRFS_PROP_HANDLERS_HT_BITS)];
fs/btrfs/props.c
53
u64 hash = btrfs_name_hash(name, strlen(name));
fs/btrfs/props.c
55
handlers = find_prop_handlers_by_hash(hash);
fs/ceph/crypto.c
256
u8 hash[SHA256_DIGEST_SIZE];
fs/ceph/crypto.c
263
sha256(extra, len - CEPH_NOHASH_NAME_MAX, hash);
fs/ceph/crypto.c
264
memcpy(extra, hash, SHA256_DIGEST_SIZE);
fs/ceph/dir.c
1419
dentry->d_name.hash);
fs/ceph/dir.c
2202
unsigned hash;
fs/ceph/dir.c
2207
return dn->d_name.hash;
fs/ceph/dir.c
2211
hash = ceph_str_hash(dci->i_dir_layout.dl_dir_hash,
fs/ceph/dir.c
2214
return hash;
fs/ceph/export.c
263
req->r_args.lookupino.hash =
fs/ceph/export.c
264
cpu_to_le32(sfh->hash);
fs/ceph/export.c
292
vino.snap, sfh->parent_ino, sfh->hash, err);
fs/ceph/export.c
33
u32 hash;
fs/ceph/export.c
71
sfh->hash = ceph_dentry_hash(dir, dentry);
fs/ceph/export.c
82
sfh->hash = 0;
fs/ceph/inode.c
1628
dname.hash = full_name_hash(parent, dname.name, dname.len);
fs/ceph/inode.c
2031
dname.hash = full_name_hash(parent, dname.name, dname.len);
fs/ceph/inode.c
2037
u32 hash = ceph_frag_value(rde->raw_hash);
fs/ceph/inode.c
2038
if (hash != last_hash)
fs/ceph/inode.c
2040
last_hash = hash;
fs/ceph/inode.c
2041
rde->offset = ceph_make_fpos(hash, fpos_offset++, true);
fs/ceph/mds_client.c
1293
u32 hash = req->r_direct_hash;
fs/ceph/mds_client.c
1353
hash = ceph_dentry_hash(dir, req->r_dentry);
fs/ceph/mds_client.c
1366
ceph_vinop(inode), (int)is_hash, hash, mode);
fs/ceph/mds_client.c
1373
ceph_choose_frag(ci, hash, &frag, &found);
fs/ceph/mds_client.c
5292
dname.hash = full_name_hash(parent, dname.name, dname.len);
fs/ceph/mds_client.c
845
u32 hash = dentry->d_name.hash;
fs/ceph/mds_client.c
853
hnode, hash) {
fs/ceph/mds_client.c
857
if (udentry->d_name.hash != hash)
fs/coda/cnode.c
127
unsigned long hash = coda_f2i(newfid);
fs/coda/cnode.c
135
inode->i_ino = hash;
fs/coda/cnode.c
136
__insert_inode_hash(inode, hash);
fs/coda/cnode.c
143
unsigned long hash = coda_f2i(fid);
fs/coda/cnode.c
145
inode = ilookup5(sb, hash, coda_test_inode, fid);
fs/coda/cnode.c
65
unsigned long hash = coda_f2i(fid);
fs/coda/cnode.c
69
inode = iget5_locked(sb, hash, coda_test_inode, coda_set_inode, fid);
fs/coda/cnode.c
76
inode->i_ino = hash;
fs/crypto/fname.c
268
u32 hash, u32 minor_hash,
fs/crypto/fname.c
299
nokey_name.dirhash[0] = hash;
fs/crypto/fname.c
404
fname->hash = nokey_name->dirhash[0];
fs/dax.c
150
unsigned long hash;
fs/dax.c
163
hash = hash_long((unsigned long)xas->xa ^ index, DAX_WAIT_TABLE_BITS);
fs/dax.c
164
return wait_table + hash;
fs/dcache.c
127
unsigned int hash)
fs/dcache.c
129
hash += (unsigned long) parent / L1_CACHE_BYTES;
fs/dcache.c
130
return in_lookup_hashtable + hash_32(hash, IN_LOOKUP_SHIFT);
fs/dcache.c
1771
dentry->__d_name.hash = name->hash;
fs/dcache.c
2256
if (dentry->d_name.hash != hashlen_hash(hashlen))
fs/dcache.c
2419
unsigned int hash = name->hash;
fs/dcache.c
2420
struct hlist_bl_head *b = d_hash(hash);
fs/dcache.c
2449
if (dentry->d_name.hash != hash)
fs/dcache.c
2487
name->hash = full_name_hash(dir, name->name, name->len);
fs/dcache.c
2541
struct hlist_bl_head *b = d_hash(entry->d_name.hash);
fs/dcache.c
2601
unsigned int hash = name->hash;
fs/dcache.c
2602
struct hlist_bl_head *b = in_lookup_hash(parent, hash);
fs/dcache.c
2662
if (dentry->d_name.hash != hash)
fs/dcache.c
2688
if (unlikely(dentry->d_name.hash != hash))
fs/dcache.c
2725
b = in_lookup_hash(dentry->d_parent, dentry->d_name.hash);
fs/dcache.c
565
b = d_hash(dentry->d_name.hash);
fs/dlm/dir.c
33
int dlm_hash2nodeid(struct dlm_ls *ls, uint32_t hash)
fs/dlm/dir.c
40
node = (hash >> 16) % ls->ls_total_weight;
fs/dlm/dir.h
16
int dlm_hash2nodeid(struct dlm_ls *ls, uint32_t hash);
fs/dlm/lock.c
1016
r->res_hash = hash;
fs/dlm/lock.c
1087
uint32_t hash;
fs/dlm/lock.c
1093
hash = jhash(name, len, 0);
fs/dlm/lock.c
1094
dir_nodeid = dlm_hash2nodeid(ls, hash);
fs/dlm/lock.c
1098
rv = find_rsb_nodir(ls, name, len, hash, dir_nodeid,
fs/dlm/lock.c
1101
rv = find_rsb_dir(ls, name, len, hash, dir_nodeid,
fs/dlm/lock.c
1265
uint32_t hash;
fs/dlm/lock.c
1278
hash = jhash(name, len, 0);
fs/dlm/lock.c
1279
dir_nodeid = dlm_hash2nodeid(ls, hash);
fs/dlm/lock.c
1282
from_nodeid, dir_nodeid, our_nodeid, hash,
fs/dlm/lock.c
1363
r->res_hash = hash;
fs/dlm/lock.c
1405
static void dlm_dump_rsb_hash(struct dlm_ls *ls, uint32_t hash)
fs/dlm/lock.c
1411
if (r->res_hash == hash)
fs/dlm/lock.c
696
uint32_t hash, int dir_nodeid, int from_nodeid,
fs/dlm/lock.c
856
r->res_hash = hash;
fs/dlm/lock.c
920
uint32_t hash, int dir_nodeid, int from_nodeid,
fs/ecryptfs/keystore.c
603
char hash[MD5_DIGEST_SIZE];
fs/ecryptfs/keystore.c
743
s->hash);
fs/ecryptfs/keystore.c
746
s->hash[s->j % MD5_DIGEST_SIZE];
fs/ecryptfs/keystore.c
748
md5(s->hash, MD5_DIGEST_SIZE, s->hash);
fs/efivarfs/super.c
172
unsigned long hash = init_name_hash(dentry);
fs/efivarfs/super.c
177
hash = partial_name_hash(*s++, hash);
fs/efivarfs/super.c
181
hash = partial_name_hash(tolower(*s++), hash);
fs/efivarfs/super.c
183
qstr->hash = end_name_hash(hash);
fs/erofs/ishare.c
46
unsigned long hash;
fs/erofs/ishare.c
53
hash = xxh32(fp.opaque, fp.size, 0);
fs/erofs/ishare.c
54
sharedinode = iget5_locked(erofs_ishare_mnt->mnt_sb, hash,
fs/exfat/namei.c
142
unsigned long hash = init_name_hash(dentry);
fs/exfat/namei.c
154
hash = partial_name_hash(u <= 0xFFFF ? exfat_toupper(sb, u) : u,
fs/exfat/namei.c
155
hash);
fs/exfat/namei.c
158
qstr->hash = end_name_hash(hash);
fs/exfat/namei.c
86
unsigned long hash = init_name_hash(dentry);
fs/exfat/namei.c
94
hash = partial_name_hash(exfat_toupper(sb, c), hash);
fs/exfat/namei.c
97
qstr->hash = end_name_hash(hash);
fs/ext2/xattr.c
1006
hash = (hash << VALUE_HASH_SHIFT) ^
fs/ext2/xattr.c
1007
(hash >> (8*sizeof(hash) - VALUE_HASH_SHIFT)) ^
fs/ext2/xattr.c
1011
entry->e_hash = cpu_to_le32(hash);
fs/ext2/xattr.c
1028
__u32 hash = 0;
fs/ext2/xattr.c
1035
hash = 0;
fs/ext2/xattr.c
1038
hash = (hash << BLOCK_HASH_SHIFT) ^
fs/ext2/xattr.c
1039
(hash >> (8*sizeof(hash) - BLOCK_HASH_SHIFT)) ^
fs/ext2/xattr.c
1043
header->h_hash = cpu_to_le32(hash);
fs/ext2/xattr.c
523
__u32 hash = le32_to_cpu(header->h_hash);
fs/ext2/xattr.c
527
hash, bh->b_blocknr);
fs/ext2/xattr.c
661
__u32 hash = le32_to_cpu(HDR(bh)->h_hash);
fs/ext2/xattr.c
668
oe = mb_cache_entry_delete_or_get(ea_block_cache, hash,
fs/ext2/xattr.c
874
__u32 hash = le32_to_cpu(HDR(bh)->h_hash);
fs/ext2/xattr.c
877
error = mb_cache_entry_create(cache, GFP_KERNEL, hash, bh->b_blocknr,
fs/ext2/xattr.c
885
ea_bdebug(bh, "inserting [%x]", (int)hash);
fs/ext2/xattr.c
940
__u32 hash = le32_to_cpu(header->h_hash);
fs/ext2/xattr.c
946
ea_idebug(inode, "looking for cached blocks [%x]", (int)hash);
fs/ext2/xattr.c
948
ce = mb_cache_entry_find_first(ea_block_cache, hash);
fs/ext2/xattr.c
991
__u32 hash = 0;
fs/ext2/xattr.c
996
hash = (hash << NAME_HASH_SHIFT) ^
fs/ext2/xattr.c
997
(hash >> (8*sizeof(hash) - NAME_HASH_SHIFT)) ^
fs/ext4/crypto.c
17
dst->hinfo.hash = src->hash;
fs/ext4/dir.c
287
u32 hash;
fs/ext4/dir.c
291
hash = EXT4_DIRENT_HASH(de);
fs/ext4/dir.c
294
hash = 0;
fs/ext4/dir.c
300
hash, minor_hash, &de_name, &fstr);
fs/ext4/dir.c
418
__u32 hash;
fs/ext4/dir.c
470
int ext4_htree_store_dirent(struct file *dir_file, __u32 hash,
fs/ext4/dir.c
486
new_fn->hash = hash;
fs/ext4/dir.c
501
if ((new_fn->hash == fname->hash) &&
fs/ext4/dir.c
508
if (new_fn->hash < fname->hash)
fs/ext4/dir.c
510
else if (new_fn->hash > fname->hash)
fs/ext4/dir.c
543
ctx->pos = hash2pos(file, fname->hash, fname->minor_hash);
fs/ext4/dir.c
614
info->curr_hash = fname->hash;
fs/ext4/dir.c
623
info->curr_hash = fname->hash;
fs/ext4/ext4.h
2411
__le32 hash;
fs/ext4/ext4.h
2436
#define EXT4_DIRENT_HASH(entry) le32_to_cpu(EXT4_DIRENT_HASHES(entry)->hash)
fs/ext4/ext4.h
2560
u32 hash;
fs/ext4/ext4.h
2897
extern int ext4_htree_store_dirent(struct file *dir_file, __u32 hash,
fs/ext4/hash.c
105
__u32 hash, hash0 = 0x12a3fe2d, hash1 = 0x37abe8f9;
fs/ext4/hash.c
109
hash = hash1 + (hash0 ^ (((int) *ucp++) * 7152373));
fs/ext4/hash.c
111
if (hash & 0x80000000)
fs/ext4/hash.c
112
hash -= 0x7fffffff;
fs/ext4/hash.c
114
hash0 = hash;
fs/ext4/hash.c
121
__u32 hash, hash0 = 0x12a3fe2d, hash1 = 0x37abe8f9;
fs/ext4/hash.c
125
hash = hash1 + (hash0 ^ (((int) *scp++) * 7152373));
fs/ext4/hash.c
127
if (hash & 0x80000000)
fs/ext4/hash.c
128
hash -= 0x7fffffff;
fs/ext4/hash.c
130
hash0 = hash;
fs/ext4/hash.c
203
__u32 hash;
fs/ext4/hash.c
229
hash = dx_hack_hash_unsigned(name, len);
fs/ext4/hash.c
232
hash = dx_hack_hash_signed(name, len);
fs/ext4/hash.c
246
hash = buf[1];
fs/ext4/hash.c
259
hash = buf[0];
fs/ext4/hash.c
274
hash = (__u32)(combined_hash >> 32);
fs/ext4/hash.c
279
hinfo->hash = 0;
fs/ext4/hash.c
286
hash = hash & ~1;
fs/ext4/hash.c
287
if (hash == (EXT4_HTREE_EOF_32BIT << 1))
fs/ext4/hash.c
288
hash = (EXT4_HTREE_EOF_32BIT - 1) << 1;
fs/ext4/hash.c
289
hinfo->hash = hash;
fs/ext4/ialloc.c
465
parent_group = hinfo.hash % ngroups;
fs/ext4/inline.c
1339
hinfo->hash = EXT4_DIRENT_HASH(de);
fs/ext4/inline.c
1348
if ((hinfo->hash < start_hash) ||
fs/ext4/inline.c
1349
((hinfo->hash == start_hash) &&
fs/ext4/inline.c
1356
err = ext4_htree_store_dirent(dir_file, hinfo->hash,
fs/ext4/mballoc.c
2271
int hash = ac->ac_inode->i_ino % sbi->s_mb_nr_global_goals;
fs/ext4/mballoc.c
2273
WRITE_ONCE(sbi->s_mb_last_groups[hash], ac->ac_f_ex.fe_group);
fs/ext4/mballoc.c
3041
int hash = ac->ac_inode->i_ino % sbi->s_mb_nr_global_goals;
fs/ext4/mballoc.c
3043
ac->ac_g_ex.fe_group = READ_ONCE(sbi->s_mb_last_groups[hash]);
fs/ext4/namei.c
1012
if ((hash & 1) == 0) {
fs/ext4/namei.c
1013
if ((bhash & ~1) != hash)
fs/ext4/namei.c
1086
hinfo->hash = EXT4_DIRENT_HASH(de);
fs/ext4/namei.c
1089
hinfo->hash = 0;
fs/ext4/namei.c
1100
if ((hinfo->hash < start_hash) ||
fs/ext4/namei.c
1101
((hinfo->hash == start_hash) &&
fs/ext4/namei.c
1110
hinfo->hash, hinfo->minor_hash, de,
fs/ext4/namei.c
1118
err = fscrypt_fname_disk_to_usr(dir, hinfo->hash,
fs/ext4/namei.c
1126
hinfo->hash, hinfo->minor_hash, de,
fs/ext4/namei.c
1193
hinfo.hash = start_hash;
fs/ext4/namei.c
1300
h.hash = EXT4_DIRENT_HASH(de);
fs/ext4/namei.c
1308
map_tail->hash = h.hash;
fs/ext4/namei.c
1331
if (p->hash < q->hash)
fs/ext4/namei.c
1339
if (q[1].hash >= q[0].hash)
fs/ext4/namei.c
1347
static void dx_insert_block(struct dx_frame *frame, u32 hash, ext4_lblk_t block)
fs/ext4/namei.c
1356
dx_set_hash(new, hash);
fs/ext4/namei.c
1435
(fname->hinfo.hash != EXT4_DIRENT_HASH(de) ||
fs/ext4/namei.c
1741
retval = ext4_htree_next_block(dir, fname->hinfo.hash, frame,
fs/ext4/namei.c
1979
hinfo->hash, hinfo->minor_hash, count, move);
fs/ext4/namei.c
1984
hash2 = map[split].hash;
fs/ext4/namei.c
1985
continued = hash2 == map[split - 1].hash;
fs/ext4/namei.c
2011
if (hinfo->hash >= hash2) {
fs/ext4/namei.c
2094
EXT4_DIRENT_HASHES(de)->hash = cpu_to_le32(hinfo->hash);
fs/ext4/namei.c
237
__le32 hash;
fs/ext4/namei.c
281
u32 hash;
fs/ext4/namei.c
549
return le32_to_cpu(entry->hash);
fs/ext4/namei.c
554
entry->hash = cpu_to_le32(value);
fs/ext4/namei.c
651
name, h.hash,
fs/ext4/namei.c
681
h.hash = EXT4_DIRENT_HASH(de);
fs/ext4/namei.c
687
h.hash, (unsigned) ((char *) de
fs/ext4/namei.c
697
printk("%*.s:%x.%u ", len, name, h.hash,
fs/ext4/namei.c
721
ext4_lblk_t hash = i ? dx_get_hash(entries): 0;
fs/ext4/namei.c
722
u32 range = i < count - 1? (dx_get_hash(entries + 1) - hash): ~hash;
fs/ext4/namei.c
724
printk("%s%3u:%03u hash %8x/%8x ",levels?"":" ", i, block, hash, range);
fs/ext4/namei.c
749
u32 hash, unsigned int n)
fs/ext4/namei.c
753
if (dx_get_hash(++at) > hash) {
fs/ext4/namei.c
763
u32 hash, unsigned int n)
fs/ext4/namei.c
786
u32 hash;
fs/ext4/namei.c
833
hash = hinfo->hash;
fs/ext4/namei.c
865
dxtrace(printk("Look up %x", hash));
fs/ext4/namei.c
882
if (dx_get_hash(m) > hash)
fs/ext4/namei.c
888
htree_rep_invariant_check(entries, p, hash, count - 1);
fs/ext4/namei.c
975
static int ext4_htree_next_block(struct inode *dir, __u32 hash,
fs/ext4/xattr.c
1276
u32 hash, ref;
fs/ext4/xattr.c
1287
hash = le32_to_cpu(BHDR(bh)->h_hash);
fs/ext4/xattr.c
1298
oe = mb_cache_entry_delete_or_get(ea_block_cache, hash,
fs/ext4/xattr.c
1327
ce = mb_cache_entry_get(ea_block_cache, hash,
fs/ext4/xattr.c
1471
struct inode *inode, u32 hash)
fs/ext4/xattr.c
1498
ext4_xattr_inode_set_hash(ea_inode, hash);
fs/ext4/xattr.c
1526
size_t value_len, u32 hash)
fs/ext4/xattr.c
1536
ce = mb_cache_entry_find_first(ea_inode_cache, hash);
fs/ext4/xattr.c
1580
u32 hash;
fs/ext4/xattr.c
1588
hash = ext4_xattr_inode_hash(EXT4_SB(inode->i_sb), value, value_len);
fs/ext4/xattr.c
1589
ea_inode = ext4_xattr_inode_cache_find(inode, value, value_len, hash);
fs/ext4/xattr.c
1598
ea_inode = ext4_xattr_inode_create(handle, inode, hash);
fs/ext4/xattr.c
1612
mb_cache_entry_create(EA_INODE_CACHE(inode), GFP_NOFS, hash,
fs/ext4/xattr.c
1817
__le32 hash = 0;
fs/ext4/xattr.c
1830
hash = ext4_xattr_hash_entry(here->e_name,
fs/ext4/xattr.c
1837
hash = ext4_xattr_hash_entry(here->e_name,
fs/ext4/xattr.c
1841
here->e_hash = hash;
fs/ext4/xattr.c
1939
__u32 hash = le32_to_cpu(BHDR(bs->bh)->h_hash);
fs/ext4/xattr.c
1950
hash, bs->bh->b_blocknr);
fs/ext4/xattr.c
3037
__u32 hash = le32_to_cpu(header->h_hash);
fs/ext4/xattr.c
3044
error = mb_cache_entry_create(ea_block_cache, GFP_NOFS, hash,
fs/ext4/xattr.c
3050
ea_bdebug(bh, "inserting [%x]", (int)hash);
fs/ext4/xattr.c
3105
__u32 hash = le32_to_cpu(header->h_hash);
fs/ext4/xattr.c
3113
ea_idebug(inode, "looking for cached blocks [%x]", (int)hash);
fs/ext4/xattr.c
3114
ce = mb_cache_entry_find_first(ea_block_cache, hash);
fs/ext4/xattr.c
3146
__u32 hash = 0;
fs/ext4/xattr.c
3149
hash = (hash << NAME_HASH_SHIFT) ^
fs/ext4/xattr.c
3150
(hash >> (8*sizeof(hash) - NAME_HASH_SHIFT)) ^
fs/ext4/xattr.c
3154
hash = (hash << VALUE_HASH_SHIFT) ^
fs/ext4/xattr.c
3155
(hash >> (8*sizeof(hash) - VALUE_HASH_SHIFT)) ^
fs/ext4/xattr.c
3158
return cpu_to_le32(hash);
fs/ext4/xattr.c
3168
__u32 hash = 0;
fs/ext4/xattr.c
3171
hash = (hash << NAME_HASH_SHIFT) ^
fs/ext4/xattr.c
3172
(hash >> (8*sizeof(hash) - NAME_HASH_SHIFT)) ^
fs/ext4/xattr.c
3176
hash = (hash << VALUE_HASH_SHIFT) ^
fs/ext4/xattr.c
3177
(hash >> (8*sizeof(hash) - VALUE_HASH_SHIFT)) ^
fs/ext4/xattr.c
3180
return cpu_to_le32(hash);
fs/ext4/xattr.c
3196
__u32 hash = 0;
fs/ext4/xattr.c
3202
hash = 0;
fs/ext4/xattr.c
3205
hash = (hash << BLOCK_HASH_SHIFT) ^
fs/ext4/xattr.c
3206
(hash >> (8*sizeof(hash) - BLOCK_HASH_SHIFT)) ^
fs/ext4/xattr.c
3210
header->h_hash = cpu_to_le32(hash);
fs/ext4/xattr.c
375
static void ext4_xattr_inode_set_hash(struct inode *ea_inode, u32 hash)
fs/ext4/xattr.c
377
inode_set_atime(ea_inode, hash, 0);
fs/ext4/xattr.c
491
u32 hash;
fs/ext4/xattr.c
494
hash = ext4_xattr_inode_hash(EXT4_SB(ea_inode->i_sb), buffer, size);
fs/ext4/xattr.c
495
if (hash != ext4_xattr_inode_get_hash(ea_inode))
fs/ext4/xattr.c
502
tmp_data = cpu_to_le32(hash);
fs/f2fs/dir.c
119
fname->hash = cpu_to_le32(crypt_name->hash);
fs/f2fs/dir.c
252
if (!use_hash || de->hash_code == fname->hash) {
fs/f2fs/dir.c
293
bucket_no = use_hash ? le32_to_cpu(fname->hash) % nbucket : 0;
fs/f2fs/dir.c
336
if (room && F2FS_I(dir)->chash != fname->hash) {
fs/f2fs/dir.c
337
F2FS_I(dir)->chash = fname->hash;
fs/f2fs/dir.c
488
put_unaligned(fname->hash, (f2fs_hash_t *)
fs/f2fs/dir.c
695
if (F2FS_I(dir)->chash == fname->hash) {
fs/f2fs/dir.c
715
(le32_to_cpu(fname->hash) % nbucket));
fs/f2fs/dir.c
747
f2fs_update_dentry(ino, mode, &d, &fname->disk_name, fname->hash,
fs/f2fs/f2fs.h
627
f2fs_hash_t hash;
fs/f2fs/hash.c
104
fname->hash = 0;
fs/f2fs/hash.c
130
fname->hash =
fs/f2fs/hash.c
136
fname->hash = cpu_to_le32(TEA_hash_name(name, len));
fs/f2fs/inline.c
518
fname.hash = de->hash_code;
fs/f2fs/inline.c
677
f2fs_update_dentry(ino, mode, &d, &fname->disk_name, fname->hash,
fs/f2fs/recovery.c
145
fname->hash = get_unaligned((f2fs_hash_t *)
fs/fat/namei_msdos.c
159
qstr->hash = full_name_hash(dentry, msdos_name, MSDOS_NAME);
fs/fat/namei_vfat.c
122
qstr->hash = full_name_hash(dentry, qstr->name, vfat_striptail_len(qstr));
fs/fat/namei_vfat.c
137
unsigned long hash;
fs/fat/namei_vfat.c
142
hash = init_name_hash(dentry);
fs/fat/namei_vfat.c
144
hash = partial_name_hash(nls_tolower(t, *name++), hash);
fs/fat/namei_vfat.c
145
qstr->hash = end_name_hash(hash);
fs/fuse/dev.c
1412
unsigned int hash;
fs/fuse/dev.c
1515
hash = fuse_req_hash(req->in.h.unique);
fs/fuse/dev.c
1516
list_move_tail(&req->list, &fpq->processing[hash]);
fs/fuse/dev.c
2133
unsigned int hash = fuse_req_hash(unique);
fs/fuse/dev.c
2136
list_for_each_entry(req, &fpq->processing[hash], list) {
fs/fuse/dev_uring.c
762
unsigned int hash;
fs/fuse/dev_uring.c
765
hash = fuse_req_hash(req->in.h.unique);
fs/fuse/dev_uring.c
766
list_move_tail(&req->list, &fpq->processing[hash]);
fs/fuse/readdir.c
196
name.hash = full_name_hash(parent, name.name, name.len);
fs/fuse/virtio_fs.c
1385
unsigned int i, hash;
fs/fuse/virtio_fs.c
1445
hash = fuse_req_hash(req->in.h.unique);
fs/fuse/virtio_fs.c
1447
list_add_tail(&req->list, &fpq->processing[hash]);
fs/gfs2/dentry.c
77
str->hash = gfs2_disk_hash(str->name, str->len);
fs/gfs2/dir.c
1012
index = name->hash >> (32 - dip->i_depth);
fs/gfs2/dir.c
1092
str.hash = be32_to_cpu(dent->de_hash);
fs/gfs2/dir.c
1534
u32 hash, index;
fs/gfs2/dir.c
1541
hash = gfs2_dir_offset2hash(ctx->pos);
fs/gfs2/dir.c
1542
index = hash >> (32 - dip->i_depth);
fs/gfs2/dir.c
1734
index = name->hash >> (32 - ip->i_depth);
fs/gfs2/dir.c
406
be32_to_cpu(dent->de_hash) == name->hash &&
fs/gfs2/dir.c
784
__be64 *hash;
fs/gfs2/dir.c
787
hash = gfs2_dir_get_hash_table(dip);
fs/gfs2/dir.c
788
error = PTR_ERR_OR_ZERO(hash);
fs/gfs2/dir.c
791
*leaf_out = be64_to_cpu(*(hash + index));
fs/gfs2/dir.c
829
index = name->hash >> (32 - ip->i_depth);
fs/gfs2/dir.h
63
name->hash = gfs2_disk_hash(name->name, name->len);
fs/gfs2/dir.h
71
dent->de_hash = cpu_to_be32(name->hash);
fs/gfs2/glock.c
111
u32 hash = jhash2((u32 *)name, ht_parms.key_len / 4, 0);
fs/gfs2/glock.c
113
return glock_wait_table + hash_32(hash, GLOCK_WAIT_TABLE_BITS);
fs/gfs2/log.c
826
u32 hash, crc;
fs/gfs2/log.c
849
hash = ~crc32(~0, lh, LH_V1_SIZE);
fs/gfs2/log.c
850
lh->lh_hash = cpu_to_be32(hash);
fs/gfs2/ops_fstype.c
551
name.hash = gfs2_disk_hash(name.name, name.len);
fs/gfs2/quota.c
101
static inline void spin_unlock_bucket(unsigned int hash)
fs/gfs2/quota.c
103
hlist_bl_unlock(&qd_hash_table[hash]);
fs/gfs2/quota.c
1396
unsigned int hash;
fs/gfs2/quota.c
1448
hash = gfs2_qd_hash(sdp, qc_id);
fs/gfs2/quota.c
1449
qd = qd_alloc(hash, sdp, qc_id);
fs/gfs2/quota.c
1460
spin_lock_bucket(hash);
fs/gfs2/quota.c
1461
old_qd = gfs2_qd_search_bucket(hash, sdp, qc_id);
fs/gfs2/quota.c
1468
spin_unlock_bucket(hash);
fs/gfs2/quota.c
1486
hlist_bl_add_head_rcu(&qd->qd_hlist, &qd_hash_table[hash]);
fs/gfs2/quota.c
1487
spin_unlock_bucket(hash);
fs/gfs2/quota.c
229
static struct gfs2_quota_data *qd_alloc(unsigned hash, struct gfs2_sbd *sdp, struct kqid qid)
fs/gfs2/quota.c
243
qd->qd_hash = hash;
fs/gfs2/quota.c
257
static struct gfs2_quota_data *gfs2_qd_search_bucket(unsigned int hash,
fs/gfs2/quota.c
264
hlist_bl_for_each_entry_rcu(qd, h, &qd_hash_table[hash], qd_hlist) {
fs/gfs2/quota.c
283
unsigned int hash = gfs2_qd_hash(sdp, qid);
fs/gfs2/quota.c
286
*qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid);
fs/gfs2/quota.c
292
new_qd = qd_alloc(hash, sdp, qid);
fs/gfs2/quota.c
297
spin_lock_bucket(hash);
fs/gfs2/quota.c
298
*qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid);
fs/gfs2/quota.c
302
hlist_bl_add_head_rcu(&new_qd->qd_hlist, &qd_hash_table[hash]);
fs/gfs2/quota.c
305
spin_unlock_bucket(hash);
fs/gfs2/quota.c
96
static inline void spin_lock_bucket(unsigned int hash)
fs/gfs2/quota.c
98
hlist_bl_lock(&qd_hash_table[hash]);
fs/gfs2/recovery.c
122
u32 hash, crc;
fs/gfs2/recovery.c
129
hash = crc32(~0, lh, LH_V1_SIZE - 4);
fs/gfs2/recovery.c
130
hash = ~crc32(hash, &zero, 4); /* assume lh_hash is zero */
fs/gfs2/recovery.c
132
if (be32_to_cpu(lh->lh_hash) != hash)
fs/hfs/bnode.c
347
int size, block, i, hash;
fs/hfs/bnode.c
370
hash = hfs_bnode_hash(cnid);
fs/hfs/bnode.c
371
node->next_hash = tree->node_hash[hash];
fs/hfs/bnode.c
372
tree->node_hash[hash] = node;
fs/hfs/string.c
59
unsigned int hash, len = this->len;
fs/hfs/string.c
64
hash = init_name_hash(dentry);
fs/hfs/string.c
66
hash = partial_name_hash(caseorder[*name++], hash);
fs/hfs/string.c
67
this->hash = end_name_hash(hash);
fs/hfs/string_test.c
53
test_name1.hash = 0;
fs/hfs/string_test.c
57
test_name2.hash = 0;
fs/hfs/string_test.c
61
test_name3.hash = 0;
fs/hfs/string_test.c
69
KUNIT_EXPECT_EQ(test, test_name1.hash, test_name2.hash);
fs/hfs/string_test.c
72
KUNIT_EXPECT_NE(test, test_name1.hash, test_name3.hash);
fs/hfsplus/bnode.c
455
int size, block, i, hash;
fs/hfsplus/bnode.c
479
hash = hfs_bnode_hash(cnid);
fs/hfsplus/bnode.c
480
node->next_hash = tree->node_hash[hash];
fs/hfsplus/bnode.c
481
tree->node_hash[hash] = node;
fs/hfsplus/unicode.c
443
unsigned long hash;
fs/hfsplus/unicode.c
450
hash = init_name_hash(dentry);
fs/hfsplus/unicode.c
469
hash = partial_name_hash(c2, hash);
fs/hfsplus/unicode.c
476
hash = partial_name_hash(c2, hash);
fs/hfsplus/unicode.c
479
str->hash = end_name_hash(hash);
fs/hfsplus/unicode_test.c
1023
KUNIT_EXPECT_NE(test, str1.hash, str2.hash);
fs/hfsplus/unicode_test.c
1037
KUNIT_EXPECT_EQ(test, str1.hash, str2.hash);
fs/hfsplus/unicode_test.c
1043
KUNIT_EXPECT_EQ(test, str1.hash, str2.hash);
fs/hfsplus/unicode_test.c
1071
KUNIT_EXPECT_EQ(test, str1.hash, str2.hash);
fs/hfsplus/unicode_test.c
1082
KUNIT_EXPECT_EQ(test, str1.hash, str2.hash);
fs/hfsplus/unicode_test.c
1118
KUNIT_EXPECT_NE(test, 0, str2.hash);
fs/hfsplus/unicode_test.c
1141
hash1 = str1.hash;
fs/hfsplus/unicode_test.c
1147
KUNIT_EXPECT_EQ(test, hash1, str2.hash);
fs/hfsplus/unicode_test.c
1154
KUNIT_EXPECT_NE(test, str1.hash, str3.hash);
fs/hfsplus/unicode_test.c
1165
KUNIT_EXPECT_NE(test, str1.hash, str2.hash);
fs/hfsplus/unicode_test.c
1195
KUNIT_EXPECT_NE(test, 0, str.hash);
fs/hfsplus/unicode_test.c
1202
KUNIT_EXPECT_NE(test, 0, str.hash);
fs/hfsplus/unicode_test.c
1207
str.hash = 0;
fs/hfsplus/unicode_test.c
1211
KUNIT_EXPECT_NE(test, 0, str.hash);
fs/hfsplus/unicode_test.c
1440
.hash = 0
fs/hfsplus/unicode_test.c
948
str->hash = 0; /* Will be set by hash function */
fs/hfsplus/unicode_test.c
969
KUNIT_EXPECT_NE(test, 0, str1.hash);
fs/hfsplus/unicode_test.c
976
KUNIT_EXPECT_EQ(test, str1.hash, str2.hash);
fs/hfsplus/unicode_test.c
990
KUNIT_EXPECT_NE(test, 0, str1.hash);
fs/hpfs/dentry.c
18
unsigned long hash;
fs/hpfs/dentry.c
30
hash = init_name_hash(dentry);
fs/hpfs/dentry.c
32
hash = partial_name_hash(hpfs_upcase(hpfs_sb(dentry->d_sb)->sb_cp_table,qstr->name[i]), hash);
fs/hpfs/dentry.c
33
qstr->hash = end_name_hash(hash);
fs/hugetlbfs/inode.c
593
u32 hash = 0;
fs/hugetlbfs/inode.c
596
hash = hugetlb_fault_mutex_hash(mapping, index);
fs/hugetlbfs/inode.c
597
mutex_lock(&hugetlb_fault_mutex_table[hash]);
fs/hugetlbfs/inode.c
606
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
fs/hugetlbfs/inode.c
745
u32 hash;
fs/hugetlbfs/inode.c
806
hash = hugetlb_fault_mutex_hash(mapping, index);
fs/hugetlbfs/inode.c
807
mutex_lock(&hugetlb_fault_mutex_table[hash]);
fs/hugetlbfs/inode.c
813
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
fs/hugetlbfs/inode.c
827
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
fs/hugetlbfs/inode.c
837
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
fs/hugetlbfs/inode.c
841
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
fs/inode.c
1308
struct hlist_head *head = inode_hashtable + hash(inode->i_sb, hashval);
fs/inode.c
1415
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1460
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1532
struct hlist_head *b = inode_hashtable + hash(sb, ino);
fs/inode.c
1622
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1682
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1734
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1779
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1817
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1837
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
695
struct hlist_head *b = inode_hashtable + hash(inode->i_sb, hashval);
fs/isofs/inode.c
180
unsigned long hash;
fs/isofs/inode.c
189
hash = init_name_hash(dentry);
fs/isofs/inode.c
192
hash = partial_name_hash(c, hash);
fs/isofs/inode.c
194
qstr->hash = end_name_hash(hash);
fs/isofs/inode.c
259
qstr->hash = full_name_hash(dentry, name, len);
fs/jbd2/revoke.c
104
struct list_head hash;
fs/jbd2/revoke.c
151
hash_list = &journal->j_revoke->hash_table[hash(journal, blocknr)];
fs/jbd2/revoke.c
153
list_add(&record->hash, hash_list);
fs/jbd2/revoke.c
166
hash_list = &journal->j_revoke->hash_table[hash(journal, blocknr)];
fs/jbd2/revoke.c
170
while (&(record->hash) != hash_list) {
fs/jbd2/revoke.c
175
record = (struct jbd2_revoke_record_s *) record->hash.next;
fs/jbd2/revoke.c
451
list_del(&record->hash);
fs/jbd2/revoke.c
557
list_del(&record->hash);
fs/jbd2/revoke.c
739
list_del(&record->hash);
fs/jfs/namei.c
1556
unsigned long hash;
fs/jfs/namei.c
1559
hash = init_name_hash(dir);
fs/jfs/namei.c
1561
hash = partial_name_hash(tolower(this->name[i]), hash);
fs/jfs/namei.c
1562
this->hash = end_name_hash(hash);
fs/kernfs/dir.c
1817
kn->hash = kernfs_name_hash(new_name ?: old_name, kn->ns);
fs/kernfs/dir.c
1836
struct kernfs_node *parent, loff_t hash, struct kernfs_node *pos)
fs/kernfs/dir.c
1841
hash == pos->hash;
fs/kernfs/dir.c
1846
if (!pos && (hash > 1) && (hash < INT_MAX)) {
fs/kernfs/dir.c
1851
if (hash < pos->hash)
fs/kernfs/dir.c
1853
else if (hash > pos->hash)
fs/kernfs/dir.c
1911
ctx->pos = pos->hash;
fs/kernfs/dir.c
318
unsigned long hash = init_name_hash(ns);
fs/kernfs/dir.c
321
hash = partial_name_hash(*name++, hash);
fs/kernfs/dir.c
322
hash = end_name_hash(hash);
fs/kernfs/dir.c
323
hash &= 0x7fffffffU;
fs/kernfs/dir.c
325
if (hash < 2)
fs/kernfs/dir.c
326
hash += 2;
fs/kernfs/dir.c
327
if (hash >= INT_MAX)
fs/kernfs/dir.c
328
hash = INT_MAX - 1;
fs/kernfs/dir.c
329
return hash;
fs/kernfs/dir.c
332
static int kernfs_name_compare(unsigned int hash, const char *name,
fs/kernfs/dir.c
335
if (hash < kn->hash)
fs/kernfs/dir.c
337
if (hash > kn->hash)
fs/kernfs/dir.c
349
return kernfs_name_compare(left->hash, kernfs_rcu_name(left), left->ns, right);
fs/kernfs/dir.c
813
kn->hash = kernfs_name_hash(kernfs_rcu_name(kn), kn->ns);
fs/kernfs/dir.c
863
unsigned int hash;
fs/kernfs/dir.c
873
hash = kernfs_name_hash(name, ns);
fs/kernfs/dir.c
879
result = kernfs_name_compare(hash, name, ns, kn);
fs/lockd/host.c
100
hash = __nlm_hash_addr6(sap);
fs/lockd/host.c
103
hash = 0;
fs/lockd/host.c
105
return hash & (NLM_HOST_NRHASH - 1);
fs/lockd/host.c
71
unsigned int hash = (__force u32)n ^ ((__force u32)n >> 16);
fs/lockd/host.c
72
return hash ^ (hash >> 8);
fs/lockd/host.c
93
unsigned int hash;
fs/lockd/host.c
97
hash = __nlm_hash_addr4(sap);
fs/lockd/svcsubs.c
109
unsigned int hash;
fs/lockd/svcsubs.c
115
hash = file_hash(&lock->fh);
fs/lockd/svcsubs.c
121
hlist_for_each_entry(file, &nlm_files[hash], f_list)
fs/lockd/svcsubs.c
144
hlist_add_head(&file->f_list, &nlm_files[hash]);
fs/namei.c
2470
nd->last.hash = fold_hash(a, y);
fs/namei.c
2492
nd->last.hash = fold_hash(x, y);
fs/namei.c
2511
unsigned long hash = init_name_hash(salt);
fs/namei.c
2513
hash = partial_name_hash((unsigned char)*name++, hash);
fs/namei.c
2514
return end_name_hash(hash);
fs/namei.c
2521
unsigned long hash = init_name_hash(salt);
fs/namei.c
2527
hash = partial_name_hash(c, hash);
fs/namei.c
2530
return hashlen_create(end_name_hash(hash), len);
fs/namei.c
2540
unsigned long hash = init_name_hash(nd->path.dentry);
fs/namei.c
2547
hash = partial_name_hash(c, hash);
fs/namei.c
2555
nd->last.hash = end_name_hash(hash);
fs/namei.c
3093
qname->hash = full_name_hash(base, name, len);
fs/nfs/delegation.c
1193
hlist_for_each_entry_rcu(delegation, head, hash) {
fs/nfs/delegation.c
365
hlist_del_init_rcu(&delegation->hash);
fs/nfs/delegation.c
520
hlist_add_head_rcu(&delegation->hash,
fs/nfs/delegation.h
17
struct hlist_node hash;
fs/nfs/dir.c
754
filename.hash = full_name_hash(parent, filename.name, filename.len);
fs/nfs/inode.c
416
unsigned long hash;
fs/nfs/inode.c
422
hash = nfs_fattr_to_ino_t(fattr);
fs/nfs/inode.c
423
inode = ilookup5(sb, hash, nfs_find_actor, &desc);
fs/nfs/inode.c
459
unsigned long hash;
fs/nfs/inode.c
470
hash = nfs_fattr_to_ino_t(fattr);
fs/nfs/inode.c
472
inode = iget5_locked(sb, hash, nfs_find_actor, nfs_init_locked, &desc);
fs/nfs/inode.c
484
inode->i_ino = hash;
fs/nfs/pnfs_dev.c
171
const struct nfs4_deviceid *id, long hash)
fs/nfs/pnfs_dev.c
177
hash);
fs/nfs/pnfs_dev.c
189
long hash = nfs4_deviceid_hash(id);
fs/nfs/pnfs_dev.c
192
d = __nfs4_find_get_deviceid(server, id, hash);
fs/nfs/pnfs_dev.c
203
d = __nfs4_find_get_deviceid(server, id, hash);
fs/nfs/pnfs_dev.c
209
hlist_add_head_rcu(&new->node, &nfs4_deviceid_cache[hash]);
fs/nfs/pnfs_dev.c
328
_deviceid_purge_client(const struct nfs_client *clp, long hash)
fs/nfs/pnfs_dev.c
335
hlist_for_each_entry_rcu(d, &nfs4_deviceid_cache[hash], node)
fs/nfs/pnfs_dev.c
81
long hash)
fs/nfs/pnfs_dev.c
85
hlist_for_each_entry_rcu(d, &nfs4_deviceid_cache[hash], node)
fs/nfsd/export.c
285
int hash = item->ek_fsidtype;
fs/nfsd/export.c
289
hash ^= hash_mem(cp, len, EXPKEY_HASHBITS);
fs/nfsd/export.c
290
hash ^= hash_ptr(item->ek_client, EXPKEY_HASHBITS);
fs/nfsd/export.c
291
hash &= EXPKEY_HASHMASK;
fs/nfsd/export.c
292
return hash;
fs/nfsd/export.c
299
int hash = svc_expkey_hash(item);
fs/nfsd/export.c
301
ch = sunrpc_cache_lookup_rcu(cd, &item->h, hash);
fs/nfsd/export.c
313
int hash = svc_expkey_hash(new);
fs/nfsd/export.c
315
ch = sunrpc_cache_update(cd, &new->h, &old->h, hash);
fs/nfsd/export.c
922
int hash;
fs/nfsd/export.c
924
hash = hash_ptr(exp->ex_client, EXPORT_HASHBITS);
fs/nfsd/export.c
925
hash ^= hash_ptr(exp->ex_path.dentry, EXPORT_HASHBITS);
fs/nfsd/export.c
926
hash ^= hash_ptr(exp->ex_path.mnt, EXPORT_HASHBITS);
fs/nfsd/export.c
927
return hash;
fs/nfsd/export.c
934
int hash = svc_export_hash(exp);
fs/nfsd/export.c
936
ch = sunrpc_cache_lookup_rcu(exp->cd, &exp->h, hash);
fs/nfsd/export.c
947
int hash = svc_export_hash(old);
fs/nfsd/export.c
949
ch = sunrpc_cache_update(old->cd, &new->h, &old->h, hash);
fs/nfsd/nfs4idmap.c
114
uint32_t hash;
fs/nfsd/nfs4idmap.c
116
hash = hash_str(ent->authname, ENT_HASHBITS);
fs/nfsd/nfs4idmap.c
117
hash = hash_long(hash ^ ent->id, ENT_HASHBITS);
fs/nfsd/nfs4idmap.c
121
hash ^= 1;
fs/nfsd/nfs4idmap.c
123
return hash;
fs/nfsd/nfs4layouts.c
103
list_for_each_entry_rcu(map, &nfsd_devid_hash[devid_hashfn(idx)], hash)
fs/nfsd/nfs4layouts.c
75
list_for_each_entry(old, &nfsd_devid_hash[i], hash) {
fs/nfsd/nfs4layouts.c
822
list_for_each_entry_safe(map, n, &nfsd_devid_hash[i], hash)
fs/nfsd/nfs4layouts.c
88
list_add_tail_rcu(&map->hash, &nfsd_devid_hash[devid_hashfn(map->idx)]);
fs/nfsd/nfs4state.c
1079
u32 hash;
fs/nfsd/nfs4state.c
1096
hash = jhash(&fh->fh_raw, fh->fh_size, 0);
fs/nfsd/nfs4state.c
1097
if (test_bit(hash&255, bd->set[0]) &&
fs/nfsd/nfs4state.c
1098
test_bit((hash>>8)&255, bd->set[0]) &&
fs/nfsd/nfs4state.c
1099
test_bit((hash>>16)&255, bd->set[0]))
fs/nfsd/nfs4state.c
1102
if (test_bit(hash&255, bd->set[1]) &&
fs/nfsd/nfs4state.c
1103
test_bit((hash>>8)&255, bd->set[1]) &&
fs/nfsd/nfs4state.c
1104
test_bit((hash>>16)&255, bd->set[1]))
fs/nfsd/nfs4state.c
1112
u32 hash;
fs/nfsd/nfs4state.c
1115
hash = jhash(&fh->fh_raw, fh->fh_size, 0);
fs/nfsd/nfs4state.c
1118
__set_bit(hash&255, bd->set[bd->new]);
fs/nfsd/nfs4state.c
1119
__set_bit((hash>>8)&255, bd->set[bd->new]);
fs/nfsd/nfs4state.c
1120
__set_bit((hash>>16)&255, bd->set[bd->new]);
fs/nfsd/nfscache.c
250
unsigned int hash = hash_32((__force u32)xid, nn->maskbits);
fs/nfsd/nfscache.c
252
return &nn->drc_hashtbl[hash];
fs/nfsd/pnfs.h
15
struct list_head hash;
fs/notify/fanotify/fanotify.c
134
if (old->hash != new->hash ||
fs/notify/fanotify/fanotify.c
413
unsigned int fh_len, unsigned int *hash,
fs/notify/fanotify/fanotify.c
474
if (hash)
fs/notify/fanotify/fanotify.c
475
*hash ^= fanotify_hash_fh(fh);
fs/notify/fanotify/fanotify.c
554
unsigned int *hash,
fs/notify/fanotify/fanotify.c
565
*hash ^= fanotify_hash_path(path);
fs/notify/fanotify/fanotify.c
615
unsigned int *hash,
fs/notify/fanotify/fanotify.c
626
*hash ^= fanotify_hash_fsid(fsid);
fs/notify/fanotify/fanotify.c
628
hash, gfp);
fs/notify/fanotify/fanotify.c
638
unsigned int *hash,
fs/notify/fanotify/fanotify.c
667
*hash ^= fanotify_hash_fsid(fsid);
fs/notify/fanotify/fanotify.c
672
len = fanotify_encode_fh(dfh, dir, dir_fh_len, hash, 0);
fs/notify/fanotify/fanotify.c
677
len = fanotify_encode_fh(dfh, dir2, dir2_fh_len, hash, 0);
fs/notify/fanotify/fanotify.c
682
len = fanotify_encode_fh(ffh, child, child_fh_len, hash, 0);
fs/notify/fanotify/fanotify.c
687
*hash ^= full_name_hash((void *)name_len, name->name, name_len);
fs/notify/fanotify/fanotify.c
691
*hash ^= full_name_hash((void *)name2_len, name2->name,
fs/notify/fanotify/fanotify.c
712
unsigned int *hash)
fs/notify/fanotify/fanotify.c
741
*hash ^= fanotify_hash_fsid(fsid);
fs/notify/fanotify/fanotify.c
764
unsigned int hash = 0;
fs/notify/fanotify/fanotify.c
849
data_type, &hash);
fs/notify/fanotify/fanotify.c
852
moved, &hash, gfp);
fs/notify/fanotify/fanotify.c
854
event = fanotify_alloc_fid_event(id, fsid, &hash, gfp);
fs/notify/fanotify/fanotify.c
856
event = fanotify_alloc_path_event(path, &hash, gfp);
fs/notify/fanotify/fanotify.c
872
hash ^= hash_long((unsigned long)pid | ondir, FANOTIFY_EVENT_HASH_BITS);
fs/notify/fanotify/fanotify.c
873
fanotify_init_event(event, hash, mask);
fs/notify/fanotify/fanotify.h
262
unsigned int hash : FANOTIFY_EVENT_HASH_BITS;
fs/notify/fanotify/fanotify.h
268
unsigned int hash, u32 mask)
fs/notify/fanotify/fanotify.h
272
event->hash = hash;
fs/notify/fanotify/fanotify.h
518
return event->hash & FANOTIFY_HTABLE_MASK;
fs/notify/fanotify/fanotify_user.c
1588
struct hlist_head *hash;
fs/notify/fanotify/fanotify_user.c
1590
hash = kmalloc(sizeof(struct hlist_head) << FANOTIFY_HTABLE_BITS,
fs/notify/fanotify/fanotify_user.c
1592
if (!hash)
fs/notify/fanotify/fanotify_user.c
1595
__hash_init(hash, FANOTIFY_HTABLE_SIZE);
fs/notify/fanotify/fanotify_user.c
1597
return hash;
fs/ntfs3/fsntfs.c
1024
u32 hash = 0;
fs/ntfs3/fsntfs.c
1029
hash = ((hash >> 0x1D) | (hash << 3)) + le32_to_cpu(*ptr++);
fs/ntfs3/fsntfs.c
1030
return cpu_to_le32(hash);
fs/ntfs3/fsntfs.c
2105
hash_key.hash = security_hash(sd, size_sd);
fs/ntfs3/fsntfs.c
2156
d_security->key.hash == hash_key.hash &&
fs/ntfs3/fsntfs.c
2170
if (!e || e->key.hash != hash_key.hash)
fs/ntfs3/fsntfs.c
2198
d_security->key.hash = hash_key.hash;
fs/ntfs3/fsntfs.c
2253
sdh_e.key.hash = d_security->key.hash;
fs/ntfs3/index.c
98
t1 = le32_to_cpu(k1->hash);
fs/ntfs3/index.c
99
t2 = le32_to_cpu(k2->hash);
fs/ntfs3/lznt.c
302
struct lznt *r = kzalloc(level ? offsetof(struct lznt, hash) :
fs/ntfs3/lznt.c
331
memset(ctx->hash, 0, sizeof(ctx->hash));
fs/ntfs3/lznt.c
35
struct lznt_hash hash[LZNT_CHUNK_SIZE];
fs/ntfs3/lznt.c
52
const u8 **hash;
fs/ntfs3/lznt.c
58
hash = &(ctx->hash[hash_index].p1);
fs/ntfs3/lznt.c
60
if (hash[0] >= ctx->unc && hash[0] < src && hash[0][0] == src[0] &&
fs/ntfs3/lznt.c
61
hash[0][1] == src[1] && hash[0][2] == src[2]) {
fs/ntfs3/lznt.c
65
hash[0] + 3, ctx->max_len - 3);
fs/ntfs3/lznt.c
68
if (hash[1] >= ctx->unc && hash[1] < src && hash[1][0] == src[0] &&
fs/ntfs3/lznt.c
69
hash[1][1] == src[1] && hash[1][2] == src[2]) {
fs/ntfs3/lznt.c
73
hash[1] + 3, ctx->max_len - 3);
fs/ntfs3/lznt.c
78
ctx->best_match = hash[1];
fs/ntfs3/lznt.c
81
ctx->best_match = hash[0];
fs/ntfs3/lznt.c
84
hash[1] = hash[0];
fs/ntfs3/lznt.c
85
hash[0] = src;
fs/ntfs3/namei.c
385
unsigned long hash;
fs/ntfs3/namei.c
391
hash = init_name_hash(dentry);
fs/ntfs3/namei.c
395
name->hash = end_name_hash(hash);
fs/ntfs3/namei.c
403
hash = partial_name_hash(toupper(c), hash);
fs/ntfs3/namei.c
425
hash = ntfs_names_hash(uni->name, uni->len, sbi->upcase,
fs/ntfs3/namei.c
427
name->hash = end_name_hash(hash);
fs/ntfs3/ntfs.h
915
__le32 hash; // Hash value for descriptor
fs/ntfs3/ntfs_fs.h
932
unsigned long hash);
fs/ntfs3/upcase.c
108
unsigned long hash)
fs/ntfs3/upcase.c
112
hash = partial_name_hash(c, hash);
fs/ntfs3/upcase.c
115
return hash;
fs/ocfs2/dlm/dlmcommon.h
867
unsigned int hash);
fs/ocfs2/dlm/dlmcommon.h
871
unsigned int hash);
fs/ocfs2/dlm/dlmdomain.c
162
bucket = dlm_lockres_hash(dlm, res->lockname.hash);
fs/ocfs2/dlm/dlmdomain.c
176
unsigned int hash)
fs/ocfs2/dlm/dlmdomain.c
185
bucket = dlm_lockres_hash(dlm, hash);
fs/ocfs2/dlm/dlmdomain.c
209
unsigned int hash)
fs/ocfs2/dlm/dlmdomain.c
217
res = __dlm_lookup_lockres_full(dlm, name, len, hash);
fs/ocfs2/dlm/dlmdomain.c
236
unsigned int hash = dlm_lockid_hash(name, len);
fs/ocfs2/dlm/dlmdomain.c
239
res = __dlm_lookup_lockres(dlm, name, len, hash);
fs/ocfs2/dlm/dlmmaster.c
1409
unsigned int namelen, hash;
fs/ocfs2/dlm/dlmmaster.c
1425
hash = dlm_lockid_hash(name, namelen);
fs/ocfs2/dlm/dlmmaster.c
1434
res = __dlm_lookup_lockres(dlm, name, namelen, hash);
fs/ocfs2/dlm/dlmmaster.c
1758
unsigned int namelen, hash;
fs/ocfs2/dlm/dlmmaster.c
1768
hash = dlm_lockid_hash(name, namelen);
fs/ocfs2/dlm/dlmmaster.c
1834
res = __dlm_lookup_lockres(dlm, name, namelen, hash);
fs/ocfs2/dlm/dlmmaster.c
2248
unsigned int hash;
fs/ocfs2/dlm/dlmmaster.c
2269
hash = dlm_lockid_hash(name, namelen);
fs/ocfs2/dlm/dlmmaster.c
2272
res = __dlm_lookup_lockres_full(dlm, name, namelen, hash);
fs/ocfs2/dlm/dlmmaster.c
2343
unsigned int hash;
fs/ocfs2/dlm/dlmmaster.c
2361
hash = dlm_lockid_hash(name, namelen);
fs/ocfs2/dlm/dlmmaster.c
2364
res = __dlm_lookup_lockres_full(dlm, name, namelen, hash);
fs/ocfs2/dlm/dlmmaster.c
280
mle->mnamehash = res->lockname.hash;
fs/ocfs2/dlm/dlmmaster.c
3096
unsigned int namelen, hash;
fs/ocfs2/dlm/dlmmaster.c
3104
hash = dlm_lockid_hash(name, namelen);
fs/ocfs2/dlm/dlmmaster.c
3116
res = __dlm_lookup_lockres(dlm, name, namelen, hash);
fs/ocfs2/dlm/dlmmaster.c
328
unsigned int hash;
fs/ocfs2/dlm/dlmmaster.c
332
hash = dlm_lockid_hash(name, namelen);
fs/ocfs2/dlm/dlmmaster.c
333
bucket = dlm_master_hash(dlm, hash);
fs/ocfs2/dlm/dlmmaster.c
537
res->lockname.hash = dlm_lockid_hash(name, namelen);
fs/ocfs2/dlm/dlmmaster.c
712
unsigned int hash;
fs/ocfs2/dlm/dlmmaster.c
718
hash = dlm_lockid_hash(lockid, namelen);
fs/ocfs2/dlm/dlmmaster.c
724
tmpres = __dlm_lookup_lockres_full(dlm, lockid, namelen, hash);
fs/ocfs2/dlm/dlmrecovery.c
1355
unsigned int hash;
fs/ocfs2/dlm/dlmrecovery.c
1392
hash = dlm_lockid_hash(mres->lockname, mres->lockname_len);
fs/ocfs2/dlm/dlmrecovery.c
1395
hash);
fs/ocfs2/dlm/dlmrecovery.c
1701
unsigned int hash;
fs/ocfs2/dlm/dlmrecovery.c
1712
hash = dlm_lockid_hash(req->name, req->namelen);
fs/ocfs2/dlm/dlmrecovery.c
1715
res = __dlm_lookup_lockres(dlm, req->name, req->namelen, hash);
fs/ocfs2/ocfs2_trace.h
1758
unsigned int hash, unsigned long long location,
fs/ocfs2/ocfs2_trace.h
1760
TP_ARGS(ino, name, name_index, hash, location, xe_index),
fs/ocfs2/ocfs2_trace.h
1765
__field(unsigned int, hash)
fs/ocfs2/ocfs2_trace.h
1773
__entry->hash = hash;
fs/ocfs2/ocfs2_trace.h
1778
__entry->name_index, __entry->hash, __entry->location,
fs/ocfs2/ocfs2_trace.h
1785
unsigned int hash, unsigned long long bucket, \
fs/ocfs2/ocfs2_trace.h
1787
TP_ARGS(ino, name, name_index, hash, bucket, xe_index))
fs/ocfs2/xattr.c
545
u32 hash = OCFS2_SB(inode->i_sb)->uuid_hash;
fs/ocfs2/xattr.c
550
hash = (hash << OCFS2_HASH_SHIFT) ^
fs/ocfs2/xattr.c
551
(hash >> (8*sizeof(hash) - OCFS2_HASH_SHIFT)) ^
fs/ocfs2/xattr.c
555
return hash;
fs/omfs/dir.c
14
int i, hash = 0;
fs/omfs/dir.c
16
hash ^= tolower(name[i]) << (i % 24);
fs/omfs/dir.c
17
return hash % mod;
fs/orangefs/inode.c
1038
unsigned long hash;
fs/orangefs/inode.c
1041
hash = orangefs_handle_hash(ref);
fs/orangefs/inode.c
1043
hash,
fs/orangefs/inode.c
1060
inode->i_ino = hash; /* needed for stat etc */
fs/orangefs/inode.c
1068
hash,
fs/orangefs/inode.c
1081
unsigned long hash = orangefs_handle_hash(ref);
fs/orangefs/inode.c
1102
inode->i_ino = hash; /* needed for stat etc */
fs/orangefs/inode.c
1124
error = insert_inode_locked4(inode, hash, orangefs_test_inode, ref);
fs/proc/proc_sysctl.c
691
qname.hash = full_name_hash(dir, qname.name, qname.len);
fs/smb/client/compress/lz77.c
155
u64 hash;
fs/smb/client/compress/lz77.c
157
hash = ((lz77_read64(srcp) << 24) * 889523592379ULL) >> (64 - LZ77_HASH_LOG);
fs/smb/client/compress/lz77.c
158
wnd = src + htable[hash];
fs/smb/client/compress/lz77.c
159
htable[hash] = srcp - src;
fs/smb/client/dfs_cache.c
320
static int cache_entry_hash(const void *data, int size, unsigned int *hash)
fs/smb/client/dfs_cache.c
336
*hash = h % CACHE_HTABLE_SIZE;
fs/smb/client/dfs_cache.c
478
unsigned int hash;
fs/smb/client/dfs_cache.c
488
rc = cache_entry_hash(refs[0].path_name, strlen(refs[0].path_name), &hash);
fs/smb/client/dfs_cache.c
499
hlist_add_head(&ce->hlist, &cache_htable[hash]);
fs/smb/client/dfs_cache.c
533
static struct cache_entry *__lookup_cache_entry(const char *path, unsigned int hash, int len)
fs/smb/client/dfs_cache.c
537
hlist_for_each_entry(ce, &cache_htable[hash], hlist) {
fs/smb/client/dfs_cache.c
561
unsigned int hash;
fs/smb/client/dfs_cache.c
568
rc = cache_entry_hash(path, strlen(path), &hash);
fs/smb/client/dfs_cache.c
571
return __lookup_cache_entry(path, hash, strlen(path));
fs/smb/client/dfs_cache.c
590
rc = cache_entry_hash(path, len, &hash);
fs/smb/client/dfs_cache.c
593
ce = __lookup_cache_entry(path, hash, len);
fs/smb/client/dir.c
898
unsigned long hash;
fs/smb/client/dir.c
902
hash = init_name_hash(dentry);
fs/smb/client/dir.c
908
hash = partial_name_hash(cifs_toupper(c), hash);
fs/smb/client/dir.c
910
q->hash = end_name_hash(hash);
fs/smb/client/inode.c
1037
__u64 hash = 0;
fs/smb/client/inode.c
1040
hash = (hash + (__u64) *str++) * hash_mult;
fs/smb/client/inode.c
1042
return hash;
fs/smb/client/inode.c
1612
unsigned long hash;
fs/smb/client/inode.c
1619
hash = cifs_uniqueid_to_ino_t(fattr->cf_uniqueid);
fs/smb/client/inode.c
1621
inode = iget5_locked(sb, hash, cifs_find_inode, cifs_init_inode, fattr);
fs/smb/client/inode.c
1640
inode->i_ino = hash;
fs/smb/common/cifs_md4.c
115
hash[0] += a;
fs/smb/common/cifs_md4.c
116
hash[1] += b;
fs/smb/common/cifs_md4.c
117
hash[2] += c;
fs/smb/common/cifs_md4.c
118
hash[3] += d;
fs/smb/common/cifs_md4.c
124
md4_transform(ctx->hash, ctx->block);
fs/smb/common/cifs_md4.c
130
mctx->hash[0] = 0x67452301;
fs/smb/common/cifs_md4.c
131
mctx->hash[1] = 0xefcdab89;
fs/smb/common/cifs_md4.c
132
mctx->hash[2] = 0x98badcfe;
fs/smb/common/cifs_md4.c
133
mctx->hash[3] = 0x10325476;
fs/smb/common/cifs_md4.c
191
md4_transform(mctx->hash, mctx->block);
fs/smb/common/cifs_md4.c
192
cpu_to_le32_array(mctx->hash, ARRAY_SIZE(mctx->hash));
fs/smb/common/cifs_md4.c
193
memcpy(out, mctx->hash, sizeof(mctx->hash));
fs/smb/common/cifs_md4.c
55
static void md4_transform(u32 *hash, u32 const *in)
fs/smb/common/cifs_md4.c
59
a = hash[0];
fs/smb/common/cifs_md4.c
60
b = hash[1];
fs/smb/common/cifs_md4.c
61
c = hash[2];
fs/smb/common/cifs_md4.c
62
d = hash[3];
fs/smb/common/md4.h
17
u32 hash[MD4_HASH_WORDS];
fs/smb/server/ksmbd_netlink.h
149
__s8 hash[KSMBD_REQ_MAX_HASH_SZ]; /* password hash */
fs/smb/server/mgmt/user_config.c
50
memcpy(user->passkey, resp->hash, resp->hash_sz);
fs/smb/server/ndr.c
429
ret = ndr_write_bytes(n, acl->hash, XATTR_SD_HASH_SIZE);
fs/smb/server/ndr.c
487
ret = ndr_read_bytes(n, acl->hash, XATTR_SD_HASH_SIZE);
fs/smb/server/vfs.c
1430
sha256(acl.sd_buf, acl.sd_size, acl.hash);
fs/smb/server/xattr.h
98
__u8 hash[XATTR_SD_HASH_SIZE]; /* 64bytes hash for ntacl */
fs/ubifs/auth.c
136
void ubifs_bad_hash(const struct ubifs_info *c, const void *node, const u8 *hash,
fs/ubifs/auth.c
148
ubifs_err(c, "hash expected: %*ph%s", len, hash, cont);
fs/ubifs/auth.c
29
u8 *hash)
fs/ubifs/auth.c
34
hash);
fs/ubifs/auth.c
45
static int ubifs_hash_calc_hmac(const struct ubifs_info *c, const u8 *hash,
fs/ubifs/auth.c
48
return crypto_shash_tfm_digest(c->hmac_tfm, hash, c->hash_len, hmac);
fs/ubifs/auth.c
66
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/auth.c
75
err = crypto_shash_final(hash_desc, hash);
fs/ubifs/auth.c
80
err = ubifs_hash_calc_hmac(c, hash, auth->hmac);
fs/ubifs/debug.c
38
static const char *get_key_hash(int hash)
fs/ubifs/debug.c
40
switch (hash) {
fs/ubifs/dir.c
246
if (nm.hash & ~UBIFS_S_KEY_HASH_MASK)
fs/ubifs/dir.c
248
dent_key_init_hash(c, &key, dir->i_ino, nm.hash);
fs/ubifs/journal.c
1065
err = ubifs_node_calc_hash(c, ino, hash);
fs/ubifs/journal.c
1089
err = ubifs_tnc_add(c, &key, lnum, offs, ilen, hash);
fs/ubifs/journal.c
1867
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/journal.c
1909
err = ubifs_node_calc_hash(c, ino, hash);
fs/ubifs/journal.c
1946
err = ubifs_tnc_add(c, &key1, lnum, xent_offs + len - hlen, hlen, hash);
fs/ubifs/journal.c
1988
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/journal.c
2015
err = ubifs_node_calc_hash(c, (void *)ino + aligned_len1, hash);
fs/ubifs/journal.c
2038
err = ubifs_tnc_add(c, &key, lnum, offs + aligned_len1, len2, hash);
fs/ubifs/journal.c
226
int len, struct shash_desc *hash)
fs/ubifs/journal.c
243
err = ubifs_shash_update(c, hash, (void *)node, nodelen);
fs/ubifs/journal.c
251
return ubifs_prepare_auth_node(c, node, hash);
fs/ubifs/journal.c
720
dent_key_init_hash(c, &dent_key, dir->i_ino, nm->hash);
fs/ubifs/journal.c
865
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/journal.c
933
err = ubifs_node_calc_hash(c, data, hash);
fs/ubifs/journal.c
942
err = ubifs_tnc_add(c, key, lnum, offs, dlen, hash);
fs/ubifs/journal.c
983
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/key.h
150
uint32_t hash = c->key_hash(fname_name(nm), fname_len(nm));
fs/ubifs/key.h
152
ubifs_assert(c, !(hash & ~UBIFS_S_KEY_HASH_MASK));
fs/ubifs/key.h
154
key->u32[1] = hash | (UBIFS_DENT_KEY << UBIFS_S_KEY_HASH_BITS);
fs/ubifs/key.h
167
uint32_t hash)
fs/ubifs/key.h
169
ubifs_assert(c, !(hash & ~UBIFS_S_KEY_HASH_MASK));
fs/ubifs/key.h
171
key->u32[1] = hash | (UBIFS_DENT_KEY << UBIFS_S_KEY_HASH_BITS);
fs/ubifs/key.h
186
uint32_t hash = c->key_hash(fname_name(nm), fname_len(nm));
fs/ubifs/key.h
188
ubifs_assert(c, !(hash & ~UBIFS_S_KEY_HASH_MASK));
fs/ubifs/key.h
190
key->j32[1] = cpu_to_le32(hash |
fs/ubifs/key.h
219
uint32_t hash = c->key_hash(fname_name(nm), fname_len(nm));
fs/ubifs/key.h
221
ubifs_assert(c, !(hash & ~UBIFS_S_KEY_HASH_MASK));
fs/ubifs/key.h
223
key->u32[1] = hash | (UBIFS_XENT_KEY << UBIFS_S_KEY_HASH_BITS);
fs/ubifs/key.h
237
uint32_t hash = c->key_hash(fname_name(nm), fname_len(nm));
fs/ubifs/key.h
239
ubifs_assert(c, !(hash & ~UBIFS_S_KEY_HASH_MASK));
fs/ubifs/key.h
241
key->j32[1] = cpu_to_le32(hash |
fs/ubifs/key.h
42
static inline uint32_t key_mask_hash(uint32_t hash)
fs/ubifs/key.h
44
hash &= UBIFS_S_KEY_HASH_MASK;
fs/ubifs/key.h
45
if (unlikely(hash <= 2))
fs/ubifs/key.h
46
hash += 3;
fs/ubifs/key.h
47
return hash;
fs/ubifs/lpt.c
1653
int ubifs_lpt_calc_hash(struct ubifs_info *c, u8 *hash)
fs/ubifs/lpt.c
1732
err = ubifs_shash_final(c, desc, hash);
fs/ubifs/lpt.c
1751
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/lpt.c
1756
err = ubifs_lpt_calc_hash(c, hash);
fs/ubifs/lpt.c
1760
if (ubifs_check_hash(c, c->mst_node->hash_lpt, hash)) {
fs/ubifs/lpt.c
600
int *lpt_lebs, int *big_lpt, u8 *hash)
fs/ubifs/lpt.c
835
err = ubifs_shash_final(c, desc, hash);
fs/ubifs/master.c
384
ubifs_copy_hash(c, c->mst_node->hash_root_idx, c->zroot.hash);
fs/ubifs/master.c
459
ubifs_copy_hash(c, c->zroot.hash, c->mst_node->hash_root_idx);
fs/ubifs/replay.c
255
r->len, r->hash, &r->nm);
fs/ubifs/replay.c
280
r->len, r->hash);
fs/ubifs/replay.c
385
const u8 *hash, union ubifs_key *key,
fs/ubifs/replay.c
405
ubifs_copy_hash(c, hash, r->hash);
fs/ubifs/replay.c
435
const u8 *hash, union ubifs_key *key,
fs/ubifs/replay.c
461
ubifs_copy_hash(c, hash, r->hash);
fs/ubifs/replay.c
49
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/replay.c
567
struct shash_desc *log_hash, u8 *hash)
fs/ubifs/replay.c
574
return crypto_shash_final(hash_desc, hash);
fs/ubifs/replay.c
600
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/replay.c
613
err = authenticate_sleb_hash(c, log_hash, hash);
fs/ubifs/replay.c
617
err = crypto_shash_tfm_digest(c->hmac_tfm, hash,
fs/ubifs/replay.c
724
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/replay.c
734
ubifs_node_calc_hash(c, snod->node, hash);
fs/ubifs/replay.c
747
err = insert_node(c, lnum, snod->offs, snod->len, hash,
fs/ubifs/replay.c
759
err = insert_node(c, lnum, snod->offs, snod->len, hash,
fs/ubifs/replay.c
773
err = insert_dent(c, lnum, snod->offs, snod->len, hash,
fs/ubifs/replay.c
799
err = insert_node(c, lnum, snod->offs, snod->len, hash,
fs/ubifs/sb.c
332
ubifs_node_calc_hash(c, ino, hash);
fs/ubifs/sb.c
333
ubifs_copy_hash(c, hash, ubifs_branch_hash(c, br));
fs/ubifs/sb.c
339
ubifs_node_calc_hash(c, idx, hash);
fs/ubifs/sb.c
340
ubifs_copy_hash(c, hash, mst->hash_root_idx);
fs/ubifs/sb.c
91
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/tnc.c
1745
err = ubifs_node_check_hash(c, buf, zbr->hash);
fs/ubifs/tnc.c
1747
ubifs_bad_hash(c, buf, zbr->hash, zbr->lnum, zbr->offs);
fs/ubifs/tnc.c
2311
int offs, int len, const u8 *hash)
fs/ubifs/tnc.c
2326
ubifs_copy_hash(c, hash, zbr.hash);
fs/ubifs/tnc.c
2337
ubifs_copy_hash(c, hash, zbr->hash);
fs/ubifs/tnc.c
2446
int lnum, int offs, int len, const u8 *hash,
fs/ubifs/tnc.c
2489
ubifs_copy_hash(c, hash, zbr->hash);
fs/ubifs/tnc.c
2501
ubifs_copy_hash(c, hash, zbr.hash);
fs/ubifs/tnc.c
515
err = ubifs_node_check_hash(c, buf, zbr->hash);
fs/ubifs/tnc.c
517
ubifs_bad_hash(c, buf, zbr->hash, lnum, offs);
fs/ubifs/tnc_commit.c
29
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/tnc_commit.c
44
ubifs_copy_hash(c, zbr->hash, ubifs_branch_hash(c, br));
fs/ubifs/tnc_commit.c
55
ubifs_node_calc_hash(c, idx, hash);
fs/ubifs/tnc_commit.c
72
ubifs_copy_hash(c, hash, zbr->hash);
fs/ubifs/tnc_commit.c
77
ubifs_copy_hash(c, hash, c->zroot.hash);
fs/ubifs/tnc_commit.c
859
u8 hash[UBIFS_HASH_ARR_SZ];
fs/ubifs/tnc_commit.c
878
ubifs_copy_hash(c, zbr->hash, ubifs_branch_hash(c, br));
fs/ubifs/tnc_commit.c
890
ubifs_node_calc_hash(c, idx, hash);
fs/ubifs/tnc_commit.c
895
ubifs_copy_hash(c, hash,
fs/ubifs/tnc_commit.c
896
znode->cparent->zbranch[znode->ciip].hash);
fs/ubifs/tnc_commit.c
900
ubifs_copy_hash(c, hash,
fs/ubifs/tnc_commit.c
901
znode->parent->zbranch[znode->iip].hash);
fs/ubifs/tnc_commit.c
903
ubifs_copy_hash(c, hash, c->zroot.hash);
fs/ubifs/tnc_misc.c
306
err = ubifs_node_check_hash(c, idx, zzbr->hash);
fs/ubifs/tnc_misc.c
308
ubifs_bad_hash(c, idx, zzbr->hash, lnum, offs);
fs/ubifs/tnc_misc.c
335
ubifs_copy_hash(c, ubifs_branch_hash(c, br), zbr->hash);
fs/ubifs/tnc_misc.c
510
err = ubifs_node_check_hash(c, node, zbr->hash);
fs/ubifs/tnc_misc.c
512
ubifs_bad_hash(c, node, zbr->hash, zbr->lnum, zbr->offs);
fs/ubifs/ubifs.h
1572
u8 *hash);
fs/ubifs/ubifs.h
1574
const void *buf, u8 *hash)
fs/ubifs/ubifs.h
1577
return __ubifs_node_calc_hash(c, buf, hash);
fs/ubifs/ubifs.h
1617
const u8 *hash, int lnum, int offs);
fs/ubifs/ubifs.h
1620
const u8 *hash, int lnum, int offs) {};
fs/ubifs/ubifs.h
1854
int offs, int len, const u8 *hash);
fs/ubifs/ubifs.h
1858
int lnum, int offs, int len, const u8 *hash,
fs/ubifs/ubifs.h
1957
int *lpt_lebs, int *big_lpt, u8 *hash);
fs/ubifs/ubifs.h
1985
int ubifs_lpt_calc_hash(struct ubifs_info *c, u8 *hash);
fs/ubifs/ubifs.h
755
u8 hash[UBIFS_HASH_ARR_SZ];
fs/unicode/utf8-core.c
126
unsigned long hash = init_name_hash(salt);
fs/unicode/utf8-core.c
134
hash = partial_name_hash((unsigned char)c, hash);
fs/unicode/utf8-core.c
136
str->hash = end_name_hash(hash);
fs/xfs/libxfs/xfs_attr_sf.h
17
xfs_dahash_t hash; /* this entry's hash value */
fs/xfs/libxfs/xfs_da_btree.c
2287
xfs_dahash_t hash;
fs/xfs/libxfs/xfs_da_btree.c
2292
for (hash = 0; namelen >= 4; namelen -= 4, name += 4)
fs/xfs/libxfs/xfs_da_btree.c
2293
hash = (name[0] << 21) ^ (name[1] << 14) ^ (name[2] << 7) ^
fs/xfs/libxfs/xfs_da_btree.c
2294
(name[3] << 0) ^ rol32(hash, 7 * 4);
fs/xfs/libxfs/xfs_da_btree.c
2302
rol32(hash, 7 * 3);
fs/xfs/libxfs/xfs_da_btree.c
2304
return (name[0] << 7) ^ (name[1] << 0) ^ rol32(hash, 7 * 2);
fs/xfs/libxfs/xfs_da_btree.c
2306
return (name[0] << 0) ^ rol32(hash, 7 * 1);
fs/xfs/libxfs/xfs_da_btree.c
2308
return hash;
fs/xfs/libxfs/xfs_dir2.c
75
xfs_dahash_t hash;
fs/xfs/libxfs/xfs_dir2.c
78
for (i = 0, hash = 0; i < name->len; i++)
fs/xfs/libxfs/xfs_dir2.c
79
hash = xfs_ascii_ci_xfrm(name->name[i]) ^ rol32(hash, 7);
fs/xfs/libxfs/xfs_dir2.c
81
return hash;
fs/xfs/libxfs/xfs_dir2_block.c
370
xfs_dahash_t hash; /* hash value of found entry */
fs/xfs/libxfs/xfs_dir2_block.c
461
if ((hash = be32_to_cpu(blp[mid].hashval)) == args->hashval)
fs/xfs/libxfs/xfs_dir2_block.c
463
if (hash < args->hashval)
fs/xfs/libxfs/xfs_dir2_block.c
695
xfs_dahash_t hash; /* found hash value */
fs/xfs/libxfs/xfs_dir2_block.c
720
if ((hash = be32_to_cpu(blp[mid].hashval)) == args->hashval)
fs/xfs/libxfs/xfs_dir2_block.c
722
if (hash < args->hashval)
fs/xfs/libxfs/xfs_dir2_block.c
764
be32_to_cpu(blp[mid].hashval) == hash);
fs/xfs/libxfs/xfs_dir2_data.c
110
xfs_dahash_t hash; /* hash of current name */
fs/xfs/libxfs/xfs_dir2_data.c
260
hash = xfs_dir2_hashname(mp, &name);
fs/xfs/libxfs/xfs_dir2_data.c
263
be32_to_cpu(lep[i].hashval) == hash)
fs/xfs/libxfs/xfs_dir2_leaf.c
1604
xfs_dahash_t hash=0; /* hash from this entry */
fs/xfs/libxfs/xfs_dir2_leaf.c
1622
if ((hash = be32_to_cpu(lep[mid].hashval)) == hashwant)
fs/xfs/libxfs/xfs_dir2_leaf.c
1624
if (hash < hashwant)
fs/xfs/libxfs/xfs_dir2_leaf.c
1632
if (hash == hashwant) {
fs/xfs/libxfs/xfs_dir2_leaf.c
1640
else if (hash < hashwant)
fs/xfs/scrub/attr.c
496
xfs_dahash_t hash;
fs/xfs/scrub/attr.c
527
hash = be32_to_cpu(ent->hashval);
fs/xfs/scrub/attr.c
563
if (calc_hash != hash)
fs/xfs/scrub/dabtree.c
121
xfs_dahash_t hash;
fs/xfs/scrub/dabtree.c
125
hash = be32_to_cpu(*hashp);
fs/xfs/scrub/dabtree.c
126
if (hash < ds->hashes[level])
fs/xfs/scrub/dabtree.c
128
ds->hashes[level] = hash;
fs/xfs/scrub/dabtree.c
136
if (parent_hash < hash)
fs/xfs/scrub/dir.c
336
xfs_dahash_t hash;
fs/xfs/scrub/dir.c
409
hash = be32_to_cpu(ent->hashval);
fs/xfs/scrub/dir.c
422
if (calc_hash != hash)
fs/xfs/scrub/dir.c
633
__u32 hash;
fs/xfs/scrub/dir.c
686
hash = be32_to_cpu(leafhdr.ents[i].hashval);
fs/xfs/scrub/dir.c
687
if (i > 0 && lasthash > hash)
fs/xfs/scrub/dir.c
689
lasthash = hash;
fs/xfs/scrub/dirtree_repair.c
312
qname.hash = full_name_hash(parent_dentry, name->name, name->len);
fs/xfs/scrub/quotacheck.c
178
.head_offset = offsetof(struct xqcheck_dqacct, hash),
fs/xfs/scrub/quotacheck.c
265
&dqa->hash, xqcheck_dqacct_hash_params);
fs/xfs/scrub/quotacheck.c
372
&dqa->hash, xqcheck_dqacct_hash_params);
fs/xfs/scrub/quotacheck.c
62
struct rhash_head hash;
fs/xfs/xfs_attr_list.c
146
sbp->hash = xfs_attr_hashval(dp->i_mount, sfe->flags,
fs/xfs/xfs_attr_list.c
167
if (sbp->hash == cursor->hashval) {
fs/xfs/xfs_attr_list.c
172
} else if (sbp->hash > cursor->hashval) {
fs/xfs/xfs_attr_list.c
183
if (cursor->hashval != sbp->hash) {
fs/xfs/xfs_attr_list.c
184
cursor->hashval = sbp->hash;
fs/xfs/xfs_attr_list.c
34
if (sa->hash < sb->hash) {
fs/xfs/xfs_attr_list.c
36
} else if (sa->hash > sb->hash) {
fs/xfs/xfs_dahash_test.c
652
xfs_dahash_t hash;
fs/xfs/xfs_dahash_test.c
654
hash = xfs_da_hashname(test_buf + test[i].start,
fs/xfs/xfs_dahash_test.c
656
if (hash != test[i].dahash)
fs/xfs/xfs_dahash_test.c
661
hash = xfs_ascii_ci_hashname(&xname);
fs/xfs/xfs_dahash_test.c
662
if (hash != test[i].ascii_ci_dahash)
include/crypto/internal/hash.h
181
static inline struct ahash_alg *crypto_ahash_alg(struct crypto_ahash *hash)
include/crypto/internal/hash.h
183
return container_of(crypto_hash_alg_common(hash), struct ahash_alg,
include/crypto/md5.h
26
u32 hash[MD5_HASH_WORDS];
include/crypto/nh.h
50
__le64 hash[NH_NUM_PASSES]);
include/crypto/streebog.h
26
struct streebog_uint512 hash;
include/keys/system_keyring.h
77
extern int mark_hash_blacklisted(const u8 *hash, size_t hash_len,
include/keys/system_keyring.h
79
extern int is_hash_blacklisted(const u8 *hash, size_t hash_len,
include/keys/system_keyring.h
81
extern int is_binary_blacklisted(const u8 *hash, size_t hash_len);
include/keys/system_keyring.h
83
static inline int is_hash_blacklisted(const u8 *hash, size_t hash_len,
include/keys/system_keyring.h
89
static inline int is_binary_blacklisted(const u8 *hash, size_t hash_len)
include/keys/trusted-type.h
49
uint32_t hash;
include/linux/bpf.h
3798
u32 hash;
include/linux/ceph/ceph_fs.h
477
__le32 hash;
include/linux/ceph/osd_client.h
374
u32 hash;
include/linux/ceph/osd_client.h
386
hoid->hash_reverse_bits = bitrev32(hoid->hash);
include/linux/cfi.h
42
u32 hash;
include/linux/cfi.h
44
if (get_kernel_nofault(hash, func - cfi_get_offset()))
include/linux/cfi.h
47
return hash;
include/linux/crush/crush.h
135
__u8 hash; /* which hash function to use, CRUSH_HASH_* */
include/linux/dcache.h
35
#define HASH_LEN_DECLARE u32 hash; u32 len
include/linux/dcache.h
38
#define HASH_LEN_DECLARE u32 len; u32 hash
include/linux/digsig.h
39
uint8_t hash;
include/linux/filter.h
1249
u32 hash);
include/linux/filter.h
1255
u32 hash)
include/linux/fscrypt.h
444
u32 hash, u32 minor_hash,
include/linux/fscrypt.h
46
u32 hash;
include/linux/fscrypt.h
723
u32 hash, u32 minor_hash,
include/linux/ftrace.h
415
void free_ftrace_hash(struct ftrace_hash *hash);
include/linux/ftrace.h
416
struct ftrace_func_entry *add_ftrace_hash_entry_direct(struct ftrace_hash *hash,
include/linux/ftrace.h
548
int update_ftrace_direct_add(struct ftrace_ops *ops, struct ftrace_hash *hash);
include/linux/ftrace.h
549
int update_ftrace_direct_del(struct ftrace_ops *ops, struct ftrace_hash *hash);
include/linux/ftrace.h
550
int update_ftrace_direct_mod(struct ftrace_ops *ops, struct ftrace_hash *hash, bool do_direct_lock);
include/linux/ftrace.h
578
static inline int update_ftrace_direct_add(struct ftrace_ops *ops, struct ftrace_hash *hash)
include/linux/ftrace.h
583
static inline int update_ftrace_direct_del(struct ftrace_ops *ops, struct ftrace_hash *hash)
include/linux/ftrace.h
588
static inline int update_ftrace_direct_mod(struct ftrace_ops *ops, struct ftrace_hash *hash, bool do_direct_lock)
include/linux/ima.h
25
bool hash, u8 *digest, size_t digest_len);
include/linux/ima.h
61
bool hash, u8 *digest,
include/linux/kernfs.h
213
unsigned int hash; /* ns + name hash */
include/linux/key.h
116
unsigned long hash; /* Hash value */
include/linux/key.h
249
unsigned long hash;
include/linux/mlx5/mlx5_ifc.h
7837
u8 hash[0x1];
include/linux/module_signature.h
35
u8 hash; /* Digest algorithm [0] */
include/linux/netdevice.h
360
struct gro_list hash[GRO_HASH_BUCKETS];
include/linux/nvme-auth.h
13
u8 hash;
include/linux/nvme-auth.h
31
struct nvme_dhchap_key *nvme_auth_alloc_key(u32 len, u8 hash);
include/linux/platform_data/cros_ec_commands.h
4956
uint8_t hash[32];
include/linux/platform_data/cros_ec_commands.h
4964
uint8_t hash[32];
include/linux/platform_data/cros_ec_commands.h
4981
uint8_t hash[32];
include/linux/rhashtable.h
1060
unsigned int hash;
include/linux/rhashtable.h
1063
hash = rht_head_hashfn(ht, tbl, obj, params);
include/linux/rhashtable.h
1064
bkt = rht_bucket_var(tbl, hash);
include/linux/rhashtable.h
1070
rht_for_each_from(he, rht_ptr(bkt, tbl, hash), tbl, hash) {
include/linux/rhashtable.h
1086
tbl, hash);
include/linux/rhashtable.h
1092
list = rht_dereference_bucket(list->next, tbl, hash);
include/linux/rhashtable.h
1098
obj = rht_dereference_bucket(obj->next, tbl, hash);
include/linux/rhashtable.h
1102
list = rht_dereference_bucket(list->next, tbl, hash);
include/linux/rhashtable.h
120
unsigned int hash)
include/linux/rhashtable.h
1213
unsigned int hash;
include/linux/rhashtable.h
1219
hash = rht_head_hashfn(ht, tbl, obj_old, params);
include/linux/rhashtable.h
122
return hash & (tbl->size - 1);
include/linux/rhashtable.h
1220
if (hash != rht_head_hashfn(ht, tbl, obj_new, params))
include/linux/rhashtable.h
1223
bkt = rht_bucket_var(tbl, hash);
include/linux/rhashtable.h
1230
rht_for_each_from(he, rht_ptr(bkt, tbl, hash), tbl, hash) {
include/linux/rhashtable.h
129
unsigned int hash;
include/linux/rhashtable.h
133
hash = ht->p.hashfn(key, ht->key_len, hash_rnd);
include/linux/rhashtable.h
138
hash = params.hashfn(key, key_len, hash_rnd);
include/linux/rhashtable.h
140
hash = jhash(key, key_len, hash_rnd);
include/linux/rhashtable.h
142
hash = jhash2(key, key_len / sizeof(u32), hash_rnd);
include/linux/rhashtable.h
147
hash = params.hashfn(key, key_len, hash_rnd);
include/linux/rhashtable.h
149
hash = jhash(key, key_len, hash_rnd);
include/linux/rhashtable.h
152
return hash;
include/linux/rhashtable.h
159
unsigned int hash = rht_key_get_hash(ht, key, params, tbl->hash_rnd);
include/linux/rhashtable.h
161
return rht_bucket_index(tbl, hash);
include/linux/rhashtable.h
228
int lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash);
include/linux/rhashtable.h
236
u32 hash)
include/linux/rhashtable.h
266
const struct bucket_table *tbl, unsigned int hash);
include/linux/rhashtable.h
268
const struct bucket_table *tbl, unsigned int hash);
include/linux/rhashtable.h
270
struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash);
include/linux/rhashtable.h
278
#define rht_dereference_bucket(p, tbl, hash) \
include/linux/rhashtable.h
279
rcu_dereference_protected(p, lockdep_rht_bucket_is_held(tbl, hash))
include/linux/rhashtable.h
281
#define rht_dereference_bucket_rcu(p, tbl, hash) \
include/linux/rhashtable.h
282
rcu_dereference_all_check(p, lockdep_rht_bucket_is_held(tbl, hash))
include/linux/rhashtable.h
288
const struct bucket_table *tbl, unsigned int hash)
include/linux/rhashtable.h
290
return unlikely(tbl->nest) ? rht_bucket_nested(tbl, hash) :
include/linux/rhashtable.h
291
&tbl->buckets[hash];
include/linux/rhashtable.h
295
struct bucket_table *tbl, unsigned int hash)
include/linux/rhashtable.h
297
return unlikely(tbl->nest) ? __rht_bucket_nested(tbl, hash) :
include/linux/rhashtable.h
298
&tbl->buckets[hash];
include/linux/rhashtable.h
302
struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash)
include/linux/rhashtable.h
304
return unlikely(tbl->nest) ? rht_bucket_nested_insert(ht, tbl, hash) :
include/linux/rhashtable.h
305
&tbl->buckets[hash];
include/linux/rhashtable.h
406
unsigned int hash)
include/linux/rhashtable.h
408
return __rht_ptr(rht_dereference_bucket(*bkt, tbl, hash), bkt,
include/linux/rhashtable.h
449
#define rht_for_each_from(pos, head, tbl, hash) \
include/linux/rhashtable.h
452
pos = rht_dereference_bucket((pos)->next, tbl, hash))
include/linux/rhashtable.h
460
#define rht_for_each(pos, tbl, hash) \
include/linux/rhashtable.h
461
rht_for_each_from(pos, rht_ptr(rht_bucket(tbl, hash), tbl, hash), \
include/linux/rhashtable.h
462
tbl, hash)
include/linux/rhashtable.h
473
#define rht_for_each_entry_from(tpos, pos, head, tbl, hash, member) \
include/linux/rhashtable.h
476
pos = rht_dereference_bucket((pos)->next, tbl, hash))
include/linux/rhashtable.h
486
#define rht_for_each_entry(tpos, pos, tbl, hash, member) \
include/linux/rhashtable.h
488
rht_ptr(rht_bucket(tbl, hash), tbl, hash), \
include/linux/rhashtable.h
489
tbl, hash, member)
include/linux/rhashtable.h
503
#define rht_for_each_entry_safe(tpos, pos, next, tbl, hash, member) \
include/linux/rhashtable.h
504
for (pos = rht_ptr(rht_bucket(tbl, hash), tbl, hash), \
include/linux/rhashtable.h
506
rht_dereference_bucket(pos->next, tbl, hash) : NULL; \
include/linux/rhashtable.h
510
rht_dereference_bucket(pos->next, tbl, hash) : NULL)
include/linux/rhashtable.h
523
#define rht_for_each_rcu_from(pos, head, tbl, hash) \
include/linux/rhashtable.h
539
#define rht_for_each_rcu(pos, tbl, hash) \
include/linux/rhashtable.h
541
pos = rht_ptr_rcu(rht_bucket(tbl, hash)); \
include/linux/rhashtable.h
558
#define rht_for_each_entry_rcu_from(tpos, pos, head, tbl, hash, member) \
include/linux/rhashtable.h
562
pos = rht_dereference_bucket_rcu(pos->next, tbl, hash))
include/linux/rhashtable.h
576
#define rht_for_each_entry_rcu(tpos, pos, tbl, hash, member) \
include/linux/rhashtable.h
578
rht_ptr_rcu(rht_bucket(tbl, hash)), \
include/linux/rhashtable.h
579
tbl, hash, member)
include/linux/rhashtable.h
629
unsigned int hash;
include/linux/rhashtable.h
634
hash = rht_key_hashfn(ht, tbl, key, params);
include/linux/rhashtable.h
635
bkt = rht_bucket(tbl, hash);
include/linux/rhashtable.h
637
rht_for_each_rcu_from(he, __rht_ptr_rcu(bkt, freq), tbl, hash) {
include/linux/rhashtable.h
774
unsigned int hash;
include/linux/rhashtable.h
781
hash = rht_head_hashfn(ht, tbl, obj, params);
include/linux/rhashtable.h
783
bkt = rht_bucket_insert(ht, tbl, hash);
include/linux/rhashtable.h
797
rht_for_each_from(head, rht_ptr(bkt, tbl, hash), tbl, hash) {
include/linux/rhashtable.h
820
head = rht_dereference_bucket(head->next, tbl, hash);
include/linux/rhashtable.h
842
head = rht_ptr(bkt, tbl, hash);
include/linux/skbuff.h
1048
__u32 hash;
include/linux/skbuff.h
1539
skb->hash = 0;
include/linux/skbuff.h
1551
__skb_set_hash(struct sk_buff *skb, __u32 hash, bool is_sw, bool is_l4)
include/linux/skbuff.h
1555
skb->hash = hash;
include/linux/skbuff.h
1559
skb_set_hash(struct sk_buff *skb, __u32 hash, enum pkt_hash_types type)
include/linux/skbuff.h
1562
__skb_set_hash(skb, hash, false, type == PKT_HASH_TYPE_L4);
include/linux/skbuff.h
1566
__skb_set_sw_hash(struct sk_buff *skb, __u32 hash, bool is_l4)
include/linux/skbuff.h
1568
__skb_set_hash(skb, hash, true, is_l4);
include/linux/skbuff.h
1656
return skb->hash;
include/linux/skbuff.h
1664
return skb->hash;
include/linux/skbuff.h
1671
__u32 hash = __get_hash_from_flowi6(fl6, &keys);
include/linux/skbuff.h
1673
__skb_set_sw_hash(skb, hash, flow_keys_have_l4(&keys));
include/linux/skbuff.h
1676
return skb->hash;
include/linux/skbuff.h
1684
return skb->hash;
include/linux/skbuff.h
1689
to->hash = from->hash;
include/linux/soc/airoha/airoha_offload.h
192
u32 entry_size, u32 hash,
include/linux/soc/airoha/airoha_offload.h
22
struct sk_buff *skb, u16 hash,
include/linux/soc/airoha/airoha_offload.h
41
u16 hash, bool rx_wlan)
include/linux/soc/airoha/airoha_offload.h
43
dev->ops.check_skb(dev, skb, hash, rx_wlan);
include/linux/soc/airoha/airoha_offload.h
62
struct sk_buff *skb, u16 hash,
include/linux/soc/mediatek/mtk_wed.h
207
u32 reason, u32 hash);
include/linux/stackdepot.h
55
u32 hash; /* Hash in hash table */
include/linux/stringhash.h
53
static inline unsigned int end_name_hash(unsigned long hash)
include/linux/stringhash.h
55
return hash_long(hash, 32);
include/linux/stringhash.h
74
#define hashlen_create(hash, len) ((u64)(len)<<32 | (u32)(hash))
include/linux/sunrpc/cache.h
143
struct hlist_node hash; /* on hash chain */
include/linux/sunrpc/cache.h
181
struct cache_head *key, int hash);
include/linux/sunrpc/cache.h
184
struct cache_head *new, struct cache_head *old, int hash);
include/linux/sunrpc/svcauth.h
81
struct hlist_node hash;
include/linux/tsm-mr.h
50
#define TSM_MR_(mr, hash) \
include/linux/tsm-mr.h
51
.mr_name = #mr, .mr_size = hash##_DIGEST_SIZE, \
include/linux/tsm-mr.h
52
.mr_hash = HASH_ALGO_##hash, .mr_flags = TSM_MR_F_READABLE
include/net/bluetooth/hci.h
1283
__u8 hash[16];
include/net/bluetooth/hci.h
978
__u8 hash[16];
include/net/bluetooth/mgmt.h
336
__u8 hash[16];
include/net/flow_dissector.h
286
u32 hash;
include/net/fq_impl.h
139
u32 hash = skb_get_hash(skb);
include/net/fq_impl.h
141
return reciprocal_scale(hash, fq->flows_cnt);
include/net/inet_hashtables.h
186
inet_lhash2_bucket(struct inet_hashinfo *h, u32 hash)
include/net/inet_hashtables.h
188
return &h->lhash2[hash & h->lhash2_mask];
include/net/inet_hashtables.h
193
unsigned int hash)
include/net/inet_hashtables.h
195
return &hashinfo->ehash[hash & hashinfo->ehash_mask];
include/net/inet_hashtables.h
200
unsigned int hash)
include/net/inet_hashtables.h
202
return &hashinfo->ehash_locks[hash & hashinfo->ehash_locks_mask];
include/net/inet_hashtables.h
256
u32 hash;
include/net/inet_hashtables.h
260
hash = ipv6_portaddr_hash(net, &sk->sk_v6_rcv_saddr, port);
include/net/inet_hashtables.h
263
hash = ipv4_portaddr_hash(net, sk->sk_rcv_saddr, port);
include/net/inet_hashtables.h
264
return &hinfo->bhash2[hash & (hinfo->bhash_size - 1)];
include/net/inet_hashtables.h
532
u32 hash));
include/net/ip_fib.h
577
void fib_select_multipath(struct fib_result *res, int hash,
include/net/ipv6.h
715
unsigned int hash, mix = net_hash_mix(net);
include/net/ipv6.h
718
hash = jhash_1word(0, mix);
include/net/ipv6.h
720
hash = jhash_1word((__force u32)addr6->s6_addr32[3], mix);
include/net/ipv6.h
722
hash = jhash2((__force u32 *)addr6->s6_addr32, 4, mix);
include/net/ipv6.h
724
return hash ^ port;
include/net/ipv6.h
879
u32 hash;
include/net/ipv6.h
894
hash = skb_get_hash_flowi6(skb, fl6);
include/net/ipv6.h
900
hash = rol32(hash, 16);
include/net/ipv6.h
902
flowlabel = (__force __be32)hash & IPV6_FLOWLABEL_MASK;
include/net/libeth/rx.h
308
static inline void libeth_rx_pt_set_hash(struct sk_buff *skb, u32 hash,
include/net/libeth/rx.h
311
skb_set_hash(skb, hash, pt.payload_layer);
include/net/libeth/xdp.h
1729
static inline int libeth_xdpmo_rx_hash(u32 *hash,
include/net/libeth/xdp.h
1736
*hash = val;
include/net/neighbour.h
141
struct hlist_node hash;
include/net/neighbour.h
214
__u32 (*hash)(const void *pkey,
include/net/neighbour.h
283
#define neigh_for_each_in_bucket(pos, head) hlist_for_each_entry(pos, head, hash)
include/net/neighbour.h
285
hlist_for_each_entry_rcu(pos, head, hash)
include/net/neighbour.h
287
hlist_for_each_entry_safe(pos, tmp, head, hash)
include/net/neighbour.h
306
__u32 (*hash)(const void *pkey,
include/net/neighbour.h
316
hash_val = hash(pkey, dev, nht->hash_rnd) >> (32 - nht->hash_shift);
include/net/neighbour.h
328
return ___neigh_lookup_noref(tbl, tbl->key_eq, tbl->hash, pkey, dev);
include/net/netfilter/nf_conntrack.h
134
nf_ct_tuplehash_to_ctrack(const struct nf_conntrack_tuple_hash *hash)
include/net/netfilter/nf_conntrack.h
136
return container_of(hash, struct nf_conn,
include/net/netfilter/nf_conntrack.h
137
tuplehash[hash->tuple.dst.dir]);
include/net/netfilter/nf_conntrack.h
337
nf_conntrack_get_ht(struct hlist_nulls_head **hash, unsigned int *hsize)
include/net/netfilter/nf_conntrack.h
348
*hash = hptr;
include/net/nexthop.h
318
struct nexthop *nexthop_select_path(struct nexthop *nh, int hash);
include/net/nexthop.h
386
static inline void nexthop_path_fib_result(struct fib_result *res, int hash)
include/net/nexthop.h
391
nh = nexthop_select_path(res->fi->nh, hash);
include/net/nexthop.h
540
static inline void nexthop_path_fib6_result(struct fib6_result *res, int hash)
include/net/nexthop.h
545
nh = nexthop_select_path(nh, hash);
include/net/nexthop.h
577
int hash)
include/net/nexthop.h
582
nhp = nexthop_select_path(nh, hash);
include/net/rps.h
124
u32 hash, index;
include/net/rps.h
126
hash = READ_ONCE(sk->sk_rxhash);
include/net/rps.h
127
if (!hash)
include/net/rps.h
133
index = hash & table->mask;
include/net/rps.h
150
static inline void sock_rps_record_flow_hash(__u32 hash)
include/net/rps.h
156
_sock_rps_record_flow_hash(hash);
include/net/rps.h
36
u32 hash;
include/net/rps.h
73
u32 hash)
include/net/rps.h
75
unsigned int index = hash & table->mask;
include/net/rps.h
76
u32 val = hash & ~net_hotdata.rps_cpu_mask;
include/net/rps.h
88
static inline void _sock_rps_record_flow_hash(__u32 hash)
include/net/rps.h
92
if (!hash)
include/net/rps.h
97
rps_record_sock_flow(sock_flow_table, hash);
include/net/sch_generic.h
634
struct hlist_head *hash;
include/net/sch_generic.h
648
qdisc_class_find(const struct Qdisc_class_hash *hash, u32 id)
include/net/sch_generic.h
656
h = qdisc_class_hash(id, hash->hashmask);
include/net/sch_generic.h
657
hlist_for_each_entry(cl, &hash->hash[h], hnode) {
include/net/sch_generic.h
96
struct hlist_node hash;
include/net/sock.h
1207
if (unlikely(READ_ONCE(sk->sk_rxhash) != skb->hash))
include/net/sock.h
1208
WRITE_ONCE(sk->sk_rxhash, skb->hash);
include/net/sock.h
1334
int (*hash)(struct sock *sk);
include/net/sock.h
1510
return sk->sk_prot->hash(sk);
include/net/sock.h
2433
skb->hash = txhash;
include/net/sock_reuseport.h
38
u32 hash,
include/net/tls_toe.h
66
int (*hash)(struct tls_toe_device *device, struct sock *sk);
include/net/udp.h
112
return &table->hash[udp_hashfn(net, num, table->mask)];
include/net/udp.h
120
unsigned int hash)
include/net/udp.h
122
return &table->hash2[hash & table->mask].hslot;
include/net/udp.h
131
unsigned int hash)
include/net/udp.h
175
unsigned int hash)
include/net/udp.h
177
return &table->hash4[hash & table->mask];
include/net/udp.h
329
static inline void udp_lib_hash4(struct sock *sk, u16 hash)
include/net/udp.h
337
void udp_lib_hash4(struct sock *sk, u16 hash);
include/net/udp.h
349
u32 hash;
include/net/udp.h
356
hash = skb_get_hash(skb);
include/net/udp.h
357
if (unlikely(!hash)) {
include/net/udp.h
362
hash = jhash(skb->data, 2 * ETH_ALEN,
include/net/udp.h
368
hash = udp_flow_hashrnd();
include/net/udp.h
377
hash ^= hash << 16;
include/net/udp.h
379
return htons((((u64) hash * (max - min)) >> 32) + min);
include/net/udp.h
98
struct udp_hslot *hash;
include/net/vxlan.h
565
u32 hash,
include/net/vxlan.h
570
nhc = nexthop_path_fdb_result(nh, hash >> 1);
include/net/xdp.h
648
int (*xmo_rx_hash)(const struct xdp_md *ctx, u32 *hash,
include/sound/sof/ext_manifest4.h
98
uint8_t hash[DEFAULT_HASH_SHA256_LEN];
include/trace/events/net.h
195
__field( u32, hash )
include/trace/events/net.h
222
__entry->hash = skb->hash;
include/trace/events/net.h
239
__entry->hash, __entry->l4_hash, __entry->len,
include/trace/events/tsm_mr.h
19
__string(hash, mr->mr_flags & TSM_MR_F_NOHASH ?
include/trace/events/tsm_mr.h
26
__assign_str(hash);
include/trace/events/tsm_mr.h
30
TP_printk("[%s] %s:%s", __get_str(mr), __get_str(hash),
include/trace/events/tsm_mr.h
62
__string(hash, mr->mr_flags & TSM_MR_F_NOHASH ?
include/trace/events/tsm_mr.h
69
__assign_str(hash);
include/trace/events/tsm_mr.h
73
TP_printk("[%s] %s:%s", __get_str(mr), __get_str(hash),
include/uapi/linux/bpf.h
6351
__u32 hash;
include/uapi/linux/bpf.h
6640
__u32 hash; /* A hash of the packet 4 tuples */
include/uapi/linux/bpf.h
6718
__aligned_u64 hash;
include/uapi/linux/virtio_crypto.h
377
struct virtio_crypto_hash_para hash;
include/xen/interface/io/netif.h
289
uint64_t hash = 0;
include/xen/interface/io/netif.h
303
hash ^= prefix;
include/xen/interface/io/netif.h
317
return hash >> 32;
include/xen/interface/io/netif.h
878
} hash;
init/initramfs.c
105
q = kmalloc_obj(struct hash);
init/initramfs.c
121
struct hash **p, *q;
init/initramfs.c
78
struct hash *next;
init/initramfs.c
93
struct hash **p, *q;
init/initramfs.c
94
for (p = head + hash(major, minor, ino); *p; p = &(*p)->next) {
io_uring/io-wq.c
1002
unsigned int hash;
io_uring/io-wq.c
1011
hash = __io_get_work_hash(work_flags);
io_uring/io-wq.c
1012
tail = wq->hash_tail[hash];
io_uring/io-wq.c
1013
wq->hash_tail[hash] = work;
io_uring/io-wq.c
1121
unsigned int hash = io_get_work_hash(work);
io_uring/io-wq.c
1124
if (io_wq_is_hashed(work) && work == wq->hash_tail[hash]) {
io_uring/io-wq.c
1127
if (prev_work && io_get_work_hash(prev_work) == hash)
io_uring/io-wq.c
1128
wq->hash_tail[hash] = prev_work;
io_uring/io-wq.c
1130
wq->hash_tail[hash] = NULL;
io_uring/io-wq.c
119
struct io_wq_hash *hash;
io_uring/io-wq.c
1262
refcount_inc(&data->hash->refs);
io_uring/io-wq.c
1263
wq->hash = data->hash;
io_uring/io-wq.c
1299
io_wq_put_hash(data->hash);
io_uring/io-wq.c
1372
spin_lock_irq(&wq->hash->wait.lock);
io_uring/io-wq.c
1374
spin_unlock_irq(&wq->hash->wait.lock);
io_uring/io-wq.c
1390
io_wq_put_hash(wq->hash);
io_uring/io-wq.c
435
unsigned int hash, work_flags;
io_uring/io-wq.c
448
hash = __io_get_work_hash(work_flags);
io_uring/io-wq.c
453
return hash == __io_get_work_hash(work_flags);
io_uring/io-wq.c
507
static bool io_wait_on_hash(struct io_wq *wq, unsigned int hash)
io_uring/io-wq.c
511
spin_lock_irq(&wq->hash->wait.lock);
io_uring/io-wq.c
513
__add_wait_queue(&wq->hash->wait, &wq->wait);
io_uring/io-wq.c
514
if (!test_bit(hash, &wq->hash->map)) {
io_uring/io-wq.c
520
spin_unlock_irq(&wq->hash->wait.lock);
io_uring/io-wq.c
534
unsigned int hash;
io_uring/io-wq.c
545
hash = __io_get_work_hash(work_flags);
io_uring/io-wq.c
547
tail = wq->hash_tail[hash];
io_uring/io-wq.c
550
if (!test_and_set_bit(hash, &wq->hash->map)) {
io_uring/io-wq.c
551
wq->hash_tail[hash] = NULL;
io_uring/io-wq.c
556
stall_hash = hash;
io_uring/io-wq.c
574
if (wq_has_sleeper(&wq->hash->wait))
io_uring/io-wq.c
575
wake_up(&wq->hash->wait);
io_uring/io-wq.c
643
unsigned int hash = __io_wq_is_hashed(work_flags)
io_uring/io-wq.c
665
if (hash != -1U && !next_hashed) {
io_uring/io-wq.c
667
spin_lock_irq(&wq->hash->wait.lock);
io_uring/io-wq.c
668
clear_bit(hash, &wq->hash->map);
io_uring/io-wq.c
670
spin_unlock_irq(&wq->hash->wait.lock);
io_uring/io-wq.c
671
if (wq_has_sleeper(&wq->hash->wait))
io_uring/io-wq.c
672
wake_up(&wq->hash->wait);
io_uring/io-wq.h
31
static inline void io_wq_put_hash(struct io_wq_hash *hash)
io_uring/io-wq.h
33
if (refcount_dec_and_test(&hash->refs))
io_uring/io-wq.h
34
kfree(hash);
io_uring/io-wq.h
38
struct io_wq_hash *hash;
io_uring/tctx.c
19
struct io_wq_hash *hash;
io_uring/tctx.c
24
hash = ctx->hash_map;
io_uring/tctx.c
25
if (!hash) {
io_uring/tctx.c
26
hash = kzalloc_obj(*hash);
io_uring/tctx.c
27
if (!hash) {
io_uring/tctx.c
31
refcount_set(&hash->refs, 1);
io_uring/tctx.c
32
init_waitqueue_head(&hash->wait);
io_uring/tctx.c
33
ctx->hash_map = hash;
io_uring/tctx.c
37
data.hash = hash;
kernel/audit_tree.c
1033
list_del_rcu(&chunk->hash);
kernel/audit_tree.c
199
INIT_LIST_HEAD(&chunk->hash);
kernel/audit_tree.c
240
list_add_rcu(&chunk->hash, list);
kernel/audit_tree.c
250
list_for_each_entry_rcu(p, list, hash) {
kernel/audit_tree.c
26
struct list_head hash;
kernel/audit_tree.c
324
list_replace_rcu(&old->hash, &new->hash);
kernel/audit_tree.c
369
list_del_rcu(&chunk->hash);
kernel/bpf/bloom_filter.c
43
h = hash(bloom, value, map->value_size, i);
kernel/bpf/bloom_filter.c
61
h = hash(bloom, value, map->value_size, i);
kernel/bpf/bpf_lru_list.c
348
u32 hash)
kernel/bpf/bpf_lru_list.c
350
*(u32 *)((void *)node + lru->hash_offset) = hash;
kernel/bpf/bpf_lru_list.c
397
u32 hash)
kernel/bpf/bpf_lru_list.c
418
*(u32 *)((void *)node + lru->hash_offset) = hash;
kernel/bpf/bpf_lru_list.c
429
u32 hash)
kernel/bpf/bpf_lru_list.c
449
__local_list_add_pending(lru, loc_l, cpu, node, hash);
kernel/bpf/bpf_lru_list.c
484
__local_list_add_pending(lru, loc_l, cpu, node, hash);
kernel/bpf/bpf_lru_list.c
491
struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash)
kernel/bpf/bpf_lru_list.c
494
return bpf_percpu_lru_pop_free(lru, hash);
kernel/bpf/bpf_lru_list.c
496
return bpf_common_lru_pop_free(lru, hash);
kernel/bpf/bpf_lru_list.h
77
struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash);
kernel/bpf/devmap.c
100
INIT_HLIST_HEAD(&hash[i]);
kernel/bpf/devmap.c
102
return hash;
kernel/bpf/devmap.c
95
struct hlist_head *hash;
kernel/bpf/devmap.c
97
hash = bpf_map_area_alloc((u64) entries * sizeof(*hash), numa_node);
kernel/bpf/devmap.c
98
if (hash != NULL)
kernel/bpf/hashtab.c
1074
void *value, u32 key_size, u32 hash,
kernel/bpf/hashtab.c
1145
l_new->hash = hash;
kernel/bpf/hashtab.c
1175
u32 key_size, hash;
kernel/bpf/hashtab.c
1186
hash = htab_map_hash(key, key_size, htab->hashrnd);
kernel/bpf/hashtab.c
1188
b = __select_bucket(htab, hash);
kernel/bpf/hashtab.c
1195
l_old = lookup_nulls_elem_raw(head, hash, key, key_size,
kernel/bpf/hashtab.c
1217
l_old = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1237
l_new = alloc_htab_elem(htab, key, value, key_size, hash, false, false,
kernel/bpf/hashtab.c
124
u32 hash;
kernel/bpf/hashtab.c
1282
u32 key_size, hash;
kernel/bpf/hashtab.c
1293
hash = htab_map_hash(key, key_size, htab->hashrnd);
kernel/bpf/hashtab.c
1295
b = __select_bucket(htab, hash);
kernel/bpf/hashtab.c
1303
l_new = prealloc_lru_pop(htab, key, hash);
kernel/bpf/hashtab.c
1312
l_old = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1359
u32 key_size, hash;
kernel/bpf/hashtab.c
1370
hash = htab_map_hash(key, key_size, htab->hashrnd);
kernel/bpf/hashtab.c
1372
b = __select_bucket(htab, hash);
kernel/bpf/hashtab.c
1379
l_old = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1398
hash, percpu, onallcpus, NULL, map_flags);
kernel/bpf/hashtab.c
1421
u32 key_size, hash;
kernel/bpf/hashtab.c
1432
hash = htab_map_hash(key, key_size, htab->hashrnd);
kernel/bpf/hashtab.c
1434
b = __select_bucket(htab, hash);
kernel/bpf/hashtab.c
1443
l_new = prealloc_lru_pop(htab, key, hash);
kernel/bpf/hashtab.c
1452
l_old = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1502
u32 hash, key_size;
kernel/bpf/hashtab.c
1509
hash = htab_map_hash(key, key_size, htab->hashrnd);
kernel/bpf/hashtab.c
1510
b = __select_bucket(htab, hash);
kernel/bpf/hashtab.c
1517
l = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1537
u32 hash, key_size;
kernel/bpf/hashtab.c
1544
hash = htab_map_hash(key, key_size, htab->hashrnd);
kernel/bpf/hashtab.c
1545
b = __select_bucket(htab, hash);
kernel/bpf/hashtab.c
1552
l = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1679
u32 hash, key_size;
kernel/bpf/hashtab.c
1685
hash = htab_map_hash(key, key_size, htab->hashrnd);
kernel/bpf/hashtab.c
1686
b = __select_bucket(htab, hash);
kernel/bpf/hashtab.c
1693
l = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
299
u32 hash)
kernel/bpf/hashtab.c
301
struct bpf_lru_node *node = bpf_lru_pop_free(&htab->lru, hash);
kernel/bpf/hashtab.c
347
offsetof(struct htab_elem, hash) -
kernel/bpf/hashtab.c
681
static inline struct bucket *__select_bucket(struct bpf_htab *htab, u32 hash)
kernel/bpf/hashtab.c
683
return &htab->buckets[hash & (htab->n_buckets - 1)];
kernel/bpf/hashtab.c
686
static inline struct hlist_nulls_head *select_bucket(struct bpf_htab *htab, u32 hash)
kernel/bpf/hashtab.c
688
return &__select_bucket(htab, hash)->head;
kernel/bpf/hashtab.c
692
static struct htab_elem *lookup_elem_raw(struct hlist_nulls_head *head, u32 hash,
kernel/bpf/hashtab.c
699
if (l->hash == hash && !memcmp(&l->key, key, key_size))
kernel/bpf/hashtab.c
710
u32 hash, void *key,
kernel/bpf/hashtab.c
718
if (l->hash == hash && !memcmp(&l->key, key, key_size))
kernel/bpf/hashtab.c
721
if (unlikely(get_nulls_value(n) != (hash & (n_buckets - 1))))
kernel/bpf/hashtab.c
737
u32 hash, key_size;
kernel/bpf/hashtab.c
743
hash = htab_map_hash(key, key_size, htab->hashrnd);
kernel/bpf/hashtab.c
745
head = select_bucket(htab, hash);
kernel/bpf/hashtab.c
747
l = lookup_nulls_elem_raw(head, hash, key, key_size, htab->n_buckets);
kernel/bpf/hashtab.c
870
b = __select_bucket(htab, tgt_l->hash);
kernel/bpf/hashtab.c
897
u32 hash, key_size;
kernel/bpf/hashtab.c
907
hash = htab_map_hash(key, key_size, htab->hashrnd);
kernel/bpf/hashtab.c
909
head = select_bucket(htab, hash);
kernel/bpf/hashtab.c
912
l = lookup_nulls_elem_raw(head, hash, key, key_size, htab->n_buckets);
kernel/bpf/hashtab.c
928
i = hash & (htab->n_buckets - 1);
kernel/bpf/stackmap.c
21
u32 hash;
kernel/bpf/stackmap.c
254
u32 hash, id, trace_nr, trace_len, i, max_depth;
kernel/bpf/stackmap.c
268
hash = jhash2((u32 *)ips, trace_len / sizeof(u32), 0);
kernel/bpf/stackmap.c
269
id = hash & (smap->n_buckets - 1);
kernel/bpf/stackmap.c
272
hash_matches = bucket && bucket->hash == hash;
kernel/bpf/stackmap.c
314
new_bucket->hash = hash;
kernel/bpf/syscall.c
5344
if (info.hash) {
kernel/bpf/syscall.c
5345
char __user *uhash = u64_to_user_ptr(info.hash);
kernel/bpf/trampoline.c
198
struct ftrace_hash *hash;
kernel/bpf/trampoline.c
203
hash = alloc_ftrace_hash(FTRACE_HASH_DEFAULT_BITS);
kernel/bpf/trampoline.c
204
if (!hash)
kernel/bpf/trampoline.c
208
if (!add_ftrace_hash_entry_direct(hash, ip, addr)) {
kernel/bpf/trampoline.c
209
free_ftrace_hash(hash);
kernel/bpf/trampoline.c
212
return hash;
kernel/bpf/trampoline.c
217
struct ftrace_hash *hash = hash_from_ip(tr, addr);
kernel/bpf/trampoline.c
220
if (!hash)
kernel/bpf/trampoline.c
222
err = update_ftrace_direct_add(tr->fops, hash);
kernel/bpf/trampoline.c
223
free_ftrace_hash(hash);
kernel/bpf/trampoline.c
229
struct ftrace_hash *hash = hash_from_ip(tr, addr);
kernel/bpf/trampoline.c
232
if (!hash)
kernel/bpf/trampoline.c
234
err = update_ftrace_direct_del(tr->fops, hash);
kernel/bpf/trampoline.c
235
free_ftrace_hash(hash);
kernel/bpf/trampoline.c
241
struct ftrace_hash *hash = hash_from_ip(tr, addr);
kernel/bpf/trampoline.c
244
if (!hash)
kernel/bpf/trampoline.c
246
err = update_ftrace_direct_mod(tr->fops, hash, lock_direct_mutex);
kernel/bpf/trampoline.c
247
free_ftrace_hash(hash);
kernel/events/core.c
10950
u64 hash = swevent_hash(type, event_id);
kernel/events/core.c
10952
return &hlist->heads[hash];
kernel/futex/core.c
184
u32 hash;
kernel/futex/core.c
194
hash = jhash2((void *)&key->private.address,
kernel/futex/core.c
197
return &fph->queues[hash & fph->hash_mask];
kernel/futex/core.c
418
u32 hash;
kernel/futex/core.c
428
hash = jhash2((u32 *)key,
kernel/futex/core.c
442
node = (hash >> futex_hashshift) % nr_node_ids;
kernel/futex/core.c
449
return &futex_queues[node][hash & futex_hashmask];
kernel/locking/lockdep.c
1215
unsigned long hash = hash_long((uintptr_t)key, KEYHASH_BITS);
kernel/locking/lockdep.c
1217
return lock_keys_hash + hash;
kernel/locking/lockdep.c
543
u32 hash;
kernel/locking/lockdep.c
557
return t1->hash == t2->hash && t1->nr_entries == t2->nr_entries &&
kernel/locking/lockdep.c
566
u32 hash;
kernel/locking/lockdep.c
589
hash = jhash(trace->entries, trace->nr_entries *
kernel/locking/lockdep.c
591
trace->hash = hash;
kernel/locking/lockdep.c
592
hash_head = stack_trace_hash + (hash & (STACK_TRACE_HASH_SIZE - 1));
kernel/locking/qspinlock_paravirt.h
203
#define for_each_hash_entry(he, offset, hash) \
kernel/locking/qspinlock_paravirt.h
204
for (hash &= ~(PV_HE_PER_LINE - 1), he = &pv_lock_hash[hash], offset = 0; \
kernel/locking/qspinlock_paravirt.h
206
offset++, he = &pv_lock_hash[(hash + offset) & ((1 << pv_lock_hash_bits) - 1)])
kernel/locking/qspinlock_paravirt.h
210
unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits);
kernel/locking/qspinlock_paravirt.h
214
for_each_hash_entry(he, offset, hash) {
kernel/locking/qspinlock_paravirt.h
238
unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits);
kernel/locking/qspinlock_paravirt.h
242
for_each_hash_entry(he, offset, hash) {
kernel/module/main.c
3606
int hash = hash_ptr(cookie, IDEM_HASH_BITS);
kernel/module/main.c
3607
struct hlist_head *head = idem_hash + hash;
kernel/module/main.c
3623
hlist_add_head(&u->entry, idem_hash + hash);
kernel/module/main.c
3638
int hash = hash_ptr(cookie, IDEM_HASH_BITS);
kernel/module/main.c
3639
struct hlist_head *head = idem_hash + hash;
kernel/module_signature.c
34
ms->hash != 0 ||
kernel/trace/ftrace.c
1156
ftrace_hash_key(struct ftrace_hash *hash, unsigned long ip)
kernel/trace/ftrace.c
1158
if (hash->size_bits > 0)
kernel/trace/ftrace.c
1159
return hash_long(ip, hash->size_bits);
kernel/trace/ftrace.c
1166
__ftrace_lookup_ip(struct ftrace_hash *hash, unsigned long ip)
kernel/trace/ftrace.c
1172
key = ftrace_hash_key(hash, ip);
kernel/trace/ftrace.c
1173
hhd = &hash->buckets[key];
kernel/trace/ftrace.c
1193
ftrace_lookup_ip(struct ftrace_hash *hash, unsigned long ip)
kernel/trace/ftrace.c
1195
if (ftrace_hash_empty(hash))
kernel/trace/ftrace.c
1198
return __ftrace_lookup_ip(hash, ip);
kernel/trace/ftrace.c
1201
static void __add_hash_entry(struct ftrace_hash *hash,
kernel/trace/ftrace.c
1207
key = ftrace_hash_key(hash, entry->ip);
kernel/trace/ftrace.c
1208
hhd = &hash->buckets[key];
kernel/trace/ftrace.c
1210
hash->count++;
kernel/trace/ftrace.c
1214
add_ftrace_hash_entry_direct(struct ftrace_hash *hash, unsigned long ip, unsigned long direct)
kernel/trace/ftrace.c
1224
__add_hash_entry(hash, entry);
kernel/trace/ftrace.c
1230
add_hash_entry(struct ftrace_hash *hash, unsigned long ip)
kernel/trace/ftrace.c
1232
return add_ftrace_hash_entry_direct(hash, ip, 0);
kernel/trace/ftrace.c
1236
free_hash_entry(struct ftrace_hash *hash,
kernel/trace/ftrace.c
1241
hash->count--;
kernel/trace/ftrace.c
1245
remove_hash_entry(struct ftrace_hash *hash,
kernel/trace/ftrace.c
1249
hash->count--;
kernel/trace/ftrace.c
1252
static void ftrace_hash_clear(struct ftrace_hash *hash)
kernel/trace/ftrace.c
1257
int size = 1 << hash->size_bits;
kernel/trace/ftrace.c
1260
if (!hash->count)
kernel/trace/ftrace.c
1264
hhd = &hash->buckets[i];
kernel/trace/ftrace.c
1266
free_hash_entry(hash, entry);
kernel/trace/ftrace.c
1268
FTRACE_WARN_ON(hash->count);
kernel/trace/ftrace.c
1293
void free_ftrace_hash(struct ftrace_hash *hash)
kernel/trace/ftrace.c
1295
if (!hash || hash == EMPTY_HASH)
kernel/trace/ftrace.c
1297
ftrace_hash_clear(hash);
kernel/trace/ftrace.c
1298
kfree(hash->buckets);
kernel/trace/ftrace.c
1299
kfree(hash);
kernel/trace/ftrace.c
1304
struct ftrace_hash *hash;
kernel/trace/ftrace.c
1306
hash = container_of(rcu, struct ftrace_hash, rcu);
kernel/trace/ftrace.c
1307
free_ftrace_hash(hash);
kernel/trace/ftrace.c
1310
static void free_ftrace_hash_rcu(struct ftrace_hash *hash)
kernel/trace/ftrace.c
1312
if (!hash || hash == EMPTY_HASH)
kernel/trace/ftrace.c
1314
call_rcu(&hash->rcu, __free_ftrace_hash_rcu);
kernel/trace/ftrace.c
1335
struct ftrace_hash *hash;
kernel/trace/ftrace.c
1338
hash = kzalloc_obj(*hash);
kernel/trace/ftrace.c
1339
if (!hash)
kernel/trace/ftrace.c
1343
hash->buckets = kzalloc_objs(*hash->buckets, size);
kernel/trace/ftrace.c
1345
if (!hash->buckets) {
kernel/trace/ftrace.c
1346
kfree(hash);
kernel/trace/ftrace.c
1350
hash->size_bits = size_bits;
kernel/trace/ftrace.c
1352
return hash;
kernel/trace/ftrace.c
1386
alloc_and_copy_ftrace_hash(int size_bits, struct ftrace_hash *hash)
kernel/trace/ftrace.c
1397
if (hash)
kernel/trace/ftrace.c
1398
new_hash->flags = hash->flags;
kernel/trace/ftrace.c
1401
if (ftrace_hash_empty(hash))
kernel/trace/ftrace.c
1404
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
1406
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
1412
FTRACE_WARN_ON(new_hash->count != hash->count);
kernel/trace/ftrace.c
1541
struct ftrace_ops_hash *hash)
kernel/trace/ftrace.c
1549
return (ftrace_hash_empty(hash->filter_hash) ||
kernel/trace/ftrace.c
1550
__ftrace_lookup_ip(hash->filter_hash, ip)) &&
kernel/trace/ftrace.c
1551
(ftrace_hash_empty(hash->notrace_hash) ||
kernel/trace/ftrace.c
1552
!__ftrace_lookup_ip(hash->notrace_hash, ip));
kernel/trace/ftrace.c
1570
struct ftrace_ops_hash hash;
kernel/trace/ftrace.c
1583
rcu_assign_pointer(hash.filter_hash, ops->func_hash->filter_hash);
kernel/trace/ftrace.c
1584
rcu_assign_pointer(hash.notrace_hash, ops->func_hash->notrace_hash);
kernel/trace/ftrace.c
1586
if (hash_contains_ip(ip, &hash))
kernel/trace/ftrace.c
1769
struct ftrace_hash *hash;
kernel/trace/ftrace.c
1785
hash = ops->func_hash->filter_hash;
kernel/trace/ftrace.c
1787
if (ftrace_hash_empty(hash))
kernel/trace/ftrace.c
1806
in_hash = !!ftrace_lookup_ip(hash, rec->ip);
kernel/trace/ftrace.c
1912
if (!all && count == hash->count)
kernel/trace/ftrace.c
2119
struct ftrace_hash *hash = ops->func_hash->filter_hash;
kernel/trace/ftrace.c
2121
if (ftrace_hash_empty(hash))
kernel/trace/ftrace.c
2122
hash = NULL;
kernel/trace/ftrace.c
2124
return __ftrace_hash_update_ipmodify(ops, EMPTY_HASH, hash, false);
kernel/trace/ftrace.c
2130
struct ftrace_hash *hash = ops->func_hash->filter_hash;
kernel/trace/ftrace.c
2132
if (ftrace_hash_empty(hash))
kernel/trace/ftrace.c
2133
hash = NULL;
kernel/trace/ftrace.c
2135
__ftrace_hash_update_ipmodify(ops, hash, EMPTY_HASH, false);
kernel/trace/ftrace.c
3270
static int append_hash(struct ftrace_hash **hash, struct ftrace_hash *new_hash,
kernel/trace/ftrace.c
3277
if (*hash) {
kernel/trace/ftrace.c
3279
if (ftrace_hash_empty(*hash))
kernel/trace/ftrace.c
3282
*hash = alloc_ftrace_hash(size_bits);
kernel/trace/ftrace.c
3283
if (!*hash)
kernel/trace/ftrace.c
3289
free_ftrace_hash(*hash);
kernel/trace/ftrace.c
3290
*hash = EMPTY_HASH;
kernel/trace/ftrace.c
3298
if (!__ftrace_lookup_ip(*hash, entry->ip) &&
kernel/trace/ftrace.c
3299
add_hash_entry(*hash, entry->ip) == NULL)
kernel/trace/ftrace.c
3309
static void remove_hash(struct ftrace_hash *hash, struct ftrace_hash *notrace_hash)
kernel/trace/ftrace.c
3320
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
3322
hlist_for_each_entry_safe(entry, tmp, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
3325
remove_hash_entry(hash, entry);
kernel/trace/ftrace.c
3337
static int intersect_hash(struct ftrace_hash **hash, struct ftrace_hash *new_hash1,
kernel/trace/ftrace.c
3349
free_ftrace_hash(*hash);
kernel/trace/ftrace.c
3350
*hash = EMPTY_HASH;
kernel/trace/ftrace.c
3359
add_hash_entry(*hash, entry->ip) == NULL)
kernel/trace/ftrace.c
3364
if (ftrace_hash_empty(*hash)) {
kernel/trace/ftrace.c
3365
free_ftrace_hash(*hash);
kernel/trace/ftrace.c
3366
*hash = EMPTY_HASH;
kernel/trace/ftrace.c
3402
struct ftrace_hash *hash,
kernel/trace/ftrace.c
3412
ret = ftrace_hash_move(ops, enable, orig_hash, hash);
kernel/trace/ftrace.c
3733
struct ftrace_hash *hash)
kernel/trace/ftrace.c
3748
*orig_subhash = __ftrace_hash_move(hash);
kernel/trace/ftrace.c
3958
struct ftrace_hash *hash;
kernel/trace/ftrace.c
3973
struct ftrace_hash *hash;
kernel/trace/ftrace.c
3997
hash = iter->probe->ops.func_hash->filter_hash;
kernel/trace/ftrace.c
4003
if (!hash || hash == EMPTY_HASH)
kernel/trace/ftrace.c
4006
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
4014
hash = iter->probe->ops.func_hash->filter_hash;
kernel/trace/ftrace.c
4015
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
4019
hhd = &hash->buckets[iter->pidx];
kernel/trace/ftrace.c
4192
!ftrace_lookup_ip(iter->hash, rec->ip)) ||
kernel/trace/ftrace.c
4274
ftrace_hash_empty(iter->hash)) {
kernel/trace/ftrace.c
433
struct hlist_head *hash;
kernel/trace/ftrace.c
4676
struct ftrace_hash *hash;
kernel/trace/ftrace.c
4703
hash = ops->func_hash->notrace_hash;
kernel/trace/ftrace.c
4706
hash = ops->func_hash->filter_hash;
kernel/trace/ftrace.c
4716
iter->hash = alloc_ftrace_hash(size_bits);
kernel/trace/ftrace.c
4719
iter->hash = alloc_and_copy_ftrace_hash(size_bits, hash);
kernel/trace/ftrace.c
4722
if (hash)
kernel/trace/ftrace.c
4723
iter->hash = alloc_and_copy_ftrace_hash(hash->size_bits, hash);
kernel/trace/ftrace.c
4725
iter->hash = EMPTY_HASH;
kernel/trace/ftrace.c
4728
if (!iter->hash) {
kernel/trace/ftrace.c
4744
free_ftrace_hash(iter->hash);
kernel/trace/ftrace.c
4837
enter_record(struct ftrace_hash *hash, struct dyn_ftrace *rec, int clear_filter)
kernel/trace/ftrace.c
4842
entry = ftrace_lookup_ip(hash, rec->ip);
kernel/trace/ftrace.c
4848
free_hash_entry(hash, entry);
kernel/trace/ftrace.c
4853
if (add_hash_entry(hash, rec->ip) == NULL)
kernel/trace/ftrace.c
4860
add_rec_by_index(struct ftrace_hash *hash, struct ftrace_glob *func_g,
kernel/trace/ftrace.c
4878
enter_record(hash, rec, clear_filter);
kernel/trace/ftrace.c
4946
match_records(struct ftrace_hash *hash, char *func, int len, char *mod)
kernel/trace/ftrace.c
4976
return add_rec_by_index(hash, &func_g, clear_filter);
kernel/trace/ftrace.c
4984
ret = enter_record(hash, rec, clear_filter);
kernel/trace/ftrace.c
4996
ftrace_match_records(struct ftrace_hash *hash, char *buff, int len)
kernel/trace/ftrace.c
4998
return match_records(hash, buff, len, NULL);
kernel/trace/ftrace.c
5034
struct ftrace_hash *hash,
kernel/trace/ftrace.c
5038
return ftrace_hash_move_and_update_subops(ops, orig_hash, hash);
kernel/trace/ftrace.c
5057
return ftrace_hash_move_and_update_subops(subops, orig_hash, hash);
kernel/trace/ftrace.c
5063
return __ftrace_hash_move_and_update_ops(ops, orig_hash, hash, enable);
kernel/trace/ftrace.c
5202
ftrace_mod_callback(struct trace_array *tr, struct ftrace_hash *hash,
kernel/trace/ftrace.c
5223
ret = match_records(hash, func, strlen(func), module);
kernel/trace/ftrace.c
5273
struct ftrace_hash hash; /* Must be first! */
kernel/trace/ftrace.c
5283
struct ftrace_hash *hash;
kernel/trace/ftrace.c
5290
hash = alloc_ftrace_hash(FTRACE_HASH_DEFAULT_BITS);
kernel/trace/ftrace.c
5291
return (struct ftrace_func_mapper *)hash;
kernel/trace/ftrace.c
5311
entry = ftrace_lookup_ip(&mapper->hash, ip);
kernel/trace/ftrace.c
5333
entry = ftrace_lookup_ip(&mapper->hash, ip);
kernel/trace/ftrace.c
5344
__add_hash_entry(&mapper->hash, &map->entry);
kernel/trace/ftrace.c
5366
entry = ftrace_lookup_ip(&mapper->hash, ip);
kernel/trace/ftrace.c
5373
remove_hash_entry(&mapper->hash, entry);
kernel/trace/ftrace.c
5398
if (free_func && mapper->hash.count) {
kernel/trace/ftrace.c
5399
size = 1 << mapper->hash.size_bits;
kernel/trace/ftrace.c
5401
hhd = &mapper->hash.buckets[i];
kernel/trace/ftrace.c
5409
free_ftrace_hash(&mapper->hash);
kernel/trace/ftrace.c
5454
struct ftrace_hash *hash;
kernel/trace/ftrace.c
5501
hash = alloc_and_copy_ftrace_hash(FTRACE_HASH_DEFAULT_BITS, old_hash);
kernel/trace/ftrace.c
5503
if (!hash) {
kernel/trace/ftrace.c
5508
ret = ftrace_match_records(hash, glob, strlen(glob));
kernel/trace/ftrace.c
5517
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
5519
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
5552
hash, 1);
kernel/trace/ftrace.c
5569
free_ftrace_hash(hash);
kernel/trace/ftrace.c
5581
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
5600
struct ftrace_hash *hash = NULL;
kernel/trace/ftrace.c
5654
hash = alloc_and_copy_ftrace_hash(FTRACE_HASH_DEFAULT_BITS, old_hash);
kernel/trace/ftrace.c
5655
if (!hash)
kernel/trace/ftrace.c
5660
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
5662
hlist_for_each_entry_safe(entry, tmp, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
5671
remove_hash_entry(hash, entry);
kernel/trace/ftrace.c
5688
if (ftrace_hash_empty(hash))
kernel/trace/ftrace.c
5692
hash, 1);
kernel/trace/ftrace.c
5695
if (ftrace_enabled && !ftrace_hash_empty(hash))
kernel/trace/ftrace.c
5710
free_ftrace_hash(hash);
kernel/trace/ftrace.c
5773
struct ftrace_hash *hash = iter->hash;
kernel/trace/ftrace.c
5782
ret = ftrace_match_records(hash, func, len);
kernel/trace/ftrace.c
5798
return p->func(tr, hash, func, command, next, enable);
kernel/trace/ftrace.c
5856
__ftrace_match_addr(struct ftrace_hash *hash, unsigned long ip, int remove)
kernel/trace/ftrace.c
5865
entry = ftrace_lookup_ip(hash, ip);
kernel/trace/ftrace.c
5868
free_hash_entry(hash, entry);
kernel/trace/ftrace.c
5870
} else if (__ftrace_lookup_ip(hash, ip) != NULL) {
kernel/trace/ftrace.c
5875
entry = add_hash_entry(hash, ip);
kernel/trace/ftrace.c
5880
ftrace_match_addr(struct ftrace_hash *hash, unsigned long *ips,
kernel/trace/ftrace.c
5887
err = __ftrace_match_addr(hash, ips[i], remove);
kernel/trace/ftrace.c
5905
struct ftrace_hash *hash;
kernel/trace/ftrace.c
5919
hash = alloc_ftrace_hash(FTRACE_HASH_DEFAULT_BITS);
kernel/trace/ftrace.c
5921
hash = alloc_and_copy_ftrace_hash(FTRACE_HASH_DEFAULT_BITS, *orig_hash);
kernel/trace/ftrace.c
5923
if (!hash) {
kernel/trace/ftrace.c
5928
if (buf && !match_records(hash, buf, len, mod)) {
kernel/trace/ftrace.c
5941
ret = ftrace_match_addr(hash, ips, cnt, remove);
kernel/trace/ftrace.c
5947
ret = ftrace_hash_move_and_update_ops(ops, orig_hash, hash, enable);
kernel/trace/ftrace.c
5953
free_ftrace_hash(hash);
kernel/trace/ftrace.c
5989
static void remove_direct_functions_hash(struct ftrace_hash *hash, unsigned long addr)
kernel/trace/ftrace.c
5994
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
5996
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
6016
struct ftrace_hash *hash = ops->func_hash->filter_hash;
kernel/trace/ftrace.c
6018
remove_direct_functions_hash(hash, addr);
kernel/trace/ftrace.c
6049
struct ftrace_hash *hash, *new_hash = NULL, *free_hash = NULL;
kernel/trace/ftrace.c
6060
hash = ops->func_hash->filter_hash;
kernel/trace/ftrace.c
6061
if (ftrace_hash_empty(hash))
kernel/trace/ftrace.c
6067
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
6069
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
6078
size = hash->count + direct_functions->count;
kernel/trace/ftrace.c
6098
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
6100
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
6175
struct ftrace_hash *hash = ops->func_hash->filter_hash;
kernel/trace/ftrace.c
6200
err = __ftrace_hash_update_ipmodify(ops, hash, hash, true);
kernel/trace/ftrace.c
6210
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
6212
hlist_for_each_entry(iter, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
626
memset(stat->hash, 0,
kernel/trace/ftrace.c
6291
static unsigned long hash_count(struct ftrace_hash *hash)
kernel/trace/ftrace.c
6293
return hash ? hash->count : 0;
kernel/trace/ftrace.c
6341
int update_ftrace_direct_add(struct ftrace_ops *ops, struct ftrace_hash *hash)
kernel/trace/ftrace.c
6352
if (!hash_count(hash))
kernel/trace/ftrace.c
6358
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
6360
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
6378
new_filter_hash = hash_add(old_filter_hash, hash);
kernel/trace/ftrace.c
6382
new_direct_functions = hash_add(direct_functions, hash);
kernel/trace/ftrace.c
6477
int update_ftrace_direct_del(struct ftrace_ops *ops, struct ftrace_hash *hash)
kernel/trace/ftrace.c
6488
if (!hash_count(hash))
kernel/trace/ftrace.c
6505
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
6507
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
6515
new_filter_hash = hash_sub(old_filter_hash, hash);
kernel/trace/ftrace.c
6519
new_direct_functions = hash_sub(direct_functions, hash);
kernel/trace/ftrace.c
6578
int update_ftrace_direct_mod(struct ftrace_ops *ops, struct ftrace_hash *hash, bool do_direct_lock)
kernel/trace/ftrace.c
6589
if (!hash_count(hash))
kernel/trace/ftrace.c
6611
tmp_ops.func_hash->filter_hash = hash;
kernel/trace/ftrace.c
6632
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
6634
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
6860
static int ftrace_graph_set_hash(struct ftrace_hash *hash, char *buffer);
kernel/trace/ftrace.c
6888
struct ftrace_hash *hash;
kernel/trace/ftrace.c
6890
hash = alloc_ftrace_hash(FTRACE_HASH_DEFAULT_BITS);
kernel/trace/ftrace.c
6891
if (MEM_FAIL(!hash, "Failed to allocate hash\n"))
kernel/trace/ftrace.c
6897
ret = ftrace_graph_set_hash(hash, func);
kernel/trace/ftrace.c
6904
ftrace_graph_hash = hash;
kernel/trace/ftrace.c
6906
ftrace_graph_notrace_hash = hash;
kernel/trace/ftrace.c
693
if (stat->hash) {
kernel/trace/ftrace.c
6978
iter->hash->flags &= ~FTRACE_HASH_FL_MOD;
kernel/trace/ftrace.c
6980
iter->hash->flags |= FTRACE_HASH_FL_MOD;
kernel/trace/ftrace.c
6987
iter->hash, filter_hash);
kernel/trace/ftrace.c
6992
free_ftrace_hash(iter->hash);
kernel/trace/ftrace.c
705
stat->hash = kzalloc_objs(struct hlist_head, size);
kernel/trace/ftrace.c
7059
struct ftrace_hash *hash;
kernel/trace/ftrace.c
707
if (!stat->hash)
kernel/trace/ftrace.c
7076
if (*pos >= fgd->hash->count)
kernel/trace/ftrace.c
7088
for (i = idx; i < 1 << fgd->hash->size_bits; i++) {
kernel/trace/ftrace.c
7089
head = &fgd->hash->buckets[i];
kernel/trace/ftrace.c
7113
fgd->hash = rcu_dereference_protected(ftrace_graph_hash,
kernel/trace/ftrace.c
7116
fgd->hash = rcu_dereference_protected(ftrace_graph_notrace_hash,
kernel/trace/ftrace.c
712
kfree(stat->hash);
kernel/trace/ftrace.c
7120
if (ftrace_hash_empty(fgd->hash) && !*pos)
kernel/trace/ftrace.c
713
stat->hash = NULL;
kernel/trace/ftrace.c
7183
fgd->hash);
kernel/trace/ftrace.c
7214
fgd->hash = NULL;
kernel/trace/ftrace.c
7234
fgd->hash = rcu_dereference_protected(ftrace_graph_hash,
kernel/trace/ftrace.c
7262
fgd->hash = rcu_dereference_protected(ftrace_graph_notrace_hash,
kernel/trace/ftrace.c
7346
ftrace_graph_set_hash(struct ftrace_hash *hash, char *buffer)
kernel/trace/ftrace.c
7372
entry = ftrace_lookup_ip(hash, rec->ip);
kernel/trace/ftrace.c
7379
if (add_hash_entry(hash, rec->ip) == NULL)
kernel/trace/ftrace.c
7383
free_hash_entry(hash, entry);
kernel/trace/ftrace.c
743
hhd = &stat->hash[key];
kernel/trace/ftrace.c
762
hlist_add_head_rcu(&rec->node, &stat->hash[key]);
kernel/trace/ftrace.c
7809
clear_mod_from_hash(struct ftrace_page *pg, struct ftrace_hash *hash)
kernel/trace/ftrace.c
7815
if (ftrace_hash_empty(hash))
kernel/trace/ftrace.c
7820
entry = __ftrace_lookup_ip(hash, rec->ip);
kernel/trace/ftrace.c
814
if (!stat->hash || !ftrace_profile_enabled)
kernel/trace/ftrace.c
8210
clear_func_from_hash(struct ftrace_init_func *func, struct ftrace_hash *hash)
kernel/trace/ftrace.c
8214
entry = ftrace_lookup_ip(hash, func->ip);
kernel/trace/ftrace.c
880
if (!stat->hash || !ftrace_profile_enabled)
kernel/trace/ftrace.c
9048
struct ftrace_hash *hash;
kernel/trace/ftrace.c
9057
hash = ops->func_hash->filter_hash;
kernel/trace/ftrace.c
9058
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
9060
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/ftrace.c
9097
struct ftrace_hash *hash;
kernel/trace/ftrace.c
9106
hash = ops->func_hash->filter_hash;
kernel/trace/ftrace.c
9107
size = 1 << hash->size_bits;
kernel/trace/ftrace.c
9109
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
kernel/trace/trace.c
8511
ftrace_trace_snapshot_callback(struct trace_array *tr, struct ftrace_hash *hash,
kernel/trace/trace.h
1054
struct ftrace_hash *hash;
kernel/trace/trace.h
1064
hash = rcu_dereference_protected(ftrace_graph_hash, !preemptible());
kernel/trace/trace.h
1066
if (ftrace_hash_empty(hash)) {
kernel/trace/trace.h
1071
if (ftrace_lookup_ip(hash, addr)) {
kernel/trace/trace.h
1184
struct ftrace_hash *hash,
kernel/trace/trace.h
943
ftrace_lookup_ip(struct ftrace_hash *hash, unsigned long ip);
kernel/trace/trace.h
945
static __always_inline bool ftrace_hash_empty(struct ftrace_hash *hash)
kernel/trace/trace.h
947
return !hash || !(hash->count || (hash->flags & FTRACE_HASH_FL_MOD));
kernel/trace/trace_events.c
4303
event_enable_func(struct trace_array *tr, struct ftrace_hash *hash,
kernel/trace/trace_functions.c
855
struct ftrace_hash *hash, char *glob,
kernel/trace/trace_functions.c
892
ftrace_trace_onoff_callback(struct trace_array *tr, struct ftrace_hash *hash,
kernel/trace/trace_functions.c
906
return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd,
kernel/trace/trace_functions.c
911
ftrace_stacktrace_callback(struct trace_array *tr, struct ftrace_hash *hash,
kernel/trace/trace_functions.c
921
return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd,
kernel/trace/trace_functions.c
926
ftrace_dump_callback(struct trace_array *tr, struct ftrace_hash *hash,
kernel/trace/trace_functions.c
937
return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd,
kernel/trace/trace_functions.c
942
ftrace_cpudump_callback(struct trace_array *tr, struct ftrace_hash *hash,
kernel/trace/trace_functions.c
953
return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd,
kernel/workqueue.c
4766
u32 hash = 0;
kernel/workqueue.c
4768
hash = jhash_1word(attrs->nice, hash);
kernel/workqueue.c
4769
hash = jhash_1word(attrs->affn_strict, hash);
kernel/workqueue.c
4770
hash = jhash(cpumask_bits(attrs->__pod_cpumask),
kernel/workqueue.c
4771
BITS_TO_LONGS(nr_cpumask_bits) * sizeof(long), hash);
kernel/workqueue.c
4773
hash = jhash(cpumask_bits(attrs->cpumask),
kernel/workqueue.c
4774
BITS_TO_LONGS(nr_cpumask_bits) * sizeof(long), hash);
kernel/workqueue.c
4775
return hash;
kernel/workqueue.c
5099
u32 hash = wqattrs_hash(attrs);
kernel/workqueue.c
5106
hash_for_each_possible(unbound_pool_hash, pool, hash_node, hash) {
kernel/workqueue.c
5138
hash_add(unbound_pool_hash, &pool->hash_node, hash);
lib/crypto/arm/nh.h
14
__le64 hash[NH_NUM_PASSES]);
lib/crypto/arm/nh.h
17
__le64 hash[NH_NUM_PASSES])
lib/crypto/arm/nh.h
22
nh_neon(key, message, message_len, hash);
lib/crypto/arm64/nh.h
15
__le64 hash[NH_NUM_PASSES]);
lib/crypto/arm64/nh.h
18
__le64 hash[NH_NUM_PASSES])
lib/crypto/arm64/nh.h
23
nh_neon(key, message, message_len, hash);
lib/crypto/mips/sha1.h
25
u64 *hash = (u64 *)&state->h[0];
lib/crypto/mips/sha1.h
31
write_octeon_64bit_hash_dword(hash[0], 0);
lib/crypto/mips/sha1.h
32
write_octeon_64bit_hash_dword(hash[1], 1);
lib/crypto/mips/sha1.h
39
u64 *hash = (u64 *)&state->h[0];
lib/crypto/mips/sha1.h
45
hash[0] = read_octeon_64bit_hash_dword(0);
lib/crypto/mips/sha1.h
46
hash[1] = read_octeon_64bit_hash_dword(1);
lib/crypto/nh.c
24
__le64 hash[NH_NUM_PASSES])
lib/crypto/nh.c
31
__le64 hash[NH_NUM_PASSES])
lib/crypto/nh.c
35
if (nh_arch(key, message, message_len, hash))
lib/crypto/nh.c
60
hash[0] = cpu_to_le64(sums[0]);
lib/crypto/nh.c
61
hash[1] = cpu_to_le64(sums[1]);
lib/crypto/nh.c
62
hash[2] = cpu_to_le64(sums[2]);
lib/crypto/nh.c
63
hash[3] = cpu_to_le64(sums[3]);
lib/crypto/sha3.c
394
u8 hash[SHA3_256_DIGEST_SIZE];
lib/crypto/sha3.c
396
sha3_256(fips_test_data, sizeof(fips_test_data), hash);
lib/crypto/sha3.c
397
if (memcmp(fips_test_sha3_256_value, hash, sizeof(hash)) != 0)
lib/crypto/tests/blake2b_kunit.c
106
u8 hash[BLAKE2B_HASH_SIZE];
lib/crypto/tests/blake2b_kunit.c
109
blake2b(NULL, 0, test_buf, data_len, hash, out_len);
lib/crypto/tests/blake2b_kunit.c
111
KUNIT_ASSERT_MEMEQ(test, hash, guarded_hash, out_len);
lib/crypto/tests/blake2b_kunit.c
46
u8 *hash = key + BLAKE2B_KEY_SIZE;
lib/crypto/tests/blake2b_kunit.c
55
blake2b(key, key_len, data, data_len, hash, out_len);
lib/crypto/tests/blake2b_kunit.c
56
blake2b_update(&main_ctx, hash, out_len);
lib/crypto/tests/blake2s_kunit.c
106
u8 hash[BLAKE2S_HASH_SIZE];
lib/crypto/tests/blake2s_kunit.c
109
blake2s(NULL, 0, test_buf, data_len, hash, out_len);
lib/crypto/tests/blake2s_kunit.c
111
KUNIT_ASSERT_MEMEQ(test, hash, guarded_hash, out_len);
lib/crypto/tests/blake2s_kunit.c
46
u8 *hash = key + BLAKE2S_KEY_SIZE;
lib/crypto/tests/blake2s_kunit.c
55
blake2s(key, key_len, data, data_len, hash, out_len);
lib/crypto/tests/blake2s_kunit.c
56
blake2s_update(&main_ctx, hash, out_len);
lib/crypto/tests/hash-test-template.h
131
u8 hash[HASH_SIZE];
lib/crypto/tests/hash-test-template.h
137
HASH(test_buf, len, hash);
lib/crypto/tests/hash-test-template.h
138
HASH_UPDATE(&ctx, hash, HASH_SIZE);
lib/crypto/tests/hash-test-template.h
140
HASH_FINAL(&ctx, hash);
lib/crypto/tests/hash-test-template.h
141
KUNIT_ASSERT_MEMEQ(test, hash, hash_testvec_consolidated, HASH_SIZE);
lib/crypto/tests/hash-test-template.h
209
u8 hash[HASH_SIZE];
lib/crypto/tests/hash-test-template.h
212
HASH(buf_end - len, len, hash);
lib/crypto/tests/hash-test-template.h
215
HASH_FINAL(&ctx, hash);
lib/crypto/tests/hash-test-template.h
226
HASH_FINAL(guarded_ctx, hash);
lib/crypto/tests/hash-test-template.h
238
u8 hash[HASH_SIZE];
lib/crypto/tests/hash-test-template.h
249
HASH(&test_buf[offset], len, hash);
lib/crypto/tests/hash-test-template.h
252
test, hash, ovl_hash, HASH_SIZE,
lib/crypto/tests/hash-test-template.h
257
HASH(&test_buf[offset], len, hash);
lib/crypto/tests/hash-test-template.h
262
test, hash, ovl_hash, HASH_SIZE,
lib/crypto/tests/hash-test-template.h
267
HASH(&test_buf[offset], len, hash);
lib/crypto/tests/hash-test-template.h
273
test, hash, ovl_hash, HASH_SIZE,
lib/crypto/tests/hash-test-template.h
543
u8 hash[HASH_SIZE];
lib/crypto/tests/hash-test-template.h
550
HASH(test_buf, TEST_BUF_LEN, hash);
lib/crypto/tests/hash-test-template.h
562
HASH(test_buf, len, hash);
lib/crypto/tests/nh_kunit.c
12
__le64 hash[NH_NUM_PASSES];
lib/crypto/tests/nh_kunit.c
18
nh(key, nh_test_msg, 16, hash);
lib/crypto/tests/nh_kunit.c
19
KUNIT_ASSERT_MEMEQ(test, hash, nh_test_val16, NH_HASH_BYTES);
lib/crypto/tests/nh_kunit.c
21
nh(key, nh_test_msg, 96, hash);
lib/crypto/tests/nh_kunit.c
22
KUNIT_ASSERT_MEMEQ(test, hash, nh_test_val96, NH_HASH_BYTES);
lib/crypto/tests/nh_kunit.c
24
nh(key, nh_test_msg, 256, hash);
lib/crypto/tests/nh_kunit.c
25
KUNIT_ASSERT_MEMEQ(test, hash, nh_test_val256, NH_HASH_BYTES);
lib/crypto/tests/nh_kunit.c
27
nh(key, nh_test_msg, 1024, hash);
lib/crypto/tests/nh_kunit.c
28
KUNIT_ASSERT_MEMEQ(test, hash, nh_test_val1024, NH_HASH_BYTES);
lib/crypto/tests/polyval_kunit.c
54
u8 hash[POLYVAL_BLOCK_SIZE];
lib/crypto/tests/polyval_kunit.c
58
polyval(&key, data, sizeof(data), hash);
lib/crypto/tests/polyval_kunit.c
59
KUNIT_ASSERT_MEMEQ(test, hash, expected_hash, sizeof(hash));
lib/crypto/tests/polyval_kunit.c
71
u8 hash[POLYVAL_BLOCK_SIZE];
lib/crypto/tests/polyval_kunit.c
79
polyval(&key, test_buf, len, hash);
lib/crypto/tests/polyval_kunit.c
80
polyval_update(&hashofhashes_ctx, hash, sizeof(hash));
lib/crypto/tests/polyval_kunit.c
82
polyval_final(&hashofhashes_ctx, hash);
lib/crypto/tests/polyval_kunit.c
83
KUNIT_ASSERT_MEMEQ(test, hash, polyval_allones_hashofhashes,
lib/crypto/tests/polyval_kunit.c
84
sizeof(hash));
lib/crypto/tests/sha256_kunit.c
136
u8 hash[SHA256_DIGEST_SIZE];
lib/crypto/tests/sha256_kunit.c
143
sha256_finup_2x(&ctx, test_buf, test_buf, data_len, hash, hash);
lib/crypto/tests/sha256_kunit.c
148
KUNIT_ASSERT_MEMEQ(test, hash, expected_hash,
lib/crypto/x86/nh.h
15
__le64 hash[NH_NUM_PASSES]);
lib/crypto/x86/nh.h
17
__le64 hash[NH_NUM_PASSES]);
lib/crypto/x86/nh.h
20
__le64 hash[NH_NUM_PASSES])
lib/crypto/x86/nh.h
26
nh_avx2(key, message, message_len, hash);
lib/crypto/x86/nh.h
28
nh_sse2(key, message, message_len, hash);
lib/debugobjects.c
605
unsigned long hash;
lib/debugobjects.c
607
hash = hash_long((addr >> ODEBUG_CHUNK_SHIFT), ODEBUG_HASH_BITS);
lib/debugobjects.c
608
return &obj_hash[hash];
lib/digsig.c
199
unsigned char hash[SHA1_DIGEST_SIZE];
lib/digsig.c
232
sha1_final(&ctx, hash);
lib/digsig.c
236
hash, sizeof(hash));
lib/oid_registry.c
31
unsigned i, j, k, hash;
lib/oid_registry.c
35
hash = datasize - 1;
lib/oid_registry.c
38
hash += octets[i] * 33;
lib/oid_registry.c
39
hash = (hash >> 24) ^ (hash >> 16) ^ (hash >> 8) ^ hash;
lib/oid_registry.c
40
hash &= 0xff;
lib/oid_registry.c
51
xhash = oid_search_table[j].hash;
lib/oid_registry.c
52
if (xhash > hash) {
lib/oid_registry.c
56
if (xhash < hash) {
lib/rhashtable.c
1190
const struct bucket_table *tbl, unsigned int hash)
lib/rhashtable.c
1193
unsigned int index = hash & ((1 << tbl->nest) - 1);
lib/rhashtable.c
1195
unsigned int subhash = hash;
lib/rhashtable.c
1199
ntbl = rht_dereference_bucket_rcu(ntbl[index].table, tbl, hash);
lib/rhashtable.c
1205
tbl, hash);
lib/rhashtable.c
1219
const struct bucket_table *tbl, unsigned int hash)
lib/rhashtable.c
1225
return __rht_bucket_nested(tbl, hash) ?: &rhnull;
lib/rhashtable.c
1230
struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash)
lib/rhashtable.c
1233
unsigned int index = hash & ((1 << tbl->nest) - 1);
lib/rhashtable.c
1238
hash >>= tbl->nest;
lib/rhashtable.c
1243
index = hash & ((1 << shift) - 1);
lib/rhashtable.c
1245
hash >>= shift;
lib/rhashtable.c
1253
return &ntbl[hash].bucket;
lib/rhashtable.c
498
struct bucket_table *tbl, unsigned int hash,
lib/rhashtable.c
510
rht_for_each_from(head, rht_ptr(bkt, tbl, hash), tbl, hash) {
lib/rhashtable.c
53
int lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash)
lib/rhashtable.c
530
head = rht_dereference_bucket(head->next, tbl, hash);
lib/rhashtable.c
549
struct bucket_table *tbl, unsigned int hash, struct rhash_head *obj,
lib/rhashtable.c
574
head = rht_ptr(bkt, tbl, hash);
lib/rhashtable.c
59
return bit_spin_is_locked(0, (unsigned long *)&tbl->buckets[hash]);
lib/rhashtable.c
599
unsigned int hash;
lib/rhashtable.c
606
hash = rht_head_hashfn(ht, tbl, obj, ht->p);
lib/rhashtable.c
609
bkt = rht_bucket_var(tbl, hash);
lib/rhashtable.c
611
bkt = rht_bucket_insert(ht, tbl, hash);
lib/rhashtable.c
620
hash, key, obj);
lib/rhashtable.c
622
hash, obj, data);
lib/stackdepot.c
434
depot_alloc_stack(unsigned long *entries, unsigned int nr_entries, u32 hash, depot_flags_t flags, void **prealloc)
lib/stackdepot.c
468
stack->hash = hash;
lib/stackdepot.c
594
u32 hash, depot_flags_t flags)
lib/stackdepot.c
610
if (stack->hash != hash || stack->size != size)
lib/stackdepot.c
654
u32 hash;
lib/stackdepot.c
672
hash = hash_stack(entries, nr_entries);
lib/stackdepot.c
673
bucket = &stack_table[hash & stack_hash_mask];
lib/stackdepot.c
676
found = find_stack(bucket, entries, nr_entries, hash, depot_flags);
lib/stackdepot.c
703
found = find_stack(bucket, entries, nr_entries, hash, depot_flags);
lib/stackdepot.c
706
depot_alloc_stack(entries, nr_entries, hash, depot_flags, &prealloc);
lib/test_bpf.c
14751
skb->hash = SKB_HASH;
lib/tests/hashtable_test.c
102
hash_add(hash, &a.node, a.key);
lib/tests/hashtable_test.c
106
hash_add(hash, &b.node, b.key);
lib/tests/hashtable_test.c
109
hash_for_each_possible(hash, x, node, b.key) {
lib/tests/hashtable_test.c
120
KUNIT_EXPECT_TRUE(test, hash_empty(hash));
lib/tests/hashtable_test.c
128
DEFINE_HASHTABLE(hash, 3);
lib/tests/hashtable_test.c
135
hash_add(hash, &entries[i].node, entries[i].key);
lib/tests/hashtable_test.c
139
hash_for_each(hash, bkt, x, node) {
lib/tests/hashtable_test.c
158
DEFINE_HASHTABLE(hash, 3);
lib/tests/hashtable_test.c
165
hash_add(hash, &entries[i].node, entries[i].key);
lib/tests/hashtable_test.c
169
hash_for_each_safe(hash, bkt, tmp, x, node) {
lib/tests/hashtable_test.c
191
DEFINE_HASHTABLE(hash, 5);
lib/tests/hashtable_test.c
198
hash_add(hash, &entries[i].node, entries[i].key);
lib/tests/hashtable_test.c
205
hash_add(hash, &entries[3].node, entries[3].key);
lib/tests/hashtable_test.c
208
hash_for_each_possible(hash, x, node, 0) {
lib/tests/hashtable_test.c
220
hash_for_each(hash, bkt, y, node) {
lib/tests/hashtable_test.c
246
DEFINE_HASHTABLE(hash, 5);
lib/tests/hashtable_test.c
253
hash_add(hash, &entries[i].node, entries[i].key);
lib/tests/hashtable_test.c
260
hash_add(hash, &entries[3].node, entries[3].key);
lib/tests/hashtable_test.c
263
hash_for_each_possible_safe(hash, x, tmp, node, 0) {
lib/tests/hashtable_test.c
278
hash_for_each(hash, bkt, y, node) {
lib/tests/hashtable_test.c
37
DEFINE_HASHTABLE(hash, 1);
lib/tests/hashtable_test.c
39
KUNIT_EXPECT_TRUE(test, hash_empty(hash));
lib/tests/hashtable_test.c
43
hash_add(hash, &a.node, a.key);
lib/tests/hashtable_test.c
46
KUNIT_EXPECT_FALSE(test, hash_empty(hash));
lib/tests/hashtable_test.c
52
DEFINE_HASHTABLE(hash, 4);
lib/tests/hashtable_test.c
56
hash_add(hash, &a.node, a.key);
lib/tests/hashtable_test.c
59
hash_add(hash, &b.node, b.key);
lib/tests/hashtable_test.c
69
DEFINE_HASHTABLE(hash, 3);
lib/tests/hashtable_test.c
74
hash_add(hash, &a.node, a.key);
lib/tests/hashtable_test.c
78
hash_add(hash, &b.node, b.key);
lib/tests/hashtable_test.c
80
hash_for_each(hash, bkt, x, node) {
lib/tests/hashtable_test.c
98
DEFINE_HASHTABLE(hash, 6);
lib/tests/siphash_kunit.c
113
#define chk(hash, vector, fmt...) \
lib/tests/siphash_kunit.c
114
KUNIT_EXPECT_EQ_MSG(test, hash, vector, fmt)
lib/xz/xz_dec_stream.c
114
struct xz_dec_hash hash;
lib/xz/xz_dec_stream.c
258
s->block.hash.unpadded += s->block_header.size
lib/xz/xz_dec_stream.c
262
s->block.hash.unpadded += check_sizes[s->check_type];
lib/xz/xz_dec_stream.c
265
s->block.hash.unpadded += 4;
lib/xz/xz_dec_stream.c
268
s->block.hash.uncompressed += s->block.uncompressed;
lib/xz/xz_dec_stream.c
269
s->block.hash.crc32 = xz_crc32(
lib/xz/xz_dec_stream.c
270
(const uint8_t *)&s->block.hash,
lib/xz/xz_dec_stream.c
271
sizeof(s->block.hash), s->block.hash.crc32);
lib/xz/xz_dec_stream.c
322
s->index.hash.unpadded += s->vli;
lib/xz/xz_dec_stream.c
327
s->index.hash.uncompressed += s->vli;
lib/xz/xz_dec_stream.c
328
s->index.hash.crc32 = xz_crc32(
lib/xz/xz_dec_stream.c
329
(const uint8_t *)&s->index.hash,
lib/xz/xz_dec_stream.c
330
sizeof(s->index.hash),
lib/xz/xz_dec_stream.c
331
s->index.hash.crc32);
lib/xz/xz_dec_stream.c
695
if (!memeq(&s->block.hash, &s->index.hash,
lib/xz/xz_dec_stream.c
696
sizeof(s->block.hash)))
lib/xz/xz_dec_stream.c
92
struct xz_dec_hash hash;
lib/zstd/compress/zstd_compress_internal.h
1007
MEM_STATIC U64 ZSTD_rollingHash_rotate(U64 hash, BYTE toRemove, BYTE toAdd, U64 primePower)
lib/zstd/compress/zstd_compress_internal.h
1009
hash -= (toRemove + ZSTD_ROLL_HASH_CHAR_OFFSET) * primePower;
lib/zstd/compress/zstd_compress_internal.h
1010
hash *= prime8bytes;
lib/zstd/compress/zstd_compress_internal.h
1011
hash += toAdd + ZSTD_ROLL_HASH_CHAR_OFFSET;
lib/zstd/compress/zstd_compress_internal.h
1012
return hash;
lib/zstd/compress/zstd_compress_internal.h
1480
size_t const hash = hashAndTag >> ZSTD_SHORT_CACHE_TAG_BITS;
lib/zstd/compress/zstd_compress_internal.h
1483
hashTable[hash] = (index << ZSTD_SHORT_CACHE_TAG_BITS) | tag;
lib/zstd/compress/zstd_compress_internal.h
328
U32 hash;
lib/zstd/compress/zstd_compress_internal.h
976
static U64 ZSTD_rollingHash_append(U64 hash, void const* buf, size_t size)
lib/zstd/compress/zstd_compress_internal.h
981
hash *= prime8bytes;
lib/zstd/compress/zstd_compress_internal.h
982
hash += istart[pos] + ZSTD_ROLL_HASH_CHAR_OFFSET;
lib/zstd/compress/zstd_compress_internal.h
984
return hash;
lib/zstd/compress/zstd_fast.c
82
size_t const hash = ZSTD_hashPtr(ip + p, hBits, mls);
lib/zstd/compress/zstd_fast.c
83
if (hashTable[hash] == 0) { /* not yet filled */
lib/zstd/compress/zstd_fast.c
84
hashTable[hash] = curr + p;
lib/zstd/compress/zstd_lazy.c
1172
U32 hash;
lib/zstd/compress/zstd_lazy.c
1208
hash = ZSTD_row_nextCachedHash(hashCache, hashTable, tagTable, base, curr, hashLog, rowLog, mls, hashSalt);
lib/zstd/compress/zstd_lazy.c
1213
hash = (U32)ZSTD_hashPtrSalted(ip, hashLog + ZSTD_ROW_HASH_TAG_BITS, mls, hashSalt);
lib/zstd/compress/zstd_lazy.c
1216
ms->hashSaltEntropy += hash; /* collect salt entropy */
lib/zstd/compress/zstd_lazy.c
1219
U32 const relRow = (hash >> ZSTD_ROW_HASH_TAG_BITS) << rowLog;
lib/zstd/compress/zstd_lazy.c
1220
U32 const tag = hash & ZSTD_ROW_HASH_TAG_MASK;
lib/zstd/compress/zstd_lazy.c
849
U32 const hash = (U32)ZSTD_hashPtrSalted(base + idx, hashLog + ZSTD_ROW_HASH_TAG_BITS, mls, ms->hashSalt);
lib/zstd/compress/zstd_lazy.c
850
U32 const row = (hash >> ZSTD_ROW_HASH_TAG_BITS) << rowLog;
lib/zstd/compress/zstd_lazy.c
852
ms->hashCache[idx & ZSTD_ROW_HASH_CACHE_MASK] = hash;
lib/zstd/compress/zstd_lazy.c
875
{ U32 const hash = cache[idx & ZSTD_ROW_HASH_CACHE_MASK];
lib/zstd/compress/zstd_lazy.c
877
return hash;
lib/zstd/compress/zstd_lazy.c
898
U32 const hash = useCache ? ZSTD_row_nextCachedHash(ms->hashCache, hashTable, tagTable, base, updateStartIdx, hashLog, rowLog, mls, ms->hashSalt)
lib/zstd/compress/zstd_lazy.c
900
U32 const relRow = (hash >> ZSTD_ROW_HASH_TAG_BITS) << rowLog;
lib/zstd/compress/zstd_lazy.c
905
assert(hash == ZSTD_hashPtrSalted(base + updateStartIdx, hashLog + ZSTD_ROW_HASH_TAG_BITS, mls, ms->hashSalt));
lib/zstd/compress/zstd_lazy.c
906
tagRow[pos] = hash & ZSTD_ROW_HASH_TAG_MASK;
lib/zstd/compress/zstd_ldm.c
102
U64 hash, mask;
lib/zstd/compress/zstd_ldm.c
104
hash = state->rolling;
lib/zstd/compress/zstd_ldm.c
109
hash = (hash << 1) + ZSTD_ldm_gearTab[data[n] & 0xff]; \
lib/zstd/compress/zstd_ldm.c
111
if (UNLIKELY((hash & mask) == 0)) { \
lib/zstd/compress/zstd_ldm.c
132
state->rolling = hash;
lib/zstd/compress/zstd_ldm.c
188
const ldmState_t* ldmState, size_t hash, U32 const bucketSizeLog)
lib/zstd/compress/zstd_ldm.c
190
return ldmState->hashTable + (hash << bucketSizeLog);
lib/zstd/compress/zstd_ldm.c
196
size_t const hash, const ldmEntry_t entry,
lib/zstd/compress/zstd_ldm.c
199
BYTE* const pOffset = ldmState->bucketOffsets + hash;
lib/zstd/compress/zstd_ldm.c
202
*(ZSTD_ldm_getBucket(ldmState, hash, bucketSizeLog) + offset) = entry;
lib/zstd/compress/zstd_ldm.c
313
U32 const hash = (U32)(xxhash & (((U32)1 << hBits) - 1));
lib/zstd/compress/zstd_ldm.c
318
ZSTD_ldm_insertEntry(ldmState, hash, entry, params->bucketSizeLog);
lib/zstd/compress/zstd_ldm.c
393
U32 const hash = (U32)(xxhash & (((U32)1 << hBits) - 1));
lib/zstd/compress/zstd_ldm.c
396
candidates[n].hash = hash;
lib/zstd/compress/zstd_ldm.c
398
candidates[n].bucket = ZSTD_ldm_getBucket(ldmState, hash, params->bucketSizeLog);
lib/zstd/compress/zstd_ldm.c
408
U32 const hash = candidates[n].hash;
lib/zstd/compress/zstd_ldm.c
421
ZSTD_ldm_insertEntry(ldmState, hash, newEntry, params->bucketSizeLog);
lib/zstd/compress/zstd_ldm.c
468
ZSTD_ldm_insertEntry(ldmState, hash, newEntry, params->bucketSizeLog);
lib/zstd/compress/zstd_ldm.c
489
ZSTD_ldm_insertEntry(ldmState, hash, newEntry, params->bucketSizeLog);
lib/zstd/compress/zstd_ldm.c
69
U64 hash = state->rolling;
lib/zstd/compress/zstd_ldm.c
73
hash = (hash << 1) + ZSTD_ldm_gearTab[data[n] & 0xff]; \
lib/zstd/decompress/zstd_decompress.c
91
const U64 hash = xxh64(&dictID, sizeof(U32), 0);
lib/zstd/decompress/zstd_decompress.c
93
return hash & (hashSet->ddictPtrTableSize - 1);
mm/hugetlb.c
5545
u32 hash;
mm/hugetlb.c
5558
hash = hugetlb_fault_mutex_hash(mapping, idx);
mm/hugetlb.c
5560
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/hugetlb.c
5564
mutex_lock(&hugetlb_fault_mutex_table[hash]);
mm/hugetlb.c
5692
u32 hash;
mm/hugetlb.c
5700
hash = hugetlb_fault_mutex_hash(mapping, vmf->pgoff);
mm/hugetlb.c
5701
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/hugetlb.c
5725
u32 hash = hugetlb_fault_mutex_hash(mapping, vmf->pgoff);
mm/hugetlb.c
5933
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/hugetlb.c
5952
u32 hash;
mm/hugetlb.c
5957
hash = jhash2((u32 *)&key, sizeof(key)/(sizeof(u32)), 0);
mm/hugetlb.c
5959
return hash & (num_fault_mutexes - 1);
mm/hugetlb.c
5976
u32 hash;
mm/hugetlb.c
6002
hash = hugetlb_fault_mutex_hash(mapping, vmf.pgoff);
mm/hugetlb.c
6003
mutex_lock(&hugetlb_fault_mutex_table[hash]);
mm/hugetlb.c
6014
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/hugetlb.c
6057
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/hugetlb.c
6097
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/hugetlb.c
6145
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/khugepaged.c
1411
hash_del(&slot->hash);
mm/khugepaged.c
472
hash_del(&slot->hash);
mm/ksm.c
1237
hash_del(&mm_slot->slot.hash);
mm/ksm.c
2737
hash_del(&mm_slot->slot.hash);
mm/ksm.c
3081
hash_del(&slot->hash);
mm/memfd.c
119
hash = hugetlb_fault_mutex_hash(memfd->f_mapping, idx);
mm/memfd.c
120
mutex_lock(&hugetlb_fault_mutex_table[hash]);
mm/memfd.c
126
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/memfd.c
98
u32 hash;
mm/mm_slot.h
16
struct hlist_node hash;
mm/mm_slot.h
40
hash_for_each_possible(_hashtable, tmp_slot, hash, (unsigned long)_mm) \
mm/mm_slot.h
52
hash_add(_hashtable, &_mm_slot->hash, (unsigned long)_mm); \
mm/userfaultfd.c
508
u32 hash;
mm/userfaultfd.c
578
hash = hugetlb_fault_mutex_hash(mapping, idx);
mm/userfaultfd.c
579
mutex_lock(&hugetlb_fault_mutex_table[hash]);
mm/userfaultfd.c
586
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/userfaultfd.c
596
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/userfaultfd.c
605
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
mm/vmscan.c
2799
u32 hash = hash_ptr(item, BLOOM_FILTER_SHIFT * 2);
mm/vmscan.c
2803
key[0] = hash & (BIT(BLOOM_FILTER_SHIFT) - 1);
mm/vmscan.c
2804
key[1] = hash >> BLOOM_FILTER_SHIFT;
net/9p/error.c
180
u32 hash;
net/9p/error.c
185
hash = jhash(c->name, c->namelen, 0);
net/9p/error.c
187
hash_add(hash_errmap, &c->list, hash);
net/9p/error.c
205
u32 hash;
net/9p/error.c
209
hash = jhash(errstr, len, 0);
net/9p/error.c
210
hash_for_each_possible(hash_errmap, c, list, hash) {
net/appletalk/aarp.c
429
int hash = sa->s_node % (AARP_HASH_SIZE - 1);
net/appletalk/aarp.c
434
a = __aarp_find_entry(proxies[hash], dev, sa);
net/appletalk/aarp.c
445
int hash = sa->s_node % (AARP_HASH_SIZE - 1);
net/appletalk/aarp.c
446
struct aarp_entry *a = __aarp_find_entry(proxies[hash], dev, sa);
net/appletalk/aarp.c
468
int hash, retval = -EPROTONOSUPPORT;
net/appletalk/aarp.c
498
hash = sa->s_node % (AARP_HASH_SIZE - 1);
net/appletalk/aarp.c
499
entry->next = proxies[hash];
net/appletalk/aarp.c
500
proxies[hash] = entry;
net/appletalk/aarp.c
534
int hash;
net/appletalk/aarp.c
590
hash = sa->s_node % (AARP_HASH_SIZE - 1);
net/appletalk/aarp.c
600
a = __aarp_find_entry(resolved[hash], dev, sa);
net/appletalk/aarp.c
610
a = __aarp_find_entry(unresolved[hash], dev, sa);
net/appletalk/aarp.c
628
a->next = unresolved[hash];
net/appletalk/aarp.c
631
unresolved[hash] = a;
net/appletalk/aarp.c
673
int hash)
net/appletalk/aarp.c
683
a->next = resolved[hash];
net/appletalk/aarp.c
684
resolved[hash] = a;
net/appletalk/aarp.c
704
int hash, ret = 0;
net/appletalk/aarp.c
730
hash = ea->pa_src_node % (AARP_HASH_SIZE - 1);
net/appletalk/aarp.c
753
a = __aarp_find_entry(proxies[hash], dev, &da);
net/appletalk/aarp.c
770
a = __aarp_find_entry(unresolved[hash], dev, &sa);
net/appletalk/aarp.c
776
__aarp_resolved(&unresolved[hash], a, hash);
net/batman-adv/bat_iv_ogm.c
1977
struct batadv_hashtable *hash = bat_priv->orig_hash;
net/batman-adv/bat_iv_ogm.c
1984
while (bucket < hash->size) {
net/batman-adv/bat_iv_ogm.c
1985
head = &hash->table[bucket];
net/batman-adv/bat_iv_ogm.c
753
struct batadv_hashtable *hash = bat_priv->orig_hash;
net/batman-adv/bat_iv_ogm.c
761
for (i = 0; i < hash->size; i++) {
net/batman-adv/bat_iv_ogm.c
762
head = &hash->table[i];
net/batman-adv/bat_v.c
423
struct batadv_hashtable *hash = bat_priv->orig_hash;
net/batman-adv/bat_v.c
430
while (bucket < hash->size) {
net/batman-adv/bat_v.c
431
head = &hash->table[bucket];
net/batman-adv/bridge_loop_avoidance.c
1224
struct batadv_hashtable *hash;
net/batman-adv/bridge_loop_avoidance.c
1228
hash = bat_priv->bla.backbone_hash;
net/batman-adv/bridge_loop_avoidance.c
1229
if (!hash)
net/batman-adv/bridge_loop_avoidance.c
1232
for (i = 0; i < hash->size; i++) {
net/batman-adv/bridge_loop_avoidance.c
1233
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
1234
list_lock = &hash->list_locks[i];
net/batman-adv/bridge_loop_avoidance.c
1279
struct batadv_hashtable *hash;
net/batman-adv/bridge_loop_avoidance.c
1282
hash = bat_priv->bla.claim_hash;
net/batman-adv/bridge_loop_avoidance.c
1283
if (!hash)
net/batman-adv/bridge_loop_avoidance.c
1286
for (i = 0; i < hash->size; i++) {
net/batman-adv/bridge_loop_avoidance.c
1287
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
1334
struct batadv_hashtable *hash;
net/batman-adv/bridge_loop_avoidance.c
1352
hash = bat_priv->bla.backbone_hash;
net/batman-adv/bridge_loop_avoidance.c
1353
if (!hash)
net/batman-adv/bridge_loop_avoidance.c
1356
for (i = 0; i < hash->size; i++) {
net/batman-adv/bridge_loop_avoidance.c
1357
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
1432
struct batadv_hashtable *hash;
net/batman-adv/bridge_loop_avoidance.c
1466
hash = bat_priv->bla.backbone_hash;
net/batman-adv/bridge_loop_avoidance.c
1467
if (!hash)
net/batman-adv/bridge_loop_avoidance.c
1470
for (i = 0; i < hash->size; i++) {
net/batman-adv/bridge_loop_avoidance.c
1471
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
1734
struct batadv_hashtable *hash = bat_priv->bla.backbone_hash;
net/batman-adv/bridge_loop_avoidance.c
1742
if (!hash)
net/batman-adv/bridge_loop_avoidance.c
1745
for (i = 0; i < hash->size; i++) {
net/batman-adv/bridge_loop_avoidance.c
1746
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
2195
struct batadv_hashtable *hash, unsigned int bucket,
net/batman-adv/bridge_loop_avoidance.c
2202
spin_lock_bh(&hash->list_locks[bucket]);
net/batman-adv/bridge_loop_avoidance.c
2203
cb->seq = atomic_read(&hash->generation) << 1 | 1;
net/batman-adv/bridge_loop_avoidance.c
2205
hlist_for_each_entry(claim, &hash->table[bucket], hash_entry) {
net/batman-adv/bridge_loop_avoidance.c
221
struct batadv_hashtable *hash = bat_priv->bla.claim_hash;
net/batman-adv/bridge_loop_avoidance.c
2219
spin_unlock_bh(&hash->list_locks[bucket]);
net/batman-adv/bridge_loop_avoidance.c
2235
struct batadv_hashtable *hash;
net/batman-adv/bridge_loop_avoidance.c
2246
hash = bat_priv->bla.claim_hash;
net/batman-adv/bridge_loop_avoidance.c
2254
while (bucket < hash->size) {
net/batman-adv/bridge_loop_avoidance.c
2256
hash, bucket, &idx))
net/batman-adv/bridge_loop_avoidance.c
227
if (!hash)
net/batman-adv/bridge_loop_avoidance.c
230
index = batadv_choose_claim(data, hash->size);
net/batman-adv/bridge_loop_avoidance.c
231
head = &hash->table[index];
net/batman-adv/bridge_loop_avoidance.c
2356
struct batadv_hashtable *hash,
net/batman-adv/bridge_loop_avoidance.c
2363
spin_lock_bh(&hash->list_locks[bucket]);
net/batman-adv/bridge_loop_avoidance.c
2364
cb->seq = atomic_read(&hash->generation) << 1 | 1;
net/batman-adv/bridge_loop_avoidance.c
2366
hlist_for_each_entry(backbone_gw, &hash->table[bucket], hash_entry) {
net/batman-adv/bridge_loop_avoidance.c
2380
spin_unlock_bh(&hash->list_locks[bucket]);
net/batman-adv/bridge_loop_avoidance.c
2396
struct batadv_hashtable *hash;
net/batman-adv/bridge_loop_avoidance.c
2407
hash = bat_priv->bla.backbone_hash;
net/batman-adv/bridge_loop_avoidance.c
2415
while (bucket < hash->size) {
net/batman-adv/bridge_loop_avoidance.c
2417
hash, bucket, &idx))
net/batman-adv/bridge_loop_avoidance.c
261
struct batadv_hashtable *hash = bat_priv->bla.backbone_hash;
net/batman-adv/bridge_loop_avoidance.c
267
if (!hash)
net/batman-adv/bridge_loop_avoidance.c
273
index = batadv_choose_backbone_gw(&search_entry, hash->size);
net/batman-adv/bridge_loop_avoidance.c
274
head = &hash->table[index];
net/batman-adv/bridge_loop_avoidance.c
300
struct batadv_hashtable *hash;
net/batman-adv/bridge_loop_avoidance.c
307
hash = backbone_gw->bat_priv->bla.claim_hash;
net/batman-adv/bridge_loop_avoidance.c
308
if (!hash)
net/batman-adv/bridge_loop_avoidance.c
311
for (i = 0; i < hash->size; i++) {
net/batman-adv/bridge_loop_avoidance.c
312
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
313
list_lock = &hash->list_locks[i];
net/batman-adv/bridge_loop_avoidance.c
595
struct batadv_hashtable *hash;
net/batman-adv/bridge_loop_avoidance.c
610
hash = bat_priv->bla.claim_hash;
net/batman-adv/bridge_loop_avoidance.c
611
for (i = 0; i < hash->size; i++) {
net/batman-adv/bridge_loop_avoidance.c
612
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
69
u32 hash = 0;
net/batman-adv/bridge_loop_avoidance.c
71
hash = jhash(&claim->addr, sizeof(claim->addr), hash);
net/batman-adv/bridge_loop_avoidance.c
72
hash = jhash(&claim->vid, sizeof(claim->vid), hash);
net/batman-adv/bridge_loop_avoidance.c
74
return hash % size;
net/batman-adv/bridge_loop_avoidance.c
87
u32 hash = 0;
net/batman-adv/bridge_loop_avoidance.c
90
hash = jhash(&gw->orig, sizeof(gw->orig), hash);
net/batman-adv/bridge_loop_avoidance.c
91
hash = jhash(&gw->vid, sizeof(gw->vid), hash);
net/batman-adv/bridge_loop_avoidance.c
93
return hash % size;
net/batman-adv/distributed-arp-table.c
166
if (!bat_priv->dat.hash)
net/batman-adv/distributed-arp-table.c
169
for (i = 0; i < bat_priv->dat.hash->size; i++) {
net/batman-adv/distributed-arp-table.c
170
head = &bat_priv->dat.hash->table[i];
net/batman-adv/distributed-arp-table.c
171
list_lock = &bat_priv->dat.hash->list_locks[i];
net/batman-adv/distributed-arp-table.c
287
u32 hash = 0;
net/batman-adv/distributed-arp-table.c
295
hash += key[i];
net/batman-adv/distributed-arp-table.c
296
hash += (hash << 10);
net/batman-adv/distributed-arp-table.c
297
hash ^= (hash >> 6);
net/batman-adv/distributed-arp-table.c
303
hash += key[i];
net/batman-adv/distributed-arp-table.c
304
hash += (hash << 10);
net/batman-adv/distributed-arp-table.c
305
hash ^= (hash >> 6);
net/batman-adv/distributed-arp-table.c
308
hash += (hash << 3);
net/batman-adv/distributed-arp-table.c
309
hash ^= (hash >> 11);
net/batman-adv/distributed-arp-table.c
310
hash += (hash << 15);
net/batman-adv/distributed-arp-table.c
312
return hash % size;
net/batman-adv/distributed-arp-table.c
330
struct batadv_hashtable *hash = bat_priv->dat.hash;
net/batman-adv/distributed-arp-table.c
333
if (!hash)
net/batman-adv/distributed-arp-table.c
339
index = batadv_hash_dat(&to_find, hash->size);
net/batman-adv/distributed-arp-table.c
340
head = &hash->table[index];
net/batman-adv/distributed-arp-table.c
395
hash_added = batadv_hash_add(bat_priv->dat.hash, batadv_compare_dat,
net/batman-adv/distributed-arp-table.c
566
struct batadv_hashtable *hash = bat_priv->orig_hash;
net/batman-adv/distributed-arp-table.c
578
for (i = 0; i < hash->size; i++) {
net/batman-adv/distributed-arp-table.c
579
head = &hash->table[i];
net/batman-adv/distributed-arp-table.c
792
if (!bat_priv->dat.hash)
net/batman-adv/distributed-arp-table.c
797
batadv_hash_destroy(bat_priv->dat.hash);
net/batman-adv/distributed-arp-table.c
799
bat_priv->dat.hash = NULL;
net/batman-adv/distributed-arp-table.c
810
if (bat_priv->dat.hash)
net/batman-adv/distributed-arp-table.c
813
bat_priv->dat.hash = batadv_hash_new(1024);
net/batman-adv/distributed-arp-table.c
815
if (!bat_priv->dat.hash)
net/batman-adv/distributed-arp-table.c
899
struct batadv_hashtable *hash, unsigned int bucket,
net/batman-adv/distributed-arp-table.c
905
spin_lock_bh(&hash->list_locks[bucket]);
net/batman-adv/distributed-arp-table.c
906
cb->seq = atomic_read(&hash->generation) << 1 | 1;
net/batman-adv/distributed-arp-table.c
908
hlist_for_each_entry(dat_entry, &hash->table[bucket], hash_entry) {
net/batman-adv/distributed-arp-table.c
913
spin_unlock_bh(&hash->list_locks[bucket]);
net/batman-adv/distributed-arp-table.c
922
spin_unlock_bh(&hash->list_locks[bucket]);
net/batman-adv/distributed-arp-table.c
939
struct batadv_hashtable *hash;
net/batman-adv/distributed-arp-table.c
950
hash = bat_priv->dat.hash;
net/batman-adv/distributed-arp-table.c
958
while (bucket < hash->size) {
net/batman-adv/distributed-arp-table.c
959
if (batadv_dat_cache_dump_bucket(msg, portid, cb, hash, bucket,
net/batman-adv/hash.c
15
static void batadv_hash_init(struct batadv_hashtable *hash)
net/batman-adv/hash.c
19
for (i = 0; i < hash->size; i++) {
net/batman-adv/hash.c
20
INIT_HLIST_HEAD(&hash->table[i]);
net/batman-adv/hash.c
21
spin_lock_init(&hash->list_locks[i]);
net/batman-adv/hash.c
24
atomic_set(&hash->generation, 0);
net/batman-adv/hash.c
31
void batadv_hash_destroy(struct batadv_hashtable *hash)
net/batman-adv/hash.c
33
kfree(hash->list_locks);
net/batman-adv/hash.c
34
kfree(hash->table);
net/batman-adv/hash.c
35
kfree(hash);
net/batman-adv/hash.c
46
struct batadv_hashtable *hash;
net/batman-adv/hash.c
48
hash = kmalloc_obj(*hash, GFP_ATOMIC);
net/batman-adv/hash.c
49
if (!hash)
net/batman-adv/hash.c
52
hash->table = kmalloc_objs(*hash->table, size, GFP_ATOMIC);
net/batman-adv/hash.c
53
if (!hash->table)
net/batman-adv/hash.c
56
hash->list_locks = kmalloc_objs(*hash->list_locks, size, GFP_ATOMIC);
net/batman-adv/hash.c
57
if (!hash->list_locks)
net/batman-adv/hash.c
60
hash->size = size;
net/batman-adv/hash.c
61
batadv_hash_init(hash);
net/batman-adv/hash.c
62
return hash;
net/batman-adv/hash.c
65
kfree(hash->table);
net/batman-adv/hash.c
67
kfree(hash);
net/batman-adv/hash.c
76
void batadv_hash_set_lock_class(struct batadv_hashtable *hash,
net/batman-adv/hash.c
81
for (i = 0; i < hash->size; i++)
net/batman-adv/hash.c
82
lockdep_set_class(&hash->list_locks[i], key);
net/batman-adv/hash.h
106
atomic_inc(&hash->generation);
net/batman-adv/hash.h
129
static inline void *batadv_hash_remove(struct batadv_hashtable *hash,
net/batman-adv/hash.h
139
index = choose(data, hash->size);
net/batman-adv/hash.h
140
head = &hash->table[index];
net/batman-adv/hash.h
142
spin_lock_bh(&hash->list_locks[index]);
net/batman-adv/hash.h
149
atomic_inc(&hash->generation);
net/batman-adv/hash.h
152
spin_unlock_bh(&hash->list_locks[index]);
net/batman-adv/hash.h
58
void batadv_hash_set_lock_class(struct batadv_hashtable *hash,
net/batman-adv/hash.h
62
void batadv_hash_destroy(struct batadv_hashtable *hash);
net/batman-adv/hash.h
75
static inline int batadv_hash_add(struct batadv_hashtable *hash,
net/batman-adv/hash.h
87
if (!hash)
net/batman-adv/hash.h
90
index = choose(data, hash->size);
net/batman-adv/hash.h
91
head = &hash->table[index];
net/batman-adv/hash.h
92
list_lock = &hash->list_locks[index];
net/batman-adv/multicast.c
2024
struct batadv_hashtable *hash,
net/batman-adv/multicast.c
2030
spin_lock_bh(&hash->list_locks[bucket]);
net/batman-adv/multicast.c
2031
cb->seq = atomic_read(&hash->generation) << 1 | 1;
net/batman-adv/multicast.c
2033
hlist_for_each_entry(orig_node, &hash->table[bucket], hash_entry) {
net/batman-adv/multicast.c
2042
spin_unlock_bh(&hash->list_locks[bucket]);
net/batman-adv/multicast.c
2051
spin_unlock_bh(&hash->list_locks[bucket]);
net/batman-adv/multicast.c
2072
struct batadv_hashtable *hash = bat_priv->orig_hash;
net/batman-adv/multicast.c
2076
while (bucket_tmp < hash->size) {
net/batman-adv/multicast.c
2077
if (batadv_mcast_flags_dump_bucket(msg, portid, cb, hash,
net/batman-adv/originator.c
1261
struct batadv_hashtable *hash = bat_priv->orig_hash;
net/batman-adv/originator.c
1268
if (!hash)
net/batman-adv/originator.c
1272
for (i = 0; i < hash->size; i++) {
net/batman-adv/originator.c
1273
head = &hash->table[i];
net/batman-adv/originator.c
1276
list_lock = &hash->list_locks[i];
net/batman-adv/originator.c
56
struct batadv_hashtable *hash = bat_priv->orig_hash;
net/batman-adv/originator.c
61
if (!hash)
net/batman-adv/originator.c
64
index = batadv_choose_orig(data, hash->size);
net/batman-adv/originator.c
65
head = &hash->table[index];
net/batman-adv/originator.c
899
struct batadv_hashtable *hash = bat_priv->orig_hash;
net/batman-adv/originator.c
906
if (!hash)
net/batman-adv/originator.c
913
for (i = 0; i < hash->size; i++) {
net/batman-adv/originator.c
914
head = &hash->table[i];
net/batman-adv/originator.c
915
list_lock = &hash->list_locks[i];
net/batman-adv/originator.c
926
batadv_hash_destroy(hash);
net/batman-adv/originator.h
79
u32 hash = 0;
net/batman-adv/originator.h
81
hash = jhash(data, ETH_ALEN, hash);
net/batman-adv/originator.h
82
return hash % size;
net/batman-adv/translation-table.c
108
u32 hash = 0;
net/batman-adv/translation-table.c
1098
struct batadv_hashtable *hash, unsigned int bucket,
net/batman-adv/translation-table.c
1104
spin_lock_bh(&hash->list_locks[bucket]);
net/batman-adv/translation-table.c
1105
cb->seq = atomic_read(&hash->generation) << 1 | 1;
net/batman-adv/translation-table.c
1107
hlist_for_each_entry(common, &hash->table[bucket], hash_entry) {
net/batman-adv/translation-table.c
111
hash = jhash(&tt->addr, ETH_ALEN, hash);
net/batman-adv/translation-table.c
1113
spin_unlock_bh(&hash->list_locks[bucket]);
net/batman-adv/translation-table.c
1118
spin_unlock_bh(&hash->list_locks[bucket]);
net/batman-adv/translation-table.c
112
hash = jhash(&tt->vid, sizeof(tt->vid), hash);
net/batman-adv/translation-table.c
1136
struct batadv_hashtable *hash;
net/batman-adv/translation-table.c
114
return hash % size;
net/batman-adv/translation-table.c
1154
hash = bat_priv->tt.local_hash;
net/batman-adv/translation-table.c
1156
while (bucket < hash->size) {
net/batman-adv/translation-table.c
1158
hash, bucket, &idx))
net/batman-adv/translation-table.c
127
batadv_tt_hash_find(struct batadv_hashtable *hash, const u8 *addr,
net/batman-adv/translation-table.c
1304
struct batadv_hashtable *hash = bat_priv->tt.local_hash;
net/batman-adv/translation-table.c
1309
for (i = 0; i < hash->size; i++) {
net/batman-adv/translation-table.c
1310
head = &hash->table[i];
net/batman-adv/translation-table.c
1311
list_lock = &hash->list_locks[i];
net/batman-adv/translation-table.c
1321
struct batadv_hashtable *hash;
net/batman-adv/translation-table.c
1332
hash = bat_priv->tt.local_hash;
net/batman-adv/translation-table.c
1334
for (i = 0; i < hash->size; i++) {
net/batman-adv/translation-table.c
1335
head = &hash->table[i];
net/batman-adv/translation-table.c
1336
list_lock = &hash->list_locks[i];
net/batman-adv/translation-table.c
134
if (!hash)
net/batman-adv/translation-table.c
1351
batadv_hash_destroy(hash);
net/batman-adv/translation-table.c
140
index = batadv_choose_tt(&to_search, hash->size);
net/batman-adv/translation-table.c
141
head = &hash->table[index];
net/batman-adv/translation-table.c
1903
struct batadv_hashtable *hash;
net/batman-adv/translation-table.c
1923
hash = bat_priv->tt.global_hash;
net/batman-adv/translation-table.c
1925
while (bucket < hash->size) {
net/batman-adv/translation-table.c
1926
head = &hash->table[bucket];
net/batman-adv/translation-table.c
2151
struct batadv_hashtable *hash = bat_priv->tt.global_hash;
net/batman-adv/translation-table.c
2157
if (!hash)
net/batman-adv/translation-table.c
2160
for (i = 0; i < hash->size; i++) {
net/batman-adv/translation-table.c
2161
head = &hash->table[i];
net/batman-adv/translation-table.c
2162
list_lock = &hash->list_locks[i];
net/batman-adv/translation-table.c
2217
struct batadv_hashtable *hash = bat_priv->tt.global_hash;
net/batman-adv/translation-table.c
2226
for (i = 0; i < hash->size; i++) {
net/batman-adv/translation-table.c
2227
head = &hash->table[i];
net/batman-adv/translation-table.c
2228
list_lock = &hash->list_locks[i];
net/batman-adv/translation-table.c
2256
struct batadv_hashtable *hash;
net/batman-adv/translation-table.c
2267
hash = bat_priv->tt.global_hash;
net/batman-adv/translation-table.c
2269
for (i = 0; i < hash->size; i++) {
net/batman-adv/translation-table.c
2270
head = &hash->table[i];
net/batman-adv/translation-table.c
2271
list_lock = &hash->list_locks[i];
net/batman-adv/translation-table.c
2285
batadv_hash_destroy(hash);
net/batman-adv/translation-table.c
2392
struct batadv_hashtable *hash = bat_priv->tt.global_hash;
net/batman-adv/translation-table.c
2401
for (i = 0; i < hash->size; i++) {
net/batman-adv/translation-table.c
2402
head = &hash->table[i];
net/batman-adv/translation-table.c
2472
struct batadv_hashtable *hash = bat_priv->tt.local_hash;
net/batman-adv/translation-table.c
2479
for (i = 0; i < hash->size; i++) {
net/batman-adv/translation-table.c
2480
head = &hash->table[i];
net/batman-adv/translation-table.c
2705
struct batadv_hashtable *hash,
net/batman-adv/translation-table.c
2727
for (i = 0; i < hash->size; i++) {
net/batman-adv/translation-table.c
2728
head = &hash->table[i];
net/batman-adv/translation-table.c
3551
struct batadv_hashtable *hash = bat_priv->tt.local_hash;
net/batman-adv/translation-table.c
3556
if (!hash)
net/batman-adv/translation-table.c
3559
for (i = 0; i < hash->size; i++) {
net/batman-adv/translation-table.c
3560
head = &hash->table[i];
net/batman-adv/translation-table.c
3588
struct batadv_hashtable *hash = bat_priv->tt.local_hash;
net/batman-adv/translation-table.c
3596
if (!hash)
net/batman-adv/translation-table.c
3599
for (i = 0; i < hash->size; i++) {
net/batman-adv/translation-table.c
3600
head = &hash->table[i];
net/batman-adv/translation-table.c
3601
list_lock = &hash->list_locks[i];
net/batman-adv/types.h
1144
struct batadv_hashtable *hash;
net/bluetooth/hci_event.c
5649
memcpy(cp.hash, data->hash192, sizeof(cp.hash));
net/bluetooth/mgmt.c
5713
memcpy(mgmt_rp.hash192, rp->hash, sizeof(rp->hash));
net/bluetooth/mgmt.c
5830
cp->addr.type, cp->hash,
net/bluetooth/mgmt.c
8218
h192 = rp->hash;
net/bluetooth/mgmt.c
8324
u8 status, flags, role, addr[7], hash[16], rand[16];
net/bluetooth/mgmt.c
8389
smp_generate_oob(hdev, hash, rand) < 0) {
net/bluetooth/mgmt.c
8436
hash, sizeof(hash));
net/bluetooth/smp.c
501
u8 hash[3];
net/bluetooth/smp.c
509
err = smp_ah(irk, &bdaddr->b[3], hash);
net/bluetooth/smp.c
513
return !crypto_memneq(bdaddr->b, hash, 3);
net/bluetooth/smp.c
538
int smp_generate_oob(struct hci_dev *hdev, u8 hash[16], u8 rand[16])
net/bluetooth/smp.c
578
smp->local_rand, 0, hash);
net/bluetooth/smp.h
195
int smp_generate_oob(struct hci_dev *hdev, u8 hash[16], u8 rand[16]);
net/can/af_can.c
330
unsigned int hash;
net/can/af_can.c
332
hash = can_id;
net/can/af_can.c
333
hash ^= can_id >> CAN_EFF_RCV_HASH_BITS;
net/can/af_can.c
334
hash ^= can_id >> (2 * CAN_EFF_RCV_HASH_BITS);
net/can/af_can.c
336
return hash & ((1 << CAN_EFF_RCV_HASH_BITS) - 1);
net/can/af_can.c
647
while (!(skb->hash))
net/can/af_can.c
648
skb->hash = atomic_inc_return(&skbcounter);
net/can/raw.c
168
this_cpu_ptr(ro->uniq)->hash == oskb->hash) {
net/can/raw.c
178
this_cpu_ptr(ro->uniq)->hash = oskb->hash;
net/can/raw.c
80
u32 hash;
net/ceph/ceph_hash.c
101
hash = (hash + (c << 4) + (c >> 4)) * 11;
net/ceph/ceph_hash.c
103
return hash;
net/ceph/ceph_hash.c
96
unsigned long hash = 0;
net/ceph/crush/hash.c
29
__u32 hash = crush_hash_seed ^ a;
net/ceph/crush/hash.c
33
crush_hashmix(b, x, hash);
net/ceph/crush/hash.c
34
crush_hashmix(y, a, hash);
net/ceph/crush/hash.c
35
return hash;
net/ceph/crush/hash.c
40
__u32 hash = crush_hash_seed ^ a ^ b;
net/ceph/crush/hash.c
43
crush_hashmix(a, b, hash);
net/ceph/crush/hash.c
44
crush_hashmix(x, a, hash);
net/ceph/crush/hash.c
45
crush_hashmix(b, y, hash);
net/ceph/crush/hash.c
46
return hash;
net/ceph/crush/hash.c
51
__u32 hash = crush_hash_seed ^ a ^ b ^ c;
net/ceph/crush/hash.c
54
crush_hashmix(a, b, hash);
net/ceph/crush/hash.c
55
crush_hashmix(c, x, hash);
net/ceph/crush/hash.c
56
crush_hashmix(y, a, hash);
net/ceph/crush/hash.c
57
crush_hashmix(b, x, hash);
net/ceph/crush/hash.c
58
crush_hashmix(y, c, hash);
net/ceph/crush/hash.c
59
return hash;
net/ceph/crush/hash.c
64
__u32 hash = crush_hash_seed ^ a ^ b ^ c ^ d;
net/ceph/crush/hash.c
67
crush_hashmix(a, b, hash);
net/ceph/crush/hash.c
68
crush_hashmix(c, d, hash);
net/ceph/crush/hash.c
69
crush_hashmix(a, x, hash);
net/ceph/crush/hash.c
70
crush_hashmix(y, b, hash);
net/ceph/crush/hash.c
71
crush_hashmix(c, x, hash);
net/ceph/crush/hash.c
72
crush_hashmix(y, d, hash);
net/ceph/crush/hash.c
73
return hash;
net/ceph/crush/hash.c
79
__u32 hash = crush_hash_seed ^ a ^ b ^ c ^ d ^ e;
net/ceph/crush/hash.c
82
crush_hashmix(a, b, hash);
net/ceph/crush/hash.c
83
crush_hashmix(c, d, hash);
net/ceph/crush/hash.c
84
crush_hashmix(e, x, hash);
net/ceph/crush/hash.c
85
crush_hashmix(y, a, hash);
net/ceph/crush/hash.c
86
crush_hashmix(b, x, hash);
net/ceph/crush/hash.c
87
crush_hashmix(y, c, hash);
net/ceph/crush/hash.c
88
crush_hashmix(d, x, hash);
net/ceph/crush/hash.c
89
crush_hashmix(y, e, hash);
net/ceph/crush/hash.c
90
return hash;
net/ceph/crush/mapper.c
113
i = crush_hash32_3(bucket->hash, x, bucket->id, p) %
net/ceph/crush/mapper.c
148
__u64 w = crush_hash32_4(bucket->h.hash, x, bucket->h.items[i],
net/ceph/crush/mapper.c
210
t = (__u64)crush_hash32_4(bucket->h.hash, x, n, r,
net/ceph/crush/mapper.c
237
draw = crush_hash32_3(bucket->h.hash, x, bucket->h.items[i], r);
net/ceph/crush/mapper.c
340
u = crush_hash32_3(bucket->h.hash, x, ids[i], r);
net/ceph/crush/mapper.c
88
s = crush_hash32_3(bucket->hash, x, bucket->id, 0) %
net/ceph/debugfs.c
308
if (hoid->snapid == 0 && hoid->hash == 0 && !hoid->is_max &&
net/ceph/osd_client.c
1863
ceph_decode_32_safe(p, end, hoid->hash, e_inval);
net/ceph/osd_client.c
1895
ceph_encode_32(p, hoid->hash);
net/ceph/osd_client.c
2003
hoid->hash = t->pgid.seed;
net/ceph/osd_client.c
2235
__le32 hash;
net/ceph/osd_client.c
2288
pgid.seed = le32_to_cpu(head.hash);
net/ceph/osd_client.c
3598
s64 hash = ceph_decode_64(p);
net/ceph/osd_client.c
3599
if (hash != -1) {
net/ceph/osdmap.c
519
b->hash = ceph_decode_8(p);
net/core/dev.c
12887
struct hlist_head *hash;
net/core/dev.c
12889
hash = kmalloc_objs(*hash, NETDEV_HASHENTRIES);
net/core/dev.c
12890
if (hash != NULL)
net/core/dev.c
12892
INIT_HLIST_HEAD(&hash[i]);
net/core/dev.c
12894
return hash;
net/core/dev.c
196
unsigned int hash = full_name_hash(net, name, strnlen(name, IFNAMSIZ));
net/core/dev.c
198
return &net->dev_name_head[hash_32(hash, NETDEV_HASHBITS)];
net/core/dev.c
3513
u32 hash;
net/core/dev.c
3532
hash = skb_get_rx_queue(skb);
net/core/dev.c
3533
if (hash >= qoffset)
net/core/dev.c
3534
hash -= qoffset;
net/core/dev.c
3535
while (unlikely(hash >= qcount))
net/core/dev.c
3536
hash -= qcount;
net/core/dev.c
3537
return hash + qoffset;
net/core/dev.c
4985
static u32 rfs_slot(u32 hash, const struct rps_dev_flow_table *flow_table)
net/core/dev.c
4987
return hash_32(hash, flow_table->log);
net/core/dev.c
5022
struct rps_dev_flow *rflow, u16 next_cpu, u32 hash)
net/core/dev.c
5049
flow_id = rfs_slot(hash, flow_table);
net/core/dev.c
5056
if (hash != READ_ONCE(tmp_rflow->hash) ||
net/core/dev.c
5070
WRITE_ONCE(rflow->hash, hash);
net/core/dev.c
5098
u32 hash;
net/core/dev.c
5121
hash = skb_get_hash(skb);
net/core/dev.c
5122
if (!hash)
net/core/dev.c
5134
ident = READ_ONCE(sock_flow_table->ents[hash & sock_flow_table->mask]);
net/core/dev.c
5135
if ((ident ^ hash) & ~net_hotdata.rps_cpu_mask)
net/core/dev.c
5143
rflow = &flow_table->flows[rfs_slot(hash, flow_table)];
net/core/dev.c
5162
rflow = set_rps_cpu(dev, skb, rflow, next_cpu, hash);
net/core/dev.c
5175
tcpu = map->cpus[reciprocal_scale(hash, map->len)];
net/core/dev.c
796
unsigned int hash = napi_id % HASH_SIZE(napi_hash);
net/core/dev.c
799
hlist_for_each_entry_rcu(napi, &napi_hash[hash], napi_hash_node)
net/core/filter.c
11424
u32 hash)
net/core/filter.c
11431
reuse_kern->hash = hash;
net/core/filter.c
11439
u32 hash)
net/core/filter.c
11444
bpf_init_reuseport_kern(&reuse_kern, reuse, sk, skb, migrating_sk, hash);
net/core/filter.c
11590
case offsetof(struct sk_reuseport_md, hash):
net/core/filter.c
11666
case offsetof(struct sk_reuseport_md, hash):
net/core/filter.c
11667
SK_REUSEPORT_LOAD_FIELD(hash);
net/core/filter.c
3196
BPF_CALL_2(bpf_set_hash, struct sk_buff *, skb, u32, hash)
net/core/filter.c
3202
__skb_set_sw_hash(skb, hash, true);
net/core/filter.c
413
BUILD_BUG_ON(sizeof_field(struct sk_buff, hash) != 4);
net/core/filter.c
416
offsetof(struct sk_buff, hash));
net/core/filter.c
9814
case offsetof(struct __sk_buff, hash):
net/core/filter.c
9816
bpf_target_off(struct sk_buff, hash, 4,
net/core/flow_dissector.c
1802
u32 hash;
net/core/flow_dissector.c
1806
hash = siphash(flow_keys_hash_start(keys),
net/core/flow_dissector.c
1808
if (!hash)
net/core/flow_dissector.c
1809
hash = 1;
net/core/flow_dissector.c
1811
return hash;
net/core/flow_dissector.c
1894
u32 hash;
net/core/flow_dissector.c
1904
hash = __flow_hash_from_keys(&keys, &hashrnd);
net/core/flow_dissector.c
1906
__skb_set_sw_hash(skb, hash, flow_keys_have_l4(&keys));
net/core/flow_dissector.c
510
key->hash = skb_get_hash_raw(skb);
net/core/gro.c
292
struct list_head *head = &gro->hash[index].list;
net/core/gro.c
300
gro->hash[index].count--;
net/core/gro.c
303
if (!gro->hash[index].count)
net/core/gro.c
347
u32 hash = skb_get_hash_raw(skb);
net/core/gro.c
353
if (hash != skb_get_hash_raw(p)) {
net/core/gro.c
467
struct gro_list *gro_list = &gro->hash[bucket];
net/core/gro.c
806
INIT_LIST_HEAD(&gro->hash[i].list);
net/core/gro.c
807
gro->hash[i].count = 0;
net/core/gro.c
822
list_for_each_entry_safe(skb, n, &gro->hash[i].list, list)
net/core/gro.c
825
gro->hash[i].count = 0;
net/core/neighbour.c
1026
hlist_del_rcu(&n->hash);
net/core/neighbour.c
242
hlist_del_rcu(&n->hash);
net/core/neighbour.c
3191
hlist_del_rcu(&n->hash);
net/core/neighbour.c
3314
hlist_for_each_entry_continue(n, hash) {
net/core/neighbour.c
383
hlist_del_rcu(&n->hash);
net/core/neighbour.c
594
unsigned int i, hash;
net/core/neighbour.c
610
hash = tbl->hash(n->primary_key, n->dev,
net/core/neighbour.c
613
hash >>= (32 - new_nht->hash_shift);
net/core/neighbour.c
615
hlist_del_rcu(&n->hash);
net/core/neighbour.c
616
hlist_add_head_rcu(&n->hash, &new_nht->hash_heads[hash]);
net/core/neighbour.c
696
hash_val = tbl->hash(n->primary_key, dev, nht->hash_rnd) >> (32 - nht->hash_shift);
net/core/neighbour.c
719
hlist_add_head_rcu(&n->hash, &nht->hash_heads[hash_val]);
net/core/net-procfs.c
236
int hash;
net/core/net-procfs.c
273
hash = 0;
net/core/net-procfs.c
276
hash = ntohs(pt->type) & PTYPE_HASH_MASK;
net/core/net-procfs.c
278
while (nxt == &ptype_base[hash]) {
net/core/net-procfs.c
279
if (++hash >= PTYPE_HASH_SIZE)
net/core/net-procfs.c
281
nxt = READ_ONCE(ptype_base[hash].next);
net/core/skbuff.c
1355
skb->hash, skb->sw_hash, skb->l4_hash,
net/core/skbuff.c
1578
CHECK_SKB_FIELD(hash);
net/core/sock.c
4344
proto_method_implemented(proto->hash),
net/core/sock_map.c
1017
hash = sock_hash_bucket_hash(key, key_size);
net/core/sock_map.c
1018
bucket = sock_hash_select_bucket(htab, hash);
net/core/sock_map.c
1021
elem = sock_hash_lookup_elem_raw(&bucket->head, hash, key, key_size);
net/core/sock_map.c
1030
elem_new = sock_hash_alloc_elem(htab, key, key_size, hash, sk, elem);
net/core/sock_map.c
1061
u32 hash, key_size = map->key_size;
net/core/sock_map.c
1067
hash = sock_hash_bucket_hash(key, key_size);
net/core/sock_map.c
1068
head = &sock_hash_select_bucket(htab, hash)->head;
net/core/sock_map.c
1069
elem = sock_hash_lookup_elem_raw(head, hash, key, key_size);
net/core/sock_map.c
1080
i = hash & (htab->buckets_num - 1);
net/core/sock_map.c
847
u32 hash;
net/core/sock_map.c
873
u32 hash)
net/core/sock_map.c
875
return &htab->buckets[hash & (htab->buckets_num - 1)];
net/core/sock_map.c
879
sock_hash_lookup_elem_raw(struct hlist_head *head, u32 hash, void *key,
net/core/sock_map.c
885
if (elem->hash == hash &&
net/core/sock_map.c
896
u32 key_size = map->key_size, hash;
net/core/sock_map.c
902
hash = sock_hash_bucket_hash(key, key_size);
net/core/sock_map.c
903
bucket = sock_hash_select_bucket(htab, hash);
net/core/sock_map.c
904
elem = sock_hash_lookup_elem_raw(&bucket->head, hash, key, key_size);
net/core/sock_map.c
924
bucket = sock_hash_select_bucket(htab, elem->hash);
net/core/sock_map.c
931
elem_probe = sock_hash_lookup_elem_raw(&bucket->head, elem->hash,
net/core/sock_map.c
944
u32 hash, key_size = map->key_size;
net/core/sock_map.c
949
hash = sock_hash_bucket_hash(key, key_size);
net/core/sock_map.c
950
bucket = sock_hash_select_bucket(htab, hash);
net/core/sock_map.c
953
elem = sock_hash_lookup_elem_raw(&bucket->head, hash, key, key_size);
net/core/sock_map.c
966
u32 hash, struct sock *sk,
net/core/sock_map.c
987
new->hash = hash;
net/core/sock_map.c
995
u32 key_size = map->key_size, hash;
net/core/sock_reuseport.c
528
u32 hash, u16 num_socks)
net/core/sock_reuseport.c
533
i = j = reciprocal_scale(hash, num_socks);
net/core/sock_reuseport.c
569
u32 hash,
net/core/sock_reuseport.c
595
sk2 = bpf_run_sk_reuseport(reuse, sk, prog, skb, NULL, hash);
net/core/sock_reuseport.c
602
sk2 = reuseport_select_sock_by_hash(reuse, hash, socks);
net/core/sock_reuseport.c
629
u32 hash;
net/core/sock_reuseport.c
644
hash = migrating_sk->sk_hash;
net/core/sock_reuseport.c
659
nsk = bpf_run_sk_reuseport(reuse, sk, prog, skb, migrating_sk, hash);
net/core/sock_reuseport.c
666
nsk = reuseport_select_sock_by_hash(reuse, hash, socks);
net/core/xdp.c
925
__bpf_kfunc int bpf_xdp_metadata_rx_hash(const struct xdp_md *ctx, u32 *hash,
net/ieee802154/socket.c
1044
if (sk->sk_prot->hash) {
net/ieee802154/socket.c
1045
rc = sk->sk_prot->hash(sk);
net/ieee802154/socket.c
402
.hash = raw_hash,
net/ieee802154/socket.c
963
.hash = dgram_hash,
net/ipv4/af_inet.c
386
err = sk->sk_prot->hash(sk);
net/ipv4/arp.c
157
.hash = arp_hash,
net/ipv4/cipso_ipv4.c
239
u32 hash;
net/ipv4/cipso_ipv4.c
244
hash = cipso_v4_map_cache_hash(key, key_len);
net/ipv4/cipso_ipv4.c
245
bkt = hash & (CIPSO_V4_CACHE_BUCKETS - 1);
net/ipv4/cipso_ipv4.c
248
if (entry->hash == hash &&
net/ipv4/cipso_ipv4.c
319
entry->hash = cipso_v4_map_cache_hash(cipso_ptr, cipso_ptr_len);
net/ipv4/cipso_ipv4.c
323
bkt = entry->hash & (CIPSO_V4_CACHE_BUCKETS - 1);
net/ipv4/cipso_ipv4.c
63
u32 hash;
net/ipv4/devinet.c
123
u32 hash = inet_addr_hash(net, ifa->ifa_local);
net/ipv4/devinet.c
126
hlist_add_head_rcu(&ifa->addr_lst, &net->ipv4.inet_addr_lst[hash]);
net/ipv4/devinet.c
176
u32 hash = inet_addr_hash(net, addr);
net/ipv4/devinet.c
179
hlist_for_each_entry_rcu(ifa, &net->ipv4.inet_addr_lst[hash], addr_lst)
net/ipv4/fib_semantics.c
157
struct fnhe_hash_bucket *hash;
net/ipv4/fib_semantics.c
160
hash = rcu_dereference_protected(nhc->nhc_exceptions, 1);
net/ipv4/fib_semantics.c
161
if (!hash)
net/ipv4/fib_semantics.c
166
fnhe = rcu_dereference_protected(hash[i].chain, 1);
net/ipv4/fib_semantics.c
180
kfree(hash);
net/ipv4/fib_semantics.c
2164
void fib_select_multipath(struct fib_result *res, int hash,
net/ipv4/fib_semantics.c
2174
nexthop_path_fib_result(res, hash);
net/ipv4/fib_semantics.c
2194
if (hash <= nh_upper_bound)
net/ipv4/fib_semantics.c
422
unsigned int hash;
net/ipv4/fib_semantics.c
424
hash = fib_info_hashfn_1(cfg->fc_nh_id,
net/ipv4/fib_semantics.c
428
hash = fib_info_hashfn_result(net, hash);
net/ipv4/fib_semantics.c
429
head = &net->ipv4.fib_info_hash[hash];
net/ipv4/igmp.c
1394
u32 hash;
net/ipv4/igmp.c
1398
hash = ip_mc_hash(im);
net/ipv4/igmp.c
1399
im->next_hash = mc_hash[hash];
net/ipv4/igmp.c
1400
rcu_assign_pointer(mc_hash[hash], im);
net/ipv4/igmp.c
1414
hash = ip_mc_hash(im);
net/ipv4/igmp.c
1415
im->next_hash = mc_hash[hash];
net/ipv4/igmp.c
1416
RCU_INIT_POINTER(mc_hash[hash], im);
net/ipv4/igmp.c
1514
u32 hash = hash_32((__force u32)addr, MC_HASH_SZ_LOG);
net/ipv4/igmp.c
1516
for (im = rtnl_dereference(mc_hash[hash]);
net/ipv4/igmp.c
2816
u32 hash = hash_32((__force u32)mc_addr, MC_HASH_SZ_LOG);
net/ipv4/igmp.c
2818
for (im = rcu_dereference(mc_hash[hash]);
net/ipv4/inet_connection_sock.c
1348
err = sk->sk_prot->hash(sk);
net/ipv4/inet_hashtables.c
1040
bool rcu_lookup, u32 hash))
net/ipv4/inet_hashtables.c
334
u32 hash;
net/ipv4/inet_hashtables.c
338
hash = ipv6_portaddr_hash(sock_net(sk),
net/ipv4/inet_hashtables.c
343
hash = ipv4_portaddr_hash(sock_net(sk),
net/ipv4/inet_hashtables.c
346
return inet_lhash2_bucket(h, hash);
net/ipv4/inet_hashtables.c
537
unsigned int hash, slot;
net/ipv4/inet_hashtables.c
541
hash = inet_ehashfn(net, daddr, hnum, saddr, sport);
net/ipv4/inet_hashtables.c
542
slot = hash & hashinfo->ehash_mask;
net/ipv4/inet_hashtables.c
547
if (sk->sk_hash != hash)
net/ipv4/inet_hashtables.c
579
u32 hash)
net/ipv4/inet_hashtables.c
590
struct inet_ehash_bucket *head = inet_ehash_bucket(hinfo, hash);
net/ipv4/inet_hashtables.c
598
if (sk2->sk_hash != hash ||
net/ipv4/inet_hashtables.c
608
lock = inet_ehash_lockp(hinfo, hash);
net/ipv4/inet_hashtables.c
612
if (sk2->sk_hash != hash)
net/ipv4/inet_hashtables.c
630
sk->sk_hash = hash;
net/ipv4/inet_hashtables.c
902
u32 hash;
net/ipv4/inet_hashtables.c
906
hash = ipv6_portaddr_hash(net, &in6addr_any, port);
net/ipv4/inet_hashtables.c
909
hash = ipv4_portaddr_hash(net, 0, port);
net/ipv4/inet_hashtables.c
911
return &hinfo->bhash2[hash & (hinfo->bhash_size - 1)];
net/ipv4/ip_tunnel.c
123
hash = ip_tunnel_hash(key, 0);
net/ipv4/ip_tunnel.c
124
head = &itn->tunnels[hash];
net/ipv4/ip_tunnel.c
89
unsigned int hash;
net/ipv4/ip_tunnel.c
91
hash = ip_tunnel_hash(key, remote);
net/ipv4/ip_tunnel.c
92
head = &itn->tunnels[hash];
net/ipv4/nexthop.c
1410
static struct nexthop *nexthop_select_path_fdb(struct nh_group *nhg, int hash)
net/ipv4/nexthop.c
1417
if (hash > atomic_read(&nhge->hthr.upper_bound))
net/ipv4/nexthop.c
1428
static struct nexthop *nexthop_select_path_hthr(struct nh_group *nhg, int hash)
net/ipv4/nexthop.c
1434
return nexthop_select_path_fdb(nhg, hash);
net/ipv4/nexthop.c
1448
if (hash > atomic_read(&nhge->hthr.upper_bound))
net/ipv4/nexthop.c
1461
static struct nexthop *nexthop_select_path_res(struct nh_group *nhg, int hash)
net/ipv4/nexthop.c
1464
u16 bucket_index = hash % res_table->num_nh_buckets;
net/ipv4/nexthop.c
1478
struct nexthop *nexthop_select_path(struct nexthop *nh, int hash)
net/ipv4/nexthop.c
1487
return nexthop_select_path_hthr(nhg, hash);
net/ipv4/nexthop.c
1489
return nexthop_select_path_res(nhg, hash);
net/ipv4/nexthop.c
2681
unsigned int hash = nh_dev_hashfn(dev->ifindex);
net/ipv4/nexthop.c
2683
struct hlist_head *head = &net->nexthop.devhash[hash];
net/ipv4/nexthop.c
3907
unsigned int hash = nh_dev_hashfn(dev->ifindex);
net/ipv4/nexthop.c
3909
struct hlist_head *head = &net->nexthop.devhash[hash];
net/ipv4/nexthop.c
471
unsigned int hash;
net/ipv4/nexthop.c
475
hash = nh_dev_hashfn(dev->ifindex);
net/ipv4/nexthop.c
476
head = &net->nexthop.devhash[hash];
net/ipv4/ping.c
1025
hslot = &ping_table.hash[state->bucket];
net/ipv4/ping.c
1185
INIT_HLIST_HEAD(&ping_table.hash[i]);
net/ipv4/ping.c
53
struct hlist_head hash[PING_HTABLE_SIZE];
net/ipv4/ping.c
72
return &table->hash[ping_hashfn(net, num, PING_HTABLE_MASK)];
net/ipv4/raw.c
164
const struct iphdr *iph, int hash)
net/ipv4/raw.c
172
hlist = &raw_v4_hashinfo.ht[hash];
net/ipv4/raw.c
280
int hash;
net/ipv4/raw.c
282
hash = raw_hashfunc(net, protocol);
net/ipv4/raw.c
283
hlist = &raw_v4_hashinfo.ht[hash];
net/ipv4/raw.c
948
.hash = raw_hash_sk,
net/ipv4/route.c
1351
struct fnhe_hash_bucket *hash;
net/ipv4/route.c
1357
hash = rcu_dereference_protected(nhc->nhc_exceptions,
net/ipv4/route.c
1359
hash += hval;
net/ipv4/route.c
1361
fnhe_p = &hash->chain;
net/ipv4/route.c
1386
struct fnhe_hash_bucket *hash = rcu_dereference(nhc->nhc_exceptions);
net/ipv4/route.c
1390
if (!hash)
net/ipv4/route.c
1395
for (fnhe = rcu_dereference(hash[hval].chain); fnhe;
net/ipv4/route.c
470
static u32 ip_idents_reserve(u32 hash, int segs)
net/ipv4/route.c
477
bucket = hash & ip_idents_mask;
net/ipv4/route.c
494
u32 hash, id;
net/ipv4/route.c
501
hash = siphash_3u32((__force u32)iph->daddr,
net/ipv4/route.c
505
id = ip_idents_reserve(hash, segs);
net/ipv4/route.c
594
static void fnhe_remove_oldest(struct fnhe_hash_bucket *hash)
net/ipv4/route.c
599
for (fnhe_p = &hash->chain; ; fnhe_p = &fnhe->fnhe_next) {
net/ipv4/route.c
648
struct fnhe_hash_bucket *hash;
net/ipv4/route.c
660
hash = rcu_dereference(nhc->nhc_exceptions);
net/ipv4/route.c
661
if (!hash) {
net/ipv4/route.c
662
hash = kzalloc_objs(*hash, FNHE_HASH_SIZE, GFP_ATOMIC);
net/ipv4/route.c
663
if (!hash)
net/ipv4/route.c
665
rcu_assign_pointer(nhc->nhc_exceptions, hash);
net/ipv4/route.c
668
hash += hval;
net/ipv4/route.c
671
for (fnhe = rcu_dereference(hash->chain); fnhe;
net/ipv4/route.c
701
fnhe_remove_oldest(hash);
net/ipv4/route.c
709
fnhe->fnhe_next = hash->chain;
net/ipv4/route.c
718
rcu_assign_pointer(hash->chain, fnhe);
net/ipv4/tcp_ao.c
528
bool exclude_options, u8 *hash,
net/ipv4/tcp_ipv4.c
3437
.hash = inet_hash,
net/ipv4/tcp_metrics.c
166
unsigned int hash)
net/ipv4/tcp_metrics.c
178
tm = __tcp_get_metrics(saddr, daddr, net, hash);
net/ipv4/tcp_metrics.c
191
oldest = deref_locked(tcp_metrics_hash[hash].chain);
net/ipv4/tcp_metrics.c
213
tm->tcpm_next = tcp_metrics_hash[hash].chain;
net/ipv4/tcp_metrics.c
214
rcu_assign_pointer(tcp_metrics_hash[hash].chain, tm);
net/ipv4/tcp_metrics.c
233
struct net *net, unsigned int hash)
net/ipv4/tcp_metrics.c
238
for (tm = rcu_dereference(tcp_metrics_hash[hash].chain); tm;
net/ipv4/tcp_metrics.c
254
unsigned int hash;
net/ipv4/tcp_metrics.c
26
struct net *net, unsigned int hash);
net/ipv4/tcp_metrics.c
263
hash = ipv4_addr_hash(inet_rsk(req)->ir_rmt_addr);
net/ipv4/tcp_metrics.c
269
hash = ipv6_addr_hash(&inet_rsk(req)->ir_v6_rmt_addr);
net/ipv4/tcp_metrics.c
277
hash ^= net_hash_mix(net);
net/ipv4/tcp_metrics.c
278
hash = hash_32(hash, tcp_metrics_hash_log);
net/ipv4/tcp_metrics.c
280
for (tm = rcu_dereference(tcp_metrics_hash[hash].chain); tm;
net/ipv4/tcp_metrics.c
297
unsigned int hash;
net/ipv4/tcp_metrics.c
303
hash = ipv4_addr_hash(inet_sk(sk)->inet_daddr);
net/ipv4/tcp_metrics.c
310
hash = ipv4_addr_hash(inet_sk(sk)->inet_daddr);
net/ipv4/tcp_metrics.c
314
hash = ipv6_addr_hash(&sk->sk_v6_daddr);
net/ipv4/tcp_metrics.c
322
hash ^= net_hash_mix(net);
net/ipv4/tcp_metrics.c
323
hash = hash_32(hash, tcp_metrics_hash_log);
net/ipv4/tcp_metrics.c
325
tm = __tcp_get_metrics(&saddr, &daddr, net, hash);
net/ipv4/tcp_metrics.c
329
tm = tcpm_new(dst, &saddr, &daddr, hash);
net/ipv4/tcp_metrics.c
803
unsigned int *hash, int optional, int v4, int v6)
net/ipv4/tcp_metrics.c
810
if (hash)
net/ipv4/tcp_metrics.c
811
*hash = ipv4_addr_hash(inetpeer_get_addr_v4(addr));
net/ipv4/tcp_metrics.c
820
if (hash)
net/ipv4/tcp_metrics.c
821
*hash = ipv6_addr_hash(inetpeer_get_addr_v6(addr));
net/ipv4/tcp_metrics.c
828
unsigned int *hash, int optional)
net/ipv4/tcp_metrics.c
830
return __parse_nl_addr(info, addr, hash, optional,
net/ipv4/tcp_metrics.c
846
unsigned int hash;
net/ipv4/tcp_metrics.c
853
ret = parse_nl_addr(info, &daddr, &hash, 0);
net/ipv4/tcp_metrics.c
870
hash ^= net_hash_mix(net);
net/ipv4/tcp_metrics.c
871
hash = hash_32(hash, tcp_metrics_hash_log);
net/ipv4/tcp_metrics.c
874
for (tm = rcu_dereference(tcp_metrics_hash[hash].chain); tm;
net/ipv4/tcp_metrics.c
934
unsigned int hash;
net/ipv4/tcp_metrics.c
939
ret = parse_nl_addr(info, &daddr, &hash, 1);
net/ipv4/tcp_metrics.c
950
hash ^= net_hash_mix(net);
net/ipv4/tcp_metrics.c
951
hash = hash_32(hash, tcp_metrics_hash_log);
net/ipv4/tcp_metrics.c
952
hb = tcp_metrics_hash + hash;
net/ipv4/tcp_sigpool.c
108
static int __cpool_try_clone(struct crypto_ahash *hash)
net/ipv4/tcp_sigpool.c
112
tmp = crypto_clone_ahash(hash);
net/ipv4/tcp_sigpool.c
140
e->hash = cpu0_hash;
net/ipv4/tcp_sigpool.c
202
crypto_free_ahash(e->hash);
net/ipv4/tcp_sigpool.c
23
struct crypto_ahash *hash;
net/ipv4/tcp_sigpool.c
262
struct crypto_ahash *hash;
net/ipv4/tcp_sigpool.c
270
hash = crypto_clone_ahash(cpool[id].hash);
net/ipv4/tcp_sigpool.c
271
if (IS_ERR(hash)) {
net/ipv4/tcp_sigpool.c
273
return PTR_ERR(hash);
net/ipv4/tcp_sigpool.c
276
c->req = ahash_request_alloc(hash, GFP_ATOMIC);
net/ipv4/tcp_sigpool.c
278
crypto_free_ahash(hash);
net/ipv4/tcp_sigpool.c
295
struct crypto_ahash *hash = crypto_ahash_reqtfm(c->req);
net/ipv4/tcp_sigpool.c
300
crypto_free_ahash(hash);
net/ipv4/udp.c
2816
hslot = &udptable->hash[slot];
net/ipv4/udp.c
3281
.hash = udp_lib_hash,
net/ipv4/udp.c
3342
struct udp_hslot *hslot = &udptable->hash[state->bucket];
net/ipv4/udp.c
3373
spin_unlock_bh(&udptable->hash[state->bucket].lock);
net/ipv4/udp.c
3421
spin_unlock_bh(&udptable->hash[state->bucket].lock);
net/ipv4/udp.c
3817
table->hash = alloc_large_system_hash(name,
net/ipv4/udp.c
3827
table->hash2 = (void *)(table->hash + (table->mask + 1));
net/ipv4/udp.c
3829
INIT_HLIST_HEAD(&table->hash[i].head);
net/ipv4/udp.c
3830
table->hash[i].count = 0;
net/ipv4/udp.c
3831
spin_lock_init(&table->hash[i].lock);
net/ipv4/udp.c
3873
udptable->hash = vmalloc_huge(hash_entries * slot_size,
net/ipv4/udp.c
3875
if (!udptable->hash)
net/ipv4/udp.c
3878
udptable->hash2 = (void *)(udptable->hash + hash_entries);
net/ipv4/udp.c
3883
INIT_HLIST_HEAD(&udptable->hash[i].head);
net/ipv4/udp.c
3884
udptable->hash[i].count = 0;
net/ipv4/udp.c
3885
spin_lock_init(&udptable->hash[i].lock);
net/ipv4/udp.c
3908
kvfree(udptable->hash);
net/ipv4/udp.c
449
struct udp_hslot *hslot = &udptable->hash[slot];
net/ipv4/udp.c
621
void udp_lib_hash4(struct sock *sk, u16 hash)
net/ipv4/udp.c
636
hslot4 = udp_hashslot4(udptable, hash);
net/ipv4/udp.c
637
udp_sk(sk)->udp_lrpa_hash = hash;
net/ipv4/udp.c
661
unsigned int hash;
net/ipv4/udp.c
666
hash = udp_ehashfn(net, sk->sk_rcv_saddr, sk->sk_num,
net/ipv4/udp.c
669
udp_lib_hash4(sk, hash);
net/ipv4/udp_diag.c
101
struct udp_hslot *hslot = &table->hash[slot];
net/ipv4/udplite.c
57
.hash = udp_lib_hash,
net/ipv6/addrconf.c
1030
struct net_device *dev, unsigned int hash)
net/ipv6/addrconf.c
1034
hlist_for_each_entry(ifp, &net->ipv6.inet6_addr_lst[hash], addr_lst) {
net/ipv6/addrconf.c
1046
unsigned int hash = inet6_addr_hash(net, &ifa->addr);
net/ipv6/addrconf.c
1052
if (ipv6_chk_same_addr(net, &ifa->addr, dev, hash)) {
net/ipv6/addrconf.c
1056
hlist_add_head_rcu(&ifa->addr_lst, &net->ipv6.inet6_addr_lst[hash]);
net/ipv6/addrconf.c
1981
unsigned int hash = inet6_addr_hash(net, addr);
net/ipv6/addrconf.c
1992
hlist_for_each_entry_rcu(ifp, &net->ipv6.inet6_addr_lst[hash], addr_lst) {
net/ipv6/addrconf.c
2093
unsigned int hash = inet6_addr_hash(net, addr);
net/ipv6/addrconf.c
2097
hlist_for_each_entry_rcu(ifp, &net->ipv6.inet6_addr_lst[hash], addr_lst) {
net/ipv6/addrconf.c
4540
unsigned int hash = inet6_addr_hash(net, addr);
net/ipv6/addrconf.c
4545
hlist_for_each_entry_rcu(ifp, &net->ipv6.inet6_addr_lst[hash], addr_lst) {
net/ipv6/addrconf.c
4574
unsigned int hash;
net/ipv6/addrconf.c
4580
hash = inet6_addr_hash(net, addr);
net/ipv6/addrconf.c
4583
hlist_for_each_entry_rcu(ifp, &net->ipv6.inet6_addr_lst[hash], addr_lst) {
net/ipv6/af_inet6.c
246
err = sk->sk_prot->hash(sk);
net/ipv6/anycast.c
244
unsigned int hash = inet6_acaddr_hash(net, &aca->aca_addr);
net/ipv6/anycast.c
247
hlist_add_head_rcu(&aca->aca_addr_lst, &inet6_acaddr_lst[hash]);
net/ipv6/anycast.c
500
unsigned int hash = inet6_acaddr_hash(net, addr);
net/ipv6/anycast.c
502
hlist_for_each_entry_rcu(aca, &inet6_acaddr_lst[hash],
net/ipv6/calipso.c
205
u32 hash;
net/ipv6/calipso.c
210
hash = calipso_map_cache_hash(key, key_len);
net/ipv6/calipso.c
211
bkt = hash & (CALIPSO_CACHE_BUCKETS - 1);
net/ipv6/calipso.c
214
if (entry->hash == hash &&
net/ipv6/calipso.c
286
entry->hash = calipso_map_cache_hash(calipso_ptr, calipso_ptr_len);
net/ipv6/calipso.c
290
bkt = entry->hash & (CALIPSO_CACHE_BUCKETS - 1);
net/ipv6/calipso.c
74
u32 hash;
net/ipv6/inet6_hashtables.c
269
u32 hash)
net/ipv6/inet6_hashtables.c
279
struct inet_ehash_bucket *head = inet_ehash_bucket(hinfo, hash);
net/ipv6/inet6_hashtables.c
287
if (sk2->sk_hash != hash ||
net/ipv6/inet6_hashtables.c
298
lock = inet_ehash_lockp(hinfo, hash);
net/ipv6/inet6_hashtables.c
302
if (sk2->sk_hash != hash)
net/ipv6/inet6_hashtables.c
321
sk->sk_hash = hash;
net/ipv6/inet6_hashtables.c
60
unsigned int hash, slot;
net/ipv6/inet6_hashtables.c
64
hash = inet6_ehashfn(net, daddr, hnum, saddr, sport);
net/ipv6/inet6_hashtables.c
65
slot = hash & hashinfo->ehash_mask;
net/ipv6/inet6_hashtables.c
69
if (sk->sk_hash != hash)
net/ipv6/ip6_flowlabel.c
60
#define for_each_fl_rcu(hash, fl) \
net/ipv6/ip6_flowlabel.c
61
for (fl = rcu_dereference(fl_ht[(hash)]); \
net/ipv6/ip6_gre.c
104
u32 hash = ipv6_addr_hash(addr);
net/ipv6/ip6_gre.c
106
return hash_32(hash, IP6_GRE_HASH_SIZE_SHIFT);
net/ipv6/ip6_tunnel.c
119
unsigned int hash = HASH(remote, local);
net/ipv6/ip6_tunnel.c
124
for_each_ip6_tunnel_rcu(ip6n->tnls_r_l[hash]) {
net/ipv6/ip6_tunnel.c
137
hash = HASH(&any, local);
net/ipv6/ip6_tunnel.c
138
for_each_ip6_tunnel_rcu(ip6n->tnls_r_l[hash]) {
net/ipv6/ip6_tunnel.c
150
hash = HASH(remote, &any);
net/ipv6/ip6_tunnel.c
151
for_each_ip6_tunnel_rcu(ip6n->tnls_r_l[hash]) {
net/ipv6/ip6_tunnel.c
74
u32 hash = ipv6_addr_hash(addr1) ^ ipv6_addr_hash(addr2);
net/ipv6/ip6_tunnel.c
76
return hash_32(hash, IP6_TUNNEL_HASH_SIZE_SHIFT);
net/ipv6/ip6_vti.c
106
hash = HASH(&any, local);
net/ipv6/ip6_vti.c
107
for_each_vti6_tunnel_rcu(ip6n->tnls_r_l[hash]) {
net/ipv6/ip6_vti.c
113
hash = HASH(remote, &any);
net/ipv6/ip6_vti.c
114
for_each_vti6_tunnel_rcu(ip6n->tnls_r_l[hash]) {
net/ipv6/ip6_vti.c
56
u32 hash = ipv6_addr_hash(addr1) ^ ipv6_addr_hash(addr2);
net/ipv6/ip6_vti.c
58
return hash_32(hash, IP6_VTI_HASH_SIZE_SHIFT);
net/ipv6/ip6_vti.c
93
unsigned int hash = HASH(remote, local);
net/ipv6/ip6_vti.c
98
for_each_vti6_tunnel_rcu(ip6n->tnls_r_l[hash]) {
net/ipv6/ndisc.c
113
.hash = ndisc_hash,
net/ipv6/raw.c
1210
.hash = raw_hash_sk,
net/ipv6/raw.c
149
__u8 hash;
net/ipv6/raw.c
153
hash = raw_hashfunc(net, nexthdr);
net/ipv6/raw.c
154
hlist = &raw_v6_hashinfo.ht[hash];
net/ipv6/raw.c
342
int hash;
net/ipv6/raw.c
344
hash = raw_hashfunc(net, nexthdr);
net/ipv6/raw.c
345
hlist = &raw_v6_hashinfo.ht[hash];
net/ipv6/route.c
444
int hash;
net/ipv6/route.c
471
hash = fl6->mp_hash;
net/ipv6/route.c
472
if (hash <= atomic_read(&first->fib6_nh->fib_nh_upper_bound)) {
net/ipv6/route.c
485
if (hash > nh_upper_bound)
net/ipv6/seg6_iptunnel.c
115
u32 hash;
net/ipv6/seg6_iptunnel.c
118
hash = skb_get_hash(skb);
net/ipv6/seg6_iptunnel.c
119
hash = rol32(hash, 16);
net/ipv6/seg6_iptunnel.c
120
flowlabel = (__force __be32)hash & IPV6_FLOWLABEL_MASK;
net/ipv6/tcp_ipv6.c
2299
.hash = inet_hash,
net/ipv6/udp.c
1920
.hash = udp_lib_hash,
net/ipv6/udp.c
202
struct udp_hslot *hslot = &udptable->hash[slot];
net/ipv6/udp.c
326
unsigned int hash;
net/ipv6/udp.c
336
hash = udp6_ehashfn(net, &sk->sk_v6_rcv_saddr, sk->sk_num,
net/ipv6/udp.c
339
udp_lib_hash4(sk, hash);
net/ipv6/udplite.c
56
.hash = udp_lib_hash,
net/l2tp/l2tp_ip.c
623
.hash = l2tp_ip_hash,
net/l2tp/l2tp_ip6.c
751
.hash = l2tp_ip6_hash,
net/mac80211/sta_info.c
3363
goto hash;
net/mac80211/sta_info.c
3380
hash:
net/mctp/af_mctp.c
120
rc = sk->sk_prot->hash(sk);
net/mctp/af_mctp.c
714
u32 hash;
net/mctp/af_mctp.c
723
hash = mctp_bind_hash(msk->bind_type, msk->bind_local_addr, remote);
net/mctp/af_mctp.c
728
sk_for_each(existing, &net->mctp.binds[hash]) {
net/mctp/af_mctp.c
747
sk_add_node_rcu(sk, &net->mctp.binds[hash]);
net/mctp/af_mctp.c
795
.hash = mctp_sk_hash,
net/mctp/route.c
50
u8 hash;
net/mctp/route.c
54
hash = mctp_bind_hash(type, dest, src);
net/mctp/route.c
56
sk_for_each_rcu(sk, &net->mctp.binds[hash]) {
net/mpls/af_mpls.c
188
u32 hash = 0;
net/mpls/af_mpls.c
204
hash = jhash_1word(dec.label, hash);
net/mpls/af_mpls.c
227
hash = jhash_3words(ntohl(v4hdr->saddr),
net/mpls/af_mpls.c
229
v4hdr->protocol, hash);
net/mpls/af_mpls.c
236
hash = __ipv6_addr_jhash(&v6hdr->saddr, hash);
net/mpls/af_mpls.c
237
hash = __ipv6_addr_jhash(&v6hdr->daddr, hash);
net/mpls/af_mpls.c
238
hash = jhash_1word(v6hdr->nexthdr, hash);
net/mpls/af_mpls.c
245
return hash;
net/mpls/af_mpls.c
261
u32 hash = 0;
net/mpls/af_mpls.c
276
hash = mpls_multipath_hash(rt, skb);
net/mpls/af_mpls.c
277
nh_index = hash % alive;
net/mptcp/protocol.c
4034
.hash = mptcp_hash,
net/netfilter/ipvs/ip_vs_conn.c
1316
unsigned int hash = get_random_u32() & ip_vs_conn_tab_mask;
net/netfilter/ipvs/ip_vs_conn.c
1318
hlist_for_each_entry_rcu(cp, &ip_vs_conn_tab[hash], c_list) {
net/netfilter/ipvs/ip_vs_conn.c
164
unsigned int hash;
net/netfilter/ipvs/ip_vs_conn.c
171
hash = ip_vs_conn_hashkey_conn(cp);
net/netfilter/ipvs/ip_vs_conn.c
173
ct_write_lock_bh(hash);
net/netfilter/ipvs/ip_vs_conn.c
179
hlist_add_head_rcu(&cp->c_list, &ip_vs_conn_tab[hash]);
net/netfilter/ipvs/ip_vs_conn.c
188
ct_write_unlock_bh(hash);
net/netfilter/ipvs/ip_vs_conn.c
200
unsigned int hash;
net/netfilter/ipvs/ip_vs_conn.c
204
hash = ip_vs_conn_hashkey_conn(cp);
net/netfilter/ipvs/ip_vs_conn.c
206
ct_write_lock_bh(hash);
net/netfilter/ipvs/ip_vs_conn.c
218
ct_write_unlock_bh(hash);
net/netfilter/ipvs/ip_vs_conn.c
228
unsigned int hash;
net/netfilter/ipvs/ip_vs_conn.c
234
hash = ip_vs_conn_hashkey_conn(cp);
net/netfilter/ipvs/ip_vs_conn.c
236
ct_write_lock_bh(hash);
net/netfilter/ipvs/ip_vs_conn.c
249
ct_write_unlock_bh(hash);
net/netfilter/ipvs/ip_vs_conn.c
264
unsigned int hash;
net/netfilter/ipvs/ip_vs_conn.c
267
hash = ip_vs_conn_hashkey_param(p, false);
net/netfilter/ipvs/ip_vs_conn.c
271
hlist_for_each_entry_rcu(cp, &ip_vs_conn_tab[hash], c_list) {
net/netfilter/ipvs/ip_vs_conn.c
350
unsigned int hash;
net/netfilter/ipvs/ip_vs_conn.c
353
hash = ip_vs_conn_hashkey_param(p, false);
net/netfilter/ipvs/ip_vs_conn.c
357
hlist_for_each_entry_rcu(cp, &ip_vs_conn_tab[hash], c_list) {
net/netfilter/ipvs/ip_vs_conn.c
402
unsigned int hash;
net/netfilter/ipvs/ip_vs_conn.c
410
hash = ip_vs_conn_hashkey_param(p, true);
net/netfilter/ipvs/ip_vs_conn.c
414
hlist_for_each_entry_rcu(cp, &ip_vs_conn_tab[hash], c_list) {
net/netfilter/ipvs/ip_vs_ctl.c
347
unsigned int hash;
net/netfilter/ipvs/ip_vs_ctl.c
359
hash = ip_vs_svc_hashkey(svc->ipvs, svc->af, svc->protocol,
net/netfilter/ipvs/ip_vs_ctl.c
361
hlist_add_head_rcu(&svc->s_list, &ip_vs_svc_table[hash]);
net/netfilter/ipvs/ip_vs_ctl.c
366
hash = ip_vs_svc_fwm_hashkey(svc->ipvs, svc->fwmark);
net/netfilter/ipvs/ip_vs_ctl.c
367
hlist_add_head_rcu(&svc->f_list, &ip_vs_svc_fwm_table[hash]);
net/netfilter/ipvs/ip_vs_ctl.c
410
unsigned int hash;
net/netfilter/ipvs/ip_vs_ctl.c
414
hash = ip_vs_svc_hashkey(ipvs, af, protocol, vaddr, vport);
net/netfilter/ipvs/ip_vs_ctl.c
416
hlist_for_each_entry_rcu(svc, &ip_vs_svc_table[hash], s_list) {
net/netfilter/ipvs/ip_vs_ctl.c
437
unsigned int hash;
net/netfilter/ipvs/ip_vs_ctl.c
441
hash = ip_vs_svc_fwm_hashkey(ipvs, fwmark);
net/netfilter/ipvs/ip_vs_ctl.c
443
hlist_for_each_entry_rcu(svc, &ip_vs_svc_fwm_table[hash], f_list) {
net/netfilter/ipvs/ip_vs_ctl.c
560
unsigned int hash;
net/netfilter/ipvs/ip_vs_ctl.c
591
hash = ip_vs_rs_hashkey(dest->af, &dest->addr, port);
net/netfilter/ipvs/ip_vs_ctl.c
593
hlist_add_head_rcu(&dest->d_list, &ipvs->rs_table[hash]);
net/netfilter/ipvs/ip_vs_ctl.c
613
unsigned int hash;
net/netfilter/ipvs/ip_vs_ctl.c
617
hash = ip_vs_rs_hashkey(af, daddr, dport);
net/netfilter/ipvs/ip_vs_ctl.c
619
hlist_for_each_entry_rcu(dest, &ipvs->rs_table[hash], d_list) {
net/netfilter/ipvs/ip_vs_ctl.c
644
unsigned int hash;
net/netfilter/ipvs/ip_vs_ctl.c
648
hash = ip_vs_rs_hashkey(af, daddr, dport);
net/netfilter/ipvs/ip_vs_ctl.c
650
hlist_for_each_entry_rcu(dest, &ipvs->rs_table[hash], d_list) {
net/netfilter/ipvs/ip_vs_ctl.c
675
unsigned int hash;
net/netfilter/ipvs/ip_vs_ctl.c
678
hash = ip_vs_rs_hashkey(af, daddr, tun_port);
net/netfilter/ipvs/ip_vs_ctl.c
680
hlist_for_each_entry_rcu(dest, &ipvs->rs_table[hash], d_list) {
net/netfilter/ipvs/ip_vs_lblc.c
168
unsigned int hash = ip_vs_lblc_hashkey(en->af, &en->addr);
net/netfilter/ipvs/ip_vs_lblc.c
170
hlist_add_head_rcu(&en->list, &tbl->bucket[hash]);
net/netfilter/ipvs/ip_vs_lblc.c
180
unsigned int hash = ip_vs_lblc_hashkey(af, addr);
net/netfilter/ipvs/ip_vs_lblc.c
183
hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list)
net/netfilter/ipvs/ip_vs_lblcr.c
331
unsigned int hash = ip_vs_lblcr_hashkey(en->af, &en->addr);
net/netfilter/ipvs/ip_vs_lblcr.c
333
hlist_add_head_rcu(&en->list, &tbl->bucket[hash]);
net/netfilter/ipvs/ip_vs_lblcr.c
343
unsigned int hash = ip_vs_lblcr_hashkey(af, addr);
net/netfilter/ipvs/ip_vs_lblcr.c
346
hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list)
net/netfilter/ipvs/ip_vs_mh.c
237
unsigned int hash = ip_vs_mh_hashkey(svc->af, addr, port, &s->hash1, 0)
net/netfilter/ipvs/ip_vs_mh.c
239
struct ip_vs_dest *dest = rcu_dereference(s->lookup[hash].dest);
net/netfilter/ipvs/ip_vs_mh.c
250
unsigned int hash, ihash;
net/netfilter/ipvs/ip_vs_mh.c
270
hash = ip_vs_mh_hashkey(svc->af, addr, port, &s->hash1,
net/netfilter/ipvs/ip_vs_mh.c
272
dest = rcu_dereference(s->lookup[hash].dest);
net/netfilter/ipvs/ip_vs_proto.c
100
pp_p = &ip_vs_proto_table[hash];
net/netfilter/ipvs/ip_vs_proto.c
120
unsigned int hash = IP_VS_PROTO_HASH(pd->pp->protocol);
net/netfilter/ipvs/ip_vs_proto.c
122
pd_p = &ipvs->proto_data_table[hash];
net/netfilter/ipvs/ip_vs_proto.c
142
unsigned int hash = IP_VS_PROTO_HASH(proto);
net/netfilter/ipvs/ip_vs_proto.c
144
for (pp = ip_vs_proto_table[hash]; pp; pp = pp->next) {
net/netfilter/ipvs/ip_vs_proto.c
160
unsigned int hash = IP_VS_PROTO_HASH(proto);
net/netfilter/ipvs/ip_vs_proto.c
162
for (pd = ipvs->proto_data_table[hash]; pd; pd = pd->next) {
net/netfilter/ipvs/ip_vs_proto.c
50
unsigned int hash = IP_VS_PROTO_HASH(pp->protocol);
net/netfilter/ipvs/ip_vs_proto.c
52
pp->next = ip_vs_proto_table[hash];
net/netfilter/ipvs/ip_vs_proto.c
53
ip_vs_proto_table[hash] = pp;
net/netfilter/ipvs/ip_vs_proto.c
67
unsigned int hash = IP_VS_PROTO_HASH(pp->protocol);
net/netfilter/ipvs/ip_vs_proto.c
75
pd->next = ipvs->proto_data_table[hash];
net/netfilter/ipvs/ip_vs_proto.c
76
ipvs->proto_data_table[hash] = pd;
net/netfilter/ipvs/ip_vs_proto.c
83
ipvs->proto_data_table[hash] = pd->next;
net/netfilter/ipvs/ip_vs_proto.c
98
unsigned int hash = IP_VS_PROTO_HASH(pp->protocol);
net/netfilter/ipvs/ip_vs_proto_sctp.c
487
__u16 hash;
net/netfilter/ipvs/ip_vs_proto_sctp.c
492
hash = sctp_app_hashkey(port);
net/netfilter/ipvs/ip_vs_proto_sctp.c
494
list_for_each_entry(i, &ipvs->sctp_apps[hash], p_list) {
net/netfilter/ipvs/ip_vs_proto_sctp.c
500
list_add_rcu(&inc->p_list, &ipvs->sctp_apps[hash]);
net/netfilter/ipvs/ip_vs_proto_sctp.c
518
int hash;
net/netfilter/ipvs/ip_vs_proto_sctp.c
526
hash = sctp_app_hashkey(cp->vport);
net/netfilter/ipvs/ip_vs_proto_sctp.c
528
list_for_each_entry_rcu(inc, &ipvs->sctp_apps[hash], p_list) {
net/netfilter/ipvs/ip_vs_proto_tcp.c
611
__u16 hash;
net/netfilter/ipvs/ip_vs_proto_tcp.c
616
hash = tcp_app_hashkey(port);
net/netfilter/ipvs/ip_vs_proto_tcp.c
618
list_for_each_entry(i, &ipvs->tcp_apps[hash], p_list) {
net/netfilter/ipvs/ip_vs_proto_tcp.c
624
list_add_rcu(&inc->p_list, &ipvs->tcp_apps[hash]);
net/netfilter/ipvs/ip_vs_proto_tcp.c
646
int hash;
net/netfilter/ipvs/ip_vs_proto_tcp.c
655
hash = tcp_app_hashkey(cp->vport);
net/netfilter/ipvs/ip_vs_proto_tcp.c
657
list_for_each_entry_rcu(inc, &ipvs->tcp_apps[hash], p_list) {
net/netfilter/ipvs/ip_vs_proto_udp.c
357
__u16 hash;
net/netfilter/ipvs/ip_vs_proto_udp.c
362
hash = udp_app_hashkey(port);
net/netfilter/ipvs/ip_vs_proto_udp.c
364
list_for_each_entry(i, &ipvs->udp_apps[hash], p_list) {
net/netfilter/ipvs/ip_vs_proto_udp.c
370
list_add_rcu(&inc->p_list, &ipvs->udp_apps[hash]);
net/netfilter/ipvs/ip_vs_proto_udp.c
391
int hash;
net/netfilter/ipvs/ip_vs_proto_udp.c
400
hash = udp_app_hashkey(cp->vport);
net/netfilter/ipvs/ip_vs_proto_udp.c
402
list_for_each_entry_rcu(inc, &ipvs->udp_apps[hash], p_list) {
net/netfilter/ipvs/ip_vs_sh.c
106
unsigned int hash = ip_vs_sh_hashkey(svc->af, addr, port, 0);
net/netfilter/ipvs/ip_vs_sh.c
107
struct ip_vs_dest *dest = rcu_dereference(s->buckets[hash].dest);
net/netfilter/ipvs/ip_vs_sh.c
124
unsigned int hash, ihash;
net/netfilter/ipvs/ip_vs_sh.c
143
hash = ip_vs_sh_hashkey(svc->af, addr, port, roffset);
net/netfilter/ipvs/ip_vs_sh.c
144
dest = rcu_dereference(s->buckets[hash].dest);
net/netfilter/nf_conncount.c
400
unsigned int hash,
net/netfilter/nf_conncount.c
413
spin_lock_bh(&nf_conncount_locks[hash]);
net/netfilter/nf_conncount.c
448
schedule_gc_worker(data, hash);
net/netfilter/nf_conncount.c
483
spin_unlock_bh(&nf_conncount_locks[hash]);
net/netfilter/nf_conncount.c
497
unsigned int hash;
net/netfilter/nf_conncount.c
499
hash = jhash2(key, data->keylen, conncount_rnd) % CONNCOUNT_SLOTS;
net/netfilter/nf_conncount.c
500
root = &data->root[hash];
net/netfilter/nf_conncount.c
547
return insert_tree(net, skb, l3num, data, root, hash, key);
net/netfilter/nf_conntrack_bpf.c
176
struct nf_conntrack_tuple_hash *hash;
net/netfilter/nf_conntrack_bpf.c
218
hash = nf_conntrack_find_get(net, &ct_zone, &tuple);
net/netfilter/nf_conntrack_bpf.c
221
if (!hash)
net/netfilter/nf_conntrack_bpf.c
224
ct = nf_ct_tuplehash_to_ctrack(hash);
net/netfilter/nf_conntrack_bpf.c
225
opts->dir = NF_CT_DIRECTION(hash);
net/netfilter/nf_conntrack_core.c
1207
unsigned int hash, reply_hash;
net/netfilter/nf_conntrack_core.c
1232
hash = *(unsigned long *)&ct->tuplehash[IP_CT_DIR_REPLY].hnnode.pprev;
net/netfilter/nf_conntrack_core.c
1233
hash = scale_hash(hash);
net/netfilter/nf_conntrack_core.c
1237
} while (nf_conntrack_double_lock(hash, reply_hash, sequence));
net/netfilter/nf_conntrack_core.c
1251
nf_conntrack_double_unlock(hash, reply_hash);
net/netfilter/nf_conntrack_core.c
1275
hlist_nulls_for_each_entry(h, n, &nf_conntrack_hash[hash], hnnode) {
net/netfilter/nf_conntrack_core.c
1308
__nf_conntrack_hash_insert(ct, hash, reply_hash);
net/netfilter/nf_conntrack_core.c
1320
nf_conntrack_double_unlock(hash, reply_hash);
net/netfilter/nf_conntrack_core.c
1344
nf_conntrack_double_unlock(hash, reply_hash);
net/netfilter/nf_conntrack_core.c
1360
unsigned int hash, hsize;
net/netfilter/nf_conntrack_core.c
1369
hash = __hash_conntrack(net, tuple, nf_ct_zone_id(zone, IP_CT_DIR_REPLY), hsize);
net/netfilter/nf_conntrack_core.c
1371
hlist_nulls_for_each_entry_rcu(h, n, &ct_hash[hash], hnnode) {
net/netfilter/nf_conntrack_core.c
1406
if (get_nulls_value(n) != hash) {
net/netfilter/nf_conntrack_core.c
1466
static noinline int early_drop(struct net *net, unsigned int hash)
net/netfilter/nf_conntrack_core.c
1477
bucket = reciprocal_scale(hash, hsize);
net/netfilter/nf_conntrack_core.c
1662
gfp_t gfp, u32 hash)
net/netfilter/nf_conntrack_core.c
1672
if (!early_drop(net, hash)) {
net/netfilter/nf_conntrack_core.c
1698
*(unsigned long *)(&ct->tuplehash[IP_CT_DIR_REPLY].hnnode.pprev) = hash;
net/netfilter/nf_conntrack_core.c
1762
unsigned int dataoff, u32 hash)
net/netfilter/nf_conntrack_core.c
1781
hash);
net/netfilter/nf_conntrack_core.c
1875
u32 hash, zone_id, rid;
net/netfilter/nf_conntrack_core.c
1887
hash = hash_conntrack_raw(&tuple, zone_id, state->net);
net/netfilter/nf_conntrack_core.c
1888
h = __nf_conntrack_find_get(state->net, zone, &tuple, hash);
net/netfilter/nf_conntrack_core.c
1901
skb, dataoff, hash);
net/netfilter/nf_conntrack_core.c
2276
const struct nf_conntrack_tuple_hash *hash;
net/netfilter/nf_conntrack_core.c
228
static u32 scale_hash(u32 hash)
net/netfilter/nf_conntrack_core.c
2293
hash = nf_conntrack_find_get(dev_net(skb->dev),
net/netfilter/nf_conntrack_core.c
2296
if (!hash)
net/netfilter/nf_conntrack_core.c
2299
ct = nf_ct_tuplehash_to_ctrack(hash);
net/netfilter/nf_conntrack_core.c
230
return reciprocal_scale(hash, nf_conntrack_htable_size);
net/netfilter/nf_conntrack_core.c
2300
src_tuple = nf_ct_tuple(ct, !hash->tuple.dst.dir);
net/netfilter/nf_conntrack_core.c
2526
struct hlist_nulls_head *hash;
net/netfilter/nf_conntrack_core.c
2538
hash = kvzalloc_objs(struct hlist_nulls_head, nr_slots);
net/netfilter/nf_conntrack_core.c
2540
if (hash && nulls)
net/netfilter/nf_conntrack_core.c
2542
INIT_HLIST_NULLS_HEAD(&hash[i], i);
net/netfilter/nf_conntrack_core.c
2544
return hash;
net/netfilter/nf_conntrack_core.c
2552
struct hlist_nulls_head *hash, *old_hash;
net/netfilter/nf_conntrack_core.c
2559
hash = nf_ct_alloc_hashtable(&hashsize, 1);
net/netfilter/nf_conntrack_core.c
2560
if (!hash)
net/netfilter/nf_conntrack_core.c
2567
kvfree(hash);
net/netfilter/nf_conntrack_core.c
2593
hlist_nulls_add_head_rcu(&h->hnnode, &hash[bucket]);
net/netfilter/nf_conntrack_core.c
2598
nf_conntrack_hash = hash;
net/netfilter/nf_conntrack_core.c
605
unsigned int hash, reply_hash;
net/netfilter/nf_conntrack_core.c
610
hash = hash_conntrack(net,
net/netfilter/nf_conntrack_core.c
616
} while (nf_conntrack_double_lock(hash, reply_hash, sequence));
net/netfilter/nf_conntrack_core.c
619
nf_conntrack_double_unlock(hash, reply_hash);
net/netfilter/nf_conntrack_core.c
736
const struct nf_conntrack_tuple *tuple, u32 hash)
net/netfilter/nf_conntrack_core.c
745
bucket = reciprocal_scale(hash, hsize);
net/netfilter/nf_conntrack_core.c
775
const struct nf_conntrack_tuple *tuple, u32 hash)
net/netfilter/nf_conntrack_core.c
780
h = ____nf_conntrack_find(net, zone, tuple, hash);
net/netfilter/nf_conntrack_core.c
830
unsigned int hash,
net/netfilter/nf_conntrack_core.c
834
&nf_conntrack_hash[hash]);
net/netfilter/nf_conntrack_core.c
871
unsigned int hash, reply_hash;
net/netfilter/nf_conntrack_core.c
887
hash = hash_conntrack(net,
net/netfilter/nf_conntrack_core.c
893
} while (nf_conntrack_double_lock(hash, reply_hash, sequence));
net/netfilter/nf_conntrack_core.c
898
hlist_nulls_for_each_entry(h, n, &nf_conntrack_hash[hash], hnnode) {
net/netfilter/nf_conntrack_core.c
932
__nf_conntrack_hash_insert(ct, hash, reply_hash);
net/netfilter/nf_conntrack_core.c
933
nf_conntrack_double_unlock(hash, reply_hash);
net/netfilter/nf_conntrack_core.c
942
nf_conntrack_double_unlock(hash, reply_hash);
net/netfilter/nf_conntrack_expect.c
104
hash = siphash(&combined, sizeof(combined), &nf_ct_expect_hashrnd);
net/netfilter/nf_conntrack_expect.c
106
return reciprocal_scale(hash, nf_ct_expect_hsize);
net/netfilter/nf_conntrack_expect.c
92
u32 hash;
net/netfilter/nf_conntrack_standalone.c
113
hlist_nulls_for_each_entry_rcu(h, n, &st->hash[i], hnnode) {
net/netfilter/nf_conntrack_standalone.c
158
nf_conntrack_get_ht(&st->hash, &st->htable_size);
net/netfilter/nf_conntrack_standalone.c
306
struct nf_conntrack_tuple_hash *hash = v;
net/netfilter/nf_conntrack_standalone.c
307
struct nf_conn *ct = nf_ct_tuplehash_to_ctrack(hash);
net/netfilter/nf_conntrack_standalone.c
328
if (NF_CT_DIRECTION(hash))
net/netfilter/nf_conntrack_standalone.c
95
struct hlist_nulls_head *hash;
net/netfilter/nf_nat_core.c
156
unsigned int hash;
net/netfilter/nf_nat_core.c
177
hash = siphash(&combined, sizeof(combined), &nf_nat_hash_rnd);
net/netfilter/nf_nat_core.c
179
return reciprocal_scale(hash, nf_nat_htable_size);
net/netfilter/nft_set_hash.c
607
u32 hash;
net/netfilter/nft_set_hash.c
609
hash = jhash(key, set->klen, priv->seed);
net/netfilter/nft_set_hash.c
610
hash = reciprocal_scale(hash, priv->buckets);
net/netfilter/nft_set_hash.c
611
hlist_for_each_entry_rcu(he, &priv->table[hash], node) {
net/netfilter/nft_set_hash.c
627
u32 hash;
net/netfilter/nft_set_hash.c
630
hash = jhash_1word(*key, priv->seed);
net/netfilter/nft_set_hash.c
632
hash = jhash(key, set->klen, priv->seed);
net/netfilter/nft_set_hash.c
634
hash = reciprocal_scale(hash, priv->buckets);
net/netfilter/nft_set_hash.c
635
hlist_for_each_entry_rcu(he, &priv->table[hash], node) {
net/netfilter/nft_set_hash.c
651
u32 hash, k1, k2;
net/netfilter/nft_set_hash.c
654
hash = jhash_1word(k1, priv->seed);
net/netfilter/nft_set_hash.c
655
hash = reciprocal_scale(hash, priv->buckets);
net/netfilter/nft_set_hash.c
656
hlist_for_each_entry_rcu(he, &priv->table[hash], node) {
net/netfilter/nft_set_hash.c
669
u32 hash, k1;
net/netfilter/nft_set_hash.c
673
hash = jhash_1word(k1, priv->seed);
net/netfilter/nft_set_hash.c
675
hash = jhash(key, set->klen, priv->seed);
net/netfilter/nft_set_hash.c
677
hash = reciprocal_scale(hash, priv->buckets);
net/netfilter/nft_set_hash.c
679
return hash;
net/netfilter/nft_set_hash.c
689
u32 hash;
net/netfilter/nft_set_hash.c
691
hash = nft_jhash(set, priv, &this->ext);
net/netfilter/nft_set_hash.c
692
hlist_for_each_entry(he, &priv->table[hash], node) {
net/netfilter/nft_set_hash.c
700
hlist_add_head_rcu(&this->node, &priv->table[hash]);
net/netfilter/nft_set_hash.c
728
u32 hash;
net/netfilter/nft_set_hash.c
730
hash = nft_jhash(set, priv, &this->ext);
net/netfilter/nft_set_hash.c
731
hlist_for_each_entry(he, &priv->table[hash], node) {
net/netfilter/xt_HMARK.c
118
u32 hash;
net/netfilter/xt_HMARK.c
125
hash = jhash_3words(src, dst, t->uports.v32, info->hashrnd);
net/netfilter/xt_HMARK.c
126
hash = hash ^ (t->proto & info->proto_mask);
net/netfilter/xt_HMARK.c
128
return reciprocal_scale(hash, info->hmodulus) + info->hoffset;
net/netfilter/xt_RATEEST.c
215
for (i = 0; i < ARRAY_SIZE(xn->hash); i++)
net/netfilter/xt_RATEEST.c
216
INIT_HLIST_HEAD(&xn->hash[i]);
net/netfilter/xt_RATEEST.c
24
struct hlist_head hash[RATEEST_HSIZE];
net/netfilter/xt_RATEEST.c
43
hlist_add_head(&est->list, &xn->hash[h]);
net/netfilter/xt_RATEEST.c
53
hlist_for_each_entry(est, &xn->hash[h], list) {
net/netfilter/xt_cluster.c
114
hash = xt_cluster_hash(ct->master, info);
net/netfilter/xt_cluster.c
116
hash = xt_cluster_hash(ct, info);
net/netfilter/xt_cluster.c
118
return !!((1 << hash) & info->node_mask) ^
net/netfilter/xt_cluster.c
42
u_int32_t hash = 0;
net/netfilter/xt_cluster.c
46
hash = xt_cluster_hash_ipv4(nf_ct_orig_ipv4_src(ct), info);
net/netfilter/xt_cluster.c
49
hash = xt_cluster_hash_ipv6(nf_ct_orig_ipv6_src(ct), info);
net/netfilter/xt_cluster.c
56
return reciprocal_scale(hash, info->total_nodes);
net/netfilter/xt_cluster.c
85
unsigned long hash;
net/netfilter/xt_hashlimit.c
1178
if (!hlist_empty(&htable->hash[*bucket])) {
net/netfilter/xt_hashlimit.c
1179
hlist_for_each_entry(ent, &htable->hash[*bucket], node)
net/netfilter/xt_hashlimit.c
1192
if (!hlist_empty(&htable->hash[*bucket])) {
net/netfilter/xt_hashlimit.c
1193
hlist_for_each_entry(ent, &htable->hash[*bucket], node)
net/netfilter/xt_hashlimit.c
1206
if (!hlist_empty(&htable->hash[*bucket])) {
net/netfilter/xt_hashlimit.c
1207
hlist_for_each_entry(ent, &htable->hash[*bucket], node)
net/netfilter/xt_hashlimit.c
134
struct hlist_head hash[]; /* hashtable itself */
net/netfilter/xt_hashlimit.c
185
u_int32_t hash = jhash2((const u32 *)dst,
net/netfilter/xt_hashlimit.c
194
return reciprocal_scale(hash, ht->cfg.size);
net/netfilter/xt_hashlimit.c
202
u_int32_t hash = hash_dst(ht, dst);
net/netfilter/xt_hashlimit.c
204
if (!hlist_empty(&ht->hash[hash])) {
net/netfilter/xt_hashlimit.c
205
hlist_for_each_entry_rcu(ent, &ht->hash[hash], node)
net/netfilter/xt_hashlimit.c
251
hlist_add_head_rcu(&ent->node, &ht->hash[hash_dst(ht, dst)]);
net/netfilter/xt_hashlimit.c
296
hinfo = kvmalloc_flex(*hinfo, hash, size);
net/netfilter/xt_hashlimit.c
315
INIT_HLIST_HEAD(&hinfo->hash[i]);
net/netfilter/xt_hashlimit.c
364
struct hlist_head *head = &ht->hash[i];
net/netlink/af_netlink.c
2619
rhashtable_walk_enter(&nl_table[iter->link].hash, &iter->hti);
net/netlink/af_netlink.c
2935
if (rhashtable_init(&nl_table[i].hash,
net/netlink/af_netlink.c
498
return rhashtable_lookup_fast(&table->hash, &arg,
net/netlink/af_netlink.c
507
return rhashtable_lookup_insert_key(&table->hash, &arg,
net/netlink/af_netlink.c
598
if (!rhashtable_remove_fast(&table->hash, &nlk_sk(sk)->node,
net/netlink/af_netlink.h
62
struct rhashtable hash;
net/netlink/diag.c
118
rhashtable_walk_enter(&tbl->hash, hti);
net/openvswitch/actions.c
1003
u32 hash = 0;
net/openvswitch/actions.c
1007
hash = skb_get_hash(skb);
net/openvswitch/actions.c
1012
hash = __skb_get_hash_symmetric(skb);
net/openvswitch/actions.c
1015
hash = jhash_1word(hash, hash_act->hash_basis);
net/openvswitch/actions.c
1016
if (!hash)
net/openvswitch/actions.c
1017
hash = 0x1;
net/openvswitch/actions.c
1019
key->ovs_flow_hash = hash;
net/openvswitch/datapath.c
463
u64 hash;
net/openvswitch/datapath.c
570
hash = skb_get_hash_raw(skb);
net/openvswitch/datapath.c
572
hash |= OVS_PACKET_HASH_SW_BIT;
net/openvswitch/datapath.c
575
hash |= OVS_PACKET_HASH_L4_BIT;
net/openvswitch/datapath.c
577
if (nla_put(user_skb, OVS_PACKET_ATTR_HASH, sizeof (u64), &hash)) {
net/openvswitch/datapath.c
622
u64 hash;
net/openvswitch/datapath.c
649
hash = nla_get_u64(a[OVS_PACKET_ATTR_HASH]);
net/openvswitch/datapath.c
651
__skb_set_hash(packet, hash & 0xFFFFFFFFULL,
net/openvswitch/datapath.c
652
!!(hash & OVS_PACKET_HASH_SW_BIT),
net/openvswitch/datapath.c
653
!!(hash & OVS_PACKET_HASH_L4_BIT));
net/openvswitch/flow.h
225
u32 hash;
net/openvswitch/flow_table.c
1038
flow->flow_table.hash = flow_hash(&flow->key, &flow->mask->range);
net/openvswitch/flow_table.c
1061
flow->ufid_table.hash = ufid_hash(&flow->id);
net/openvswitch/flow_table.c
550
static struct hlist_head *find_bucket(struct table_instance *ti, u32 hash)
net/openvswitch/flow_table.c
552
hash = jhash_1word(hash, ti->hash_seed);
net/openvswitch/flow_table.c
553
return &ti->buckets[hash & (ti->n_buckets - 1)];
net/openvswitch/flow_table.c
561
head = find_bucket(ti, flow->flow_table.hash);
net/openvswitch/flow_table.c
570
head = find_bucket(ti, flow->ufid_table.hash);
net/openvswitch/flow_table.c
703
u32 hash;
net/openvswitch/flow_table.c
707
hash = flow_hash(&masked_key, &mask->range);
net/openvswitch/flow_table.c
708
head = find_bucket(ti, hash);
net/openvswitch/flow_table.c
713
if (flow->mask == mask && flow->flow_table.hash == hash &&
net/openvswitch/flow_table.c
792
u32 hash;
net/openvswitch/flow_table.c
812
hash = skb_hash;
net/openvswitch/flow_table.c
817
int index = hash & (mc->cache_size - 1);
net/openvswitch/flow_table.c
832
hash >>= MC_HASH_SHIFT;
net/openvswitch/flow_table.c
921
u32 hash;
net/openvswitch/flow_table.c
923
hash = ufid_hash(ufid);
net/openvswitch/flow_table.c
924
head = find_bucket(ti, hash);
net/openvswitch/flow_table.c
927
if (flow->ufid_table.hash == hash &&
net/openvswitch/meter.c
119
u32 hash;
net/openvswitch/meter.c
121
hash = meter_hash(ti, meter->id);
net/openvswitch/meter.c
122
rcu_assign_pointer(ti->dp_meters[hash], meter);
net/openvswitch/meter.c
128
u32 hash;
net/openvswitch/meter.c
130
hash = meter_hash(ti, meter->id);
net/openvswitch/meter.c
131
RCU_INIT_POINTER(ti->dp_meters[hash], NULL);
net/openvswitch/meter.c
137
u32 hash = meter_hash(ti, meter->id);
net/openvswitch/meter.c
143
if (unlikely(rcu_dereference_ovsl(ti->dp_meters[hash])))
net/openvswitch/meter.c
58
u32 hash = meter_hash(ti, meter_id);
net/openvswitch/meter.c
61
meter = rcu_dereference_ovsl(ti->dp_meters[hash]);
net/openvswitch/vport.c
471
u32 hash;
net/openvswitch/vport.c
479
hash = skb_get_hash(skb);
net/openvswitch/vport.c
480
ids_index = hash - ids->n_ids * reciprocal_divide(hash, ids->rn_ids);
net/openvswitch/vport.c
56
unsigned int hash = jhash(name, strlen(name), (unsigned long) net);
net/openvswitch/vport.c
57
return &dev_table[hash & (VPORT_HASH_BUCKETS - 1)];
net/phonet/datagram.c
174
.hash = pn_sock_hash,
net/phonet/pep.c
1360
.hash = pn_sock_hash,
net/phonet/socket.c
194
err = sk->sk_prot->hash(sk);
net/psp/psp_main.c
191
u32 hash = sk->sk_hash;
net/psp/psp_main.c
204
hash ^= hash << 16;
net/psp/psp_main.c
205
uh->source = htons((((u64)hash * (max - min)) >> 32) + min);
net/rds/connection.c
61
u32 hash;
net/rds/connection.c
72
hash = __inet_ehashfn(lhash, 0, fhash, 0, rds_hash_secret);
net/rds/connection.c
74
return &rds_conn_hash[hash & RDS_CONNECTION_HASH_MASK];
net/sched/act_skbedit.c
32
u32 hash = skb_get_hash(skb);
net/sched/act_skbedit.c
34
queue_mapping += hash % params->mapping_mod;
net/sched/cls_api.c
200
u32 hash = destroy_obj_hashfn(tp);
net/sched/cls_api.c
206
destroy_ht_node, hash) {
net/sched/cls_flower.c
2090
fl_set_key_val(tb, &key->hash.hash, TCA_FLOWER_KEY_HASH,
net/sched/cls_flower.c
2091
&mask->hash.hash, TCA_FLOWER_KEY_HASH_MASK,
net/sched/cls_flower.c
2092
sizeof(key->hash.hash));
net/sched/cls_flower.c
2219
FLOW_DISSECTOR_KEY_HASH, hash);
net/sched/cls_flower.c
3670
if (fl_dump_key_val(skb, &key->hash.hash, TCA_FLOWER_KEY_HASH,
net/sched/cls_flower.c
3671
&mask->hash.hash, TCA_FLOWER_KEY_HASH_MASK,
net/sched/cls_flower.c
3672
sizeof(key->hash.hash)))
net/sched/cls_flower.c
82
struct flow_dissector_key_hash hash;
net/sched/sch_api.c
1856
hash_for_each(qdisc_dev(root)->qdisc_hash, b, q, hash) {
net/sched/sch_api.c
2383
hash_for_each(qdisc_dev(root)->qdisc_hash, b, q, hash) {
net/sched/sch_api.c
274
hash_for_each_possible_rcu(qdisc_dev(root)->qdisc_hash, q, hash, handle,
net/sched/sch_api.c
286
hash_add_rcu(qdisc_dev(q)->qdisc_hash, &q->hash, q->handle);
net/sched/sch_api.c
297
hash_del_rcu(&q->hash);
net/sched/sch_api.c
697
ohash = clhash->hash;
net/sched/sch_api.c
707
clhash->hash = nhash;
net/sched/sch_api.c
720
clhash->hash = qdisc_class_hash_alloc(size);
net/sched/sch_api.c
721
if (!clhash->hash)
net/sched/sch_api.c
732
kvfree(clhash->hash);
net/sched/sch_api.c
743
hlist_add_head(&cl->hnode, &clhash->hash[h]);
net/sched/sch_cake.c
1723
goto hash;
net/sched/sch_cake.c
1745
hash:
net/sched/sch_cake.c
781
flow_hash = skb->hash;
net/sched/sch_drr.c
293
hlist_for_each_entry(cl, &q->clhash.hash[i], common.hnode) {
net/sched/sch_drr.c
436
hlist_for_each_entry(cl, &q->clhash.hash[i], common.hnode) {
net/sched/sch_drr.c
454
hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
net/sched/sch_fq.c
372
unsigned long hash = skb_get_hash(skb) & q->orphan_mask;
net/sched/sch_fq.c
377
sk = (struct sock *)((hash << 1) | 1UL);
net/sched/sch_fq.c
380
unsigned long hash = skb_get_hash(skb) & q->orphan_mask;
net/sched/sch_fq.c
389
sk = (struct sock *)((hash << 1) | 1UL);
net/sched/sch_hfsc.c
1367
hlist_for_each_entry(cl, &q->clhash.hash[i],
net/sched/sch_hfsc.c
1490
hlist_for_each_entry(cl, &q->clhash.hash[i], cl_common.hnode)
net/sched/sch_hfsc.c
1506
hlist_for_each_entry(cl, &q->clhash.hash[i], cl_common.hnode) {
net/sched/sch_hfsc.c
1512
hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
net/sched/sch_hhf.c
180
static struct hh_flow_state *seek_list(const u32 hash,
net/sched/sch_hhf.c
202
} else if (flow->hash_id == hash) {
net/sched/sch_hhf.c
250
u32 tmp_hash, hash;
net/sched/sch_hhf.c
267
hash = skb_get_hash_perturb(skb, &q->perturbation);
net/sched/sch_hhf.c
270
flow_pos = hash & HHF_BIT_MASK;
net/sched/sch_hhf.c
271
flow = seek_list(hash, &q->hh_flows[flow_pos], q);
net/sched/sch_hhf.c
278
tmp_hash = hash;
net/sched/sch_hhf.c
310
flow->hash_id = hash;
net/sched/sch_htb.c
1007
hlist_for_each_entry(cl, &q->clhash.hash[i], common.hnode) {
net/sched/sch_htb.c
1301
hlist_for_each_entry(c, &q->clhash.hash[i], common.hnode) {
net/sched/sch_htb.c
1634
hlist_for_each_entry(cl, &q->clhash.hash[i], common.hnode) {
net/sched/sch_htb.c
1644
hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
net/sched/sch_htb.c
2116
hlist_for_each_entry(cl, &q->clhash.hash[i], common.hnode) {
net/sched/sch_netem.c
997
hash_for_each(qdisc_dev(root)->qdisc_hash, i, q, hash) {
net/sched/sch_qfq.c
1485
hlist_for_each_entry(cl, &q->clhash.hash[i], common.hnode) {
net/sched/sch_qfq.c
1504
hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
net/sched/sch_qfq.c
687
hlist_for_each_entry(cl, &q->clhash.hash[i], common.hnode) {
net/sched/sch_sfb.c
129
u32 hash = sfbhash & SFB_BUCKET_MASK;
net/sched/sch_sfb.c
132
if (b[hash].qlen < 0xFFFF)
net/sched/sch_sfb.c
133
b[hash].qlen++;
net/sched/sch_sfb.c
158
u32 hash = sfbhash & SFB_BUCKET_MASK;
net/sched/sch_sfb.c
161
if (b[hash].qlen > 0)
net/sched/sch_sfb.c
162
b[hash].qlen--;
net/sched/sch_sfb.c
332
u32 hash = sfbhash & SFB_BUCKET_MASK;
net/sched/sch_sfb.c
333
struct sfb_bucket *b = &q->bins[slot].bins[i][hash];
net/sched/sch_sfb.c
365
u32 hash = sfbhash & SFB_BUCKET_MASK;
net/sched/sch_sfb.c
366
struct sfb_bucket *b = &q->bins[slot].bins[i][hash];
net/sched/sch_sfq.c
100
unsigned short hash; /* hash value (index in ht[]) */
net/sched/sch_sfq.c
317
q->ht[slot->hash] = SFQ_EMPTY_SLOT;
net/sched/sch_sfq.c
345
unsigned int hash, dropped;
net/sched/sch_sfq.c
352
hash = sfq_classify(skb, sch, &ret);
net/sched/sch_sfq.c
353
if (hash == 0) {
net/sched/sch_sfq.c
359
hash--;
net/sched/sch_sfq.c
361
x = q->ht[hash];
net/sched/sch_sfq.c
367
q->ht[hash] = x;
net/sched/sch_sfq.c
369
slot->hash = hash;
net/sched/sch_sfq.c
503
q->ht[slot->hash] = SFQ_EMPTY_SLOT;
net/sched/sch_sfq.c
554
q->ht[slot->hash] = SFQ_EMPTY_SLOT;
net/sched/sch_sfq.c
559
unsigned int hash = sfq_hash(q, skb);
net/sched/sch_sfq.c
560
sfq_index x = q->ht[hash];
net/sched/sch_sfq.c
573
q->ht[hash] = x;
net/sched/sch_sfq.c
575
slot->hash = hash;
net/sctp/input.c
850
int hash;
net/sctp/input.c
853
hash = sctp_ep_hashfn(net, ntohs(lport));
net/sctp/input.c
854
head = &sctp_ep_hashtable[hash];
net/sctp/proc.c
166
int hash = *(loff_t *)v;
net/sctp/proc.c
168
if (hash >= sctp_ep_hashsize)
net/sctp/proc.c
171
head = &sctp_ep_hashtable[hash];
net/sctp/proc.c
178
sctp_sk(sk)->type, sk->sk_state, hash,
net/sctp/socket.c
5364
int hash = 0;
net/sctp/socket.c
5368
for (head = sctp_ep_hashtable; hash < sctp_ep_hashsize;
net/sctp/socket.c
5369
hash++, head++) {
net/sctp/socket.c
9644
.hash = sctp_hash,
net/sctp/socket.c
9699
.hash = sctp_hash,
net/smc/af_smc.c
240
.hash = smc_hash_sk,
net/smc/af_smc.c
253
.hash = smc_hash_sk,
net/smc/af_smc.c
410
sk->sk_prot->hash(sk);
net/smc/smc_inet.c
25
.hash = smc_hash_sk,
net/smc/smc_inet.c
71
.hash = smc_hash_sk,
net/sunrpc/auth_gss/auth_gss.c
1034
INIT_HLIST_NODE(&gss_auth->hash);
net/sunrpc/auth_gss/auth_gss.c
1142
if (hash_hashed(&gss_auth->hash)) {
net/sunrpc/auth_gss/auth_gss.c
1144
hash_del(&gss_auth->hash);
net/sunrpc/auth_gss/auth_gss.c
1177
hash,
net/sunrpc/auth_gss/auth_gss.c
1196
hash_add(gss_auth_hash_table, &new->hash, hashval);
net/sunrpc/auth_gss/auth_gss.c
88
struct hlist_node hash;
net/sunrpc/auth_gss/svcauth_gss.c
324
int hash = rsi_hash(item);
net/sunrpc/auth_gss/svcauth_gss.c
326
ch = sunrpc_cache_lookup_rcu(cd, &item->h, hash);
net/sunrpc/auth_gss/svcauth_gss.c
336
int hash = rsi_hash(new);
net/sunrpc/auth_gss/svcauth_gss.c
339
&old->h, hash);
net/sunrpc/auth_gss/svcauth_gss.c
601
int hash = rsc_hash(item);
net/sunrpc/auth_gss/svcauth_gss.c
603
ch = sunrpc_cache_lookup_rcu(cd, &item->h, hash);
net/sunrpc/auth_gss/svcauth_gss.c
613
int hash = rsc_hash(new);
net/sunrpc/auth_gss/svcauth_gss.c
616
&old->h, hash);
net/sunrpc/cache.c
102
int hash)
net/sunrpc/cache.c
105
struct hlist_head *head = &detail->hash_table[hash];
net/sunrpc/cache.c
1369
unsigned int hash, entry;
net/sunrpc/cache.c
1375
hash = n >> 32;
net/sunrpc/cache.c
1378
hlist_for_each_entry_rcu(ch, &cd->hash_table[hash], cache_list)
net/sunrpc/cache.c
1383
hash++;
net/sunrpc/cache.c
1385
} while(hash < cd->hash_size &&
net/sunrpc/cache.c
1386
hlist_empty(&cd->hash_table[hash]));
net/sunrpc/cache.c
1387
if (hash >= cd->hash_size)
net/sunrpc/cache.c
1391
hlist_first_rcu(&cd->hash_table[hash])),
net/sunrpc/cache.c
1398
int hash = (*pos >> 32);
net/sunrpc/cache.c
1402
hash = 0;
net/sunrpc/cache.c
1404
hash++;
net/sunrpc/cache.c
1413
while (hash < cd->hash_size &&
net/sunrpc/cache.c
1414
hlist_empty(&cd->hash_table[hash])) {
net/sunrpc/cache.c
1415
hash++;
net/sunrpc/cache.c
1418
if (hash >= cd->hash_size)
net/sunrpc/cache.c
1422
hlist_first_rcu(&cd->hash_table[hash])),
net/sunrpc/cache.c
150
struct cache_head *key, int hash)
net/sunrpc/cache.c
154
ret = sunrpc_cache_find_rcu(detail, key, hash);
net/sunrpc/cache.c
158
return sunrpc_cache_add_entry(detail, key, hash);
net/sunrpc/cache.c
206
struct cache_head *new, struct cache_head *old, int hash)
net/sunrpc/cache.c
236
hlist_add_head(&tmp->cache_list, &detail->hash_table[hash]);
net/sunrpc/cache.c
600
hlist_del_init(&dreq->hash);
net/sunrpc/cache.c
609
int hash = DFR_HASH(item);
net/sunrpc/cache.c
612
hlist_add_head(&dreq->hash, &cache_defer_hash[hash]);
net/sunrpc/cache.c
64
int hash)
net/sunrpc/cache.c
66
struct hlist_head *head = &detail->hash_table[hash];
net/sunrpc/cache.c
664
if (!hlist_unhashed(&sleeper.handle.hash)) {
net/sunrpc/cache.c
748
int hash = DFR_HASH(item);
net/sunrpc/cache.c
753
hlist_for_each_entry_safe(dreq, tmp, &cache_defer_hash[hash], hash)
net/sunrpc/svcauth.c
226
hlist_del_rcu(&dom->hash);
net/sunrpc/svcauth.c
247
hlist_for_each_entry(hp, head, hash) {
net/sunrpc/svcauth.c
255
hlist_add_head_rcu(&new->hash, head);
net/sunrpc/svcauth.c
269
hlist_for_each_entry_rcu(hp, head, hash) {
net/sunrpc/svcauth.c
300
hlist_for_each_entry(hp, &auth_domain_table[h], hash)
net/tipc/monitor.c
188
hlist_for_each_entry(peer, &mon->peers[thash], hash) {
net/tipc/monitor.c
368
hlist_del(&peer->hash);
net/tipc/monitor.c
404
hlist_add_head(&p->hash, &mon->peers[tipc_hashfn(addr)]);
net/tipc/monitor.c
697
hlist_del(&peer->hash);
net/tipc/monitor.c
78
struct hlist_node hash;
net/tipc/name_table.c
1063
i = hash(*last_type);
net/tipc/name_table.c
275
hd = &nt->services[hash(ua->sr.type)];
net/tipc/name_table.c
470
service_head = &nt->services[hash(ua->sr.type)];
net/tipc/node.c
120
struct hlist_node hash;
net/tipc/node.c
337
hlist_for_each_entry_rcu(node, &tn->node_htable[thash], hash) {
net/tipc/node.c
501
hlist_del_rcu(&n->hash);
net/tipc/node.c
502
hlist_add_head_rcu(&n->hash,
net/tipc/node.c
563
INIT_HLIST_NODE(&n->hash);
net/tipc/node.c
595
hlist_add_head_rcu(&n->hash, &tn->node_htable[tipc_hashfn(addr)]);
net/tipc/node.c
632
hlist_del_rcu(&node->hash);
net/tls/tls_main.c
1042
prot[TLS_HW_RECORD][TLS_HW_RECORD].hash = tls_toe_hash;
net/tls/tls_toe.c
109
err = ctx->sk_proto->hash(sk);
net/tls/tls_toe.c
112
if (dev->hash) {
net/tls/tls_toe.c
115
err |= dev->hash(dev, sk);
net/unix/af_unix.c
1255
unsigned int hash = unix_abstract_hash(sunaddr, addr_len, type);
net/unix/af_unix.c
1259
sk = unix_find_socket_byname(net, sunaddr, addr_len, hash);
net/unix/af_unix.c
201
unsigned long hash = (unsigned long)sk;
net/unix/af_unix.c
203
hash ^= hash >> 16;
net/unix/af_unix.c
204
hash ^= hash >> 8;
net/unix/af_unix.c
205
hash ^= sk->sk_type;
net/unix/af_unix.c
207
return hash & UNIX_HASH_MOD;
net/unix/af_unix.c
219
unsigned int hash;
net/unix/af_unix.c
221
hash = (__force unsigned int)csum_fold(csum);
net/unix/af_unix.c
222
hash ^= hash >> 8;
net/unix/af_unix.c
223
hash ^= type;
net/unix/af_unix.c
225
return UNIX_HASH_MOD + 1 + (hash & UNIX_HASH_MOD);
net/unix/af_unix.c
384
struct unix_address *addr, unsigned int hash)
net/unix/af_unix.c
389
sk->sk_hash = hash;
net/unix/af_unix.c
427
int len, unsigned int hash)
net/unix/af_unix.c
431
sk_for_each(s, &net->unx.table.buckets[hash]) {
net/unix/af_unix.c
443
int len, unsigned int hash)
net/unix/af_unix.c
447
spin_lock(&net->unx.table.locks[hash]);
net/unix/af_unix.c
448
s = __unix_find_socket_byname(net, sunname, len, hash);
net/unix/af_unix.c
451
spin_unlock(&net->unx.table.locks[hash]);
net/unix/af_unix.c
457
unsigned int hash = unix_bsd_hash(i);
net/unix/af_unix.c
460
spin_lock(&bsd_socket_locks[hash]);
net/unix/af_unix.c
461
sk_for_each_bound(s, &bsd_socket_buckets[hash]) {
net/unix/af_unix.c
466
spin_unlock(&bsd_socket_locks[hash]);
net/unix/af_unix.c
470
spin_unlock(&bsd_socket_locks[hash]);
net/xfrm/xfrm_policy.c
541
unsigned int hash;
net/xfrm/xfrm_policy.c
546
hash = __sel_hash(sel, family, hmask, dbits, sbits);
net/xfrm/xfrm_policy.c
548
if (hash == hmask + 1)
net/xfrm/xfrm_policy.c
552
lockdep_is_held(&net->xfrm.xfrm_policy_lock)) + hash;
net/xfrm/xfrm_policy.c
561
unsigned int hash;
net/xfrm/xfrm_policy.c
566
hash = __addr_hash(daddr, saddr, family, hmask, dbits, sbits);
net/xfrm/xfrm_policy.c
569
lockdep_is_held(&net->xfrm.xfrm_policy_lock)) + hash;
samples/bpf/hash_func01.h
11
__u32 hash = initval;
samples/bpf/hash_func01.h
23
hash += get16bits (data);
samples/bpf/hash_func01.h
24
tmp = (get16bits (data+2) << 11) ^ hash;
samples/bpf/hash_func01.h
25
hash = (hash << 16) ^ tmp;
samples/bpf/hash_func01.h
27
hash += hash >> 11;
samples/bpf/hash_func01.h
32
case 3: hash += get16bits (data);
samples/bpf/hash_func01.h
33
hash ^= hash << 16;
samples/bpf/hash_func01.h
34
hash ^= ((signed char)data[sizeof (__u16)]) << 18;
samples/bpf/hash_func01.h
35
hash += hash >> 11;
samples/bpf/hash_func01.h
37
case 2: hash += get16bits (data);
samples/bpf/hash_func01.h
38
hash ^= hash << 11;
samples/bpf/hash_func01.h
39
hash += hash >> 17;
samples/bpf/hash_func01.h
41
case 1: hash += (signed char)*data;
samples/bpf/hash_func01.h
42
hash ^= hash << 10;
samples/bpf/hash_func01.h
43
hash += hash >> 1;
samples/bpf/hash_func01.h
47
hash ^= hash << 3;
samples/bpf/hash_func01.h
48
hash += hash >> 5;
samples/bpf/hash_func01.h
49
hash ^= hash << 4;
samples/bpf/hash_func01.h
50
hash += hash >> 17;
samples/bpf/hash_func01.h
51
hash ^= hash << 25;
samples/bpf/hash_func01.h
52
hash += hash >> 6;
samples/bpf/hash_func01.h
54
return hash;
samples/bpf/test_lwt_bpf.c
39
printk("len %d hash %d protocol %d", skb->len, skb->hash,
samples/tsm-mr/tsm_mr_sample.c
69
#define MR_(mr, hash) .mr_value = &sample_report.mr, TSM_MR_(mr, hash)
scripts/basic/fixdep.c
113
unsigned int hash;
scripts/basic/fixdep.c
123
unsigned int i, hash = 2166136261U;
scripts/basic/fixdep.c
126
hash = (hash ^ str[i]) * 0x01000193;
scripts/basic/fixdep.c
127
return hash;
scripts/basic/fixdep.c
133
static void add_to_hashtable(const char *name, int len, unsigned int hash,
scripts/basic/fixdep.c
141
aux->hash = hash;
scripts/basic/fixdep.c
142
aux->next = hashtab[hash % HASHSZ];
scripts/basic/fixdep.c
143
hashtab[hash % HASHSZ] = aux;
scripts/basic/fixdep.c
153
unsigned int hash = strhash(name, len);
scripts/basic/fixdep.c
155
for (aux = hashtab[hash % HASHSZ]; aux; aux = aux->next) {
scripts/basic/fixdep.c
156
if (aux->hash == hash && aux->len == len &&
scripts/basic/fixdep.c
161
add_to_hashtable(name, len, hash, hashtab);
scripts/gcc-plugins/randomize_layout_plugin.c
47
unsigned long hash = 0;
scripts/gcc-plugins/randomize_layout_plugin.c
50
hash = partial_name_hash(*name++, hash);
scripts/gcc-plugins/randomize_layout_plugin.c
51
return (unsigned int)hash;
scripts/gendwarfksyms/cache.c
11
struct hlist_node hash;
scripts/gendwarfksyms/cache.c
21
hash_add(cache->cache, &ci->hash, hash_32(key));
scripts/gendwarfksyms/cache.c
28
hash_for_each_possible(cache->cache, ci, hash, hash_32(key)) {
scripts/gendwarfksyms/cache.c
46
hash_for_each_safe(cache->cache, ci, tmp, hash) {
scripts/gendwarfksyms/die.c
106
hash_for_each_safe(die_map, cd, tmp, hash) {
scripts/gendwarfksyms/die.c
40
hash_add(die_map, &cd->hash, die_hash(cd->addr, state));
scripts/gendwarfksyms/die.c
48
hash_for_each_possible(die_map, cd, hash, die_hash(addr, state)) {
scripts/gendwarfksyms/die.c
92
hash_for_each_safe(die_map, cd, tmp, hash) {
scripts/gendwarfksyms/gendwarfksyms.h
190
struct hlist_node hash;
scripts/gendwarfksyms/kabi.c
237
hash_add(rules, &rule->hash, rule_hash(rule));
scripts/gendwarfksyms/kabi.c
268
hash_for_each_possible(rules, rule, hash,
scripts/gendwarfksyms/kabi.c
366
hash_for_each_safe(rules, rule, tmp, hash) {
scripts/gendwarfksyms/kabi.c
86
struct hlist_node hash;
scripts/gendwarfksyms/types.c
109
hash_for_each_possible(type_map, e, hash, hash_str(name)) {
scripts/gendwarfksyms/types.c
129
hash_add(type_map, &e->hash, hash_str(e->name));
scripts/gendwarfksyms/types.c
203
hash_for_each_safe(type_map, e, tmp, hash)
scripts/gendwarfksyms/types.c
206
hash_for_each_safe(type_map, e, tmp, hash)
scripts/gendwarfksyms/types.c
226
hash_for_each_safe(type_map, e, tmp, hash) {
scripts/gendwarfksyms/types.c
73
struct hlist_node hash;
scripts/include/hash.h
11
hash = (hash ^ *s) * 0x01000193;
scripts/include/hash.h
12
return hash;
scripts/include/hash.h
8
unsigned int hash = 2166136261U;
scripts/kconfig/expr.c
36
int hash;
scripts/kconfig/expr.c
38
hash = hash_32((unsigned int)type ^ hash_ptr(l) ^ hash_ptr(r));
scripts/kconfig/expr.c
40
hash_for_each_possible(expr_hashtable, e, node, hash) {
scripts/kconfig/expr.c
52
hash_add(expr_hashtable, &e->node, hash);
scripts/kconfig/symbol.c
923
int hash;
scripts/kconfig/symbol.c
933
hash = hash_str(name);
scripts/kconfig/symbol.c
935
hash_for_each_possible(sym_hashtable, symbol, node, hash) {
scripts/kconfig/symbol.c
945
hash = 0;
scripts/kconfig/symbol.c
956
hash_add(sym_hashtable, &symbol->node, hash);
scripts/kconfig/symbol.c
964
int hash = 0;
scripts/kconfig/symbol.c
976
hash = hash_str(name);
scripts/kconfig/symbol.c
978
hash_for_each_possible(sym_hashtable, symbol, node, hash) {
scripts/kconfig/util.c
29
int hash = hash_str(name);
scripts/kconfig/util.c
31
hash_for_each_possible(file_hashtable, file, node, hash)
scripts/kconfig/util.c
41
hash_add(file_hashtable, &file->node, hash);
scripts/mod/sumversion.c
150
hash[0] += a;
scripts/mod/sumversion.c
151
hash[1] += b;
scripts/mod/sumversion.c
152
hash[2] += c;
scripts/mod/sumversion.c
153
hash[3] += d;
scripts/mod/sumversion.c
159
md4_transform(ctx->hash, ctx->block);
scripts/mod/sumversion.c
164
mctx->hash[0] = 0x67452301;
scripts/mod/sumversion.c
165
mctx->hash[1] = 0xefcdab89;
scripts/mod/sumversion.c
166
mctx->hash[2] = 0x98badcfe;
scripts/mod/sumversion.c
167
mctx->hash[3] = 0x10325476;
scripts/mod/sumversion.c
220
md4_transform(mctx->hash, mctx->block);
scripts/mod/sumversion.c
221
cpu_to_le32_array(mctx->hash, ARRAY_SIZE(mctx->hash));
scripts/mod/sumversion.c
224
mctx->hash[0], mctx->hash[1], mctx->hash[2], mctx->hash[3]);
scripts/mod/sumversion.c
43
uint32_t hash[MD4_HASH_WORDS];
scripts/mod/sumversion.c
90
static void md4_transform(uint32_t *hash, uint32_t const *in)
scripts/mod/sumversion.c
94
a = hash[0];
scripts/mod/sumversion.c
95
b = hash[1];
scripts/mod/sumversion.c
96
c = hash[2];
scripts/mod/sumversion.c
97
d = hash[3];
scripts/sign-file.c
45
uint8_t hash; /* Digest algorithm [0] */
security/apparmor/apparmorfs.c
1207
if (profile->hash) {
security/apparmor/apparmorfs.c
1209
seq_printf(seq, "%.2x", profile->hash[i]);
security/apparmor/apparmorfs.c
1220
SEQ_PROFILE_FOPS(hash);
security/apparmor/apparmorfs.c
1384
if (data->hash) {
security/apparmor/apparmorfs.c
1386
seq_printf(seq, "%.2x", data->hash[i]);
security/apparmor/apparmorfs.c
1404
SEQ_RAWDATA_FOPS(hash);
security/apparmor/apparmorfs.c
1837
if (profile->hash) {
security/apparmor/crypto.c
26
char *hash;
security/apparmor/crypto.c
28
hash = kzalloc(SHA256_DIGEST_SIZE, GFP_KERNEL);
security/apparmor/crypto.c
29
if (!hash)
security/apparmor/crypto.c
32
sha256(data, len, hash);
security/apparmor/crypto.c
33
return hash;
security/apparmor/crypto.c
45
profile->hash = kzalloc(SHA256_DIGEST_SIZE, GFP_KERNEL);
security/apparmor/crypto.c
46
if (!profile->hash)
security/apparmor/crypto.c
52
sha256_final(&sctx, profile->hash);
security/apparmor/include/policy.h
274
unsigned char *hash;
security/apparmor/include/policy_unpack.h
122
unsigned char *hash;
security/apparmor/policy.c
352
kfree_sensitive(profile->hash);
security/apparmor/policy_unpack.c
107
if (aa_g_hash_policy && memcmp(l->hash, r->hash, aa_hash_size()) != 0)
security/apparmor/policy_unpack.c
114
kfree_sensitive(d->hash);
security/apparmor/policy_unpack.c
1775
udata->hash = aa_calc_hash(udata->data, udata->size);
security/apparmor/policy_unpack.c
1776
if (IS_ERR(udata->hash)) {
security/apparmor/policy_unpack.c
1777
error = PTR_ERR(udata->hash);
security/apparmor/policy_unpack.c
1778
udata->hash = NULL;
security/integrity/ima/ima.h
265
int ima_calc_file_hash(struct file *file, struct ima_digest_data *hash);
security/integrity/ima/ima.h
267
struct ima_digest_data *hash);
security/integrity/ima/ima.h
270
int ima_calc_boot_aggregate(struct ima_digest_data *hash);
security/integrity/ima/ima_api.c
208
struct ima_max_digest_data *hash)
security/integrity/ima/ima_api.c
217
digest_len = fsverity_get_digest(inode, hash->digest, NULL, &alg);
security/integrity/ima/ima_api.c
228
hash->hdr.algo = alg;
security/integrity/ima/ima_api.c
229
hash->hdr.length = digest_len;
security/integrity/ima/ima_api.c
250
struct ima_max_digest_data hash;
security/integrity/ima/ima_api.c
251
struct ima_digest_data *hash_hdr = container_of(&hash.hdr,
security/integrity/ima/ima_api.c
281
hash.hdr.algo = algo;
security/integrity/ima/ima_api.c
282
hash.hdr.length = hash_digest_size[algo];
security/integrity/ima/ima_api.c
285
memset(&hash.digest, 0, sizeof(hash.digest));
security/integrity/ima/ima_api.c
288
if (!ima_get_verity_digest(iint, inode, &hash)) {
security/integrity/ima/ima_api.c
301
length = sizeof(hash.hdr) + hash.hdr.length;
security/integrity/ima/ima_api.c
309
memcpy(iint->ima_hash, &hash, length);
security/integrity/ima/ima_api.c
398
char *hash;
security/integrity/ima/ima_api.c
405
hash = kzalloc((iint->ima_hash->length * 2) + 1, GFP_KERNEL);
security/integrity/ima/ima_api.c
406
if (!hash)
security/integrity/ima/ima_api.c
410
hex_byte_pack(hash + (i * 2), iint->ima_hash->digest[i]);
security/integrity/ima/ima_api.c
411
hash[i * 2] = '\0';
security/integrity/ima/ima_api.c
420
audit_log_format(ab, " hash=\"%s:%s\"", algo_name, hash);
security/integrity/ima/ima_api.c
427
kfree(hash);
security/integrity/ima/ima_appraise.c
254
struct ima_digest_data *hash)
security/integrity/ima/ima_appraise.c
263
memcpy(file_id.hash, digest, hash_digest_size[algo]);
security/integrity/ima/ima_appraise.c
265
hash->algo = algo;
security/integrity/ima/ima_appraise.c
266
hash->length = hash_digest_size[algo];
security/integrity/ima/ima_appraise.c
268
return ima_calc_buffer_hash(&file_id, sizeof(file_id) - unused, hash);
security/integrity/ima/ima_appraise.c
282
struct ima_max_digest_data hash;
security/integrity/ima/ima_appraise.c
383
container_of(&hash.hdr,
security/integrity/ima/ima_appraise.c
393
xattr_len, hash.digest,
security/integrity/ima/ima_appraise.c
394
hash.hdr.length);
security/integrity/ima/ima_crypto.c
325
struct ima_digest_data *hash,
security/integrity/ima/ima_crypto.c
336
hash->length = crypto_ahash_digestsize(tfm);
security/integrity/ima/ima_crypto.c
427
ahash_request_set_crypt(req, NULL, hash->digest, 0);
security/integrity/ima/ima_crypto.c
435
static int ima_calc_file_ahash(struct file *file, struct ima_digest_data *hash)
security/integrity/ima/ima_crypto.c
440
tfm = ima_alloc_atfm(hash->algo);
security/integrity/ima/ima_crypto.c
444
rc = ima_calc_file_hash_atfm(file, hash, tfm);
security/integrity/ima/ima_crypto.c
452
struct ima_digest_data *hash,
security/integrity/ima/ima_crypto.c
462
hash->length = crypto_shash_digestsize(tfm);
security/integrity/ima/ima_crypto.c
498
rc = crypto_shash_final(shash, hash->digest);
security/integrity/ima/ima_crypto.c
502
static int ima_calc_file_shash(struct file *file, struct ima_digest_data *hash)
security/integrity/ima/ima_crypto.c
507
tfm = ima_alloc_tfm(hash->algo);
security/integrity/ima/ima_crypto.c
511
rc = ima_calc_file_hash_tfm(file, hash, tfm);
security/integrity/ima/ima_crypto.c
531
int ima_calc_file_hash(struct file *file, struct ima_digest_data *hash)
security/integrity/ima/ima_crypto.c
543
hash->length = hash_digest_size[ima_hash_algo];
security/integrity/ima/ima_crypto.c
544
hash->algo = ima_hash_algo;
security/integrity/ima/ima_crypto.c
563
rc = ima_calc_file_ahash(f, hash);
security/integrity/ima/ima_crypto.c
568
rc = ima_calc_file_shash(f, hash);
security/integrity/ima/ima_crypto.c
659
struct ima_digest_data *hash,
security/integrity/ima/ima_crypto.c
667
hash->length = crypto_ahash_digestsize(tfm);
security/integrity/ima/ima_crypto.c
690
ahash_request_set_crypt(req, NULL, hash->digest, 0);
security/integrity/ima/ima_crypto.c
699
struct ima_digest_data *hash)
security/integrity/ima/ima_crypto.c
704
tfm = ima_alloc_atfm(hash->algo);
security/integrity/ima/ima_crypto.c
708
rc = calc_buffer_ahash_atfm(buf, len, hash, tfm);
security/integrity/ima/ima_crypto.c
716
struct ima_digest_data *hash,
security/integrity/ima/ima_crypto.c
725
hash->length = crypto_shash_digestsize(tfm);
security/integrity/ima/ima_crypto.c
741
rc = crypto_shash_final(shash, hash->digest);
security/integrity/ima/ima_crypto.c
746
struct ima_digest_data *hash)
security/integrity/ima/ima_crypto.c
751
tfm = ima_alloc_tfm(hash->algo);
security/integrity/ima/ima_crypto.c
755
rc = calc_buffer_shash_tfm(buf, len, hash, tfm);
security/integrity/ima/ima_crypto.c
762
struct ima_digest_data *hash)
security/integrity/ima/ima_crypto.c
767
rc = calc_buffer_ahash(buf, len, hash);
security/integrity/ima/ima_crypto.c
772
return calc_buffer_shash(buf, len, hash);
security/integrity/ima/ima_crypto.c
839
int ima_calc_boot_aggregate(struct ima_digest_data *hash)
security/integrity/ima/ima_crypto.c
847
if (crypto_id == hash->algo) {
security/integrity/ima/ima_crypto.c
864
hash->algo = ima_tpm_chip->allocated_banks[bank_idx].crypto_id;
security/integrity/ima/ima_crypto.c
866
tfm = ima_alloc_tfm(hash->algo);
security/integrity/ima/ima_crypto.c
870
hash->length = crypto_shash_digestsize(tfm);
security/integrity/ima/ima_crypto.c
872
rc = ima_calc_boot_aggregate_tfm(hash->digest, alg_id, tfm);
security/integrity/ima/ima_init.c
50
struct ima_max_digest_data hash;
security/integrity/ima/ima_init.c
51
struct ima_digest_data *hash_hdr = container_of(&hash.hdr,
security/integrity/ima/ima_init.c
57
memset(&hash, 0, sizeof(hash));
security/integrity/ima/ima_main.c
1059
struct ima_max_digest_data hash;
security/integrity/ima/ima_main.c
1060
struct ima_digest_data *hash_hdr = container_of(&hash.hdr,
security/integrity/ima/ima_main.c
1195
bool hash, u8 *digest, size_t digest_len)
security/integrity/ima/ima_main.c
1202
event_label, hash, digest,
security/integrity/ima/ima_template_lib.c
347
struct ima_max_digest_data hash;
security/integrity/ima/ima_template_lib.c
348
struct ima_digest_data *hash_hdr = container_of(&hash.hdr,
security/integrity/ima/ima_template_lib.c
355
memset(&hash, 0, sizeof(hash));
security/integrity/ima/ima_template_lib.c
368
hash.hdr.algo = HASH_ALGO_SHA1;
security/integrity/ima/ima_template_lib.c
372
if (!result && hash.hdr.algo != HASH_ALGO_SHA1)
security/integrity/ima/ima_template_lib.c
376
memset(&hash, 0, sizeof(hash));
security/integrity/ima/ima_template_lib.c
388
hash.hdr.algo = ima_template_hash_algo_allowed(ima_hash_algo) ?
security/integrity/ima/ima_template_lib.c
398
cur_digestsize = hash.hdr.length;
security/integrity/integrity.h
112
__u8 hash[HASH_MAX_DIGESTSIZE];
security/keys/dh.c
101
size_t outbuf_len = roundup(buflen, crypto_shash_digestsize(hash));
security/keys/dh.c
109
ret = crypto_kdf108_ctr_generate(hash, &kbuf_iov, 1, outbuf, outbuf_len);
security/keys/dh.c
138
struct crypto_shash *hash = NULL;
security/keys/dh.c
171
ret = kdf_alloc(&hash, hashname);
security/keys/dh.c
277
ret = keyctl_dh_compute_kdf(hash, buffer, buflen, outbuf,
security/keys/dh.c
296
kdf_dealloc(hash);
security/keys/dh.c
67
static int kdf_alloc(struct crypto_shash **hash, char *hashname)
security/keys/dh.c
83
*hash = tfm;
security/keys/dh.c
88
static void kdf_dealloc(struct crypto_shash *hash)
security/keys/dh.c
90
if (hash)
security/keys/dh.c
91
crypto_free_shash(hash);
security/keys/dh.c
94
static int keyctl_dh_compute_kdf(struct crypto_shash *hash,
security/keys/keyring.c
168
unsigned long hash, type;
security/keys/keyring.c
195
hash = acc;
security/keys/keyring.c
197
hash ^= acc >> 32;
security/keys/keyring.c
203
if (index_key->type != &key_type_keyring && (hash & fan_mask) == 0)
security/keys/keyring.c
204
hash |= (hash >> (ASSOC_ARRAY_KEY_CHUNK_SIZE - level_shift)) | 1;
security/keys/keyring.c
205
else if (index_key->type == &key_type_keyring && (hash & fan_mask) != 0)
security/keys/keyring.c
206
hash = (hash + (hash << level_shift)) & ~fan_mask;
security/keys/keyring.c
207
index_key->hash = hash;
security/keys/keyring.c
278
return index_key->hash;
security/keys/keyring.c
334
seg_a = a->hash;
security/keys/keyring.c
335
seg_b = b->hash;
security/keys/trusted-keys/trusted_tpm1.c
714
opt->hash = tpm2 ? HASH_ALGO_SHA256 : HASH_ALGO_SHA1;
security/keys/trusted-keys/trusted_tpm1.c
796
opt->hash = i;
security/keys/trusted-keys/trusted_tpm1.c
808
digest_len = hash_digest_size[opt->hash];
security/keys/trusted-keys/trusted_tpm2.c
239
int hash;
security/keys/trusted-keys/trusted_tpm2.c
243
hash = tpm2_find_hash_alg(options->hash);
security/keys/trusted-keys/trusted_tpm2.c
244
if (hash < 0)
security/keys/trusted-keys/trusted_tpm2.c
245
return hash;
security/keys/trusted-keys/trusted_tpm2.c
291
tpm_buf_append_u16(&sized, hash);
security/selinux/include/hash.h
19
u32 hash = 0;
security/selinux/include/hash.h
27
hash ^= v; \
security/selinux/include/hash.h
28
hash = (hash << r2) | (hash >> (32 - r2)); \
security/selinux/include/hash.h
29
hash = hash * m + n; \
security/selinux/include/hash.h
38
hash ^= hash >> 16;
security/selinux/include/hash.h
39
hash *= 0x85ebca6b;
security/selinux/include/hash.h
40
hash ^= hash >> 13;
security/selinux/include/hash.h
41
hash *= 0xc2b2ae35;
security/selinux/include/hash.h
42
hash ^= hash >> 16;
security/selinux/include/hash.h
44
return hash & mask;
security/selinux/ss/context.c
16
u32 hash = 0;
security/selinux/ss/context.c
29
hash = jhash_3words(c->user, c->role, c->type, hash);
security/selinux/ss/context.c
30
hash = mls_range_hash(&c->range, hash);
security/selinux/ss/context.c
31
return hash;
security/selinux/ss/ebitmap.c
560
u32 ebitmap_hash(const struct ebitmap *e, u32 hash)
security/selinux/ss/ebitmap.c
565
hash = jhash_1word(e->highbit, hash);
security/selinux/ss/ebitmap.c
567
hash = jhash_1word(node->startbit, hash);
security/selinux/ss/ebitmap.c
568
hash = jhash(node->maps, sizeof(node->maps), hash);
security/selinux/ss/ebitmap.c
570
return hash;
security/selinux/ss/ebitmap.h
135
u32 ebitmap_hash(const struct ebitmap *e, u32 hash);
security/selinux/ss/hashtab.h
106
hvalue = key_params.hash(key) & (h->size - 1);
security/selinux/ss/hashtab.h
22
u32 (*hash)(const void *key); /* hash func */
security/selinux/ss/hashtab.h
73
hvalue = key_params.hash(key) & (h->size - 1);
security/selinux/ss/mls.h
88
static inline u32 mls_range_hash(const struct mls_range *r, u32 hash)
security/selinux/ss/mls.h
90
hash = jhash_2words(r->level[0].sens, r->level[1].sens, hash);
security/selinux/ss/mls.h
91
hash = ebitmap_hash(&r->level[0].cat, hash);
security/selinux/ss/mls.h
92
hash = ebitmap_hash(&r->level[1].cat, hash);
security/selinux/ss/mls.h
93
return hash;
security/selinux/ss/policydb.c
444
.hash = filenametr_hash,
security/selinux/ss/policydb.c
481
.hash = rangetr_hash,
security/selinux/ss/policydb.c
516
.hash = role_trans_hash,
security/selinux/ss/sidtab.c
106
if (!context_to_sid(s, context, hash)) {
security/selinux/ss/sidtab.c
108
isid->entry.hash = hash;
security/selinux/ss/sidtab.c
109
hash_add(s->context_to_sid, &isid->entry.list, hash);
security/selinux/ss/sidtab.c
271
u32 count, hash = context_compute_hash(context);
security/selinux/ss/sidtab.c
276
*sid = context_to_sid(s, context, hash);
security/selinux/ss/sidtab.c
284
*sid = context_to_sid(s, context, hash);
security/selinux/ss/sidtab.c
311
dst->hash = hash;
security/selinux/ss/sidtab.c
340
dst_convert->hash = context_compute_hash(&dst_convert->context);
security/selinux/ss/sidtab.c
344
dst_convert->hash);
security/selinux/ss/sidtab.c
355
hash_add_rcu(s->context_to_sid, &dst->list, dst->hash);
security/selinux/ss/sidtab.c
371
entry->hash = context_compute_hash(&entry->context);
security/selinux/ss/sidtab.c
373
hash_add_rcu(s->context_to_sid, &entry->list, entry->hash);
security/selinux/ss/sidtab.c
60
static u32 context_to_sid(struct sidtab *s, struct context *context, u32 hash)
security/selinux/ss/sidtab.c
66
hash_for_each_possible_rcu(s->context_to_sid, entry, list, hash) {
security/selinux/ss/sidtab.c
67
if (entry->hash != hash)
security/selinux/ss/sidtab.c
81
u32 hash;
security/selinux/ss/sidtab.c
98
hash = context_compute_hash(context);
security/selinux/ss/sidtab.h
23
u32 hash;
security/selinux/ss/symtab.c
19
unsigned int hash = 5381;
security/selinux/ss/symtab.c
23
hash = ((hash << 5) + hash) ^ c;
security/selinux/ss/symtab.c
25
return hash;
security/selinux/ss/symtab.c
38
.hash = symhash,
security/smack/smack_access.c
412
unsigned int hash;
security/smack/smack_access.c
415
hash = full_name_hash(NULL, skp->smk_known, strlen(skp->smk_known));
security/smack/smack_access.c
416
head = &smack_known_hash[hash & (SMACK_HASH_SLOTS - 1)];
security/smack/smack_access.c
431
unsigned int hash;
security/smack/smack_access.c
435
hash = full_name_hash(NULL, string, strlen(string));
security/smack/smack_access.c
436
head = &smack_known_hash[hash & (SMACK_HASH_SLOTS - 1)];
security/tomoyo/common.h
1170
return a->hash != b->hash || strcmp(a->name, b->name);
security/tomoyo/common.h
506
u32 hash; /* = full_name_hash(name, strlen(name)) */
security/tomoyo/memory.c
155
unsigned int hash;
security/tomoyo/memory.c
162
hash = full_name_hash(NULL, (const unsigned char *) name, len - 1);
security/tomoyo/memory.c
163
head = &tomoyo_name_list[hash_long(hash, TOMOYO_HASH_BITS)];
security/tomoyo/memory.c
167
if (hash != ptr->entry.hash || strcmp(name, ptr->entry.name) ||
security/tomoyo/util.c
685
ptr->hash = full_name_hash(NULL, name, len);
sound/soc/codecs/cros_ec_codec.c
701
memcpy(pp->hash, digest, SHA256_DIGEST_SIZE);
sound/soc/codecs/cros_ec_codec.c
725
memcpy(pp->hash, digest, SHA256_DIGEST_SIZE);
sound/soc/codecs/cros_ec_codec.c
774
if (memcmp(digest, r.hash, SHA256_DIGEST_SIZE) == 0) {
sound/soc/intel/avs/messages.h
580
u8 hash[32];
tools/bpf/bpftool/btf.c
661
static __u64 hasher(__u64 hash, __u64 val)
tools/bpf/bpftool/btf.c
663
return hash * 31 + val;
tools/bpf/bpftool/btf.c
666
static __u64 btf_name_hasher(__u64 hash, const struct btf *btf, __u32 name_off)
tools/bpf/bpftool/btf.c
669
return hash;
tools/bpf/bpftool/btf.c
671
return hasher(hash, str_hash(btf__name_by_offset(btf, name_off)));
tools/bpf/bpftool/btf.c
678
size_t hash = 0;
tools/bpf/bpftool/btf.c
680
hash = btf_name_hasher(hash, btf, t->name_off);
tools/bpf/bpftool/btf.c
690
hash = btf_name_hasher(hash, btf, name_off);
tools/bpf/bpftool/btf.c
700
hash = btf_name_hasher(hash, btf, m->name_off);
tools/bpf/bpftool/btf.c
702
hash = hasher(hash, btf_type_disambig_hash(btf, m->type, false));
tools/bpf/bpftool/btf.c
712
hash = hasher(hash, btf_type_disambig_hash(btf, t->type, include_members));
tools/bpf/bpftool/btf.c
717
hash = hasher(hash, arr->nelems);
tools/bpf/bpftool/btf.c
718
hash = hasher(hash, btf_type_disambig_hash(btf, arr->type, include_members));
tools/bpf/bpftool/btf.c
724
return hash;
tools/include/uapi/linux/bpf.h
6351
__u32 hash;
tools/include/uapi/linux/bpf.h
6640
__u32 hash; /* A hash of the packet 4 tuples */
tools/include/uapi/linux/bpf.h
6718
__aligned_u64 hash;
tools/lib/bpf/btf.c
3728
#define for_each_dedup_cand(d, node, hash) \
tools/lib/bpf/btf.c
3729
hashmap__for_each_key_entry(d->dedup_table, node, hash)
tools/lib/bpf/btf.c
3731
static int btf_dedup_table_add(struct btf_dedup *d, long hash, __u32 type_id)
tools/lib/bpf/btf.c
3733
return hashmap__append(d->dedup_table, hash, type_id);
tools/lib/bpf/hashmap.c
131
const long key, size_t hash,
tools/lib/bpf/hashmap.c
140
for (prev_ptr = &map->buckets[hash], cur = *prev_ptr;
tools/lib/bpf/libbpf.c
4558
prog->hash);
tools/lib/bpf/libbpf.c
513
__u8 hash[SHA256_DIGEST_LENGTH];
tools/lib/bpf/libbpf.c
5328
create_attr.excl_prog_hash = map->excl_prog->hash;
tools/lib/bpf/skel_internal.h
327
info.hash = (long) &sha;
tools/lib/bpf/strset.c
42
struct hashmap *hash;
tools/lib/bpf/strset.c
48
hash = hashmap__new(strset_hash_fn, strset_equal_fn, set);
tools/lib/bpf/strset.c
49
if (IS_ERR(hash))
tools/lib/bpf/strset.c
53
set->strs_hash = hash;
tools/lib/bpf/strset.c
70
err = hashmap__add(hash, off, off);
tools/lib/perf/evlist.c
338
int hash;
tools/lib/perf/evlist.c
343
hash = hash_64(sid->id, PERF_EVLIST__HLIST_BITS);
tools/lib/perf/evlist.c
344
hlist_add_head(&sid->node, &evlist->heads[hash]);
tools/lib/perf/evsel.c
571
int hash;
tools/lib/perf/evsel.c
576
hash = hash_32(tid, PERF_SAMPLE_ID__HLIST_BITS);
tools/lib/perf/evsel.c
577
head = &sid->periods[hash];
tools/lib/perf/evsel.c
594
hlist_add_head(&res->hnode, &sid->periods[hash]);
tools/objtool/check.c
351
return memcmp((void *)cfi1 + sizeof(cfi1->hash),
tools/objtool/check.c
352
(void *)cfi2 + sizeof(cfi2->hash),
tools/objtool/check.c
358
return jhash((void *)cfi + sizeof(cfi->hash),
tools/objtool/check.c
359
sizeof(*cfi) - sizeof(cfi->hash), 0);
tools/objtool/check.c
367
hlist_for_each_entry(obj, head, hash) {
tools/objtool/check.c
376
hlist_add_head(&obj->hash, head);
tools/objtool/check.c
385
hlist_add_head(&cfi->hash, head);
tools/objtool/check.c
46
hash_for_each_possible(file->insn_hash, insn, hash, sec_offset_hash(sec, offset)) {
tools/objtool/check.c
478
hash_add(file->insn_hash, &insn->hash, sec_offset_hash(sec, insn->offset));
tools/objtool/elf.c
1007
elf_hash_add(reloc, &reloc->hash, reloc_hash(reloc));
tools/objtool/elf.c
1104
elf_hash_add(reloc, &reloc->hash, reloc_hash(reloc));
tools/objtool/elf.c
1457
elf_hash_add(section, &sec->hash, sec->idx);
tools/objtool/elf.c
150
elf_hash_for_each_possible(section, sec, hash, idx) {
tools/objtool/elf.c
1562
elf_hash_del(reloc, &old->hash, key);
tools/objtool/elf.c
1563
elf_hash_add(reloc, &new->hash, key);
tools/objtool/elf.c
162
elf_hash_for_each_possible(symbol, sym, hash, idx) {
tools/objtool/elf.c
330
elf_hash_for_each_possible(reloc, reloc, hash,
tools/objtool/elf.c
422
elf_hash_add(section, &sec->hash, sec->idx);
tools/objtool/elf.c
507
elf_hash_add(symbol, &sym->hash, sym->idx);
tools/objtool/elf.c
922
elf_hash_del(symbol, &old->hash, old->idx);
tools/objtool/elf.c
923
elf_hash_add(symbol, &old->hash, new_idx);
tools/objtool/include/objtool/cfi.h
28
struct hlist_node hash; /* must be first, cficmp() */
tools/objtool/include/objtool/check.h
59
struct hlist_node hash;
tools/objtool/include/objtool/elf.h
100
struct elf_hash_node hash;
tools/objtool/include/objtool/elf.h
44
struct elf_hash_node hash;
tools/objtool/include/objtool/elf.h
64
struct elf_hash_node hash;
tools/objtool/klp-diff.c
159
hash_add(exports, &export->hash, str_hash(sym));
tools/objtool/klp-diff.c
32
struct hlist_node hash;
tools/objtool/klp-diff.c
764
hash_for_each_possible(exports, export, hash, str_hash(sym->name)) {
tools/perf/arch/arm/util/cs-etm.c
272
u32 hash;
tools/perf/arch/arm/util/cs-etm.c
282
ret = perf_pmu__scan_file(pmu, path, "%x", &hash);
tools/perf/arch/arm/util/cs-etm.c
296
evsel__set_config_if_unset(evsel, "sinkid", hash);
tools/perf/builtin-inject.c
1137
int hash;
tools/perf/builtin-inject.c
1144
hash = hash_32(guest_tid->tid, PERF_EVLIST__HLIST_BITS);
tools/perf/builtin-inject.c
1145
hlist_add_head(&guest_tid->node, &gs->tids[hash]);
tools/perf/builtin-inject.c
1213
int hash;
tools/perf/builtin-inject.c
1221
hash = hash_64(guest_id->id, PERF_EVLIST__HLIST_BITS);
tools/perf/builtin-inject.c
1222
hlist_add_head(&guest_id->node, &gs->heads[hash]);
tools/perf/builtin-inject.c
1278
int hash;
tools/perf/builtin-inject.c
1280
hash = hash_64(id, PERF_EVLIST__HLIST_BITS);
tools/perf/builtin-inject.c
1281
head = &gs->heads[hash];
tools/perf/builtin-inject.c
1385
int hash;
tools/perf/builtin-inject.c
1387
hash = hash_32(tid, PERF_EVLIST__HLIST_BITS);
tools/perf/builtin-inject.c
1388
head = &gs->tids[hash];
tools/perf/builtin-lock.c
904
u64 hash = 0;
tools/perf/builtin-lock.c
937
hash ^= hash_long((unsigned long)node->ip, 64);
tools/perf/builtin-lock.c
942
return hash;
tools/perf/tests/util.c
34
u8 hash[HASH_LEN];
tools/perf/tests/util.c
58
blake2s_init(&main_ctx, sizeof(hash));
tools/perf/tests/util.c
65
blake2s_final(&ctx, hash);
tools/perf/tests/util.c
73
memcmp(hash, hash2, HASH_LEN) == 0);
tools/perf/tests/util.c
79
blake2s_update(&main_ctx, hash, HASH_LEN);
tools/perf/tests/util.c
83
blake2s_final(&main_ctx, hash);
tools/perf/tests/util.c
85
memcmp(hash, expected_hash_of_hashes, HASH_LEN) == 0);
tools/perf/util/auxtrace.c
2101
hlist_for_each_entry_safe(entry, tmp, &c->hashtable[i], hash) {
tools/perf/util/auxtrace.c
2102
hlist_del(&entry->hash);
tools/perf/util/auxtrace.c
2138
hlist_add_head(&entry->hash, &c->hashtable[hash_32(key, c->bits)]);
tools/perf/util/auxtrace.c
2154
hlist_for_each_entry_safe(entry, n, hlist, hash) {
tools/perf/util/auxtrace.c
2156
hlist_del(&entry->hash);
tools/perf/util/auxtrace.c
2180
hlist_for_each_entry(entry, hlist, hash) {
tools/perf/util/auxtrace.h
554
struct hlist_node hash;
tools/perf/util/bpf-trace-summary.c
172
static int update_thread_stats(struct hashmap *hash, struct syscall_key *map_key,
tools/perf/util/bpf-trace-summary.c
178
if (!hashmap__find(hash, map_key->cpu_or_tid, &data)) {
tools/perf/util/bpf-trace-summary.c
184
if (hashmap__add(hash, data->key, data) < 0) {
tools/perf/util/bpf-trace-summary.c
236
static int update_total_stats(struct hashmap *hash, struct syscall_key *map_key,
tools/perf/util/bpf-trace-summary.c
242
if (!hashmap__find(hash, map_key->nr, &data)) {
tools/perf/util/bpf-trace-summary.c
257
if (hashmap__add(hash, data->key, data) < 0) {
tools/perf/util/bpf-trace-summary.c
308
static int update_cgroup_stats(struct hashmap *hash, struct syscall_key *map_key,
tools/perf/util/bpf-trace-summary.c
314
if (!hashmap__find(hash, map_key->cgroup, &data)) {
tools/perf/util/bpf-trace-summary.c
320
if (hashmap__add(hash, data->key, data) < 0) {
tools/perf/util/evlist.c
662
int hash;
tools/perf/util/evlist.c
664
hash = hash_64(id, PERF_EVLIST__HLIST_BITS);
tools/perf/util/evlist.c
665
head = &evlist->core.heads[hash];
tools/perf/util/evlist.c
730
int hash;
tools/perf/util/evlist.c
747
hash = hash_64(id, PERF_EVLIST__HLIST_BITS);
tools/perf/util/evlist.c
748
head = &evlist->core.heads[hash];
tools/perf/util/expr.c
51
size_t hash = 0;
tools/perf/util/expr.c
54
hash *= 31;
tools/perf/util/expr.c
55
hash += *str;
tools/perf/util/expr.c
58
return hash;
tools/perf/util/expr.c
68
struct hashmap *hash;
tools/perf/util/expr.c
70
hash = hashmap__new(key_hash, key_equal, NULL);
tools/perf/util/expr.c
71
if (IS_ERR(hash))
tools/perf/util/expr.c
73
return hash;
tools/perf/util/hashmap.c
131
const long key, size_t hash,
tools/perf/util/hashmap.c
140
for (prev_ptr = &map->buckets[hash], cur = *prev_ptr;
tools/perf/util/sharded_mutex.h
24
static inline struct mutex *sharded_mutex__get_mutex(struct sharded_mutex *sm, size_t hash)
tools/perf/util/sharded_mutex.h
26
return &sm->mutexes[hash_bits(hash, sm->cap_bits)];
tools/testing/cxl/test/mem.c
1829
u8 hash[SHA256_DIGEST_SIZE];
tools/testing/cxl/test/mem.c
1831
sha256(mdata->fw, mdata->fw_size, hash);
tools/testing/cxl/test/mem.c
1833
return sysfs_emit(buf, "%*phN\n", SHA256_DIGEST_SIZE, hash);
tools/testing/selftests/bpf/bpf_arena_htab.h
19
static inline htab_bucket_t *__select_bucket(htab_t *htab, __u32 hash)
tools/testing/selftests/bpf/bpf_arena_htab.h
24
return &b[hash & (htab->n_buckets - 1)];
tools/testing/selftests/bpf/bpf_arena_htab.h
27
static inline arena_list_head_t *select_bucket(htab_t *htab, __u32 hash)
tools/testing/selftests/bpf/bpf_arena_htab.h
29
return &__select_bucket(htab, hash)->head;
tools/testing/selftests/bpf/bpf_arena_htab.h
33
int hash;
tools/testing/selftests/bpf/bpf_arena_htab.h
40
static hashtab_elem_t *lookup_elem_raw(arena_list_head_t *head, __u32 hash, int key)
tools/testing/selftests/bpf/bpf_arena_htab.h
45
if (l->hash == hash && l->key == key)
tools/testing/selftests/bpf/bpf_arena_htab.h
82
l_new->hash = htab_hash(key);
tools/testing/selftests/bpf/prog_tests/select_reuseport.c
291
RET_IF(!result.hash, "data_check result.hash empty",
tools/testing/selftests/bpf/prog_tests/select_reuseport.c
292
"result.hash:%u", result.hash);
tools/testing/selftests/bpf/prog_tests/skb_ctx.c
67
skb.hash = 1;
tools/testing/selftests/bpf/prog_tests/skb_ctx.c
70
skb.hash = 0;
tools/testing/selftests/bpf/prog_tests/unpriv_bpf_disabled.c
249
map_fds[2] = bpf_map__fd(skel->maps.hash);
tools/testing/selftests/bpf/prog_tests/verify_pkcs7_sig.c
51
__u8 hash; /* Digest algorithm [0] */
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
237
static int fq_new_flow(void *flow_map, struct fq_stashed_flow **sflow, u64 hash)
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
252
ret = bpf_map_update_elem(flow_map, &hash, &tmp, 0);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
255
bpf_map_update_elem(&fq_nonprio_flows, &hash, &tmp, 0);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
258
*sflow = bpf_map_lookup_elem(flow_map, &hash);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
273
u64 hash = 0;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
276
*sflow = bpf_map_lookup_elem(&fq_prio_flows, &hash);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
280
hash = bpf_skb_get_hash(skb) & q.orphan_mask;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
285
hash |= (1ULL << 32);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
287
hash = bpf_skb_get_hash(skb) & q.orphan_mask;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
288
hash |= (1ULL << 32);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
290
hash = sk->__sk_common.skc_hash;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
292
*sflow = bpf_map_lookup_elem(&fq_nonprio_flows, &hash);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
296
ret = fq_new_flow(&fq_nonprio_flows, sflow, hash) < 0 ?
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
554
u64 hash = 0;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
556
sflow = bpf_map_lookup_elem(&fq_prio_flows, &hash);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
661
fq_remove_flows(struct bpf_map *flow_map, u64 *hash,
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
666
bpf_map_delete_elem(flow_map, hash);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
696
u64 hash = 0;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
705
fq_new_flow(&fq_prio_flows, &sflow, hash);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
721
u64 hash = 0;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
723
if (fq_new_flow(&fq_prio_flows, &sflow, hash) < 0)
tools/testing/selftests/bpf/progs/lsm.c
129
value = bpf_map_lookup_elem(&hash, &key);
tools/testing/selftests/bpf/progs/lsm.c
26
} hash SEC(".maps");
tools/testing/selftests/bpf/progs/map_ptr_kern.c
110
struct bpf_htab *hash = (struct bpf_htab *)&m_hash;
tools/testing/selftests/bpf/progs/map_ptr_kern.c
114
VERIFY(check_default_noinline(&hash->map, map));
tools/testing/selftests/bpf/progs/map_ptr_kern.c
116
VERIFY(hash->n_buckets == MAX_ENTRIES);
tools/testing/selftests/bpf/progs/map_ptr_kern.c
117
VERIFY(hash->elem_size == 64);
tools/testing/selftests/bpf/progs/map_ptr_kern.c
119
VERIFY(hash->count.counter == 0);
tools/testing/selftests/bpf/progs/map_ptr_kern.c
126
if (bpf_map_update_elem(hash, &key, &val, 0))
tools/testing/selftests/bpf/progs/map_ptr_kern.c
129
VERIFY(hash->count.counter == HALF_ENTRIES);
tools/testing/selftests/bpf/progs/sock_iter_batch.c
35
unsigned int hash;
tools/testing/selftests/bpf/progs/sock_iter_batch.c
63
hash = jhash2(sk->sk_v6_rcv_saddr.s6_addr32, 4, net->hash_mix);
tools/testing/selftests/bpf/progs/sock_iter_batch.c
64
hash ^= sk->sk_num;
tools/testing/selftests/bpf/progs/sock_iter_batch.c
66
bucket[idx] = hash & hinfo->lhash2_mask;
tools/testing/selftests/bpf/progs/test_l4lb.c
220
__u32 hash = get_packet_hash(pckt, is_ipv6) % RING_SIZE;
tools/testing/selftests/bpf/progs/test_l4lb.c
221
__u32 key = RING_SIZE * vip_info->vip_num + hash;
tools/testing/selftests/bpf/progs/test_l4lb_noinline.c
215
__u32 hash = get_packet_hash(pckt, is_ipv6);
tools/testing/selftests/bpf/progs/test_l4lb_noinline.c
216
__u32 key = RING_SIZE * vip_info->vip_num + hash % RING_SIZE;
tools/testing/selftests/bpf/progs/test_l4lb_noinline.c
219
if (hash != 0x358459b7 /* jhash of ipv4 packet */ &&
tools/testing/selftests/bpf/progs/test_l4lb_noinline.c
220
hash != 0x2f4bc6bb /* jhash of ipv6 packet */)
tools/testing/selftests/bpf/progs/test_l4lb_noinline_dynptr.c
217
__u32 hash = get_packet_hash(pckt, is_ipv6);
tools/testing/selftests/bpf/progs/test_l4lb_noinline_dynptr.c
218
__u32 key = RING_SIZE * vip_info->vip_num + hash % RING_SIZE;
tools/testing/selftests/bpf/progs/test_l4lb_noinline_dynptr.c
221
if (hash != 0x358459b7 /* jhash of ipv4 packet */ &&
tools/testing/selftests/bpf/progs/test_l4lb_noinline_dynptr.c
222
hash != 0x2f4bc6bb /* jhash of ipv6 packet */)
tools/testing/selftests/bpf/progs/test_pkt_md_access.c
36
TEST_FIELD(__u8, hash, 0xFF);
tools/testing/selftests/bpf/progs/test_pkt_md_access.c
37
TEST_FIELD(__u16, hash, 0xFFFF);
tools/testing/selftests/bpf/progs/test_pkt_md_access.c
38
TEST_FIELD(__u32, hash, 0xFFFFFFFF);
tools/testing/selftests/bpf/progs/test_select_reuseport_kern.c
80
data_check.hash = reuse_md->hash;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
321
u32 hash;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
343
hash = siphash_2u64(first, second, &test_key_siphash);
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
348
ctx->attrs.rcv_tsecr |= hash & COOKIE_MASK;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
351
hash &= ~COOKIE_MASK;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
352
hash |= mssind << 6;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
355
hash |= ctx->attrs.snd_wscale & BPF_SYNCOOKIE_WSCALE_MASK;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
358
hash |= BPF_SYNCOOKIE_SACK;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
361
hash |= BPF_SYNCOOKIE_ECN;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
363
ctx->cookie = hash;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
470
u32 hash;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
479
hash = siphash_2u64(first, second, &test_key_siphash);
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
482
hash -= ctx->attrs.rcv_tsecr & COOKIE_MASK;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
484
hash &= ~COOKIE_MASK;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
486
hash -= cookie & ~COOKIE_MASK;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
487
if (hash)
tools/testing/selftests/bpf/progs/test_tcpnotify_kern.c
67
.hash = 0xef,
tools/testing/selftests/bpf/progs/test_unpriv_bpf_disabled.c
34
} hash SEC(".maps");
tools/testing/selftests/bpf/progs/test_xdp_noinline.c
528
__u32 *real_pos, hash, key;
tools/testing/selftests/bpf/progs/test_xdp_noinline.c
537
hash = get_packet_hash(pckt, hash_16bytes);
tools/testing/selftests/bpf/progs/test_xdp_noinline.c
538
if (hash != 0x358459b7 /* jhash of ipv4 packet */ &&
tools/testing/selftests/bpf/progs/test_xdp_noinline.c
539
hash != 0x2f4bc6bb /* jhash of ipv6 packet */)
tools/testing/selftests/bpf/progs/test_xdp_noinline.c
541
key = 2 * vip_info->vip_num + hash % 2;
tools/testing/selftests/bpf/progs/verifier_cgroup_skb.c
59
: __imm_const(__sk_buff_hash, offsetof(struct __sk_buff, hash)),
tools/testing/selftests/bpf/progs/verifier_ctx.c
293
padding_access("sk_reuseport", sk_reuseport_md, hash, 4);
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
21
extern int bpf_xdp_metadata_rx_hash(const struct xdp_md *ctx, __u32 *hash,
tools/testing/selftests/bpf/progs/xdp_metadata.c
31
extern int bpf_xdp_metadata_rx_hash(const struct xdp_md *ctx, __u32 *hash,
tools/testing/selftests/bpf/progs/xdp_metadata2.c
17
u32 hash = 0;
tools/testing/selftests/bpf/progs/xdp_metadata2.c
19
bpf_xdp_metadata_rx_hash(ctx, &hash, &type);
tools/testing/selftests/bpf/progs/xdp_metadata2.c
8
extern int bpf_xdp_metadata_rx_hash(const struct xdp_md *ctx, __u32 *hash,
tools/testing/selftests/bpf/test_select_reuseport_common.h
33
__u32 hash;
tools/testing/selftests/bpf/test_tcpnotify.h
15
__u8 hash;
tools/testing/selftests/bpf/test_tcpnotify_user.c
37
t->source != 0xbe || t->hash != 0xef)
tools/testing/selftests/bpf/verifier/ctx_skb.c
484
offsetof(struct __sk_buff, hash)),
tools/testing/selftests/bpf/verifier/ctx_skb.c
507
offsetof(struct __sk_buff, hash)),
tools/testing/selftests/bpf/verifier/ctx_skb.c
510
offsetof(struct __sk_buff, hash) + 3),
tools/testing/selftests/bpf/verifier/ctx_skb.c
521
offsetof(struct __sk_buff, hash) + 1),
tools/testing/selftests/bpf/verifier/ctx_skb.c
531
offsetof(struct __sk_buff, hash) + 2),
tools/testing/selftests/bpf/verifier/ctx_skb.c
542
offsetof(struct __sk_buff, hash) + 3),
tools/testing/selftests/bpf/verifier/ctx_skb.c
545
offsetof(struct __sk_buff, hash)),
tools/testing/selftests/bpf/verifier/ctx_skb.c
628
offsetof(struct __sk_buff, hash)),
tools/testing/selftests/bpf/verifier/ctx_skb.c
651
offsetof(struct __sk_buff, hash)),
tools/testing/selftests/bpf/verifier/ctx_skb.c
654
offsetof(struct __sk_buff, hash) + 2),
tools/testing/selftests/bpf/verifier/ctx_skb.c
666
offsetof(struct __sk_buff, hash) + 2),
tools/testing/selftests/bpf/verifier/ctx_skb.c
669
offsetof(struct __sk_buff, hash)),
tools/testing/selftests/bpf/verifier/ctx_skb.c
681
offsetof(struct __sk_buff, hash) + 1),
tools/testing/selftests/bpf/verifier/ctx_skb.c
684
offsetof(struct __sk_buff, hash) + 3),
tools/testing/selftests/bpf/verifier/ctx_skb.c
698
offsetof(struct __sk_buff, hash) + 3),
tools/testing/selftests/bpf/verifier/ctx_skb.c
701
offsetof(struct __sk_buff, hash) + 1),
tools/testing/selftests/powerpc/dexcr/dexcr.c
167
unsigned long hash = 0;
tools/testing/selftests/powerpc/dexcr/dexcr.c
169
hashst(0, &hash);
tools/testing/selftests/powerpc/dexcr/dexcr.c
170
hash += 1;
tools/testing/selftests/powerpc/dexcr/dexcr.c
171
hashchk(0, &hash);
tools/testing/selftests/vDSO/parse_vdso.c
136
ELF_HASH_ENTRY *hash = 0;
tools/testing/selftests/vDSO/parse_vdso.c
155
hash = (ELF_HASH_ENTRY *)
tools/testing/selftests/vDSO/parse_vdso.c
177
(!hash && !vdso_info.gnu_hash))
tools/testing/selftests/vDSO/parse_vdso.c
192
vdso_info.nbucket = hash[0];
tools/testing/selftests/vDSO/parse_vdso.c
193
vdso_info.nchain = hash[1];
tools/testing/selftests/vDSO/parse_vdso.c
194
vdso_info.bucket = &hash[2];
tools/testing/selftests/vDSO/parse_vdso.c
195
vdso_info.chain = &hash[vdso_info.nbucket + 2];
tools/testing/selftests/vDSO/parse_vdso.c
203
const char *name, ELF(Word) hash)
tools/testing/selftests/vDSO/parse_vdso.c
236
return def->vd_hash == hash
tools/testing/vsock/util.c
589
unsigned long hash = 5381;
tools/testing/vsock/util.c
593
hash = ((hash << 5) + hash) + ((unsigned char *)data)[i];
tools/testing/vsock/util.c
597
return hash;
tools/testing/vsock/util.c
613
unsigned long hash;
tools/testing/vsock/util.c
632
hash = hash_djb2(tmp, iov_bytes);
tools/testing/vsock/util.c
635
return hash;
tools/testing/vsock/vsock_test_zerocopy.c
369
unsigned long hash;
tools/testing/vsock/vsock_test_zerocopy.c
389
hash = hash_djb2(sbuf2, sizeof(sbuf2));
tools/testing/vsock/vsock_test_zerocopy.c
390
control_writeulong(hash);