Symbol: head
arch/arm/crypto/ghash-ce-glue.c
191
struct gcm_key const *k, char *head,
arch/arm/crypto/ghash-ce-glue.c
199
struct gcm_key const *k, char *head,
arch/arm/crypto/ghash-ce-glue.c
54
u64 const h[][2], const char *head);
arch/arm/crypto/ghash-ce-glue.c
57
u64 const h[][2], const char *head);
arch/arm/crypto/ghash-ce-glue.c
70
struct ghash_key *key, const char *head)
arch/arm/crypto/ghash-ce-glue.c
74
pmull_ghash_update_p64(blocks, dg, src, key->h, head);
arch/arm/crypto/ghash-ce-glue.c
76
pmull_ghash_update_p8(blocks, dg, src, key->h, head);
arch/arm/kernel/bios32.c
428
struct list_head *head)
arch/arm/kernel/bios32.c
486
list_add(&sys->node, head);
arch/arm/kernel/bios32.c
498
LIST_HEAD(head);
arch/arm/kernel/bios32.c
503
pcibios_init_hw(parent, hw, &head);
arch/arm/kernel/bios32.c
507
list_for_each_entry(sys, &head, node) {
arch/arm/kernel/machine_kexec.c
156
page_list = image->head & PAGE_MASK;
arch/arm64/crypto/ghash-ce-glue.c
49
u64 const h[][2], const char *head);
arch/arm64/crypto/ghash-ce-glue.c
52
u64 const h[][2], const char *head);
arch/arm64/crypto/ghash-ce-glue.c
72
struct ghash_key *key, const char *head,
arch/arm64/crypto/ghash-ce-glue.c
76
const char *head))
arch/arm64/crypto/ghash-ce-glue.c
79
simd_update(blocks, dg, src, key->h, head);
arch/arm64/include/asm/kvm_host.h
101
*p = mc->head;
arch/arm64/include/asm/kvm_host.h
102
mc->head = to_pa(p);
arch/arm64/include/asm/kvm_host.h
109
phys_addr_t *p = to_va(mc->head & PAGE_MASK);
arch/arm64/include/asm/kvm_host.h
114
mc->head = *p;
arch/arm64/include/asm/kvm_host.h
89
phys_addr_t head;
arch/arm64/include/uapi/asm/sigcontext.h
104
struct _aarch64_ctx head;
arch/arm64/include/uapi/asm/sigcontext.h
137
struct _aarch64_ctx head;
arch/arm64/include/uapi/asm/sigcontext.h
146
struct _aarch64_ctx head;
arch/arm64/include/uapi/asm/sigcontext.h
158
struct _aarch64_ctx head;
arch/arm64/include/uapi/asm/sigcontext.h
166
struct _aarch64_ctx head;
arch/arm64/include/uapi/asm/sigcontext.h
173
struct _aarch64_ctx head;
arch/arm64/include/uapi/asm/sigcontext.h
181
struct _aarch64_ctx head;
arch/arm64/include/uapi/asm/sigcontext.h
189
struct _aarch64_ctx head;
arch/arm64/include/uapi/asm/sigcontext.h
78
struct _aarch64_ctx head;
arch/arm64/include/uapi/asm/sigcontext.h
97
struct _aarch64_ctx head;
arch/arm64/kernel/asm-offsets.c
175
DEFINE(KIMAGE_HEAD, offsetof(struct kimage, head));
arch/arm64/kernel/machine_kexec.c
117
if (kimage->head & IND_DONE) {
arch/arm64/kernel/machine_kexec.c
190
if (kimage->head & IND_DONE) {
arch/arm64/kernel/machine_kexec.c
38
kexec_dprintk(" head: %lx\n", kimage->head);
arch/arm64/kernel/signal.c
1272
__put_user_error(ESR_MAGIC, &esr_ctx->head.magic, err);
arch/arm64/kernel/signal.c
1273
__put_user_error(sizeof(*esr_ctx), &esr_ctx->head.size, err);
arch/arm64/kernel/signal.c
1350
__put_user_error(EXTRA_MAGIC, &extra->head.magic, err);
arch/arm64/kernel/signal.c
1351
__put_user_error(EXTRA_CONTEXT_SIZE, &extra->head.size, err);
arch/arm64/kernel/signal.c
265
__put_user_error(FPSIMD_MAGIC, &ctx->head.magic, err);
arch/arm64/kernel/signal.c
266
__put_user_error(sizeof(struct fpsimd_context), &ctx->head.size, err);
arch/arm64/kernel/signal.c
311
__put_user_error(FPMR_MAGIC, &ctx->head.magic, err);
arch/arm64/kernel/signal.c
312
__put_user_error(sizeof(*ctx), &ctx->head.size, err);
arch/arm64/kernel/signal.c
338
__put_user_error(POE_MAGIC, &ctx->head.magic, err);
arch/arm64/kernel/signal.c
339
__put_user_error(sizeof(*ctx), &ctx->head.size, err);
arch/arm64/kernel/signal.c
381
__put_user_error(SVE_MAGIC, &ctx->head.magic, err);
arch/arm64/kernel/signal.c
383
&ctx->head.size, err);
arch/arm64/kernel/signal.c
511
__put_user_error(TPIDR2_MAGIC, &ctx->head.magic, err);
arch/arm64/kernel/signal.c
512
__put_user_error(sizeof(*ctx), &ctx->head.size, err);
arch/arm64/kernel/signal.c
547
__put_user_error(ZA_MAGIC, &ctx->head.magic, err);
arch/arm64/kernel/signal.c
549
&ctx->head.size, err);
arch/arm64/kernel/signal.c
623
__put_user_error(ZT_MAGIC, &ctx->head.magic, err);
arch/arm64/kernel/signal.c
625
&ctx->head.size, err);
arch/arm64/kernel/signal.c
694
__put_user_error(GCS_MAGIC, &ctx->head.magic, err);
arch/arm64/kernel/signal.c
695
__put_user_error(sizeof(*ctx), &ctx->head.size, err);
arch/arm64/kernel/signal.c
757
struct _aarch64_ctx __user *head;
arch/arm64/kernel/signal.c
786
if (limit - offset < sizeof(*head))
arch/arm64/kernel/signal.c
792
head = (struct _aarch64_ctx __user *)(base + offset);
arch/arm64/kernel/signal.c
793
__get_user_error(magic, &head->magic, err);
arch/arm64/kernel/signal.c
794
__get_user_error(size, &head->size, err);
arch/arm64/kernel/signal.c
814
user->fpsimd = (struct fpsimd_context __user *)head;
arch/arm64/kernel/signal.c
829
user->poe = (struct poe_context __user *)head;
arch/arm64/kernel/signal.c
840
user->sve = (struct sve_context __user *)head;
arch/arm64/kernel/signal.c
851
user->tpidr2 = (struct tpidr2_context __user *)head;
arch/arm64/kernel/signal.c
862
user->za = (struct za_context __user *)head;
arch/arm64/kernel/signal.c
873
user->zt = (struct zt_context __user *)head;
arch/arm64/kernel/signal.c
884
user->fpmr = (struct fpmr_context __user *)head;
arch/arm64/kernel/signal.c
895
user->gcs = (struct gcs_context __user *)head;
arch/arm64/kernel/signal.c
906
userp = (char const __user *)head;
arch/arm64/kernel/signal.c
965
if (size < sizeof(*head))
arch/arm64/kvm/hyp/nvhe/mm.c
486
if (__pkvm_host_donate_hyp(hyp_phys_to_pfn(host_mc->head), 1))
arch/arm64/kvm/hyp/nvhe/page_alloc.c
80
static inline void page_add_to_list(struct hyp_page *p, struct list_head *head)
arch/arm64/kvm/hyp/nvhe/page_alloc.c
85
list_add_tail(node, head);
arch/arm64/kvm/mmu.c
231
static void stage2_free_unlinked_table_rcu_cb(struct rcu_head *head)
arch/arm64/kvm/mmu.c
233
struct page *page = container_of(head, struct page, rcu_head);
arch/loongarch/kernel/machine_kexec.c
258
(unsigned long)phys_to_virt(image->head & PAGE_MASK) : 0;
arch/loongarch/kernel/machine_kexec.c
265
for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE);
arch/m68k/kernel/machine_kexec.c
57
((relocate_kernel_t) reboot_code_buffer)(image->head & PAGE_MASK,
arch/m68k/mac/via.c
550
void via1_set_head(int head)
arch/m68k/mac/via.c
552
if (head == 0)
arch/mips/include/asm/octeon/cvmx-pow.h
669
uint64_t head:1;
arch/mips/include/asm/octeon/cvmx-pow.h
692
uint64_t head:1;
arch/mips/include/asm/octeon/cvmx-pow.h
724
uint64_t head:1;
arch/mips/include/asm/octeon/cvmx-pow.h
747
uint64_t head:1;
arch/mips/kernel/machine_kexec.c
223
(unsigned long) phys_to_virt(image->head & PAGE_MASK);
arch/mips/kernel/machine_kexec.c
225
kexec_indirection_page = (unsigned long)&image->head;
arch/mips/kernel/machine_kexec.c
237
for (ptr = &image->head; (entry = *ptr) && !(entry &IND_DONE);
arch/mips/kernel/machine_kexec.c
41
pr_debug(" head: %lx\n", kimage->head);
arch/parisc/kernel/kexec.c
107
reloc(image->head & PAGE_MASK, image->start, phys);
arch/parisc/kernel/kexec.c
12
extern void relocate_new_kernel(unsigned long head,
arch/parisc/kernel/kexec.c
40
pr_debug(" head: %lx\n", kimage->head);
arch/parisc/kernel/kexec.c
76
void (*reloc)(unsigned long head,
arch/powerpc/include/asm/guest-state-buffer.h
556
#define kvmppc_gse_for_each_elem(i, max, pos, head, len, rem) \
arch/powerpc/include/asm/guest-state-buffer.h
557
for (i = 0, pos = head, rem = len; kvmppc_gse_ok(pos, rem) && i < max; \
arch/powerpc/include/asm/hvcserver.h
38
extern int hvcs_free_partner_info(struct list_head *head);
arch/powerpc/include/asm/hvcserver.h
40
struct list_head *head, unsigned long *pi_buff);
arch/powerpc/include/asm/ps3.h
95
struct list_head head;
arch/powerpc/include/asm/ps3gpu.h
34
static inline int lv1_gpu_display_sync(u64 context_handle, u64 head,
arch/powerpc/include/asm/ps3gpu.h
39
head, ddr_offset, 0, 0);
arch/powerpc/include/asm/ps3gpu.h
42
static inline int lv1_gpu_display_flip(u64 context_handle, u64 head,
arch/powerpc/include/asm/ps3gpu.h
47
head, ddr_offset, 0, 0);
arch/powerpc/kexec/core_32.c
44
page_list = image->head;
arch/powerpc/kexec/core_64.c
128
copy_segments(image->head);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1008
unsigned long head, i, j;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1021
i = head = *rmapp & KVMPPC_RMAP_INDEX;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1080
} while ((i = j) != head);
arch/powerpc/kvm/book3s_64_mmu_hv.c
886
unsigned long head, i, j;
arch/powerpc/kvm/book3s_64_mmu_hv.c
903
i = head = *rmapp & KVMPPC_RMAP_INDEX;
arch/powerpc/kvm/book3s_64_mmu_hv.c
931
} while ((i = j) != head);
arch/powerpc/kvm/book3s_64_mmu_hv.c
957
unsigned long head, i, j;
arch/powerpc/kvm/book3s_64_mmu_hv.c
971
i = head = *rmapp & KVMPPC_RMAP_INDEX;
arch/powerpc/kvm/book3s_64_mmu_hv.c
977
} while ((i = j) != head);
arch/powerpc/kvm/book3s_64_vio.c
195
static void release_spapr_tce_table(struct rcu_head *head)
arch/powerpc/kvm/book3s_64_vio.c
197
struct kvmppc_spapr_tce_table *stt = container_of(head,
arch/powerpc/kvm/book3s_64_vio.c
61
static void kvm_spapr_tce_iommu_table_free(struct rcu_head *head)
arch/powerpc/kvm/book3s_64_vio.c
63
struct kvmppc_spapr_tce_iommu_table *stit = container_of(head,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
152
unsigned long ptel, head;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
165
head = *rmap & KVMPPC_RMAP_INDEX;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
170
if (head == pte_index) {
arch/powerpc/kvm/book3s_hv_rm_mmu.c
171
head = rev->forw;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
172
if (head == pte_index)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
175
*rmap = (*rmap & ~KVMPPC_RMAP_INDEX) | head;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
71
struct revmap_entry *head, *tail;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
76
head = &kvm->arch.hpt.rev[i];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
78
head = real_vmalloc_addr(head);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
79
tail = &kvm->arch.hpt.rev[head->back];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
83
rev->back = head->back;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
85
head->back = pte_index;
arch/powerpc/kvm/test-guest-state-buffer.c
28
const struct kvmppc_gs_elem *head, *curr;
arch/powerpc/kvm/test-guest-state-buffer.c
47
head = kvmppc_gsb_data(gsb);
arch/powerpc/kvm/test-guest-state-buffer.c
48
KUNIT_EXPECT_EQ(test, kvmppc_gse_iden(head), KVMPPC_GSID_GPR(0));
arch/powerpc/kvm/test-guest-state-buffer.c
49
KUNIT_EXPECT_EQ(test, kvmppc_gse_len(head), 8);
arch/powerpc/kvm/test-guest-state-buffer.c
51
memcpy(&data, kvmppc_gse_data(head), 8);
arch/powerpc/mm/book3s64/iommu_api.c
235
static void mm_iommu_free(struct rcu_head *head)
arch/powerpc/mm/book3s64/iommu_api.c
237
struct mm_iommu_table_group_mem_t *mem = container_of(head,
arch/powerpc/mm/pgtable-frag.c
107
static void pte_free_now(struct rcu_head *head)
arch/powerpc/mm/pgtable-frag.c
111
ptdesc = container_of(head, struct ptdesc, pt_rcu_head);
arch/powerpc/perf/vpa-dtl.c
131
if (buf->head + *n_entries >= buf->threshold) {
arch/powerpc/perf/vpa-dtl.c
132
*n_entries = buf->threshold - buf->head;
arch/powerpc/perf/vpa-dtl.c
139
memcpy(aux_copy_buf + buf->head, &dtl->buf[index], *n_entries * sizeof(struct dtl_entry));
arch/powerpc/perf/vpa-dtl.c
147
buf->head = 0;
arch/powerpc/perf/vpa-dtl.c
151
buf->head += *n_entries;
arch/powerpc/perf/vpa-dtl.c
237
aux_buf->head += 1;
arch/powerpc/perf/vpa-dtl.c
254
aux_buf->head = 0;
arch/powerpc/perf/vpa-dtl.c
271
aux_buf->head = 0;
arch/powerpc/perf/vpa-dtl.c
541
buf->head = 0;
arch/powerpc/perf/vpa-dtl.c
87
u64 head;
arch/powerpc/platforms/cell/spufs/file.c
2261
return (ctx->switch_log->head - ctx->switch_log->tail) %
arch/powerpc/platforms/cell/spufs/file.c
2292
ctx->switch_log->head = ctx->switch_log->tail = 0;
arch/powerpc/platforms/cell/spufs/file.c
2445
p = ctx->switch_log->log + ctx->switch_log->head;
arch/powerpc/platforms/cell/spufs/file.c
2452
ctx->switch_log->head =
arch/powerpc/platforms/cell/spufs/file.c
2453
(ctx->switch_log->head + 1) % SWITCH_LOG_BUFSIZE;
arch/powerpc/platforms/cell/spufs/spufs.h
56
unsigned long head;
arch/powerpc/platforms/powernv/opal.c
334
if (opal_msg_notifier_head[msg_type].head == NULL) {
arch/powerpc/platforms/ps3/mm.c
405
list_for_each_entry(c, &r->chunk_list.head, link) {
arch/powerpc/platforms/ps3/mm.c
434
list_for_each_entry(c, &r->chunk_list.head, link) {
arch/powerpc/platforms/ps3/mm.c
540
list_add(&c->link, &r->chunk_list.head);
arch/powerpc/platforms/ps3/mm.c
574
if (list_empty(&r->chunk_list.head)) {
arch/powerpc/platforms/ps3/mm.c
579
last = list_entry(r->chunk_list.head.next,
arch/powerpc/platforms/ps3/mm.c
610
list_add(&c->link, &r->chunk_list.head);
arch/powerpc/platforms/ps3/mm.c
660
INIT_LIST_HEAD(&r->chunk_list.head);
arch/powerpc/platforms/ps3/mm.c
682
INIT_LIST_HEAD(&r->chunk_list.head);
arch/powerpc/platforms/ps3/mm.c
722
list_for_each_entry_safe(c, tmp, &r->chunk_list.head, link) {
arch/powerpc/platforms/ps3/mm.c
745
list_for_each_entry_safe(c, n, &r->chunk_list.head, link) {
arch/powerpc/platforms/pseries/hvcserver.c
119
int hvcs_get_partner_info(uint32_t unit_address, struct list_head *head,
arch/powerpc/platforms/pseries/hvcserver.c
133
if (!head || !pi_buff)
arch/powerpc/platforms/pseries/hvcserver.c
138
INIT_LIST_HEAD(head);
arch/powerpc/platforms/pseries/hvcserver.c
148
if (!list_empty(head))
arch/powerpc/platforms/pseries/hvcserver.c
169
hvcs_free_partner_info(head);
arch/powerpc/platforms/pseries/hvcserver.c
183
list_add_tail(&(next_partner_info->node), head);
arch/powerpc/platforms/pseries/hvcserver.c
62
int hvcs_free_partner_info(struct list_head *head)
arch/powerpc/platforms/pseries/hvcserver.c
67
if (!head)
arch/powerpc/platforms/pseries/hvcserver.c
70
while (!list_empty(head)) {
arch/powerpc/platforms/pseries/hvcserver.c
71
element = head->next;
arch/riscv/kernel/machine_kexec.c
153
unsigned long first_ind_entry = (unsigned long) &image->head;
arch/riscv/kernel/module.c
20
struct list_head head;
arch/riscv/kernel/module.c
31
struct list_head head;
arch/riscv/kernel/module.c
630
used_buckets_list, head) {
arch/riscv/kernel/module.c
638
head) {
arch/riscv/kernel/module.c
671
INIT_LIST_HEAD(&entry->head);
arch/riscv/kernel/module.c
720
INIT_LIST_HEAD(&bucket->head);
arch/riscv/kernel/module.c
722
list_add(&bucket->head, used_buckets_list);
arch/riscv/kernel/module.c
728
list_add_tail(&entry->head, &rel_head->rel_entry);
arch/riscv/kernel/signal.c
246
struct __riscv_ctx_hdr __user *head = sc_ext_ptr;
arch/riscv/kernel/signal.c
248
err |= __get_user(magic, &head->magic);
arch/riscv/kernel/signal.c
249
err |= __get_user(size, &head->size);
arch/riscv/kernel/signal.c
253
sc_ext_ptr += sizeof(*head);
arch/riscv/kernel/signal.c
276
sc_ext_ptr = (void __user *)head + size;
arch/s390/include/uapi/asm/ipl.h
132
__u8 head;
arch/s390/kernel/ipl.c
523
!ipb->br_chr.head && \
arch/s390/kernel/ipl.c
529
ipb->br_chr.head, \
arch/s390/kernel/ipl.c
571
ipb->br_chr.head = args[1]; \
arch/s390/kernel/machine_kexec.c
233
entry = virt_to_phys(&image->head);
arch/s390/kernel/perf_cpum_cf.c
1070
struct list_head head; /* Head of list of active processes */
arch/s390/kernel/perf_cpum_cf.c
1072
.head = LIST_HEAD_INIT(cfset_session.head)
arch/s390/kernel/perf_cpum_cf.c
1256
INIT_LIST_HEAD(&cfset_session.head);
arch/s390/kernel/perf_cpum_cf.c
1274
list_add(&p->node, &cfset_session.head);
arch/s390/kernel/perf_cpum_cf.c
1708
if (!list_empty(&cfset_session.head)) {
arch/s390/kernel/perf_cpum_cf.c
1709
list_for_each_entry(rp, &cfset_session.head, node) {
arch/s390/kernel/perf_cpum_cf.c
1727
if (!list_empty(&cfset_session.head)) {
arch/s390/kernel/perf_cpum_cf.c
1728
list_for_each_entry(rp, &cfset_session.head, node) {
arch/s390/kernel/perf_cpum_sf.c
1271
return aux_sdb_num(aux->head, aux->alert_mark);
arch/s390/kernel/perf_cpum_sf.c
1276
return aux_sdb_num(aux->head, aux->empty_mark);
arch/s390/kernel/perf_cpum_sf.c
1312
for (i = 0, idx = aux->head; i < range_scan; i++, idx++) {
arch/s390/kernel/perf_cpum_sf.c
1338
unsigned long range, i, range_scan, idx, head, base, offset;
arch/s390/kernel/perf_cpum_sf.c
1341
if (handle->head & ~PAGE_MASK)
arch/s390/kernel/perf_cpum_sf.c
1344
aux->head = handle->head >> PAGE_SHIFT;
arch/s390/kernel/perf_cpum_sf.c
1363
aux->empty_mark = aux->head + range - 1;
arch/s390/kernel/perf_cpum_sf.c
1367
aux->alert_mark = aux->head + range/2 - 1;
arch/s390/kernel/perf_cpum_sf.c
1372
head = aux_sdb_index(aux, aux->head);
arch/s390/kernel/perf_cpum_sf.c
1373
base = aux->sdbt_index[head / CPUM_SF_SDB_PER_TABLE];
arch/s390/kernel/perf_cpum_sf.c
1374
offset = head % CPUM_SF_SDB_PER_TABLE;
arch/s390/kernel/perf_cpum_sf.c
1376
cpuhw->lsctl.dear = virt_to_phys((void *)aux->sdb_index[head]);
arch/s390/kernel/perf_cpum_sf.c
1482
aux->empty_mark = aux->head + range - 1;
arch/s390/kernel/perf_cpum_sf.c
1523
aux->head = handle->head >> PAGE_SHIFT;
arch/s390/kernel/perf_cpum_sf.c
1526
aux->alert_mark = aux->head;
arch/s390/kernel/perf_cpum_sf.c
1528
aux->alert_mark = aux->head + range/2 - 1;
arch/s390/kernel/perf_cpum_sf.c
181
unsigned long *sdbt, *curr, *head;
arch/s390/kernel/perf_cpum_sf.c
188
head = sdbt;
arch/s390/kernel/perf_cpum_sf.c
201
} while (curr != head);
arch/s390/kernel/perf_cpum_sf.c
95
unsigned long head; /* index of SDB of buffer head */
arch/s390/kernel/vtime.c
246
static void list_add_sorted(struct vtimer_list *timer, struct list_head *head)
arch/s390/kernel/vtime.c
250
list_for_each_entry(tmp, head, entry) {
arch/s390/kernel/vtime.c
256
list_add_tail(&timer->entry, head);
arch/s390/kvm/gmap.c
1111
struct vsie_rmap *rmap, *rnext, *head;
arch/s390/kvm/gmap.c
1123
head = radix_tree_delete(&sg->host_to_rmap, gfn);
arch/s390/kvm/gmap.c
1124
gmap_for_each_rmap_safe(rmap, rnext, head)
arch/s390/kvm/gmap.c
151
struct vsie_rmap *rmap, *rnext, *head;
arch/s390/kvm/gmap.c
169
head = radix_tree_delete(root, index);
arch/s390/kvm/gmap.c
170
gmap_for_each_rmap_safe(rmap, rnext, head)
arch/s390/kvm/gmap.h
81
#define gmap_for_each_rmap_safe(pos, n, head) \
arch/s390/kvm/gmap.h
82
for (pos = (head); n = pos ? pos->next : NULL, pos; pos = n)
arch/s390/mm/pgalloc.c
149
static void pte_free_now(struct rcu_head *head)
arch/s390/mm/pgalloc.c
151
struct ptdesc *ptdesc = container_of(head, struct ptdesc, pt_rcu_head);
arch/sh/kernel/machine_kexec.c
101
page_list = image->head;
arch/sh/kernel/machine_kexec.c
128
for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE);
arch/sh/kernel/machine_kexec.c
83
for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE);
arch/sh/mm/pmb.c
655
static void __init pmb_merge(struct pmb_entry *head)
arch/sh/mm/pmb.c
661
span = newsize = head->size;
arch/sh/mm/pmb.c
663
tail = head->link;
arch/sh/mm/pmb.c
686
head->flags &= ~PMB_SZ_MASK;
arch/sh/mm/pmb.c
687
head->flags |= pmb_size_to_flags(newsize);
arch/sh/mm/pmb.c
689
head->size = newsize;
arch/sh/mm/pmb.c
691
__pmb_unmap_entry(head->link, depth);
arch/sh/mm/pmb.c
692
__set_pmb_entry(head);
arch/sparc/kernel/ldc.c
1706
static int rx_set_head(struct ldc_channel *lp, unsigned long head)
arch/sparc/kernel/ldc.c
1708
int err = __set_rx_head(lp, head);
arch/sparc/kernel/ldc.c
1713
lp->rx_head = head;
arch/sparc/kernel/ldc.c
320
static int __set_rx_head(struct ldc_channel *lp, unsigned long head)
arch/sparc/kernel/ldc.c
327
err = sun4v_ldc_rx_set_qhead(lp->id, head);
arch/sparc/kernel/ldc.c
760
unsigned long head = lp->tx_acked;
arch/sparc/kernel/ldc.c
764
struct ldc_packet *p = lp->tx_base + (head / LDC_PACKET_SIZE);
arch/sparc/kernel/ldc.c
766
head = tx_advance(lp, head);
arch/sparc/kernel/ldc.c
769
lp->tx_acked = head;
arch/sparc/kernel/ldc.c
772
if (head == lp->tx_tail)
arch/sparc/kernel/pci_fire.c
152
unsigned long *head)
arch/sparc/kernel/pci_fire.c
154
*head = upa_readq(pbm->pbm_regs + EVENT_QUEUE_HEAD(msiqid));
arch/sparc/kernel/pci_fire.c
159
unsigned long *head, unsigned long *msi)
arch/sparc/kernel/pci_fire.c
165
ep = &base[*head];
arch/sparc/kernel/pci_fire.c
186
(*head)++;
arch/sparc/kernel/pci_fire.c
187
if (*head >= pbm->msiq_ent_count)
arch/sparc/kernel/pci_fire.c
188
*head = 0;
arch/sparc/kernel/pci_fire.c
194
unsigned long head)
arch/sparc/kernel/pci_fire.c
196
upa_writeq(head, pbm->pbm_regs + EVENT_QUEUE_HEAD(msiqid));
arch/sparc/kernel/pci_impl.h
37
unsigned long *head);
arch/sparc/kernel/pci_impl.h
39
unsigned long *head, unsigned long *msi);
arch/sparc/kernel/pci_impl.h
41
unsigned long head);
arch/sparc/kernel/pci_msi.c
21
unsigned long orig_head, head;
arch/sparc/kernel/pci_msi.c
26
err = ops->get_head(pbm, msiqid, &head);
arch/sparc/kernel/pci_msi.c
30
orig_head = head;
arch/sparc/kernel/pci_msi.c
34
err = ops->dequeue_msi(pbm, msiqid, &head, &msi);
arch/sparc/kernel/pci_msi.c
48
if (likely(head != orig_head)) {
arch/sparc/kernel/pci_msi.c
49
err = ops->set_head(pbm, msiqid, head);
arch/sparc/kernel/pci_msi.c
63
head, msiqid, err);
arch/sparc/kernel/pci_msi.c
69
head, msiqid, err);
arch/sparc/kernel/pci_sun4v.c
1017
(*head) += sizeof(struct pci_sun4v_msiq_entry);
arch/sparc/kernel/pci_sun4v.c
1018
if (*head >=
arch/sparc/kernel/pci_sun4v.c
1020
*head = 0;
arch/sparc/kernel/pci_sun4v.c
1026
unsigned long head)
arch/sparc/kernel/pci_sun4v.c
1030
err = pci_sun4v_msiq_sethead(pbm->devhandle, msiqid, head);
arch/sparc/kernel/pci_sun4v.c
970
unsigned long *head)
arch/sparc/kernel/pci_sun4v.c
974
err = pci_sun4v_msiq_gethead(pbm->devhandle, msiqid, head);
arch/sparc/kernel/pci_sun4v.c
979
if (unlikely(*head >= limit))
arch/sparc/kernel/pci_sun4v.c
986
unsigned long msiqid, unsigned long *head,
arch/sparc/kernel/pci_sun4v.c
996
*head);
arch/sparc/kernel/pci_sun4v.h
54
unsigned long *head);
arch/sparc/kernel/pci_sun4v.h
57
unsigned long head);
arch/sparc/kernel/pci_sun4v.h
60
unsigned long *head);
arch/sparc/mm/init_64.c
2918
static void pte_free_now(struct rcu_head *head)
arch/sparc/mm/init_64.c
2922
page = container_of(head, struct page, rcu_head);
arch/sparc/mm/srmmu.c
398
static inline void add_to_ctx_list(struct ctx_list *head, struct ctx_list *entry)
arch/sparc/mm/srmmu.c
400
entry->next = head;
arch/sparc/mm/srmmu.c
401
(entry->prev = head->prev)->next = entry;
arch/sparc/mm/srmmu.c
402
head->prev = entry;
arch/um/drivers/line.c
134
if ((line->buffer == NULL) || (line->head == line->tail))
arch/um/drivers/line.c
137
if (line->tail < line->head) {
arch/um/drivers/line.c
139
count = line->buffer + LINE_BUFSIZE - line->head;
arch/um/drivers/line.c
141
n = write_chan(line->chan_out, line->head, count,
arch/um/drivers/line.c
150
line->head = line->buffer;
arch/um/drivers/line.c
152
line->head += n;
arch/um/drivers/line.c
157
count = line->tail - line->head;
arch/um/drivers/line.c
158
n = write_chan(line->chan_out, line->head, count,
arch/um/drivers/line.c
164
line->head += n;
arch/um/drivers/line.c
165
return line->head == line->tail;
arch/um/drivers/line.c
194
if (line->head != line->tail)
arch/um/drivers/line.c
247
line->head = line->buffer;
arch/um/drivers/line.c
43
n = line->head - line->tail;
arch/um/drivers/line.c
97
line->head = line->buffer;
arch/um/drivers/line.h
51
u8 *head;
arch/um/drivers/vector_kern.c
262
qi->head =
arch/um/drivers/vector_kern.c
263
(qi->head + advance)
arch/um/drivers/vector_kern.c
377
for (skb_index = qi->head; skb_index < qi->head + count; skb_index++) {
arch/um/drivers/vector_kern.c
411
send_from += qi->head;
arch/um/drivers/vector_kern.c
413
if (send_len + qi->head > qi->max_depth)
arch/um/drivers/vector_kern.c
414
send_len = qi->max_depth - qi->head;
arch/um/drivers/vector_kern.c
574
result->head = 0;
arch/um/drivers/vector_kern.c
670
qi->head = qi->tail = 0;
arch/um/drivers/vector_kern.h
49
int head, tail, max_depth, max_iov_frags;
arch/x86/events/intel/bts.c
153
index = local_read(&bb->head);
arch/x86/events/intel/bts.c
177
static void bts_buffer_pad_out(struct bts_phys *phys, unsigned long head)
arch/x86/events/intel/bts.c
179
unsigned long index = head - phys->offset;
arch/x86/events/intel/bts.c
189
unsigned long index = ds->bts_index - ds->bts_buffer_base, old, head;
arch/x86/events/intel/bts.c
194
head = index + bts_buffer_offset(bb, bb->cur_buf);
arch/x86/events/intel/bts.c
195
old = local_xchg(&bb->head, head);
arch/x86/events/intel/bts.c
198
if (old == head)
arch/x86/events/intel/bts.c
209
local_add(head - old, &bb->data_size);
arch/x86/events/intel/bts.c
211
local_set(&bb->data_size, head);
arch/x86/events/intel/bts.c
326
bts->handle.head =
arch/x86/events/intel/bts.c
388
unsigned long head, space, next_space, pad, gap, skip, wakeup;
arch/x86/events/intel/bts.c
396
head = handle->head & ((bb->nr_pages << PAGE_SHIFT) - 1);
arch/x86/events/intel/bts.c
399
space = phys->offset + phys->displacement + phys->size - head;
arch/x86/events/intel/bts.c
422
bts_buffer_pad_out(phys, head);
arch/x86/events/intel/bts.c
429
head = phys->offset + phys->displacement;
arch/x86/events/intel/bts.c
436
local_set(&bb->head, head);
arch/x86/events/intel/bts.c
443
handle->head;
arch/x86/events/intel/bts.c
449
bb->end = head + space;
arch/x86/events/intel/bts.c
501
old_head = local_read(&bb->head);
arch/x86/events/intel/bts.c
505
if (old_head == local_read(&bb->head))
arch/x86/events/intel/bts.c
59
local_t head;
arch/x86/events/intel/pt.c
1113
unsigned long head = local64_read(&buf->head);
arch/x86/events/intel/pt.c
1143
if (!offset_in_page(head + handle->size + 1))
arch/x86/events/intel/pt.c
1146
idx = (head >> PAGE_SHIFT) + npages;
arch/x86/events/intel/pt.c
1158
idx = (head >> PAGE_SHIFT) + npages - 1;
arch/x86/events/intel/pt.c
1191
static void pt_buffer_reset_offsets(struct pt_buffer *buf, unsigned long head)
arch/x86/events/intel/pt.c
1198
head &= (buf->nr_pages << PAGE_SHIFT) - 1;
arch/x86/events/intel/pt.c
1201
pg = (head >> PAGE_SHIFT) & (buf->nr_pages - 1);
arch/x86/events/intel/pt.c
1207
buf->output_off = head & (pt_buffer_region_size(buf) - 1);
arch/x86/events/intel/pt.c
1209
buf->output_off = head;
arch/x86/events/intel/pt.c
1212
local64_set(&buf->head, head);
arch/x86/events/intel/pt.c
1454
struct perf_addr_filters_head *head = perf_event_addr_filters(event);
arch/x86/events/intel/pt.c
1464
list_for_each_entry(filter, &head->list, entry) {
arch/x86/events/intel/pt.c
1546
pt_buffer_reset_offsets(buf, pt->handle.head);
arch/x86/events/intel/pt.c
1630
pt_buffer_reset_offsets(buf, pt->handle.head);
arch/x86/events/intel/pt.c
1698
pt->handle.head =
arch/x86/events/intel/pt.c
894
old = (local64_xchg(&buf->head, base) &
arch/x86/events/intel/pt.c
963
void *head = pt_buffer_region(buf);
arch/x86/events/intel/pt.c
966
memset(head + buf->output_off, 0,
arch/x86/events/intel/pt.h
80
local64_t head;
arch/x86/kernel/apic/io_apic.c
80
#define for_each_irq_pin(entry, head) \
arch/x86/kernel/apic/io_apic.c
81
list_for_each_entry(entry, &head, list)
arch/x86/kernel/apic/vector.c
1051
hlist_for_each_entry_safe(apicd, tmp, &cl->head, clist) {
arch/x86/kernel/apic/vector.c
1101
hlist_add_head(&apicd->clist, &cl->head);
arch/x86/kernel/apic/vector.c
51
struct hlist_head head;
arch/x86/kernel/apic/vector.c
56
.head = HLIST_HEAD_INIT,
arch/x86/kernel/apic/vector.c
874
WARN_ON_ONCE(!hlist_empty(&cl->head));
arch/x86/kernel/cpu/mce/genpool.c
57
struct llist_node *head;
arch/x86/kernel/cpu/mce/genpool.c
61
head = llist_del_all(&mce_event_llist);
arch/x86/kernel/cpu/mce/genpool.c
62
if (!head)
arch/x86/kernel/cpu/mce/genpool.c
66
llist_for_each_entry_safe(node, t, head, llnode) {
arch/x86/kernel/cpu/mce/genpool.c
77
struct llist_node *head;
arch/x86/kernel/cpu/mce/genpool.c
80
head = llist_del_all(&mce_event_llist);
arch/x86/kernel/cpu/mce/genpool.c
81
if (!head)
arch/x86/kernel/cpu/mce/genpool.c
84
head = llist_reverse_order(head);
arch/x86/kernel/cpu/mce/genpool.c
85
llist_for_each_entry_safe(node, tmp, head, llnode) {
arch/x86/kernel/machine_kexec_32.c
223
image->start = relocate_kernel_ptr((unsigned long)image->head,
arch/x86/kernel/machine_kexec_64.c
470
image->start = relocate_kernel_ptr((unsigned long)image->head,
arch/x86/kernel/nmi.c
158
list_for_each_entry_rcu(a, &desc->head, list) {
arch/x86/kernel/nmi.c
192
WARN_ON_ONCE(type == NMI_SERR && !list_empty(&desc->head));
arch/x86/kernel/nmi.c
193
WARN_ON_ONCE(type == NMI_IO_CHECK && !list_empty(&desc->head));
arch/x86/kernel/nmi.c
200
list_add_rcu(&action->list, &desc->head);
arch/x86/kernel/nmi.c
202
list_add_tail_rcu(&action->list, &desc->head);
arch/x86/kernel/nmi.c
217
list_for_each_entry_rcu(n, &desc->head, list) {
arch/x86/kernel/nmi.c
50
struct list_head head;
arch/x86/kernel/nmi.c
55
.head = LIST_HEAD_INIT(nmi_desc[type].head), \
arch/x86/kvm/mmu/mmu.c
1214
struct rmap_head *head;
arch/x86/kvm/mmu/page_track.c
154
struct kvm_page_track_notifier_head *head;
arch/x86/kvm/mmu/page_track.c
156
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
157
cleanup_srcu_struct(&head->track_srcu);
arch/x86/kvm/mmu/page_track.c
162
struct kvm_page_track_notifier_head *head;
arch/x86/kvm/mmu/page_track.c
164
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
165
INIT_HLIST_HEAD(&head->track_notifier_list);
arch/x86/kvm/mmu/page_track.c
166
return init_srcu_struct(&head->track_srcu);
arch/x86/kvm/mmu/page_track.c
223
struct kvm_page_track_notifier_head *head;
arch/x86/kvm/mmu/page_track.c
237
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
240
hlist_add_head_rcu(&n->node, &head->track_notifier_list);
arch/x86/kvm/mmu/page_track.c
253
struct kvm_page_track_notifier_head *head;
arch/x86/kvm/mmu/page_track.c
255
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
260
synchronize_srcu(&head->track_srcu);
arch/x86/kvm/mmu/page_track.c
275
struct kvm_page_track_notifier_head *head;
arch/x86/kvm/mmu/page_track.c
279
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
281
if (hlist_empty(&head->track_notifier_list))
arch/x86/kvm/mmu/page_track.c
284
idx = srcu_read_lock(&head->track_srcu);
arch/x86/kvm/mmu/page_track.c
285
hlist_for_each_entry_srcu(n, &head->track_notifier_list, node,
arch/x86/kvm/mmu/page_track.c
286
srcu_read_lock_held(&head->track_srcu))
arch/x86/kvm/mmu/page_track.c
289
srcu_read_unlock(&head->track_srcu, idx);
arch/x86/kvm/mmu/page_track.c
298
struct kvm_page_track_notifier_head *head;
arch/x86/kvm/mmu/page_track.c
302
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
304
if (hlist_empty(&head->track_notifier_list))
arch/x86/kvm/mmu/page_track.c
307
idx = srcu_read_lock(&head->track_srcu);
arch/x86/kvm/mmu/page_track.c
308
hlist_for_each_entry_srcu(n, &head->track_notifier_list, node,
arch/x86/kvm/mmu/page_track.c
309
srcu_read_lock_held(&head->track_srcu))
arch/x86/kvm/mmu/page_track.c
312
srcu_read_unlock(&head->track_srcu, idx);
arch/x86/kvm/mmu/tdp_mmu.c
71
static void tdp_mmu_free_sp_rcu_callback(struct rcu_head *head)
arch/x86/kvm/mmu/tdp_mmu.c
73
struct kvm_mmu_page *sp = container_of(head, struct kvm_mmu_page,
arch/x86/kvm/svm/sev.c
2754
struct list_head *head = &sev->regions_list;
arch/x86/kvm/svm/sev.c
2757
list_for_each_entry(i, head, list) {
arch/x86/kvm/svm/sev.c
2904
struct list_head *head = &sev->regions_list;
arch/x86/kvm/svm/sev.c
2934
if (!list_empty(head)) {
arch/x86/kvm/svm/sev.c
2935
list_for_each_safe(pos, q, head) {
arch/x86/mm/kmmio.c
117
struct list_head *head;
arch/x86/mm/kmmio.c
125
head = kmmio_page_list(addr);
arch/x86/mm/kmmio.c
126
list_for_each_entry_rcu(f, head, list) {
arch/x86/mm/kmmio.c
479
static void rcu_free_kmmio_fault_pages(struct rcu_head *head)
arch/x86/mm/kmmio.c
482
head,
arch/x86/mm/kmmio.c
495
static void remove_kmmio_fault_pages(struct rcu_head *head)
arch/x86/mm/kmmio.c
498
container_of(head, struct kmmio_delayed_release, rcu);
arch/x86/platform/uv/uv_time.c
147
struct uv_rtc_timer_head *head = blade_info[bid];
arch/x86/platform/uv/uv_time.c
149
if (!head) {
arch/x86/platform/uv/uv_time.c
150
head = kmalloc_node(struct_size(head, cpu,
arch/x86/platform/uv/uv_time.c
153
if (!head) {
arch/x86/platform/uv/uv_time.c
157
spin_lock_init(&head->lock);
arch/x86/platform/uv/uv_time.c
158
head->ncpus = uv_blade_nr_possible_cpus(bid);
arch/x86/platform/uv/uv_time.c
159
head->next_cpu = -1;
arch/x86/platform/uv/uv_time.c
160
blade_info[bid] = head;
arch/x86/platform/uv/uv_time.c
163
head->cpu[bcpu].lcpu = cpu;
arch/x86/platform/uv/uv_time.c
164
head->cpu[bcpu].expires = ULLONG_MAX;
arch/x86/platform/uv/uv_time.c
171
static void uv_rtc_find_next_timer(struct uv_rtc_timer_head *head, int pnode)
arch/x86/platform/uv/uv_time.c
176
head->next_cpu = -1;
arch/x86/platform/uv/uv_time.c
177
for (c = 0; c < head->ncpus; c++) {
arch/x86/platform/uv/uv_time.c
178
u64 exp = head->cpu[c].expires;
arch/x86/platform/uv/uv_time.c
185
head->next_cpu = bcpu;
arch/x86/platform/uv/uv_time.c
186
c = head->cpu[bcpu].lcpu;
arch/x86/platform/uv/uv_time.c
205
struct uv_rtc_timer_head *head = blade_info[bid];
arch/x86/platform/uv/uv_time.c
207
u64 *t = &head->cpu[bcpu].expires;
arch/x86/platform/uv/uv_time.c
211
spin_lock_irqsave(&head->lock, flags);
arch/x86/platform/uv/uv_time.c
213
next_cpu = head->next_cpu;
arch/x86/platform/uv/uv_time.c
218
expires < head->cpu[next_cpu].expires) {
arch/x86/platform/uv/uv_time.c
219
head->next_cpu = bcpu;
arch/x86/platform/uv/uv_time.c
222
uv_rtc_find_next_timer(head, pnode);
arch/x86/platform/uv/uv_time.c
223
spin_unlock_irqrestore(&head->lock, flags);
arch/x86/platform/uv/uv_time.c
228
spin_unlock_irqrestore(&head->lock, flags);
arch/x86/platform/uv/uv_time.c
241
struct uv_rtc_timer_head *head = blade_info[bid];
arch/x86/platform/uv/uv_time.c
243
u64 *t = &head->cpu[bcpu].expires;
arch/x86/platform/uv/uv_time.c
247
spin_lock_irqsave(&head->lock, flags);
arch/x86/platform/uv/uv_time.c
249
if ((head->next_cpu == bcpu && uv_read_rtc(NULL) >= *t) || force)
arch/x86/platform/uv/uv_time.c
255
if (head->next_cpu == bcpu)
arch/x86/platform/uv/uv_time.c
256
uv_rtc_find_next_timer(head, pnode);
arch/x86/platform/uv/uv_time.c
259
spin_unlock_irqrestore(&head->lock, flags);
block/blk-cgroup.c
219
if (bios.head && bios.head->bi_next) {
block/blk-crypto-profile.c
191
const struct hlist_head *head =
block/blk-crypto-profile.c
195
hlist_for_each_entry(slotp, head, hash_node) {
block/blk-mq-tag.c
585
static void blk_mq_free_tags_callback(struct rcu_head *head)
block/blk-mq-tag.c
587
struct blk_mq_tags *tags = container_of(head, struct blk_mq_tags,
block/blk-mq.c
2866
struct request **prev = &rqs->head;
block/blk-mq.c
3429
list->head = rq->bio;
block/blk-mq.c
3875
static void __blk_mq_remove_cpuhp_list(struct list_head *head)
block/blk-mq.c
3881
list_for_each_entry(hctx, head, hctx_list)
block/blk-mq.c
3942
static void blk_free_flush_queue_callback(struct rcu_head *head)
block/blk-mq.c
3945
container_of(head, struct blk_flush_queue, rcu_head);
block/blk-stat.c
167
static void blk_stat_free_callback_rcu(struct rcu_head *head)
block/blk-stat.c
171
cb = container_of(head, struct blk_stat_callback, rcu);
block/kyber-iosched.c
586
struct list_head *head = &kcq->rq_list[sched_domain];
block/kyber-iosched.c
591
list_move(&rq->queuelist, head);
block/kyber-iosched.c
593
list_move_tail(&rq->queuelist, head);
block/partitions/ibm.c
34
__u16 head;
block/partitions/ibm.c
40
head = ptr->hh & 0x000F;
block/partitions/ibm.c
42
head * geo->sectors;
block/partitions/ibm.c
52
__u16 head;
block/partitions/ibm.c
58
head = ptr->hh & 0x000F;
block/partitions/ibm.c
60
head * geo->sectors +
drivers/accel/amdxdna/amdxdna_mailbox.c
191
u32 head, tail;
drivers/accel/amdxdna/amdxdna_mailbox.c
196
head = mailbox_get_headptr(mb_chann, CHAN_RES_X2I);
drivers/accel/amdxdna/amdxdna_mailbox.c
204
if (tail >= head && tmp_tail > ringbuf_size) {
drivers/accel/amdxdna/amdxdna_mailbox.c
213
if (tail < head && tmp_tail >= head) {
drivers/accel/amdxdna/amdxdna_mailbox.c
214
ret = read_poll_timeout(mailbox_get_headptr, head,
drivers/accel/amdxdna/amdxdna_mailbox.c
215
tmp_tail < head || tail >= head,
drivers/accel/amdxdna/amdxdna_mailbox.c
220
if (tail >= head)
drivers/accel/amdxdna/amdxdna_mailbox.c
274
u32 head, tail;
drivers/accel/amdxdna/amdxdna_mailbox.c
279
head = mb_chann->i2x_head;
drivers/accel/amdxdna/amdxdna_mailbox.c
289
if (head == tail)
drivers/accel/amdxdna/amdxdna_mailbox.c
292
if (head == ringbuf_size)
drivers/accel/amdxdna/amdxdna_mailbox.c
293
head = 0;
drivers/accel/amdxdna/amdxdna_mailbox.c
296
read_addr = mb_chann->mb->res.ringbuf_base + start_addr + head;
drivers/accel/amdxdna/amdxdna_mailbox.c
300
if (head < tail) {
drivers/accel/amdxdna/amdxdna_mailbox.c
302
head, tail);
drivers/accel/amdxdna/amdxdna_mailbox.c
315
if (msg_size > ringbuf_size - head || msg_size > tail - head) {
drivers/accel/amdxdna/amdxdna_mailbox.c
317
msg_size, tail, head);
drivers/accel/amdxdna/amdxdna_mailbox.c
328
mailbox_set_headptr(mb_chann, head + msg_size);
drivers/accel/habanalabs/common/debugfs.c
1017
u32 i, head, count = 0;
drivers/accel/habanalabs/common/debugfs.c
1024
head = dbgfs->head;
drivers/accel/habanalabs/common/debugfs.c
1025
if (head == 0)
drivers/accel/habanalabs/common/debugfs.c
1028
i = head - 1;
drivers/accel/habanalabs/common/debugfs.c
1078
new_entry = &dbgfs_cfg_accesses->cfg_access_list[dbgfs_cfg_accesses->head];
drivers/accel/habanalabs/common/debugfs.c
1083
dbgfs_cfg_accesses->head = (dbgfs_cfg_accesses->head + 1)
drivers/accel/habanalabs/common/debugfs.c
2079
hdev->debugfs_cfg_accesses.head = 0; /* already zero by alloc but explicit init is fine */
drivers/accel/habanalabs/common/habanalabs.h
2472
u32 head;
drivers/accel/ivpu/ivpu_job.c
248
jobq->header.head = 0;
drivers/accel/ivpu/ivpu_job.c
432
if (next_entry == header->head) {
drivers/accel/ivpu/ivpu_job.c
434
job->file_priv->ctx.id, cmdq->id, cmdq->db_id, header->head, tail);
drivers/accel/ivpu/ivpu_job.c
726
cmdq->jobq->header.head = cmdq->jobq->header.tail;
drivers/accel/ivpu/vpu_jsm_api.h
320
u32 head;
drivers/accel/qaic/qaic.h
336
void qaic_data_get_fifo_info(struct dma_bridge_chan *dbc, u32 *head, u32 *tail);
drivers/accel/qaic/qaic_control.c
1490
free_wrapper_from_list(wrapper->head, wrapper);
drivers/accel/qaic/qaic_control.c
351
w->head = wrappers;
drivers/accel/qaic/qaic_control.c
79
struct wrapper_list *head;
drivers/accel/qaic/qaic_data.c
1093
static inline u32 fifo_space_avail(u32 head, u32 tail, u32 q_size)
drivers/accel/qaic/qaic_data.c
1095
u32 avail = head - tail - 1;
drivers/accel/qaic/qaic_data.c
1097
if (head <= tail)
drivers/accel/qaic/qaic_data.c
1104
u32 head, u32 *ptail)
drivers/accel/qaic/qaic_data.c
1111
avail = fifo_space_avail(head, tail, dbc->nelem);
drivers/accel/qaic/qaic_data.c
1133
u64 resize, struct dma_bridge_chan *dbc, u32 head,
drivers/accel/qaic/qaic_data.c
1143
avail = fifo_space_avail(head, tail, dbc->nelem);
drivers/accel/qaic/qaic_data.c
1201
bool is_partial, struct dma_bridge_chan *dbc, u32 head,
drivers/accel/qaic/qaic_data.c
1254
ret = copy_partial_exec_reqs(qdev, slice, 0, dbc, head, tail);
drivers/accel/qaic/qaic_data.c
1259
head, tail);
drivers/accel/qaic/qaic_data.c
1261
ret = copy_exec_reqs(qdev, slice, dbc->id, head, tail);
drivers/accel/qaic/qaic_data.c
1336
u32 head;
drivers/accel/qaic/qaic_data.c
1387
head = readl(dbc->dbc_base + REQHP_OFF);
drivers/accel/qaic/qaic_data.c
1390
if (head == U32_MAX || tail == U32_MAX) {
drivers/accel/qaic/qaic_data.c
1396
queue_level = head <= tail ? tail - head : dbc->nelem - (head - tail);
drivers/accel/qaic/qaic_data.c
1399
head, &tail);
drivers/accel/qaic/qaic_data.c
1474
u32 head;
drivers/accel/qaic/qaic_data.c
1494
head = readl(dbc->dbc_base + RSPHP_OFF);
drivers/accel/qaic/qaic_data.c
1495
if (head == U32_MAX) { /* PCI link error */
drivers/accel/qaic/qaic_data.c
1506
if (head == tail) { /* queue empty */
drivers/accel/qaic/qaic_data.c
1522
u32 head;
drivers/accel/qaic/qaic_data.c
1544
head = readl(dbc->dbc_base + RSPHP_OFF);
drivers/accel/qaic/qaic_data.c
1545
if (head == U32_MAX) { /* PCI link error */
drivers/accel/qaic/qaic_data.c
1556
if (head != tail) {
drivers/accel/qaic/qaic_data.c
1579
u32 head;
drivers/accel/qaic/qaic_data.c
1585
head = readl(dbc->dbc_base + RSPHP_OFF);
drivers/accel/qaic/qaic_data.c
1586
if (head == U32_MAX) /* PCI link error */
drivers/accel/qaic/qaic_data.c
1607
if (head == tail) { /* queue empty */
drivers/accel/qaic/qaic_data.c
1617
while (head != tail) {
drivers/accel/qaic/qaic_data.c
1621
rsp = dbc->rsp_q_base + head * sizeof(*rsp);
drivers/accel/qaic/qaic_data.c
1655
head = (head + 1) % dbc->nelem;
drivers/accel/qaic/qaic_data.c
1662
writel(head, dbc->dbc_base + RSPHP_OFF);
drivers/accel/qaic/qaic_data.c
1674
if (tail != U32_MAX && head != tail) {
drivers/accel/qaic/qaic_data.c
2038
void qaic_data_get_fifo_info(struct dma_bridge_chan *dbc, u32 *head, u32 *tail)
drivers/accel/qaic/qaic_data.c
2040
if (!dbc || !head || !tail)
drivers/accel/qaic/qaic_data.c
2043
*head = readl(dbc->dbc_base + REQHP_OFF);
drivers/accel/qaic/qaic_debugfs.c
82
u32 tail = 0, head = 0;
drivers/accel/qaic/qaic_debugfs.c
84
qaic_data_get_fifo_info(dbc, &head, &tail);
drivers/accel/qaic/qaic_debugfs.c
86
if (head == U32_MAX || tail == U32_MAX)
drivers/accel/qaic/qaic_debugfs.c
88
else if (head > tail)
drivers/accel/qaic/qaic_debugfs.c
89
seq_printf(s, "%u\n", dbc->nelem - head + tail);
drivers/accel/qaic/qaic_debugfs.c
91
seq_printf(s, "%u\n", tail - head);
drivers/acpi/acpi_dbg.c
262
p = &crc->buf[crc->head];
drivers/acpi/acpi_dbg.c
267
crc->head = (crc->head + n) & (ACPI_AML_BUF_SIZE - 1);
drivers/acpi/acpi_dbg.c
28
(CIRC_CNT((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE))
drivers/acpi/acpi_dbg.c
30
(CIRC_CNT_TO_END((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE))
drivers/acpi/acpi_dbg.c
32
(CIRC_SPACE((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE))
drivers/acpi/acpi_dbg.c
34
(CIRC_SPACE_TO_END((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE))
drivers/acpi/acpi_dbg.c
513
acpi_aml_io.out_crc.head = acpi_aml_io.out_crc.tail = 0;
drivers/acpi/acpi_dbg.c
514
acpi_aml_io.in_crc.head = acpi_aml_io.in_crc.tail = 0;
drivers/acpi/acpi_dbg.c
654
p = &crc->buf[crc->head];
drivers/acpi/acpi_dbg.c
662
crc->head = (crc->head + n) & (ACPI_AML_BUF_SIZE - 1);
drivers/acpi/acpi_ipmi.c
125
INIT_LIST_HEAD(&ipmi_device->head);
drivers/acpi/acpi_ipmi.c
161
list_del(&ipmi_device->head);
drivers/acpi/acpi_ipmi.c
208
INIT_LIST_HEAD(&ipmi_msg->head);
drivers/acpi/acpi_ipmi.c
30
struct list_head head;
drivers/acpi/acpi_ipmi.c
344
head);
drivers/acpi/acpi_ipmi.c
345
list_del(&tx_msg->head);
drivers/acpi/acpi_ipmi.c
363
list_for_each_entry_safe(iter, temp, &ipmi->tx_msg_list, head) {
drivers/acpi/acpi_ipmi.c
366
list_del(&iter->head);
drivers/acpi/acpi_ipmi.c
391
list_for_each_entry_safe(iter, temp, &ipmi_device->tx_msg_list, head) {
drivers/acpi/acpi_ipmi.c
394
list_del(&iter->head);
drivers/acpi/acpi_ipmi.c
461
list_for_each_entry(temp, &driver_data.ipmi_devices, head) {
drivers/acpi/acpi_ipmi.c
473
list_add_tail(&ipmi_device->head, &driver_data.ipmi_devices);
drivers/acpi/acpi_ipmi.c
492
&driver_data.ipmi_devices, head) {
drivers/acpi/acpi_ipmi.c
502
struct acpi_ipmi_device, head);
drivers/acpi/acpi_ipmi.c
562
list_add_tail(&tx_msg->head, &ipmi_device->tx_msg_list);
drivers/acpi/acpi_ipmi.c
63
struct list_head head;
drivers/acpi/acpi_ipmi.c
649
head);
drivers/acpi/acpica/acconvert.h
73
struct acpi_file_node *head);
drivers/acpi/apei/ghes.c
1070
static void ghes_estatus_cache_rcu_free(struct rcu_head *head)
drivers/acpi/apei/ghes.c
1075
cache = container_of(head, struct ghes_estatus_cache, rcu);
drivers/acpi/arm64/iort.c
1011
struct list_head *head)
drivers/acpi/arm64/iort.c
1048
list_add_tail(&rmr_data->rr.list, head);
drivers/acpi/arm64/iort.c
1102
struct device *dev, struct list_head *head)
drivers/acpi/arm64/iort.c
1155
iort_get_rmrs(node, smmu, sids, num_sids, head);
drivers/acpi/arm64/iort.c
1160
struct list_head *head)
drivers/acpi/arm64/iort.c
1183
iort_node_get_rmr_info(iort_node, iommu, dev, head);
drivers/acpi/arm64/iort.c
1197
struct list_head *head)
drivers/acpi/arm64/iort.c
1205
iort_find_rmrs(iommu, dev, head);
drivers/acpi/arm64/iort.c
1232
struct list_head *head)
drivers/acpi/arm64/iort.c
1277
list_add_tail(&region->list, head);
drivers/acpi/arm64/iort.c
1287
void iort_iommu_get_resv_regions(struct device *dev, struct list_head *head)
drivers/acpi/arm64/iort.c
1291
iort_iommu_msi_get_resv_regions(dev, head);
drivers/acpi/arm64/iort.c
1292
iort_iommu_rmr_get_resv_regions(fwspec->iommu_fwnode, dev, head);
drivers/acpi/arm64/iort.c
1302
struct list_head *head)
drivers/acpi/arm64/iort.c
1304
iort_iommu_rmr_get_resv_regions(iommu_fwnode, NULL, head);
drivers/acpi/arm64/iort.c
1314
struct list_head *head)
drivers/acpi/arm64/iort.c
1318
list_for_each_entry_safe(entry, next, head, list)
drivers/acpi/arm64/iort.c
1491
void iort_iommu_get_resv_regions(struct device *dev, struct list_head *head)
drivers/acpi/property.c
1355
const struct list_head *head;
drivers/acpi/property.c
1368
head = &adev->data.subnodes;
drivers/acpi/property.c
1370
head = &data->data.subnodes;
drivers/acpi/property.c
1374
if (list_empty(head))
drivers/acpi/property.c
1380
if (next == head)
drivers/acpi/property.c
1385
dn = list_first_entry(head, struct acpi_data_node, sibling);
drivers/acpi/utils.c
157
head = buffer->pointer;
drivers/acpi/utils.c
173
*((u64 *) head) =
drivers/acpi/utils.c
175
head += sizeof(u64);
drivers/acpi/utils.c
178
pointer = (u8 **) head;
drivers/acpi/utils.c
182
head += sizeof(u64 *);
drivers/acpi/utils.c
198
pointer = (u8 **) head;
drivers/acpi/utils.c
202
head += sizeof(char *);
drivers/acpi/utils.c
209
pointer = (u8 **) head;
drivers/acpi/utils.c
213
head += sizeof(u8 *);
drivers/acpi/utils.c
224
*(void **)head =
drivers/acpi/utils.c
226
head += sizeof(void *);
drivers/acpi/utils.c
40
u8 *head = NULL;
drivers/ata/libata-core.c
682
u32 cyl, head, sect;
drivers/ata/libata-core.c
685
head = tf->device & 0xf;
drivers/ata/libata-core.c
694
block = (cyl * dev->heads + head) * dev->sectors + sect - 1;
drivers/ata/libata-core.c
825
u32 sect, head, cyl, track;
drivers/ata/libata-core.c
837
head = track % dev->heads;
drivers/ata/libata-core.c
844
if ((cyl >> 16) || (head >> 4) || (sect >> 8) || (!sect))
drivers/ata/libata-core.c
851
tf->device |= head;
drivers/ata/libata-scsi.c
1473
u32 sect, head, cyl, track;
drivers/ata/libata-scsi.c
1481
head = track % dev->heads;
drivers/ata/libata-scsi.c
1488
if ((cyl >> 16) || (head >> 4) || (sect >> 8) || (!sect))
drivers/ata/libata-scsi.c
1496
tf->device |= head;
drivers/ata/sata_nv.c
1690
WARN_ON(dq->tail - dq->head == ATA_MAX_QUEUE);
drivers/ata/sata_nv.c
1701
if (dq->head == dq->tail) /* null queue */
drivers/ata/sata_nv.c
1704
tag = dq->tag[dq->head & (ATA_MAX_QUEUE - 1)];
drivers/ata/sata_nv.c
1705
dq->tag[dq->head++ & (ATA_MAX_QUEUE - 1)] = ATA_TAG_POISON;
drivers/ata/sata_nv.c
1727
dq->head = 0;
drivers/ata/sata_nv.c
246
unsigned int head;
drivers/atm/atmtcp.c
172
struct hlist_head *head = &vcc_hash[i];
drivers/atm/atmtcp.c
174
sk_for_each(s, head) {
drivers/atm/atmtcp.c
265
struct hlist_head *head;
drivers/atm/atmtcp.c
269
head = &vcc_hash[vci & (VCC_HTABLE_SIZE -1)];
drivers/atm/atmtcp.c
271
sk_for_each(s, head) {
drivers/atm/eni.c
2159
struct hlist_head *head = &vcc_hash[i];
drivers/atm/eni.c
2161
sk_for_each(s, head) {
drivers/atm/fore200e.c
1095
entry = &rxq->host_entry[ rxq->head ];
drivers/atm/fore200e.c
1126
FORE200E_NEXT_ENTRY(rxq->head, QUEUE_SIZE_RX);
drivers/atm/fore200e.c
1227
struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ];
drivers/atm/fore200e.c
1234
FORE200E_NEXT_ENTRY(cmdq->head, QUEUE_SIZE_CMD);
drivers/atm/fore200e.c
1562
entry = &txq->host_entry[ txq->head ];
drivers/atm/fore200e.c
1614
FORE200E_NEXT_ENTRY(txq->head, QUEUE_SIZE_TX);
drivers/atm/fore200e.c
1673
struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ];
drivers/atm/fore200e.c
1689
FORE200E_NEXT_ENTRY(cmdq->head, QUEUE_SIZE_CMD);
drivers/atm/fore200e.c
1719
struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ];
drivers/atm/fore200e.c
1726
FORE200E_NEXT_ENTRY(cmdq->head, QUEUE_SIZE_CMD);
drivers/atm/fore200e.c
1759
struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ];
drivers/atm/fore200e.c
1765
FORE200E_NEXT_ENTRY(cmdq->head, QUEUE_SIZE_CMD);
drivers/atm/fore200e.c
2163
rxq->head = 0;
drivers/atm/fore200e.c
2224
txq->head = 0;
drivers/atm/fore200e.c
2266
cmdq->head = 0;
drivers/atm/fore200e.c
558
struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ];
drivers/atm/fore200e.c
563
FORE200E_NEXT_ENTRY(cmdq->head, QUEUE_SIZE_CMD);
drivers/atm/fore200e.c
924
entry = &bsq->host_entry[ bsq->head ];
drivers/atm/fore200e.c
947
FORE200E_NEXT_ENTRY(bsq->head, QUEUE_SIZE_BS);
drivers/atm/fore200e.h
603
int head; /* head of cmd queue */
drivers/atm/fore200e.h
612
int head; /* head of tx queue */
drivers/atm/fore200e.h
624
int head; /* head of rx queue */
drivers/atm/fore200e.h
634
int head; /* head of buffer supply queue */
drivers/atm/he.c
330
struct hlist_head *head;
drivers/atm/he.c
338
head = &vcc_hash[vci & (VCC_HTABLE_SIZE -1)];
drivers/atm/he.c
340
sk_for_each(s, head) {
drivers/atm/idt77252.c
1243
u32 head, tail;
drivers/atm/idt77252.c
1257
head = IDT77252_PRV_PADDR(queue) + (queue->data - queue->head - 16);
drivers/atm/idt77252.c
1264
while (head != tail) {
drivers/atm/idt77252.c
1337
head = IDT77252_PRV_PADDR(queue)
drivers/atm/idt77252.c
1338
+ (queue->data - queue->head - 16);
drivers/atm/idt77252.c
1344
head = le32_to_cpu(*(u32 *) &queue->data[0]);
drivers/atm/idt77252.c
1802
skb->data = skb->head;
drivers/atm/nicstar.c
2096
iovb->data = iovb->head;
drivers/atm/nicstar.c
2108
iovb->data = iovb->head;
drivers/atm/solos-pci.c
899
struct hlist_head *head;
drivers/atm/solos-pci.c
904
head = &vcc_hash[vci & (VCC_HTABLE_SIZE -1)];
drivers/atm/solos-pci.c
905
sk_for_each(s, head) {
drivers/base/attribute_container.c
180
#define klist_for_each_entry(pos, head, member, iter) \
drivers/base/attribute_container.c
181
for (klist_iter_init(head, iter); (pos = ({ \
drivers/base/firmware_loader/main.c
1416
struct list_head *head = data;
drivers/base/firmware_loader/main.c
1421
list_add(&fce->list, head);
drivers/base/firmware_loader/main.c
161
list_for_each_entry(tmp, &fwc->head, list)
drivers/base/firmware_loader/main.c
1634
INIT_LIST_HEAD(&fw_cache.head);
drivers/base/firmware_loader/main.c
194
list_add(&tmp->list, &fwc->head);
drivers/base/firmware_loader/main.c
54
struct list_head head;
drivers/base/power/wakeup.c
22
#define list_for_each_entry_rcu_locked(pos, head, member) \
drivers/base/power/wakeup.c
23
list_for_each_entry_rcu(pos, head, member, \
drivers/block/amiflop.c
1584
getprm.head=p->type->heads;
drivers/block/aoe/aoe.h
120
struct list_head head;
drivers/block/aoe/aoecmd.c
102
head = &d->factive[n];
drivers/block/aoe/aoecmd.c
103
list_for_each_safe(pos, nx, head) {
drivers/block/aoe/aoecmd.c
104
f = list_entry(pos, struct frame, head);
drivers/block/aoe/aoecmd.c
1205
if (list_empty(&iocq[id].head))
drivers/block/aoe/aoecmd.c
1207
pos = iocq[id].head.next;
drivers/block/aoe/aoecmd.c
1209
f = list_entry(pos, struct frame, head);
drivers/block/aoe/aoecmd.c
1298
list_add_tail(&f->head, &iocq[id].head);
drivers/block/aoe/aoecmd.c
1667
list_splice_init(&iocq[id].head, &flist);
drivers/block/aoe/aoecmd.c
1672
f = list_entry(pos, struct frame, head);
drivers/block/aoe/aoecmd.c
1719
INIT_LIST_HEAD(&iocq[i].head);
drivers/block/aoe/aoecmd.c
201
list_add(&f->head, &t->ffree);
drivers/block/aoe/aoecmd.c
222
f = list_entry(pos, struct frame, head);
drivers/block/aoe/aoecmd.c
313
list_add_tail(&f->head, &d->factive[n]);
drivers/block/aoe/aoecmd.c
54
struct list_head head;
drivers/block/aoe/aoecmd.c
650
struct list_head *pos, *nx, *head;
drivers/block/aoe/aoecmd.c
656
head = &d->rexmitq;
drivers/block/aoe/aoecmd.c
657
list_for_each_safe(pos, nx, head) {
drivers/block/aoe/aoecmd.c
658
f = list_entry(pos, struct frame, head);
drivers/block/aoe/aoecmd.c
669
list_replace(&f->head, &nf->head);
drivers/block/aoe/aoecmd.c
670
pos = &nf->head;
drivers/block/aoe/aoecmd.c
740
struct list_head *head, *pos, *nx;
drivers/block/aoe/aoecmd.c
764
head = &d->factive[i];
drivers/block/aoe/aoecmd.c
765
list_for_each_safe(pos, nx, head) {
drivers/block/aoe/aoecmd.c
766
f = list_entry(pos, struct frame, head);
drivers/block/aoe/aoecmd.c
777
f = list_entry(pos, struct frame, head);
drivers/block/aoe/aoecmd.c
80
struct list_head *head, *pos, *nx;
drivers/block/aoe/aoecmd.c
83
head = &d->rexmitq;
drivers/block/aoe/aoecmd.c
84
list_for_each_safe(pos, nx, head) {
drivers/block/aoe/aoecmd.c
85
f = list_entry(pos, struct frame, head);
drivers/block/aoe/aoecmd.c
98
struct list_head *head, *pos, *nx;
drivers/block/aoe/aoedev.c
187
f = list_entry(pos, struct frame, head);
drivers/block/aoe/aoedev.c
200
struct list_head *head, *pos, *nx;
drivers/block/aoe/aoedev.c
211
head = &d->factive[i];
drivers/block/aoe/aoedev.c
212
list_for_each_safe(pos, nx, head)
drivers/block/aoe/aoedev.c
215
head = &d->rexmitq;
drivers/block/aoe/aoedev.c
216
list_for_each_safe(pos, nx, head)
drivers/block/aoe/aoedev.c
513
struct list_head *pos, *nx, *head;
drivers/block/aoe/aoedev.c
522
head = &t->ffree;
drivers/block/aoe/aoedev.c
523
list_for_each_safe(pos, nx, head) {
drivers/block/aoe/aoedev.c
525
f = list_entry(pos, struct frame, head);
drivers/block/ataflop.c
1621
getprm.head = 2;
drivers/block/ataflop.c
164
int head; /* "" "" */
drivers/block/ataflop.c
1728
setprm.head != 2) {
drivers/block/ataflop.c
775
if (!UDT || desc->track >= UDT->blocks/UDT->spt/2 || desc->head >= 2) {
drivers/block/ataflop.c
795
*p++ = desc->head;
drivers/block/ataflop.c
812
ReqSide = desc->head;
drivers/block/drbd/drbd_nl.c
4929
LIST_HEAD(head);
drivers/block/drbd/drbd_nl.c
4939
list_add(&head, &state_change->list);
drivers/block/drbd/drbd_nl.c
4940
free_state_changes(&head);
drivers/block/drbd/drbd_nl.c
4952
if (!list_empty(&head))
drivers/block/drbd/drbd_nl.c
4953
free_state_changes(&head);
drivers/block/drbd/drbd_nl.c
4958
list_add_tail(&state_change->list, &head);
drivers/block/drbd/drbd_nl.c
4963
if (!list_empty(&head)) {
drivers/block/drbd/drbd_nl.c
4965
list_entry(head.next, struct drbd_state_change, list);
drivers/block/drbd/drbd_nl.c
4968
list_del(&head); /* detach list from head */
drivers/block/drbd/drbd_nla.c
21
nla_for_each_attr(nla, head, len, rem) {
drivers/block/drbd/drbd_nla.c
9
struct nlattr *head = nla_data(nla);
drivers/block/drbd/drbd_receiver.c
289
struct list_head *head)
drivers/block/drbd/drbd_receiver.c
295
while (!list_empty(head)) {
drivers/block/drbd/drbd_receiver.c
305
struct list_head *head)
drivers/block/drbd/drbd_receiver.c
308
_drbd_wait_ee_list_empty(device, head);
drivers/block/floppy.c
2147
unsigned char track, head, sect, size;
drivers/block/floppy.c
2158
raw_cmd->cmd[DR_SELECT] = UNIT(current_drive) + PH_HEAD(_floppy, format_req.head);
drivers/block/floppy.c
2177
n = (track_shift * format_req.track + head_shift * format_req.head)
drivers/block/floppy.c
2188
here[count].head = format_req.head;
drivers/block/floppy.c
2237
tmp_format_req->head >= _floppy->head ||
drivers/block/floppy.c
2566
max_sector = _floppy->sect * _floppy->head;
drivers/block/floppy.c
305
#define PH_HEAD(floppy, head) (((((floppy)->stretch & 2) >> 1) ^ head) << 2)
drivers/block/floppy.c
3229
(int)g->head <= 0 ||
drivers/block/floppy.c
3231
(int)(g->sect * g->head) <= 0 ||
drivers/block/floppy.c
3372
geo->heads = g->head;
drivers/block/mtip32xx/mtip32xx.h
216
unsigned char head;
drivers/block/swim.c
357
static inline void swim_head(struct swim __iomem *base, enum head head)
drivers/block/swim.c
361
if (head == UPPER_HEAD)
drivers/block/swim.c
363
else if (head == LOWER_HEAD)
drivers/block/swim.c
603
fs->secpercyl = g->head * g->sect;
drivers/block/swim.c
724
geo->heads = g->head;
drivers/block/swim3.c
185
int head; /* head number ditto */
drivers/block/swim3.c
349
fs->head = x / fs->secpertrack;
drivers/block/swim3.c
452
fs->req_sector, fs->secpertrack, fs->head, n);
drivers/block/swim3.c
455
swim3_select(fs, fs->head? READ_DATA_1: READ_DATA_0);
drivers/block/swim3.c
786
if (++fs->head > 1) {
drivers/block/swim3.c
787
fs->head = 0;
drivers/block/xen-blkfront.c
2099
merge_bio.head = shadow[j].request->bio;
drivers/bluetooth/hci_bcm4377.c
658
u16 head;
drivers/bluetooth/hci_bcm4377.c
695
head = le16_to_cpu(bcm4377->ring_state->xfer_ring_head[ring->ring_id]);
drivers/bluetooth/hci_bcm4377.c
696
head = (head + 1) % ring->n_entries;
drivers/bluetooth/hci_bcm4377.c
697
bcm4377->ring_state->xfer_ring_head[ring->ring_id] = cpu_to_le16(head);
drivers/bluetooth/hci_bcm4377.c
699
bcm4377_ring_doorbell(bcm4377, ring->doorbell, head);
drivers/bluetooth/hci_bcm4377.c
877
u16 head, tail, new_head;
drivers/bluetooth/hci_bcm4377.c
897
head = le16_to_cpu(bcm4377->ring_state->xfer_ring_head[ring->ring_id]);
drivers/bluetooth/hci_bcm4377.c
900
new_head = (head + 1) % ring->n_entries;
drivers/bluetooth/hci_bcm4377.c
921
offset = head * (sizeof(*entry) + ring->payload_size);
drivers/bus/mhi/ep/main.c
809
LIST_HEAD(head);
drivers/bus/mhi/ep/main.c
813
list_splice_tail_init(&mhi_cntrl->ch_db_list, &head);
drivers/bus/mhi/ep/main.c
817
list_for_each_entry_safe(itr, tmp, &head, node) {
drivers/bus/mhi/ep/main.c
871
LIST_HEAD(head);
drivers/bus/mhi/ep/main.c
875
list_splice_tail_init(&mhi_cntrl->st_transition_list, &head);
drivers/bus/mhi/ep/main.c
878
list_for_each_entry_safe(itr, tmp, &head, node) {
drivers/bus/mhi/ep/main.c
908
LIST_HEAD(head);
drivers/bus/mhi/ep/main.c
922
list_add_tail(&item->node, &head);
drivers/bus/mhi/ep/main.c
928
list_splice_tail_init(&head, &mhi_cntrl->ch_db_list);
drivers/bus/mhi/host/pm.c
801
LIST_HEAD(head);
drivers/bus/mhi/host/pm.c
807
list_splice_tail_init(&mhi_cntrl->transition_list, &head);
drivers/bus/mhi/host/pm.c
810
list_for_each_entry_safe(itr, tmp, &head, node) {
drivers/char/agp/isoch.c
136
list_for_each(pos, head) {
drivers/char/agp/isoch.c
23
static void agp_3_5_dev_list_insert(struct list_head *head, struct list_head *new)
drivers/char/agp/isoch.c
28
list_for_each(pos, head) {
drivers/char/agp/isoch.c
287
struct list_head *head = &dev_list->list, *pos;
drivers/char/agp/isoch.c
299
for (pos=head->next; cdev<ndevs; cdev++, pos=pos->next) {
drivers/char/agp/isoch.c
322
struct list_head *head, *pos;
drivers/char/agp/isoch.c
340
head = &dev_list->list;
drivers/char/agp/isoch.c
341
INIT_LIST_HEAD(head);
drivers/char/agp/isoch.c
372
list_add(pos, head);
drivers/char/agp/isoch.c
387
list_for_each(pos, head) {
drivers/char/agp/isoch.c
40
struct list_head *pos, *tmp, *head = &list->list, *start = head->next;
drivers/char/agp/isoch.c
43
INIT_LIST_HEAD(head);
drivers/char/agp/isoch.c
45
for (pos=start; pos!=head; ) {
drivers/char/agp/isoch.c
455
for (pos=head->next; pos!=head; ) {
drivers/char/agp/isoch.c
54
agp_3_5_dev_list_insert(head, tmp);
drivers/char/agp/isoch.c
81
struct list_head *head = &dev_list->list, *pos;
drivers/char/hw_random/cctrng.c
204
return CIRC_SPACE(drvdata->circ.head,
drivers/char/hw_random/cctrng.c
228
cnt_w = CIRC_CNT_TO_END(drvdata->circ.head,
drivers/char/hw_random/cctrng.c
237
cnt_w = CIRC_CNT(drvdata->circ.head,
drivers/char/hw_random/cctrng.c
359
buf[drvdata->circ.head] = cc_ioread(drvdata,
drivers/char/hw_random/cctrng.c
365
if (buf[drvdata->circ.head] == 0) {
drivers/char/hw_random/cctrng.c
371
circ_idx_inc(&drvdata->circ.head, 1<<2);
drivers/char/random.c
165
ret = raw_notifier_chain_register((struct raw_notifier_head *)&random_ready_notifier.head, nb);
drivers/char/tpm/tpm-buf.c
37
struct tpm_header *head = (struct tpm_header *)buf->data;
drivers/char/tpm/tpm-buf.c
43
buf->length = sizeof(*head);
drivers/char/tpm/tpm-buf.c
44
head->tag = cpu_to_be16(tag);
drivers/char/tpm/tpm-buf.c
45
head->length = cpu_to_be32(sizeof(*head));
drivers/char/tpm/tpm-buf.c
46
head->ordinal = cpu_to_be32(ordinal);
drivers/char/tpm/tpm2-cmd.c
260
struct tpm_header *head;
drivers/char/tpm/tpm2-cmd.c
292
head = (struct tpm_header *)buf.data;
drivers/char/tpm/tpm2-cmd.c
294
head->tag = cpu_to_be16(TPM2_ST_NO_SESSIONS);
drivers/char/tpm/tpm2-cmd.c
314
head = (struct tpm_header *)buf.data;
drivers/char/tpm/tpm2-cmd.c
317
if (be16_to_cpu(head->tag) == TPM2_ST_SESSIONS)
drivers/char/tpm/tpm2-sessions.c
1082
struct tpm_header *head = (struct tpm_header *)buf->data;
drivers/char/tpm/tpm2-sessions.c
1085
u32 total_len = be32_to_cpu(head->length);
drivers/char/tpm/tpm2-sessions.c
587
struct tpm_header *head = (struct tpm_header *)buf->data;
drivers/char/tpm/tpm2-sessions.c
602
auth->ordinal = head->ordinal;
drivers/char/tpm/tpm2-sessions.c
604
cc = be32_to_cpu(head->ordinal);
drivers/char/tpm/tpm2-sessions.c
690
sha256_update(&sctx, (u8 *)&head->ordinal, sizeof(head->ordinal));
drivers/char/tpm/tpm2-sessions.c
761
struct tpm_header *head = (struct tpm_header *)buf->data;
drivers/char/tpm/tpm2-sessions.c
768
u16 tag = be16_to_cpu(head->tag);
drivers/char/tpm/tpm2-sessions.c
829
sha256_update(&sctx, (u8 *)&head->return_code,
drivers/char/tpm/tpm2-sessions.c
830
sizeof(head->return_code));
drivers/char/tpm/tpm2-sessions.c
910
struct tpm_header *head = (struct tpm_header *)buf->data;
drivers/char/tpm/tpm2-sessions.c
911
u32 tot_len = be32_to_cpu(head->length);
drivers/char/xillybus/xillybus_core.c
1303
unsigned char *head =
drivers/char/xillybus/xillybus_core.c
1316
*head++ = channel->rd_leftovers[i];
drivers/clk/clk.c
4865
if (!cn->notifier_head.head) {
drivers/cpufreq/cpufreq.c
448
for (nb = cpufreq_transition_notifier_list.head; nb; nb = nb->next)
drivers/crypto/amcc/crypto4xx_core.c
1035
u32 head = core_dev->dev->pdr_head;
drivers/crypto/amcc/crypto4xx_core.c
1050
} while (head != tail);
drivers/crypto/caam/intern.h
78
int head; /* entinfo (s/w ring) head index */
drivers/crypto/caam/jr.c
261
int hw_idx, sw_idx, i, head, tail;
drivers/crypto/caam/jr.c
273
head = READ_ONCE(jrp->head);
drivers/crypto/caam/jr.c
278
for (i = 0; CIRC_CNT(head, tail + i, JOBR_DEPTH) >= 1; i++) {
drivers/crypto/caam/jr.c
286
BUG_ON(CIRC_CNT(head, tail + i, JOBR_DEPTH) <= 0);
drivers/crypto/caam/jr.c
326
} while (CIRC_CNT(head, tail, JOBR_DEPTH) >= 1 &&
drivers/crypto/caam/jr.c
429
int head, tail, desc_size;
drivers/crypto/caam/jr.c
441
head = jrp->head;
drivers/crypto/caam/jr.c
445
CIRC_SPACE(head, tail, JOBR_DEPTH) <= 0) {
drivers/crypto/caam/jr.c
451
head_entry = &jrp->entinfo[head];
drivers/crypto/caam/jr.c
458
jr_inpentry_set(jrp->inpring, head, cpu_to_caam_dma(desc_dma));
drivers/crypto/caam/jr.c
475
jrp->head = (head + 1) & (JOBR_DEPTH - 1);
drivers/crypto/caam/jr.c
516
jrp->head = 0;
drivers/crypto/cavium/cpt/cptvf.h
45
u8 *head;
drivers/crypto/cavium/cpt/cptvf.h
77
struct pending_entry *head; /* head of the queue */
drivers/crypto/cavium/cpt/cptvf_main.c
114
queue->head = kzalloc_objs(*queue->head, qlen);
drivers/crypto/cavium/cpt/cptvf_main.c
115
if (!queue->head) {
drivers/crypto/cavium/cpt/cptvf_main.c
185
chunk->head,
drivers/crypto/cavium/cpt/cptvf_main.c
187
chunk->head = NULL;
drivers/crypto/cavium/cpt/cptvf_main.c
234
curr->head = dma_alloc_coherent(&pdev->dev,
drivers/crypto/cavium/cpt/cptvf_main.c
238
if (!curr->head) {
drivers/crypto/cavium/cpt/cptvf_main.c
258
*((u64 *)(&last->head[last->size])) = (u64)curr->dma_addr;
drivers/crypto/cavium/cpt/cptvf_main.c
266
*((u64 *)(&last->head[last->size])) = (u64)curr->dma_addr;
drivers/crypto/cavium/cpt/cptvf_main.c
87
if (!queue->head)
drivers/crypto/cavium/cpt/cptvf_main.c
91
kfree_sensitive((queue->head));
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
20
ent = &q->head[q->rear];
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
243
ent = &queue->qhead->head[queue->idx * qinfo->cmd_size];
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
337
pentry = &pqueue->head[pqueue->front];
drivers/crypto/ccp/ccp-crypto-aes-cmac.c
349
int ccp_register_aes_cmac_algs(struct list_head *head)
drivers/crypto/ccp/ccp-crypto-aes-cmac.c
399
list_add(&ccp_alg->entry, head);
drivers/crypto/ccp/ccp-crypto-aes-galois.c
208
static int ccp_register_aes_aead(struct list_head *head,
drivers/crypto/ccp/ccp-crypto-aes-galois.c
239
list_add(&ccp_aead->entry, head);
drivers/crypto/ccp/ccp-crypto-aes-galois.c
244
int ccp_register_aes_aeads(struct list_head *head)
drivers/crypto/ccp/ccp-crypto-aes-galois.c
252
ret = ccp_register_aes_aead(head, &aes_aead_algs[i]);
drivers/crypto/ccp/ccp-crypto-aes-xts.c
227
static int ccp_register_aes_xts_alg(struct list_head *head,
drivers/crypto/ccp/ccp-crypto-aes-xts.c
272
list_add(&ccp_alg->entry, head);
drivers/crypto/ccp/ccp-crypto-aes-xts.c
277
int ccp_register_aes_xts_algs(struct list_head *head)
drivers/crypto/ccp/ccp-crypto-aes-xts.c
282
ret = ccp_register_aes_xts_alg(head, &aes_xts_algs[i]);
drivers/crypto/ccp/ccp-crypto-aes.c
290
static int ccp_register_aes_alg(struct list_head *head,
drivers/crypto/ccp/ccp-crypto-aes.c
322
list_add(&ccp_alg->entry, head);
drivers/crypto/ccp/ccp-crypto-aes.c
327
int ccp_register_aes_algs(struct list_head *head)
drivers/crypto/ccp/ccp-crypto-aes.c
335
ret = ccp_register_aes_alg(head, &aes_algs[i]);
drivers/crypto/ccp/ccp-crypto-des3.c
178
static int ccp_register_des3_alg(struct list_head *head,
drivers/crypto/ccp/ccp-crypto-des3.c
210
list_add(&ccp_alg->entry, head);
drivers/crypto/ccp/ccp-crypto-des3.c
215
int ccp_register_des3_algs(struct list_head *head)
drivers/crypto/ccp/ccp-crypto-des3.c
223
ret = ccp_register_des3_alg(head, &des3_algs[i]);
drivers/crypto/ccp/ccp-crypto-rsa.c
245
static int ccp_register_rsa_alg(struct list_head *head,
drivers/crypto/ccp/ccp-crypto-rsa.c
271
list_add(&ccp_alg->entry, head);
drivers/crypto/ccp/ccp-crypto-rsa.c
276
int ccp_register_rsa_algs(struct list_head *head)
drivers/crypto/ccp/ccp-crypto-rsa.c
287
ret = ccp_register_rsa_alg(head, &rsa_algs[i]);
drivers/crypto/ccp/ccp-crypto-sha.c
411
static int ccp_register_hmac_alg(struct list_head *head,
drivers/crypto/ccp/ccp-crypto-sha.c
451
list_add(&ccp_alg->entry, head);
drivers/crypto/ccp/ccp-crypto-sha.c
456
static int ccp_register_sha_alg(struct list_head *head,
drivers/crypto/ccp/ccp-crypto-sha.c
509
list_add(&ccp_alg->entry, head);
drivers/crypto/ccp/ccp-crypto-sha.c
511
ret = ccp_register_hmac_alg(head, def, ccp_alg);
drivers/crypto/ccp/ccp-crypto-sha.c
516
int ccp_register_sha_algs(struct list_head *head)
drivers/crypto/ccp/ccp-crypto-sha.c
524
ret = ccp_register_sha_alg(head, &sha_algs[i]);
drivers/crypto/ccp/ccp-crypto.h
278
int ccp_register_aes_algs(struct list_head *head);
drivers/crypto/ccp/ccp-crypto.h
279
int ccp_register_aes_cmac_algs(struct list_head *head);
drivers/crypto/ccp/ccp-crypto.h
280
int ccp_register_aes_xts_algs(struct list_head *head);
drivers/crypto/ccp/ccp-crypto.h
281
int ccp_register_aes_aeads(struct list_head *head);
drivers/crypto/ccp/ccp-crypto.h
282
int ccp_register_sha_algs(struct list_head *head);
drivers/crypto/ccp/ccp-crypto.h
283
int ccp_register_des3_algs(struct list_head *head);
drivers/crypto/ccp/ccp-crypto.h
284
int ccp_register_rsa_algs(struct list_head *head);
drivers/crypto/ccree/cc_request_mgr.c
562
unsigned int *head = &request_mgr_handle->req_queue_head;
drivers/crypto/ccree/cc_request_mgr.c
570
if (*head == *tail) {
drivers/crypto/ccree/cc_request_mgr.c
576
*head);
drivers/crypto/hisilicon/qm.c
3667
static void qm_insert_sorted(struct list_head *head, struct hisi_qm_resource *res)
drivers/crypto/hisilicon/qm.c
3670
struct list_head *n = head;
drivers/crypto/hisilicon/qm.c
3672
list_for_each_entry(tmp, head, list) {
drivers/crypto/hisilicon/qm.c
3681
static void free_list(struct list_head *head)
drivers/crypto/hisilicon/qm.c
3685
list_for_each_entry_safe(res, tmp, head, list) {
drivers/crypto/hisilicon/qm.c
3733
static int hisi_qm_sort_devices(int node, struct list_head *head,
drivers/crypto/hisilicon/qm.c
3763
list_splice_tail(&non_full_list, head);
drivers/crypto/hisilicon/qm.c
3764
list_splice_tail(&full_list, head);
drivers/crypto/hisilicon/qm.c
3786
LIST_HEAD(head);
drivers/crypto/hisilicon/qm.c
3792
if (hisi_qm_sort_devices(node, &head, qm_list)) {
drivers/crypto/hisilicon/qm.c
3797
list_for_each_entry(tmp, &head, list) {
drivers/crypto/hisilicon/qm.c
3809
free_list(&head);
drivers/crypto/hisilicon/qm_common.h
27
__le16 head;
drivers/crypto/hisilicon/qm_common.h
42
__le16 head;
drivers/crypto/hisilicon/qm_common.h
56
__le16 head;
drivers/crypto/hisilicon/qm_common.h
66
__le16 head;
drivers/crypto/hisilicon/sec/sec_algs.c
390
list_for_each_entry_safe(el, temp, &sec_req->elements, head) {
drivers/crypto/hisilicon/sec/sec_algs.c
437
head);
drivers/crypto/hisilicon/sec/sec_algs.c
515
list_del(&sec_req_el->head);
drivers/crypto/hisilicon/sec/sec_algs.c
788
list_add_tail(&el->head, &sec_req->elements);
drivers/crypto/hisilicon/sec/sec_algs.c
840
list_for_each_entry_safe(el, temp, &sec_req->elements, head) {
drivers/crypto/hisilicon/sec/sec_algs.c
841
list_del(&el->head);
drivers/crypto/hisilicon/sec/sec_drv.h
303
struct list_head head;
drivers/crypto/intel/keembay/ocs-hcu.c
123
struct ocs_hcu_dma_entry *head;
drivers/crypto/intel/keembay/ocs-hcu.c
499
dma_list->head = dma_alloc_coherent(hcu_dev->dev,
drivers/crypto/intel/keembay/ocs-hcu.c
500
sizeof(*dma_list->head) * max_nents,
drivers/crypto/intel/keembay/ocs-hcu.c
502
if (!dma_list->head) {
drivers/crypto/intel/keembay/ocs-hcu.c
519
sizeof(*dma_list->head) * dma_list->max_nents,
drivers/crypto/intel/keembay/ocs-hcu.c
520
dma_list->head, dma_list->dma_addr);
drivers/crypto/intel/keembay/ocs-hcu.c
547
new_tail = old_tail ? old_tail + 1 : dma_list->head;
drivers/crypto/intel/keembay/ocs-hcu.c
550
if (new_tail - dma_list->head >= dma_list->max_nents)
drivers/crypto/intel/keembay/ocs-hcu.c
565
dma_list->head);
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
55
state->rings[i].head = ops->read_csr_ring_head(base, bank, i);
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
79
ops->write_csr_ring_head(base, bank, tx, state->rings[tx].head);
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
91
ops->write_csr_ring_head(base, bank, tx, state->rings[tx].head);
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
99
ops->write_csr_ring_head(base, bank, rx, state->rings[rx].head);
drivers/crypto/intel/qat/qat_common/adf_bank_state.h
13
u32 head;
drivers/crypto/intel/qat/qat_common/adf_cfg.c
102
static void adf_cfg_section_del_all(struct list_head *head);
drivers/crypto/intel/qat/qat_common/adf_cfg.c
103
static void adf_cfg_section_del_all_except(struct list_head *head,
drivers/crypto/intel/qat/qat_common/adf_cfg.c
160
struct list_head *head = &sec->param_head;
drivers/crypto/intel/qat/qat_common/adf_cfg.c
163
list_for_each_prev_safe(list_ptr, tmp, head) {
drivers/crypto/intel/qat/qat_common/adf_cfg.c
176
static void adf_cfg_keyval_del_all(struct list_head *head)
drivers/crypto/intel/qat/qat_common/adf_cfg.c
180
list_for_each_prev_safe(list_ptr, tmp, head) {
drivers/crypto/intel/qat/qat_common/adf_cfg.c
188
static void adf_cfg_section_del_all(struct list_head *head)
drivers/crypto/intel/qat/qat_common/adf_cfg.c
193
list_for_each_prev_safe(list, tmp, head) {
drivers/crypto/intel/qat/qat_common/adf_cfg.c
201
static void adf_cfg_section_del_all_except(struct list_head *head,
drivers/crypto/intel/qat/qat_common/adf_cfg.c
207
list_for_each_prev_safe(list, tmp, head) {
drivers/crypto/intel/qat/qat_common/adf_transport.c
116
u32 *msg = (u32 *)((uintptr_t)ring->base_addr + ring->head);
drivers/crypto/intel/qat/qat_common/adf_transport.c
122
ring->head = adf_modulo(ring->head +
drivers/crypto/intel/qat/qat_common/adf_transport.c
126
msg = (u32 *)((uintptr_t)ring->base_addr + ring->head);
drivers/crypto/intel/qat/qat_common/adf_transport.c
131
ring->ring_number, ring->head);
drivers/crypto/intel/qat/qat_common/adf_transport.c
273
ring->head = 0;
drivers/crypto/intel/qat/qat_common/adf_transport_debug.c
164
int head, tail, empty;
drivers/crypto/intel/qat/qat_common/adf_transport_debug.c
169
head = csr_ops->read_csr_ring_head(csr, bank->bank_number,
drivers/crypto/intel/qat/qat_common/adf_transport_debug.c
177
ring->ring_number, head, tail,
drivers/crypto/intel/qat/qat_common/adf_transport_debug.c
56
int head, tail, empty;
drivers/crypto/intel/qat/qat_common/adf_transport_debug.c
58
head = csr_ops->read_csr_ring_head(csr, bank->bank_number,
drivers/crypto/intel/qat/qat_common/adf_transport_debug.c
70
head, tail, (empty & 1 << ring->ring_number)
drivers/crypto/intel/qat/qat_common/adf_transport_internal.h
23
u16 head;
drivers/crypto/marvell/octeontx/otx_cptvf.h
29
u8 *head;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
103
queue->head = kzalloc_objs(*queue->head, qlen);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
104
if (!queue->head) {
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
173
chunk->head,
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
175
chunk->head = NULL;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
221
curr->head = dma_alloc_coherent(&pdev->dev,
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
224
if (!curr->head) {
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
242
*((u64 *)(&last->head[last->size])) =
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
252
*((u64 *)(&last->head[last->size])) = (u64)curr->dma_addr;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
81
if (!queue->head)
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
85
kfree_sensitive((queue->head));
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
280
ent = &queue->qhead->head[queue->idx * OTX_CPT_INST_SIZE];
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
518
pentry = &pqueue->head[pqueue->front];
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
569
resume_pentry = &pqueue->head[resume_index];
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
63
ent = &q->head[q->rear];
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.h
149
struct otx_cpt_pending_entry *head; /* Head of the queue */
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
97
struct otx2_cpt_pending_entry *head; /* Head of the queue */
drivers/crypto/marvell/octeontx2/otx2_cptvf_main.c
176
kfree(lfs->lf[i].pqueue.head);
drivers/crypto/marvell/octeontx2/otx2_cptvf_main.c
177
lfs->lf[i].pqueue.head = NULL;
drivers/crypto/marvell/octeontx2/otx2_cptvf_main.c
193
lfs->lf[i].pqueue.head = kzalloc(size, GFP_KERNEL);
drivers/crypto/marvell/octeontx2/otx2_cptvf_main.c
194
if (!lfs->lf[i].pqueue.head) {
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
301
pentry = &pqueue->head[pqueue->front];
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
351
resume_pentry = &pqueue->head[resume_index];
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
51
ent = &q->head[q->rear];
drivers/crypto/talitos.c
278
int head;
drivers/crypto/talitos.c
289
head = priv->chan[ch].head;
drivers/crypto/talitos.c
290
request = &priv->chan[ch].fifo[head];
drivers/crypto/talitos.c
307
priv->chan[ch].head = (priv->chan[ch].head + 1) & (priv->fifo_len - 1);
drivers/crypto/talitos.h
101
int head;
drivers/devfreq/hisi_uncore_freq.c
214
shmem.head = (struct acpi_pcct_shared_memory) {
drivers/devfreq/hisi_uncore_freq.c
229
rc = readw_poll_timeout(&addr->head.status, status,
drivers/devfreq/hisi_uncore_freq.c
37
struct acpi_pcct_shared_memory head;
drivers/dma-buf/dma-fence-array.c
296
struct dma_fence *dma_fence_array_first(struct dma_fence *head)
drivers/dma-buf/dma-fence-array.c
300
if (!head)
drivers/dma-buf/dma-fence-array.c
303
array = to_dma_fence_array(head);
drivers/dma-buf/dma-fence-array.c
305
return head;
drivers/dma-buf/dma-fence-array.c
314
struct dma_fence *dma_fence_array_next(struct dma_fence *head,
drivers/dma-buf/dma-fence-array.c
317
struct dma_fence_array *array = to_dma_fence_array(head);
drivers/dma-buf/dma-fence-chain.c
147
struct dma_fence_chain *head = to_dma_fence_chain(fence);
drivers/dma-buf/dma-fence-chain.c
149
dma_fence_get(&head->base);
drivers/dma-buf/dma-fence-chain.c
150
dma_fence_chain_for_each(fence, &head->base) {
drivers/dma-buf/dma-fence-chain.c
154
if (!dma_fence_add_callback(f, &head->cb, dma_fence_chain_cb)) {
drivers/dma-buf/dma-fence-chain.c
160
dma_fence_put(&head->base);
drivers/dma-buf/dma-fence-unwrap.c
34
struct dma_fence *dma_fence_unwrap_first(struct dma_fence *head,
drivers/dma-buf/dma-fence-unwrap.c
37
cursor->chain = dma_fence_get(head);
drivers/dma-buf/udmabuf.c
370
struct udmabuf_create_list *head,
drivers/dma-buf/udmabuf.c
386
for (i = 0; i < head->count; i++) {
drivers/dma-buf/udmabuf.c
415
for (i = 0; i < head->count; i++) {
drivers/dma-buf/udmabuf.c
441
flags = head->flags & UDMABUF_FLAGS_CLOEXEC ? O_CLOEXEC : 0;
drivers/dma-buf/udmabuf.c
471
struct udmabuf_create_list head;
drivers/dma-buf/udmabuf.c
478
head.flags = create.flags;
drivers/dma-buf/udmabuf.c
479
head.count = 1;
drivers/dma-buf/udmabuf.c
484
return udmabuf_create(filp->private_data, &head, &list);
drivers/dma-buf/udmabuf.c
489
struct udmabuf_create_list head;
drivers/dma-buf/udmabuf.c
494
if (copy_from_user(&head, (void __user *)arg, sizeof(head)))
drivers/dma-buf/udmabuf.c
496
if (head.count > list_limit)
drivers/dma-buf/udmabuf.c
498
lsize = sizeof(struct udmabuf_create_item) * head.count;
drivers/dma-buf/udmabuf.c
499
list = memdup_user((void __user *)(arg + sizeof(head)), lsize);
drivers/dma-buf/udmabuf.c
503
ret = udmabuf_create(filp->private_data, &head, list);
drivers/dma/amba-pl08x.c
1523
LIST_HEAD(head);
drivers/dma/amba-pl08x.c
1525
vchan_get_all_descriptors(&plchan->vc, &head);
drivers/dma/amba-pl08x.c
1526
vchan_dma_desc_free_list(&plchan->vc, &head);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
544
LIST_HEAD(head);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
553
vchan_get_all_descriptors(&chan->vc, &head);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
556
vchan_dma_desc_free_list(&chan->vc, &head);
drivers/dma/apple-admac.c
540
LIST_HEAD(head);
drivers/dma/apple-admac.c
543
list_splice_tail_init(&adchan->to_free, &head);
drivers/dma/apple-admac.c
548
list_for_each_entry_safe(adtx, _adtx, &head, node) {
drivers/dma/bcm2835-dma.c
775
LIST_HEAD(head);
drivers/dma/bcm2835-dma.c
786
vchan_get_all_descriptors(&c->vc, &head);
drivers/dma/bcm2835-dma.c
788
vchan_dma_desc_free_list(&c->vc, &head);
drivers/dma/dma-axi-dmac.c
494
LIST_HEAD(head);
drivers/dma/dma-axi-dmac.c
499
vchan_get_all_descriptors(&chan->vchan, &head);
drivers/dma/dma-axi-dmac.c
500
list_splice_tail_init(&chan->active_descs, &head);
drivers/dma/dma-axi-dmac.c
503
vchan_dma_desc_free_list(&chan->vchan, &head);
drivers/dma/dma-jz4780.c
579
LIST_HEAD(head);
drivers/dma/dma-jz4780.c
592
vchan_get_all_descriptors(&jzchan->vchan, &head);
drivers/dma/dma-jz4780.c
596
vchan_dma_desc_free_list(&jzchan->vchan, &head);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
1186
LIST_HEAD(head);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
1203
vchan_get_all_descriptors(&chan->vc, &head);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
1208
vchan_dma_desc_free_list(&chan->vc, &head);
drivers/dma/dw/core.c
315
struct list_head *head, *active = dwc->tx_node_active;
drivers/dma/dw/core.c
323
head = &desc->tx_list;
drivers/dma/dw/core.c
324
if (active != head) {
drivers/dma/dw/core.c
326
if (active == head->next)
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
582
struct list_head *head)
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
588
head, list) {
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
71
LIST_HEAD(head);
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
74
vchan_get_all_descriptors(&dpaa2_chan->vchan, &head);
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
77
vchan_dma_desc_free_list(&dpaa2_chan->vchan, &head);
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.h
153
struct list_head *head);
drivers/dma/fsl-edma-common.c
240
LIST_HEAD(head);
drivers/dma/fsl-edma-common.c
246
vchan_get_all_descriptors(&fsl_chan->vchan, &head);
drivers/dma/fsl-edma-common.c
248
vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
drivers/dma/fsl-edma-common.c
886
LIST_HEAD(head);
drivers/dma/fsl-edma-common.c
893
vchan_get_all_descriptors(&fsl_chan->vchan, &head);
drivers/dma/fsl-edma-common.c
902
vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
drivers/dma/fsl-qdma.c
1058
LIST_HEAD(head);
drivers/dma/fsl-qdma.c
1063
vchan_get_all_descriptors(&fsl_chan->vchan, &head);
drivers/dma/fsl-qdma.c
1065
vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
drivers/dma/fsl-qdma.c
319
LIST_HEAD(head);
drivers/dma/fsl-qdma.c
322
vchan_get_all_descriptors(&fsl_chan->vchan, &head);
drivers/dma/fsl-qdma.c
325
vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
drivers/dma/hisi_dma.c
556
LIST_HEAD(head);
drivers/dma/hisi_dma.c
566
vchan_get_all_descriptors(&chan->vc, &head);
drivers/dma/hisi_dma.c
570
vchan_dma_desc_free_list(&chan->vc, &head);
drivers/dma/hsu/hsu.c
398
LIST_HEAD(head);
drivers/dma/hsu/hsu.c
408
vchan_get_all_descriptors(&hsuc->vchan, &head);
drivers/dma/hsu/hsu.c
410
vchan_dma_desc_free_list(&hsuc->vchan, &head);
drivers/dma/idma64.c
482
LIST_HEAD(head);
drivers/dma/idma64.c
491
vchan_get_all_descriptors(&idma64c->vchan, &head);
drivers/dma/idma64.c
494
vchan_dma_desc_free_list(&idma64c->vchan, &head);
drivers/dma/idxd/cdev.c
339
h = status.head;
drivers/dma/idxd/debugfs.c
73
h = evl_status.head;
drivers/dma/idxd/debugfs.c
77
evl_status.head, evl_status.tail, evl_status.int_pending);
drivers/dma/idxd/device.c
1275
struct llist_node *head;
drivers/dma/idxd/device.c
1280
head = llist_del_all(&ie->pending_llist);
drivers/dma/idxd/device.c
1281
if (head) {
drivers/dma/idxd/device.c
1282
llist_for_each_entry_safe(desc, itr, head, llnode)
drivers/dma/idxd/irq.c
109
struct llist_node *head;
drivers/dma/idxd/irq.c
112
head = llist_del_all(&ie->pending_llist);
drivers/dma/idxd/irq.c
113
if (head) {
drivers/dma/idxd/irq.c
114
llist_for_each_entry_safe(d, t, head, llnode)
drivers/dma/idxd/irq.c
372
h = evl_status.head;
drivers/dma/idxd/irq.c
381
evl_status.head = h;
drivers/dma/idxd/irq.c
599
struct llist_node *head;
drivers/dma/idxd/irq.c
601
head = llist_del_all(&irq_entry->pending_llist);
drivers/dma/idxd/irq.c
602
if (!head)
drivers/dma/idxd/irq.c
605
llist_for_each_entry_safe(desc, t, head, llnode) {
drivers/dma/idxd/registers.h
578
u32 head:16;
drivers/dma/idxd/submit.c
101
struct llist_node *head;
drivers/dma/idxd/submit.c
110
head = llist_del_all(&ie->pending_llist);
drivers/dma/idxd/submit.c
111
if (head) {
drivers/dma/idxd/submit.c
112
llist_for_each_entry_safe(d, t, head, llnode) {
drivers/dma/img-mdc-dma.c
694
LIST_HEAD(head);
drivers/dma/img-mdc-dma.c
705
vchan_get_all_descriptors(&mchan->vc, &head);
drivers/dma/img-mdc-dma.c
711
vchan_dma_desc_free_list(&mchan->vc, &head);
drivers/dma/ioat/dma.c
174
ioat_chan->issued = ioat_chan->head;
drivers/dma/ioat/dma.c
179
__func__, ioat_chan->head, ioat_chan->tail,
drivers/dma/ioat/dma.c
220
__func__, ioat_chan->head, ioat_chan->tail, ioat_chan->issued);
drivers/dma/ioat/dma.c
221
desc = ioat_get_ring_ent(ioat_chan, ioat_chan->head);
drivers/dma/ioat/dma.c
237
ioat_chan->head += 1;
drivers/dma/ioat/dma.c
258
__func__, ioat_chan->head, ioat_chan->tail,
drivers/dma/ioat/dma.c
328
ioat_chan->head += ioat_chan->produce;
drivers/dma/ioat/dma.c
473
__func__, num_descs, ioat_chan->head,
drivers/dma/ioat/dma.c
482
__func__, num_descs, ioat_chan->head,
drivers/dma/ioat/dma.c
600
__func__, ioat_chan->head, ioat_chan->tail, ioat_chan->issued);
drivers/dma/ioat/dma.h
129
u16 head;
drivers/dma/ioat/dma.h
317
return CIRC_CNT(ioat_chan->head, ioat_chan->tail,
drivers/dma/ioat/dma.h
324
return CIRC_CNT(ioat_chan->head, ioat_chan->issued,
drivers/dma/ioat/init.c
628
desc = ioat_get_ring_ent(ioat_chan, ioat_chan->head + i);
drivers/dma/ioat/init.c
704
ioat_chan->head = 0;
drivers/dma/ioat/prep.c
118
idx = ioat_chan->head;
drivers/dma/ioat/prep.c
187
idx = ioat_chan->head;
drivers/dma/ioat/prep.c
383
idx = ioat_chan->head;
drivers/dma/ioat/prep.c
489
idx = ioat_chan->head;
drivers/dma/ioat/prep.c
715
desc = ioat_get_ring_ent(ioat_chan, ioat_chan->head);
drivers/dma/k3dma.c
720
LIST_HEAD(head);
drivers/dma/k3dma.c
731
vchan_get_all_descriptors(&c->vc, &head);
drivers/dma/k3dma.c
744
vchan_dma_desc_free_list(&c->vc, &head);
drivers/dma/lgm/lgm-dma.c
1060
LIST_HEAD(head);
drivers/dma/lgm/lgm-dma.c
1063
vchan_get_all_descriptors(&c->vchan, &head);
drivers/dma/lgm/lgm-dma.c
1065
vchan_dma_desc_free_list(&c->vchan, &head);
drivers/dma/lgm/lgm-dma.c
1314
LIST_HEAD(head);
drivers/dma/lgm/lgm-dma.c
1317
list_splice_tail_init(&vc->desc_completed, &head);
drivers/dma/lgm/lgm-dma.c
1323
list_for_each_entry_safe(vd, _vd, &head, node) {
drivers/dma/loongson1-apb-dma.c
401
LIST_HEAD(head);
drivers/dma/loongson1-apb-dma.c
410
vchan_get_all_descriptors(&chan->vc, &head);
drivers/dma/loongson1-apb-dma.c
413
vchan_dma_desc_free_list(&chan->vc, &head);
drivers/dma/loongson2-apb-dma.c
482
LIST_HEAD(head);
drivers/dma/loongson2-apb-dma.c
492
vchan_get_all_descriptors(&lchan->vchan, &head);
drivers/dma/loongson2-apb-dma.c
495
vchan_dma_desc_free_list(&lchan->vchan, &head);
drivers/dma/mediatek/mtk-cqdma.c
552
LIST_HEAD(head);
drivers/dma/mediatek/mtk-cqdma.c
559
list_splice_tail_init(&vc->desc_allocated, &head);
drivers/dma/mediatek/mtk-cqdma.c
560
list_splice_tail_init(&vc->desc_submitted, &head);
drivers/dma/mediatek/mtk-cqdma.c
561
list_splice_tail_init(&vc->desc_issued, &head);
drivers/dma/mediatek/mtk-cqdma.c
565
vchan_dma_desc_free_list(vc, &head);
drivers/dma/mediatek/mtk-hsdma.c
741
LIST_HEAD(head);
drivers/dma/mediatek/mtk-hsdma.c
744
list_splice_tail_init(&vc->desc_allocated, &head);
drivers/dma/mediatek/mtk-hsdma.c
745
list_splice_tail_init(&vc->desc_submitted, &head);
drivers/dma/mediatek/mtk-hsdma.c
746
list_splice_tail_init(&vc->desc_issued, &head);
drivers/dma/mediatek/mtk-hsdma.c
750
vchan_dma_desc_free_list(vc, &head);
drivers/dma/mediatek/mtk-uart-apdma.c
400
LIST_HEAD(head);
drivers/dma/mediatek/mtk-uart-apdma.c
435
vchan_get_all_descriptors(&c->vc, &head);
drivers/dma/mediatek/mtk-uart-apdma.c
438
vchan_dma_desc_free_list(&c->vc, &head);
drivers/dma/milbeaut-hdmac.c
294
LIST_HEAD(head);
drivers/dma/milbeaut-hdmac.c
307
vchan_get_all_descriptors(vc, &head);
drivers/dma/milbeaut-hdmac.c
311
vchan_dma_desc_free_list(vc, &head);
drivers/dma/milbeaut-xdmac.c
213
LIST_HEAD(head);
drivers/dma/milbeaut-xdmac.c
228
vchan_get_all_descriptors(vc, &head);
drivers/dma/milbeaut-xdmac.c
232
vchan_dma_desc_free_list(vc, &head);
drivers/dma/moxart-dma.c
182
LIST_HEAD(head);
drivers/dma/moxart-dma.c
198
vchan_get_all_descriptors(&ch->vc, &head);
drivers/dma/moxart-dma.c
200
vchan_dma_desc_free_list(&ch->vc, &head);
drivers/dma/nbpfaxi.c
696
LIST_HEAD(head);
drivers/dma/nbpfaxi.c
729
list_add_tail(&desc->node, &head);
drivers/dma/nbpfaxi.c
738
list_splice_tail(&head, &chan->free);
drivers/dma/nbpfaxi.c
773
LIST_HEAD(head);
drivers/dma/nbpfaxi.c
778
list_move(&desc->node, &head);
drivers/dma/nbpfaxi.c
783
list_for_each_entry_safe(desc, tmp, &head, node) {
drivers/dma/nbpfaxi.c
856
LIST_HEAD(head);
drivers/dma/nbpfaxi.c
860
list_splice_init(&chan->done, &head);
drivers/dma/nbpfaxi.c
861
list_splice_init(&chan->active, &head);
drivers/dma/nbpfaxi.c
862
list_splice_init(&chan->queued, &head);
drivers/dma/nbpfaxi.c
868
list_for_each_entry_safe(desc, tmp, &head, node) {
drivers/dma/owl-dma.c
702
LIST_HEAD(head);
drivers/dma/owl-dma.c
714
vchan_get_all_descriptors(&vchan->vc, &head);
drivers/dma/owl-dma.c
718
vchan_dma_desc_free_list(&vchan->vc, &head);
drivers/dma/plx_dma.c
112
int head;
drivers/dma/plx_dma.c
142
while (plxdev->tail != plxdev->head) {
drivers/dma/plx_dma.c
180
while (plxdev->tail != plxdev->head) {
drivers/dma/plx_dma.c
263
if (!CIRC_SPACE(plxdev->head, plxdev->tail, PLX_DMA_RING_COUNT))
drivers/dma/plx_dma.c
269
plxdesc = plx_dma_get_desc(plxdev, plxdev->head);
drivers/dma/plx_dma.c
270
plxdev->head++;
drivers/dma/plx_dma.c
412
plxdev->head = plxdev->tail = 0;
drivers/dma/pxa_dma.c
1081
LIST_HEAD(head);
drivers/dma/pxa_dma.c
1087
vchan_get_all_descriptors(&chan->vc, &head);
drivers/dma/pxa_dma.c
1089
list_for_each_entry(vd, &head, node) {
drivers/dma/pxa_dma.c
1105
vchan_dma_desc_free_list(&chan->vc, &head);
drivers/dma/qcom/bam_dma.c
1034
avail = CIRC_SPACE(bchan->tail, bchan->head,
drivers/dma/qcom/bam_dma.c
345
#define IS_BUSY(chan) (CIRC_SPACE(bchan->tail, bchan->head,\
drivers/dma/qcom/bam_dma.c
364
unsigned short head; /* start of active descriptor entries */
drivers/dma/qcom/bam_dma.c
529
bchan->head = 0;
drivers/dma/qcom/bam_dma.c
721
LIST_HEAD(head);
drivers/dma/qcom/bam_dma.c
750
vchan_get_all_descriptors(&bchan->vc, &head);
drivers/dma/qcom/bam_dma.c
753
vchan_dma_desc_free_list(&bchan->vc, &head);
drivers/dma/qcom/bam_dma.c
844
avail = CIRC_CNT(offset, bchan->head, MAX_DESCRIPTORS + 1);
drivers/dma/qcom/bam_dma.c
846
if (offset < bchan->head)
drivers/dma/qcom/bam_dma.c
856
bchan->head += async_desc->xfer_len;
drivers/dma/qcom/bam_dma.c
857
bchan->head %= MAX_DESCRIPTORS;
drivers/dma/qcom/qcom_adm.c
478
LIST_HEAD(head);
drivers/dma/qcom/qcom_adm.c
481
vchan_get_all_descriptors(&achan->vc, &head);
drivers/dma/qcom/qcom_adm.c
489
vchan_dma_desc_free_list(&achan->vc, &head);
drivers/dma/sa11x0-dma.c
755
LIST_HEAD(head);
drivers/dma/sa11x0-dma.c
761
vchan_get_all_descriptors(&c->vc, &head);
drivers/dma/sa11x0-dma.c
774
list_add_tail(&p->txd_load->vd.node, &head);
drivers/dma/sa11x0-dma.c
778
list_add_tail(&p->txd_done->vd.node, &head);
drivers/dma/sa11x0-dma.c
788
vchan_dma_desc_free_list(&c->vc, &head);
drivers/dma/sf-pdma/sf-pdma.c
144
LIST_HEAD(head);
drivers/dma/sf-pdma/sf-pdma.c
150
vchan_get_all_descriptors(&chan->vchan, &head);
drivers/dma/sf-pdma/sf-pdma.c
153
vchan_dma_desc_free_list(&chan->vchan, &head);
drivers/dma/sf-pdma/sf-pdma.c
214
LIST_HEAD(head);
drivers/dma/sf-pdma/sf-pdma.c
221
vchan_get_all_descriptors(&chan->vchan, &head);
drivers/dma/sf-pdma/sf-pdma.c
223
vchan_dma_desc_free_list(&chan->vchan, &head);
drivers/dma/sh/rz-dmac.c
233
channel->lmdesc.head = lmdesc;
drivers/dma/sh/rz-dmac.c
254
struct rz_lmdesc *lmdesc = channel->lmdesc.head;
drivers/dma/sh/rz-dmac.c
262
channel->lmdesc.head = lmdesc;
drivers/dma/sh/rz-dmac.c
281
(sizeof(struct rz_lmdesc) * (channel->lmdesc.head -
drivers/dma/sh/rz-dmac.c
566
LIST_HEAD(head);
drivers/dma/sh/rz-dmac.c
575
vchan_get_all_descriptors(&channel->vc, &head);
drivers/dma/sh/rz-dmac.c
577
vchan_dma_desc_free_list(&channel->vc, &head);
drivers/dma/sh/rz-dmac.c
85
struct rz_lmdesc *head;
drivers/dma/sh/usb-dmac.c
450
LIST_HEAD(head);
drivers/dma/sh/usb-dmac.c
455
vchan_get_all_descriptors(&uchan->vc, &head);
drivers/dma/sh/usb-dmac.c
462
vchan_dma_desc_free_list(&uchan->vc, &head);
drivers/dma/sprd-dma.c
1076
LIST_HEAD(head);
drivers/dma/sprd-dma.c
1084
vchan_get_all_descriptors(&schan->vc, &head);
drivers/dma/sprd-dma.c
1090
vchan_dma_desc_free_list(&schan->vc, &head);
drivers/dma/st_fdma.c
661
LIST_HEAD(head);
drivers/dma/st_fdma.c
671
vchan_get_all_descriptors(&fchan->vchan, &head);
drivers/dma/st_fdma.c
673
vchan_dma_desc_free_list(&fchan->vchan, &head);
drivers/dma/stm32/stm32-dma.c
503
LIST_HEAD(head);
drivers/dma/stm32/stm32-dma.c
515
vchan_get_all_descriptors(&chan->vchan, &head);
drivers/dma/stm32/stm32-dma.c
517
vchan_dma_desc_free_list(&chan->vchan, &head);
drivers/dma/stm32/stm32-dma3.c
1527
LIST_HEAD(head);
drivers/dma/stm32/stm32-dma3.c
1538
vchan_get_all_descriptors(&chan->vchan, &head);
drivers/dma/stm32/stm32-dma3.c
1541
vchan_dma_desc_free_list(&chan->vchan, &head);
drivers/dma/stm32/stm32-mdma.c
1273
LIST_HEAD(head);
drivers/dma/stm32/stm32-mdma.c
1282
vchan_get_all_descriptors(&chan->vchan, &head);
drivers/dma/stm32/stm32-mdma.c
1285
vchan_dma_desc_free_list(&chan->vchan, &head);
drivers/dma/sun4i-dma.c
1005
vchan_dma_desc_free_list(&vchan->vc, &head);
drivers/dma/sun4i-dma.c
979
LIST_HEAD(head);
drivers/dma/sun4i-dma.c
983
vchan_get_all_descriptors(&vchan->vc, &head);
drivers/dma/sun6i-dma.c
940
LIST_HEAD(head);
drivers/dma/sun6i-dma.c
958
vchan_get_all_descriptors(&vchan->vc, &head);
drivers/dma/sun6i-dma.c
972
vchan_dma_desc_free_list(&vchan->vc, &head);
drivers/dma/tegra186-gpc-dma.c
708
LIST_HEAD(head);
drivers/dma/tegra186-gpc-dma.c
727
vchan_get_all_descriptors(&tdc->vc, &head);
drivers/dma/tegra186-gpc-dma.c
730
vchan_dma_desc_free_list(&tdc->vc, &head);
drivers/dma/tegra210-adma.c
594
LIST_HEAD(head);
drivers/dma/tegra210-adma.c
602
vchan_get_all_descriptors(&tdc->vc, &head);
drivers/dma/tegra210-adma.c
604
vchan_dma_desc_free_list(&tdc->vc, &head);
drivers/dma/ti/edma.c
815
LIST_HEAD(head);
drivers/dma/ti/edma.c
834
vchan_get_all_descriptors(&echan->vchan, &head);
drivers/dma/ti/edma.c
836
vchan_dma_desc_free_list(&echan->vchan, &head);
drivers/dma/ti/k3-udma.c
3955
LIST_HEAD(head);
drivers/dma/ti/k3-udma.c
3971
vchan_get_all_descriptors(&uc->vc, &head);
drivers/dma/ti/k3-udma.c
3973
vchan_dma_desc_free_list(&uc->vc, &head);
drivers/dma/ti/k3-udma.c
4053
LIST_HEAD(head);
drivers/dma/ti/k3-udma.c
4056
list_splice_tail_init(&vc->desc_completed, &head);
drivers/dma/ti/k3-udma.c
4069
list_for_each_entry_safe(vd, _vd, &head, node) {
drivers/dma/ti/k3-udma.c
560
LIST_HEAD(head);
drivers/dma/ti/k3-udma.c
563
list_splice_tail_init(&ud->desc_to_purge, &head);
drivers/dma/ti/k3-udma.c
566
list_for_each_entry_safe(vd, _vd, &head, node) {
drivers/dma/ti/omap-dma.c
1391
LIST_HEAD(head);
drivers/dma/ti/omap-dma.c
1411
vchan_get_all_descriptors(&c->vc, &head);
drivers/dma/ti/omap-dma.c
1413
vchan_dma_desc_free_list(&c->vc, &head);
drivers/dma/uniphier-mdmac.c
258
LIST_HEAD(head);
drivers/dma/uniphier-mdmac.c
267
vchan_get_all_descriptors(vc, &head);
drivers/dma/uniphier-mdmac.c
271
vchan_dma_desc_free_list(vc, &head);
drivers/dma/uniphier-xdmac.c
411
LIST_HEAD(head);
drivers/dma/uniphier-xdmac.c
421
vchan_get_all_descriptors(vc, &head);
drivers/dma/uniphier-xdmac.c
425
vchan_dma_desc_free_list(vc, &head);
drivers/dma/virt-dma.c
103
list_for_each_entry_safe(vd, _vd, &head, node) {
drivers/dma/virt-dma.c
112
void vchan_dma_desc_free_list(struct virt_dma_chan *vc, struct list_head *head)
drivers/dma/virt-dma.c
116
list_for_each_entry_safe(vd, _vd, head, node) {
drivers/dma/virt-dma.c
88
LIST_HEAD(head);
drivers/dma/virt-dma.c
91
list_splice_tail_init(&vc->desc_completed, &head);
drivers/dma/virt-dma.h
186
struct list_head *head)
drivers/dma/virt-dma.h
190
list_splice_tail_init(&vc->desc_allocated, head);
drivers/dma/virt-dma.h
191
list_splice_tail_init(&vc->desc_submitted, head);
drivers/dma/virt-dma.h
192
list_splice_tail_init(&vc->desc_issued, head);
drivers/dma/virt-dma.h
193
list_splice_tail_init(&vc->desc_completed, head);
drivers/dma/virt-dma.h
194
list_splice_tail_init(&vc->desc_terminated, head);
drivers/dma/virt-dma.h
201
LIST_HEAD(head);
drivers/dma/virt-dma.h
204
vchan_get_all_descriptors(vc, &head);
drivers/dma/virt-dma.h
205
list_for_each_entry(vd, &head, node)
drivers/dma/virt-dma.h
209
vchan_dma_desc_free_list(vc, &head);
drivers/dma/virt-dma.h
223
LIST_HEAD(head);
drivers/dma/virt-dma.h
230
list_splice_tail_init(&vc->desc_terminated, &head);
drivers/dma/virt-dma.h
234
vchan_dma_desc_free_list(vc, &head);
drivers/dma/virt-dma.h
44
void vchan_dma_desc_free_list(struct virt_dma_chan *vc, struct list_head *head);
drivers/dma/xgene-dma.c
220
u16 head;
drivers/dma/xgene-dma.c
601
desc_hw = &ring->desc_hw[ring->head];
drivers/dma/xgene-dma.c
607
if (++ring->head == ring->slots)
drivers/dma/xgene-dma.c
608
ring->head = 0;
drivers/dma/xgene-dma.c
618
desc_hw = &ring->desc_hw[ring->head];
drivers/dma/xgene-dma.c
620
if (++ring->head == ring->slots)
drivers/dma/xgene-dma.c
621
ring->head = 0;
drivers/dma/xgene-dma.c
705
desc_hw = &ring->desc_hw[ring->head];
drivers/dma/xgene-dma.c
712
if (++ring->head == ring->slots)
drivers/dma/xgene-dma.c
713
ring->head = 0;
drivers/dma/xilinx/xdma.c
512
LIST_HEAD(head);
drivers/dma/xilinx/xdma.c
526
vchan_get_all_descriptors(&xdma_chan->vchan, &head);
drivers/dma/xilinx/xdma.c
527
list_splice_tail(&head, &xdma_chan->vchan.desc_terminated);
drivers/dma/xilinx/xilinx_dma.c
2228
struct xilinx_axidma_tx_segment *segment, *head, *prev = NULL;
drivers/dma/xilinx/xilinx_dma.c
2282
head = list_first_entry(&desc->segments, struct xilinx_axidma_tx_segment, node);
drivers/dma/xilinx/xilinx_dma.c
2283
desc->async_tx.phys = head->phys;
drivers/edac/igen6_edac.c
960
struct llist_node *head;
drivers/edac/igen6_edac.c
964
head = llist_del_all(&ecclog_llist);
drivers/edac/igen6_edac.c
965
if (!head)
drivers/edac/igen6_edac.c
968
llist_for_each_entry_safe(node, tmp, head, llnode) {
drivers/edac/imh_base.c
367
static bool imh_2lm_enabled(struct res_config *cfg, struct list_head *head)
drivers/edac/imh_base.c
372
list_for_each_entry(d, head, list) {
drivers/edac/thunderx_edac.c
1076
unsigned long head = ring_pos(ocx->com_ring_head,
drivers/edac/thunderx_edac.c
1078
struct ocx_com_err_ctx *ctx = &ocx->com_err_ctx[head];
drivers/edac/thunderx_edac.c
1167
unsigned long head = ring_pos(ocx->link_ring_head,
drivers/edac/thunderx_edac.c
1169
struct ocx_link_err_ctx *ctx = &ocx->link_err_ctx[head];
drivers/edac/thunderx_edac.c
1759
unsigned long head = ring_pos(tad->ring_head, ARRAY_SIZE(tad->err_ctx));
drivers/edac/thunderx_edac.c
1760
struct l2c_err_ctx *ctx = &tad->err_ctx[head];
drivers/edac/thunderx_edac.c
1791
unsigned long head = ring_pos(cbc->ring_head, ARRAY_SIZE(cbc->err_ctx));
drivers/edac/thunderx_edac.c
1792
struct l2c_err_ctx *ctx = &cbc->err_ctx[head];
drivers/edac/thunderx_edac.c
1820
unsigned long head = ring_pos(mci->ring_head, ARRAY_SIZE(mci->err_ctx));
drivers/edac/thunderx_edac.c
1821
struct l2c_err_ctx *ctx = &mci->err_ctx[head];
drivers/edac/thunderx_edac.c
544
unsigned long head = ring_pos(lmc->ring_head, ARRAY_SIZE(lmc->err_ctx));
drivers/edac/thunderx_edac.c
545
struct lmc_err_ctx *ctx = &lmc->err_ctx[head];
drivers/firewire/nosy.c
122
buffer->head = (struct packet *) buffer->data;
drivers/firewire/nosy.c
153
length = buffer->head->length;
drivers/firewire/nosy.c
160
if (&buffer->head->data[length] < end) {
drivers/firewire/nosy.c
161
if (copy_to_user(data, buffer->head->data, length))
drivers/firewire/nosy.c
163
buffer->head = (struct packet *) &buffer->head->data[length];
drivers/firewire/nosy.c
165
size_t split = end - buffer->head->data;
drivers/firewire/nosy.c
167
if (copy_to_user(data, buffer->head->data, split))
drivers/firewire/nosy.c
171
buffer->head = (struct packet *) &buffer->data[length - split];
drivers/firewire/nosy.c
65
struct packet *head, *tail;
drivers/firmware/arm_ffa/driver.c
1635
struct list_head *head, *phead;
drivers/firmware/arm_ffa/driver.c
1640
head = phead;
drivers/firmware/arm_ffa/driver.c
1641
list_for_each_entry(info, head, node) {
drivers/firmware/arm_scmi/bus.c
57
struct list_head *head, *phead = NULL;
drivers/firmware/arm_scmi/bus.c
77
head = phead;
drivers/firmware/arm_scmi/bus.c
78
list_for_each_entry(rdev, head, node) {
drivers/firmware/arm_scmi/raw_mode.c
262
struct list_head *head = &q->free_bufs;
drivers/firmware/arm_scmi/raw_mode.c
265
if (!list_empty(head)) {
drivers/firmware/arm_scmi/raw_mode.c
266
rb = list_first_entry(head, struct scmi_raw_buffer, node);
drivers/firmware/dmi_scan.c
1000
const struct list_head *head = from ? &from->list : &dmi_devices;
drivers/firmware/dmi_scan.c
1003
for (d = head->next; d != &dmi_devices; d = d->next) {
drivers/gpib/common/gpib_os.c
1153
static int increment_open_device_count(struct gpib_board *board, struct list_head *head,
drivers/gpib/common/gpib_os.c
1163
for (list_ptr = head->next; list_ptr != head; list_ptr = list_ptr->next) {
drivers/gpib/common/gpib_os.c
1182
list_add(&device->list, head);
drivers/gpib/common/gpib_os.c
1189
static int subtract_open_device_count(struct gpib_board *board, struct list_head *head,
drivers/gpib/common/gpib_os.c
1195
for (list_ptr = head->next; list_ptr != head; list_ptr = list_ptr->next) {
drivers/gpib/common/gpib_os.c
1218
static inline int decrement_open_device_count(struct gpib_board *board, struct list_head *head,
drivers/gpib/common/gpib_os.c
1221
return subtract_open_device_count(board, head, pad, sad, 1);
drivers/gpib/common/gpib_os.c
188
struct list_head *head = &device->status_bytes;
drivers/gpib/common/gpib_os.c
1914
struct list_head *head = &queue->event_head;
drivers/gpib/common/gpib_os.c
1938
list_add_tail(&event->list, head);
drivers/gpib/common/gpib_os.c
1970
struct list_head *head = &queue->event_head;
drivers/gpib/common/gpib_os.c
1971
struct list_head *front = head->next;
drivers/gpib/common/gpib_os.c
1979
if (front == head)
drivers/gpib/common/gpib_os.c
209
list_add_tail(&status->list, head);
drivers/gpib/common/gpib_os.c
222
struct list_head *head = &device->status_bytes;
drivers/gpib/common/gpib_os.c
223
struct list_head *front = head->next;
drivers/gpib/common/gpib_os.c
229
if (front == head)
drivers/gpib/common/gpib_os.c
255
const struct list_head *head = &board->device_list;
drivers/gpib/common/gpib_os.c
257
for (list_ptr = head->next; list_ptr != head; list_ptr = list_ptr->next) {
drivers/gpib/common/gpib_os.c
435
const struct list_head *head = NULL;
drivers/gpib/common/gpib_os.c
440
head = &board->device_list;
drivers/gpib/common/gpib_os.c
441
if (head->next == head)
drivers/gpib/common/gpib_os.c
448
for (cur = head->next; cur != head; cur = cur->next) {
drivers/gpib/lpvo_usb_gpib/lpvo_usb_gpib.c
83
#define WQH head
drivers/gpio/gpio-rockchip.c
761
struct rockchip_pin_deferred, head);
drivers/gpio/gpio-rockchip.c
762
list_del(&cfg->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_acpi.c
1310
head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c
1641
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c
347
struct drm_display_mode, head);
drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c
609
list_for_each_entry_safe(mode, t, &connector->probed_modes, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c
619
list_for_each_entry_safe(mode, t, &connector->probed_modes, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
6241
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
7087
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
7135
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
7190
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1510
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1775
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1814
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
331
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head)
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
453
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_encoders.c
135
list_for_each_entry(other_encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_encoders.c
48
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
1038
ih_data.head = *ras_if;
drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
68
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head)
drivers/gpu/drm/amd/amdgpu/amdgpu_jpeg.c
282
ih_data.head = *ras_if;
drivers/gpu/drm/amd/amdgpu/amdgpu_pll.c
276
list_for_each_entry(test_crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_pll.c
302
list_for_each_entry(test_crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_pll.c
337
list_for_each_entry(test_crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
1549
list_for_each_entry(mirror_adev, &hive->device_list, gmc.xgmi.head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1011
struct ras_common_if head = {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1025
if (__amdgpu_ras_feature_enable(adev, &head, 1))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1028
if (amdgpu_ras_feature_enable(adev, &head, 1))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1034
struct ras_common_if head = {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1045
if (__amdgpu_ras_feature_enable(adev, &head, 1))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1048
if (amdgpu_ras_feature_enable(adev, &head, 1))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1226
struct ras_manager *ras_mgr = amdgpu_ras_find_obj(adev, &query_if->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1227
const char *blk_name = get_ras_block_str(&query_if->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1306
struct ras_manager *obj = amdgpu_ras_find_obj(adev, &query_if->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1307
const char *blk_name = get_ras_block_str(&query_if->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1370
struct ras_common_if head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1372
memset(&head, 0, sizeof(head));
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1373
head.block = blk;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1375
return amdgpu_ras_find_obj(adev, &head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1425
.head = obj->head,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1444
enum amdgpu_ras_block blk = info ? info->head.block : AMDGPU_RAS_BLOCK_COUNT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1457
if (info->head.block == AMDGPU_RAS_BLOCK__UMC) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1460
block_obj = amdgpu_ras_get_ras_block(adev, info->head.block, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1463
get_ras_block_str(&info->head));
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1470
if ((info->head.block == AMDGPU_RAS_BLOCK__SDMA) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1471
(info->head.block == AMDGPU_RAS_BLOCK__GFX) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1472
(info->head.block == AMDGPU_RAS_BLOCK__MMHUB)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1505
struct ras_manager *obj = amdgpu_ras_find_obj(adev, &info->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1585
req.block_id = info->head.block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1586
req.subblock_id = info->head.sub_block_index;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1666
inject_req.block_id = info->head.block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1667
inject_req.subblock_id = info->head.sub_block_index;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1669
inject_req.error_type = info->head.type;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1681
struct ras_manager *obj = amdgpu_ras_find_obj(adev, &info->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1683
.block_id = amdgpu_ras_block_to_ta(info->head.block),
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1684
.inject_error_type = amdgpu_ras_error_to_ta(info->head.type),
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1685
.sub_block_index = info->head.sub_block_index,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1691
info->head.block,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1692
info->head.sub_block_index);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1706
get_ras_block_str(&info->head));
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1712
info->head.block != AMDGPU_RAS_BLOCK__GFX) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1719
if (info->head.block == AMDGPU_RAS_BLOCK__GFX)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1731
get_ras_block_str(&info->head), ret);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1768
if (amdgpu_ras_reset_error_status(adev, query_info->head.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1812
list_for_each_entry(obj, &con->head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1814
.head = obj->head,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2073
struct ras_common_if *head)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2075
struct ras_manager *obj = amdgpu_ras_find_obj(adev, head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2083
if (amdgpu_sriov_vf(adev) && !amdgpu_virt_ras_telemetry_block_en(adev, head->block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2089
"%s_err_count", head->name);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2113
struct ras_common_if *head)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2115
struct ras_manager *obj = amdgpu_ras_find_obj(adev, head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2138
list_for_each_entry_safe(obj, tmp, &con->head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2139
amdgpu_ras_sysfs_remove(adev, &obj->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2214
struct ras_fs_if *head,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2217
struct ras_manager *obj = amdgpu_ras_find_obj(adev, &head->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2225
head->debugfs_name,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2266
list_for_each_entry(obj, &con->head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2267
if (amdgpu_ras_is_supported(adev, obj->head.block) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2270
get_ras_block_str(&obj->head));
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2271
fs_info.head = obj->head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2355
list_for_each_entry_safe(con_obj, tmp, &con->head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2356
ip_obj = amdgpu_ras_find_obj(adev, &con_obj->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2410
amdgpu_ras_get_ras_block(adev, obj->head.block, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2439
amdgpu_umc_poison_handler(adev, obj->head.block, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2535
if (obj->head.block == AMDGPU_RAS_BLOCK__UMC)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2540
if (obj->head.block == AMDGPU_RAS_BLOCK__UMC)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2569
ih_info.block = info->head.block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2575
obj = amdgpu_ras_find_obj(adev, &info->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2598
struct ras_common_if *head)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2600
struct ras_manager *obj = amdgpu_ras_find_obj(adev, head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2620
struct ras_common_if *head)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2622
struct ras_manager *obj = amdgpu_ras_find_obj(adev, head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2628
obj = amdgpu_ras_create_obj(adev, head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2634
ras_obj = container_of(head, struct amdgpu_ras_block_object, ras_comm);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2668
list_for_each_entry_safe(obj, tmp, &con->head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2669
amdgpu_ras_interrupt_remove_handler(adev, &obj->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2685
list_for_each_entry(obj, &con->head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2687
.head = obj->head,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2696
if (info.head.block == AMDGPU_RAS_BLOCK__PCIE_BIF)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2705
if ((info.head.block == AMDGPU_RAS_BLOCK__UMC) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2718
if (amdgpu_ras_reset_error_status(adev, info.head.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2733
if ((info->head.block != AMDGPU_RAS_BLOCK__GFX) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2734
(info->head.block != AMDGPU_RAS_BLOCK__MMHUB))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2738
info->head.block,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2739
info->head.sub_block_index);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2743
get_ras_block_str(&info->head));
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2760
list_for_each_entry(obj, &con->head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2762
.head = obj->head,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
278
.head = obj->head,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2858
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
289
if (amdgpu_ras_reset_error_status(obj->adev, info.head.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2910
gmc.xgmi.head)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2923
list_add_tail(&adev->gmc.xgmi.head, &device_list);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2936
device_list_handle, gmc.xgmi.head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
3621
info.head.block = AMDGPU_RAS_BLOCK__UMC;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
393
data->head.block = block_id;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
396
data->head.type = AMDGPU_RAS_ERROR__MULTI_UNCORRECTABLE;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
398
data->head.type = AMDGPU_RAS_ERROR__SINGLE_CORRECTABLE;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
400
data->head.type = AMDGPU_RAS_ERROR__POISON;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
416
data->head.sub_block_index = sub_block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4306
INIT_LIST_HEAD(&con->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4426
.head = *ras_block,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
448
switch (data->head.block) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4505
memcpy(&query_info->head, ras_block, sizeof(struct ras_common_if));
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4579
list_for_each_entry_safe(obj, tmp, &con->head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4580
if (!amdgpu_ras_is_supported(adev, obj->head.block)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4581
amdgpu_ras_feature_enable(adev, &obj->head, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
582
if (!amdgpu_ras_is_supported(adev, data.head.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
587
ret = amdgpu_ras_feature_enable(adev, &data.head, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
590
ret = amdgpu_ras_feature_enable(adev, &data.head, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
594
if (data.head.block == AMDGPU_RAS_BLOCK__UMC)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
706
.head = obj->head,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
717
if (amdgpu_ras_reset_error_status(obj->adev, info.head.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
721
if (info.head.block == AMDGPU_RAS_BLOCK__UMC)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
742
DRM_ERROR("RAS ERROR: Unbalance obj(%s) use\n", get_ras_block_str(&obj->head));
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
747
struct ras_common_if *head)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
755
if (head->block >= AMDGPU_RAS_BLOCK_COUNT)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
758
if (head->block == AMDGPU_RAS_BLOCK__MCA) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
759
if (head->sub_block_index >= AMDGPU_RAS_MCA_BLOCK__LAST)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
762
obj = &con->objs[AMDGPU_RAS_BLOCK__LAST + head->sub_block_index];
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
764
obj = &con->objs[head->block];
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
773
obj->head = *head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
775
list_add(&obj->node, &con->head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
783
struct ras_common_if *head)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
792
if (head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
793
if (head->block >= AMDGPU_RAS_BLOCK_COUNT)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
796
if (head->block == AMDGPU_RAS_BLOCK__MCA) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
797
if (head->sub_block_index >= AMDGPU_RAS_MCA_BLOCK__LAST)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
800
obj = &con->objs[AMDGPU_RAS_BLOCK__LAST + head->sub_block_index];
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
802
obj = &con->objs[head->block];
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
820
struct ras_common_if *head)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
822
return adev->ras_hw_enabled & BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
826
struct ras_common_if *head)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
830
return con->features & BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
838
struct ras_common_if *head, int enable)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
841
struct ras_manager *obj = amdgpu_ras_find_obj(adev, head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
849
if (!amdgpu_ras_is_feature_allowed(adev, head))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
854
obj = amdgpu_ras_create_obj(adev, head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
861
con->features |= BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
863
if (obj && amdgpu_ras_is_feature_enabled(adev, head)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
864
con->features &= ~BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
874
struct ras_common_if *head, bool enable)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
886
if (head->block != AMDGPU_RAS_BLOCK__GFX &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
887
!amdgpu_ras_is_feature_allowed(adev, head))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
891
if (head->block == AMDGPU_RAS_BLOCK__GFX &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
900
.block_id = amdgpu_ras_block_to_ta(head->block),
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
901
.error_type = amdgpu_ras_error_to_ta(head->type),
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
905
.block_id = amdgpu_ras_block_to_ta(head->block),
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
906
.error_type = amdgpu_ras_error_to_ta(head->type),
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
914
get_ras_block_str(head),
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
924
__amdgpu_ras_feature_enable(adev, head, enable);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
931
struct ras_common_if *head, bool enable)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
947
ret = amdgpu_ras_feature_enable(adev, head, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
953
ret = __amdgpu_ras_feature_enable(adev, head, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
957
get_ras_block_str(head));
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
961
ret = __amdgpu_ras_feature_enable(adev, head, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
966
if (head->block == AMDGPU_RAS_BLOCK__GFX)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
967
con->features |= BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
969
ret = amdgpu_ras_feature_enable(adev, head, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
972
if (adev->ras_enabled && head->block == AMDGPU_RAS_BLOCK__GFX)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
973
con->features &= ~BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
976
ret = amdgpu_ras_feature_enable(adev, head, enable);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
987
list_for_each_entry_safe(obj, tmp, &con->head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
992
if (__amdgpu_ras_feature_enable(adev, &obj->head, 0))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
995
if (amdgpu_ras_feature_enable(adev, &obj->head, 0))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
538
struct list_head head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
687
struct ras_common_if head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
717
struct ras_common_if head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
723
struct ras_common_if head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
730
struct ras_common_if head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
737
struct ras_common_if head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
742
struct ras_common_if head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
747
struct ras_common_if head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
753
struct ras_common_if head;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
895
struct ras_common_if *head, bool enable);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
898
struct ras_common_if *head, bool enable);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
901
struct ras_common_if *head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
904
struct ras_common_if *head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
920
struct ras_common_if *head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
923
struct ras_common_if *head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
929
struct ras_common_if *head);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
59
struct list_head *head, *next;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
71
head = &to_amdgpu_vram_mgr_resource(res)->blocks;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
73
block = list_first_entry_or_null(head,
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
83
if (next != head)
drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c
144
ih_data.head = *ras_if;
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
259
struct ras_common_if head = {
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
262
struct ras_manager *obj = amdgpu_ras_find_obj(adev, &head);
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
395
ih_data.head = *ras_if;
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c
1275
ih_data.head = *ras_if;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
61
static inline bool amdgpu_is_vram_mgr_blocks_contiguous(struct list_head *head)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
66
block = amdgpu_vram_mgr_first_block(head);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
70
while (head != block->link.next) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
82
static inline u64 amdgpu_vram_mgr_blocks_size(struct list_head *head)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
87
list_for_each_entry(block, head, link)
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1019
list_add_tail(&adev->gmc.xgmi.head, &hive->device_list);
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1020
list_for_each_entry(entry, &hive->device_list, head)
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1028
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1056
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1079
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1134
list_del(&adev->gmc.xgmi.head);
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1636
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head)
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1654
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1716
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1727
tmp_adev, &hive->device_list, gmc.xgmi.head)
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
955
list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.h
89
struct list_head head;
drivers/gpu/drm/amd/amdgpu/atombios_encoders.c
1646
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/dce_v10_0.c
2628
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/dce_v10_0.c
3509
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/dce_v6_0.c
2595
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/dce_v6_0.c
3407
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/dce_v8_0.c
2554
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/dce_v8_0.c
3417
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/amd/amdgpu/gfx_v11_0_3.c
68
ih_data.head = *ras_if;
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6762
if (info->head.sub_block_index >= ARRAY_SIZE(ras_gfx_subblocks))
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6765
if (!ras_gfx_subblocks[info->head.sub_block_index].name)
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6768
if (!(ras_gfx_subblocks[info->head.sub_block_index].hw_supported_error_type &
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6769
info->head.type)) {
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6771
ras_gfx_subblocks[info->head.sub_block_index].name,
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6772
info->head.type);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6776
if (!(ras_gfx_subblocks[info->head.sub_block_index].sw_supported_error_type &
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6777
info->head.type)) {
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6779
ras_gfx_subblocks[info->head.sub_block_index].name,
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6780
info->head.type);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6784
block_info.block_id = amdgpu_ras_block_to_ta(info->head.block);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6786
ras_gfx_subblocks[info->head.sub_block_index].ta_subblock;
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6787
block_info.inject_error_type = amdgpu_ras_error_to_ta(info->head.type);
drivers/gpu/drm/amd/amdkfd/kfd_events.c
268
list_for_each_entry(waiter, &ev->wq.head, wait.entry)
drivers/gpu/drm/amd/amdkfd/kfd_events.c
646
list_for_each_entry(waiter, &ev->wq.head, wait.entry)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1175
struct svm_range *head = NULL;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1178
r = svm_range_split(prange, new_start, prange->last, &head);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1183
list_add(&head->list, insert_list);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1185
if (huge_page_mapping && head->last + 1 > start_align &&
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1186
head->last + 1 < last_align_down && (!IS_ALIGNED(head->last, 512)))
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1187
list_add(&head->update_list, remap_list);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2506
struct svm_range *head;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2517
head = tail = prange;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2521
svm_range_split(tail, last + 1, tail->last, &head);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2523
if (head != prange && tail != prange) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2524
svm_range_add_child(parent, head, SVM_OP_UNMAP_RANGE);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2528
} else if (head != prange) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2529
svm_range_add_child(parent, head, SVM_OP_UNMAP_RANGE);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
1085
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
11285
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
348
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
4344
&dev->mode_config.connector_list, head) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
5158
list_for_each_entry(connector, &dm->ddev->mode_config.connector_list, head) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
7023
list_for_each_entry(m, list_head, head) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
7033
&aconnector->base.modes, struct drm_display_mode, head);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
7047
list_for_each_entry(m, list_head, head) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
7317
list_for_each_entry(preferred_mode, &connector->modes, head) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
7328
head);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
8628
head) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
8707
list_for_each_entry(curmode, &connector->probed_modes, head) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
8797
list_for_each_entry(m, &aconnector->base.probed_modes, head) {
drivers/gpu/drm/amd/pm/amdgpu_dpm_internal.c
45
list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
drivers/gpu/drm/amd/ras/rascore/ras_cmd.c
33
INIT_LIST_HEAD(&ras_core->ras_cmd.head);
drivers/gpu/drm/amd/ras/rascore/ras_cmd.h
109
struct list_head head;
drivers/gpu/drm/arm/display/komeda/komeda_private_obj.c
430
list_for_each_entry_safe(obj, next, &config->privobj_list, head)
drivers/gpu/drm/arm/malidp_drv.c
833
list_for_each_entry(encoder, &drm->mode_config.encoder_list, head) {
drivers/gpu/drm/drm_atomic.c
2044
list_for_each_entry(colorop, &config->colorop_list, head) {
drivers/gpu/drm/drm_atomic.c
2052
list_for_each_entry(plane, &config->plane_list, head) {
drivers/gpu/drm/drm_atomic.c
2060
list_for_each_entry(crtc, &config->crtc_list, head) {
drivers/gpu/drm/drm_atomic.c
2077
list_for_each_entry(obj, &config->privobj_list, head) {
drivers/gpu/drm/drm_atomic.c
943
list_add_tail(&obj->head, &dev->mode_config.privobj_list);
drivers/gpu/drm/drm_atomic.c
958
list_del(&obj->head);
drivers/gpu/drm/drm_client_modeset.c
128
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/drm_client_modeset.c
141
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/drm_client_modeset.c
155
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/drm_client_modeset.c
169
struct drm_display_mode, head);
drivers/gpu/drm/drm_client_modeset.c
185
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/drm_client_modeset.c
201
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/drm_client_modeset.c
359
list_for_each_entry(mode, &connectors[i]->modes, head) {
drivers/gpu/drm/drm_colorop.c
113
list_add_tail(&colorop->head, &config->colorop_list);
drivers/gpu/drm/drm_colorop.c
169
list_del(&colorop->head);
drivers/gpu/drm/drm_colorop.c
193
list_for_each_entry_safe(colorop, next, &config->colorop_list, head) {
drivers/gpu/drm/drm_connector.c
1093
lhead = old_conn ? &old_conn->head : &config->connector_list;
drivers/gpu/drm/drm_connector.c
1102
iter->conn = list_entry(lhead, struct drm_connector, head);
drivers/gpu/drm/drm_connector.c
278
INIT_LIST_HEAD(&connector->head);
drivers/gpu/drm/drm_connector.c
3302
list_for_each_entry(mode_itr, modes, head) {
drivers/gpu/drm/drm_connector.c
338
if (drm_WARN_ON(dev, !list_empty(&connector->head)))
drivers/gpu/drm/drm_connector.c
3380
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/drm_connector.c
3397
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/drm_connector.c
3419
list_for_each_entry_continue(mode, &connector->modes, head)
drivers/gpu/drm/drm_connector.c
342
list_add_tail(&connector->head, &config->connector_list);
drivers/gpu/drm/drm_connector.c
3430
list_for_each_entry(mode, &connector->modes, head)
drivers/gpu/drm/drm_connector.c
355
if (list_empty(&connector->head))
drivers/gpu/drm/drm_connector.c
359
list_del_init(&connector->head);
drivers/gpu/drm/drm_connector.c
708
list_del(&mode->head);
drivers/gpu/drm/drm_connector.c
782
list_for_each_entry_safe(mode, t, &connector->probed_modes, head)
drivers/gpu/drm/drm_connector.c
785
list_for_each_entry_safe(mode, t, &connector->modes, head)
drivers/gpu/drm/drm_crtc.c
302
list_add_tail(&crtc->head, &config->crtc_list);
drivers/gpu/drm/drm_crtc.c
519
list_del(&crtc->head);
drivers/gpu/drm/drm_debugfs_crc.c
182
return CIRC_CNT(crc->head, crc->tail, DRM_CRC_ENTRIES_NR);
drivers/gpu/drm/drm_debugfs_crc.c
190
crc->head = 0;
drivers/gpu/drm/drm_debugfs_crc.c
399
int head, tail;
drivers/gpu/drm/drm_debugfs_crc.c
410
head = crc->head;
drivers/gpu/drm/drm_debugfs_crc.c
413
if (CIRC_SPACE(head, tail, DRM_CRC_ENTRIES_NR) < 1) {
drivers/gpu/drm/drm_debugfs_crc.c
425
entry = &crc->entries[head];
drivers/gpu/drm/drm_debugfs_crc.c
430
head = (head + 1) & (DRM_CRC_ENTRIES_NR - 1);
drivers/gpu/drm/drm_debugfs_crc.c
431
crc->head = head;
drivers/gpu/drm/drm_edid.c
3002
struct drm_display_mode, head);
drivers/gpu/drm/drm_edid.c
3004
list_for_each_entry_safe(cur_mode, t, &connector->probed_modes, head) {
drivers/gpu/drm/drm_edid.c
3405
list_for_each_entry(m, &connector->probed_modes, head)
drivers/gpu/drm/drm_edid.c
3692
list_for_each_entry(m, &connector->probed_modes, head) {
drivers/gpu/drm/drm_edid.c
4517
list_for_each_entry(mode, &connector->probed_modes, head) {
drivers/gpu/drm/drm_edid.c
4561
list_add_tail(&newmode->head, &list);
drivers/gpu/drm/drm_edid.c
4564
list_for_each_entry_safe(mode, tmp, &list, head) {
drivers/gpu/drm/drm_edid.c
4565
list_del(&mode->head);
drivers/gpu/drm/drm_edid.c
4720
list_for_each_entry(mode, &connector->probed_modes, head) {
drivers/gpu/drm/drm_edid.c
4735
list_add_tail(&new_mode->head, &stereo_modes);
drivers/gpu/drm/drm_encoder.c
132
list_add_tail(&encoder->head, &dev->mode_config.encoder_list);
drivers/gpu/drm/drm_encoder.c
203
list_del(&encoder->head);
drivers/gpu/drm/drm_framebuffer.c
895
list_add(&fb->head, &dev->mode_config.fb_list);
drivers/gpu/drm/drm_framebuffer.c
993
list_del(&fb->head);
drivers/gpu/drm/drm_gpusvm.c
450
struct list_head *head;
drivers/gpu/drm/drm_gpusvm.c
456
head = &(to_drm_gpusvm_notifier(node))->entry;
drivers/gpu/drm/drm_gpusvm.c
458
head = &gpusvm->notifier_list;
drivers/gpu/drm/drm_gpusvm.c
460
list_add(&notifier->entry, head);
drivers/gpu/drm/drm_gpusvm.c
580
struct list_head *head;
drivers/gpu/drm/drm_gpusvm.c
587
head = &(to_drm_gpusvm_range(node))->entry;
drivers/gpu/drm/drm_gpusvm.c
589
head = &notifier->range_list;
drivers/gpu/drm/drm_gpusvm.c
591
list_add(&range->entry, head);
drivers/gpu/drm/drm_gpuvm.c
1986
struct list_head *head;
drivers/gpu/drm/drm_gpuvm.c
1999
head = &(to_drm_gpuva(node))->rb.entry;
drivers/gpu/drm/drm_gpuvm.c
2001
head = &gpuvm->rb.list;
drivers/gpu/drm/drm_gpuvm.c
2003
list_add(&va->rb.entry, head);
drivers/gpu/drm/drm_lease.c
176
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/drm_mode_config.c
529
head) {
drivers/gpu/drm/drm_mode_config.c
552
head) {
drivers/gpu/drm/drm_mode_config.c
557
head) {
drivers/gpu/drm/drm_mode_config.c
561
list_for_each_entry_safe(crtc, ct, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/drm_mode_config.c
579
list_for_each_entry_safe(fb, fbt, &dev->mode_config.fb_list, head) {
drivers/gpu/drm/drm_modes.c
116
list_add_tail(&mode->head, &connector->probed_modes);
drivers/gpu/drm/drm_modes.c
1422
struct list_head head = dst->head;
drivers/gpu/drm/drm_modes.c
1425
dst->head = head;
drivers/gpu/drm/drm_modes.c
1811
list_for_each_entry_safe(mode, t, mode_list, head) {
drivers/gpu/drm/drm_modes.c
1813
list_del(&mode->head);
drivers/gpu/drm/drm_modes.c
1844
struct drm_display_mode *a = list_entry(lh_a, struct drm_display_mode, head);
drivers/gpu/drm/drm_modes.c
1845
struct drm_display_mode *b = list_entry(lh_b, struct drm_display_mode, head);
drivers/gpu/drm/drm_modes.c
1893
list_for_each_entry_safe(pmode, pt, &connector->probed_modes, head) {
drivers/gpu/drm/drm_modes.c
1898
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/drm_modes.c
1926
list_del(&pmode->head);
drivers/gpu/drm/drm_modes.c
1932
list_move_tail(&pmode->head, &connector->modes);
drivers/gpu/drm/drm_modes.c
2772
list_for_each_entry(mode, &connector->probed_modes, head) {
drivers/gpu/drm/drm_modeset_helper.c
61
&dev->mode_config.connector_list, head) {
drivers/gpu/drm/drm_modeset_helper.c
65
list_move_tail(&connector->head, &panel_list);
drivers/gpu/drm/drm_modeset_lock.c
285
struct drm_modeset_lock, head);
drivers/gpu/drm/drm_modeset_lock.c
319
WARN_ON(!list_empty(&lock->head));
drivers/gpu/drm/drm_modeset_lock.c
320
list_add(&lock->head, &ctx->locked);
drivers/gpu/drm/drm_modeset_lock.c
371
INIT_LIST_HEAD(&lock->head);
drivers/gpu/drm/drm_modeset_lock.c
426
list_del_init(&lock->head);
drivers/gpu/drm/drm_plane.c
458
list_add_tail(&plane->head, &config->plane_list);
drivers/gpu/drm/drm_plane.c
693
BUG_ON(list_empty(&plane->head));
drivers/gpu/drm/drm_plane.c
700
list_del(&plane->head);
drivers/gpu/drm/drm_probe_helper.c
156
list_for_each_entry(mode, &connector->probed_modes, head) {
drivers/gpu/drm/drm_probe_helper.c
463
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/drm_probe_helper.c
584
list_for_each_entry(mode, &connector->modes, head)
drivers/gpu/drm/drm_probe_helper.c
704
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/drm_property.c
133
list_add_tail(&property->head, &dev->mode_config.property_list);
drivers/gpu/drm/drm_property.c
411
list_for_each_entry(prop_enum, &property->enum_list, head) {
drivers/gpu/drm/drm_property.c
428
list_add_tail(&prop_enum->head, &property->enum_list);
drivers/gpu/drm/drm_property.c
445
list_for_each_entry_safe(prop_enum, pt, &property->enum_list, head) {
drivers/gpu/drm/drm_property.c
446
list_del(&prop_enum->head);
drivers/gpu/drm/drm_property.c
453
list_del(&property->head);
drivers/gpu/drm/drm_property.c
496
list_for_each_entry(prop_enum, &property->enum_list, head) {
drivers/gpu/drm/drm_sysfs.c
288
list_for_each_entry(mode, &connector->modes, head) {
drivers/gpu/drm/exynos/exynos_drm_drv.c
266
list_for_each_entry(encoder, &drm->mode_config.encoder_list, head)
drivers/gpu/drm/exynos/exynos_drm_drv.c
269
list_for_each_entry(encoder, &drm->mode_config.encoder_list, head)
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1279
cmdlist->head = cmdlist->last / 2;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
157
u32 head;
drivers/gpu/drm/exynos/exynos_drm_ipp.c
119
list_for_each_entry(ipp, &ipp_list, head) {
drivers/gpu/drm/exynos/exynos_drm_ipp.c
134
list_for_each_entry(ipp, &ipp_list, head)
drivers/gpu/drm/exynos/exynos_drm_ipp.c
71
list_add_tail(&ipp->head, &ipp_list);
drivers/gpu/drm/exynos/exynos_drm_ipp.c
804
head);
drivers/gpu/drm/exynos/exynos_drm_ipp.c
805
list_del_init(&task->head);
drivers/gpu/drm/exynos/exynos_drm_ipp.c
825
list_add(&task->head, &ipp->todo_list);
drivers/gpu/drm/exynos/exynos_drm_ipp.c
842
list_del_init(&task->head);
drivers/gpu/drm/exynos/exynos_drm_ipp.c
89
list_del(&ipp->head);
drivers/gpu/drm/exynos/exynos_drm_ipp.h
53
struct list_head head;
drivers/gpu/drm/exynos/exynos_drm_ipp.h
85
struct list_head head;
drivers/gpu/drm/exynos/exynos_hdmi.c
1028
list_for_each_entry(m, &connector->modes, head) {
drivers/gpu/drm/gma500/cdv_intel_dp.c
1749
head) {
drivers/gpu/drm/gma500/cdv_intel_dp.c
997
list_for_each_entry(encoder, &mode_config->encoder_list, head) {
drivers/gpu/drm/gma500/cdv_intel_lvds.c
193
head) {
drivers/gpu/drm/gma500/cdv_intel_lvds.c
590
list_for_each_entry(scan, &connector->probed_modes, head) {
drivers/gpu/drm/gma500/oaktrail_lvds.c
391
list_for_each_entry(scan, &connector->probed_modes, head) {
drivers/gpu/drm/gma500/psb_device.c
124
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/gma500/psb_device.c
170
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head)
drivers/gpu/drm/gma500/psb_intel_display.c
533
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/gma500/psb_intel_lvds.c
388
head) {
drivers/gpu/drm/gma500/psb_intel_lvds.c
734
list_for_each_entry(scan, &connector->probed_modes, head) {
drivers/gpu/drm/gma500/psb_intel_sdvo.c
1537
list_for_each_entry(newmode, &connector->probed_modes, head) {
drivers/gpu/drm/i915/display/intel_display.h
226
base.head)
drivers/gpu/drm/i915/display/intel_display.h
231
base.head) \
drivers/gpu/drm/i915/display/intel_display.h
238
base.head) \
drivers/gpu/drm/i915/display/intel_display.h
244
base.head)
drivers/gpu/drm/i915/display/intel_display.h
249
base.head) \
drivers/gpu/drm/i915/display/intel_display.h
255
base.head) \
drivers/gpu/drm/i915/display/intel_display.h
261
base.head)
drivers/gpu/drm/i915/display/intel_display.h
266
base.head) \
drivers/gpu/drm/i915/display/intel_display.h
271
list_for_each_entry((intel_encoder), &(dev)->mode_config.encoder_list, base.head) \
drivers/gpu/drm/i915/display/intel_display.h
287
list_for_each_entry((intel_encoder), &(dev)->mode_config.encoder_list, base.head) \
drivers/gpu/drm/i915/display/intel_display_debugfs.c
210
list_for_each_entry(fixed_mode, &connector->panel.fixed_modes, head)
drivers/gpu/drm/i915/display/intel_display_debugfs.c
292
list_for_each_entry(mode, &connector->modes, head)
drivers/gpu/drm/i915/display/intel_dp.c
6874
list_for_each_entry(fixed_mode, &connector->panel.fixed_modes, head)
drivers/gpu/drm/i915/display/intel_dsb.c
555
unsigned int head, tail;
drivers/gpu/drm/i915/display/intel_dsb.c
566
head = intel_dsb_head(sub_dsb);
drivers/gpu/drm/i915/display/intel_dsb.c
585
head_tail = ((u64)(DSB_GOSUB_CONVERT_ADDR(head)) << DSB_GOSUB_HEAD_SHIFT) |
drivers/gpu/drm/i915/display/intel_flipq.c
333
int head, size = intel_flipq_size_entries(flipq_id);
drivers/gpu/drm/i915/display/intel_flipq.c
335
head = intel_flipq_current_head(crtc, flipq_id);
drivers/gpu/drm/i915/display/intel_flipq.c
338
(flipq->tail + size - head) % size >= size - 1,
drivers/gpu/drm/i915/display/intel_flipq.c
341
head, flipq->tail, size);
drivers/gpu/drm/i915/display/intel_global_state.c
124
list_add_tail(&obj->head, &display->global.obj_list);
drivers/gpu/drm/i915/display/intel_global_state.c
131
list_for_each_entry_safe(obj, next, &display->global.obj_list, head) {
drivers/gpu/drm/i915/display/intel_global_state.c
132
list_del(&obj->head);
drivers/gpu/drm/i915/display/intel_global_state.c
152
list_for_each_entry(l, &ctx->locked, head) {
drivers/gpu/drm/i915/display/intel_global_state.h
25
struct list_head head;
drivers/gpu/drm/i915/display/intel_panel.c
133
list_for_each_entry(fixed_mode, &connector->panel.fixed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
153
list_for_each_entry(fixed_mode, &connector->panel.fixed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
166
list_for_each_entry(fixed_mode, &connector->panel.fixed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
183
list_for_each_entry(mode1, &connector->panel.fixed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
186
list_for_each_entry_continue(mode2, &connector->panel.fixed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
257
list_for_each_entry_safe(mode, next, &connector->base.probed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
266
list_move_tail(&mode->head, &connector->panel.fixed_modes);
drivers/gpu/drm/i915/display/intel_panel.c
279
list_for_each_entry(scan, &connector->base.probed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
288
typeof(*fixed_mode), head);
drivers/gpu/drm/i915/display/intel_panel.c
298
list_move_tail(&fixed_mode->head, &connector->panel.fixed_modes);
drivers/gpu/drm/i915/display/intel_panel.c
306
list_for_each_entry_safe(mode, next, &connector->base.probed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
311
list_del(&mode->head);
drivers/gpu/drm/i915/display/intel_panel.c
344
list_add_tail(&fixed_mode->head, &connector->panel.fixed_modes);
drivers/gpu/drm/i915/display/intel_panel.c
460
list_for_each_entry_safe(fixed_mode, next, &panel->fixed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
461
list_del(&fixed_mode->head);
drivers/gpu/drm/i915/display/intel_panel.c
59
struct drm_display_mode, head);
drivers/gpu/drm/i915/display/intel_panel.c
92
list_for_each_entry(fixed_mode, &connector->panel.fixed_modes, head) {
drivers/gpu/drm/i915/display/intel_sdvo.c
3059
&display->drm->mode_config.connector_list, head) {
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1058
struct hlist_head *head;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1061
head = &eb->buckets[hash_32(handle, eb->lut_size)];
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1062
hlist_for_each_entry(ev, head, node) {
drivers/gpu/drm/i915/gem/i915_gem_object.c
296
void __i915_gem_free_object_rcu(struct rcu_head *head)
drivers/gpu/drm/i915/gem/i915_gem_object.c
299
container_of(head, typeof(*obj), rcu);
drivers/gpu/drm/i915/gem/i915_gem_object.h
818
void __i915_gem_free_object_rcu(struct rcu_head *head);
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
510
struct list_head *head)
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
526
list_add_tail(&obj->mm.link, head);
drivers/gpu/drm/i915/gt/gen8_engine_cs.c
607
GEM_BUG_ON(intel_ring_direction(ring, rq->wa_tail, rq->head) <= 0);
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
166
slist_add(struct llist_node *node, struct llist_node *head)
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
168
node->next = head;
drivers/gpu/drm/i915/gt/intel_context.c
280
ce->ring->head, ce->ring->tail);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2224
rq->head, rq->postfix, rq->tail,
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2228
size = rq->tail - rq->head;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2229
if (rq->tail < rq->head)
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2235
unsigned int head = rq->head;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2238
if (rq->tail < head) {
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2239
len = rq->ring->size - head;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2240
memcpy(ring, vaddr + head, len);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2241
head = 0;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2243
memcpy(ring + len, vaddr + head, size - len);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2294
rq->ring->head);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1820
u8 head, tail;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1840
head = execlists->csb_head;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1842
if (unlikely(head == tail))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1862
ENGINE_TRACE(engine, "cs-irq head=%d, tail=%d\n", head, tail);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1882
if (++head == num_entries)
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1883
head = 0;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1903
csb = csb_read(engine, buf + head);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1905
head, upper_32_bits(csb), lower_32_bits(csb));
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1981
rq->head, rq->tail,
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1997
} while (head != tail);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3028
u32 head;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3044
head = intel_ring_wrap(ce->ring, rq->tail);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3055
head = intel_ring_wrap(ce->ring, rq->head);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3056
GEM_BUG_ON(head == ce->ring->tail);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3096
head, ce->ring->tail);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3098
ce->lrc.lrca = lrc_update_regs(ce, engine, head);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
432
u32 head;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
454
head = rq->tail;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
456
head = __active_request(ce->timeline, rq, -EIO)->head;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
457
head = intel_ring_wrap(ce->ring, head);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
463
ce->lrc.lrca = lrc_update_regs(ce, engine, head);
drivers/gpu/drm/i915/gt/intel_lrc.c
1539
u32 head)
drivers/gpu/drm/i915/gt/intel_lrc.c
1544
GEM_BUG_ON(!intel_ring_offset_valid(ring, head));
drivers/gpu/drm/i915/gt/intel_lrc.c
1548
regs[CTX_RING_HEAD] = head;
drivers/gpu/drm/i915/gt/intel_lrc.h
66
u32 head);
drivers/gpu/drm/i915/gt/intel_ring.c
23
space = __intel_ring_space(ring->head, ring->emit, ring->size);
drivers/gpu/drm/i915/gt/intel_ring.c
89
ring->head = tail;
drivers/gpu/drm/i915/gt/intel_ring.h
111
GEM_BUG_ON(cacheline(tail) == cacheline(head) && tail < head);
drivers/gpu/drm/i915/gt/intel_ring.h
130
__intel_ring_space(unsigned int head, unsigned int tail, unsigned int size)
drivers/gpu/drm/i915/gt/intel_ring.h
138
return (head - tail - CACHELINE_BYTES) & (size - 1);
drivers/gpu/drm/i915/gt/intel_ring.h
92
unsigned int head = READ_ONCE(ring->head);
drivers/gpu/drm/i915/gt/intel_ring_submission.c
200
ring->head, ring->tail);
drivers/gpu/drm/i915/gt/intel_ring_submission.c
229
GEM_BUG_ON(!intel_ring_offset_valid(ring, ring->head));
drivers/gpu/drm/i915/gt/intel_ring_submission.c
249
ENGINE_WRITE_FW(engine, RING_HEAD, ring->head);
drivers/gpu/drm/i915/gt/intel_ring_submission.c
250
if (ENGINE_READ_FW(engine, RING_HEAD) == ring->head)
drivers/gpu/drm/i915/gt/intel_ring_submission.c
254
ENGINE_WRITE_FW(engine, RING_TAIL, ring->head);
drivers/gpu/drm/i915/gt/intel_ring_submission.c
259
ring->head);
drivers/gpu/drm/i915/gt/intel_ring_submission.c
282
if (ring->tail != ring->head) {
drivers/gpu/drm/i915/gt/intel_ring_submission.c
297
ENGINE_READ(engine, RING_HEAD), ring->head,
drivers/gpu/drm/i915/gt/intel_ring_submission.c
392
u32 head;
drivers/gpu/drm/i915/gt/intel_ring_submission.c
446
head = rq->head;
drivers/gpu/drm/i915/gt/intel_ring_submission.c
448
head = engine->legacy.ring->tail;
drivers/gpu/drm/i915/gt/intel_ring_submission.c
450
engine->legacy.ring->head = intel_ring_wrap(engine->legacy.ring, head);
drivers/gpu/drm/i915/gt/intel_ring_types.h
41
u32 head; /* updated during retire, loosely tracks RING_HEAD */
drivers/gpu/drm/i915/gt/selftest_execlists.c
229
lrc_update_regs(ce[1], engine, ce[1]->ring->head);
drivers/gpu/drm/i915/gt/selftest_execlists.c
2594
ring_size = rq->wa_tail - rq->head;
drivers/gpu/drm/i915/gt/selftest_execlists.c
530
GEM_BUG_ON(ring->head);
drivers/gpu/drm/i915/gt/selftest_execlists.c
543
GEM_BUG_ON(!rq->head);
drivers/gpu/drm/i915/gt/selftest_execlists.c
936
struct i915_request *head;
drivers/gpu/drm/i915/gt/selftest_execlists.c
940
head = semaphore_queue(outer, vma, n++);
drivers/gpu/drm/i915/gt/selftest_execlists.c
941
if (IS_ERR(head))
drivers/gpu/drm/i915/gt/selftest_execlists.c
942
return PTR_ERR(head);
drivers/gpu/drm/i915/gt/selftest_execlists.c
965
if (i915_request_wait(head, 0,
drivers/gpu/drm/i915/gt/selftest_execlists.c
975
i915_request_put(head);
drivers/gpu/drm/i915/gt/uc/abi/guc_communication_ctb_abi.h
49
u32 head;
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
1374
ct->ctbs.send.desc->head);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
1380
ct->ctbs.recv.desc->head);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
167
ctb->head = 0;
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
168
space = CIRC_SPACE(ctb->tail, ctb->head, ctb->size) - ctb->resv_space;
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
455
if (unlikely(READ_ONCE(desc->head) >= size)) {
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
457
desc->head, size);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
516
desc->head, desc->tail, desc->status);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
581
CT_ERROR(ct, "Head: %u (Dwords)\n", ct->ctbs.send.desc->head);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
585
CT_ERROR(ct, "Head: %u\n (Dwords)", ct->ctbs.recv.desc->head);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
625
u32 head;
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
631
head = READ_ONCE(desc->head);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
632
if (unlikely(head > ctb->size)) {
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
634
head, ctb->size);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
641
space = CIRC_SPACE(ctb->tail, head, ctb->size);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
888
u32 head = ctb->head;
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
917
GEM_BUG_ON(head > size);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
920
if (unlikely(head != READ_ONCE(desc->head))) {
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
922
desc->head, head);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
935
available = tail - head;
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
944
CT_DEBUG(ct, "available %d (%u:%u:%u)\n", available, head, tail, size);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
947
header = cmds[head];
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
948
head = (head + 1) % size;
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
955
4 * (head + available - 1 > size ?
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
956
size - head : available - 1), &cmds[head],
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
957
4 * (head + available - 1 > size ?
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
958
available - 1 - size + head : 0), &cmds[0]);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
967
4 * (head + available - 1 > size ?
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
968
size - head : available - 1), &cmds[head],
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
969
4 * (head + available - 1 > size ?
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
970
available - 1 - size + head : 0), &cmds[0]);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
977
(*msg)->msg[i] = cmds[head];
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
978
head = (head + 1) % size;
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
983
ctb->head = head;
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
986
WRITE_ONCE(desc->head, head);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
994
desc->head, desc->tail, desc->status);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.h
52
u32 head;
drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h
210
u32 head;
drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h
222
u32 head;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1777
static void guc_reset_state(struct intel_context *ce, u32 head, bool scrub)
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1798
lrc_update_regs(ce, engine, head);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1862
u32 head;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1890
head = ce->ring->tail;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1898
head = intel_ring_wrap(ce->ring, rq->head);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1903
guc_reset_state(ce, head, guilty);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2844
ce->parallel.guc.wq_head = &pdesc->head;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2918
ce->parallel.guc.wq_head = &wq_desc->head;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3224
guc_reset_state(ce, intel_ring_wrap(ce->ring, rq->head),
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5546
ce->ring->head,
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5554
ce->ring->head);
drivers/gpu/drm/i915/gvt/sched_policy.c
181
struct list_head *head = &sched_data->lru_runq_head;
drivers/gpu/drm/i915/gvt/sched_policy.c
185
list_for_each(pos, head) {
drivers/gpu/drm/i915/gvt/scheduler.c
1645
u32 head, tail, start, ctl, ctx_ctl, per_ctx, indirect_ctx;
drivers/gpu/drm/i915/gvt/scheduler.c
1657
RING_CTX_OFF(ring_header.val), &head, 4);
drivers/gpu/drm/i915/gvt/scheduler.c
1662
guest_head = head;
drivers/gpu/drm/i915/gvt/scheduler.c
1664
head &= RB_HEAD_OFF_MASK;
drivers/gpu/drm/i915/gvt/scheduler.c
1672
gvt_dbg_el("ctx head %x real head %lx\n", head,
drivers/gpu/drm/i915/gvt/scheduler.c
1678
head = last_workload->rb_tail;
drivers/gpu/drm/i915/gvt/scheduler.c
1706
workload->rb_head = head;
drivers/gpu/drm/i915/gvt/scheduler.c
1751
workload, engine->name, head, tail, start, ctl);
drivers/gpu/drm/i915/gvt/scheduler.c
958
u32 head, tail;
drivers/gpu/drm/i915/gvt/scheduler.c
966
head = workload->rb_head;
drivers/gpu/drm/i915/gvt/scheduler.c
970
if (tail < head) {
drivers/gpu/drm/i915/gvt/scheduler.c
977
head = (wrap_count << RB_HEAD_WRAP_CNT_OFF) | tail;
drivers/gpu/drm/i915/gvt/scheduler.c
981
vgpu_vreg_t(vgpu, RING_HEAD(ring_base)) = head;
drivers/gpu/drm/i915/i915_active.c
369
struct llist_node *head = NULL, *tail = NULL;
drivers/gpu/drm/i915/i915_active.c
395
pos->next = head;
drivers/gpu/drm/i915/i915_active.c
396
head = pos;
drivers/gpu/drm/i915/i915_active.c
400
if (head)
drivers/gpu/drm/i915/i915_active.c
401
llist_add_batch(head, tail, &engine->barrier_tasks);
drivers/gpu/drm/i915/i915_gpu_error.c
1350
ee->head = ENGINE_READ(engine, RING_HEAD);
drivers/gpu/drm/i915/i915_gpu_error.c
1423
erq->head = request->head;
drivers/gpu/drm/i915/i915_gpu_error.c
1656
ee->rq_head = rq->head;
drivers/gpu/drm/i915/i915_gpu_error.c
1785
saved->raw_head = ctb->desc->head;
drivers/gpu/drm/i915/i915_gpu_error.c
1787
saved->head = ctb->head;
drivers/gpu/drm/i915/i915_gpu_error.c
498
erq->head, erq->tail);
drivers/gpu/drm/i915/i915_gpu_error.c
539
err_printf(m, " HEAD: 0x%08x [0x%08x]\n", ee->head, ee->rq_head);
drivers/gpu/drm/i915/i915_gpu_error.c
685
ctb->head, ctb->tail, ctb->desc_offset, ctb->cmds_offset, ctb->size);
drivers/gpu/drm/i915/i915_gpu_error.h
131
u32 raw_head, head;
drivers/gpu/drm/i915/i915_gpu_error.h
52
u32 head;
drivers/gpu/drm/i915/i915_gpu_error.h
73
u32 head;
drivers/gpu/drm/i915/i915_list_util.h
10
static inline void __list_del_many(struct list_head *head,
drivers/gpu/drm/i915/i915_list_util.h
13
first->prev = head;
drivers/gpu/drm/i915/i915_list_util.h
14
WRITE_ONCE(head->next, first);
drivers/gpu/drm/i915/i915_list_util.h
18
const struct list_head *head)
drivers/gpu/drm/i915/i915_list_util.h
20
return READ_ONCE(list->next) == head;
drivers/gpu/drm/i915/i915_perf.c
1050
u32 head, tail;
drivers/gpu/drm/i915/i915_perf.c
1058
head = stream->oa_buffer.head;
drivers/gpu/drm/i915/i915_perf.c
1070
head > OA_BUFFER_SIZE || head % report_size ||
drivers/gpu/drm/i915/i915_perf.c
1073
head, tail))
drivers/gpu/drm/i915/i915_perf.c
1078
OA_TAKEN(tail, head);
drivers/gpu/drm/i915/i915_perf.c
1079
head = (head + report_size) & mask) {
drivers/gpu/drm/i915/i915_perf.c
1080
u8 *report = oa_buf_base + head;
drivers/gpu/drm/i915/i915_perf.c
1092
(OA_BUFFER_SIZE - head) < report_size)) {
drivers/gpu/drm/i915/i915_perf.c
1126
((head + gtt_offset) & GEN7_OASTATUS2_HEAD_MASK) |
drivers/gpu/drm/i915/i915_perf.c
1128
stream->oa_buffer.head = head;
drivers/gpu/drm/i915/i915_perf.c
1713
stream->oa_buffer.head = 0;
drivers/gpu/drm/i915/i915_perf.c
1755
stream->oa_buffer.head = 0;
drivers/gpu/drm/i915/i915_perf.c
1809
stream->oa_buffer.head = 0;
drivers/gpu/drm/i915/i915_perf.c
230
#define OA_TAKEN(tail, head) ((tail - head) & (OA_BUFFER_SIZE - 1))
drivers/gpu/drm/i915/i915_perf.c
601
stream->oa_buffer.head, tail, hw_tail);
drivers/gpu/drm/i915/i915_perf.c
606
stream->oa_buffer.head) >= report_size;
drivers/gpu/drm/i915/i915_perf.c
740
u32 head, tail;
drivers/gpu/drm/i915/i915_perf.c
748
head = stream->oa_buffer.head;
drivers/gpu/drm/i915/i915_perf.c
760
head > OA_BUFFER_SIZE ||
drivers/gpu/drm/i915/i915_perf.c
763
head, tail))
drivers/gpu/drm/i915/i915_perf.c
768
OA_TAKEN(tail, head);
drivers/gpu/drm/i915/i915_perf.c
769
head = (head + report_size) & mask) {
drivers/gpu/drm/i915/i915_perf.c
770
u8 *report = oa_buf_base + head;
drivers/gpu/drm/i915/i915_perf.c
918
(head + gtt_offset) & GEN12_OAG_OAHEADPTR_MASK);
drivers/gpu/drm/i915/i915_perf.c
919
stream->oa_buffer.head = head;
drivers/gpu/drm/i915/i915_perf_types.h
323
u32 head;
drivers/gpu/drm/i915/i915_pmu.c
396
u32 tail, head, acthd;
drivers/gpu/drm/i915/i915_pmu.c
399
head = ENGINE_READ_FW(engine, RING_HEAD);
drivers/gpu/drm/i915/i915_pmu.c
402
if (head & HEAD_WAIT_I8XX)
drivers/gpu/drm/i915/i915_pmu.c
405
if (head & HEAD_WAIT_I8XX || head != acthd ||
drivers/gpu/drm/i915/i915_pmu.c
406
(head & HEAD_ADDR) != (tail & TAIL_ADDR))
drivers/gpu/drm/i915/i915_request.c
1002
rq->head = rq->ring->emit;
drivers/gpu/drm/i915/i915_request.c
1016
ce->ring->emit = rq->head;
drivers/gpu/drm/i915/i915_request.c
225
u32 head;
drivers/gpu/drm/i915/i915_request.c
227
head = rq->infix;
drivers/gpu/drm/i915/i915_request.c
228
if (rq->postfix < head) {
drivers/gpu/drm/i915/i915_request.c
229
memset(vaddr + head, val, rq->ring->size - head);
drivers/gpu/drm/i915/i915_request.c
230
head = 0;
drivers/gpu/drm/i915/i915_request.c
232
memset(vaddr + head, val, rq->postfix - head);
drivers/gpu/drm/i915/i915_request.c
385
rq->ring->head = rq->postfix;
drivers/gpu/drm/i915/i915_request.h
282
u32 head;
drivers/gpu/drm/i915/i915_sw_fence.c
160
list_for_each_entry_safe(pos, next, &x->head, entry) {
drivers/gpu/drm/i915/i915_sw_fence.c
170
list_for_each_entry_safe(pos, next, &x->head, entry) {
drivers/gpu/drm/i915/i915_sw_fence.c
183
list_splice_tail_init(&extra, &x->head);
drivers/gpu/drm/i915/i915_sw_fence.c
260
I915_SW_FENCE_BUG_ON(!list_empty(&fence->wait.head));
drivers/gpu/drm/i915/i915_sw_fence.c
293
list_for_each_entry(wq, &fence->wait.head, entry) {
drivers/gpu/drm/i915/i915_sw_fence.c
311
list_for_each_entry(wq, &fence->wait.head, entry) {
drivers/gpu/drm/i915/i915_vma_resource.c
79
static void unbind_fence_free_rcu(struct rcu_head *head)
drivers/gpu/drm/i915/i915_vma_resource.c
82
container_of(head, typeof(*vma_res), unbind_fence.rcu);
drivers/gpu/drm/i915/selftests/i915_request.c
1715
sz = rq->ring->emit - rq->head;
drivers/gpu/drm/imx/ipuv3/ipuv3-crtc.c
276
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
1650
list_for_each_entry(enc, &crtc->dev->mode_config.encoder_list, head) {
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
514
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
152
list_for_each_entry(crtc, &config->crtc_list, head) {
drivers/gpu/drm/msm/disp/mdp4/mdp4_lcdc_encoder.c
38
list_for_each_entry(connector, &dev->mode_config.connector_list, head)
drivers/gpu/drm/msm/disp/mdp5/mdp5_encoder.c
52
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/msm/msm_debugfs.c
134
list_for_each_entry(fb, &dev->mode_config.fb_list, head) {
drivers/gpu/drm/msm/msm_rd.c
103
char *fptr = &fifo->buf[fifo->head];
drivers/gpu/drm/msm/msm_rd.c
117
smp_store_release(&fifo->head, (fifo->head + n) & (BUF_SZ - 1));
drivers/gpu/drm/msm/msm_rd.c
195
rd->fifo.head = rd->fifo.tail = 0;
drivers/gpu/drm/msm/msm_rd.c
72
(CIRC_CNT((circ)->head, (circ)->tail, BUF_SZ))
drivers/gpu/drm/msm/msm_rd.c
74
(CIRC_CNT_TO_END((circ)->head, (circ)->tail, BUF_SZ))
drivers/gpu/drm/msm/msm_rd.c
77
(CIRC_SPACE((circ)->head, (circ)->tail, BUF_SZ))
drivers/gpu/drm/msm/msm_rd.c
79
(CIRC_SPACE_TO_END((circ)->head, (circ)->tail, BUF_SZ))
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1031
struct list_head head;
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1056
s = list_first_entry(&fctx->flip, struct nv04_page_flip_state, head);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1064
list_del(&s->head);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1107
list_add_tail(&s->head, &fctx->flip);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1130
list_del(&s->head);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1152
int head = nouveau_crtc(crtc)->index;
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1206
PUSH_NVSQ(push, NV05F, 0x0134, head);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1211
if (dispnv04->image[head])
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1212
drm_gem_object_put(&dispnv04->image[head]->bo.base);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1215
dispnv04->image[head] = new_bo;
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1328
ret = nvif_head_ctor(&disp->disp, nv_crtc->base.name, nv_crtc->index, &nv_crtc->head);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1332
return nvif_head_vblank_event_ctor(&nv_crtc->head, "kmsVbl", nv04_crtc_vblank_handler,
drivers/gpu/drm/nouveau/dispnv04/crtc.c
261
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/nouveau/dispnv04/crtc.c
473
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/nouveau/dispnv04/crtc.c
696
int head = nv_crtc->index;
drivers/gpu/drm/nouveau/dispnv04/crtc.c
697
uint8_t saved_cr21 = nv04_display(dev)->saved_reg.crtc_reg[head].CRTC[NV_CIO_CRE_21];
drivers/gpu/drm/nouveau/dispnv04/crtc.c
700
NVSetOwner(crtc->dev, head);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
702
nouveau_hw_load_state(crtc->dev, head, &nv04_display(dev)->saved_reg);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
703
nv_lock_vga_crtc_shadow(crtc->dev, head, saved_cr21);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
773
nvif_head_dtor(&nv_crtc->head);
drivers/gpu/drm/nouveau/dispnv04/dac.c
245
int head;
drivers/gpu/drm/nouveau/dispnv04/dac.c
282
head = (saved_routput & 0x100) >> 8;
drivers/gpu/drm/nouveau/dispnv04/dac.c
285
if (!(NVReadVgaCrtc(dev, head, NV_CIO_CRE_RPC1_INDEX) & 0xC0))
drivers/gpu/drm/nouveau/dispnv04/dac.c
286
head ^= 1;
drivers/gpu/drm/nouveau/dispnv04/dac.c
289
routput = (saved_routput & 0xfffffece) | head << 8;
drivers/gpu/drm/nouveau/dispnv04/dac.c
304
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TESTPOINT_DATA,
drivers/gpu/drm/nouveau/dispnv04/dac.c
306
temp = NVReadRAMDAC(dev, head, NV_PRAMDAC_TEST_CONTROL);
drivers/gpu/drm/nouveau/dispnv04/dac.c
307
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TEST_CONTROL,
drivers/gpu/drm/nouveau/dispnv04/dac.c
315
temp = NVReadRAMDAC(dev, head, NV_PRAMDAC_TEST_CONTROL);
drivers/gpu/drm/nouveau/dispnv04/dac.c
316
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TEST_CONTROL,
drivers/gpu/drm/nouveau/dispnv04/dac.c
318
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TESTPOINT_DATA, 0);
drivers/gpu/drm/nouveau/dispnv04/dac.c
368
int head = nouveau_crtc(encoder->crtc)->index;
drivers/gpu/drm/nouveau/dispnv04/dac.c
372
nv04_dfp_disable(dev, head);
drivers/gpu/drm/nouveau/dispnv04/dac.c
381
int head = nouveau_crtc(encoder->crtc)->index;
drivers/gpu/drm/nouveau/dispnv04/dac.c
391
head << 8 | NV_PRAMDAC_DACCLK_SEL_DACCLK);
drivers/gpu/drm/nouveau/dispnv04/dac.c
393
list_for_each_entry(rebind, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/nouveau/dispnv04/dac.c
401
(otherdac & ~0x0100) | (head ^ 1) << 8);
drivers/gpu/drm/nouveau/dispnv04/dfp.c
103
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL,
drivers/gpu/drm/nouveau/dispnv04/dfp.c
108
crtcstate[head].fp_control = FP_TG_CONTROL_OFF;
drivers/gpu/drm/nouveau/dispnv04/dfp.c
109
crtcstate[head].CRTC[NV_CIO_CRE_LCD__INDEX] &=
drivers/gpu/drm/nouveau/dispnv04/dfp.c
135
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/nouveau/dispnv04/dfp.c
171
list_for_each_entry(slave, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/nouveau/dispnv04/dfp.c
205
struct nouveau_encoder *nv_encoder, int head)
drivers/gpu/drm/nouveau/dispnv04/dfp.c
217
if (head)
drivers/gpu/drm/nouveau/dispnv04/dfp.c
241
state->sel_clk |= (head ? 0x40 : 0x10) << shift;
drivers/gpu/drm/nouveau/dispnv04/dfp.c
250
int head = nouveau_crtc(encoder->crtc)->index;
drivers/gpu/drm/nouveau/dispnv04/dfp.c
252
uint8_t *cr_lcd = &crtcstate[head].CRTC[NV_CIO_CRE_LCD__INDEX];
drivers/gpu/drm/nouveau/dispnv04/dfp.c
253
uint8_t *cr_lcd_oth = &crtcstate[head ^ 1].CRTC[NV_CIO_CRE_LCD__INDEX];
drivers/gpu/drm/nouveau/dispnv04/dfp.c
257
nv04_dfp_prepare_sel_clk(dev, nv_encoder, head);
drivers/gpu/drm/nouveau/dispnv04/dfp.c
263
*cr_lcd |= head ? 0x0 : 0x8;
drivers/gpu/drm/nouveau/dispnv04/dfp.c
271
NVWriteVgaCrtc(dev, head ^ 1,
drivers/gpu/drm/nouveau/dispnv04/dfp.c
454
int head = nouveau_crtc(encoder->crtc)->index;
drivers/gpu/drm/nouveau/dispnv04/dfp.c
458
run_tmds_table(dev, dcbe, head, nv_encoder->mode.clock);
drivers/gpu/drm/nouveau/dispnv04/dfp.c
460
call_lvds_script(dev, dcbe, head, LVDS_RESET, nv_encoder->mode.clock);
drivers/gpu/drm/nouveau/dispnv04/dfp.c
464
nv04_display(dev)->mode_reg.crtc_reg[head].fp_control =
drivers/gpu/drm/nouveau/dispnv04/dfp.c
465
NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL);
drivers/gpu/drm/nouveau/dispnv04/dfp.c
537
int head = crtc ? nouveau_crtc(crtc)->index :
drivers/gpu/drm/nouveau/dispnv04/dfp.c
541
call_lvds_script(dev, nv_encoder->dcb, head,
drivers/gpu/drm/nouveau/dispnv04/dfp.c
547
call_lvds_script(dev, nv_encoder->dcb, head,
drivers/gpu/drm/nouveau/dispnv04/dfp.c
585
nv_encoder->restore.head =
drivers/gpu/drm/nouveau/dispnv04/dfp.c
593
int head = nv_encoder->restore.head;
drivers/gpu/drm/nouveau/dispnv04/dfp.c
600
call_lvds_script(dev, nv_encoder->dcb, head,
drivers/gpu/drm/nouveau/dispnv04/dfp.c
606
(&nv04_display(dev)->saved_reg.crtc_reg[head].pllvals);
drivers/gpu/drm/nouveau/dispnv04/dfp.c
608
run_tmds_table(dev, nv_encoder->dcb, head, clock);
drivers/gpu/drm/nouveau/dispnv04/dfp.c
69
int head, bool dl)
drivers/gpu/drm/nouveau/dispnv04/dfp.c
81
if (head != ramdac)
drivers/gpu/drm/nouveau/dispnv04/dfp.c
93
void nv04_dfp_disable(struct drm_device *dev, int head)
drivers/gpu/drm/nouveau/dispnv04/dfp.c
97
if (NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL) &
drivers/gpu/drm/nouveau/dispnv04/disp.c
115
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/nouveau/dispnv04/disp.c
120
list_for_each_entry(encoder, &dev->mode_config.encoder_list, base.base.head)
drivers/gpu/drm/nouveau/dispnv04/disp.c
130
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/nouveau/dispnv04/disp.c
142
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/nouveau/dispnv04/disp.c
156
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/nouveau/dispnv04/disp.c
172
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/nouveau/dispnv04/disp.c
196
list_for_each_entry(encoder, &dev->mode_config.encoder_list, base.base.head)
drivers/gpu/drm/nouveau/dispnv04/disp.c
199
list_for_each_entry(nv_crtc, &dev->mode_config.crtc_list, base.head)
drivers/gpu/drm/nouveau/dispnv04/disp.c
282
&dev->mode_config.connector_list, head) {
drivers/gpu/drm/nouveau/dispnv04/disp.c
290
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/nouveau/dispnv04/disp.c
298
list_for_each_entry(crtc, &dev->mode_config.crtc_list, base.head)
drivers/gpu/drm/nouveau/dispnv04/disp.c
301
list_for_each_entry(nv_encoder, &dev->mode_config.encoder_list, base.base.head)
drivers/gpu/drm/nouveau/dispnv04/disp.c
78
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/nouveau/dispnv04/disp.c
88
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/nouveau/dispnv04/disp.h
118
int head, bool dl);
drivers/gpu/drm/nouveau/dispnv04/disp.h
119
void nv04_dfp_disable(struct drm_device *dev, int head);
drivers/gpu/drm/nouveau/dispnv04/disp.h
181
init.head = crtc;
drivers/gpu/drm/nouveau/dispnv04/disp.h
74
int head;
drivers/gpu/drm/nouveau/dispnv04/hw.c
110
NVBlankScreen(struct drm_device *dev, int head, bool blank)
drivers/gpu/drm/nouveau/dispnv04/hw.c
115
NVSetOwner(dev, head);
drivers/gpu/drm/nouveau/dispnv04/hw.c
117
seq1 = NVReadVgaSeq(dev, head, NV_VIO_SR_CLOCK_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
119
NVVgaSeqReset(dev, head, true);
drivers/gpu/drm/nouveau/dispnv04/hw.c
121
NVWriteVgaSeq(dev, head, NV_VIO_SR_CLOCK_INDEX, seq1 | 0x20);
drivers/gpu/drm/nouveau/dispnv04/hw.c
123
NVWriteVgaSeq(dev, head, NV_VIO_SR_CLOCK_INDEX, seq1 & ~0x20);
drivers/gpu/drm/nouveau/dispnv04/hw.c
124
NVVgaSeqReset(dev, head, false);
drivers/gpu/drm/nouveau/dispnv04/hw.c
252
nouveau_hw_fix_bad_vpll(struct drm_device *dev, int head)
drivers/gpu/drm/nouveau/dispnv04/hw.c
265
enum nvbios_pll_type pll = head ? PLL_VPLL1 : PLL_VPLL0;
drivers/gpu/drm/nouveau/dispnv04/hw.c
276
NV_WARN(drm, "VPLL %d outwith limits, attempting to fix\n", head + 1);
drivers/gpu/drm/nouveau/dispnv04/hw.c
379
rd_cio_state(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
382
crtcstate->CRTC[index] = NVReadVgaCrtc(dev, head, index);
drivers/gpu/drm/nouveau/dispnv04/hw.c
386
wr_cio_state(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
389
NVWriteVgaCrtc(dev, head, index, crtcstate->CRTC[index]);
drivers/gpu/drm/nouveau/dispnv04/hw.c
39
NVWriteVgaSeq(struct drm_device *dev, int head, uint8_t index, uint8_t value)
drivers/gpu/drm/nouveau/dispnv04/hw.c
393
nv_save_state_ramdac(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
397
struct nv04_crtc_reg *regp = &state->crtc_reg[head];
drivers/gpu/drm/nouveau/dispnv04/hw.c
401
regp->nv10_cursync = NVReadRAMDAC(dev, head, NV_RAMDAC_NV10_CURSYNC);
drivers/gpu/drm/nouveau/dispnv04/hw.c
403
nouveau_hw_get_pllvals(dev, head ? PLL_VPLL1 : PLL_VPLL0, &regp->pllvals);
drivers/gpu/drm/nouveau/dispnv04/hw.c
408
regp->dither = NVReadRAMDAC(dev, head, NV_RAMDAC_DITHER_NV11);
drivers/gpu/drm/nouveau/dispnv04/hw.c
41
NVWritePRMVIO(dev, head, NV_PRMVIO_SRX, index);
drivers/gpu/drm/nouveau/dispnv04/hw.c
410
regp->ramdac_gen_ctrl = NVReadRAMDAC(dev, head, NV_PRAMDAC_GENERAL_CONTROL);
drivers/gpu/drm/nouveau/dispnv04/hw.c
413
regp->ramdac_630 = NVReadRAMDAC(dev, head, NV_PRAMDAC_630);
drivers/gpu/drm/nouveau/dispnv04/hw.c
415
regp->ramdac_634 = NVReadRAMDAC(dev, head, NV_PRAMDAC_634);
drivers/gpu/drm/nouveau/dispnv04/hw.c
417
regp->tv_setup = NVReadRAMDAC(dev, head, NV_PRAMDAC_TV_SETUP);
drivers/gpu/drm/nouveau/dispnv04/hw.c
418
regp->tv_vtotal = NVReadRAMDAC(dev, head, NV_PRAMDAC_TV_VTOTAL);
drivers/gpu/drm/nouveau/dispnv04/hw.c
419
regp->tv_vskew = NVReadRAMDAC(dev, head, NV_PRAMDAC_TV_VSKEW);
drivers/gpu/drm/nouveau/dispnv04/hw.c
42
NVWritePRMVIO(dev, head, NV_PRMVIO_SR, value);
drivers/gpu/drm/nouveau/dispnv04/hw.c
420
regp->tv_vsync_delay = NVReadRAMDAC(dev, head, NV_PRAMDAC_TV_VSYNC_DELAY);
drivers/gpu/drm/nouveau/dispnv04/hw.c
421
regp->tv_htotal = NVReadRAMDAC(dev, head, NV_PRAMDAC_TV_HTOTAL);
drivers/gpu/drm/nouveau/dispnv04/hw.c
422
regp->tv_hskew = NVReadRAMDAC(dev, head, NV_PRAMDAC_TV_HSKEW);
drivers/gpu/drm/nouveau/dispnv04/hw.c
423
regp->tv_hsync_delay = NVReadRAMDAC(dev, head, NV_PRAMDAC_TV_HSYNC_DELAY);
drivers/gpu/drm/nouveau/dispnv04/hw.c
424
regp->tv_hsync_delay2 = NVReadRAMDAC(dev, head, NV_PRAMDAC_TV_HSYNC_DELAY2);
drivers/gpu/drm/nouveau/dispnv04/hw.c
428
regp->fp_vert_regs[i] = NVReadRAMDAC(dev, head, ramdac_reg);
drivers/gpu/drm/nouveau/dispnv04/hw.c
429
regp->fp_horiz_regs[i] = NVReadRAMDAC(dev, head, ramdac_reg + 0x20);
drivers/gpu/drm/nouveau/dispnv04/hw.c
433
regp->dither = NVReadRAMDAC(dev, head, NV_RAMDAC_FP_DITHER);
drivers/gpu/drm/nouveau/dispnv04/hw.c
435
regp->dither_regs[i] = NVReadRAMDAC(dev, head, NV_PRAMDAC_850 + i * 4);
drivers/gpu/drm/nouveau/dispnv04/hw.c
436
regp->dither_regs[i + 3] = NVReadRAMDAC(dev, head, NV_PRAMDAC_85C + i * 4);
drivers/gpu/drm/nouveau/dispnv04/hw.c
440
regp->fp_control = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL);
drivers/gpu/drm/nouveau/dispnv04/hw.c
441
regp->fp_debug_0 = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_DEBUG_0);
drivers/gpu/drm/nouveau/dispnv04/hw.c
442
if (!nv_gf4_disp_arch(dev) && head == 0) {
drivers/gpu/drm/nouveau/dispnv04/hw.c
448
regp->fp_debug_1 = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_DEBUG_1);
drivers/gpu/drm/nouveau/dispnv04/hw.c
449
regp->fp_debug_2 = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_DEBUG_2);
drivers/gpu/drm/nouveau/dispnv04/hw.c
451
regp->fp_margin_color = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_MARGIN_COLOR);
drivers/gpu/drm/nouveau/dispnv04/hw.c
454
regp->ramdac_8c0 = NVReadRAMDAC(dev, head, NV_PRAMDAC_8C0);
drivers/gpu/drm/nouveau/dispnv04/hw.c
457
regp->ramdac_a20 = NVReadRAMDAC(dev, head, NV_PRAMDAC_A20);
drivers/gpu/drm/nouveau/dispnv04/hw.c
458
regp->ramdac_a24 = NVReadRAMDAC(dev, head, NV_PRAMDAC_A24);
drivers/gpu/drm/nouveau/dispnv04/hw.c
459
regp->ramdac_a34 = NVReadRAMDAC(dev, head, NV_PRAMDAC_A34);
drivers/gpu/drm/nouveau/dispnv04/hw.c
46
NVReadVgaSeq(struct drm_device *dev, int head, uint8_t index)
drivers/gpu/drm/nouveau/dispnv04/hw.c
462
regp->ctv_regs[i] = NVReadRAMDAC(dev, head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
468
nv_load_state_ramdac(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
473
struct nv04_crtc_reg *regp = &state->crtc_reg[head];
drivers/gpu/drm/nouveau/dispnv04/hw.c
474
uint32_t pllreg = head ? NV_RAMDAC_VPLL2 : NV_PRAMDAC_VPLL_COEFF;
drivers/gpu/drm/nouveau/dispnv04/hw.c
478
NVWriteRAMDAC(dev, head, NV_RAMDAC_NV10_CURSYNC, regp->nv10_cursync);
drivers/gpu/drm/nouveau/dispnv04/hw.c
48
NVWritePRMVIO(dev, head, NV_PRMVIO_SRX, index);
drivers/gpu/drm/nouveau/dispnv04/hw.c
485
NVWriteRAMDAC(dev, head, NV_RAMDAC_DITHER_NV11, regp->dither);
drivers/gpu/drm/nouveau/dispnv04/hw.c
487
NVWriteRAMDAC(dev, head, NV_PRAMDAC_GENERAL_CONTROL, regp->ramdac_gen_ctrl);
drivers/gpu/drm/nouveau/dispnv04/hw.c
49
return NVReadPRMVIO(dev, head, NV_PRMVIO_SR);
drivers/gpu/drm/nouveau/dispnv04/hw.c
490
NVWriteRAMDAC(dev, head, NV_PRAMDAC_630, regp->ramdac_630);
drivers/gpu/drm/nouveau/dispnv04/hw.c
492
NVWriteRAMDAC(dev, head, NV_PRAMDAC_634, regp->ramdac_634);
drivers/gpu/drm/nouveau/dispnv04/hw.c
494
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TV_SETUP, regp->tv_setup);
drivers/gpu/drm/nouveau/dispnv04/hw.c
495
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TV_VTOTAL, regp->tv_vtotal);
drivers/gpu/drm/nouveau/dispnv04/hw.c
496
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TV_VSKEW, regp->tv_vskew);
drivers/gpu/drm/nouveau/dispnv04/hw.c
497
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TV_VSYNC_DELAY, regp->tv_vsync_delay);
drivers/gpu/drm/nouveau/dispnv04/hw.c
498
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TV_HTOTAL, regp->tv_htotal);
drivers/gpu/drm/nouveau/dispnv04/hw.c
499
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TV_HSKEW, regp->tv_hskew);
drivers/gpu/drm/nouveau/dispnv04/hw.c
500
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TV_HSYNC_DELAY, regp->tv_hsync_delay);
drivers/gpu/drm/nouveau/dispnv04/hw.c
501
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TV_HSYNC_DELAY2, regp->tv_hsync_delay2);
drivers/gpu/drm/nouveau/dispnv04/hw.c
506
NVWriteRAMDAC(dev, head, ramdac_reg, regp->fp_vert_regs[i]);
drivers/gpu/drm/nouveau/dispnv04/hw.c
507
NVWriteRAMDAC(dev, head, ramdac_reg + 0x20, regp->fp_horiz_regs[i]);
drivers/gpu/drm/nouveau/dispnv04/hw.c
511
NVWriteRAMDAC(dev, head, NV_RAMDAC_FP_DITHER, regp->dither);
drivers/gpu/drm/nouveau/dispnv04/hw.c
513
NVWriteRAMDAC(dev, head, NV_PRAMDAC_850 + i * 4, regp->dither_regs[i]);
drivers/gpu/drm/nouveau/dispnv04/hw.c
514
NVWriteRAMDAC(dev, head, NV_PRAMDAC_85C + i * 4, regp->dither_regs[i + 3]);
drivers/gpu/drm/nouveau/dispnv04/hw.c
518
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL, regp->fp_control);
drivers/gpu/drm/nouveau/dispnv04/hw.c
519
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_DEBUG_0, regp->fp_debug_0);
drivers/gpu/drm/nouveau/dispnv04/hw.c
520
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_DEBUG_1, regp->fp_debug_1);
drivers/gpu/drm/nouveau/dispnv04/hw.c
521
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_DEBUG_2, regp->fp_debug_2);
drivers/gpu/drm/nouveau/dispnv04/hw.c
523
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_MARGIN_COLOR, regp->fp_margin_color);
drivers/gpu/drm/nouveau/dispnv04/hw.c
526
NVWriteRAMDAC(dev, head, NV_PRAMDAC_8C0, regp->ramdac_8c0);
drivers/gpu/drm/nouveau/dispnv04/hw.c
529
NVWriteRAMDAC(dev, head, NV_PRAMDAC_A20, regp->ramdac_a20);
drivers/gpu/drm/nouveau/dispnv04/hw.c
53
NVWriteVgaGr(struct drm_device *dev, int head, uint8_t index, uint8_t value)
drivers/gpu/drm/nouveau/dispnv04/hw.c
530
NVWriteRAMDAC(dev, head, NV_PRAMDAC_A24, regp->ramdac_a24);
drivers/gpu/drm/nouveau/dispnv04/hw.c
531
NVWriteRAMDAC(dev, head, NV_PRAMDAC_A34, regp->ramdac_a34);
drivers/gpu/drm/nouveau/dispnv04/hw.c
534
NVWriteRAMDAC(dev, head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
540
nv_save_state_vga(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
543
struct nv04_crtc_reg *regp = &state->crtc_reg[head];
drivers/gpu/drm/nouveau/dispnv04/hw.c
546
regp->MiscOutReg = NVReadPRMVIO(dev, head, NV_PRMVIO_MISC__READ);
drivers/gpu/drm/nouveau/dispnv04/hw.c
549
rd_cio_state(dev, head, regp, i);
drivers/gpu/drm/nouveau/dispnv04/hw.c
55
NVWritePRMVIO(dev, head, NV_PRMVIO_GRX, index);
drivers/gpu/drm/nouveau/dispnv04/hw.c
551
NVSetEnablePalette(dev, head, true);
drivers/gpu/drm/nouveau/dispnv04/hw.c
553
regp->Attribute[i] = NVReadVgaAttr(dev, head, i);
drivers/gpu/drm/nouveau/dispnv04/hw.c
554
NVSetEnablePalette(dev, head, false);
drivers/gpu/drm/nouveau/dispnv04/hw.c
557
regp->Graphics[i] = NVReadVgaGr(dev, head, i);
drivers/gpu/drm/nouveau/dispnv04/hw.c
56
NVWritePRMVIO(dev, head, NV_PRMVIO_GX, value);
drivers/gpu/drm/nouveau/dispnv04/hw.c
560
regp->Sequencer[i] = NVReadVgaSeq(dev, head, i);
drivers/gpu/drm/nouveau/dispnv04/hw.c
564
nv_load_state_vga(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
567
struct nv04_crtc_reg *regp = &state->crtc_reg[head];
drivers/gpu/drm/nouveau/dispnv04/hw.c
570
NVWritePRMVIO(dev, head, NV_PRMVIO_MISC__WRITE, regp->MiscOutReg);
drivers/gpu/drm/nouveau/dispnv04/hw.c
573
NVWriteVgaSeq(dev, head, i, regp->Sequencer[i]);
drivers/gpu/drm/nouveau/dispnv04/hw.c
575
nv_lock_vga_crtc_base(dev, head, false);
drivers/gpu/drm/nouveau/dispnv04/hw.c
577
wr_cio_state(dev, head, regp, i);
drivers/gpu/drm/nouveau/dispnv04/hw.c
578
nv_lock_vga_crtc_base(dev, head, true);
drivers/gpu/drm/nouveau/dispnv04/hw.c
581
NVWriteVgaGr(dev, head, i, regp->Graphics[i]);
drivers/gpu/drm/nouveau/dispnv04/hw.c
583
NVSetEnablePalette(dev, head, true);
drivers/gpu/drm/nouveau/dispnv04/hw.c
585
NVWriteVgaAttr(dev, head, i, regp->Attribute[i]);
drivers/gpu/drm/nouveau/dispnv04/hw.c
586
NVSetEnablePalette(dev, head, false);
drivers/gpu/drm/nouveau/dispnv04/hw.c
590
nv_save_state_ext(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
594
struct nv04_crtc_reg *regp = &state->crtc_reg[head];
drivers/gpu/drm/nouveau/dispnv04/hw.c
597
rd_cio_state(dev, head, regp, NV_CIO_CRE_LCD__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
598
rd_cio_state(dev, head, regp, NV_CIO_CRE_RPC0_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
599
rd_cio_state(dev, head, regp, NV_CIO_CRE_RPC1_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
60
NVReadVgaGr(struct drm_device *dev, int head, uint8_t index)
drivers/gpu/drm/nouveau/dispnv04/hw.c
600
rd_cio_state(dev, head, regp, NV_CIO_CRE_LSR_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
601
rd_cio_state(dev, head, regp, NV_CIO_CRE_PIXEL_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
602
rd_cio_state(dev, head, regp, NV_CIO_CRE_HEB__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
603
rd_cio_state(dev, head, regp, NV_CIO_CRE_ENH_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
605
rd_cio_state(dev, head, regp, NV_CIO_CRE_FF_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
606
rd_cio_state(dev, head, regp, NV_CIO_CRE_FFLWM__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
607
rd_cio_state(dev, head, regp, NV_CIO_CRE_21);
drivers/gpu/drm/nouveau/dispnv04/hw.c
610
rd_cio_state(dev, head, regp, NV_CIO_CRE_47);
drivers/gpu/drm/nouveau/dispnv04/hw.c
613
rd_cio_state(dev, head, regp, 0x9f);
drivers/gpu/drm/nouveau/dispnv04/hw.c
615
rd_cio_state(dev, head, regp, NV_CIO_CRE_49);
drivers/gpu/drm/nouveau/dispnv04/hw.c
616
rd_cio_state(dev, head, regp, NV_CIO_CRE_HCUR_ADDR0_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
617
rd_cio_state(dev, head, regp, NV_CIO_CRE_HCUR_ADDR1_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
618
rd_cio_state(dev, head, regp, NV_CIO_CRE_HCUR_ADDR2_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
619
rd_cio_state(dev, head, regp, NV_CIO_CRE_ILACE__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
62
NVWritePRMVIO(dev, head, NV_PRMVIO_GRX, index);
drivers/gpu/drm/nouveau/dispnv04/hw.c
622
regp->crtc_830 = NVReadCRTC(dev, head, NV_PCRTC_830);
drivers/gpu/drm/nouveau/dispnv04/hw.c
623
regp->crtc_834 = NVReadCRTC(dev, head, NV_PCRTC_834);
drivers/gpu/drm/nouveau/dispnv04/hw.c
626
regp->gpio_ext = NVReadCRTC(dev, head, NV_PCRTC_GPIO_EXT);
drivers/gpu/drm/nouveau/dispnv04/hw.c
629
regp->crtc_850 = NVReadCRTC(dev, head, NV_PCRTC_850);
drivers/gpu/drm/nouveau/dispnv04/hw.c
63
return NVReadPRMVIO(dev, head, NV_PRMVIO_GX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
632
regp->crtc_eng_ctrl = NVReadCRTC(dev, head, NV_PCRTC_ENGINE_CTRL);
drivers/gpu/drm/nouveau/dispnv04/hw.c
633
regp->cursor_cfg = NVReadCRTC(dev, head, NV_PCRTC_CURSOR_CONFIG);
drivers/gpu/drm/nouveau/dispnv04/hw.c
636
regp->crtc_cfg = NVReadCRTC(dev, head, NV_PCRTC_CONFIG);
drivers/gpu/drm/nouveau/dispnv04/hw.c
638
rd_cio_state(dev, head, regp, NV_CIO_CRE_SCRATCH3__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
639
rd_cio_state(dev, head, regp, NV_CIO_CRE_SCRATCH4__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
641
rd_cio_state(dev, head, regp, NV_CIO_CRE_EBR_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
642
rd_cio_state(dev, head, regp, NV_CIO_CRE_CSB);
drivers/gpu/drm/nouveau/dispnv04/hw.c
643
rd_cio_state(dev, head, regp, NV_CIO_CRE_4B);
drivers/gpu/drm/nouveau/dispnv04/hw.c
644
rd_cio_state(dev, head, regp, NV_CIO_CRE_TVOUT_LATENCY);
drivers/gpu/drm/nouveau/dispnv04/hw.c
648
rd_cio_state(dev, head, regp, NV_CIO_CRE_42);
drivers/gpu/drm/nouveau/dispnv04/hw.c
649
rd_cio_state(dev, head, regp, NV_CIO_CRE_53);
drivers/gpu/drm/nouveau/dispnv04/hw.c
650
rd_cio_state(dev, head, regp, NV_CIO_CRE_54);
drivers/gpu/drm/nouveau/dispnv04/hw.c
653
regp->CR58[i] = NVReadVgaCrtc5758(dev, head, i);
drivers/gpu/drm/nouveau/dispnv04/hw.c
654
rd_cio_state(dev, head, regp, NV_CIO_CRE_59);
drivers/gpu/drm/nouveau/dispnv04/hw.c
655
rd_cio_state(dev, head, regp, NV_CIO_CRE_5B);
drivers/gpu/drm/nouveau/dispnv04/hw.c
657
rd_cio_state(dev, head, regp, NV_CIO_CRE_85);
drivers/gpu/drm/nouveau/dispnv04/hw.c
658
rd_cio_state(dev, head, regp, NV_CIO_CRE_86);
drivers/gpu/drm/nouveau/dispnv04/hw.c
661
regp->fb_start = NVReadCRTC(dev, head, NV_PCRTC_START);
drivers/gpu/drm/nouveau/dispnv04/hw.c
665
nv_load_state_ext(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
670
struct nv04_crtc_reg *regp = &state->crtc_reg[head];
drivers/gpu/drm/nouveau/dispnv04/hw.c
680
NVWriteCRTC(dev, head, NV_PCRTC_ENGINE_CTRL, regp->crtc_eng_ctrl);
drivers/gpu/drm/nouveau/dispnv04/hw.c
692
NVWriteCRTC(dev, head, NV_PCRTC_CURSOR_CONFIG, regp->cursor_cfg);
drivers/gpu/drm/nouveau/dispnv04/hw.c
693
NVWriteCRTC(dev, head, NV_PCRTC_830, regp->crtc_830);
drivers/gpu/drm/nouveau/dispnv04/hw.c
694
NVWriteCRTC(dev, head, NV_PCRTC_834, regp->crtc_834);
drivers/gpu/drm/nouveau/dispnv04/hw.c
697
NVWriteCRTC(dev, head, NV_PCRTC_GPIO_EXT, regp->gpio_ext);
drivers/gpu/drm/nouveau/dispnv04/hw.c
700
NVWriteCRTC(dev, head, NV_PCRTC_850, regp->crtc_850);
drivers/gpu/drm/nouveau/dispnv04/hw.c
702
reg900 = NVReadRAMDAC(dev, head, NV_PRAMDAC_900);
drivers/gpu/drm/nouveau/dispnv04/hw.c
704
NVWriteRAMDAC(dev, head, NV_PRAMDAC_900, reg900 | 0x10000);
drivers/gpu/drm/nouveau/dispnv04/hw.c
706
NVWriteRAMDAC(dev, head, NV_PRAMDAC_900, reg900 & ~0x10000);
drivers/gpu/drm/nouveau/dispnv04/hw.c
710
NVWriteCRTC(dev, head, NV_PCRTC_CONFIG, regp->crtc_cfg);
drivers/gpu/drm/nouveau/dispnv04/hw.c
712
wr_cio_state(dev, head, regp, NV_CIO_CRE_RPC0_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
713
wr_cio_state(dev, head, regp, NV_CIO_CRE_RPC1_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
714
wr_cio_state(dev, head, regp, NV_CIO_CRE_LSR_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
715
wr_cio_state(dev, head, regp, NV_CIO_CRE_PIXEL_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
716
wr_cio_state(dev, head, regp, NV_CIO_CRE_LCD__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
717
wr_cio_state(dev, head, regp, NV_CIO_CRE_HEB__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
718
wr_cio_state(dev, head, regp, NV_CIO_CRE_ENH_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
719
wr_cio_state(dev, head, regp, NV_CIO_CRE_FF_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
720
wr_cio_state(dev, head, regp, NV_CIO_CRE_FFLWM__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
723
wr_cio_state(dev, head, regp, NV_CIO_CRE_47);
drivers/gpu/drm/nouveau/dispnv04/hw.c
726
wr_cio_state(dev, head, regp, 0x9f);
drivers/gpu/drm/nouveau/dispnv04/hw.c
728
wr_cio_state(dev, head, regp, NV_CIO_CRE_49);
drivers/gpu/drm/nouveau/dispnv04/hw.c
729
wr_cio_state(dev, head, regp, NV_CIO_CRE_HCUR_ADDR0_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
730
wr_cio_state(dev, head, regp, NV_CIO_CRE_HCUR_ADDR1_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
731
wr_cio_state(dev, head, regp, NV_CIO_CRE_HCUR_ADDR2_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
733
nv_fix_nv40_hw_cursor(dev, head);
drivers/gpu/drm/nouveau/dispnv04/hw.c
734
wr_cio_state(dev, head, regp, NV_CIO_CRE_ILACE__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
736
wr_cio_state(dev, head, regp, NV_CIO_CRE_SCRATCH3__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
737
wr_cio_state(dev, head, regp, NV_CIO_CRE_SCRATCH4__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
739
wr_cio_state(dev, head, regp, NV_CIO_CRE_EBR_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.c
740
wr_cio_state(dev, head, regp, NV_CIO_CRE_CSB);
drivers/gpu/drm/nouveau/dispnv04/hw.c
741
wr_cio_state(dev, head, regp, NV_CIO_CRE_4B);
drivers/gpu/drm/nouveau/dispnv04/hw.c
742
wr_cio_state(dev, head, regp, NV_CIO_CRE_TVOUT_LATENCY);
drivers/gpu/drm/nouveau/dispnv04/hw.c
759
wr_cio_state(dev, head, regp, NV_CIO_CRE_42);
drivers/gpu/drm/nouveau/dispnv04/hw.c
760
wr_cio_state(dev, head, regp, NV_CIO_CRE_53);
drivers/gpu/drm/nouveau/dispnv04/hw.c
761
wr_cio_state(dev, head, regp, NV_CIO_CRE_54);
drivers/gpu/drm/nouveau/dispnv04/hw.c
764
NVWriteVgaCrtc5758(dev, head, i, regp->CR58[i]);
drivers/gpu/drm/nouveau/dispnv04/hw.c
765
wr_cio_state(dev, head, regp, NV_CIO_CRE_59);
drivers/gpu/drm/nouveau/dispnv04/hw.c
766
wr_cio_state(dev, head, regp, NV_CIO_CRE_5B);
drivers/gpu/drm/nouveau/dispnv04/hw.c
768
wr_cio_state(dev, head, regp, NV_CIO_CRE_85);
drivers/gpu/drm/nouveau/dispnv04/hw.c
769
wr_cio_state(dev, head, regp, NV_CIO_CRE_86);
drivers/gpu/drm/nouveau/dispnv04/hw.c
772
NVWriteCRTC(dev, head, NV_PCRTC_START, regp->fb_start);
drivers/gpu/drm/nouveau/dispnv04/hw.c
776
nv_save_state_palette(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
780
int head_offset = head * NV_PRMDIO_SIZE, i;
drivers/gpu/drm/nouveau/dispnv04/hw.c
787
state->crtc_reg[head].DAC[i] = nvif_rd08(device,
drivers/gpu/drm/nouveau/dispnv04/hw.c
791
NVSetEnablePalette(dev, head, false);
drivers/gpu/drm/nouveau/dispnv04/hw.c
795
nouveau_hw_load_state_palette(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
799
int head_offset = head * NV_PRMDIO_SIZE, i;
drivers/gpu/drm/nouveau/dispnv04/hw.c
807
state->crtc_reg[head].DAC[i]);
drivers/gpu/drm/nouveau/dispnv04/hw.c
810
NVSetEnablePalette(dev, head, false);
drivers/gpu/drm/nouveau/dispnv04/hw.c
813
void nouveau_hw_save_state(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
820
nouveau_hw_fix_bad_vpll(dev, head);
drivers/gpu/drm/nouveau/dispnv04/hw.c
821
nv_save_state_ramdac(dev, head, state);
drivers/gpu/drm/nouveau/dispnv04/hw.c
822
nv_save_state_vga(dev, head, state);
drivers/gpu/drm/nouveau/dispnv04/hw.c
823
nv_save_state_palette(dev, head, state);
drivers/gpu/drm/nouveau/dispnv04/hw.c
824
nv_save_state_ext(dev, head, state);
drivers/gpu/drm/nouveau/dispnv04/hw.c
827
void nouveau_hw_load_state(struct drm_device *dev, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.c
830
NVVgaProtect(dev, head, true);
drivers/gpu/drm/nouveau/dispnv04/hw.c
831
nv_load_state_ramdac(dev, head, state);
drivers/gpu/drm/nouveau/dispnv04/hw.c
832
nv_load_state_ext(dev, head, state);
drivers/gpu/drm/nouveau/dispnv04/hw.c
833
nouveau_hw_load_state_palette(dev, head, state);
drivers/gpu/drm/nouveau/dispnv04/hw.c
834
nv_load_state_vga(dev, head, state);
drivers/gpu/drm/nouveau/dispnv04/hw.c
835
NVVgaProtect(dev, head, false);
drivers/gpu/drm/nouveau/dispnv04/hw.h
120
int head, uint8_t index, uint8_t value)
drivers/gpu/drm/nouveau/dispnv04/hw.h
123
nvif_wr08(device, NV_PRMCIO_CRX__COLOR + head * NV_PRMCIO_SIZE, index);
drivers/gpu/drm/nouveau/dispnv04/hw.h
124
nvif_wr08(device, NV_PRMCIO_CR__COLOR + head * NV_PRMCIO_SIZE, value);
drivers/gpu/drm/nouveau/dispnv04/hw.h
128
int head, uint8_t index)
drivers/gpu/drm/nouveau/dispnv04/hw.h
132
nvif_wr08(device, NV_PRMCIO_CRX__COLOR + head * NV_PRMCIO_SIZE, index);
drivers/gpu/drm/nouveau/dispnv04/hw.h
133
val = nvif_rd08(device, NV_PRMCIO_CR__COLOR + head * NV_PRMCIO_SIZE);
drivers/gpu/drm/nouveau/dispnv04/hw.h
152
NVWriteVgaCrtc5758(struct drm_device *dev, int head, uint8_t index, uint8_t value)
drivers/gpu/drm/nouveau/dispnv04/hw.h
154
NVWriteVgaCrtc(dev, head, NV_CIO_CRE_57, index);
drivers/gpu/drm/nouveau/dispnv04/hw.h
155
NVWriteVgaCrtc(dev, head, NV_CIO_CRE_58, value);
drivers/gpu/drm/nouveau/dispnv04/hw.h
158
static inline uint8_t NVReadVgaCrtc5758(struct drm_device *dev, int head, uint8_t index)
drivers/gpu/drm/nouveau/dispnv04/hw.h
160
NVWriteVgaCrtc(dev, head, NV_CIO_CRE_57, index);
drivers/gpu/drm/nouveau/dispnv04/hw.h
161
return NVReadVgaCrtc(dev, head, NV_CIO_CRE_58);
drivers/gpu/drm/nouveau/dispnv04/hw.h
165
int head, uint32_t reg)
drivers/gpu/drm/nouveau/dispnv04/hw.h
173
if (head && drm->client.device.info.family == NV_DEVICE_INFO_V0_CURIE)
drivers/gpu/drm/nouveau/dispnv04/hw.h
181
int head, uint32_t reg, uint8_t value)
drivers/gpu/drm/nouveau/dispnv04/hw.h
188
if (head && drm->client.device.info.family == NV_DEVICE_INFO_V0_CURIE)
drivers/gpu/drm/nouveau/dispnv04/hw.h
194
static inline void NVSetEnablePalette(struct drm_device *dev, int head, bool enable)
drivers/gpu/drm/nouveau/dispnv04/hw.h
197
nvif_rd08(device, NV_PRMCIO_INP0__COLOR + head * NV_PRMCIO_SIZE);
drivers/gpu/drm/nouveau/dispnv04/hw.h
198
nvif_wr08(device, NV_PRMCIO_ARX + head * NV_PRMCIO_SIZE, enable ? 0 : 0x20);
drivers/gpu/drm/nouveau/dispnv04/hw.h
201
static inline bool NVGetEnablePalette(struct drm_device *dev, int head)
drivers/gpu/drm/nouveau/dispnv04/hw.h
204
nvif_rd08(device, NV_PRMCIO_INP0__COLOR + head * NV_PRMCIO_SIZE);
drivers/gpu/drm/nouveau/dispnv04/hw.h
205
return !(nvif_rd08(device, NV_PRMCIO_ARX + head * NV_PRMCIO_SIZE) & 0x20);
drivers/gpu/drm/nouveau/dispnv04/hw.h
209
int head, uint8_t index, uint8_t value)
drivers/gpu/drm/nouveau/dispnv04/hw.h
212
if (NVGetEnablePalette(dev, head))
drivers/gpu/drm/nouveau/dispnv04/hw.h
217
nvif_rd08(device, NV_PRMCIO_INP0__COLOR + head * NV_PRMCIO_SIZE);
drivers/gpu/drm/nouveau/dispnv04/hw.h
218
nvif_wr08(device, NV_PRMCIO_ARX + head * NV_PRMCIO_SIZE, index);
drivers/gpu/drm/nouveau/dispnv04/hw.h
219
nvif_wr08(device, NV_PRMCIO_AR__WRITE + head * NV_PRMCIO_SIZE, value);
drivers/gpu/drm/nouveau/dispnv04/hw.h
223
int head, uint8_t index)
drivers/gpu/drm/nouveau/dispnv04/hw.h
227
if (NVGetEnablePalette(dev, head))
drivers/gpu/drm/nouveau/dispnv04/hw.h
232
nvif_rd08(device, NV_PRMCIO_INP0__COLOR + head * NV_PRMCIO_SIZE);
drivers/gpu/drm/nouveau/dispnv04/hw.h
233
nvif_wr08(device, NV_PRMCIO_ARX + head * NV_PRMCIO_SIZE, index);
drivers/gpu/drm/nouveau/dispnv04/hw.h
234
val = nvif_rd08(device, NV_PRMCIO_AR__READ + head * NV_PRMCIO_SIZE);
drivers/gpu/drm/nouveau/dispnv04/hw.h
238
static inline void NVVgaSeqReset(struct drm_device *dev, int head, bool start)
drivers/gpu/drm/nouveau/dispnv04/hw.h
240
NVWriteVgaSeq(dev, head, NV_VIO_SR_RESET_INDEX, start ? 0x1 : 0x3);
drivers/gpu/drm/nouveau/dispnv04/hw.h
243
static inline void NVVgaProtect(struct drm_device *dev, int head, bool protect)
drivers/gpu/drm/nouveau/dispnv04/hw.h
245
uint8_t seq1 = NVReadVgaSeq(dev, head, NV_VIO_SR_CLOCK_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.h
248
NVVgaSeqReset(dev, head, true);
drivers/gpu/drm/nouveau/dispnv04/hw.h
249
NVWriteVgaSeq(dev, head, NV_VIO_SR_CLOCK_INDEX, seq1 | 0x20);
drivers/gpu/drm/nouveau/dispnv04/hw.h
252
NVWriteVgaSeq(dev, head, NV_VIO_SR_CLOCK_INDEX, seq1 & ~0x20); /* reenable display */
drivers/gpu/drm/nouveau/dispnv04/hw.h
253
NVVgaSeqReset(dev, head, false);
drivers/gpu/drm/nouveau/dispnv04/hw.h
255
NVSetEnablePalette(dev, head, protect);
drivers/gpu/drm/nouveau/dispnv04/hw.h
272
nv_lock_vga_crtc_base(struct drm_device *dev, int head, bool lock)
drivers/gpu/drm/nouveau/dispnv04/hw.h
274
uint8_t cr11 = NVReadVgaCrtc(dev, head, NV_CIO_CR_VRE_INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.h
281
NVWriteVgaCrtc(dev, head, NV_CIO_CR_VRE_INDEX, cr11);
drivers/gpu/drm/nouveau/dispnv04/hw.h
287
nv_lock_vga_crtc_shadow(struct drm_device *dev, int head, int lock)
drivers/gpu/drm/nouveau/dispnv04/hw.h
303
cr21 = NVReadVgaCrtc(dev, head, NV_CIO_CRE_21) | 0xfa;
drivers/gpu/drm/nouveau/dispnv04/hw.h
305
NVWriteVgaCrtc(dev, head, NV_CIO_CRE_21, cr21);
drivers/gpu/drm/nouveau/dispnv04/hw.h
341
nv_fix_nv40_hw_cursor(struct drm_device *dev, int head)
drivers/gpu/drm/nouveau/dispnv04/hw.h
348
uint32_t curpos = NVReadRAMDAC(dev, head, NV_PRAMDAC_CU_START_POS);
drivers/gpu/drm/nouveau/dispnv04/hw.h
349
NVWriteRAMDAC(dev, head, NV_PRAMDAC_CU_START_POS, curpos);
drivers/gpu/drm/nouveau/dispnv04/hw.h
353
nv_set_crtc_base(struct drm_device *dev, int head, uint32_t offset)
drivers/gpu/drm/nouveau/dispnv04/hw.h
357
NVWriteCRTC(dev, head, NV_PCRTC_START, offset);
drivers/gpu/drm/nouveau/dispnv04/hw.h
364
int cre_heb = NVReadVgaCrtc(dev, head, NV_CIO_CRE_HEB__INDEX);
drivers/gpu/drm/nouveau/dispnv04/hw.h
366
NVWriteVgaCrtc(dev, head, NV_CIO_CRE_HEB__INDEX,
drivers/gpu/drm/nouveau/dispnv04/hw.h
37
void NVWriteVgaSeq(struct drm_device *, int head, uint8_t index, uint8_t value);
drivers/gpu/drm/nouveau/dispnv04/hw.h
372
nv_show_cursor(struct drm_device *dev, int head, bool show)
drivers/gpu/drm/nouveau/dispnv04/hw.h
376
&nv04_display(dev)->mode_reg.crtc_reg[head].CRTC[NV_CIO_CRE_HCUR_ADDR1_INDEX];
drivers/gpu/drm/nouveau/dispnv04/hw.h
38
uint8_t NVReadVgaSeq(struct drm_device *, int head, uint8_t index);
drivers/gpu/drm/nouveau/dispnv04/hw.h
382
NVWriteVgaCrtc(dev, head, NV_CIO_CRE_HCUR_ADDR1_INDEX, *curctl1);
drivers/gpu/drm/nouveau/dispnv04/hw.h
385
nv_fix_nv40_hw_cursor(dev, head);
drivers/gpu/drm/nouveau/dispnv04/hw.h
39
void NVWriteVgaGr(struct drm_device *, int head, uint8_t index, uint8_t value);
drivers/gpu/drm/nouveau/dispnv04/hw.h
40
uint8_t NVReadVgaGr(struct drm_device *, int head, uint8_t index);
drivers/gpu/drm/nouveau/dispnv04/hw.h
42
void NVBlankScreen(struct drm_device *, int head, bool blank);
drivers/gpu/drm/nouveau/dispnv04/hw.h
48
void nouveau_hw_save_state(struct drm_device *, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.h
50
void nouveau_hw_load_state(struct drm_device *, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.h
52
void nouveau_hw_load_state_palette(struct drm_device *, int head,
drivers/gpu/drm/nouveau/dispnv04/hw.h
60
int head, uint32_t reg)
drivers/gpu/drm/nouveau/dispnv04/hw.h
64
if (head)
drivers/gpu/drm/nouveau/dispnv04/hw.h
71
int head, uint32_t reg, uint32_t val)
drivers/gpu/drm/nouveau/dispnv04/hw.h
74
if (head)
drivers/gpu/drm/nouveau/dispnv04/hw.h
80
int head, uint32_t reg)
drivers/gpu/drm/nouveau/dispnv04/hw.h
84
if (head)
drivers/gpu/drm/nouveau/dispnv04/hw.h
91
int head, uint32_t reg, uint32_t val)
drivers/gpu/drm/nouveau/dispnv04/hw.h
94
if (head)
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
545
int head = nouveau_crtc(encoder->crtc)->index;
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
546
struct nv04_crtc_reg *regs = &nv04_display(dev)->mode_reg.crtc_reg[head];
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
581
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_HVALID_START,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
583
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_HVALID_END,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
585
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_VVALID_START,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
587
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_VVALID_END,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
589
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_DEBUG_1, regs->fp_debug_1);
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
105
static void nv04_tv_bind(struct drm_device *dev, int head, bool bind)
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
107
struct nv04_crtc_reg *state = &nv04_display(dev)->mode_reg.crtc_reg[head];
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
116
NVWriteVgaCrtc(dev, head, NV_CIO_CRE_LCD__INDEX,
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
118
NVWriteVgaCrtc(dev, head, NV_CIO_CRE_49,
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
120
NVWriteRAMDAC(dev, head, NV_PRAMDAC_TV_SETUP,
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
127
int head = nouveau_crtc(encoder->crtc)->index;
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
132
nv04_dfp_disable(dev, head);
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
135
nv04_tv_bind(dev, head ^ 1, false);
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
137
nv04_tv_bind(dev, head, true);
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
88
int head = nouveau_crtc(encoder->crtc)->index;
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
89
crtc1A = NVReadVgaCrtc(dev, head, NV_CIO_CRE_RPC1_INDEX);
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
91
state->pllsel |= head ? PLLSEL_TV_CRTC2_MASK :
drivers/gpu/drm/nouveau/dispnv04/tvnv04.c
97
NVWriteVgaCrtc(dev, head, NV_CIO_CRE_RPC1_INDEX, crtc1A);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
100
NVWriteRAMDAC(dev, head, 0x680c14, 4 << 16);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
103
NVWriteRAMDAC(dev, head, 0x680c6c, testval >> 10 & 0x3ff);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
109
NVWriteRAMDAC(dev, head, 0x680c6c, testval & 0x3ff);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
115
NVWriteRAMDAC(dev, head, 0x680c1c, ctv_1c);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
116
NVWriteRAMDAC(dev, head, 0x680c14, ctv_14);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
117
NVWriteRAMDAC(dev, head, 0x680c6c, ctv_6c);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
120
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL, fp_control);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
121
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_HSYNC_END, fp_hsync_end);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
122
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_HSYNC_START, fp_hsync_start);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
123
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_HTOTAL, fp_htotal);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
409
int head = nouveau_crtc(encoder->crtc)->index;
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
410
uint8_t *cr_lcd = &nv04_display(dev)->mode_reg.crtc_reg[head].CRTC[
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
418
nv04_dfp_disable(dev, head);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
425
list_for_each_entry(enc, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
431
nv04_dfp_get_bound_head(dev, dcb) == head) {
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
432
nv04_dfp_bind_head(dev, dcb, head ^ 1,
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
440
*cr_lcd |= 0x1 | (head ? 0x0 : 0x8);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
451
if (head)
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
470
int head = nouveau_crtc(encoder->crtc)->index;
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
471
struct nv04_crtc_reg *regs = &nv04_display(dev)->mode_reg.crtc_reg[head];
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
484
if (head)
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
55
int head;
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
63
head = (dacclk & 0x100) >> 8;
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
68
fp_htotal = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_HTOTAL);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
69
fp_hsync_start = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_HSYNC_START);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
70
fp_hsync_end = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_HSYNC_END);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
71
fp_control = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
73
ctv_1c = NVReadRAMDAC(dev, head, 0x680c1c);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
74
ctv_14 = NVReadRAMDAC(dev, head, 0x680c14);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
75
ctv_6c = NVReadRAMDAC(dev, head, 0x680c6c);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
81
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_HTOTAL, 1343);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
82
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_HSYNC_START, 1047);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
83
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_HSYNC_END, 1183);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
84
NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL,
drivers/gpu/drm/nouveau/dispnv04/tvnv17.c
99
NVWriteRAMDAC(dev, head, 0x680c1c, 1 << 20);
drivers/gpu/drm/nouveau/dispnv50/base.c
27
nv50_base_new(struct nouveau_drm *drm, int head, struct nv50_wndw **pwndw)
drivers/gpu/drm/nouveau/dispnv50/base.c
52
return bases[cid].new(drm, head, bases[cid].oclass, pwndw);
drivers/gpu/drm/nouveau/dispnv50/base.h
26
int nv50_base_new(struct nouveau_drm *, int head, struct nv50_wndw **);
drivers/gpu/drm/nouveau/dispnv50/base.h
7
struct nouveau_drm *, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/base507c.c
304
struct nouveau_drm *drm, int head, s32 oclass, u32 interlock_data,
drivers/gpu/drm/nouveau/dispnv50/base507c.c
308
.id = head,
drivers/gpu/drm/nouveau/dispnv50/base507c.c
315
"base", head, format, BIT(head),
drivers/gpu/drm/nouveau/dispnv50/base507c.c
321
&oclass, head, &args, sizeof(args),
drivers/gpu/drm/nouveau/dispnv50/base507c.c
335
base507c_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/base507c.c
338
return base507c_new_(&base507c, base507c_format, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/base507c.c
339
0x00000002 << (head * 8), pwndw);
drivers/gpu/drm/nouveau/dispnv50/base827c.c
102
return base507c_new_(&base827c, base507c_format, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/base827c.c
103
0x00000002 << (head * 8), pwndw);
drivers/gpu/drm/nouveau/dispnv50/base827c.c
99
base827c_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/base907c.c
211
base907c_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/base907c.c
214
return base507c_new_(&base907c, base507c_format, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/base907c.c
215
0x00000002 << (head * 4), pwndw);
drivers/gpu/drm/nouveau/dispnv50/base917c.c
45
base917c_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/base917c.c
48
return base507c_new_(&base907c, base917c_format, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/base917c.c
49
0x00000002 << (head * 4), pwndw);
drivers/gpu/drm/nouveau/dispnv50/core.h
33
const struct nv50_head_func *head;
drivers/gpu/drm/nouveau/dispnv50/core507d.c
150
.head = &head507d,
drivers/gpu/drm/nouveau/dispnv50/core827d.c
32
.head = &head827d,
drivers/gpu/drm/nouveau/dispnv50/core907d.c
66
.head = &head907d,
drivers/gpu/drm/nouveau/dispnv50/core917d.c
32
.head = &head917d,
drivers/gpu/drm/nouveau/dispnv50/corec37d.c
169
.head = &headc37d,
drivers/gpu/drm/nouveau/dispnv50/corec57d.c
71
.head = &headc57d,
drivers/gpu/drm/nouveau/dispnv50/coreca7d.c
111
.head = &headca7d,
drivers/gpu/drm/nouveau/dispnv50/crc.c
108
nv50_crc_program_ctx(head, NULL);
drivers/gpu/drm/nouveau/dispnv50/crc.c
109
nv50_crc_program_ctx(head, &crc->ctx[new_idx]);
drivers/gpu/drm/nouveau/dispnv50/crc.c
129
nv50_crc_get_entries(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/crc.c
133
struct drm_crtc *crtc = &head->base.base;
drivers/gpu/drm/nouveau/dispnv50/crc.c
134
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
143
output_crc = func->get_entry(head, &crc->ctx[crc->ctx_idx],
drivers/gpu/drm/nouveau/dispnv50/crc.c
154
void nv50_crc_handle_vblank(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/crc.c
156
struct drm_crtc *crtc = &head->base.base;
drivers/gpu/drm/nouveau/dispnv50/crc.c
157
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
159
nv50_disp(head->base.base.dev)->core->func->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
178
if (crc->ctx_changed && func->ctx_finished(head, ctx)) {
drivers/gpu/drm/nouveau/dispnv50/crc.c
179
nv50_crc_get_entries(head, func, crc->src);
drivers/gpu/drm/nouveau/dispnv50/crc.c
198
drm_dbg_kms(head->base.base.dev,
drivers/gpu/drm/nouveau/dispnv50/crc.c
200
head->base.index, crc->frame);
drivers/gpu/drm/nouveau/dispnv50/crc.c
207
nv50_crc_get_entries(head, func, crc->src);
drivers/gpu/drm/nouveau/dispnv50/crc.c
220
static void nv50_crc_wait_ctx_finished(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/crc.c
224
struct drm_device *dev = head->base.base.dev;
drivers/gpu/drm/nouveau/dispnv50/crc.c
229
if (func->ctx_finished(head, ctx)) break;);
drivers/gpu/drm/nouveau/dispnv50/crc.c
233
head->base.index);
drivers/gpu/drm/nouveau/dispnv50/crc.c
237
head->base.index, ret);
drivers/gpu/drm/nouveau/dispnv50/crc.c
247
struct nv50_head *head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/crc.c
249
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
263
head->base.index);
drivers/gpu/drm/nouveau/dispnv50/crc.c
279
struct nv50_head *head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/crc.c
281
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
303
struct nv50_head *head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/crc.c
305
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
312
nv50_crc_wait_ctx_finished(head, func, ctx);
drivers/gpu/drm/nouveau/dispnv50/crc.c
315
nv50_crc_wait_ctx_finished(head, func, ctx);
drivers/gpu/drm/nouveau/dispnv50/crc.c
326
struct nv50_head *head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/crc.c
328
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
347
head->base.index);
drivers/gpu/drm/nouveau/dispnv50/crc.c
351
int nv50_crc_atomic_check_head(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/crc.c
415
list_for_each_entry(outp_atom, &atom->outp, head) {
drivers/gpu/drm/nouveau/dispnv50/crc.c
450
void nv50_crc_atomic_set(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/crc.c
453
struct drm_crtc *crtc = &head->base.base;
drivers/gpu/drm/nouveau/dispnv50/crc.c
455
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
468
func->set_src(head, outp->outp.or.id, nv50_crc_source_type(outp, asyh->crc.src),
drivers/gpu/drm/nouveau/dispnv50/crc.c
472
void nv50_crc_atomic_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/crc.c
475
nv50_disp(head->base.base.dev)->core->func->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
477
func->set_src(head, 0, NV50_CRC_SOURCE_TYPE_NONE, NULL);
drivers/gpu/drm/nouveau/dispnv50/crc.c
502
nv50_crc_ctx_init(struct nv50_head *head, struct nvif_mmu *mmu,
drivers/gpu/drm/nouveau/dispnv50/crc.c
505
struct nv50_core *core = nv50_disp(head->base.base.dev)->core;
drivers/gpu/drm/nouveau/dispnv50/crc.c
517
NV50_DISP_HANDLE_CRC_CTX(head, idx),
drivers/gpu/drm/nouveau/dispnv50/crc.c
549
struct nv50_head *head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/crc.c
550
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
578
for (i = 0; i < ARRAY_SIZE(head->crc.ctx); i++) {
drivers/gpu/drm/nouveau/dispnv50/crc.c
579
ret = nv50_crc_ctx_init(head, mmu, &crc->ctx[i],
drivers/gpu/drm/nouveau/dispnv50/crc.c
587
crtc_state = drm_atomic_get_crtc_state(state, &head->base.base);
drivers/gpu/drm/nouveau/dispnv50/crc.c
634
struct nv50_head *head = m->private;
drivers/gpu/drm/nouveau/dispnv50/crc.c
635
struct drm_crtc *crtc = &head->base.base;
drivers/gpu/drm/nouveau/dispnv50/crc.c
636
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
662
struct nv50_head *head = m->private;
drivers/gpu/drm/nouveau/dispnv50/crc.c
664
struct drm_crtc *crtc = &head->base.base;
drivers/gpu/drm/nouveau/dispnv50/crc.c
666
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
694
head->base.index, value);
drivers/gpu/drm/nouveau/dispnv50/crc.c
711
int nv50_head_crc_late_register(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/crc.c
713
struct drm_crtc *crtc = &head->base.base;
drivers/gpu/drm/nouveau/dispnv50/crc.c
72
nv50_crc_program_ctx(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/crc.c
722
debugfs_create_file("flip_threshold", 0644, root, head,
drivers/gpu/drm/nouveau/dispnv50/crc.c
730
struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/crc.c
732
struct nv50_crc *crc = &head->crc;
drivers/gpu/drm/nouveau/dispnv50/crc.c
736
drm_vblank_work_init(&crc->flip_work, &head->base.base,
drivers/gpu/drm/nouveau/dispnv50/crc.c
75
struct nv50_disp *disp = nv50_disp(head->base.base.dev);
drivers/gpu/drm/nouveau/dispnv50/crc.c
79
core->func->crc->set_ctx(head, ctx);
drivers/gpu/drm/nouveau/dispnv50/crc.c
87
struct nv50_head *head = container_of(crc, struct nv50_head, crc);
drivers/gpu/drm/nouveau/dispnv50/crc.c
88
struct drm_crtc *crtc = &head->base.base;
drivers/gpu/drm/nouveau/dispnv50/crc.h
110
nv50_head_crc_late_register(struct nv50_head *head) { return 0; }
drivers/gpu/drm/nouveau/dispnv50/crc.h
111
static inline void nv50_crc_handle_vblank(struct nv50_head *head) {}
drivers/gpu/drm/nouveau/dispnv50/crc.h
114
nv50_crc_atomic_check_head(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/crc.h
127
nv50_crc_atomic_set(struct nv50_head *head, struct nv50_head_atom *state) {}
drivers/gpu/drm/nouveau/dispnv50/crc.h
129
nv50_crc_atomic_clr(struct nv50_head *head) {}
drivers/gpu/drm/nouveau/dispnv50/crc.h
78
void nv50_crc_handle_vblank(struct nv50_head *head);
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
100
struct nouveau_drm *drm = nouveau_drm(head->base.base.dev);
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
121
head->base.index, engine, status);
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
125
head->base.index, status);
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
129
head->base.index, status);
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
26
crc907d_set_src(struct nv50_head *head, int or, enum nv50_crc_source_type source,
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
29
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
30
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
75
crc907d_set_ctx(struct nv50_head *head, struct nv50_crc_notifier_ctx *ctx)
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
77
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
78
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
88
static u32 crc907d_get_entry(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
97
static bool crc907d_ctx_finished(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
106
head->base.index, engine, status);
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
110
head->base.index, status);
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
114
head->base.index, status);
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
15
crcc37d_set_src(struct nv50_head *head, int or, enum nv50_crc_source_type source,
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
18
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
19
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
54
int crcc37d_set_ctx(struct nv50_head *head, struct nv50_crc_notifier_ctx *ctx)
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
56
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
57
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
67
u32 crcc37d_get_entry(struct nv50_head *head, struct nv50_crc_notifier_ctx *ctx,
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
82
bool crcc37d_ctx_finished(struct nv50_head *head, struct nv50_crc_notifier_ctx *ctx)
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
84
struct nouveau_drm *drm = nouveau_drm(head->base.base.dev);
drivers/gpu/drm/nouveau/dispnv50/crcc37d.h
35
int crcc37d_set_ctx(struct nv50_head *head, struct nv50_crc_notifier_ctx *ctx);
drivers/gpu/drm/nouveau/dispnv50/crcc37d.h
36
u32 crcc37d_get_entry(struct nv50_head *head, struct nv50_crc_notifier_ctx *ctx,
drivers/gpu/drm/nouveau/dispnv50/crcc37d.h
38
bool crcc37d_ctx_finished(struct nv50_head *head, struct nv50_crc_notifier_ctx *ctx);
drivers/gpu/drm/nouveau/dispnv50/crcc57d.c
13
static int crcc57d_set_src(struct nv50_head *head, int or, enum nv50_crc_source_type source,
drivers/gpu/drm/nouveau/dispnv50/crcc57d.c
16
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/crcc57d.c
17
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
14
crcca7d_set_ctx(struct nv50_head *head, struct nv50_crc_notifier_ctx *ctx)
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
16
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
17
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
43
crcca7d_set_src(struct nv50_head *head, int or, enum nv50_crc_source_type source,
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
46
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
47
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
57
return crcca7d_set_ctx(head, NULL);
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
71
ret = crcca7d_set_ctx(head, ctx);
drivers/gpu/drm/nouveau/dispnv50/curs.c
27
nv50_curs_new(struct nouveau_drm *drm, int head, struct nv50_wndw **pwndw)
drivers/gpu/drm/nouveau/dispnv50/curs.c
54
return curses[cid].new(drm, head, curses[cid].oclass, pwndw);
drivers/gpu/drm/nouveau/dispnv50/curs.h
13
int nv50_curs_new(struct nouveau_drm *, int head, struct nv50_wndw **);
drivers/gpu/drm/nouveau/dispnv50/curs.h
7
int head, s32 oclass, u32 interlock_data,
drivers/gpu/drm/nouveau/dispnv50/curs507a.c
103
struct nv50_head *head = nv50_head(asyw->state.crtc);
drivers/gpu/drm/nouveau/dispnv50/curs507a.c
137
ret = head->func->curs_layout(head, asyw, asyh);
drivers/gpu/drm/nouveau/dispnv50/curs507a.c
145
ret = head->func->curs_format(head, asyw, asyh);
drivers/gpu/drm/nouveau/dispnv50/curs507a.c
171
int head, s32 oclass, u32 interlock_data,
drivers/gpu/drm/nouveau/dispnv50/curs507a.c
175
.id = head,
drivers/gpu/drm/nouveau/dispnv50/curs507a.c
182
"curs", head, curs507a_format, BIT(head),
drivers/gpu/drm/nouveau/dispnv50/curs507a.c
201
curs507a_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/curs507a.c
204
return curs507a_new_(&curs507a, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/curs507a.c
205
0x00000001 << (head * 8), pwndw);
drivers/gpu/drm/nouveau/dispnv50/curs907a.c
25
curs907a_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/curs907a.c
28
return curs507a_new_(&curs507a, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/curs907a.c
29
0x00000001 << (head * 4), pwndw);
drivers/gpu/drm/nouveau/dispnv50/cursc37a.c
57
cursc37a_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/cursc37a.c
60
return curs507a_new_(&cursc37a, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/cursc37a.c
61
0x00000001 << head, pwndw);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1030
struct nv50_head *head = msto->head;
drivers/gpu/drm/nouveau/dispnv50/disp.c
1032
nv50_head_atom(drm_atomic_get_new_crtc_state(state, &head->base.base));
drivers/gpu/drm/nouveau/dispnv50/disp.c
1057
if (head->func->display_id) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
1059
head->func->display_id(head, msto->display_id);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1067
mstm->outp->update(mstm->outp, head->base.index, asyh, proto,
drivers/gpu/drm/nouveau/dispnv50/disp.c
1082
if (msto->head->func->display_id)
drivers/gpu/drm/nouveau/dispnv50/disp.c
1083
msto->head->func->display_id(msto->head, 0);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1085
mstm->outp->update(mstm->outp, msto->head->base.index, NULL, 0, 0);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1113
nv50_msto_new(struct drm_device *dev, struct nv50_head *head, int id)
drivers/gpu/drm/nouveau/dispnv50/disp.c
1130
msto->encoder.possible_crtcs = drm_crtc_mask(&head->base.base);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1131
msto->head = head;
drivers/gpu/drm/nouveau/dispnv50/disp.c
1540
nv50_sor_update(struct nouveau_encoder *nv_encoder, u8 head,
drivers/gpu/drm/nouveau/dispnv50/disp.c
1547
nv_encoder->ctrl &= ~BIT(head);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1552
nv_encoder->ctrl |= BIT(head);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1568
struct nv50_head *head = nv50_head(nv_encoder->crtc);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1585
nvif_outp_hdmi(&nv_encoder->outp, head->base.index,
drivers/gpu/drm/nouveau/dispnv50/disp.c
1593
if (head->func->display_id)
drivers/gpu/drm/nouveau/dispnv50/disp.c
1594
head->func->display_id(head, 0);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1596
nv_encoder->update(nv_encoder, head->base.index, NULL, 0, 0);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1597
nv50_audio_disable(encoder, &head->base);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1609
struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/disp.c
1740
return nvif_outp_dp_sst(&outp->outp, head->base.index, waterMark, hBlankSym, vBlankSym);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1752
struct nv50_head *head = nv50_head(&nv_crtc->base);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1830
nv50_sor_dp_watermark_sst(nv_encoder, head, asyh);
drivers/gpu/drm/nouveau/dispnv50/disp.c
1851
if (head->func->display_id)
drivers/gpu/drm/nouveau/dispnv50/disp.c
1852
head->func->display_id(head, BIT(nv_encoder->outp.id));
drivers/gpu/drm/nouveau/dispnv50/disp.c
2134
list_for_each_entry(outp, &atom->outp, head) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2199
struct nv50_head *head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2210
nv50_head_flush_clr(head, asyh, atom->flush_disable);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2229
list_for_each_entry(outp, &atom->outp, head) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2262
list_for_each_entry(outp, &atom->outp, head) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2282
struct nv50_head *head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2288
nv50_head_flush_set(head, asyh);
drivers/gpu/drm/nouveau/dispnv50/disp.c
233
const s32 *oclass, u8 head, void *data, u32 size, s64 syncbuf,
drivers/gpu/drm/nouveau/dispnv50/disp.c
2333
struct nv50_head *head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2339
nv50_head_flush_set_wndw(head, asyh);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2374
list_for_each_entry_safe(outp, outt, &atom->outp, head) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2375
list_del(&outp->head);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2494
list_for_each_entry(outp, &atom->outp, head) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2503
list_add(&outp->head, &atom->outp);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2571
struct nv50_head *head;
drivers/gpu/drm/nouveau/dispnv50/disp.c
2575
if (core->assign_windows && core->func->head->static_wndw_map) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2582
head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2584
core->func->head->static_wndw_map(head, asyh);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2626
list_for_each_entry_safe(outp, outt, &atom->outp, head) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2627
list_del(&outp->head);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2680
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
274
ret = nv50_chan_create(device, disp, oclass, head, data, size,
drivers/gpu/drm/nouveau/dispnv50/disp.c
2796
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2973
list_for_each_entry_safe(connector, tmp, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2984
struct nv50_head *head;
drivers/gpu/drm/nouveau/dispnv50/disp.c
2986
head = nv50_head_create(dev, i);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2987
if (IS_ERR(head)) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2988
ret = PTR_ERR(head);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2993
head->msto = nv50_msto_new(dev, head, i);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2994
if (IS_ERR(head->msto)) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2995
ret = PTR_ERR(head->msto);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2996
head->msto = NULL;
drivers/gpu/drm/nouveau/dispnv50/disp.c
3010
head->msto->encoder.possible_crtcs = disp->disp->head_mask;
drivers/gpu/drm/nouveau/dispnv50/disp.c
78
const s32 *oclass, u8 head, void *data, u32 size,
drivers/gpu/drm/nouveau/dispnv50/disp.c
827
args->head = nv_crtc->index;
drivers/gpu/drm/nouveau/dispnv50/disp.c
873
struct nv50_head *head;
drivers/gpu/drm/nouveau/dispnv50/disp.c
913
if (msto->head->func->display_id) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
945
nvif_outp_dp_mst_vcpi(&mstm->outp->outp, msto->head->base.index, 0, 0, 0, 0);
drivers/gpu/drm/nouveau/dispnv50/disp.c
953
nvif_outp_dp_mst_vcpi(&mstm->outp->outp, msto->head->base.index,
drivers/gpu/drm/nouveau/dispnv50/disp.c
958
nvif_outp_dp_mst_vcpi(&mstm->outp->outp, msto->head->base.index, 0, 0, 0, 0);
drivers/gpu/drm/nouveau/dispnv50/disp.h
76
struct list_head head;
drivers/gpu/drm/nouveau/dispnv50/disp.h
92
const s32 *oclass, u8 head, void *data, u32 size,
drivers/gpu/drm/nouveau/dispnv50/handles.h
14
#define NV50_DISP_HANDLE_CRC_CTX(head, i) (0xfc000000 | head->base.index << 1 | i)
drivers/gpu/drm/nouveau/dispnv50/head.c
225
nv50_head_atomic_check_lut(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/head.c
228
struct drm_device *dev = head->base.base.dev;
drivers/gpu/drm/nouveau/dispnv50/head.c
229
struct drm_crtc *crtc = &head->base.base;
drivers/gpu/drm/nouveau/dispnv50/head.c
239
if (!head->func->ilut_check(size)) {
drivers/gpu/drm/nouveau/dispnv50/head.c
262
if (!head->func->olut_identity) {
drivers/gpu/drm/nouveau/dispnv50/head.c
271
if (!head->func->olut(head, asyh, size)) {
drivers/gpu/drm/nouveau/dispnv50/head.c
283
nv50_head_atomic_check_mode(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head.c
328
asyh->set.or = head->func->or != NULL;
drivers/gpu/drm/nouveau/dispnv50/head.c
340
struct nv50_head *head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/head.c
353
ret = nv50_head_atomic_check_lut(head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
377
asyh->set.or = head->func->or != NULL;
drivers/gpu/drm/nouveau/dispnv50/head.c
381
nv50_head_atomic_check_mode(head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
395
if (head->func->core_calc) {
drivers/gpu/drm/nouveau/dispnv50/head.c
396
head->func->core_calc(head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
40
nv50_head_flush_clr(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/head.c
444
ret = nv50_crc_atomic_check_head(head, asyh, armh);
drivers/gpu/drm/nouveau/dispnv50/head.c
47
lockdep_assert_held(&head->disp->mutex);
drivers/gpu/drm/nouveau/dispnv50/head.c
49
if (clr.crc) nv50_crc_atomic_clr(head);
drivers/gpu/drm/nouveau/dispnv50/head.c
50
if (clr.olut) head->func->olut_clr(head);
drivers/gpu/drm/nouveau/dispnv50/head.c
51
if (clr.core) head->func->core_clr(head);
drivers/gpu/drm/nouveau/dispnv50/head.c
517
struct nv50_head *head = nv50_head(crtc);
drivers/gpu/drm/nouveau/dispnv50/head.c
519
nvif_event_dtor(&head->base.vblank);
drivers/gpu/drm/nouveau/dispnv50/head.c
52
if (clr.curs) head->func->curs_clr(head);
drivers/gpu/drm/nouveau/dispnv50/head.c
520
nvif_head_dtor(&head->base.head);
drivers/gpu/drm/nouveau/dispnv50/head.c
521
nv50_lut_fini(&head->olut);
drivers/gpu/drm/nouveau/dispnv50/head.c
523
kfree(head);
drivers/gpu/drm/nouveau/dispnv50/head.c
56
nv50_head_flush_set_wndw(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head.c
573
struct nv50_head *head;
drivers/gpu/drm/nouveau/dispnv50/head.c
58
if (asyh->set.curs ) head->func->curs_set(head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
580
head = kzalloc_obj(*head);
drivers/gpu/drm/nouveau/dispnv50/head.c
581
if (!head)
drivers/gpu/drm/nouveau/dispnv50/head.c
584
head->func = disp->core->func->head;
drivers/gpu/drm/nouveau/dispnv50/head.c
585
head->disp = disp;
drivers/gpu/drm/nouveau/dispnv50/head.c
586
head->base.index = index;
drivers/gpu/drm/nouveau/dispnv50/head.c
594
ret = nv50_base_new(drm, head->base.index, &base);
drivers/gpu/drm/nouveau/dispnv50/head.c
595
ret = nv50_ovly_new(drm, head->base.index, &ovly);
drivers/gpu/drm/nouveau/dispnv50/head.c
598
head->base.index * 2 + 0, &base);
drivers/gpu/drm/nouveau/dispnv50/head.c
60
asyh->olut.offset = nv50_lut_load(&head->olut,
drivers/gpu/drm/nouveau/dispnv50/head.c
600
head->base.index * 2 + 1, &ovly);
drivers/gpu/drm/nouveau/dispnv50/head.c
603
ret = nv50_curs_new(drm, head->base.index, &curs);
drivers/gpu/drm/nouveau/dispnv50/head.c
605
kfree(head);
drivers/gpu/drm/nouveau/dispnv50/head.c
609
nv_crtc = &head->base;
drivers/gpu/drm/nouveau/dispnv50/head.c
612
funcs, "head-%d", head->base.index);
drivers/gpu/drm/nouveau/dispnv50/head.c
618
head->func->olut_size);
drivers/gpu/drm/nouveau/dispnv50/head.c
620
if (head->func->olut_set) {
drivers/gpu/drm/nouveau/dispnv50/head.c
621
ret = nv50_lut_init(disp, &drm->client.mmu, &head->olut);
drivers/gpu/drm/nouveau/dispnv50/head.c
628
ret = nvif_head_ctor(disp->disp, head->base.base.name, head->base.index, &head->base.head);
drivers/gpu/drm/nouveau/dispnv50/head.c
632
ret = nvif_head_vblank_event_ctor(&head->base.head, "kmsVbl", nv50_head_vblank_handler,
drivers/gpu/drm/nouveau/dispnv50/head.c
637
return head;
drivers/gpu/drm/nouveau/dispnv50/head.c
64
head->func->olut_set(head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
69
nv50_head_flush_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head.c
71
lockdep_assert_held(&head->disp->mutex);
drivers/gpu/drm/nouveau/dispnv50/head.c
73
if (asyh->set.view ) head->func->view (head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
74
if (asyh->set.mode ) head->func->mode (head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
75
if (asyh->set.core ) head->func->core_set(head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
76
if (asyh->set.base ) head->func->base (head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
77
if (asyh->set.ovly ) head->func->ovly (head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
78
if (asyh->set.dither ) head->func->dither (head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
79
if (asyh->set.procamp) head->func->procamp (head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
80
if (asyh->set.crc ) nv50_crc_atomic_set (head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.c
81
if (asyh->set.or ) head->func->or (head, asyh);
drivers/gpu/drm/nouveau/dispnv50/head.h
25
void nv50_head_flush_set(struct nv50_head *head, struct nv50_head_atom *asyh);
drivers/gpu/drm/nouveau/dispnv50/head.h
26
void nv50_head_flush_set_wndw(struct nv50_head *head, struct nv50_head_atom *asyh);
drivers/gpu/drm/nouveau/dispnv50/head.h
27
void nv50_head_flush_clr(struct nv50_head *head,
drivers/gpu/drm/nouveau/dispnv50/head507d.c
123
head507d_curs_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
125
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
126
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
140
head507d_curs_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
142
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
143
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
163
head507d_curs_format(struct nv50_head *head, struct nv50_wndw_atom *asyw,
drivers/gpu/drm/nouveau/dispnv50/head507d.c
176
head507d_curs_layout(struct nv50_head *head, struct nv50_wndw_atom *asyw,
drivers/gpu/drm/nouveau/dispnv50/head507d.c
189
head507d_core_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
191
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
192
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
203
head507d_core_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
205
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
206
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
247
head507d_core_calc(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
249
struct nv50_disp *disp = nv50_disp(head->base.base.dev);
drivers/gpu/drm/nouveau/dispnv50/head507d.c
279
head507d_olut_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
281
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
282
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
294
head507d_olut_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
296
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
297
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
30
head507d_procamp(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
32
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
33
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
331
head507d_olut(struct nv50_head *head, struct nv50_head_atom *asyh, int size)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
346
head507d_mode(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
348
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
350
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
401
head507d_view(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
403
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
404
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
49
head507d_dither(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
51
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
52
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
67
head507d_ovly(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
69
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
70
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
95
head507d_base(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
97
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head507d.c
98
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head827d.c
111
head827d_olut_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/head827d.c
113
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head827d.c
114
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head827d.c
128
head827d_olut_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head827d.c
130
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head827d.c
131
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head827d.c
30
head827d_curs_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/head827d.c
32
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head827d.c
33
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head827d.c
49
head827d_curs_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head827d.c
51
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head827d.c
52
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head827d.c
74
head827d_core_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head827d.c
76
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head827d.c
77
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
125
head907d_base(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
127
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
128
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
153
head907d_curs_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
155
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
156
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
172
head907d_curs_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
174
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
175
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
196
head907d_core_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
198
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
199
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
210
head907d_core_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
212
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
213
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
247
head907d_olut_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
249
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
250
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
264
head907d_olut_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
266
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
267
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
303
head907d_olut(struct nv50_head *head, struct nv50_head_atom *asyh, int size)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
323
head907d_mode(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
325
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
327
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
37
head907d_or(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
379
head907d_view(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
381
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
382
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
39
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
40
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
52
HEAD_SET_CONTROL(i), 0x31ec6000 | head->base.index << 25 |
drivers/gpu/drm/nouveau/dispnv50/head907d.c
58
head907d_procamp(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
60
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
61
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
78
head907d_dither(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
80
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
81
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
96
head907d_ovly(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
98
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head907d.c
99
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head917d.c
103
head917d_curs_layout(struct nv50_head *head, struct nv50_wndw_atom *asyw,
drivers/gpu/drm/nouveau/dispnv50/head917d.c
31
head917d_dither(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head917d.c
33
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head917d.c
34
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head917d.c
49
head917d_base(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head917d.c
51
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head917d.c
52
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/head917d.c
78
head917d_curs_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/head917d.c
80
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/head917d.c
81
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
105
headc37d_curs_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
107
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
108
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
123
headc37d_curs_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
125
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
126
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
154
headc37d_curs_format(struct nv50_head *head, struct nv50_wndw_atom *asyw,
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
162
headc37d_olut_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
164
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
165
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
176
headc37d_olut_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
178
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
179
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
196
headc37d_olut(struct nv50_head *head, struct nv50_head_atom *asyh, int size)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
210
headc37d_mode(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
212
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
214
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
255
headc37d_view(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
257
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
258
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
275
headc37d_static_wndw_map(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
279
for (i = head->base.index * 2, end = i + 2; i < end; i++)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
31
headc37d_or(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
33
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
34
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
65
headc37d_procamp(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
67
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
68
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
86
headc37d_dither(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
88
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc37d.c
89
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
100
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
111
headc57d_olut_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
113
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
114
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
186
headc57d_olut(struct nv50_head *head, struct nv50_head_atom *asyh, int size)
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
202
headc57d_mode(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
204
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
206
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
31
headc57d_display_id(struct nv50_head *head, u32 display_id)
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
33
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
39
PUSH_NVSQ(push, NVC57D, 0x2020 + (head->base.index * 0x400), display_id);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
44
headc57d_or(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
46
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
47
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
79
headc57d_procamp(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
81
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
82
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
97
headc57d_olut_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
99
struct nvif_push *push = &nv50_disp(head->base.base.dev)->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
111
headca7d_curs_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
113
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
114
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
132
headca7d_curs_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
134
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
137
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
14
headca7d_display_id(struct nv50_head *head, u32 display_id)
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
16
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
17
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
170
headca7d_olut_clr(struct nv50_head *head)
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
172
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
173
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
187
headca7d_olut_set(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
189
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
192
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
218
headca7d_mode(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
220
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
222
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
258
headca7d_view(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
260
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
261
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
30
headca7d_or(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
32
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
33
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
71
headca7d_procamp(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
73
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
74
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
90
headca7d_dither(struct nv50_head *head, struct nv50_head_atom *asyh)
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
92
struct nvif_push *push = &head->disp->core->chan.push;
drivers/gpu/drm/nouveau/dispnv50/headca7d.c
93
const int i = head->base.index;
drivers/gpu/drm/nouveau/dispnv50/ovly.c
28
nv50_ovly_new(struct nouveau_drm *drm, int head, struct nv50_wndw **pwndw)
drivers/gpu/drm/nouveau/dispnv50/ovly.c
52
ret = ovlys[cid].new(drm, head, ovlys[cid].oclass, pwndw);
drivers/gpu/drm/nouveau/dispnv50/ovly.h
25
int nv50_ovly_new(struct nouveau_drm *, int head, struct nv50_wndw **);
drivers/gpu/drm/nouveau/dispnv50/ovly.h
7
struct nouveau_drm *, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/ovly507e.c
145
struct nouveau_drm *drm, int head, s32 oclass, u32 interlock_data,
drivers/gpu/drm/nouveau/dispnv50/ovly507e.c
149
.id = head,
drivers/gpu/drm/nouveau/dispnv50/ovly507e.c
156
"ovly", head, format, BIT(head),
drivers/gpu/drm/nouveau/dispnv50/ovly507e.c
177
ovly507e_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/ovly507e.c
180
return ovly507e_new_(&ovly507e, ovly507e_format, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/ovly507e.c
181
0x00000004 << (head * 8), pwndw);
drivers/gpu/drm/nouveau/dispnv50/ovly827e.c
115
ovly827e_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/ovly827e.c
118
return ovly507e_new_(&ovly827e, ovly827e_format, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/ovly827e.c
119
0x00000004 << (head * 8), pwndw);
drivers/gpu/drm/nouveau/dispnv50/ovly907e.c
91
ovly907e_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/ovly907e.c
94
return ovly507e_new_(&ovly907e, ovly907e_format, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/ovly907e.c
95
0x00000004 << (head * 4), pwndw);
drivers/gpu/drm/nouveau/dispnv50/ovly917e.c
37
ovly917e_new(struct nouveau_drm *drm, int head, s32 oclass,
drivers/gpu/drm/nouveau/dispnv50/ovly917e.c
40
return ovly507e_new_(&ovly907e, ovly917e_format, drm, head, oclass,
drivers/gpu/drm/nouveau/dispnv50/ovly917e.c
41
0x00000004 << (head * 4), pwndw);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
52
list_del(&ctxdma->head);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
775
list_for_each_entry_safe(ctxdma, ctxtmp, &wndw->ctxdma.list, head) {
drivers/gpu/drm/nouveau/dispnv50/wndw.c
78
list_for_each_entry(ctxdma, &wndw->ctxdma.list, head) {
drivers/gpu/drm/nouveau/dispnv50/wndw.c
85
list_add(&ctxdma->head, &wndw->ctxdma.list);
drivers/gpu/drm/nouveau/dispnv50/wndw.h
9
struct list_head head;
drivers/gpu/drm/nouveau/include/nvfw/ls.h
29
struct nvfw_ls_desc_head head;
drivers/gpu/drm/nouveau/include/nvfw/ls.h
41
struct nvfw_ls_desc_head head;
drivers/gpu/drm/nouveau/include/nvif/head.h
16
nvif_head_id(struct nvif_head *head)
drivers/gpu/drm/nouveau/include/nvif/head.h
18
return head->object.handle;
drivers/gpu/drm/nouveau/include/nvif/if0012.h
131
__u8 head;
drivers/gpu/drm/nouveau/include/nvif/if0012.h
171
__u8 head;
drivers/gpu/drm/nouveau/include/nvif/if0012.h
188
__u8 head;
drivers/gpu/drm/nouveau/include/nvif/if0012.h
197
__u8 head;
drivers/gpu/drm/nouveau/include/nvif/if0012.h
260
__u8 head;
drivers/gpu/drm/nouveau/include/nvif/if0012.h
287
__u8 head;
drivers/gpu/drm/nouveau/include/nvif/log.h
29
struct list_head head;
drivers/gpu/drm/nouveau/include/nvif/log.h
33
struct nvif_logs logs = { LIST_HEAD_INIT(logs.head) }
drivers/gpu/drm/nouveau/include/nvif/log.h
37
if (!list_empty(&logs->head)) {
drivers/gpu/drm/nouveau/include/nvif/log.h
40
list_for_each_entry_safe(log, n, &logs->head, entry) {
drivers/gpu/drm/nouveau/include/nvif/outp.h
111
int nvif_outp_dp_sst(struct nvif_outp *, int head, u32 watermark, u32 hblanksym, u32 vblanksym);
drivers/gpu/drm/nouveau/include/nvif/outp.h
114
int nvif_outp_dp_mst_vcpi(struct nvif_outp *, int head,
drivers/gpu/drm/nouveau/include/nvif/outp.h
92
int nvif_outp_hdmi(struct nvif_outp *, int head, bool enable, u8 max_ac_packet, u8 rekey, u32 khz,
drivers/gpu/drm/nouveau/include/nvif/outp.h
96
int nvif_outp_hda_eld(struct nvif_outp *, int head, void *data, u32 size);
drivers/gpu/drm/nouveau/include/nvkm/core/device.h
26
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/core/event.h
62
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/core/intr.h
41
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/core/intr.h
66
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/core/object.h
17
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/core/subdev.h
32
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/engine/disp.h
53
} wndw, head, dac, sor;
drivers/gpu/drm/nouveau/include/nvkm/engine/falcon.h
86
u32 head;
drivers/gpu/drm/nouveau/include/nvkm/engine/fifo.h
49
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/acr.h
77
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/init.h
12
int head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/init.h
28
.head = -1, \
drivers/gpu/drm/nouveau/include/nvkm/subdev/clk.h
56
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/clk.h
63
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/fb.h
113
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
474
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/i2c.h
25
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/i2c.h
45
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
8
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/timer.h
16
INIT_LIST_HEAD(&alarm->head);
drivers/gpu/drm/nouveau/include/nvkm/subdev/timer.h
7
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/top.h
21
struct list_head head;
drivers/gpu/drm/nouveau/include/nvkm/subdev/vga.h
11
u8 nvkm_rdvgas(struct nvkm_device *, int head, u8 index);
drivers/gpu/drm/nouveau/include/nvkm/subdev/vga.h
12
void nvkm_wrvgas(struct nvkm_device *, int head, u8 index, u8 value);
drivers/gpu/drm/nouveau/include/nvkm/subdev/vga.h
15
u8 nvkm_rdvgag(struct nvkm_device *, int head, u8 index);
drivers/gpu/drm/nouveau/include/nvkm/subdev/vga.h
16
void nvkm_wrvgag(struct nvkm_device *, int head, u8 index, u8 value);
drivers/gpu/drm/nouveau/include/nvkm/subdev/vga.h
19
u8 nvkm_rdvgac(struct nvkm_device *, int head, u8 index);
drivers/gpu/drm/nouveau/include/nvkm/subdev/vga.h
20
void nvkm_wrvgac(struct nvkm_device *, int head, u8 index, u8 value);
drivers/gpu/drm/nouveau/include/nvkm/subdev/vga.h
23
u8 nvkm_rdvgai(struct nvkm_device *, int head, u16 port, u8 index);
drivers/gpu/drm/nouveau/include/nvkm/subdev/vga.h
24
void nvkm_wrvgai(struct nvkm_device *, int head, u16 port, u8 index, u8 value);
drivers/gpu/drm/nouveau/include/nvkm/subdev/vga.h
7
u8 nvkm_rdport(struct nvkm_device *, int head, u16 port);
drivers/gpu/drm/nouveau/include/nvkm/subdev/vga.h
8
void nvkm_wrport(struct nvkm_device *, int head, u16 port, u8 value);
drivers/gpu/drm/nouveau/nouveau_abi16.c
103
list_for_each_entry(obj, &abi16->objects, head) {
drivers/gpu/drm/nouveau/nouveau_abi16.c
114
list_del(&obj->head);
drivers/gpu/drm/nouveau/nouveau_abi16.c
133
list_add_tail(&obj->head, &abi16->objects);
drivers/gpu/drm/nouveau/nouveau_abi16.c
167
list_del(&ntfy->head);
drivers/gpu/drm/nouveau/nouveau_abi16.c
188
list_for_each_entry_safe(ntfy, temp, &chan->notifiers, head) {
drivers/gpu/drm/nouveau/nouveau_abi16.c
207
list_del(&chan->head);
drivers/gpu/drm/nouveau/nouveau_abi16.c
219
list_for_each_entry_safe(obj, tmp, &abi16->objects, head) {
drivers/gpu/drm/nouveau/nouveau_abi16.c
224
list_for_each_entry_safe(chan, temp, &abi16->channels, head) {
drivers/gpu/drm/nouveau/nouveau_abi16.c
405
list_add(&chan->head, &abi16->channels);
drivers/gpu/drm/nouveau/nouveau_abi16.c
501
list_for_each_entry(chan, &abi16->channels, head) {
drivers/gpu/drm/nouveau/nouveau_abi16.c
604
list_add(&ntfy->head, &chan->notifiers);
drivers/gpu/drm/nouveau/nouveau_abi16.c
642
list_add(&ntfy->head, &chan->notifiers);
drivers/gpu/drm/nouveau/nouveau_abi16.c
701
list_for_each_entry(ntfy, &chan->notifiers, head) {
drivers/gpu/drm/nouveau/nouveau_abi16.c
95
struct list_head head; /* protected by nouveau_abi16.cli.mutex */
drivers/gpu/drm/nouveau/nouveau_abi16.h
17
struct list_head head;
drivers/gpu/drm/nouveau/nouveau_abi16.h
22
struct list_head head;
drivers/gpu/drm/nouveau/nouveau_bios.c
100
nouveau_bios_run_init_table(dev, scriptptr, dcbent, head);
drivers/gpu/drm/nouveau/nouveau_bios.c
102
nv04_dfp_bind_head(dev, dcbent, head, dl);
drivers/gpu/drm/nouveau/nouveau_bios.c
105
static int call_lvds_manufacturer_script(struct drm_device *dev, struct dcb_output *dcbent, int head, enum LVDS_script script)
drivers/gpu/drm/nouveau/nouveau_bios.c
118
run_digital_op_script(dev, scriptofs, dcbent, head, bios->fp.dual_link);
drivers/gpu/drm/nouveau/nouveau_bios.c
135
static int run_lvds_table(struct drm_device *dev, struct dcb_output *dcbent, int head, enum LVDS_script script, int pxclk)
drivers/gpu/drm/nouveau/nouveau_bios.c
203
run_digital_op_script(dev, scriptptr, dcbent, head, bios->fp.dual_link);
drivers/gpu/drm/nouveau/nouveau_bios.c
208
int call_lvds_script(struct drm_device *dev, struct dcb_output *dcbent, int head, enum LVDS_script script, int pxclk)
drivers/gpu/drm/nouveau/nouveau_bios.c
223
if (bios->fp.last_script_invoc == (script << 1 | head) || !lvds_ver ||
drivers/gpu/drm/nouveau/nouveau_bios.c
229
call_lvds_script(dev, dcbent, head, LVDS_INIT, pxclk);
drivers/gpu/drm/nouveau/nouveau_bios.c
233
call_lvds_script(dev, dcbent, head, LVDS_RESET, pxclk);
drivers/gpu/drm/nouveau/nouveau_bios.c
235
call_lvds_script(dev, dcbent, head, LVDS_PANEL_OFF, pxclk);
drivers/gpu/drm/nouveau/nouveau_bios.c
243
ret = call_lvds_manufacturer_script(dev, dcbent, head, script);
drivers/gpu/drm/nouveau/nouveau_bios.c
245
ret = run_lvds_table(dev, dcbent, head, script, pxclk);
drivers/gpu/drm/nouveau/nouveau_bios.c
247
bios->fp.last_script_invoc = (script << 1 | head);
drivers/gpu/drm/nouveau/nouveau_bios.c
625
int run_tmds_table(struct drm_device *dev, struct dcb_output *dcbent, int head, int pxclk)
drivers/gpu/drm/nouveau/nouveau_bios.c
672
run_digital_op_script(dev, scriptptr, dcbent, head, pxclk >= 165000);
drivers/gpu/drm/nouveau/nouveau_bios.c
92
struct dcb_output *dcbent, int head, bool dl)
drivers/gpu/drm/nouveau/nouveau_bios.c
98
NVWriteVgaCrtc(dev, 0, NV_CIO_CRE_44, head ? NV_CIO_CRE_44_HEADB :
drivers/gpu/drm/nouveau/nouveau_bios.h
174
int head, int pxclk);
drivers/gpu/drm/nouveau/nouveau_bios.h
175
int call_lvds_script(struct drm_device *, struct dcb_output *, int head,
drivers/gpu/drm/nouveau/nouveau_bo.c
1082
list_for_each_entry(vma, &nvbo->vma_list, head) {
drivers/gpu/drm/nouveau/nouveau_bo.c
1087
list_for_each_entry(vma, &nvbo->vma_list, head) {
drivers/gpu/drm/nouveau/nouveau_bo.c
229
INIT_LIST_HEAD(&nvbo->head);
drivers/gpu/drm/nouveau/nouveau_bo.h
21
struct list_head head;
drivers/gpu/drm/nouveau/nouveau_connector.c
61
list_for_each_entry(mode, &connector->probed_modes, head) {
drivers/gpu/drm/nouveau/nouveau_crtc.h
37
struct nvif_head head;
drivers/gpu/drm/nouveau/nouveau_debugfs.c
227
list_for_each_entry(cli, &drm->clients, head) {
drivers/gpu/drm/nouveau/nouveau_display.c
89
struct nvif_head *head = &nouveau_crtc(crtc)->head;
drivers/gpu/drm/nouveau/nouveau_display.c
97
ret = nvif_mthd(&head->object, NVIF_HEAD_V0_SCANOUTPOS, &args, sizeof(args));
drivers/gpu/drm/nouveau/nouveau_dmem.c
280
static void nouveau_dmem_folio_split(struct folio *head, struct folio *tail)
drivers/gpu/drm/nouveau/nouveau_dmem.c
284
tail->pgmap = head->pgmap;
drivers/gpu/drm/nouveau/nouveau_dmem.c
285
tail->mapping = head->mapping;
drivers/gpu/drm/nouveau/nouveau_dmem.c
286
folio_set_zone_device_data(tail, folio_zone_device_data(head));
drivers/gpu/drm/nouveau/nouveau_drm.c
1218
list_add(&cli->head, &drm->clients);
drivers/gpu/drm/nouveau/nouveau_drm.c
1256
list_del(&cli->head);
drivers/gpu/drm/nouveau/nouveau_drm.c
177
list_for_each_entry_safe(work, wtmp, &cli->worker, head) {
drivers/gpu/drm/nouveau/nouveau_drm.c
179
list_del(&work->head);
drivers/gpu/drm/nouveau/nouveau_drm.c
200
list_add_tail(&work->head, &cli->worker);
drivers/gpu/drm/nouveau/nouveau_drm.c
612
list_for_each_entry_safe(cli, temp_cli, &drm->clients, head) {
drivers/gpu/drm/nouveau/nouveau_drm.c
613
list_del(&cli->head);
drivers/gpu/drm/nouveau/nouveau_drv.h
109
struct list_head head;
drivers/gpu/drm/nouveau/nouveau_drv.h
122
struct list_head head;
drivers/gpu/drm/nouveau/nouveau_encoder.h
117
void (*update)(struct nouveau_encoder *, u8 head,
drivers/gpu/drm/nouveau/nouveau_exec.c
369
list_for_each_entry(chan16, &abi16->channels, head) {
drivers/gpu/drm/nouveau/nouveau_fence.c
129
list_for_each_entry_safe(fence, tmp, &fctx->pending, head) {
drivers/gpu/drm/nouveau/nouveau_fence.c
151
fence = list_first_entry_or_null(&fctx->pending, typeof(*fence), head);
drivers/gpu/drm/nouveau/nouveau_fence.c
236
list_add_tail(&fence->head, &fctx->pending);
drivers/gpu/drm/nouveau/nouveau_fence.c
489
list_del(&fence->head);
drivers/gpu/drm/nouveau/nouveau_fence.c
53
list_del(&fence->head);
drivers/gpu/drm/nouveau/nouveau_fence.c
84
list_for_each_entry_safe(fence, tmp, &fctx->pending, head) {
drivers/gpu/drm/nouveau/nouveau_fence.h
14
struct list_head head;
drivers/gpu/drm/nouveau/nouveau_gem.c
164
list_del_init(&vma->head);
drivers/gpu/drm/nouveau/nouveau_gem.c
767
list_for_each_entry(temp, &abi16->channels, head) {
drivers/gpu/drm/nouveau/nouveau_sched.c
308
list_add(&job->entry, &sched->job.list.head);
drivers/gpu/drm/nouveau/nouveau_sched.c
448
INIT_LIST_HEAD(&sched->job.list.head);
drivers/gpu/drm/nouveau/nouveau_sched.c
489
empty = list_empty(&sched->job.list.head);
drivers/gpu/drm/nouveau/nouveau_sched.h
107
struct list_head head;
drivers/gpu/drm/nouveau/nouveau_svm.c
213
list_del(&ivmm->head);
drivers/gpu/drm/nouveau/nouveau_svm.c
232
list_add(&ivmm->head, &svmm->vmm->cli->drm->svm->inst);
drivers/gpu/drm/nouveau/nouveau_svm.c
90
struct list_head head;
drivers/gpu/drm/nouveau/nouveau_svm.c
97
list_for_each_entry(ivmm, &svm->inst, head) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1091
list_for_each_entry(__job, &sched->job.list.head, entry) {
drivers/gpu/drm/nouveau/nouveau_vmm.c
53
list_for_each_entry(vma, &nvbo->vma_list, head) {
drivers/gpu/drm/nouveau/nouveau_vmm.c
70
list_del(&vma->head);
drivers/gpu/drm/nouveau/nouveau_vmm.c
97
list_add_tail(&vma->head, &nvbo->vma_list);
drivers/gpu/drm/nouveau/nouveau_vmm.h
10
struct list_head head;
drivers/gpu/drm/nouveau/nvif/head.c
30
nvif_head_vblank_event_ctor(struct nvif_head *head, const char *name, nvif_event_func func,
drivers/gpu/drm/nouveau/nvif/head.c
33
int ret = nvif_event_ctor(&head->object, name ?: "nvifHeadVBlank", nvif_head_id(head),
drivers/gpu/drm/nouveau/nvif/head.c
35
NVIF_ERRON(ret, &head->object, "[NEW EVENT:VBLANK]");
drivers/gpu/drm/nouveau/nvif/head.c
40
nvif_head_dtor(struct nvif_head *head)
drivers/gpu/drm/nouveau/nvif/head.c
42
nvif_object_dtor(&head->object);
drivers/gpu/drm/nouveau/nvif/head.c
46
nvif_head_ctor(struct nvif_disp *disp, const char *name, int id, struct nvif_head *head)
drivers/gpu/drm/nouveau/nvif/head.c
55
&args, sizeof(args), &head->object);
drivers/gpu/drm/nouveau/nvif/outp.c
196
nvif_outp_hda_eld(struct nvif_outp *outp, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvif/outp.c
205
mthd->head = head;
drivers/gpu/drm/nouveau/nvif/outp.c
209
NVIF_ERRON(ret, &outp->object, "[HDA_ELD head:%d size:%d]", head, size);
drivers/gpu/drm/nouveau/nvif/outp.c
226
nvif_outp_hdmi(struct nvif_outp *outp, int head, bool enable, u8 max_ac_packet, u8 rekey,
drivers/gpu/drm/nouveau/nvif/outp.c
233
args.head = head;
drivers/gpu/drm/nouveau/nvif/outp.c
246
args.head, args.enable, args.max_ac_packet, args.rekey, args.khz,
drivers/gpu/drm/nouveau/nvif/outp.c
29
nvif_outp_dp_mst_vcpi(struct nvif_outp *outp, int head,
drivers/gpu/drm/nouveau/nvif/outp.c
36
args.head = head;
drivers/gpu/drm/nouveau/nvif/outp.c
382
return ret ?: args.head;
drivers/gpu/drm/nouveau/nvif/outp.c
393
return ret ?: args.head;
drivers/gpu/drm/nouveau/nvif/outp.c
407
return ret ?: args.head;
drivers/gpu/drm/nouveau/nvif/outp.c
418
return ret ?: args.head;
drivers/gpu/drm/nouveau/nvif/outp.c
45
args.head, args.start_slot, args.num_slots, args.pbn, args.aligned_pbn);
drivers/gpu/drm/nouveau/nvif/outp.c
79
nvif_outp_dp_sst(struct nvif_outp *outp, int head, u32 watermark, u32 hblanksym, u32 vblanksym)
drivers/gpu/drm/nouveau/nvif/outp.c
85
args.head = head;
drivers/gpu/drm/nouveau/nvif/outp.c
92
args.head, args.watermark, args.hblanksym, args.vblanksym);
drivers/gpu/drm/nouveau/nvkm/core/event.c
158
INIT_LIST_HEAD(&ntfy->head);
drivers/gpu/drm/nouveau/nvkm/core/event.c
181
list_for_each_entry_safe(ntfy, ntmp, &event->ntfy, head) {
drivers/gpu/drm/nouveau/nvkm/core/event.c
85
list_del_init(&ntfy->head);
drivers/gpu/drm/nouveau/nvkm/core/event.c
93
list_add_tail(&ntfy->head, &ntfy->event->ntfy);
drivers/gpu/drm/nouveau/nvkm/core/intr.c
150
list_for_each_entry(intr, &device->intr.intr, head)
drivers/gpu/drm/nouveau/nvkm/core/intr.c
159
list_for_each_entry(intr, &device->intr.intr, head)
drivers/gpu/drm/nouveau/nvkm/core/intr.c
182
list_for_each_entry(intr, &device->intr.intr, head) {
drivers/gpu/drm/nouveau/nvkm/core/intr.c
196
list_for_each_entry(inth, &device->intr.prio[prio], head) {
drivers/gpu/drm/nouveau/nvkm/core/intr.c
212
list_for_each_entry(intr, &device->intr.intr, head) {
drivers/gpu/drm/nouveau/nvkm/core/intr.c
255
list_add_tail(&intr->head, &device->intr.intr);
drivers/gpu/drm/nouveau/nvkm/core/intr.c
302
list_for_each_entry(tdev, &device->top->device, head) {
drivers/gpu/drm/nouveau/nvkm/core/intr.c
322
list_for_each_entry(intr, &device->intr.intr, head)
drivers/gpu/drm/nouveau/nvkm/core/intr.c
328
list_for_each_entry(intr, &device->intr.intr, head) {
drivers/gpu/drm/nouveau/nvkm/core/intr.c
371
list_for_each_entry_safe(intr, intt, &device->intr.intr, head) {
drivers/gpu/drm/nouveau/nvkm/core/intr.c
372
list_del(&intr->head);
drivers/gpu/drm/nouveau/nvkm/core/intr.c
41
list_for_each_entry(tdev, &device->top->device, head) {
drivers/gpu/drm/nouveau/nvkm/core/intr.c
440
list_add_tail(&inth->head, &device->intr.prio[prio]);
drivers/gpu/drm/nouveau/nvkm/core/intr.c
81
list_for_each_entry(intr, &subdev->device->intr.intr, head) {
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
137
list_add_tail(&object->head, &parent->tree);
drivers/gpu/drm/nouveau/nvkm/core/object.c
166
list_for_each_entry_reverse(child, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
193
list_for_each_entry_continue_reverse(child, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
215
list_for_each_entry(child, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
226
list_for_each_entry_continue_reverse(child, &object->tree, head)
drivers/gpu/drm/nouveau/nvkm/core/object.c
244
list_for_each_entry_safe(child, ctemp, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
265
list_del(&object->head);
drivers/gpu/drm/nouveau/nvkm/core/object.c
281
INIT_LIST_HEAD(&object->head);
drivers/gpu/drm/nouveau/nvkm/core/subdev.c
236
list_del(&subdev->head);
drivers/gpu/drm/nouveau/nvkm/core/subdev.c
251
list_for_each_entry(subdev, &device->subdev, head) {
drivers/gpu/drm/nouveau/nvkm/core/subdev.c
276
list_add_tail(&subdev->head, &device->subdev);
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2921
list_for_each_entry(subdev, &device->subdev, head) {
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2963
list_for_each_entry_reverse(subdev, &device->subdev, head) {
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2981
list_for_each_entry_from(subdev, &device->subdev, head) {
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
3009
list_for_each_entry(subdev, &device->subdev, head) {
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
3060
list_for_each_entry(subdev, &device->subdev, head) {
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
3074
list_for_each_entry_from(subdev, &device->subdev, head)
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
3093
list_for_each_entry_safe_reverse(subdev, subtmp, &device->subdev, head)
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
3098
list_del(&device->head);
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
3165
list_add_tail(&device->head, &nv_devices);
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
39
list_for_each_entry(device, &nv_devices, head) {
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
106
list_for_each_entry(pstate, &clk->states, head) {
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
113
list_for_each_entry(cstate, &pstate->list, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
110
list_for_each_entry(outp, &disp->outps, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
125
list_for_each_entry(outp, &disp->outps, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
139
list_for_each_entry(ior, &disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
152
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
162
list_for_each_entry(head, &disp->heads, head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
163
i = max(i, head->id + 1);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
175
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
191
conn = list_first_entry(&disp->conns, typeof(*conn), head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
192
list_del(&conn->head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
197
outp = list_first_entry(&disp->outps, typeof(*outp), head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
198
list_del(&outp->head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
203
ior = list_first_entry(&disp->iors, typeof(*ior), head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
208
head = list_first_entry(&disp->heads, typeof(*head), head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
209
nvkm_head_del(&head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
42
struct nvkm_head *head = nvkm_head_find(disp, id);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
43
if (head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
44
head->func->vblank_put(head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
51
struct nvkm_head *head = nvkm_head_find(disp, id);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
52
if (head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
53
head->func->vblank_get(head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
63
nvkm_disp_vblank(struct nvkm_disp *disp, int head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c
65
nvkm_event_ntfy(&disp->vblank, head, NVKM_DISP_HEAD_EVENT_VBLANK);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
208
chan->head = args->v0.id;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
241
return nvkm_disp_chan_new_(disp, disp->head.nr, oclass, argv, argc, pobject);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.h
17
int head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/conn.h
14
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c
329
.head = { .cnt = nv50_head_cnt, .new = nv50_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c
33
g84_sor_hdmi_infoframe_vsi(struct nvkm_ior *ior, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c
37
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c
56
g84_sor_hdmi_infoframe_avi(struct nvkm_ior *ior, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c
60
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c
79
g84_sor_hdmi_ctrl(struct nvkm_ior *ior, int head, bool enable, u8 max_ac_packet, u8 rekey)
drivers/gpu/drm/nouveau/nvkm/engine/disp/g84.c
86
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/g94.c
179
list_for_each_entry(ior, &disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/g94.c
287
state->head = ctrl & 0x00000003;
drivers/gpu/drm/nouveau/nvkm/engine/disp/g94.c
34
g94_sor_dp_watermark(struct nvkm_ior *sor, int head, u8 watermark)
drivers/gpu/drm/nouveau/nvkm/engine/disp/g94.c
359
.head = { .cnt = nv50_head_cnt, .new = nv50_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/g94.c
43
g94_sor_dp_activesym(struct nvkm_ior *sor, int head,
drivers/gpu/drm/nouveau/nvkm/engine/disp/g94.c
54
g94_sor_dp_audio_sym(struct nvkm_ior *sor, int head, u16 h, u32 v)
drivers/gpu/drm/nouveau/nvkm/engine/disp/ga102.c
133
.head = { .cnt = gv100_head_cnt, .new = gv100_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1016
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1022
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1023
mask[head->id] = nvkm_rd32(device, 0x6101d4 + (head->id * 0x800));
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1024
HEAD_DBG(head, "%08x", mask[head->id]);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1030
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1031
if (!(mask[head->id] & 0x00001000))
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1033
nv50_disp_super_1_0(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1037
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1038
if (!(mask[head->id] & 0x00001000))
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
104
gf119_sor_dp_audio(struct nvkm_ior *sor, int head, bool enable)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1040
nv50_disp_super_2_0(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1042
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1043
if (!(mask[head->id] & 0x00010000))
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1045
nv50_disp_super_2_1(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1047
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1048
if (!(mask[head->id] & 0x00001000))
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1050
nv50_disp_super_2_2(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1054
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1055
if (!(mask[head->id] & 0x00001000))
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1057
nv50_disp_super_3_0(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1061
list_for_each_entry(head, &disp->heads, head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1062
nvkm_wr32(device, 0x6101d4 + (head->id * 0x800), 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
107
const u32 hoff = 0x800 * head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1105
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1143
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1144
const u32 hoff = head->id * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1145
u32 mask = 0x01000000 << head->id;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1149
nvkm_disp_vblank(disp, head->id);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1168
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1178
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1179
const u32 hoff = head->id * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
119
gf119_sor_dp_vcpi(struct nvkm_ior *sor, int head, u8 slot, u8 slot_nr, u16 pbn, u16 aligned)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
122
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1225
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1226
const u32 hoff = head->id * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
1242
.head = { .cnt = gf119_head_cnt, .new = gf119_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
205
gf119_sor_hdmi_infoframe_vsi(struct nvkm_ior *ior, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
209
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
231
gf119_sor_hdmi_infoframe_avi(struct nvkm_ior *ior, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
235
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
253
gf119_sor_hdmi_ctrl(struct nvkm_ior *ior, int head, bool enable, u8 max_ac_packet, u8 rekey)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
259
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
323
state->head = ctrl & 0x0000000f;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
37
gf119_sor_hda_device_entry(struct nvkm_ior *ior, int head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
374
state->head = ctrl & 0x0000000f;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
40
const u32 hoff = 0x800 * head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
400
gf119_head_vblank_put(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
402
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
403
const u32 hoff = head->id * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
408
gf119_head_vblank_get(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
410
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
411
const u32 hoff = head->id * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
416
gf119_head_rgclk(struct nvkm_head *head, int div)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
418
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
419
nvkm_mask(device, 0x612200 + (head->id * 0x800), 0x0000000f, div);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
42
nvkm_mask(device, 0x616548 + hoff, 0x00000070, head << 4);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
423
gf119_head_state(struct nvkm_head *head, struct nvkm_head_state *state)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
425
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
426
const u32 hoff = (state == &head->asy) * 0x20000 + head->id * 0x300;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
46
gf119_sor_hda_eld(struct nvkm_ior *ior, int head, u8 *data, u8 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
49
const u32 soff = 0x030 * ior->id + (head * 0x04);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
60
gf119_sor_hda_hpd(struct nvkm_ior *ior, int head, bool present)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
63
const u32 soff = 0x030 * ior->id + (head * 0x04);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
68
ior->func->hda->device_entry(ior, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
85
gf119_sor_dp_watermark(struct nvkm_ior *sor, int head, u8 watermark)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
88
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
94
gf119_sor_dp_audio_sym(struct nvkm_ior *sor, int head, u16 h, u32 v)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
97
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c
314
.head = { .cnt = gf119_head_cnt, .new = gf119_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c
33
gk104_sor_hdmi_infoframe_vsi(struct nvkm_ior *ior, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c
37
const u32 hoff = head * 0x400;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c
54
gk104_sor_hdmi_infoframe_avi(struct nvkm_ior *ior, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c
58
const u32 hoff = head * 0x400;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c
77
gk104_sor_hdmi_ctrl(struct nvkm_ior *ior, int head, bool enable, u8 max_ac_packet, u8 rekey)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c
83
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gk104.c
84
const u32 hdmi = head * 0x400;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gk110.c
40
.head = { .cnt = gf119_head_cnt, .new = gf119_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gm107.c
94
.head = { .cnt = gf119_head_cnt, .new = gf119_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gm200.c
177
.head = { .cnt = gf119_head_cnt, .new = gf119_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gp100.c
67
.head = { .cnt = gf119_head_cnt, .new = gf119_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gp102.c
182
.head = { .cnt = gf119_head_cnt, .new = gf119_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gt200.c
89
.head = { .cnt = nv50_head_cnt, .new = nv50_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c
119
gt215_sor_hdmi_infoframe_avi(struct nvkm_ior *ior, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c
141
gt215_sor_hdmi_ctrl(struct nvkm_ior *ior, int head, bool enable, u8 max_ac_packet, u8 rekey)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c
247
.head = { .cnt = nv50_head_cnt, .new = nv50_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c
35
gt215_sor_hda_eld(struct nvkm_ior *ior, int head, u8 *data, u8 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c
49
gt215_sor_hda_hpd(struct nvkm_ior *ior, int head, bool present)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c
68
gt215_sor_dp_audio(struct nvkm_ior *sor, int head, bool enable)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gt215.c
96
gt215_sor_hdmi_infoframe_vsi(struct nvkm_ior *ior, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
100
gv100_sor_hdmi_infoframe_vsi(struct nvkm_ior *ior, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1000
nvkm_wr32(device, 0x611854, 0x00010000 << head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1001
gv100_disp_exception(disp, 73 + head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1002
stat &= ~(0x00010000 << head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
104
const u32 hoff = head * 0x400;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1053
gv100_disp_intr_head_timing(struct nvkm_disp *disp, int head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1057
u32 stat = nvkm_rd32(device, 0x611800 + (head * 0x04));
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1061
nvkm_wr32(device, 0x611800 + (head * 0x04), stat & 0x00000003);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1066
nvkm_disp_vblank(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1067
nvkm_wr32(device, 0x611800 + (head * 0x04), 0x00000004);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1073
nvkm_wr32(device, 0x611800 + (head * 0x04), stat);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1084
int head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1087
for_each_set_bit(head, &mask, 8) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1088
gv100_disp_intr_head_timing(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1089
stat &= ~BIT(head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1128
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1154
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1155
const int id = head->id;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1201
nvkm_wr32(device, 0x611cec, disp->head.mask << 16 |
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1214
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1215
const u32 hoff = head->id * 4;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1235
.head = { .cnt = gv100_head_cnt, .new = gv100_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
124
gv100_sor_hdmi_infoframe_avi(struct nvkm_ior *ior, int head, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
128
const u32 hoff = head * 0x400;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
146
gv100_sor_hdmi_ctrl(struct nvkm_ior *ior, int head, bool enable, u8 max_ac_packet, u8 rekey)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
152
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
153
const u32 hdmi = head * 0x400;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
203
state->head = ctrl & 0x000000ff;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
243
gv100_head_vblank_put(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
245
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
246
nvkm_mask(device, 0x611d80 + (head->id * 4), 0x00000004, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
250
gv100_head_vblank_get(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
252
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
253
nvkm_mask(device, 0x611d80 + (head->id * 4), 0x00000004, 0x00000004);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
257
gv100_head_rgpos(struct nvkm_head *head, u16 *hline, u16 *vline)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
259
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
260
const u32 hoff = head->id * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
267
gv100_head_state(struct nvkm_head *head, struct nvkm_head_state *state)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
269
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
270
const u32 hoff = (state == &head->arm) * 0x8000 + head->id * 0x400;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
38
gv100_sor_hda_device_entry(struct nvkm_ior *ior, int head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
396
const u32 mask = 0x00000001 << chan->head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
41
const u32 hoff = 0x800 * head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
43
nvkm_mask(device, 0x616528 + hoff, 0x00000070, head << 4);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
522
const u32 mask = 0x00000001 << chan->head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
54
gv100_sor_dp_watermark(struct nvkm_ior *sor, int head, u8 watermark)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
57
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
571
const u32 mask = 0x00010000 << chan->head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
63
gv100_sor_dp_audio_sym(struct nvkm_ior *sor, int head, u16 h, u32 v)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
66
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
73
gv100_sor_dp_audio(struct nvkm_ior *sor, int head, bool enable)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
76
const u32 hoff = 0x800 * head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
840
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
847
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
848
mask[head->id] = nvkm_rd32(device, 0x6107ac + (head->id * 4));
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
849
HEAD_DBG(head, "%08x", mask[head->id]);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
855
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
856
if (!(mask[head->id] & 0x00001000))
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
858
nv50_disp_super_1_0(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
862
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
863
if (!(mask[head->id] & 0x00001000))
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
865
nv50_disp_super_2_0(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
867
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
868
if (!(mask[head->id] & 0x00010000))
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
870
nv50_disp_super_2_1(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
872
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
873
if (!(mask[head->id] & 0x00001000))
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
875
nv50_disp_super_2_2(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
879
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
880
if (!(mask[head->id] & 0x00001000))
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
882
nv50_disp_super_3_0(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
886
list_for_each_entry(head, &disp->heads, head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
887
nvkm_wr32(device, 0x6107ac + (head->id * 4), 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
990
int head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
999
for_each_set_bit(head, &mask, disp->wndw.nr) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
34
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
35
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
36
if (head->id == id)
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
37
return head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
45
struct nvkm_head *head = *phead;
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
46
if (head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
47
HEAD_DBG(head, "dtor");
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
48
list_del(&head->head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
58
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
59
if (!(head = kzalloc_obj(*head)))
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
61
head->func = func;
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
62
head->disp = disp;
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
63
head->id = id;
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
64
list_add_tail(&head->head, &disp->heads);
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c
65
HEAD_DBG(head, "ctor");
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.h
12
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.c
37
list_for_each_entry(ior, &disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.c
50
list_del(&ior->head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.c
69
list_add_tail(&ior->head, &disp->iors);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
100
void (*eld)(struct nvkm_ior *, int head, u8 *data, u8 size);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
101
void (*device_entry)(struct nvkm_ior *, int head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
19
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
35
unsigned head:8;
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
72
void (*ctrl)(struct nvkm_ior *, int head, bool enable, u8 max_ac_packet, u8 rekey);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
75
void (*infoframe_avi)(struct nvkm_ior *, int head, void *data, u32 size);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
76
void (*infoframe_vsi)(struct nvkm_ior *, int head, void *data, u32 size);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
77
void (*audio)(struct nvkm_ior *, int head, bool enable);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
87
int (*sst)(struct nvkm_ior *, int head, bool ef,
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
89
void (*vcpi)(struct nvkm_ior *, int head, u8 slot,
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
91
void (*audio)(struct nvkm_ior *, int head, bool enable);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
92
void (*audio_sym)(struct nvkm_ior *, int head, u16 h, u32 v);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
93
void (*activesym)(struct nvkm_ior *, int head,
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
95
void (*watermark)(struct nvkm_ior *, int head, u8 watermark);
drivers/gpu/drm/nouveau/nvkm/engine/disp/ior.h
99
void (*hpd)(struct nvkm_ior *, int head, bool present);
drivers/gpu/drm/nouveau/nvkm/engine/disp/mcp77.c
53
.head = { .cnt = nv50_head_cnt, .new = nv50_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/mcp89.c
67
.head = { .cnt = nv50_head_cnt, .new = nv50_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
30
nv04_head_vblank_put(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
32
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
33
nvkm_wr32(device, 0x600140 + (head->id * 0x2000) , 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
37
nv04_head_vblank_get(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
39
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
40
nvkm_wr32(device, 0x600140 + (head->id * 0x2000) , 0x00000001);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
44
nv04_head_rgpos(struct nvkm_head *head, u16 *hline, u16 *vline)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
46
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
47
u32 data = nvkm_rd32(device, 0x600868 + (head->id * 0x2000));
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
53
nv04_head_state(struct nvkm_head *head, struct nvkm_head_state *state)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
55
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv04.c
56
const u32 hoff = head->id * 0x0200;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
100
struct nvkm_head *head = nvkm_head_find(ior->disp, __ffs(state->head));
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1014
nv50_disp_super_iedt(struct nvkm_head *head, struct nvkm_outp *outp,
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1018
struct nvkm_bios *bios = head->disp->engine.subdev.device->bios;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
102
if (!WARN_ON(!head)) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1021
const u16 m = (0x0100 << head->id) | (l << 6) | outp->info.or;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1029
nv50_disp_super_ied_on(struct nvkm_head *head,
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
103
struct nvkm_head_state *state = &head->asy;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1032
struct nvkm_subdev *subdev = &head->disp->engine.subdev;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1046
data = nv50_disp_super_iedt(head, outp, &ver, &hdr, &cnt, &len, &iedt);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1053
if (head->asy.or.depth == 24)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1080
init.head = head->id;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1085
nv50_disp_super_ied_off(struct nvkm_head *head, struct nvkm_ior *ior, int id)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1097
data = nv50_disp_super_iedt(head, outp, &ver, &hdr, &cnt, &len, &iedt);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1101
nvbios_init(&head->disp->engine.subdev, iedt.script[id],
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1105
init.head = head->id;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1110
nv50_disp_super_ior_asy(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1113
list_for_each_entry(ior, &head->disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1114
if (ior->asy.head & (1 << head->id)) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1115
HEAD_DBG(head, "to %s", ior->name);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1119
HEAD_DBG(head, "nothing to attach");
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1124
nv50_disp_super_ior_arm(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1127
list_for_each_entry(ior, &head->disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1128
if (ior->arm.head & (1 << head->id)) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1129
HEAD_DBG(head, "on %s", ior->name);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1133
HEAD_DBG(head, "nothing attached");
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1138
nv50_disp_super_3_0(struct nvkm_disp *disp, struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1143
HEAD_DBG(head, "supervisor 3.0");
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1144
ior = nv50_disp_super_ior_asy(head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1149
nv50_disp_super_ied_on(head, ior, 1, head->asy.hz / 1000);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1157
nv50_disp_super_2_2_dp(struct nvkm_head *head, struct nvkm_ior *ior)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1159
struct nvkm_subdev *subdev = &head->disp->engine.subdev;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1160
const u32 khz = head->asy.hz / 1000;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1170
h = head->asy.hblanke + head->asy.htotal - head->asy.hblanks - 7;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1176
v = head->asy.vblanks - head->asy.vblanke - 25;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1181
ior->func->dp->audio_sym(ior, head->id, h, v);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1184
link_data_rate = (khz * head->asy.or.depth / 8) / ior->dp.nr;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1247
ior->func->dp->activesym(ior, head->id, bestTU, bestVTUa, bestVTUf, bestVTUi);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1259
ior->func->dp->watermark(ior, head->id, unk);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1263
nv50_disp_super_2_2(struct nvkm_disp *disp, struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1265
const u32 khz = head->asy.hz / 1000;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1270
HEAD_DBG(head, "supervisor 2.2");
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1271
ior = nv50_disp_super_ior_asy(head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1287
head->asy.or.depth = outp->lvds.bpc8 ? 24 : 18;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1292
nv50_disp_super_ied_on(head, ior, 0, khz);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1295
head->func->rgclk(head, ior->asy.rgdiv);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1299
nv50_disp_super_2_2_dp(head, ior);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1308
nv50_disp_super_2_1(struct nvkm_disp *disp, struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1311
const u32 khz = head->asy.hz / 1000;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1312
HEAD_DBG(head, "supervisor 2.1 - %d khz", khz);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1314
nvkm_devinit_pll_set(devinit, PLL_VPLL0 + head->id, khz);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1318
nv50_disp_super_2_0(struct nvkm_disp *disp, struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1323
HEAD_DBG(head, "supervisor 2.0");
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1324
ior = nv50_disp_super_ior_arm(head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1329
nv50_disp_super_ied_off(head, ior, 2);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1333
nv50_disp_super_1_0(struct nvkm_disp *disp, struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1338
HEAD_DBG(head, "supervisor 1.0");
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1339
ior = nv50_disp_super_ior_arm(head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
134
state->head = ctrl & 0x00000003;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1344
nv50_disp_super_ied_off(head, ior, 1);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1350
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1353
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1354
head->func->state(head, &head->arm);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1355
head->func->state(head, &head->asy);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1358
list_for_each_entry(ior, &disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1370
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1381
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1382
if (!(super & (0x00000020 << head->id)))
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1384
if (!(super & (0x00000080 << head->id)))
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1386
nv50_disp_super_1_0(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1390
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1391
if (!(super & (0x00000080 << head->id)))
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1393
nv50_disp_super_2_0(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1395
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1396
if (!(super & (0x00000200 << head->id)))
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1398
nv50_disp_super_2_1(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1400
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1401
if (!(super & (0x00000080 << head->id)))
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1403
nv50_disp_super_2_2(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1407
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1408
if (!(super & (0x00000080 << head->id)))
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1410
nv50_disp_super_3_0(disp, head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1519
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1531
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1532
tmp = nvkm_rd32(device, 0x616100 + (head->id * 0x800));
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1533
nvkm_wr32(device, 0x610190 + (head->id * 0x10), tmp);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1534
tmp = nvkm_rd32(device, 0x616104 + (head->id * 0x800));
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1535
nvkm_wr32(device, 0x610194 + (head->id * 0x10), tmp);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1536
tmp = nvkm_rd32(device, 0x616108 + (head->id * 0x800));
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1537
nvkm_wr32(device, 0x610198 + (head->id * 0x10), tmp);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1538
tmp = nvkm_rd32(device, 0x61610c + (head->id * 0x800));
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1539
nvkm_wr32(device, 0x61019c + (head->id * 0x10), tmp);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1601
disp->head.nr = func->head.cnt(disp, &disp->head.mask);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1602
nvkm_debug(subdev, " Head(s): %d (%02lx)\n", disp->head.nr, disp->head.mask);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1603
for_each_set_bit(i, &disp->head.mask, disp->head.nr) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1604
ret = func->head.new(disp, i);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1690
list_add_tail(&outp->head, &disp->outps);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1694
list_for_each_entry_safe(outp, outt, &disp->outps, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1708
list_for_each_entry(pair, &disp->outps, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1728
list_for_each_entry(conn, &disp->conns, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1743
list_del(&outp->head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1748
list_add_tail(&outp->conn->head, &disp->conns);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1754
list_for_each_entry(outp, &disp->outps, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1775
.head = { .cnt = nv50_head_cnt, .new = nv50_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
248
state->head = ctrl & 0x00000003;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
344
state->head = ctrl & 0x00000003;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
371
nv50_head_vblank_put(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
373
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
375
nvkm_mask(device, 0x61002c, (4 << head->id), 0);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
379
nv50_head_vblank_get(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
381
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
383
nvkm_mask(device, 0x61002c, (4 << head->id), (4 << head->id));
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
387
nv50_head_rgclk(struct nvkm_head *head, int div)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
389
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
391
nvkm_mask(device, 0x614200 + (head->id * 0x800), 0x0000000f, div);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
395
nv50_head_rgpos(struct nvkm_head *head, u16 *hline, u16 *vline)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
397
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
398
const u32 hoff = head->id * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
406
nv50_head_state(struct nvkm_head *head, struct nvkm_head_state *state)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
408
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
409
const u32 hoff = head->id * 0x540 + (state == &head->arm) * 4;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
494
u32 base = chan->head * mthd->addr;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
99
if (state->head && state == &ior->asy) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.c
122
list_for_each_entry(ior, &outp->disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.c
132
list_for_each_entry(ior, &outp->disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.c
171
list_for_each_entry(ior, &outp->disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.c
343
if (!ior->arm.head || ior->arm.proto != proto) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.c
344
OUTP_DBG(outp, "no heads (%x %d %d)", ior->arm.head,
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.c
351
if (ior->func->route.get && !ior->arm.head && outp->info.type == DCB_OUTPUT_DP)
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.c
40
list_for_each_entry(ior, &disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.c
49
list_for_each_entry(ior, &disp->iors, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.h
19
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.h
68
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/priv.h
18
void nvkm_disp_vblank(struct nvkm_disp *, int head);
drivers/gpu/drm/nouveau/nvkm/engine/disp/priv.h
35
} wndw, head, dac, sor, pior;
drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c
111
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c
137
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c
138
const int id = head->id;
drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c
185
nvkm_wr32(device, 0x611cec, disp->head.mask << 16 |
drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c
198
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c
199
const u32 hoff = head->id * 4;
drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c
219
.head = { .cnt = gv100_head_cnt, .new = gv100_head_new },
drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c
34
tu102_sor_dp_vcpi(struct nvkm_ior *sor, int head, u8 slot, u8 slot_nr, u16 pbn, u16 aligned)
drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c
37
const u32 hoff = head * 0x800;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
113
list_for_each_entry(outp, &conn->disp->outps, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
118
if (&outp->head == &conn->disp->outps)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
180
list_for_each_entry(cont, &disp->conns, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
102
list_for_each_entry(conn, &disp->conns, head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
106
list_for_each_entry(outp, &disp->outps, head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
110
list_for_each_entry(head, &disp->heads, head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
111
args->v0.head_mask |= BIT(head->id);
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
86
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
109
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
115
if (!(head = nvkm_head_find(disp, args->v0.id)))
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
120
if (!head->object.func) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
121
nvkm_object_ctor(&nvkm_uhead, oclass, &head->object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
122
*pobject = &head->object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
33
struct nvkm_head *head = nvkm_uhead(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
41
return nvkm_uevent_add(uevent, &head->disp->vblank, head->id,
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
46
nvkm_uhead_mthd_scanoutpos(struct nvkm_head *head, void *argv, u32 argc)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
53
head->func->state(head, &head->arm);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
54
args->v0.vtotal = head->arm.vtotal;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
55
args->v0.vblanks = head->arm.vblanks;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
56
args->v0.vblanke = head->arm.vblanke;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
57
args->v0.htotal = head->arm.htotal;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
58
args->v0.hblanks = head->arm.hblanks;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
59
args->v0.hblanke = head->arm.hblanke;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
69
head->func->rgpos(head, &args->v0.hline, &args->v0.vline);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
77
struct nvkm_head *head = nvkm_uhead(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
80
case NVIF_HEAD_V0_SCANOUTPOS: return nvkm_uhead_mthd_scanoutpos(head, argv, argc);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
89
struct nvkm_head *head = nvkm_uhead(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
90
struct nvkm_disp *disp = head->disp;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
93
head->object.func = NULL;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
189
if (!ior->hda || !nvkm_head_find(outp->disp, args->v0.head))
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
196
ior->func->dp->audio(ior, args->v0.head, true);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
199
ior->func->hdmi->audio(ior, args->v0.head, true);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
201
ior->func->hda->hpd(ior, args->v0.head, true);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
202
ior->func->hda->eld(ior, args->v0.head, args->v0.data, argc);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
204
ior->func->hda->hpd(ior, args->v0.head, false);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
207
ior->func->dp->audio(ior, args->v0.head, false);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
210
ior->func->hdmi->audio(ior, args->v0.head, false);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
225
if (!nvkm_head_find(outp->disp, args->v0.head))
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
230
ior->func->hdmi->infoframe_avi(ior, args->v0.head, &args->v0.data, size);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
233
ior->func->hdmi->infoframe_vsi(ior, args->v0.head, &args->v0.data, size);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
251
if (!(outp->asy.head = nvkm_head_find(outp->disp, args->v0.head)))
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
261
ior->func->hdmi->infoframe_avi(ior, args->v0.head, NULL, 0);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
262
ior->func->hdmi->infoframe_vsi(ior, args->v0.head, NULL, 0);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
263
ior->func->hdmi->ctrl(ior, args->v0.head, false, 0, 0);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
267
ior->func->hdmi->ctrl(ior, args->v0.head, args->v0.enable,
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
388
if (!ior || !ior->arm.head)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
40
if (!ior->func->dp || !ior->func->dp->vcpi || !nvkm_head_find(outp->disp, args->v0.head))
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
43
ior->func->dp->vcpi(ior, args->v0.head, args->v0.start_slot, args->v0.num_slots,
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
433
args->v0.head = ffs(ior->arm.head) - 1;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
596
list_for_each_entry(outt, &disp->outps, head) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
84
if (!ior->func->dp || !nvkm_head_find(disp, args->v0.head))
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
89
return ior->func->dp->sst(ior, args->v0.head,
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
100
return nvkm_rdport(device, head, 0x03d5);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
104
nvkm_wrvgac(struct nvkm_device *device, int head, u8 index, u8 value)
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
106
nvkm_wrport(device, head, 0x03d4, index);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
107
nvkm_wrport(device, head, 0x03d5, value);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
111
nvkm_rdvgai(struct nvkm_device *device, int head, u16 port, u8 index)
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
113
if (port == 0x03c4) return nvkm_rdvgas(device, head, index);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
114
if (port == 0x03ce) return nvkm_rdvgag(device, head, index);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
115
if (port == 0x03d4) return nvkm_rdvgac(device, head, index);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
120
nvkm_wrvgai(struct nvkm_device *device, int head, u16 port, u8 index, u8 value)
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
122
if (port == 0x03c4) nvkm_wrvgas(device, head, index, value);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
123
else if (port == 0x03ce) nvkm_wrvgag(device, head, index, value);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
124
else if (port == 0x03d4) nvkm_wrvgac(device, head, index, value);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
27
nvkm_rdport(struct nvkm_device *device, int head, u16 port)
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
35
return nvkm_rd08(device, 0x601000 + (head * 0x2000) + port);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
41
head = 0; /* CR44 selects head */
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
42
return nvkm_rd08(device, 0x0c0000 + (head * 0x2000) + port);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
49
nvkm_wrport(struct nvkm_device *device, int head, u16 port, u8 data)
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
57
nvkm_wr08(device, 0x601000 + (head * 0x2000) + port, data);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
63
head = 0; /* CR44 selects head */
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
64
nvkm_wr08(device, 0x0c0000 + (head * 0x2000) + port, data);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
69
nvkm_rdvgas(struct nvkm_device *device, int head, u8 index)
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
71
nvkm_wrport(device, head, 0x03c4, index);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
72
return nvkm_rdport(device, head, 0x03c5);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
76
nvkm_wrvgas(struct nvkm_device *device, int head, u8 index, u8 value)
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
78
nvkm_wrport(device, head, 0x03c4, index);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
79
nvkm_wrport(device, head, 0x03c5, value);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
83
nvkm_rdvgag(struct nvkm_device *device, int head, u8 index)
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
85
nvkm_wrport(device, head, 0x03ce, index);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
86
return nvkm_rdport(device, head, 0x03cf);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
90
nvkm_wrvgag(struct nvkm_device *device, int head, u8 index, u8 value)
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
92
nvkm_wrport(device, head, 0x03ce, index);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
93
nvkm_wrport(device, head, 0x03cf, value);
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
97
nvkm_rdvgac(struct nvkm_device *device, int head, u8 index)
drivers/gpu/drm/nouveau/nvkm/engine/disp/vga.c
99
nvkm_wrport(device, head, 0x03d4, index);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c
344
list_for_each_entry_safe(runl, runt, &fifo->runls, head)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c
346
list_for_each_entry_safe(runq, rtmp, &fifo->runqs, head)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.c
110
list_del(&vctx->head);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.c
127
vctx = nvkm_list_find(vctx, &cgrp->vctxs, head,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.c
152
list_add_tail(&vctx->head, &cgrp->vctxs);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.c
42
list_del(&ectx->head);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.c
63
ectx = nvkm_list_find(ectx, &cgrp->ectxs, head, ectx->engn == engn);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.c
78
list_add_tail(&ectx->head, &cgrp->ectxs);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.h
16
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.h
26
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.h
54
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.h
67
#define nvkm_cgrp_foreach_chan(chan,cgrp) list_for_each_entry((chan), &(cgrp)->chans, head)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.h
69
list_for_each_entry_safe((chan), (ctmp), &(cgrp)->chans, head)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
102
cctx = nvkm_list_find(cctx, &chan->cctxs, head,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
129
list_add_tail(&cctx->head, &chan->cctxs);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
168
if (list_empty(&chan->head))
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
174
list_del(&cgrp->head);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
177
list_del_init(&chan->head);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
201
if (WARN_ON(!list_empty(&chan->head))) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
207
list_add_tail(&chan->head, &cgrp->chans);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
210
list_add_tail(&cgrp->head, &cgrp->runl->cgrps);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
381
INIT_LIST_HEAD(&chan->head);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
82
list_del(&cctx->head);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.h
16
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ga100.c
473
nvkm_list_foreach(tdev, &device->top->device, head, tdev->runlist == runl->addr) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ga100.c
542
struct nvkm_engn *engn = list_first_entry(&runl->engns, typeof(*engn), head);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ga100.c
587
nvkm_list_foreach(tdev, &device->top->device, head, tdev->runlist >= 0) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
764
nvkm_list_foreach(tdev, &device->top->device, head, tdev->runlist >= 0) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
324
list_for_each_entry_safe(engn, engt, &runl->engns, head) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
325
list_del(&engn->head);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
332
list_del(&runl->head);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
360
list_add_tail(&engn->head, &runl->engns);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
415
list_add_tail(&runl->head, &fifo->runls);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
120
#define nvkm_runl_find_engn(engn,runl,cond) nvkm_list_find(engn, &(runl)->engns, head, (cond))
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
122
#define nvkm_runl_first(fifo) list_first_entry(&(fifo)->runls, struct nvkm_runl, head)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
123
#define nvkm_runl_foreach(runl,fifo) list_for_each_entry((runl), &(fifo)->runls, head)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
124
#define nvkm_runl_foreach_cond(runl,fifo,cond) nvkm_list_foreach(runl, &(fifo)->runls, head, (cond))
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
125
#define nvkm_runl_foreach_engn(engn,runl) list_for_each_entry((engn), &(runl)->engns, head)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
127
nvkm_list_foreach(engn, &(runl)->engns, head, (cond))
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
128
#define nvkm_runl_foreach_cgrp(cgrp,runl) list_for_each_entry((cgrp), &(runl)->cgrps, head)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
130
list_for_each_entry_safe((cgrp), (gtmp), &(runl)->cgrps, head)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
37
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
98
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runq.c
28
list_del(&runq->head);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runq.c
43
list_add_tail(&runq->head, &fifo->runqs);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runq.h
18
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runq.h
24
#define nvkm_runq_foreach(runq,fifo) list_for_each_entry((runq), &(fifo)->runqs, head)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runq.h
25
#define nvkm_runq_foreach_cond(runq,fifo,cond) nvkm_list_foreach(runq, &(fifo)->runqs, head, (cond))
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1726
u32 head = init->addr - base;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1727
u32 tail = head + init->count * init->pitch;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1728
while (head < tail) {
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1729
if (head != prev + 4 || xfer >= 32) {
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1735
addr = head;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1738
prev = head;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1740
head = head + init->pitch;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.h
312
#define pack_for_each_init(init, pack, head) \
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.h
313
for (pack = head; pack && pack->init; pack++) \
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
135
list_del(&chan->head);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
163
list_add(&chan->head, &gr->chan);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
252
list_for_each_entry(temp, &gr->chan, head) {
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
255
list_del(&chan->head);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
256
list_add(&chan->head, &gr->chan);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.h
27
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
118
list_add(&chan->head, &mpeg->chan);
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
157
list_for_each_entry(temp, &mpeg->chan, head) {
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
160
list_del(&chan->head);
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
161
list_add(&chan->head, &mpeg->chan);
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
47
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
90
list_del(&chan->head);
drivers/gpu/drm/nouveau/nvkm/engine/sw/base.c
37
list_for_each_entry(chan, &sw->chan, head) {
drivers/gpu/drm/nouveau/nvkm/engine/sw/base.c
40
list_del(&chan->head);
drivers/gpu/drm/nouveau/nvkm/engine/sw/base.c
41
list_add(&chan->head, &sw->chan);
drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.c
65
list_del(&chan->head);
drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.c
87
list_add(&chan->head, &sw->chan);
drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.h
15
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/falcon/cmdq.c
180
cmdq->head_reg = func->cmdq.head + index * func->cmdq.stride;
drivers/gpu/drm/nouveau/nvkm/falcon/cmdq.c
28
u32 head = nvkm_falcon_rd32(cmdq->qmgr->falcon, cmdq->head_reg);
drivers/gpu/drm/nouveau/nvkm/falcon/cmdq.c
34
if (head >= tail) {
drivers/gpu/drm/nouveau/nvkm/falcon/cmdq.c
35
free = cmdq->offset + cmdq->size - head;
drivers/gpu/drm/nouveau/nvkm/falcon/cmdq.c
40
head = cmdq->offset;
drivers/gpu/drm/nouveau/nvkm/falcon/cmdq.c
44
if (head < tail)
drivers/gpu/drm/nouveau/nvkm/falcon/cmdq.c
45
free = tail - head - 1;
drivers/gpu/drm/nouveau/nvkm/falcon/msgq.c
161
msgq->head_reg = falcon->func->msgq.head;
drivers/gpu/drm/nouveau/nvkm/falcon/msgq.c
182
msgq->head_reg = func->msgq.head + index * func->msgq.stride;
drivers/gpu/drm/nouveau/nvkm/falcon/msgq.c
46
u32 head = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->head_reg);
drivers/gpu/drm/nouveau/nvkm/falcon/msgq.c
48
return head == tail;
drivers/gpu/drm/nouveau/nvkm/falcon/msgq.c
55
u32 head, tail, available;
drivers/gpu/drm/nouveau/nvkm/falcon/msgq.c
57
head = nvkm_falcon_rd32(falcon, msgq->head_reg);
drivers/gpu/drm/nouveau/nvkm/falcon/msgq.c
59
if (head < msgq->position)
drivers/gpu/drm/nouveau/nvkm/falcon/msgq.c
64
available = head - tail;
drivers/gpu/drm/nouveau/nvkm/nvfw/ls.c
77
nvfw_ls_desc_head(subdev, &hdr->head);
drivers/gpu/drm/nouveau/nvkm/nvfw/ls.c
95
nvfw_ls_desc_head(subdev, &hdr->head);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
236
list_for_each_entry_safe(lsfw, lsft, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
259
list_add_tail(&lsf->head, &acr->lsf);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
267
list_move(&rtos->head, &acr->lsf);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
275
list_for_each_entry_safe(lsfw, lsft, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
339
list_for_each_entry(hsfw, &acr->hsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
366
list_for_each_entry_safe(hsfw, hsft, &acr->hsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
368
list_del(&hsfw->head);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
37
list_for_each_entry(hsfw, &acr->hsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
378
list_for_each_entry_safe(lsf, lst, &acr->lsf, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
379
list_del(&lsf->head);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/base.c
65
list_for_each_entry(lsf, &acr->lsf, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/ga100.c
46
list_add_tail(&hsfw->head, &acr->hsfw);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/ga102.c
148
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/ga102.c
199
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/ga102.c
43
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/gm200.c
136
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/gm200.c
190
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/gm200.c
256
list_add_tail(&hsfw->head, &acr->hsfw);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/gm200.c
77
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/gp102.c
144
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/gp102.c
44
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/gp102.c
84
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
155
nvkm_acr_lsfw_from_desc(&nvfw_ls_desc(subdev, fw->data)->head, lsfw);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
175
nvkm_acr_lsfw_from_desc(&nvfw_ls_desc_v1(subdev, fw->data)->head, lsfw);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
34
list_del(&lsfw->head);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
42
list_for_each_entry_safe(lsfw, lsft, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
51
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
78
list_add_tail(&lsfw->head, &acr->lsfw);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/priv.h
110
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/acr/priv.h
68
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/acr/tu102.c
53
list_for_each_entry(lsfw, &acr->lsfw, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
114
if (init->head >= 0)
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
115
return init->head;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
217
return nvkm_rdport(init->subdev->device, init->head, port);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
225
nvkm_wrport(init->subdev->device, init->head, port, value);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
233
int head = init->head < 0 ? 0 : init->head;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
234
return nvkm_rdvgai(subdev->device, head, port, index);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
247
init->head = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
251
int head = init->head < 0 ? 0 : init->head;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
252
nvkm_wrvgai(device, head, port, index, value);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
258
init->head = 1;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
136
list_for_each_entry_from_reverse(cstate, &pstate->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
149
return list_last_entry(&pstate->list, typeof(*cstate), head);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
151
list_for_each_entry(cstate, &pstate->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
220
list_del(&cstate->head);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
259
list_add(&cstate->head, &pstate->list);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
275
list_for_each_entry(pstate, &clk->states, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
365
list_for_each_entry(cstate, &pstate->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
393
list_for_each_entry_safe(cstate, temp, &pstate->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
397
list_del(&pstate->head);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
464
list_add_tail(&pstate->head, &clk->states);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
482
list_for_each_entry(pstate, &clk->states, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
633
list_for_each_entry_safe(pstate, temp, &clk->states, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
686
list_add_tail(&func->pstates[idx].head, &clk->states);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ga100.c
35
int head = type - PLL_VPLL0;
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ga100.c
52
nvkm_wr32(device, 0x00ef00 + (head * 0x40), 0x02080004);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ga100.c
53
nvkm_wr32(device, 0x00ef18 + (head * 0x40), (N << 16) | fN);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ga100.c
54
nvkm_wr32(device, 0x00ef04 + (head * 0x40), (P << 16) | M);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ga100.c
55
nvkm_wr32(device, 0x00e9c0 + (head * 0x04), 0x00000001);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gv100.c
34
int head = type - PLL_VPLL0;
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gv100.c
51
nvkm_wr32(device, 0x00ef10 + (head * 0x40), fN << 16);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gv100.c
52
nvkm_wr32(device, 0x00ef04 + (head * 0x40), (P << 16) |
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/tu102.c
35
int head = type - PLL_VPLL0;
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/tu102.c
52
nvkm_wr32(device, 0x00ef10 + (head * 0x40), fN << 16);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/tu102.c
53
nvkm_wr32(device, 0x00ef04 + (head * 0x40), (P << 16) |
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/tu102.c
57
nvkm_wr32(device, 0x00ef0c + (head * 0x40), 0x00000900);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/tu102.c
58
nvkm_wr32(device, 0x00ef00 + (head * 0x40), 0x02000014);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1162
list_for_each_entry(cfg, &ram->cfg, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1168
if (&cfg->head == &ram->cfg)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1451
p = &list_last_entry(&ram->cfg, typeof(*cfg), head)->bios;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1480
list_add_tail(&cfg->head, &ram->cfg);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1513
list_for_each_entry_safe(cfg, tmp, &ram->cfg, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
968
list_for_each_entry(cfg, &ram->cfg, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/gh100.c
130
u32 head = nvkm_rd32(device, NV_PFSP_QUEUE_HEAD(0));
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/gh100.c
133
if (tail == head)
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/gh100.c
72
u32 head, tail;
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/gh100.c
74
head = nvkm_rd32(device, NV_PFSP_MSGQ_HEAD(0));
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/gh100.c
77
if (head == tail)
drivers/gpu/drm/nouveau/nvkm/subdev/fsp/gh100.c
80
return (tail - head) + sizeof(u32); /* TAIL points at last DWORD written. */
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
36
list_del(&event->head);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
97
list_add(&event->head, &client->events);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
120
ret = rmapi->disp->chan.set_pushbuf(chan->disp, chan->object.oclass, chan->head, NULL);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
125
(chan->object.oclass << 16) | chan->head,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
130
args->channelInstance = chan->head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1333
dcbE.heads = disp->head.mask;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1363
list_add_tail(&outp->head, &disp->outps);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1412
r535_disp_intr_head_timing(struct nvkm_disp *disp, int head)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1416
u32 stat = nvkm_rd32(device, 0x611c00 + (head * 0x04));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1419
nvkm_disp_vblank(disp, head);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1421
nvkm_wr32(device, 0x611800 + (head * 0x04), 0x00000002);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1432
int head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1434
for_each_set_bit(head, &mask, 8)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1435
r535_disp_intr_head_timing(disp, head);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1648
disp->head.nr = ctrl->numHeads;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1662
disp->head.mask = ctrl->headMask;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1665
for_each_set_bit(i, &disp->head.mask, disp->head.nr) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
190
ret = rmapi->disp->chan.set_pushbuf(chan->disp, chan->object.oclass, chan->head, chan->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
194
return rmapi->disp->chan.dmac_alloc(chan->disp, chan->object.oclass, chan->head,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
315
r535_sor_hda_eld(struct nvkm_ior *sor, int head, u8 *data, u8 size)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
334
ctrl->deviceEntry = head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
340
r535_sor_hda_hpd(struct nvkm_ior *sor, int head, bool present)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
354
ctrl->deviceEntry = head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
382
r535_sor_dp_audio(struct nvkm_ior *sor, int head, bool enable)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
404
r535_sor_dp_vcpi(struct nvkm_ior *sor, int head, u8 slot, u8 slot_nr, u16 pbn, u16 aligned_pbn)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
415
ctrl->head = head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
436
r535_sor_dp_sst(struct nvkm_ior *sor, int head, bool ef,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
448
ctrl->head = head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
50
case 0x7e: *psize = 0x01000; return 0x690000 + (chan->head * *psize);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
51
case 0x7b: *psize = 0x01000; return 0x6b0000 + (chan->head * *psize);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
52
case 0x7a: *psize = 0x01000; return 0x6d8000 + (chan->head * *psize);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
546
r535_sor_hdmi_audio(struct nvkm_ior *sor, int head, bool enable)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
549
const u32 hdmi = head * 0x400;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
561
r535_sor_hdmi_ctrl(struct nvkm_ior *sor, int head, bool enable, u8 max_ac_packet, u8 rekey)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
612
r535_head_vblank_put(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
614
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
616
nvkm_mask(device, 0x611d80 + (head->id * 4), 0x00000002, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
620
r535_head_vblank_get(struct nvkm_head *head)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
622
struct nvkm_device *device = head->disp->engine.subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
624
nvkm_wr32(device, 0x611800 + (head->id * 4), 0x00000002);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
625
nvkm_mask(device, 0x611d80 + (head->id * 4), 0x00000002, 0x00000002);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
629
r535_head_state(struct nvkm_head *head, struct nvkm_head_state *state)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
662
list_for_each_entry(conn, &disp->conns, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
677
list_add_tail(&conn->head, &disp->conns);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
735
r535_disp_get_active(struct nvkm_disp *disp, unsigned head, u32 *displayid)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
746
ctrl->head = head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
763
struct nvkm_head *head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
767
list_for_each_entry(head, &disp->heads, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
770
ret = rmapi->disp->get_active(disp, head->id, &displayid);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
829
ior->arm.head = BIT(head->id);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
193
list_add_tail(&item->head, &fbsr->items);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
226
list_for_each_entry(iobj, &imem->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
233
list_for_each_entry(iobj, &imem->boot, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
242
list_for_each_entry(item, &fbsr.items, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
270
list_for_each_entry(item, &fbsr.items, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
281
list_for_each_entry_safe(item, temp, &fbsr.items, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
282
list_del(&item->head);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
35
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
2041
list_add(&log->log.entry, &gsp_logs.head);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
383
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
447
list_add_tail(&reg->head, &gsp->registry_list);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
511
list_for_each_entry_safe(reg, n, &gsp->registry_list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
536
list_del(&reg->head);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
557
list_for_each_entry_safe(reg, n, &gsp->registry_list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
558
list_del(&reg->head);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
976
list_for_each_entry(event, &client->events, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/nvrm/disp.h
274
NvU32 head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/nvrm/disp.h
646
NvU32 head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
168
r570_disp_get_active(struct nvkm_disp *disp, unsigned head, u32 *displayid)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
179
ctrl->head = head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fbsr.c
108
list_for_each_entry(iobj, &imem->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fbsr.c
116
list_for_each_entry(iobj, &imem->boot, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fbsr.c
38
list_for_each_entry(iobj, &imem->boot, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fbsr.c
49
list_for_each_entry(iobj, &imem->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/nvrm/disp.h
216
NvU32 head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/nvrm/disp.h
261
NvU32 head;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/rm.h
89
int (*get_active)(struct nvkm_disp *, unsigned head, u32 *display_id);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/auxch.c
160
list_del(&aux->head);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/auxch.c
196
list_add_tail(&aux->head, &pad->i2c->aux);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
126
list_for_each_entry(aux, &i2c->aux, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
146
list_for_each_entry(aux, &i2c->aux, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
150
list_for_each_entry(bus, &i2c->bus, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
159
list_for_each_entry(pad, &i2c->pad, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
177
list_for_each_entry(pad, &i2c->pad, head)
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
179
list_for_each_entry(bus, &i2c->bus, head)
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
193
list_for_each_entry(pad, &i2c->pad, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
197
list_for_each_entry(bus, &i2c->bus, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
201
list_for_each_entry(aux, &i2c->aux, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
217
list_first_entry(&i2c->aux, typeof(*aux), head);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
223
list_first_entry(&i2c->bus, typeof(*bus), head);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
229
list_first_entry(&i2c->pad, typeof(*pad), head);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
39
list_for_each_entry(pad, &i2c->pad, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
411
list_for_each_entry(aux, &i2c->aux, head)
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
67
list_for_each_entry(bus, &i2c->bus, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
80
list_for_each_entry(aux, &i2c->aux, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/bus.c
199
list_del(&bus->head);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/bus.c
225
list_add_tail(&bus->head, &pad->i2c->bus);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/pad.c
104
list_add_tail(&pad->head, &i2c->pad);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/pad.c
89
list_del(&pad->head);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/pad.h
20
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c
117
list_for_each_entry(rail, &iccsense->rails, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c
137
list_for_each_entry_safe(sensor, tmps, &iccsense->sensors, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c
138
list_del(&sensor->head);
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c
141
list_for_each_entry_safe(rail, tmpr, &iccsense->rails, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c
142
list_del(&rail->head);
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c
195
list_add_tail(&sensor->head, &iccsense->sensors);
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c
208
list_for_each_entry(sensor, &iccsense->sensors, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c
291
list_add_tail(&rail->head, &iccsense->rails);
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c
302
list_for_each_entry(sensor, &iccsense->sensors, head)
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/priv.h
18
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/priv.h
9
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
172
list_for_each_entry_safe(iobj, itmp, &imem->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
173
list_move_tail(&iobj->head, &imem->boot);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
77
list_del(&iobj->head);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
88
list_add_tail(&iobj->head, &imem->list);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
163
list_for_each_entry(iobj, &imem->boot, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
170
list_for_each_entry(iobj, &imem->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
181
list_for_each_entry(iobj, &imem->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
191
list_for_each_entry(iobj, &imem->boot, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/priv.h
44
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
104
list_del(&ptp->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
113
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
124
list_for_each_entry(ptc, &mmu->ptc.list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
134
list_add(&ptc->head, &mmu->ptc.list);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
156
list_add_tail(&pt->head, &pt->ptc->item);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
190
pt = list_first_entry_or_null(&ptc->item, typeof(*pt), head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
194
list_del(&pt->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
223
list_for_each_entry(ptc, &mmu->ptc.list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
225
list_for_each_entry_safe(pt, tt, &ptc->item, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
227
list_del(&pt->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
238
list_for_each_entry_safe(ptc, ptct, &mmu->ptc.list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
240
list_del(&ptc->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
35
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
51
list_add(&ptp->head, &mmu->ptp.list);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
57
list_del(&ptp->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
74
ptp = list_first_entry_or_null(&mmu->ptp.list, typeof(*ptp), head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
93
list_add(&ptp->head, &mmu->ptp.list);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/priv.h
67
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
128
list_del_init(&umem->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
169
INIT_LIST_HEAD(&umem->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
183
list_add(&umem->head, &umem->object.client->umem);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
46
list_for_each_entry(umem, &master->umem, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.h
15
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1046
list_for_each_entry(vma, &vmm->list, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1079
vma = list_first_entry(&vmm->list, typeof(*vma), head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1080
list_del(&vma->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1105
list_add_tail(&vma->head, &vmm->list);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1193
list_add_tail(&vma->head, &vmm->list);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1217
list_add(&vma->head, &vmm->list);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
855
list_add(&new->head, &vma->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
869
list_del(&vma->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
911
list_del(&vma->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
954
#define node(root, dir) (((root)->head.dir == &vmm->list) ? NULL : \
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
955
list_entry((root)->head.dir, struct nvkm_vma, head))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
169
struct list_head head;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
154
list_for_each_entry(join, &vmm->join, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
331
list_for_each_entry(join, &vmm->join, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
333
list_del(&join->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
352
list_add_tail(&join->head, &vmm->join);
drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c
28
#define pack_for_each_init(init, pack, head) \
drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c
29
for (pack = head; pack && pack->init; pack++) \
drivers/gpu/drm/nouveau/nvkm/subdev/timer/base.c
117
list_del_init(&alarm->head);
drivers/gpu/drm/nouveau/nvkm/subdev/timer/base.c
122
list_for_each_entry(list, &tmr->alarms, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/timer/base.c
127
list_add_tail(&alarm->head, &list->head);
drivers/gpu/drm/nouveau/nvkm/subdev/timer/base.c
130
list = list_first_entry(&tmr->alarms, typeof(*list), head);
drivers/gpu/drm/nouveau/nvkm/subdev/timer/base.c
77
list_for_each_entry_safe(alarm, atemp, &tmr->alarms, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/timer/base.c
89
list_del_init(&alarm->head);
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
112
list_for_each_entry(info, &top->device, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
137
list_for_each_entry_safe(info, temp, &top->device, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
138
list_del(&info->head);
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
39
list_add_tail(&info->head, &top->device);
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
51
list_for_each_entry(info, &top->device, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
67
list_for_each_entry(info, &top->device, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
83
list_for_each_entry(info, &top->device, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
98
list_for_each_entry(info, &top->device, head) {
drivers/gpu/drm/omapdrm/omap_debugfs.c
58
list_for_each_entry(fb, &dev->mode_config.fb_list, head) {
drivers/gpu/drm/omapdrm/omap_encoder.c
83
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/qxl/qxl_display.c
1056
struct qxl_head *head;
drivers/gpu/drm/qxl/qxl_display.c
1057
head = &qdev->client_monitors_config->heads[output->index];
drivers/gpu/drm/qxl/qxl_display.c
1058
if (head->width)
drivers/gpu/drm/qxl/qxl_display.c
1059
pwidth = head->width;
drivers/gpu/drm/qxl/qxl_display.c
1060
if (head->height)
drivers/gpu/drm/qxl/qxl_display.c
1061
pheight = head->height;
drivers/gpu/drm/qxl/qxl_display.c
155
struct qxl_head *head;
drivers/gpu/drm/qxl/qxl_display.c
157
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/qxl/qxl_display.c
160
head = &qdev->client_monitors_config->heads[output->index];
drivers/gpu/drm/qxl/qxl_display.c
163
dev->mode_config.suggested_x_property, head->x);
drivers/gpu/drm/qxl/qxl_display.c
165
dev->mode_config.suggested_y_property, head->y);
drivers/gpu/drm/qxl/qxl_display.c
259
struct qxl_head *head;
drivers/gpu/drm/qxl/qxl_display.c
270
head = &qdev->client_monitors_config->heads[h];
drivers/gpu/drm/qxl/qxl_display.c
271
DRM_DEBUG_KMS("head %d is %dx%d\n", h, head->width, head->height);
drivers/gpu/drm/qxl/qxl_display.c
273
return qxl_add_mode(connector, head->width, head->height, true);
drivers/gpu/drm/qxl/qxl_display.c
307
struct qxl_head *head = &qdev->monitors_config->heads[i];
drivers/gpu/drm/qxl/qxl_display.c
309
if (head->y > 8192 || head->x > 8192 ||
drivers/gpu/drm/qxl/qxl_display.c
310
head->width > 8192 || head->height > 8192) {
drivers/gpu/drm/qxl/qxl_display.c
312
i, head->width, head->height,
drivers/gpu/drm/qxl/qxl_display.c
313
head->x, head->y);
drivers/gpu/drm/qxl/qxl_display.c
326
struct qxl_head head;
drivers/gpu/drm/qxl/qxl_display.c
337
head.id = i;
drivers/gpu/drm/qxl/qxl_display.c
338
head.flags = 0;
drivers/gpu/drm/qxl/qxl_display.c
339
head.surface_id = 0;
drivers/gpu/drm/qxl/qxl_display.c
344
head.width = mode->hdisplay;
drivers/gpu/drm/qxl/qxl_display.c
345
head.height = mode->vdisplay;
drivers/gpu/drm/qxl/qxl_display.c
346
head.x = crtc->x;
drivers/gpu/drm/qxl/qxl_display.c
347
head.y = crtc->y;
drivers/gpu/drm/qxl/qxl_display.c
351
head.x += qdev->dumb_heads[i].x;
drivers/gpu/drm/qxl/qxl_display.c
353
head.width = 0;
drivers/gpu/drm/qxl/qxl_display.c
354
head.height = 0;
drivers/gpu/drm/qxl/qxl_display.c
355
head.x = 0;
drivers/gpu/drm/qxl/qxl_display.c
356
head.y = 0;
drivers/gpu/drm/qxl/qxl_display.c
364
if (head.width == qdev->monitors_config->heads[i].width &&
drivers/gpu/drm/qxl/qxl_display.c
365
head.height == qdev->monitors_config->heads[i].height &&
drivers/gpu/drm/qxl/qxl_display.c
366
head.x == qdev->monitors_config->heads[i].x &&
drivers/gpu/drm/qxl/qxl_display.c
367
head.y == qdev->monitors_config->heads[i].y &&
drivers/gpu/drm/qxl/qxl_display.c
372
i, head.width, head.height, head.x, head.y,
drivers/gpu/drm/qxl/qxl_display.c
379
qdev->monitors_config->heads[i] = head;
drivers/gpu/drm/qxl/qxl_display.c
47
static bool qxl_head_enabled(struct qxl_head *head)
drivers/gpu/drm/qxl/qxl_display.c
49
return head->width && head->height;
drivers/gpu/drm/qxl/qxl_display.c
797
struct qxl_head *head;
drivers/gpu/drm/qxl/qxl_display.c
802
head = qdev->dumb_heads + i;
drivers/gpu/drm/qxl/qxl_display.c
803
head->x = surf->width;
drivers/gpu/drm/qxl/qxl_display.c
804
surf->width += head->width;
drivers/gpu/drm/qxl/qxl_display.c
805
if (surf->height < head->height)
drivers/gpu/drm/qxl/qxl_display.c
806
surf->height = head->height;
drivers/gpu/drm/qxl/qxl_drv.h
158
struct list_head head;
drivers/gpu/drm/qxl/qxl_image.c
122
drv_chunk = list_first_entry(&dimage->chunk_list, struct qxl_drm_chunk, head);
drivers/gpu/drm/qxl/qxl_image.c
53
list_add_tail(&chunk->head, &image->chunk_list);
drivers/gpu/drm/qxl/qxl_image.c
92
list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) {
drivers/gpu/drm/radeon/atombios_crtc.c
1695
list_for_each_entry(test_crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/atombios_crtc.c
1722
list_for_each_entry(test_crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/atombios_crtc.c
1762
list_for_each_entry(test_crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/atombios_crtc.c
2058
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/atombios_encoders.c
2141
list_for_each_entry(test_encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/atombios_encoders.c
2185
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/atombios_encoders.c
2496
list_for_each_entry(other_encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/atombios_encoders.c
2669
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/dce6_afmt.c
94
list_for_each_entry(encoder, &rdev_to_drm(rdev)->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/evergreen.c
1683
list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/evergreen.c
1708
list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/evergreen.c
1773
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/evergreen.c
1812
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/mkregtable.c
101
#define list_for_each_entry(pos, head, member) \
drivers/gpu/drm/radeon/mkregtable.c
102
for (pos = list_entry((head)->next, typeof(*pos), member); \
drivers/gpu/drm/radeon/mkregtable.c
103
&pos->member != (head); \
drivers/gpu/drm/radeon/mkregtable.c
81
static inline void list_add_tail(struct list_head *new, struct list_head *head)
drivers/gpu/drm/radeon/mkregtable.c
83
__list_add(new, head->prev, head);
drivers/gpu/drm/radeon/r100.c
468
list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/r100.c
499
list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/r100.c
610
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/r100.c
633
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/r600.c
1024
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/r600.c
957
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/r600_dpm.c
163
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/r600_dpm.c
189
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/r600_hdmi.c
133
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon.h
2838
struct list_head *head);
drivers/gpu/drm/radeon/radeon_acpi.c
739
head) {
drivers/gpu/drm/radeon/radeon_agp.c
133
struct radeon_agp_head *head;
drivers/gpu/drm/radeon/radeon_agp.c
135
head = kzalloc_obj(*head);
drivers/gpu/drm/radeon/radeon_agp.c
136
if (!head)
drivers/gpu/drm/radeon/radeon_agp.c
138
head->bridge = agp_find_bridge(pdev);
drivers/gpu/drm/radeon/radeon_agp.c
139
if (!head->bridge) {
drivers/gpu/drm/radeon/radeon_agp.c
140
head->bridge = agp_backend_acquire(pdev);
drivers/gpu/drm/radeon/radeon_agp.c
141
if (!head->bridge) {
drivers/gpu/drm/radeon/radeon_agp.c
142
kfree(head);
drivers/gpu/drm/radeon/radeon_agp.c
145
agp_copy_info(head->bridge, &head->agp_info);
drivers/gpu/drm/radeon/radeon_agp.c
146
agp_backend_release(head->bridge);
drivers/gpu/drm/radeon/radeon_agp.c
148
agp_copy_info(head->bridge, &head->agp_info);
drivers/gpu/drm/radeon/radeon_agp.c
150
if (head->agp_info.chipset == NOT_SUPPORTED) {
drivers/gpu/drm/radeon/radeon_agp.c
151
kfree(head);
drivers/gpu/drm/radeon/radeon_agp.c
154
INIT_LIST_HEAD(&head->memory);
drivers/gpu/drm/radeon/radeon_agp.c
155
head->cant_use_aperture = head->agp_info.cant_use_aperture;
drivers/gpu/drm/radeon/radeon_agp.c
156
head->page_mask = head->agp_info.page_mask;
drivers/gpu/drm/radeon/radeon_agp.c
157
head->base = head->agp_info.aper_base;
drivers/gpu/drm/radeon/radeon_agp.c
159
return head;
drivers/gpu/drm/radeon/radeon_audio.c
199
list_for_each_entry(encoder, &rdev_to_drm(rdev)->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_audio.c
763
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_connectors.c
1293
list_for_each_entry(list_connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_connectors.c
1856
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_connectors.c
1876
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_connectors.c
2374
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_connectors.c
363
struct drm_display_mode, head);
drivers/gpu/drm/radeon/radeon_connectors.c
387
list_for_each_entry(conflict, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_connectors.c
744
list_for_each_entry_safe(mode, t, &connector->probed_modes, head) {
drivers/gpu/drm/radeon/radeon_connectors.c
754
list_for_each_entry_safe(mode, t, &connector->probed_modes, head) {
drivers/gpu/drm/radeon/radeon_cursor.c
185
list_for_each_entry(crtc_p, &crtc->dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/radeon_device.c
1568
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_device.c
1574
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/radeon_device.c
1691
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/radeon_device.c
1730
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_display.c
1708
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_display.c
644
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head)
drivers/gpu/drm/radeon/radeon_display.c
777
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_display.c
811
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_drv.c
483
list_for_each_entry(crtc, &drm_dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/radeon_encoders.c
218
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_encoders.c
220
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_encoders.c
237
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_encoders.c
257
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_encoders.c
273
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/radeon_encoders.c
291
list_for_each_entry(other_encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_encoders.c
60
list_for_each_entry(clone_encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_encoders.c
81
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_irq_kms.c
109
list_for_each_entry(connector, &mode_config->connector_list, head)
drivers/gpu/drm/radeon/radeon_irq_kms.c
93
list_for_each_entry(connector, &mode_config->connector_list, head)
drivers/gpu/drm/radeon/radeon_legacy_crtc.c
1060
list_for_each_entry(crtci, &dev->mode_config.crtc_list, head)
drivers/gpu/drm/radeon/radeon_legacy_crtc.c
1072
list_for_each_entry(crtci, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/radeon_legacy_crtc.c
581
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_legacy_crtc.c
775
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_legacy_encoders.c
1539
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/radeon_legacy_encoders.c
1745
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/radeon/radeon_object.c
468
struct list_head *head, int ring)
drivers/gpu/drm/radeon/radeon_object.c
477
list_for_each_entry(lobj, head, list) {
drivers/gpu/drm/radeon/radeon_object.c
486
list_for_each_entry(lobj, head, list) {
drivers/gpu/drm/radeon/radeon_object.h
156
struct list_head *head, int ring);
drivers/gpu/drm/radeon/radeon_pm.c
1707
&ddev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/radeon_pm.c
1783
&ddev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/radeon_vm.c
131
struct list_head *head)
drivers/gpu/drm/radeon/radeon_vm.c
146
list_add(&list[0].list, head);
drivers/gpu/drm/radeon/radeon_vm.c
157
list_add(&list[idx++].list, head);
drivers/gpu/drm/radeon/rs600.c
330
list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/rs600.c
348
list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
drivers/gpu/drm/radeon/rs600.c
415
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/radeon/rs600.c
442
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/renesas/rcar-du/rcar_du_kms.c
976
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/renesas/rz-du/rzg2l_du_kms.c
465
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/renesas/shmobile/shmob_drm_crtc.c
429
struct drm_display_mode, head);
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
1966
head)
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
1990
head)
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
2507
list_for_each_entry_safe(plane, tmpp, plane_list, head)
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
2514
list_for_each_entry_safe(crtc, tmpc, crtc_list, head) {
drivers/gpu/drm/sti/sti_crtc.c
147
list_for_each_entry(p, &drm_dev->mode_config.plane_list, head) {
drivers/gpu/drm/sti/sti_crtc.c
268
head) {
drivers/gpu/drm/sti/sti_drv.c
46
list_for_each_entry(p, &drm_dev->mode_config.plane_list, head) {
drivers/gpu/drm/sti/sti_drv.c
62
list_for_each_entry(p, &drm_dev->mode_config.plane_list, head) {
drivers/gpu/drm/sti/sti_drv.c
83
list_for_each_entry(p, &dev->mode_config.plane_list, head) {
drivers/gpu/drm/sti/sti_dvo.c
427
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/sti/sti_hda.c
671
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/sti/sti_hdmi.c
1155
list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
drivers/gpu/drm/stm/lvds.c
704
struct drm_display_mode, head);
drivers/gpu/drm/stm/lvds.c
836
struct drm_display_mode, head);
drivers/gpu/drm/stm/lvds.c
909
struct drm_display_mode, head);
drivers/gpu/drm/tegra/drm.c
840
list_for_each_entry(fb, &drm->mode_config.fb_list, head) {
drivers/gpu/drm/tests/drm_connector_test.c
268
KUNIT_ASSERT_PTR_EQ(test, connector->head.next, &connector->head);
drivers/gpu/drm/tests/drm_connector_test.c
418
KUNIT_ASSERT_TRUE(test, list_empty(&connector->head));
drivers/gpu/drm/tests/drm_connector_test.c
423
KUNIT_ASSERT_PTR_EQ(test, connector->head.next, &priv->drm.mode_config.connector_list);
drivers/gpu/drm/tests/drm_connector_test.c
525
KUNIT_ASSERT_TRUE(test, list_empty(&priv->connector.head));
drivers/gpu/drm/tests/drm_connector_test.c
530
KUNIT_ASSERT_PTR_EQ(test, priv->connector.head.next, &priv->drm.mode_config.connector_list);
drivers/gpu/drm/tests/drm_framebuffer_test.c
525
KUNIT_ASSERT_TRUE(test, list_is_singular(&fb2.head));
drivers/gpu/drm/tests/drm_framebuffer_test.c
605
KUNIT_EXPECT_PTR_EQ(test, dev->mode_config.fb_list.prev, &fb1.head);
drivers/gpu/drm/tests/drm_framebuffer_test.c
606
KUNIT_EXPECT_PTR_EQ(test, dev->mode_config.fb_list.next, &fb1.head);
drivers/gpu/drm/tests/drm_hdmi_state_helper_test.c
54
preferred = list_first_entry_or_null(&connector->modes, struct drm_display_mode, head);
drivers/gpu/drm/tests/drm_hdmi_state_helper_test.c
55
list_for_each_entry(mode, &connector->modes, head)
drivers/gpu/drm/tests/drm_probe_helper_test.c
123
list_for_each_entry(mode, &connector->probed_modes, head)
drivers/gpu/drm/tests/drm_probe_helper_test.c
129
struct drm_display_mode, head);
drivers/gpu/drm/tests/drm_probe_helper_test.c
143
mode = list_next_entry(mode, head);
drivers/gpu/drm/tilcdc/tilcdc_external.c
49
list_for_each_entry(connector, &ddev->mode_config.connector_list, head) {
drivers/gpu/drm/tilcdc/tilcdc_external.c
65
list_for_each_entry(iter, &ddev->mode_config.encoder_list, head)
drivers/gpu/drm/ttm/ttm_execbuf_util.c
135
list_del(&entry->head);
drivers/gpu/drm/ttm/ttm_execbuf_util.c
136
list_add(&entry->head, list);
drivers/gpu/drm/ttm/ttm_execbuf_util.c
152
list_for_each_entry(entry, list, head) {
drivers/gpu/drm/ttm/ttm_execbuf_util.c
37
list_for_each_entry_continue_reverse(entry, list, head) {
drivers/gpu/drm/ttm/ttm_execbuf_util.c
52
list_for_each_entry(entry, list, head) {
drivers/gpu/drm/ttm/ttm_execbuf_util.c
89
list_for_each_entry(entry, list, head) {
drivers/gpu/drm/ttm/ttm_execbuf_util.c
96
entry = list_prev_entry(entry, head);
drivers/gpu/drm/ttm/ttm_execbuf_util.c
97
list_del(&safe->head);
drivers/gpu/drm/ttm/ttm_execbuf_util.c
98
list_add(&safe->head, dups);
drivers/gpu/drm/ttm/ttm_resource.c
708
struct ttm_resource *ttm_lru_first_res_or_null(struct list_head *head)
drivers/gpu/drm/ttm/ttm_resource.c
712
list_for_each_entry(lru, head, link) {
drivers/gpu/drm/vboxvideo/vbox_mode.c
136
list_for_each_entry(crtci, &vbox->ddev.mode_config.crtc_list, head) {
drivers/gpu/drm/vboxvideo/vbox_mode.c
160
head) {
drivers/gpu/drm/vboxvideo/vbox_mode.c
203
head) {
drivers/gpu/drm/vboxvideo/vbox_mode.c
455
list_for_each_entry(crtci, &vbox->ddev.mode_config.crtc_list, head) {
drivers/gpu/drm/vboxvideo/vbox_mode.c
783
list_for_each_entry_safe(mode, iterator, &connector->modes, head) {
drivers/gpu/drm/vboxvideo/vbox_mode.c
784
list_del(&mode->head);
drivers/gpu/drm/vc4/vc4_drv.h
701
struct list_head head;
drivers/gpu/drm/vc4/vc4_drv.h
806
struct vc4_exec_info, head);
drivers/gpu/drm/vc4/vc4_drv.h
813
struct vc4_exec_info, head);
drivers/gpu/drm/vc4/vc4_drv.h
822
struct vc4_exec_info, head);
drivers/gpu/drm/vc4/vc4_gem.c
551
list_move_tail(&exec->head, &vc4->render_job_list);
drivers/gpu/drm/vc4/vc4_gem.c
647
list_add_tail(&exec->head, &vc4->bin_job_list);
drivers/gpu/drm/vc4/vc4_gem.c
907
struct vc4_exec_info, head);
drivers/gpu/drm/vc4/vc4_gem.c
908
list_del(&exec->head);
drivers/gpu/drm/vc4/vc4_hdmi.c
446
list_for_each_entry(mode, &connector->probed_modes, head) {
drivers/gpu/drm/vc4/vc4_irq.c
152
list_move_tail(&exec->head, &vc4->bin_job_list);
drivers/gpu/drm/vc4/vc4_irq.c
169
list_move_tail(&exec->head, &vc4->job_done_list);
drivers/gpu/drm/virtio/virtgpu_drv.h
220
struct list_head head;
drivers/gpu/drm/virtio/virtgpu_ioctl.c
406
list_for_each_entry(cache_ent, &vgdev->cap_cache, head) {
drivers/gpu/drm/virtio/virtgpu_kms.c
275
list_for_each_entry_safe(cache_ent, tmp, &vgdev->cap_cache, head) {
drivers/gpu/drm/virtio/virtgpu_vq.c
1021
list_for_each_entry(search_ent, &vgdev->cap_cache, head) {
drivers/gpu/drm/virtio/virtgpu_vq.c
1029
list_add_tail(&cache_ent->head, &vgdev->cap_cache);
drivers/gpu/drm/virtio/virtgpu_vq.c
879
list_for_each_entry(cache_ent, &vgdev->cap_cache, head) {
drivers/gpu/drm/vmwgfx/ttm_object.c
124
struct list_head head;
drivers/gpu/drm/vmwgfx/ttm_object.c
145
hash_for_each_possible_rcu(tfile->ref_hash, hash, head, key) {
drivers/gpu/drm/vmwgfx/ttm_object.c
160
hash_for_each_possible(tfile->ref_hash, hash, head, key) {
drivers/gpu/drm/vmwgfx/ttm_object.c
332
hash_add_rcu(tfile->ref_hash, &ref->hash.head, ref->hash.key);
drivers/gpu/drm/vmwgfx/ttm_object.c
335
list_add_tail(&ref->head, &tfile->ref_list);
drivers/gpu/drm/vmwgfx/ttm_object.c
352
hash_del_rcu(&ref->hash.head);
drivers/gpu/drm/vmwgfx/ttm_object.c
353
list_del(&ref->head);
drivers/gpu/drm/vmwgfx/ttm_object.c
396
ref = list_entry(list, struct ttm_ref_object, head);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
470
void vmw_binding_res_list_kill(struct list_head *head)
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
474
vmw_binding_res_list_scrub(head);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
475
list_for_each_entry_safe(entry, next, head, res_list)
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
488
void vmw_binding_res_list_scrub(struct list_head *head)
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
492
list_for_each_entry(entry, head, res_list) {
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
500
list_for_each_entry(entry, head, res_list) {
drivers/gpu/drm/vmwgfx/vmwgfx_binding.h
227
extern void vmw_binding_res_list_kill(struct list_head *head);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.h
228
extern void vmw_binding_res_list_scrub(struct list_head *head);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
108
list_del(&entry->head);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
109
hash_del_rcu(&entry->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
128
list_for_each_entry_safe(entry, next, list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
129
list_del(&entry->head);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
136
list_add_tail(&entry->head, &entry->man->list);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
164
list_for_each_entry_safe(entry, next, list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
170
hash_add_rcu(entry->man->resources, &entry->hash.head,
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
172
list_move_tail(&entry->head, &entry->man->list);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
208
hash_add_rcu(man->resources, &cres->hash.head, cres->hash.key);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
213
list_add_tail(&cres->head, list);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
244
hash_for_each_possible_rcu(man->resources, hash, head, key) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
259
hash_del_rcu(&entry->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
260
list_del(&entry->head);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
262
list_add_tail(&entry->head, list);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
311
list_for_each_entry_safe(entry, next, &man->list, head)
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
48
struct list_head head;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
89
hash_for_each_possible_rcu(man->resources, hash, head, key) {
drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c
675
void vmw_cotable_add_resource(struct vmw_resource *res, struct list_head *head)
drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c
680
list_add_tail(head, &vcotbl->resource_list);
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
808
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
1274
struct list_head *head);
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
93
struct hlist_node head;
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
101
struct list_head head;
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
1163
list_add_tail(&reloc->head, &sw_context->bo_relocations);
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
1219
list_add_tail(&reloc->head, &sw_context->bo_relocations);
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
161
list_for_each_entry(entry, &sw_context->ctx_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
230
list_add_tail(&node->head, &sw_context->ctx_list);
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
3761
list_for_each_entry(reloc, &sw_context->bo_relocations, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
50
struct list_head head;
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
503
list_add_tail(&rel->head, &sw_context->res_relocations);
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
536
list_for_each_entry(rel, list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
729
list_for_each_entry(val, &sw_context->ctx_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
86
struct list_head head;
drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
178
list_add_tail(&fence->head, &fman->fence_list);
drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
192
list_for_each_entry_safe(fence, next_fence, &fman->fence_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
194
list_del_init(&fence->head);
drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
362
head);
drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
370
list_del_init(&fence->head);
drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
374
BUG_ON(!list_empty(&fence->head));
drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
59
if (!list_empty(&fence->head)) {
drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
72
list_del_init(&fence->head);
drivers/gpu/drm/vmwgfx/vmwgfx_fence.h
45
struct list_head head;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1293
list_for_each_entry(con, &dev->mode_config.connector_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1526
head) {
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
537
list_add_tail(&val_buf->head, &val_list);
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
620
list_add_tail(&val_buf->head, &val_list);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
169
hash_for_each_possible_rcu(ctx->sw_context->res_ht, hash, head, key) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
178
list_for_each_entry(entry, &ctx->bo_list, base.head) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
211
hash_for_each_possible_rcu(ctx->sw_context->res_ht, hash, head, key) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
220
list_for_each_entry(entry, &ctx->resource_ctx_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
227
list_for_each_entry(entry, &ctx->resource_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
261
hash_add_rcu(ctx->sw_context->res_ht, &bo_node->hash.head,
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
268
list_add_tail(&val_buf->head, &ctx->bo_list);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
308
hash_add_rcu(ctx->sw_context->res_ht, &node->hash.head, node->hash.key);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
312
hash_del_rcu(&node->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
318
list_add_tail(&node->head, &ctx->resource_list);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
323
list_add(&node->head, &ctx->resource_ctx_list);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
326
list_add_tail(&node->head, &ctx->resource_ctx_list);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
329
list_add_tail(&node->head, &ctx->resource_list);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
416
list_for_each_entry(val, &ctx->resource_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
470
list_for_each_entry(val, &ctx->resource_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
477
list_for_each_entry(val, &ctx->resource_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
539
list_for_each_entry(entry, &ctx->bo_list, base.head) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
588
list_for_each_entry(val, &ctx->resource_list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
634
list_for_each_entry(entry, &ctx->bo_list, base.head)
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
635
hash_del_rcu(&entry->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
637
list_for_each_entry(val, &ctx->resource_list, head)
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
638
hash_del_rcu(&val->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
640
list_for_each_entry(val, &ctx->resource_ctx_list, head)
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
641
hash_del_rcu(&val->hash.head);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
659
list_for_each_entry(entry, &ctx->bo_list, base.head) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
665
list_for_each_entry(val, &ctx->resource_list, head)
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
73
struct list_head head;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
837
list_for_each_entry(entry, &ctx->bo_list, base.head) {
drivers/gpu/drm/xe/abi/guc_communication_ctb_abi.h
48
u32 head;
drivers/gpu/drm/xe/xe_guc_ct.c
1010
desc_read(xe, h2g, head), h2g->info.tail);
drivers/gpu/drm/xe/xe_guc_ct.c
1122
trace_xe_guc_ct_h2g_flow_control(xe, h2g->info.head, h2g->info.tail,
drivers/gpu/drm/xe/xe_guc_ct.c
1133
trace_xe_guc_ct_g2h_flow_control(xe, g2h->info.head,
drivers/gpu/drm/xe/xe_guc_ct.c
1142
(desc_read(ct_to_xe(ct), (&ct->ctbs.g2h), tail) != ct->ctbs.g2h.info.head)
drivers/gpu/drm/xe/xe_guc_ct.c
1665
u32 tail, head, len, desc_status;
drivers/gpu/drm/xe/xe_guc_ct.c
1726
if (g2h->info.head > g2h->info.size) {
drivers/gpu/drm/xe/xe_guc_ct.c
1729
g2h->info.head, g2h->info.size);
drivers/gpu/drm/xe/xe_guc_ct.c
1743
avail = tail - g2h->info.head;
drivers/gpu/drm/xe/xe_guc_ct.c
1751
xe_map_memcpy_from(xe, msg, &g2h->cmds, sizeof(u32) * g2h->info.head,
drivers/gpu/drm/xe/xe_guc_ct.c
1760
head = (g2h->info.head + 1) % g2h->info.size;
drivers/gpu/drm/xe/xe_guc_ct.c
1764
if (avail + head > g2h->info.size) {
drivers/gpu/drm/xe/xe_guc_ct.c
1765
u32 avail_til_wrap = g2h->info.size - head;
drivers/gpu/drm/xe/xe_guc_ct.c
1768
&g2h->cmds, sizeof(u32) * head,
drivers/gpu/drm/xe/xe_guc_ct.c
1775
&g2h->cmds, sizeof(u32) * head,
drivers/gpu/drm/xe/xe_guc_ct.c
1797
g2h->info.head = (head + avail) % g2h->info.size;
drivers/gpu/drm/xe/xe_guc_ct.c
1798
desc_write(xe, g2h, head, g2h->info.head);
drivers/gpu/drm/xe/xe_guc_ct.c
1801
action, len, g2h->info.head, tail);
drivers/gpu/drm/xe/xe_guc_ct.c
1986
drm_printf(p, "\thead: %d\n", snapshot->info.head);
drivers/gpu/drm/xe/xe_guc_ct.c
1990
drm_printf(p, "\thead (memory): %d\n", snapshot->desc.head);
drivers/gpu/drm/xe/xe_guc_ct.c
415
h2g->info.head = 0;
drivers/gpu/drm/xe/xe_guc_ct.c
416
h2g->info.space = CIRC_SPACE(h2g->info.tail, h2g->info.head,
drivers/gpu/drm/xe/xe_guc_ct.c
432
g2h->info.head = 0;
drivers/gpu/drm/xe/xe_guc_ct.c
434
g2h->info.space = CIRC_SPACE(g2h->info.tail, g2h->info.head,
drivers/gpu/drm/xe/xe_guc_ct.c
771
h2g->info.head = desc_read(ct_to_xe(ct), h2g, head);
drivers/gpu/drm/xe/xe_guc_ct.c
773
if (h2g->info.head > h2g->info.size) {
drivers/gpu/drm/xe/xe_guc_ct.c
780
h2g->info.head, h2g->info.size);
drivers/gpu/drm/xe/xe_guc_ct.c
785
h2g->info.space = CIRC_SPACE(h2g->info.tail, h2g->info.head,
drivers/gpu/drm/xe/xe_guc_ct.c
940
u32 desc_head = desc_read(xe, h2g, head);
drivers/gpu/drm/xe/xe_guc_ct_types.h
29
u32 head;
drivers/gpu/drm/xe/xe_guc_submit.c
1013
q->guc->wqi_head = parallel_read(xe, map, wq_desc.head);
drivers/gpu/drm/xe/xe_guc_submit.c
2480
q->lrc[i]->ring.tail = job->ptrs[i].head;
drivers/gpu/drm/xe/xe_guc_submit.c
2541
q->lrc[i]->ring.tail = job->ptrs[i].head;
drivers/gpu/drm/xe/xe_guc_submit.c
3113
snapshot->parallel.wq_desc.head = parallel_read(xe, map, wq_desc.head);
drivers/gpu/drm/xe/xe_guc_submit.c
3118
if (snapshot->parallel.wq_desc.head !=
drivers/gpu/drm/xe/xe_guc_submit.c
3120
for (i = snapshot->parallel.wq_desc.head;
drivers/gpu/drm/xe/xe_guc_submit.c
3135
snapshot->guc.wqi_head, snapshot->parallel.wq_desc.head);
drivers/gpu/drm/xe/xe_guc_submit.c
3140
if (snapshot->parallel.wq_desc.head !=
drivers/gpu/drm/xe/xe_guc_submit.c
3142
for (i = snapshot->parallel.wq_desc.head;
drivers/gpu/drm/xe/xe_guc_submit_types.h
122
u32 head;
drivers/gpu/drm/xe/xe_guc_submit_types.h
31
u32 head;
drivers/gpu/drm/xe/xe_lrc.c
1688
void xe_lrc_set_ring_head(struct xe_lrc *lrc, u32 head)
drivers/gpu/drm/xe/xe_lrc.c
1691
xe_lrc_write_indirect_ctx_reg(lrc, INDIRECT_CTX_RING_HEAD, head);
drivers/gpu/drm/xe/xe_lrc.c
1693
xe_lrc_write_ctx_reg(lrc, CTX_RING_HEAD, head);
drivers/gpu/drm/xe/xe_lrc.c
1706
const u32 head = xe_lrc_ring_head(lrc);
drivers/gpu/drm/xe/xe_lrc.c
1710
return ((head - tail - 1) & (size - 1)) + 1;
drivers/gpu/drm/xe/xe_lrc.c
2280
snapshot->head = xe_lrc_ring_head(lrc);
drivers/gpu/drm/xe/xe_lrc.c
2339
drm_printf(p, "\tLRC Head: (memory) %u\n", snapshot->head);
drivers/gpu/drm/xe/xe_lrc.h
32
u32 head;
drivers/gpu/drm/xe/xe_lrc.h
99
void xe_lrc_set_ring_head(struct xe_lrc *lrc, u32 head);
drivers/gpu/drm/xe/xe_oa.c
147
static u32 xe_oa_circ_diff(struct xe_oa_stream *stream, u32 tail, u32 head)
drivers/gpu/drm/xe/xe_oa.c
149
return tail >= head ? tail - head :
drivers/gpu/drm/xe/xe_oa.c
150
tail + stream->oa_buffer.circ_size - head;
drivers/gpu/drm/xe/xe_oa.c
289
stream->oa_buffer.head, tail, hw_tail);
drivers/gpu/drm/xe/xe_oa.c
293
available = xe_oa_circ_diff(stream, stream->oa_buffer.tail, stream->oa_buffer.head);
drivers/gpu/drm/xe/xe_oa.c
354
u32 head, tail;
drivers/gpu/drm/xe/xe_oa.c
358
head = stream->oa_buffer.head;
drivers/gpu/drm/xe/xe_oa.c
363
head < stream->oa_buffer.circ_size && tail < stream->oa_buffer.circ_size);
drivers/gpu/drm/xe/xe_oa.c
365
for (; xe_oa_circ_diff(stream, tail, head);
drivers/gpu/drm/xe/xe_oa.c
366
head = xe_oa_circ_incr(stream, head, report_size)) {
drivers/gpu/drm/xe/xe_oa.c
367
u8 *report = oa_buf_base + head;
drivers/gpu/drm/xe/xe_oa.c
396
(head + gtt_offset) & OAG_OAHEADPTR_MASK);
drivers/gpu/drm/xe/xe_oa.c
397
stream->oa_buffer.head = head;
drivers/gpu/drm/xe/xe_oa.c
424
stream->oa_buffer.head = 0;
drivers/gpu/drm/xe/xe_oa_types.h
172
u32 head;
drivers/gpu/drm/xe/xe_pagefault.c
219
if (pf_queue->tail != pf_queue->head) {
drivers/gpu/drm/xe/xe_pagefault.c
384
for (i = pf_queue->tail; i != pf_queue->head;
drivers/gpu/drm/xe/xe_pagefault.c
414
return CIRC_SPACE(pf_queue->head, pf_queue->tail, pf_queue->size) <=
drivers/gpu/drm/xe/xe_pagefault.c
438
memcpy(pf_queue->data + pf_queue->head, pf, sizeof(*pf));
drivers/gpu/drm/xe/xe_pagefault.c
439
pf_queue->head = (pf_queue->head + xe_pagefault_entry_size()) %
drivers/gpu/drm/xe/xe_pagefault_types.h
127
u32 head;
drivers/gpu/drm/xe/xe_reg_whitelist.c
85
#define WHITELIST_OA_MMIO_TRG(trg, status, head) \
drivers/gpu/drm/xe/xe_reg_whitelist.c
88
WHITELIST(head, RING_FORCE_TO_NONPRIV_ACCESS_RD | RING_FORCE_TO_NONPRIV_RANGE_4)
drivers/gpu/drm/xe/xe_res_cursor.h
108
struct list_head *head, *next;
drivers/gpu/drm/xe/xe_res_cursor.h
111
head = &to_xe_ttm_vram_mgr_resource(res)->blocks;
drivers/gpu/drm/xe/xe_res_cursor.h
113
block = list_first_entry_or_null(head,
drivers/gpu/drm/xe/xe_res_cursor.h
123
if (next != head)
drivers/gpu/drm/xe/xe_ring_ops.c
261
u64 batch_addr, u32 *head, u32 seqno)
drivers/gpu/drm/xe/xe_ring_ops.c
267
*head = lrc->ring.tail;
drivers/gpu/drm/xe/xe_ring_ops.c
317
u64 batch_addr, u32 *head, u32 seqno)
drivers/gpu/drm/xe/xe_ring_ops.c
325
*head = lrc->ring.tail;
drivers/gpu/drm/xe/xe_ring_ops.c
372
u64 batch_addr, u32 *head,
drivers/gpu/drm/xe/xe_ring_ops.c
382
*head = lrc->ring.tail;
drivers/gpu/drm/xe/xe_ring_ops.c
426
struct xe_lrc *lrc, u32 *head,
drivers/gpu/drm/xe/xe_ring_ops.c
434
*head = lrc->ring.tail;
drivers/gpu/drm/xe/xe_ring_ops.c
469
&job->ptrs[0].head,
drivers/gpu/drm/xe/xe_ring_ops.c
479
&job->ptrs[0].head,
drivers/gpu/drm/xe/xe_ring_ops.c
487
&job->ptrs[i].head,
drivers/gpu/drm/xe/xe_ring_ops.c
499
&job->ptrs[i].head,
drivers/gpu/drm/xe/xe_ring_ops.c
510
&job->ptrs[i].head,
drivers/gpu/drm/xe/xe_sched_job_types.h
31
u32 head;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
26
struct list_head *head)
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
31
block = xe_ttm_vram_mgr_first_block(head);
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
35
while (head != block->link.next) {
drivers/gpu/vga/vga_switcheroo.c
394
find_client_from_pci(struct list_head *head, struct pci_dev *pdev)
drivers/gpu/vga/vga_switcheroo.c
398
list_for_each_entry(client, head, list)
drivers/gpu/vga/vga_switcheroo.c
405
find_client_from_id(struct list_head *head,
drivers/gpu/vga/vga_switcheroo.c
410
list_for_each_entry(client, head, list)
drivers/gpu/vga/vga_switcheroo.c
417
find_active_client(struct list_head *head)
drivers/gpu/vga/vga_switcheroo.c
421
list_for_each_entry(client, head, list)
drivers/greybus/gb-beagleplay.c
231
int head = smp_load_acquire(&bg->tx_circ_buf.head);
drivers/greybus/gb-beagleplay.c
233
int count = CIRC_CNT_TO_END(head, tail, TX_CIRC_BUF_SIZE);
drivers/greybus/gb-beagleplay.c
253
int tail, head = bg->tx_circ_buf.head;
drivers/greybus/gb-beagleplay.c
258
if (CIRC_SPACE(head, tail, TX_CIRC_BUF_SIZE) >= 1) {
drivers/greybus/gb-beagleplay.c
259
bg->tx_circ_buf.buf[head] = value;
drivers/greybus/gb-beagleplay.c
262
smp_store_release(&bg->tx_circ_buf.head,
drivers/greybus/gb-beagleplay.c
263
(head + 1) & (TX_CIRC_BUF_SIZE - 1));
drivers/greybus/gb-beagleplay.c
417
bg->tx_circ_buf.head = 0;
drivers/greybus/manifest.c
60
static void release_cport_descriptors(struct list_head *head, u8 bundle_id)
drivers/greybus/manifest.c
65
list_for_each_entry_safe(desc, tmp, head, links) {
drivers/hid/hid-kye.c
531
struct list_head *head;
drivers/hid/hid-kye.c
536
list_for_each(head, list) {
drivers/hid/hid-kye.c
537
report = list_entry(head, struct hid_report, list);
drivers/hid/hid-kye.c
542
if (head == list) {
drivers/hid/hid-wiimote-core.c
100
wdata->queue.head = newhead;
drivers/hid/hid-wiimote-core.c
103
wdata->queue.head = newhead;
drivers/hid/hid-wiimote-core.c
52
while (wdata->queue.head != wdata->queue.tail) {
drivers/hid/hid-wiimote-core.c
95
memcpy(wdata->queue.outq[wdata->queue.head].data, buffer, count);
drivers/hid/hid-wiimote-core.c
96
wdata->queue.outq[wdata->queue.head].size = count;
drivers/hid/hid-wiimote-core.c
97
newhead = (wdata->queue.head + 1) % WIIMOTE_BUFSIZE;
drivers/hid/hid-wiimote-core.c
99
if (wdata->queue.head == wdata->queue.tail) {
drivers/hid/hid-wiimote.h
111
__u8 head;
drivers/hid/hidraw.c
269
if (list->head != list->tail)
drivers/hid/hidraw.c
375
while (list->tail != list->head) {
drivers/hid/hidraw.c
579
int new_head = (list->head + 1) & (HIDRAW_BUFFER_SIZE - 1);
drivers/hid/hidraw.c
58
if (list->head == list->tail) {
drivers/hid/hidraw.c
584
if (!(list->buffer[list->head].value = kmemdup(data, len, GFP_ATOMIC))) {
drivers/hid/hidraw.c
588
list->buffer[list->head].len = len;
drivers/hid/hidraw.c
589
list->head = new_head;
drivers/hid/hidraw.c
62
while (list->head == list->tail) {
drivers/hid/uhid.c
100
uhid->head = newhead;
drivers/hid/uhid.c
52
__u8 head;
drivers/hid/uhid.c
686
if (uhid->head == uhid->tail)
drivers/hid/uhid.c
690
uhid->head != uhid->tail);
drivers/hid/uhid.c
699
if (uhid->head == uhid->tail) {
drivers/hid/uhid.c
793
if (uhid->head != uhid->tail)
drivers/hid/uhid.c
96
newhead = (uhid->head + 1) % UHID_BUFSIZE;
drivers/hid/uhid.c
99
uhid->outq[uhid->head] = ev;
drivers/hid/usbhid/hid-core.c
526
int head;
drivers/hid/usbhid/hid-core.c
534
if ((head = (usbhid->outhead + 1) & (HID_OUTPUT_FIFO_SIZE - 1)) == usbhid->outtail) {
drivers/hid/usbhid/hid-core.c
546
usbhid->outhead = head;
drivers/hid/usbhid/hid-core.c
582
if ((head = (usbhid->ctrlhead + 1) & (HID_CONTROL_FIFO_SIZE - 1)) == usbhid->ctrltail) {
drivers/hid/usbhid/hid-core.c
597
usbhid->ctrlhead = head;
drivers/hid/usbhid/hiddev.c
151
list->buffer[list->head] = *uref;
drivers/hid/usbhid/hiddev.c
152
list->head = (list->head + 1) &
drivers/hid/usbhid/hiddev.c
338
if (list->head == list->tail) {
drivers/hid/usbhid/hiddev.c
341
while (list->head == list->tail) {
drivers/hid/usbhid/hiddev.c
374
while (list->head != list->tail &&
drivers/hid/usbhid/hiddev.c
40
int head;
drivers/hid/usbhid/hiddev.c
417
if (list->head != list->tail)
drivers/hsi/clients/cmt_speech.c
131
static void cs_notify(u32 message, struct list_head *head)
drivers/hsi/clients/cmt_speech.c
1319
static void cs_free_char_queue(struct list_head *head)
drivers/hsi/clients/cmt_speech.c
1324
if (!list_empty(head)) {
drivers/hsi/clients/cmt_speech.c
1325
list_for_each_safe(cursor, next, head) {
drivers/hsi/clients/cmt_speech.c
151
list_add_tail(&entry->list, head);
drivers/hsi/clients/cmt_speech.c
162
static u32 cs_pop_entry(struct list_head *head)
drivers/hsi/clients/cmt_speech.c
167
entry = list_entry(head->next, struct char_queue, list);
drivers/hsi/clients/ssi_protocol.c
388
struct list_head *head, *tmp;
drivers/hsi/clients/ssi_protocol.c
411
list_for_each_safe(head, tmp, &ssi->txqueue) {
drivers/hsi/clients/ssi_protocol.c
412
msg = list_entry(head, struct hsi_msg, link);
drivers/hsi/clients/ssi_protocol.c
414
list_del(head);
drivers/hsi/controllers/omap_ssi_port.c
171
struct list_head *head, *tmp;
drivers/hsi/controllers/omap_ssi_port.c
176
list_for_each_safe(head, tmp, &omap_port->errqueue) {
drivers/hsi/controllers/omap_ssi_port.c
177
msg = list_entry(head, struct hsi_msg, link);
drivers/hsi/controllers/omap_ssi_port.c
179
list_del(head);
drivers/hwtracing/coresight/coresight-etb10.c
407
unsigned long head;
drivers/hwtracing/coresight/coresight-etb10.c
414
head = handle->head & ((buf->nr_pages << PAGE_SHIFT) - 1);
drivers/hwtracing/coresight/coresight-etb10.c
417
buf->cur = head / PAGE_SIZE;
drivers/hwtracing/coresight/coresight-etb10.c
420
buf->offset = head % PAGE_SIZE;
drivers/hwtracing/coresight/coresight-etb10.c
564
handle->head += to_read;
drivers/hwtracing/coresight/coresight-etm-perf.c
790
struct perf_addr_filters_head *head = perf_event_addr_filters(event);
drivers/hwtracing/coresight/coresight-etm-perf.c
798
list_for_each_entry(filter, &head->list, entry) {
drivers/hwtracing/coresight/coresight-tmc-etf.c
454
unsigned long head;
drivers/hwtracing/coresight/coresight-tmc-etf.c
461
head = handle->head & (((unsigned long)buf->nr_pages << PAGE_SHIFT) - 1);
drivers/hwtracing/coresight/coresight-tmc-etf.c
464
buf->cur = head / PAGE_SIZE;
drivers/hwtracing/coresight/coresight-tmc-etf.c
467
buf->offset = head % PAGE_SIZE;
drivers/hwtracing/coresight/coresight-tmc-etf.c
584
handle->head += to_read;
drivers/hwtracing/coresight/coresight-tmc-etr.c
1580
unsigned long head,
drivers/hwtracing/coresight/coresight-tmc-etr.c
1589
head = PERF_IDX2OFF(head, etr_perf);
drivers/hwtracing/coresight/coresight-tmc-etr.c
1590
pg_idx = head >> PAGE_SHIFT;
drivers/hwtracing/coresight/coresight-tmc-etr.c
1591
pg_offset = head & (PAGE_SIZE - 1);
drivers/hwtracing/coresight/coresight-tmc-etr.c
1697
tmc_etr_sync_perf_buffer(etr_perf, handle->head, offset, size);
drivers/hwtracing/coresight/coresight-tmc-etr.c
1705
handle->head += size;
drivers/hwtracing/coresight/coresight-trbe.c
1076
handle->head += size;
drivers/hwtracing/coresight/coresight-trbe.c
355
u64 head = PERF_IDX2OFF(handle->head, buf);
drivers/hwtracing/coresight/coresight-trbe.c
357
__trbe_pad_buf(buf, head, len);
drivers/hwtracing/coresight/coresight-trbe.c
409
u64 head, tail, wakeup;
drivers/hwtracing/coresight/coresight-trbe.c
411
head = PERF_IDX2OFF(handle->head, buf);
drivers/hwtracing/coresight/coresight-trbe.c
429
if (!IS_ALIGNED(head, cpudata->trbe_align)) {
drivers/hwtracing/coresight/coresight-trbe.c
430
unsigned long delta = roundup(head, cpudata->trbe_align) - head;
drivers/hwtracing/coresight/coresight-trbe.c
434
head = PERF_IDX2OFF(handle->head, buf);
drivers/hwtracing/coresight/coresight-trbe.c
450
tail = PERF_IDX2OFF(handle->head + handle->size, buf);
drivers/hwtracing/coresight/coresight-trbe.c
497
if (head < tail)
drivers/hwtracing/coresight/coresight-trbe.c
514
if (handle->wakeup < (handle->head + handle->size) && head <= wakeup)
drivers/hwtracing/coresight/coresight-trbe.c
545
if (limit > head)
drivers/hwtracing/coresight/coresight-trbe.c
556
u64 head = PERF_IDX2OFF(handle->head, buf);
drivers/hwtracing/coresight/coresight-trbe.c
566
while (limit && ((limit - head) < trbe_min_trace_buf_size(handle))) {
drivers/hwtracing/coresight/coresight-trbe.c
567
trbe_pad_buf(handle, limit - head);
drivers/hwtracing/coresight/coresight-trbe.c
569
head = PERF_IDX2OFF(handle->head, buf);
drivers/hwtracing/coresight/coresight-trbe.c
715
start_off = PERF_IDX2OFF(handle->head, buf);
drivers/hwtracing/coresight/coresight-trbe.c
869
handle->head += size;
drivers/hwtracing/coresight/coresight-trbe.c
995
buf->trbe_write = buf->trbe_base + PERF_IDX2OFF(handle->head, buf);
drivers/hwtracing/coresight/ultrasoc-smb.c
335
unsigned long head)
drivers/hwtracing/coresight/ultrasoc-smb.c
342
pg_idx = head >> PAGE_SHIFT;
drivers/hwtracing/coresight/ultrasoc-smb.c
343
pg_offset = head & (PAGE_SIZE - 1);
drivers/hwtracing/coresight/ultrasoc-smb.c
402
smb_sync_perf_buffer(drvdata, buf, handle->head);
drivers/hwtracing/ptt/hisi_ptt.c
1119
buf->pos = handle->head % buf->length;
drivers/hwtracing/ptt/hisi_ptt.c
289
buf->pos = handle->head % buf->length;
drivers/hwtracing/stm/policy.c
494
struct list_head *head = &policy->group.cg_children;
drivers/hwtracing/stm/policy.c
498
if (list_empty(head))
drivers/hwtracing/stm/policy.c
510
list_for_each_entry(item, head, ci_entry) {
drivers/hwtracing/stm/policy.c
520
head = &policy_node->group.cg_children;
drivers/i2c/busses/i2c-ismt.c
176
u8 head; /* ring buffer head pointer */
drivers/i2c/busses/i2c-ismt.c
227
struct ismt_desc *desc = &priv->hw[priv->head];
drivers/i2c/busses/i2c-ismt.c
229
dev_dbg(dev, "Dump of the descriptor struct: 0x%X\n", priv->head);
drivers/i2c/busses/i2c-ismt.c
305
fmhp = ((priv->head + 1) % ISMT_DESC_ENTRIES) << 16;
drivers/i2c/busses/i2c-ismt.c
417
desc = &priv->hw[priv->head];
drivers/i2c/busses/i2c-ismt.c
645
priv->head++;
drivers/i2c/busses/i2c-ismt.c
646
priv->head %= ISMT_DESC_ENTRIES;
drivers/i2c/busses/i2c-ismt.c
818
priv->head = 0;
drivers/iio/adc/nxp-sar-adc.c
614
dma_buf->head = (NXP_SAR_ADC_DMA_BUFF_SZ - state.residue) /
drivers/iio/adc/nxp-sar-adc.c
619
dma_buf->head--;
drivers/iio/adc/nxp-sar-adc.c
624
dma_buf->head -= dma_buf->head % info->channels_used;
drivers/iio/adc/nxp-sar-adc.c
630
while (dma_buf->tail != dma_buf->head) {
drivers/iio/adc/nxp-sar-adc.c
652
dma_buf->tail = dma_buf->head;
drivers/iio/adc/nxp-sar-adc.c
666
info->dma_buf.head = 0;
drivers/iio/buffer/industrialio-buffer-dma.c
142
list_for_each_entry_safe(block, _block, &block_list, head)
drivers/iio/buffer/industrialio-buffer-dma.c
154
list_add_tail(&block->head, &iio_dma_buffer_dead_blocks);
drivers/iio/buffer/industrialio-buffer-dma.c
192
INIT_LIST_HEAD(&block->head);
drivers/iio/buffer/industrialio-buffer-dma.c
266
list_for_each_entry_safe(block, _block, list, head) {
drivers/iio/buffer/industrialio-buffer-dma.c
267
list_del(&block->head);
drivers/iio/buffer/industrialio-buffer-dma.c
404
list_add_tail(&block->head, &queue->incoming);
drivers/iio/buffer/industrialio-buffer-dma.c
489
list_for_each_entry_safe(block, _block, &queue->incoming, head) {
drivers/iio/buffer/industrialio-buffer-dma.c
490
list_del(&block->head);
drivers/iio/buffer/industrialio-buffer-dma.c
529
list_add_tail(&block->head, &queue->incoming);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
133
list_add_tail(&block->head, &dmaengine_buffer->active);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
54
list_del(&block->head);
drivers/iio/buffer/industrialio-hw-consumer.c
116
list_for_each_entry(buf, &hwc->buffers, head)
drivers/iio/buffer/industrialio-hw-consumer.c
134
list_for_each_entry_safe(buf, n, &hwc->buffers, head)
drivers/iio/buffer/industrialio-hw-consumer.c
183
list_for_each_entry(buf, &hwc->buffers, head) {
drivers/iio/buffer/industrialio-hw-consumer.c
192
list_for_each_entry_continue_reverse(buf, &hwc->buffers, head)
drivers/iio/buffer/industrialio-hw-consumer.c
206
list_for_each_entry(buf, &hwc->buffers, head)
drivers/iio/buffer/industrialio-hw-consumer.c
28
struct list_head head;
drivers/iio/buffer/industrialio-hw-consumer.c
58
list_for_each_entry(buf, &hwc->buffers, head) {
drivers/iio/buffer/industrialio-hw-consumer.c
72
list_add_tail(&buf->head, &hwc->buffers);
drivers/infiniband/core/cache.c
223
static void put_gid_ndev(struct rcu_head *head)
drivers/infiniband/core/cache.c
226
container_of(head, struct roce_gid_ndev_storage, rcu_head);
drivers/infiniband/core/sa_query.c
938
const struct nlattr *head, *curr;
drivers/infiniband/core/sa_query.c
950
head = (const struct nlattr *) nlmsg_data(nlh);
drivers/infiniband/core/sa_query.c
968
nla_for_each_attr(curr, head, len, rem) {
drivers/infiniband/hw/cxgb4/qp.c
903
static void add_to_fc_list(struct list_head *head, struct list_head *entry)
drivers/infiniband/hw/cxgb4/qp.c
906
list_add_tail(entry, head);
drivers/infiniband/hw/hfi1/chip.c
11870
u32 head, tail;
drivers/infiniband/hw/hfi1/chip.c
11872
head = (read_uctxt_csr(rcd->dd, rcd->ctxt, RCV_HDR_HEAD)
drivers/infiniband/hw/hfi1/chip.c
11880
return head == tail;
drivers/infiniband/hw/hfi1/driver.c
1839
rcd->head);
drivers/infiniband/hw/hfi1/hfi.h
1465
return rcd->head;
drivers/infiniband/hw/hfi1/hfi.h
1473
static inline void hfi1_set_rcd_head(struct hfi1_ctxtdata *rcd, u32 head)
drivers/infiniband/hw/hfi1/hfi.h
1475
rcd->head = head;
drivers/infiniband/hw/hfi1/hfi.h
1481
return (__le32 *)rcd->rcvhdrq + rcd->head + rcd->rhf_offset;
drivers/infiniband/hw/hfi1/hfi.h
203
u32 head;
drivers/infiniband/hw/hfi1/ipoib.h
92
u32 head;
drivers/infiniband/hw/hfi1/ipoib_tx.c
147
tx_ring->head = 0;
drivers/infiniband/hw/hfi1/ipoib_tx.c
159
u32 head = tx_ring->head;
drivers/infiniband/hw/hfi1/ipoib_tx.c
162
struct ipoib_txreq *tx = hfi1_txreq_from_idx(tx_ring, head);
drivers/infiniband/hw/hfi1/ipoib_tx.c
170
trace_hfi1_tx_produce(tx, head);
drivers/infiniband/hw/hfi1/ipoib_tx.c
172
head = CIRC_NEXT(head, max_tx);
drivers/infiniband/hw/hfi1/ipoib_tx.c
173
tx = hfi1_txreq_from_idx(tx_ring, head);
drivers/infiniband/hw/hfi1/ipoib_tx.c
178
smp_store_release(&tx_ring->head, head);
drivers/infiniband/hw/hfi1/ipoib_tx.c
360
u32 head;
drivers/infiniband/hw/hfi1/ipoib_tx.c
366
head = smp_load_acquire(&tx_ring->head);
drivers/infiniband/hw/hfi1/ipoib_tx.c
369
CIRC_CNT(head, tail, tx_ring->max_items));
drivers/infiniband/hw/hfi1/ipoib_tx.c
862
txq->tx_ring.head, txq->tx_ring.tail);
drivers/infiniband/hw/hfi1/pio.c
1407
u32 head, next;
drivers/infiniband/hw/hfi1/pio.c
1443
head = sc->sr_head;
drivers/infiniband/hw/hfi1/pio.c
1459
pbuf = &sc->sr[head].pbuf;
drivers/infiniband/hw/hfi1/pio.c
1467
next = head + 1;
drivers/infiniband/hw/hfi1/pio.c
1649
u32 head, tail;
drivers/infiniband/hw/hfi1/pio.c
1671
head = READ_ONCE(sc->sr_head); /* snapshot the head */
drivers/infiniband/hw/hfi1/pio.c
1673
while (head != tail) {
drivers/infiniband/hw/hfi1/rc.c
1698
u32 opcode, head, tail;
drivers/infiniband/hw/hfi1/rc.c
1730
head = priv->s_tid_head;
drivers/infiniband/hw/hfi1/rc.c
1742
if (head == tail && req->comp_seg < req->total_segs) {
drivers/infiniband/hw/hfi1/rc.c
1749
head = qp->s_tail;
drivers/infiniband/hw/hfi1/rc.c
1757
if ((psn & IB_BTH_REQ_ACK) && tail != head &&
drivers/infiniband/hw/hfi1/sdma.c
2091
u16 head, tail, cnt;
drivers/infiniband/hw/hfi1/sdma.c
2093
head = sde->descq_head & sde->sdma_mask;
drivers/infiniband/hw/hfi1/sdma.c
2099
sde->this_idx, head, tail, cnt,
drivers/infiniband/hw/hfi1/sdma.c
2103
while (head != tail) {
drivers/infiniband/hw/hfi1/sdma.c
2106
descqp = &sde->descq[head];
drivers/infiniband/hw/hfi1/sdma.c
2122
head, flags, addr, gen, len);
drivers/infiniband/hw/hfi1/sdma.c
2138
head++;
drivers/infiniband/hw/hfi1/sdma.c
2139
head &= sde->sdma_mask;
drivers/infiniband/hw/hfi1/sdma.c
2154
u16 head, tail;
drivers/infiniband/hw/hfi1/sdma.c
2161
head = sde->descq_head & sde->sdma_mask;
drivers/infiniband/hw/hfi1/sdma.c
2170
(unsigned long long)read_sde_csr(sde, SD(HEAD)), head,
drivers/infiniband/hw/hfi1/sdma.c
2186
while (head != tail) {
drivers/infiniband/hw/hfi1/sdma.c
2189
descqp = &sde->descq[head];
drivers/infiniband/hw/hfi1/sdma.c
2205
head, flags, addr, gen, len);
drivers/infiniband/hw/hfi1/sdma.c
2214
head = (head + 1) & sde->sdma_mask;
drivers/infiniband/hw/hfi1/sdma.c
539
u16 head, tail;
drivers/infiniband/hw/hfi1/sdma.c
548
head = sde->descq_head & sde->sdma_mask;
drivers/infiniband/hw/hfi1/sdma.c
550
while (head != tail) {
drivers/infiniband/hw/hfi1/sdma.c
552
head = ++sde->descq_head & sde->sdma_mask;
drivers/infiniband/hw/hfi1/sdma.c
554
if (txp && txp->next_descq_idx == head) {
drivers/infiniband/hw/hfi1/sdma.c
558
trace_hfi1_sdma_progress(sde, head, tail, txp);
drivers/infiniband/hw/hfi1/tid_rdma.c
1684
u16 head, tail;
drivers/infiniband/hw/hfi1/tid_rdma.c
1687
head = req->setup_head;
drivers/infiniband/hw/hfi1/tid_rdma.c
1689
for ( ; CIRC_CNT(head, tail, MAX_FLOWS);
drivers/infiniband/hw/hfi1/trace_tx.h
903
__field(ulong, head)
drivers/infiniband/hw/hfi1/trace_tx.h
916
__entry->head = txq->tx_ring.head;
drivers/infiniband/hw/hfi1/trace_tx.h
936
__entry->head,
drivers/infiniband/hw/hns/hns_roce_device.h
356
u32 head;
drivers/infiniband/hw/hns/hns_roce_device.h
450
u32 head;
drivers/infiniband/hw/hns/hns_roce_hem.c
1263
int unit, int max_ba_num, struct hns_roce_hem_head *head,
drivers/infiniband/hw/hns/hns_roce_hem.c
1273
root_hem = list_first_entry(&head->root,
drivers/infiniband/hw/hns/hns_roce_hem.c
1293
&head->branch[i], &head->leaf);
drivers/infiniband/hw/hns/hns_roce_hem.c
1304
list_splice(&head->leaf, &hem_list->btm_bt);
drivers/infiniband/hw/hns/hns_roce_hem.c
1305
list_splice(&head->root, &hem_list->root_bt);
drivers/infiniband/hw/hns/hns_roce_hem.c
1307
list_splice(&head->branch[i], &hem_list->mid_bt[i][0]);
drivers/infiniband/hw/hns/hns_roce_hem.c
1318
struct hns_roce_hem_head head;
drivers/infiniband/hw/hns/hns_roce_hem.c
1334
INIT_LIST_HEAD(&head.root);
drivers/infiniband/hw/hns/hns_roce_hem.c
1335
INIT_LIST_HEAD(&head.leaf);
drivers/infiniband/hw/hns/hns_roce_hem.c
1337
INIT_LIST_HEAD(&head.branch[i]);
drivers/infiniband/hw/hns/hns_roce_hem.c
1340
list_add(&root_hem->list, &head.root);
drivers/infiniband/hw/hns/hns_roce_hem.c
1341
ret = setup_root_hem(hr_dev, hem_list, unit, max_ba_num, &head, regions,
drivers/infiniband/hw/hns/hns_roce_hem.c
1345
hem_list_free_all(hr_dev, &head.branch[i]);
drivers/infiniband/hw/hns/hns_roce_hem.c
1347
hem_list_free_all(hr_dev, &head.root);
drivers/infiniband/hw/hns/hns_roce_hem.c
1441
struct list_head *head = &hem_list->btm_bt;
drivers/infiniband/hw/hns/hns_roce_hem.c
1447
list_for_each_entry_safe(hem, temp_hem, head, sibling) {
drivers/infiniband/hw/hns/hns_roce_hem.c
972
struct list_head *head)
drivers/infiniband/hw/hns/hns_roce_hem.c
976
list_for_each_entry_safe(hem, temp_hem, head, list) {
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1069
return idx_que->head - idx_que->tail >= srq->wqe_cnt;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1110
unsigned int head;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1113
head = idx_que->head & (srq->wqe_cnt - 1);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1115
buf = get_idx_buf(idx_que, head);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1118
idx_que->head++;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1128
hr_reg_write(&db, DB_PI, srq->idx_que.head);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1172
*srq->rdb.db_record = srq->idx_que.head &
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1361
csq->head = 0;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1419
return tail == priv->cmq.csq.head;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1492
tail = csq->head;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1497
csq->desc[csq->head++] = desc[i];
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1498
if (csq->head == csq->desc_num)
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1499
csq->head = 0;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1503
roce_write(hr_dev, ROCEE_TX_CMQ_PI_REG, csq->head);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1526
csq->head, tail);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1527
csq->head = tail;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1556
rsv_tail = csq->head;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3911
if (hr_qp->sq.head != hr_qp->sq.tail)
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3916
if (hr_qp->rq.head != hr_qp->rq.tail)
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3961
left = wq->head - wq->tail;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
5582
hr_qp->rq.head = 0;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
5584
hr_qp->sq.head = 0;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
5601
trace_hns_sq_flush_cqe(hr_qp->qpn, hr_qp->sq.head, TRACE_SQ);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
5602
hr_reg_write(context, QPC_SQ_PRODUCER_IDX, hr_qp->sq.head);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
5611
trace_hns_rq_flush_cqe(hr_qp->qpn, hr_qp->rq.head, TRACE_RQ);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
5612
hr_reg_write(context, QPC_RQ_PRODUCER_IDX, hr_qp->rq.head);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
630
hr_reg_write(&sq_db, DB_PI, qp->sq.head);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
645
qp->rq.head & V2_DB_PRODUCER_IDX_M;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
651
hr_reg_write(&rq_db, DB_PI, qp->rq.head);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
688
hr_reg_write(rc_sq_wqe, RC_SEND_WQE_WQE_INDEX, qp->sq.head);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
726
wqe_idx = (qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
739
~(((qp->sq.head + nreq) >> ilog2(qp->sq.wqe_cnt)) & 0x1);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
760
qp->sq.head += nreq;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
863
wqe_idx = (hr_qp->rq.head + nreq) & (hr_qp->rq.wqe_cnt - 1);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
870
hr_qp->rq.head += nreq;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
891
wq->wrid[wq->head & (wq->wqe_cnt - 1)] = wr_id;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
892
wq->head++;
drivers/infiniband/hw/hns/hns_roce_hw_v2.h
1319
u32 head;
drivers/infiniband/hw/hns/hns_roce_main.c
1138
if (hr_qp->sq.tail != hr_qp->sq.head)
drivers/infiniband/hw/hns/hns_roce_main.c
1143
if ((!hr_qp->ibqp.srq) && (hr_qp->rq.tail != hr_qp->rq.head))
drivers/infiniband/hw/hns/hns_roce_qp.c
1465
hr_qp->sq.head = *(int *)(hr_qp->sdb.virt_addr);
drivers/infiniband/hw/hns/hns_roce_qp.c
1468
hr_qp->rq.head = *(int *)(hr_qp->rdb.virt_addr);
drivers/infiniband/hw/hns/hns_roce_qp.c
1587
cur = hr_wq->head - hr_wq->tail;
drivers/infiniband/hw/hns/hns_roce_qp.c
1593
cur = hr_wq->head - hr_wq->tail;
drivers/infiniband/hw/hns/hns_roce_srq.c
199
idx_que->head = 0;
drivers/infiniband/hw/irdma/ctrl.c
17
struct irdma_sc_qp *irdma_get_qp_from_list(struct list_head *head,
drivers/infiniband/hw/irdma/ctrl.c
23
if (list_empty(head))
drivers/infiniband/hw/irdma/ctrl.c
2497
cqp->sq_ring.head, cqp->sq_ring.tail, cqp->sq_ring.size);
drivers/infiniband/hw/irdma/ctrl.c
27
entry = head->next;
drivers/infiniband/hw/irdma/ctrl.c
31
if (entry == head)
drivers/infiniband/hw/irdma/ctrl.c
3793
cqp->sq_ring.head, cqp->sq_ring.tail, cqp->sq_ring.size);
drivers/infiniband/hw/irdma/ctrl.c
3812
cqp->sq_ring.head, cqp->sq_ring.tail,
drivers/infiniband/hw/irdma/defs.h
1004
(_ring).head = ((_ring).head + (_count)) % size; \
drivers/infiniband/hw/irdma/defs.h
1015
(_ring).head = ((_ring).head + 1) % size; \
drivers/infiniband/hw/irdma/defs.h
1026
(_ring).head = ((_ring).head + (_count)) % size; \
drivers/infiniband/hw/irdma/defs.h
1033
(_ring).head = ((_ring).head + (_count)) % (_ring).size
drivers/infiniband/hw/irdma/defs.h
1039
(_ring).head = ((_ring).head + 1) % (_ring).size
drivers/infiniband/hw/irdma/defs.h
1082
(((_ring).head + (_ring).size - (_ring).tail) % (_ring).size) \
drivers/infiniband/hw/irdma/defs.h
980
(_ring).head = 0; \
drivers/infiniband/hw/irdma/defs.h
985
#define IRDMA_RING_CURRENT_HEAD(_ring) ((_ring).head)
drivers/infiniband/hw/irdma/defs.h
993
(_ring).head = ((_ring).head + 1) % size; \
drivers/infiniband/hw/irdma/protos.h
63
struct irdma_sc_qp *irdma_get_qp_from_list(struct list_head *head,
drivers/infiniband/hw/irdma/puda.c
240
peek_head = (cq_uk->cq_ring.head + 1) % cq_uk->cq_ring.size;
drivers/infiniband/hw/irdma/uk.c
1178
peek_head = (cq->cq_ring.head + 1) % cq->cq_ring.size;
drivers/infiniband/hw/irdma/uk.c
1343
if (wqe_idx < qp->conn_wqes && qp->sq_ring.head == qp->sq_ring.tail) {
drivers/infiniband/hw/irdma/uk.c
1783
cq_head = cq->cq_ring.head;
drivers/infiniband/hw/irdma/uk.c
65
if (!qp->sq_ring.head)
drivers/infiniband/hw/irdma/user.h
271
u32 head;
drivers/infiniband/hw/mlx4/alias_GUID.c
498
struct list_head *head =
drivers/infiniband/hw/mlx4/alias_GUID.c
540
list_add_tail(&callback_context->list, head);
drivers/infiniband/hw/mlx4/cq.c
620
cur = wq->head - wq->tail;
drivers/infiniband/hw/mlx4/main.c
3085
if (mqp->sq.tail != mqp->sq.head) {
drivers/infiniband/hw/mlx4/main.c
3103
if (mqp->rq.tail != mqp->rq.head) {
drivers/infiniband/hw/mlx4/mlx4_ib.h
171
unsigned head;
drivers/infiniband/hw/mlx4/mlx4_ib.h
381
int head;
drivers/infiniband/hw/mlx4/qp.c
2590
qp->rq.head = 0;
drivers/infiniband/hw/mlx4/qp.c
2592
qp->sq.head = 0;
drivers/infiniband/hw/mlx4/qp.c
3290
cur = wq->head - wq->tail;
drivers/infiniband/hw/mlx4/qp.c
3296
cur = wq->head - wq->tail;
drivers/infiniband/hw/mlx4/qp.c
3596
qp->sq.wrid[(qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1)] = wr->wr_id;
drivers/infiniband/hw/mlx4/qp.c
3823
qp->sq.head += nreq;
drivers/infiniband/hw/mlx4/qp.c
3874
ind = qp->rq.head & (qp->rq.wqe_cnt - 1);
drivers/infiniband/hw/mlx4/qp.c
3922
qp->rq.head += nreq;
drivers/infiniband/hw/mlx4/qp.c
3930
*qp->db.db = cpu_to_be32(qp->rq.head & 0xffff);
drivers/infiniband/hw/mlx4/srq.c
148
srq->head = 0;
drivers/infiniband/hw/mlx4/srq.c
333
if (unlikely(srq->head == srq->tail)) {
drivers/infiniband/hw/mlx4/srq.c
339
srq->wrid[srq->head] = wr->wr_id;
drivers/infiniband/hw/mlx4/srq.c
341
next = get_wqe(srq, srq->head);
drivers/infiniband/hw/mlx4/srq.c
342
srq->head = be16_to_cpu(next->next_wqe_index);
drivers/infiniband/hw/mlx5/cq.c
345
u16 tail, u16 head)
drivers/infiniband/hw/mlx5/cq.c
351
if (idx == head)
drivers/infiniband/hw/mlx5/cq.c
406
cur = wq->head - wq->tail;
drivers/infiniband/hw/mlx5/devx.c
2697
int head = 0;
drivers/infiniband/hw/mlx5/devx.c
2729
if (tail - head == MAX_ASYNC_CMDS) {
drivers/infiniband/hw/mlx5/devx.c
2730
devx_wait_async_destroy(&async_cmd[head % MAX_ASYNC_CMDS]);
drivers/infiniband/hw/mlx5/devx.c
2731
head++;
drivers/infiniband/hw/mlx5/devx.c
2735
while (head != tail) {
drivers/infiniband/hw/mlx5/devx.c
2736
devx_wait_async_destroy(&async_cmd[head % MAX_ASYNC_CMDS]);
drivers/infiniband/hw/mlx5/devx.c
2737
head++;
drivers/infiniband/hw/mlx5/main.c
2947
if (mqp->sq.tail != mqp->sq.head) {
drivers/infiniband/hw/mlx5/main.c
2964
if (mqp->rq.tail != mqp->rq.head) {
drivers/infiniband/hw/mlx5/mlx5_ib.h
377
unsigned head;
drivers/infiniband/hw/mlx5/mlx5_ib.h
606
int head;
drivers/infiniband/hw/mlx5/odp.c
1007
struct pf_frame *head = NULL, *frame;
drivers/infiniband/hw/mlx5/odp.c
1106
frame->next = head;
drivers/infiniband/hw/mlx5/odp.c
1107
head = frame;
drivers/infiniband/hw/mlx5/odp.c
1120
if (head) {
drivers/infiniband/hw/mlx5/odp.c
1121
frame = head;
drivers/infiniband/hw/mlx5/odp.c
1122
head = frame->next;
drivers/infiniband/hw/mlx5/odp.c
1137
while (head) {
drivers/infiniband/hw/mlx5/odp.c
1138
frame = head;
drivers/infiniband/hw/mlx5/odp.c
1139
head = frame->next;
drivers/infiniband/hw/mlx5/qp.c
4448
qp->rq.head = 0;
drivers/infiniband/hw/mlx5/qp.c
4450
qp->sq.head = 0;
drivers/infiniband/hw/mlx5/srq.c
128
srq->head = 0;
drivers/infiniband/hw/mlx5/srq.c
437
if (unlikely(srq->head == srq->tail)) {
drivers/infiniband/hw/mlx5/srq.c
443
srq->wrid[srq->head] = wr->wr_id;
drivers/infiniband/hw/mlx5/srq.c
445
next = get_wqe(srq, srq->head);
drivers/infiniband/hw/mlx5/srq.c
446
srq->head = be16_to_cpu(next->next_wqe_index);
drivers/infiniband/hw/mlx5/wr.c
1030
qp->sq.head += nreq;
drivers/infiniband/hw/mlx5/wr.c
1231
ind = qp->rq.head & (qp->rq.wqe_cnt - 1);
drivers/infiniband/hw/mlx5/wr.c
1271
qp->rq.head += nreq;
drivers/infiniband/hw/mlx5/wr.c
1278
*qp->db.db = cpu_to_be32(qp->rq.head & 0xffff);
drivers/infiniband/hw/mlx5/wr.c
34
cur = wq->head - wq->tail;
drivers/infiniband/hw/mlx5/wr.c
40
cur = wq->head - wq->tail;
drivers/infiniband/hw/mlx5/wr.c
773
qp->sq.wqe_head[idx] = qp->sq.head + nreq;
drivers/infiniband/hw/mthca/mthca_provider.h
233
unsigned head;
drivers/infiniband/hw/mthca/mthca_qp.c
1574
cur = wq->head - wq->tail;
drivers/infiniband/hw/mthca/mthca_qp.c
1580
cur = wq->head - wq->tail;
drivers/infiniband/hw/mthca/mthca_qp.c
1659
qp->sq.head, qp->sq.tail,
drivers/infiniband/hw/mthca/mthca_qp.c
1817
qp->sq.head += nreq;
drivers/infiniband/hw/mthca/mthca_qp.c
1855
qp->rq.head, qp->rq.tail,
drivers/infiniband/hw/mthca/mthca_qp.c
1908
qp->rq.head += MTHCA_TAVOR_MAX_WQES_PER_RECV_DB;
drivers/infiniband/hw/mthca/mthca_qp.c
1922
qp->rq.head += nreq;
drivers/infiniband/hw/mthca/mthca_qp.c
1957
ind = qp->sq.head & (qp->sq.max - 1);
drivers/infiniband/hw/mthca/mthca_qp.c
1964
((qp->sq.head & 0xffff) << 8) | f0 | op0;
drivers/infiniband/hw/mthca/mthca_qp.c
1966
qp->sq.head += MTHCA_ARBEL_MAX_WQES_PER_SEND_DB;
drivers/infiniband/hw/mthca/mthca_qp.c
1973
*qp->sq.db = cpu_to_be32(qp->sq.head & 0xffff);
drivers/infiniband/hw/mthca/mthca_qp.c
1989
qp->sq.head, qp->sq.tail,
drivers/infiniband/hw/mthca/mthca_qp.c
2137
dbhi = (nreq << 24) | ((qp->sq.head & 0xffff) << 8) | f0 | op0;
drivers/infiniband/hw/mthca/mthca_qp.c
2139
qp->sq.head += nreq;
drivers/infiniband/hw/mthca/mthca_qp.c
2146
*qp->sq.db = cpu_to_be32(qp->sq.head & 0xffff);
drivers/infiniband/hw/mthca/mthca_qp.c
2178
ind = qp->rq.head & (qp->rq.max - 1);
drivers/infiniband/hw/mthca/mthca_qp.c
2184
qp->rq.head, qp->rq.tail,
drivers/infiniband/hw/mthca/mthca_qp.c
2219
qp->rq.head += nreq;
drivers/infiniband/hw/mthca/mthca_qp.c
2226
*qp->rq.db = cpu_to_be32(qp->rq.head & 0xffff);
drivers/infiniband/hw/mthca/mthca_qp.c
234
wq->head = 0;
drivers/infiniband/hw/ocrdma/ocrdma.h
135
u16 head, tail;
drivers/infiniband/hw/ocrdma/ocrdma.h
362
u32 head, tail;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1043
dev->mqe_ctx.tag = dev->mq.sq.head;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1046
cmd->hdr.tag_lo = dev->mq.sq.head;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
138
return dev->mq.sq.va + (dev->mq.sq.head * sizeof(struct ocrdma_mqe));
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
143
dev->mq.sq.head = (dev->mq.sq.head + 1) & (OCRDMA_MQ_LEN - 1);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
2123
qp->sq.head = 0;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
2125
qp->rq.head = 0;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
908
struct list_head *head = sq?(&cq->sq_head):(&cq->rq_head);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
910
list_for_each(cur, head) {
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1556
return ((q->max_wqe_idx - q->head) + q->tail) % q->max_cnt;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1561
return (qp->sq.tail == qp->sq.head);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1566
return (qp->rq.tail == qp->rq.head);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1571
return q->va + (q->head * q->entry_size);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1582
q->head = (q->head + 1) & q->max_wqe_idx;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2174
qp->wqe_wr_id_tbl[qp->sq.head].signaled = 1;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2176
qp->wqe_wr_id_tbl[qp->sq.head].signaled = 0;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2177
qp->wqe_wr_id_tbl[qp->sq.head].wrid = wr->wr_id;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2245
qp->rqe_wr_id_tbl[qp->rq.head] = wr->wr_id;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
281
unsigned int head;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
289
cq->ibcq.cqe, &head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
298
items = (tail > head) ? (tail - head) :
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
299
(cq->ibcq.cqe - head + tail);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
328
unsigned int head;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
334
cq->ibcq.cqe, &head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
348
cqe = get_cqe(cq, head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
80
unsigned int head;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
83
cq->ibcq.cqe, &head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
427
unsigned int head;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
438
while (pvrdma_idx_ring_has_data(ring, ring_slots, &head) > 0) {
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
441
eqe = get_eqe(dev, head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
503
unsigned int head;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
507
while (pvrdma_idx_ring_has_data(ring, ring_slots, &head) > 0) {
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
511
cqne = get_cqne(dev, head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h
104
const __u32 head = atomic_read(&r->cons_head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h
107
pvrdma_idx_valid(head, max_elems)) {
drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h
108
*out_head = head & (max_elems - 1);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h
109
return tail != head;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h
90
const __u32 head = atomic_read(&r->cons_head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h
93
pvrdma_idx_valid(head, max_elems)) {
drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h
95
return tail != (head ^ max_elems);
drivers/infiniband/sw/rdmavt/cq.c
320
if (RDMA_READ_UAPI_ATOMIC(cq->queue->head) !=
drivers/infiniband/sw/rdmavt/cq.c
324
if (cq->kqueue->head != cq->kqueue->tail)
drivers/infiniband/sw/rdmavt/cq.c
33
u32 head;
drivers/infiniband/sw/rdmavt/cq.c
343
u32 head, tail, n;
drivers/infiniband/sw/rdmavt/cq.c
387
head = RDMA_READ_UAPI_ATOMIC(old_u_wc->head);
drivers/infiniband/sw/rdmavt/cq.c
391
head = old_k_wc->head;
drivers/infiniband/sw/rdmavt/cq.c
395
if (head > (u32)cq->ibcq.cqe)
drivers/infiniband/sw/rdmavt/cq.c
396
head = (u32)cq->ibcq.cqe;
drivers/infiniband/sw/rdmavt/cq.c
399
if (head < tail)
drivers/infiniband/sw/rdmavt/cq.c
400
n = cq->ibcq.cqe + 1 + head - tail;
drivers/infiniband/sw/rdmavt/cq.c
402
n = head - tail;
drivers/infiniband/sw/rdmavt/cq.c
407
for (n = 0; tail != head; n++) {
drivers/infiniband/sw/rdmavt/cq.c
419
RDMA_WRITE_UAPI_ATOMIC(u_wc->head, n);
drivers/infiniband/sw/rdmavt/cq.c
42
head = RDMA_READ_UAPI_ATOMIC(u_wc->head);
drivers/infiniband/sw/rdmavt/cq.c
423
k_wc->head = n;
drivers/infiniband/sw/rdmavt/cq.c
47
head = k_wc->head;
drivers/infiniband/sw/rdmavt/cq.c
497
if (tail == wc->head)
drivers/infiniband/sw/rdmavt/cq.c
55
if (head >= (unsigned)cq->ibcq.cqe) {
drivers/infiniband/sw/rdmavt/cq.c
56
head = cq->ibcq.cqe;
drivers/infiniband/sw/rdmavt/cq.c
59
next = head + 1;
drivers/infiniband/sw/rdmavt/cq.c
79
trace_rvt_cq_enter(cq, entry, head);
drivers/infiniband/sw/rdmavt/cq.c
81
uqueue[head].wr_id = entry->wr_id;
drivers/infiniband/sw/rdmavt/cq.c
82
uqueue[head].status = entry->status;
drivers/infiniband/sw/rdmavt/cq.c
83
uqueue[head].opcode = entry->opcode;
drivers/infiniband/sw/rdmavt/cq.c
84
uqueue[head].vendor_err = entry->vendor_err;
drivers/infiniband/sw/rdmavt/cq.c
85
uqueue[head].byte_len = entry->byte_len;
drivers/infiniband/sw/rdmavt/cq.c
86
uqueue[head].ex.imm_data = entry->ex.imm_data;
drivers/infiniband/sw/rdmavt/cq.c
87
uqueue[head].qp_num = entry->qp->qp_num;
drivers/infiniband/sw/rdmavt/cq.c
88
uqueue[head].src_qp = entry->src_qp;
drivers/infiniband/sw/rdmavt/cq.c
89
uqueue[head].wc_flags = entry->wc_flags;
drivers/infiniband/sw/rdmavt/cq.c
90
uqueue[head].pkey_index = entry->pkey_index;
drivers/infiniband/sw/rdmavt/cq.c
91
uqueue[head].slid = ib_lid_cpu16(entry->slid);
drivers/infiniband/sw/rdmavt/cq.c
92
uqueue[head].sl = entry->sl;
drivers/infiniband/sw/rdmavt/cq.c
93
uqueue[head].dlid_path_bits = entry->dlid_path_bits;
drivers/infiniband/sw/rdmavt/cq.c
94
uqueue[head].port_num = entry->port_num;
drivers/infiniband/sw/rdmavt/cq.c
96
RDMA_WRITE_UAPI_ATOMIC(u_wc->head, next);
drivers/infiniband/sw/rdmavt/cq.c
98
kqueue[head] = *entry;
drivers/infiniband/sw/rdmavt/cq.c
99
k_wc->head = next;
drivers/infiniband/sw/rdmavt/qp.c
1325
u32 head;
drivers/infiniband/sw/rdmavt/qp.c
1334
head = RDMA_READ_UAPI_ATOMIC(wq->head);
drivers/infiniband/sw/rdmavt/qp.c
1338
head = kwq->head;
drivers/infiniband/sw/rdmavt/qp.c
1342
if (head >= qp->r_rq.size)
drivers/infiniband/sw/rdmavt/qp.c
1343
head = 0;
drivers/infiniband/sw/rdmavt/qp.c
1346
while (tail != head) {
drivers/infiniband/sw/rdmavt/qp.c
1805
next = wq->head + 1;
drivers/infiniband/sw/rdmavt/qp.c
1823
wqe = rvt_get_rwqe_ptr(&qp->r_rq, wq->head);
drivers/infiniband/sw/rdmavt/qp.c
1835
smp_store_release(&wq->head, next);
drivers/infiniband/sw/rdmavt/qp.c
2221
next = wq->head + 1;
drivers/infiniband/sw/rdmavt/qp.c
2230
wqe = rvt_get_rwqe_ptr(&srq->rq, wq->head);
drivers/infiniband/sw/rdmavt/qp.c
2239
smp_store_release(&wq->head, next);
drivers/infiniband/sw/rdmavt/qp.c
2321
u32 head;
drivers/infiniband/sw/rdmavt/qp.c
2324
head = RDMA_READ_UAPI_ATOMIC(rq->wq->head);
drivers/infiniband/sw/rdmavt/qp.c
2326
head = rq->kwq->head;
drivers/infiniband/sw/rdmavt/qp.c
2328
return head;
drivers/infiniband/sw/rdmavt/qp.c
2351
u32 head;
drivers/infiniband/sw/rdmavt/qp.c
2385
head = get_rvt_head(rq, ip);
drivers/infiniband/sw/rdmavt/qp.c
2386
kwq->count = rvt_get_rq_count(rq, head, tail);
drivers/infiniband/sw/rdmavt/rc.c
65
u32 head;
drivers/infiniband/sw/rdmavt/rc.c
72
head = RDMA_READ_UAPI_ATOMIC(qp->r_rq.wq->head);
drivers/infiniband/sw/rdmavt/rc.c
75
head = READ_ONCE(qp->r_rq.kwq->head);
drivers/infiniband/sw/rdmavt/rc.c
78
if (head >= qp->r_rq.size)
drivers/infiniband/sw/rdmavt/rc.c
79
head = 0;
drivers/infiniband/sw/rdmavt/rc.c
88
credits = rvt_get_rq_count(&qp->r_rq, head, tail);
drivers/infiniband/sw/rdmavt/srq.c
137
u32 sz, size, n, head, tail;
drivers/infiniband/sw/rdmavt/srq.c
174
head = RDMA_READ_UAPI_ATOMIC(owq->head);
drivers/infiniband/sw/rdmavt/srq.c
178
head = okwq->head;
drivers/infiniband/sw/rdmavt/srq.c
181
if (head >= srq->rq.size || tail >= srq->rq.size) {
drivers/infiniband/sw/rdmavt/srq.c
185
n = head;
drivers/infiniband/sw/rdmavt/srq.c
196
while (tail != head) {
drivers/infiniband/sw/rdmavt/srq.c
213
RDMA_WRITE_UAPI_ATOMIC(tmp_rq.wq->head, n);
drivers/infiniband/sw/rdmavt/srq.c
216
tmp_rq.kwq->head = n;
drivers/infiniband/sw/rdmavt/trace_tx.h
106
__entry->head,
drivers/infiniband/sw/rdmavt/trace_tx.h
63
__field(u32, head)
drivers/infiniband/sw/rdmavt/trace_tx.h
83
__entry->head = qp->s_head;
drivers/infiniband/ulp/ipoib/ipoib_ib.c
570
void *head, int hlen)
drivers/infiniband/ulp/ipoib/ipoib_ib.c
580
if (head) {
drivers/infiniband/ulp/ipoib/ipoib_ib.c
582
priv->tx_wr.header = head;
drivers/infiniband/ulp/ipoib/ipoib_main.c
1619
static void neigh_hash_free_rcu(struct rcu_head *head)
drivers/infiniband/ulp/ipoib/ipoib_main.c
1621
struct ipoib_neigh_hash *htbl = container_of(head,
drivers/infiniband/ulp/ipoib/ipoib_main.c
2697
LIST_HEAD(head);
drivers/infiniband/ulp/ipoib/ipoib_main.c
2705
unregister_netdevice_queue(cpriv->dev, &head);
drivers/infiniband/ulp/ipoib/ipoib_main.c
2707
unregister_netdevice_queue(priv->dev, &head);
drivers/infiniband/ulp/ipoib/ipoib_main.c
2708
unregister_netdevice_many(&head);
drivers/infiniband/ulp/ipoib/ipoib_netlink.c
154
static void ipoib_del_child_link(struct net_device *dev, struct list_head *head)
drivers/infiniband/ulp/ipoib/ipoib_netlink.c
161
unregister_netdevice_queue(dev, head);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
762
rtrs_clt_get_next_path_or_null(struct list_head *head, struct rtrs_clt_path *clt_path)
drivers/infiniband/ulp/rtrs/rtrs-clt.c
764
return list_next_or_null_rcu(head, &clt_path->s.entry, typeof(*clt_path), s.entry) ?:
drivers/infiniband/ulp/rtrs/rtrs-clt.c
765
list_next_or_null_rcu(head,
drivers/infiniband/ulp/rtrs/rtrs-pri.h
315
struct ib_send_wr *head);
drivers/infiniband/ulp/rtrs/rtrs-pri.h
320
struct ib_send_wr *head,
drivers/infiniband/ulp/rtrs/rtrs.c
109
static int rtrs_post_send(struct ib_qp *qp, struct ib_send_wr *head,
drivers/infiniband/ulp/rtrs/rtrs.c
112
if (head) {
drivers/infiniband/ulp/rtrs/rtrs.c
113
struct ib_send_wr *next = head;
drivers/infiniband/ulp/rtrs/rtrs.c
119
head = wr;
drivers/infiniband/ulp/rtrs/rtrs.c
125
return ib_post_send(qp, head, NULL);
drivers/infiniband/ulp/rtrs/rtrs.c
129
struct ib_send_wr *head)
drivers/infiniband/ulp/rtrs/rtrs.c
150
return rtrs_post_send(con->qp, head, &wr, NULL);
drivers/infiniband/ulp/rtrs/rtrs.c
158
struct ib_send_wr *head,
drivers/infiniband/ulp/rtrs/rtrs.c
183
return rtrs_post_send(con->qp, head, &wr.wr, tail);
drivers/infiniband/ulp/rtrs/rtrs.c
190
struct ib_send_wr *head)
drivers/infiniband/ulp/rtrs/rtrs.c
207
return rtrs_post_send(con->qp, head, &wr.wr, NULL);
drivers/input/evdev.c
103
unsigned int i, head, num;
drivers/input/evdev.c
110
head = client->tail;
drivers/input/evdev.c
116
for (i = client->tail; i != client->head; i = (i + 1) & mask) {
drivers/input/evdev.c
126
} else if (head != i) {
drivers/input/evdev.c
128
client->buffer[head] = *ev;
drivers/input/evdev.c
132
head = (head + 1) & mask;
drivers/input/evdev.c
136
client->packet_head = head;
drivers/input/evdev.c
140
client->head = head;
drivers/input/evdev.c
155
client->buffer[client->head++] = ev;
drivers/input/evdev.c
156
client->head &= client->bufsize - 1;
drivers/input/evdev.c
158
if (unlikely(client->head == client->tail)) {
drivers/input/evdev.c
160
client->tail = (client->head - 1) & (client->bufsize - 1);
drivers/input/evdev.c
203
if (client->head != client->tail) {
drivers/input/evdev.c
204
client->packet_head = client->head = client->tail;
drivers/input/evdev.c
217
client->buffer[client->head++] = *event;
drivers/input/evdev.c
218
client->head &= client->bufsize - 1;
drivers/input/evdev.c
220
if (unlikely(client->head == client->tail)) {
drivers/input/evdev.c
225
client->tail = (client->head - 2) & (client->bufsize - 1);
drivers/input/evdev.c
239
client->packet_head = client->head;
drivers/input/evdev.c
269
if (client->packet_head == client->head)
drivers/input/evdev.c
41
unsigned int head;
drivers/input/joydev.c
101
client->head++;
drivers/input/joydev.c
102
client->head &= JOYDEV_BUFFER_SIZE - 1;
drivers/input/joydev.c
103
if (client->tail == client->head)
drivers/input/joydev.c
326
have_event = client->head != client->tail;
drivers/input/joydev.c
364
client->tail = client->head;
drivers/input/joydev.c
380
client->head != client->tail;
drivers/input/joydev.c
59
int head;
drivers/input/joydev.c
98
client->buffer[client->head] = *event;
drivers/input/joystick/iforce/iforce-packets.c
34
int head, tail;
drivers/input/joystick/iforce/iforce-packets.c
40
head = iforce->xmit.head;
drivers/input/joystick/iforce/iforce-packets.c
43
if (CIRC_SPACE(head, tail, XMIT_SIZE) < n + 2) {
drivers/input/joystick/iforce/iforce-packets.c
49
empty = head == tail;
drivers/input/joystick/iforce/iforce-packets.c
50
XMIT_INC(iforce->xmit.head, n + 2);
drivers/input/joystick/iforce/iforce-packets.c
55
iforce->xmit.buf[head] = HI(cmd);
drivers/input/joystick/iforce/iforce-packets.c
56
XMIT_INC(head, 1);
drivers/input/joystick/iforce/iforce-packets.c
57
iforce->xmit.buf[head] = LO(cmd);
drivers/input/joystick/iforce/iforce-packets.c
58
XMIT_INC(head, 1);
drivers/input/joystick/iforce/iforce-packets.c
60
c = CIRC_SPACE_TO_END(head, tail, XMIT_SIZE);
drivers/input/joystick/iforce/iforce-packets.c
64
memcpy(&iforce->xmit.buf[head], data, c);
drivers/input/joystick/iforce/iforce-packets.c
68
XMIT_INC(head, n);
drivers/input/joystick/iforce/iforce-serio.c
40
if (iforce->xmit.head == iforce->xmit.tail)
drivers/input/joystick/iforce/iforce-usb.c
31
if (iforce->xmit.head == iforce->xmit.tail) {
drivers/input/joystick/iforce/iforce-usb.c
45
c = CIRC_CNT_TO_END(iforce->xmit.head, iforce->xmit.tail, XMIT_SIZE);
drivers/input/misc/uinput.c
62
unsigned char head;
drivers/input/misc/uinput.c
676
have_event = udev->head != udev->tail;
drivers/input/misc/uinput.c
721
else if (udev->head == udev->tail &&
drivers/input/misc/uinput.c
734
udev->head != udev->tail ||
drivers/input/misc/uinput.c
748
if (udev->head != udev->tail)
drivers/input/misc/uinput.c
80
udev->buff[udev->head] = (struct input_event) {
drivers/input/misc/uinput.c
88
udev->head = (udev->head + 1) % UINPUT_BUFFER_SIZE;
drivers/input/mouse/cyapa_gen6.c
565
struct pip_app_cmd_head head;
drivers/input/mouse/cyapa_gen6.c
575
put_unaligned_le16(PIP_OUTPUT_REPORT_ADDR, &cmd.head.addr);
drivers/input/mouse/cyapa_gen6.c
576
put_unaligned_le16(sizeof(cmd) - 2, &cmd.head.length);
drivers/input/mouse/cyapa_gen6.c
577
cmd.head.report_id = PIP_APP_CMD_REPORT_ID;
drivers/input/mouse/cyapa_gen6.c
578
cmd.head.cmd_code = PIP_RETRIEVE_DATA_STRUCTURE;
drivers/input/mousedev.c
100
unsigned int head, tail;
drivers/input/mousedev.c
275
p = &client->packets[client->head];
drivers/input/mousedev.c
277
new_head = (client->head + 1) % PACKET_QUEUE_LEN;
drivers/input/mousedev.c
279
p = &client->packets[client->head = new_head];
drivers/input/mousedev.c
620
if (client->tail == client->head) {
drivers/input/serio/sa1111ps2.c
116
unsigned int head;
drivers/input/serio/sa1111ps2.c
126
if (ps2if->head == ps2if->tail)
drivers/input/serio/sa1111ps2.c
128
head = (ps2if->head + 1) & (sizeof(ps2if->buf) - 1);
drivers/input/serio/sa1111ps2.c
129
if (head != ps2if->tail) {
drivers/input/serio/sa1111ps2.c
130
ps2if->buf[ps2if->head] = val;
drivers/input/serio/sa1111ps2.c
131
ps2if->head = head;
drivers/input/serio/sa1111ps2.c
51
unsigned int head;
drivers/input/serio/sa1111ps2.c
98
if (ps2if->head == ps2if->tail) {
drivers/input/serio/serio_raw.c
133
if (serio_raw->head == serio_raw->tail)
drivers/input/serio/serio_raw.c
155
if (serio_raw->head == serio_raw->tail &&
drivers/input/serio/serio_raw.c
173
serio_raw->head != serio_raw->tail ||
drivers/input/serio/serio_raw.c
225
if (serio_raw->head != serio_raw->tail)
drivers/input/serio/serio_raw.c
252
unsigned int head = serio_raw->head;
drivers/input/serio/serio_raw.c
255
serio_raw->queue[head] = data;
drivers/input/serio/serio_raw.c
256
head = (head + 1) % SERIO_RAW_QUEUE_LEN;
drivers/input/serio/serio_raw.c
257
if (likely(head != serio_raw->tail)) {
drivers/input/serio/serio_raw.c
258
serio_raw->head = head;
drivers/input/serio/serio_raw.c
30
unsigned int tail, head;
drivers/input/serio/userio.c
122
available = CIRC_CNT_TO_END(userio->head, userio->tail,
drivers/input/serio/userio.c
166
userio->head != userio->tail);
drivers/input/serio/userio.c
256
if (userio->head != userio->tail)
drivers/input/serio/userio.c
41
u8 head;
drivers/input/serio/userio.c
60
userio->buf[userio->head] = val;
drivers/input/serio/userio.c
61
userio->head = (userio->head + 1) % USERIO_BUFSIZE;
drivers/input/serio/userio.c
63
if (userio->head == userio->tail)
drivers/iommu/amd/debugfs.c
119
u32 head, tail;
drivers/iommu/amd/debugfs.c
123
head = readl(iommu->mmio_base + MMIO_CMD_HEAD_OFFSET);
drivers/iommu/amd/debugfs.c
126
(head >> 4) & 0x7fff, (tail >> 4) & 0x7fff);
drivers/iommu/amd/iommu.c
1003
u32 head, tail;
drivers/iommu/amd/iommu.c
1005
head = readl(iommu->mmio_base + MMIO_EVT_HEAD_OFFSET);
drivers/iommu/amd/iommu.c
1008
while (head != tail) {
drivers/iommu/amd/iommu.c
1009
iommu_print_event(iommu, iommu->evt_buf + head);
drivers/iommu/amd/iommu.c
1012
head = (head + EVENT_ENTRY_SIZE) % EVT_BUFFER_SIZE;
drivers/iommu/amd/iommu.c
1013
writel(head, iommu->mmio_base + MMIO_EVT_HEAD_OFFSET);
drivers/iommu/amd/iommu.c
1039
u32 head, tail;
drivers/iommu/amd/iommu.c
1044
head = readl(iommu->mmio_base + MMIO_GA_HEAD_OFFSET);
drivers/iommu/amd/iommu.c
1047
while (head != tail) {
drivers/iommu/amd/iommu.c
1051
raw = (u64 *)(iommu->ga_log + head);
drivers/iommu/amd/iommu.c
1057
head = (head + GA_ENTRY_SIZE) % GA_LOG_SIZE;
drivers/iommu/amd/iommu.c
1058
writel(head, iommu->mmio_base + MMIO_GA_HEAD_OFFSET);
drivers/iommu/amd/iommu.c
1180
u32 head, tail;
drivers/iommu/amd/iommu.c
1183
head = readl(iommu->mmio_base + MMIO_CMD_HEAD_OFFSET);
drivers/iommu/amd/iommu.c
1186
pr_err("CMD Buffer head=%llu tail=%llu\n", MMIO_CMD_BUFFER_HEAD(head),
drivers/iommu/amd/iommu.c
3050
struct list_head *head)
drivers/iommu/amd/iommu.c
3090
list_add_tail(&region->list, head);
drivers/iommu/amd/iommu.c
3098
list_add_tail(&region->list, head);
drivers/iommu/amd/iommu.c
3108
list_add_tail(&region->list, head);
drivers/iommu/amd/ppr.c
164
u32 head, tail;
drivers/iommu/amd/ppr.c
169
head = readl(iommu->mmio_base + MMIO_PPR_HEAD_OFFSET);
drivers/iommu/amd/ppr.c
172
while (head != tail) {
drivers/iommu/amd/ppr.c
177
raw = (u64 *)(iommu->ppr_log + head);
drivers/iommu/amd/ppr.c
204
head = (head + PPR_ENTRY_SIZE) % PPR_LOG_SIZE;
drivers/iommu/amd/ppr.c
205
writel(head, iommu->mmio_base + MMIO_PPR_HEAD_OFFSET);
drivers/iommu/apple-dart.c
1000
iommu_dma_get_resv_regions(dev, head);
drivers/iommu/apple-dart.c
985
struct list_head *head)
drivers/iommu/apple-dart.c
997
list_add_tail(&region->list, head);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3703
struct list_head *head)
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3713
list_add_tail(&region->list, head);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3715
iommu_dma_get_resv_regions(dev, head);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
809
struct arm_smmu_ll_queue llq, head;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
827
head.cons = llq.cons;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
828
head.prod = queue_inc_prod_n(&llq, n + sync) |
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
831
old = cmpxchg_relaxed(&cmdq->q.llq.val, llq.val, head.val);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
838
head.prod &= ~CMDQ_PROD_OWNED_FLAG;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
862
arm_smmu_cmdq_set_valid_map(cmdq, llq.prod, head.prod);
drivers/iommu/arm/arm-smmu/arm-smmu.c
1607
struct list_head *head)
drivers/iommu/arm/arm-smmu/arm-smmu.c
1617
list_add_tail(&region->list, head);
drivers/iommu/arm/arm-smmu/arm-smmu.c
1619
iommu_dma_get_resv_regions(dev, head);
drivers/iommu/dma-iommu.c
116
unsigned int head, tail;
drivers/iommu/dma-iommu.c
122
for ((i) = (fq)->head; (i) != (fq)->tail; (i) = ((i) + 1) & (fq)->mod_mask)
drivers/iommu/dma-iommu.c
127
return (((fq->tail + 1) & fq->mod_mask) == fq->head);
drivers/iommu/dma-iommu.c
160
fq->head = (fq->head + 1) & fq->mod_mask;
drivers/iommu/dma-iommu.c
289
fq->head = 0;
drivers/iommu/intel/dmar.c
1242
unsigned int head = dmar_readl(iommu->reg + DMAR_IQH_REG);
drivers/iommu/intel/dmar.c
1244
struct qi_desc *desc = iommu->qi->desc + head;
drivers/iommu/intel/dmar.c
1261
head = ((head >> qi_shift(iommu)) + QI_LENGTH - 1) % QI_LENGTH;
drivers/iommu/intel/dmar.c
1262
head <<= qi_shift(iommu);
drivers/iommu/intel/dmar.c
1263
desc = iommu->qi->desc + head;
drivers/iommu/intel/dmar.c
1274
int head, tail;
drivers/iommu/intel/dmar.c
1293
head = readl(iommu->reg + DMAR_IQH_REG);
drivers/iommu/intel/dmar.c
1294
if ((head >> shift) == index) {
drivers/iommu/intel/dmar.c
1295
struct qi_desc *desc = qi->desc + head;
drivers/iommu/intel/dmar.c
1315
head = readl(iommu->reg + DMAR_IQH_REG);
drivers/iommu/intel/dmar.c
1316
head = ((head >> shift) - 1 + QI_LENGTH) % QI_LENGTH;
drivers/iommu/intel/dmar.c
1331
if (qi->desc_status[head] == QI_IN_USE)
drivers/iommu/intel/dmar.c
1332
qi->desc_status[head] = QI_ABORT;
drivers/iommu/intel/dmar.c
1333
head = (head - 1 + QI_LENGTH) % QI_LENGTH;
drivers/iommu/intel/dmar.c
1334
} while (head != tail);
drivers/iommu/intel/iommu.c
3380
struct list_head *head)
drivers/iommu/intel/iommu.c
3411
list_add_tail(&resv->list, head);
drivers/iommu/intel/iommu.c
3425
list_add_tail(&reg->list, head);
drivers/iommu/intel/iommu.c
3435
list_add_tail(&reg->list, head);
drivers/iommu/intel/prq.c
201
int head, tail, handled;
drivers/iommu/intel/prq.c
212
head = dmar_readq(iommu->reg + DMAR_PQH_REG) & PRQ_RING_MASK;
drivers/iommu/intel/prq.c
213
handled = (head != tail);
drivers/iommu/intel/prq.c
214
while (head != tail) {
drivers/iommu/intel/prq.c
215
req = &iommu->prq[head / sizeof(*req)];
drivers/iommu/intel/prq.c
259
head = (head + sizeof(*req)) & PRQ_RING_MASK;
drivers/iommu/intel/prq.c
271
head = dmar_readq(iommu->reg + DMAR_PQH_REG) & PRQ_RING_MASK;
drivers/iommu/intel/prq.c
273
if (head == tail) {
drivers/iommu/intel/prq.c
66
int head, tail;
drivers/iommu/intel/prq.c
85
head = dmar_readq(iommu->reg + DMAR_PQH_REG) & PRQ_RING_MASK;
drivers/iommu/intel/prq.c
86
while (head != tail) {
drivers/iommu/intel/prq.c
89
req = &iommu->prq[head / sizeof(*req)];
drivers/iommu/intel/prq.c
93
head = (head + sizeof(*req)) & PRQ_RING_MASK;
drivers/iommu/iommu.c
928
struct list_head *head)
drivers/iommu/iommu.c
946
ret = iommu_insert_device_resv_regions(&dev_resv_regions, head);
drivers/iommu/mtk_iommu.c
1028
struct list_head *head)
drivers/iommu/mtk_iommu.c
1053
list_add_tail(&region->list, head);
drivers/iommu/mtk_iommu.c
345
#define for_each_m4u(data, head) list_for_each_entry(data, head, list)
drivers/iommu/mtk_iommu.c
407
struct list_head *head = bank->parent_data->hw_list;
drivers/iommu/mtk_iommu.c
416
for_each_m4u(data, head) {
drivers/iommu/riscv/iommu.c
314
unsigned int head = atomic_read(&queue->head);
drivers/iommu/riscv/iommu.c
317
int available = (int)(tail - head);
drivers/iommu/riscv/iommu.c
319
*index = head;
drivers/iommu/riscv/iommu.c
337
return (int)(atomic_add_return((tail - last) & queue->mask, &queue->tail) - head);
drivers/iommu/riscv/iommu.c
345
const unsigned int head = atomic_add_return(count, &queue->head);
drivers/iommu/riscv/iommu.c
347
riscv_iommu_writel(queue->iommu, Q_HEAD(queue), Q_ITEM(queue, head));
drivers/iommu/riscv/iommu.c
353
const unsigned int cons = atomic_read(&queue->head);
drivers/iommu/riscv/iommu.c
355
unsigned int head;
drivers/iommu/riscv/iommu.c
357
if (riscv_iommu_readl_timeout(queue->iommu, Q_HEAD(queue), head,
drivers/iommu/riscv/iommu.c
358
!(head & ~queue->mask),
drivers/iommu/riscv/iommu.c
362
return cons + ((head - last) & queue->mask);
drivers/iommu/riscv/iommu.c
370
unsigned int cons = atomic_read(&queue->head);
drivers/iommu/riscv/iommu.c
391
unsigned int head;
drivers/iommu/riscv/iommu.c
400
head = atomic_read(&queue->head);
drivers/iommu/riscv/iommu.c
403
if ((prod - head) > queue->mask) {
drivers/iommu/riscv/iommu.c
404
if (readx_poll_timeout(atomic_read, &queue->head,
drivers/iommu/riscv/iommu.c
405
head, (prod - head) < queue->mask,
drivers/iommu/riscv/iommu.c
408
} else if ((prod - head) == queue->mask) {
drivers/iommu/riscv/iommu.c
409
const unsigned int last = Q_ITEM(queue, head);
drivers/iommu/riscv/iommu.c
411
if (riscv_iommu_readl_timeout(queue->iommu, Q_HEAD(queue), head,
drivers/iommu/riscv/iommu.c
412
!(head & ~queue->mask) && head != last,
drivers/iommu/riscv/iommu.c
415
atomic_add((head - last) & queue->mask, &queue->head);
drivers/iommu/riscv/iommu.h
24
atomic_t head; /* unbounded shadow ring buffer consumer index */
drivers/iommu/s390-iommu.c
571
static void s390_iommu_rcu_free_domain(struct rcu_head *head)
drivers/iommu/s390-iommu.c
573
struct s390_domain *s390_domain = container_of(head, struct s390_domain, rcu);
drivers/iommu/virtio-iommu.c
1000
iommu_dma_get_resv_regions(dev, head);
drivers/iommu/virtio-iommu.c
463
.head.type = VIRTIO_IOMMU_T_MAP,
drivers/iommu/virtio-iommu.c
551
probe->head.type = VIRTIO_IOMMU_T_PROBE;
drivers/iommu/virtio-iommu.c
642
} else if (!(evt->head & VIOMMU_FAULT_RESV_MASK)) {
drivers/iommu/virtio-iommu.c
760
.head.type = VIRTIO_IOMMU_T_ATTACH,
drivers/iommu/virtio-iommu.c
794
.head.type = VIRTIO_IOMMU_T_ATTACH,
drivers/iommu/virtio-iommu.c
830
.head.type = VIRTIO_IOMMU_T_DETACH,
drivers/iommu/virtio-iommu.c
866
.head.type = VIRTIO_IOMMU_T_MAP,
drivers/iommu/virtio-iommu.c
905
.head.type = VIRTIO_IOMMU_T_UNMAP,
drivers/iommu/virtio-iommu.c
94
u32 head;
drivers/iommu/virtio-iommu.c
970
static void viommu_get_resv_regions(struct device *dev, struct list_head *head)
drivers/iommu/virtio-iommu.c
983
list_add_tail(&new_entry->list, head);
drivers/iommu/virtio-iommu.c
997
list_add_tail(&msi->list, head);
drivers/irqchip/irq-loongarch-avec.c
159
if (!list_empty(&plist->head))
drivers/irqchip/irq-loongarch-avec.c
176
list_for_each_entry_safe(adata, tdata, &plist->head, entry) {
drivers/irqchip/irq-loongarch-avec.c
31
struct list_head head;
drivers/irqchip/irq-loongarch-avec.c
87
INIT_LIST_HEAD(&plist->head);
drivers/irqchip/irq-loongarch-avec.c
96
list_add_tail(&adata->entry, &plist->head);
drivers/irqchip/irq-riscv-aplic-main.c
106
list_for_each_entry(priv, &aplics, head)
drivers/irqchip/irq-riscv-aplic-main.c
152
list_del(&priv->head);
drivers/irqchip/irq-riscv-aplic-main.c
167
list_add(&priv->head, &aplics);
drivers/irqchip/irq-riscv-aplic-main.c
184
list_del(&priv->head);
drivers/irqchip/irq-riscv-aplic-main.c
96
list_for_each_entry(priv, &aplics, head)
drivers/irqchip/irq-riscv-aplic-main.h
42
struct list_head head;
drivers/isdn/mISDN/socket.c
49
sk_add_node(sk, &l->head);
drivers/isdn/mISDN/socket.c
493
sk_for_each(csk, &data_sockets.head) {
drivers/isdn/mISDN/stack.c
151
if (!hlist_empty(&st->l1sock.head)) {
drivers/isdn/mISDN/stack.c
157
if (!hlist_empty(&st->l1sock.head))
drivers/isdn/mISDN/stack.c
376
INIT_HLIST_HEAD(&newst->l1sock.head);
drivers/isdn/mISDN/stack.c
445
sk_add_node(&msk->sk, &dev->D.st->l1sock.head);
drivers/isdn/mISDN/stack.c
644
if (!hlist_empty(&st->l1sock.head))
drivers/isdn/mISDN/stack.c
65
sk_for_each(sk, &sl->head) {
drivers/md/dm-bio-prison-v1.c
346
static void __sweep(struct dm_deferred_set *ds, struct list_head *head)
drivers/md/dm-bio-prison-v1.c
350
list_splice_init(&ds->entries[ds->sweeper].work_items, head);
drivers/md/dm-bio-prison-v1.c
355
list_splice_init(&ds->entries[ds->sweeper].work_items, head);
drivers/md/dm-bio-prison-v1.c
358
void dm_deferred_entry_dec(struct dm_deferred_entry *entry, struct list_head *head)
drivers/md/dm-bio-prison-v1.c
365
__sweep(entry->ds, head);
drivers/md/dm-bio-prison-v1.h
125
void dm_deferred_entry_dec(struct dm_deferred_entry *entry, struct list_head *head);
drivers/md/dm-cache-policy-smq.c
114
unsigned int head, tail;
drivers/md/dm-cache-policy-smq.c
120
l->head = l->tail = INDEXER_NULL;
drivers/md/dm-cache-policy-smq.c
125
return to_entry(es, l->head);
drivers/md/dm-cache-policy-smq.c
145
return l->head == INDEXER_NULL;
drivers/md/dm-cache-policy-smq.c
150
struct entry *head = l_head(es, l);
drivers/md/dm-cache-policy-smq.c
152
e->next = l->head;
drivers/md/dm-cache-policy-smq.c
155
if (head)
drivers/md/dm-cache-policy-smq.c
156
head->prev = l->head = to_index(es, e);
drivers/md/dm-cache-policy-smq.c
158
l->head = l->tail = to_index(es, e);
drivers/md/dm-cache-policy-smq.c
174
l->head = l->tail = to_index(es, e);
drivers/md/dm-cache-policy-smq.c
206
l->head = e->next;
drivers/md/dm-clone-target.c
1012
batch->head = hd;
drivers/md/dm-clone-target.c
1067
.head = NULL,
drivers/md/dm-clone-target.c
1106
if (batch.head)
drivers/md/dm-clone-target.c
1107
hydration_copy(batch.head, batch.nr_batched_regions);
drivers/md/dm-clone-target.c
558
struct hlist_head head;
drivers/md/dm-clone-target.c
590
INIT_HLIST_HEAD(&bucket->head);
drivers/md/dm-clone-target.c
618
hlist_for_each_entry(hd, &bucket->head, h) {
drivers/md/dm-clone-target.c
634
hlist_add_head(&hd->h, &bucket->head);
drivers/md/dm-clone-target.c
974
struct dm_clone_region_hydration *head;
drivers/md/dm-clone-target.c
984
if (batch->head) {
drivers/md/dm-clone-target.c
987
(batch->head->region_nr + batch->nr_batched_regions) == hd->region_nr) {
drivers/md/dm-clone-target.c
988
list_add_tail(&hd->list, &batch->head->list);
drivers/md/dm-clone-target.c
995
hydration_copy(batch->head, batch->nr_batched_regions);
drivers/md/dm-clone-target.c
996
batch->head = NULL;
drivers/md/dm-raid1.c
134
should_wake = !(bl->head);
drivers/md/dm-raid1.c
645
if (!ms->failures.head)
drivers/md/dm-raid1.c
695
if (!writes->head)
drivers/md/dm-raid1.c
744
if (unlikely(requeue.head)) {
drivers/md/dm-raid1.c
798
if (likely(!failures->head))
drivers/md/dm-region-hash.c
533
for (bio = bios->head; bio; bio = bio->bi_next) {
drivers/md/dm-snap.c
44
struct hlist_head head;
drivers/md/dm-snap.c
671
INIT_HLIST_HEAD(&et->table[i].head);
drivers/md/dm-snap.c
690
hlist_for_each_entry_safe(ex, pos, &slot->head, hash_list) {
drivers/md/dm-snap.c
720
slot = &et->table[exception_hash(et, chunk)].head;
drivers/md/dm-snap.c
771
l = &eh->table[exception_hash(eh, new_e->old_chunk)].head;
drivers/md/dm-stats.c
172
static void dm_stat_free(struct rcu_head *head)
drivers/md/dm-stats.c
175
struct dm_stat *s = container_of(head, struct dm_stat, rcu_head);
drivers/md/dm-thin.c
1216
static void process_prepared(struct pool *pool, struct list_head *head,
drivers/md/dm-thin.c
1224
list_splice_init(head, &maps);
drivers/md/dm-vdo/dm-vdo-target.c
2614
vdo_forget(vdo->next_layout.head);
drivers/md/dm-vdo/encodings.c
808
partition->next = layout->head;
drivers/md/dm-vdo/encodings.c
809
layout->head = partition;
drivers/md/dm-vdo/encodings.c
888
.head = NULL,
drivers/md/dm-vdo/encodings.c
926
while (layout->head != NULL) {
drivers/md/dm-vdo/encodings.c
927
struct partition *part = layout->head;
drivers/md/dm-vdo/encodings.c
929
layout->head = part->next;
drivers/md/dm-vdo/encodings.c
949
for (partition = layout->head; partition != NULL; partition = partition->next) {
drivers/md/dm-vdo/encodings.c
997
for (partition = layout->head; partition != NULL; partition = partition->next) {
drivers/md/dm-vdo/encodings.h
1150
packed->head = __cpu_to_le64(header->head);
drivers/md/dm-vdo/encodings.h
1171
.head = __le64_to_cpu(packed->head),
drivers/md/dm-vdo/encodings.h
472
sequence_number_t head;
drivers/md/dm-vdo/encodings.h
493
__le64 head;
drivers/md/dm-vdo/encodings.h
577
struct partition *head;
drivers/md/dm-vdo/io-submitter.c
149
vio->bios_merged.head->bi_iter.bi_sector);
drivers/md/dm-vdo/io-submitter.c
152
bio = vio->bios_merged.head;
drivers/md/dm-vdo/io-submitter.c
222
return (vio_merge->bios_merged.head->bi_iter.bi_sector == merge_sector ?
drivers/md/dm-vdo/io-submitter.c
231
bio_sector = vio->bios_merged.head->bi_iter.bi_sector;
drivers/md/dm-vdo/io-submitter.c
256
vdo_int_map_remove(bio_map, next_vio->bios_merged.head->bi_iter.bi_sector);
drivers/md/dm-vdo/repair.c
1555
sequence_number_t i, head;
drivers/md/dm-vdo/repair.c
1561
head = min(repair->block_map_head, repair->slab_journal_head);
drivers/md/dm-vdo/repair.c
1562
header = get_recovery_journal_block_header(journal, repair->journal_data, head);
drivers/md/dm-vdo/repair.c
1564
for (i = head; i <= repair->highest_tail; i++) {
drivers/md/dm-vdo/slab-depot.c
102
journal->unreapable = journal->head;
drivers/md/dm-vdo/slab-depot.c
1830
VDO_ASSERT_LOG_ONLY((journal->head + journal->size) == journal->tail,
drivers/md/dm-vdo/slab-depot.c
2513
journal->head = (slab->allocator->summary_entries[slab->slab_number].is_dirty ?
drivers/md/dm-vdo/slab-depot.c
2514
header.head : journal->tail);
drivers/md/dm-vdo/slab-depot.c
2922
sequence_number_t head = __le64_to_cpu(end_block->header.head);
drivers/md/dm-vdo/slab-depot.c
2923
tail_block_offset_t head_index = head % journal->size;
drivers/md/dm-vdo/slab-depot.c
2930
for (sequence = head; sequence < tail; sequence++) {
drivers/md/dm-vdo/slab-depot.c
3680
(unsigned long long) journal->head,
drivers/md/dm-vdo/slab-depot.c
3762
journal->head = 1;
drivers/md/dm-vdo/slab-depot.c
403
journal->head = journal->unreapable;
drivers/md/dm-vdo/slab-depot.c
654
(journal->head > 1), false, free_block_count);
drivers/md/dm-vdo/slab-depot.c
668
journal->head = journal->tail;
drivers/md/dm-vdo/slab-depot.c
755
header->head = journal->head;
drivers/md/dm-vdo/slab-depot.c
80
return (journal->head != journal->unreapable);
drivers/md/dm-vdo/slab-depot.c
942
return journal->tail - journal->head;
drivers/md/dm-vdo/slab-depot.c
992
journal->head++;
drivers/md/dm-vdo/slab-depot.h
96
sequence_number_t head;
drivers/md/dm.c
1692
struct dm_io **head = dm_poll_list_head(bio);
drivers/md/dm.c
1711
io->data = (*head)->data;
drivers/md/dm.c
1712
io->next = *head;
drivers/md/dm.c
1715
*head = io;
drivers/md/dm.c
2111
struct dm_io **head = dm_poll_list_head(bio);
drivers/md/dm.c
2112
struct dm_io *list = *head;
drivers/md/dm.c
2151
*head = tmp;
drivers/md/md-bitmap.c
2976
.head = {
drivers/md/md-bitmap.c
3026
return register_md_submodule(&bitmap_ops.head);
drivers/md/md-bitmap.c
3032
unregister_md_submodule(&bitmap_ops.head);
drivers/md/md-bitmap.h
84
struct md_submodule_head head;
drivers/md/md-cluster.c
1621
.head = {
drivers/md/md-cluster.c
1657
return register_md_submodule(&cluster_ops.head);
drivers/md/md-cluster.c
1662
unregister_md_submodule(&cluster_ops.head);
drivers/md/md-cluster.h
13
struct md_submodule_head head;
drivers/md/md-linear.c
317
.head = {
drivers/md/md-linear.c
336
return register_md_submodule(&linear_personality.head);
drivers/md/md-linear.c
341
unregister_md_submodule(&linear_personality.head);
drivers/md/md-llbitmap.c
1565
.head = {
drivers/md/md-llbitmap.c
1618
return register_md_submodule(&llbitmap_ops.head);
drivers/md/md-llbitmap.c
1627
unregister_md_submodule(&llbitmap_ops.head);
drivers/md/md.c
1021
struct md_submodule_head *head;
drivers/md/md.c
1025
xa_for_each(&md_submodule, i, head) {
drivers/md/md.c
1026
if (head->type != MD_PERSONALITY)
drivers/md/md.c
1028
if ((level != LEVEL_NONE && head->id == level) ||
drivers/md/md.c
1029
!strcmp(head->name, clevel)) {
drivers/md/md.c
1030
if (try_module_get(head->owner))
drivers/md/md.c
1031
ret = (void *)head;
drivers/md/md.c
1051
module_put(pers->head.owner);
drivers/md/md.c
1345
mdname(mddev), mddev->pers->head.name);
drivers/md/md.c
4019
ret = sprintf(page, "%s\n", p->head.name);
drivers/md/md.c
4076
mdname(mddev), mddev->pers->head.name);
drivers/md/md.c
4139
strscpy(mddev->clevel, pers->head.name, sizeof(mddev->clevel));
drivers/md/md.c
4260
struct md_submodule_head *head;
drivers/md/md.c
4270
xa_for_each(&md_submodule, i, head) {
drivers/md/md.c
4271
if (head->type != MD_BITMAP)
drivers/md/md.c
4274
if (mddev->bitmap_id == head->id)
drivers/md/md.c
4275
len += sprintf(page + len, "[%s] ", head->name);
drivers/md/md.c
4277
len += sprintf(page + len, "%s ", head->name);
drivers/md/md.c
4288
struct md_submodule_head *head;
drivers/md/md.c
4305
xa_for_each(&md_submodule, i, head) {
drivers/md/md.c
4306
if (head->type == MD_BITMAP && cmd_match(buf, head->name)) {
drivers/md/md.c
4307
mddev->bitmap_id = head->id;
drivers/md/md.c
4321
head = xa_load(&md_submodule, id);
drivers/md/md.c
4322
if (head && head->type == MD_BITMAP) {
drivers/md/md.c
5895
if (mddev->pers == NULL || (mddev->pers->head.id != ID_RAID1))
drivers/md/md.c
5922
if (mddev->pers == NULL || (mddev->pers->head.id != ID_RAID1)) {
drivers/md/md.c
6552
if (mddev->level != pers->head.id) {
drivers/md/md.c
6553
mddev->level = pers->head.id;
drivers/md/md.c
6554
mddev->new_level = pers->head.id;
drivers/md/md.c
6556
strscpy(mddev->clevel, pers->head.name, sizeof(mddev->clevel));
drivers/md/md.c
684
struct md_submodule_head *head;
drivers/md/md.c
687
(old && old->head.id == mddev->bitmap_id))
drivers/md/md.c
691
head = xa_load(&md_submodule, mddev->bitmap_id);
drivers/md/md.c
693
if (!head) {
drivers/md/md.c
698
if (head->type != MD_BITMAP) {
drivers/md/md.c
703
mddev->bitmap_ops = (void *)head;
drivers/md/md.c
7894
!try_module_get(mddev->cluster_ops->head.owner))
drivers/md/md.c
7907
module_put(mddev->cluster_ops->head.owner);
drivers/md/md.c
8591
if (mddev->pers->head.id == ID_RAID0 ||
drivers/md/md.c
8592
mddev->pers->head.id == ID_LINEAR)
drivers/md/md.c
8630
struct md_submodule_head *head;
drivers/md/md.c
8636
xa_for_each(&md_submodule, i, head)
drivers/md/md.c
8637
if (head->type == MD_PERSONALITY)
drivers/md/md.c
8638
seq_printf(seq, "[%s] ", head->name);
drivers/md/md.c
8866
seq_printf(seq, " %s", mddev->pers->head.name);
drivers/md/md.h
748
struct md_submodule_head head;
drivers/md/md.h
848
#define rdev_for_each_list(rdev, tmp, head) \
drivers/md/md.h
849
list_for_each_entry_safe(rdev, tmp, head, same_set)
drivers/md/raid0.c
808
.head = {
drivers/md/raid0.c
827
return register_md_submodule(&raid0_personality.head);
drivers/md/raid0.c
832
unregister_md_submodule(&raid0_personality.head);
drivers/md/raid1.c
2661
struct list_head *head = &conf->retry_list;
drivers/md/raid1.c
2692
if (list_empty(head)) {
drivers/md/raid1.c
2696
r1_bio = list_entry(head->prev, struct r1bio, retry_list);
drivers/md/raid1.c
2697
list_del(head->prev);
drivers/md/raid1.c
3461
.head = {
drivers/md/raid1.c
3486
return register_md_submodule(&raid1_personality.head);
drivers/md/raid1.c
3491
unregister_md_submodule(&raid1_personality.head);
drivers/md/raid1.c
921
if (conf->pending_bio_list.head) {
drivers/md/raid10.c
2966
struct list_head *head = &conf->retry_list;
drivers/md/raid10.c
3000
if (list_empty(head)) {
drivers/md/raid10.c
3004
r10_bio = list_entry(head->prev, struct r10bio, retry_list);
drivers/md/raid10.c
3005
list_del(head->prev);
drivers/md/raid10.c
5046
.head = {
drivers/md/raid10.c
5074
return register_md_submodule(&raid10_personality.head);
drivers/md/raid10.c
5079
unregister_md_submodule(&raid10_personality.head);
drivers/md/raid10.c
868
if (conf->pending_bio_list.head) {
drivers/md/raid5.c
1007
unlock_two_stripes(head, sh);
drivers/md/raid5.c
1009
raid5_release_stripe(head);
drivers/md/raid5.c
374
struct llist_node *head;
drivers/md/raid5.c
376
head = llist_del_all(&conf->released_stripes);
drivers/md/raid5.c
377
head = llist_reverse_order(head);
drivers/md/raid5.c
378
llist_for_each_entry_safe(sh, t, head, release_list) {
drivers/md/raid5.c
5323
struct list_head head;
drivers/md/raid5.c
5324
list_add(&head, &conf->bitmap_list);
drivers/md/raid5.c
5326
while (!list_empty(&head)) {
drivers/md/raid5.c
5327
struct stripe_head *sh = list_entry(head.next, struct stripe_head, lru);
drivers/md/raid5.c
9006
.head = {
drivers/md/raid5.c
9036
.head = {
drivers/md/raid5.c
9067
.head = {
drivers/md/raid5.c
9112
ret = register_md_submodule(&raid6_personality.head);
drivers/md/raid5.c
9116
ret = register_md_submodule(&raid5_personality.head);
drivers/md/raid5.c
9120
ret = register_md_submodule(&raid4_personality.head);
drivers/md/raid5.c
9127
unregister_md_submodule(&raid5_personality.head);
drivers/md/raid5.c
9129
unregister_md_submodule(&raid6_personality.head);
drivers/md/raid5.c
9139
unregister_md_submodule(&raid6_personality.head);
drivers/md/raid5.c
9140
unregister_md_submodule(&raid5_personality.head);
drivers/md/raid5.c
9141
unregister_md_submodule(&raid4_personality.head);
drivers/md/raid5.c
920
struct stripe_head *head;
drivers/md/raid5.c
932
head = last_sh;
drivers/md/raid5.c
933
atomic_inc(&head->count);
drivers/md/raid5.c
937
head = find_get_stripe(conf, head_sector, conf->generation,
drivers/md/raid5.c
940
if (!head)
drivers/md/raid5.c
942
if (!stripe_can_batch(head))
drivers/md/raid5.c
946
lock_two_stripes(head, sh);
drivers/md/raid5.c
948
if (!stripe_can_batch(head) || !stripe_can_batch(sh))
drivers/md/raid5.c
957
if (head->dev[dd_idx].towrite->bi_opf != sh->dev[dd_idx].towrite->bi_opf ||
drivers/md/raid5.c
958
bio_op(head->dev[dd_idx].towrite) != bio_op(sh->dev[dd_idx].towrite))
drivers/md/raid5.c
961
if (head->batch_head) {
drivers/md/raid5.c
962
spin_lock(&head->batch_head->batch_lock);
drivers/md/raid5.c
964
if (!stripe_can_batch(head)) {
drivers/md/raid5.c
965
spin_unlock(&head->batch_head->batch_lock);
drivers/md/raid5.c
975
sh->batch_head = head->batch_head;
drivers/md/raid5.c
981
list_add(&sh->batch_list, &head->batch_list);
drivers/md/raid5.c
982
spin_unlock(&head->batch_head->batch_lock);
drivers/md/raid5.c
984
head->batch_head = head;
drivers/md/raid5.c
985
sh->batch_head = head->batch_head;
drivers/md/raid5.c
986
spin_lock(&head->batch_lock);
drivers/md/raid5.c
987
list_add_tail(&sh->batch_list, &head->batch_list);
drivers/md/raid5.c
988
spin_unlock(&head->batch_lock);
drivers/media/cec/core/cec-notifier.c
24
struct list_head head;
drivers/media/cec/core/cec-notifier.c
57
list_for_each_entry(n, &cec_notifiers, head) {
drivers/media/cec/core/cec-notifier.c
82
list_add_tail(&n->head, &cec_notifiers);
drivers/media/cec/core/cec-notifier.c
93
list_del(&n->head);
drivers/media/common/siano/smscoreapi.c
504
static void list_add_locked(struct list_head *new, struct list_head *head,
drivers/media/common/siano/smscoreapi.c
511
list_add(new, head);
drivers/media/dvb-core/dmxdev.c
104
struct list_head *head, *pos;
drivers/media/dvb-core/dmxdev.c
106
head = demux->get_frontends(demux);
drivers/media/dvb-core/dmxdev.c
107
if (!head)
drivers/media/dvb-core/dmxdev.c
109
list_for_each(pos, head)
drivers/media/dvb-core/dvb_demux.c
1173
struct list_head *head = &dvbdemux->frontend_list;
drivers/media/dvb-core/dvb_demux.c
1175
list_add(&(frontend->connectivity_list), head);
drivers/media/dvb-core/dvb_demux.c
1184
struct list_head *pos, *n, *head = &dvbdemux->frontend_list;
drivers/media/dvb-core/dvb_demux.c
1186
list_for_each_safe(pos, n, head) {
drivers/media/mc/mc-entity.c
1060
static struct media_link *media_add_link(struct list_head *head)
drivers/media/mc/mc-entity.c
1068
list_add_tail(&link->list, head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
118
list_for_each_entry_safe(ib, ib_safe, &bl->head, head) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
128
list_del(&ib->head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
130
list_add(&ib->head, &aq->active);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
132
list_add_tail(&ib->head, &aq->incoming);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
166
list_for_each_entry_safe(ib, ib_safe, &aq->active, head) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
170
list_del(&ib->head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
176
list_add(&ib->head, &aq->incoming);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
205
INIT_LIST_HEAD(&bl->head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
213
if (!list_empty(&bl->head))
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
219
struct ipu6_isys_buffer, head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
224
list_del(&ib->head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
225
list_add(&ib->head, &bl->head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
281
list_for_each_entry(ib, &bl->head, head) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
366
list_add(&ib->head, &aq->incoming);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
485
head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
487
list_del(&ib->head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
504
head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
507
list_del(&ib->head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
764
list_for_each_entry_reverse(ib, &aq->active, head) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
794
list_del(&ib->head);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.h
30
struct list_head head;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.h
44
struct list_head head;
drivers/media/pci/intel/ipu6/ipu6-isys.c
1002
msg = list_last_entry(&isys->framebuflist, struct isys_fw_msgs, head);
drivers/media/pci/intel/ipu6/ipu6-isys.c
1003
list_move(&msg->head, &isys->framebuflist_fw);
drivers/media/pci/intel/ipu6/ipu6-isys.c
1016
list_for_each_entry_safe(fwmsg, fwmsg0, &isys->framebuflist_fw, head)
drivers/media/pci/intel/ipu6/ipu6-isys.c
1017
list_move(&fwmsg->head, &isys->framebuflist);
drivers/media/pci/intel/ipu6/ipu6-isys.c
1032
list_move(&msg->head, &isys->framebuflist);
drivers/media/pci/intel/ipu6/ipu6-isys.c
932
list_for_each_entry_safe(fwmsg, safe, &isys->framebuflist, head)
drivers/media/pci/intel/ipu6/ipu6-isys.c
936
list_for_each_entry_safe(fwmsg, safe, &isys->framebuflist_fw, head)
drivers/media/pci/intel/ipu6/ipu6-isys.c
956
list_add(&addr->head, &isys->framebuflist);
drivers/media/pci/intel/ipu6/ipu6-isys.c
966
struct isys_fw_msgs, head);
drivers/media/pci/intel/ipu6/ipu6-isys.c
967
list_del(&addr->head);
drivers/media/pci/intel/ipu6/ipu6-isys.h
174
struct list_head head;
drivers/media/platform/allegro-dvt/allegro-core.c
121
struct list_head head;
drivers/media/platform/allegro-dvt/allegro-core.c
129
unsigned int head;
drivers/media/platform/allegro-dvt/allegro-core.c
1502
list_for_each_entry(al_buffer, list, head)
drivers/media/platform/allegro-dvt/allegro-core.c
1517
list_for_each_entry(al_buffer, list, head) {
drivers/media/platform/allegro-dvt/allegro-core.c
1559
INIT_LIST_HEAD(&buffer->head);
drivers/media/platform/allegro-dvt/allegro-core.c
1566
list_add(&buffer->head, list);
drivers/media/platform/allegro-dvt/allegro-core.c
1572
list_for_each_entry_safe(buffer, tmp, list, head) {
drivers/media/platform/allegro-dvt/allegro-core.c
1573
list_del(&buffer->head);
drivers/media/platform/allegro-dvt/allegro-core.c
1586
list_for_each_entry_safe(buffer, tmp, list, head) {
drivers/media/platform/allegro-dvt/allegro-core.c
1587
list_del(&buffer->head);
drivers/media/platform/allegro-dvt/allegro-core.c
1979
list_add_tail(&shadow->head, list);
drivers/media/platform/allegro-dvt/allegro-core.c
1993
list_for_each_entry_safe(shadow, tmp, list, head) {
drivers/media/platform/allegro-dvt/allegro-core.c
1996
list_del_init(&shadow->head);
drivers/media/platform/allegro-dvt/allegro-core.c
2927
&channel->source_shadow_list, head) {
drivers/media/platform/allegro-dvt/allegro-core.c
2928
list_del(&shadow->head);
drivers/media/platform/allegro-dvt/allegro-core.c
2938
&channel->stream_shadow_list, head) {
drivers/media/platform/allegro-dvt/allegro-core.c
2939
list_del(&shadow->head);
drivers/media/platform/allegro-dvt/allegro-core.c
357
struct list_head head;
drivers/media/platform/allegro-dvt/allegro-core.c
811
mbox->head = base;
drivers/media/platform/allegro-dvt/allegro-core.c
817
regmap_write(dev->sram, mbox->head, 0);
drivers/media/platform/allegro-dvt/allegro-core.c
861
unsigned int head, tail;
drivers/media/platform/allegro-dvt/allegro-core.c
863
regmap_read(sram, mbox->head, &head);
drivers/media/platform/allegro-dvt/allegro-core.c
866
if (tail >= head)
drivers/media/platform/allegro-dvt/allegro-core.c
867
return tail - head;
drivers/media/platform/allegro-dvt/allegro-core.c
869
return mbox->size - (head - tail);
drivers/media/platform/allegro-dvt/allegro-core.c
880
unsigned int available, head;
drivers/media/platform/allegro-dvt/allegro-core.c
889
regmap_read(sram, mbox->head, &head);
drivers/media/platform/allegro-dvt/allegro-core.c
890
if (head > mbox->size)
drivers/media/platform/allegro-dvt/allegro-core.c
894
regmap_bulk_read(sram, mbox->data + head,
drivers/media/platform/allegro-dvt/allegro-core.c
915
(size_t)(mbox->size - (head + sizeof(*header))));
drivers/media/platform/allegro-dvt/allegro-core.c
916
regmap_bulk_read(sram, mbox->data + head + sizeof(*header),
drivers/media/platform/allegro-dvt/allegro-core.c
923
regmap_write(sram, mbox->head, (head + size) % mbox->size);
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
121
struct list_head *head;
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
144
head = vdec_get_buf_list(msg_ctx->hardware_index, buf);
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
145
if (!head) {
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
150
list_del(head);
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
76
struct list_head *head;
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
78
head = vdec_get_buf_list(msg_ctx->hardware_index, buf);
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
79
if (!head) {
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
85
list_add_tail(head, &msg_ctx->ready_queue);
drivers/media/platform/nuvoton/npcm-video.c
1528
struct list_head *head, *pos, *nx;
drivers/media/platform/nuvoton/npcm-video.c
1539
head = &video->list[vb->index];
drivers/media/platform/nuvoton/npcm-video.c
1540
list_for_each_safe(pos, nx, head) {
drivers/media/platform/nuvoton/npcm-video.c
391
struct list_head *head, *pos, *nx;
drivers/media/platform/nuvoton/npcm-video.c
396
head = &video->list[i];
drivers/media/platform/nuvoton/npcm-video.c
397
list_for_each_safe(pos, nx, head) {
drivers/media/platform/nuvoton/npcm-video.c
410
struct list_head *head = &video->list[index];
drivers/media/platform/nuvoton/npcm-video.c
424
list_add_tail(&list->list, head);
drivers/media/platform/nuvoton/npcm-video.c
431
struct list_head *head = info->head;
drivers/media/platform/nuvoton/npcm-video.c
438
list_add_tail(&list->list, head);
drivers/media/platform/nuvoton/npcm-video.c
454
list_add_tail(&list->list, head);
drivers/media/platform/nuvoton/npcm-video.c
533
info.head = &video->list[index];
drivers/media/platform/nuvoton/npcm-video.c
88
struct list_head *head;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
754
int vsp1_dl_list_add_chain(struct vsp1_dl_list *head,
drivers/media/platform/renesas/vsp1/vsp1_dl.c
757
head->has_chain = true;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
758
list_add_tail(&dl->chain, &head->chain);
drivers/media/platform/renesas/vsp1/vsp1_dl.h
77
int vsp1_dl_list_add_chain(struct vsp1_dl_list *head, struct vsp1_dl_list *dl);
drivers/media/rc/meson-ir-tx.c
239
unsigned int len, unsigned int head)
drivers/media/rc/meson-ir-tx.c
243
ir->buf_head = head;
drivers/media/test-drivers/vidtv/vidtv_channel.c
131
if (head) {
drivers/media/test-drivers/vidtv/vidtv_channel.c
132
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_channel.c
133
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
135
head->next = s302m;
drivers/media/test-drivers/vidtv/vidtv_channel.c
164
struct vidtv_psi_table_eit_event *head = NULL;
drivers/media/test-drivers/vidtv/vidtv_channel.c
184
vidtv_psi_eit_event_destroy(head);
drivers/media/test-drivers/vidtv/vidtv_channel.c
191
if (!head)
drivers/media/test-drivers/vidtv/vidtv_channel.c
192
head = tail;
drivers/media/test-drivers/vidtv/vidtv_channel.c
200
return head;
drivers/media/test-drivers/vidtv/vidtv_channel.c
210
struct vidtv_psi_table_sdt_service *head = NULL;
drivers/media/test-drivers/vidtv/vidtv_channel.c
241
if (!head)
drivers/media/test-drivers/vidtv/vidtv_channel.c
242
head = tail;
drivers/media/test-drivers/vidtv/vidtv_channel.c
250
return head;
drivers/media/test-drivers/vidtv/vidtv_channel.c
255
vidtv_psi_sdt_service_destroy(head);
drivers/media/test-drivers/vidtv/vidtv_channel.c
265
struct vidtv_psi_table_pat_program *head = NULL;
drivers/media/test-drivers/vidtv/vidtv_channel.c
288
vidtv_psi_pat_program_destroy(head);
drivers/media/test-drivers/vidtv/vidtv_channel.c
292
if (!head)
drivers/media/test-drivers/vidtv/vidtv_channel.c
293
head = tail;
drivers/media/test-drivers/vidtv/vidtv_channel.c
303
return head;
drivers/media/test-drivers/vidtv/vidtv_channel.c
316
struct vidtv_psi_table_pmt_stream *head = NULL;
drivers/media/test-drivers/vidtv/vidtv_channel.c
345
if (!head)
drivers/media/test-drivers/vidtv/vidtv_channel.c
346
head = tail;
drivers/media/test-drivers/vidtv/vidtv_channel.c
355
vidtv_psi_pmt_stream_assign(curr_section, head);
drivers/media/test-drivers/vidtv/vidtv_channel.c
54
*vidtv_channel_s302m_init(struct vidtv_channel *head, u16 transport_stream_id)
drivers/media/test-drivers/vidtv/vidtv_channel.h
78
*vidtv_channel_s302m_init(struct vidtv_channel *head, u16 transport_stream_id);
drivers/media/test-drivers/vidtv/vidtv_psi.c
1060
vidtv_psi_pmt_stream_init(struct vidtv_psi_table_pmt_stream *head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
1085
if (head) {
drivers/media/test-drivers/vidtv/vidtv_psi.c
1086
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_psi.c
1087
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1089
head->next = stream;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1434
*vidtv_psi_sdt_service_init(struct vidtv_psi_table_sdt_service *head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
1458
if (head) {
drivers/media/test-drivers/vidtv/vidtv_psi.c
1459
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_psi.c
1460
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1462
head->next = service;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1974
*vidtv_psi_eit_event_init(struct vidtv_psi_table_eit_event *head, u16 event_id)
drivers/media/test-drivers/vidtv/vidtv_psi.c
2024
if (head) {
drivers/media/test-drivers/vidtv/vidtv_psi.c
2025
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_psi.c
2026
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
2028
head->next = e;
drivers/media/test-drivers/vidtv/vidtv_psi.c
269
static void vidtv_psi_desc_chain(struct vidtv_psi_desc *head, struct vidtv_psi_desc *desc)
drivers/media/test-drivers/vidtv/vidtv_psi.c
271
if (head) {
drivers/media/test-drivers/vidtv/vidtv_psi.c
272
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_psi.c
273
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
275
head->next = desc;
drivers/media/test-drivers/vidtv/vidtv_psi.c
279
struct vidtv_psi_desc_service *vidtv_psi_service_desc_init(struct vidtv_psi_desc *head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
318
vidtv_psi_desc_chain(head, (struct vidtv_psi_desc *)desc);
drivers/media/test-drivers/vidtv/vidtv_psi.c
330
*vidtv_psi_registration_desc_init(struct vidtv_psi_desc *head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
353
vidtv_psi_desc_chain(head, (struct vidtv_psi_desc *)desc);
drivers/media/test-drivers/vidtv/vidtv_psi.c
358
*vidtv_psi_network_name_desc_init(struct vidtv_psi_desc *head, char *network_name)
drivers/media/test-drivers/vidtv/vidtv_psi.c
379
vidtv_psi_desc_chain(head, (struct vidtv_psi_desc *)desc);
drivers/media/test-drivers/vidtv/vidtv_psi.c
384
*vidtv_psi_service_list_desc_init(struct vidtv_psi_desc *head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
429
vidtv_psi_desc_chain(head, (struct vidtv_psi_desc *)desc);
drivers/media/test-drivers/vidtv/vidtv_psi.c
434
*vidtv_psi_short_event_desc_init(struct vidtv_psi_desc *head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
478
vidtv_psi_desc_chain(head, (struct vidtv_psi_desc *)desc);
drivers/media/test-drivers/vidtv/vidtv_psi.c
497
struct vidtv_psi_desc *head = NULL;
drivers/media/test-drivers/vidtv/vidtv_psi.c
506
vidtv_psi_service_desc_init(head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
515
vidtv_psi_network_name_desc_init(head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
522
vidtv_psi_service_list_desc_init(head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
529
vidtv_psi_short_event_desc_init(head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
546
if (!head)
drivers/media/test-drivers/vidtv/vidtv_psi.c
547
head = curr;
drivers/media/test-drivers/vidtv/vidtv_psi.c
555
return head;
drivers/media/test-drivers/vidtv/vidtv_psi.c
891
vidtv_psi_pat_program_init(struct vidtv_psi_table_pat_program *head,
drivers/media/test-drivers/vidtv/vidtv_psi.c
908
if (head) {
drivers/media/test-drivers/vidtv/vidtv_psi.c
909
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_psi.c
910
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
912
head->next = program;
drivers/media/test-drivers/vidtv/vidtv_psi.h
140
*vidtv_psi_short_event_desc_init(struct vidtv_psi_desc *head,
drivers/media/test-drivers/vidtv/vidtv_psi.h
336
struct vidtv_psi_desc_service *vidtv_psi_service_desc_init(struct vidtv_psi_desc *head,
drivers/media/test-drivers/vidtv/vidtv_psi.h
342
*vidtv_psi_registration_desc_init(struct vidtv_psi_desc *head,
drivers/media/test-drivers/vidtv/vidtv_psi.h
348
*vidtv_psi_network_name_desc_init(struct vidtv_psi_desc *head, char *network_name);
drivers/media/test-drivers/vidtv/vidtv_psi.h
351
*vidtv_psi_service_list_desc_init(struct vidtv_psi_desc *head,
drivers/media/test-drivers/vidtv/vidtv_psi.h
355
*vidtv_psi_pat_program_init(struct vidtv_psi_table_pat_program *head,
drivers/media/test-drivers/vidtv/vidtv_psi.h
360
vidtv_psi_pmt_stream_init(struct vidtv_psi_table_pmt_stream *head,
drivers/media/test-drivers/vidtv/vidtv_psi.h
373
vidtv_psi_sdt_service_init(struct vidtv_psi_table_sdt_service *head,
drivers/media/test-drivers/vidtv/vidtv_psi.h
805
*vidtv_psi_eit_event_init(struct vidtv_psi_table_eit_event *head, u16 event_id);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
147
static struct vidtv_access_unit *vidtv_s302m_access_unit_init(struct vidtv_access_unit *head)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
155
if (head) {
drivers/media/test-drivers/vidtv/vidtv_s302m.c
156
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
157
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
159
head->next = au;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
167
struct vidtv_access_unit *head = e->access_units;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
170
while (head) {
drivers/media/test-drivers/vidtv/vidtv_s302m.c
171
tmp = head;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
172
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
492
if (args.head) {
drivers/media/test-drivers/vidtv/vidtv_s302m.c
493
while (args.head->next)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
494
args.head = args.head->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
496
args.head->next = e;
drivers/media/test-drivers/vidtv/vidtv_s302m.h
89
struct vidtv_encoder *head;
drivers/media/usb/go7007/go7007-fw.c
773
const unsigned char head[] = { 0x00, 0x00, 0x01, 0xb0, go->pali,
drivers/media/usb/go7007/go7007-fw.c
779
CODE_GEN(c, buf + 2 + sizeof(head));
drivers/media/usb/go7007/go7007-fw.c
793
memcpy(buf + 2, head, sizeof(head));
drivers/media/usb/go7007/go7007-fw.c
811
i = CODE_LENGTH(c) + sizeof(head) * 8;
drivers/media/usb/uvc/uvc_video.c
515
if (clock->head == clock->last_sof_overflow)
drivers/media/usb/uvc/uvc_video.c
527
clock->count = (clock->head - clock->last_sof_overflow
drivers/media/usb/uvc/uvc_video.c
529
clock->last_sof_overflow = clock->head;
drivers/media/usb/uvc/uvc_video.c
533
clock->samples[clock->head] = *sample;
drivers/media/usb/uvc/uvc_video.c
534
clock->head = (clock->head + 1) % clock->size;
drivers/media/usb/uvc/uvc_video.c
673
clock->head = 0;
drivers/media/usb/uvc/uvc_video.c
820
first = &clock->samples[(clock->head - clock->count + clock->size) % clock->size];
drivers/media/usb/uvc/uvc_video.c
821
last = &clock->samples[(clock->head - 1 + clock->size) % clock->size];
drivers/media/usb/uvc/uvcvideo.h
518
unsigned int head;
drivers/media/v4l2-core/v4l2-fwnode.c
693
list_for_each_entry_safe(link, tmp, &connector->links, head) {
drivers/media/v4l2-core/v4l2-fwnode.c
695
list_del(&link->head);
drivers/media/v4l2-core/v4l2-fwnode.c
800
list_add(&link->head, &connector->links);
drivers/memstick/core/ms_block.c
1756
msb->geometry.heads = chs_table[i].head;
drivers/memstick/core/ms_block.h
268
unsigned char head;
drivers/message/fusion/mptsas.c
1186
struct list_head *head = &hd->target_reset_list;
drivers/message/fusion/mptsas.c
1193
if (list_empty(head))
drivers/message/fusion/mptsas.c
1196
target_reset_list = list_entry(head->next,
drivers/message/fusion/mptsas.c
1223
struct list_head *head = &hd->target_reset_list;
drivers/message/fusion/mptsas.c
1268
if (list_empty(head))
drivers/message/fusion/mptsas.c
1271
target_reset_list = list_entry(head->next,
drivers/mfd/ezx-pcap.c
236
u8 head;
drivers/mfd/ezx-pcap.c
239
head = pcap->adc_head;
drivers/mfd/ezx-pcap.c
240
if (!pcap->adc_queue[head]) {
drivers/mfd/ezx-pcap.c
249
tmp |= pcap->adc_queue[head]->flags | PCAP_ADC_ADEN;
drivers/mfd/ezx-pcap.c
251
if (pcap->adc_queue[head]->bank == PCAP_ADC_BANK_1)
drivers/misc/bcm-vk/bcm_vk_dev.c
520
list_for_each_entry(ctx, &vk->pid_ht[i].head, node) {
drivers/misc/bcm-vk/bcm_vk_msg.c
244
list_add_tail(&ctx->node, &vk->pid_ht[hash_idx].head);
drivers/misc/bcm-vk/bcm_vk_msg.c
317
list_for_each_entry(entry, &vk->pid_ht[hash_idx].head, node) {
drivers/misc/bcm-vk/bcm_vk_msg.c
950
INIT_LIST_HEAD(&vk->pid_ht[i].head);
drivers/misc/bcm-vk/bcm_vk_msg.h
97
struct list_head head;
drivers/misc/hpilo.c
100
c = fifo_q->fifobar[fifo_q->head & fifo_q->imask];
drivers/misc/hpilo.c
105
fifo_q->fifobar[fifo_q->head & fifo_q->imask] =
drivers/misc/hpilo.c
107
fifo_q->head += 1;
drivers/misc/hpilo.c
123
c = fifo_q->fifobar[fifo_q->head & fifo_q->imask];
drivers/misc/hpilo.c
221
fifo_q->head = 0;
drivers/misc/hpilo.h
172
u64 head;
drivers/misc/mei/client.c
316
struct list_head *head)
drivers/misc/mei/client.c
318
list_add_tail(&cb->list, head);
drivers/misc/mei/client.c
394
static void mei_io_list_flush_cl(struct list_head *head,
drivers/misc/mei/client.c
399
list_for_each_entry_safe(cb, next, head, list) {
drivers/misc/mei/client.c
415
static void mei_io_tx_list_free_cl(struct list_head *head,
drivers/misc/mei/client.c
421
list_for_each_entry_safe(cb, next, head, list) {
drivers/misc/mei/client.c
433
static void mei_io_list_free_fp(struct list_head *head, const struct file *fp)
drivers/misc/mei/client.c
437
list_for_each_entry_safe(cb, next, head, list)
drivers/misc/ocxl/pasid.c
13
static void dump_list(struct list_head *head, char *type_str)
drivers/misc/ocxl/pasid.c
18
list_for_each_entry(cur, head, list) {
drivers/misc/ocxl/pasid.c
24
static int range_alloc(struct list_head *head, u32 size, int max_id,
drivers/misc/ocxl/pasid.c
35
pos = head;
drivers/misc/ocxl/pasid.c
37
list_for_each_entry(cur, head, list) {
drivers/misc/ocxl/pasid.c
56
dump_list(head, type_str);
drivers/misc/ocxl/pasid.c
61
static void range_free(struct list_head *head, u32 start, u32 size,
drivers/misc/ocxl/pasid.c
67
list_for_each_entry_safe(cur, tmp, head, list) {
drivers/misc/ocxl/pasid.c
77
dump_list(head, type_str);
drivers/misc/sgi-gru/gru_instructions.h
295
unsigned int head;
drivers/misc/sgi-gru/gru_instructions.h
593
static inline union gru_mesqhead gru_mesq_head(int head, int limit)
drivers/misc/sgi-gru/gru_instructions.h
597
mqh.head = head;
drivers/misc/sgi-gru/grukservices.c
116
union gru_mesqhead head __gru_cacheline_aligned__; /* CL 0 */
drivers/misc/sgi-gru/grukservices.c
558
mq->head = gru_mesq_head(2, qlines / 2 + 1);
drivers/misc/sgi-gru/grukservices.c
639
unsigned int limit, head;
drivers/misc/sgi-gru/grukservices.c
645
head = gru_get_amo_value_head(cb);
drivers/misc/sgi-gru/grukservices.c
666
if (head != limit) {
drivers/misc/sgi-xp/xpc_uv.c
872
xpc_init_fifo_uv(struct xpc_fifo_head_uv *head)
drivers/misc/sgi-xp/xpc_uv.c
874
head->first = NULL;
drivers/misc/sgi-xp/xpc_uv.c
875
head->last = NULL;
drivers/misc/sgi-xp/xpc_uv.c
876
spin_lock_init(&head->lock);
drivers/misc/sgi-xp/xpc_uv.c
877
head->n_entries = 0;
drivers/misc/sgi-xp/xpc_uv.c
881
xpc_get_fifo_entry_uv(struct xpc_fifo_head_uv *head)
drivers/misc/sgi-xp/xpc_uv.c
886
spin_lock_irqsave(&head->lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
887
first = head->first;
drivers/misc/sgi-xp/xpc_uv.c
888
if (head->first != NULL) {
drivers/misc/sgi-xp/xpc_uv.c
889
head->first = first->next;
drivers/misc/sgi-xp/xpc_uv.c
890
if (head->first == NULL)
drivers/misc/sgi-xp/xpc_uv.c
891
head->last = NULL;
drivers/misc/sgi-xp/xpc_uv.c
893
head->n_entries--;
drivers/misc/sgi-xp/xpc_uv.c
894
BUG_ON(head->n_entries < 0);
drivers/misc/sgi-xp/xpc_uv.c
898
spin_unlock_irqrestore(&head->lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
903
xpc_put_fifo_entry_uv(struct xpc_fifo_head_uv *head,
drivers/misc/sgi-xp/xpc_uv.c
909
spin_lock_irqsave(&head->lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
910
if (head->last != NULL)
drivers/misc/sgi-xp/xpc_uv.c
911
head->last->next = last;
drivers/misc/sgi-xp/xpc_uv.c
913
head->first = last;
drivers/misc/sgi-xp/xpc_uv.c
914
head->last = last;
drivers/misc/sgi-xp/xpc_uv.c
915
head->n_entries++;
drivers/misc/sgi-xp/xpc_uv.c
916
spin_unlock_irqrestore(&head->lock, irq_flags);
drivers/misc/sgi-xp/xpc_uv.c
920
xpc_n_of_fifo_entries_uv(struct xpc_fifo_head_uv *head)
drivers/misc/sgi-xp/xpc_uv.c
922
return head->n_entries;
drivers/misc/sgi-xp/xpnet.c
234
"skb->end=0x%p skb->len=%d\n", (void *)skb->head,
drivers/misc/sgi-xp/xpnet.c
244
(void *)skb->head, (void *)skb->data, skb_tail_pointer(skb),
drivers/misc/sgi-xp/xpnet.c
355
(void *)queued_msg->skb->head);
drivers/misc/sgi-xp/xpnet.c
420
"skb->end=0x%p skb->len=%d\n", (void *)skb->head,
drivers/misc/vmw_vmci/vmci_context.c
170
list_add_tail_rcu(&context->list_item, &ctx_list.head);
drivers/misc/vmw_vmci/vmci_context.c
222
list_for_each_entry_rcu(sub_ctx, &ctx_list.head, list_item) {
drivers/misc/vmw_vmci/vmci_context.c
34
struct list_head head;
drivers/misc/vmw_vmci/vmci_context.c
357
list_for_each_entry_rcu(context, &ctx_list.head, list_item) {
drivers/misc/vmw_vmci/vmci_context.c
37
.head = LIST_HEAD_INIT(ctx_list.head),
drivers/misc/vmw_vmci/vmci_context.c
379
list_for_each_entry_rcu(c, &ctx_list.head, list_item) {
drivers/misc/vmw_vmci/vmci_queue_pair.c
1006
list_add(&entry->list_item, &qp_list->head);
drivers/misc/vmw_vmci/vmci_queue_pair.c
1896
if (!list_empty(&qp_list->head)) {
drivers/misc/vmw_vmci/vmci_queue_pair.c
1898
list_first_entry(&qp_list->head, struct qp_entry,
drivers/misc/vmw_vmci/vmci_queue_pair.c
222
struct list_head head;
drivers/misc/vmw_vmci/vmci_queue_pair.c
227
.head = LIST_HEAD_INIT(qp_broker_list.head),
drivers/misc/vmw_vmci/vmci_queue_pair.c
232
.head = LIST_HEAD_INIT(qp_guest_endpoints.head),
drivers/misc/vmw_vmci/vmci_queue_pair.c
2614
u64 head;
drivers/misc/vmw_vmci/vmci_queue_pair.c
2638
head = vmci_q_header_consumer_head(produce_q->q_header);
drivers/misc/vmw_vmci/vmci_queue_pair.c
2639
if (likely(head + read < consume_q_size)) {
drivers/misc/vmw_vmci/vmci_queue_pair.c
2640
result = qp_memcpy_from_queue_iter(to, consume_q, head, read);
drivers/misc/vmw_vmci/vmci_queue_pair.c
2644
const size_t tmp = (size_t) (consume_q_size - head);
drivers/misc/vmw_vmci/vmci_queue_pair.c
2646
result = qp_memcpy_from_queue_iter(to, consume_q, head, tmp);
drivers/misc/vmw_vmci/vmci_queue_pair.c
812
list_for_each_entry(entry, &qp_list->head, list_item) {
drivers/mmc/host/usdhi6rol0.c
359
size_t head = PAGE_SIZE - sg->offset;
drivers/mmc/host/usdhi6rol0.c
360
size_t blk_head = head % data->blksz;
drivers/mmc/host/usdhi6rol0.c
378
if (head < data->blksz)
drivers/mmc/host/vub300.c
1067
vub300->cmnd.head.header_size = 20;
drivers/mmc/host/vub300.c
1068
vub300->cmnd.head.header_type = 0x00;
drivers/mmc/host/vub300.c
1069
vub300->cmnd.head.port_number = 0; /* "0" means port 1 */
drivers/mmc/host/vub300.c
1070
vub300->cmnd.head.command_type = 0x00; /* standard read command */
drivers/mmc/host/vub300.c
1071
vub300->cmnd.head.response_type = response_type;
drivers/mmc/host/vub300.c
1072
vub300->cmnd.head.command_index = cmd->opcode;
drivers/mmc/host/vub300.c
1073
vub300->cmnd.head.arguments[0] = cmd->arg >> 24;
drivers/mmc/host/vub300.c
1074
vub300->cmnd.head.arguments[1] = cmd->arg >> 16;
drivers/mmc/host/vub300.c
1075
vub300->cmnd.head.arguments[2] = cmd->arg >> 8;
drivers/mmc/host/vub300.c
1076
vub300->cmnd.head.arguments[3] = cmd->arg >> 0;
drivers/mmc/host/vub300.c
1079
vub300->cmnd.head.block_count[0] = 0;
drivers/mmc/host/vub300.c
1080
vub300->cmnd.head.block_count[1] = 0;
drivers/mmc/host/vub300.c
1081
vub300->cmnd.head.block_size[0] = (vub300->fbs[fn] >> 8) & 0xFF;
drivers/mmc/host/vub300.c
1082
vub300->cmnd.head.block_size[1] = (vub300->fbs[fn] >> 0) & 0xFF;
drivers/mmc/host/vub300.c
1083
vub300->cmnd.head.command_type = 0x00;
drivers/mmc/host/vub300.c
1084
vub300->cmnd.head.transfer_size[0] = 0;
drivers/mmc/host/vub300.c
1085
vub300->cmnd.head.transfer_size[1] = 0;
drivers/mmc/host/vub300.c
1086
vub300->cmnd.head.transfer_size[2] = 0;
drivers/mmc/host/vub300.c
1087
vub300->cmnd.head.transfer_size[3] = 0;
drivers/mmc/host/vub300.c
1089
vub300->cmnd.head.block_count[0] = 0;
drivers/mmc/host/vub300.c
1090
vub300->cmnd.head.block_count[1] = 0;
drivers/mmc/host/vub300.c
1091
vub300->cmnd.head.block_size[0] = (vub300->fbs[0] >> 8) & 0xFF;
drivers/mmc/host/vub300.c
1092
vub300->cmnd.head.block_size[1] = (vub300->fbs[0] >> 0) & 0xFF;
drivers/mmc/host/vub300.c
1093
vub300->cmnd.head.command_type = 0x00;
drivers/mmc/host/vub300.c
1094
vub300->cmnd.head.transfer_size[0] = 0;
drivers/mmc/host/vub300.c
1095
vub300->cmnd.head.transfer_size[1] = 0;
drivers/mmc/host/vub300.c
1096
vub300->cmnd.head.transfer_size[2] = 0;
drivers/mmc/host/vub300.c
1097
vub300->cmnd.head.transfer_size[3] = 0;
drivers/mmc/host/vub300.c
1100
if (0x08 & vub300->cmnd.head.arguments[0]) { /* BLOCK MODE */
drivers/mmc/host/vub300.c
1101
vub300->cmnd.head.block_count[0] =
drivers/mmc/host/vub300.c
1103
vub300->cmnd.head.block_count[1] =
drivers/mmc/host/vub300.c
1105
vub300->cmnd.head.block_size[0] =
drivers/mmc/host/vub300.c
1107
vub300->cmnd.head.block_size[1] =
drivers/mmc/host/vub300.c
1110
vub300->cmnd.head.block_count[0] = 0;
drivers/mmc/host/vub300.c
1111
vub300->cmnd.head.block_count[1] = 0;
drivers/mmc/host/vub300.c
1112
vub300->cmnd.head.block_size[0] =
drivers/mmc/host/vub300.c
1114
vub300->cmnd.head.block_size[1] =
drivers/mmc/host/vub300.c
1117
vub300->cmnd.head.command_type =
drivers/mmc/host/vub300.c
1119
vub300->cmnd.head.transfer_size[0] =
drivers/mmc/host/vub300.c
1121
vub300->cmnd.head.transfer_size[1] =
drivers/mmc/host/vub300.c
1123
vub300->cmnd.head.transfer_size[2] =
drivers/mmc/host/vub300.c
1125
vub300->cmnd.head.transfer_size[3] =
drivers/mmc/host/vub300.c
1128
vub300->cmnd.head.block_count[0] = 0;
drivers/mmc/host/vub300.c
1129
vub300->cmnd.head.block_count[1] = 0;
drivers/mmc/host/vub300.c
1132
vub300->cmnd.head.block_count[0] = (data->blocks >> 8) & 0xFF;
drivers/mmc/host/vub300.c
1133
vub300->cmnd.head.block_count[1] = (data->blocks >> 0) & 0xFF;
drivers/mmc/host/vub300.c
1134
vub300->cmnd.head.block_size[0] = (data->blksz >> 8) & 0xFF;
drivers/mmc/host/vub300.c
1135
vub300->cmnd.head.block_size[1] = (data->blksz >> 0) & 0xFF;
drivers/mmc/host/vub300.c
1136
vub300->cmnd.head.command_type =
drivers/mmc/host/vub300.c
1138
vub300->cmnd.head.transfer_size[0] =
drivers/mmc/host/vub300.c
1140
vub300->cmnd.head.transfer_size[1] =
drivers/mmc/host/vub300.c
1142
vub300->cmnd.head.transfer_size[2] =
drivers/mmc/host/vub300.c
1144
vub300->cmnd.head.transfer_size[3] =
drivers/mmc/host/vub300.c
1147
vub300->cmnd.head.block_count[0] = 0;
drivers/mmc/host/vub300.c
1148
vub300->cmnd.head.block_count[1] = 0;
drivers/mmc/host/vub300.c
1151
if (vub300->cmnd.head.block_size[0] || vub300->cmnd.head.block_size[1]) {
drivers/mmc/host/vub300.c
1152
u16 block_size = vub300->cmnd.head.block_size[1] |
drivers/mmc/host/vub300.c
1153
(vub300->cmnd.head.block_size[0] << 8);
drivers/mmc/host/vub300.c
1156
vub300->cmnd.head.block_boundary[0] =
drivers/mmc/host/vub300.c
1158
vub300->cmnd.head.block_boundary[1] =
drivers/mmc/host/vub300.c
1161
vub300->cmnd.head.block_boundary[0] = 0;
drivers/mmc/host/vub300.c
1162
vub300->cmnd.head.block_boundary[1] = 0;
drivers/mmc/host/vub300.c
164
struct sd_command_header head;
drivers/mmc/host/vub300.c
2178
vub300->cmnd.head.block_size[0] = 0x00;
drivers/mmc/host/vub300.c
2179
vub300->cmnd.head.block_size[1] = 0x00;
drivers/mtd/sm_ftl.c
644
ftl->heads = chs_table[i].head;
drivers/mtd/sm_ftl.h
64
unsigned char head;
drivers/mtd/ssfdc.c
56
unsigned char head;
drivers/mtd/ssfdc.c
73
static int get_chs(unsigned long size, unsigned short *cyl, unsigned char *head,
drivers/mtd/ssfdc.c
86
if (head)
drivers/mtd/ssfdc.c
87
*head = chs_table[k].head;
drivers/net/amt.c
3288
static void amt_dellink(struct net_device *dev, struct list_head *head)
drivers/net/amt.c
3292
unregister_netdevice_queue(dev, head);
drivers/net/bareudp.c
158
nh = skb_network_header(skb) - skb->head;
drivers/net/bareudp.c
169
oiph = skb->head + nh;
drivers/net/bareudp.c
694
static void bareudp_dellink(struct net_device *dev, struct list_head *head)
drivers/net/bareudp.c
699
unregister_netdevice_queue(dev, head);
drivers/net/bonding/bond_alb.c
104
index = SLAVE_TLB_INFO(slave).head;
drivers/net/bonding/bond_alb.c
206
u32 next_index = slave_info->head;
drivers/net/bonding/bond_alb.c
215
slave_info->head = hash_index;
drivers/net/bonding/bond_alb.c
90
SLAVE_TLB_INFO(slave).head = TLB_NULL_INDEX;
drivers/net/caif/caif_serial.c
208
while ((skb = skb_peek(&ser->head)) != NULL) {
drivers/net/caif/caif_serial.c
237
struct sk_buff *tmp = skb_dequeue(&ser->head);
drivers/net/caif/caif_serial.c
243
if (ser->head.qlen <= SEND_QUEUE_LOW &&
drivers/net/caif/caif_serial.c
261
if (ser->head.qlen > SEND_QUEUE_HIGH &&
drivers/net/caif/caif_serial.c
267
skb_queue_tail(&ser->head, skb);
drivers/net/caif/caif_serial.c
398
skb_queue_head_init(&serdev->head);
drivers/net/caif/caif_serial.c
69
struct sk_buff_head head;
drivers/net/caif/caif_virtio.c
267
if (cfv->ctx.head != USHRT_MAX) {
drivers/net/caif/caif_virtio.c
269
cfv->ctx.head,
drivers/net/caif/caif_virtio.c
271
cfv->ctx.head = USHRT_MAX;
drivers/net/caif/caif_virtio.c
278
&cfv->ctx.head,
drivers/net/caif/caif_virtio.c
50
unsigned short head;
drivers/net/caif/caif_virtio.c
716
cfv->ctx.head = USHRT_MAX;
drivers/net/can/c_can/c_can.h
181
unsigned int head;
drivers/net/can/c_can/c_can.h
230
return ring->head & (ring->obj_num - 1);
drivers/net/can/c_can/c_can.h
241
u8 head = c_can_get_tx_head(ring);
drivers/net/can/c_can/c_can.h
245
return ring->obj_num - (ring->head - ring->tail);
drivers/net/can/c_can/c_can.h
250
if (head < tail)
drivers/net/can/c_can/c_can.h
253
return ring->obj_num - head;
drivers/net/can/c_can/c_can_main.c
1247
priv->tx.head = 0;
drivers/net/can/c_can/c_can_main.c
443
tx_ring->head, tx_ring->tail,
drivers/net/can/c_can/c_can_main.c
444
tx_ring->head - tx_ring->tail);
drivers/net/can/c_can/c_can_main.c
467
tx_ring->head++;
drivers/net/can/c_can/c_can_main.c
624
tx_ring->head = 0;
drivers/net/can/c_can/c_can_main.c
764
u8 head = c_can_get_tx_head(tx_ring);
drivers/net/can/c_can/c_can_main.c
767
for (idx = tail; idx < head; idx++) {
drivers/net/can/dev/netlink.c
1086
static void can_dellink(struct net_device *dev, struct list_head *head)
drivers/net/can/dev/rx-offload.c
101
__skb_queue_head(head, new);
drivers/net/can/dev/rx-offload.c
103
__skb_queue_after(head, insert, new);
drivers/net/can/dev/rx-offload.c
77
__skb_queue_add_sort(struct sk_buff_head *head, struct sk_buff *new,
drivers/net/can/dev/rx-offload.c
82
skb_queue_reverse_walk(head, pos) {
drivers/net/can/dev/rx-offload.c
93
skb_queue_len(head));
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
125
.val = tef->head,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
158
.val = rx->head,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
201
.val = tx->head,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
179
tx_ring->head = 0;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
210
rx_ring->head = 0;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
78
tef_ring->head = 0;
drivers/net/can/spi/mcp251xfd/mcp251xfd-rx.c
239
ring->head += len;
drivers/net/can/spi/mcp251xfd/mcp251xfd-tef.c
260
ring->head += len;
drivers/net/can/spi/mcp251xfd/mcp251xfd-tx.c
143
tx_ring->head--;
drivers/net/can/spi/mcp251xfd/mcp251xfd-tx.c
187
tx_ring->head, tx_ring->tail,
drivers/net/can/spi/mcp251xfd/mcp251xfd-tx.c
188
tx_ring->head - tx_ring->tail);
drivers/net/can/spi/mcp251xfd/mcp251xfd-tx.c
221
mcp251xfd_tx_obj_from_skb(priv, tx_obj, skb, tx_ring->head);
drivers/net/can/spi/mcp251xfd/mcp251xfd-tx.c
225
tx_ring->head++;
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
529
unsigned int head;
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
546
unsigned int head;
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
561
unsigned int head;
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
887
return priv->tef->head & (priv->tx->obj_num - 1);
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
902
return ring->head & (ring->obj_num - 1);
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
912
return ring->obj_num - (ring->head - ring->tail);
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
931
return ring->head & (ring->obj_num - 1);
drivers/net/can/vxcan.c
282
static void vxcan_dellink(struct net_device *dev, struct list_head *head)
drivers/net/can/vxcan.c
295
unregister_netdevice_queue(dev, head);
drivers/net/can/vxcan.c
300
unregister_netdevice_queue(peer, head);
drivers/net/eql.c
147
struct list_head *this, *tmp, *head;
drivers/net/eql.c
150
head = &eql->queue.all_slaves;
drivers/net/eql.c
151
list_for_each_safe(this, tmp, head) {
drivers/net/eql.c
234
struct list_head *head, *tmp, *this;
drivers/net/eql.c
238
head = &queue->all_slaves;
drivers/net/eql.c
239
list_for_each_safe(this, tmp, head) {
drivers/net/eql.c
305
struct list_head *this, *tmp, *head;
drivers/net/eql.c
311
head = &queue->all_slaves;
drivers/net/eql.c
312
list_for_each_safe(this, tmp, head) {
drivers/net/eql.c
370
struct list_head *this, *head;
drivers/net/eql.c
372
head = &queue->all_slaves;
drivers/net/eql.c
373
list_for_each(this, head) {
drivers/net/ethernet/3com/3c509.c
166
int head, size;
drivers/net/ethernet/3com/3c515.c
1375
skb->head, temp);
drivers/net/ethernet/actions/owl-emac.c
142
return CIRC_SPACE(ring->head, ring->tail, ring->size);
drivers/net/ethernet/actions/owl-emac.c
153
ring->head = owl_emac_ring_get_next(ring, ring->head);
drivers/net/ethernet/actions/owl-emac.c
212
ring->head = 0;
drivers/net/ethernet/actions/owl-emac.c
237
ring->head = 0;
drivers/net/ethernet/actions/owl-emac.c
515
tx_head = ring->head;
drivers/net/ethernet/actions/owl-emac.c
586
tx_head = ring->head;
drivers/net/ethernet/actions/owl-emac.c
704
while (ring->tail != ring->head) {
drivers/net/ethernet/actions/owl-emac.c
724
while ((tx_next = owl_emac_ring_get_next(ring, tx_next)) != ring->head) {
drivers/net/ethernet/actions/owl-emac.c
850
ring->skbs[ring->head] = new_skb;
drivers/net/ethernet/actions/owl-emac.c
851
ring->skbs_dma[ring->head] = new_dma;
drivers/net/ethernet/actions/owl-emac.h
250
unsigned int head;
drivers/net/ethernet/airoha/airoha_eth.c
547
struct airoha_queue_entry *e = &q->entry[q->head];
drivers/net/ethernet/airoha/airoha_eth.c
548
struct airoha_qdma_desc *desc = &q->desc[q->head];
drivers/net/ethernet/airoha/airoha_eth.c
558
q->head = (q->head + 1) % q->ndesc;
drivers/net/ethernet/airoha/airoha_eth.c
569
val = FIELD_PREP(QDMA_DESC_NEXT_ID_MASK, q->head);
drivers/net/ethernet/airoha/airoha_eth.c
578
FIELD_PREP(RX_RING_CPU_IDX_MASK, q->head));
drivers/net/ethernet/airoha/airoha_eth.c
786
FIELD_PREP(RX_RING_DMA_IDX_MASK, q->head));
drivers/net/ethernet/airoha/airoha_eth.c
821
q->head = q->tail;
drivers/net/ethernet/airoha/airoha_eth.c
853
u32 status, head;
drivers/net/ethernet/airoha/airoha_eth.c
861
head = FIELD_GET(IRQ_HEAD_IDX_MASK, status);
drivers/net/ethernet/airoha/airoha_eth.c
862
head = head % irq_q->size;
drivers/net/ethernet/airoha/airoha_eth.c
866
u32 qid, val = irq_q->q[head];
drivers/net/ethernet/airoha/airoha_eth.c
876
irq_q->q[head] = 0xff; /* mark as done */
drivers/net/ethernet/airoha/airoha_eth.c
877
head = (head + 1) % irq_q->size;
drivers/net/ethernet/airoha/airoha_eth.h
189
u16 head;
drivers/net/ethernet/airoha/airoha_ppe.c
745
struct hlist_head *head = &e->l2_flows;
drivers/net/ethernet/airoha/airoha_ppe.c
752
hlist_for_each_entry_safe(e, n, head, l2_subflow_node)
drivers/net/ethernet/amazon/ena/ena_com.c
106
sq->head = 0;
drivers/net/ethernet/amazon/ena/ena_com.c
128
cq->head = 0;
drivers/net/ethernet/amazon/ena/ena_com.c
1480
WARN(ena_dev->aenq.head != depth, "Invalid AENQ state\n");
drivers/net/ethernet/amazon/ena/ena_com.c
150
aenq->head = aenq->q_depth;
drivers/net/ethernet/amazon/ena/ena_com.c
2303
masked_head = aenq->head & (aenq->q_depth - 1);
drivers/net/ethernet/amazon/ena/ena_com.c
2338
aenq->head += processed;
drivers/net/ethernet/amazon/ena/ena_com.c
2347
writel_relaxed((u32)aenq->head, ena_dev->reg_bar + ENA_REGS_AENQ_HEAD_DB_OFF);
drivers/net/ethernet/amazon/ena/ena_com.c
424
io_cq->head = 0;
drivers/net/ethernet/amazon/ena/ena_com.c
463
head_masked = admin_queue->cq.head & (admin_queue->q_depth - 1);
drivers/net/ethernet/amazon/ena/ena_com.c
487
admin_queue->cq.head += comp_num;
drivers/net/ethernet/amazon/ena/ena_com.c
489
admin_queue->sq.head += comp_num;
drivers/net/ethernet/amazon/ena/ena_com.h
135
u16 head;
drivers/net/ethernet/amazon/ena/ena_com.h
189
u16 head;
drivers/net/ethernet/amazon/ena/ena_com.h
199
u16 head;
drivers/net/ethernet/amazon/ena/ena_com.h
251
u16 head;
drivers/net/ethernet/amazon/ena/ena_eth_com.c
15
head_masked = io_cq->head & (io_cq->q_depth - 1);
drivers/net/ethernet/amazon/ena/ena_eth_com.c
266
head_masked = io_cq->head & (io_cq->q_depth - 1);
drivers/net/ethernet/amazon/ena/ena_eth_com.h
193
io_cq->head++;
drivers/net/ethernet/amazon/ena/ena_eth_com.h
196
if (unlikely((io_cq->head & (io_cq->q_depth - 1)) == 0))
drivers/net/ethernet/amazon/ena/ena_eth_com.h
207
masked_head = io_cq->head & (io_cq->q_depth - 1);
drivers/net/ethernet/amd/atarilance.c
1007
head->flag |= RMD1_OWN_CHIP;
drivers/net/ethernet/amd/atarilance.c
1014
u_char *data = PKTBUF_ADDR(head);
drivers/net/ethernet/amd/atarilance.c
1024
lp->memcpy_f( skb->data, PKTBUF_ADDR(head), pkt_len );
drivers/net/ethernet/amd/atarilance.c
1032
head->flag |= RMD1_OWN_CHIP;
drivers/net/ethernet/amd/atarilance.c
242
#define PKTBUF_ADDR(head) (((unsigned char *)(MEM)) + (head)->base)
drivers/net/ethernet/amd/atarilance.c
779
struct lance_tx_head *head;
drivers/net/ethernet/amd/atarilance.c
816
head = &(MEM->tx_head[entry]);
drivers/net/ethernet/amd/atarilance.c
823
head->length = -len;
drivers/net/ethernet/amd/atarilance.c
824
head->misc = 0;
drivers/net/ethernet/amd/atarilance.c
825
lp->memcpy_f( PKTBUF_ADDR(head), (void *)skb->data, skb->len );
drivers/net/ethernet/amd/atarilance.c
826
head->flag = TMD1_OWN_CHIP | TMD1_ENP | TMD1_STP;
drivers/net/ethernet/amd/atarilance.c
973
struct lance_rx_head *head = &(MEM->rx_head[entry]);
drivers/net/ethernet/amd/atarilance.c
974
int status = head->flag;
drivers/net/ethernet/amd/atarilance.c
987
head->flag &= (RMD1_ENP|RMD1_STP);
drivers/net/ethernet/amd/atarilance.c
990
short pkt_len = head->msg_length & 0xfff;
drivers/net/ethernet/amd/sun3lance.c
107
#define PKTBUF_ADDR(head) (void *)((unsigned long)(MEM) | (head)->base)
drivers/net/ethernet/amd/sun3lance.c
515
struct lance_tx_head *head;
drivers/net/ethernet/amd/sun3lance.c
609
head = &(MEM->tx_head[entry]);
drivers/net/ethernet/amd/sun3lance.c
620
head->length = (-len) | 0xf000;
drivers/net/ethernet/amd/sun3lance.c
621
head->misc = 0;
drivers/net/ethernet/amd/sun3lance.c
623
skb_copy_from_linear_data(skb, PKTBUF_ADDR(head), skb->len);
drivers/net/ethernet/amd/sun3lance.c
625
memset(PKTBUF_ADDR(head) + skb->len, 0, len-skb->len);
drivers/net/ethernet/amd/sun3lance.c
627
head->flag = TMD1_OWN_CHIP | TMD1_ENP | TMD1_STP;
drivers/net/ethernet/amd/sun3lance.c
687
struct lance_tx_head *head = &(MEM->tx_head[old_tx]);
drivers/net/ethernet/amd/sun3lance.c
691
if (head->flag & TMD1_OWN_CHIP)
drivers/net/ethernet/amd/sun3lance.c
694
if (head->flag & TMD1_ERR) {
drivers/net/ethernet/amd/sun3lance.c
695
int status = head->misc;
drivers/net/ethernet/amd/sun3lance.c
710
} else if(head->flag & (TMD1_ENP | TMD1_STP)) {
drivers/net/ethernet/amd/sun3lance.c
712
head->flag &= ~(TMD1_ENP | TMD1_STP);
drivers/net/ethernet/amd/sun3lance.c
713
if(head->flag & (TMD1_ONE | TMD1_MORE))
drivers/net/ethernet/amd/sun3lance.c
773
struct lance_rx_head *head = &(MEM->rx_head[entry]);
drivers/net/ethernet/amd/sun3lance.c
774
int status = head->flag;
drivers/net/ethernet/amd/sun3lance.c
787
head->flag &= (RMD1_ENP|RMD1_STP);
drivers/net/ethernet/amd/sun3lance.c
791
short pkt_len = (head->msg_length & 0xfff) - 4;
drivers/net/ethernet/amd/sun3lance.c
802
head->msg_length = 0;
drivers/net/ethernet/amd/sun3lance.c
803
head->flag |= RMD1_OWN_CHIP;
drivers/net/ethernet/amd/sun3lance.c
810
u_char *data = PKTBUF_ADDR(head);
drivers/net/ethernet/amd/sun3lance.c
824
u_char *data = PKTBUF_ADDR(head);
drivers/net/ethernet/amd/sun3lance.c
832
PKTBUF_ADDR(head),
drivers/net/ethernet/amd/sun3lance.c
843
head->msg_length = 0;
drivers/net/ethernet/amd/sun3lance.c
844
head->flag = RMD1_OWN_CHIP;
drivers/net/ethernet/apm/xgene-v2/main.c
244
u8 head;
drivers/net/ethernet/apm/xgene-v2/main.c
247
head = tx_ring->head;
drivers/net/ethernet/apm/xgene-v2/main.c
254
raw_desc = &tx_ring->raw_desc[head];
drivers/net/ethernet/apm/xgene-v2/main.c
261
skb = tx_ring->pkt_info[head].skb;
drivers/net/ethernet/apm/xgene-v2/main.c
262
dma_addr = tx_ring->pkt_info[head].dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
263
pkt_buf = tx_ring->pkt_info[head].pkt_buf;
drivers/net/ethernet/apm/xgene-v2/main.c
274
head = (head + 1) & (XGENE_ENET_NUM_DESC - 1);
drivers/net/ethernet/apm/xgene-v2/main.c
280
tx_ring->head = head;
drivers/net/ethernet/apm/xgene-v2/main.c
292
u8 head, rx_error;
drivers/net/ethernet/apm/xgene-v2/main.c
298
head = rx_ring->head;
drivers/net/ethernet/apm/xgene-v2/main.c
305
raw_desc = &rx_ring->raw_desc[head];
drivers/net/ethernet/apm/xgene-v2/main.c
312
skb = rx_ring->pkt_info[head].skb;
drivers/net/ethernet/apm/xgene-v2/main.c
313
rx_ring->pkt_info[head].skb = NULL;
drivers/net/ethernet/apm/xgene-v2/main.c
314
dma_addr = rx_ring->pkt_info[head].dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
340
head = (head + 1) & (XGENE_ENET_NUM_DESC - 1);
drivers/net/ethernet/apm/xgene-v2/main.c
344
rx_ring->head = head;
drivers/net/ethernet/apm/xgene-v2/ring.c
42
ring->head = 0;
drivers/net/ethernet/apm/xgene-v2/ring.c
54
ring->head = 0;
drivers/net/ethernet/apm/xgene-v2/ring.h
76
u8 head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
616
u16 slots, head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
626
head = buf_pool->head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
636
page = buf_pool->frag_page[head];
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
639
buf_pool->frag_page[head] = NULL;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
640
head = (head + 1) & slots;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
642
buf_pool->head = head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
680
u16 slots, head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
730
head = page_pool->head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
741
page = page_pool->frag_page[head];
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
747
page_pool->frag_page[head] = NULL;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
748
head = (head + 1) & slots;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
751
page_pool->head = head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
789
u16 head = ring->head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
795
raw_desc = &ring->raw_desc[head];
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
805
head = (head + 1) & slots;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
806
exp_desc = &ring->raw_desc[head];
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
809
head = (head - 1) & slots;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
826
head = (head + 1) & slots;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
839
ring->head = head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.h
95
u16 head;
drivers/net/ethernet/apple/macmace.c
680
int left, head;
drivers/net/ethernet/apple/macmace.c
701
head = N_RX_RING - left;
drivers/net/ethernet/apple/macmace.c
705
while (mp->rx_tail < head) {
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
121
unsigned int next_head = (ring->head + 1) % ring->size;
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
126
ring->buff[ring->head] = skb_get(skb);
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
127
ring->head = next_head;
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
148
if (ring->tail == ring->head)
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
175
len = (ring->head >= ring->tail) ?
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
176
ring->head - ring->tail :
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
177
ring->size - ring->tail + ring->head;
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
194
ring->head = 0;
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
42
unsigned int head;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1715
struct hlist_head *head;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1721
head = &bn->l2_fltr_hash_tbl[idx];
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1722
hlist_add_head_rcu(&fltr->base.hash, head);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1731
struct hlist_head *head;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1733
head = &bn->l2_fltr_hash_tbl[idx];
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1734
hlist_for_each_entry_rcu(fltr, head, base.hash) {
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
1234
if (list_empty(&o->head)) {
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
1241
*ppos = list_first_entry(&o->head,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
1250
if (list_is_last(&pos->link, &o->head))
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
1825
list_add(&reg_elem->link, &cam_obj->head);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
2072
list_for_each_entry(pos, &o->head, link) {
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
2120
INIT_LIST_HEAD(&o->head);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
622
list_for_each_entry(pos, &o->head, link) {
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
653
list_for_each_entry(pos, &o->head, link)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
669
list_for_each_entry(pos, &o->head, link)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
685
list_for_each_entry(pos, &o->head, link)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
706
list_for_each_entry(pos, &o->head, link)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
723
list_for_each_entry(pos, &o->head, link)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
740
list_for_each_entry(pos, &o->head, link)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h
298
struct list_head head;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15560
struct hlist_head *head;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15573
head = &bp->ntp_fltr_hash_tbl[idx];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15574
hlist_add_head_rcu(&fltr->base.hash, head);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15625
struct hlist_head *head;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15627
head = &bp->ntp_fltr_hash_tbl[idx];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15628
hlist_for_each_entry_rcu(f, head, base.hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15745
struct hlist_head *head;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15750
head = &bp->ntp_fltr_hash_tbl[i];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15751
hlist_for_each_entry_safe(fltr, tmp, head, base.hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3609
void *buf, size_t offset, size_t head,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3615
head_page = head / rmem->page_size;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3616
source_offset = head % rmem->page_size;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3617
total_len = (tail - head) & MAX_CTX_BYTES_MASK;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5387
struct hlist_head *head;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5391
head = &bp->ntp_fltr_hash_tbl[i];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5392
hlist_for_each_entry_safe(fltr, tmp, head, base.hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5432
struct hlist_head *head;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5436
head = &bp->l2_fltr_hash_tbl[i];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5437
hlist_for_each_entry_safe(fltr, tmp, head, base.hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5939
struct hlist_head *head = &bp->l2_fltr_hash_tbl[idx];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
5942
hlist_for_each_entry_rcu(fltr, head, base.hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6056
struct hlist_head *head;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6071
head = &bp->l2_fltr_hash_tbl[idx];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6072
hlist_add_head_rcu(&fltr->base.hash, head);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9093
void *buf, size_t offset, size_t head,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9105
int i = head / (page_size * MAX_CTX_PAGES);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9111
len = __bnxt_copy_ring(bp, rmem, buf, offset, head, tail);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9112
head += len;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9115
if (head >= nr_pages * page_size)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9116
head = 0;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9117
} while (head != tail);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9298
size_t offset, size_t head, size_t tail)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9310
len = bnxt_copy_ctx_pg_tbls(bp, &ctx_pg[i], buf, offset, head,
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1042
struct hlist_head *head;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1045
head = &tbl[i];
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1046
hlist_for_each_entry_rcu(fltr, head, hash) {
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1065
struct hlist_head *head;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1068
head = &tbl[i];
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1069
hlist_for_each_entry_rcu(fltr, head, hash) {
drivers/net/ethernet/brocade/bna/bna.h
270
struct bna_mac *bna_cam_mod_mac_get(struct list_head *head);
drivers/net/ethernet/brocade/bna/bna_enet.c
2050
bna_cam_mod_mac_get(struct list_head *head)
drivers/net/ethernet/brocade/bna/bna_enet.c
2054
mac = list_first_entry_or_null(head, struct bna_mac, qe);
drivers/net/ethernet/cadence/macb_main.c
1270
unsigned int head;
drivers/net/ethernet/cadence/macb_main.c
1275
head = queue->tx_head;
drivers/net/ethernet/cadence/macb_main.c
1276
for (tail = queue->tx_tail; tail != head && packets < budget; tail++) {
drivers/net/ethernet/cadence/macb_main.c
2399
queue_index, skb->len, skb->head, skb->data,
drivers/net/ethernet/cadence/macb_main.c
3491
unsigned int tail, head;
drivers/net/ethernet/cadence/macb_main.c
3498
head = macb_tx_ring_wrap(bp, bp->queues[0].tx_head);
drivers/net/ethernet/cadence/macb_main.c
3510
regs_buff[9] = head;
drivers/net/ethernet/cadence/macb_main.c
3512
regs_buff[11] = macb_tx_dma(&bp->queues[0], head);
drivers/net/ethernet/cadence/macb_main.c
675
unsigned int head, tail, count, ring_size, desc_size;
drivers/net/ethernet/cadence/macb_main.c
689
head = queue->tx_head;
drivers/net/ethernet/cadence/macb_main.c
692
count = CIRC_CNT(head, tail, ring_size);
drivers/net/ethernet/cavium/liquidio/octeon_iq.h
316
struct list_head head;
drivers/net/ethernet/cavium/liquidio/request_manager.c
405
[OCTEON_ORDERED_SC_LIST].head);
drivers/net/ethernet/cavium/liquidio/request_manager.c
744
INIT_LIST_HEAD(&oct->sc_buf_pool.head);
drivers/net/ethernet/cavium/liquidio/request_manager.c
761
list_add_tail(&sc->node, &oct->sc_buf_pool.head);
drivers/net/ethernet/cavium/liquidio/request_manager.c
785
list_for_each_safe(tmp, tmp2, &done_sc_list->head) {
drivers/net/ethernet/cavium/liquidio/request_manager.c
794
list_add_tail(&sc->node, &zombie_sc_list->head);
drivers/net/ethernet/cavium/liquidio/request_manager.c
820
list_for_each_safe(tmp, tmp2, &zombie_sc_list->head) {
drivers/net/ethernet/cavium/liquidio/request_manager.c
842
list_for_each_safe(tmp, tmp2, &oct->sc_buf_pool.head) {
drivers/net/ethernet/cavium/liquidio/request_manager.c
850
INIT_LIST_HEAD(&oct->sc_buf_pool.head);
drivers/net/ethernet/cavium/liquidio/request_manager.c
877
if (list_empty(&oct->sc_buf_pool.head)) {
drivers/net/ethernet/cavium/liquidio/request_manager.c
882
list_for_each(tmp, &oct->sc_buf_pool.head)
drivers/net/ethernet/cavium/liquidio/request_manager.c
935
list_add_tail(&sc->node, &oct->sc_buf_pool.head);
drivers/net/ethernet/cavium/liquidio/response_manager.c
151
[OCTEON_DONE_SC_LIST].head);
drivers/net/ethernet/cavium/liquidio/response_manager.c
192
head);
drivers/net/ethernet/cavium/liquidio/response_manager.c
36
INIT_LIST_HEAD(&oct->response_list[i].head);
drivers/net/ethernet/cavium/liquidio/response_manager.c
82
if (list_empty(&ordered_sc_list->head)) {
drivers/net/ethernet/cavium/liquidio/response_manager.c
87
sc = list_first_entry(&ordered_sc_list->head,
drivers/net/ethernet/cavium/liquidio/response_manager.h
42
struct list_head head;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1148
sq->head += desc_cnt;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1149
sq->head &= (sq->dmem.q_len - 1);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1182
u64 head;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1187
head = nicvf_queue_reg_read(nic, NIC_QSET_SQ_0_7_HEAD, qidx) >> 4;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1188
while (sq->head != head) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1189
hdr = (struct sq_hdr_subdesc *)GET_SQ_DESC(sq, sq->head);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1194
skb = (struct sk_buff *)sq->skbuff[sq->head];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
272
rbdr->head = 0;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
320
int head, tail;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
332
head = rbdr->head;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
336
while (head != tail) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
337
desc = GET_RBDR_DESC(rbdr, head);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
344
head++;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
345
head &= (rbdr->dmem.q_len - 1);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
360
head = 0;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
361
while (head < rbdr->pgcnt) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
362
pgcache = &rbdr->pgcache[head];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
370
head++;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
518
sq->head = 0;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
592
while (sq->head != sq->tail) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
593
skb = (struct sk_buff *)sq->skbuff[sq->head];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
597
page = (struct page *)sq->xdp_page[sq->head];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
603
hdr = (struct sq_hdr_subdesc *)GET_SQ_DESC(sq, sq->head);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
612
nicvf_unmap_sndq_buffers(nic, sq, sq->head,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
618
sq->head++;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
619
sq->head &= (sq->dmem.q_len - 1);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
666
rbdr->head = nicvf_queue_reg_read(nic,
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
227
u32 head;
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
270
u32 head;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1242
static void clean_l2_data(struct rcu_head *head)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1244
struct l2t_data *d = container_of(head, struct l2t_data, rcu_head);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1586
dui = (struct deferred_unmap_info *)skb->head;
drivers/net/ethernet/chelsio/cxgb3/sge.c
1606
dui = (struct deferred_unmap_info *)skb->head;
drivers/net/ethernet/chelsio/cxgb3/sge.c
1714
map_skb(adap->pdev, skb, (dma_addr_t *)skb->head)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
1729
write_ofld_wr(adap, skb, q, pidx, gen, ndesc, (dma_addr_t *)skb->head);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1769
map_skb(adap->pdev, skb, (dma_addr_t *)skb->head))
drivers/net/ethernet/chelsio/cxgb3/sge.c
1785
(dma_addr_t *)skb->head);
drivers/net/ethernet/chelsio/cxgb4/sge.c
2979
(dma_addr_t *)skb->head)) {
drivers/net/ethernet/chelsio/cxgb4/sge.c
299
unmap_skb(skb->dev->dev.parent, skb, (dma_addr_t *)skb->head);
drivers/net/ethernet/chelsio/cxgb4/sge.c
3012
(dma_addr_t *)skb->head);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1285
req = (struct cpl_rx_data_ack *)skb->head;
drivers/net/ethernet/cisco/enic/enic_clsf.c
235
struct hlist_head *head;
drivers/net/ethernet/cisco/enic/enic_clsf.c
237
head = &enic->rfs_h.ht_head[tbl_idx];
drivers/net/ethernet/cisco/enic/enic_clsf.c
242
hlist_add_head(&d->node, head);
drivers/net/ethernet/dec/tulip/interrupt.c
241
skb->head, temp);
drivers/net/ethernet/dec/tulip/interrupt.c
471
skb->head, temp);
drivers/net/ethernet/dlink/sundance.c
1065
unsigned head = np->cur_task % TX_RING_SIZE;
drivers/net/ethernet/dlink/sundance.c
1083
iowrite32 (np->tx_ring_dma + head * sizeof(struct netdev_desc),
drivers/net/ethernet/emulex/benet/be.h
127
u32 tail, head;
drivers/net/ethernet/emulex/benet/be.h
152
return q->dma_mem.va + q->head * q->entry_size;
drivers/net/ethernet/emulex/benet/be.h
167
index_inc(&q->head, q->len);
drivers/net/ethernet/emulex/benet/be_cmds.c
608
u32 index = mcc_obj->q.head;
drivers/net/ethernet/emulex/benet/be_main.c
1001
head = be_tx_get_wrb_hdr(txo);
drivers/net/ethernet/emulex/benet/be_main.c
1025
be_tx_setup_wrb_hdr(adapter, txo, wrb_params, skb, head);
drivers/net/ethernet/emulex/benet/be_main.c
1032
be_xmit_restore(adapter, txo, head, map_single, copied);
drivers/net/ethernet/emulex/benet/be_main.c
1434
i, txo->q.head, txo->q.tail,
drivers/net/ethernet/emulex/benet/be_main.c
1449
i, txo->cq.head, txo->cq.tail,
drivers/net/ethernet/emulex/benet/be_main.c
2604
page_info = &rxo->page_info_tbl[rxq->head];
drivers/net/ethernet/emulex/benet/be_main.c
2646
page_info = &rxo->page_info_tbl[rxq->head];
drivers/net/ethernet/emulex/benet/be_main.c
2828
rxq->head = 0;
drivers/net/ethernet/emulex/benet/be_main.c
2932
txq->head = notified_idx;
drivers/net/ethernet/emulex/benet/be_main.c
919
u32 head = txo->q.head;
drivers/net/ethernet/emulex/benet/be_main.c
922
return head;
drivers/net/ethernet/emulex/benet/be_main.c
929
struct sk_buff *skb, u16 head)
drivers/net/ethernet/emulex/benet/be_main.c
933
struct be_eth_hdr_wrb *hdr = queue_index_node(txq, head);
drivers/net/ethernet/emulex/benet/be_main.c
938
BUG_ON(txo->sent_skb_list[head]);
drivers/net/ethernet/emulex/benet/be_main.c
939
txo->sent_skb_list[head] = skb;
drivers/net/ethernet/emulex/benet/be_main.c
940
txo->last_req_hdr = head;
drivers/net/ethernet/emulex/benet/be_main.c
963
struct be_tx_obj *txo, u32 head, bool map_single,
drivers/net/ethernet/emulex/benet/be_main.c
971
txq->head = head;
drivers/net/ethernet/emulex/benet/be_main.c
983
txq->head = head;
drivers/net/ethernet/emulex/benet/be_main.c
997
u32 head;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
1081
if (aligned_start >= skb->head)
drivers/net/ethernet/freescale/fec_main.c
464
*(__sum16 *)(skb->head + skb->csum_start + skb->csum_offset) = 0;
drivers/net/ethernet/fungible/funeth/funeth_tx.c
426
unsigned int head, limit, reclaim_idx;
drivers/net/ethernet/fungible/funeth/funeth_tx.c
431
for (head = txq_hw_head(q), reclaim_idx = q->cons_cnt & q->mask;
drivers/net/ethernet/fungible/funeth/funeth_tx.c
432
head != reclaim_idx && npkts < limit; head = txq_hw_head(q)) {
drivers/net/ethernet/fungible/funeth/funeth_tx.c
444
trace_funeth_tx_free(q, reclaim_idx, pkt_desc, head);
drivers/net/ethernet/fungible/funeth/funeth_tx.c
451
} while (reclaim_idx != head && npkts < limit);
drivers/net/ethernet/fungible/funeth/funeth_tx.c
464
return reclaim_idx != head;
drivers/net/ethernet/fungible/funeth/funeth_tx.c
486
unsigned int npkts = 0, ndesc = 0, head, reclaim_idx;
drivers/net/ethernet/fungible/funeth/funeth_tx.c
488
for (head = txq_hw_head(q), reclaim_idx = q->cons_cnt & q->mask;
drivers/net/ethernet/fungible/funeth/funeth_tx.c
489
head != reclaim_idx && npkts < budget; head = txq_hw_head(q)) {
drivers/net/ethernet/fungible/funeth/funeth_tx.c
502
trace_funeth_tx_free(q, reclaim_idx, pkt_desc, head);
drivers/net/ethernet/fungible/funeth/funeth_tx.c
507
} while (reclaim_idx != head && npkts < budget);
drivers/net/ethernet/google/gve/gve.h
154
u32 head; /* Pointer to start cleaning buffers at. */
drivers/net/ethernet/google/gve/gve.h
179
u32 head;
drivers/net/ethernet/google/gve/gve.h
217
s16 head;
drivers/net/ethernet/google/gve/gve.h
388
u32 head; /* offset to write at */
drivers/net/ethernet/google/gve/gve.h
498
u32 head;
drivers/net/ethernet/google/gve/gve.h
548
u32 head; /* Last read on compl_desc */
drivers/net/ethernet/google/gve/gve_adminq.c
463
int tail, head;
drivers/net/ethernet/google/gve/gve_adminq.c
469
head = priv->adminq_prod_cnt;
drivers/net/ethernet/google/gve/gve_adminq.c
471
gve_adminq_kick_cmd(priv, head);
drivers/net/ethernet/google/gve/gve_adminq.c
472
if (!gve_adminq_wait_for_cmd(priv, head)) {
drivers/net/ethernet/google/gve/gve_adminq.c
478
for (i = tail; i < head; i++) {
drivers/net/ethernet/google/gve/gve_adminq.c
601
u32 tail, head;
drivers/net/ethernet/google/gve/gve_adminq.c
606
head = priv->adminq_prod_cnt;
drivers/net/ethernet/google/gve/gve_adminq.c
607
if (tail != head) {
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
108
if (unlikely(rx->dqo.used_buf_states.head == -1))
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
63
buffer_id = list->head;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
70
list->head = buf_state->next;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
87
if (list->head == -1) {
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
88
list->head = buffer_id;
drivers/net/ethernet/google/gve/gve_ethtool.c
419
(tx->dqo_tx.tail - tx->dqo_tx.head) &
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1029
&complq->desc_ring[complq->head];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1037
prefetch(&complq->desc_ring[(complq->head + 1) & complq->mask]);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1038
prefetch(&complq->desc_ring[(complq->head + 2) & complq->mask]);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1043
err = gve_rx_dqo(napi, rx, compl_desc, complq->head, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1054
complq->head = (complq->head + 1) & complq->mask;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1058
complq->cur_gen_bit ^= (complq->head == 0);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1066
bufq->head = (bufq->head + 1) & bufq->mask;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
372
num_full_slots = (bufq->tail - bufq->head) & bufq->mask;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
41
rx->dqo.bufq.head = 0;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
48
rx->dqo.complq.head = 0;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
62
rx->dqo.recycled_buf_states.head = -1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
64
rx->dqo.used_buf_states.head = -1;
drivers/net/ethernet/google/gve/gve_tx.c
105
iov[0].iov_offset = fifo->head;
drivers/net/ethernet/google/gve/gve_tx.c
107
fifo->head += bytes;
drivers/net/ethernet/google/gve/gve_tx.c
109
if (fifo->head > fifo->size) {
drivers/net/ethernet/google/gve/gve_tx.c
114
overflow = fifo->head - fifo->size;
drivers/net/ethernet/google/gve/gve_tx.c
119
fifo->head = overflow;
drivers/net/ethernet/google/gve/gve_tx.c
123
aligned_head = L1_CACHE_ALIGN(fifo->head);
drivers/net/ethernet/google/gve/gve_tx.c
124
padding = aligned_head - fifo->head;
drivers/net/ethernet/google/gve/gve_tx.c
127
fifo->head = aligned_head;
drivers/net/ethernet/google/gve/gve_tx.c
129
if (fifo->head == fifo->size)
drivers/net/ethernet/google/gve/gve_tx.c
130
fifo->head = 0;
drivers/net/ethernet/google/gve/gve_tx.c
51
fifo->head = 0;
drivers/net/ethernet/google/gve/gve_tx.c
66
return (fifo->head + bytes < fifo->size) ? 0 : fifo->size - fifo->head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1029
u32 head = tx->dqo_compl.xsk_reorder_queue_head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1031
if (head == tx->dqo_compl.xsk_reorder_queue_tail) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1035
if (head == tx->dqo_compl.xsk_reorder_queue_tail)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1039
return &tx->dqo.pending_packets[tx->dqo.xsk_reorder_queue[head]];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1136
list->head = index;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1155
list->head = next_index;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1294
next_index = tx->dqo_compl.miss_completions.head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1344
next_index = tx->dqo_compl.timed_out_completions.head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1398
&tx->dqo.compl_ring[tx->dqo_compl.head];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1405
prefetch(&tx->dqo.compl_ring[(tx->dqo_compl.head + 1) &
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1447
tx->dqo_compl.head =
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1448
(tx->dqo_compl.head + 1) & tx->dqo.complq_mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1450
tx->dqo_compl.cur_gen_bit ^= tx->dqo_compl.head == 0;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1493
compl_desc = &tx->dqo.compl_ring[tx->dqo_compl.head];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1519
compl_desc = &tx->dqo.compl_ring[tx->dqo_compl.head];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
371
tx->dqo_compl.miss_completions.head = -1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
373
tx->dqo_compl.timed_out_completions.head = -1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
471
u32 num_used = (tx->dqo_tx.tail - tx->dqo_tx.head) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
487
tx->dqo_tx.head = atomic_read_acquire(&tx->dqo_compl.hw_tx_head);
drivers/net/ethernet/hisilicon/hibmcge/hbg_common.h
75
u32 head;
drivers/net/ethernet/hisilicon/hibmcge/hbg_debugfs.c
33
seq_printf(s, "ring head: %u, tail: %u\n", ring->head, ring->tail);
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
19
#define hbg_queue_used_num(head, tail, ring) ({ \
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
21
((tail) + _ring->len - (head)) % _ring->len; })
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
22
#define hbg_queue_left_num(head, tail, ring) ({ \
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
24
_r->len - hbg_queue_used_num((head), (tail), _r) - 1; })
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
25
#define hbg_queue_is_empty(head, tail, ring) \
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
26
(hbg_queue_used_num((head), (tail), (ring)) == 0)
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
27
#define hbg_queue_is_full(head, tail, ring) \
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
28
(hbg_queue_left_num((head), (tail), (ring)) == 0)
drivers/net/ethernet/hisilicon/hip04_eth.c
251
static inline unsigned int tx_count(unsigned int head, unsigned int tail)
drivers/net/ethernet/hisilicon/hip04_eth.c
253
return (head - tail) % TX_DESC_NUM;
drivers/net/ethernet/hisilicon/hisi_femac.c
104
unsigned int head;
drivers/net/ethernet/hisilicon/hisi_femac.c
218
pos = rxq->head;
drivers/net/ethernet/hisilicon/hisi_femac.c
242
rxq->head = pos;
drivers/net/ethernet/hisilicon/hisi_femac.c
362
queue->head = 0;
drivers/net/ethernet/hisilicon/hisi_femac.c
394
while (pos != rxq->head) {
drivers/net/ethernet/hisilicon/hisi_femac.c
398
pos, rxq->head);
drivers/net/ethernet/hisilicon/hisi_femac.c
413
while (pos != txq->head) {
drivers/net/ethernet/hisilicon/hisi_femac.c
417
pos, txq->head);
drivers/net/ethernet/hisilicon/hisi_femac.c
517
if (unlikely(!CIRC_SPACE(txq->head, txq->tail,
drivers/net/ethernet/hisilicon/hisi_femac.c
533
txq->dma_phys[txq->head] = addr;
drivers/net/ethernet/hisilicon/hisi_femac.c
535
txq->skb[txq->head] = skb;
drivers/net/ethernet/hisilicon/hisi_femac.c
536
txq->head = (txq->head + 1) % txq->num;
drivers/net/ethernet/hisilicon/hns/hnae.c
20
hnae_list_add(spinlock_t *lock, struct list_head *node, struct list_head *head)
drivers/net/ethernet/hisilicon/hns/hnae.c
25
list_add_tail_rcu(node, head);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
63
u32 head, tail;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
69
head = dsaf_read_dev(&qs->tx_ring, RCB_REG_HEAD);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
70
if (tail == head)
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1029
int head;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1033
head = readl_relaxed(ring->io_base + RCB_REG_HEAD);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1035
if (head != ring->next_to_clean) {
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1048
int head = readl_relaxed(ring->io_base + RCB_REG_HEAD);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1050
if (head == ring->next_to_clean)
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1061
int head;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1064
head = ring->next_to_use; /* ntu :soft setted ring position*/
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1067
while (head != ring->next_to_clean)
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1634
int head, tail;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1665
head = readl_relaxed(ring->io_base + RCB_REG_HEAD);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1668
fetch_num = ring_dist(ring, head, tail);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1670
while (head != tail) {
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1671
if (ring->desc_cb[head].page_offset != 0) {
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1676
head++;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1677
if (head == ring->desc_num)
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1678
head = 0;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
978
int head;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
981
head = readl_relaxed(ring->io_base + RCB_REG_HEAD);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
984
if (is_ring_empty(ring) || head == ring->next_to_clean)
drivers/net/ethernet/hisilicon/hns/hns_enet.c
987
if (!is_valid_clean_head(ring, head)) {
drivers/net/ethernet/hisilicon/hns/hns_enet.c
988
netdev_err(ndev, "wrong head (%d, %d-%d)\n", head,
drivers/net/ethernet/hisilicon/hns/hns_enet.c
996
while (head != ring->next_to_clean) {
drivers/net/ethernet/hisilicon/hns3/hclge_mbx.h
230
u32 head;
drivers/net/ethernet/hisilicon/hns3/hclge_mbx.h
252
(arq.head = (arq.head + 1) % HCLGE_MBX_MAX_ARQ_MSG_NUM)
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
312
int head)
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
318
return head >= ntc && head <= ntu;
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
320
return head >= ntc || head <= ntu;
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
327
u32 head;
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
329
head = hclge_comm_read_dev(hw, HCLGE_COMM_NIC_CSQ_HEAD_REG);
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
332
if (!hclge_comm_is_valid_csq_clean_head(csq, head)) {
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
334
head, csq->next_to_use, csq->next_to_clean);
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
343
clean = (head - csq->next_to_clean + csq->desc_num) % csq->desc_num;
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
344
csq->next_to_clean = head;
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
350
u32 head = hclge_comm_read_dev(hw, HCLGE_COMM_NIC_CSQ_HEAD_REG);
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.c
351
return head == (u32)hw->cmq.csq.next_to_use;
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.h
418
u32 head;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3912
skb->csum_start = (unsigned char *)th - skb->head;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4568
struct hnae3_ring_chain_node **head,
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4573
struct hnae3_ring_chain_node *cur_chain = *head;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4592
*head = chain;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4634
struct hnae3_ring_chain_node *head)
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4639
chain = head;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.h
696
#define hns3_for_each_ring(pos, head) \
drivers/net/ethernet/hisilicon/hns3/hns3_enet.h
697
for ((pos) = (head).ring; (pos); (pos) = (pos)->next)
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c
157
static void hclge_free_vector_ring_chain(struct hnae3_ring_chain_node *head)
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c
161
chain = head->next;
drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c
94
hdev->arq.head = 0;
drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c
317
while (tail != hdev->arq.head) {
drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c
325
msg_q = hdev->arq.msg_q[hdev->arq.head];
drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c
386
msg_q = hdev->arq.msg_q[hdev->arq.head];
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
641
if (err || cmdq_ctxt.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
643
err, cmdq_ctxt.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_eqs.c
382
if (err || ceq_ctrl.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_eqs.c
384
q_id, err, ceq_ctrl.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_filter.c
154
static void hinic3_cleanup_filter_list(const struct list_head *head)
drivers/net/ethernet/huawei/hinic3/hinic3_filter.c
159
list_for_each_entry_safe(f, ftmp, head, list) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_cfg.c
60
if (err || dev_cap.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_cfg.c
63
err, dev_cap.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
111
if (err || func_reset.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
113
reset_flag, err, func_reset.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
138
if (err || feature_nego.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
140
err, feature_nego.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
176
if (err || get_attr.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
179
err, get_attr.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
205
if (err || used_state.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
208
err, used_state.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
234
if (err || dma_attr.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
236
err, dma_attr.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
259
if (err || page_size_info.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
262
err, page_size_info.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
27
if (err || msix_cfg.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
285
if (err || root_ctxt.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
288
err, root_ctxt.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
29
err, msix_cfg.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
362
if (err || clear_db.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
364
err, clear_db.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
406
if (err || bdf_info->head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
409
err, bdf_info->head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
429
if (err || time_info.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
432
err, time_info.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
510
if (err || root_ctxt.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
513
err, root_ctxt.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
533
if (err || root_ctxt.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
536
err, root_ctxt.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
63
if (err || msix_cfg.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
66
err, msix_cfg.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
122
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
155
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
163
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
181
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
191
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
199
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
211
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
220
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
229
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
247
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
255
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
261
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
268
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
62
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_mgmt_interface.h
135
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_mgmt_interface.h
266
struct mgmt_msg_head head;
drivers/net/ethernet/huawei/hinic3/hinic3_nic_cfg.c
585
if (err || en_state.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_nic_cfg.c
587
err, en_state.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_nic_cfg.c
609
if (err || dcb_state.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_nic_cfg.c
612
err, dcb_state.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_nic_cfg.c
631
if (err || get_link.head.status) {
drivers/net/ethernet/huawei/hinic3/hinic3_nic_cfg.c
633
err, get_link.head.status);
drivers/net/ethernet/huawei/hinic3/hinic3_nic_cfg.h
45
struct mgmt_msg_head head;
drivers/net/ethernet/intel/e1000e/e1000.h
164
void __iomem *head;
drivers/net/ethernet/intel/e1000e/netdev.c
1153
readl(tx_ring->head), readl(tx_ring->tail), tx_ring->next_to_use,
drivers/net/ethernet/intel/e1000e/netdev.c
2926
tx_ring->head = adapter->hw.hw_addr + E1000_TDH(0);
drivers/net/ethernet/intel/e1000e/netdev.c
3248
rx_ring->head = adapter->hw.hw_addr + E1000_RDH(0);
drivers/net/ethernet/intel/fm10k/fm10k.h
173
#define fm10k_for_each_ring(pos, head) \
drivers/net/ethernet/intel/fm10k/fm10k.h
174
for (pos = &(head).ring[(head).count]; (--pos) >= (head).ring;)
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1116
u32 head, tail;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1119
head = ring->next_to_clean;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1122
head = fm10k_read_reg(hw, FM10K_TDH(ring->reg_idx));
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1126
return ((head <= tail) ? tail : tail + ring->count) - head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1018
struct fm10k_mbx_info *mbx, u16 head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1024
fm10k_mbx_update_local_crc(mbx, head);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1027
fm10k_mbx_pull_head(hw, mbx, head);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
106
fifo->head += len;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1060
u16 len, head, ack;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1066
head = FM10K_MSG_HDR_FIELD_GET(mbx->mbx_hdr, HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1067
ack = fm10k_mbx_index_len(mbx, head, mbx->tail);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1086
mbx->rx.head = 0;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1152
u16 size, head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1156
head = FM10K_MSG_HDR_FIELD_GET(*hdr, HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1180
mbx->tail = head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1182
return fm10k_mbx_create_reply(hw, mbx, head);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1198
u16 head, tail;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
120
fifo->head = fifo->tail;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1202
head = FM10K_MSG_HDR_FIELD_GET(*hdr, HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1207
mbx->tail = head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1224
return fm10k_mbx_create_reply(hw, mbx, head);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1241
u16 head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1245
head = FM10K_MSG_HDR_FIELD_GET(*hdr, HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1267
if (head != mbx->tail)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1277
return fm10k_mbx_create_reply(hw, mbx, head);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1293
u16 head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1296
head = FM10K_MSG_HDR_FIELD_GET(*hdr, HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1309
mbx->tail = head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
132
static u16 fm10k_mbx_index_len(struct fm10k_mbx_info *mbx, u16 head, u16 tail)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
134
u16 len = tail - head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1597
mbx->head = 1;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
16
fifo->head = 0;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1638
FM10K_MSG_HDR_FIELD_SET(mbx->head, SM_HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1655
FM10K_MSG_HDR_FIELD_SET(mbx->head, SM_HEAD) |
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1676
mbx->head = 1;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1775
u16 tail, head, ver;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1779
head = FM10K_MSG_HDR_FIELD_GET(*hdr, SM_HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1785
if (!head || head > FM10K_SM_MBX_FIFO_LEN)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1789
if (mbx->tail < head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1790
head += mbx->mbmem_len - 1;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1791
if (tail < mbx->head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1793
if (fm10k_mbx_index_len(mbx, head, mbx->tail) > mbx->tail_len)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1795
if (fm10k_mbx_index_len(mbx, mbx->head, tail) < mbx->mbmem_len)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
185
u16 head = (mbx->head + offset + 1) & ((mbx->mbmem_len << 1) - 1);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
188
return (head > mbx->head) ? --head : ++head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1887
if (tail < mbx->head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1899
mbx->head = fm10k_mbx_head_sub(mbx, mbx->pushed);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1903
if (mbx->head > mbmem_len)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1904
mbx->head -= mbmem_len;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1919
struct fm10k_mbx_info *mbx, u16 head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1927
if (mbx->tail < head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1928
head += mbmem_len;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1930
fm10k_mbx_pull_head(hw, mbx, head);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1964
struct fm10k_mbx_info *mbx, u16 head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1970
fm10k_sm_mbx_transmit(hw, mbx, head);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
201
u16 head = (mbx->head - offset - 1) & ((mbx->mbmem_len << 1) - 1);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
204
return (head < mbx->head) ? ++head : --head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
2043
u16 head, tail;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
2048
head = FM10K_MSG_HDR_FIELD_GET(*hdr, SM_HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
2070
fm10k_sm_mbx_create_reply(hw, mbx, head);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
28
return fifo->tail - fifo->head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
335
u32 *head = fifo->buffer;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
350
head += end;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
356
for (end = fifo->size - end; len; head = fifo->buffer) {
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
366
fm10k_write_reg(hw, mbmem + tail++, *(head++));
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
383
struct fm10k_mbx_info *mbx, u16 head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
385
u16 mbmem_len, len, ack = fm10k_mbx_index_len(mbx, head, mbx->tail);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
39
return fifo->size + fifo->head - fifo->tail;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
429
u16 end, len, head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
433
head = fm10k_mbx_head_sub(mbx, len);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
434
if (head >= mbx->mbmem_len)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
435
head++;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
445
head &= mbx->mbmem_len - 1;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
446
if (!head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
447
head++;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
452
*(tail++) = fm10k_read_reg(hw, mbmem + head++);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
476
u16 len, seq = fm10k_mbx_index_len(mbx, mbx->head, tail);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
484
mbx->head = fm10k_mbx_head_add(mbx, len);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
50
return fifo->head == fifo->tail;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
617
static void fm10k_mbx_update_local_crc(struct fm10k_mbx_info *mbx, u16 head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
619
u16 len = mbx->tail_len - fm10k_mbx_index_len(mbx, head, mbx->tail);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
62
return (fifo->head + offset) & (fifo->size - 1);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
622
head = fm10k_fifo_head_offset(&mbx->tx, mbx->pulled);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
625
mbx->local = fm10k_fifo_crc(&mbx->tx, head, len, mbx->local);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
711
err = fm10k_tlv_msg_parse(hw, fifo->buffer + fifo->head,
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
723
fifo->tail -= fifo->head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
724
fifo->head = 0;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
841
FM10K_MSG_HDR_FIELD_SET(mbx->head, HEAD) |
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
85
u32 *head = fifo->buffer + fm10k_fifo_head_offset(fifo, 0);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
855
FM10K_MSG_HDR_FIELD_SET(mbx->head, HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
881
FM10K_MSG_HDR_FIELD_SET(mbx->head, HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
901
FM10K_MSG_HDR_FIELD_SET(mbx->head, TAIL) |
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
92
return FM10K_TLV_DWORD_LEN(*head);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
939
FM10K_MSG_HDR_FIELD_SET(mbx->head, HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
952
u16 type, rsvd0, head, tail, size;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
958
head = FM10K_MSG_HDR_FIELD_GET(*hdr, HEAD);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
967
if (tail != mbx->head)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
973
if (!head || (head == FM10K_MSG_HDR_MASK(HEAD)))
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
975
if (fm10k_mbx_index_len(mbx, head, mbx->tail) > mbx->tail_len)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
981
if (fm10k_mbx_index_len(mbx, mbx->head, tail) < mbx->mbmem_len)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
992
if (!head || (head == FM10K_MSG_HDR_MASK(HEAD)))
drivers/net/ethernet/intel/fm10k/fm10k_mbx.h
231
u16 head;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.h
267
u16 head, head_len, pushed;
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
671
{I40E_HMC_STORE(i40e_hmc_obj_txq, head), 13, 0 },
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
697
{ I40E_HMC_STORE(i40e_hmc_obj_rxq, head), 13, 0 },
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.h
22
u16 head;
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.h
55
u16 head;
drivers/net/ethernet/intel/i40e/i40e_main.c
361
u32 head, val;
drivers/net/ethernet/intel/i40e/i40e_main.c
387
head = i40e_get_head(tx_ring);
drivers/net/ethernet/intel/i40e/i40e_main.c
398
head, tx_ring->next_to_use,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
846
u32 head, tail;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
849
head = i40e_get_head(ring);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
852
head = ring->next_to_clean;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
856
if (head != tail)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
857
return (head < tail) ?
drivers/net/ethernet/intel/i40e/i40e_txrx.c
858
tail - head : (tail + ring->count - head);
drivers/net/ethernet/intel/i40e/i40e_txrx.h
447
#define i40e_for_each_ring(pos, head) \
drivers/net/ethernet/intel/i40e/i40e_txrx.h
448
for (pos = (head).ring; pos != NULL; pos = pos->next)
drivers/net/ethernet/intel/i40e/i40e_txrx.h
491
void *head = (struct i40e_tx_desc *)tx_ring->desc + tx_ring->count;
drivers/net/ethernet/intel/i40e/i40e_txrx.h
493
return le32_to_cpu(*(volatile __le32 *)head);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
137
u32 head, tail;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
142
head = ring->next_to_clean;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
145
if (head != tail)
drivers/net/ethernet/intel/iavf/iavf_txrx.c
146
return (head < tail) ?
drivers/net/ethernet/intel/iavf/iavf_txrx.c
147
tail - head : (tail + ring->count - head);
drivers/net/ethernet/intel/iavf/iavf_txrx.h
309
#define iavf_for_each_ring(pos, head) \
drivers/net/ethernet/intel/iavf/iavf_txrx.h
310
for (pos = (head).ring; pos != NULL; pos = pos->next)
drivers/net/ethernet/intel/ice/devlink/health.c
379
u32 head;
drivers/net/ethernet/intel/ice/devlink/health.c
400
ICE_DEVLINK_FMSG_PUT_FIELD(fmsg, event, head);
drivers/net/ethernet/intel/ice/devlink/health.c
419
u16 vsi_num, u32 head, u32 intr)
drivers/net/ethernet/intel/ice/devlink/health.c
425
buf->head = head;
drivers/net/ethernet/intel/ice/devlink/health.c
435
.head = buf->head,
drivers/net/ethernet/intel/ice/devlink/health.h
50
u32 head;
drivers/net/ethernet/intel/ice/devlink/health.h
66
u16 vsi_num, u32 head, u32 intr);
drivers/net/ethernet/intel/ice/ice_common.c
1386
ICE_CTX_STORE(ice_rlan_ctx, head, 13, 0),
drivers/net/ethernet/intel/ice/ice_controlq.c
1037
val = rd32(hw, cq->sq.head);
drivers/net/ethernet/intel/ice/ice_controlq.c
1199
ntu = (u16)(rd32(hw, cq->rq.head) & cq->rq.head_mask);
drivers/net/ethernet/intel/ice/ice_controlq.c
1254
ntu = (u16)(rd32(hw, cq->rq.head) & cq->rq.head_mask);
drivers/net/ethernet/intel/ice/ice_controlq.c
17
(qinfo)->rq.head = prefix##_ARQH; \
drivers/net/ethernet/intel/ice/ice_controlq.c
265
wr32(hw, ring->head, 0);
drivers/net/ethernet/intel/ice/ice_controlq.c
473
wr32(hw, cq->sq.head, 0);
drivers/net/ethernet/intel/ice/ice_controlq.c
548
wr32(hw, cq->rq.head, 0);
drivers/net/ethernet/intel/ice/ice_controlq.c
8
(qinfo)->sq.head = prefix##_ATQH; \
drivers/net/ethernet/intel/ice/ice_controlq.c
865
while (rd32(hw, cq->sq.head) != ntc) {
drivers/net/ethernet/intel/ice/ice_controlq.c
866
ice_debug(hw, ICE_DBG_AQ_MSG, "ntc %d head %d.\n", ntc, rd32(hw, cq->sq.head));
drivers/net/ethernet/intel/ice/ice_controlq.c
969
u32 head;
drivers/net/ethernet/intel/ice/ice_controlq.c
976
return !rd32_poll_timeout(hw, cq->sq.head,
drivers/net/ethernet/intel/ice/ice_controlq.c
977
head, head == cq->sq.next_to_use,
drivers/net/ethernet/intel/ice/ice_controlq.h
66
u32 head;
drivers/net/ethernet/intel/ice/ice_lan_tx_rx.h
383
u16 head;
drivers/net/ethernet/intel/ice/ice_main.c
104
u16 head, tail;
drivers/net/ethernet/intel/ice/ice_main.c
106
head = ring->next_to_clean;
drivers/net/ethernet/intel/ice/ice_main.c
109
if (head != tail)
drivers/net/ethernet/intel/ice/ice_main.c
110
return (head < tail) ?
drivers/net/ethernet/intel/ice/ice_main.c
111
tail - head : (tail + ring->count - head);
drivers/net/ethernet/intel/ice/ice_main.c
1578
ntu = (u16)(rd32(hw, cq->rq.head) & cq->rq.head_mask);
drivers/net/ethernet/intel/ice/ice_main.c
8316
u32 head, intr = 0;
drivers/net/ethernet/intel/ice/ice_main.c
8318
head = FIELD_GET(QTX_COMM_HEAD_HEAD_M,
drivers/net/ethernet/intel/ice/ice_main.c
8325
head, tx_ring->next_to_use, intr);
drivers/net/ethernet/intel/ice/ice_main.c
8327
ice_prep_tx_hang_report(pf, tx_ring, vsi->vsi_num, head, intr);
drivers/net/ethernet/intel/ice/ice_switch.c
6618
struct list_head *head;
drivers/net/ethernet/intel/ice/ice_switch.c
6620
head = &sw->recp_list[i].filt_replay_rules;
drivers/net/ethernet/intel/ice/ice_switch.c
6622
status = ice_replay_vsi_fltr(hw, vsi_handle, i, head);
drivers/net/ethernet/intel/ice/ice_switch.c
6624
status = ice_replay_vsi_adv_rule(hw, vsi_handle, head);
drivers/net/ethernet/intel/ice/ice_txrx.h
440
#define ice_for_each_rx_ring(pos, head) \
drivers/net/ethernet/intel/ice/ice_txrx.h
441
for (pos = (head).rx_ring; pos; pos = pos->next)
drivers/net/ethernet/intel/ice/ice_txrx.h
443
#define ice_for_each_tx_ring(pos, head) \
drivers/net/ethernet/intel/ice/ice_txrx.h
444
for (pos = (head).tx_ring; pos; pos = pos->next)
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
325
struct ice_tx_buf *head = tx_buf;
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
348
ice_clean_xdp_tx_buf(dev, head, &bq);
drivers/net/ethernet/intel/ice/ice_type.h
815
u32 head;
drivers/net/ethernet/intel/ice/ice_vf_mbx.c
179
if (num_iterations == snap_buf->head ||
drivers/net/ethernet/intel/ice/ice_vf_mbx.c
322
snap_buf->head = ICE_RQ_DATA_MASK(cq->rq.next_to_clean +
drivers/net/ethernet/intel/ice/ice_xsk.c
441
struct xdp_buff *head;
drivers/net/ethernet/intel/ice/ice_xsk.c
462
head = xdp;
drivers/net/ethernet/intel/ice/ice_xsk.c
486
xdp = xsk_buff_get_frag(head);
drivers/net/ethernet/intel/idpf/idpf_controlq.c
15
cq->reg.head = q_create_info->reg.head;
drivers/net/ethernet/intel/idpf/idpf_controlq.c
46
idpf_mbx_wr32(hw, cq->reg.head, 0);
drivers/net/ethernet/intel/idpf/idpf_controlq_api.h
26
u32 head;
drivers/net/ethernet/intel/idpf/idpf_dev.c
28
ccq->reg.head = PF_FW_ATQH - mbx_start;
drivers/net/ethernet/intel/idpf/idpf_dev.c
39
ccq->reg.head = PF_FW_ARQH - mbx_start;
drivers/net/ethernet/intel/idpf/idpf_ptp.c
651
struct list_head *head;
drivers/net/ethernet/intel/idpf/idpf_ptp.c
656
head = &tx_q->cached_tstamp_caps->latches_free;
drivers/net/ethernet/intel/idpf/idpf_ptp.c
657
if (list_empty(head)) {
drivers/net/ethernet/intel/idpf/idpf_ptp.c
662
ptp_tx_tstamp = list_first_entry(head, struct idpf_ptp_tx_tstamp,
drivers/net/ethernet/intel/idpf/idpf_ptp.c
849
struct list_head *head;
drivers/net/ethernet/intel/idpf/idpf_ptp.c
856
head = &vport->tx_tstamp_caps->latches_free;
drivers/net/ethernet/intel/idpf/idpf_ptp.c
857
list_for_each_entry_safe(ptp_tx_tstamp, tmp, head, list_member) {
drivers/net/ethernet/intel/idpf/idpf_ptp.c
863
head = &vport->tx_tstamp_caps->latches_in_use;
drivers/net/ethernet/intel/idpf/idpf_ptp.c
865
list_for_each_entry_safe(ptp_tx_tstamp, tmp, head, list_member) {
drivers/net/ethernet/intel/idpf/idpf_vf_dev.c
27
ccq->reg.head = VF_ATQH - mbx_start;
drivers/net/ethernet/intel/idpf/idpf_vf_dev.c
38
ccq->reg.head = VF_ARQH - mbx_start;
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
342
struct list_head *head;
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
428
head = &tstamp_caps->latches_free;
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
429
list_for_each_entry_safe(ptp_tx_tstamp, tmp, head, list_member) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
553
struct list_head *head;
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
578
head = &tx_tstamp_caps->latches_in_use;
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
586
if (list_empty(head)) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
591
list_for_each_entry_safe(tx_tstamp, tmp, head, list_member) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
632
struct list_head *head;
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
637
head = &tx_tstamp_caps->latches_in_use;
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
646
list_for_each_entry(ptp_tx_tstamp, head, list_member) {
drivers/net/ethernet/intel/igb/igb_main.c
1147
struct igb_ring_container *head)
drivers/net/ethernet/intel/igb/igb_main.c
1149
head->ring = ring;
drivers/net/ethernet/intel/igb/igb_main.c
1150
head->count++;
drivers/net/ethernet/intel/igbvf/igbvf.h
135
u16 head;
drivers/net/ethernet/intel/igbvf/netdev.c
1303
tx_ring->head = E1000_TDH(0);
drivers/net/ethernet/intel/igbvf/netdev.c
1384
rx_ring->head = E1000_RDH(0);
drivers/net/ethernet/intel/igbvf/netdev.c
521
writel(0, adapter->hw.hw_addr + tx_ring->head);
drivers/net/ethernet/intel/igbvf/netdev.c
604
writel(0, adapter->hw.hw_addr + rx_ring->head);
drivers/net/ethernet/intel/igc/igc_main.c
2387
struct igc_tx_buffer *head = &ring->tx_buffer_info[index];
drivers/net/ethernet/intel/igc/igc_main.c
2388
struct igc_tx_buffer *buffer = head;
drivers/net/ethernet/intel/igc/igc_main.c
2404
head->bytecount = xdp_get_frame_len(xdpf);
drivers/net/ethernet/intel/igc/igc_main.c
2405
head->type = IGC_TX_BUFFER_TYPE_XDP;
drivers/net/ethernet/intel/igc/igc_main.c
2406
head->gso_segs = 1;
drivers/net/ethernet/intel/igc/igc_main.c
2407
head->xdpf = xdpf;
drivers/net/ethernet/intel/igc/igc_main.c
2409
olinfo_status = head->bytecount << IGC_ADVTXD_PAYLEN_SHIFT;
drivers/net/ethernet/intel/igc/igc_main.c
2449
netdev_tx_sent_queue(txring_txq(ring), head->bytecount);
drivers/net/ethernet/intel/igc/igc_main.c
2451
head->time_stamp = jiffies;
drivers/net/ethernet/intel/igc/igc_main.c
2453
head->next_to_watch = desc;
drivers/net/ethernet/intel/igc/igc_main.c
2467
if (buffer == head)
drivers/net/ethernet/intel/igc/igc_main.c
4771
struct igc_ring_container *head)
drivers/net/ethernet/intel/igc/igc_main.c
4773
head->ring = ring;
drivers/net/ethernet/intel/igc/igc_main.c
4774
head->count++;
drivers/net/ethernet/intel/ixgbe/ixgbe.h
492
#define ixgbe_for_each_ring(pos, head) \
drivers/net/ethernet/intel/ixgbe/ixgbe.h
493
for (pos = (head).ring; pos != NULL; pos = pos->next)
drivers/net/ethernet/intel/ixgbe/ixgbe.h
917
struct sk_buff *head;
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
810
struct ixgbe_ring_container *head)
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
812
ring->next = head->ring;
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
813
head->ring = ring;
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
814
head->count++;
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
815
head->next_update = jiffies + 1;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1033
unsigned int head, tail;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1035
head = ring->next_to_clean;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1038
return ((head <= tail) ? tail : tail + ring->count) - head;
drivers/net/ethernet/intel/ixgbevf/ixgbevf.h
225
#define ixgbevf_for_each_ring(pos, head) \
drivers/net/ethernet/intel/ixgbevf/ixgbevf.h
226
for (pos = (head).ring; pos != NULL; pos = pos->next)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
204
u32 head = IXGBE_READ_REG(hw, IXGBE_VFTDH(ring->reg_idx));
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
207
if (head != tail)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
208
return (head < tail) ?
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
209
tail - head : (tail + ring->count - head);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
2734
struct ixgbevf_ring_container *head)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
2736
ring->next = head->ring;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
2737
head->ring = ring;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
2738
head->count++;
drivers/net/ethernet/intel/libeth/xdp.c
338
goto head;
drivers/net/ethernet/intel/libeth/xdp.c
344
head:
drivers/net/ethernet/intel/libeth/xsk.c
53
struct libeth_xdp_buff *libeth_xsk_buff_add_frag(struct libeth_xdp_buff *head,
drivers/net/ethernet/intel/libeth/xsk.c
56
if (!xsk_buff_add_frag(&head->base, &xdp->base))
drivers/net/ethernet/intel/libeth/xsk.c
59
return head;
drivers/net/ethernet/intel/libeth/xsk.c
63
libeth_xsk_buff_free_slow(head);
drivers/net/ethernet/intel/libie/fwlog.c
1124
libie_fwlog_ring_increment(&fwlog->ring.head, fwlog->ring.size);
drivers/net/ethernet/intel/libie/fwlog.c
176
fwlog->ring.head = 0;
drivers/net/ethernet/intel/libie/fwlog.c
770
log = &fwlog->ring.rings[fwlog->ring.head];
drivers/net/ethernet/intel/libie/fwlog.c
789
libie_fwlog_ring_increment(&fwlog->ring.head, fwlog->ring.size);
drivers/net/ethernet/intel/libie/fwlog.c
80
return rings->head == rings->tail;
drivers/net/ethernet/intel/libie/fwlog.c
818
fwlog->ring.head = 0;
drivers/net/ethernet/intel/libie/fwlog.c
907
u16 head, tail;
drivers/net/ethernet/intel/libie/fwlog.c
909
head = rings->head;
drivers/net/ethernet/intel/libie/fwlog.c
912
if (head < tail && (tail - head == (rings->size - 1)))
drivers/net/ethernet/intel/libie/fwlog.c
914
else if (head > tail && (tail == (head - 1)))
drivers/net/ethernet/marvell/octeontx2/af/cn20k/debugfs.c
109
cq_ctx->head, cq_ctx->tail);
drivers/net/ethernet/marvell/octeontx2/af/cn20k/struct.h
122
u64 head : 20;
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
159
struct hlist_head head;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
2279
cq_ctx->head, cq_ctx->tail);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1305
aq_req.cq_mask.head = 0;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
171
INIT_HLIST_HEAD(&list->head);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3214
hlist_for_each_entry_safe(mce, tmp, &mce_list->head, node) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3239
hlist_for_each_entry(mce, &mce_list->head, node) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3291
hlist_for_each_entry(mce, &mce_list->head, node) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3339
hlist_for_each_entry(mce, &mce_list->head, node) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3385
hlist_add_head(&mce->node, &mce_list->head);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3399
hlist_for_each_entry_safe(mce, tmp, &mce_list->head, node) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3415
hlist_for_each_entry(mce, &mce_list->head, node) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3443
hlist_add_head(&mce->node, &mce_list->head);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3495
hlist_for_each_entry(mce, &mce_list->head, node) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5228
hlist_for_each_entry(mce, &mce_list->head, node) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
6368
hlist_for_each_entry_safe(mce, tmp, &mce_list->head, node) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
6427
INIT_HLIST_HEAD(&elem->mcast_mce_list.head);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
972
u64 reg, head;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
979
head = (reg >> 4) & AQ_PTR_MASK;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
981
memcpy((void *)(aq->inst->base + (head * aq->inst->entry_sz)),
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
21
u64 reg, head;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
27
head = (reg >> 4) & AQ_PTR_MASK;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
29
memcpy((void *)(aq->inst->base + (head * aq->inst->entry_sz)),
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1067
struct list_head *head = &mcam->mcam_rules;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1074
head = &iter->list;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1077
list_add(&rule->list, head);
drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h
357
u64 head : 20;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
178
sq->head++;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
179
sq->head &= (sq->sqe_cnt - 1);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
1016
dptr_iova = (sq->sqe_ring->iova + (sq->head * (sq->sqe_size * 2)));
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
1035
sq->head++;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
1036
sq->head &= (sq->sqe_cnt - 1);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
876
sq->sg[sq->head].num_segs = 0;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
908
sq->sg[sq->head].dma_addr[seg] = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
909
sq->sg[sq->head].size[seg] = len;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
910
sq->sg[sq->head].num_segs++;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
916
sq->sg[sq->head].skb = (u64)skb;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
988
res = (struct cpt_res_s *)(sq->cpt_resp->base + (64 * sq->head));
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
990
inst.res_addr = sq->cpt_resp->iova + (64 * sq->head);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1022
sq->head = 0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
471
struct list_head *head = &pfvf->flow_cfg->flow_list;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
477
head = &iter->list;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
480
list_add(&flow->list, head);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1181
iova = sq->timestamps->iova + (sq->head * sizeof(u64));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1255
otx2_dma_unmap_skb_frags(pfvf, &sq->sg[sq->head]);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1432
sq->sg[sq->head].dma_addr[0] = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1433
sq->sg[sq->head].size[0] = len;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1434
sq->sg[sq->head].num_segs = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1435
sq->sg[sq->head].flags = flags;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1436
sq->sg[sq->head].skb = (u64)xdpf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1477
sqe_hdr->sqe_id = sq->head;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
36
return (sq->cons_head - sq->head - 1 + sq->sqe_cnt)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
52
(sq->head * (sq->sqe_size * 2));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
657
sq->head++;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
658
sq->head &= (sq->sqe_cnt - 1);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
670
sq->sg[sq->head].num_segs = 0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
697
sq->sg[sq->head].dma_addr[seg] = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
698
sq->sg[sq->head].size[seg] = len;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
699
sq->sg[sq->head].num_segs++;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
702
sq->sg[sq->head].skb = (u64)skb;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
818
sqe_hdr->sqe_id = sq->head;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
934
int first_sqe = sq->head;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
964
hdr = sq->tso_hdrs->base + (sq->head * TSO_HEADER_SIZE);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
967
sq->tso_hdrs->iova + (sq->head * TSO_HEADER_SIZE);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.h
88
u16 head;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
219
sqe_hdr->sqe_id = sq->head;
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
436
struct list_head *head = &parent->child_list;
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
443
for (tmp = head->next; tmp != head; tmp = tmp->next) {
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
454
list_add_tail(&node->list, head);
drivers/net/ethernet/marvell/prestera/prestera_router.c
559
list_del(&fib_cache->kern_neigh_cache_head[i].head);
drivers/net/ethernet/marvell/prestera/prestera_router.c
57
struct list_head head;
drivers/net/ethernet/marvell/prestera/prestera_router.c
606
list_add(&fc->kern_neigh_cache_head[i].head,
drivers/net/ethernet/marvell/prestera/prestera_router.c
849
list_for_each_entry(nhead, &nc->kern_fib_cache_list, head) {
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
364
list_for_each_entry(nh_head, &neigh->nexthop_group_list, head) {
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
382
&nh_neigh->nexthop_group_list, head) {
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
416
list_add(&nh_grp->nh_neigh_head[nh_cnt].head,
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
446
list_del(&nh_grp->nh_neigh_head[nh_cnt].head);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
471
list_del(&nh_grp->nh_neigh_head[nh_cnt].head);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.h
82
struct list_head head;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
379
struct prestera_sdma_buf *head, *tail, *next, *prev;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
382
ring->bufs = kmalloc_objs(*head, bnum);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
389
head = &ring->bufs[0];
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
390
next = head;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
413
prestera_sdma_rx_desc_set_next(sdma, tail->desc, head->desc_dma);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
416
prestera_sdma_map(sdma, head->desc_dma));
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
524
struct prestera_sdma_buf *head, *tail, *next, *prev;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
532
tx_ring->bufs = kmalloc_objs(*head, bnum);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
537
head = &tx_ring->bufs[0];
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
538
next = head;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
562
prestera_sdma_tx_desc_set_next(sdma, tail->desc, head->desc_dma);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
568
prestera_sdma_map(sdma, head->desc_dma));
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
102
list_for_each_entry(bridge, &sw->swdev->bridge_list, head)
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
1025
list_for_each_entry(br_port, &br_dev->port_list, head) {
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
1119
list_for_each_entry(br_port, &br_dev->port_list, head)
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
115
list_for_each_entry(br_port, &bridge->port_list, head)
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
189
list_add(&br_vlan->head, &br_port->vlan_list);
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
196
list_del(&br_vlan->head);
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
206
list_for_each_entry(br_vlan, &br_port->vlan_list, head) {
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
221
list_for_each_entry(br_port, &bridge->port_list, head) {
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
222
list_for_each_entry(br_vlan, &br_port->vlan_list, head) {
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
343
list_for_each_entry(br_port, &br_mdb->bridge->port_list, head)
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
38
struct list_head head;
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
470
list_add(&bridge->head, &swdev->bridge_list);
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
479
list_del(&bridge->head);
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
50
struct list_head head;
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
503
list_for_each_entry(bridge, &swdev->bridge_list, head)
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
516
list_for_each_entry(br_port, &bridge->port_list, head) {
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
577
list_add(&br_port->head, &bridge->port_list);
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
586
list_del(&br_port->head);
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
62
struct list_head head;
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
851
list_for_each_entry(br_vlan, &br_port->vlan_list, head) {
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
864
list_for_each_entry_continue_reverse(br_vlan, &br_port->vlan_list, head)
drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
915
list_for_each_entry(br_port, &br_dev->port_list, head) {
drivers/net/ethernet/mediatek/mtk_ppe.c
503
struct hlist_head *head;
drivers/net/ethernet/mediatek/mtk_ppe.c
510
head = &entry->l2_flows;
drivers/net/ethernet/mediatek/mtk_ppe.c
511
hlist_for_each_entry_safe(entry, tmp, head, l2_data.list)
drivers/net/ethernet/mediatek/mtk_ppe.c
741
struct hlist_head *head = &ppe->foe_flow[hash / soc->hash_offset];
drivers/net/ethernet/mediatek/mtk_ppe.c
755
hlist_for_each_entry_safe(entry, n, head, list) {
drivers/net/ethernet/mediatek/mtk_star_emac.c
1164
unsigned int head = ring->head;
drivers/net/ethernet/mediatek/mtk_star_emac.c
1168
while (entry != head && count < (MTK_STAR_RING_NUM_DESCS - 1)) {
drivers/net/ethernet/mediatek/mtk_star_emac.c
243
unsigned int head;
drivers/net/ethernet/mediatek/mtk_star_emac.c
306
ring->head = 0;
drivers/net/ethernet/mediatek/mtk_star_emac.c
344
struct mtk_star_ring_desc *desc = &ring->descs[ring->head];
drivers/net/ethernet/mediatek/mtk_star_emac.c
349
ring->skbs[ring->head] = desc_data->skb;
drivers/net/ethernet/mediatek/mtk_star_emac.c
350
ring->dma_addrs[ring->head] = desc_data->dma_addr;
drivers/net/ethernet/mediatek/mtk_star_emac.c
363
ring->head = (ring->head + 1) % MTK_STAR_RING_NUM_DESCS;
drivers/net/ethernet/mediatek/mtk_star_emac.c
388
if (ring->tail > ring->head)
drivers/net/ethernet/mediatek/mtk_star_emac.c
389
avail = ring->tail - ring->head - 1;
drivers/net/ethernet/mediatek/mtk_star_emac.c
391
avail = MTK_STAR_RING_NUM_DESCS - ring->head + ring->tail - 1;
drivers/net/ethernet/mediatek/mtk_wed.c
1543
void *head = (void *)ring->desc;
drivers/net/ethernet/mediatek/mtk_wed.c
1549
desc = (struct mtk_wdma_desc *)(head + i * ring->desc_size);
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
55
memset(skb->head, 0, length);
drivers/net/ethernet/mediatek/mtk_wed_wo.c
156
q->head = (q->head + 1) % q->n_desc;
drivers/net/ethernet/mediatek/mtk_wed_wo.c
157
entry = &q->entry[q->head];
drivers/net/ethernet/mediatek/mtk_wed_wo.c
160
q->entry[q->head].buf = buf;
drivers/net/ethernet/mediatek/mtk_wed_wo.c
163
struct mtk_wed_wo_queue_desc *desc = &q->desc[q->head];
drivers/net/ethernet/mediatek/mtk_wed_wo.c
218
u32 index = (q->head - 1) % q->n_desc;
drivers/net/ethernet/mediatek/mtk_wed_wo.c
338
index = (q->head + 1) % q->n_desc;
drivers/net/ethernet/mediatek/mtk_wed_wo.c
351
q->head = index;
drivers/net/ethernet/mediatek/mtk_wed_wo.c
364
mtk_wed_wo_queue_kick(wo, q, q->head);
drivers/net/ethernet/mediatek/mtk_wed_wo.h
223
u16 head;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1466
struct mlx5_cmd_mailbox *tmp, *head = NULL;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1489
tmp->next = head;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1493
head = tmp;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1495
msg->next = head;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1499
while (head) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1500
tmp = head->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1501
free_cmd_box(dev, head);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1502
head = tmp;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1512
struct mlx5_cmd_mailbox *head = msg->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1515
while (head) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1516
next = head->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1517
free_cmd_box(dev, head);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1518
head = next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1725
list_add_tail(&msg->list, &msg->parent->head);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1910
if (list_empty(&ch->head)) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1914
msg = list_entry(ch->head.next, typeof(*msg), list);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
2344
list_for_each_entry_safe(msg, n, &ch->head, list) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
2375
INIT_LIST_HEAD(&ch->head);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
2384
list_add_tail(&msg->list, &ch->head);
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
411
struct hlist_head *head =
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
419
hlist_add_head(&cur_string->hlist, head);
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
491
static struct tracer_string_format *mlx5_tracer_message_find(struct hlist_head *head,
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
496
hlist_for_each_entry(message, head, hlist)
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
506
struct hlist_head *head =
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
509
return mlx5_tracer_message_find(head, tracer_event->event_id, tracer_event->string_event.tmsn);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c
142
(skb->csum_start + skb->head - skb->data) / 2;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.c
87
(skb->csum_start + skb->head - skb->data) / 2;
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
719
struct hlist_head *head;
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
721
head = arfs_hash_bucket(arfs_t, fk->ports.src, fk->ports.dst);
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
722
hlist_for_each_entry(arfs_rule, head, hlist) {
drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c
399
struct list_head *head = &ethtool->rules;
drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c
405
head = &iter->list;
drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c
408
list_add(&rule->list, head);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1261
u16 head;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1268
head = wq->head;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1276
rq->dealloc_wqe(rq, head);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1277
head = mlx5_wq_ll_get_wqe_next_ix(wq, head);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1280
rq->mpwqe.actual_wq_head = wq->head;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1308
u16 head = mlx5_wq_cyc_get_head(wq);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1319
wqe_ix = mlx5_wq_cyc_ctr2ix(wq, head++);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1047
skb->csum_start = (unsigned char *)uh - skb->head;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1066
skb->csum_start = (unsigned char *)uh - skb->head;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1136
skb->csum_start = (unsigned char *)tcp - skb->head;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1157
skb->csum_start = (unsigned char *)tcp - skb->head;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1917
va = skb->head;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
603
u16 next_wqe_index = mlx5_wq_ll_get_wqe_next_ix(wq, wq->head);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
712
u16 head;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
724
head = mlx5_wq_cyc_get_head(wq);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
729
wqe_bulk -= (head + wqe_bulk) & rq->wqe.info.wqe_index_mask;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
732
count = mlx5e_refill_rx_wqes(rq, head, wqe_bulk);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
734
mlx5e_xsk_free_rx_wqes(rq, head, wqe_bulk);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
735
count = mlx5e_xsk_alloc_rx_wqes_batched(rq, head, wqe_bulk);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
737
mlx5e_xsk_free_rx_wqes(rq, head, wqe_bulk);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
743
count = mlx5e_xsk_alloc_rx_wqes(rq, head, wqe_bulk);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
872
static void mlx5e_reclaim_mpwqe_pages(struct mlx5e_rq *rq, int head,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
878
head = mlx5_wq_ll_get_wqe_next_ix(wq, head);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
880
mlx5e_dealloc_rx_mpwqe(rq, head);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
892
u16 head;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
916
head = rq->mpwqe.actual_wq_head;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
919
struct mlx5e_mpw_info *wi = mlx5e_get_mpw_info(rq, head);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
925
alloc_err = rq->xsk_pool ? mlx5e_xsk_alloc_rx_mpwqe(rq, head) :
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
926
mlx5e_alloc_rx_mpwqe(rq, head);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
933
mlx5e_reclaim_mpwqe_pages(rq, head, reclaim);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
938
head = mlx5_wq_ll_get_wqe_next_ix(wq, head);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
948
rq->mpwqe.actual_wq_head = head;
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
2064
static void free_match_list(struct match_list *head, bool ft_locked)
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
2068
list_for_each_entry_safe(iter, match_tmp, &head->list,
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
984
#define list_for_each_advance_continue(pos, head, reverse) \
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
986
&pos->list != (head); \
drivers/net/ethernet/mellanox/mlx5/core/lib/crypto.c
526
struct list_head *head)
drivers/net/ethernet/mellanox/mlx5/core/lib/crypto.c
529
list_splice_init(list, head);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/table.c
210
INIT_LIST_HEAD(&tbl->default_miss.head);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/table.c
279
if (!list_empty(&tbl->default_miss.head)) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/table.c
372
if (list_empty(&dst_tbl->default_miss.head))
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/table.c
375
list_for_each_entry(src_tbl, &dst_tbl->default_miss.head, default_miss.next) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/table.c
489
list_add(&tbl->default_miss.next, &miss_tbl->default_miss.head);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/table.h
12
struct list_head head;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
273
dr_qp->sq.head = 0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
478
dr_qp->sq.wqe_head[idx] = dr_qp->sq.head++;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1421
unsigned int head;
drivers/net/ethernet/mellanox/mlx5/core/uar.c
209
struct list_head *head;
drivers/net/ethernet/mellanox/mlx5/core/uar.c
217
head = &bfregs->wc_head.list;
drivers/net/ethernet/mellanox/mlx5/core/uar.c
220
head = &bfregs->reg_head.list;
drivers/net/ethernet/mellanox/mlx5/core/uar.c
224
if (list_empty(head)) {
drivers/net/ethernet/mellanox/mlx5/core/uar.c
230
list_add(&up->list, head);
drivers/net/ethernet/mellanox/mlx5/core/uar.c
232
up = list_entry(head->next, struct mlx5_uars_page, list);
drivers/net/ethernet/mellanox/mlx5/core/uar.c
298
struct list_head *head;
drivers/net/ethernet/mellanox/mlx5/core/uar.c
302
head = &bfregs->wc_head.list;
drivers/net/ethernet/mellanox/mlx5/core/uar.c
305
head = &bfregs->reg_head.list;
drivers/net/ethernet/mellanox/mlx5/core/uar.c
322
list_add_tail(&up->list, head);
drivers/net/ethernet/mellanox/mlx5/core/wq.c
249
wq->head = 0;
drivers/net/ethernet/mellanox/mlx5/core/wq.h
297
wq->head = head_next;
drivers/net/ethernet/mellanox/mlx5/core/wq.h
317
return wq->head;
drivers/net/ethernet/mellanox/mlx5/core/wq.h
74
u16 head;
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
28
u32 head = 0, tail = 0;
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
39
ring->head, ring->tail);
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
44
head = FIELD_GET(FBNIC_QUEUE_TWQ_PTRS_HEAD_MASK, tail);
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
48
head = FIELD_GET(FBNIC_QUEUE_TWQ_PTRS_HEAD_MASK, tail);
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
51
head = readl(csr_base + FBNIC_QUEUE_TCQ_PTRS);
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
52
tail = FIELD_GET(FBNIC_QUEUE_TCQ_PTRS_TAIL_MASK, head);
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
543
mbx->ready, mbx->head, mbx->tail);
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
56
head = FIELD_GET(FBNIC_QUEUE_BDQ_PTRS_HEAD_MASK, tail);
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
60
head = FIELD_GET(FBNIC_QUEUE_BDQ_PTRS_HEAD_MASK, tail);
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
63
head = readl(csr_base + FBNIC_QUEUE_RCQ_PTRS);
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
64
tail = FIELD_GET(FBNIC_QUEUE_RCQ_PTRS_TAIL_MASK, head);
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
69
head &= FBNIC_QUEUE_RCQ_PTRS_HEAD_MASK;
drivers/net/ethernet/meta/fbnic/fbnic_debugfs.c
71
seq_printf(s, "HW: head: %#06x tail: %#06x\n", head, tail);
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
129
if (mbx->head == ((tail + 1) % FBNIC_IPC_MBX_DESC_LEN))
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1595
u8 head = rx_mbx->head, tail = rx_mbx->tail;
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1598
while (head != rx_mbx->tail) {
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1602
desc = __fbnic_mbx_rd_desc(fbd, FBNIC_IPC_MBX_RX_IDX, head);
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1606
dma_sync_single_for_cpu(fbd->dev, rx_mbx->buf_info[head].addr,
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1609
msg = rx_mbx->buf_info[head].msg;
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1640
fw_wr32(fbd, FBNIC_IPC_MBX(FBNIC_IPC_MBX_RX_IDX, head), 0);
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1642
rx_mbx->buf_info[tail] = rx_mbx->buf_info[head];
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1643
rx_mbx->buf_info[head].msg = NULL;
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1644
rx_mbx->buf_info[head].addr = 0;
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1653
head = (head + 1) & (FBNIC_IPC_MBX_DESC_LEN - 1);
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1658
rx_mbx->head = head;
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1779
u8 head = tx_mbx->head;
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
1782
if (head == tail)
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
194
u8 tail = rx_mbx->tail, head = rx_mbx->head, count;
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
204
count = (head - tail - 1) & (FBNIC_IPC_MBX_DESC_LEN - 1);
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
280
u8 head = tx_mbx->head;
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
283
while (head != tx_mbx->tail) {
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
284
desc = __fbnic_mbx_rd_desc(fbd, FBNIC_IPC_MBX_TX_IDX, head);
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
288
fbnic_mbx_unmap_and_free_msg(fbd, FBNIC_IPC_MBX_TX_IDX, head);
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
290
head++;
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
291
head %= FBNIC_IPC_MBX_DESC_LEN;
drivers/net/ethernet/meta/fbnic/fbnic_fw.c
295
tx_mbx->head = head;
drivers/net/ethernet/meta/fbnic/fbnic_fw.h
15
u8 ready, head, tail;
drivers/net/ethernet/meta/fbnic/fbnic_fw_log.c
74
struct fbnic_fw_log_entry *entry, *head, *tail, *next;
drivers/net/ethernet/meta/fbnic/fbnic_fw_log.c
90
head = list_first_entry(&log->entries, typeof(*head), list);
drivers/net/ethernet/meta/fbnic/fbnic_fw_log.c
91
entry_end = head->msg + head->len + 1;
drivers/net/ethernet/meta/fbnic/fbnic_pci.c
200
if (tx_mbx->head != tx_mbx->tail)
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
107
return (ring->head - ring->tail - 1) & ring->size_mask;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
112
return (ring->tail - ring->head) & ring->size_mask;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1238
u32 head = rcq->head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1240
done = (head & (rcq->size_mask + 1)) ? cpu_to_le64(FBNIC_RCD_DONE) : 0;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1241
raw_rcd = &rcq->desc[head & rcq->size_mask];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1316
head++;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1317
if (!(head & rcq->size_mask)) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1346
if (rcq->head != head) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1347
rcq->head = head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1348
writel(head & rcq->size_mask, rcq->doorbell);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2492
twq->head = 0;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2514
twq->head = 0;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2537
tcq->head = 0;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2563
ppq->head = 0;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2565
hpq->head = 0;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2700
rcq->head = 0;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
549
unsigned int head = ring->head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
553
clean_desc = (hw_head - head) & ring->size_mask;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
556
struct sk_buff *skb = ring->tx_buf[head];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
570
ring->tx_buf[head] = NULL;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
574
while (!(ring->desc[head] & FBNIC_TWD_TYPE(AL))) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
575
head++;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
576
head &= ring->size_mask;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
580
fbnic_unmap_single_twd(nv->dev, &ring->desc[head]);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
581
head++;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
582
head &= ring->size_mask;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
586
if (ring->tx_buf[head] != FBNIC_XMIT_NOUNMAP)
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
588
&ring->desc[head]);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
590
ring->tx_buf[head] = NULL;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
591
head++;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
592
head &= ring->size_mask;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
604
ring->head = head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
637
unsigned int head = ring->head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
639
while (hw_head != head) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
643
if (unlikely(!(ring->desc[head] & FBNIC_TWD_TYPE(AL))))
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
646
twd = le64_to_cpu(ring->desc[head]);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
647
page = ring->tx_buf[head];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
659
head++;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
660
head &= ring->size_mask;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
666
ring->head = head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
688
int head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
691
head = (*ts_head < 0) ? ring->head : *ts_head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
696
if (head == ring->tail) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
703
skb = ring->tx_buf[head];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
706
head += desc_cnt;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
707
head &= ring->size_mask;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
716
*ts_head = head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
720
head = FBNIC_XMIT_CB(skb)->hw_head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
721
if (head >= 0)
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
722
*head0 = head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
801
u32 head = cmpl->head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
803
done = (head & (cmpl->size_mask + 1)) ? 0 : cpu_to_le64(FBNIC_TCD_DONE);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
804
raw_tcd = &cmpl->desc[head & cmpl->size_mask];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
838
head++;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
839
if (!(head & cmpl->size_mask)) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
846
if (cmpl->head != head) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
847
cmpl->head = head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
848
writel(head & cmpl->size_mask, cmpl->doorbell);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
858
unsigned int head = ring->head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
860
if (head == hw_head)
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
864
fbnic_page_pool_drain(ring, head, napi_budget);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
866
head++;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
867
head &= ring->size_mask;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
868
} while (head != hw_head);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
870
ring->head = head;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.h
125
u32 head, tail; /* Head/Tail of ring */
drivers/net/ethernet/micrel/ks8842.c
473
__func__, skb->len, skb->head, skb->data,
drivers/net/ethernet/microsoft/mana/gdma_main.c
1081
queue->head = 0;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1134
queue->head = 0;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1298
u32 used_space = (wq->head - wq->tail) * GDMA_WQE_BU_SIZE;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1421
wqe_ptr = mana_gd_get_wqe_ptr(wq, wq->head);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1429
wq->head += wqe_size / GDMA_WQE_BU_SIZE;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1461
cqe = &cq_cqe[cq->head % num_cqe];
drivers/net/ethernet/microsoft/mana/gdma_main.c
1464
old_bits = (cq->head / num_cqe - 1) & GDMA_CQE_OWNER_MASK;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1469
new_bits = (cq->head / num_cqe) & GDMA_CQE_OWNER_MASK;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1495
cq->head -= cqe_idx;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1502
cq->head++;
drivers/net/ethernet/microsoft/mana/gdma_main.c
385
queue->id, queue->head * GDMA_WQE_BU_SIZE, 0);
drivers/net/ethernet/microsoft/mana/gdma_main.c
395
u32 head = cq->head % (num_cqe << GDMA_CQE_OWNER_BITS);
drivers/net/ethernet/microsoft/mana/gdma_main.c
398
head, arm_bit);
drivers/net/ethernet/microsoft/mana/gdma_main.c
563
u32 head = eq->head % (eq->queue_size / GDMA_EQE_SIZE);
drivers/net/ethernet/microsoft/mana/gdma_main.c
574
eqe = &eq_eqe_ptr[head];
drivers/net/ethernet/microsoft/mana/gdma_main.c
665
u32 head, num_eqe;
drivers/net/ethernet/microsoft/mana/gdma_main.c
675
eqe = &eq_eqe_ptr[eq->head % num_eqe];
drivers/net/ethernet/microsoft/mana/gdma_main.c
679
old_bits = (eq->head / num_eqe - 1) & GDMA_EQE_OWNER_MASK;
drivers/net/ethernet/microsoft/mana/gdma_main.c
688
new_bits = (eq->head / num_eqe) & GDMA_EQE_OWNER_MASK;
drivers/net/ethernet/microsoft/mana/gdma_main.c
701
eq->head++;
drivers/net/ethernet/microsoft/mana/gdma_main.c
704
head = eq->head % (num_eqe << GDMA_EQE_OWNER_BITS);
drivers/net/ethernet/microsoft/mana/gdma_main.c
707
head, SET_ARM_BIT);
drivers/net/ethernet/microsoft/mana/gdma_main.c
872
queue->head |= INITIALIZED_OWNER_BIT(log2_num_entries);
drivers/net/ethernet/microsoft/mana/gdma_main.c
897
queue->head |= INITIALIZED_OWNER_BIT(log2_num_entries);
drivers/net/ethernet/microsoft/mana/gdma_main.c
938
queue->head = 0;
drivers/net/ethernet/microsoft/mana/mana_en.c
1629
debugfs_create_u32("head", 0400, eq.mana_eq_debugfs, &eq.eq->head);
drivers/net/ethernet/microsoft/mana/mana_en.c
1731
used_space_old = wq->head - wq->tail;
drivers/net/ethernet/microsoft/mana/mana_en.c
1732
used_space_new = wq->head - (wq->tail + num_units);
drivers/net/ethernet/microsoft/mana/mana_en.c
1743
struct mana_skb_head *ash = (struct mana_skb_head *)skb->head;
drivers/net/ethernet/microsoft/mana/mana_en.c
187
struct mana_skb_head *ash = (struct mana_skb_head *)skb->head;
drivers/net/ethernet/microsoft/mana/mana_en.c
2316
&tx_qp->txq.gdma_sq->head);
drivers/net/ethernet/microsoft/mana/mana_en.c
2322
&tx_qp->tx_cq.gdma_cq->head);
drivers/net/ethernet/microsoft/mana/mana_en.c
2777
debugfs_create_u32("rq_head", 0400, rxq->mana_rx_debugfs, &rxq->gdma_rq->head);
drivers/net/ethernet/microsoft/mana/mana_en.c
2781
&rxq->rx_cq.gdma_cq->head);
drivers/net/ethernet/netronome/nfp/crypto/tls.c
359
if (!WARN_ON_ONCE((u8 *)back < skb->head ||
drivers/net/ethernet/netronome/nfp/flower/action.c
109
pop_vlan->head.jump_id = NFP_FL_ACTION_OPCODE_POP_VLAN;
drivers/net/ethernet/netronome/nfp/flower/action.c
110
pop_vlan->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
121
push_vlan->head.jump_id = NFP_FL_ACTION_OPCODE_PUSH_VLAN;
drivers/net/ethernet/netronome/nfp/flower/action.c
122
push_vlan->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
163
pre_lag->head.jump_id = NFP_FL_ACTION_OPCODE_PRE_LAG;
drivers/net/ethernet/netronome/nfp/flower/action.c
164
pre_lag->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
184
output->head.jump_id = NFP_FL_ACTION_OPCODE_OUTPUT;
drivers/net/ethernet/netronome/nfp/flower/action.c
185
output->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
323
pre_tun_act->head.jump_id = NFP_FL_ACTION_OPCODE_PRE_TUNNEL;
drivers/net/ethernet/netronome/nfp/flower/action.c
324
pre_tun_act->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
385
push->head.jump_id = NFP_FL_ACTION_OPCODE_PUSH_GENEVE;
drivers/net/ethernet/netronome/nfp/flower/action.c
386
push->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
39
push_mpls->head.jump_id = NFP_FL_ACTION_OPCODE_PUSH_MPLS;
drivers/net/ethernet/netronome/nfp/flower/action.c
40
push_mpls->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
451
set_tun->head.jump_id = NFP_FL_ACTION_OPCODE_SET_TUNNEL;
drivers/net/ethernet/netronome/nfp/flower/action.c
452
set_tun->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
557
set_eth->head.jump_id = NFP_FL_ACTION_OPCODE_SET_ETHERNET;
drivers/net/ethernet/netronome/nfp/flower/action.c
558
set_eth->head.len_lw = sizeof(*set_eth) >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
595
set_ip_addr->head.jump_id = NFP_FL_ACTION_OPCODE_SET_IPV4_ADDRS;
drivers/net/ethernet/netronome/nfp/flower/action.c
596
set_ip_addr->head.len_lw = sizeof(*set_ip_addr) >>
drivers/net/ethernet/netronome/nfp/flower/action.c
603
set_ip_addr->head.jump_id = NFP_FL_ACTION_OPCODE_SET_IPV4_ADDRS;
drivers/net/ethernet/netronome/nfp/flower/action.c
604
set_ip_addr->head.len_lw = sizeof(*set_ip_addr) >>
drivers/net/ethernet/netronome/nfp/flower/action.c
619
set_ip_ttl_tos->head.jump_id =
drivers/net/ethernet/netronome/nfp/flower/action.c
621
set_ip_ttl_tos->head.len_lw = sizeof(*set_ip_ttl_tos) >>
drivers/net/ethernet/netronome/nfp/flower/action.c
637
set_ip_ttl_tos->head.jump_id =
drivers/net/ethernet/netronome/nfp/flower/action.c
639
set_ip_ttl_tos->head.len_lw = sizeof(*set_ip_ttl_tos) >>
drivers/net/ethernet/netronome/nfp/flower/action.c
659
ip6->head.jump_id = opcode_tag;
drivers/net/ethernet/netronome/nfp/flower/action.c
660
ip6->head.len_lw = sizeof(*ip6) >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
69
pop_mpls->head.jump_id = NFP_FL_ACTION_OPCODE_POP_MPLS;
drivers/net/ethernet/netronome/nfp/flower/action.c
70
pop_mpls->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
705
ip_hl_fl->head.jump_id = NFP_FL_ACTION_OPCODE_SET_IPV6_TC_HL_FL;
drivers/net/ethernet/netronome/nfp/flower/action.c
706
ip_hl_fl->head.len_lw = sizeof(*ip_hl_fl) >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
775
set_tport->head.jump_id = opcode;
drivers/net/ethernet/netronome/nfp/flower/action.c
776
set_tport->head.len_lw = sizeof(*set_tport) >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
81
set_mpls->head.jump_id = NFP_FL_ACTION_OPCODE_SET_MPLS;
drivers/net/ethernet/netronome/nfp/flower/action.c
82
set_mpls->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/action.c
823
if (set_act->set_eth.head.len_lw) {
drivers/net/ethernet/netronome/nfp/flower/action.c
829
if (set_act->set_ip_ttl_tos.head.len_lw) {
drivers/net/ethernet/netronome/nfp/flower/action.c
840
if (set_act->set_ip_addr.head.len_lw) {
drivers/net/ethernet/netronome/nfp/flower/action.c
851
if (set_act->set_ip6_tc_hl_fl.head.len_lw) {
drivers/net/ethernet/netronome/nfp/flower/action.c
861
if (set_act->set_ip6_dst.head.len_lw &&
drivers/net/ethernet/netronome/nfp/flower/action.c
862
set_act->set_ip6_src.head.len_lw) {
drivers/net/ethernet/netronome/nfp/flower/action.c
878
} else if (set_act->set_ip6_dst.head.len_lw) {
drivers/net/ethernet/netronome/nfp/flower/action.c
886
} else if (set_act->set_ip6_src.head.len_lw) {
drivers/net/ethernet/netronome/nfp/flower/action.c
895
if (set_act->set_tport.head.len_lw) {
drivers/net/ethernet/netronome/nfp/flower/action.c
951
meter_act->head.jump_id = NFP_FL_ACTION_OPCODE_METER;
drivers/net/ethernet/netronome/nfp/flower/action.c
952
meter_act->head.len_lw = act_size >> NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
131
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
138
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
147
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
156
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
167
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
176
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
183
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
189
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
196
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
201
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
210
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
221
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
236
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
245
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
251
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
256
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/cmsg.h
263
struct nfp_fl_act_head head;
drivers/net/ethernet/netronome/nfp/flower/metadata.c
146
if (CIRC_SPACE(ring->head, ring->tail, NFP_FLOWER_MASK_ENTRY_RS) == 0)
drivers/net/ethernet/netronome/nfp/flower/metadata.c
152
memcpy(&ring->buf[ring->head * NFP_FLOWER_MASK_ELEMENT_RS], &mask_id,
drivers/net/ethernet/netronome/nfp/flower/metadata.c
154
ring->head = (ring->head + 1) & (NFP_FLOWER_MASK_ENTRY_RS - 1);
drivers/net/ethernet/netronome/nfp/flower/metadata.c
178
if (ring->head == ring->tail)
drivers/net/ethernet/netronome/nfp/flower/metadata.c
47
if (!CIRC_SPACE(ring->head, ring->tail, priv->stats_ring_size))
drivers/net/ethernet/netronome/nfp/flower/metadata.c
51
memcpy(&ring->buf[ring->head * NFP_FL_STATS_ELEM_RS],
drivers/net/ethernet/netronome/nfp/flower/metadata.c
53
ring->head = (ring->head + 1) & (priv->stats_ring_size - 1);
drivers/net/ethernet/netronome/nfp/flower/metadata.c
83
if (ring->head == ring->tail) {
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
100
config->head.flags_opts |= cpu_to_be32(NFP_FL_QOS_PPS);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
102
config->head.flags_opts |= cpu_to_be32(NFP_FL_QOS_METER);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
105
config->head.port = cpu_to_be32(id);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
107
config->head.meter_id = cpu_to_be32(id);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
337
config->head.flags_opts = cpu_to_be32(NFP_FL_QOS_PPS);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
338
config->head.port = cpu_to_be32(netdev_port_id);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
357
if (be32_to_cpu(msg->head.flags_opts) & NFP_FL_QOS_METER)
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
360
netdev_port_id = be32_to_cpu(msg->head.port);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
393
struct nfp_police_cfg_head *head;
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
402
head = nfp_flower_cmsg_get_data(skb);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
404
memset(head, 0, sizeof(struct nfp_police_cfg_head));
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
406
head->port = cpu_to_be32(id);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
408
head->flags_opts = cpu_to_be32(NFP_FL_QOS_METER);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
409
head->meter_id = cpu_to_be32(id);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
69
struct nfp_police_cfg_head head;
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
784
config->head.flags_opts = cpu_to_be32(NFP_FL_QOS_METER);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
785
config->head.meter_id = cpu_to_be32(meter_id);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
787
config->head.flags_opts |= cpu_to_be32(NFP_FL_QOS_PPS);
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
79
struct nfp_police_cfg_head head;
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
803
meter_id = be32_to_cpu(msg->head.meter_id);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
827
if (skb && rxbuf && skb->head == rxbuf->frag)
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
220
skb->head, skb->data);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
827
if (skb && rxbuf && skb->head == rxbuf->frag)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
107
static void __resource_add(struct list_head *head, struct nfp_cpp_resource *res)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
112
list_for_each(pos, head) {
drivers/net/ethernet/pensando/ionic/ionic_debugfs.c
233
struct hlist_head *head;
drivers/net/ethernet/pensando/ionic/ionic_debugfs.c
240
head = &lif->rx_filters.by_id[i];
drivers/net/ethernet/pensando/ionic/ionic_debugfs.c
241
hlist_for_each_entry_safe(f, tmp, head, by_id) {
drivers/net/ethernet/pensando/ionic/ionic_dev.c
1071
unsigned int mask, tail, head;
drivers/net/ethernet/pensando/ionic/ionic_dev.c
1075
head = q->head_idx;
drivers/net/ethernet/pensando/ionic/ionic_dev.c
1077
return ((pos - tail) & mask) < ((head - tail) & mask);
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
110
struct hlist_head *head;
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
116
head = &lif->rx_filters.by_id[i];
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
117
hlist_for_each_entry_safe(f, tmp, head, by_id)
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
130
struct hlist_head *head;
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
175
head = &lif->rx_filters.by_hash[key];
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
176
hlist_add_head(&f->by_hash, head);
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
179
head = &lif->rx_filters.by_id[key];
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
180
hlist_add_head(&f->by_id, head);
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
188
struct hlist_head *head;
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
192
head = &lif->rx_filters.by_hash[key];
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
194
hlist_for_each_entry(f, head, by_hash) {
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
208
struct hlist_head *head;
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
212
head = &lif->rx_filters.by_hash[key];
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
214
hlist_for_each_entry(f, head, by_hash) {
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
227
struct hlist_head *head;
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
231
head = &lif->rx_filters.by_hash[key];
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
233
hlist_for_each_entry(f, head, by_hash) {
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
28
struct hlist_head *head;
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
38
head = &lif->rx_filters.by_id[i];
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
39
hlist_for_each_entry_safe(f, tmp, head, by_id) {
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
565
struct hlist_head *head;
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
580
head = &lif->rx_filters.by_id[i];
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
581
hlist_for_each_entry_safe(f, tmp, head, by_id) {
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
85
head = &lif->rx_filters.by_id[key];
drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c
86
hlist_add_head(&f->by_id, head);
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
624
struct list_head *head;
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
628
list_for_each(head, del_list) {
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
629
cur = list_entry(head, nx_mac_list_t, list);
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
632
list_move_tail(head, &adapter->mac_list);
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
656
struct list_head *head;
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
685
head = &del_list;
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
686
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
687
cur = list_entry(head->next, nx_mac_list_t, list);
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
718
struct list_head *head = &adapter->mac_list;
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
720
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
721
cur = list_entry(head->next, nx_mac_list_t, list);
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1639
#define netxen_merge_rx_buffers(list, head) \
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1640
do { list_splice_tail_init(list, head); } while (0);
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1812
struct list_head *head;
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1816
head = &rds_ring->free_list;
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1817
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1819
buffer = list_entry(head->next, struct netxen_rx_buffer, list);
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1869
struct list_head *head;
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1876
head = &rds_ring->free_list;
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1877
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1879
buffer = list_entry(head->next, struct netxen_rx_buffer, list);
drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
3185
struct list_head *head;
drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
3194
list_for_each(head, &adapter->ip_list) {
drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
3195
cur = list_entry(head, struct nx_ip_list, list);
drivers/net/ethernet/qlogic/qed/qed_spq.c
754
struct list_head *head, u32 keep_reserve)
drivers/net/ethernet/qlogic/qed/qed_spq.c
760
!list_empty(head)) {
drivers/net/ethernet/qlogic/qed/qed_spq.c
762
list_first_entry(head, struct qed_spq_entry, list);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1269
qede_get_arfs_fltr_by_loc(struct hlist_head *head, u64 location)
drivers/net/ethernet/qlogic/qede/qede_filter.c
1273
hlist_for_each_entry(fltr, head, node)
drivers/net/ethernet/qlogic/qede/qede_filter.c
1284
struct hlist_head *head;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1296
head = QEDE_ARFS_BUCKET_HEAD(edev, 0);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1298
hlist_for_each_entry(fltr, head, node) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1589
struct hlist_head *head;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1591
head = QEDE_ARFS_BUCKET_HEAD(edev, 0);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1593
hlist_for_each_entry_safe(fltr, temp, head, node) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
237
struct hlist_head *head;
drivers/net/ethernet/qlogic/qede/qede_filter.c
240
head = &edev->arfs->arfs_hl_head[i];
drivers/net/ethernet/qlogic/qede/qede_filter.c
242
hlist_for_each_entry_safe(fltr, temp, head, node) {
drivers/net/ethernet/qlogic/qede/qede_rdma.c
51
struct list_head *head = &edev->rdma_info.rdma_event_list;
drivers/net/ethernet/qlogic/qede/qede_rdma.c
55
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/qede/qede_rdma.c
56
event_node = list_entry(head->next, struct qede_rdma_event_work,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
3897
struct list_head *head = &mbx->cmd_q;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
3902
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
3903
cmd = list_entry(head->next, struct qlcnic_cmd_args, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
4104
struct list_head *head = &mbx->cmd_q;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
4123
if (list_empty(head)) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
4127
cmd = list_entry(head->next, struct qlcnic_cmd_args, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
514
struct list_head *head, *tmp;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
516
list_for_each_safe(head, tmp, &adapter->mac_list) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
517
cur = list_entry(head, struct qlcnic_mac_vlan_list, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
617
struct list_head *head = &adapter->mac_list;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
620
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
621
cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
633
struct hlist_head *head;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
639
head = &(adapter->fhash.fhead[i]);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
640
hlist_for_each_entry_safe(tmp_fil, n, head, fnode) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
658
head = &(adapter->rx_fhash.fhead[i]);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
660
hlist_for_each_entry_safe(tmp_fil, n, head, fnode)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
678
struct hlist_head *head;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
683
head = &(adapter->fhash.fhead[i]);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
684
hlist_for_each_entry_safe(tmp_fil, n, head, fnode) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
1444
struct list_head *head;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
1447
head = &rds_ring->free_list;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
1449
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
1451
buffer = list_entry(head->next, struct qlcnic_rx_buffer, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
176
static struct qlcnic_filter *qlcnic_find_mac_filter(struct hlist_head *head,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
182
hlist_for_each_entry_safe(tmp_fil, n, head, fnode) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
196
struct hlist_head *head;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
213
head = &(adapter->rx_fhash.fhead[hindex]);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
215
tmp_fil = qlcnic_find_mac_filter(head, &src_addr, vlan_id);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
231
hlist_add_head(&(fil->fnode), head);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
235
head = &adapter->fhash.fhead[hindex];
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
239
tmp_fil = qlcnic_find_mac_filter(head, &src_addr, vlan_id);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
257
head = &adapter->rx_fhash.fhead[hindex];
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
261
tmp_fil = qlcnic_find_mac_filter(head, &src_addr, vlan_id);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
310
struct hlist_head *head;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
331
head = &(adapter->fhash.fhead[hindex]);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
333
hlist_for_each_entry_safe(tmp_fil, n, head, fnode) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
358
hlist_add_head(&(fil->fnode), head);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
851
struct list_head *head;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
857
head = &rds_ring->free_list;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
858
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
859
buffer = list_entry(head->next, struct qlcnic_rx_buffer, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c
2824
void *head;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c
2848
head = kzalloc_objs(struct hlist_head, adapter->fhash.fbucket_size,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c
2851
if (!head)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c
2855
adapter->fhash.fhead = head;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c
2865
head = kzalloc_objs(struct hlist_head, adapter->rx_fhash.fbucket_size,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c
2868
if (!head)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c
2872
adapter->rx_fhash.fhead = head;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
1530
struct list_head *head = &bc->async_cmd_list;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
1537
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
1538
entry = list_entry(head->next, struct qlcnic_async_cmd,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
1607
struct list_head *head;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
1611
head = &bc->async_cmd_list;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
1614
list_splice_init(head, &del_list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
1624
if (!list_empty(head))
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
2105
struct list_head *head = &adapter->mac_list;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
2108
while (!list_empty(head)) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
2109
cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list);
drivers/net/ethernet/qualcomm/qca_debug.c
69
if (qca->txr.skb[qca->txr.head] == NULL)
drivers/net/ethernet/qualcomm/qca_spi.c
276
if (qca->txr.skb[qca->txr.head] == NULL)
drivers/net/ethernet/qualcomm/qca_spi.c
289
while (qca->txr.skb[qca->txr.head]) {
drivers/net/ethernet/qualcomm/qca_spi.c
290
pkt_len = qca->txr.skb[qca->txr.head]->len + QCASPI_HW_PKT_LEN;
drivers/net/ethernet/qualcomm/qca_spi.c
298
if (qcaspi_tx_frame(qca, qca->txr.skb[qca->txr.head]) == -1) {
drivers/net/ethernet/qualcomm/qca_spi.c
305
n_stats->tx_bytes += qca->txr.skb[qca->txr.head]->len;
drivers/net/ethernet/qualcomm/qca_spi.c
313
dev_kfree_skb(qca->txr.skb[qca->txr.head]);
drivers/net/ethernet/qualcomm/qca_spi.c
314
qca->txr.skb[qca->txr.head] = NULL;
drivers/net/ethernet/qualcomm/qca_spi.c
316
new_head = qca->txr.head + 1;
drivers/net/ethernet/qualcomm/qca_spi.c
319
qca->txr.head = new_head;
drivers/net/ethernet/qualcomm/qca_spi.c
477
qca->txr.head = 0;
drivers/net/ethernet/qualcomm/qca_spi.c
591
!qca->txr.skb[qca->txr.head])
drivers/net/ethernet/qualcomm/qca_spi.c
598
qca->txr.skb[qca->txr.head]);
drivers/net/ethernet/qualcomm/qca_spi.h
46
u16 head;
drivers/net/ethernet/qualcomm/rmnet/rmnet_config.c
193
static void rmnet_dellink(struct net_device *dev, struct list_head *head)
drivers/net/ethernet/qualcomm/rmnet/rmnet_config.c
222
unregister_netdevice_queue(dev, head);
drivers/net/ethernet/rocker/rocker.h
30
u32 head;
drivers/net/ethernet/rocker/rocker_main.c
370
u32 head = __pos_inc(info->head, info->size);
drivers/net/ethernet/rocker/rocker_main.c
372
desc_info = &info->desc_info[info->head];
drivers/net/ethernet/rocker/rocker_main.c
373
if (head == info->tail)
drivers/net/ethernet/rocker/rocker_main.c
389
u32 head = __pos_inc(info->head, info->size);
drivers/net/ethernet/rocker/rocker_main.c
391
BUG_ON(head == info->tail);
drivers/net/ethernet/rocker/rocker_main.c
393
info->head = head;
drivers/net/ethernet/rocker/rocker_main.c
394
rocker_write32(rocker, DMA_DESC_HEAD(info->type), head);
drivers/net/ethernet/rocker/rocker_main.c
402
if (info->tail == info->head)
drivers/net/ethernet/rocker/rocker_main.c
436
info->head = 0;
drivers/net/ethernet/rocker/rocker_main.c
477
BUG_ON(info->head || info->tail);
drivers/net/ethernet/rocker/rocker_tlv.c
19
const struct rocker_tlv *head = (const struct rocker_tlv *) buf;
drivers/net/ethernet/rocker/rocker_tlv.c
24
rocker_tlv_for_each(tlv, head, buf_len, rem) {
drivers/net/ethernet/rocker/rocker_tlv.h
45
#define rocker_tlv_for_each(pos, head, len, rem) \
drivers/net/ethernet/rocker/rocker_tlv.h
46
for (pos = head, rem = len; \
drivers/net/ethernet/sfc/ef100_rep.c
397
struct list_head head;
drivers/net/ethernet/sfc/ef100_rep.c
402
INIT_LIST_HEAD(&head);
drivers/net/ethernet/sfc/ef100_rep.c
409
list_add_tail(&skb->list, &head);
drivers/net/ethernet/sfc/ef100_rep.c
414
netif_receive_skb_list(&head);
drivers/net/ethernet/sfc/rx_common.c
678
struct hlist_head *head;
drivers/net/ethernet/sfc/rx_common.c
681
head = efx_rps_hash_bucket(efx, spec);
drivers/net/ethernet/sfc/rx_common.c
682
if (!head)
drivers/net/ethernet/sfc/rx_common.c
684
hlist_for_each(node, head) {
drivers/net/ethernet/sfc/rx_common.c
697
struct hlist_head *head;
drivers/net/ethernet/sfc/rx_common.c
700
head = efx_rps_hash_bucket(efx, spec);
drivers/net/ethernet/sfc/rx_common.c
701
if (!head)
drivers/net/ethernet/sfc/rx_common.c
703
hlist_for_each(node, head) {
drivers/net/ethernet/sfc/rx_common.c
714
hlist_add_head(&rule->node, head);
drivers/net/ethernet/sfc/rx_common.c
722
struct hlist_head *head;
drivers/net/ethernet/sfc/rx_common.c
725
head = efx_rps_hash_bucket(efx, spec);
drivers/net/ethernet/sfc/rx_common.c
726
if (WARN_ON(!head))
drivers/net/ethernet/sfc/rx_common.c
728
hlist_for_each(node, head) {
drivers/net/ethernet/sfc/siena/rx_common.c
665
struct hlist_head *head;
drivers/net/ethernet/sfc/siena/rx_common.c
668
head = efx_rps_hash_bucket(efx, spec);
drivers/net/ethernet/sfc/siena/rx_common.c
669
if (!head)
drivers/net/ethernet/sfc/siena/rx_common.c
671
hlist_for_each(node, head) {
drivers/net/ethernet/sfc/siena/rx_common.c
684
struct hlist_head *head;
drivers/net/ethernet/sfc/siena/rx_common.c
687
head = efx_rps_hash_bucket(efx, spec);
drivers/net/ethernet/sfc/siena/rx_common.c
688
if (!head)
drivers/net/ethernet/sfc/siena/rx_common.c
690
hlist_for_each(node, head) {
drivers/net/ethernet/sfc/siena/rx_common.c
701
hlist_add_head(&rule->node, head);
drivers/net/ethernet/sfc/siena/rx_common.c
710
struct hlist_head *head;
drivers/net/ethernet/sfc/siena/rx_common.c
713
head = efx_rps_hash_bucket(efx, spec);
drivers/net/ethernet/sfc/siena/rx_common.c
714
if (WARN_ON(!head))
drivers/net/ethernet/sfc/siena/rx_common.c
716
hlist_for_each(node, head) {
drivers/net/ethernet/sgi/meth.c
234
priv->rx_ring[i]=(rx_packet*)(priv->rx_skbs[i]->head);
drivers/net/ethernet/sgi/meth.c
452
priv->rx_ring[priv->rx_write] = (rx_packet*)skb->head;
drivers/net/ethernet/socionext/netsec.c
1108
if (dring->head >= dring->tail)
drivers/net/ethernet/socionext/netsec.c
1109
used = dring->head - dring->tail;
drivers/net/ethernet/socionext/netsec.c
1111
used = dring->head + DESC_NUM - dring->tail;
drivers/net/ethernet/socionext/netsec.c
1236
dring->head = 0;
drivers/net/ethernet/socionext/netsec.c
284
u16 head, tail;
drivers/net/ethernet/socionext/netsec.c
794
int idx = dring->head;
drivers/net/ethernet/socionext/netsec.c
824
dring->head = (dring->head + 1) % DESC_NUM;
drivers/net/ethernet/socionext/netsec.c
839
if (tx_ring->head >= tx_ring->tail)
drivers/net/ethernet/socionext/netsec.c
840
filled = tx_ring->head - tx_ring->tail;
drivers/net/ethernet/socionext/netsec.c
842
filled = tx_ring->head + DESC_NUM - tx_ring->tail;
drivers/net/ethernet/spacemit/k1_emac.c
285
if (tx_ring->tail > tx_ring->head)
drivers/net/ethernet/spacemit/k1_emac.c
286
avail = tx_ring->tail - tx_ring->head - 1;
drivers/net/ethernet/spacemit/k1_emac.c
288
avail = tx_ring->total_cnt - tx_ring->head + tx_ring->tail - 1;
drivers/net/ethernet/spacemit/k1_emac.c
360
tx_ring->head = 0;
drivers/net/ethernet/spacemit/k1_emac.c
386
rx_ring->head = 0;
drivers/net/ethernet/spacemit/k1_emac.c
411
tx_ring->head = 0;
drivers/net/ethernet/spacemit/k1_emac.c
439
rx_ring->head = 0;
drivers/net/ethernet/spacemit/k1_emac.c
488
while (i != tx_ring->head) {
drivers/net/ethernet/spacemit/k1_emac.c
551
i = rx_ring->head;
drivers/net/ethernet/spacemit/k1_emac.c
593
rx_ring->head = i;
drivers/net/ethernet/spacemit/k1_emac.c
736
u32 head, old_head, frag_num, f, i;
drivers/net/ethernet/spacemit/k1_emac.c
740
head = tx_ring->head;
drivers/net/ethernet/spacemit/k1_emac.c
741
old_head = head;
drivers/net/ethernet/spacemit/k1_emac.c
751
tx_buf = &tx_ring->tx_desc_buf[head];
drivers/net/ethernet/spacemit/k1_emac.c
753
&((struct emac_desc *)tx_ring->desc_addr)[head];
drivers/net/ethernet/spacemit/k1_emac.c
761
if (head != old_head)
drivers/net/ethernet/spacemit/k1_emac.c
764
if (++head == tx_ring->total_cnt) {
drivers/net/ethernet/spacemit/k1_emac.c
767
head = 0;
drivers/net/ethernet/spacemit/k1_emac.c
798
tx_ring->head = head;
drivers/net/ethernet/spacemit/k1_emac.c
806
while (i != head) {
drivers/net/ethernet/spacemit/k1_emac.c
95
u32 head;
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
413
static void dwmac4_display_ring(void *head, unsigned int size, bool rx,
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
422
struct dma_desc *p = (struct dma_desc *)head;
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
433
struct dma_extended_desc *extp = (struct dma_extended_desc *)head;
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
446
struct dma_edesc *ep = (struct dma_edesc *)head;
drivers/net/ethernet/stmicro/stmmac/enh_desc.c
412
static void enh_desc_display_ring(void *head, unsigned int size, bool rx,
drivers/net/ethernet/stmicro/stmmac/enh_desc.c
415
struct dma_extended_desc *ep = (struct dma_extended_desc *)head;
drivers/net/ethernet/stmicro/stmmac/hwif.h
89
void (*display_ring)(void *head, unsigned int size, bool rx,
drivers/net/ethernet/stmicro/stmmac/norm_desc.c
258
static void ndesc_display_ring(void *head, unsigned int size, bool rx,
drivers/net/ethernet/stmicro/stmmac/norm_desc.c
261
struct dma_desc *p = (struct dma_desc *)head;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6430
static void sysfs_display_ring(void *head, int size, int extend_desc,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6433
struct dma_extended_desc *ep = (struct dma_extended_desc *)head;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6434
struct dma_desc *p = (struct dma_desc *)head;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
203
skb->csum_start = skb_transport_header(skb) - skb->head;
drivers/net/ethernet/sun/sunvnet_common.c
1301
curr->csum_start = skb_transport_header(curr) - curr->head;
drivers/net/ethernet/ti/davinci_cpdma.c
109
struct cpdma_desc __iomem *head, *tail;
drivers/net/ethernet/ti/davinci_cpdma.c
1231
desc = chan->head;
drivers/net/ethernet/ti/davinci_cpdma.c
1253
chan->head = desc_from_phys(pool, desc_read(desc, hw_next));
drivers/net/ethernet/ti/davinci_cpdma.c
1258
if ((status & CPDMA_DESC_EOQ) && chan->head) {
drivers/net/ethernet/ti/davinci_cpdma.c
1260
chan_write(chan, hdp, desc_phys(pool, chan->head));
drivers/net/ethernet/ti/davinci_cpdma.c
1354
while (chan->head) {
drivers/net/ethernet/ti/davinci_cpdma.c
1355
struct cpdma_desc __iomem *desc = chan->head;
drivers/net/ethernet/ti/davinci_cpdma.c
1359
chan->head = desc_from_phys(pool, next_dma);
drivers/net/ethernet/ti/davinci_cpdma.c
392
if (chan->head) {
drivers/net/ethernet/ti/davinci_cpdma.c
393
chan_write(chan, hdp, desc_phys(pool, chan->head));
drivers/net/ethernet/ti/davinci_cpdma.c
992
if (!chan->head) {
drivers/net/ethernet/ti/davinci_cpdma.c
994
chan->head = desc;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1082
chain->head = descr->next;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
153
if (gelic_descr_get_status(card->rx_chain.head) !=
drivers/net/ethernet/toshiba/ps3_gelic_net.c
156
be32_to_cpu(card->rx_chain.head->hw_regs.dmac_cmd_status));
drivers/net/ethernet/toshiba/ps3_gelic_net.c
158
be32_to_cpu(card->rx_chain.head->hw_regs.next_descr_addr));
drivers/net/ethernet/toshiba/ps3_gelic_net.c
160
card->rx_chain.head);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
164
card->rx_chain.head->link.cpu_addr, 0);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1776
card->tx_top = card->tx_chain.head;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1777
card->rx_top = card->rx_chain.head;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1812
gelic_card_free_chain(card, card->rx_chain.head);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1814
gelic_card_free_chain(card, card->tx_chain.head);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
232
chain->head = start_descr;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
354
chain->head = start_descr;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
438
struct gelic_descr *descr = card->rx_chain.head;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
454
} while (descr != card->rx_chain.head);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
467
struct gelic_descr *descr = card->rx_chain.head;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
477
} while (descr != card->rx_chain.head);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
562
tx_chain->head != tx_chain->tail && tx_chain->tail;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
694
if (!card->tx_chain.head)
drivers/net/ethernet/toshiba/ps3_gelic_net.c
697
if (card->tx_chain.tail != card->tx_chain.head->next &&
drivers/net/ethernet/toshiba/ps3_gelic_net.c
698
gelic_descr_get_status(card->tx_chain.head) ==
drivers/net/ethernet/toshiba/ps3_gelic_net.c
700
return card->tx_chain.head;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
815
card->tx_chain.head = descr->next;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
906
card->tx_chain.head = descr;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
989
struct gelic_descr *descr = chain->head;
drivers/net/ethernet/toshiba/ps3_gelic_net.h
259
struct gelic_descr *head;
drivers/net/ethernet/toshiba/tc35815.c
1784
int head = (lp->tfd_start + TX_FD_NUM - 1) % TX_FD_NUM;
drivers/net/ethernet/toshiba/tc35815.c
1785
struct TxFD *txhead = &lp->tfd_base[head];
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2067
static void wx_add_ring(struct wx_ring *ring, struct wx_ring_container *head)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2069
ring->next = head->ring;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2070
head->ring = ring;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2071
head->count++;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
779
u32 head = *tx_ring->headwb_mem;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
781
if (head == tx_ring->next_to_clean)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
783
else if (head > tx_ring->next_to_clean &&
drivers/net/ethernet/wangxun/libwx/wx_lib.c
785
tx_buffer->next_eop < head))
drivers/net/ethernet/wangxun/libwx/wx_lib.c
788
tx_buffer->next_eop < head))
drivers/net/fjes/fjes_hw.c
157
info->v1i.head = 0;
drivers/net/fjes/fjes_hw.c
931
return EP_RING_EMPTY(info->v1i.head, info->v1i.tail,
drivers/net/fjes/fjes_hw.c
943
(info->v1i.head,
drivers/net/fjes/fjes_hw.c
961
EP_RING_INDEX_INC(epbh->info->v1i.head, info->v1i.count_max);
drivers/net/fjes/fjes_hw.c
970
if (EP_RING_FULL(info->v1i.head, info->v1i.tail, info->v1i.count_max))
drivers/net/fjes/fjes_hw.h
218
u32 head;
drivers/net/fjes/fjes_main.c
1109
if (EP_RING_FULL(info->v1i.head, info->v1i.tail,
drivers/net/geneve.c
2247
static void geneve_dellink(struct net_device *dev, struct list_head *head)
drivers/net/geneve.c
2252
unregister_netdevice_queue(dev, head);
drivers/net/geneve.c
324
nh = skb_network_header(skb) - skb->head;
drivers/net/geneve.c
335
oiph = skb->head + nh;
drivers/net/geneve.c
869
struct list_head *head,
drivers/net/geneve.c
917
list_for_each_entry(p, head, list) {
drivers/net/geneve.c
931
return call_gro_receive(eth_gro_receive, head, skb);
drivers/net/geneve.c
937
pp = call_gro_receive(ptype->callbacks.gro_receive, head, skb);
drivers/net/gtp.c
149
struct hlist_head *head;
drivers/net/gtp.c
152
head = &gtp->tid_hash[gtp0_hashfn(tid) % gtp->hash_size];
drivers/net/gtp.c
154
hlist_for_each_entry_rcu(pdp, head, hlist_tid) {
drivers/net/gtp.c
1550
static void gtp_dellink(struct net_device *dev, struct list_head *head)
drivers/net/gtp.c
1562
unregister_netdevice_queue(dev, head);
drivers/net/gtp.c
166
struct hlist_head *head;
drivers/net/gtp.c
169
head = &gtp->tid_hash[gtp1u_hashfn(tid) % gtp->hash_size];
drivers/net/gtp.c
171
hlist_for_each_entry_rcu(pdp, head, hlist_tid) {
drivers/net/gtp.c
183
struct hlist_head *head;
drivers/net/gtp.c
186
head = &gtp->addr_hash[ipv4_hashfn(ms_addr) % gtp->hash_size];
drivers/net/gtp.c
188
hlist_for_each_entry_rcu(pdp, head, hlist_addr) {
drivers/net/gtp.c
1988
static void pdp_context_free(struct rcu_head *head)
drivers/net/gtp.c
1990
struct pdp_ctx *pctx = container_of(head, struct pdp_ctx, rcu_head);
drivers/net/gtp.c
218
struct hlist_head *head;
drivers/net/gtp.c
221
head = &gtp->addr_hash[ipv6_hashfn(ms_addr) % gtp->hash_size];
drivers/net/gtp.c
223
hlist_for_each_entry_rcu(pdp, head, hlist_addr) {
drivers/net/hyperv/netvsc.c
150
static void free_netvsc_device(struct rcu_head *head)
drivers/net/hyperv/netvsc.c
153
= container_of(head, struct netvsc_device, rcu);
drivers/net/hyperv/netvsc_drv.c
505
rndis_msg = (struct rndis_message *)skb->head;
drivers/net/ipvlan/ipvlan.h
171
void ipvlan_link_delete(struct net_device *dev, struct list_head *head);
drivers/net/ipvlan/ipvlan_main.c
656
void ipvlan_link_delete(struct net_device *dev, struct list_head *head)
drivers/net/ipvlan/ipvlan_main.c
671
unregister_netdevice_queue(dev, head);
drivers/net/ipvlan/ipvtap.c
112
struct list_head *head)
drivers/net/ipvlan/ipvtap.c
118
ipvlan_link_delete(dev, head);
drivers/net/macsec.c
158
static void free_rx_sc_rcu(struct rcu_head *head)
drivers/net/macsec.c
160
struct macsec_rx_sc *rx_sc = container_of(head, struct macsec_rx_sc, rcu_head);
drivers/net/macsec.c
177
static void free_rxsa(struct rcu_head *head)
drivers/net/macsec.c
179
struct macsec_rx_sa *sa = container_of(head, struct macsec_rx_sa, rcu);
drivers/net/macsec.c
205
static void free_txsa(struct rcu_head *head)
drivers/net/macsec.c
207
struct macsec_tx_sa *sa = container_of(head, struct macsec_tx_sa, rcu);
drivers/net/macsec.c
3968
static void macsec_common_dellink(struct net_device *dev, struct list_head *head)
drivers/net/macsec.c
3985
unregister_netdevice_queue(dev, head);
drivers/net/macsec.c
3993
static void macsec_dellink(struct net_device *dev, struct list_head *head)
drivers/net/macsec.c
3999
macsec_common_dellink(dev, head);
drivers/net/macsec.c
4397
LIST_HEAD(head);
drivers/net/macsec.c
4416
macsec_common_dellink(m->secy.netdev, &head);
drivers/net/macsec.c
4422
unregister_netdevice_many(&head);
drivers/net/macvlan.c
1317
struct nlattr *nla, *head;
drivers/net/macvlan.c
1369
head = nla_data(data[IFLA_MACVLAN_MACADDR_DATA]);
drivers/net/macvlan.c
1372
nla_for_each_attr(nla, head, len, rem) {
drivers/net/macvlan.c
1399
struct nlattr *nla, *head;
drivers/net/macvlan.c
1434
head = nla_data(data[IFLA_MACVLAN_MACADDR_DATA]);
drivers/net/macvlan.c
1437
nla_for_each_attr(nla, head, len, rem) {
drivers/net/macvlan.c
1591
void macvlan_dellink(struct net_device *dev, struct list_head *head)
drivers/net/macvlan.c
1599
unregister_netdevice_queue(dev, head);
drivers/net/macvtap.c
120
struct list_head *head)
drivers/net/macvtap.c
126
macvlan_dellink(dev, head);
drivers/net/mhi_net.c
146
struct sk_buff *head = mhi_netdev->skbagg_head;
drivers/net/mhi_net.c
150
if (!head) {
drivers/net/mhi_net.c
155
if (!skb_shinfo(head)->frag_list)
drivers/net/mhi_net.c
156
skb_shinfo(head)->frag_list = skb;
drivers/net/mhi_net.c
160
head->len += skb->len;
drivers/net/mhi_net.c
161
head->data_len += skb->len;
drivers/net/mhi_net.c
162
head->truesize += skb->truesize;
drivers/net/netkit.c
849
static void netkit_del_link(struct net_device *dev, struct list_head *head)
drivers/net/netkit.c
855
unregister_netdevice_queue(dev, head);
drivers/net/netkit.c
859
unregister_netdevice_queue(peer, head);
drivers/net/ovpn/crypto.c
21
static void ovpn_ks_destroy_rcu(struct rcu_head *head)
drivers/net/ovpn/crypto.c
25
ks = container_of(head, struct ovpn_crypto_key_slot, rcu);
drivers/net/ovpn/main.c
211
static void ovpn_dellink(struct net_device *dev, struct list_head *head)
drivers/net/ovpn/main.c
217
unregister_netdevice_queue(dev, head);
drivers/net/ovpn/peer.c
336
static void ovpn_peer_release_rcu(struct rcu_head *head)
drivers/net/ovpn/peer.c
338
struct ovpn_peer *peer = container_of(head, struct ovpn_peer, rcu);
drivers/net/ovpn/peer.c
654
struct hlist_head *head;
drivers/net/ovpn/peer.c
659
head = ovpn_get_hash_head(ovpn->peers->by_id, &peer_id,
drivers/net/ovpn/peer.c
663
hlist_for_each_entry_rcu(tmp, head, hash_entry_id) {
drivers/net/pfcp.c
224
static void pfcp_dellink(struct net_device *dev, struct list_head *head)
drivers/net/pfcp.c
229
unregister_netdevice_queue(dev, head);
drivers/net/phy/phy_device.c
1651
list_add(&port->head, &phydev->ports);
drivers/net/phy/phy_device.c
1663
list_del(&port->head);
drivers/net/phy/phy_device.c
3430
list_for_each_entry_safe(port, tmp, &phydev->ports, head) {
drivers/net/phy/phy_device.c
3542
list_for_each_entry(port, &phydev->ports, head)
drivers/net/phy/phy_device.c
3552
list_for_each_entry(port, &phydev->ports, head)
drivers/net/phy/phy_device.c
3576
list_for_each_entry(port, &phydev->ports, head)
drivers/net/phy/phy_port.c
27
INIT_LIST_HEAD(&port->head);
drivers/net/ppp/ppp_generic.c
1127
static void ppp_nl_dellink(struct net_device *dev, struct list_head *head);
drivers/net/ppp/ppp_generic.c
1364
static void ppp_nl_dellink(struct net_device *dev, struct list_head *head)
drivers/net/ppp/ppp_generic.c
1366
unregister_netdevice_queue(dev, head);
drivers/net/ppp/ppp_generic.c
2765
struct sk_buff *head, *tail;
drivers/net/ppp/ppp_generic.c
2771
head = __skb_peek(list);
drivers/net/ppp/ppp_generic.c
2814
head = p;
drivers/net/ppp/ppp_generic.c
2823
(PPP_MP_CB(head)->BEbits & B)) {
drivers/net/ppp/ppp_generic.c
2852
head = skb_peek(list);
drivers/net/ppp/ppp_generic.c
2853
if (!head)
drivers/net/ppp/ppp_generic.c
2863
if (PPP_MP_CB(head)->sequence != ppp->nextseq) {
drivers/net/ppp/ppp_generic.c
2865
if (p == head)
drivers/net/ppp/ppp_generic.c
2879
PPP_MP_CB(head)->sequence-1);
drivers/net/ppp/ppp_generic.c
2884
skb = head;
drivers/net/ppp/ppp_generic.c
2885
if (head != tail) {
drivers/net/ppp/ppp_generic.c
2887
p = skb_queue_next(list, head);
drivers/net/ppp/pptp.c
340
skb_set_network_header(skb, skb->head-skb->data);
drivers/net/tun.c
306
static struct tun_flow_entry *tun_flow_find(struct hlist_head *head, u32 rxhash)
drivers/net/tun.c
310
hlist_for_each_entry_rcu(e, head, hash_link) {
drivers/net/tun.c
318
struct hlist_head *head,
drivers/net/tun.c
332
hlist_add_head_rcu(&e->hash_link, head);
drivers/net/tun.c
414
struct hlist_head *head;
drivers/net/tun.c
419
head = &tun->flows[tun_hashfn(rxhash)];
drivers/net/tun.c
423
e = tun_flow_find(head, rxhash);
drivers/net/tun.c
433
if (!tun_flow_find(head, rxhash) &&
drivers/net/tun.c
435
tun_flow_create(tun, head, rxhash, queue_index);
drivers/net/tun_vnet.h
226
16, 1, skb->head,
drivers/net/tun_vnet.h
259
16, 1, skb->head, min(hdr_len, 64),
drivers/net/usb/asix_common.c
295
skb->data = memmove(skb->head + 4, skb->data, skb->len);
drivers/net/usb/cdc_eem.c
116
skb->data = memmove(skb->head +
drivers/net/usb/gl620a.c
151
skb->data = memmove(skb->head + (4 + 4*1),
drivers/net/usb/int51x1.c
67
skb->data = memmove(skb->head + INT51X1_HEADER_SIZE,
drivers/net/usb/kalmia.c
183
skb->data = memmove(skb->head + KALMIA_HEADER_LENGTH,
drivers/net/usb/lan78xx.c
4098
rx_buf->data = rx_buf->head;
drivers/net/usb/lan78xx.c
487
buf->data = buf->head;
drivers/net/usb/lg-vl600.c
259
skb->data = memmove(skb->head + sizeof(*frame),
drivers/net/usb/net1080.c
436
skb->data = memmove(skb->head
drivers/net/usb/qmi_wwan.c
294
struct list_head *head)
drivers/net/usb/qmi_wwan.c
300
unregister_netdevice_queue(dev, head);
drivers/net/usb/r8152.c
2101
tp->tx_info[i].head = NULL;
drivers/net/usb/r8152.c
2161
tp->tx_info[i].head = tx_agg_align(buf);
drivers/net/usb/r8152.c
2366
tx_data = agg->head;
drivers/net/usb/r8152.c
2415
remain = agg_buf_sz - (int)(tx_agg_align(tx_data) - agg->head);
drivers/net/usb/r8152.c
2440
agg->head, (int)(tx_data - (u8 *)agg->head),
drivers/net/usb/r8152.c
858
void *head;
drivers/net/usb/rndis_host.c
570
skb->data = memmove(skb->head + sizeof *hdr,
drivers/net/usb/sr9800.c
124
skb->data = memmove(skb->head + 4, skb->data,
drivers/net/veth.c
1931
static void veth_dellink(struct net_device *dev, struct list_head *head)
drivers/net/veth.c
1944
unregister_netdevice_queue(dev, head);
drivers/net/veth.c
1949
unregister_netdevice_queue(peer, head);
drivers/net/veth.c
768
frame_sz = skb_end_pointer(skb) - skb->head;
drivers/net/veth.c
771
xdp_prepare_buff(xdp, skb->head, skb_headroom(skb),
drivers/net/virtio_net.c
1004
void *buf, *head;
drivers/net/virtio_net.c
1009
head = page_address(alloc_frag->page);
drivers/net/virtio_net.c
1011
dma = head;
drivers/net/virtio_net.c
1047
buf = head + alloc_frag->offset;
drivers/net/virtio_net.c
936
void *head;
drivers/net/virtio_net.c
941
head = page_address(page);
drivers/net/virtio_net.c
943
dma = head;
drivers/net/virtio_net.c
948
offset = buf - (head + sizeof(*dma));
drivers/net/virtio_net.c
983
void *head;
drivers/net/virtio_net.c
987
head = page_address(rq->alloc_frag.page);
drivers/net/virtio_net.c
989
offset = buf - head;
drivers/net/virtio_net.c
991
dma = head;
drivers/net/vrf.c
1666
static void vrf_dellink(struct net_device *dev, struct list_head *head)
drivers/net/vrf.c
1676
unregister_netdevice_queue(dev, head);
drivers/net/vxlan/vxlan_core.c
1762
nh = skb_network_header(skb) - skb->head;
drivers/net/vxlan/vxlan_core.c
1776
oiph = skb->head + nh;
drivers/net/vxlan/vxlan_core.c
4508
static void vxlan_dellink(struct net_device *dev, struct list_head *head)
drivers/net/vxlan/vxlan_core.c
4516
unregister_netdevice_queue(dev, head);
drivers/net/vxlan/vxlan_core.c
657
struct list_head *head,
drivers/net/vxlan/vxlan_core.c
691
list_for_each_entry(p, head, list) {
drivers/net/vxlan/vxlan_core.c
707
struct list_head *head,
drivers/net/vxlan/vxlan_core.c
714
if (vxlan_gro_prepare_receive(sk, head, skb, &grc)) {
drivers/net/vxlan/vxlan_core.c
715
pp = call_gro_receive(eth_gro_receive, head, skb);
drivers/net/vxlan/vxlan_core.c
723
struct list_head *head,
drivers/net/vxlan/vxlan_core.c
733
vh = vxlan_gro_prepare_receive(sk, head, skb, &grc);
drivers/net/vxlan/vxlan_core.c
740
pp = call_gro_receive(ptype->callbacks.gro_receive, head, skb);
drivers/net/vxlan/vxlan_core.c
928
static void vxlan_fdb_free(struct rcu_head *head)
drivers/net/vxlan/vxlan_core.c
930
struct vxlan_fdb *f = container_of(head, struct vxlan_fdb, rcu);
drivers/net/vxlan/vxlan_core.c
960
static void vxlan_dst_free(struct rcu_head *head)
drivers/net/vxlan/vxlan_core.c
962
struct vxlan_rdst *rd = container_of(head, struct vxlan_rdst, rcu);
drivers/net/vxlan/vxlan_mdb.c
686
static void vxlan_mdb_rdst_free(struct rcu_head *head)
drivers/net/vxlan/vxlan_mdb.c
688
struct vxlan_rdst *rd = container_of(head, struct vxlan_rdst, rcu);
drivers/net/wireguard/device.h
35
struct sk_buff *head, *tail, *peeked;
drivers/net/wireguard/queueing.c
56
queue->head = queue->tail = STUB(queue);
drivers/net/wireguard/queueing.c
69
WRITE_ONCE(NEXT(xchg_release(&queue->head, skb)), skb);
drivers/net/wireguard/queueing.c
96
if (tail != READ_ONCE(queue->head))
drivers/net/wireguard/receive.c
367
if (unlikely(skb_network_header(skb) < skb->head))
drivers/net/wireguard/receive.c
53
skb_transport_header(skb) < skb->head ||
drivers/net/wireguard/selftest/allowedips.c
100
hlist_for_each_entry_safe(node, h, &table->head, table) {
drivers/net/wireguard/selftest/allowedips.c
158
hlist_for_each_entry(other, &table->head, table) {
drivers/net/wireguard/selftest/allowedips.c
167
hlist_for_each_entry(other, &table->head, table) {
drivers/net/wireguard/selftest/allowedips.c
173
hlist_add_head(&node->table, &table->head);
drivers/net/wireguard/selftest/allowedips.c
219
hlist_for_each_entry(node, &table->head, table) {
drivers/net/wireguard/selftest/allowedips.c
231
hlist_for_each_entry(node, &table->head, table) {
drivers/net/wireguard/selftest/allowedips.c
245
hlist_for_each_entry_safe(node, h, &table->head, table) {
drivers/net/wireguard/selftest/allowedips.c
79
struct hlist_head head;
drivers/net/wireguard/selftest/allowedips.c
92
INIT_HLIST_HEAD(&table->head);
drivers/net/wireless/ath/ath10k/debug.c
206
static void ath10k_fw_stats_pdevs_free(struct list_head *head)
drivers/net/wireless/ath/ath10k/debug.c
210
list_for_each_entry_safe(i, tmp, head, list) {
drivers/net/wireless/ath/ath10k/debug.c
216
static void ath10k_fw_stats_vdevs_free(struct list_head *head)
drivers/net/wireless/ath/ath10k/debug.c
220
list_for_each_entry_safe(i, tmp, head, list) {
drivers/net/wireless/ath/ath10k/debug.c
226
static void ath10k_fw_stats_peers_free(struct list_head *head)
drivers/net/wireless/ath/ath10k/debug.c
230
list_for_each_entry_safe(i, tmp, head, list) {
drivers/net/wireless/ath/ath10k/debug.c
236
static void ath10k_fw_extd_stats_peers_free(struct list_head *head)
drivers/net/wireless/ath/ath10k/debug.c
240
list_for_each_entry_safe(i, tmp, head, list) {
drivers/net/wireless/ath/ath11k/cfr.c
59
static void ath11k_cfr_rfs_write(struct ath11k *ar, const void *head,
drivers/net/wireless/ath/ath11k/cfr.c
68
relay_write(cfr->rfs_cfr_capture, head, head_len);
drivers/net/wireless/ath/ath11k/core.c
1074
void ath11k_fw_stats_pdevs_free(struct list_head *head)
drivers/net/wireless/ath/ath11k/core.c
1078
list_for_each_entry_safe(i, tmp, head, list) {
drivers/net/wireless/ath/ath11k/core.c
1084
void ath11k_fw_stats_vdevs_free(struct list_head *head)
drivers/net/wireless/ath/ath11k/core.c
1088
list_for_each_entry_safe(i, tmp, head, list) {
drivers/net/wireless/ath/ath11k/core.c
1094
void ath11k_fw_stats_bcn_free(struct list_head *head)
drivers/net/wireless/ath/ath11k/core.c
1098
list_for_each_entry_safe(i, tmp, head, list) {
drivers/net/wireless/ath/ath11k/core.h
1252
void ath11k_fw_stats_pdevs_free(struct list_head *head);
drivers/net/wireless/ath/ath11k/core.h
1253
void ath11k_fw_stats_vdevs_free(struct list_head *head);
drivers/net/wireless/ath/ath11k/core.h
1254
void ath11k_fw_stats_bcn_free(struct list_head *head);
drivers/net/wireless/ath/ath11k/dp.h
127
struct sk_buff *head;
drivers/net/wireless/ath/ath11k/dp_rx.c
5484
struct sk_buff *head,
drivers/net/wireless/ath/ath11k/dp_rx.c
5492
mon_mpdu->head = head;
drivers/net/wireless/ath/ath11k/dp_rx.c
5510
tmp_msdu = mon_mpdu->head;
drivers/net/wireless/ath/ath11k/dp_rx.c
5538
head_msdu = mon_mpdu->head;
drivers/net/wireless/ath/ath12k/core.c
1211
static void ath12k_fw_stats_pdevs_free(struct list_head *head)
drivers/net/wireless/ath/ath12k/core.c
1215
list_for_each_entry_safe(i, tmp, head, list) {
drivers/net/wireless/ath/ath12k/core.c
1221
void ath12k_fw_stats_bcn_free(struct list_head *head)
drivers/net/wireless/ath/ath12k/core.c
1225
list_for_each_entry_safe(i, tmp, head, list) {
drivers/net/wireless/ath/ath12k/core.c
1231
static void ath12k_fw_stats_vdevs_free(struct list_head *head)
drivers/net/wireless/ath/ath12k/core.c
1235
list_for_each_entry_safe(i, tmp, head, list) {
drivers/net/wireless/ath/ath12k/core.h
1271
void ath12k_fw_stats_bcn_free(struct list_head *head);
drivers/net/wireless/ath/ath12k/dp.h
124
struct sk_buff *head;
drivers/net/wireless/ath/ath12k/dp_mon.c
224
head_msdu = mon_mpdu->head;
drivers/net/wireless/ath/ath12k/dp_mon.c
642
if (!pmon->mon_mpdu->head)
drivers/net/wireless/ath/ath12k/dp_mon.c
643
pmon->mon_mpdu->head = msdu;
drivers/net/wireless/ath/ath12k/dp_rx.c
24
struct list_head *head,
drivers/net/wireless/ath/ath12k/dp_rx.c
36
list_for_each(cur, head) {
drivers/net/wireless/ath/ath12k/dp_rx.c
47
list_cut_before(list, head, cur);
drivers/net/wireless/ath/ath12k/dp_tx.c
134
bool head)
drivers/net/wireless/ath/ath12k/dp_tx.c
138
if (head) {
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
1944
tx_ppdu_info->tx_mon_mpdu->head = skb;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
1971
tx_ppdu_info->tx_mon_mpdu->head = skb;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
1999
tx_ppdu_info->tx_mon_mpdu->head = skb;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
2028
tx_ppdu_info->tx_mon_mpdu->head = skb;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
2056
tx_ppdu_info->tx_mon_mpdu->head = skb;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
2480
mon_skb = mon_mpdu->head;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
2500
if (mon_mpdu->head)
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
2860
tmp_mpdu->head = head_msdu;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
2998
if (mon_mpdu->head && mon_mpdu->tail)
drivers/net/wireless/ath/ath6kl/cfg80211.c
2804
if (info->beacon.head == NULL)
drivers/net/wireless/ath/ath6kl/cfg80211.c
2806
mgmt = (struct ieee80211_mgmt *) info->beacon.head;
drivers/net/wireless/ath/ath6kl/cfg80211.c
2808
if (ies > info->beacon.head + info->beacon.head_len)
drivers/net/wireless/ath/ath6kl/txrx.c
397
(skb_network_header(skb) - skb->head) +
drivers/net/wireless/ath/ath6kl/txrx.c
900
packet = (struct htc_packet *) skb->head;
drivers/net/wireless/ath/ath6kl/txrx.c
926
packet = (struct htc_packet *) skb->head;
drivers/net/wireless/ath/ath9k/ath9k.h
78
struct list_head *head, const char *name,
drivers/net/wireless/ath/ath9k/init.c
283
struct list_head *head, const char *name,
drivers/net/wireless/ath/ath9k/init.c
293
INIT_LIST_HEAD(head);
drivers/net/wireless/ath/ath9k/init.c
368
list_add_tail(&bf->list, head);
drivers/net/wireless/ath/ath9k/init.c
399
list_add_tail(&bf->list, head);
drivers/net/wireless/ath/ath9k/xmit.c
2042
struct list_head *head, bool internal)
drivers/net/wireless/ath/ath9k/xmit.c
2055
if (list_empty(head))
drivers/net/wireless/ath/ath9k/xmit.c
2059
bf = list_first_entry(head, struct ath_buf, list);
drivers/net/wireless/ath/ath9k/xmit.c
2060
bf_last = list_entry(head->prev, struct ath_buf, list);
drivers/net/wireless/ath/ath9k/xmit.c
2066
list_splice_tail_init(head, &txq->txq_fifo[txq->txq_headidx]);
drivers/net/wireless/ath/ath9k/xmit.c
2070
list_splice_tail_init(head, &txq->axq_q);
drivers/net/wireless/ath/ath9k/xmit.c
66
struct list_head *head, bool internal);
drivers/net/wireless/ath/carl9170/carl9170.h
477
struct rcu_head head;
drivers/net/wireless/ath/carl9170/fw.c
40
const struct carl9170fw_desc_head *head, unsigned int max_len)
drivers/net/wireless/ath/carl9170/fw.c
51
pos = head;
drivers/net/wireless/ath/carl9170/fw.c
56
if (pos_addr + sizeof(*head) > end_addr)
drivers/net/wireless/ath/carl9170/fw.c
61
if (pos_length < sizeof(*head))
drivers/net/wireless/ath/carl9170/fwdesc.h
113
struct carl9170fw_desc_head head;
drivers/net/wireless/ath/carl9170/fwdesc.h
134
struct carl9170fw_desc_head head;
drivers/net/wireless/ath/carl9170/fwdesc.h
151
struct carl9170fw_desc_head head;
drivers/net/wireless/ath/carl9170/fwdesc.h
160
struct carl9170fw_desc_head head;
drivers/net/wireless/ath/carl9170/fwdesc.h
176
struct carl9170fw_desc_head head;
drivers/net/wireless/ath/carl9170/fwdesc.h
186
struct carl9170fw_desc_head head;
drivers/net/wireless/ath/carl9170/fwdesc.h
196
struct carl9170fw_desc_head head;
drivers/net/wireless/ath/carl9170/fwdesc.h
206
struct carl9170fw_desc_head head;
drivers/net/wireless/ath/carl9170/fwdesc.h
214
.head = { \
drivers/net/wireless/ath/carl9170/fwdesc.h
221
static inline void carl9170fw_fill_desc(struct carl9170fw_desc_head *head,
drivers/net/wireless/ath/carl9170/fwdesc.h
225
head->magic[0] = magic[0];
drivers/net/wireless/ath/carl9170/fwdesc.h
226
head->magic[1] = magic[1];
drivers/net/wireless/ath/carl9170/fwdesc.h
227
head->magic[2] = magic[2];
drivers/net/wireless/ath/carl9170/fwdesc.h
228
head->magic[3] = magic[3];
drivers/net/wireless/ath/carl9170/fwdesc.h
230
head->length = length;
drivers/net/wireless/ath/carl9170/fwdesc.h
231
head->min_ver = min_ver;
drivers/net/wireless/ath/carl9170/fwdesc.h
232
head->cur_ver = cur_ver;
drivers/net/wireless/ath/carl9170/fwdesc.h
242
#define CHECK_HDR_VERSION(head, _min_ver) \
drivers/net/wireless/ath/carl9170/fwdesc.h
243
(((head)->cur_ver < _min_ver) || ((head)->min_ver > _min_ver)) \
drivers/net/wireless/ath/carl9170/fwdesc.h
250
static inline bool carl9170fw_desc_cmp(const struct carl9170fw_desc_head *head,
drivers/net/wireless/ath/carl9170/fwdesc.h
254
if (descid[0] == head->magic[0] && descid[1] == head->magic[1] &&
drivers/net/wireless/ath/carl9170/fwdesc.h
255
descid[2] == head->magic[2] && descid[3] == head->magic[3] &&
drivers/net/wireless/ath/carl9170/fwdesc.h
256
!CHECK_HDR_VERSION(head, compatible_revision) &&
drivers/net/wireless/ath/carl9170/fwdesc.h
257
(le16_to_cpu(head->length) >= min_len))
drivers/net/wireless/ath/carl9170/rx.c
290
struct ar9170_rx_head *head, struct ar9170_rx_macstatus *mac,
drivers/net/wireless/ath/carl9170/rx.c
362
switch (head->plcp[0]) {
drivers/net/wireless/ath/carl9170/rx.c
378
"rate (%x).\n", head->plcp[0]);
drivers/net/wireless/ath/carl9170/rx.c
387
switch (head->plcp[0] & 0xf) {
drivers/net/wireless/ath/carl9170/rx.c
415
"rate (%x).\n", head->plcp[0]);
drivers/net/wireless/ath/carl9170/rx.c
425
if (head->plcp[3] & 0x80)
drivers/net/wireless/ath/carl9170/rx.c
427
if (head->plcp[6] & 0x80)
drivers/net/wireless/ath/carl9170/rx.c
430
status->rate_idx = clamp(head->plcp[3] & 0x7f, 0, 75);
drivers/net/wireless/ath/carl9170/rx.c
611
kfree_rcu(entry, head);
drivers/net/wireless/ath/carl9170/rx.c
699
struct ar9170_rx_head *head;
drivers/net/wireless/ath/carl9170/rx.c
723
head = (void *) buf;
drivers/net/wireless/ath/carl9170/rx.c
783
head = &ar->rx_plcp;
drivers/net/wireless/ath/carl9170/rx.c
788
head = (void *) buf;
drivers/net/wireless/ath/carl9170/rx.c
806
if (unlikely(carl9170_rx_mac_status(ar, head, mac, &status)))
drivers/net/wireless/ath/carl9170/tx.c
463
kfree_rcu(entry, head);
drivers/net/wireless/ath/carl9170/wlan.h
395
struct ar9170_rx_frame_head head;
drivers/net/wireless/ath/dfs_pattern_detector.c
162
struct list_head head;
drivers/net/wireless/ath/dfs_pattern_detector.c
185
list_del(&cd->head);
drivers/net/wireless/ath/dfs_pattern_detector.c
206
INIT_LIST_HEAD(&cd->head);
drivers/net/wireless/ath/dfs_pattern_detector.c
216
list_add(&cd->head, &dpd->channel_detectors);
drivers/net/wireless/ath/dfs_pattern_detector.c
240
list_for_each_entry(cd, &dpd->channel_detectors, head) {
drivers/net/wireless/ath/dfs_pattern_detector.c
255
list_for_each_entry(cd, &dpd->channel_detectors, head)
drivers/net/wireless/ath/dfs_pattern_detector.c
262
list_for_each_entry_safe(cd, cd0, &dpd->channel_detectors, head)
drivers/net/wireless/ath/dfs_pattern_detector.c
331
list_for_each_entry_safe(cd, cd0, &dpd->channel_detectors, head)
drivers/net/wireless/ath/dfs_pri_detector.c
104
list_for_each_entry_safe(p, p0, &pulse_pool, head) {
drivers/net/wireless/ath/dfs_pri_detector.c
105
list_del(&p->head);
drivers/net/wireless/ath/dfs_pri_detector.c
109
list_for_each_entry_safe(ps, ps0, &pseq_pool, head) {
drivers/net/wireless/ath/dfs_pri_detector.c
110
list_del(&ps->head);
drivers/net/wireless/ath/dfs_pri_detector.c
121
list_add(&pe->head, &pulse_pool);
drivers/net/wireless/ath/dfs_pri_detector.c
129
list_add(&pse->head, &pseq_pool);
drivers/net/wireless/ath/dfs_pri_detector.c
139
pse = list_first_entry(&pseq_pool, struct pri_sequence, head);
drivers/net/wireless/ath/dfs_pri_detector.c
140
list_del(&pse->head);
drivers/net/wireless/ath/dfs_pri_detector.c
152
pe = list_first_entry(&pulse_pool, struct pulse_elem, head);
drivers/net/wireless/ath/dfs_pri_detector.c
153
list_del(&pe->head);
drivers/net/wireless/ath/dfs_pri_detector.c
165
return list_entry(l->prev, struct pulse_elem, head);
drivers/net/wireless/ath/dfs_pri_detector.c
172
list_del_init(&p->head);
drivers/net/wireless/ath/dfs_pri_detector.c
213
INIT_LIST_HEAD(&p->head);
drivers/net/wireless/ath/dfs_pri_detector.c
215
list_add(&p->head, &pde->pulses);
drivers/net/wireless/ath/dfs_pri_detector.c
228
list_for_each_entry(p, &pde->pulses, head) {
drivers/net/wireless/ath/dfs_pri_detector.c
257
list_for_each_entry_continue(p2, &pde->pulses, head) {
drivers/net/wireless/ath/dfs_pri_detector.c
296
INIT_LIST_HEAD(&new_ps->head);
drivers/net/wireless/ath/dfs_pri_detector.c
297
list_add(&new_ps->head, &pde->sequences);
drivers/net/wireless/ath/dfs_pri_detector.c
308
list_for_each_entry_safe(ps, ps2, &pde->sequences, head) {
drivers/net/wireless/ath/dfs_pri_detector.c
314
list_del_init(&ps->head);
drivers/net/wireless/ath/dfs_pri_detector.c
343
list_for_each_entry(ps, &pde->sequences, head) {
drivers/net/wireless/ath/dfs_pri_detector.c
36
struct list_head head;
drivers/net/wireless/ath/dfs_pri_detector.c
362
list_for_each_entry_safe(ps, ps0, &pde->sequences, head) {
drivers/net/wireless/ath/dfs_pri_detector.c
363
list_del_init(&ps->head);
drivers/net/wireless/ath/dfs_pri_detector.c
366
list_for_each_entry_safe(p, p0, &pde->pulses, head) {
drivers/net/wireless/ath/dfs_pri_detector.c
367
list_del_init(&p->head);
drivers/net/wireless/ath/dfs_pri_detector.h
36
struct list_head head;
drivers/net/wireless/ath/wil6210/cfg80211.c
1848
b->head, b->head_len, true);
drivers/net/wireless/ath/wil6210/debugfs.c
297
seq_printf(s, " head = 0x%08x\n", r.head);
drivers/net/wireless/ath/wil6210/debugfs.c
308
!wmi_addr(wil, r.head)) {
drivers/net/wireless/ath/wil6210/debugfs.c
324
(r.head - r.base == delta) ? "h" : " ",
drivers/net/wireless/ath/wil6210/main.c
1286
le32_to_cpus(&r->head);
drivers/net/wireless/ath/wil6210/wil6210.h
460
u32 head;
drivers/net/wireless/ath/wil6210/wmi.c
1932
r->head = wil_r(wil, RGF_MBOX +
drivers/net/wireless/ath/wil6210/wmi.c
1933
offsetof(struct wil6210_mbox_ctl, rx.head));
drivers/net/wireless/ath/wil6210/wmi.c
1934
if (r->tail == r->head)
drivers/net/wireless/ath/wil6210/wmi.c
1938
r->head, r->tail);
drivers/net/wireless/ath/wil6210/wmi.c
3443
r->head = wil_r(wil, RGF_MBOX +
drivers/net/wireless/ath/wil6210/wmi.c
3444
offsetof(struct wil6210_mbox_ctl, rx.head));
drivers/net/wireless/ath/wil6210/wmi.c
3445
if (r->tail != r->head)
drivers/net/wireless/ath/wil6210/wmi.c
654
void __iomem *head = wmi_addr(wil, r->head);
drivers/net/wireless/ath/wil6210/wmi.c
681
if (!head) {
drivers/net/wireless/ath/wil6210/wmi.c
682
wil_err(wil, "WMI head is garbage: 0x%08x\n", r->head);
drivers/net/wireless/ath/wil6210/wmi.c
690
wil_memcpy_fromio_32(&d_head, head, sizeof(d_head));
drivers/net/wireless/ath/wil6210/wmi.c
701
next_head = r->base + ((r->head - r->base + sizeof(d_head)) % r->size);
drivers/net/wireless/ath/wil6210/wmi.c
702
wil_dbg_wmi(wil, "Head 0x%08x -> 0x%08x\n", r->head, next_head);
drivers/net/wireless/ath/wil6210/wmi.c
739
wil_w(wil, r->head + offsetof(struct wil6210_mbox_ring_desc, sync), 1);
drivers/net/wireless/ath/wil6210/wmi.c
741
wil_w(wil, RGF_MBOX + offsetof(struct wil6210_mbox_ctl, tx.head),
drivers/net/wireless/ath/wil6210/wmi.c
742
r->head = next_head);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
5169
(u8 *)&settings->beacon.head[ie_offset],
drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.h
112
static inline void dma_spin_for_len(uint len, struct sk_buff *head)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.h
116
while (!(len = *(u16 *) KSEG1ADDR(head->data)))
drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.h
119
*(u16 *) (head->data) = cpu_to_le16((u16) len);
drivers/net/wireless/intel/iwlegacy/4965-rs.c
235
tl->total -= tl->packet_count[tl->head];
drivers/net/wireless/intel/iwlegacy/4965-rs.c
236
tl->packet_count[tl->head] = 0;
drivers/net/wireless/intel/iwlegacy/4965-rs.c
239
tl->head++;
drivers/net/wireless/intel/iwlegacy/4965-rs.c
240
if (tl->head >= TID_QUEUE_MAX_SIZE)
drivers/net/wireless/intel/iwlegacy/4965-rs.c
241
tl->head = 0;
drivers/net/wireless/intel/iwlegacy/4965-rs.c
276
tl->head = 0;
drivers/net/wireless/intel/iwlegacy/4965-rs.c
289
idx = (tl->head + idx) % TID_QUEUE_MAX_SIZE;
drivers/net/wireless/intel/iwlegacy/common.h
2777
u8 head; /* start of the circular buffer */
drivers/net/wireless/intel/iwlwifi/dvm/rs.c
234
tl->total -= tl->packet_count[tl->head];
drivers/net/wireless/intel/iwlwifi/dvm/rs.c
235
tl->packet_count[tl->head] = 0;
drivers/net/wireless/intel/iwlwifi/dvm/rs.c
238
tl->head++;
drivers/net/wireless/intel/iwlwifi/dvm/rs.c
239
if (tl->head >= TID_QUEUE_MAX_SIZE)
drivers/net/wireless/intel/iwlwifi/dvm/rs.c
240
tl->head = 0;
drivers/net/wireless/intel/iwlwifi/dvm/rs.c
275
tl->head = 0;
drivers/net/wireless/intel/iwlwifi/dvm/rs.c
288
index = (tl->head + index) % TID_QUEUE_MAX_SIZE;
drivers/net/wireless/intel/iwlwifi/dvm/rs.h
311
u8 head; /* start of the circular buffer */
drivers/net/wireless/intel/iwlwifi/mld/mac80211.c
2009
struct sk_buff *head,
drivers/net/wireless/intel/iwlwifi/mld/mac80211.c
2022
return iwl_mld_can_hw_csum(skb) == iwl_mld_can_hw_csum(head);
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
6310
struct sk_buff *head,
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
6322
return iwl_mvm_can_hw_csum(skb) == iwl_mvm_can_hw_csum(head);
drivers/net/wireless/intersil/p54/fwio.c
395
struct p54_scan_head *head;
drivers/net/wireless/intersil/p54/fwio.c
405
skb = p54_alloc_skb(priv, P54_HDR_FLAG_CONTROL_OPSET, sizeof(*head) +
drivers/net/wireless/intersil/p54/fwio.c
412
head = skb_put(skb, sizeof(*head));
drivers/net/wireless/intersil/p54/fwio.c
413
memset(head->scan_params, 0, sizeof(head->scan_params));
drivers/net/wireless/intersil/p54/fwio.c
414
head->mode = cpu_to_le16(mode);
drivers/net/wireless/intersil/p54/fwio.c
415
head->dwell = cpu_to_le16(dwell);
drivers/net/wireless/intersil/p54/fwio.c
416
head->freq = freq;
drivers/net/wireless/marvell/mwifiex/init.c
545
struct list_head *head;
drivers/net/wireless/marvell/mwifiex/init.c
549
head = &adapter->bss_prio_tbl[i].bss_prio_head;
drivers/net/wireless/marvell/mwifiex/init.c
555
priv->bss_type, priv->bss_num, i, head);
drivers/net/wireless/marvell/mwifiex/init.c
559
list_for_each_entry_safe(bssprio_node, tmp_node, head,
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
1132
struct mwifiex_ie_types_header *head;
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
1142
while (tlv_buf_left >= sizeof(*head)) {
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
1143
head = (struct mwifiex_ie_types_header *)tlv_buf;
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
1144
tlv = le16_to_cpu(head->type);
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
1145
tlv_buf_len = le16_to_cpu(head->len);
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
1147
if (tlv_buf_left < (sizeof(*head) + tlv_buf_len))
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
1153
(u8 *)head + sizeof(*head),
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
1156
(u8 *)head + sizeof(*head), tlv_buf_len);
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
1162
tlv_buf += (sizeof(*head) + tlv_buf_len);
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
1163
tlv_buf_left -= (sizeof(*head) + tlv_buf_len);
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
261
struct mwifiex_ie_types_header *head;
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
269
while (tlv_buf_left >= sizeof(*head)) {
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
270
head = (struct mwifiex_ie_types_header *)tlv_buf;
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
271
tlv = le16_to_cpu(head->type);
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
272
tlv_buf_len = le16_to_cpu(head->len);
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
274
if (tlv_buf_left < (sizeof(*head) + tlv_buf_len))
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
303
tlv_buf += (sizeof(*head) + tlv_buf_len);
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
304
tlv_buf_left -= (sizeof(*head) + tlv_buf_len);
drivers/net/wireless/marvell/mwifiex/uap_cmd.c
246
const u8 *var_pos = params->beacon.head + var_offset;
drivers/net/wireless/marvell/mwl8k.c
1173
rxq->head = 0;
drivers/net/wireless/marvell/mwl8k.c
1344
skb = rxq->buf[rxq->head].skb;
drivers/net/wireless/marvell/mwl8k.c
1348
rxd = rxq->rxd + (rxq->head * priv->rxd_ops->rxd_size);
drivers/net/wireless/marvell/mwl8k.c
1355
rxq->buf[rxq->head].skb = NULL;
drivers/net/wireless/marvell/mwl8k.c
1358
dma_unmap_addr(&rxq->buf[rxq->head], dma),
drivers/net/wireless/marvell/mwl8k.c
1360
dma_unmap_addr_set(&rxq->buf[rxq->head], dma, 0);
drivers/net/wireless/marvell/mwl8k.c
1362
rxq->head++;
drivers/net/wireless/marvell/mwl8k.c
1363
if (rxq->head == MWL8K_RX_DESCS)
drivers/net/wireless/marvell/mwl8k.c
1364
rxq->head = 0;
drivers/net/wireless/marvell/mwl8k.c
141
int head;
drivers/net/wireless/marvell/mwl8k.c
1469
txq->head = 0;
drivers/net/wireless/marvell/mwl8k.c
1543
txq->len, txq->head, txq->tail,
drivers/net/wireless/marvell/mwl8k.c
156
int head;
drivers/net/wireless/marvell/mwl8k.c
1701
tx = txq->head;
drivers/net/wireless/marvell/mwl8k.c
1713
txq->head = (tx + 1) % MWL8K_TX_DESCS;
drivers/net/wireless/mediatek/mt76/agg-rx.c
155
u16 seqno, head, size, idx;
drivers/net/wireless/mediatek/mt76/agg-rx.c
188
head = tid->head;
drivers/net/wireless/mediatek/mt76/agg-rx.c
191
sn_less = ieee80211_sn_less(seqno, head);
drivers/net/wireless/mediatek/mt76/agg-rx.c
20
tid->head = ieee80211_sn_inc(tid->head);
drivers/net/wireless/mediatek/mt76/agg-rx.c
206
if (seqno == head) {
drivers/net/wireless/mediatek/mt76/agg-rx.c
207
tid->head = ieee80211_sn_inc(head);
drivers/net/wireless/mediatek/mt76/agg-rx.c
219
if (!ieee80211_sn_less(seqno, head + size)) {
drivers/net/wireless/mediatek/mt76/agg-rx.c
220
head = ieee80211_sn_inc(ieee80211_sn_sub(seqno, size));
drivers/net/wireless/mediatek/mt76/agg-rx.c
221
mt76_rx_aggr_release_frames(tid, frames, head);
drivers/net/wireless/mediatek/mt76/agg-rx.c
256
tid->head = ssn;
drivers/net/wireless/mediatek/mt76/agg-rx.c
34
u16 head)
drivers/net/wireless/mediatek/mt76/agg-rx.c
38
while (ieee80211_sn_less(tid->head, head)) {
drivers/net/wireless/mediatek/mt76/agg-rx.c
39
idx = tid->head % tid->size;
drivers/net/wireless/mediatek/mt76/agg-rx.c
47
int idx = tid->head % tid->size;
drivers/net/wireless/mediatek/mt76/agg-rx.c
51
idx = tid->head % tid->size;
drivers/net/wireless/mediatek/mt76/agg-rx.c
67
start = tid->head % tid->size;
drivers/net/wireless/mediatek/mt76/agg-rx.c
70
for (idx = (tid->head + 1) % tid->size;
drivers/net/wireless/mediatek/mt76/debugfs.c
67
i, q->queued, q->head, q->tail);
drivers/net/wireless/mediatek/mt76/debugfs.c
85
i, queued, q->head, q->tail);
drivers/net/wireless/mediatek/mt76/dma.c
201
q->head = Q_READ(q, dma_idx);
drivers/net/wireless/mediatek/mt76/dma.c
202
q->tail = q->head;
drivers/net/wireless/mediatek/mt76/dma.c
235
struct mt76_queue_entry *entry = &q->entry[q->head];
drivers/net/wireless/mediatek/mt76/dma.c
239
int idx = q->head;
drivers/net/wireless/mediatek/mt76/dma.c
246
data = &rro_desc[q->head];
drivers/net/wireless/mediatek/mt76/dma.c
249
data = &q->desc[q->head];
drivers/net/wireless/mediatek/mt76/dma.c
253
desc = &q->desc[q->head];
drivers/net/wireless/mediatek/mt76/dma.c
284
if ((q->head + 1) == q->ndesc)
drivers/net/wireless/mediatek/mt76/dma.c
300
q->head = (q->head + 1) % q->ndesc;
drivers/net/wireless/mediatek/mt76/dma.c
317
q->entry[q->head].txwi = DMA_DUMMY_DATA;
drivers/net/wireless/mediatek/mt76/dma.c
318
q->entry[q->head].skip_buf0 = true;
drivers/net/wireless/mediatek/mt76/dma.c
324
idx = q->head;
drivers/net/wireless/mediatek/mt76/dma.c
325
next = (q->head + 1) % q->ndesc;
drivers/net/wireless/mediatek/mt76/dma.c
364
q->head = next;
drivers/net/wireless/mediatek/mt76/dma.c
401
*q->emi_cpu_idx = cpu_to_le16(q->head);
drivers/net/wireless/mediatek/mt76/dma.c
403
Q_WRITE(q, cpu_idx, q->head);
drivers/net/wireless/mediatek/mt76/mac80211.c
883
struct sk_buff *skb = phy->rx_amsdu[q].head;
drivers/net/wireless/mediatek/mt76/mac80211.c
887
phy->rx_amsdu[q].head = NULL;
drivers/net/wireless/mediatek/mt76/mac80211.c
922
if (phy->rx_amsdu[q].head &&
drivers/net/wireless/mediatek/mt76/mac80211.c
927
if (!phy->rx_amsdu[q].head) {
drivers/net/wireless/mediatek/mt76/mac80211.c
930
phy->rx_amsdu[q].head = skb;
drivers/net/wireless/mediatek/mt76/mcu.c
22
memset(skb->head, 0, len);
drivers/net/wireless/mediatek/mt76/mt76.h
243
u16 head;
drivers/net/wireless/mediatek/mt76/mt76.h
462
u16 head;
drivers/net/wireless/mediatek/mt76/mt76.h
898
struct sk_buff *head;
drivers/net/wireless/mediatek/mt76/mt7615/debugfs.c
409
queue_map[i].queue, q->queued, q->head,
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
156
list_add(&seq->head, &dfs_pd->seq_pool);
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
173
head);
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
174
list_del(&seq->head);
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
220
list_for_each_entry_safe(seq, tmp_seq, &dfs_pd->sequences, head) {
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
221
list_del_init(&seq->head);
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
505
INIT_LIST_HEAD(&seq_p->head);
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
506
list_add(&seq_p->head, &dfs_pd->sequences);
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
523
list_for_each_entry_safe(seq, tmp_seq, &dfs_pd->sequences, head) {
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
525
list_del_init(&seq->head);
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
553
list_for_each_entry(seq, &dfs_pd->sequences, head) {
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.h
76
struct list_head head;
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
823
u32 ctrl, head, tail, queued;
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
831
head = mt76_get_field(dev, MT_FL_Q2_CTRL,
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
840
queued, head, tail);
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
908
u32 val, head, tail;
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
912
head = mt76_get_field(dev, MT_PLE_FREEPG_HEAD_TAIL, GENMASK(11, 0));
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
917
val, head, tail);
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
920
head = mt76_get_field(dev, MT_PLE_HIF_PG_INFO, GENMASK(11, 0));
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
923
val, head, tail);
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
966
queue_map[i].queue, q->queued, q->head,
drivers/net/wireless/mediatek/mt76/mt792x_debugfs.c
119
queue_map[i].queue, q->queued, q->head,
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
603
u32 ctrl, head, tail, queued;
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
611
head = mt76_get_field(dev, MT_FL_Q2_CTRL,
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
620
queued, head, tail);
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
711
u32 val, head, tail;
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
715
head = mt76_get_field(dev, MT_PLE_FREEPG_HEAD_TAIL, GENMASK(11, 0));
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
720
val, head, tail);
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
723
head = mt76_get_field(dev, MT_PLE_HIF_PG_INFO, GENMASK(11, 0));
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
726
val, head, tail);
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
788
queue_map[i].queue, q->queued, q->head,
drivers/net/wireless/mediatek/mt76/npu.c
215
memcpy(desc[q->head].txwi, txwi_ptr, txwi_len);
drivers/net/wireless/mediatek/mt76/npu.c
216
desc[q->head].addr = buf->addr;
drivers/net/wireless/mediatek/mt76/npu.c
217
desc[q->head].ctrl = FIELD_PREP(NPU_TX_DMA_DESC_VEND_LEN_MASK, txwi_len) |
drivers/net/wireless/mediatek/mt76/npu.c
221
ret = q->head;
drivers/net/wireless/mediatek/mt76/npu.c
222
q->entry[q->head].skip_buf0 = true;
drivers/net/wireless/mediatek/mt76/npu.c
223
q->entry[q->head].skip_buf1 = true;
drivers/net/wireless/mediatek/mt76/npu.c
224
q->entry[q->head].txwi = NULL;
drivers/net/wireless/mediatek/mt76/npu.c
225
q->entry[q->head].skb = NULL;
drivers/net/wireless/mediatek/mt76/npu.c
226
q->entry[q->head].wcid = 0xffff;
drivers/net/wireless/mediatek/mt76/npu.c
228
q->head = (q->head + 1) % q->ndesc;
drivers/net/wireless/mediatek/mt76/npu.c
23
struct mt76_queue_entry *e = &q->entry[q->head];
drivers/net/wireless/mediatek/mt76/npu.c
35
memset(&desc[q->head], 0, sizeof(*desc));
drivers/net/wireless/mediatek/mt76/npu.c
36
desc[q->head].addr = e->dma_addr[0];
drivers/net/wireless/mediatek/mt76/npu.c
38
q->head = (q->head + 1) % q->ndesc;
drivers/net/wireless/mediatek/mt76/sdio.c
316
q->head = q->tail = 0;
drivers/net/wireless/mediatek/mt76/sdio.c
527
u16 idx = q->head;
drivers/net/wireless/mediatek/mt76/sdio.c
537
q->entry[q->head].skb = tx_info.skb;
drivers/net/wireless/mediatek/mt76/sdio.c
538
q->entry[q->head].buf_sz = len;
drivers/net/wireless/mediatek/mt76/sdio.c
539
q->entry[q->head].wcid = 0xffff;
drivers/net/wireless/mediatek/mt76/sdio.c
543
q->head = (q->head + 1) % q->ndesc;
drivers/net/wireless/mediatek/mt76/sdio.c
568
q->entry[q->head].buf_sz = len;
drivers/net/wireless/mediatek/mt76/sdio.c
569
q->entry[q->head].skb = skb;
drivers/net/wireless/mediatek/mt76/sdio.c
574
q->head = (q->head + 1) % q->ndesc;
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
124
int index = (q->head + i) % q->ndesc;
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
147
q->head = (q->head + i) % q->ndesc;
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
252
while (q->first != q->head) {
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
384
if (q->first != q->head)
drivers/net/wireless/mediatek/mt76/testmode.c
107
struct sk_buff **frag_tail, *head;
drivers/net/wireless/mediatek/mt76/testmode.c
125
head = alloc_skb(head_len, GFP_KERNEL);
drivers/net/wireless/mediatek/mt76/testmode.c
126
if (!head)
drivers/net/wireless/mediatek/mt76/testmode.c
129
hdr = __skb_put_zero(head, sizeof(*hdr));
drivers/net/wireless/mediatek/mt76/testmode.c
134
skb_set_queue_mapping(head, IEEE80211_AC_BE);
drivers/net/wireless/mediatek/mt76/testmode.c
135
get_random_bytes(__skb_put(head, head_len - sizeof(*hdr)),
drivers/net/wireless/mediatek/mt76/testmode.c
138
info = IEEE80211_SKB_CB(head);
drivers/net/wireless/mediatek/mt76/testmode.c
144
frag_tail = &skb_shinfo(head)->frag_list;
drivers/net/wireless/mediatek/mt76/testmode.c
158
dev_kfree_skb(head);
drivers/net/wireless/mediatek/mt76/testmode.c
163
head->len += frag->len;
drivers/net/wireless/mediatek/mt76/testmode.c
164
head->data_len += frag->len;
drivers/net/wireless/mediatek/mt76/testmode.c
171
td->tx_skb = head;
drivers/net/wireless/mediatek/mt76/tx.c
336
struct sk_buff_head *head;
drivers/net/wireless/mediatek/mt76/tx.c
355
head = &wcid->tx_offchannel;
drivers/net/wireless/mediatek/mt76/tx.c
357
head = &wcid->tx_pending;
drivers/net/wireless/mediatek/mt76/tx.c
359
spin_lock_bh(&head->lock);
drivers/net/wireless/mediatek/mt76/tx.c
360
__skb_queue_tail(head, skb);
drivers/net/wireless/mediatek/mt76/tx.c
361
spin_unlock_bh(&head->lock);
drivers/net/wireless/mediatek/mt76/tx.c
603
struct sk_buff_head *head)
drivers/net/wireless/mediatek/mt76/tx.c
611
spin_lock(&head->lock);
drivers/net/wireless/mediatek/mt76/tx.c
612
while ((skb = skb_peek(head)) != NULL) {
drivers/net/wireless/mediatek/mt76/tx.c
622
head == &wcid->tx_offchannel))
drivers/net/wireless/mediatek/mt76/tx.c
631
__skb_unlink(skb, head);
drivers/net/wireless/mediatek/mt76/tx.c
632
spin_unlock(&head->lock);
drivers/net/wireless/mediatek/mt76/tx.c
640
spin_lock(&head->lock);
drivers/net/wireless/mediatek/mt76/tx.c
642
spin_unlock(&head->lock);
drivers/net/wireless/mediatek/mt76/usb.c
578
if (WARN_ONCE(q->entry[q->head].urb != urb, "rx urb mismatch"))
drivers/net/wireless/mediatek/mt76/usb.c
581
q->head = (q->head + 1) % q->ndesc;
drivers/net/wireless/mediatek/mt76/usb.c
653
q->head = q->tail = 0;
drivers/net/wireless/mediatek/mt76/usb.c
861
u16 idx = q->head;
drivers/net/wireless/mediatek/mt76/usb.c
879
q->head = (q->head + 1) % q->ndesc;
drivers/net/wireless/mediatek/mt76/usb.c
892
while (q->first != q->head) {
drivers/net/wireless/mediatek/mt76/wed.c
138
q->head = q->ndesc - 1;
drivers/net/wireless/mediatek/mt76/wed.c
139
q->queued = q->head;
drivers/net/wireless/mediatek/mt76/wed.c
145
q->head = q->ndesc - 1;
drivers/net/wireless/mediatek/mt76/wed.c
146
q->queued = q->head;
drivers/net/wireless/microchip/wilc1000/hif.c
1788
memcpy(cur_byte, params->head, params->head_len);
drivers/net/wireless/quantenna/qtnfmac/commands.c
309
s->beacon.head, s->beacon.head_len);
drivers/net/wireless/quantenna/qtnfmac/core.h
51
struct list_head head;
drivers/net/wireless/quantenna/qtnfmac/util.c
100
list_for_each_entry_safe(node, tmp, &list->head, list) {
drivers/net/wireless/quantenna/qtnfmac/util.c
105
INIT_LIST_HEAD(&list->head);
drivers/net/wireless/quantenna/qtnfmac/util.c
12
INIT_LIST_HEAD(&list->head);
drivers/net/wireless/quantenna/qtnfmac/util.c
24
list_for_each_entry(node, &list->head, list) {
drivers/net/wireless/quantenna/qtnfmac/util.c
40
list_for_each_entry(node, &list->head, list) {
drivers/net/wireless/quantenna/qtnfmac/util.c
67
list_add_tail(&node->list, &list->head);
drivers/net/wireless/quantenna/qtnfmac/util.h
31
return list_empty(&list->head);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1021
struct rcu_head head;
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
296
kfree_rcu(bar_entry, head);
drivers/net/wireless/realtek/rtw88/mac80211.c
663
struct sk_buff *head,
drivers/net/wireless/realtek/rtw88/pci.c
116
u8 *head = tx_ring->r.head;
drivers/net/wireless/realtek/rtw88/pci.c
123
dma_free_coherent(&pdev->dev, ring_sz, head, tx_ring->r.dma);
drivers/net/wireless/realtek/rtw88/pci.c
124
tx_ring->r.head = NULL;
drivers/net/wireless/realtek/rtw88/pci.c
152
u8 *head = rx_ring->r.head;
drivers/net/wireless/realtek/rtw88/pci.c
157
dma_free_coherent(&pdev->dev, ring_sz, head, rx_ring->r.dma);
drivers/net/wireless/realtek/rtw88/pci.c
185
u8 *head;
drivers/net/wireless/realtek/rtw88/pci.c
192
head = dma_alloc_coherent(&pdev->dev, ring_sz, &dma, GFP_KERNEL);
drivers/net/wireless/realtek/rtw88/pci.c
193
if (!head) {
drivers/net/wireless/realtek/rtw88/pci.c
199
tx_ring->r.head = head;
drivers/net/wireless/realtek/rtw88/pci.c
226
buf_desc = (struct rtw_pci_rx_buffer_desc *)(rx_ring->r.head +
drivers/net/wireless/realtek/rtw88/pci.c
245
buf_desc = (struct rtw_pci_rx_buffer_desc *)(rx_ring->r.head +
drivers/net/wireless/realtek/rtw88/pci.c
259
u8 *head;
drivers/net/wireless/realtek/rtw88/pci.c
265
head = dma_alloc_coherent(&pdev->dev, ring_sz, &dma, GFP_KERNEL);
drivers/net/wireless/realtek/rtw88/pci.c
266
if (!head) {
drivers/net/wireless/realtek/rtw88/pci.c
270
rx_ring->r.head = head;
drivers/net/wireless/realtek/rtw88/pci.c
308
dma_free_coherent(&pdev->dev, ring_sz, head, dma);
drivers/net/wireless/realtek/rtw88/pci.c
693
buf_desc = (struct rtw_pci_rx_buffer_desc *)(rx_ring->r.head +
drivers/net/wireless/realtek/rtw88/pci.h
177
u8 *head;
drivers/net/wireless/realtek/rtw88/pci.h
276
buf_desc = ring->r.head + ring->r.wp * size;
drivers/net/wireless/realtek/rtw89/pci.c
3408
u8 *head = wd_ring->head;
drivers/net/wireless/realtek/rtw89/pci.c
3414
dma_free_coherent(&pdev->dev, ring_sz, head, dma);
drivers/net/wireless/realtek/rtw89/pci.c
3415
wd_ring->head = NULL;
drivers/net/wireless/realtek/rtw89/pci.c
3422
tx_ring->bd_ring.head = NULL;
drivers/net/wireless/realtek/rtw89/pci.c
3442
dma_free_coherent(&pdev->dev, bd_pool->size, bd_pool->head, bd_pool->dma);
drivers/net/wireless/realtek/rtw89/pci.c
3468
rx_ring->bd_ring.head = NULL;
drivers/net/wireless/realtek/rtw89/pci.c
3484
dma_free_coherent(&pdev->dev, bd_pool->size, bd_pool->head, bd_pool->dma);
drivers/net/wireless/realtek/rtw89/pci.c
3530
u8 *head;
drivers/net/wireless/realtek/rtw89/pci.c
3542
head = dma_alloc_coherent(&pdev->dev, ring_sz, &dma, GFP_KERNEL);
drivers/net/wireless/realtek/rtw89/pci.c
3543
if (!head)
drivers/net/wireless/realtek/rtw89/pci.c
3547
wd_ring->head = head;
drivers/net/wireless/realtek/rtw89/pci.c
3556
cur_vaddr = head + page_offset;
drivers/net/wireless/realtek/rtw89/pci.c
3577
void *head, dma_addr_t dma)
drivers/net/wireless/realtek/rtw89/pci.c
3595
tx_ring->bd_ring.head = head;
drivers/net/wireless/realtek/rtw89/pci.c
3625
void *head;
drivers/net/wireless/realtek/rtw89/pci.c
3637
head = dma_alloc_coherent(&pdev->dev, pool_sz, &dma, GFP_KERNEL);
drivers/net/wireless/realtek/rtw89/pci.c
3638
if (!head)
drivers/net/wireless/realtek/rtw89/pci.c
3641
bd_pool->head = head;
drivers/net/wireless/realtek/rtw89/pci.c
3650
desc_size, len, i, head, dma);
drivers/net/wireless/realtek/rtw89/pci.c
3656
head += ring_sz;
drivers/net/wireless/realtek/rtw89/pci.c
3669
dma_free_coherent(&pdev->dev, bd_pool->size, bd_pool->head, bd_pool->dma);
drivers/net/wireless/realtek/rtw89/pci.c
3678
void *head, dma_addr_t dma)
drivers/net/wireless/realtek/rtw89/pci.c
3693
rx_ring->bd_ring.head = head;
drivers/net/wireless/realtek/rtw89/pci.c
3741
rx_ring->bd_ring.head = NULL;
drivers/net/wireless/realtek/rtw89/pci.c
3757
void *head;
drivers/net/wireless/realtek/rtw89/pci.c
3766
head = dma_alloc_coherent(&pdev->dev, pool_sz, &dma, GFP_KERNEL);
drivers/net/wireless/realtek/rtw89/pci.c
3767
if (!head)
drivers/net/wireless/realtek/rtw89/pci.c
3770
bd_pool->head = head;
drivers/net/wireless/realtek/rtw89/pci.c
3779
head, dma);
drivers/net/wireless/realtek/rtw89/pci.c
3785
head += ring_sz;
drivers/net/wireless/realtek/rtw89/pci.c
3798
dma_free_coherent(&pdev->dev, bd_pool->size, bd_pool->head, bd_pool->dma);
drivers/net/wireless/realtek/rtw89/pci.h
1544
void *head;
drivers/net/wireless/realtek/rtw89/pci.h
1556
void *head;
drivers/net/wireless/realtek/rtw89/pci.h
1562
void *head;
drivers/net/wireless/realtek/rtw89/pci.h
1649
u8 *head = bd_ring->head;
drivers/net/wireless/realtek/rtw89/pci.h
1653
return (struct rtw89_pci_rx_bd_32 *)(head + offset);
drivers/net/wireless/realtek/rtw89/pci.h
1682
struct rtw89_pci_tx_bd_32 *tx_bd, *head;
drivers/net/wireless/realtek/rtw89/pci.h
1684
head = bd_ring->head;
drivers/net/wireless/realtek/rtw89/pci.h
1685
tx_bd = head + bd_ring->wp;
drivers/net/wireless/realtek/rtw89/phy.c
2384
struct rtw89_txpwr_byrate *head,
drivers/net/wireless/realtek/rtw89/phy.c
2389
return &head->cck[desc->idx];
drivers/net/wireless/realtek/rtw89/phy.c
2391
return &head->ofdm[desc->idx];
drivers/net/wireless/realtek/rtw89/phy.c
2393
return &head->mcs[desc->ofdma][desc->nss][desc->idx];
drivers/net/wireless/realtek/rtw89/phy.c
2395
return &head->hedcm[desc->ofdma][desc->nss][desc->idx];
drivers/net/wireless/realtek/rtw89/phy.c
2397
return &head->offset[desc->idx];
drivers/net/wireless/realtek/rtw89/phy.c
2400
return &head->trap;
drivers/net/wireless/realtek/rtw89/phy.h
871
struct rtw89_txpwr_byrate *head,
drivers/net/wireless/st/cw1200/queue.c
110
cw1200_queue_register_post_gc(head, iter);
drivers/net/wireless/st/cw1200/queue.c
112
list_move_tail(&iter->head, &queue->free_pool);
drivers/net/wireless/st/cw1200/queue.c
18
struct list_head head;
drivers/net/wireless/st/cw1200/queue.c
194
list_add_tail(&queue->pool[i].head, &queue->free_pool);
drivers/net/wireless/st/cw1200/queue.c
209
list_for_each_entry_safe(item, tmp, &queue->pending, head) {
drivers/net/wireless/st/cw1200/queue.c
213
list_move_tail(&item->head, &queue->free_pool);
drivers/net/wireless/st/cw1200/queue.c
290
&queue->free_pool, struct cw1200_queue_item, head);
drivers/net/wireless/st/cw1200/queue.c
293
list_move_tail(&item->head, &queue->queue);
drivers/net/wireless/st/cw1200/queue.c
340
list_for_each_entry(item, &queue->queue, head) {
drivers/net/wireless/st/cw1200/queue.c
352
list_move_tail(&item->head, &queue->pending);
drivers/net/wireless/st/cw1200/queue.c
405
list_move(&item->head, &queue->queue);
drivers/net/wireless/st/cw1200/queue.c
446
list_move(&item->head, &queue->free_pool);
drivers/net/wireless/st/cw1200/queue.c
516
list_for_each_entry(item, &queue->pending, head) {
drivers/net/wireless/st/cw1200/queue.c
72
list_for_each_entry_safe(item, tmp, gc_list, head) {
drivers/net/wireless/st/cw1200/queue.c
73
list_del(&item->head);
drivers/net/wireless/st/cw1200/queue.c
86
list_add_tail(&gc_item->head, gc_list);
drivers/net/wireless/st/cw1200/queue.c
90
struct list_head *head,
drivers/net/wireless/st/cw1200/queue.c
97
list_for_each_entry_safe(iter, tmp, &queue->queue, head) {
drivers/net/wireless/virtual/mac80211_hwsim.c
6900
skb->data = skb->head;
drivers/net/wireless/virtual/mac80211_hwsim.c
6911
sg_init_one(sg, skb->head, skb_end_offset(skb));
drivers/net/wireless/virtual/mac80211_hwsim.c
6950
sg_init_one(sg, skb->head, skb_end_offset(skb));
drivers/net/wireless/virtual/mac80211_hwsim.c
964
sg_init_one(sg, skb->head, skb_end_offset(skb));
drivers/net/wireless/virtual/virt_wifi.c
605
struct list_head *head)
drivers/net/wireless/virtual/virt_wifi.c
619
unregister_netdevice_queue(dev, head);
drivers/net/wwan/iosm/iosm_ipc_imem.c
339
u32 head = 0, tail = 0;
drivers/net/wwan/iosm/iosm_ipc_imem.c
345
ipc_protocol_get_head_tail_index(ipc_imem->ipc_protocol, pipe, &head,
drivers/net/wwan/iosm/iosm_ipc_imem.c
398
u32 tail = 0, head = 0;
drivers/net/wwan/iosm/iosm_ipc_imem.c
405
ipc_protocol_get_head_tail_index(ipc_imem->ipc_protocol, pipe, &head,
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
13
u32 head = le32_to_cpu(ipc_protocol->p_ap_shm->msg_head);
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
14
u32 new_head = (head + 1) % IPC_MEM_MSG_ENTRIES;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
226
u32 head;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
237
head = le32_to_cpu(ipc_protocol->p_ap_shm->head_array[pipe->pipe_nr]);
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
241
if (head < tail)
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
242
free_elements = tail - head - 1;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
245
pipe->nr_of_entries - head + ((s32)tail - 1);
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
25
msg = &ipc_protocol->p_ap_shm->msg_ring[head];
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
255
td = &pipe->tdr_start[head];
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
265
pipe->skbr_start[head] = skb;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
274
head++;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
275
if (head >= pipe->nr_of_entries)
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
276
head = 0;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
279
cpu_to_le32(head);
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
282
if (pipe->old_head != head) {
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
285
pipe->old_head = head;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
29
*index = head;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
328
u32 head, new_head;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
335
head = le32_to_cpu(ipc_protocol->p_ap_shm->head_array[pipe->pipe_nr]);
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
338
new_head = head + 1;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
346
td = &pipe->tdr_start[head];
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
364
pipe->skbr_start[head] = skb;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
38
u32 head = le32_to_cpu(ipc_protocol->p_ap_shm->msg_head);
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
39
u32 new_head = (head + 1) % IPC_MEM_MSG_ENTRIES;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
434
struct ipc_pipe *pipe, u32 *head,
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
439
if (head)
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
440
*head = le32_to_cpu(ipc_ap_shm->head_array[pipe->pipe_nr]);
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
451
u32 head;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
455
head = le32_to_cpu(ipc_protocol->p_ap_shm->head_array[pipe->pipe_nr]);
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
464
while (head != tail) {
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.h
407
struct ipc_pipe *pipe, u32 *head,
drivers/net/wwan/iosm/iosm_ipc_wwan.c
194
struct list_head *head)
drivers/net/wwan/iosm/iosm_ipc_wwan.c
209
unregister_netdevice_queue(dev, head);
drivers/net/wwan/mhi_wwan_mbim.c
369
struct sk_buff *head = mbim->skbagg_head;
drivers/net/wwan/mhi_wwan_mbim.c
373
if (!head) {
drivers/net/wwan/mhi_wwan_mbim.c
378
if (!skb_shinfo(head)->frag_list)
drivers/net/wwan/mhi_wwan_mbim.c
379
skb_shinfo(head)->frag_list = skb;
drivers/net/wwan/mhi_wwan_mbim.c
383
head->len += skb->len;
drivers/net/wwan/mhi_wwan_mbim.c
384
head->data_len += skb->len;
drivers/net/wwan/mhi_wwan_mbim.c
385
head->truesize += skb->truesize;
drivers/net/wwan/mhi_wwan_mbim.c
574
struct list_head *head)
drivers/net/wwan/mhi_wwan_mbim.c
581
unregister_netdevice_queue(ndev, head);
drivers/net/wwan/qcom_bam_dmux.c
308
unsigned int head = needed_room(skb_headroom(skb), BAM_DMUX_HDR_SIZE);
drivers/net/wwan/qcom_bam_dmux.c
314
if (head || tail || skb_cloned(skb)) {
drivers/net/wwan/qcom_bam_dmux.c
315
ret = pskb_expand_head(skb, head, tail, GFP_ATOMIC);
drivers/net/wwan/t7xx/t7xx_netdev.c
315
static void t7xx_ccmni_wwan_dellink(void *ctxt, struct net_device *dev, struct list_head *head)
drivers/net/wwan/wwan_core.c
1174
static void wwan_rtnl_dellink(struct net_device *dev, struct list_head *head)
drivers/net/wwan/wwan_core.c
1186
wwandev->ops->dellink(wwandev->ops_ctxt, dev, head);
drivers/net/wwan/wwan_core.c
1188
unregister_netdevice_queue(dev, head);
drivers/net/wwan/wwan_core.c
519
struct sk_buff *skb, *head = NULL, *tail = NULL;
drivers/net/wwan/wwan_core.c
537
if (!head) {
drivers/net/wwan/wwan_core.c
538
head = skb;
drivers/net/wwan/wwan_core.c
541
skb_shinfo(head)->frag_list = skb;
drivers/net/wwan/wwan_core.c
546
head->data_len += skb->len;
drivers/net/wwan/wwan_core.c
547
head->len += skb->len;
drivers/net/wwan/wwan_core.c
548
head->truesize += skb->truesize;
drivers/net/wwan/wwan_core.c
552
ret = wwan_port_op_tx(port, head, false);
drivers/net/wwan/wwan_core.c
557
kfree_skb(head);
drivers/net/wwan/wwan_core.c
900
struct sk_buff *skb, *head = NULL, *tail = NULL;
drivers/net/wwan/wwan_core.c
918
if (!head) {
drivers/net/wwan/wwan_core.c
919
head = skb;
drivers/net/wwan/wwan_core.c
921
skb_shinfo(head)->frag_list = skb;
drivers/net/wwan/wwan_core.c
933
if (skb != head) {
drivers/net/wwan/wwan_core.c
934
head->data_len += skb->len;
drivers/net/wwan/wwan_core.c
935
head->len += skb->len;
drivers/net/wwan/wwan_core.c
936
head->truesize += skb->truesize;
drivers/net/wwan/wwan_core.c
940
ret = wwan_port_op_tx(port, head, !!(filp->f_flags & O_NONBLOCK));
drivers/net/wwan/wwan_core.c
945
kfree_skb(head);
drivers/net/xen-netfront.c
195
static void add_id_to_list(unsigned *head, unsigned short *list,
drivers/net/xen-netfront.c
198
list[id] = *head;
drivers/net/xen-netfront.c
199
*head = id;
drivers/net/xen-netfront.c
202
static unsigned short get_id_from_list(unsigned *head, unsigned short *list)
drivers/net/xen-netfront.c
204
unsigned int id = *head;
drivers/net/xen-netfront.c
207
*head = list[id];
drivers/net/xen-netfront.c
687
if (!IS_ALIGNED((uintptr_t)n->head, XEN_PAGE_SIZE)) {
drivers/net/xen-netfront.c
698
BUG_ON(skb_copy_bits(skb, -headerlen, n->head, headerlen + skb->len));
drivers/nfc/pn544/pn544.c
814
u8 head;
drivers/ntb/ntb_transport.c
2487
unsigned int head = qp->tx_index;
drivers/ntb/ntb_transport.c
2490
return tail >= head ? tail - head : qp->tx_max_entry + tail - head;
drivers/nvme/host/core.c
1005
if (op == nvme_cmd_write && ns->head->nr_plids) {
drivers/nvme/host/core.c
1008
if (WARN_ON_ONCE(write_stream > ns->head->nr_plids))
drivers/nvme/host/core.c
1012
dsmgmt |= ns->head->plids[write_stream - 1] << 16;
drivers/nvme/host/core.c
1022
cmnd->rw.nsid = cpu_to_le32(ns->head->ns_id);
drivers/nvme/host/core.c
1027
cpu_to_le64(nvme_sect_to_lba(ns->head, blk_rq_pos(req)));
drivers/nvme/host/core.c
1029
cpu_to_le16((blk_rq_bytes(req) >> ns->head->lba_shift) - 1);
drivers/nvme/host/core.c
1034
if (ns->head->ms) {
drivers/nvme/host/core.c
1042
if (WARN_ON_ONCE(!nvme_ns_has_pi(ns->head)))
drivers/nvme/host/core.c
1211
effects = le32_to_cpu(ns->head->effects->iocs[opcode]);
drivers/nvme/host/core.c
1779
if (WARN_ON_ONCE(nvme_ns_head_multipath(ns->head)))
drivers/nvme/host/core.c
1820
static bool nvme_init_integrity(struct nvme_ns_head *head,
drivers/nvme/host/core.c
1827
if (!head->ms)
drivers/nvme/host/core.c
1835
!(head->features & NVME_NS_METADATA_SUPPORTED))
drivers/nvme/host/core.c
1836
return nvme_ns_has_pi(head);
drivers/nvme/host/core.c
1838
switch (head->pi_type) {
drivers/nvme/host/core.c
1840
switch (head->guard_type) {
drivers/nvme/host/core.c
1857
switch (head->guard_type) {
drivers/nvme/host/core.c
1878
bi->metadata_size = head->ms;
drivers/nvme/host/core.c
1880
bi->pi_tuple_size = head->pi_size;
drivers/nvme/host/core.c
1890
if (ctrl->dmrsl && ctrl->dmrsl <= nvme_sect_to_lba(ns->head, UINT_MAX))
drivers/nvme/host/core.c
1892
nvme_lba_to_sect(ns->head, ctrl->dmrsl);
drivers/nvme/host/core.c
1938
static void nvme_configure_pi_elbas(struct nvme_ns_head *head,
drivers/nvme/host/core.c
1953
head->guard_type = guard_type;
drivers/nvme/host/core.c
1954
switch (head->guard_type) {
drivers/nvme/host/core.c
1956
head->pi_size = sizeof(struct crc64_pi_tuple);
drivers/nvme/host/core.c
1959
head->pi_size = sizeof(struct t10_pi_tuple);
drivers/nvme/host/core.c
1967
struct nvme_ns_head *head, struct nvme_id_ns *id,
drivers/nvme/host/core.c
1970
head->features &= ~(NVME_NS_METADATA_SUPPORTED | NVME_NS_EXT_LBAS);
drivers/nvme/host/core.c
1971
head->pi_type = 0;
drivers/nvme/host/core.c
1972
head->pi_size = 0;
drivers/nvme/host/core.c
1973
head->ms = le16_to_cpu(id->lbaf[nvme_lbaf_index(id->flbas)].ms);
drivers/nvme/host/core.c
1974
if (!head->ms || !(ctrl->ops->flags & NVME_F_METADATA_SUPPORTED))
drivers/nvme/host/core.c
1978
nvme_configure_pi_elbas(head, id, nvm);
drivers/nvme/host/core.c
1980
head->pi_size = sizeof(struct t10_pi_tuple);
drivers/nvme/host/core.c
1981
head->guard_type = NVME_NVM_NS_16B_GUARD;
drivers/nvme/host/core.c
1984
if (head->pi_size && head->ms >= head->pi_size)
drivers/nvme/host/core.c
1985
head->pi_type = id->dps & NVME_NS_DPS_PI_MASK;
drivers/nvme/host/core.c
1988
head->pi_type = 0;
drivers/nvme/host/core.c
1990
info->pi_offset = head->ms - head->pi_size;
drivers/nvme/host/core.c
2002
head->features |= NVME_NS_EXT_LBAS;
drivers/nvme/host/core.c
2013
if (ctrl->max_integrity_segments && nvme_ns_has_pi(head))
drivers/nvme/host/core.c
2014
head->features |= NVME_NS_METADATA_SUPPORTED;
drivers/nvme/host/core.c
2023
head->features |= NVME_NS_EXT_LBAS;
drivers/nvme/host/core.c
2025
head->features |= NVME_NS_METADATA_SUPPORTED;
drivers/nvme/host/core.c
2083
struct nvme_ns_head *head = ns->head;
drivers/nvme/host/core.c
2084
u32 bs = 1U << head->lba_shift;
drivers/nvme/host/core.c
2147
iob = nvme_lba_to_sect(ns->head, le16_to_cpu(id->noiob));
drivers/nvme/host/core.c
2265
struct nvme_ns_head *head = ns->head;
drivers/nvme/host/core.c
2278
if (head->nr_plids)
drivers/nvme/host/core.c
2301
c.imr.nsid = cpu_to_le32(head->ns_id);
drivers/nvme/host/core.c
2310
head->nr_plids = le16_to_cpu(ruhs->nruhsd);
drivers/nvme/host/core.c
2311
if (!head->nr_plids)
drivers/nvme/host/core.c
2314
head->plids = kcalloc(head->nr_plids, sizeof(*head->plids),
drivers/nvme/host/core.c
2316
if (!head->plids) {
drivers/nvme/host/core.c
2319
head->nr_plids);
drivers/nvme/host/core.c
2320
head->nr_plids = 0;
drivers/nvme/host/core.c
2325
for (i = 0; i < head->nr_plids; i++)
drivers/nvme/host/core.c
2326
head->plids[i] = le16_to_cpu(ruhs->ruhsd[i].pid);
drivers/nvme/host/core.c
2363
ns->head->ids.csi == NVME_CSI_ZNS) {
drivers/nvme/host/core.c
2378
ns->head->lba_shift = id->lbaf[lbaf].ds;
drivers/nvme/host/core.c
2379
ns->head->nuse = le64_to_cpu(id->nuse);
drivers/nvme/host/core.c
2380
capacity = nvme_lba_to_sect(ns->head, le64_to_cpu(id->nsze));
drivers/nvme/host/core.c
2382
nvme_configure_metadata(ns->ctrl, ns->head, id, nvm, info);
drivers/nvme/host/core.c
2389
ns->head->ids.csi == NVME_CSI_ZNS)
drivers/nvme/host/core.c
2406
if (!nvme_init_integrity(ns->head, &lim, info))
drivers/nvme/host/core.c
2409
lim.max_write_streams = ns->head->nr_plids;
drivers/nvme/host/core.c
2422
ns->head->features |= NVME_NS_DEAC;
drivers/nvme/host/core.c
2488
if (!ret && nvme_ns_head_multipath(ns->head)) {
drivers/nvme/host/core.c
2493
lim = queue_limits_start_update(ns->head->disk->queue);
drivers/nvme/host/core.c
2494
memflags = blk_mq_freeze_queue(ns->head->disk->queue);
drivers/nvme/host/core.c
2515
ns->head->disk->disk_name);
drivers/nvme/host/core.c
2517
ns->head->disk->flags |= GENHD_FL_HIDDEN;
drivers/nvme/host/core.c
2519
nvme_init_integrity(ns->head, &lim, info);
drivers/nvme/host/core.c
2522
ret = queue_limits_commit_update(ns->head->disk->queue, &lim);
drivers/nvme/host/core.c
2524
set_capacity_and_notify(ns->head->disk, get_capacity(ns->disk));
drivers/nvme/host/core.c
2525
set_disk_ro(ns->head->disk, nvme_ns_is_readonly(ns, info));
drivers/nvme/host/core.c
2528
blk_mq_unfreeze_queue(ns->head->disk->queue, memflags);
drivers/nvme/host/core.c
2537
struct nvme_ns_ids *ids = &ns->head->ids;
drivers/nvme/host/core.c
347
nvme_sect_to_lba(ns->head, blk_rq_pos(req)),
drivers/nvme/host/core.c
348
blk_rq_bytes(req) >> ns->head->lba_shift,
drivers/nvme/host/core.c
3873
ns->ctrl->instance, ns->head->instance);
drivers/nvme/host/core.c
3884
struct nvme_ns_head *head;
drivers/nvme/host/core.c
3885
size_t size = sizeof(*head);
drivers/nvme/host/core.c
3892
head = kzalloc(size, GFP_KERNEL);
drivers/nvme/host/core.c
3893
if (!head)
drivers/nvme/host/core.c
3898
head->instance = ret;
drivers/nvme/host/core.c
3899
INIT_LIST_HEAD(&head->list);
drivers/nvme/host/core.c
3900
ret = init_srcu_struct(&head->srcu);
drivers/nvme/host/core.c
3903
head->subsys = ctrl->subsys;
drivers/nvme/host/core.c
3904
head->ns_id = info->nsid;
drivers/nvme/host/core.c
3905
head->ids = info->ids;
drivers/nvme/host/core.c
3906
head->shared = info->is_shared;
drivers/nvme/host/core.c
3907
head->rotational = info->is_rotational;
drivers/nvme/host/core.c
3908
ratelimit_state_init(&head->rs_nuse, 5 * HZ, 1);
drivers/nvme/host/core.c
3909
ratelimit_set_flags(&head->rs_nuse, RATELIMIT_MSG_ON_RELEASE);
drivers/nvme/host/core.c
3910
kref_init(&head->ref);
drivers/nvme/host/core.c
3912
if (head->ids.csi) {
drivers/nvme/host/core.c
3913
ret = nvme_get_effects_log(ctrl, head->ids.csi, &head->effects);
drivers/nvme/host/core.c
3917
head->effects = ctrl->effects;
drivers/nvme/host/core.c
3919
ret = nvme_mpath_alloc_disk(ctrl, head);
drivers/nvme/host/core.c
3923
list_add_tail(&head->entry, &ctrl->subsys->nsheads);
drivers/nvme/host/core.c
3927
return head;
drivers/nvme/host/core.c
3929
cleanup_srcu_struct(&head->srcu);
drivers/nvme/host/core.c
3931
ida_free(&ctrl->subsys->ns_ida, head->instance);
drivers/nvme/host/core.c
3933
kfree(head);
drivers/nvme/host/core.c
3969
struct nvme_ns_head *head = NULL;
drivers/nvme/host/core.c
4011
head = nvme_find_ns_head(ctrl, info->nsid);
drivers/nvme/host/core.c
4012
if (!head) {
drivers/nvme/host/core.c
4020
head = nvme_alloc_ns_head(ctrl, info);
drivers/nvme/host/core.c
4021
if (IS_ERR(head)) {
drivers/nvme/host/core.c
4022
ret = PTR_ERR(head);
drivers/nvme/host/core.c
4027
if ((!info->is_shared || !head->shared) &&
drivers/nvme/host/core.c
4028
!list_empty(&head->list)) {
drivers/nvme/host/core.c
4034
if (!nvme_ns_ids_equal(&head->ids, &info->ids)) {
drivers/nvme/host/core.c
4050
list_add_tail_rcu(&ns->siblings, &head->list);
drivers/nvme/host/core.c
4051
ns->head = head;
drivers/nvme/host/core.c
4055
cancel_delayed_work(&head->remove_work);
drivers/nvme/host/core.c
4060
nvme_put_ns_head(head);
drivers/nvme/host/core.c
4074
if (ns->head->ns_id == nsid) {
drivers/nvme/host/core.c
4080
if (ns->head->ns_id > nsid)
drivers/nvme/host/core.c
4096
if (tmp->head->ns_id < ns->head->ns_id) {
drivers/nvme/host/core.c
4147
if (nvme_ns_head_multipath(ns->head)) {
drivers/nvme/host/core.c
4149
ctrl->instance, ns->head->instance);
drivers/nvme/host/core.c
4153
ns->head->instance);
drivers/nvme/host/core.c
4156
ns->head->instance);
drivers/nvme/host/core.c
4179
if (!nvme_ns_head_multipath(ns->head))
drivers/nvme/host/core.c
4196
if (list_empty(&ns->head->list)) {
drivers/nvme/host/core.c
4197
list_del_init(&ns->head->entry);
drivers/nvme/host/core.c
4206
if (ns->head->disk)
drivers/nvme/host/core.c
4211
nvme_put_ns_head(ns->head);
drivers/nvme/host/core.c
4212
nvme_put_ns_head(ns->head);
drivers/nvme/host/core.c
4234
synchronize_srcu(&ns->head->srcu);
drivers/nvme/host/core.c
4238
synchronize_srcu(&ns->head->srcu);
drivers/nvme/host/core.c
4242
if (list_empty(&ns->head->list)) {
drivers/nvme/host/core.c
4243
if (!nvme_mpath_queue_if_no_path(ns->head))
drivers/nvme/host/core.c
4244
list_del_init(&ns->head->entry);
drivers/nvme/host/core.c
4250
synchronize_srcu(&ns->head->srcu);
drivers/nvme/host/core.c
4252
if (!nvme_ns_head_multipath(ns->head))
drivers/nvme/host/core.c
4265
nvme_mpath_remove_disk(ns->head);
drivers/nvme/host/core.c
4283
if (!nvme_ns_ids_equal(&ns->head->ids, &info->ids)) {
drivers/nvme/host/core.c
4285
"identifiers changed for nsid %d\n", ns->head->ns_id);
drivers/nvme/host/core.c
430
req->__sector = nvme_lba_to_sect(ns->head,
drivers/nvme/host/core.c
4384
if (ns->head->ns_id > nsid) {
drivers/nvme/host/core.c
667
struct nvme_ns_head *head =
drivers/nvme/host/core.c
670
nvme_mpath_put_disk(head);
drivers/nvme/host/core.c
671
ida_free(&head->subsys->ns_ida, head->instance);
drivers/nvme/host/core.c
672
cleanup_srcu_struct(&head->srcu);
drivers/nvme/host/core.c
673
nvme_put_subsystem(head->subsys);
drivers/nvme/host/core.c
674
kfree(head->plids);
drivers/nvme/host/core.c
675
kfree(head);
drivers/nvme/host/core.c
678
bool nvme_tryget_ns_head(struct nvme_ns_head *head)
drivers/nvme/host/core.c
680
return kref_get_unless_zero(&head->ref);
drivers/nvme/host/core.c
683
void nvme_put_ns_head(struct nvme_ns_head *head)
drivers/nvme/host/core.c
685
kref_put(&head->ref, nvme_free_ns_head);
drivers/nvme/host/core.c
693
nvme_put_ns_head(ns->head);
drivers/nvme/host/core.c
726
logging_enabled = ns->head->passthru_err_log_enabled;
drivers/nvme/host/core.c
821
cmnd->common.nsid = cpu_to_le32(ns->head->ns_id);
drivers/nvme/host/core.c
852
u64 slba = nvme_sect_to_lba(ns->head, blk_rq_pos(req));
drivers/nvme/host/core.c
853
u32 nlb = blk_rq_sectors(req) >> (ns->head->lba_shift - 9);
drivers/nvme/host/core.c
861
u64 slba = nvme_sect_to_lba(ns->head,
drivers/nvme/host/core.c
863
u32 nlb = bio->bi_iter.bi_size >> ns->head->lba_shift;
drivers/nvme/host/core.c
884
cmnd->dsm.nsid = cpu_to_le32(ns->head->ns_id);
drivers/nvme/host/core.c
907
switch (ns->head->pi_type) {
drivers/nvme/host/core.c
916
switch (ns->head->guard_type) {
drivers/nvme/host/core.c
942
cmnd->write_zeroes.nsid = cpu_to_le32(ns->head->ns_id);
drivers/nvme/host/core.c
944
cpu_to_le64(nvme_sect_to_lba(ns->head, blk_rq_pos(req)));
drivers/nvme/host/core.c
946
cpu_to_le16((blk_rq_bytes(req) >> ns->head->lba_shift) - 1);
drivers/nvme/host/core.c
949
(ns->head->features & NVME_NS_DEAC))
drivers/nvme/host/core.c
952
if (nvme_ns_has_pi(ns->head)) {
drivers/nvme/host/ioctl.c
227
length = (io.nblocks + 1) << ns->head->lba_shift;
drivers/nvme/host/ioctl.c
230
(ns->head->ms == ns->head->pi_size)) {
drivers/nvme/host/ioctl.c
240
meta_len = (io.nblocks + 1) * ns->head->ms;
drivers/nvme/host/ioctl.c
244
if (ns->head->features & NVME_NS_EXT_LBAS) {
drivers/nvme/host/ioctl.c
255
c.rw.nsid = cpu_to_le32(ns->head->ns_id);
drivers/nvme/host/ioctl.c
271
if (ns && nsid != ns->head->ns_id) {
drivers/nvme/host/ioctl.c
274
current->comm, nsid, ns->head->ns_id);
drivers/nvme/host/ioctl.c
587
return ns->head->ns_id;
drivers/nvme/host/ioctl.c
694
void __user *argp, struct nvme_ns_head *head, int srcu_idx,
drivers/nvme/host/ioctl.c
696
__releases(&head->srcu)
drivers/nvme/host/ioctl.c
702
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/ioctl.c
712
struct nvme_ns_head *head = bdev->bd_disk->private_data;
drivers/nvme/host/ioctl.c
722
srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/ioctl.c
723
ns = nvme_find_path(head);
drivers/nvme/host/ioctl.c
733
return nvme_ns_head_ctrl_ioctl(ns, cmd, argp, head, srcu_idx,
drivers/nvme/host/ioctl.c
738
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/ioctl.c
747
struct nvme_ns_head *head =
drivers/nvme/host/ioctl.c
753
srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/ioctl.c
754
ns = nvme_find_path(head);
drivers/nvme/host/ioctl.c
759
return nvme_ns_head_ctrl_ioctl(ns, cmd, argp, head, srcu_idx,
drivers/nvme/host/ioctl.c
764
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/ioctl.c
772
struct nvme_ns_head *head = container_of(cdev, struct nvme_ns_head, cdev);
drivers/nvme/host/ioctl.c
773
int srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/ioctl.c
774
struct nvme_ns *ns = nvme_find_path(head);
drivers/nvme/host/ioctl.c
779
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/multipath.c
1094
if (ns->head->subsys->iopolicy != NVME_IOPOLICY_QD)
drivers/nvme/host/multipath.c
1108
struct nvme_ns_head *head = ns->head;
drivers/nvme/host/multipath.c
1110
if (head->subsys->iopolicy != NVME_IOPOLICY_NUMA)
drivers/nvme/host/multipath.c
1115
srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/multipath.c
1117
current_ns = srcu_dereference(head->current_path[node],
drivers/nvme/host/multipath.c
1118
&head->srcu);
drivers/nvme/host/multipath.c
1122
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/multipath.c
1132
struct nvme_ns_head *head = disk->private_data;
drivers/nvme/host/multipath.c
1135
mutex_lock(&head->subsys->lock);
drivers/nvme/host/multipath.c
1136
ret = sysfs_emit(buf, "%u\n", head->delayed_removal_secs);
drivers/nvme/host/multipath.c
1137
mutex_unlock(&head->subsys->lock);
drivers/nvme/host/multipath.c
1145
struct nvme_ns_head *head = disk->private_data;
drivers/nvme/host/multipath.c
1153
mutex_lock(&head->subsys->lock);
drivers/nvme/host/multipath.c
1154
head->delayed_removal_secs = sec;
drivers/nvme/host/multipath.c
1156
set_bit(NVME_NSHEAD_QUEUE_IF_NO_PATH, &head->flags);
drivers/nvme/host/multipath.c
1158
clear_bit(NVME_NSHEAD_QUEUE_IF_NO_PATH, &head->flags);
drivers/nvme/host/multipath.c
1159
mutex_unlock(&head->subsys->lock);
drivers/nvme/host/multipath.c
1164
synchronize_srcu(&head->srcu);
drivers/nvme/host/multipath.c
1183
void nvme_mpath_add_sysfs_link(struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
1194
if (!test_bit(GD_ADDED, &head->disk->state))
drivers/nvme/host/multipath.c
1197
kobj = &disk_to_dev(head->disk)->kobj;
drivers/nvme/host/multipath.c
1203
srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/multipath.c
1205
list_for_each_entry_srcu(ns, &head->list, siblings,
drivers/nvme/host/multipath.c
1206
srcu_read_lock_held(&head->srcu)) {
drivers/nvme/host/multipath.c
1237
dev_err(disk_to_dev(ns->head->disk),
drivers/nvme/host/multipath.c
1244
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/multipath.c
1256
kobj = &disk_to_dev(ns->head->disk)->kobj;
drivers/nvme/host/multipath.c
1288
if (blk_queue_is_zoned(ns->queue) && ns->head->disk)
drivers/nvme/host/multipath.c
1289
ns->head->disk->nr_zones = ns->disk->nr_zones;
drivers/nvme/host/multipath.c
1293
void nvme_mpath_remove_disk(struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
1297
if (!head->disk)
drivers/nvme/host/multipath.c
1300
mutex_lock(&head->subsys->lock);
drivers/nvme/host/multipath.c
1310
if (!list_empty(&head->list))
drivers/nvme/host/multipath.c
1317
if (head->delayed_removal_secs && try_module_get(THIS_MODULE)) {
drivers/nvme/host/multipath.c
1318
mod_delayed_work(nvme_wq, &head->remove_work,
drivers/nvme/host/multipath.c
1319
head->delayed_removal_secs * HZ);
drivers/nvme/host/multipath.c
1321
list_del_init(&head->entry);
drivers/nvme/host/multipath.c
1325
mutex_unlock(&head->subsys->lock);
drivers/nvme/host/multipath.c
1327
nvme_remove_head(head);
drivers/nvme/host/multipath.c
1330
void nvme_mpath_put_disk(struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
1332
if (!head->disk)
drivers/nvme/host/multipath.c
1335
kblockd_schedule_work(&head->requeue_work);
drivers/nvme/host/multipath.c
1336
flush_work(&head->requeue_work);
drivers/nvme/host/multipath.c
1337
flush_work(&head->partition_scan_work);
drivers/nvme/host/multipath.c
1338
put_disk(head->disk);
drivers/nvme/host/multipath.c
156
spin_lock_irqsave(&ns->head->requeue_lock, flags);
drivers/nvme/host/multipath.c
158
bio_set_dev(bio, ns->head->disk->part0);
drivers/nvme/host/multipath.c
172
blk_steal_bios(&ns->head->requeue_list, req);
drivers/nvme/host/multipath.c
173
spin_unlock_irqrestore(&ns->head->requeue_lock, flags);
drivers/nvme/host/multipath.c
177
kblockd_schedule_work(&ns->head->requeue_work);
drivers/nvme/host/multipath.c
183
struct gendisk *disk = ns->head->disk;
drivers/nvme/host/multipath.c
185
if ((READ_ONCE(ns->head->subsys->iopolicy) == NVME_IOPOLICY_QD) &&
drivers/nvme/host/multipath.c
210
bdev_end_io_acct(ns->head->disk->part0, req_op(rq),
drivers/nvme/host/multipath.c
223
if (!ns->head->disk)
drivers/nvme/host/multipath.c
225
kblockd_schedule_work(&ns->head->requeue_work);
drivers/nvme/host/multipath.c
227
disk_uevent(ns->head->disk, KOBJ_CHANGE);
drivers/nvme/host/multipath.c
243
struct nvme_ns_head *head = ns->head;
drivers/nvme/host/multipath.c
247
if (!head)
drivers/nvme/host/multipath.c
251
if (ns == rcu_access_pointer(head->current_path[node])) {
drivers/nvme/host/multipath.c
252
rcu_assign_pointer(head->current_path[node], NULL);
drivers/nvme/host/multipath.c
269
kblockd_schedule_work(&ns->head->requeue_work);
drivers/nvme/host/multipath.c
276
struct nvme_ns_head *head = ns->head;
drivers/nvme/host/multipath.c
277
sector_t capacity = get_capacity(head->disk);
drivers/nvme/host/multipath.c
281
srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/multipath.c
282
list_for_each_entry_srcu(ns, &head->list, siblings,
drivers/nvme/host/multipath.c
283
srcu_read_lock_held(&head->srcu)) {
drivers/nvme/host/multipath.c
287
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/multipath.c
290
rcu_assign_pointer(head->current_path[node], NULL);
drivers/nvme/host/multipath.c
291
kblockd_schedule_work(&head->requeue_work);
drivers/nvme/host/multipath.c
311
static struct nvme_ns *__nvme_find_path(struct nvme_ns_head *head, int node)
drivers/nvme/host/multipath.c
316
list_for_each_entry_srcu(ns, &head->list, siblings,
drivers/nvme/host/multipath.c
317
srcu_read_lock_held(&head->srcu)) {
drivers/nvme/host/multipath.c
322
READ_ONCE(head->subsys->iopolicy) == NVME_IOPOLICY_NUMA)
drivers/nvme/host/multipath.c
348
rcu_assign_pointer(head->current_path[node], found);
drivers/nvme/host/multipath.c
352
static struct nvme_ns *nvme_next_ns(struct nvme_ns_head *head,
drivers/nvme/host/multipath.c
355
ns = list_next_or_null_rcu(&head->list, &ns->siblings, struct nvme_ns,
drivers/nvme/host/multipath.c
359
return list_first_or_null_rcu(&head->list, struct nvme_ns, siblings);
drivers/nvme/host/multipath.c
362
static struct nvme_ns *nvme_round_robin_path(struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
366
struct nvme_ns *old = srcu_dereference(head->current_path[node],
drivers/nvme/host/multipath.c
367
&head->srcu);
drivers/nvme/host/multipath.c
370
return __nvme_find_path(head, node);
drivers/nvme/host/multipath.c
372
if (list_is_singular(&head->list)) {
drivers/nvme/host/multipath.c
378
for (ns = nvme_next_ns(head, old);
drivers/nvme/host/multipath.c
380
ns = nvme_next_ns(head, ns)) {
drivers/nvme/host/multipath.c
406
rcu_assign_pointer(head->current_path[node], found);
drivers/nvme/host/multipath.c
410
static struct nvme_ns *nvme_queue_depth_path(struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
416
list_for_each_entry_srcu(ns, &head->list, siblings,
drivers/nvme/host/multipath.c
417
srcu_read_lock_held(&head->srcu)) {
drivers/nvme/host/multipath.c
453
static struct nvme_ns *nvme_numa_path(struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
458
ns = srcu_dereference(head->current_path[node], &head->srcu);
drivers/nvme/host/multipath.c
460
return __nvme_find_path(head, node);
drivers/nvme/host/multipath.c
462
return __nvme_find_path(head, node);
drivers/nvme/host/multipath.c
466
inline struct nvme_ns *nvme_find_path(struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
468
switch (READ_ONCE(head->subsys->iopolicy)) {
drivers/nvme/host/multipath.c
470
return nvme_queue_depth_path(head);
drivers/nvme/host/multipath.c
472
return nvme_round_robin_path(head);
drivers/nvme/host/multipath.c
474
return nvme_numa_path(head);
drivers/nvme/host/multipath.c
478
static bool nvme_available_path(struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
482
if (!test_bit(NVME_NSHEAD_DISK_LIVE, &head->flags))
drivers/nvme/host/multipath.c
485
list_for_each_entry_srcu(ns, &head->list, siblings,
drivers/nvme/host/multipath.c
486
srcu_read_lock_held(&head->srcu)) {
drivers/nvme/host/multipath.c
508
return nvme_mpath_queue_if_no_path(head);
drivers/nvme/host/multipath.c
513
struct nvme_ns_head *head = bio->bi_bdev->bd_disk->private_data;
drivers/nvme/host/multipath.c
514
struct device *dev = disk_to_dev(head->disk);
drivers/nvme/host/multipath.c
527
srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/multipath.c
528
ns = nvme_find_path(head);
drivers/nvme/host/multipath.c
532
trace_block_bio_remap(bio, disk_devt(ns->head->disk),
drivers/nvme/host/multipath.c
535
} else if (nvme_available_path(head)) {
drivers/nvme/host/multipath.c
538
spin_lock_irq(&head->requeue_lock);
drivers/nvme/host/multipath.c
539
bio_list_add(&head->requeue_list, bio);
drivers/nvme/host/multipath.c
540
spin_unlock_irq(&head->requeue_lock);
drivers/nvme/host/multipath.c
547
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/multipath.c
565
struct nvme_ns_head *head = disk->private_data;
drivers/nvme/host/multipath.c
569
srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/multipath.c
570
ns = nvme_find_path(head);
drivers/nvme/host/multipath.c
573
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/multipath.c
581
struct nvme_ns_head *head = disk->private_data;
drivers/nvme/host/multipath.c
585
srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/multipath.c
586
ns = nvme_find_path(head);
drivers/nvme/host/multipath.c
589
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/multipath.c
637
static int nvme_add_ns_head_cdev(struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
641
head->cdev_device.parent = &head->subsys->dev;
drivers/nvme/host/multipath.c
642
ret = dev_set_name(&head->cdev_device, "ng%dn%d",
drivers/nvme/host/multipath.c
643
head->subsys->instance, head->instance);
drivers/nvme/host/multipath.c
646
ret = nvme_cdev_add(&head->cdev, &head->cdev_device,
drivers/nvme/host/multipath.c
653
struct nvme_ns_head *head =
drivers/nvme/host/multipath.c
657
&head->disk->state)))
drivers/nvme/host/multipath.c
660
mutex_lock(&head->disk->open_mutex);
drivers/nvme/host/multipath.c
661
bdev_disk_changed(head->disk, false);
drivers/nvme/host/multipath.c
662
mutex_unlock(&head->disk->open_mutex);
drivers/nvme/host/multipath.c
667
struct nvme_ns_head *head =
drivers/nvme/host/multipath.c
671
spin_lock_irq(&head->requeue_lock);
drivers/nvme/host/multipath.c
672
next = bio_list_get(&head->requeue_list);
drivers/nvme/host/multipath.c
673
spin_unlock_irq(&head->requeue_lock);
drivers/nvme/host/multipath.c
683
static void nvme_remove_head(struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
685
if (test_and_clear_bit(NVME_NSHEAD_DISK_LIVE, &head->flags)) {
drivers/nvme/host/multipath.c
690
kblockd_schedule_work(&head->requeue_work);
drivers/nvme/host/multipath.c
692
nvme_cdev_del(&head->cdev, &head->cdev_device);
drivers/nvme/host/multipath.c
693
synchronize_srcu(&head->srcu);
drivers/nvme/host/multipath.c
694
del_gendisk(head->disk);
drivers/nvme/host/multipath.c
696
nvme_put_ns_head(head);
drivers/nvme/host/multipath.c
701
struct nvme_ns_head *head = container_of(to_delayed_work(work),
drivers/nvme/host/multipath.c
705
mutex_lock(&head->subsys->lock);
drivers/nvme/host/multipath.c
706
if (list_empty(&head->list)) {
drivers/nvme/host/multipath.c
707
list_del_init(&head->entry);
drivers/nvme/host/multipath.c
710
mutex_unlock(&head->subsys->lock);
drivers/nvme/host/multipath.c
712
nvme_remove_head(head);
drivers/nvme/host/multipath.c
717
int nvme_mpath_alloc_disk(struct nvme_ctrl *ctrl, struct nvme_ns_head *head)
drivers/nvme/host/multipath.c
721
mutex_init(&head->lock);
drivers/nvme/host/multipath.c
722
bio_list_init(&head->requeue_list);
drivers/nvme/host/multipath.c
723
spin_lock_init(&head->requeue_lock);
drivers/nvme/host/multipath.c
724
INIT_WORK(&head->requeue_work, nvme_requeue_work);
drivers/nvme/host/multipath.c
725
INIT_WORK(&head->partition_scan_work, nvme_partition_scan_work);
drivers/nvme/host/multipath.c
726
INIT_DELAYED_WORK(&head->remove_work, nvme_remove_head_work);
drivers/nvme/host/multipath.c
727
head->delayed_removal_secs = 0;
drivers/nvme/host/multipath.c
744
if (!nvme_is_unique_nsid(ctrl, head))
drivers/nvme/host/multipath.c
751
if (head->ids.csi == NVME_CSI_ZNS)
drivers/nvme/host/multipath.c
754
head->disk = blk_alloc_disk(&lim, ctrl->numa_node);
drivers/nvme/host/multipath.c
755
if (IS_ERR(head->disk))
drivers/nvme/host/multipath.c
756
return PTR_ERR(head->disk);
drivers/nvme/host/multipath.c
757
head->disk->fops = &nvme_ns_head_ops;
drivers/nvme/host/multipath.c
758
head->disk->private_data = head;
drivers/nvme/host/multipath.c
768
set_bit(GD_SUPPRESS_PART_SCAN, &head->disk->state);
drivers/nvme/host/multipath.c
769
sprintf(head->disk->disk_name, "nvme%dn%d",
drivers/nvme/host/multipath.c
770
ctrl->subsys->instance, head->instance);
drivers/nvme/host/multipath.c
771
nvme_tryget_ns_head(head);
drivers/nvme/host/multipath.c
777
struct nvme_ns_head *head = ns->head;
drivers/nvme/host/multipath.c
780
if (!head->disk)
drivers/nvme/host/multipath.c
788
if (!test_and_set_bit(NVME_NSHEAD_DISK_LIVE, &head->flags)) {
drivers/nvme/host/multipath.c
789
rc = device_add_disk(&head->subsys->dev, head->disk,
drivers/nvme/host/multipath.c
792
clear_bit(NVME_NSHEAD_DISK_LIVE, &head->flags);
drivers/nvme/host/multipath.c
795
nvme_add_ns_head_cdev(head);
drivers/nvme/host/multipath.c
796
queue_work(nvme_wq, &head->partition_scan_work);
drivers/nvme/host/multipath.c
799
nvme_mpath_add_sysfs_link(ns->head);
drivers/nvme/host/multipath.c
801
mutex_lock(&head->lock);
drivers/nvme/host/multipath.c
805
srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/multipath.c
807
__nvme_find_path(head, node);
drivers/nvme/host/multipath.c
808
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/multipath.c
810
mutex_unlock(&head->lock);
drivers/nvme/host/multipath.c
812
synchronize_srcu(&head->srcu);
drivers/nvme/host/multipath.c
813
kblockd_schedule_work(&head->requeue_work);
drivers/nvme/host/multipath.c
899
if (test_bit(NVME_NSHEAD_DISK_LIVE, &ns->head->flags))
drivers/nvme/host/multipath.c
900
nvme_mpath_add_sysfs_link(ns->head);
drivers/nvme/host/multipath.c
928
if (ns->head->ns_id < nsid)
drivers/nvme/host/multipath.c
930
if (ns->head->ns_id == nsid)
drivers/nvme/host/multipath.c
934
if (ns->head->ns_id > nsid)
drivers/nvme/host/nvme.h
1014
struct nvme_ns *nvme_find_path(struct nvme_ns_head *head);
drivers/nvme/host/nvme.h
1027
int nvme_mpath_alloc_disk(struct nvme_ctrl *ctrl,struct nvme_ns_head *head);
drivers/nvme/host/nvme.h
1031
void nvme_mpath_put_disk(struct nvme_ns_head *head);
drivers/nvme/host/nvme.h
1040
void nvme_mpath_remove_disk(struct nvme_ns_head *head);
drivers/nvme/host/nvme.h
1049
trace_block_bio_complete(ns->head->disk->queue, req->bio);
drivers/nvme/host/nvme.h
1064
static inline bool nvme_mpath_queue_if_no_path(struct nvme_ns_head *head)
drivers/nvme/host/nvme.h
1066
if (test_bit(NVME_NSHEAD_QUEUE_IF_NO_PATH, &head->flags))
drivers/nvme/host/nvme.h
1083
struct nvme_ns_head *head)
drivers/nvme/host/nvme.h
1090
static inline void nvme_mpath_put_disk(struct nvme_ns_head *head)
drivers/nvme/host/nvme.h
1109
static inline void nvme_mpath_remove_disk(struct nvme_ns_head *head)
drivers/nvme/host/nvme.h
1157
static inline bool nvme_mpath_queue_if_no_path(struct nvme_ns_head *head)
drivers/nvme/host/nvme.h
572
static inline bool nvme_ns_head_multipath(struct nvme_ns_head *head)
drivers/nvme/host/nvme.h
574
return IS_ENABLED(CONFIG_NVME_MULTIPATH) && head->disk;
drivers/nvme/host/nvme.h
595
struct nvme_ns_head *head;
drivers/nvme/host/nvme.h
611
static inline bool nvme_ns_has_pi(struct nvme_ns_head *head)
drivers/nvme/host/nvme.h
613
return head->pi_type && head->ms == head->pi_size;
drivers/nvme/host/nvme.h
744
static inline u64 nvme_sect_to_lba(struct nvme_ns_head *head, sector_t sector)
drivers/nvme/host/nvme.h
746
return sector >> (head->lba_shift - SECTOR_SHIFT);
drivers/nvme/host/nvme.h
752
static inline sector_t nvme_lba_to_sect(struct nvme_ns_head *head, u64 lba)
drivers/nvme/host/nvme.h
754
return lba << (head->lba_shift - SECTOR_SHIFT);
drivers/nvme/host/nvme.h
936
struct nvme_ns_head *head)
drivers/nvme/host/nvme.h
938
return head->shared ||
drivers/nvme/host/nvme.h
979
bool nvme_tryget_ns_head(struct nvme_ns_head *head);
drivers/nvme/host/nvme.h
980
void nvme_put_ns_head(struct nvme_ns_head *head);
drivers/nvme/host/pci.c
1517
u16 head = nvmeq->cq_head;
drivers/nvme/host/pci.c
1519
if (nvme_dbbuf_update_and_check_event(head, nvmeq->dbbuf_cq_db,
drivers/nvme/host/pci.c
1521
writel(head, nvmeq->q_db + nvmeq->dev->db_stride);
drivers/nvme/host/pr.c
55
struct nvme_ns_head *head = bdev->bd_disk->private_data;
drivers/nvme/host/pr.c
56
int srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/pr.c
57
struct nvme_ns *ns = nvme_find_path(head);
drivers/nvme/host/pr.c
61
c->common.nsid = cpu_to_le32(ns->head->ns_id);
drivers/nvme/host/pr.c
64
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/pr.c
71
c->common.nsid = cpu_to_le32(ns->head->ns_id);
drivers/nvme/host/rdma.c
1430
nvme_rdma_set_sig_attrs(bi, c, req->mr->sig_attrs, ns->head->pi_type);
drivers/nvme/host/rdma.c
1451
xfer_len += (xfer_len >> bi->interval_exp) * ns->head->pi_size;
drivers/nvme/host/rdma.c
2033
nvme_ns_has_pi(ns->head))
drivers/nvme/host/sysfs.c
106
struct nvme_ns_head *head = dev_to_ns_head(dev);
drivers/nvme/host/sysfs.c
107
struct nvme_ns_ids *ids = &head->ids;
drivers/nvme/host/sysfs.c
108
struct nvme_subsystem *subsys = head->subsys;
drivers/nvme/host/sysfs.c
130
head->ns_id);
drivers/nvme/host/sysfs.c
186
static int ns_head_update_nuse(struct nvme_ns_head *head)
drivers/nvme/host/sysfs.c
193
if (!__ratelimit(&head->rs_nuse))
drivers/nvme/host/sysfs.c
196
srcu_idx = srcu_read_lock(&head->srcu);
drivers/nvme/host/sysfs.c
197
ns = nvme_find_path(head);
drivers/nvme/host/sysfs.c
201
ret = nvme_identify_ns(ns->ctrl, head->ns_id, &id);
drivers/nvme/host/sysfs.c
205
head->nuse = le64_to_cpu(id->nuse);
drivers/nvme/host/sysfs.c
209
srcu_read_unlock(&head->srcu, srcu_idx);
drivers/nvme/host/sysfs.c
219
if (!__ratelimit(&ns->head->rs_nuse))
drivers/nvme/host/sysfs.c
222
ret = nvme_identify_ns(ns->ctrl, ns->head->ns_id, &id);
drivers/nvme/host/sysfs.c
226
ns->head->nuse = le64_to_cpu(id->nuse);
drivers/nvme/host/sysfs.c
234
struct nvme_ns_head *head = dev_to_ns_head(dev);
drivers/nvme/host/sysfs.c
239
ret = ns_head_update_nuse(head);
drivers/nvme/host/sysfs.c
245
return sysfs_emit(buf, "%llu\n", head->nuse);
drivers/nvme/host/sysfs.c
69
return nvme_get_ns_from_dev(dev)->head;
drivers/nvme/host/sysfs.c
75
struct nvme_ns_head *head = dev_to_ns_head(dev);
drivers/nvme/host/sysfs.c
77
return sysfs_emit(buf, head->passthru_err_log_enabled ? "on\n" : "off\n");
drivers/nvme/host/sysfs.c
83
struct nvme_ns_head *head = dev_to_ns_head(dev);
drivers/nvme/host/sysfs.c
90
head->passthru_err_log_enabled = passthru_err_log_enabled;
drivers/nvme/host/zns.c
115
lim->chunk_sectors = ns->head->zsze =
drivers/nvme/host/zns.c
116
nvme_lba_to_sect(ns->head, zi->zone_size);
drivers/nvme/host/zns.c
130
get_capacity(ns->disk) >> ilog2(ns->head->zsze));
drivers/nvme/host/zns.c
154
struct nvme_ns_head *head = ns->head;
drivers/nvme/host/zns.c
164
zone.len = head->zsze;
drivers/nvme/host/zns.c
165
zone.capacity = nvme_lba_to_sect(head, le64_to_cpu(entry->zcap));
drivers/nvme/host/zns.c
166
zone.start = nvme_lba_to_sect(head, le64_to_cpu(entry->zslba));
drivers/nvme/host/zns.c
170
zone.wp = nvme_lba_to_sect(head, le64_to_cpu(entry->wp));
drivers/nvme/host/zns.c
184
if (ns->head->ids.csi != NVME_CSI_ZNS)
drivers/nvme/host/zns.c
192
c.zmr.nsid = cpu_to_le32(ns->head->ns_id);
drivers/nvme/host/zns.c
198
sector &= ~(ns->head->zsze - 1);
drivers/nvme/host/zns.c
202
c.zmr.slba = cpu_to_le64(nvme_sect_to_lba(ns->head, sector));
drivers/nvme/host/zns.c
222
sector += ns->head->zsze * nz;
drivers/nvme/host/zns.c
240
c->zms.nsid = cpu_to_le32(ns->head->ns_id);
drivers/nvme/host/zns.c
241
c->zms.slba = cpu_to_le64(nvme_sect_to_lba(ns->head, blk_rq_pos(req)));
drivers/nvme/host/zns.c
41
struct nvme_effects_log *log = ns->head->effects;
drivers/nvme/host/zns.c
52
ns->head->ns_id);
drivers/nvme/host/zns.c
57
ns->head->ns_id);
drivers/nvme/host/zns.c
72
c.identify.nsid = cpu_to_le32(ns->head->ns_id);
drivers/nvme/host/zns.c
87
le16_to_cpu(id->zoc), ns->head->ns_id);
drivers/nvme/host/zns.c
96
zi->zone_size, ns->head->ns_id);
drivers/nvme/target/pci-epf.c
1284
cq->head = 0;
drivers/nvme/target/pci-epf.c
1383
sq->head = 0;
drivers/nvme/target/pci-epf.c
1668
u16 head = sq->head;
drivers/nvme/target/pci-epf.c
1671
while (head != sq->tail && (!ctrl->sq_ab || n < ctrl->sq_ab)) {
drivers/nvme/target/pci-epf.c
1678
sq->pci_addr + head * sq->qes,
drivers/nvme/target/pci-epf.c
1687
sq->qid, head, sq->tail,
drivers/nvme/target/pci-epf.c
1690
head++;
drivers/nvme/target/pci-epf.c
1691
if (head == sq->depth)
drivers/nvme/target/pci-epf.c
1692
head = 0;
drivers/nvme/target/pci-epf.c
1693
WRITE_ONCE(sq->head, head);
drivers/nvme/target/pci-epf.c
1767
cq->head = nvmet_pci_epf_bar_read32(ctrl, cq->db);
drivers/nvme/target/pci-epf.c
1768
if (cq->head == cq->tail + 1) {
drivers/nvme/target/pci-epf.c
1791
cqe->sq_head = cpu_to_le16(READ_ONCE(iod->sq->head));
drivers/nvme/target/pci-epf.c
1799
le64_to_cpu(cqe->result.u64), cq->head, cq->tail,
drivers/nvme/target/pci-epf.c
94
u16 head;
drivers/nvmem/rave-sp-eeprom.c
227
unsigned int head;
drivers/nvmem/rave-sp-eeprom.c
232
head = offset % RAVE_SP_EEPROM_PAGE_SIZE;
drivers/nvmem/rave-sp-eeprom.c
242
if (unlikely(head)) {
drivers/nvmem/rave-sp-eeprom.c
243
chunk = RAVE_SP_EEPROM_PAGE_SIZE - head;
drivers/nvmem/rave-sp-eeprom.c
250
head = 0;
drivers/of/pdt.c
116
struct property *head, *tail;
drivers/of/pdt.c
118
head = tail = of_pdt_build_one_prop(node, NULL,
drivers/of/pdt.c
129
return head;
drivers/opp/core.c
1566
BLOCKING_INIT_NOTIFIER_HEAD(&opp_table->head);
drivers/opp/core.c
1772
blocking_notifier_call_chain(&opp_table->head, OPP_EVENT_REMOVE, opp);
drivers/opp/core.c
2015
struct list_head **head)
drivers/opp/core.c
2031
*head = &opp->node;
drivers/opp/core.c
2080
struct list_head *head;
drivers/opp/core.c
2084
head = &opp_table->opp_list;
drivers/opp/core.c
2086
ret = _opp_is_duplicate(dev, new_opp, opp_table, &head);
drivers/opp/core.c
2090
list_add(&new_opp->node, head);
drivers/opp/core.c
2171
blocking_notifier_call_chain(&opp_table->head, OPP_EVENT_ADD, new_opp);
drivers/opp/core.c
2910
blocking_notifier_call_chain(&opp_table->head, OPP_EVENT_ENABLE,
drivers/opp/core.c
2913
blocking_notifier_call_chain(&opp_table->head,
drivers/opp/core.c
2975
blocking_notifier_call_chain(&opp_table->head, OPP_EVENT_ADJUST_VOLTAGE,
drivers/opp/core.c
3075
return blocking_notifier_chain_register(&opp_table->head, nb);
drivers/opp/core.c
3095
return blocking_notifier_chain_unregister(&opp_table->head, nb);
drivers/opp/of.c
943
blocking_notifier_call_chain(&opp_table->head, OPP_EVENT_ADD, new_opp);
drivers/opp/opp.h
209
struct blocking_notifier_head head;
drivers/pci/controller/pcie-iproc-msi.c
304
static inline u32 decode_msi_hwirq(struct iproc_msi *msi, u32 eq, u32 head)
drivers/pci/controller/pcie-iproc-msi.c
310
offs = iproc_msi_eq_offset(msi, eq) + head * sizeof(u32);
drivers/pci/controller/pcie-iproc-msi.c
328
u32 eq, head, tail, nr_events;
drivers/pci/controller/pcie-iproc-msi.c
347
head = iproc_msi_read_reg(msi, IPROC_MSI_EQ_HEAD,
drivers/pci/controller/pcie-iproc-msi.c
357
nr_events = (tail < head) ?
drivers/pci/controller/pcie-iproc-msi.c
358
(EQ_LEN - (head - tail)) : (tail - head);
drivers/pci/controller/pcie-iproc-msi.c
364
hwirq = decode_msi_hwirq(msi, eq, head);
drivers/pci/controller/pcie-iproc-msi.c
367
head++;
drivers/pci/controller/pcie-iproc-msi.c
368
head %= EQ_LEN;
drivers/pci/controller/pcie-iproc-msi.c
375
iproc_msi_write_reg(msi, IPROC_MSI_EQ_HEAD, eq, head);
drivers/pci/endpoint/functions/pci-epf-mhi.c
458
LIST_HEAD(head);
drivers/pci/endpoint/functions/pci-epf-mhi.c
461
list_splice_tail_init(&epf_mhi->dma_list, &head);
drivers/pci/endpoint/functions/pci-epf-mhi.c
464
list_for_each_entry_safe(itr, tmp, &head, node) {
drivers/pci/hotplug/cpqphp.h
413
int cpqhp_resource_sort_and_combine(struct pci_resource **head);
drivers/pci/hotplug/cpqphp.h
461
static inline void return_resource(struct pci_resource **head,
drivers/pci/hotplug/cpqphp.h
464
if (!node || !head)
drivers/pci/hotplug/cpqphp.h
466
node->next = *head;
drivers/pci/hotplug/cpqphp.h
467
*head = node;
drivers/pci/hotplug/cpqphp_ctrl.c
298
static int sort_by_size(struct pci_resource **head)
drivers/pci/hotplug/cpqphp_ctrl.c
304
if (!(*head))
drivers/pci/hotplug/cpqphp_ctrl.c
307
if (!((*head)->next))
drivers/pci/hotplug/cpqphp_ctrl.c
314
if (((*head)->next) &&
drivers/pci/hotplug/cpqphp_ctrl.c
315
((*head)->length > (*head)->next->length)) {
drivers/pci/hotplug/cpqphp_ctrl.c
317
current_res = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
318
*head = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
319
current_res->next = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
320
(*head)->next = current_res;
drivers/pci/hotplug/cpqphp_ctrl.c
323
current_res = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
346
static int sort_by_max_size(struct pci_resource **head)
drivers/pci/hotplug/cpqphp_ctrl.c
352
if (!(*head))
drivers/pci/hotplug/cpqphp_ctrl.c
355
if (!((*head)->next))
drivers/pci/hotplug/cpqphp_ctrl.c
362
if (((*head)->next) &&
drivers/pci/hotplug/cpqphp_ctrl.c
363
((*head)->length < (*head)->next->length)) {
drivers/pci/hotplug/cpqphp_ctrl.c
365
current_res = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
366
*head = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
367
current_res->next = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
368
(*head)->next = current_res;
drivers/pci/hotplug/cpqphp_ctrl.c
371
current_res = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
396
static struct pci_resource *do_pre_bridge_resource_split(struct pci_resource **head,
drivers/pci/hotplug/cpqphp_ctrl.c
406
if (!(*head) || !(*orig_head))
drivers/pci/hotplug/cpqphp_ctrl.c
409
rc = cpqhp_resource_sort_and_combine(head);
drivers/pci/hotplug/cpqphp_ctrl.c
414
if ((*head)->base != (*orig_head)->base)
drivers/pci/hotplug/cpqphp_ctrl.c
417
if ((*head)->length == (*orig_head)->length)
drivers/pci/hotplug/cpqphp_ctrl.c
425
node = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
445
*head = split_node;
drivers/pci/hotplug/cpqphp_ctrl.c
453
if (*head == node) {
drivers/pci/hotplug/cpqphp_ctrl.c
454
*head = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
456
prevnode = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
473
static struct pci_resource *do_bridge_resource_split(struct pci_resource **head, u32 alignment)
drivers/pci/hotplug/cpqphp_ctrl.c
480
rc = cpqhp_resource_sort_and_combine(head);
drivers/pci/hotplug/cpqphp_ctrl.c
485
node = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
526
static struct pci_resource *get_io_resource(struct pci_resource **head, u32 size)
drivers/pci/hotplug/cpqphp_ctrl.c
533
if (!(*head))
drivers/pci/hotplug/cpqphp_ctrl.c
536
if (cpqhp_resource_sort_and_combine(head))
drivers/pci/hotplug/cpqphp_ctrl.c
539
if (sort_by_size(head))
drivers/pci/hotplug/cpqphp_ctrl.c
542
for (node = *head; node; node = node->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
597
if (*head == node) {
drivers/pci/hotplug/cpqphp_ctrl.c
598
*head = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
600
prevnode = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
623
static struct pci_resource *get_max_resource(struct pci_resource **head, u32 size)
drivers/pci/hotplug/cpqphp_ctrl.c
630
if (cpqhp_resource_sort_and_combine(head))
drivers/pci/hotplug/cpqphp_ctrl.c
633
if (sort_by_max_size(head))
drivers/pci/hotplug/cpqphp_ctrl.c
636
for (max = *head; max; max = max->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
690
temp = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
692
*head = max->next;
drivers/pci/hotplug/cpqphp_ctrl.c
720
static struct pci_resource *get_resource(struct pci_resource **head, u32 size)
drivers/pci/hotplug/cpqphp_ctrl.c
727
if (cpqhp_resource_sort_and_combine(head))
drivers/pci/hotplug/cpqphp_ctrl.c
730
if (sort_by_size(head))
drivers/pci/hotplug/cpqphp_ctrl.c
733
for (node = *head; node; node = node->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
787
if (*head == node) {
drivers/pci/hotplug/cpqphp_ctrl.c
788
*head = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
790
prevnode = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
813
int cpqhp_resource_sort_and_combine(struct pci_resource **head)
drivers/pci/hotplug/cpqphp_ctrl.c
819
dbg("%s: head = %p, *head = %p\n", __func__, head, *head);
drivers/pci/hotplug/cpqphp_ctrl.c
821
if (!(*head))
drivers/pci/hotplug/cpqphp_ctrl.c
824
dbg("*head->next = %p\n", (*head)->next);
drivers/pci/hotplug/cpqphp_ctrl.c
826
if (!(*head)->next)
drivers/pci/hotplug/cpqphp_ctrl.c
829
dbg("*head->base = 0x%x\n", (*head)->base);
drivers/pci/hotplug/cpqphp_ctrl.c
830
dbg("*head->next->base = 0x%x\n", (*head)->next->base);
drivers/pci/hotplug/cpqphp_ctrl.c
835
if (((*head)->next) &&
drivers/pci/hotplug/cpqphp_ctrl.c
836
((*head)->base > (*head)->next->base)) {
drivers/pci/hotplug/cpqphp_ctrl.c
837
node1 = *head;
drivers/pci/hotplug/cpqphp_ctrl.c
838
(*head) = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
839
node1->next = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
840
(*head)->next = node1;
drivers/pci/hotplug/cpqphp_ctrl.c
844
node1 = (*head);
drivers/pci/hotplug/cpqphp_ctrl.c
859
node1 = *head;
drivers/pci/pci.h
510
int pci_dev_res_add_to_list(struct list_head *head, struct pci_dev *dev,
drivers/pci/setup-bus.c
107
static struct pci_dev_resource *res_to_dev_res(struct list_head *head,
drivers/pci/setup-bus.c
112
list_for_each_entry(dev_res, head, list) {
drivers/pci/setup-bus.c
120
static resource_size_t get_res_add_size(struct list_head *head,
drivers/pci/setup-bus.c
125
dev_res = res_to_dev_res(head, res);
drivers/pci/setup-bus.c
329
static void pdev_sort_resources(struct pci_dev *dev, struct list_head *head)
drivers/pci/setup-bus.c
363
n = head;
drivers/pci/setup-bus.c
364
list_for_each_entry(dev_res, head, list) {
drivers/pci/setup-bus.c
423
struct list_head *head)
drivers/pci/setup-bus.c
446
if (!res_to_dev_res(head, res))
drivers/pci/setup-bus.c
485
static void assign_requested_resources_sorted(struct list_head *head,
drivers/pci/setup-bus.c
495
list_for_each_entry(dev_res, head, list) {
drivers/pci/setup-bus.c
52
static void pci_dev_res_free_list(struct list_head *head)
drivers/pci/setup-bus.c
56
list_for_each_entry_safe(dev_res, tmp, head, list) {
drivers/pci/setup-bus.c
577
static void __assign_resources_sorted(struct list_head *head,
drivers/pci/setup-bus.c
622
list_for_each_entry(dev_res, head, list) {
drivers/pci/setup-bus.c
631
list_for_each_entry_safe(dev_res, tmp_res, head, list) {
drivers/pci/setup-bus.c
660
list_for_each_entry(dev_res2, head, list) {
drivers/pci/setup-bus.c
672
assign_requested_resources_sorted(head, &local_fail_head, true);
drivers/pci/setup-bus.c
677
list_for_each_entry(dev_res, head, list)
drivers/pci/setup-bus.c
70
int pci_dev_res_add_to_list(struct list_head *head, struct pci_dev *dev,
drivers/pci/setup-bus.c
703
list_for_each_entry_safe(dev_res, tmp_res, head, list) {
drivers/pci/setup-bus.c
718
list_for_each_entry(dev_res, head, list) {
drivers/pci/setup-bus.c
731
assign_requested_resources_sorted(head, NULL, false);
drivers/pci/setup-bus.c
735
reassign_resources_sorted(realloc_head, head);
drivers/pci/setup-bus.c
739
list_for_each_entry(dev_res, head, list) {
drivers/pci/setup-bus.c
755
pci_dev_res_free_list(head);
drivers/pci/setup-bus.c
762
LIST_HEAD(head);
drivers/pci/setup-bus.c
764
pdev_sort_resources(dev, &head);
drivers/pci/setup-bus.c
765
__assign_resources_sorted(&head, add_head, fail_head);
drivers/pci/setup-bus.c
774
LIST_HEAD(head);
drivers/pci/setup-bus.c
777
pdev_sort_resources(dev, &head);
drivers/pci/setup-bus.c
779
__assign_resources_sorted(&head, realloc_head, fail_head);
drivers/pci/setup-bus.c
88
list_add(&tmp->list, head);
drivers/pci/setup-bus.c
93
static void pci_dev_res_remove_from_list(struct list_head *head,
drivers/pci/setup-bus.c
98
list_for_each_entry_safe(dev_res, tmp, head, list) {
drivers/perf/arm_spe_pmu.c
487
u64 head = PERF_IDX2OFF(handle->head, buf);
drivers/perf/arm_spe_pmu.c
489
memset(buf->base + head, ARM_SPE_BUF_PAD_BYTE, len);
drivers/perf/arm_spe_pmu.c
498
u64 head = PERF_IDX2OFF(handle->head, buf);
drivers/perf/arm_spe_pmu.c
507
if (head < limit >> 1)
drivers/perf/arm_spe_pmu.c
514
if (limit - head < spe_pmu->max_record_sz) {
drivers/perf/arm_spe_pmu.c
515
arm_spe_pmu_pad_buf(handle, limit - head);
drivers/perf/arm_spe_pmu.c
516
handle->head = PERF_IDX2OFF(limit, buf);
drivers/perf/arm_spe_pmu.c
517
limit = ((buf->nr_pages * PAGE_SIZE) >> 1) + handle->head;
drivers/perf/arm_spe_pmu.c
529
u64 head, tail, wakeup;
drivers/perf/arm_spe_pmu.c
545
head = PERF_IDX2OFF(handle->head, buf);
drivers/perf/arm_spe_pmu.c
546
if (!IS_ALIGNED(head, spe_pmu->align)) {
drivers/perf/arm_spe_pmu.c
547
unsigned long delta = roundup(head, spe_pmu->align) - head;
drivers/perf/arm_spe_pmu.c
551
head = PERF_IDX2OFF(handle->head, buf);
drivers/perf/arm_spe_pmu.c
559
tail = PERF_IDX2OFF(handle->head + handle->size, buf);
drivers/perf/arm_spe_pmu.c
568
if (head < tail)
drivers/perf/arm_spe_pmu.c
580
if (handle->wakeup < (handle->head + handle->size) && head <= wakeup)
drivers/perf/arm_spe_pmu.c
583
if (limit > head)
drivers/perf/arm_spe_pmu.c
598
u64 head = PERF_IDX2OFF(handle->head, buf);
drivers/perf/arm_spe_pmu.c
604
if (limit && (limit - head < spe_pmu->max_record_sz)) {
drivers/perf/arm_spe_pmu.c
605
arm_spe_pmu_pad_buf(handle, limit - head);
drivers/perf/arm_spe_pmu.c
641
base = (u64)buf->base + PERF_IDX2OFF(handle->head, buf);
drivers/perf/arm_spe_pmu.c
655
size = offset - PERF_IDX2OFF(handle->head, buf);
drivers/perf/arm_spe_pmu.c
658
handle->head = offset;
drivers/pinctrl/pinctrl-rockchip.c
3615
list_add_tail(&cfg->head, &bank->deferred_pins);
drivers/pinctrl/pinctrl-rockchip.c
4272
struct rockchip_pin_deferred, head);
drivers/pinctrl/pinctrl-rockchip.c
4273
list_del(&cfg->head);
drivers/pinctrl/pinctrl-rockchip.h
427
struct list_head head;
drivers/platform/chrome/cros_ec_debugfs.c
113
cb->buf[cb->head] = ec_buffer[idx];
drivers/platform/chrome/cros_ec_debugfs.c
114
cb->head = CIRC_ADD(cb->head, LOG_SIZE, 1);
drivers/platform/chrome/cros_ec_debugfs.c
145
while (!CIRC_CNT(cb->head, cb->tail, LOG_SIZE)) {
drivers/platform/chrome/cros_ec_debugfs.c
154
CIRC_CNT(cb->head, cb->tail, LOG_SIZE));
drivers/platform/chrome/cros_ec_debugfs.c
164
ret = min_t(size_t, CIRC_CNT_TO_END(cb->head, cb->tail, LOG_SIZE),
drivers/platform/chrome/cros_ec_debugfs.c
188
if (CIRC_CNT(debug_info->log_buffer.head,
drivers/platform/chrome/cros_ec_debugfs.c
373
debug_info->log_buffer.head = 0;
drivers/platform/chrome/cros_ec_debugfs.c
91
buf_space = CIRC_SPACE(cb->head, cb->tail, LOG_SIZE);
drivers/platform/chrome/cros_ec_typec.c
282
struct list_head *head;
drivers/platform/chrome/cros_ec_typec.c
284
head = is_partner ? &port->partner_mode_list : &port->plug_mode_list;
drivers/platform/chrome/cros_ec_typec.c
285
list_for_each_entry_safe(node, tmp, head, list) {
drivers/platform/chrome/cros_ec_typec.c
555
struct list_head *head = &port->plug_mode_list;
drivers/platform/chrome/cros_ec_typec.c
559
list_for_each_entry(node, head, list) {
drivers/platform/chrome/wilco_ec/event.c
121
return q->head == q->tail && !q->entries[q->head];
drivers/platform/chrome/wilco_ec/event.c
127
return q->head == q->tail && q->entries[q->head];
drivers/platform/chrome/wilco_ec/event.c
155
q->entries[q->head] = ev;
drivers/platform/chrome/wilco_ec/event.c
156
q->head = (q->head + 1) % q->capacity;
drivers/platform/chrome/wilco_ec/event.c
96
int head;
drivers/platform/mellanox/mlxbf-tmfifo.c
332
unsigned int idx, head;
drivers/platform/mellanox/mlxbf-tmfifo.c
341
head = virtio16_to_cpu(vdev, vr->avail->ring[idx]);
drivers/platform/mellanox/mlxbf-tmfifo.c
342
if (WARN_ON(head >= vr->num))
drivers/platform/mellanox/mlxbf-tmfifo.c
347
return &vr->desc[head];
drivers/platform/mellanox/mlxbf-tmfifo.c
467
seg = CIRC_SPACE_TO_END(cons->tx_buf.head, cons->tx_buf.tail,
drivers/platform/mellanox/mlxbf-tmfifo.c
470
memcpy(cons->tx_buf.buf + cons->tx_buf.head, addr, len);
drivers/platform/mellanox/mlxbf-tmfifo.c
472
memcpy(cons->tx_buf.buf + cons->tx_buf.head, addr, seg);
drivers/platform/mellanox/mlxbf-tmfifo.c
476
cons->tx_buf.head = (cons->tx_buf.head + len) %
drivers/platform/mellanox/mlxbf-tmfifo.c
497
avail = CIRC_SPACE(cons->tx_buf.head, cons->tx_buf.tail,
drivers/platform/mellanox/mlxbf-tmfifo.c
556
size = CIRC_CNT(cons->tx_buf.head, cons->tx_buf.tail,
drivers/platform/mellanox/mlxbf-tmfifo.c
576
seg = CIRC_CNT_TO_END(cons->tx_buf.head, cons->tx_buf.tail,
drivers/platform/surface/aggregator/controller.c
155
list_for_each_entry_rcu(nf, &nh->head, base.node,
drivers/platform/surface/aggregator/controller.c
186
list_for_each(h, &nh->head) {
drivers/platform/surface/aggregator/controller.c
220
list_for_each_entry(p, &nh->head, node) {
drivers/platform/surface/aggregator/controller.c
2339
nf_head = &nf->head[ssh_rqid_to_event(rqid)];
drivers/platform/surface/aggregator/controller.c
2408
nf_head = &nf->head[ssh_rqid_to_event(rqid)];
drivers/platform/surface/aggregator/controller.c
255
INIT_LIST_HEAD(&nh->head);
drivers/platform/surface/aggregator/controller.c
475
nf_head = &nf->head[ssh_rqid_to_event(rqid)];
drivers/platform/surface/aggregator/controller.c
501
status = ssam_nf_head_init(&nf->head[i]);
drivers/platform/surface/aggregator/controller.c
508
ssam_nf_head_destroy(&nf->head[i]);
drivers/platform/surface/aggregator/controller.c
526
ssam_nf_head_destroy(&nf->head[i]);
drivers/platform/surface/aggregator/controller.c
648
list_add_tail(&item->node, &q->head);
drivers/platform/surface/aggregator/controller.c
664
item = list_first_entry_or_null(&q->head, struct ssam_event_item, node);
drivers/platform/surface/aggregator/controller.c
681
empty = list_empty(&q->head);
drivers/platform/surface/aggregator/controller.c
812
INIT_LIST_HEAD(&evq->head);
drivers/platform/surface/aggregator/controller.h
109
struct list_head head;
drivers/platform/surface/aggregator/controller.h
56
struct list_head head;
drivers/platform/surface/aggregator/controller.h
71
struct ssam_nf_head head[SSH_NUM_EVENTS];
drivers/platform/surface/aggregator/ssh_packet_layer.c
1199
list_for_each_entry_safe(p, n, &ptl->pending.head, pending_node) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1417
list_for_each_entry(p, &ptl->pending.head, pending_node) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1523
list_for_each_entry_safe(p, n, &ptl->pending.head, pending_node) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1968
list_for_each_entry_safe(p, n, &ptl->queue.head, queue_node) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1980
list_for_each_entry_safe(p, n, &ptl->pending.head, pending_node) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
2035
INIT_LIST_HEAD(&ptl->queue.head);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2038
INIT_LIST_HEAD(&ptl->pending.head);
drivers/platform/surface/aggregator/ssh_packet_layer.c
699
struct list_head *head;
drivers/platform/surface/aggregator/ssh_packet_layer.c
718
list_for_each(head, &p->ptl->queue.head) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
719
q = list_entry(head, struct ssh_packet, queue_node);
drivers/platform/surface/aggregator/ssh_packet_layer.c
725
list_for_each_prev(head, &p->ptl->queue.head) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
726
q = list_entry(head, struct ssh_packet, queue_node);
drivers/platform/surface/aggregator/ssh_packet_layer.c
729
head = head->next;
drivers/platform/surface/aggregator/ssh_packet_layer.c
735
return head;
drivers/platform/surface/aggregator/ssh_packet_layer.c
742
struct list_head *head;
drivers/platform/surface/aggregator/ssh_packet_layer.c
757
head = __ssh_ptl_queue_find_entrypoint(packet);
drivers/platform/surface/aggregator/ssh_packet_layer.c
759
list_add_tail(&ssh_packet_get(packet)->queue_node, head);
drivers/platform/surface/aggregator/ssh_packet_layer.c
822
list_add_tail(&ssh_packet_get(p)->pending_node, &ptl->pending.head);
drivers/platform/surface/aggregator/ssh_packet_layer.c
908
list_for_each_entry_safe(p, n, &ptl->queue.head, queue_node) {
drivers/platform/surface/aggregator/ssh_packet_layer.h
83
struct list_head head;
drivers/platform/surface/aggregator/ssh_packet_layer.h
88
struct list_head head;
drivers/platform/surface/aggregator/ssh_request_layer.c
1038
INIT_LIST_HEAD(&rtl->queue.head);
drivers/platform/surface/aggregator/ssh_request_layer.c
1041
INIT_LIST_HEAD(&rtl->pending.head);
drivers/platform/surface/aggregator/ssh_request_layer.c
1214
list_for_each_entry_safe(r, n, &rtl->queue.head, node) {
drivers/platform/surface/aggregator/ssh_request_layer.c
122
empty = list_empty(&rtl->queue.head);
drivers/platform/surface/aggregator/ssh_request_layer.c
1247
list_for_each_entry_safe(r, n, &rtl->pending.head, node) {
drivers/platform/surface/aggregator/ssh_request_layer.c
164
list_add_tail(&ssh_request_get(rqst)->node, &rtl->pending.head);
drivers/platform/surface/aggregator/ssh_request_layer.c
215
list_for_each_entry_safe(p, n, &rtl->queue.head, node) {
drivers/platform/surface/aggregator/ssh_request_layer.c
418
list_add_tail(&ssh_request_get(rqst)->node, &rtl->queue.head);
drivers/platform/surface/aggregator/ssh_request_layer.c
482
list_for_each_entry_safe(p, n, &rtl->pending.head, node) {
drivers/platform/surface/aggregator/ssh_request_layer.c
840
list_for_each_entry_safe(r, n, &rtl->pending.head, node) {
drivers/platform/surface/aggregator/ssh_request_layer.h
71
struct list_head head;
drivers/platform/surface/aggregator/ssh_request_layer.h
76
struct list_head head;
drivers/platform/x86/amd/pmf/acpi.c
348
rb->data[rb->head].val[i] = pmf_dev->req1.custom_policy[i];
drivers/platform/x86/amd/pmf/acpi.c
349
rb->data[rb->head].preq = pmf_dev->req1.pending_req;
drivers/platform/x86/amd/pmf/acpi.c
354
rb->data[rb->head].val[i] = pmf_dev->req.custom_policy[i];
drivers/platform/x86/amd/pmf/acpi.c
355
rb->data[rb->head].preq = pmf_dev->req.pending_req;
drivers/platform/x86/amd/pmf/acpi.c
361
if (CIRC_SPACE(rb->head, rb->tail, CUSTOM_BIOS_INPUT_RING_ENTRIES) == 0) {
drivers/platform/x86/amd/pmf/acpi.c
367
rb->head = (rb->head + 1) & (CUSTOM_BIOS_INPUT_RING_ENTRIES - 1);
drivers/platform/x86/amd/pmf/pmf.h
390
int head;
drivers/platform/x86/amd/pmf/spc.c
162
if (CIRC_CNT(rb->head, rb->tail, CUSTOM_BIOS_INPUT_RING_ENTRIES) == 0)
drivers/platform/x86/amd/pmf/tee-if.c
588
dev->cbi_buf.head = 0;
drivers/platform/x86/dell/dell-rbtn.c
350
first = !rbtn_chain_head.head;
drivers/platform/x86/dell/dell-rbtn.c
372
if (auto_remove_rfkill && !rbtn_chain_head.head)
drivers/platform/x86/dell/dell-rbtn.c
415
if (auto_remove_rfkill && rbtn_chain_head.head)
drivers/platform/x86/uniwill/uniwill-acpi.c
1310
list_add(&entry->head, &data->batteries);
drivers/platform/x86/uniwill/uniwill-acpi.c
1321
list_for_each_entry_safe(entry, tmp, &data->batteries, head) {
drivers/platform/x86/uniwill/uniwill-acpi.c
1323
list_del(&entry->head);
drivers/platform/x86/uniwill/uniwill-acpi.c
1365
list_for_each_entry(entry, &data->batteries, head) {
drivers/platform/x86/uniwill/uniwill-acpi.c
346
struct list_head head;
drivers/power/reset/reboot-mode.c
110
list_add_tail(&info->list, &reboot->head);
drivers/power/reset/reboot-mode.c
119
list_for_each_entry(info, &reboot->head, list)
drivers/power/reset/reboot-mode.c
136
list_for_each_entry(info, &reboot->head, list)
drivers/power/reset/reboot-mode.c
32
list_for_each_entry(info, &reboot->head, list)
drivers/power/reset/reboot-mode.c
44
list_for_each_entry(info, &reboot->head, list)
drivers/power/reset/reboot-mode.c
79
INIT_LIST_HEAD(&reboot->head);
drivers/ps3/ps3-vuart.c
1027
INIT_LIST_HEAD(&priv->tx_list.head);
drivers/ps3/ps3-vuart.c
1030
INIT_LIST_HEAD(&priv->rx_list.head);
drivers/ps3/ps3-vuart.c
458
const unsigned char *head;
drivers/ps3/ps3-vuart.c
488
if (list_empty(&priv->tx_list.head)) {
drivers/ps3/ps3-vuart.c
519
lb->head = lb->data;
drivers/ps3/ps3-vuart.c
524
list_add_tail(&lb->link, &priv->tx_list.head);
drivers/ps3/ps3-vuart.c
574
lb->head = lb->data;
drivers/ps3/ps3-vuart.c
578
list_add_tail(&lb->link, &priv->rx_list.head);
drivers/ps3/ps3-vuart.c
627
list_for_each_entry_safe(lb, n, &priv->rx_list.head, link) {
drivers/ps3/ps3-vuart.c
628
bytes_read = min((unsigned int)(lb->tail - lb->head), bytes);
drivers/ps3/ps3-vuart.c
630
memcpy(buf, lb->head, bytes_read);
drivers/ps3/ps3-vuart.c
635
if (bytes_read < lb->tail - lb->head) {
drivers/ps3/ps3-vuart.c
636
lb->head += bytes_read;
drivers/ps3/ps3-vuart.c
71
struct list_head head;
drivers/ps3/ps3-vuart.c
730
list_for_each_entry_safe(lb, n, &priv->tx_list.head, link) {
drivers/ps3/ps3-vuart.c
734
result = ps3_vuart_raw_write(dev, lb->head, lb->tail - lb->head,
drivers/ps3/ps3-vuart.c
746
if (bytes_written < lb->tail - lb->head) {
drivers/ps3/ps3-vuart.c
747
lb->head += bytes_written;
drivers/ps3/ps3-vuart.c
77
struct list_head head;
drivers/ps3/ps3av.c
43
u32 head[PS3AV_HEAD_MAX];
drivers/ps3/ps3av.c
500
ps3av->head[i], video_mode->vid,
drivers/ps3/ps3av.c
810
ps3av->head[i] = PS3AV_CMD_VIDEO_HEAD_A + i;
drivers/ps3/ps3av_cmd.c
366
u32 ps3av_cmd_set_video_mode(void *p, u32 head, int video_vid, int video_fmt,
drivers/ps3/ps3av_cmd.c
385
video_mode->video_head = head;
drivers/ps3/ps3av_cmd.c
387
&& head == PS3AV_CMD_VIDEO_HEAD_B)
drivers/ps3/ps3av_cmd.c
409
int ps3av_cmd_video_format_black(u32 head, u32 video_fmt, u32 mute)
drivers/ps3/ps3av_cmd.c
415
video_format.video_head = head;
drivers/ptp/ptp_chardev.c
629
event[i] = queue->buf[queue->head];
drivers/ptp/ptp_chardev.c
631
WRITE_ONCE(queue->head, (queue->head + 1) % PTP_MAX_TIMESTAMPS);
drivers/ptp/ptp_clock.c
79
WRITE_ONCE(queue->head, (queue->head + 1) % PTP_MAX_TIMESTAMPS);
drivers/ptp/ptp_private.h
100
int cnt = READ_ONCE(q->tail) - READ_ONCE(q->head);
drivers/ptp/ptp_private.h
35
int head;
drivers/ptp/ptp_sysfs.c
95
event = queue->buf[queue->head];
drivers/ptp/ptp_sysfs.c
97
WRITE_ONCE(queue->head, (queue->head + 1) % PTP_MAX_TIMESTAMPS);
drivers/rapidio/rio_cm.c
1319
ch->rx_ring.head = 0;
drivers/rapidio/rio_cm.c
167
int head;
drivers/rapidio/rio_cm.c
560
ch->rx_ring.buf[ch->rx_ring.head] = buf;
drivers/rapidio/rio_cm.c
561
ch->rx_ring.head++;
drivers/rapidio/rio_cm.c
563
ch->rx_ring.head %= RIOCM_RX_RING_SIZE;
drivers/rpmsg/qcom_glink_rpm.c
126
unsigned int head;
drivers/rpmsg/qcom_glink_rpm.c
129
head = readl(pipe->head);
drivers/rpmsg/qcom_glink_rpm.c
132
if (tail <= head)
drivers/rpmsg/qcom_glink_rpm.c
133
return pipe->native.length - head + tail;
drivers/rpmsg/qcom_glink_rpm.c
135
return tail - head;
drivers/rpmsg/qcom_glink_rpm.c
139
unsigned int head,
drivers/rpmsg/qcom_glink_rpm.c
144
len = min_t(size_t, count, pipe->native.length - head);
drivers/rpmsg/qcom_glink_rpm.c
146
__iowrite32_copy(pipe->fifo + head, data,
drivers/rpmsg/qcom_glink_rpm.c
155
head += count;
drivers/rpmsg/qcom_glink_rpm.c
156
if (head >= pipe->native.length)
drivers/rpmsg/qcom_glink_rpm.c
157
head -= pipe->native.length;
drivers/rpmsg/qcom_glink_rpm.c
159
return head;
drivers/rpmsg/qcom_glink_rpm.c
169
unsigned int head;
drivers/rpmsg/qcom_glink_rpm.c
185
head = readl(pipe->head);
drivers/rpmsg/qcom_glink_rpm.c
186
head = glink_rpm_tx_write_one(pipe, head, hdr, hlen);
drivers/rpmsg/qcom_glink_rpm.c
187
head = glink_rpm_tx_write_one(pipe, head, data, aligned_dlen);
drivers/rpmsg/qcom_glink_rpm.c
191
head = glink_rpm_tx_write_one(pipe, head, padding, pad);
drivers/rpmsg/qcom_glink_rpm.c
192
writel(head, pipe->head);
drivers/rpmsg/qcom_glink_rpm.c
262
rx->head = msg_ram + offset + sizeof(u32);
drivers/rpmsg/qcom_glink_rpm.c
269
tx->head = msg_ram + offset + sizeof(u32);
drivers/rpmsg/qcom_glink_rpm.c
342
writel(0, rpm->tx_pipe.head);
drivers/rpmsg/qcom_glink_rpm.c
52
void __iomem *head;
drivers/rpmsg/qcom_glink_rpm.c
72
unsigned int head;
drivers/rpmsg/qcom_glink_rpm.c
75
head = readl(pipe->head);
drivers/rpmsg/qcom_glink_rpm.c
78
if (head < tail)
drivers/rpmsg/qcom_glink_rpm.c
79
return pipe->native.length - tail + head;
drivers/rpmsg/qcom_glink_rpm.c
81
return head - tail;
drivers/rpmsg/qcom_glink_smem.c
132
u32 head;
drivers/rpmsg/qcom_glink_smem.c
136
head = le32_to_cpu(*pipe->head);
drivers/rpmsg/qcom_glink_smem.c
139
if (tail <= head)
drivers/rpmsg/qcom_glink_smem.c
140
avail = pipe->native.length - head + tail;
drivers/rpmsg/qcom_glink_smem.c
142
avail = tail - head;
drivers/rpmsg/qcom_glink_smem.c
153
unsigned int head,
drivers/rpmsg/qcom_glink_smem.c
158
len = min_t(size_t, count, pipe->native.length - head);
drivers/rpmsg/qcom_glink_smem.c
160
memcpy(pipe->fifo + head, data, len);
drivers/rpmsg/qcom_glink_smem.c
165
head += count;
drivers/rpmsg/qcom_glink_smem.c
166
if (head >= pipe->native.length)
drivers/rpmsg/qcom_glink_smem.c
167
head -= pipe->native.length;
drivers/rpmsg/qcom_glink_smem.c
169
return head;
drivers/rpmsg/qcom_glink_smem.c
177
unsigned int head;
drivers/rpmsg/qcom_glink_smem.c
179
head = le32_to_cpu(*pipe->head);
drivers/rpmsg/qcom_glink_smem.c
181
head = glink_smem_tx_write_one(pipe, head, hdr, hlen);
drivers/rpmsg/qcom_glink_smem.c
182
head = glink_smem_tx_write_one(pipe, head, data, dlen);
drivers/rpmsg/qcom_glink_smem.c
185
head = ALIGN(head, 8);
drivers/rpmsg/qcom_glink_smem.c
186
if (head >= pipe->native.length)
drivers/rpmsg/qcom_glink_smem.c
187
head -= pipe->native.length;
drivers/rpmsg/qcom_glink_smem.c
192
*pipe->head = cpu_to_le32(head);
drivers/rpmsg/qcom_glink_smem.c
288
tx_pipe->head = &descs[1];
drivers/rpmsg/qcom_glink_smem.c
290
rx_pipe->head = &descs[3];
drivers/rpmsg/qcom_glink_smem.c
336
*tx_pipe->head = 0;
drivers/rpmsg/qcom_glink_smem.c
54
__le32 *head;
drivers/rpmsg/qcom_glink_smem.c
69
u32 head;
drivers/rpmsg/qcom_glink_smem.c
85
head = le32_to_cpu(*pipe->head);
drivers/rpmsg/qcom_glink_smem.c
88
if (head < tail)
drivers/rpmsg/qcom_glink_smem.c
89
return pipe->native.length - tail + head;
drivers/rpmsg/qcom_glink_smem.c
91
return head - tail;
drivers/rpmsg/qcom_smd.c
245
__le32 head;
drivers/rpmsg/qcom_smd.c
267
__le32 head;
drivers/rpmsg/qcom_smd.c
401
SET_TX_CHANNEL_INFO(channel, head, 0);
drivers/rpmsg/qcom_smd.c
429
unsigned head;
drivers/rpmsg/qcom_smd.c
432
head = GET_RX_CHANNEL_INFO(channel, head);
drivers/rpmsg/qcom_smd.c
435
return (head - tail) & (channel->fifo_size - 1);
drivers/rpmsg/qcom_smd.c
685
unsigned head;
drivers/rpmsg/qcom_smd.c
689
head = GET_TX_CHANNEL_INFO(channel, head);
drivers/rpmsg/qcom_smd.c
692
return mask - ((head - tail) & mask);
drivers/rpmsg/qcom_smd.c
703
unsigned head;
drivers/rpmsg/qcom_smd.c
707
head = GET_TX_CHANNEL_INFO(channel, head);
drivers/rpmsg/qcom_smd.c
709
len = min_t(size_t, count, channel->fifo_size - head);
drivers/rpmsg/qcom_smd.c
711
smd_copy_to_fifo(channel->tx_fifo + head,
drivers/rpmsg/qcom_smd.c
724
head += count;
drivers/rpmsg/qcom_smd.c
725
head &= (channel->fifo_size - 1);
drivers/rpmsg/qcom_smd.c
726
SET_TX_CHANNEL_INFO(channel, head, head);
drivers/rtc/class.c
28
struct timerqueue_head *head = &rtc->timerqueue;
drivers/rtc/class.c
32
while ((node = timerqueue_getnext(head)))
drivers/rtc/class.c
33
timerqueue_del(head, node);
drivers/rtc/rtc-imxdi.c
692
if (list_empty_careful(&imxdi->write_wait.head))
drivers/s390/block/dasd_eckd.c
199
static void set_ch_t(struct ch_t *geo, __u32 cyl, __u8 head)
drivers/s390/block/dasd_eckd.c
202
geo->head = cyl >> 16;
drivers/s390/block/dasd_eckd.c
203
geo->head <<= 4;
drivers/s390/block/dasd_eckd.c
204
geo->head |= head;
drivers/s390/block/dasd_eckd.c
217
u8 head;
drivers/s390/block/dasd_eckd.c
2380
private->count_area[i].head != count_area_head[i] ||
drivers/s390/block/dasd_eckd.c
2395
private->count_area[i].head != count_area_head[i] ||
drivers/s390/block/dasd_eckd.c
241
head = sense[31] & 0x0F;
drivers/s390/block/dasd_eckd.c
242
*track = cyl * private->rdc_data.trk_per_cyl + head;
drivers/s390/block/dasd_eckd.c
2841
ect->head = address.head;
drivers/s390/block/dasd_eckd.c
2856
ect->head = address.head;
drivers/s390/block/dasd_eckd.c
2870
ect->head = address.head;
drivers/s390/block/dasd_eckd.c
2879
address.cyl == 0 && address.head == 0) {
drivers/s390/block/dasd_eckd.c
2886
address.cyl == 0 && address.head == 1) {
drivers/s390/block/dasd_eckd.c
3332
int head;
drivers/s390/block/dasd_eckd.c
3335
head = fmt_buffer[start].head;
drivers/s390/block/dasd_eckd.c
3348
if ((fmt_buffer[i].head == head &&
drivers/s390/block/dasd_eckd.c
3350
fmt_buffer[i].head != head ||
drivers/s390/block/dasd_eckd.c
3420
geo.cyl == 0 && geo.head == 0) {
drivers/s390/block/dasd_eckd.c
3427
geo.cyl == 0 && geo.head == 1) {
drivers/s390/block/dasd_eckd.c
3444
fmt_buffer[pos].head != geo.head ||
drivers/s390/block/dasd_eckd.c
4463
lredata->search_arg.head = lredata->seek_addr.head;
drivers/s390/block/dasd_eckd.c
519
data->search_arg.head = data->seek_addr.head;
drivers/s390/block/dasd_eckd.c
696
data->search_arg.head = data->seek_addr.head;
drivers/s390/block/dasd_eckd.h
143
__u16 head;
drivers/s390/block/dasd_eckd.h
151
__u16 head;
drivers/s390/block/dasd_eckd.h
156
__u16 head;
drivers/s390/block/dasd_eer.c
100
if (eerb->head < eerb->tail)
drivers/s390/block/dasd_eer.c
101
return eerb->tail - eerb->head - 1;
drivers/s390/block/dasd_eer.c
102
return eerb->buffersize - eerb->head + eerb->tail -1;
drivers/s390/block/dasd_eer.c
112
if (eerb->head >= eerb->tail)
drivers/s390/block/dasd_eer.c
113
return eerb->head - eerb->tail;
drivers/s390/block/dasd_eer.c
114
return eerb->buffersize - eerb->tail + eerb->head;
drivers/s390/block/dasd_eer.c
134
headindex = eerb->head / PAGE_SIZE;
drivers/s390/block/dasd_eer.c
135
localhead = eerb->head % PAGE_SIZE;
drivers/s390/block/dasd_eer.c
140
eerb->head += len;
drivers/s390/block/dasd_eer.c
141
if (eerb->head == eerb->buffersize)
drivers/s390/block/dasd_eer.c
142
eerb->head = 0; /* wrap around */
drivers/s390/block/dasd_eer.c
143
BUG_ON(eerb->head > eerb->buffersize);
drivers/s390/block/dasd_eer.c
633
eerb->head != eerb->tail);
drivers/s390/block/dasd_eer.c
669
if (eerb->head != eerb->tail)
drivers/s390/block/dasd_eer.c
85
int head;
drivers/s390/char/con3215.c
193
req->start = (raw->head - raw->count + raw->written) &
drivers/s390/char/con3215.c
202
while (lines < RAW3215_MAX_NEWLINE && ix != raw->head) {
drivers/s390/char/con3215.c
214
req->delayable = (ix == raw->head) && (len < RAW3215_MIN_WRITE);
drivers/s390/char/con3215.c
468
raw->head = (raw->head - raw->count + raw->written) &
drivers/s390/char/con3215.c
564
raw->buffer[raw->head] = _ascebc[ch];
drivers/s390/char/con3215.c
565
raw->head = (raw->head + 1) & (RAW3215_BUFFER_SIZE - 1);
drivers/s390/char/con3215.c
85
int head; /* first free byte in output buffer */
drivers/s390/cio/airq.c
92
struct hlist_head *head;
drivers/s390/cio/airq.c
96
head = &airq_lists[tpi_info->isc];
drivers/s390/cio/airq.c
98
hlist_for_each_entry_rcu(airq, head, list)
drivers/s390/cio/eadm_sch.c
165
list_for_each_entry(private, &eadm_list, head) {
drivers/s390/cio/eadm_sch.c
170
list_move_tail(&private->head, &eadm_list);
drivers/s390/cio/eadm_sch.c
222
INIT_LIST_HEAD(&private->head);
drivers/s390/cio/eadm_sch.c
240
list_add(&private->head, &eadm_list);
drivers/s390/cio/eadm_sch.c
281
list_del(&private->head);
drivers/s390/cio/eadm_sch.h
17
struct list_head head;
drivers/s390/cio/vfio_ccw_cp.c
302
static inline int is_cpa_within_range(dma32_t cpa, u32 head, int len)
drivers/s390/cio/vfio_ccw_cp.c
304
u32 tail = head + (len - 1) * sizeof(struct ccw1);
drivers/s390/cio/vfio_ccw_cp.c
307
return head <= gcpa && gcpa <= tail;
drivers/s390/cio/vfio_ccw_cp.c
310
static inline int is_tic_within_range(struct ccw1 *ccw, u32 head, int len)
drivers/s390/cio/vfio_ccw_cp.c
315
return is_cpa_within_range(ccw->cda, head, len);
drivers/s390/crypto/pkey_ep11.c
449
*keybitsize = kb->head.bitlen;
drivers/s390/crypto/zcrypt_ep11misc.c
1060
struct pl_head head;
drivers/s390/crypto/zcrypt_ep11misc.c
1082
struct pl_head head;
drivers/s390/crypto/zcrypt_ep11misc.c
1116
prep_head(&req_pl->head, req_pl_size, api, 34); /* UnwrapKey */
drivers/s390/crypto/zcrypt_ep11misc.c
1240
struct pl_head head;
drivers/s390/crypto/zcrypt_ep11misc.c
1255
struct pl_head head;
drivers/s390/crypto/zcrypt_ep11misc.c
1282
prep_head(&req_pl->head, req_pl_size, api, 33); /* WrapKey */
drivers/s390/crypto/zcrypt_ep11misc.c
312
if (kb->head.type != TOKTYPE_NON_CCA) {
drivers/s390/crypto/zcrypt_ep11misc.c
315
__func__, (int)kb->head.type, TOKTYPE_NON_CCA);
drivers/s390/crypto/zcrypt_ep11misc.c
318
if (kb->head.version != TOKVER_EP11_AES) {
drivers/s390/crypto/zcrypt_ep11misc.c
321
__func__, (int)kb->head.version, TOKVER_EP11_AES);
drivers/s390/crypto/zcrypt_ep11misc.c
324
if (kb->head.len > keylen) {
drivers/s390/crypto/zcrypt_ep11misc.c
327
__func__, (int)kb->head.len, keylen);
drivers/s390/crypto/zcrypt_ep11misc.c
330
if (kb->head.len < sizeof(*kb)) {
drivers/s390/crypto/zcrypt_ep11misc.c
333
__func__, (int)kb->head.len, sizeof(*kb));
drivers/s390/crypto/zcrypt_ep11misc.c
555
struct pl_head head;
drivers/s390/crypto/zcrypt_ep11misc.c
564
struct pl_head head;
drivers/s390/crypto/zcrypt_ep11misc.c
582
prep_head(&req_pl->head, sizeof(*req_pl), api, 38); /* get xcp info */
drivers/s390/crypto/zcrypt_ep11misc.c
749
struct pl_head head;
drivers/s390/crypto/zcrypt_ep11misc.c
769
struct pl_head head;
drivers/s390/crypto/zcrypt_ep11misc.c
813
prep_head(&req_pl->head, req_pl_size, api, 21); /* GenerateKey */
drivers/s390/crypto/zcrypt_ep11misc.c
931
struct pl_head head;
drivers/s390/crypto/zcrypt_ep11misc.c
945
struct pl_head head;
drivers/s390/crypto/zcrypt_ep11misc.c
971
prep_head(&req_pl->head, req_pl_size, api, (mode ? 20 : 19));
drivers/s390/crypto/zcrypt_ep11misc.h
34
struct ep11kblob_header head;
drivers/s390/crypto/zcrypt_error.h
106
} __packed * head = reply->msg;
drivers/s390/crypto/zcrypt_error.h
107
unsigned int apfs = *((u32 *)head->fmt2.apfs);
drivers/s390/crypto/zcrypt_error.h
131
} __packed * head = reply->msg;
drivers/s390/crypto/zcrypt_error.h
132
unsigned int apfs = *((u32 *)head->fmt2.apfs);
drivers/s390/scsi/zfcp_fc.c
260
struct fc_els_rscn *head;
drivers/s390/scsi/zfcp_fc.c
266
head = (struct fc_els_rscn *) status_buffer->payload.data;
drivers/s390/scsi/zfcp_fc.c
267
page = (struct fc_els_rscn_page *) head;
drivers/s390/scsi/zfcp_fc.c
270
no_entries = be16_to_cpu(head->rscn_plen) /
drivers/scsi/aic7xxx/queue.h
112
#define SLIST_HEAD_INITIALIZER(head) \
drivers/scsi/aic7xxx/queue.h
123
#define SLIST_EMPTY(head) ((head)->slh_first == NULL)
drivers/scsi/aic7xxx/queue.h
125
#define SLIST_FIRST(head) ((head)->slh_first)
drivers/scsi/aic7xxx/queue.h
127
#define SLIST_FOREACH(var, head, field) \
drivers/scsi/aic7xxx/queue.h
128
for ((var) = SLIST_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
132
#define SLIST_INIT(head) do { \
drivers/scsi/aic7xxx/queue.h
133
SLIST_FIRST((head)) = NULL; \
drivers/scsi/aic7xxx/queue.h
141
#define SLIST_INSERT_HEAD(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
142
SLIST_NEXT((elm), field) = SLIST_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
143
SLIST_FIRST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
148
#define SLIST_REMOVE(head, elm, type, field) do { \
drivers/scsi/aic7xxx/queue.h
149
if (SLIST_FIRST((head)) == (elm)) { \
drivers/scsi/aic7xxx/queue.h
150
SLIST_REMOVE_HEAD((head), field); \
drivers/scsi/aic7xxx/queue.h
153
struct type *curelm = SLIST_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
161
#define SLIST_REMOVE_HEAD(head, field) do { \
drivers/scsi/aic7xxx/queue.h
162
SLIST_FIRST((head)) = SLIST_NEXT(SLIST_FIRST((head)), field); \
drivers/scsi/aic7xxx/queue.h
174
#define STAILQ_HEAD_INITIALIZER(head) \
drivers/scsi/aic7xxx/queue.h
175
{ NULL, &(head).stqh_first }
drivers/scsi/aic7xxx/queue.h
185
#define STAILQ_EMPTY(head) ((head)->stqh_first == NULL)
drivers/scsi/aic7xxx/queue.h
187
#define STAILQ_FIRST(head) ((head)->stqh_first)
drivers/scsi/aic7xxx/queue.h
189
#define STAILQ_FOREACH(var, head, field) \
drivers/scsi/aic7xxx/queue.h
190
for((var) = STAILQ_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
194
#define STAILQ_INIT(head) do { \
drivers/scsi/aic7xxx/queue.h
195
STAILQ_FIRST((head)) = NULL; \
drivers/scsi/aic7xxx/queue.h
196
(head)->stqh_last = &STAILQ_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
199
#define STAILQ_INSERT_AFTER(head, tqelm, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
201
(head)->stqh_last = &STAILQ_NEXT((elm), field); \
drivers/scsi/aic7xxx/queue.h
205
#define STAILQ_INSERT_HEAD(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
206
if ((STAILQ_NEXT((elm), field) = STAILQ_FIRST((head))) == NULL) \
drivers/scsi/aic7xxx/queue.h
207
(head)->stqh_last = &STAILQ_NEXT((elm), field); \
drivers/scsi/aic7xxx/queue.h
208
STAILQ_FIRST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
211
#define STAILQ_INSERT_TAIL(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
213
STAILQ_LAST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
214
(head)->stqh_last = &STAILQ_NEXT((elm), field); \
drivers/scsi/aic7xxx/queue.h
217
#define STAILQ_LAST(head) (*(head)->stqh_last)
drivers/scsi/aic7xxx/queue.h
221
#define STAILQ_REMOVE(head, elm, type, field) do { \
drivers/scsi/aic7xxx/queue.h
222
if (STAILQ_FIRST((head)) == (elm)) { \
drivers/scsi/aic7xxx/queue.h
223
STAILQ_REMOVE_HEAD(head, field); \
drivers/scsi/aic7xxx/queue.h
226
struct type *curelm = STAILQ_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
231
(head)->stqh_last = &STAILQ_NEXT((curelm), field);\
drivers/scsi/aic7xxx/queue.h
235
#define STAILQ_REMOVE_HEAD(head, field) do { \
drivers/scsi/aic7xxx/queue.h
236
if ((STAILQ_FIRST((head)) = \
drivers/scsi/aic7xxx/queue.h
237
STAILQ_NEXT(STAILQ_FIRST((head)), field)) == NULL) \
drivers/scsi/aic7xxx/queue.h
238
(head)->stqh_last = &STAILQ_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
241
#define STAILQ_REMOVE_HEAD_UNTIL(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
242
if ((STAILQ_FIRST((head)) = STAILQ_NEXT((elm), field)) == NULL) \
drivers/scsi/aic7xxx/queue.h
243
(head)->stqh_last = &STAILQ_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
254
#define LIST_HEAD_INITIALIZER(head) \
drivers/scsi/aic7xxx/queue.h
267
#define LIST_EMPTY(head) ((head)->lh_first == NULL)
drivers/scsi/aic7xxx/queue.h
269
#define LIST_FIRST(head) ((head)->lh_first)
drivers/scsi/aic7xxx/queue.h
271
#define LIST_FOREACH(var, head, field) \
drivers/scsi/aic7xxx/queue.h
272
for ((var) = LIST_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
276
#define LIST_INIT(head) do { \
drivers/scsi/aic7xxx/queue.h
277
LIST_FIRST((head)) = NULL; \
drivers/scsi/aic7xxx/queue.h
295
#define LIST_INSERT_HEAD(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
296
if ((LIST_NEXT((elm), field) = LIST_FIRST((head))) != NULL) \
drivers/scsi/aic7xxx/queue.h
297
LIST_FIRST((head))->field.le_prev = &LIST_NEXT((elm), field);\
drivers/scsi/aic7xxx/queue.h
298
LIST_FIRST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
299
(elm)->field.le_prev = &LIST_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
320
#define TAILQ_HEAD_INITIALIZER(head) \
drivers/scsi/aic7xxx/queue.h
321
{ NULL, &(head).tqh_first }
drivers/scsi/aic7xxx/queue.h
332
#define TAILQ_EMPTY(head) ((head)->tqh_first == NULL)
drivers/scsi/aic7xxx/queue.h
334
#define TAILQ_FIRST(head) ((head)->tqh_first)
drivers/scsi/aic7xxx/queue.h
336
#define TAILQ_FOREACH(var, head, field) \
drivers/scsi/aic7xxx/queue.h
337
for ((var) = TAILQ_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
341
#define TAILQ_FOREACH_REVERSE(var, head, headname, field) \
drivers/scsi/aic7xxx/queue.h
342
for ((var) = TAILQ_LAST((head), headname); \
drivers/scsi/aic7xxx/queue.h
346
#define TAILQ_INIT(head) do { \
drivers/scsi/aic7xxx/queue.h
347
TAILQ_FIRST((head)) = NULL; \
drivers/scsi/aic7xxx/queue.h
348
(head)->tqh_last = &TAILQ_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
351
#define TAILQ_INSERT_AFTER(head, listelm, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
356
(head)->tqh_last = &TAILQ_NEXT((elm), field); \
drivers/scsi/aic7xxx/queue.h
368
#define TAILQ_INSERT_HEAD(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
369
if ((TAILQ_NEXT((elm), field) = TAILQ_FIRST((head))) != NULL) \
drivers/scsi/aic7xxx/queue.h
370
TAILQ_FIRST((head))->field.tqe_prev = \
drivers/scsi/aic7xxx/queue.h
373
(head)->tqh_last = &TAILQ_NEXT((elm), field); \
drivers/scsi/aic7xxx/queue.h
374
TAILQ_FIRST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
375
(elm)->field.tqe_prev = &TAILQ_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
378
#define TAILQ_INSERT_TAIL(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
380
(elm)->field.tqe_prev = (head)->tqh_last; \
drivers/scsi/aic7xxx/queue.h
381
*(head)->tqh_last = (elm); \
drivers/scsi/aic7xxx/queue.h
382
(head)->tqh_last = &TAILQ_NEXT((elm), field); \
drivers/scsi/aic7xxx/queue.h
385
#define TAILQ_LAST(head, headname) \
drivers/scsi/aic7xxx/queue.h
386
(*(((struct headname *)((head)->tqh_last))->tqh_last))
drivers/scsi/aic7xxx/queue.h
393
#define TAILQ_REMOVE(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
398
(head)->tqh_last = (elm)->field.tqe_prev; \
drivers/scsi/aic7xxx/queue.h
411
#define CIRCLEQ_HEAD_INITIALIZER(head) \
drivers/scsi/aic7xxx/queue.h
412
{ (void *)&(head), (void *)&(head) }
drivers/scsi/aic7xxx/queue.h
423
#define CIRCLEQ_EMPTY(head) ((head)->cqh_first == (void *)(head))
drivers/scsi/aic7xxx/queue.h
425
#define CIRCLEQ_FIRST(head) ((head)->cqh_first)
drivers/scsi/aic7xxx/queue.h
427
#define CIRCLEQ_FOREACH(var, head, field) \
drivers/scsi/aic7xxx/queue.h
428
for ((var) = CIRCLEQ_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
429
(var) != (void *)(head); \
drivers/scsi/aic7xxx/queue.h
432
#define CIRCLEQ_FOREACH_REVERSE(var, head, field) \
drivers/scsi/aic7xxx/queue.h
433
for ((var) = CIRCLEQ_LAST((head)); \
drivers/scsi/aic7xxx/queue.h
434
(var) != (void *)(head); \
drivers/scsi/aic7xxx/queue.h
437
#define CIRCLEQ_INIT(head) do { \
drivers/scsi/aic7xxx/queue.h
438
CIRCLEQ_FIRST((head)) = (void *)(head); \
drivers/scsi/aic7xxx/queue.h
439
CIRCLEQ_LAST((head)) = (void *)(head); \
drivers/scsi/aic7xxx/queue.h
442
#define CIRCLEQ_INSERT_AFTER(head, listelm, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
445
if (CIRCLEQ_NEXT((listelm), field) == (void *)(head)) \
drivers/scsi/aic7xxx/queue.h
446
CIRCLEQ_LAST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
452
#define CIRCLEQ_INSERT_BEFORE(head, listelm, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
455
if (CIRCLEQ_PREV((listelm), field) == (void *)(head)) \
drivers/scsi/aic7xxx/queue.h
456
CIRCLEQ_FIRST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
462
#define CIRCLEQ_INSERT_HEAD(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
463
CIRCLEQ_NEXT((elm), field) = CIRCLEQ_FIRST((head)); \
drivers/scsi/aic7xxx/queue.h
464
CIRCLEQ_PREV((elm), field) = (void *)(head); \
drivers/scsi/aic7xxx/queue.h
465
if (CIRCLEQ_LAST((head)) == (void *)(head)) \
drivers/scsi/aic7xxx/queue.h
466
CIRCLEQ_LAST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
468
CIRCLEQ_PREV(CIRCLEQ_FIRST((head)), field) = (elm); \
drivers/scsi/aic7xxx/queue.h
469
CIRCLEQ_FIRST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
472
#define CIRCLEQ_INSERT_TAIL(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
473
CIRCLEQ_NEXT((elm), field) = (void *)(head); \
drivers/scsi/aic7xxx/queue.h
474
CIRCLEQ_PREV((elm), field) = CIRCLEQ_LAST((head)); \
drivers/scsi/aic7xxx/queue.h
475
if (CIRCLEQ_FIRST((head)) == (void *)(head)) \
drivers/scsi/aic7xxx/queue.h
476
CIRCLEQ_FIRST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
478
CIRCLEQ_NEXT(CIRCLEQ_LAST((head)), field) = (elm); \
drivers/scsi/aic7xxx/queue.h
479
CIRCLEQ_LAST((head)) = (elm); \
drivers/scsi/aic7xxx/queue.h
482
#define CIRCLEQ_LAST(head) ((head)->cqh_last)
drivers/scsi/aic7xxx/queue.h
488
#define CIRCLEQ_REMOVE(head, elm, field) do { \
drivers/scsi/aic7xxx/queue.h
489
if (CIRCLEQ_NEXT((elm), field) == (void *)(head)) \
drivers/scsi/aic7xxx/queue.h
490
CIRCLEQ_LAST((head)) = CIRCLEQ_PREV((elm), field); \
drivers/scsi/aic7xxx/queue.h
494
if (CIRCLEQ_PREV((elm), field) == (void *)(head)) \
drivers/scsi/aic7xxx/queue.h
495
CIRCLEQ_FIRST((head)) = CIRCLEQ_NEXT((elm), field); \
drivers/scsi/arcmsr/arcmsr_attr.c
82
unsigned int head = acb->rqbuf_putIndex;
drivers/scsi/arcmsr/arcmsr_attr.c
83
unsigned int cnt_to_end = CIRC_CNT_TO_END(head, tail, ARCMSR_MAX_QBUFFER);
drivers/scsi/arcmsr/arcmsr_attr.c
85
allxfer_len = CIRC_CNT(head, tail, ARCMSR_MAX_QBUFFER);
drivers/scsi/arcmsr/arcmsr_hba.c
3036
unsigned int head = acb->rqbuf_putIndex;
drivers/scsi/arcmsr/arcmsr_hba.c
3037
unsigned int cnt_to_end = CIRC_CNT_TO_END(head, tail, ARCMSR_MAX_QBUFFER);
drivers/scsi/arcmsr/arcmsr_hba.c
3039
allxfer_len = CIRC_CNT(head, tail, ARCMSR_MAX_QBUFFER);
drivers/scsi/arcmsr/arcmsr_hba.c
3249
struct list_head *head;
drivers/scsi/arcmsr/arcmsr_hba.c
3254
head = &acb->ccb_free_list;
drivers/scsi/arcmsr/arcmsr_hba.c
3255
if (!list_empty(head)) {
drivers/scsi/arcmsr/arcmsr_hba.c
3256
ccb = list_entry(head->next, struct CommandControlBlock, list);
drivers/scsi/arm/queue.c
107
int __queue_add(Queue_t *queue, struct scsi_cmnd *SCpnt, int head)
drivers/scsi/arm/queue.c
127
if (head)
drivers/scsi/arm/queue.c
128
list_add(l, &queue->head);
drivers/scsi/arm/queue.c
130
list_add_tail(l, &queue->head);
drivers/scsi/arm/queue.c
169
list_for_each(l, &queue->head) {
drivers/scsi/arm/queue.c
194
if (!list_empty(&queue->head))
drivers/scsi/arm/queue.c
195
SCpnt = __queue_remove(queue, queue->head.next);
drivers/scsi/arm/queue.c
218
list_for_each(l, &queue->head) {
drivers/scsi/arm/queue.c
244
list_for_each(l, &queue->head) {
drivers/scsi/arm/queue.c
268
list_for_each(l, &queue->head) {
drivers/scsi/arm/queue.c
294
list_for_each(l, &queue->head) {
drivers/scsi/arm/queue.c
65
INIT_LIST_HEAD(&queue->head);
drivers/scsi/arm/queue.c
93
if (!list_empty(&queue->head))
drivers/scsi/arm/queue.h
11
struct list_head head;
drivers/scsi/arm/queue.h
61
extern int __queue_add(Queue_t *queue, struct scsi_cmnd *SCpnt, int head);
drivers/scsi/be2iscsi/be.h
35
u16 tail, head;
drivers/scsi/be2iscsi/be.h
53
return q->dma_mem.va + q->head * q->entry_size;
drivers/scsi/be2iscsi/be.h
68
index_inc(&q->head, q->len);
drivers/scsi/be2iscsi/be_cmds.c
132
wrb->tag0 |= (mccq->head << MCC_Q_WRB_IDX_SHIFT) & MCC_Q_WRB_IDX_MASK;
drivers/scsi/bfa/bfa_core.c
85
if (trcm->tail == trcm->head)
drivers/scsi/bfa/bfa_core.c
86
trcm->head = (trcm->head + 1) & (BFA_TRC_MAX - 1);
drivers/scsi/bfa/bfa_cs.h
59
u32 head;
drivers/scsi/bfa/bfa_cs.h
93
trcm->head = trcm->tail = trcm->stopped = 0;
drivers/scsi/bfa/bfa_plog.h
131
u16 head;
drivers/scsi/bfa/bfa_svc.c
292
if (plog->head == plog->tail)
drivers/scsi/bfa/bfa_svc.c
293
BFA_PL_LOG_REC_INCR(plog->head);
drivers/scsi/bfa/bfa_svc.c
302
plog->head = plog->tail = 0;
drivers/scsi/bfa/bfad_bsg.c
2060
bfad->plog_buf.head = bfad->plog_buf.tail = 0;
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
1080
struct ulp_mem_io *req = (struct ulp_mem_io *)skb->head;
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
1115
req = (struct ulp_mem_io *)skb->head;
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
1167
req = (struct cpl_set_tcb_field *)skb->head;
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
1203
req = (struct cpl_set_tcb_field *)skb->head;
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
163
struct cpl_act_open_req *req = (struct cpl_act_open_req *)skb->head;
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
204
struct cpl_close_con_req *req = (struct cpl_close_con_req *)skb->head;
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
253
req = (struct cpl_abort_req *)skb->head;
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
279
struct cpl_abort_rpl *rpl = (struct cpl_abort_rpl *)skb->head;
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
314
req = (struct cpl_rx_data_ack *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
2010
ulp_mem_io_set_hdr(cdev, (struct ulp_mem_io *)skb->head, wr_len, dlen,
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
2033
req = (struct ulp_mem_io *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
2091
req = (struct cpl_set_tcb_field *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
2126
req = (struct cpl_set_tcb_field *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
233
(struct cpl_act_open_req *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
256
(struct cpl_t5_act_open_req *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
284
(struct cpl_t6_act_open_req *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
352
(struct cpl_act_open_req6 *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
377
(struct cpl_t5_act_open_req6 *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
400
(struct cpl_t6_act_open_req6 *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
444
struct cpl_close_con_req *req = (struct cpl_close_con_req *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
492
req = (struct cpl_abort_req *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
512
struct cpl_abort_rpl *rpl = (struct cpl_abort_rpl *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
545
req = (struct cpl_rx_data_ack *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
625
flowc = (struct fw_flowc_wr *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
681
struct cxgbi_iso_info *info = (struct cxgbi_iso_info *)skb->head;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
848
iso_cpl = (struct cxgbi_iso_info *)skb->head;
drivers/scsi/cxgbi/libcxgbi.c
2064
struct cxgbi_iso_info *iso_info = (struct cxgbi_iso_info *)skb->head;
drivers/scsi/cxgbi/libcxgbi.c
2330
iso_cpl = (struct cxgbi_iso_info *)skb->head;
drivers/scsi/cxgbi/libcxgbi.h
371
memset(skb->head, 0, wrlen + dlen);
drivers/scsi/dc395x.c
611
static unsigned int list_size(struct list_head *head)
drivers/scsi/dc395x.c
615
list_for_each(pos, head)
drivers/scsi/dc395x.c
621
static struct DeviceCtlBlk *dcb_get_next(struct list_head *head,
drivers/scsi/dc395x.c
628
if (list_empty(head))
drivers/scsi/dc395x.c
632
list_for_each_entry(i, head, list)
drivers/scsi/dc395x.c
641
list_for_each_entry(i, head, list) {
drivers/scsi/dc395x.c
661
struct list_head *head)
drivers/scsi/dc395x.c
664
list_for_each_entry(i, head, list)
drivers/scsi/esp_scsi.c
880
struct list_head *head = &esp->esp_cmd_pool;
drivers/scsi/esp_scsi.c
883
if (list_empty(head)) {
drivers/scsi/esp_scsi.c
886
ret = list_entry(head->next, struct esp_cmd_entry, list);
drivers/scsi/fcoe/fcoe.c
1338
skb->len, skb->data_len, skb->head, skb->data,
drivers/scsi/fcoe/fcoe.c
1647
skb->head, skb->data, skb_tail_pointer(skb),
drivers/scsi/fnic/fnic.h
544
void fnic_free_txq(struct list_head *head);
drivers/scsi/fnic/fnic_fcs.c
839
void fnic_free_txq(struct list_head *head)
drivers/scsi/fnic/fnic_fcs.c
843
list_for_each_entry_safe(cur_frame, next, head, links) {
drivers/scsi/hpsa.c
8177
if (!h->reply_queue[i].head)
drivers/scsi/hpsa.c
8181
h->reply_queue[i].head,
drivers/scsi/hpsa.c
8183
h->reply_queue[i].head = NULL;
drivers/scsi/hpsa.c
9215
memset(h->reply_queue[i].head, 0, h->reply_queue_size);
drivers/scsi/hpsa.c
9280
memset(h->reply_queue[i].head,
drivers/scsi/hpsa.c
9488
h->reply_queue[i].head = dma_alloc_coherent(&h->pdev->dev,
drivers/scsi/hpsa.c
9492
if (!h->reply_queue[i].head) {
drivers/scsi/hpsa.c
998
if ((rq->head[rq->current_entry] & 1) == rq->wraparound) {
drivers/scsi/hpsa.c
999
a = rq->head[rq->current_entry];
drivers/scsi/hpsa.h
117
u64 *head;
drivers/scsi/hpsa.h
507
if ((((u32) rq->head[rq->current_entry]) & 1) == rq->wraparound) {
drivers/scsi/hpsa.h
508
register_value = rq->head[rq->current_entry];
drivers/scsi/hpsa.h
597
register_value = rq->head[rq->current_entry];
drivers/scsi/hpsa.h
599
rq->head[rq->current_entry] = IOACCEL_MODE1_REPLY_UNUSED;
drivers/scsi/hptiop.c
164
u32 head = inbound_head + 1;
drivers/scsi/hptiop.c
166
if (head == MVIOP_QUEUE_LEN)
drivers/scsi/hptiop.c
167
head = 0;
drivers/scsi/hptiop.c
170
writel(head, &hba->u.mv.mu->inbound_head);
drivers/scsi/ips.c
2546
(ha->copp_waitlist.head) && (scb = ips_getscb(ha))) {
drivers/scsi/ips.c
2613
p = ha->scb_waitlist.head;
drivers/scsi/ips.c
2746
item->q_next = queue->head;
drivers/scsi/ips.c
2747
queue->head = item;
drivers/scsi/ips.c
2773
item = queue->head;
drivers/scsi/ips.c
2779
queue->head = item->q_next;
drivers/scsi/ips.c
2811
if (item == queue->head) {
drivers/scsi/ips.c
2815
p = queue->head;
drivers/scsi/ips.c
2861
if (!queue->head)
drivers/scsi/ips.c
2862
queue->head = item;
drivers/scsi/ips.c
2884
item = queue->head;
drivers/scsi/ips.c
2890
queue->head = (struct scsi_cmnd *) item->host_scribble;
drivers/scsi/ips.c
2922
if (item == queue->head) {
drivers/scsi/ips.c
2926
p = queue->head;
drivers/scsi/ips.c
2973
if (!queue->head)
drivers/scsi/ips.c
2974
queue->head = item;
drivers/scsi/ips.c
2997
item = queue->head;
drivers/scsi/ips.c
3003
queue->head = item->next;
drivers/scsi/ips.c
3035
if (item == queue->head) {
drivers/scsi/ips.c
3039
p = queue->head;
drivers/scsi/ips.c
800
item = ha->copp_waitlist.head;
drivers/scsi/ips.h
963
struct ips_scb *head;
drivers/scsi/ips.h
972
struct scsi_cmnd *head;
drivers/scsi/ips.h
983
struct ips_copp_wait_item *head;
drivers/scsi/isci/host.c
2560
u16 head = ihost->tci_head & (SCI_MAX_IO_REQUESTS-1);
drivers/scsi/isci/host.c
2561
u16 tci = ihost->tci_pool[head];
drivers/scsi/isci/host.c
2563
ihost->tci_head = head + 1;
drivers/scsi/lpfc/lpfc_bsg.c
1126
if (!list_empty(&head))
drivers/scsi/lpfc/lpfc_bsg.c
1127
list_del(&head);
drivers/scsi/lpfc/lpfc_bsg.c
2898
struct list_head head, *curr, *next;
drivers/scsi/lpfc/lpfc_bsg.c
2933
INIT_LIST_HEAD(&head);
drivers/scsi/lpfc/lpfc_bsg.c
2934
list_add_tail(&head, &dmp->list);
drivers/scsi/lpfc/lpfc_bsg.c
2935
list_for_each_safe(curr, next, &head) {
drivers/scsi/lpfc/lpfc_bsg.c
3000
list_del(&head);
drivers/scsi/lpfc/lpfc_bsg.c
3053
struct list_head head;
drivers/scsi/lpfc/lpfc_bsg.c
3191
INIT_LIST_HEAD(&head);
drivers/scsi/lpfc/lpfc_bsg.c
3192
list_add_tail(&head, &txbuffer->dma.list);
drivers/scsi/lpfc/lpfc_bsg.c
3193
list_for_each_entry(curr, &head, list) {
drivers/scsi/lpfc/lpfc_bsg.c
3216
list_del(&head);
drivers/scsi/lpfc/lpfc_bsg.c
873
struct list_head head, *curr, *next;
drivers/scsi/lpfc/lpfc_bsg.c
881
list_add_tail(&head, &mlist->dma.list);
drivers/scsi/lpfc/lpfc_bsg.c
883
list_for_each_safe(curr, next, &head) {
drivers/scsi/lpfc/lpfc_bsg.c
913
struct list_head head;
drivers/scsi/lpfc/lpfc_bsg.c
927
INIT_LIST_HEAD(&head);
drivers/scsi/lpfc/lpfc_bsg.c
928
list_add_tail(&head, &piocbq->list);
drivers/scsi/lpfc/lpfc_bsg.c
954
iocbq = list_entry(head.prev, typeof(*iocbq), list);
drivers/scsi/lpfc/lpfc_bsg.c
960
list_for_each_entry(iocbq, &head, list) {
drivers/scsi/lpfc/lpfc_bsg.c
982
list_for_each_entry(iocbq, &head, list) {
drivers/scsi/lpfc/lpfc_ct.c
347
struct list_head head;
drivers/scsi/lpfc/lpfc_ct.c
399
INIT_LIST_HEAD(&head);
drivers/scsi/lpfc/lpfc_ct.c
400
list_add_tail(&head, &ctiocbq->list);
drivers/scsi/lpfc/lpfc_ct.c
401
list_for_each_entry(iocb, &head, list) {
drivers/scsi/lpfc/lpfc_ct.c
429
list_del(&head);
drivers/scsi/lpfc/lpfc_ct.c
431
INIT_LIST_HEAD(&head);
drivers/scsi/lpfc/lpfc_ct.c
432
list_add_tail(&head, &ctiocbq->list);
drivers/scsi/lpfc/lpfc_ct.c
433
list_for_each_entry(iocbq, &head, list) {
drivers/scsi/lpfc/lpfc_ct.c
448
list_del(&head);
drivers/scsi/lpfc/lpfc_ct.c
853
struct list_head head;
drivers/scsi/lpfc/lpfc_ct.c
861
list_add_tail(&head, &mp->list);
drivers/scsi/lpfc/lpfc_ct.c
862
list_for_each_entry_safe(mp, next_mp, &head, list) {
drivers/scsi/lpfc/lpfc_ct.c
904
list_del(&head);
drivers/scsi/megaraid.c
428
struct list_head *head = &adapter->free_list;
drivers/scsi/megaraid.c
432
if( !list_empty(head) ) {
drivers/scsi/megaraid.c
434
scb = list_entry(head->next, scb_t, list);
drivers/scsi/megaraid.c
436
list_del_init(head->next);
drivers/scsi/megaraid/megaraid_mbox.c
1278
struct list_head *head = &adapter->kscb_pool;
drivers/scsi/megaraid/megaraid_mbox.c
1285
if (list_empty(head)) {
drivers/scsi/megaraid/megaraid_mbox.c
1290
scb = list_entry(head->next, scb_t, list);
drivers/scsi/megaraid/megaraid_mbox.c
3557
struct list_head *head = &adapter->uscb_pool;
drivers/scsi/megaraid/megaraid_mbox.c
3567
if (list_empty(head)) { // should never happen because of CMM
drivers/scsi/megaraid/megaraid_mbox.c
3577
scb = list_entry(head->next, scb_t, list);
drivers/scsi/megaraid/megaraid_mm.c
591
struct list_head* head;
drivers/scsi/megaraid/megaraid_mm.c
598
head = &adp->kioc_pool;
drivers/scsi/megaraid/megaraid_mm.c
600
if (list_empty(head)) {
drivers/scsi/megaraid/megaraid_mm.c
608
kioc = list_entry(head->next, uioc_t, list);
drivers/scsi/mpt3sas/mpt3sas_scsih.c
11461
LIST_HEAD(head);
drivers/scsi/mpt3sas/mpt3sas_scsih.c
11486
list_move_tail(&sas_device->list, &head);
drivers/scsi/mpt3sas/mpt3sas_scsih.c
11495
list_for_each_entry_safe(sas_device, sas_device_next, &head, list) {
drivers/scsi/mpt3sas/mpt3sas_scsih.c
11502
INIT_LIST_HEAD(&head);
drivers/scsi/mpt3sas/mpt3sas_scsih.c
11517
list_move_tail(&pcie_device->list, &head);
drivers/scsi/mpt3sas/mpt3sas_scsih.c
11523
list_for_each_entry_safe(pcie_device, pcie_device_next, &head, list) {
drivers/scsi/ncr53c8xx.c
1477
struct head header;
drivers/scsi/ncr53c8xx.c
156
static inline struct list_head *ncr_list_pop(struct list_head *head)
drivers/scsi/ncr53c8xx.c
158
if (!list_empty(head)) {
drivers/scsi/ncr53c8xx.c
159
struct list_head *elem = head->next;
drivers/scsi/ncr53c8xx.c
1590
struct head header;
drivers/scsi/ncr53c8xx.c
2087
SCR_COPY (sizeof (struct head)),
drivers/scsi/ncr53c8xx.c
2384
SCR_COPY (sizeof (struct head)),
drivers/scsi/ncr53c8xx.c
2603
SCR_COPY (sizeof (struct head)),
drivers/scsi/ncr53c8xx.c
2828
SCR_COPY (sizeof (struct head)),
drivers/scsi/qla2xxx/qla_def.h
4833
struct list_head head;
drivers/scsi/qla2xxx/qla_def.h
4837
struct list_head head;
drivers/scsi/qla2xxx/qla_def.h
5124
struct list_head head;
drivers/scsi/qla2xxx/qla_edif.c
1275
if (list_empty(&vha->e_dbell.head) && DBELL_ACTIVE(vha)) {
drivers/scsi/qla2xxx/qla_edif.c
1871
INIT_LIST_HEAD(&vha->pur_cinfo.head);
drivers/scsi/qla2xxx/qla_edif.c
1902
list_for_each_entry_safe(node, q, &vha->pur_cinfo.head, list) {
drivers/scsi/qla2xxx/qla_edif.c
1925
list_for_each_entry_safe(e, tmp, &vha->pur_cinfo.head, list) {
drivers/scsi/qla2xxx/qla_edif.c
1980
list_add_tail(&ptr->list, &vha->pur_cinfo.head);
drivers/scsi/qla2xxx/qla_edif.c
1998
list_for_each_entry_safe(list_node, q, &vha->pur_cinfo.head, list) {
drivers/scsi/qla2xxx/qla_edif.c
2103
INIT_LIST_HEAD(&vha->e_dbell.head);
drivers/scsi/qla2xxx/qla_edif.c
2122
list_for_each_entry_safe(e, tmp, &vha->e_dbell.head, list) {
drivers/scsi/qla2xxx/qla_edif.c
2174
list_for_each_entry_safe(node, q, &vha->e_dbell.head, list) {
drivers/scsi/qla2xxx/qla_edif.c
2217
list_add_tail(&ptr->list, &vha->e_dbell.head);
drivers/scsi/qla2xxx/qla_edif.c
63
if (!list_empty(&vha->e_dbell.head)) {
drivers/scsi/qla2xxx/qla_edif.c
64
edbnode = list_first_entry(&vha->e_dbell.head,
drivers/scsi/qla2xxx/qla_edif.h
40
struct list_head head;
drivers/scsi/qla2xxx/qla_edif.h
53
struct list_head head;
drivers/scsi/qla2xxx/qla_isr.c
1120
list_add_tail(&pkt->list, &list->head);
drivers/scsi/qla2xxx/qla_os.c
3983
list_for_each_entry_safe(item, next, &list->head, list) {
drivers/scsi/qla2xxx/qla_os.c
4244
INIT_LIST_HEAD(&ha->pool.good.head);
drivers/scsi/qla2xxx/qla_os.c
4245
INIT_LIST_HEAD(&ha->pool.unusable.head);
drivers/scsi/qla2xxx/qla_os.c
4279
&ha->pool.unusable.head);
drivers/scsi/qla2xxx/qla_os.c
4283
&ha->pool.good.head);
drivers/scsi/qla2xxx/qla_os.c
4290
&ha->pool.good.head, list) {
drivers/scsi/qla2xxx/qla_os.c
4539
list_for_each_entry_safe(dsd, nxt, &ha->pool.unusable.head,
drivers/scsi/qla2xxx/qla_os.c
4978
list_for_each_entry_safe(dsd, nxt, &ha->pool.unusable.head,
drivers/scsi/qla2xxx/qla_os.c
4988
list_for_each_entry_safe(dsd, nxt, &ha->pool.good.head, list) {
drivers/scsi/qla2xxx/qla_os.c
5078
INIT_LIST_HEAD(&vha->purex_list.head);
drivers/scsi/qla2xxx/qla_os.c
6473
struct list_head head = LIST_HEAD_INIT(head);
drivers/scsi/qla2xxx/qla_os.c
6478
list_splice_init(&list->head, &head);
drivers/scsi/qla2xxx/qla_os.c
6481
list_for_each_entry_safe(item, next, &head, list) {
drivers/scsi/scsi_debug.c
1057
static void sdebug_err_free(struct rcu_head *head)
drivers/scsi/scsi_debug.c
1060
container_of(head, typeof(*inject), rcu);
drivers/scsi/scsi_debug.c
7110
pp->head = (start_sec - (pp->cyl * heads_by_sects))
drivers/scsi/scsi_error.c
281
static void scsi_eh_inc_host_failed(struct rcu_head *head)
drivers/scsi/scsi_error.c
283
struct scsi_cmnd *scmd = container_of(head, typeof(*scmd), rcu);
drivers/scsi/scsi_transport_fc.c
2033
#define get_list_head_entry(pos, head, member) \
drivers/scsi/scsi_transport_fc.c
2034
pos = list_entry((head)->next, typeof(*pos), member)
drivers/scsi/sym53c8xx_2/sym_fw2.h
1003
offsetof (struct sym_ccb, phys.head.status),
drivers/scsi/sym53c8xx_2/sym_fw2.h
1014
offsetof(struct sym_lcb, head.itl_task_sa),
drivers/scsi/sym53c8xx_2/sym_fw2.h
1019
offsetof(struct sym_ccb, phys.head.go.restart),
drivers/scsi/sym53c8xx_2/sym_fw2.h
1692
offsetof(struct sym_ccb, phys.head.lastp),
drivers/scsi/sym53c8xx_2/sym_fw2.h
305
offsetof (struct sym_ccb, phys.head.status),
drivers/scsi/sym53c8xx_2/sym_fw2.h
455
offsetof (struct sym_ccb, phys.head.lastp),
drivers/scsi/sym53c8xx_2/sym_fw2.h
514
offsetof (struct sym_ccb, phys.head.lastp),
drivers/scsi/sym53c8xx_2/sym_fw2.h
546
offsetof (struct sym_ccb, phys.head.lastp),
drivers/scsi/sym53c8xx_2/sym_fw2.h
562
offsetof (struct sym_ccb, phys.head.lastp),
drivers/scsi/sym53c8xx_2/sym_fw2.h
654
offsetof (struct sym_ccb, phys.head.status),
drivers/scsi/sym53c8xx_2/sym_fw2.h
664
offsetof (struct sym_ccb, phys.head.status),
drivers/scsi/sym53c8xx_2/sym_fw2.h
740
offsetof (struct sym_ccb, phys.head.lastp),
drivers/scsi/sym53c8xx_2/sym_fw2.h
742
offsetof (struct sym_ccb, phys.head.savep),
drivers/scsi/sym53c8xx_2/sym_fw2.h
762
offsetof (struct sym_ccb, phys.head.savep),
drivers/scsi/sym53c8xx_2/sym_fw2.h
764
offsetof (struct sym_ccb, phys.head.lastp),
drivers/scsi/sym53c8xx_2/sym_fw2.h
792
offsetof (struct sym_ccb, phys.head.status),
drivers/scsi/sym53c8xx_2/sym_fw2.h
879
offsetof(struct sym_tcb, head.wval),
drivers/scsi/sym53c8xx_2/sym_fw2.h
881
offsetof(struct sym_tcb, head.sval),
drivers/scsi/sym53c8xx_2/sym_fw2.h
888
offsetof(struct sym_tcb, head.uval),
drivers/scsi/sym53c8xx_2/sym_fw2.h
911
offsetof(struct sym_tcb, head.luntbl_sa),
drivers/scsi/sym53c8xx_2/sym_fw2.h
927
offsetof(struct sym_tcb, head.lun0_sa),
drivers/scsi/sym53c8xx_2/sym_fw2.h
932
offsetof(struct sym_lcb, head.resel_sa),
drivers/scsi/sym53c8xx_2/sym_fw2.h
955
offsetof(struct sym_lcb, head.itlq_tbl_sa),
drivers/scsi/sym53c8xx_2/sym_fw2.h
988
offsetof(struct sym_ccb, phys.head.go.restart),
drivers/scsi/sym53c8xx_2/sym_glue.c
262
if ((len & 1) && (tp->head.wval & EWS)) {
drivers/scsi/sym53c8xx_2/sym_glue.c
383
cp->phys.head.lastp = cpu_to_scr(lastp);
drivers/scsi/sym53c8xx_2/sym_glue.c
384
cp->phys.head.savep = cpu_to_scr(lastp);
drivers/scsi/sym53c8xx_2/sym_glue.c
385
cp->startp = cp->phys.head.savep;
drivers/scsi/sym53c8xx_2/sym_glue.c
892
tp->head.sval = 0;
drivers/scsi/sym53c8xx_2/sym_glue.c
893
tp->head.wval = np->rv_scntl3;
drivers/scsi/sym53c8xx_2/sym_glue.c
894
tp->head.uval = 0;
drivers/scsi/sym53c8xx_2/sym_hipd.c
1554
lp->head.resel_sa =
drivers/scsi/sym53c8xx_2/sym_hipd.c
1562
lp->head.itl_task_sa = cpu_to_scr(cp->ccb_ba);
drivers/scsi/sym53c8xx_2/sym_hipd.c
1563
lp->head.resel_sa =
drivers/scsi/sym53c8xx_2/sym_hipd.c
1880
tp->head.sval = 0;
drivers/scsi/sym53c8xx_2/sym_hipd.c
1881
tp->head.wval = np->rv_scntl3;
drivers/scsi/sym53c8xx_2/sym_hipd.c
1882
tp->head.uval = 0;
drivers/scsi/sym53c8xx_2/sym_hipd.c
1940
sval = tp->head.sval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
1941
wval = tp->head.wval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
1942
uval = tp->head.uval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
1994
if (tp->head.sval == sval &&
drivers/scsi/sym53c8xx_2/sym_hipd.c
1995
tp->head.wval == wval &&
drivers/scsi/sym53c8xx_2/sym_hipd.c
1996
tp->head.uval == uval)
drivers/scsi/sym53c8xx_2/sym_hipd.c
1998
tp->head.sval = sval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
1999
tp->head.wval = wval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
2000
tp->head.uval = uval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
2012
OUTB(np, nc_sxfer, tp->head.sval);
drivers/scsi/sym53c8xx_2/sym_hipd.c
2013
OUTB(np, nc_scntl3, tp->head.wval);
drivers/scsi/sym53c8xx_2/sym_hipd.c
2016
OUTB(np, nc_scntl4, tp->head.uval);
drivers/scsi/sym53c8xx_2/sym_hipd.c
2027
cp->phys.select.sel_scntl3 = tp->head.wval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
2028
cp->phys.select.sel_sxfer = tp->head.sval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
2030
cp->phys.select.sel_scntl4 = tp->head.uval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
2096
u_char wide = (tp->head.wval & EWS) ? BUS_16_BIT : BUS_8_BIT;
drivers/scsi/sym53c8xx_2/sym_hipd.c
3145
cp->phys.head.savep = cpu_to_scr(startp);
drivers/scsi/sym53c8xx_2/sym_hipd.c
3146
cp->phys.head.lastp = cpu_to_scr(startp);
drivers/scsi/sym53c8xx_2/sym_hipd.c
3157
cp->phys.head.go.start = cpu_to_scr(SCRIPTA_BA(np, select));
drivers/scsi/sym53c8xx_2/sym_hipd.c
3330
np->abrt_sel.sel_scntl3 = tp->head.wval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
3331
np->abrt_sel.sel_sxfer = tp->head.sval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
3542
tp->head.sval = 0;
drivers/scsi/sym53c8xx_2/sym_hipd.c
3543
tp->head.wval = np->rv_scntl3;
drivers/scsi/sym53c8xx_2/sym_hipd.c
3544
tp->head.uval = 0;
drivers/scsi/sym53c8xx_2/sym_hipd.c
3869
if (cp->phys.head.lastp == cp->goalp)
drivers/scsi/sym53c8xx_2/sym_hipd.c
3876
if (cp->startp == cp->phys.head.lastp ||
drivers/scsi/sym53c8xx_2/sym_hipd.c
3877
sym_evaluate_dp(np, cp, scr_to_cpu(cp->phys.head.lastp),
drivers/scsi/sym53c8xx_2/sym_hipd.c
4697
lp->head.resel_sa =
drivers/scsi/sym53c8xx_2/sym_hipd.c
4730
lp->head.itl_task_sa = cpu_to_scr(cp->ccb_ba);
drivers/scsi/sym53c8xx_2/sym_hipd.c
4731
lp->head.resel_sa =
drivers/scsi/sym53c8xx_2/sym_hipd.c
4809
lp->head.itl_task_sa = cpu_to_scr(np->bad_itl_ba);
drivers/scsi/sym53c8xx_2/sym_hipd.c
4816
lp->head.resel_sa =
drivers/scsi/sym53c8xx_2/sym_hipd.c
4902
cp->phys.head.go.start = cpu_to_scr(SCRIPTA_BA(np, idle));
drivers/scsi/sym53c8xx_2/sym_hipd.c
4903
cp->phys.head.go.restart = cpu_to_scr(SCRIPTB_BA(np, bad_i_t_l));
drivers/scsi/sym53c8xx_2/sym_hipd.c
4958
offsetof(struct sym_tcb, head.sval)) &3) == 0);
drivers/scsi/sym53c8xx_2/sym_hipd.c
4960
offsetof(struct sym_tcb, head.wval)) &3) == 0);
drivers/scsi/sym53c8xx_2/sym_hipd.c
4986
tp->head.luntbl_sa = cpu_to_scr(vtobus(tp->luntbl));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5012
tp->head.lun0_sa = cpu_to_scr(vtobus(lp));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5019
lp->head.itl_task_sa = cpu_to_scr(np->bad_itl_ba);
drivers/scsi/sym53c8xx_2/sym_hipd.c
5024
lp->head.resel_sa = cpu_to_scr(SCRIPTB_BA(np, resel_bad_lun));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5083
lp->head.itlq_tbl_sa = cpu_to_scr(vtobus(lp->itlq_tbl));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5107
tp->head.luntbl_sa = cpu_to_scr(vtobus(np->badluntbl));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5114
tp->head.lun0_sa = cpu_to_scr(vtobus(&np->badlun_sa));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5227
cp->phys.head.go.start = cpu_to_scr(SCRIPTA_BA(np, select));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5228
cp->phys.head.go.restart = cpu_to_scr(SCRIPTA_BA(np, resel_dsa));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5234
cp->phys.select.sel_scntl3 = tp->head.wval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
5235
cp->phys.select.sel_sxfer = tp->head.sval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
5236
cp->phys.select.sel_scntl4 = tp->head.uval;
drivers/scsi/sym53c8xx_2/sym_hipd.c
5519
if (cp->phys.head.lastp != cp->goalp)
drivers/scsi/sym53c8xx_2/sym_hipd.c
5774
np->target[i].head.luntbl_sa =
drivers/scsi/sym53c8xx_2/sym_hipd.c
5776
np->target[i].head.lun0_sa =
drivers/scsi/sym53c8xx_2/sym_hipd.h
384
/*0*/ struct sym_tcbh head;
drivers/scsi/sym53c8xx_2/sym_hipd.h
474
/*0*/ struct sym_lcbh head;
drivers/scsi/sym53c8xx_2/sym_hipd.h
598
#define host_xflags phys.head.status[0]
drivers/scsi/sym53c8xx_2/sym_hipd.h
599
#define host_status phys.head.status[1]
drivers/scsi/sym53c8xx_2/sym_hipd.h
600
#define ssss_status phys.head.status[2]
drivers/scsi/sym53c8xx_2/sym_hipd.h
601
#define host_flags phys.head.status[3]
drivers/scsi/sym53c8xx_2/sym_hipd.h
667
cp->phys.head.lastp = cpu_to_scr(dp); \
drivers/scsi/sym53c8xx_2/sym_hipd.h
673
cp->phys.head.lastp : np->ccb_head.lastp)
drivers/scsi/sym53c8xx_2/sym_hipd.h
677
cp->phys.head.lastp = cpu_to_scr(dp); \
drivers/scsi/sym53c8xx_2/sym_hipd.h
680
#define sym_get_script_dp(np, cp) (cp->phys.head.lastp)
drivers/scsi/sym53c8xx_2/sym_hipd.h
695
/*0*/ struct sym_ccbh head;
drivers/scsi/sym53c8xx_2/sym_misc.h
117
#define sym_insque_head(new, head) __sym_que_add(new, head, (head)->flink)
drivers/scsi/sym53c8xx_2/sym_misc.h
119
static inline struct sym_quehead *sym_remque_head(struct sym_quehead *head)
drivers/scsi/sym53c8xx_2/sym_misc.h
121
struct sym_quehead *elem = head->flink;
drivers/scsi/sym53c8xx_2/sym_misc.h
123
if (elem != head)
drivers/scsi/sym53c8xx_2/sym_misc.h
124
__sym_que_del(head, elem->flink);
drivers/scsi/sym53c8xx_2/sym_misc.h
130
#define sym_insque_tail(new, head) __sym_que_add(new, (head)->blink, head)
drivers/scsi/sym53c8xx_2/sym_misc.h
132
static inline struct sym_quehead *sym_remque_tail(struct sym_quehead *head)
drivers/scsi/sym53c8xx_2/sym_misc.h
134
struct sym_quehead *elem = head->blink;
drivers/scsi/sym53c8xx_2/sym_misc.h
136
if (elem != head)
drivers/scsi/sym53c8xx_2/sym_misc.h
137
__sym_que_del(elem->blink, head);
drivers/scsi/sym53c8xx_2/sym_misc.h
146
#define FOR_EACH_QUEUED_ELEMENT(head, qp) \
drivers/scsi/sym53c8xx_2/sym_misc.h
147
for (qp = (head)->flink; qp != (head); qp = qp->flink)
drivers/scsi/sym53c8xx_2/sym_misc.h
42
static inline struct sym_quehead *sym_que_first(struct sym_quehead *head)
drivers/scsi/sym53c8xx_2/sym_misc.h
44
return (head->flink == head) ? 0 : head->flink;
drivers/scsi/sym53c8xx_2/sym_misc.h
47
static inline struct sym_quehead *sym_que_last(struct sym_quehead *head)
drivers/scsi/sym53c8xx_2/sym_misc.h
49
return (head->blink == head) ? 0 : head->blink;
drivers/scsi/sym53c8xx_2/sym_misc.h
69
static inline int sym_que_empty(struct sym_quehead *head)
drivers/scsi/sym53c8xx_2/sym_misc.h
71
return head->flink == head;
drivers/scsi/sym53c8xx_2/sym_misc.h
75
struct sym_quehead *head)
drivers/scsi/sym53c8xx_2/sym_misc.h
81
struct sym_quehead *at = head->flink;
drivers/scsi/sym53c8xx_2/sym_misc.h
83
first->blink = head;
drivers/scsi/sym53c8xx_2/sym_misc.h
84
head->flink = first;
drivers/sh/intc/virq.c
27
#define for_each_virq(entry, head) \
drivers/sh/intc/virq.c
28
for (entry = head; entry; entry = entry->next)
drivers/soc/fsl/qe/qe_common.c
133
list_add(&entry->head, &muram_block_list);
drivers/soc/fsl/qe/qe_common.c
178
list_for_each_entry(tmp, &muram_block_list, head) {
drivers/soc/fsl/qe/qe_common.c
181
list_del(&tmp->head);
drivers/soc/fsl/qe/qe_common.c
34
struct list_head head;
drivers/soc/hisilicon/kunpeng_hccs.c
534
struct hccs_rsp_head *head;
drivers/soc/hisilicon/kunpeng_hccs.c
543
head = &rsp->rsp_head;
drivers/soc/hisilicon/kunpeng_hccs.c
544
if (head->data_len > buf_len) {
drivers/soc/hisilicon/kunpeng_hccs.c
547
buf_len, head->data_len);
drivers/soc/hisilicon/kunpeng_hccs.c
551
memcpy(buf, rsp->data, head->data_len);
drivers/soc/hisilicon/kunpeng_hccs.c
552
*rsp_head = *head;
drivers/spi/spi-topcliff-pch.c
1031
head = PCH_MAX_FIFO_DEPTH - PCH_DMA_TRANS_SIZE;
drivers/spi/spi-topcliff-pch.c
1034
rem = data->bpw_len % PCH_DMA_TRANS_SIZE - head;
drivers/spi/spi-topcliff-pch.c
1038
PCH_DMA_TRANS_SIZE - head;
drivers/spi/spi-topcliff-pch.c
1045
head = 0;
drivers/spi/spi-topcliff-pch.c
1058
sg_set_page(sg, virt_to_page(dma->tx_buf_virt), size + head,
drivers/spi/spi-topcliff-pch.c
1060
sg_dma_len(sg) = size + head;
drivers/spi/spi-topcliff-pch.c
1062
sg->offset = head + size * i;
drivers/spi/spi-topcliff-pch.c
1068
sg->offset = head + size * i;
drivers/spi/spi-topcliff-pch.c
908
int head;
drivers/staging/media/atomisp/pci/hive_isp_css_common/debug_global.h
44
u32 head;
drivers/staging/media/atomisp/pci/hive_isp_css_common/debug_global.h
58
u32 head;
drivers/staging/media/atomisp/pci/hive_isp_css_common/host/debug.c
31
debug_data.head = 0;
drivers/staging/media/atomisp/pci/hive_isp_css_common/host/debug.c
39
u32 head = 0;
drivers/staging/media/atomisp/pci/hive_isp_css_common/host/debug.c
46
&head, sizeof(uint32_t));
drivers/staging/media/atomisp/pci/hive_isp_css_common/host/debug.c
53
debug_data.head = 0;
drivers/staging/media/atomisp/pci/hive_isp_css_common/host/debug_private.h
21
return (debug_data_ptr->head == debug_data_ptr->tail);
drivers/staging/media/atomisp/pci/hive_isp_css_common/host/debug_private.h
33
value = debug_data_ptr->buf[debug_data_ptr->head];
drivers/staging/media/atomisp/pci/hive_isp_css_common/host/debug_private.h
34
debug_data_ptr->head = (debug_data_ptr->head + 1) & DEBUG_BUF_MASK;
drivers/staging/media/atomisp/pci/hive_isp_css_common/host/debug_private.h
36
debug_data_ptr->head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
125
list_for_each_entry_safe(ib, ib_safe, &bl->head, head) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
134
list_del(&ib->head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
136
list_add(&ib->head, &aq->active);
drivers/staging/media/ipu7/ipu7-isys-queue.c
138
list_add_tail(&ib->head, &aq->incoming);
drivers/staging/media/ipu7/ipu7-isys-queue.c
175
list_for_each_entry_safe(ib, ib_safe, &aq->active, head) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
179
list_del(&ib->head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
185
list_add(&ib->head, &aq->incoming);
drivers/staging/media/ipu7/ipu7-isys-queue.c
212
INIT_LIST_HEAD(&bl->head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
220
if (!list_empty(&bl->head))
drivers/staging/media/ipu7/ipu7-isys-queue.c
226
struct ipu7_isys_buffer, head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
231
list_del(&ib->head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
232
list_add(&ib->head, &bl->head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
276
list_for_each_entry(ib, &bl->head, head) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
373
list_add(&ib->head, &aq->incoming);
drivers/staging/media/ipu7/ipu7-isys-queue.c
494
head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
497
list_del(&ib->head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
507
head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
509
list_del(&ib->head);
drivers/staging/media/ipu7/ipu7-isys-queue.c
750
list_for_each_entry_reverse(ib, &aq->active, head) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
774
list_del(&ib->head);
drivers/staging/media/ipu7/ipu7-isys-queue.h
32
struct list_head head;
drivers/staging/media/ipu7/ipu7-isys-queue.h
47
struct list_head head;
drivers/staging/media/ipu7/ipu7-isys.c
132
fw_log->head = log_buf;
drivers/staging/media/ipu7/ipu7-isys.c
582
list_for_each_entry_safe(fwmsg, safe, &isys->framebuflist, head)
drivers/staging/media/ipu7/ipu7-isys.c
586
list_for_each_entry_safe(fwmsg, safe, &isys->framebuflist_fw, head)
drivers/staging/media/ipu7/ipu7-isys.c
615
list_add(&addr->head, &isys->framebuflist);
drivers/staging/media/ipu7/ipu7-isys.c
625
struct isys_fw_msgs, head);
drivers/staging/media/ipu7/ipu7-isys.c
626
list_del(&addr->head);
drivers/staging/media/ipu7/ipu7-isys.c
661
msg = list_last_entry(&isys->framebuflist, struct isys_fw_msgs, head);
drivers/staging/media/ipu7/ipu7-isys.c
662
list_move(&msg->head, &isys->framebuflist_fw);
drivers/staging/media/ipu7/ipu7-isys.c
675
list_for_each_entry_safe(fwmsg, fwmsg0, &isys->framebuflist_fw, head)
drivers/staging/media/ipu7/ipu7-isys.c
676
list_move(&fwmsg->head, &isys->framebuflist);
drivers/staging/media/ipu7/ipu7-isys.c
691
list_move(&msg->head, &isys->framebuflist);
drivers/staging/media/ipu7/ipu7-isys.h
120
struct list_head head;
drivers/staging/media/ipu7/ipu7-isys.h
52
void *head;
drivers/staging/most/dim2/dim2.c
166
struct list_head *head = &hdm_ch->pending_list;
drivers/staging/most/dim2/dim2.c
175
if (list_empty(head)) {
drivers/staging/most/dim2/dim2.c
185
mbo = list_first_entry(head, struct mbo, list);
drivers/staging/most/dim2/dim2.c
195
list_del(head->next);
drivers/staging/most/dim2/dim2.c
203
list_move_tail(head->next, &hdm_ch->started_list);
drivers/staging/most/dim2/dim2.c
268
struct list_head *head;
drivers/staging/most/dim2/dim2.c
291
head = &hdm_ch->started_list;
drivers/staging/most/dim2/dim2.c
295
if (list_empty(head)) {
drivers/staging/most/dim2/dim2.c
301
mbo = list_first_entry(head, struct mbo, list);
drivers/staging/most/dim2/dim2.c
302
list_del(head->next);
drivers/staging/most/dim2/dim2.c
416
static void complete_all_mbos(struct list_head *head)
drivers/staging/most/dim2/dim2.c
423
if (list_empty(head)) {
drivers/staging/most/dim2/dim2.c
428
mbo = list_first_entry(head, struct mbo, list);
drivers/staging/most/dim2/dim2.c
429
list_del(head->next);
drivers/staging/octeon/ethernet-rx.c
263
prefetch(&skb->head);
drivers/staging/octeon/ethernet-rx.c
286
skb->data = skb->head + work->packet_ptr.s.addr -
drivers/staging/octeon/ethernet-rx.c
287
cvmx_ptr_to_phys(skb->head);
drivers/staging/octeon/ethernet-tx.c
297
fpa_head = skb->head + 256 - ((unsigned long)skb->head & 0x7f);
drivers/staging/rtl8723bs/hal/rtl8723bs_recv.c
304
precvframe->u.hdr.rx_head = pkt_copy->head;
drivers/staging/rtl8723bs/hal/sdio_ops.c
832
recvbuf->phead = recvbuf->pskb->head;
drivers/staging/rtl8723bs/include/rtw_event.h
90
volatile int head;
drivers/staging/rtl8723bs/include/rtw_xmit.h
322
volatile signed int head;
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
2245
static int rtw_add_beacon(struct adapter *adapter, const u8 *head, size_t head_len, const u8 *tail, size_t tail_len)
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
2262
memcpy(pbuf, (void *)head + 24, head_len - 24);/* 24 =beacon header len. */
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
2290
ret = rtw_add_beacon(adapter, settings->beacon.head,
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
2314
return rtw_add_beacon(adapter, info->beacon.head,
drivers/target/iscsi/cxgbit/cxgbit_target.c
1392
skb, skb->head, skb->data, skb->len, skb->data_len,
drivers/target/target_core_device.c
412
static void target_free_dev_entry(struct rcu_head *head)
drivers/target/target_core_device.c
414
struct se_dev_entry *deve = container_of(head, struct se_dev_entry,
drivers/target/target_core_internal.h
128
void target_tpg_free_lun(struct rcu_head *head);
drivers/target/target_core_tpg.c
644
void target_tpg_free_lun(struct rcu_head *head)
drivers/target/target_core_tpg.c
646
struct se_lun *lun = container_of(head, struct se_lun, rcu_head);
drivers/target/target_core_ua.c
201
int head = 1;
drivers/target/target_core_ua.c
244
if (head) {
drivers/target/target_core_ua.c
247
head = 0;
drivers/target/target_core_ua.c
263
return head == 0;
drivers/target/target_core_ua.c
275
int head = 1;
drivers/target/target_core_ua.c
306
if (head) {
drivers/target/target_core_ua.c
309
head = 0;
drivers/target/target_core_ua.c
322
return (head) ? -EPERM : 0;
drivers/target/target_core_user.c
676
static inline size_t spc_used(size_t head, size_t tail, size_t size)
drivers/target/target_core_user.c
678
int diff = head - tail;
drivers/target/target_core_user.c
686
static inline size_t spc_free(size_t head, size_t tail, size_t size)
drivers/target/target_core_user.c
689
return (size - spc_used(head, tail, size) - 1);
drivers/target/target_core_user.c
692
static inline size_t head_to_end(size_t head, size_t size)
drivers/target/target_core_user.c
694
return size - head;
drivers/target/target_core_user.c
697
#define UPDATE_HEAD(head, used, size) smp_store_release(&head, ((head % size) + used) % size)
drivers/target/tcm_fc/tfc_sess.c
161
struct hlist_head *head;
drivers/target/tcm_fc/tfc_sess.c
172
head = &tport->hash[ft_sess_hash(port_id)];
drivers/target/tcm_fc/tfc_sess.c
173
hlist_for_each_entry_rcu(sess, head, hash) {
drivers/target/tcm_fc/tfc_sess.c
194
struct hlist_head *head = &tport->hash[ft_sess_hash(sess->port_id)];
drivers/target/tcm_fc/tfc_sess.c
197
hlist_add_head_rcu(&sess->hash, head);
drivers/target/tcm_fc/tfc_sess.c
212
struct hlist_head *head;
drivers/target/tcm_fc/tfc_sess.c
217
head = &tport->hash[ft_sess_hash(port_id)];
drivers/target/tcm_fc/tfc_sess.c
218
hlist_for_each_entry_rcu(sess, head, hash)
drivers/target/tcm_fc/tfc_sess.c
263
struct hlist_head *head;
drivers/target/tcm_fc/tfc_sess.c
266
head = &tport->hash[ft_sess_hash(port_id)];
drivers/target/tcm_fc/tfc_sess.c
267
hlist_for_each_entry_rcu(sess, head, hash) {
drivers/target/tcm_fc/tfc_sess.c
289
struct hlist_head *head;
drivers/target/tcm_fc/tfc_sess.c
292
for (head = tport->hash;
drivers/target/tcm_fc/tfc_sess.c
293
head < &tport->hash[FT_SESS_HASH_SIZE]; head++) {
drivers/target/tcm_fc/tfc_sess.c
294
hlist_for_each_entry_rcu(sess, head, hash) {
drivers/tee/optee/call.c
226
struct list_head *head = &optee->shm_arg_cache.shm_args;
drivers/tee/optee/call.c
230
while (!list_empty(head)) {
drivers/tee/optee/call.c
231
entry = list_first_entry(head, struct optee_shm_arg_entry,
drivers/thermal/renesas/rcar_thermal.c
101
list_for_each_entry(pos, &common->head, list)
drivers/thermal/renesas/rcar_thermal.c
412
INIT_LIST_HEAD(&common->head);
drivers/thermal/renesas/rcar_thermal.c
45
struct list_head head;
drivers/thermal/renesas/rcar_thermal.c
517
list_move_tail(&priv->list, &common->head);
drivers/thermal/renesas/rcar_thermal.c
540
struct rcar_thermal_priv *priv = list_first_entry(&common->head,
drivers/thermal/renesas/rcar_thermal.c
555
struct rcar_thermal_priv *priv = list_first_entry(&common->head,
drivers/thunderbolt/nhi.c
227
return ((ring->head + 1) % ring->size) == ring->tail;
drivers/thunderbolt/nhi.c
232
return ring->head == ring->tail;
drivers/thunderbolt/nhi.c
248
descriptor = &ring->descriptors[ring->head];
drivers/thunderbolt/nhi.c
257
ring->head = (ring->head + 1) % ring->size;
drivers/thunderbolt/nhi.c
259
ring_iowrite_prod(ring, ring->head);
drivers/thunderbolt/nhi.c
261
ring_iowrite_cons(ring, ring->head);
drivers/thunderbolt/nhi.c
607
ring->head = 0;
drivers/thunderbolt/nhi.c
790
ring->head = 0;
drivers/tty/amiserial.c
156
if (info->xmit.head != info->xmit.tail
drivers/tty/amiserial.c
274
if (info->xmit.head == info->xmit.tail
drivers/tty/amiserial.c
288
if (CIRC_CNT(info->xmit.head,
drivers/tty/amiserial.c
296
if (info->xmit.head == info->xmit.tail) {
drivers/tty/amiserial.c
496
info->xmit.head = info->xmit.tail = 0;
drivers/tty/amiserial.c
708
if (CIRC_SPACE(info->xmit.head,
drivers/tty/amiserial.c
715
info->xmit.buf[info->xmit.head++] = ch;
drivers/tty/amiserial.c
716
info->xmit.head &= UART_XMIT_SIZE - 1;
drivers/tty/amiserial.c
726
if (info->xmit.head == info->xmit.tail
drivers/tty/amiserial.c
753
c = CIRC_SPACE_TO_END(info->xmit.head,
drivers/tty/amiserial.c
761
memcpy(info->xmit.buf + info->xmit.head, buf, c);
drivers/tty/amiserial.c
762
info->xmit.head = (info->xmit.head + c) & (UART_XMIT_SIZE - 1);
drivers/tty/amiserial.c
769
if (info->xmit.head != info->xmit.tail
drivers/tty/amiserial.c
789
return CIRC_SPACE(info->xmit.head, info->xmit.tail, UART_XMIT_SIZE);
drivers/tty/amiserial.c
796
return CIRC_CNT(info->xmit.head, info->xmit.tail, UART_XMIT_SIZE);
drivers/tty/amiserial.c
805
info->xmit.head = info->xmit.tail = 0;
drivers/tty/ehv_bytechan.c
419
CIRC_CNT_TO_END(bc->head, bc->tail, BUF_SIZE),
drivers/tty/ehv_bytechan.c
428
count = CIRC_CNT(bc->head, bc->tail, BUF_SIZE);
drivers/tty/ehv_bytechan.c
433
if (CIRC_CNT(bc->head, bc->tail, BUF_SIZE))
drivers/tty/ehv_bytechan.c
481
len = CIRC_SPACE_TO_END(bc->head, bc->tail, BUF_SIZE);
drivers/tty/ehv_bytechan.c
485
memcpy(bc->buf + bc->head, s, len);
drivers/tty/ehv_bytechan.c
486
bc->head = (bc->head + len) & (BUF_SIZE - 1);
drivers/tty/ehv_bytechan.c
53
unsigned int head; /* circular buffer head */
drivers/tty/ehv_bytechan.c
548
count = CIRC_SPACE(bc->head, bc->tail, BUF_SIZE);
drivers/tty/ehv_bytechan.c
695
bc->head = 0;
drivers/tty/hvc/hvcs.c
867
struct list_head head;
drivers/tty/hvc/hvcs.c
875
retval = hvcs_get_partner_info(unit_address, &head, hvcs_pi_buff);
drivers/tty/hvc/hvcs.c
887
list_for_each_entry(pi, &head, node)
drivers/tty/hvc/hvcs.c
890
hvcs_free_partner_info(&head);
drivers/tty/moxa.c
1918
u16 head, tail, tx_mask, spage, epage;
drivers/tty/moxa.c
1927
head = readw(ofsAddr + TXrptr);
drivers/tty/moxa.c
1928
c = (head > tail) ? (head - tail - 1) : (head - tail + tx_mask);
drivers/tty/moxa.c
1936
if (head > tail)
drivers/tty/moxa.c
1937
len = head - tail - 1;
drivers/tty/moxa.c
1976
u16 pageno, pageofs, bufhead, head;
drivers/tty/moxa.c
1980
head = readw(ofsAddr + RXrptr);
drivers/tty/moxa.c
1985
count = (tail >= head) ? (tail - head) : (tail - head + rx_mask + 1);
drivers/tty/moxa.c
1994
ofs = baseAddr + DynPage_addr + bufhead + head;
drivers/tty/moxa.c
1995
len = (tail >= head) ? (tail - head) :
drivers/tty/moxa.c
1996
(rx_mask + 1 - head);
drivers/tty/moxa.c
2000
head = (head + len) & rx_mask;
drivers/tty/moxa.c
2004
pageno = spage + (head >> 13);
drivers/tty/moxa.c
2005
pageofs = head & Page_mask;
drivers/tty/moxa.c
2018
head = (head + total) & rx_mask;
drivers/tty/moxa.c
2020
writew(head, ofsAddr + RXrptr);
drivers/tty/n_tty.c
1505
size_t head = MASK(ldata->read_head);
drivers/tty/n_tty.c
1506
size_t n = min(count, N_TTY_BUF_SIZE - head);
drivers/tty/n_tty.c
1508
memcpy(read_buf_addr(ldata, head), cp, n);
drivers/tty/n_tty.c
1943
size_t head = smp_load_acquire(&ldata->commit_head);
drivers/tty/n_tty.c
1946
n = min3(head - ldata->read_tail, N_TTY_BUF_SIZE - tail, *nr);
drivers/tty/n_tty.c
1959
head == ldata->read_tail)
drivers/tty/n_tty.c
1966
return head != ldata->read_tail;
drivers/tty/n_tty.c
216
size_t head = ldata->icanon ? ldata->canon_head : ldata->commit_head;
drivers/tty/n_tty.c
218
return head - ldata->read_tail;
drivers/tty/n_tty.c
2462
size_t nr, head, tail;
drivers/tty/n_tty.c
2466
head = ldata->canon_head;
drivers/tty/n_tty.c
2468
nr = head - tail;
drivers/tty/n_tty.c
2470
while (MASK(head) != MASK(tail)) {
drivers/tty/n_tty.c
734
size_t head;
drivers/tty/n_tty.c
737
head = ldata->echo_head;
drivers/tty/n_tty.c
738
ldata->echo_mark = head;
drivers/tty/n_tty.c
745
nr = head - ldata->echo_tail;
drivers/tty/n_tty.c
749
ldata->echo_commit = head;
drivers/tty/n_tty.c
929
size_t head;
drivers/tty/n_tty.c
960
head = ldata->read_head;
drivers/tty/n_tty.c
964
head--;
drivers/tty/n_tty.c
965
c = read_buf(ldata, head);
drivers/tty/n_tty.c
967
MASK(head) != MASK(ldata->canon_head));
drivers/tty/n_tty.c
980
cnt = ldata->read_head - head;
drivers/tty/n_tty.c
981
ldata->read_head = head;
drivers/tty/n_tty.c
991
head++;
drivers/tty/n_tty.c
992
echo_char_raw(read_buf(ldata, head), ldata);
drivers/tty/serial/8250/8250_core.c
112
if (!list_empty(i->head)) {
drivers/tty/serial/8250/8250_core.c
113
if (i->head == &up->list)
drivers/tty/serial/8250/8250_core.c
114
i->head = i->head->next;
drivers/tty/serial/8250/8250_core.c
117
BUG_ON(i->head != &up->list);
drivers/tty/serial/8250/8250_core.c
118
i->head = NULL;
drivers/tty/serial/8250/8250_core.c
122
if (i->head == NULL) {
drivers/tty/serial/8250/8250_core.c
164
if (i->head) {
drivers/tty/serial/8250/8250_core.c
165
list_add(&up->list, i->head);
drivers/tty/serial/8250/8250_core.c
171
i->head = &up->list;
drivers/tty/serial/8250/8250_core.c
189
if (WARN_ON(i->head == NULL))
drivers/tty/serial/8250/8250_core.c
192
if (list_empty(i->head))
drivers/tty/serial/8250/8250_core.c
48
struct list_head *head;
drivers/tty/serial/8250/8250_core.c
81
l = i->head;
drivers/tty/serial/8250/8250_core.c
94
if (l == i->head && pass_counter++ > PASS_LIMIT)
drivers/tty/serial/atmel_serial.c
1135
ring->head = ATMEL_SERIAL_RX_SIZE - state.residue;
drivers/tty/serial/atmel_serial.c
1136
BUG_ON(ring->head > ATMEL_SERIAL_RX_SIZE);
drivers/tty/serial/atmel_serial.c
1149
if (ring->head < ring->tail) {
drivers/tty/serial/atmel_serial.c
1158
if (ring->tail < ring->head) {
drivers/tty/serial/atmel_serial.c
1159
count = ring->head - ring->tail;
drivers/tty/serial/atmel_serial.c
1163
if (ring->head >= ATMEL_SERIAL_RX_SIZE)
drivers/tty/serial/atmel_serial.c
1164
ring->head = 0;
drivers/tty/serial/atmel_serial.c
1165
ring->tail = ring->head;
drivers/tty/serial/atmel_serial.c
1513
while (ring->head != ring->tail) {
drivers/tty/serial/atmel_serial.c
1590
unsigned int head;
drivers/tty/serial/atmel_serial.c
1599
head = atmel_uart_readl(port, ATMEL_PDC_RPR) - pdc->dma_addr;
drivers/tty/serial/atmel_serial.c
1612
head = min(head, pdc->dma_size);
drivers/tty/serial/atmel_serial.c
1614
if (likely(head != tail)) {
drivers/tty/serial/atmel_serial.c
1624
count = head - tail;
drivers/tty/serial/atmel_serial.c
1633
pdc->ofs = head;
drivers/tty/serial/atmel_serial.c
1640
if (head >= pdc->dma_size) {
drivers/tty/serial/atmel_serial.c
1648
} while (head >= pdc->dma_size);
drivers/tty/serial/atmel_serial.c
2052
atmel_port->rx_ring.head = 0;
drivers/tty/serial/atmel_serial.c
744
if (!CIRC_SPACE(ring->head, ring->tail, ATMEL_SERIAL_RINGSIZE))
drivers/tty/serial/atmel_serial.c
748
c = &((struct atmel_uart_char *)ring->buf)[ring->head];
drivers/tty/serial/atmel_serial.c
755
ring->head = (ring->head + 1) & (ATMEL_SERIAL_RINGSIZE - 1);
drivers/tty/serial/fsl_lpuart.c
1079
if (ring->head < ring->tail) {
drivers/tty/serial/fsl_lpuart.c
1086
if (ring->head > ring->tail) {
drivers/tty/serial/fsl_lpuart.c
1087
count = ring->head - ring->tail;
drivers/tty/serial/fsl_lpuart.c
1090
ring->tail = ring->head;
drivers/tty/serial/fsl_lpuart.c
1189
ring->head = sport->rx_sgl.length - state.residue;
drivers/tty/serial/fsl_lpuart.c
1190
BUG_ON(ring->head > sport->rx_sgl.length);
drivers/tty/serial/fsl_lpuart.c
1212
if (ring->head < ring->tail) {
drivers/tty/serial/fsl_lpuart.c
1224
if (ring->tail < ring->head) {
drivers/tty/serial/fsl_lpuart.c
1225
count = ring->head - ring->tail;
drivers/tty/serial/fsl_lpuart.c
1231
if (ring->head >= sport->rx_sgl.length)
drivers/tty/serial/fsl_lpuart.c
1232
ring->head = 0;
drivers/tty/serial/fsl_lpuart.c
1233
ring->tail = ring->head;
drivers/tty/serial/fsl_lpuart.c
1271
ring->head = sport->rx_sgl.length - state.residue;
drivers/tty/serial/fsl_lpuart.c
1272
count = CIRC_CNT(ring->head, ring->tail, sport->rx_sgl.length);
drivers/tty/serial/fsl_lpuart.c
1324
ring->head = sport->rx_sgl.length - state.residue;
drivers/tty/serial/fsl_lpuart.c
1325
count = CIRC_CNT(ring->head, ring->tail, sport->rx_sgl.length);
drivers/tty/serial/fsl_lpuart.c
1444
sport->rx_ring.head = 0;
drivers/tty/serial/imx.c
1223
rx_ring->head = sg_dma_len(sgl) - state.residue;
drivers/tty/serial/imx.c
1227
rx_ring->tail = ((rx_ring->head-1) / bd_size) * bd_size;
drivers/tty/serial/imx.c
1229
if (rx_ring->head <= sg_dma_len(sgl) &&
drivers/tty/serial/imx.c
1230
rx_ring->head > rx_ring->tail) {
drivers/tty/serial/imx.c
1233
r_bytes = rx_ring->head - rx_ring->tail;
drivers/tty/serial/imx.c
1259
WARN_ON(rx_ring->head > sg_dma_len(sgl));
drivers/tty/serial/imx.c
1260
WARN_ON(rx_ring->head <= rx_ring->tail);
drivers/tty/serial/imx.c
1277
sport->rx_ring.head = 0;
drivers/tty/serial/jsm/jsm_cls.c
355
u16 head;
drivers/tty/serial/jsm/jsm_cls.c
365
head = ch->ch_r_head & RQUEUEMASK;
drivers/tty/serial/jsm/jsm_cls.c
371
qleft = tail - head - 1;
drivers/tty/serial/jsm/jsm_cls.c
417
ch->ch_equeue[head] = linestatus & (UART_LSR_BI | UART_LSR_PE
drivers/tty/serial/jsm/jsm_cls.c
419
ch->ch_rqueue[head] = readb(&ch->ch_cls_uart->txrx);
drivers/tty/serial/jsm/jsm_cls.c
423
if (ch->ch_equeue[head] & UART_LSR_PE)
drivers/tty/serial/jsm/jsm_cls.c
425
if (ch->ch_equeue[head] & UART_LSR_BI)
drivers/tty/serial/jsm/jsm_cls.c
427
if (ch->ch_equeue[head] & UART_LSR_FE)
drivers/tty/serial/jsm/jsm_cls.c
431
head = (head + 1) & RQUEUEMASK;
drivers/tty/serial/jsm/jsm_cls.c
438
ch->ch_r_head = head & RQUEUEMASK;
drivers/tty/serial/jsm/jsm_cls.c
439
ch->ch_e_head = head & EQUEUEMASK;
drivers/tty/serial/jsm/jsm_neo.c
282
u16 head;
drivers/tty/serial/jsm/jsm_neo.c
286
head = ch->ch_r_head & RQUEUEMASK;
drivers/tty/serial/jsm/jsm_neo.c
294
qleft = tail - head - 1;
drivers/tty/serial/jsm/jsm_neo.c
345
n = min(((u32) total), (RQUEUESIZE - (u32) head));
drivers/tty/serial/jsm/jsm_neo.c
366
memcpy_fromio(ch->ch_rqueue + head, &ch->ch_neo_uart->txrxburst, n);
drivers/tty/serial/jsm/jsm_neo.c
372
memset(ch->ch_equeue + head, 0, n);
drivers/tty/serial/jsm/jsm_neo.c
375
head = (head + n) & RQUEUEMASK;
drivers/tty/serial/jsm/jsm_neo.c
451
memcpy_fromio(ch->ch_rqueue + head, &ch->ch_neo_uart->txrxburst, 1);
drivers/tty/serial/jsm/jsm_neo.c
452
ch->ch_equeue[head] = (u8) linestatus;
drivers/tty/serial/jsm/jsm_neo.c
455
ch->ch_rqueue[head], ch->ch_equeue[head]);
drivers/tty/serial/jsm/jsm_neo.c
461
head = (head + 1) & RQUEUEMASK;
drivers/tty/serial/jsm/jsm_neo.c
470
ch->ch_r_head = head & RQUEUEMASK;
drivers/tty/serial/jsm/jsm_neo.c
471
ch->ch_e_head = head & EQUEUEMASK;
drivers/tty/serial/jsm/jsm_tty.c
519
u16 head;
drivers/tty/serial/jsm/jsm_tty.c
545
head = ch->ch_r_head & rmask;
drivers/tty/serial/jsm/jsm_tty.c
548
data_len = (head - tail) & rmask;
drivers/tty/serial/jsm/jsm_tty.c
581
ch->ch_portnum, head, tail);
drivers/tty/serial/jsm/jsm_tty.c
595
s = ((head >= tail) ? head : RQUEUESIZE) - tail;
drivers/tty/synclink_gt.c
165
static void flush_cond_wait(struct cond_wait **head);
drivers/tty/synclink_gt.c
2925
static void add_cond_wait(struct cond_wait **head, struct cond_wait *w)
drivers/tty/synclink_gt.c
2929
w->next = *head;
drivers/tty/synclink_gt.c
2930
*head = w;
drivers/tty/synclink_gt.c
2933
static void remove_cond_wait(struct cond_wait **head, struct cond_wait *cw)
drivers/tty/synclink_gt.c
2938
for (w = *head, prev = NULL ; w != NULL ; prev = w, w = w->next) {
drivers/tty/synclink_gt.c
2943
*head = w->next;
drivers/tty/synclink_gt.c
2949
static void flush_cond_wait(struct cond_wait **head)
drivers/tty/synclink_gt.c
2951
while (*head != NULL) {
drivers/tty/synclink_gt.c
2952
wake_up_interruptible(&(*head)->q);
drivers/tty/synclink_gt.c
2953
*head = (*head)->next;
drivers/tty/tty_buffer.c
128
while ((p = buf->head) != NULL) {
drivers/tty/tty_buffer.c
129
buf->head = p->next;
drivers/tty/tty_buffer.c
139
buf->head = &buf->sentinel;
drivers/tty/tty_buffer.c
233
while ((next = smp_load_acquire(&buf->head->next)) != NULL) {
drivers/tty/tty_buffer.c
234
tty_buffer_free(port, buf->head);
drivers/tty/tty_buffer.c
235
buf->head = next;
drivers/tty/tty_buffer.c
237
buf->head->read = buf->head->commit;
drivers/tty/tty_buffer.c
238
buf->head->lookahead = buf->head->read;
drivers/tty/tty_buffer.c
397
static void lookahead_bufs(struct tty_port *port, struct tty_buffer *head)
drivers/tty/tty_buffer.c
399
head->lookahead = max(head->lookahead, head->read);
drivers/tty/tty_buffer.c
401
while (head) {
drivers/tty/tty_buffer.c
410
next = smp_load_acquire(&head->next);
drivers/tty/tty_buffer.c
415
count = smp_load_acquire(&head->commit) - head->lookahead;
drivers/tty/tty_buffer.c
417
head = next;
drivers/tty/tty_buffer.c
424
p = char_buf_ptr(head, head->lookahead);
drivers/tty/tty_buffer.c
425
if (head->flags)
drivers/tty/tty_buffer.c
426
f = flag_buf_ptr(head, head->lookahead);
drivers/tty/tty_buffer.c
431
head->lookahead += count;
drivers/tty/tty_buffer.c
436
receive_buf(struct tty_port *port, struct tty_buffer *head, size_t count)
drivers/tty/tty_buffer.c
438
u8 *p = char_buf_ptr(head, head->read);
drivers/tty/tty_buffer.c
442
if (head->flags)
drivers/tty/tty_buffer.c
443
f = flag_buf_ptr(head, head->read);
drivers/tty/tty_buffer.c
470
struct tty_buffer *head = buf->head;
drivers/tty/tty_buffer.c
482
next = smp_load_acquire(&head->next);
drivers/tty/tty_buffer.c
486
count = smp_load_acquire(&head->commit) - head->read;
drivers/tty/tty_buffer.c
490
buf->head = next;
drivers/tty/tty_buffer.c
491
tty_buffer_free(port, head);
drivers/tty/tty_buffer.c
495
rcvd = receive_buf(port, head, count);
drivers/tty/tty_buffer.c
496
head->read += rcvd;
drivers/tty/tty_buffer.c
498
lookahead_bufs(port, head);
drivers/tty/tty_buffer.c
581
buf->head = &buf->sentinel;
drivers/tty/tty_buffer.c
73
bool restart = buf->head->commit != buf->head->read;
drivers/tty/vt/vt_ioctl.c
113
struct list_head *pos, *head;
drivers/tty/vt/vt_ioctl.c
118
head = &vt_events;
drivers/tty/vt/vt_ioctl.c
120
list_for_each(pos, head) {
drivers/ufs/core/ufshcd.c
1104
struct list_head *head = &hba->clk_list_head;
drivers/ufs/core/ufshcd.c
1106
if (list_empty(head))
drivers/ufs/core/ufshcd.c
1109
list_for_each_entry(clki, head, list) {
drivers/ufs/core/ufshcd.c
1160
struct list_head *head = &hba->clk_list_head;
drivers/ufs/core/ufshcd.c
1166
list_for_each_entry(clki, head, list) {
drivers/ufs/core/ufshcd.c
1263
struct list_head *head = &hba->clk_list_head;
drivers/ufs/core/ufshcd.c
1265
if (list_empty(head))
drivers/ufs/core/ufshcd.c
1271
list_for_each_entry(clki, head, list) {
drivers/ufs/core/ufshcd.c
533
struct list_head *head = &hba->clk_list_head;
drivers/ufs/core/ufshcd.c
535
if (list_empty(head))
drivers/ufs/core/ufshcd.c
538
list_for_each_entry(clki, head, list) {
drivers/ufs/core/ufshcd.c
9513
struct list_head *head = &hba->clk_list_head;
drivers/ufs/core/ufshcd.c
9517
if (list_empty(head))
drivers/ufs/core/ufshcd.c
9524
list_for_each_entry(clki, head, list) {
drivers/ufs/core/ufshcd.c
9559
list_for_each_entry(clki, head, list) {
drivers/ufs/core/ufshcd.c
9595
struct list_head *head = &hba->clk_list_head;
drivers/ufs/core/ufshcd.c
9597
if (list_empty(head))
drivers/ufs/core/ufshcd.c
9600
list_for_each_entry(clki, head, list) {
drivers/ufs/host/cdns-pltfrm.c
112
struct list_head *head = &hba->clk_list_head;
drivers/ufs/host/cdns-pltfrm.c
116
if (list_empty(head))
drivers/ufs/host/cdns-pltfrm.c
119
list_for_each_entry(clki, head, list) {
drivers/ufs/host/ufs-exynos.c
453
struct list_head *head = &hba->clk_list_head;
drivers/ufs/host/ufs-exynos.c
460
if (list_empty(head))
drivers/ufs/host/ufs-exynos.c
463
list_for_each_entry(clki, head, list) {
drivers/ufs/host/ufs-mediatek.c
1002
list_for_each_entry(clki, head, list) {
drivers/ufs/host/ufs-mediatek.c
967
struct list_head *head = &hba->clk_list_head;
drivers/ufs/host/ufs-mediatek.c
978
list_for_each_entry_safe(clki, clki_tmp, head, list) {
drivers/ufs/host/ufs-qcom.c
1595
struct list_head *head = &hba->clk_list_head;
drivers/ufs/host/ufs-qcom.c
1610
list_for_each_entry(clki, head, list) {
drivers/ufs/host/ufshcd-pltfrm.c
224
struct list_head *head = &hba->clk_list_head;
drivers/ufs/host/ufshcd-pltfrm.c
230
list_for_each_entry(clki, head, list) {
drivers/usb/core/hcd.c
1682
list_replace_init(&bh->head, &local_list);
drivers/usb/core/hcd.c
1700
if (!list_empty(&bh->head)) {
drivers/usb/core/hcd.c
1751
list_add_tail(&urb->urb_list, &bh->head);
drivers/usb/core/hcd.c
2556
INIT_LIST_HEAD(&bh->head);
drivers/usb/core/phy.c
153
struct list_head *head;
drivers/usb/core/phy.c
159
head = &phy_roothub->list;
drivers/usb/core/phy.c
161
list_for_each_entry(roothub_entry, head, list) {
drivers/usb/core/phy.c
170
list_for_each_entry_continue_reverse(roothub_entry, head, list)
drivers/usb/core/phy.c
180
struct list_head *head;
drivers/usb/core/phy.c
186
head = &phy_roothub->list;
drivers/usb/core/phy.c
188
list_for_each_entry(roothub_entry, head, list) {
drivers/usb/core/phy.c
202
struct list_head *head;
drivers/usb/core/phy.c
208
head = &phy_roothub->list;
drivers/usb/core/phy.c
210
list_for_each_entry(roothub_entry, head, list) {
drivers/usb/core/phy.c
223
struct list_head *head;
drivers/usb/core/phy.c
229
head = &phy_roothub->list;
drivers/usb/core/phy.c
231
list_for_each_entry(roothub_entry, head, list) {
drivers/usb/core/phy.c
252
struct list_head *head;
drivers/usb/core/phy.c
258
head = &phy_roothub->list;
drivers/usb/core/phy.c
260
list_for_each_entry(roothub_entry, head, list) {
drivers/usb/core/phy.c
281
struct list_head *head;
drivers/usb/core/phy.c
287
head = &phy_roothub->list;
drivers/usb/core/phy.c
289
list_for_each_entry(roothub_entry, head, list) {
drivers/usb/core/phy.c
302
struct list_head *head;
drivers/usb/core/phy.c
308
head = &phy_roothub->list;
drivers/usb/core/phy.c
310
list_for_each_entry(roothub_entry, head, list) {
drivers/usb/core/phy.c
319
list_for_each_entry_continue_reverse(roothub_entry, head, list)
drivers/usb/fotg210/fotg210-hcd.c
2510
struct list_head *head)
drivers/usb/fotg210/fotg210-hcd.c
2514
list_for_each_entry_safe(qtd, temp, head, qtd_list) {
drivers/usb/fotg210/fotg210-hcd.c
2523
struct urb *urb, struct list_head *head, gfp_t flags)
drivers/usb/fotg210/fotg210-hcd.c
2539
list_add_tail(&qtd->qtd_list, head);
drivers/usb/fotg210/fotg210-hcd.c
2562
list_add_tail(&qtd->qtd_list, head);
drivers/usb/fotg210/fotg210-hcd.c
2633
list_add_tail(&qtd->qtd_list, head);
drivers/usb/fotg210/fotg210-hcd.c
2669
list_add_tail(&qtd->qtd_list, head);
drivers/usb/fotg210/fotg210-hcd.c
2679
return head;
drivers/usb/fotg210/fotg210-hcd.c
2682
qtd_list_free(fotg210, urb, head);
drivers/usb/fotg210/fotg210-hcd.c
2904
struct fotg210_qh *head;
drivers/usb/fotg210/fotg210-hcd.c
2916
head = fotg210->async;
drivers/usb/fotg210/fotg210-hcd.c
2917
qh->qh_next = head->qh_next;
drivers/usb/fotg210/fotg210-hcd.c
2918
qh->hw->hw_next = head->hw->hw_next;
drivers/usb/fotg210/fotg210-hcd.c
2921
head->qh_next.qh = qh;
drivers/usb/fotg210/fotg210-hcd.c
2922
head->hw->hw_next = dma;
drivers/usb/gadget/function/u_serial.c
497
static void gs_free_requests(struct usb_ep *ep, struct list_head *head,
drivers/usb/gadget/function/u_serial.c
502
while (!list_empty(head)) {
drivers/usb/gadget/function/u_serial.c
503
req = list_entry(head->next, struct usb_request, list);
drivers/usb/gadget/function/u_serial.c
511
static int gs_alloc_requests(struct usb_ep *ep, struct list_head *head,
drivers/usb/gadget/function/u_serial.c
526
return list_empty(head) ? -ENOMEM : 0;
drivers/usb/gadget/function/u_serial.c
528
list_add_tail(&req->list, head);
drivers/usb/gadget/function/u_serial.c
546
struct list_head *head = &port->read_pool;
drivers/usb/gadget/function/u_serial.c
557
status = gs_alloc_requests(ep, head, gs_read_complete,
drivers/usb/gadget/function/u_serial.c
565
gs_free_requests(ep, head, &port->read_allocated);
drivers/usb/gadget/function/u_serial.c
581
gs_free_requests(ep, head, &port->read_allocated);
drivers/usb/gadget/udc/fsl_udc_core.c
1608
curr_td = curr_req->head;
drivers/usb/gadget/udc/fsl_udc_core.c
175
next_td = req->head;
drivers/usb/gadget/udc/fsl_udc_core.c
740
cpu_to_hc32(req->head->td_dma & DTD_ADDR_MASK);
drivers/usb/gadget/udc/fsl_udc_core.c
765
fsl_prime_ep(ep, req->head);
drivers/usb/gadget/udc/fsl_udc_core.c
849
req->head = dtd;
drivers/usb/gadget/udc/fsl_udc_core.c
971
fsl_prime_ep(ep, next_req->head);
drivers/usb/gadget/udc/fsl_usb2_udc.h
451
struct ep_td_struct *head, *tail; /* For dTD List
drivers/usb/host/ehci-q.c
1172
struct list_head *head;
drivers/usb/host/ehci-q.c
1180
head = &qtd_list;
drivers/usb/host/ehci-q.c
1186
list_add_tail(&qtd->qtd_list, head);
drivers/usb/host/ehci-q.c
1237
list_add_tail(&qtd->qtd_list, head);
drivers/usb/host/ehci-q.c
1247
qtd_list_free(ehci, urb, head);
drivers/usb/host/ehci-q.c
575
struct list_head *head,
drivers/usb/host/ehci-q.c
592
list_add_tail (&qtd->qtd_list, head);
drivers/usb/host/ehci-q.c
615
list_add_tail (&qtd->qtd_list, head);
drivers/usb/host/ehci-q.c
686
list_add_tail (&qtd->qtd_list, head);
drivers/usb/host/ehci-q.c
722
list_add_tail (&qtd->qtd_list, head);
drivers/usb/host/ehci-q.c
732
return head;
drivers/usb/host/ehci-q.c
735
qtd_list_free (ehci, urb, head);
drivers/usb/host/ehci-q.c
981
struct ehci_qh *head;
drivers/usb/host/ehci-q.c
993
head = ehci->async;
drivers/usb/host/ehci-q.c
994
qh->qh_next = head->qh_next;
drivers/usb/host/ehci-q.c
995
qh->hw->hw_next = head->hw->hw_next;
drivers/usb/host/ehci-q.c
998
head->qh_next.qh = qh;
drivers/usb/host/ehci-q.c
999
head->hw->hw_next = dma;
drivers/usb/host/octeon-hcd.c
1210
while (available && (fifo->head != fifo->tail)) {
drivers/usb/host/octeon-hcd.c
1249
return fifo->head != fifo->tail;
drivers/usb/host/octeon-hcd.c
1259
if (usb->periodic.head != usb->periodic.tail) {
drivers/usb/host/octeon-hcd.c
1273
if (usb->nonperiodic.head != usb->nonperiodic.tail) {
drivers/usb/host/octeon-hcd.c
1328
fifo->entry[fifo->head].channel = channel;
drivers/usb/host/octeon-hcd.c
1329
fifo->entry[fifo->head].address =
drivers/usb/host/octeon-hcd.c
1332
fifo->entry[fifo->head].size = (usbc_hctsiz.s.xfersize + 3) >> 2;
drivers/usb/host/octeon-hcd.c
1333
fifo->head++;
drivers/usb/host/octeon-hcd.c
1334
if (fifo->head > MAX_CHANNELS)
drivers/usb/host/octeon-hcd.c
1335
fifo->head = 0;
drivers/usb/host/octeon-hcd.c
344
int head;
drivers/usb/host/ohci-hcd.c
753
u32 head;
drivers/usb/host/ohci-hcd.c
811
head = hc32_to_cpu(ohci, READ_ONCE(ed->hwHeadP)) & TD_MASK;
drivers/usb/host/ohci-hcd.c
815
if (head == (u32) td_next->td_dma)
drivers/usb/host/ohci-q.c
1002
head = hc32_to_cpu(ohci, ed->hwHeadP) & TD_MASK;
drivers/usb/host/ohci-q.c
1003
if (td->td_dma != head &&
drivers/usb/host/ohci-q.c
997
u32 head;
drivers/usb/host/oxu210hp-hcd.c
1577
struct urb *urb, struct list_head *head)
drivers/usb/host/oxu210hp-hcd.c
1581
list_for_each_entry_safe(qtd, temp, head, qtd_list) {
drivers/usb/host/oxu210hp-hcd.c
1591
struct list_head *head,
drivers/usb/host/oxu210hp-hcd.c
1608
list_add_tail(&qtd->qtd_list, head);
drivers/usb/host/oxu210hp-hcd.c
1639
list_add_tail(&qtd->qtd_list, head);
drivers/usb/host/oxu210hp-hcd.c
1700
list_add_tail(&qtd->qtd_list, head);
drivers/usb/host/oxu210hp-hcd.c
1733
list_add_tail(&qtd->qtd_list, head);
drivers/usb/host/oxu210hp-hcd.c
1742
return head;
drivers/usb/host/oxu210hp-hcd.c
1745
qtd_list_free(oxu, urb, head);
drivers/usb/host/oxu210hp-hcd.c
1895
struct ehci_qh *head;
drivers/usb/host/oxu210hp-hcd.c
1898
head = oxu->async;
drivers/usb/host/oxu210hp-hcd.c
1900
if (!head->qh_next.qh) {
drivers/usb/host/oxu210hp-hcd.c
1919
qh->qh_next = head->qh_next;
drivers/usb/host/oxu210hp-hcd.c
1920
qh->hw_next = head->hw_next;
drivers/usb/host/oxu210hp-hcd.c
1923
head->qh_next.qh = qh;
drivers/usb/host/oxu210hp-hcd.c
1924
head->hw_next = dma;
drivers/usb/host/uhci-debug.c
380
struct list_head *tmp, *head;
drivers/usb/host/uhci-debug.c
431
head = &td->fl_list;
drivers/usb/host/uhci-debug.c
432
tmp = head;
drivers/usb/host/uhci-debug.c
452
} while (tmp != head);
drivers/usb/host/uhci-debug.c
506
head = &qh->node;
drivers/usb/host/uhci-debug.c
507
tmp = head->next;
drivers/usb/host/uhci-debug.c
509
while (tmp != head) {
drivers/usb/host/xhci-dbgtty.c
228
struct list_head *head,
drivers/usb/host/xhci-dbgtty.c
247
list_add_tail(&req->list_pool, head);
drivers/usb/host/xhci-dbgtty.c
250
return list_empty(head) ? -ENOMEM : 0;
drivers/usb/host/xhci-dbgtty.c
254
xhci_dbc_free_requests(struct list_head *head)
drivers/usb/host/xhci-dbgtty.c
258
while (!list_empty(head)) {
drivers/usb/host/xhci-dbgtty.c
259
req = list_entry(head->next, struct dbc_request, list_pool);
drivers/usb/host/xhci-mtk-sch.c
526
u32 head, tail;
drivers/usb/host/xhci-mtk-sch.c
534
head = tt->fs_frame_bw[j];
drivers/usb/host/xhci-mtk-sch.c
537
head = tt->fs_frame_bw[j];
drivers/usb/host/xhci-mtk-sch.c
544
head += sch_ep->bw_budget_table[i];
drivers/usb/host/xhci-mtk-sch.c
549
if (head > FS_BW_BOUNDARY || tail > FS_BW_BOUNDARY)
drivers/usb/host/xhci.h
1278
#define xhci_for_each_ring_seg(head, seg) \
drivers/usb/host/xhci.h
1279
for (seg = head; seg != NULL; seg = (seg->next != head ? seg->next : NULL))
drivers/usb/isp1760/isp1760-hcd.c
1771
struct urb *urb, struct list_head *head, gfp_t flags)
drivers/usb/isp1760/isp1760-hcd.c
1803
list_add_tail(&qtd->qtd_list, head);
drivers/usb/isp1760/isp1760-hcd.c
1833
list_add_tail(&qtd->qtd_list, head);
drivers/usb/isp1760/isp1760-hcd.c
1868
list_add_tail(&qtd->qtd_list, head);
drivers/usb/isp1760/isp1760-hcd.c
1875
qtd_list_free(head);
drivers/usb/musb/cppi_dma.h
98
struct cppi_descriptor *head;
drivers/usb/musb/musb_host.c
2002
struct list_head *head = NULL;
drivers/usb/musb/musb_host.c
2009
head = &musb->control;
drivers/usb/musb/musb_host.c
2073
head = &musb->in_bulk;
drivers/usb/musb/musb_host.c
2075
head = &musb->out_bulk;
drivers/usb/musb/musb_host.c
2101
if (head) {
drivers/usb/musb/musb_host.c
2102
idle = list_empty(head);
drivers/usb/musb/musb_host.c
2103
list_add_tail(&qh->ring, head);
drivers/usb/musb/musb_host.c
335
struct list_head *head;
drivers/usb/musb/musb_host.c
364
head = qh->ring.prev;
drivers/usb/musb/musb_host.c
367
qh = first_qh(head);
drivers/usb/phy/phy.c
56
list_for_each_entry(phy, list, head) {
drivers/usb/phy/phy.c
649
INIT_LIST_HEAD(&x->head);
drivers/usb/phy/phy.c
660
list_for_each_entry(phy, &phy_list, head) {
drivers/usb/phy/phy.c
670
list_add_tail(&x->head, &phy_list);
drivers/usb/phy/phy.c
701
INIT_LIST_HEAD(&x->head);
drivers/usb/phy/phy.c
713
list_add_tail(&x->head, &phy_list);
drivers/usb/phy/phy.c
73
list_for_each_entry(phy, &phy_list, head) {
drivers/usb/phy/phy.c
732
list_del(&x->head);
drivers/usb/phy/phy.c
87
list_for_each_entry(usb_phy, &phy_list, head) {
drivers/usb/serial/io_edgeport.c
1155
bytesleft = fifo->size - fifo->head;
drivers/usb/serial/io_edgeport.c
1161
memcpy(&fifo->fifo[fifo->head], data, firsthalf);
drivers/usb/serial/io_edgeport.c
1162
usb_serial_debug_data(&port->dev, __func__, firsthalf, &fifo->fifo[fifo->head]);
drivers/usb/serial/io_edgeport.c
1165
fifo->head += firsthalf;
drivers/usb/serial/io_edgeport.c
1169
if (fifo->head == fifo->size)
drivers/usb/serial/io_edgeport.c
1170
fifo->head = 0;
drivers/usb/serial/io_edgeport.c
1176
memcpy(&fifo->fifo[fifo->head], &data[firsthalf], secondhalf);
drivers/usb/serial/io_edgeport.c
1177
usb_serial_debug_data(&port->dev, __func__, secondhalf, &fifo->fifo[fifo->head]);
drivers/usb/serial/io_edgeport.c
1180
fifo->head += secondhalf;
drivers/usb/serial/io_edgeport.c
155
unsigned int head; /* index to head pointer (write) */
drivers/usb/serial/io_edgeport.c
913
edge_port->txfifo.head = 0;
drivers/usb/storage/isd200.c
1233
unsigned char sectnum, head;
drivers/usb/storage/isd200.c
1324
head = ATA_ADDRESS_DEVHEAD_LBA_MODE | (unsigned char)(lba>>24 & 0x0F);
drivers/usb/storage/isd200.c
1329
head = (u8)((lba / id[ATA_ID_SECTORS]) %
drivers/usb/storage/isd200.c
1343
ataCdb->write.DeviceHeadByte = (head | ATA_ADDRESS_DEVHEAD_STD);
drivers/usb/storage/isd200.c
1356
head = ATA_ADDRESS_DEVHEAD_LBA_MODE | (unsigned char)(lba>>24 & 0x0F);
drivers/usb/storage/isd200.c
1361
head = (u8)((lba / id[ATA_ID_SECTORS]) %
drivers/usb/storage/isd200.c
1375
ataCdb->write.DeviceHeadByte = (head | ATA_ADDRESS_DEVHEAD_STD);
drivers/vdpa/mlx5/core/mlx5_vdpa.h
31
struct list_head head;
drivers/vdpa/mlx5/core/mlx5_vdpa.h
65
unsigned short head;
drivers/vdpa/mlx5/core/mr.c
140
s = list_first_entry_or_null(&mkey->head, struct mlx5_vdpa_direct_mr, list);
drivers/vdpa/mlx5/core/mr.c
152
s = list_first_entry_or_null(&mkey->head, struct mlx5_vdpa_direct_mr, list);
drivers/vdpa/mlx5/core/mr.c
156
e = list_last_entry(&mkey->head, struct mlx5_vdpa_direct_mr, list);
drivers/vdpa/mlx5/core/mr.c
180
list_for_each_entry(dmr, &mkey->head, list) {
drivers/vdpa/mlx5/core/mr.c
222
list_for_each_entry(dmr, &mr->head, list) {
drivers/vdpa/mlx5/core/mr.c
252
list_for_each_entry(dmr, &mr->head, list) {
drivers/vdpa/mlx5/core/mr.c
295
list_for_each_entry(dmr, &mr->head, list) {
drivers/vdpa/mlx5/core/mr.c
312
list_for_each_entry(dmr, &mr->head, list) {
drivers/vdpa/mlx5/core/mr.c
480
list_splice_tail(&tmp, &mr->head);
drivers/vdpa/mlx5/core/mr.c
484
list_for_each_entry_safe(dmr, n, &mr->head, list) {
drivers/vdpa/mlx5/core/mr.c
513
INIT_LIST_HEAD(&mr->head);
drivers/vdpa/mlx5/core/mr.c
558
list_for_each_entry_safe_reverse(dmr, n, &mr->head, list) {
drivers/vdpa/mlx5/core/mr.c
635
list_for_each_entry_safe_reverse(dmr, n, &mr->head, list) {
drivers/vdpa/mlx5/net/mlx5_vnet.c
2393
err = vringh_getdesc_iotlb(&cvq->vring, &cvq->riov, &cvq->wiov, &cvq->head,
drivers/vdpa/mlx5/net/mlx5_vnet.c
2421
vringh_complete_iotlb(&cvq->vring, cvq->head, write);
drivers/vdpa/mlx5/net/mlx5_vnet.c
369
vqp->head += n;
drivers/vdpa/mlx5/net/mlx5_vnet.c
370
vqp->db.db[0] = cpu_to_be32(vqp->head);
drivers/vdpa/mlx5/net/mlx5_vnet.c
88
u16 head;
drivers/vdpa/vdpa_sim/vdpa_sim.h
26
unsigned short head;
drivers/vdpa/vdpa_sim/vdpa_sim_blk.c
124
&vq->head, GFP_ATOMIC);
drivers/vdpa/vdpa_sim/vdpa_sim_blk.c
319
vringh_complete_iotlb(&vq->vring, vq->head, pushed);
drivers/vdpa/vdpa_sim/vdpa_sim_net.c
146
&cvq->head, GFP_ATOMIC);
drivers/vdpa/vdpa_sim/vdpa_sim_net.c
176
vringh_complete_iotlb(&cvq->vring, cvq->head, write);
drivers/vdpa/vdpa_sim/vdpa_sim_net.c
221
&txq->head, GFP_ATOMIC);
drivers/vdpa/vdpa_sim/vdpa_sim_net.c
241
&rxq->head, GFP_ATOMIC);
drivers/vdpa/vdpa_sim/vdpa_sim_net.c
74
vringh_complete_iotlb(&vq->vring, vq->head, len);
drivers/vdpa/vdpa_user/vduse_dev.c
191
static struct vduse_dev_msg *vduse_find_msg(struct list_head *head,
drivers/vdpa/vdpa_user/vduse_dev.c
196
list_for_each_entry(msg, head, list) {
drivers/vdpa/vdpa_user/vduse_dev.c
206
static struct vduse_dev_msg *vduse_dequeue_msg(struct list_head *head)
drivers/vdpa/vdpa_user/vduse_dev.c
210
if (!list_empty(head)) {
drivers/vdpa/vdpa_user/vduse_dev.c
211
msg = list_first_entry(head, struct vduse_dev_msg, list);
drivers/vdpa/vdpa_user/vduse_dev.c
218
static void vduse_enqueue_msg(struct list_head *head,
drivers/vdpa/vdpa_user/vduse_dev.c
221
list_add_tail(&msg->list, head);
drivers/vfio/group.c
459
WARN_ON(group->notifier.head);
drivers/vfio/group.c
723
WARN_ON(group->notifier.head);
drivers/vfio/vfio_iommu_type1.c
2015
static int vfio_iommu_iova_insert(struct list_head *head,
drivers/vfio/vfio_iommu_type1.c
2028
list_add_tail(&region->list, head);
drivers/vhost/net.c
1001
head = skb_peek(&sk->sk_receive_queue);
drivers/vhost/net.c
1002
if (likely(head)) {
drivers/vhost/net.c
1003
len = head->len;
drivers/vhost/net.c
1004
if (skb_vlan_tag_present(head))
drivers/vhost/net.c
109
int head;
drivers/vhost/net.c
156
if (rxq->tail != rxq->head)
drivers/vhost/net.c
157
return rxq->queue[rxq->head];
drivers/vhost/net.c
164
return rxq->tail - rxq->head;
drivers/vhost/net.c
169
return rxq->tail == rxq->head;
drivers/vhost/net.c
175
++rxq->head;
drivers/vhost/net.c
183
rxq->head = 0;
drivers/vhost/net.c
194
ptr_ring_unconsume(nvq->rx_ring, rxq->queue + rxq->head,
drivers/vhost/net.c
197
rxq->head = rxq->tail = 0;
drivers/vhost/net.c
228
rxq->head = rxq->tail = 0;
drivers/vhost/net.c
757
int head;
drivers/vhost/net.c
778
head = get_tx_bufs(net, nvq, &msg, &out, &in, &len,
drivers/vhost/net.c
781
if (unlikely(head < 0))
drivers/vhost/net.c
784
if (head == vq->num) {
drivers/vhost/net.c
844
vq->heads[0].id = cpu_to_vhost32(vq, head);
drivers/vhost/net.c
846
vq->heads[nvq->done_idx].id = cpu_to_vhost32(vq, head);
drivers/vhost/net.c
860
int head;
drivers/vhost/net.c
884
head = get_tx_bufs(net, nvq, &msg, &out, &in, &len,
drivers/vhost/net.c
887
if (unlikely(head < 0))
drivers/vhost/net.c
890
if (head == vq->num) {
drivers/vhost/net.c
907
vq->heads[nvq->upend_idx].id = cpu_to_vhost32(vq, head);
drivers/vhost/net.c
956
vhost_add_used_and_signal(&net->dev, vq, head, 0);
drivers/vhost/net.c
993
struct sk_buff *head;
drivers/vhost/scsi.c
1092
vhost_add_used_and_signal(&vs->dev, vq, vc->head, 0);
drivers/vhost/scsi.c
1134
vhost_add_used_and_signal(&vs->dev, vq, vc->head, 0);
drivers/vhost/scsi.c
1149
vc->head = vhost_get_vq_desc(vq, vq->iov,
drivers/vhost/scsi.c
1154
vc->head, vc->out, vc->in);
drivers/vhost/scsi.c
1157
if (unlikely(vc->head < 0))
drivers/vhost/scsi.c
1161
if (vc->head == vq->num) {
drivers/vhost/scsi.c
1498
cmd->tvc_vq_desc = vc.head;
drivers/vhost/scsi.c
1612
tmf->vq_desc = vc->head;
drivers/vhost/scsi.c
1639
vhost_scsi_send_tmf_resp(vs, vq, vc->in, vc->head, &vq->iov[vc->out],
drivers/vhost/scsi.c
1661
vhost_add_used_and_signal(&vs->dev, vq, vc->head, 0);
drivers/vhost/scsi.c
281
int head;
drivers/vhost/scsi.c
578
int head, ret;
drivers/vhost/scsi.c
597
head = vhost_get_vq_desc(vq, vq->iov,
drivers/vhost/scsi.c
600
if (head < 0) {
drivers/vhost/scsi.c
604
if (head == vq->num) {
drivers/vhost/scsi.c
626
vhost_add_used_and_signal(&vs->dev, vq, head, 0);
drivers/vhost/test.c
53
int head;
drivers/vhost/test.c
67
head = vhost_get_vq_desc(vq, vq->iov,
drivers/vhost/test.c
72
if (unlikely(head < 0))
drivers/vhost/test.c
75
if (head == vq->num) {
drivers/vhost/test.c
93
vhost_add_used_and_signal(&n->dev, vq, head, 0);
drivers/vhost/vdpa.c
104
struct hlist_head *head = &v->as[asid % VHOST_VDPA_IOTLB_BUCKETS];
drivers/vhost/vdpa.c
119
hlist_add_head(&as->hash_link, head);
drivers/vhost/vdpa.c
82
struct hlist_head *head = &v->as[asid % VHOST_VDPA_IOTLB_BUCKETS];
drivers/vhost/vdpa.c
85
hlist_for_each_entry(as, head, hash_link)
drivers/vhost/vhost.c
1463
struct vring_used_elem *head, int idx,
drivers/vhost/vhost.c
1466
return vhost_copy_to_user(vq, vq->used->ring + idx, head,
drivers/vhost/vhost.c
1467
count * sizeof(*head));
drivers/vhost/vhost.c
1555
__virtio16 *head, int idx)
drivers/vhost/vhost.c
1557
return vhost_get_avail(vq, *head,
drivers/vhost/vhost.c
2823
unsigned int i, head, found = 0;
drivers/vhost/vhost.c
2838
head = vq->next_avail_head & (vq->num - 1);
drivers/vhost/vhost.c
2849
head = vhost16_to_cpu(vq, ring_head);
drivers/vhost/vhost.c
2853
if (unlikely(head >= vq->num)) {
drivers/vhost/vhost.c
2855
head, vq->num);
drivers/vhost/vhost.c
2864
i = head;
drivers/vhost/vhost.c
2869
i, vq->num, head);
drivers/vhost/vhost.c
2875
i, vq->num, head);
drivers/vhost/vhost.c
2943
return head;
drivers/vhost/vhost.c
2985
int vhost_add_used(struct vhost_virtqueue *vq, unsigned int head, int len)
drivers/vhost/vhost.c
2988
cpu_to_vhost32(vq, head),
drivers/vhost/vhost.c
3179
unsigned int head, int len)
drivers/vhost/vhost.c
3181
vhost_add_used(vq, head, len);
drivers/vhost/vhost.c
3280
void vhost_enqueue_msg(struct vhost_dev *dev, struct list_head *head,
drivers/vhost/vhost.c
3284
list_add_tail(&node->node, head);
drivers/vhost/vhost.c
3292
struct list_head *head)
drivers/vhost/vhost.c
3297
if (!list_empty(head)) {
drivers/vhost/vhost.c
3298
node = list_first_entry(head, struct vhost_msg_node,
drivers/vhost/vhost.h
248
int vhost_add_used(struct vhost_virtqueue *, unsigned int head, int len);
drivers/vhost/vhost.h
268
struct list_head *head,
drivers/vhost/vhost.h
271
struct list_head *head);
drivers/vhost/vringh.c
1407
u16 *head,
drivers/vhost/vringh.c
1420
*head = err;
drivers/vhost/vringh.c
1421
err = __vringh_iov(vrh, *head, riov, wiov, no_range_check, NULL,
drivers/vhost/vringh.c
1473
int vringh_complete_iotlb(struct vringh *vrh, u16 head, u32 len)
drivers/vhost/vringh.c
1477
used.id = cpu_to_vringh32(vrh, head);
drivers/vhost/vringh.c
43
u16 avail_idx, i, head;
drivers/vhost/vringh.c
61
err = getu16(vrh, &head, &vrh->vring.avail->ring[i]);
drivers/vhost/vringh.c
68
if (head >= vrh->vring.num) {
drivers/vhost/vringh.c
70
head, vrh->vring.num);
drivers/vhost/vringh.c
707
u16 *head)
drivers/vhost/vringh.c
711
*head = vrh->vring.num;
drivers/vhost/vringh.c
740
*head = err;
drivers/vhost/vringh.c
741
err = __vringh_iov(vrh, *head, (struct vringh_kiov *)riov,
drivers/vhost/vringh.c
75
return head;
drivers/vhost/vringh.c
791
int vringh_complete_user(struct vringh *vrh, u16 head, u32 len)
drivers/vhost/vringh.c
795
used.id = cpu_to_vringh32(vrh, head);
drivers/vhost/vringh.c
948
u16 *head,
drivers/vhost/vringh.c
961
*head = err;
drivers/vhost/vringh.c
962
err = __vringh_iov(vrh, *head, riov, wiov, no_range_check, NULL,
drivers/vhost/vringh.c
980
int vringh_complete_kern(struct vringh *vrh, u16 head, u32 len)
drivers/vhost/vringh.c
984
used.id = cpu_to_vringh32(vrh, head);
drivers/vhost/vsock.c
123
int head;
drivers/vhost/vsock.c
132
head = vhost_get_vq_desc(vq, vq->iov, ARRAY_SIZE(vq->iov),
drivers/vhost/vsock.c
134
if (head < 0) {
drivers/vhost/vsock.c
139
if (head == vq->num) {
drivers/vhost/vsock.c
221
vhost_add_used(vq, head, sizeof(*hdr) + payload_len);
drivers/vhost/vsock.c
500
int head, pkts = 0, total_len = 0;
drivers/vhost/vsock.c
525
head = vhost_get_vq_desc(vq, vq->iov, ARRAY_SIZE(vq->iov),
drivers/vhost/vsock.c
527
if (head < 0)
drivers/vhost/vsock.c
530
if (head == vq->num) {
drivers/vhost/vsock.c
560
vhost_add_used(vq, head, 0);
drivers/video/fbdev/core/fbcon.c
2813
void fbcon_delete_modelist(struct list_head *head)
drivers/video/fbdev/core/fbcon.c
2817
list_for_each_entry(modelist, head, list)
drivers/video/fbdev/core/modedb.c
1000
struct list_head *head)
drivers/video/fbdev/core/modedb.c
1006
list_for_each_entry(modelist, head, list) {
drivers/video/fbdev/core/modedb.c
1037
struct list_head *head)
drivers/video/fbdev/core/modedb.c
1043
list_for_each_entry(modelist, head, list) {
drivers/video/fbdev/core/modedb.c
1059
int fb_add_videomode(const struct fb_videomode *mode, struct list_head *head)
drivers/video/fbdev/core/modedb.c
1065
list_for_each_entry(modelist, head, list) {
drivers/video/fbdev/core/modedb.c
1078
list_add(&modelist->list, head);
drivers/video/fbdev/core/modedb.c
1092
struct list_head *head)
drivers/video/fbdev/core/modedb.c
1098
list_for_each_safe(pos, n, head) {
drivers/video/fbdev/core/modedb.c
1112
void fb_destroy_modelist(struct list_head *head)
drivers/video/fbdev/core/modedb.c
1116
list_for_each_safe(pos, n, head) {
drivers/video/fbdev/core/modedb.c
1130
struct list_head *head)
drivers/video/fbdev/core/modedb.c
1134
INIT_LIST_HEAD(head);
drivers/video/fbdev/core/modedb.c
1137
if (fb_add_videomode(&modedb[i], head))
drivers/video/fbdev/core/modedb.c
1143
struct list_head *head)
drivers/video/fbdev/core/modedb.c
1149
if (!head->prev || !head->next || list_empty(head))
drivers/video/fbdev/core/modedb.c
1153
list_for_each_entry(modelist, head, list) {
drivers/video/fbdev/core/modedb.c
1180
m = fb_find_best_mode(&var, head);
drivers/video/fbdev/core/modedb.c
965
struct list_head *head)
drivers/video/fbdev/core/modedb.c
971
list_for_each_entry(modelist, head, list) {
drivers/video/fbdev/i810/i810_accel.c
62
u32 head, count = WAIT_COUNT, tail;
drivers/video/fbdev/i810/i810_accel.c
67
head = i810_readl(IRING + 4, mmio) & RBUFFER_HEAD_MASK;
drivers/video/fbdev/i810/i810_accel.c
68
if ((tail == head) ||
drivers/video/fbdev/i810/i810_accel.c
69
(tail > head &&
drivers/video/fbdev/i810/i810_accel.c
70
(par->iring.size - tail + head) >= space) ||
drivers/video/fbdev/i810/i810_accel.c
71
(tail < head && (head - tail) >= space)) {
drivers/video/fbdev/nvidia/nv_hw.c
1519
NV_WR32(par->PCRTC0, 0x0860, state->head);
drivers/video/fbdev/nvidia/nv_hw.c
1661
state->head = NV_RD32(par->PCRTC0, 0x0860);
drivers/video/fbdev/nvidia/nv_setup.c
182
static void NVSelectHeadRegisters(struct nvidia_par *par, int head)
drivers/video/fbdev/nvidia/nv_setup.c
184
if (head) {
drivers/video/fbdev/nvidia/nv_type.h
78
u32 head;
drivers/video/fbdev/nvidia/nvidia.c
441
state->head = NV_RD32(par->PCRTC0, 0x00000860) & ~0x00001000;
drivers/video/fbdev/nvidia/nvidia.c
449
state->head = NV_RD32(par->PCRTC0, 0x00000860) | 0x00001000;
drivers/video/fbdev/omap/hwa742.c
303
static void submit_req_list(struct list_head *head)
drivers/video/fbdev/omap/hwa742.c
311
list_splice_init(head, hwa742.pending_req_list.prev);
drivers/video/fbdev/ps3fb.c
916
struct display_head *head = &ps3fb.dinfo->display_head[1];
drivers/video/fbdev/ps3fb.c
927
ps3fb.vblank_count = head->vblank_count;
drivers/video/fbdev/riva/fbdev.c
786
newmode.ext.head = NV_RD32(par->riva.PCRTC0, 0x00000860) &
drivers/video/fbdev/riva/fbdev.c
794
newmode.ext.head = NV_RD32(par->riva.PCRTC0, 0x00000860) |
drivers/video/fbdev/riva/riva_hw.c
1531
NV_WR32(chip->PCRTC0, 0x00000860, state->head);
drivers/video/fbdev/riva/riva_hw.c
1777
state->head = NV_RD32(chip->PCRTC0, 0x00000860);
drivers/video/fbdev/riva/riva_hw.h
514
U032 head;
drivers/video/fbdev/sm501fb.c
105
enum sm501_controller head;
drivers/video/fbdev/sm501fb.c
1056
if (par->head == HEAD_CRT)
drivers/video/fbdev/sm501fb.c
1190
enum sm501_controller head;
drivers/video/fbdev/sm501fb.c
1197
head = HEAD_CRT;
drivers/video/fbdev/sm501fb.c
1199
head = HEAD_PANEL;
drivers/video/fbdev/sm501fb.c
1203
dev_info(dev, "setting crt source to head %d\n", head);
drivers/video/fbdev/sm501fb.c
1207
if (head == HEAD_CRT) {
drivers/video/fbdev/sm501fb.c
1686
static int sm501fb_init_fb(struct fb_info *fb, enum sm501_controller head,
drivers/video/fbdev/sm501fb.c
1696
switch (head) {
drivers/video/fbdev/sm501fb.c
1728
if (head == HEAD_CRT && info->pdata->fb_route == SM501_FB_CRT_PANEL) {
drivers/video/fbdev/sm501fb.c
1737
(head == HEAD_CRT) ? &sm501fb_ops_crt : &sm501fb_ops_pnl,
drivers/video/fbdev/sm501fb.c
1872
enum sm501_controller head)
drivers/video/fbdev/sm501fb.c
1874
unsigned char *name = (head == HEAD_CRT) ? "crt" : "panel";
drivers/video/fbdev/sm501fb.c
1879
pd = (head == HEAD_CRT) ? info->pdata->fb_crt : info->pdata->fb_pnl;
drivers/video/fbdev/sm501fb.c
1893
par->head = head;
drivers/video/fbdev/sm501fb.c
1896
info->fb[head] = fbi;
drivers/video/fbdev/sm501fb.c
1904
enum sm501_controller head)
drivers/video/fbdev/sm501fb.c
1906
struct fb_info *fbi = info->fb[head];
drivers/video/fbdev/sm501fb.c
1915
enum sm501_controller head, const char *drvname)
drivers/video/fbdev/sm501fb.c
1917
struct fb_info *fbi = info->fb[head];
drivers/video/fbdev/sm501fb.c
1923
mutex_init(&info->fb[head]->mm_lock);
drivers/video/fbdev/sm501fb.c
1925
ret = sm501fb_init_fb(info->fb[head], head, drvname);
drivers/video/fbdev/sm501fb.c
1931
ret = register_framebuffer(info->fb[head]);
drivers/video/fbdev/sm501fb.c
1934
sm501_free_init_fb(info, head);
drivers/video/fbdev/sm501fb.c
2084
enum sm501_controller head)
drivers/video/fbdev/sm501fb.c
2086
struct fb_info *fbi = info->fb[head];
drivers/video/fbdev/sm501fb.c
2135
enum sm501_controller head)
drivers/video/fbdev/sm501fb.c
2137
struct fb_info *fbi = info->fb[head];
drivers/video/fbdev/sm501fb.c
450
switch (par->head) {
drivers/video/fbdev/sm501fb.c
539
if (par->head == HEAD_CRT)
drivers/video/fbdev/sm501fb.c
549
smc501_writel(reg, fbi->regs + (par->head == HEAD_CRT ?
drivers/video/fbdev/sm501fb.c
928
if (par->head == HEAD_CRT)
drivers/virt/fsl_hypervisor.c
470
unsigned int head;
drivers/virt/fsl_hypervisor.c
498
if (dbq->head != nextp(dbq->tail)) {
drivers/virt/fsl_hypervisor.c
579
mask = (dbq->head == dbq->tail) ? 0 : (EPOLLIN | EPOLLRDNORM);
drivers/virt/fsl_hypervisor.c
612
if (dbq->head == dbq->tail) {
drivers/virt/fsl_hypervisor.c
619
dbq->head != dbq->tail))
drivers/virt/fsl_hypervisor.c
637
dbell = dbq->q[dbq->head];
drivers/virt/fsl_hypervisor.c
638
dbq->head = nextp(dbq->head);
drivers/virtio/virtio_ring.c
1501
u16 head;
drivers/virtio/virtio_ring.c
1504
head = vq->packed.next_avail_idx;
drivers/virtio/virtio_ring.c
1551
vq->packed.vring.desc[head].addr = cpu_to_le64(addr);
drivers/virtio/virtio_ring.c
1552
vq->packed.vring.desc[head].len = cpu_to_le32(total_sg *
drivers/virtio/virtio_ring.c
1554
vq->packed.vring.desc[head].id = cpu_to_le16(id);
drivers/virtio/virtio_ring.c
1570
vq->packed.vring.desc[head].flags = cpu_to_le16(VRING_DESC_F_INDIRECT |
drivers/virtio/virtio_ring.c
1577
n = head + 1;
drivers/virtio/virtio_ring.c
1598
pr_debug("Added buffer head %i to %p\n", head, vq);
drivers/virtio/virtio_ring.c
1630
u16 head, id, prev, curr, avail_used_flags;
drivers/virtio/virtio_ring.c
1661
head = vq->packed.next_avail_idx;
drivers/virtio/virtio_ring.c
1667
i = head;
drivers/virtio/virtio_ring.c
1694
if (i == head)
drivers/virtio/virtio_ring.c
1722
if (i <= head)
drivers/virtio/virtio_ring.c
1744
vq->packed.vring.desc[head].flags = head_flags;
drivers/virtio/virtio_ring.c
1747
pr_debug("Added buffer head %i to %p\n", head, vq);
drivers/virtio/virtio_ring.c
1754
i = head;
drivers/virtio/virtio_ring.c
1788
u16 head, avail_used_flags;
drivers/virtio/virtio_ring.c
1819
head = vq->packed.next_avail_idx;
drivers/virtio/virtio_ring.c
1826
i = head;
drivers/virtio/virtio_ring.c
1854
if (i == head)
drivers/virtio/virtio_ring.c
1861
desc[i].id = cpu_to_le16(head);
drivers/virtio/virtio_ring.c
1891
vq->packed.desc_state[head].num = total_sg;
drivers/virtio/virtio_ring.c
1892
vq->packed.desc_state[head].data = data;
drivers/virtio/virtio_ring.c
1893
vq->packed.desc_state[head].indir_desc = ctx;
drivers/virtio/virtio_ring.c
1894
vq->packed.desc_state[head].total_in_len = total_in_len;
drivers/virtio/virtio_ring.c
1902
vq->packed.vring.desc[head].flags = head_flags;
drivers/virtio/virtio_ring.c
1905
pr_debug("Added buffer head %i to %p\n", head, vq);
drivers/virtio/virtio_ring.c
1912
i = head;
drivers/virtio/virtio_ring.c
616
int head;
drivers/virtio/virtio_ring.c
633
head = vq->free_head;
drivers/virtio/virtio_ring.c
653
i = head;
drivers/virtio/virtio_ring.c
723
head, addr,
drivers/virtio/virtio_ring.c
736
vq->split.desc_state[head].total_in_len = total_in_len;
drivers/virtio/virtio_ring.c
738
vq->free_head = vq->split.desc_extra[head].next;
drivers/virtio/virtio_ring.c
743
vq->split.desc_state[head].data = data;
drivers/virtio/virtio_ring.c
745
vq->split.desc_state[head].indir_desc = desc;
drivers/virtio/virtio_ring.c
747
vq->split.desc_state[head].indir_desc = ctx;
drivers/virtio/virtio_ring.c
752
vq->split.vring.avail->ring[avail] = cpu_to_virtio16(vq->vq.vdev, head);
drivers/virtio/virtio_ring.c
762
pr_debug("Added buffer head %i to %p\n", head, vq);
drivers/virtio/virtio_ring.c
778
i = head;
drivers/virtio/virtio_ring.c
825
unsigned int head)
drivers/virtio/virtio_ring.c
828
struct vring_desc *indir_desc = vq->split.desc_state[head].indir_desc;
drivers/virtio/virtio_ring.c
835
len = vq->split.desc_extra[head].len;
drivers/virtio/virtio_ring.c
837
BUG_ON(!(vq->split.desc_extra[head].flags &
drivers/virtio/virtio_ring.c
851
vq->split.desc_state[head].indir_desc = NULL;
drivers/virtio/virtio_ring.c
855
unsigned int head,
drivers/virtio/virtio_ring.c
863
vq->split.desc_state[head].data = NULL;
drivers/virtio/virtio_ring.c
868
i = head;
drivers/virtio/virtio_ring.c
881
detach_indirect_split(vq, head);
drivers/virtio/virtio_ring.c
883
*ctx = vq->split.desc_state[head].indir_desc;
drivers/virtio/virtio_ring.c
888
static void detach_buf_split(struct vring_virtqueue *vq, unsigned int head,
drivers/virtio/virtio_ring.c
891
unsigned int i = detach_buf_split_in_order(vq, head, ctx);
drivers/virtio/virtio_ring.c
894
vq->free_head = head;
drivers/virtio/virtio_rtc_driver.c
321
struct virtio_rtc_notif_head *head = token;
drivers/virtio/virtio_rtc_driver.c
326
if (len < sizeof(*head)) {
drivers/virtio/virtio_rtc_driver.c
333
if (virtio_le_to_cpu(head->msg_type) != VIRTIO_RTC_NOTIF_ALARM) {
drivers/virtio/virtio_rtc_driver.c
336
__func__, virtio_le_to_cpu(head->msg_type));
drivers/xen/events/events_fifo.c
112
q->head[i] = 0;
drivers/xen/events/events_fifo.c
278
uint32_t head;
drivers/xen/events/events_fifo.c
282
head = q->head[priority];
drivers/xen/events/events_fifo.c
288
if (head == 0) {
drivers/xen/events/events_fifo.c
290
head = control_block->head[priority];
drivers/xen/events/events_fifo.c
293
port = head;
drivers/xen/events/events_fifo.c
295
head = clear_linked(word);
drivers/xen/events/events_fifo.c
304
if (head == 0)
drivers/xen/events/events_fifo.c
314
q->head[priority] = head;
drivers/xen/events/events_fifo.c
61
uint32_t head[EVTCHN_FIFO_MAX_QUEUES];
drivers/xen/grant-table.c
183
grant_ref_t head;
drivers/xen/grant-table.c
193
ref = head = gnttab_free_head;
drivers/xen/grant-table.c
196
bitmap_clear(gnttab_free_bitmap, head, 1);
drivers/xen/grant-table.c
197
if (gnttab_free_tail_ptr == __gnttab_entry(head))
drivers/xen/grant-table.c
200
head = gnttab_entry(head);
drivers/xen/grant-table.c
202
gnttab_free_head = gnttab_entry(head);
drivers/xen/grant-table.c
203
gnttab_entry(head) = GNTTAB_LIST_END;
drivers/xen/grant-table.c
619
void gnttab_free_grant_references(grant_ref_t head)
drivers/xen/grant-table.c
625
while (head != GNTTAB_LIST_END) {
drivers/xen/grant-table.c
626
ref = gnttab_entry(head);
drivers/xen/grant-table.c
627
put_free_entry_locked(head);
drivers/xen/grant-table.c
628
head = ref;
drivers/xen/grant-table.c
635
void gnttab_free_grant_reference_seq(grant_ref_t head, unsigned int count)
drivers/xen/grant-table.c
642
put_free_entry_locked(head + i - 1);
drivers/xen/grant-table.c
648
int gnttab_alloc_grant_references(u16 count, grant_ref_t *head)
drivers/xen/grant-table.c
655
*head = h;
drivers/xen/xen-pciback/vpci.c
26
static inline struct list_head *list_first(struct list_head *head)
drivers/xen/xen-pciback/vpci.c
28
return head->next;
drivers/xen/xen-scsiback.c
1236
struct list_head *head = &(info->v2p_entry_lists);
drivers/xen/xen-scsiback.c
1242
list_cut_before(&tmp_list, head, head);
drivers/xen/xen-scsiback.c
652
struct list_head *head = &(info->v2p_entry_lists);
drivers/xen/xen-scsiback.c
656
list_for_each_entry(entry, head, l) {
drivers/xen/xen-scsiback.c
906
struct list_head *head = &(info->v2p_entry_lists);
drivers/xen/xen-scsiback.c
909
list_for_each_entry(entry, head, l)
fs/9p/vfs_dentry.c
51
struct hlist_head head;
fs/9p/vfs_dentry.c
57
hlist_move_list((struct hlist_head *)&dentry->d_fsdata, &head);
fs/9p/vfs_dentry.c
60
hlist_for_each_safe(p, n, &head)
fs/9p/vfs_dir.c
106
if (rdir->tail == rdir->head) {
fs/9p/vfs_dir.c
118
rdir->head = 0;
fs/9p/vfs_dir.c
121
while (rdir->head < rdir->tail) {
fs/9p/vfs_dir.c
122
err = p9stat_read(fid->clnt, rdir->buf + rdir->head,
fs/9p/vfs_dir.c
123
rdir->tail - rdir->head, &st);
fs/9p/vfs_dir.c
135
rdir->head += err;
fs/9p/vfs_dir.c
165
if (rdir->tail == rdir->head) {
fs/9p/vfs_dir.c
171
rdir->head = 0;
fs/9p/vfs_dir.c
175
while (rdir->head < rdir->tail) {
fs/9p/vfs_dir.c
177
err = p9dirent_read(fid->clnt, rdir->buf + rdir->head,
fs/9p/vfs_dir.c
178
rdir->tail - rdir->head,
fs/9p/vfs_dir.c
192
rdir->head += err;
fs/9p/vfs_dir.c
37
int head;
fs/adfs/dir_f.c
126
struct adfs_dirheader *head = dir->dirhead;
fs/adfs/dir_f.c
129
if (head->startmasseq != tail->endmasseq ||
fs/adfs/dir_f.c
131
(memcmp(&head->startname, "Nick", 4) &&
fs/adfs/dir_f.c
132
memcmp(&head->startname, "Hugo", 4)) ||
fs/adfs/dir_f.c
133
memcmp(&head->startname, &tail->endname, 4) ||
fs/aio.c
1018
unsigned head;
fs/aio.c
1030
head = ring->head;
fs/aio.c
1032
refill_reqs_available(ctx, head, ctx->tail);
fs/aio.c
1125
unsigned tail, pos, head, avail;
fs/aio.c
1160
head = ring->head;
fs/aio.c
1166
refill_reqs_available(ctx, head, tail);
fs/aio.c
1168
avail = tail > head
fs/aio.c
1169
? tail - head
fs/aio.c
1170
: tail + ctx->nr_events - head;
fs/aio.c
1196
list_for_each_entry_safe(curr, next, &ctx->wait.head, w.entry)
fs/aio.c
1221
unsigned head, tail, pos;
fs/aio.c
1236
head = ring->head;
fs/aio.c
1245
pr_debug("h%u t%u m%u\n", head, tail, ctx->nr_events);
fs/aio.c
1247
if (head == tail)
fs/aio.c
1250
head %= ctx->nr_events;
fs/aio.c
1258
avail = (head <= tail ? tail : ctx->nr_events) - head;
fs/aio.c
1259
if (head == tail)
fs/aio.c
1262
pos = head + AIO_EVENTS_OFFSET;
fs/aio.c
1279
head += avail;
fs/aio.c
1280
head %= ctx->nr_events;
fs/aio.c
1284
ring->head = head;
fs/aio.c
1287
pr_debug("%li h%u t%u\n", ret, head, tail);
fs/aio.c
1688
wait_queue_head_t *head;
fs/aio.c
1706
head = smp_load_acquire(&req->head);
fs/aio.c
1707
if (head) {
fs/aio.c
1708
spin_lock(&head->lock);
fs/aio.c
1711
spin_unlock(&head->lock);
fs/aio.c
1719
spin_unlock(&req->head->lock);
fs/aio.c
182
struct wait_queue_head *head;
fs/aio.c
1862
smp_store_release(&req->head, NULL);
fs/aio.c
1876
aio_poll_queue_proc(struct file *file, struct wait_queue_head *head,
fs/aio.c
1889
pt->iocb->poll.head = head;
fs/aio.c
1890
add_wait_queue(head, &pt->iocb->poll.wait);
fs/aio.c
1911
req->head = NULL;
fs/aio.c
569
ring->head = ring->tail = 0;
fs/aio.c
59
unsigned head; /* Written to by userland or under ring_lock
fs/aio.c
984
static void refill_reqs_available(struct kioctx *ctx, unsigned head,
fs/aio.c
990
head %= ctx->nr_events;
fs/aio.c
991
if (head <= tail)
fs/aio.c
992
events_in_ring = tail - head;
fs/aio.c
994
events_in_ring = ctx->nr_events - (head - tail);
fs/autofs/root.c
139
struct list_head *p, *head;
fs/autofs/root.c
141
head = &sbi->active_list;
fs/autofs/root.c
142
if (list_empty(head))
fs/autofs/root.c
145
list_for_each(p, head) {
fs/autofs/root.c
194
struct list_head *p, *head;
fs/autofs/root.c
196
head = &sbi->expiring_list;
fs/autofs/root.c
197
if (list_empty(head))
fs/autofs/root.c
200
list_for_each(p, head) {
fs/befs/btree.c
214
node->head.left = fs64_to_cpu(sb, node->od_node->left);
fs/befs/btree.c
215
node->head.right = fs64_to_cpu(sb, node->od_node->right);
fs/befs/btree.c
216
node->head.overflow = fs64_to_cpu(sb, node->od_node->overflow);
fs/befs/btree.c
217
node->head.all_key_count =
fs/befs/btree.c
219
node->head.all_key_length =
fs/befs/btree.c
282
node_off = this_node->head.overflow;
fs/befs/btree.c
344
last = node->head.all_key_count - 1;
fs/befs/btree.c
458
while (key_sum + this_node->head.all_key_count <= key_no) {
fs/befs/btree.c
461
if (this_node->head.right == BEFS_BT_INVAL) {
fs/befs/btree.c
467
key_sum + this_node->head.all_key_count);
fs/befs/btree.c
473
key_sum += this_node->head.all_key_count;
fs/befs/btree.c
474
node_off = this_node->head.right;
fs/befs/btree.c
556
if (this_node->head.all_key_count == 0 && befs_leafnode(this_node)) {
fs/befs/btree.c
563
if (this_node->head.all_key_count == 0) {
fs/befs/btree.c
567
this_node->head.overflow);
fs/befs/btree.c
568
*node_off = this_node->head.overflow;
fs/befs/btree.c
601
if (node->head.overflow == BEFS_BT_INVAL)
fs/befs/btree.c
625
(sizeof (befs_btree_nodehead) + node->head.all_key_length);
fs/befs/btree.c
645
size_t keylen_index_size = node->head.all_key_count * sizeof (fs16);
fs/befs/btree.c
681
if (index < 0 || index > node->head.all_key_count) {
fs/befs/btree.c
82
befs_host_btree_nodehead head; /* head of node converted to cpu byteorder */
fs/btrfs/backref.c
1384
struct btrfs_delayed_ref_head *head;
fs/btrfs/backref.c
1426
head = NULL;
fs/btrfs/backref.c
1450
head = btrfs_find_delayed_ref_head(ctx->fs_info, delayed_refs,
fs/btrfs/backref.c
1452
if (head) {
fs/btrfs/backref.c
1453
if (!mutex_trylock(&head->mutex)) {
fs/btrfs/backref.c
1454
refcount_inc(&head->refs);
fs/btrfs/backref.c
1463
mutex_lock(&head->mutex);
fs/btrfs/backref.c
1464
mutex_unlock(&head->mutex);
fs/btrfs/backref.c
1465
btrfs_put_delayed_ref_head(head);
fs/btrfs/backref.c
1469
ret = add_delayed_refs(ctx->fs_info, head, ctx->time_seq,
fs/btrfs/backref.c
1471
mutex_unlock(&head->mutex);
fs/btrfs/backref.c
887
struct btrfs_delayed_ref_head *head, u64 seq,
fs/btrfs/backref.c
896
spin_lock(&head->lock);
fs/btrfs/backref.c
897
for (n = rb_first_cached(&head->ref_tree); n; n = rb_next(n)) {
fs/btrfs/backref.c
924
if (head->extent_op && head->extent_op->update_key) {
fs/btrfs/backref.c
925
btrfs_disk_key_to_cpu(&key, &head->extent_op->key);
fs/btrfs/backref.c
996
spin_unlock(&head->lock);
fs/btrfs/block-group.c
1031
struct list_head *head = &fs_info->space_info;
fs/btrfs/block-group.c
1034
list_for_each_entry_rcu(sinfo, head, list) {
fs/btrfs/block-group.c
4017
struct list_head *head = &info->space_info;
fs/btrfs/block-group.c
4020
list_for_each_entry(found, head, list) {
fs/btrfs/delayed-ref.c
1239
bool btrfs_find_delayed_tree_ref(struct btrfs_delayed_ref_head *head,
fs/btrfs/delayed-ref.c
1245
lockdep_assert_held(&head->mutex);
fs/btrfs/delayed-ref.c
1247
spin_lock(&head->lock);
fs/btrfs/delayed-ref.c
1248
node = head->ref_tree.rb_root.rb_node;
fs/btrfs/delayed-ref.c
1269
spin_unlock(&head->lock);
fs/btrfs/delayed-ref.c
1280
struct btrfs_delayed_ref_head *head;
fs/btrfs/delayed-ref.c
1284
head = find_first_ref_head(delayed_refs);
fs/btrfs/delayed-ref.c
1285
if (!head)
fs/btrfs/delayed-ref.c
1288
if (!btrfs_delayed_ref_lock(delayed_refs, head))
fs/btrfs/delayed-ref.c
1291
spin_lock(&head->lock);
fs/btrfs/delayed-ref.c
1292
while ((n = rb_first_cached(&head->ref_tree)) != NULL) {
fs/btrfs/delayed-ref.c
1296
drop_delayed_ref(fs_info, delayed_refs, head, ref);
fs/btrfs/delayed-ref.c
1298
if (head->must_insert_reserved)
fs/btrfs/delayed-ref.c
1300
btrfs_free_delayed_extent_op(head->extent_op);
fs/btrfs/delayed-ref.c
1301
btrfs_delete_ref_head(fs_info, delayed_refs, head);
fs/btrfs/delayed-ref.c
1302
spin_unlock(&head->lock);
fs/btrfs/delayed-ref.c
1304
mutex_unlock(&head->mutex);
fs/btrfs/delayed-ref.c
1309
bg = btrfs_lookup_block_group(fs_info, head->bytenr);
fs/btrfs/delayed-ref.c
1319
head->bytenr);
fs/btrfs/delayed-ref.c
1323
bg->pinned += head->num_bytes;
fs/btrfs/delayed-ref.c
1325
head->num_bytes);
fs/btrfs/delayed-ref.c
1326
bg->reserved -= head->num_bytes;
fs/btrfs/delayed-ref.c
1327
bg->space_info->bytes_reserved -= head->num_bytes;
fs/btrfs/delayed-ref.c
1334
btrfs_error_unpin_extent_range(fs_info, head->bytenr,
fs/btrfs/delayed-ref.c
1335
head->bytenr + head->num_bytes - 1);
fs/btrfs/delayed-ref.c
1338
btrfs_cleanup_ref_head_accounting(fs_info, delayed_refs, head);
fs/btrfs/delayed-ref.c
1339
btrfs_put_delayed_ref_head(head);
fs/btrfs/delayed-ref.c
350
struct btrfs_delayed_ref_head *head)
fs/btrfs/delayed-ref.c
353
if (mutex_trylock(&head->mutex))
fs/btrfs/delayed-ref.c
356
refcount_inc(&head->refs);
fs/btrfs/delayed-ref.c
359
mutex_lock(&head->mutex);
fs/btrfs/delayed-ref.c
361
if (!head->tracked) {
fs/btrfs/delayed-ref.c
362
mutex_unlock(&head->mutex);
fs/btrfs/delayed-ref.c
363
btrfs_put_delayed_ref_head(head);
fs/btrfs/delayed-ref.c
366
btrfs_put_delayed_ref_head(head);
fs/btrfs/delayed-ref.c
372
struct btrfs_delayed_ref_head *head,
fs/btrfs/delayed-ref.c
375
lockdep_assert_held(&head->lock);
fs/btrfs/delayed-ref.c
376
rb_erase_cached(&ref->ref_node, &head->ref_tree);
fs/btrfs/delayed-ref.c
386
struct btrfs_delayed_ref_head *head,
fs/btrfs/delayed-ref.c
414
drop_delayed_ref(fs_info, delayed_refs, head, next);
fs/btrfs/delayed-ref.c
417
drop_delayed_ref(fs_info, delayed_refs, head, ref);
fs/btrfs/delayed-ref.c
433
struct btrfs_delayed_ref_head *head)
fs/btrfs/delayed-ref.c
439
lockdep_assert_held(&head->lock);
fs/btrfs/delayed-ref.c
441
if (RB_EMPTY_ROOT(&head->ref_tree.rb_root))
fs/btrfs/delayed-ref.c
445
if (head->is_data)
fs/btrfs/delayed-ref.c
450
for (node = rb_first_cached(&head->ref_tree); node;
fs/btrfs/delayed-ref.c
455
if (merge_ref(fs_info, delayed_refs, head, ref, seq))
fs/btrfs/delayed-ref.c
479
struct btrfs_delayed_ref_head *head;
fs/btrfs/delayed-ref.c
488
xa_for_each_start(&delayed_refs->head_refs, found_index, head, start_index) {
fs/btrfs/delayed-ref.c
489
if (!head->processing) {
fs/btrfs/delayed-ref.c
503
head->processing = true;
fs/btrfs/delayed-ref.c
506
delayed_refs->run_delayed_start = head->bytenr +
fs/btrfs/delayed-ref.c
507
head->num_bytes;
fs/btrfs/delayed-ref.c
509
locked = btrfs_delayed_ref_lock(delayed_refs, head);
fs/btrfs/delayed-ref.c
520
return head;
fs/btrfs/delayed-ref.c
524
struct btrfs_delayed_ref_head *head)
fs/btrfs/delayed-ref.c
527
head->processing = false;
fs/btrfs/delayed-ref.c
530
btrfs_delayed_ref_unlock(head);
fs/btrfs/delayed-ref.c
535
struct btrfs_delayed_ref_head *head)
fs/btrfs/delayed-ref.c
537
const unsigned long index = (head->bytenr >> fs_info->sectorsize_bits);
fs/btrfs/delayed-ref.c
540
lockdep_assert_held(&head->lock);
fs/btrfs/delayed-ref.c
543
head->tracked = false;
fs/btrfs/delayed-ref.c
545
if (!head->processing)
fs/btrfs/delayed-ref.c
549
struct btrfs_delayed_ref_node *btrfs_select_delayed_ref(struct btrfs_delayed_ref_head *head)
fs/btrfs/delayed-ref.c
553
lockdep_assert_held(&head->mutex);
fs/btrfs/delayed-ref.c
554
lockdep_assert_held(&head->lock);
fs/btrfs/delayed-ref.c
556
if (RB_EMPTY_ROOT(&head->ref_tree.rb_root))
fs/btrfs/delayed-ref.c
565
if (!list_empty(&head->ref_add_list))
fs/btrfs/delayed-ref.c
566
return list_first_entry(&head->ref_add_list,
fs/btrfs/delayed-ref.c
569
ref = rb_entry(rb_first_cached(&head->ref_tree),
fs/btrfs/delayed-ref.h
371
static inline void btrfs_put_delayed_ref_head(struct btrfs_delayed_ref_head *head)
fs/btrfs/delayed-ref.h
373
if (refcount_dec_and_test(&head->refs))
fs/btrfs/delayed-ref.h
374
kmem_cache_free(btrfs_delayed_ref_head_cachep, head);
fs/btrfs/delayed-ref.h
388
struct btrfs_delayed_ref_head *head);
fs/btrfs/delayed-ref.h
394
static inline void btrfs_delayed_ref_unlock(struct btrfs_delayed_ref_head *head)
fs/btrfs/delayed-ref.h
396
mutex_unlock(&head->mutex);
fs/btrfs/delayed-ref.h
400
struct btrfs_delayed_ref_head *head);
fs/btrfs/delayed-ref.h
406
struct btrfs_delayed_ref_head *head);
fs/btrfs/delayed-ref.h
407
struct btrfs_delayed_ref_node *btrfs_select_delayed_ref(struct btrfs_delayed_ref_head *head);
fs/btrfs/delayed-ref.h
420
bool btrfs_find_delayed_tree_ref(struct btrfs_delayed_ref_head *head,
fs/btrfs/disk-io.c
3972
struct list_head *head;
fs/btrfs/disk-io.c
3978
head = &info->fs_devices->devices;
fs/btrfs/disk-io.c
3979
list_for_each_entry(dev, head, dev_list) {
fs/btrfs/disk-io.c
3992
list_for_each_entry(dev, head, dev_list) {
fs/btrfs/disk-io.c
4048
struct list_head *head;
fs/btrfs/disk-io.c
4075
head = &fs_info->fs_devices->devices;
fs/btrfs/disk-io.c
4089
list_for_each_entry(dev, head, dev_list) {
fs/btrfs/disk-io.c
4140
list_for_each_entry(dev, head, dev_list) {
fs/btrfs/extent-tree.c
109
struct btrfs_delayed_ref_head *head;
fs/btrfs/extent-tree.c
1691
const struct btrfs_delayed_ref_head *head,
fs/btrfs/extent-tree.c
1714
key.objectid = head->bytenr;
fs/btrfs/extent-tree.c
1718
key.offset = head->level;
fs/btrfs/extent-tree.c
1721
key.offset = head->num_bytes;
fs/btrfs/extent-tree.c
1741
if (key.objectid == head->bytenr &&
fs/btrfs/extent-tree.c
1743
key.offset == head->num_bytes)
fs/btrfs/extent-tree.c
1750
key.objectid = head->bytenr;
fs/btrfs/extent-tree.c
1752
key.offset = head->num_bytes;
fs/btrfs/extent-tree.c
1759
head->bytenr, head->num_bytes, head->level);
fs/btrfs/extent-tree.c
1875
struct btrfs_delayed_ref_head *head)
fs/btrfs/extent-tree.c
1877
struct btrfs_delayed_extent_op *extent_op = head->extent_op;
fs/btrfs/extent-tree.c
1882
if (head->must_insert_reserved) {
fs/btrfs/extent-tree.c
1883
head->extent_op = NULL;
fs/btrfs/extent-tree.c
1891
struct btrfs_delayed_ref_head *head)
fs/btrfs/extent-tree.c
1896
extent_op = cleanup_extent_op(head);
fs/btrfs/extent-tree.c
1899
head->extent_op = NULL;
fs/btrfs/extent-tree.c
1900
spin_unlock(&head->lock);
fs/btrfs/extent-tree.c
1901
ret = run_delayed_extent_op(trans, head, extent_op);
fs/btrfs/extent-tree.c
1908
struct btrfs_delayed_ref_head *head)
fs/btrfs/extent-tree.c
1916
if (head->total_ref_mod < 0 && head->is_data) {
fs/btrfs/extent-tree.c
1920
delayed_refs->pending_csums -= head->num_bytes;
fs/btrfs/extent-tree.c
1922
nr_csums = btrfs_csum_bytes_to_leaves(fs_info, head->num_bytes);
fs/btrfs/extent-tree.c
1929
if (head->must_insert_reserved)
fs/btrfs/extent-tree.c
1930
free_head_ref_squota_rsv(fs_info, head);
fs/btrfs/extent-tree.c
1936
struct btrfs_delayed_ref_head *head,
fs/btrfs/extent-tree.c
1946
ret = run_and_cleanup_extent_op(trans, head);
fs/btrfs/extent-tree.c
1948
btrfs_unselect_ref_head(delayed_refs, head);
fs/btrfs/extent-tree.c
1959
spin_unlock(&head->lock);
fs/btrfs/extent-tree.c
196
head = btrfs_find_delayed_ref_head(fs_info, delayed_refs, bytenr);
fs/btrfs/extent-tree.c
1961
spin_lock(&head->lock);
fs/btrfs/extent-tree.c
1962
if (!RB_EMPTY_ROOT(&head->ref_tree.rb_root) || head->extent_op) {
fs/btrfs/extent-tree.c
1963
spin_unlock(&head->lock);
fs/btrfs/extent-tree.c
1967
btrfs_delete_ref_head(fs_info, delayed_refs, head);
fs/btrfs/extent-tree.c
1968
spin_unlock(&head->lock);
fs/btrfs/extent-tree.c
197
if (head) {
fs/btrfs/extent-tree.c
1971
if (head->must_insert_reserved) {
fs/btrfs/extent-tree.c
1972
btrfs_pin_extent(trans, head->bytenr, head->num_bytes);
fs/btrfs/extent-tree.c
1973
if (head->is_data) {
fs/btrfs/extent-tree.c
1976
csum_root = btrfs_csum_root(fs_info, head->bytenr);
fs/btrfs/extent-tree.c
198
if (!mutex_trylock(&head->mutex)) {
fs/btrfs/extent-tree.c
1980
head->bytenr);
fs/btrfs/extent-tree.c
1983
ret = btrfs_del_csums(trans, csum_root, head->bytenr,
fs/btrfs/extent-tree.c
1984
head->num_bytes);
fs/btrfs/extent-tree.c
1989
*bytes_released += btrfs_cleanup_ref_head_accounting(fs_info, delayed_refs, head);
fs/btrfs/extent-tree.c
199
refcount_inc(&head->refs);
fs/btrfs/extent-tree.c
1991
trace_run_delayed_ref_head(fs_info, head, 0);
fs/btrfs/extent-tree.c
1992
btrfs_delayed_ref_unlock(head);
fs/btrfs/extent-tree.c
1993
btrfs_put_delayed_ref_head(head);
fs/btrfs/extent-tree.c
208
mutex_lock(&head->mutex);
fs/btrfs/extent-tree.c
209
mutex_unlock(&head->mutex);
fs/btrfs/extent-tree.c
210
btrfs_put_delayed_ref_head(head);
fs/btrfs/extent-tree.c
213
spin_lock(&head->lock);
fs/btrfs/extent-tree.c
214
if (head->extent_op && head->extent_op->update_flags)
fs/btrfs/extent-tree.c
215
extent_flags |= head->extent_op->flags_to_set;
fs/btrfs/extent-tree.c
217
num_refs += head->ref_mod;
fs/btrfs/extent-tree.c
218
spin_unlock(&head->lock);
fs/btrfs/extent-tree.c
219
mutex_unlock(&head->mutex);
fs/btrfs/extent-tree.c
2294
struct btrfs_delayed_ref_head *head;
fs/btrfs/extent-tree.c
2311
head = btrfs_find_delayed_ref_head(root->fs_info, delayed_refs, bytenr);
fs/btrfs/extent-tree.c
2312
if (!head) {
fs/btrfs/extent-tree.c
2318
if (!mutex_trylock(&head->mutex)) {
fs/btrfs/extent-tree.c
2325
refcount_inc(&head->refs);
fs/btrfs/extent-tree.c
2334
mutex_lock(&head->mutex);
fs/btrfs/extent-tree.c
2335
mutex_unlock(&head->mutex);
fs/btrfs/extent-tree.c
2336
btrfs_put_delayed_ref_head(head);
fs/btrfs/extent-tree.c
2342
spin_lock(&head->lock);
fs/btrfs/extent-tree.c
2347
for (node = rb_first_cached(&head->ref_tree); node;
fs/btrfs/extent-tree.c
2372
spin_unlock(&head->lock);
fs/btrfs/extent-tree.c
2373
mutex_unlock(&head->mutex);
fs/btrfs/extent-tree.c
3571
struct btrfs_delayed_ref_head *head;
fs/btrfs/extent-tree.c
3577
head = btrfs_find_delayed_ref_head(fs_info, delayed_refs, bytenr);
fs/btrfs/extent-tree.c
3578
if (!head)
fs/btrfs/extent-tree.c
3581
spin_lock(&head->lock);
fs/btrfs/extent-tree.c
3582
if (!RB_EMPTY_ROOT(&head->ref_tree.rb_root))
fs/btrfs/extent-tree.c
3585
if (cleanup_extent_op(head) != NULL)
fs/btrfs/extent-tree.c
3592
if (!mutex_trylock(&head->mutex))
fs/btrfs/extent-tree.c
3595
btrfs_delete_ref_head(fs_info, delayed_refs, head);
fs/btrfs/extent-tree.c
3596
head->processing = false;
fs/btrfs/extent-tree.c
3598
spin_unlock(&head->lock);
fs/btrfs/extent-tree.c
3601
BUG_ON(head->extent_op);
fs/btrfs/extent-tree.c
3602
if (head->must_insert_reserved)
fs/btrfs/extent-tree.c
3605
btrfs_cleanup_ref_head_accounting(fs_info, delayed_refs, head);
fs/btrfs/extent-tree.c
3606
mutex_unlock(&head->mutex);
fs/btrfs/extent-tree.c
3607
btrfs_put_delayed_ref_head(head);
fs/btrfs/extent-tree.c
3610
spin_unlock(&head->lock);
fs/btrfs/extent-tree.c
5712
struct btrfs_delayed_ref_head *head;
fs/btrfs/extent-tree.c
5740
head = btrfs_find_delayed_ref_head(root->fs_info, delayed_refs, bytenr);
fs/btrfs/extent-tree.c
5741
if (!head)
fs/btrfs/extent-tree.c
5743
if (!mutex_trylock(&head->mutex)) {
fs/btrfs/extent-tree.c
5749
refcount_inc(&head->refs);
fs/btrfs/extent-tree.c
5754
mutex_lock(&head->mutex);
fs/btrfs/extent-tree.c
5755
mutex_unlock(&head->mutex);
fs/btrfs/extent-tree.c
5756
btrfs_put_delayed_ref_head(head);
fs/btrfs/extent-tree.c
5760
exists = btrfs_find_delayed_tree_ref(head, btrfs_root_id(root), parent);
fs/btrfs/extent-tree.c
5761
mutex_unlock(&head->mutex);
fs/btrfs/extent-tree.h
107
struct btrfs_delayed_ref_head *head);
fs/btrfs/extent_io.c
3611
static inline void btrfs_release_extent_buffer_rcu(struct rcu_head *head)
fs/btrfs/extent_io.c
3614
container_of(head, struct extent_buffer, rcu_head);
fs/btrfs/file.c
2829
static int add_falloc_range(struct list_head *head, u64 start, u64 len)
fs/btrfs/file.c
2833
if (!list_empty(head)) {
fs/btrfs/file.c
2838
range = list_last_entry(head, struct falloc_range, list);
fs/btrfs/file.c
2850
list_add_tail(&range->list, head);
fs/btrfs/free-space-cache.c
3034
struct list_head *head;
fs/btrfs/free-space-cache.c
3037
while ((head = block_group->cluster_list.next) !=
fs/btrfs/free-space-cache.c
3039
cluster = list_entry(head, struct btrfs_free_cluster,
fs/btrfs/lru_cache.c
111
struct list_head *head;
fs/btrfs/lru_cache.c
114
head = kmalloc_obj(*head, gfp);
fs/btrfs/lru_cache.c
115
if (!head)
fs/btrfs/lru_cache.c
118
ret = mtree_insert(&cache->entries, key, head, gfp);
fs/btrfs/lru_cache.c
120
INIT_LIST_HEAD(head);
fs/btrfs/lru_cache.c
121
list_add_tail(&new_entry->list, head);
fs/btrfs/lru_cache.c
123
kfree(head);
fs/btrfs/lru_cache.c
124
head = mtree_load(&cache->entries, key);
fs/btrfs/lru_cache.c
125
ASSERT(head != NULL);
fs/btrfs/lru_cache.c
126
if (match_entry(head, key, new_entry->gen) != NULL)
fs/btrfs/lru_cache.c
128
list_add_tail(&new_entry->list, head);
fs/btrfs/lru_cache.c
130
kfree(head);
fs/btrfs/lru_cache.c
23
static struct btrfs_lru_cache_entry *match_entry(struct list_head *head, u64 key,
fs/btrfs/lru_cache.c
28
list_for_each_entry(entry, head, list) {
fs/btrfs/lru_cache.c
48
struct list_head *head;
fs/btrfs/lru_cache.c
51
head = mtree_load(&cache->entries, key);
fs/btrfs/lru_cache.c
52
if (!head)
fs/btrfs/lru_cache.c
55
entry = match_entry(head, key, gen);
fs/btrfs/lru_cache.c
82
struct list_head *head;
fs/btrfs/lru_cache.c
89
head = mtree_erase(&cache->entries, entry->key);
fs/btrfs/lru_cache.c
90
ASSERT(head == prev);
fs/btrfs/lru_cache.c
91
kfree(head);
fs/btrfs/qgroup.c
1435
static void qgroup_iterator_add(struct list_head *head, struct btrfs_qgroup *qgroup)
fs/btrfs/qgroup.c
1440
list_add_tail(&qgroup->iterator, head);
fs/btrfs/qgroup.c
1443
static void qgroup_iterator_clean(struct list_head *head)
fs/btrfs/qgroup.c
1445
while (!list_empty(head)) {
fs/btrfs/qgroup.c
1448
qgroup = list_first_entry(head, struct btrfs_qgroup, iterator);
fs/btrfs/qgroup.c
2725
static void qgroup_iterator_nested_add(struct list_head *head, struct btrfs_qgroup *qgroup)
fs/btrfs/qgroup.c
2730
list_add_tail(&qgroup->nested_iterator, head);
fs/btrfs/qgroup.c
2733
static void qgroup_iterator_nested_clean(struct list_head *head)
fs/btrfs/qgroup.c
2735
while (!list_empty(head)) {
fs/btrfs/qgroup.c
2738
qgroup = list_first_entry(head, struct btrfs_qgroup, nested_iterator);
fs/btrfs/raid56.c
1790
u64 a_sector = ra->bio_list.head->bi_iter.bi_sector;
fs/btrfs/raid56.c
1791
u64 b_sector = rb->bio_list.head->bi_iter.bi_sector;
fs/btrfs/raid56.c
913
struct list_head *head = rbio->plug_list.next;
fs/btrfs/raid56.c
915
next = list_entry(head, struct btrfs_raid_bio,
fs/btrfs/send.c
3017
static void __free_recorded_refs(struct list_head *head)
fs/btrfs/send.c
3021
while (!list_empty(head)) {
fs/btrfs/send.c
3022
cur = list_first_entry(head, struct recorded_ref, list);
fs/btrfs/space-info.c
191
struct list_head *head = &info->space_info;
fs/btrfs/space-info.c
194
list_for_each_entry(found, head, list)
fs/btrfs/space-info.c
409
struct list_head *head = &info->space_info;
fs/btrfs/space-info.c
414
list_for_each_entry(found, head, list) {
fs/btrfs/space-info.c
580
struct list_head *head;
fs/btrfs/space-info.c
586
head = &space_info->priority_tickets;
fs/btrfs/space-info.c
588
while (!list_empty(head)) {
fs/btrfs/space-info.c
592
ticket = list_first_entry(head, struct reserve_ticket, list);
fs/btrfs/space-info.c
607
if (head == &space_info->priority_tickets) {
fs/btrfs/space-info.c
608
head = &space_info->tickets;
fs/btrfs/tests/delayed-refs-tests.c
173
struct btrfs_delayed_ref_head *head;
fs/btrfs/tests/delayed-refs-tests.c
201
head = btrfs_select_ref_head(fs_info, delayed_refs);
fs/btrfs/tests/delayed-refs-tests.c
202
if (IS_ERR_OR_NULL(head)) {
fs/btrfs/tests/delayed-refs-tests.c
203
if (IS_ERR(head))
fs/btrfs/tests/delayed-refs-tests.c
205
PTR_ERR(head));
fs/btrfs/tests/delayed-refs-tests.c
212
if (validate_ref_head(head, head_check))
fs/btrfs/tests/delayed-refs-tests.c
215
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
216
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
217
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
227
btrfs_unselect_ref_head(delayed_refs, head);
fs/btrfs/tests/delayed-refs-tests.c
341
struct btrfs_delayed_ref_head *head = NULL;
fs/btrfs/tests/delayed-refs-tests.c
402
head = btrfs_select_ref_head(fs_info, &trans->transaction->delayed_refs);
fs/btrfs/tests/delayed-refs-tests.c
403
if (IS_ERR_OR_NULL(head)) {
fs/btrfs/tests/delayed-refs-tests.c
404
if (IS_ERR(head))
fs/btrfs/tests/delayed-refs-tests.c
406
PTR_ERR(head));
fs/btrfs/tests/delayed-refs-tests.c
413
if (validate_ref_head(head, &head_check)) {
fs/btrfs/tests/delayed-refs-tests.c
418
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
419
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
420
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
426
delete_delayed_ref_head(trans, head);
fs/btrfs/tests/delayed-refs-tests.c
427
head = NULL;
fs/btrfs/tests/delayed-refs-tests.c
44
struct btrfs_delayed_ref_head *head)
fs/btrfs/tests/delayed-refs-tests.c
452
head = btrfs_select_ref_head(fs_info, &trans->transaction->delayed_refs);
fs/btrfs/tests/delayed-refs-tests.c
453
if (IS_ERR_OR_NULL(head)) {
fs/btrfs/tests/delayed-refs-tests.c
454
if (IS_ERR(head))
fs/btrfs/tests/delayed-refs-tests.c
456
PTR_ERR(head));
fs/btrfs/tests/delayed-refs-tests.c
465
if (validate_ref_head(head, &head_check)) {
fs/btrfs/tests/delayed-refs-tests.c
470
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
471
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
472
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
483
delete_delayed_ref_node(head, node);
fs/btrfs/tests/delayed-refs-tests.c
485
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
486
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
487
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
492
delete_delayed_ref_head(trans, head);
fs/btrfs/tests/delayed-refs-tests.c
493
head = NULL;
fs/btrfs/tests/delayed-refs-tests.c
51
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
515
head = btrfs_select_ref_head(fs_info, &trans->transaction->delayed_refs);
fs/btrfs/tests/delayed-refs-tests.c
516
if (IS_ERR_OR_NULL(head)) {
fs/btrfs/tests/delayed-refs-tests.c
517
if (IS_ERR(head))
fs/btrfs/tests/delayed-refs-tests.c
519
PTR_ERR(head));
fs/btrfs/tests/delayed-refs-tests.c
52
btrfs_delete_ref_head(fs_info, delayed_refs, head);
fs/btrfs/tests/delayed-refs-tests.c
528
if (validate_ref_head(head, &head_check)) {
fs/btrfs/tests/delayed-refs-tests.c
53
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
534
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
535
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
536
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
547
delete_delayed_ref_node(head, node);
fs/btrfs/tests/delayed-refs-tests.c
549
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
550
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
551
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
556
delete_delayed_ref_head(trans, head);
fs/btrfs/tests/delayed-refs-tests.c
557
head = NULL;
fs/btrfs/tests/delayed-refs-tests.c
56
btrfs_delayed_ref_unlock(head);
fs/btrfs/tests/delayed-refs-tests.c
57
btrfs_put_delayed_ref_head(head);
fs/btrfs/tests/delayed-refs-tests.c
584
head = btrfs_select_ref_head(fs_info, &trans->transaction->delayed_refs);
fs/btrfs/tests/delayed-refs-tests.c
585
if (IS_ERR_OR_NULL(head)) {
fs/btrfs/tests/delayed-refs-tests.c
586
if (IS_ERR(head))
fs/btrfs/tests/delayed-refs-tests.c
588
PTR_ERR(head));
fs/btrfs/tests/delayed-refs-tests.c
598
if (validate_ref_head(head, &head_check)) {
fs/btrfs/tests/delayed-refs-tests.c
60
static void delete_delayed_ref_node(struct btrfs_delayed_ref_head *head,
fs/btrfs/tests/delayed-refs-tests.c
603
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
604
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
605
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
616
delete_delayed_ref_node(head, node);
fs/btrfs/tests/delayed-refs-tests.c
618
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
619
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
620
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
626
delete_delayed_ref_head(trans, head);
fs/btrfs/tests/delayed-refs-tests.c
627
head = NULL;
fs/btrfs/tests/delayed-refs-tests.c
63
rb_erase_cached(&node->ref_node, &head->ref_tree);
fs/btrfs/tests/delayed-refs-tests.c
654
head = btrfs_select_ref_head(fs_info, &trans->transaction->delayed_refs);
fs/btrfs/tests/delayed-refs-tests.c
655
if (IS_ERR_OR_NULL(head)) {
fs/btrfs/tests/delayed-refs-tests.c
656
if (IS_ERR(head))
fs/btrfs/tests/delayed-refs-tests.c
658
PTR_ERR(head));
fs/btrfs/tests/delayed-refs-tests.c
668
if (validate_ref_head(head, &head_check)) {
fs/btrfs/tests/delayed-refs-tests.c
674
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
675
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
676
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
687
delete_delayed_ref_node(head, node);
fs/btrfs/tests/delayed-refs-tests.c
689
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
690
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
691
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
696
delete_delayed_ref_head(trans, head);
fs/btrfs/tests/delayed-refs-tests.c
697
head = NULL;
fs/btrfs/tests/delayed-refs-tests.c
70
static int validate_ref_head(struct btrfs_delayed_ref_head *head,
fs/btrfs/tests/delayed-refs-tests.c
73
if (head->bytenr != check->bytenr) {
fs/btrfs/tests/delayed-refs-tests.c
739
head = btrfs_select_ref_head(fs_info, &trans->transaction->delayed_refs);
fs/btrfs/tests/delayed-refs-tests.c
74
test_err("invalid bytenr have: %llu want: %llu", head->bytenr,
fs/btrfs/tests/delayed-refs-tests.c
740
if (IS_ERR_OR_NULL(head)) {
fs/btrfs/tests/delayed-refs-tests.c
741
if (IS_ERR(head))
fs/btrfs/tests/delayed-refs-tests.c
743
PTR_ERR(head));
fs/btrfs/tests/delayed-refs-tests.c
753
if (validate_ref_head(head, &head_check)) {
fs/btrfs/tests/delayed-refs-tests.c
758
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
759
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
760
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
767
if (!IS_ERR_OR_NULL(head))
fs/btrfs/tests/delayed-refs-tests.c
768
btrfs_unselect_ref_head(&trans->transaction->delayed_refs, head);
fs/btrfs/tests/delayed-refs-tests.c
782
struct btrfs_delayed_ref_head *head = NULL;
fs/btrfs/tests/delayed-refs-tests.c
79
if (head->num_bytes != check->num_bytes) {
fs/btrfs/tests/delayed-refs-tests.c
81
head->num_bytes, check->num_bytes);
fs/btrfs/tests/delayed-refs-tests.c
830
head = btrfs_select_ref_head(fs_info, delayed_refs);
fs/btrfs/tests/delayed-refs-tests.c
831
if (IS_ERR_OR_NULL(head)) {
fs/btrfs/tests/delayed-refs-tests.c
832
if (IS_ERR(head))
fs/btrfs/tests/delayed-refs-tests.c
834
PTR_ERR(head));
fs/btrfs/tests/delayed-refs-tests.c
838
head = NULL;
fs/btrfs/tests/delayed-refs-tests.c
843
if (validate_ref_head(head, &head_check)) {
fs/btrfs/tests/delayed-refs-tests.c
848
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
849
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
85
if (head->ref_mod != check->ref_mod) {
fs/btrfs/tests/delayed-refs-tests.c
850
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
86
test_err("invalid ref_mod have: %d want: %d", head->ref_mod,
fs/btrfs/tests/delayed-refs-tests.c
861
delete_delayed_ref_node(head, node);
fs/btrfs/tests/delayed-refs-tests.c
863
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
864
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
865
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
877
delete_delayed_ref_node(head, node);
fs/btrfs/tests/delayed-refs-tests.c
878
delete_delayed_ref_head(trans, head);
fs/btrfs/tests/delayed-refs-tests.c
879
head = NULL;
fs/btrfs/tests/delayed-refs-tests.c
91
if (head->total_ref_mod != check->total_ref_mod) {
fs/btrfs/tests/delayed-refs-tests.c
917
head = btrfs_select_ref_head(fs_info, delayed_refs);
fs/btrfs/tests/delayed-refs-tests.c
918
if (IS_ERR_OR_NULL(head)) {
fs/btrfs/tests/delayed-refs-tests.c
919
if (IS_ERR(head))
fs/btrfs/tests/delayed-refs-tests.c
921
PTR_ERR(head));
fs/btrfs/tests/delayed-refs-tests.c
925
head = NULL;
fs/btrfs/tests/delayed-refs-tests.c
93
head->total_ref_mod, check->total_ref_mod);
fs/btrfs/tests/delayed-refs-tests.c
930
if (validate_ref_head(head, &head_check)) {
fs/btrfs/tests/delayed-refs-tests.c
935
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
936
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
937
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
949
delete_delayed_ref_node(head, node);
fs/btrfs/tests/delayed-refs-tests.c
951
spin_lock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
952
node = btrfs_select_delayed_ref(head);
fs/btrfs/tests/delayed-refs-tests.c
953
spin_unlock(&head->lock);
fs/btrfs/tests/delayed-refs-tests.c
965
delete_delayed_ref_node(head, node);
fs/btrfs/tests/delayed-refs-tests.c
968
if (head)
fs/btrfs/tests/delayed-refs-tests.c
969
btrfs_unselect_ref_head(delayed_refs, head);
fs/btrfs/tests/delayed-refs-tests.c
97
if (head->must_insert_reserved != check->must_insert) {
fs/btrfs/tests/delayed-refs-tests.c
99
head->must_insert_reserved, check->must_insert);
fs/btrfs/transaction.c
1953
struct list_head *head = &trans->transaction->pending_snapshots;
fs/btrfs/transaction.c
1956
list_for_each_entry_safe(pending, next, head, list) {
fs/buffer.c
1004
struct buffer_head *head = folio_buffers(folio);
fs/buffer.c
1005
struct buffer_head *bh = head;
fs/buffer.c
1023
} while (bh != head);
fs/buffer.c
105
bh = head;
fs/buffer.c
114
} while (bh != head);
fs/buffer.c
1630
struct buffer_head *head, *bh, *next;
fs/buffer.c
1641
head = folio_buffers(folio);
fs/buffer.c
1642
if (!head)
fs/buffer.c
1645
bh = head;
fs/buffer.c
1663
} while (bh != head);
fs/buffer.c
1685
struct buffer_head *bh, *head, *tail;
fs/buffer.c
1688
head = folio_alloc_buffers(folio, blocksize, gfp);
fs/buffer.c
1689
bh = head;
fs/buffer.c
1695
tail->b_this_page = head;
fs/buffer.c
1699
bh = head;
fs/buffer.c
1706
} while (bh != head);
fs/buffer.c
1708
folio_attach_private(folio, head);
fs/buffer.c
1711
return head;
fs/buffer.c
1744
struct buffer_head *head;
fs/buffer.c
1762
head = folio_buffers(folio);
fs/buffer.c
1763
if (!head)
fs/buffer.c
1765
bh = head;
fs/buffer.c
1776
} while (bh != head);
fs/buffer.c
1839
struct buffer_head *bh, *head;
fs/buffer.c
1844
head = folio_create_buffers(folio, inode,
fs/buffer.c
1857
bh = head;
fs/buffer.c
188
struct buffer_head *head;
fs/buffer.c
1894
} while (bh != head);
fs/buffer.c
1918
} while ((bh = bh->b_this_page) != head);
fs/buffer.c
1935
} while (bh != head);
fs/buffer.c
1962
bh = head;
fs/buffer.c
1977
} while ((bh = bh->b_this_page) != head);
fs/buffer.c
1990
} while (bh != head);
fs/buffer.c
2004
struct buffer_head *head, *bh;
fs/buffer.c
2007
head = folio_buffers(folio);
fs/buffer.c
2008
if (!head)
fs/buffer.c
2011
bh = head;
fs/buffer.c
2035
} while (bh != head);
fs/buffer.c
208
head = folio_buffers(folio);
fs/buffer.c
209
if (!head)
fs/buffer.c
2117
struct buffer_head *bh, *head, *wait[2], **wait_bh=wait;
fs/buffer.c
2123
head = folio_create_buffers(folio, inode, 0);
fs/buffer.c
2124
blocksize = head->b_size;
fs/buffer.c
2127
for (bh = head, block_start = 0; bh != head || !block_start;
fs/buffer.c
215
if (test_bit_acquire(BH_Migrate, &head->b_state)) {
fs/buffer.c
220
bh = head;
fs/buffer.c
2200
struct buffer_head *bh, *head;
fs/buffer.c
2202
bh = head = folio_buffers(folio);
fs/buffer.c
2222
} while (bh != head);
fs/buffer.c
230
} while (bh != head);
fs/buffer.c
2347
struct buffer_head *bh, *head;
fs/buffer.c
2350
head = folio_buffers(folio);
fs/buffer.c
2351
if (!head)
fs/buffer.c
2353
blocksize = head->b_size;
fs/buffer.c
2359
bh = head;
fs/buffer.c
2373
} while (bh != head);
fs/buffer.c
2390
struct buffer_head *bh, *head, *prev = NULL;
fs/buffer.c
2400
head = folio_create_buffers(folio, inode, 0);
fs/buffer.c
2401
blocksize = head->b_size;
fs/buffer.c
2405
bh = head;
fs/buffer.c
2446
} while (iblock++, (bh = bh->b_this_page) != head);
fs/buffer.c
2890
struct buffer_head *head = folio_buffers(folio);
fs/buffer.c
2893
bh = head;
fs/buffer.c
2898
} while (bh != head);
fs/buffer.c
2906
} while (bh != head);
fs/buffer.c
2907
*buffers_to_free = head;
fs/buffer.c
730
struct buffer_head *head;
fs/buffer.c
734
head = folio_buffers(folio);
fs/buffer.c
735
if (head) {
fs/buffer.c
736
struct buffer_head *bh = head;
fs/buffer.c
741
} while (bh != head);
fs/buffer.c
92
struct buffer_head *head, *bh;
fs/buffer.c
921
struct buffer_head *bh, *head;
fs/buffer.c
929
head = NULL;
fs/buffer.c
936
bh->b_this_page = head;
fs/buffer.c
938
head = bh;
fs/buffer.c
947
return head;
fs/buffer.c
952
if (head) {
fs/buffer.c
954
bh = head;
fs/buffer.c
955
head = head->b_this_page;
fs/buffer.c
957
} while (head);
fs/buffer.c
973
struct buffer_head *head)
fs/buffer.c
977
bh = head;
fs/buffer.c
98
head = folio_buffers(folio);
fs/buffer.c
982
tail->b_this_page = head;
fs/buffer.c
983
folio_attach_private(folio, head);
fs/buffer.c
99
if (!head)
fs/ceph/dir.c
1085
if (!err && !req->r_reply_info.head->is_dentry)
fs/ceph/dir.c
1167
!req->r_reply_info.head->is_target &&
fs/ceph/dir.c
1168
!req->r_reply_info.head->is_dentry)
fs/ceph/dir.c
1236
} else if (!req->r_reply_info.head->is_dentry) {
fs/ceph/dir.c
1252
le32_to_cpu(req->r_reply_info.head->result);
fs/ceph/dir.c
1460
if (!err && !req->r_reply_info.head->is_dentry)
fs/ceph/dir.c
1529
if (!err && !req->r_reply_info.head->is_dentry) {
fs/ceph/dir.c
764
if (!req->r_reply_info.head->is_dentry) {
fs/ceph/dir.c
965
if (!err && !req->r_reply_info.head->is_dentry)
fs/ceph/file.c
604
le32_to_cpu(req->r_reply_info.head->result);
fs/ceph/file.c
847
!list_is_singular(&as_ctx.pagelist->head))
fs/ceph/file.c
946
if (!err && (flags & O_CREAT) && !req->r_reply_info.head->is_dentry)
fs/ceph/inode.c
1566
rinfo->head->is_dentry, rinfo->head->is_target);
fs/ceph/inode.c
1568
if (!rinfo->head->is_target && !rinfo->head->is_dentry) {
fs/ceph/inode.c
1570
if (rinfo->head->result == 0 && req->r_parent)
fs/ceph/inode.c
1575
if (rinfo->head->is_dentry) {
fs/ceph/inode.c
1608
BUG_ON(!rinfo->head->is_target);
fs/ceph/inode.c
1669
if (rinfo->head->is_target) {
fs/ceph/inode.c
1678
rinfo->head->result == 0) ? req->r_fmode : -1,
fs/ceph/inode.c
1698
if (rinfo->head->is_dentry &&
fs/ceph/inode.c
1701
(rinfo->head->is_target || strncmp(req->r_dentry->d_name.name,
fs/ceph/inode.c
1768
if (!rinfo->head->is_target) {
fs/ceph/inode.c
1836
} else if (rinfo->head->is_dentry && req->r_dentry) {
fs/ceph/inode.c
1841
if (rinfo->head->is_target) {
fs/ceph/inode.c
2000
if (le32_to_cpu(rinfo->head->op) == CEPH_MDS_OP_LSSNAP) {
fs/ceph/mds_client.c
2314
struct ceph_mds_cap_release *head;
fs/ceph/mds_client.c
2339
head = msg->front.iov_base;
fs/ceph/mds_client.c
2340
head->num = cpu_to_le32(0);
fs/ceph/mds_client.c
2341
msg->front.iov_len = sizeof(*head);
fs/ceph/mds_client.c
2352
head = msg->front.iov_base;
fs/ceph/mds_client.c
2353
put_unaligned_le32(get_unaligned_le32(&head->num) + 1,
fs/ceph/mds_client.c
2354
&head->num);
fs/ceph/mds_client.c
2364
if (le32_to_cpu(head->num) == CEPH_CAPS_PER_RELEASE) {
fs/ceph/mds_client.c
2984
struct ceph_mds_request_head *head;
fs/ceph/mds_client.c
2988
head = (struct ceph_mds_request_head *)p;
fs/ceph/mds_client.c
2989
return (struct ceph_mds_request_head_legacy *)&head->oldest_client_tid;
fs/ceph/mds_client.c
3657
struct list_head *head)
fs/ceph/mds_client.c
3663
list_splice_init(head, &tmp_list);
fs/ceph/mds_client.c
371
if (info->head->is_dentry) {
fs/ceph/mds_client.c
3779
err = le32_to_cpu(req->r_reply_info.head->result);
fs/ceph/mds_client.c
3859
struct ceph_mds_reply_head *head = msg->front.iov_base;
fs/ceph/mds_client.c
3867
if (msg->front.iov_len < sizeof(*head)) {
fs/ceph/mds_client.c
3894
if ((test_bit(CEPH_MDS_R_GOT_UNSAFE, &req->r_req_flags) && !head->safe) ||
fs/ceph/mds_client.c
3895
(test_bit(CEPH_MDS_R_GOT_SAFE, &req->r_req_flags) && head->safe)) {
fs/ceph/mds_client.c
3897
head->safe ? "safe" : "unsafe", tid, mds);
fs/ceph/mds_client.c
3908
result = le32_to_cpu(head->result);
fs/ceph/mds_client.c
391
if (info->head->is_target) {
fs/ceph/mds_client.c
3910
if (head->safe) {
fs/ceph/mds_client.c
3946
if ((err >= 0) && rinfo->head->is_target) {
fs/ceph/mds_client.c
3988
le32_to_cpu(head->op) == CEPH_MDS_OP_RMSNAP,
fs/ceph/mds_client.c
4567
page = list_first_entry(&recon_state->pagelist->head, struct page, lru);
fs/ceph/mds_client.c
5018
list_first_entry(&recon_state.pagelist->head,
fs/ceph/mds_client.c
65
struct list_head *head);
fs/ceph/mds_client.c
731
u32 op = le32_to_cpu(info->head->op);
fs/ceph/mds_client.c
757
info->head = msg->front.iov_base;
fs/ceph/mds_client.h
145
struct ceph_mds_reply_head *head;
fs/ceph/metric.c
175
put_unaligned_le32(items, &head->num);
fs/ceph/metric.c
20
struct ceph_metric_head *head;
fs/ceph/metric.c
48
len = sizeof(*head) + sizeof(*cap) + sizeof(*read) + sizeof(*write)
fs/ceph/metric.c
60
head = msg->front.iov_base;
fs/ceph/metric.c
63
cap = (struct ceph_metric_cap *)(head + 1);
fs/ceph/xattr.c
1416
if (list_is_singular(&pagelist->head)) {
fs/ceph/xattr.c
1419
struct page *page = list_first_entry(&pagelist->head,
fs/crypto/keyring.c
55
static void fscrypt_free_master_key(struct rcu_head *head)
fs/crypto/keyring.c
58
container_of(head, struct fscrypt_master_key, mk_rcu_head);
fs/dcache.c
2866
kfree_rcu(old_name, head);
fs/dcache.c
333
struct rcu_head head;
fs/dcache.c
342
static void __d_free(struct rcu_head *head)
fs/dcache.c
344
struct dentry *dentry = container_of(head, struct dentry, d_u.d_rcu);
fs/dcache.c
349
static void __d_free_external(struct rcu_head *head)
fs/dcache.c
351
struct dentry *dentry = container_of(head, struct dentry, d_u.d_rcu);
fs/dcache.c
396
kfree_rcu(p, head);
fs/direct-io.c
110
unsigned head; /* next page to process */
fs/direct-io.c
160
return sdio->tail - sdio->head;
fs/direct-io.c
184
sdio->head = 0;
fs/direct-io.c
193
sdio->head = 0;
fs/direct-io.c
218
return dio->pages[sdio->head];
fs/direct-io.c
455
unpin_user_pages(dio->pages + sdio->head,
fs/direct-io.c
456
sdio->tail - sdio->head);
fs/direct-io.c
457
sdio->head = sdio->tail;
fs/direct-io.c
920
from = sdio->head ? 0 : sdio->from;
fs/direct-io.c
921
to = (sdio->head == sdio->tail - 1) ? sdio->to : PAGE_SIZE;
fs/direct-io.c
922
sdio->head++;
fs/dlm/lock.c
1638
static void lkb_add_ordered(struct list_head *new, struct list_head *head,
fs/dlm/lock.c
1643
list_for_each_entry(iter, head, lkb_statequeue)
fs/dlm/lock.c
1651
list_add_tail(new, head);
fs/dlm/lock.c
2139
static inline int first_in_list(struct dlm_lkb *lkb, struct list_head *head)
fs/dlm/lock.c
2141
struct dlm_lkb *first = list_entry(head->next, struct dlm_lkb,
fs/dlm/lock.c
2151
static int queue_conflict(struct list_head *head, struct dlm_lkb *lkb)
fs/dlm/lock.c
2155
list_for_each_entry(this, head, lkb_statequeue) {
fs/dlm/lock.c
2610
static void send_bast_queue(struct dlm_rsb *r, struct list_head *head,
fs/dlm/lock.c
2615
list_for_each_entry(gr, head, lkb_statequeue) {
fs/dlm/lock.c
5497
static struct dlm_lkb *search_remid_list(struct list_head *head, int nodeid,
fs/dlm/lock.c
5502
list_for_each_entry(lkb, head, lkb_statequeue) {
fs/dlm/member.c
291
struct list_head *head = &ls->ls_nodes;
fs/dlm/member.c
293
list_for_each(tmp, head) {
fs/dlm/member.c
300
list_add_tail(newlist, head);
fs/dlm/member.c
345
static struct dlm_member *find_memb(struct list_head *head, int nodeid)
fs/dlm/member.c
349
list_for_each_entry(memb, head, list) {
fs/dlm/member.c
372
static void clear_memb_list(struct list_head *head,
fs/dlm/member.c
377
while (!list_empty(head)) {
fs/dlm/member.c
378
memb = list_entry(head->next, struct dlm_member, list);
fs/dlm/recover.c
618
static int recover_locks_queue(struct dlm_rsb *r, struct list_head *head,
fs/dlm/recover.c
624
list_for_each_entry(lkb, head, lkb_statequeue) {
fs/erofs/decompressor_lzma.c
123
last->next = head;
fs/erofs/decompressor_lzma.c
124
head = strm;
fs/erofs/decompressor_lzma.c
129
for (strm = head; strm; strm = strm->next) {
fs/erofs/decompressor_lzma.c
140
z_erofs_lzma_head = head;
fs/erofs/decompressor_lzma.c
78
struct z_erofs_lzma *strm, *head = NULL;
fs/erofs/decompressor_zstd.c
108
strm->next = head;
fs/erofs/decompressor_zstd.c
109
head = strm;
fs/erofs/decompressor_zstd.c
116
for (strm = head; strm; strm = strm->next) {
fs/erofs/decompressor_zstd.c
128
z_erofs_zstd_head = head;
fs/erofs/decompressor_zstd.c
81
struct z_erofs_zstd *strm, *head = NULL;
fs/erofs/namei.c
147
head = mid + 1;
fs/erofs/namei.c
50
int head, back;
fs/erofs/namei.c
55
head = 1;
fs/erofs/namei.c
59
while (head <= back) {
fs/erofs/namei.c
60
const int mid = head + (back - head) / 2;
fs/erofs/namei.c
78
head = mid + 1;
fs/erofs/namei.c
93
int head = 0, back = erofs_iblks(dir) - 1;
fs/erofs/namei.c
97
while (head <= back) {
fs/erofs/namei.c
98
const int mid = head + (back - head) / 2;
fs/erofs/zdata.c
1402
.pcl = io->head,
fs/erofs/zdata.c
1421
DBG_BUGON(bgq->head == Z_EROFS_PCLUSTER_TAIL);
fs/erofs/zdata.c
1616
q->head = Z_EROFS_PCLUSTER_TAIL;
fs/erofs/zdata.c
1682
qtail[JQ_BYPASS] = &q[JQ_BYPASS]->head;
fs/erofs/zdata.c
1683
qtail[JQ_SUBMIT] = &q[JQ_SUBMIT]->head;
fs/erofs/zdata.c
1686
q[JQ_SUBMIT]->head = next = f->head;
fs/erofs/zdata.c
1809
if (f->head == Z_EROFS_PCLUSTER_TAIL)
fs/erofs/zdata.c
1916
struct folio *head = NULL, *folio;
fs/erofs/zdata.c
1922
folio->private = head;
fs/erofs/zdata.c
1923
head = folio;
fs/erofs/zdata.c
1927
while (head) {
fs/erofs/zdata.c
1928
folio = head;
fs/erofs/zdata.c
1929
head = folio_get_private(folio);
fs/erofs/zdata.c
503
struct z_erofs_pcluster *pcl, *head;
fs/erofs/zdata.c
513
.inode = i, .sharedinode = si, .head = Z_EROFS_PCLUSTER_TAIL, \
fs/erofs/zdata.c
759
pcl->next = fe->head;
fs/erofs/zdata.c
795
fe->head = fe->pcl = pcl;
fs/erofs/zdata.c
814
DBG_BUGON(!fe->head);
fs/erofs/zdata.c
851
if (!cmpxchg(&fe->pcl->next, NULL, fe->head)) {
fs/erofs/zdata.c
853
fe->head = fe->pcl;
fs/erofs/zdata.c
878
static void z_erofs_rcu_callback(struct rcu_head *head)
fs/erofs/zdata.c
880
z_erofs_free_pcluster(container_of(head, struct z_erofs_pcluster, rcu));
fs/erofs/zdata.c
95
struct z_erofs_pcluster *head;
fs/eventpoll.c
1530
struct hlist_head *head = NULL;
fs/eventpoll.c
1537
head = &ep->refs;
fs/eventpoll.c
1543
head = &to_free->epitems;
fs/eventpoll.c
1547
if (unlikely(!head)) {
fs/eventpoll.c
1552
WRITE_ONCE(file->f_ep, head);
fs/eventpoll.c
2157
struct epitems_head *head = tfile_check_list;
fs/eventpoll.c
2158
tfile_check_list = head->next;
fs/eventpoll.c
2159
unlist_file(head);
fs/eventpoll.c
283
static inline void free_ephead(struct epitems_head *head)
fs/eventpoll.c
285
if (head)
fs/eventpoll.c
286
kmem_cache_free(ephead_cache, head);
fs/eventpoll.c
291
struct epitems_head *head;
fs/eventpoll.c
293
head = container_of(file->f_ep, struct epitems_head, epitems);
fs/eventpoll.c
294
if (!head->next) {
fs/eventpoll.c
295
head->next = tfile_check_list;
fs/eventpoll.c
296
tfile_check_list = head;
fs/eventpoll.c
300
static void unlist_file(struct epitems_head *head)
fs/eventpoll.c
302
struct epitems_head *to_free = head;
fs/eventpoll.c
303
struct hlist_node *p = rcu_dereference(hlist_first_rcu(&head->epitems));
fs/eventpoll.c
307
if (!hlist_empty(&head->epitems))
fs/eventpoll.c
309
head->next = NULL;
fs/eventpoll.c
837
struct hlist_head *head;
fs/eventpoll.c
854
head = file->f_ep;
fs/eventpoll.c
855
if (head->first == &epi->fllink && !epi->fllink.next) {
fs/eventpoll.c
860
v = container_of(head, struct epitems_head, epitems);
fs/exfat/inode.c
565
struct hlist_head *head = sbi->inode_hashtable + exfat_hash(i_pos);
fs/exfat/inode.c
569
hlist_add_head(&EXFAT_I(inode)->i_hash_fat, head);
fs/exfat/inode.c
587
struct hlist_head *head = sbi->inode_hashtable + exfat_hash(i_pos);
fs/exfat/inode.c
591
hlist_for_each_entry(info, head, i_hash_fat) {
fs/ext4/ext4.h
3046
struct buffer_head *head,
fs/ext4/fast_commit.c
1047
struct ext4_fc_head head;
fs/ext4/fast_commit.c
1131
head.fc_features = cpu_to_le32(EXT4_FC_SUPPORTED_FEATURES);
fs/ext4/fast_commit.c
1132
head.fc_tid = cpu_to_le32(
fs/ext4/fast_commit.c
1134
if (!ext4_fc_add_tlv(sb, EXT4_FC_TAG_HEAD, sizeof(head),
fs/ext4/fast_commit.c
1135
(u8 *)&head, &crc)) {
fs/ext4/fast_commit.c
2086
struct ext4_fc_head head;
fs/ext4/fast_commit.c
2166
memcpy(&head, val, sizeof(head));
fs/ext4/fast_commit.c
2167
if (le32_to_cpu(head.fc_features) &
fs/ext4/fast_commit.c
2172
if (le32_to_cpu(head.fc_tid) != expected_tid) {
fs/ext4/fsmap.c
703
int ext4_getfsmap(struct super_block *sb, struct ext4_fsmap_head *head,
fs/ext4/fsmap.c
712
if (head->fmh_iflags & ~FMH_IF_VALID)
fs/ext4/fsmap.c
714
if (!ext4_getfsmap_is_valid_device(sb, &head->fmh_keys[0]) ||
fs/ext4/fsmap.c
715
!ext4_getfsmap_is_valid_device(sb, &head->fmh_keys[1]))
fs/ext4/fsmap.c
718
head->fmh_entries = 0;
fs/ext4/fsmap.c
744
dkeys[0] = head->fmh_keys[0];
fs/ext4/fsmap.c
750
if (!ext4_getfsmap_check_keys(dkeys, &head->fmh_keys[1]))
fs/ext4/fsmap.c
753
info.gfi_next_fsblk = head->fmh_keys[0].fmr_physical +
fs/ext4/fsmap.c
754
head->fmh_keys[0].fmr_length;
fs/ext4/fsmap.c
757
info.gfi_head = head;
fs/ext4/fsmap.c
764
if (head->fmh_keys[0].fmr_device > handlers[i].gfd_dev)
fs/ext4/fsmap.c
766
if (head->fmh_keys[1].fmr_device < handlers[i].gfd_dev)
fs/ext4/fsmap.c
776
if (handlers[i].gfd_dev == head->fmh_keys[1].fmr_device)
fs/ext4/fsmap.c
777
dkeys[1] = head->fmh_keys[1];
fs/ext4/fsmap.c
778
if (handlers[i].gfd_dev > head->fmh_keys[0].fmr_device)
fs/ext4/fsmap.c
790
head->fmh_oflags = FMH_OF_DEV_T;
fs/ext4/fsmap.h
39
int ext4_getfsmap(struct super_block *sb, struct ext4_fsmap_head *head,
fs/ext4/inode.c
1108
struct buffer_head *head,
fs/ext4/inode.c
1117
unsigned blocksize = head->b_size;
fs/ext4/inode.c
1121
for (bh = head, block_start = 0;
fs/ext4/inode.c
1122
ret == 0 && (bh != head || !block_start);
fs/ext4/inode.c
1176
struct buffer_head *bh, *head, *wait[2];
fs/ext4/inode.c
1186
head = folio_buffers(folio);
fs/ext4/inode.c
1187
if (!head)
fs/ext4/inode.c
1188
head = create_empty_buffers(folio, blocksize, 0);
fs/ext4/inode.c
1191
for (bh = head, block_start = 0; bh != head || !block_start;
fs/ext4/inode.c
1518
struct buffer_head *head, *bh;
fs/ext4/inode.c
1520
bh = head = folio_buffers(folio);
fs/ext4/inode.c
1539
} while (bh != head);
fs/ext4/inode.c
2173
struct buffer_head *head,
fs/ext4/inode.c
2198
} while (lblk++, (bh = bh->b_this_page) != head);
fs/ext4/inode.c
2201
err = mpage_submit_folio(mpd, head->b_folio);
fs/ext4/inode.c
2204
mpage_folio_done(mpd, head->b_folio);
fs/ext4/inode.c
2233
struct buffer_head *head, *bh;
fs/ext4/inode.c
2241
bh = head = folio_buffers(folio);
fs/ext4/inode.c
2254
err = mpage_process_page_bufs(mpd, head, bh, lblk);
fs/ext4/inode.c
2275
} while (lblk++, (bh = bh->b_this_page) != head);
fs/ext4/inode.c
2626
struct buffer_head *head;
fs/ext4/inode.c
2739
head = folio_buffers(folio);
fs/ext4/inode.c
2740
err = mpage_process_page_bufs(mpd, head, head,
fs/ext4/ioctl.c
893
struct fsmap_head head;
fs/ext4/ioctl.c
897
if (copy_from_user(&head, arg, sizeof(struct fsmap_head)))
fs/ext4/ioctl.c
899
if (memchr_inv(head.fmh_reserved, 0, sizeof(head.fmh_reserved)) ||
fs/ext4/ioctl.c
900
memchr_inv(head.fmh_keys[0].fmr_reserved, 0,
fs/ext4/ioctl.c
901
sizeof(head.fmh_keys[0].fmr_reserved)) ||
fs/ext4/ioctl.c
902
memchr_inv(head.fmh_keys[1].fmr_reserved, 0,
fs/ext4/ioctl.c
903
sizeof(head.fmh_keys[1].fmr_reserved)))
fs/ext4/ioctl.c
909
if (head.fmh_keys[0].fmr_offset ||
fs/ext4/ioctl.c
910
(head.fmh_keys[1].fmr_offset != 0 &&
fs/ext4/ioctl.c
911
head.fmh_keys[1].fmr_offset != -1ULL))
fs/ext4/ioctl.c
914
xhead.fmh_iflags = head.fmh_iflags;
fs/ext4/ioctl.c
915
xhead.fmh_count = head.fmh_count;
fs/ext4/ioctl.c
916
ext4_fsmap_to_internal(sb, &xhead.fmh_keys[0], &head.fmh_keys[0]);
fs/ext4/ioctl.c
917
ext4_fsmap_to_internal(sb, &xhead.fmh_keys[1], &head.fmh_keys[1]);
fs/ext4/ioctl.c
940
head.fmh_entries = xhead.fmh_entries;
fs/ext4/ioctl.c
941
head.fmh_oflags = xhead.fmh_oflags;
fs/ext4/ioctl.c
942
if (copy_to_user(arg, &head, sizeof(struct fsmap_head)))
fs/ext4/mballoc.c
5153
static void ext4_mb_pa_callback(struct rcu_head *head)
fs/ext4/mballoc.c
5157
pa = container_of(head, struct ext4_prealloc_space, u.pa_rcu);
fs/ext4/move_extent.c
121
struct buffer_head *bh, *head;
fs/ext4/move_extent.c
133
head = folio_buffers(folio);
fs/ext4/move_extent.c
134
if (!head)
fs/ext4/move_extent.c
135
head = create_empty_buffers(folio, blocksize, 0);
fs/ext4/move_extent.c
139
bh = head;
fs/ext4/move_extent.c
167
} while (block++, (bh = bh->b_this_page) != head);
fs/ext4/move_extent.c
173
bh = head;
fs/ext4/move_extent.c
183
} while ((bh = bh->b_this_page) != head);
fs/ext4/move_extent.c
267
struct buffer_head *bh, *head;
fs/ext4/move_extent.c
272
head = folio_buffers(folio);
fs/ext4/move_extent.c
273
if (!head)
fs/ext4/move_extent.c
274
head = create_empty_buffers(folio, blocksize, 0);
fs/ext4/move_extent.c
278
bh = head;
fs/ext4/move_extent.c
288
} while (block++, (bh = bh->b_this_page) != head);
fs/ext4/page-io.c
108
struct buffer_head *bh, *head;
fs/ext4/page-io.c
123
bh = head = folio_buffers(folio);
fs/ext4/page-io.c
128
spin_lock_irqsave(&head->b_uptodate_lock, flags);
fs/ext4/page-io.c
141
} while ((bh = bh->b_this_page) != head);
fs/ext4/page-io.c
142
spin_unlock_irqrestore(&head->b_uptodate_lock, flags);
fs/ext4/page-io.c
215
static void dump_completed_IO(struct inode *inode, struct list_head *head)
fs/ext4/page-io.c
221
if (list_empty(head))
fs/ext4/page-io.c
225
list_for_each_entry(io_end, head, list) {
fs/ext4/page-io.c
273
struct list_head *head)
fs/ext4/page-io.c
282
dump_completed_IO(inode, head);
fs/ext4/page-io.c
283
list_replace_init(head, &unwritten);
fs/ext4/page-io.c
464
struct buffer_head *bh, *head;
fs/ext4/page-io.c
491
bh = head = folio_buffers(folio);
fs/ext4/page-io.c
525
} while ((bh = bh->b_this_page) != head);
fs/ext4/page-io.c
537
bh = head = folio_buffers(folio);
fs/ext4/page-io.c
583
} while (bh != head);
fs/ext4/page-io.c
597
} while ((bh = bh->b_this_page) != head);
fs/ext4/resize.c
24
static void ext4_rcu_ptr_callback(struct rcu_head *head)
fs/ext4/resize.c
28
ptr = container_of(head, struct ext4_rcu_ptr, rcu);
fs/ext4/super.c
522
struct buffer_head *bh, *head;
fs/ext4/super.c
525
bh = head = folio_buffers(folio);
fs/ext4/super.c
545
} while ((bh = bh->b_this_page) != head);
fs/f2fs/checkpoint.c
1000
head = &im->ino_list;
fs/f2fs/checkpoint.c
1003
list_for_each_entry(orphan, head, list) {
fs/f2fs/checkpoint.c
1278
struct list_head *head;
fs/f2fs/checkpoint.c
1297
head = &sbi->inode_list[type];
fs/f2fs/checkpoint.c
1298
if (list_empty(head)) {
fs/f2fs/checkpoint.c
1305
fi = list_first_entry(head, struct f2fs_inode_info, dirty_list);
fs/f2fs/checkpoint.c
1340
struct list_head *head = &sbi->inode_list[DIRTY_META];
fs/f2fs/checkpoint.c
1350
if (list_empty(head)) {
fs/f2fs/checkpoint.c
1354
fi = list_first_entry(head, struct f2fs_inode_info,
fs/f2fs/checkpoint.c
984
struct list_head *head;
fs/f2fs/data.c
847
struct list_head *head = &io->bio_list;
fs/f2fs/data.c
851
list_for_each_entry(be, head, list) {
fs/f2fs/data.c
895
struct list_head *head = &io->bio_list;
fs/f2fs/data.c
898
if (list_empty(head))
fs/f2fs/data.c
902
list_for_each_entry(be, head, list) {
fs/f2fs/data.c
919
list_for_each_entry(be, head, list) {
fs/f2fs/node.c
2210
struct list_head *head = &sbi->fsync_node_list;
fs/f2fs/node.c
2218
if (list_empty(head)) {
fs/f2fs/node.c
2222
fn = list_first_entry(head, struct fsync_node_entry, list);
fs/f2fs/node.c
256
struct nat_entry_set *head;
fs/f2fs/node.c
258
head = radix_tree_lookup(&nm_i->nat_set_root, set);
fs/f2fs/node.c
259
if (!head) {
fs/f2fs/node.c
260
head = f2fs_kmem_cache_alloc(nat_entry_set_slab,
fs/f2fs/node.c
263
INIT_LIST_HEAD(&head->entry_list);
fs/f2fs/node.c
264
INIT_LIST_HEAD(&head->set_list);
fs/f2fs/node.c
265
head->set = set;
fs/f2fs/node.c
266
head->entry_cnt = 0;
fs/f2fs/node.c
267
f2fs_radix_tree_insert(&nm_i->nat_set_root, set, head);
fs/f2fs/node.c
269
return head;
fs/f2fs/node.c
275
struct nat_entry_set *head;
fs/f2fs/node.c
279
head = __grab_nat_entry_set(nm_i, ne);
fs/f2fs/node.c
288
head->entry_cnt++;
fs/f2fs/node.c
304
list_move_tail(&ne->list, &head->entry_list);
fs/f2fs/node.c
3048
struct list_head *head, int max)
fs/f2fs/node.c
3055
list_for_each_entry(cur, head, set_list) {
fs/f2fs/node.c
3062
list_add_tail(&nes->set_list, head);
fs/f2fs/recovery.c
401
static int find_fsync_dnodes(struct f2fs_sb_info *sbi, struct list_head *head,
fs/f2fs/recovery.c
435
entry = get_fsync_inode(head, ino_of_node(folio));
fs/f2fs/recovery.c
450
entry = add_fsync_inode(sbi, head, ino_of_node(folio),
fs/f2fs/recovery.c
484
static void destroy_fsync_dnodes(struct list_head *head, int drop)
fs/f2fs/recovery.c
488
list_for_each_entry_safe(entry, tmp, head, list)
fs/f2fs/recovery.c
62
static struct fsync_inode_entry *get_fsync_inode(struct list_head *head,
fs/f2fs/recovery.c
67
list_for_each_entry(entry, head, list)
fs/f2fs/recovery.c
75
struct list_head *head, nid_t ino, bool quota_inode)
fs/f2fs/recovery.c
98
list_add_tail(&entry->list, head);
fs/f2fs/segment.c
2085
struct list_head *head = &SM_I(sbi)->dcc_info->entry_list;
fs/f2fs/segment.c
2126
list_add_tail(&de->list, head);
fs/f2fs/segment.c
2145
struct list_head *head = &(SM_I(sbi)->dcc_info->entry_list);
fs/f2fs/segment.c
2149
list_for_each_entry_safe(entry, this, head, list)
fs/f2fs/segment.c
2171
struct list_head *head = &dcc->entry_list;
fs/f2fs/segment.c
2241
list_for_each_entry_safe(entry, this, head, list) {
fs/f2fs/segment.c
284
static void __complete_revoke_list(struct inode *inode, struct list_head *head,
fs/f2fs/segment.c
291
list_for_each_entry_safe(cur, tmp, head, list) {
fs/f2fs/segment.c
4608
struct list_head *head)
fs/f2fs/segment.c
4612
if (list_is_last(&ses->set_list, head))
fs/f2fs/segment.c
4615
list_for_each_entry_continue(next, head, set_list)
fs/f2fs/segment.c
4621
list_move_tail(&ses->set_list, head);
fs/f2fs/segment.c
4624
static void add_sit_entry(unsigned int segno, struct list_head *head)
fs/f2fs/segment.c
4629
list_for_each_entry(ses, head, set_list) {
fs/f2fs/segment.c
4632
adjust_sit_entry_set(ses, head);
fs/f2fs/segment.c
4641
list_add(&ses->set_list, head);
fs/f2fs/segment.c
4687
struct list_head *head = &SM_I(sbi)->sit_entry_set;
fs/f2fs/segment.c
4716
list_for_each_entry_safe(ses, tmp, head, set_list) {
fs/f2fs/segment.c
4791
f2fs_bug_on(sbi, !list_empty(head));
fs/fat/inode.c
407
struct hlist_head *head = sbi->inode_hashtable
fs/fat/inode.c
412
hlist_add_head(&MSDOS_I(inode)->i_fat_hash, head);
fs/fat/inode.c
450
struct hlist_head *head = sbi->inode_hashtable + fat_hash(i_pos);
fs/fat/inode.c
455
hlist_for_each_entry(i, head, i_fat_hash) {
fs/fat/nfs.c
26
struct hlist_head *head;
fs/fat/nfs.c
30
head = sbi->dir_hashtable + fat_dir_hash(i_logstart);
fs/fat/nfs.c
32
hlist_for_each_entry(i, head, i_dir_hash) {
fs/fs-writeback.c
116
struct list_head *head)
fs/fs-writeback.c
122
list_move(&inode->i_io_list, head);
fs/fs-writeback.c
125
if (head != &wb->b_dirty_time)
fs/fs-writeback.c
66
static inline struct inode *wb_inode(struct list_head *head)
fs/fs-writeback.c
68
return list_entry(head, struct inode, i_io_list);
fs/fs_context.c
443
index = log->head & (logsize - 1);
fs/fs_context.c
444
BUILD_BUG_ON(sizeof(log->head) != sizeof(u8) ||
fs/fs_context.c
446
if ((u8)(log->head - log->tail) == logsize) {
fs/fs_context.c
458
log->head++;
fs/fsopen.c
27
if (unlikely(log->head == log->tail))
fs/fuse/dev.c
1282
struct fuse_forget_link *head = fiq->forget_list_head.next;
fs/fuse/dev.c
1283
struct fuse_forget_link **newhead = &head;
fs/fuse/dev.c
1297
return head;
fs/fuse/dev.c
1340
struct fuse_forget_link *head;
fs/fuse/dev.c
1354
head = fuse_dequeue_forget(fiq, max_forgets, &count);
fs/fuse/dev.c
1363
while (head) {
fs/fuse/dev.c
1364
struct fuse_forget_link *forget = head;
fs/fuse/dev.c
1370
head = forget->next;
fs/fuse/dev.c
2295
unsigned int head, tail, count;
fs/fuse/dev.c
2309
head = pipe->head;
fs/fuse/dev.c
2311
count = pipe_occupancy(head, tail);
fs/fuse/dev.c
2321
for (idx = tail; !pipe_empty(head, idx) && rem < len; idx++)
fs/fuse/dev.c
2333
if (WARN_ON(nbuf >= count || pipe_empty(head, tail)))
fs/fuse/dev.c
2406
void fuse_dev_end_requests(struct list_head *head)
fs/fuse/dev.c
2408
while (!list_empty(head)) {
fs/fuse/dev.c
2410
req = list_entry(head->next, struct fuse_req, list);
fs/fuse/dev_uring.c
362
static void fuse_uring_stop_list_entries(struct list_head *head,
fs/fuse/dev_uring.c
372
list_for_each_entry_safe(ent, next, head, list) {
fs/fuse/fuse_dev_i.h
61
void fuse_dev_end_requests(struct list_head *head);
fs/gfs2/aops.c
611
struct buffer_head *bh, *head;
fs/gfs2/aops.c
617
head = folio_buffers(folio);
fs/gfs2/aops.c
618
if (!head)
fs/gfs2/aops.c
621
bh = head;
fs/gfs2/aops.c
630
} while (bh != head);
fs/gfs2/aops.c
651
struct buffer_head *bh, *head;
fs/gfs2/aops.c
654
head = folio_buffers(folio);
fs/gfs2/aops.c
655
if (!head)
fs/gfs2/aops.c
670
bh = head;
fs/gfs2/aops.c
680
} while (bh != head);
fs/gfs2/aops.c
682
bh = head;
fs/gfs2/aops.c
700
} while (bh != head);
fs/gfs2/glops.c
602
struct gfs2_log_header_host head;
fs/gfs2/glops.c
608
error = gfs2_find_jhead(sdp->sd_jdesc, &head);
fs/gfs2/glops.c
611
if (gfs2_assert_withdraw(sdp, head.lh_flags &
fs/gfs2/glops.c
614
gfs2_log_pointers_init(sdp, &head);
fs/gfs2/glops.c
62
struct list_head *head = &gl->gl_ail_list;
fs/gfs2/glops.c
69
list_for_each_entry_safe_reverse(bd, tmp, head, bd_ail_gl_list) {
fs/gfs2/incore.h
60
struct gfs2_log_header_host *head, int pass);
fs/gfs2/log.c
1002
head = &tr->tr_buf;
fs/gfs2/log.c
1003
while (!list_empty(head)) {
fs/gfs2/log.c
1004
bd = list_first_entry(head, struct gfs2_bufdata, bd_list);
fs/gfs2/log.c
1010
head = &tr->tr_databuf;
fs/gfs2/log.c
1011
while (!list_empty(head)) {
fs/gfs2/log.c
1012
bd = list_first_entry(head, struct gfs2_bufdata, bd_list);
fs/gfs2/log.c
198
struct list_head *head = &sdp->sd_ail1_list;
fs/gfs2/log.c
215
list_for_each_entry_reverse(tr, head, tr_list) {
fs/gfs2/log.c
279
struct list_head *head)
fs/gfs2/log.c
283
while (!list_empty(head)) {
fs/gfs2/log.c
284
bd = list_first_entry(head, struct gfs2_bufdata,
fs/gfs2/log.c
997
struct list_head *head;
fs/gfs2/lops.c
1070
struct list_head *head;
fs/gfs2/lops.c
1076
head = &tr->tr_databuf;
fs/gfs2/lops.c
1077
while (!list_empty(head)) {
fs/gfs2/lops.c
1078
bd = list_first_entry(head, struct gfs2_bufdata, bd_list);
fs/gfs2/lops.c
417
struct gfs2_log_header_host *head,
fs/gfs2/lops.c
430
if (lh.lh_sequence >= head->lh_sequence)
fs/gfs2/lops.c
431
*head = lh;
fs/gfs2/lops.c
463
struct gfs2_log_header_host *head,
fs/gfs2/lops.c
475
*done = gfs2_jhead_folio_search(jd, head, folio);
fs/gfs2/lops.c
504
int gfs2_find_jhead(struct gfs2_jdesc *jd, struct gfs2_log_header_host *head)
fs/gfs2/lops.c
520
memset(head, 0, sizeof(*head));
fs/gfs2/lops.c
575
gfs2_jhead_process_page(jd, blocks_read >> shift, head, &done);
fs/gfs2/lops.c
586
gfs2_jhead_process_page(jd, blocks_read >> shift, head, &done);
fs/gfs2/lops.c
724
struct list_head *head;
fs/gfs2/lops.c
730
head = &tr->tr_buf;
fs/gfs2/lops.c
731
while (!list_empty(head)) {
fs/gfs2/lops.c
732
bd = list_first_entry(head, struct gfs2_bufdata, bd_list);
fs/gfs2/lops.c
739
struct gfs2_log_header_host *head, int pass)
fs/gfs2/lops.c
851
struct list_head *head = &sdp->sd_log_revokes;
fs/gfs2/lops.c
864
list_for_each_entry(bd, head, bd_list) {
fs/gfs2/lops.c
888
struct list_head *head = &sdp->sd_log_revokes;
fs/gfs2/lops.c
892
while (!list_empty(head)) {
fs/gfs2/lops.c
893
bd = list_first_entry(head, struct gfs2_bufdata, bd_list);
fs/gfs2/lops.c
907
struct gfs2_log_header_host *head, int pass)
fs/gfs2/lops.c
913
jd->jd_replay_tail = head->lh_tail;
fs/gfs2/lops.h
23
struct gfs2_log_header_host *head);
fs/gfs2/lops.h
55
struct gfs2_log_header_host *head,
fs/gfs2/lops.h
61
gfs2_log_ops[x]->lo_before_scan(jd, head, pass);
fs/gfs2/meta_io.c
36
struct buffer_head *bh, *head;
fs/gfs2/meta_io.c
42
head = folio_buffers(folio);
fs/gfs2/meta_io.c
43
bh = head;
fs/gfs2/meta_io.c
66
} while ((bh = bh->b_this_page) != head);
fs/gfs2/meta_io.c
82
} while (bh != head);
fs/gfs2/meta_io.h
19
static inline void gfs2_buffer_clear_tail(struct buffer_head *bh, int head)
fs/gfs2/meta_io.h
21
BUG_ON(head > bh->b_size);
fs/gfs2/meta_io.h
22
memset(bh->b_data + head, 0, bh->b_size - head);
fs/gfs2/recovery.c
108
struct list_head *head = &jd->jd_revoke_list;
fs/gfs2/recovery.c
111
while (!list_empty(head)) {
fs/gfs2/recovery.c
112
rr = list_first_entry(head, struct gfs2_revoke_replay, rr_list);
fs/gfs2/recovery.c
119
unsigned int blkno, struct gfs2_log_header_host *head)
fs/gfs2/recovery.c
141
head->lh_sequence = be64_to_cpu(lh->lh_sequence);
fs/gfs2/recovery.c
142
head->lh_flags = be32_to_cpu(lh->lh_flags);
fs/gfs2/recovery.c
143
head->lh_tail = be32_to_cpu(lh->lh_tail);
fs/gfs2/recovery.c
144
head->lh_blkno = be32_to_cpu(lh->lh_blkno);
fs/gfs2/recovery.c
146
head->lh_local_total = be64_to_cpu(lh->lh_local_total);
fs/gfs2/recovery.c
147
head->lh_local_free = be64_to_cpu(lh->lh_local_free);
fs/gfs2/recovery.c
148
head->lh_local_dinodes = be64_to_cpu(lh->lh_local_dinodes);
fs/gfs2/recovery.c
167
struct gfs2_log_header_host *head)
fs/gfs2/recovery.c
178
blk, head);
fs/gfs2/recovery.c
264
struct gfs2_log_header_host *head)
fs/gfs2/recovery.c
268
gfs2_replay_incr_blk(jd, &head->lh_blkno);
fs/gfs2/recovery.c
269
head->lh_sequence++;
fs/gfs2/recovery.c
270
gfs2_write_log_header(sdp, jd, head->lh_sequence, 0, head->lh_blkno,
fs/gfs2/recovery.c
305
struct gfs2_log_header_host *head,
fs/gfs2/recovery.c
323
if (head) { /* Update the master statfs inode */
fs/gfs2/recovery.c
325
sc.sc_total += head->lh_local_total;
fs/gfs2/recovery.c
326
sc.sc_free += head->lh_local_free;
fs/gfs2/recovery.c
327
sc.sc_dinodes += head->lh_local_dinodes;
fs/gfs2/recovery.c
333
sc.sc_free, sc.sc_dinodes, head->lh_local_total,
fs/gfs2/recovery.c
334
head->lh_local_free, head->lh_local_dinodes);
fs/gfs2/recovery.c
374
struct gfs2_log_header_host *head)
fs/gfs2/recovery.c
379
if (!head->lh_local_total && !head->lh_local_free
fs/gfs2/recovery.c
380
&& !head->lh_local_dinodes) /* No change */
fs/gfs2/recovery.c
385
error = update_statfs_inode(jd, head, sdp->sd_statfs_inode);
fs/gfs2/recovery.c
403
struct gfs2_log_header_host head;
fs/gfs2/recovery.c
459
error = gfs2_find_jhead(jd, &head);
fs/gfs2/recovery.c
466
if (!(head.lh_flags & GFS2_LOG_HEAD_UNMOUNT)) {
fs/gfs2/recovery.c
503
jd->jd_jid, head.lh_tail, head.lh_blkno);
fs/gfs2/recovery.c
510
lops_before_scan(jd, &head, pass);
fs/gfs2/recovery.c
511
error = foreach_descriptor(jd, head.lh_tail,
fs/gfs2/recovery.c
512
head.lh_blkno, pass);
fs/gfs2/recovery.c
520
recover_local_statfs(jd, &head);
fs/gfs2/recovery.c
521
clean_journal(jd, &head);
fs/gfs2/recovery.c
536
gfs2_log_pointers_init(sdp, &head);
fs/gfs2/recovery.c
57
struct list_head *head = &jd->jd_revoke_list;
fs/gfs2/recovery.c
586
struct gfs2_log_header_host *head)
fs/gfs2/recovery.c
588
sdp->sd_log_sequence = head->lh_sequence + 1;
fs/gfs2/recovery.c
589
gfs2_replay_incr_blk(sdp->sd_jdesc, &head->lh_blkno);
fs/gfs2/recovery.c
590
sdp->sd_log_tail = head->lh_blkno;
fs/gfs2/recovery.c
591
sdp->sd_log_flush_head = head->lh_blkno;
fs/gfs2/recovery.c
592
sdp->sd_log_flush_tail = head->lh_blkno;
fs/gfs2/recovery.c
593
sdp->sd_log_head = head->lh_blkno;
fs/gfs2/recovery.c
60
list_for_each_entry(iter, head, rr_list) {
fs/gfs2/recovery.c
78
list_add(&rr->rr_list, head);
fs/gfs2/recovery.h
31
struct gfs2_log_header_host *head);
fs/gfs2/recovery.h
33
struct gfs2_log_header_host *head);
fs/gfs2/super.c
85
static struct gfs2_jdesc *jdesc_find_i(struct list_head *head, unsigned int jid)
fs/gfs2/super.c
89
list_for_each_entry(jd, head, jd_list) {
fs/gfs2/trans.c
237
struct buffer_head *head = folio_buffers(folio);
fs/gfs2/trans.c
238
unsigned int bsize = head->b_size;
fs/gfs2/trans.c
243
for (bh = head, start = 0; bh != head || !start;
fs/gfs2/util.c
57
struct gfs2_log_header_host head;
fs/gfs2/util.c
76
error = gfs2_find_jhead(jd, &head);
fs/gfs2/util.c
83
if (!(head.lh_flags & GFS2_LOG_HEAD_UNMOUNT)) {
fs/hfs/btree.c
115
head = (struct hfs_btree_header_rec *)(kmap_local_folio(folio, 0) +
fs/hfs/btree.c
117
tree->root = be32_to_cpu(head->root);
fs/hfs/btree.c
118
tree->leaf_count = be32_to_cpu(head->leaf_count);
fs/hfs/btree.c
119
tree->leaf_head = be32_to_cpu(head->leaf_head);
fs/hfs/btree.c
120
tree->leaf_tail = be32_to_cpu(head->leaf_tail);
fs/hfs/btree.c
121
tree->node_count = be32_to_cpu(head->node_count);
fs/hfs/btree.c
122
tree->free_nodes = be32_to_cpu(head->free_nodes);
fs/hfs/btree.c
123
tree->attributes = be32_to_cpu(head->attributes);
fs/hfs/btree.c
124
tree->node_size = be16_to_cpu(head->node_size);
fs/hfs/btree.c
125
tree->max_key_len = be16_to_cpu(head->max_key_len);
fs/hfs/btree.c
126
tree->depth = be16_to_cpu(head->depth);
fs/hfs/btree.c
155
kunmap_local(head);
fs/hfs/btree.c
161
kunmap_local(head);
fs/hfs/btree.c
199
struct hfs_btree_header_rec *head;
fs/hfs/btree.c
209
head = (struct hfs_btree_header_rec *)(kmap_local_page(page) +
fs/hfs/btree.c
212
head->root = cpu_to_be32(tree->root);
fs/hfs/btree.c
213
head->leaf_count = cpu_to_be32(tree->leaf_count);
fs/hfs/btree.c
214
head->leaf_head = cpu_to_be32(tree->leaf_head);
fs/hfs/btree.c
215
head->leaf_tail = cpu_to_be32(tree->leaf_tail);
fs/hfs/btree.c
216
head->node_count = cpu_to_be32(tree->node_count);
fs/hfs/btree.c
217
head->free_nodes = cpu_to_be32(tree->free_nodes);
fs/hfs/btree.c
218
head->attributes = cpu_to_be32(tree->attributes);
fs/hfs/btree.c
219
head->depth = cpu_to_be16(tree->depth);
fs/hfs/btree.c
22
struct hfs_btree_header_rec *head;
fs/hfs/btree.c
221
kunmap_local(head);
fs/hfsplus/btree.c
136
struct hfs_btree_header_rec *head;
fs/hfsplus/btree.c
166
head = (struct hfs_btree_header_rec *)(kmap_local_page(page) +
fs/hfsplus/btree.c
168
tree->root = be32_to_cpu(head->root);
fs/hfsplus/btree.c
169
tree->leaf_count = be32_to_cpu(head->leaf_count);
fs/hfsplus/btree.c
170
tree->leaf_head = be32_to_cpu(head->leaf_head);
fs/hfsplus/btree.c
171
tree->leaf_tail = be32_to_cpu(head->leaf_tail);
fs/hfsplus/btree.c
172
tree->node_count = be32_to_cpu(head->node_count);
fs/hfsplus/btree.c
173
tree->free_nodes = be32_to_cpu(head->free_nodes);
fs/hfsplus/btree.c
174
tree->attributes = be32_to_cpu(head->attributes);
fs/hfsplus/btree.c
175
tree->node_size = be16_to_cpu(head->node_size);
fs/hfsplus/btree.c
176
tree->max_key_len = be16_to_cpu(head->max_key_len);
fs/hfsplus/btree.c
177
tree->depth = be16_to_cpu(head->depth);
fs/hfsplus/btree.c
206
(head->key_type == HFSPLUS_KEY_BINARY))
fs/hfsplus/btree.c
243
kunmap_local(head);
fs/hfsplus/btree.c
248
kunmap_local(head);
fs/hfsplus/btree.c
285
struct hfs_btree_header_rec *head;
fs/hfsplus/btree.c
295
head = (struct hfs_btree_header_rec *)(kmap_local_page(page) +
fs/hfsplus/btree.c
298
head->root = cpu_to_be32(tree->root);
fs/hfsplus/btree.c
299
head->leaf_count = cpu_to_be32(tree->leaf_count);
fs/hfsplus/btree.c
300
head->leaf_head = cpu_to_be32(tree->leaf_head);
fs/hfsplus/btree.c
301
head->leaf_tail = cpu_to_be32(tree->leaf_tail);
fs/hfsplus/btree.c
302
head->node_count = cpu_to_be32(tree->node_count);
fs/hfsplus/btree.c
303
head->free_nodes = cpu_to_be32(tree->free_nodes);
fs/hfsplus/btree.c
304
head->attributes = cpu_to_be32(tree->attributes);
fs/hfsplus/btree.c
305
head->depth = cpu_to_be16(tree->depth);
fs/hfsplus/btree.c
307
kunmap_local(head);
fs/hfsplus/xattr.c
101
map_nodes = (be32_to_cpu(head->node_count) -
fs/hfsplus/xattr.c
104
be32_add_cpu(&head->free_nodes, 0 - map_nodes);
fs/hfsplus/xattr.c
109
be32_to_cpu(head->node_count) - be32_to_cpu(head->free_nodes);
fs/hfsplus/xattr.c
58
struct hfs_btree_header_rec *head;
fs/hfsplus/xattr.c
79
head = (struct hfs_btree_header_rec *)(buf + offset);
fs/hfsplus/xattr.c
80
head->node_size = cpu_to_be16(node_size);
fs/hfsplus/xattr.c
83
head->node_count = cpu_to_be32(tmp);
fs/hfsplus/xattr.c
84
head->free_nodes = cpu_to_be32(be32_to_cpu(head->node_count) - 1);
fs/hfsplus/xattr.c
85
head->clump_size = cpu_to_be32(clump_size);
fs/hfsplus/xattr.c
86
head->attributes |= cpu_to_be32(HFS_TREE_BIGKEYS | HFS_TREE_VARIDXKEYS);
fs/hfsplus/xattr.c
87
head->max_key_len = cpu_to_be16(HFSPLUS_ATTR_KEYLEN - sizeof(u16));
fs/hfsplus/xattr.c
94
if (be32_to_cpu(head->node_count) > hdr_node_map_rec_bits) {
fs/hfsplus/xattr.c
98
desc->next = cpu_to_be32(be32_to_cpu(head->leaf_tail) + 1);
fs/inode.c
1046
struct hlist_head *head,
fs/inode.c
1060
hlist_for_each_entry_rcu(inode, head, i_hash) {
fs/inode.c
1090
struct hlist_head *head, unsigned long ino,
fs/inode.c
1102
hlist_for_each_entry_rcu(inode, head, i_hash) {
fs/inode.c
1308
struct hlist_head *head = inode_hashtable + hash(inode->i_sb, hashval);
fs/inode.c
1316
old = find_inode(inode->i_sb, head, test, data, true, &isnew);
fs/inode.c
1345
hlist_add_head_rcu(&inode->i_hash, head);
fs/inode.c
1415
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1422
inode = find_inode(sb, head, test, data, false, &isnew);
fs/inode.c
1460
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1467
inode = find_inode_fast(sb, head, ino, false, &isnew);
fs/inode.c
1486
old = find_inode_fast(sb, head, ino, true, &isnew);
fs/inode.c
1491
hlist_add_head_rcu(&inode->i_hash, head);
fs/inode.c
1622
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1626
inode = find_inode(sb, head, test, data, true, isnew);
fs/inode.c
1682
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1689
inode = find_inode_fast(sb, head, ino, false, &isnew);
fs/inode.c
1734
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1739
hlist_for_each_entry(inode, head, i_hash) {
fs/inode.c
1779
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1785
hlist_for_each_entry_rcu(inode, head, i_hash) {
fs/inode.c
1817
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1823
hlist_for_each_entry_rcu(inode, head, i_hash) {
fs/inode.c
1837
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1846
hlist_for_each_entry(old, head, i_hash) {
fs/inode.c
1857
hlist_add_head_rcu(&inode->i_hash, head);
fs/inode.c
322
static void i_callback(struct rcu_head *head)
fs/inode.c
324
struct inode *inode = container_of(head, struct inode, i_rcu);
fs/inode.c
880
static void dispose_list(struct list_head *head)
fs/inode.c
882
while (!list_empty(head)) {
fs/inode.c
885
inode = list_first_entry(head, struct inode, i_lru);
fs/jbd2/transaction.c
2130
struct buffer_head *head;
fs/jbd2/transaction.c
2136
head = folio_buffers(folio);
fs/jbd2/transaction.c
2137
bh = head;
fs/jbd2/transaction.c
2162
} while ((bh = bh->b_this_page) != head);
fs/jbd2/transaction.c
2432
struct buffer_head *head, *bh, *next;
fs/jbd2/transaction.c
2441
head = folio_buffers(folio);
fs/jbd2/transaction.c
2442
if (!head)
fs/jbd2/transaction.c
2451
bh = head;
fs/jbd2/transaction.c
2471
} while (bh != head);
fs/jffs2/nodemgmt.c
572
static inline int on_list(struct list_head *obj, struct list_head *head)
fs/jffs2/nodemgmt.c
576
list_for_each(this, head) {
fs/jffs2/nodemgmt.c
578
jffs2_dbg(1, "%p is on list at %p\n", obj, head);
fs/jffs2/scan.c
1135
static void rotate_list(struct list_head *head, uint32_t count)
fs/jffs2/scan.c
1137
struct list_head *n = head->next;
fs/jffs2/scan.c
1139
list_del(head);
fs/jffs2/scan.c
1143
list_add(head, n);
fs/jfs/jfs_imap.c
2854
int i, n, head;
fs/jfs/jfs_imap.c
2916
if ((head = imap->im_agctl[n].inofree) == -1) {
fs/jfs/jfs_imap.c
2920
if ((rc = diIAGRead(imap, head, &hbp))) {
fs/jfs/jfs_imap.c
2926
iagp->inofreefwd = cpu_to_le32(head);
fs/jfs/jfs_imap.c
2942
if ((head = imap->im_agctl[n].extfree) == -1) {
fs/jfs/jfs_imap.c
2946
if ((rc = diIAGRead(imap, head, &hbp))) {
fs/jfs/jfs_imap.c
2952
iagp->extfreefwd = cpu_to_le32(head);
fs/locks.c
2980
struct list_head *head, int *id,
fs/locks.c
2985
list_for_each_entry(fl, head, flc_list) {
fs/mbcache.c
106
head = mb_cache_entry_head(cache, key);
fs/mbcache.c
107
hlist_bl_lock(head);
fs/mbcache.c
108
hlist_bl_for_each_entry(dup, dup_node, head, e_hash_list) {
fs/mbcache.c
110
hlist_bl_unlock(head);
fs/mbcache.c
115
hlist_bl_add_head(&entry->e_hash_list, head);
fs/mbcache.c
116
hlist_bl_unlock(head);
fs/mbcache.c
129
struct hlist_bl_head *head;
fs/mbcache.c
131
head = mb_cache_entry_head(cache, entry->e_key);
fs/mbcache.c
132
hlist_bl_lock(head);
fs/mbcache.c
134
hlist_bl_unlock(head);
fs/mbcache.c
158
struct hlist_bl_head *head;
fs/mbcache.c
160
head = mb_cache_entry_head(cache, key);
fs/mbcache.c
161
hlist_bl_lock(head);
fs/mbcache.c
165
node = hlist_bl_first(head);
fs/mbcache.c
177
hlist_bl_unlock(head);
fs/mbcache.c
226
struct hlist_bl_head *head;
fs/mbcache.c
229
head = mb_cache_entry_head(cache, key);
fs/mbcache.c
230
hlist_bl_lock(head);
fs/mbcache.c
231
hlist_bl_for_each_entry(entry, node, head, e_hash_list) {
fs/mbcache.c
238
hlist_bl_unlock(head);
fs/mbcache.c
79
struct hlist_bl_head *head;
fs/mpage.c
101
struct buffer_head *page_bh, *head;
fs/mpage.c
104
head = folio_buffers(folio);
fs/mpage.c
105
if (!head) {
fs/mpage.c
115
head = create_empty_buffers(folio, i_blocksize(inode), 0);
fs/mpage.c
118
page_bh = head;
fs/mpage.c
128
} while (page_bh != head);
fs/mpage.c
432
struct buffer_head *bh, *head = folio_buffers(folio);
fs/mpage.c
434
if (!head)
fs/mpage.c
436
bh = head;
fs/mpage.c
443
} while (bh != head);
fs/mpage.c
475
struct buffer_head *head = folio_buffers(folio);
fs/mpage.c
477
if (head) {
fs/mpage.c
478
struct buffer_head *bh = head;
fs/mpage.c
514
} while ((bh = bh->b_this_page) != head);
fs/namespace.c
1317
static void __cleanup_mnt(struct rcu_head *head)
fs/namespace.c
1319
cleanup_mnt(container_of(head, struct mount, mnt_rcu));
fs/namespace.c
1686
struct hlist_head head;
fs/namespace.c
1692
hlist_move_list(&unmounted, &head);
fs/namespace.c
1716
if (likely(hlist_empty(&head)))
fs/namespace.c
1721
hlist_for_each_entry_safe(m, p, &head, mnt_umount) {
fs/namespace.c
734
static void delayed_free_vfsmnt(struct rcu_head *head)
fs/namespace.c
736
free_vfsmnt(container_of(head, struct mount, mnt_rcu));
fs/namespace.c
791
struct hlist_head *head = m_hash(mnt, dentry);
fs/namespace.c
794
hlist_for_each_entry_rcu(p, head, mnt_hash)
fs/netfs/rolling_buffer.c
109
smp_store_release(&head->next, fq);
fs/netfs/rolling_buffer.c
128
fq = roll->head;
fs/netfs/rolling_buffer.c
164
slot = folioq_append(roll->head, folio);
fs/netfs/rolling_buffer.c
166
folioq_mark(roll->head, slot);
fs/netfs/rolling_buffer.c
168
folioq_mark2(roll->head, slot);
fs/netfs/rolling_buffer.c
71
roll->head = fq;
fs/netfs/rolling_buffer.c
82
struct folio_queue *fq, *head = roll->head;
fs/netfs/rolling_buffer.c
84
if (!folioq_full(head))
fs/netfs/rolling_buffer.c
87
fq = netfs_folioq_alloc(head->rreq_id, GFP_NOFS, netfs_trace_folioq_make_space);
fs/netfs/rolling_buffer.c
90
fq->prev = head;
fs/netfs/rolling_buffer.c
92
roll->head = fq;
fs/netfs/rolling_buffer.c
93
if (folioq_full(head)) {
fs/netfs/rolling_buffer.c
97
if (roll->iter.folioq == head &&
fs/netfs/rolling_buffer.c
98
roll->iter.folioq_slot == folioq_nr_slots(head)) {
fs/nfs/blocklayout/extent_tree.c
126
static void __ext_put_deviceids(struct list_head *head)
fs/nfs/blocklayout/extent_tree.c
130
list_for_each_entry_safe(be, tmp, head, be_list) {
fs/nfs/delegation.c
1188
struct hlist_head *head = nfs_delegation_hash(server, fhandle);
fs/nfs/delegation.c
1193
hlist_for_each_entry_rcu(delegation, head, hash) {
fs/nfs/dir.c
2892
static void nfs_access_free_list(struct list_head *head)
fs/nfs/dir.c
2896
while (!list_empty(head)) {
fs/nfs/dir.c
2897
cache = list_entry(head->next, struct nfs_access_entry, lru);
fs/nfs/dir.c
2906
LIST_HEAD(head);
fs/nfs/dir.c
2923
list_move(&cache->lru, &head);
fs/nfs/dir.c
2939
nfs_access_free_list(&head);
fs/nfs/dir.c
2977
static void __nfs_access_zap_cache(struct nfs_inode *nfsi, struct list_head *head)
fs/nfs/dir.c
2987
list_move(&entry->lru, head);
fs/nfs/dir.c
2994
LIST_HEAD(head);
fs/nfs/dir.c
3004
__nfs_access_zap_cache(NFS_I(inode), &head);
fs/nfs/dir.c
3007
nfs_access_free_list(&head);
fs/nfs/direct.c
309
static void nfs_read_sync_pgio_error(struct list_head *head, int error)
fs/nfs/direct.c
313
while (!list_empty(head)) {
fs/nfs/direct.c
314
req = nfs_list_entry(head->next);
fs/nfs/direct.c
512
struct nfs_page *head = req->wb_head;
fs/nfs/direct.c
514
if (!list_empty(&head->wb_list) || !nfs_lock_request(head))
fs/nfs/direct.c
516
if (!list_empty(&head->wb_list)) {
fs/nfs/direct.c
517
nfs_unlock_request(head);
fs/nfs/direct.c
520
list_add(&head->wb_list, list);
fs/nfs/direct.c
521
kref_get(&head->wb_kref);
fs/nfs/direct.c
522
kref_get(&head->wb_kref);
fs/nfs/direct.c
811
static void nfs_write_sync_pgio_error(struct list_head *head, int error)
fs/nfs/direct.c
815
while (!list_empty(head)) {
fs/nfs/direct.c
816
req = nfs_list_entry(head->next);
fs/nfs/dns_resolve.c
104
static void nfs_dns_ent_free_rcu(struct rcu_head *head)
fs/nfs/dns_resolve.c
108
item = container_of(head, struct nfs_dns_ent, rcu_head);
fs/nfs/flexfilelayout/flexfilelayout.c
2582
.head = {
fs/nfs/flexfilelayout/flexfilelayout.c
2667
LIST_HEAD(head);
fs/nfs/flexfilelayout/flexfilelayout.c
2671
ff_layout_fetch_ds_ioerr(lo, &lseg->pls_range, &head, -1);
fs/nfs/flexfilelayout/flexfilelayout.c
2672
if (list_empty(&head))
fs/nfs/flexfilelayout/flexfilelayout.c
2680
list_for_each_entry(pos, &head, list) {
fs/nfs/flexfilelayout/flexfilelayout.c
2688
if (!list_is_last(&pos->list, &head) &&
fs/nfs/flexfilelayout/flexfilelayout.c
2697
ff_layout_free_ds_ioerr(&head);
fs/nfs/flexfilelayout/flexfilelayout.h
222
int ff_layout_encode_ds_ioerr(struct xdr_stream *xdr, const struct list_head *head);
fs/nfs/flexfilelayout/flexfilelayout.h
223
void ff_layout_free_ds_ioerr(struct list_head *head);
fs/nfs/flexfilelayout/flexfilelayout.h
226
struct list_head *head,
fs/nfs/flexfilelayout/flexfilelayoutdev.c
228
struct list_head *head = &flo->error_list;
fs/nfs/flexfilelayout/flexfilelayoutdev.c
238
head = &err->list;
fs/nfs/flexfilelayout/flexfilelayoutdev.c
248
list_add_tail(&dserr->list, head);
fs/nfs/flexfilelayout/flexfilelayoutdev.c
473
void ff_layout_free_ds_ioerr(struct list_head *head)
fs/nfs/flexfilelayout/flexfilelayoutdev.c
477
while (!list_empty(head)) {
fs/nfs/flexfilelayout/flexfilelayoutdev.c
478
err = list_first_entry(head,
fs/nfs/flexfilelayout/flexfilelayoutdev.c
487
int ff_layout_encode_ds_ioerr(struct xdr_stream *xdr, const struct list_head *head)
fs/nfs/flexfilelayout/flexfilelayoutdev.c
492
list_for_each_entry(err, head, list) {
fs/nfs/flexfilelayout/flexfilelayoutdev.c
522
struct list_head *head,
fs/nfs/flexfilelayout/flexfilelayoutdev.c
539
list_move(&err->list, head);
fs/nfs/flexfilelayout/flexfilelayoutdev.c
549
struct list_head *head,
fs/nfs/flexfilelayout/flexfilelayoutdev.c
554
ret = do_layout_fetch_ds_ioerr(lo, range, head, maxnum);
fs/nfs/internal.h
600
struct list_head *head,
fs/nfs/internal.h
614
int nfs_generic_commit_list(struct inode *inode, struct list_head *head,
fs/nfs/nfs42proc.c
1161
LIST_HEAD(head);
fs/nfs/nfs42proc.c
1167
pnfs_mark_layout_stateid_invalid(lo, &head);
fs/nfs/nfs42proc.c
1169
pnfs_free_lseg_list(&head);
fs/nfs/nfs42proc.c
991
LIST_HEAD(head);
fs/nfs/nfs42proc.c
997
pnfs_mark_layout_stateid_invalid(lo, &head);
fs/nfs/nfs42proc.c
999
pnfs_free_lseg_list(&head);
fs/nfs/nfs42xdr.c
1131
bufpos = xdr->buf->head[0].iov_len + (offset - args->offset);
fs/nfs/nfs4_fs.h
502
extern void nfs4_free_state_owners(struct list_head *head);
fs/nfs/nfs4proc.c
9588
LIST_HEAD(head);
fs/nfs/nfs4proc.c
9653
pnfs_mark_layout_stateid_invalid(lo, &head);
fs/nfs/nfs4proc.c
9656
pnfs_free_lseg_list(&head);
fs/nfs/nfs4state.c
580
void nfs4_purge_state_owners(struct nfs_server *server, struct list_head *head)
fs/nfs/nfs4state.c
587
list_move(&sp->so_lru, head);
fs/nfs/nfs4state.c
600
void nfs4_free_state_owners(struct list_head *head)
fs/nfs/nfs4state.c
604
list_for_each_entry_safe(sp, tmp, head, so_lru) {
fs/nfs/pagelist.c
1162
LIST_HEAD(head);
fs/nfs/pagelist.c
1164
nfs_list_move_request(req, &head);
fs/nfs/pagelist.c
1165
desc->pg_completion_ops->error_cleanup(&head, desc->pg_error);
fs/nfs/pagelist.c
1240
LIST_HEAD(head);
fs/nfs/pagelist.c
1243
list_splice_init(&mirror->pg_list, &head);
fs/nfs/pagelist.c
1246
while (!list_empty(&head)) {
fs/nfs/pagelist.c
1249
req = list_first_entry(&head, struct nfs_page, wb_list);
fs/nfs/pagelist.c
1253
list_splice_tail(&head, &mirror->pg_list);
fs/nfs/pagelist.c
265
struct nfs_page *head = req->wb_head;
fs/nfs/pagelist.c
268
WARN_ON_ONCE(!test_bit(PG_HEADLOCK, &head->wb_flags));
fs/nfs/pagelist.c
356
struct nfs_page *head = req->wb_head;
fs/nfs/pagelist.c
373
if (head != req)
fs/nfs/pagelist.c
374
nfs_release_request(head);
fs/nfs/pagelist.c
883
struct list_head *head = &mirror->pg_list;
fs/nfs/pagelist.c
910
while (!list_empty(head)) {
fs/nfs/pagelist.c
914
req = nfs_list_entry(head->next);
fs/nfs/pnfs.c
1234
struct list_head *head)
fs/nfs/pnfs.c
1244
pnfs_lseg_dec_and_remove_zero(lseg, head);
fs/nfs/pnfs.c
2780
struct list_head *head;
fs/nfs/pnfs.c
2788
head = &NFS_I(lo->plh_inode)->open_files;
fs/nfs/pnfs.c
2789
list_for_each_entry_rcu(ctx, head, list) {
fs/nfs/pnfs.c
421
LIST_HEAD(head);
fs/nfs/pnfs.c
435
err = pnfs_mark_matching_lsegs_return(lo, &head, &range, 0);
fs/nfs/pnfs.c
444
pnfs_free_lseg_list(&head);
fs/nfs/pnfs.c
523
LIST_HEAD(head);
fs/nfs/pnfs.c
527
pnfs_mark_matching_lsegs_return(lo, &head, &range, 0);
fs/nfs/pnfs.c
529
pnfs_free_lseg_list(&head);
fs/nfs/pnfs.h
561
pnfs_recover_commit_reqs(struct list_head *head, struct nfs_commit_info *cinfo)
fs/nfs/pnfs.h
566
fl_cinfo->ops->recover_commit_reqs(head, cinfo);
fs/nfs/pnfs.h
864
pnfs_recover_commit_reqs(struct list_head *head, struct nfs_commit_info *cinfo)
fs/nfs/pnfs_nfs.c
357
pnfs_bucket_get_committing(struct list_head *head,
fs/nfs/pnfs_nfs.c
366
list_splice_init(&bucket->committing, head);
fs/nfs/read.c
200
nfs_async_read_error(struct list_head *head, int error)
fs/nfs/read.c
204
while (!list_empty(head)) {
fs/nfs/read.c
205
req = nfs_list_entry(head->next);
fs/nfs/write.c
1372
static void nfs_async_write_error(struct list_head *head, int error)
fs/nfs/write.c
1376
while (!list_empty(head)) {
fs/nfs/write.c
1377
req = nfs_list_entry(head->next);
fs/nfs/write.c
1671
static loff_t nfs_get_lwb(struct list_head *head)
fs/nfs/write.c
1676
list_for_each_entry(req, head, wb_list)
fs/nfs/write.c
1687
struct list_head *head,
fs/nfs/write.c
1698
if (head)
fs/nfs/write.c
1699
list_splice_init(head, &data->pages);
fs/nfs/write.c
1756
nfs_commit_list(struct inode *inode, struct list_head *head, int how,
fs/nfs/write.c
1764
if (list_empty(head))
fs/nfs/write.c
1769
nfs_retry_commit(head, NULL, cinfo, -1);
fs/nfs/write.c
1774
nfs_init_commit(data, head, NULL, cinfo);
fs/nfs/write.c
1870
int nfs_generic_commit_list(struct inode *inode, struct list_head *head,
fs/nfs/write.c
1875
status = pnfs_commit_list(inode, head, how, cinfo);
fs/nfs/write.c
1877
status = nfs_commit_list(inode, head, how, cinfo);
fs/nfs/write.c
1884
LIST_HEAD(head);
fs/nfs/write.c
1893
ret = nscan = nfs_scan_commit(inode, &head, &cinfo);
fs/nfs/write.c
1896
ret = nfs_generic_commit_list(inode, &head, how, &cinfo);
fs/nfs/write.c
383
void nfs_join_page_group(struct nfs_page *head, struct nfs_commit_info *cinfo,
fs/nfs/write.c
390
pgbase = head->wb_pgbase;
fs/nfs/write.c
391
bytes = head->wb_bytes;
fs/nfs/write.c
392
off = head->wb_offset;
fs/nfs/write.c
393
for (subreq = head->wb_this_page; subreq != head;
fs/nfs/write.c
406
head->wb_pgbase = pgbase;
fs/nfs/write.c
407
head->wb_bytes = bytes;
fs/nfs/write.c
408
head->wb_offset = off;
fs/nfs/write.c
412
subreq = head;
fs/nfs/write.c
416
} while (subreq != head);
fs/nfs/write.c
419
if (head->wb_this_page != head) {
fs/nfs/write.c
421
destroy_list = head->wb_this_page;
fs/nfs/write.c
422
head->wb_this_page = head;
fs/nfs/write.c
425
nfs_destroy_unlinked_subrequests(destroy_list, head, inode);
fs/nfs/write.c
454
nfs_unroll_locks(struct nfs_page *head, struct nfs_page *req)
fs/nfs/write.c
459
for (tmp = head->wb_this_page ; tmp != req; tmp = tmp->wb_this_page) {
fs/nfs/write.c
476
nfs_page_group_lock_subreq(struct nfs_page *head, struct nfs_page *subreq)
fs/nfs/write.c
483
nfs_page_group_unlock(head);
fs/nfs/write.c
486
ret = nfs_page_group_lock(head);
fs/nfs/write.c
488
nfs_unroll_locks(head, subreq);
fs/nfs/write.c
514
struct nfs_page *head, *subreq;
fs/nfs/write.c
524
head = nfs_folio_find_head_request(folio);
fs/nfs/write.c
525
if (!head)
fs/nfs/write.c
528
while (!nfs_lock_request(head)) {
fs/nfs/write.c
529
ret = nfs_wait_on_request(head);
fs/nfs/write.c
531
nfs_release_request(head);
fs/nfs/write.c
536
ret = nfs_page_group_lock(head);
fs/nfs/write.c
541
if (head != folio->private) {
fs/nfs/write.c
542
nfs_page_group_unlock(head);
fs/nfs/write.c
543
nfs_unlock_and_release_request(head);
fs/nfs/write.c
547
nfs_cancel_remove_inode(head, inode);
fs/nfs/write.c
550
for (subreq = head->wb_this_page;
fs/nfs/write.c
551
subreq != head;
fs/nfs/write.c
553
ret = nfs_page_group_lock_subreq(head, subreq);
fs/nfs/write.c
558
nfs_page_group_unlock(head);
fs/nfs/write.c
561
nfs_join_page_group(head, &cinfo, inode);
fs/nfs/write.c
562
return head;
fs/nfs/write.c
565
nfs_unlock_and_release_request(head);
fs/nfsd/filecache.c
498
struct list_head *head = arg;
fs/nfsd/filecache.c
531
list_add(&nf->nf_gc, head);
fs/nfsd/nfs3xdr.c
613
struct kvec *head = rqstp->rq_arg.head;
fs/nfsd/nfs3xdr.c
623
args->first.iov_len = head->iov_len - xdr_stream_pos(xdr);
fs/nfsd/nfs3xdr.c
818
struct kvec *head = rqstp->rq_res.head;
fs/nfsd/nfs3xdr.c
830
if (svc_encode_result_payload(rqstp, head->iov_len, resp->len) < 0)
fs/nfsd/nfs3xdr.c
846
struct kvec *head = rqstp->rq_res.head;
fs/nfsd/nfs3xdr.c
863
if (svc_encode_result_payload(rqstp, head->iov_len, resp->count) < 0)
fs/nfsd/nfs4proc.c
1570
struct list_head *head = &nn->conf_id_hashtbl[idhashval];
fs/nfsd/nfs4proc.c
1572
list_for_each_entry(clp, head, cl_idhash) {
fs/nfsd/nfs4proc.c
3221
buflen -= rqstp->rq_res.head[0].iov_len;
fs/nfsd/nfs4state.c
1783
struct list_head *head = &nn->conf_id_hashtbl[idhashval];
fs/nfsd/nfs4state.c
1786
list_for_each_entry(clp, head, cl_idhash) {
fs/nfsd/nfs4xdr.c
2291
struct kvec *head = xdr->head;
fs/nfsd/nfs4xdr.c
2295
if (buflen <= head->iov_len) {
fs/nfsd/nfs4xdr.c
2300
*bufp = head->iov_base;
fs/nfsd/nfs4xdr.c
2309
memcpy(dp, head->iov_base, head->iov_len);
fs/nfsd/nfs4xdr.c
2310
buflen -= head->iov_len;
fs/nfsd/nfs4xdr.c
2311
dp += head->iov_len;
fs/nfsd/nfs4xdr.c
4105
buf->head[0].iov_base = p;
fs/nfsd/nfs4xdr.c
4106
buf->head[0].iov_len = 0;
fs/nfsd/nfs4xdr.c
4109
xdr->iov = buf->head;
fs/nfsd/nfs4xdr.c
4771
if (xdr->iov != xdr->buf->head || xdr->end - xdr->p < 1)
fs/nfsd/nfs4xdr.c
4783
buf->head[0].iov_len, maxcount);
fs/nfsd/nfscache.c
226
struct list_head *head = &nn->drc_hashtbl[i].lru_head;
fs/nfsd/nfscache.c
227
while (!list_empty(head)) {
fs/nfsd/nfscache.c
228
rp = list_first_entry(head, struct nfsd_cacherep, c_lru);
fs/nfsd/nfscache.c
369
if (subbuf.head[0].iov_len) {
fs/nfsd/nfscache.c
370
len = min_t(unsigned int, subbuf.head[0].iov_len, remaining);
fs/nfsd/nfscache.c
371
csum = csum_partial(subbuf.head[0].iov_base, len, csum);
fs/nfsd/nfscache.c
576
struct kvec *resv = &rqstp->rq_res.head[0], *cachv;
fs/nfsd/nfsxdr.c
380
struct kvec *head = rqstp->rq_arg.head;
fs/nfsd/nfsxdr.c
389
args->first.iov_len = head->iov_len - xdr_stream_pos(xdr);
fs/nfsd/nfsxdr.c
463
struct kvec *head = rqstp->rq_res.head;
fs/nfsd/nfsxdr.c
473
if (svc_encode_result_payload(rqstp, head->iov_len, resp->len) < 0)
fs/nfsd/nfsxdr.c
485
struct kvec *head = rqstp->rq_res.head;
fs/nfsd/nfsxdr.c
498
if (svc_encode_result_payload(rqstp, head->iov_len, resp->count) < 0)
fs/nilfs2/btree.c
2135
struct list_head *head;
fs/nilfs2/btree.c
2156
list_for_each(head, &lists[level]) {
fs/nilfs2/btree.c
2157
cbh = list_entry(head, struct buffer_head, b_assoc_buffers);
fs/nilfs2/btree.c
2163
list_add_tail(&bh->b_assoc_buffers, head);
fs/nilfs2/btree.c
2173
struct buffer_head *bh, *head;
fs/nilfs2/btree.c
2187
bh = head = folio_buffers(fbatch.folios[i]);
fs/nilfs2/btree.c
2192
} while ((bh = bh->b_this_page) != head);
fs/nilfs2/file.c
52
struct buffer_head *bh, *head;
fs/nilfs2/file.c
74
head = folio_buffers(folio);
fs/nilfs2/file.c
75
if (head) {
fs/nilfs2/file.c
78
bh = head;
fs/nilfs2/file.c
84
} while (bh = bh->b_this_page, bh != head);
fs/nilfs2/gcinode.c
176
struct list_head *head = &nilfs->ns_gc_inodes;
fs/nilfs2/gcinode.c
179
while (!list_empty(head)) {
fs/nilfs2/gcinode.c
180
ii = list_first_entry(head, struct nilfs_inode_info, i_dirty);
fs/nilfs2/inode.c
181
struct buffer_head *head;
fs/nilfs2/inode.c
189
head = folio_buffers(folio);
fs/nilfs2/inode.c
190
if (head) {
fs/nilfs2/inode.c
191
struct buffer_head *bh = head;
fs/nilfs2/inode.c
200
} while (bh = bh->b_this_page, bh != head);
fs/nilfs2/mdt.c
627
struct list_head *head = &shadow->frozen_buffers;
fs/nilfs2/mdt.c
630
while (!list_empty(head)) {
fs/nilfs2/mdt.c
631
bh = list_first_entry(head, struct buffer_head,
fs/nilfs2/page.c
142
struct buffer_head *bh, *head;
fs/nilfs2/page.c
144
bh = head = folio_buffers(folio);
fs/nilfs2/page.c
149
} while (bh != head);
fs/nilfs2/page.c
155
struct buffer_head *bh, *head;
fs/nilfs2/page.c
172
head = folio_buffers(folio);
fs/nilfs2/page.c
173
if (head) {
fs/nilfs2/page.c
176
bh = head;
fs/nilfs2/page.c
183
} while (bh != head);
fs/nilfs2/page.c
402
struct buffer_head *bh, *head;
fs/nilfs2/page.c
406
head = folio_buffers(folio);
fs/nilfs2/page.c
407
if (head) {
fs/nilfs2/page.c
417
bh = head;
fs/nilfs2/page.c
423
} while (bh = bh->b_this_page, bh != head);
fs/nilfs2/page.c
433
bh = head;
fs/nilfs2/page.c
438
} while (bh = bh->b_this_page, bh != head);
fs/nilfs2/page.c
451
struct buffer_head *bh, *head;
fs/nilfs2/page.c
454
for (bh = head = folio_buffers(folio), block_start = 0;
fs/nilfs2/page.c
455
bh != head || !block_start;
fs/nilfs2/page.c
535
struct buffer_head *bh, *head;
fs/nilfs2/page.c
539
bh = head = folio_buffers(folio);
fs/nilfs2/page.c
550
} while (++b, bh = bh->b_this_page, bh != head);
fs/nilfs2/recovery.c
330
struct list_head *head)
fs/nilfs2/recovery.c
383
list_add_tail(&rb->list, head);
fs/nilfs2/recovery.c
399
static void dispose_recovery_list(struct list_head *head)
fs/nilfs2/recovery.c
401
while (!list_empty(head)) {
fs/nilfs2/recovery.c
404
rb = list_first_entry(head, struct nilfs_recovery_block, list);
fs/nilfs2/recovery.c
415
static int nilfs_segment_list_add(struct list_head *head, __u64 segnum)
fs/nilfs2/recovery.c
424
list_add_tail(&ent->list, head);
fs/nilfs2/recovery.c
428
void nilfs_dispose_segment_list(struct list_head *head)
fs/nilfs2/recovery.c
430
while (!list_empty(head)) {
fs/nilfs2/recovery.c
433
ent = list_first_entry(head, struct nilfs_segment_entry, list);
fs/nilfs2/recovery.c
443
struct list_head *head = &ri->ri_used_segments;
fs/nilfs2/recovery.c
473
err = nilfs_segment_list_add(head, segnum[i]);
fs/nilfs2/recovery.c
482
list_for_each_entry_safe(ent, n, head, list) {
fs/nilfs2/recovery.c
525
struct list_head *head,
fs/nilfs2/recovery.c
535
list_for_each_entry_safe(rb, n, head, list) {
fs/nilfs2/recovery.c
760
LIST_HEAD(head);
fs/nilfs2/recovery.c
764
list_splice_init(&nilfs->ns_dirty_files, &head);
fs/nilfs2/recovery.c
766
if (list_empty(&head))
fs/nilfs2/recovery.c
770
list_for_each_entry_safe(ii, n, &head, i_dirty) {
fs/nilfs2/segbuf.h
102
#define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next == head)
fs/nilfs2/segbuf.h
85
#define NILFS_LIST_SEGBUF(head) \
fs/nilfs2/segbuf.h
86
list_entry((head), struct nilfs_segment_buffer, sb_list)
fs/nilfs2/segbuf.h
89
#define NILFS_LAST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->prev)
fs/nilfs2/segbuf.h
90
#define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next)
fs/nilfs2/segbuf.h
91
#define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head))
fs/nilfs2/segbuf.h
97
#define NILFS_SEGBUF_FIRST_BH(head) \
fs/nilfs2/segbuf.h
98
(list_entry((head)->next, struct buffer_head, b_assoc_buffers))
fs/nilfs2/segment.c
1166
struct list_head *head;
fs/nilfs2/segment.c
1194
head = &sci->sc_gc_inodes;
fs/nilfs2/segment.c
1196
head, i_dirty);
fs/nilfs2/segment.c
1197
list_for_each_entry_continue(ii, head, i_dirty) {
fs/nilfs2/segment.c
1215
head = &sci->sc_dirty_files;
fs/nilfs2/segment.c
1216
ii = list_prepare_entry(sci->sc_stage.dirty_file_ptr, head,
fs/nilfs2/segment.c
1218
list_for_each_entry_continue(ii, head, i_dirty) {
fs/nilfs2/segment.c
2287
list_for_each_entry_safe(wrq, n, &sci->sc_wait_request.head, wq.entry) {
fs/nilfs2/segment.c
2494
nilfs_remove_written_gcinodes(struct the_nilfs *nilfs, struct list_head *head)
fs/nilfs2/segment.c
2498
list_for_each_entry_safe(ii, n, head, i_dirty) {
fs/nilfs2/segment.c
726
struct buffer_head *bh, *head;
fs/nilfs2/segment.c
735
head = folio_buffers(folio);
fs/nilfs2/segment.c
736
if (!head)
fs/nilfs2/segment.c
737
head = create_empty_buffers(folio,
fs/nilfs2/segment.c
740
bh = head;
fs/nilfs2/segment.c
753
} while (bh = bh->b_this_page, bh != head);
fs/nilfs2/segment.c
768
struct buffer_head *bh, *head;
fs/nilfs2/segment.c
779
bh = head = folio_buffers(fbatch.folios[i]);
fs/nilfs2/segment.c
788
} while (bh != head);
fs/nilfs2/segment.c
796
struct list_head *head, int force)
fs/nilfs2/segment.c
802
while (!list_empty(head)) {
fs/nilfs2/segment.c
804
list_for_each_entry_safe(ii, n, head, i_dirty) {
fs/nilfs2/segment.c
974
static void nilfs_redirty_inodes(struct list_head *head)
fs/nilfs2/segment.c
978
list_for_each_entry(ii, head, i_dirty) {
fs/nilfs2/segment.c
984
static void nilfs_drop_collected_inodes(struct list_head *head)
fs/nilfs2/segment.c
988
list_for_each_entry(ii, head, i_dirty) {
fs/notify/mark.c
1003
head = &group->marks_list;
fs/notify/mark.c
1025
if (list_empty(head)) {
fs/notify/mark.c
1029
mark = list_first_entry(head, struct fsnotify_mark, g_list);
fs/notify/mark.c
999
struct list_head *head = &to_free;
fs/ntfs3/fslog.c
585
const struct CLIENT_REC *cr, __le16 *head)
fs/ntfs3/fslog.c
588
*head = cr->next_client;
fs/ntfs3/fslog.c
599
static inline void add_client(struct CLIENT_REC *ca, u16 index, __le16 *head)
fs/ntfs3/fslog.c
604
cr->next_client = *head;
fs/ntfs3/fslog.c
606
if (*head != LFS_NO_CLIENT_LE)
fs/ntfs3/fslog.c
607
ca[le16_to_cpu(*head)].prev_client = cpu_to_le16(index);
fs/ntfs3/fslog.c
609
*head = cpu_to_le16(index);
fs/ocfs2/alloc.c
6420
struct ocfs2_cached_block_free *head)
fs/ocfs2/alloc.c
6444
while (head) {
fs/ocfs2/alloc.c
6445
if (head->free_bg)
fs/ocfs2/alloc.c
6446
bg_blkno = head->free_bg;
fs/ocfs2/alloc.c
6448
bg_blkno = ocfs2_which_suballoc_group(head->free_blk,
fs/ocfs2/alloc.c
6449
head->free_bit);
fs/ocfs2/alloc.c
6458
(unsigned long long)head->free_blk, head->free_bit);
fs/ocfs2/alloc.c
6461
head->free_bit, bg_blkno, 1);
fs/ocfs2/alloc.c
6467
tmp = head;
fs/ocfs2/alloc.c
6468
head = head->free_next;
fs/ocfs2/alloc.c
6479
while(head) {
fs/ocfs2/alloc.c
6481
tmp = head;
fs/ocfs2/alloc.c
6482
head = head->free_next;
fs/ocfs2/alloc.c
6513
struct ocfs2_cached_block_free *head)
fs/ocfs2/alloc.c
6522
while (head) {
fs/ocfs2/alloc.c
6538
ret = ocfs2_truncate_log_append(osb, handle, head->free_blk,
fs/ocfs2/alloc.c
6539
head->free_bit);
fs/ocfs2/alloc.c
6542
tmp = head;
fs/ocfs2/alloc.c
6543
head = head->free_next;
fs/ocfs2/alloc.c
6554
while (head) {
fs/ocfs2/alloc.c
6556
tmp = head;
fs/ocfs2/alloc.c
6557
head = head->free_next;
fs/ocfs2/aops.c
395
struct buffer_head *head,
fs/ocfs2/aops.c
404
unsigned blocksize = head->b_size;
fs/ocfs2/aops.c
408
for ( bh = head, block_start = 0;
fs/ocfs2/aops.c
409
ret == 0 && (bh != head || !block_start);
fs/ocfs2/aops.c
578
struct buffer_head *head, *bh, *wait[2], **wait_bh = wait;
fs/ocfs2/aops.c
582
head = folio_buffers(folio);
fs/ocfs2/aops.c
583
if (!head)
fs/ocfs2/aops.c
584
head = create_empty_buffers(folio, bsize, 0);
fs/ocfs2/aops.c
586
for (bh = head, block_start = 0; bh != head || !block_start;
fs/ocfs2/aops.c
643
bh = head;
fs/ocfs2/aops.c
659
} while (bh != head);
fs/ocfs2/aops.c
795
struct list_head *head)
fs/ocfs2/aops.c
800
list_for_each_entry_safe(ue, tmp, head, ue_node) {
fs/ocfs2/aops.c
859
struct buffer_head *head, *bh;
fs/ocfs2/aops.c
862
head = folio_buffers(folio);
fs/ocfs2/aops.c
863
if (!head)
fs/ocfs2/aops.c
866
bh = head;
fs/ocfs2/aops.c
890
} while (bh != head);
fs/ocfs2/aops.h
18
struct buffer_head *head,
fs/ocfs2/dlm/dlmast.c
266
struct list_head *head = NULL;
fs/ocfs2/dlm/dlmast.c
345
head = &res->converting;
fs/ocfs2/dlm/dlmast.c
347
list_for_each_entry(lock, head, list) {
fs/ocfs2/dlm/dlmast.c
354
head = &res->blocked;
fs/ocfs2/dlm/dlmast.c
356
head = &res->granted;
fs/ocfs2/dlm/dlmast.c
358
list_for_each_entry(lock, head, list) {
fs/ocfs2/dlm/dlmcommon.h
1064
static inline int dlm_lock_on_list(struct list_head *head,
fs/ocfs2/dlm/dlmcommon.h
1069
list_for_each_entry(tmplock, head, list) {
fs/ocfs2/journal.c
2132
struct inode *head;
fs/ocfs2/journal.c
2176
OCFS2_I(iter)->ip_next_orphan = p->head;
fs/ocfs2/journal.c
2177
p->head = iter;
fs/ocfs2/journal.c
2184
struct inode **head,
fs/ocfs2/journal.c
2192
.head = *head,
fs/ocfs2/journal.c
2218
*head = priv.head;
fs/ocfs2/quota_local.c
237
static void ocfs2_release_local_quota_bitmaps(struct list_head *head)
fs/ocfs2/quota_local.c
241
list_for_each_entry_safe(pos, next, head, qc_chunk) {
fs/ocfs2/quota_local.c
251
struct list_head *head)
fs/ocfs2/quota_local.c
256
INIT_LIST_HEAD(head);
fs/ocfs2/quota_local.c
260
ocfs2_release_local_quota_bitmaps(head);
fs/ocfs2/quota_local.c
271
ocfs2_release_local_quota_bitmaps(head);
fs/ocfs2/quota_local.c
274
list_add_tail(&newchunk->qc_chunk, head);
fs/ocfs2/quota_local.c
297
struct list_head *head)
fs/ocfs2/quota_local.c
312
list_add_tail(&rc->rc_list, head);
fs/ocfs2/quota_local.c
316
static void free_recovery_list(struct list_head *head)
fs/ocfs2/quota_local.c
321
list_for_each_entry_safe(rchunk, next, head, rc_list) {
fs/ocfs2/quota_local.c
341
struct list_head *head)
fs/ocfs2/quota_local.c
360
status = ocfs2_add_recovery_chunk(sb, dchunk, i, head);
fs/ocfs2/quota_local.c
366
free_recovery_list(head);
fs/orangefs/devorangefs-req.c
370
} head;
fs/orangefs/devorangefs-req.c
373
int head_size = sizeof(head);
fs/orangefs/devorangefs-req.c
388
if (!copy_from_iter_full(&head, head_size, iter)) {
fs/orangefs/devorangefs-req.c
393
if (head.version < ORANGEFS_MINIMUM_USERSPACE_VERSION) {
fs/orangefs/devorangefs-req.c
397
head.version,
fs/orangefs/devorangefs-req.c
402
if (head.magic != ORANGEFS_DEVREQ_MAGIC) {
fs/orangefs/devorangefs-req.c
408
orangefs_userspace_version = head.version;
fs/orangefs/devorangefs-req.c
409
} else if (orangefs_userspace_version != head.version) {
fs/orangefs/devorangefs-req.c
415
op = orangefs_devreq_remove_op(head.tag);
fs/orangefs/devorangefs-req.c
419
__func__, llu(head.tag));
fs/pipe.c
1294
unsigned int head, tail, mask, n;
fs/pipe.c
1306
head = pipe->head;
fs/pipe.c
1309
n = pipe_occupancy(head, tail);
fs/pipe.c
1321
unsigned int h = head & mask;
fs/pipe.c
1336
head = n;
fs/pipe.c
1345
pipe->head = head;
fs/pipe.c
235
return !pipe_empty(idx.head, idx.tail) || !writers;
fs/pipe.c
294
unsigned int head = smp_load_acquire(&pipe->head);
fs/pipe.c
321
if (!pipe_empty(head, tail)) {
fs/pipe.c
360
wake_writer |= pipe_full(head, tail, pipe->max_usage);
fs/pipe.c
366
if (!pipe_empty(head, tail)) /* More to do? */
fs/pipe.c
426
return !pipe_full(idx.head, idx.tail, max_usage) ||
fs/pipe.c
435
unsigned int head;
fs/pipe.c
475
head = pipe->head;
fs/pipe.c
476
was_empty = pipe_empty(head, pipe->tail);
fs/pipe.c
479
struct pipe_buffer *buf = pipe_buf(pipe, head - 1);
fs/pipe.c
509
head = pipe->head;
fs/pipe.c
510
if (!pipe_full(head, pipe->tail, pipe->max_usage)) {
fs/pipe.c
530
pipe->head = head + 1;
fs/pipe.c
532
buf = pipe_buf(pipe, head);
fs/pipe.c
622
unsigned int count, head, tail;
fs/pipe.c
628
head = pipe->head;
fs/pipe.c
631
while (!pipe_empty(head, tail)) {
fs/pipe.c
689
if (!pipe_empty(idx.head, idx.tail))
fs/pipe.c
696
if (!pipe_full(idx.head, idx.tail, pipe->max_usage))
fs/pnode.c
440
struct list_head *head;
fs/pnode.c
446
head = &child->mnt_mounts;
fs/pnode.c
447
if (!list_empty(head)) {
fs/pnode.c
452
if (!list_is_singular(head) || !child->overmount)
fs/proc/inode.c
33
struct ctl_table_header *head;
fs/proc/inode.c
43
head = ei->sysctl;
fs/proc/inode.c
44
if (head) {
fs/proc/inode.c
46
proc_sys_evict_inode(inode, head);
fs/proc/internal.h
354
struct ctl_table_header *head);
fs/proc/internal.h
358
struct ctl_table_header *head) { }
fs/proc/kcore.c
142
static int kcore_ram_list(struct list_head *head)
fs/proc/kcore.c
152
list_add(&ent->list, head);
fs/proc/kcore.c
161
get_sparsemem_vmemmap_info(struct kcore_list *ent, struct list_head *head)
fs/proc/kcore.c
173
list_for_each_entry(tmp, head, list) {
fs/proc/kcore.c
187
list_add_tail(&vmm->list, head);
fs/proc/kcore.c
194
get_sparsemem_vmemmap_info(struct kcore_list *ent, struct list_head *head)
fs/proc/kcore.c
204
struct list_head *head = (struct list_head *)arg;
fs/proc/kcore.c
237
list_add_tail(&ent->list, head);
fs/proc/kcore.c
239
if (!get_sparsemem_vmemmap_info(ent, head)) {
fs/proc/proc_sysctl.c
1061
struct ctl_table_header *head;
fs/proc/proc_sysctl.c
1076
head = NULL;
fs/proc/proc_sysctl.c
1077
entry = find_entry(&head, dir, procname, strlen(procname));
fs/proc/proc_sysctl.c
1079
if (entry && use_table(head)) {
fs/proc/proc_sysctl.c
1081
*phead = head;
fs/proc/proc_sysctl.c
116
struct ctl_table_header *head;
fs/proc/proc_sysctl.c
1180
static struct ctl_table_header *new_links(struct ctl_dir *dir, struct ctl_table_header *head)
fs/proc/proc_sysctl.c
1190
list_for_each_table_entry(entry, head) {
fs/proc/proc_sysctl.c
1195
sizeof(struct ctl_node)*head->ctl_table_size +
fs/proc/proc_sysctl.c
1196
sizeof(struct ctl_table)*head->ctl_table_size +
fs/proc/proc_sysctl.c
1204
link_table = (struct ctl_table *)(node + head->ctl_table_size);
fs/proc/proc_sysctl.c
1205
link_name = (char *)(link_table + head->ctl_table_size);
fs/proc/proc_sysctl.c
1208
list_for_each_table_entry(entry, head) {
fs/proc/proc_sysctl.c
1213
link->data = head->root;
fs/proc/proc_sysctl.c
1218
head->ctl_table_size);
fs/proc/proc_sysctl.c
1219
links->nreg = head->ctl_table_size;
fs/proc/proc_sysctl.c
1257
static int insert_links(struct ctl_table_header *head)
fs/proc/proc_sysctl.c
1264
if (head->set == root_set)
fs/proc/proc_sysctl.c
1267
core_parent = xlate_dir(root_set, head->parent);
fs/proc/proc_sysctl.c
1271
if (get_links(core_parent, head, head->root))
fs/proc/proc_sysctl.c
1277
links = new_links(core_parent, head);
fs/proc/proc_sysctl.c
1285
if (get_links(core_parent, head, head->root)) {
fs/proc/proc_sysctl.c
129
head = ctl_node->header;
fs/proc/proc_sysctl.c
130
entry = &head->ctl_table[ctl_node - head->node];
fs/proc/proc_sysctl.c
139
*phead = head;
fs/proc/proc_sysctl.c
146
static int insert_entry(struct ctl_table_header *head, const struct ctl_table *entry)
fs/proc/proc_sysctl.c
148
struct rb_node *node = &head->node[entry - head->ctl_table].node;
fs/proc/proc_sysctl.c
149
struct rb_node **p = &head->parent->root.rb_node;
fs/proc/proc_sysctl.c
174
sysctl_print_dir(head->parent);
fs/proc/proc_sysctl.c
181
rb_insert_color(node, &head->parent->root);
fs/proc/proc_sysctl.c
185
static void erase_entry(struct ctl_table_header *head, const struct ctl_table *entry)
fs/proc/proc_sysctl.c
187
struct rb_node *node = &head->node[entry - head->ctl_table].node;
fs/proc/proc_sysctl.c
189
rb_erase(node, &head->parent->root);
fs/proc/proc_sysctl.c
192
static void init_header(struct ctl_table_header *head,
fs/proc/proc_sysctl.c
196
head->ctl_table = table;
fs/proc/proc_sysctl.c
197
head->ctl_table_size = table_size;
fs/proc/proc_sysctl.c
198
head->ctl_table_arg = table;
fs/proc/proc_sysctl.c
199
head->used = 0;
fs/proc/proc_sysctl.c
200
head->count = 1;
fs/proc/proc_sysctl.c
201
head->nreg = 1;
fs/proc/proc_sysctl.c
202
head->unregistering = NULL;
fs/proc/proc_sysctl.c
203
head->root = root;
fs/proc/proc_sysctl.c
204
head->set = set;
fs/proc/proc_sysctl.c
205
head->parent = NULL;
fs/proc/proc_sysctl.c
206
head->node = node;
fs/proc/proc_sysctl.c
207
INIT_HLIST_HEAD(&head->inodes);
fs/proc/proc_sysctl.c
211
list_for_each_table_entry(entry, head) {
fs/proc/proc_sysctl.c
212
node->header = head;
fs/proc/proc_sysctl.c
217
sysctl_set_perm_empty_ctl_header(head);
fs/proc/proc_sysctl.c
220
static void erase_header(struct ctl_table_header *head)
fs/proc/proc_sysctl.c
224
list_for_each_table_entry(entry, head)
fs/proc/proc_sysctl.c
225
erase_entry(head, entry);
fs/proc/proc_sysctl.c
287
static void proc_sys_invalidate_dcache(struct ctl_table_header *head)
fs/proc/proc_sysctl.c
289
proc_invalidate_siblings_dcache(&head->inodes, &sysctl_lock);
fs/proc/proc_sysctl.c
325
static struct ctl_table_header *sysctl_head_grab(struct ctl_table_header *head)
fs/proc/proc_sysctl.c
327
BUG_ON(!head);
fs/proc/proc_sysctl.c
329
if (!use_table(head))
fs/proc/proc_sysctl.c
330
head = ERR_PTR(-ENOENT);
fs/proc/proc_sysctl.c
332
return head;
fs/proc/proc_sysctl.c
335
static void sysctl_head_finish(struct ctl_table_header *head)
fs/proc/proc_sysctl.c
337
if (!head)
fs/proc/proc_sysctl.c
340
unuse_table(head);
fs/proc/proc_sysctl.c
357
struct ctl_table_header *head;
fs/proc/proc_sysctl.c
361
entry = find_entry(&head, dir, name, namelen);
fs/proc/proc_sysctl.c
362
if (entry && use_table(head))
fs/proc/proc_sysctl.c
363
*phead = head;
fs/proc/proc_sysctl.c
385
struct ctl_table_header *head = NULL;
fs/proc/proc_sysctl.c
393
head = ctl_node->header;
fs/proc/proc_sysctl.c
394
entry = &head->ctl_table[ctl_node - head->node];
fs/proc/proc_sysctl.c
396
*phead = head;
fs/proc/proc_sysctl.c
402
struct ctl_table_header *head = *phead;
fs/proc/proc_sysctl.c
404
struct ctl_node *ctl_node = &head->node[entry - head->ctl_table];
fs/proc/proc_sysctl.c
407
unuse_table(head);
fs/proc/proc_sysctl.c
411
head = NULL;
fs/proc/proc_sysctl.c
413
head = ctl_node->header;
fs/proc/proc_sysctl.c
414
entry = &head->ctl_table[ctl_node - head->node];
fs/proc/proc_sysctl.c
416
*phead = head;
fs/proc/proc_sysctl.c
436
static int sysctl_perm(struct ctl_table_header *head, const struct ctl_table *table, int op)
fs/proc/proc_sysctl.c
438
struct ctl_table_root *root = head->root;
fs/proc/proc_sysctl.c
442
mode = root->permissions(head, table);
fs/proc/proc_sysctl.c
450
struct ctl_table_header *head, const struct ctl_table *table)
fs/proc/proc_sysctl.c
452
struct ctl_table_root *root = head->root;
fs/proc/proc_sysctl.c
465
if (unlikely(head->unregistering)) {
fs/proc/proc_sysctl.c
470
ei->sysctl = head;
fs/proc/proc_sysctl.c
472
hlist_add_head_rcu(&ei->sibling_inodes, &head->inodes);
fs/proc/proc_sysctl.c
473
head->count++;
fs/proc/proc_sysctl.c
486
if (sysctl_is_perm_empty_ctl_header(head))
fs/proc/proc_sysctl.c
493
root->set_ownership(head, &inode->i_uid, &inode->i_gid);
fs/proc/proc_sysctl.c
498
void proc_sys_evict_inode(struct inode *inode, struct ctl_table_header *head)
fs/proc/proc_sysctl.c
502
if (!--head->count)
fs/proc/proc_sysctl.c
503
kfree_rcu(head, rcu);
fs/proc/proc_sysctl.c
509
struct ctl_table_header *head = PROC_I(inode)->sysctl;
fs/proc/proc_sysctl.c
510
if (!head)
fs/proc/proc_sysctl.c
511
head = &sysctl_table_root.default_set.dir.header;
fs/proc/proc_sysctl.c
512
return sysctl_head_grab(head);
fs/proc/proc_sysctl.c
518
struct ctl_table_header *head = grab_header(dir);
fs/proc/proc_sysctl.c
527
if (IS_ERR(head))
fs/proc/proc_sysctl.c
528
return ERR_CAST(head);
fs/proc/proc_sysctl.c
530
ctl_dir = container_of(head, struct ctl_dir, header);
fs/proc/proc_sysctl.c
543
inode = proc_sys_make_inode(dir->i_sb, h ? h : head, p);
fs/proc/proc_sysctl.c
549
sysctl_head_finish(head);
fs/proc/proc_sysctl.c
557
struct ctl_table_header *head = grab_header(inode);
fs/proc/proc_sysctl.c
563
if (IS_ERR(head))
fs/proc/proc_sysctl.c
564
return PTR_ERR(head);
fs/proc/proc_sysctl.c
571
if (sysctl_perm(head, table, write ? MAY_WRITE : MAY_READ))
fs/proc/proc_sysctl.c
594
error = BPF_CGROUP_RUN_PROG_SYSCTL(head, table, write, &kbuf, &count,
fs/proc/proc_sysctl.c
614
sysctl_head_finish(head);
fs/proc/proc_sysctl.c
631
struct ctl_table_header *head = grab_header(inode);
fs/proc/proc_sysctl.c
635
if (IS_ERR(head))
fs/proc/proc_sysctl.c
636
return PTR_ERR(head);
fs/proc/proc_sysctl.c
641
sysctl_head_finish(head);
fs/proc/proc_sysctl.c
649
struct ctl_table_header *head = grab_header(inode);
fs/proc/proc_sysctl.c
655
if (IS_ERR(head))
fs/proc/proc_sysctl.c
673
sysctl_head_finish(head);
fs/proc/proc_sysctl.c
680
struct ctl_table_header *head,
fs/proc/proc_sysctl.c
701
inode = proc_sys_make_inode(dir->d_sb, head, table);
fs/proc/proc_sysctl.c
724
struct ctl_table_header *head,
fs/proc/proc_sysctl.c
729
head = sysctl_head_grab(head);
fs/proc/proc_sysctl.c
730
if (IS_ERR(head))
fs/proc/proc_sysctl.c
734
if (sysctl_follow_link(&head, &table))
fs/proc/proc_sysctl.c
737
ret = proc_sys_fill_cache(file, ctx, head, table);
fs/proc/proc_sysctl.c
739
sysctl_head_finish(head);
fs/proc/proc_sysctl.c
743
static int scan(struct ctl_table_header *head, const struct ctl_table *table,
fs/proc/proc_sysctl.c
753
res = proc_sys_link_fill_cache(file, ctx, head, table);
fs/proc/proc_sysctl.c
755
res = proc_sys_fill_cache(file, ctx, head, table);
fs/proc/proc_sysctl.c
765
struct ctl_table_header *head = grab_header(file_inode(file));
fs/proc/proc_sysctl.c
771
if (IS_ERR(head))
fs/proc/proc_sysctl.c
772
return PTR_ERR(head);
fs/proc/proc_sysctl.c
774
ctl_dir = container_of(head, struct ctl_dir, header);
fs/proc/proc_sysctl.c
788
sysctl_head_finish(head);
fs/proc/proc_sysctl.c
799
struct ctl_table_header *head;
fs/proc/proc_sysctl.c
807
head = grab_header(inode);
fs/proc/proc_sysctl.c
808
if (IS_ERR(head))
fs/proc/proc_sysctl.c
809
return PTR_ERR(head);
fs/proc/proc_sysctl.c
815
error = sysctl_perm(head, table, mask & ~MAY_NOT_BLOCK);
fs/proc/proc_sysctl.c
817
sysctl_head_finish(head);
fs/proc/proc_sysctl.c
843
struct ctl_table_header *head = grab_header(inode);
fs/proc/proc_sysctl.c
846
if (IS_ERR(head))
fs/proc/proc_sysctl.c
847
return PTR_ERR(head);
fs/proc/proc_sysctl.c
853
sysctl_head_finish(head);
fs/proc/proc_sysctl.c
917
struct ctl_table_header *head;
fs/proc/proc_sysctl.c
93
static int insert_links(struct ctl_table_header *head);
fs/proc/proc_sysctl.c
934
head = READ_ONCE(PROC_I(inode)->sysctl);
fs/proc/proc_sysctl.c
935
return !head || !sysctl_is_seen(head);
fs/proc/proc_sysctl.c
947
struct ctl_table_header *head;
fs/proc/proc_sysctl.c
950
entry = find_entry(&head, dir, name, namelen);
fs/proc/proc_sysctl.c
955
return container_of(head, struct ctl_dir, header);
fs/quota/dquot.c
292
struct hlist_head *head;
fs/quota/dquot.c
293
head = dquot_hash + hashfn(dquot->dq_sb, dquot->dq_id);
fs/quota/dquot.c
294
hlist_add_head(&dquot->dq_hash, head);
fs/resctrl/monitor.c
594
struct list_head *head;
fs/resctrl/monitor.c
606
head = &rdtgrp->mon.crdtgrp_list;
fs/resctrl/monitor.c
609
list_for_each_entry(entry, head, mon.crdtgrp_list) {
fs/resctrl/monitor.c
679
struct list_head *head;
fs/resctrl/monitor.c
707
head = &rgrp->mon.crdtgrp_list;
fs/resctrl/monitor.c
708
list_for_each_entry(entry, head, mon.crdtgrp_list) {
fs/resctrl/monitor.c
839
struct list_head *head;
fs/resctrl/monitor.c
858
head = &prgrp->mon.crdtgrp_list;
fs/resctrl/monitor.c
859
list_for_each_entry(crgrp, head, mon.crdtgrp_list)
fs/resctrl/rdtgroup.c
3027
struct list_head *head;
fs/resctrl/rdtgroup.c
3029
head = &rdtgrp->mon.crdtgrp_list;
fs/resctrl/rdtgroup.c
3030
list_for_each_entry_safe(sentry, stmp, head, mon.crdtgrp_list) {
fs/resctrl/rdtgroup.c
3400
struct list_head *head;
fs/resctrl/rdtgroup.c
3406
head = &prgrp->mon.crdtgrp_list;
fs/resctrl/rdtgroup.c
3407
list_for_each_entry(crgrp, head, mon.crdtgrp_list) {
fs/resctrl/rdtgroup.c
407
struct list_head *head;
fs/resctrl/rdtgroup.c
430
head = &prgrp->mon.crdtgrp_list;
fs/resctrl/rdtgroup.c
431
list_for_each_entry(crgrp, head, mon.crdtgrp_list) {
fs/resctrl/rdtgroup.c
460
struct list_head *head;
fs/resctrl/rdtgroup.c
501
head = &rdtgrp->mon.crdtgrp_list;
fs/resctrl/rdtgroup.c
502
list_for_each_entry(crgrp, head, mon.crdtgrp_list) {
fs/select.c
1004
fdcount = do_poll(head, &table, end_time);
fs/select.c
1010
for (walk = head; walk; walk = walk->next) {
fs/select.c
1021
walk = head->next;
fs/select.c
972
struct poll_list *const head = (struct poll_list *)stack_pps;
fs/select.c
973
struct poll_list *walk = head;
fs/seq_file.c
1003
struct hlist_node *seq_hlist_next(void *v, struct hlist_head *head,
fs/seq_file.c
1010
return head->first;
fs/seq_file.c
1027
struct hlist_node *seq_hlist_start_rcu(struct hlist_head *head,
fs/seq_file.c
1032
__hlist_for_each_rcu(node, head)
fs/seq_file.c
1051
struct hlist_node *seq_hlist_start_head_rcu(struct hlist_head *head,
fs/seq_file.c
1057
return seq_hlist_start_rcu(head, pos - 1);
fs/seq_file.c
1074
struct hlist_head *head,
fs/seq_file.c
1081
return rcu_dereference(head->first);
fs/seq_file.c
1096
seq_hlist_start_percpu(struct hlist_head __percpu *head, int *cpu, loff_t pos)
fs/seq_file.c
1101
hlist_for_each(node, per_cpu_ptr(head, *cpu)) {
fs/seq_file.c
1120
seq_hlist_next_percpu(void *v, struct hlist_head __percpu *head,
fs/seq_file.c
1132
struct hlist_head *bucket = per_cpu_ptr(head, *cpu);
fs/seq_file.c
897
struct list_head *seq_list_start(struct list_head *head, loff_t pos)
fs/seq_file.c
901
list_for_each(lh, head)
fs/seq_file.c
909
struct list_head *seq_list_start_head(struct list_head *head, loff_t pos)
fs/seq_file.c
912
return head;
fs/seq_file.c
914
return seq_list_start(head, pos - 1);
fs/seq_file.c
918
struct list_head *seq_list_next(void *v, struct list_head *head, loff_t *ppos)
fs/seq_file.c
924
return lh == head ? NULL : lh;
fs/seq_file.c
928
struct list_head *seq_list_start_rcu(struct list_head *head, loff_t pos)
fs/seq_file.c
932
list_for_each_rcu(lh, head)
fs/seq_file.c
940
struct list_head *seq_list_start_head_rcu(struct list_head *head, loff_t pos)
fs/seq_file.c
943
return head;
fs/seq_file.c
945
return seq_list_start_rcu(head, pos - 1);
fs/seq_file.c
949
struct list_head *seq_list_next_rcu(void *v, struct list_head *head,
fs/seq_file.c
956
return lh == head ? NULL : lh;
fs/seq_file.c
967
struct hlist_node *seq_hlist_start(struct hlist_head *head, loff_t pos)
fs/seq_file.c
971
hlist_for_each(node, head)
fs/seq_file.c
986
struct hlist_node *seq_hlist_start_head(struct hlist_head *head, loff_t pos)
fs/seq_file.c
991
return seq_hlist_start(head, pos - 1);
fs/smb/client/dfs.h
182
static inline void dfs_put_root_smb_sessions(struct list_head *head)
fs/smb/client/dfs.h
186
list_for_each_entry_safe(ses, n, head, dlist) {
fs/smb/client/dfs_cache.c
791
struct list_head *head = &tl->tl_list;
fs/smb/client/dfs_cache.c
796
INIT_LIST_HEAD(head);
fs/smb/client/dfs_cache.c
814
list_add(&it->it_list, head);
fs/smb/client/dfs_cache.c
816
list_add_tail(&it->it_list, head);
fs/smb/client/dfs_cache.c
824
list_for_each_entry_safe(it, nit, head, it_list) {
fs/smb/server/vfs_cache.c
171
struct hlist_head *head = inode_hashtable +
fs/smb/server/vfs_cache.c
175
hlist_for_each_entry(ci, head, m_hash) {
fs/smb/server/vfs_cache.c
876
static void ksmbd_scavenger_dispose_dh(struct list_head *head)
fs/smb/server/vfs_cache.c
878
while (!list_empty(head)) {
fs/smb/server/vfs_cache.c
881
fp = list_first_entry(head, struct ksmbd_file, node);
fs/splice.c
1120
pipe->tail = pipe->head = 0;
fs/splice.c
1744
o_head = opipe->head;
fs/splice.c
1756
i_head = ipipe->head;
fs/splice.c
1801
opipe->head = o_head;
fs/splice.c
1826
opipe->head = o_head;
fs/splice.c
1867
o_head = opipe->head;
fs/splice.c
1877
i_head = ipipe->head;
fs/splice.c
1916
opipe->head = o_head;
fs/splice.c
202
unsigned int head = pipe->head;
fs/splice.c
215
while (!pipe_full(head, tail, pipe->max_usage)) {
fs/splice.c
216
struct pipe_buffer *buf = pipe_buf(pipe, head);
fs/splice.c
225
head++;
fs/splice.c
226
pipe->head = head;
fs/splice.c
247
unsigned int head = pipe->head;
fs/splice.c
254
} else if (pipe_full(head, tail, pipe->max_usage)) {
fs/splice.c
257
*pipe_buf(pipe, head) = *buf;
fs/splice.c
258
pipe->head = head + 1;
fs/splice.c
392
pipe->head++;
fs/splice.c
445
unsigned int head = pipe->head;
fs/splice.c
449
while (!pipe_empty(head, tail)) {
fs/splice.c
688
unsigned int head, tail;
fs/splice.c
706
head = pipe->head;
fs/splice.c
711
for (n = 0; !pipe_empty(head, tail) && left && n < nbufs; tail++) {
fs/splice.c
808
unsigned int head, tail, bc = 0;
fs/splice.c
843
head = pipe->head;
fs/splice.c
846
while (!pipe_empty(head, tail)) {
fs/splice.c
877
if (remain && pipe_occupancy(pipe->head, tail) > 0)
fs/super.c
288
static void destroy_super_rcu(struct rcu_head *head)
fs/super.c
290
struct super_block *s = container_of(head, struct super_block, rcu);
fs/ubifs/debug.c
2328
int dbg_check_data_nodes_order(struct ubifs_info *c, struct list_head *head)
fs/ubifs/debug.c
2336
for (cur = head->next; cur->next != head; cur = cur->next) {
fs/ubifs/debug.c
2395
int dbg_check_nondata_nodes_order(struct ubifs_info *c, struct list_head *head)
fs/ubifs/debug.c
2403
for (cur = head->next; cur->next != head; cur = cur->next) {
fs/ubifs/debug.c
748
int head = 0;
fs/ubifs/debug.c
758
head = 1;
fs/ubifs/debug.c
761
if (!head)
fs/ubifs/debug.h
290
int dbg_check_data_nodes_order(struct ubifs_info *c, struct list_head *head);
fs/ubifs/debug.h
291
int dbg_check_nondata_nodes_order(struct ubifs_info *c, struct list_head *head);
fs/ubifs/scan.c
357
struct list_head *head;
fs/ubifs/scan.c
359
head = &sleb->nodes;
fs/ubifs/scan.c
360
while (!list_empty(head)) {
fs/ubifs/scan.c
361
node = list_entry(head->next, struct ubifs_scan_node, list);
fs/ufs/balloc.c
246
struct buffer_head *head, *bh;
fs/ufs/balloc.c
273
head = folio_buffers(folio);
fs/ufs/balloc.c
274
bh = head;
fs/ufs/balloc.c
306
} while (bh != head);
fs/userfaultfd.c
2088
list_for_each_entry(wq, &ctx->fault_pending_wqh.head, entry) {
fs/userfaultfd.c
2092
list_for_each_entry(wq, &ctx->fault_wqh.head, entry) {
fs/userfaultfd.c
920
wq = list_last_entry(&wqh->head, typeof(*wq), entry);
fs/vboxsf/dir.c
107
list_for_each_entry(b, &sf_d->info_list, head) {
fs/vboxsf/file.c
135
list_del(&sf_handle->head);
fs/vboxsf/file.c
20
struct list_head head;
fs/vboxsf/file.c
251
list_for_each_entry(h, &sf_i->handle_list, head) {
fs/vboxsf/file.c
43
list_add(&sf_handle->head, &sf_i->handle_list);
fs/vboxsf/utils.c
494
list_add(&b->head, list);
fs/vboxsf/utils.c
501
list_del(&b->head);
fs/vboxsf/utils.c
526
b = list_entry(pos, struct vboxsf_dir_buf, head);
fs/vboxsf/vfsmod.h
73
struct list_head head;
fs/xfs/scrub/scrub.c
1002
copy_to_user(uhead, &head, sizeof(head))) {
fs/xfs/scrub/scrub.c
873
const struct xfs_scrub_vec_head *head)
fs/xfs/scrub/scrub.c
880
error = xfs_iget(mp, tp, head->svh_ino, XCHK_IGET_FLAGS, 0, &ip);
fs/xfs/scrub/scrub.c
885
if (VFS_I(ip)->i_generation != head->svh_gen) {
fs/xfs/scrub/scrub.c
899
struct xfs_scrub_vec_head head;
fs/xfs/scrub/scrub.c
914
if (copy_from_user(&head, uhead, sizeof(head)))
fs/xfs/scrub/scrub.c
917
if (head.svh_reserved)
fs/xfs/scrub/scrub.c
919
if (head.svh_flags & ~XFS_SCRUB_VEC_FLAGS_ALL)
fs/xfs/scrub/scrub.c
921
if (head.svh_nr == 0)
fs/xfs/scrub/scrub.c
924
vec_bytes = array_size(head.svh_nr, sizeof(struct xfs_scrub_vec));
fs/xfs/scrub/scrub.c
928
uvectors = u64_to_user_ptr(head.svh_vectors);
fs/xfs/scrub/scrub.c
933
trace_xchk_scrubv_start(ip_in, &head);
fs/xfs/scrub/scrub.c
935
for (i = 0, v = vectors; i < head.svh_nr; i++, v++) {
fs/xfs/scrub/scrub.c
947
trace_xchk_scrubv_item(mp, &head, i, v);
fs/xfs/scrub/scrub.c
958
if (head.svh_ino && head.svh_ino != ip_in->i_ino)
fs/xfs/scrub/scrub.c
959
handle_ip = xchk_scrubv_open_by_handle(mp, &head);
fs/xfs/scrub/scrub.c
962
for (i = 0, v = vectors; i < head.svh_nr; i++, v++) {
fs/xfs/scrub/scrub.c
966
.sm_ino = head.svh_ino,
fs/xfs/scrub/scrub.c
967
.sm_gen = head.svh_gen,
fs/xfs/scrub/scrub.c
968
.sm_agno = head.svh_agno,
fs/xfs/scrub/scrub.c
974
trace_xchk_scrubv_barrier_fail(mp, &head, i, v);
fs/xfs/scrub/scrub.c
984
trace_xchk_scrubv_outcome(mp, &head, i, v);
fs/xfs/scrub/scrub.c
986
if (head.svh_rest_us) {
fs/xfs/scrub/scrub.c
990
head.svh_rest_us * 1000);
fs/xfs/xfs_fsmap.c
1155
struct xfs_fsmap_head *head,
fs/xfs/xfs_fsmap.c
1163
.head = head,
fs/xfs/xfs_fsmap.c
1169
if (head->fmh_iflags & ~FMH_IF_VALID)
fs/xfs/xfs_fsmap.c
1171
if (!xfs_getfsmap_is_valid_device(mp, &head->fmh_keys[0]) ||
fs/xfs/xfs_fsmap.c
1172
!xfs_getfsmap_is_valid_device(mp, &head->fmh_keys[1]))
fs/xfs/xfs_fsmap.c
1174
if (!xfs_getfsmap_check_keys(&head->fmh_keys[0], &head->fmh_keys[1]))
fs/xfs/xfs_fsmap.c
1179
head->fmh_entries = 0;
fs/xfs/xfs_fsmap.c
1231
dkeys[0] = head->fmh_keys[0];
fs/xfs/xfs_fsmap.c
1234
info.next_daddr = head->fmh_keys[0].fmr_physical +
fs/xfs/xfs_fsmap.c
1235
head->fmh_keys[0].fmr_length;
fs/xfs/xfs_fsmap.c
1242
if (head->fmh_keys[0].fmr_device > handlers[i].dev)
fs/xfs/xfs_fsmap.c
1244
if (head->fmh_keys[1].fmr_device < handlers[i].dev)
fs/xfs/xfs_fsmap.c
1253
if (handlers[i].dev == head->fmh_keys[1].fmr_device) {
fs/xfs/xfs_fsmap.c
1254
dkeys[1] = head->fmh_keys[1];
fs/xfs/xfs_fsmap.c
1265
if (handlers[i].dev > head->fmh_keys[0].fmr_device)
fs/xfs/xfs_fsmap.c
1296
head->fmh_oflags = FMH_OF_DEV_T;
fs/xfs/xfs_fsmap.c
1306
struct fsmap_head head;
fs/xfs/xfs_fsmap.c
1313
if (copy_from_user(&head, arg, sizeof(struct fsmap_head)))
fs/xfs/xfs_fsmap.c
1315
if (memchr_inv(head.fmh_reserved, 0, sizeof(head.fmh_reserved)) ||
fs/xfs/xfs_fsmap.c
1316
memchr_inv(head.fmh_keys[0].fmr_reserved, 0,
fs/xfs/xfs_fsmap.c
1317
sizeof(head.fmh_keys[0].fmr_reserved)) ||
fs/xfs/xfs_fsmap.c
1318
memchr_inv(head.fmh_keys[1].fmr_reserved, 0,
fs/xfs/xfs_fsmap.c
1319
sizeof(head.fmh_keys[1].fmr_reserved)))
fs/xfs/xfs_fsmap.c
1327
count = min_t(unsigned int, head.fmh_count,
fs/xfs/xfs_fsmap.c
1331
count = min_t(unsigned int, head.fmh_count,
fs/xfs/xfs_fsmap.c
1338
xhead.fmh_iflags = head.fmh_iflags;
fs/xfs/xfs_fsmap.c
1339
xfs_fsmap_to_internal(&xhead.fmh_keys[0], &head.fmh_keys[0]);
fs/xfs/xfs_fsmap.c
1340
xfs_fsmap_to_internal(&xhead.fmh_keys[1], &head.fmh_keys[1]);
fs/xfs/xfs_fsmap.c
1345
head.fmh_entries = 0;
fs/xfs/xfs_fsmap.c
1350
user_recs = &arg->fmh_recs[head.fmh_entries];
fs/xfs/xfs_fsmap.c
1353
head.fmh_count - head.fmh_entries);
fs/xfs/xfs_fsmap.c
1376
head.fmh_entries += xhead.fmh_entries;
fs/xfs/xfs_fsmap.c
1377
head.fmh_oflags = xhead.fmh_oflags;
fs/xfs/xfs_fsmap.c
1383
if (head.fmh_count == 0 || xhead.fmh_entries == 0)
fs/xfs/xfs_fsmap.c
1400
} while (!done && head.fmh_entries < head.fmh_count);
fs/xfs/xfs_fsmap.c
1406
if (done && head.fmh_count > 0 && head.fmh_entries > 0) {
fs/xfs/xfs_fsmap.c
1410
user_rec = &arg->fmh_recs[head.fmh_entries - 1];
fs/xfs/xfs_fsmap.c
1420
if (copy_to_user(arg, &head, sizeof(struct fsmap_head))) {
fs/xfs/xfs_fsmap.c
161
struct xfs_fsmap_head *head;
fs/xfs/xfs_fsmap.c
252
rec = &info->fsmap_recs[info->head->fmh_entries++];
fs/xfs/xfs_fsmap.c
303
if (info->head->fmh_count == 0) {
fs/xfs/xfs_fsmap.c
304
if (info->head->fmh_entries == UINT_MAX)
fs/xfs/xfs_fsmap.c
308
info->head->fmh_entries++;
fs/xfs/xfs_fsmap.c
313
info->head->fmh_entries++;
fs/xfs/xfs_fsmap.c
323
if (info->head->fmh_entries >= info->head->fmh_count)
fs/xfs/xfs_fsmap.c
339
if (info->head->fmh_entries >= info->head->fmh_count)
fs/xfs/xfs_hooks.c
20
BLOCKING_INIT_NOTIFIER_HEAD(&chain->head);
fs/xfs/xfs_hooks.c
32
return blocking_notifier_chain_register(&chain->head, &hook->nb);
fs/xfs/xfs_hooks.c
41
blocking_notifier_chain_unregister(&chain->head, &hook->nb);
fs/xfs/xfs_hooks.c
51
return blocking_notifier_call_chain(&chain->head, val, priv);
fs/xfs/xfs_hooks.h
11
struct blocking_notifier_head head;
fs/xfs/xfs_icache.c
139
struct rcu_head *head)
fs/xfs/xfs_icache.c
141
struct inode *inode = container_of(head, struct inode, i_rcu);
fs/xfs/xfs_log.c
101
struct xlog_grant_head *head)
fs/xfs/xfs_log.c
103
atomic64_set(&head->grant, 0);
fs/xfs/xfs_log.c
104
INIT_LIST_HEAD(&head->waiters);
fs/xfs/xfs_log.c
105
spin_lock_init(&head->lock);
fs/xfs/xfs_log.c
131
struct xlog_grant_head *head)
fs/xfs/xfs_log.c
137
atomic64_read(&head->grant);
fs/xfs/xfs_log.c
145
struct xlog_grant_head *head)
fs/xfs/xfs_log.c
149
spin_lock(&head->lock);
fs/xfs/xfs_log.c
150
list_for_each_entry(tic, &head->waiters, t_queue)
fs/xfs/xfs_log.c
152
spin_unlock(&head->lock);
fs/xfs/xfs_log.c
158
struct xlog_grant_head *head,
fs/xfs/xfs_log.c
161
if (head == &log->l_write_head) {
fs/xfs/xfs_log.c
175
struct xlog_grant_head *head,
fs/xfs/xfs_log.c
181
list_for_each_entry(tic, &head->waiters, t_queue) {
fs/xfs/xfs_log.c
182
need_bytes = xlog_ticket_reservation(log, head, tic);
fs/xfs/xfs_log.c
197
struct xlog_grant_head *head,
fs/xfs/xfs_log.c
199
int need_bytes) __releases(&head->lock)
fs/xfs/xfs_log.c
200
__acquires(&head->lock)
fs/xfs/xfs_log.c
202
list_add_tail(&tic->t_queue, &head->waiters);
fs/xfs/xfs_log.c
209
spin_unlock(&head->lock);
fs/xfs/xfs_log.c
220
spin_lock(&head->lock);
fs/xfs/xfs_log.c
223
} while (xlog_grant_space_left(log, head) < need_bytes);
fs/xfs/xfs_log.c
2518
struct xlog_rec_header *head;
fs/xfs/xfs_log.c
252
struct xlog_grant_head *head,
fs/xfs/xfs_log.c
2537
head = iclog->ic_header;
fs/xfs/xfs_log.c
2551
head->h_cycle = cpu_to_be32(log->l_curr_cycle);
fs/xfs/xfs_log.c
2552
head->h_lsn = cpu_to_be64(
fs/xfs/xfs_log.c
267
*need_bytes = xlog_ticket_reservation(log, head, tic);
fs/xfs/xfs_log.c
268
free_bytes = xlog_grant_space_left(log, head);
fs/xfs/xfs_log.c
269
if (!list_empty_careful(&head->waiters)) {
fs/xfs/xfs_log.c
270
spin_lock(&head->lock);
fs/xfs/xfs_log.c
271
if (!xlog_grant_head_wake(log, head, &free_bytes) ||
fs/xfs/xfs_log.c
273
error = xlog_grant_head_wait(log, head, tic,
fs/xfs/xfs_log.c
276
spin_unlock(&head->lock);
fs/xfs/xfs_log.c
278
spin_lock(&head->lock);
fs/xfs/xfs_log.c
279
error = xlog_grant_head_wait(log, head, tic, *need_bytes);
fs/xfs/xfs_log.c
280
spin_unlock(&head->lock);
fs/xfs/xfs_log.c
85
struct xlog_grant_head *head,
fs/xfs/xfs_log.c
88
atomic64_sub(bytes, &head->grant);
fs/xfs/xfs_log.c
93
struct xlog_grant_head *head,
fs/xfs/xfs_log.c
96
atomic64_add(bytes, &head->grant);
fs/xfs/xfs_log_recover.c
194
struct xlog_rec_header *head)
fs/xfs/xfs_log_recover.c
199
&head->h_fs_uuid, be32_to_cpu(head->h_fmt));
fs/xfs/xfs_log_recover.c
202
#define xlog_header_check_dump(mp, head)
fs/xfs/xfs_log_recover.c
2091
struct list_head *head)
fs/xfs/xfs_log_recover.c
2097
list_add_tail(&item->ri_list, head);
fs/xfs/xfs_log_recover.c
211
struct xlog_rec_header *head)
fs/xfs/xfs_log_recover.c
213
ASSERT(head->h_magicno == cpu_to_be32(XLOG_HEADER_MAGIC_NUM));
fs/xfs/xfs_log_recover.c
220
if (XFS_IS_CORRUPT(mp, head->h_fmt != cpu_to_be32(XLOG_FMT))) {
fs/xfs/xfs_log_recover.c
223
xlog_header_check_dump(mp, head);
fs/xfs/xfs_log_recover.c
227
&head->h_fs_uuid))) {
fs/xfs/xfs_log_recover.c
230
xlog_header_check_dump(mp, head);
fs/xfs/xfs_log_recover.c
242
struct xlog_rec_header *head)
fs/xfs/xfs_log_recover.c
244
ASSERT(head->h_magicno == cpu_to_be32(XLOG_HEADER_MAGIC_NUM));
fs/xfs/xfs_log_recover.c
246
if (uuid_is_null(&head->h_fs_uuid)) {
fs/xfs/xfs_log_recover.c
254
&head->h_fs_uuid))) {
fs/xfs/xfs_log_recover.c
256
xlog_header_check_dump(mp, head);
fs/xfs/xfs_log_recover.c
403
struct xlog_rec_header *head = NULL;
fs/xfs/xfs_log_recover.c
440
head = (struct xlog_rec_header *)offset;
fs/xfs/xfs_log_recover.c
442
if (head->h_magicno == cpu_to_be32(XLOG_HEADER_MAGIC_NUM))
fs/xfs/xfs_log_recover.c
463
if ((error = xlog_header_check_mount(log->l_mp, head)))
fs/xfs/xfs_log_recover.c
473
xhdrs = xlog_logrec_hblks(log, head);
fs/xfs/xfs_log_recover.c
476
BTOBB(be32_to_cpu(head->h_len)) + xhdrs)
include/asm-generic/vmlinux.lds.h
666
#define HEAD_TEXT KEEP(*(.head.text))
include/asm-generic/vmlinux.lds.h
669
.head.text : AT(ADDR(.head.text) - LOAD_OFFSET) { \
include/crypto/internal/aead.h
22
char head[offsetof(struct aead_alg, base)];
include/crypto/internal/akcipher.h
17
char head[offsetof(struct akcipher_alg, base)];
include/crypto/internal/hash.h
52
char head[offsetof(struct ahash_alg, halg.base)];
include/crypto/internal/hash.h
63
char head[offsetof(struct shash_alg, base)];
include/crypto/internal/kpp.h
24
char head[offsetof(struct kpp_alg, base)];
include/crypto/internal/sig.h
17
char head[offsetof(struct sig_alg, base)];
include/crypto/internal/skcipher.h
31
char head[offsetof(struct skcipher_alg, base)];
include/crypto/internal/skcipher.h
42
char head[offsetof(struct lskcipher_alg, co.base)];
include/crypto/scatterwalk.h
20
static inline void scatterwalk_crypto_chain(struct scatterlist *head,
include/crypto/scatterwalk.h
24
sg_chain(head, num, sg);
include/crypto/scatterwalk.h
26
sg_mark_end(head);
include/drm/drm_atomic.h
350
struct list_head head;
include/drm/drm_atomic.h
381
list_for_each_entry(privobj, &(dev)->mode_config.privobj_list, head)
include/drm/drm_colorop.h
212
struct list_head head;
include/drm/drm_colorop.h
436
list_for_each_entry(colorop, &(dev)->mode_config.colorop_list, head)
include/drm/drm_connector.h
1962
struct list_head head;
include/drm/drm_crtc.h
1329
list_for_each_entry(crtc, &(dev)->mode_config.crtc_list, head)
include/drm/drm_crtc.h
1339
list_for_each_entry_reverse(crtc, &(dev)->mode_config.crtc_list, head)
include/drm/drm_crtc.h
955
struct list_head head;
include/drm/drm_debugfs_crc.h
64
int head, tail;
include/drm/drm_encoder.h
107
struct list_head head;
include/drm/drm_encoder.h
333
list_for_each_entry((encoder), &(dev)->mode_config.encoder_list, head) \
include/drm/drm_encoder.h
344
list_for_each_entry(encoder, &(dev)->mode_config.encoder_list, head)
include/drm/drm_framebuffer.h
129
struct list_head head;
include/drm/drm_framebuffer.h
286
struct drm_framebuffer, head); \
include/drm/drm_framebuffer.h
287
&fb->head != (&(dev)->mode_config.fb_list); \
include/drm/drm_framebuffer.h
288
fb = list_next_entry(fb, head))
include/drm/drm_modes.h
396
struct list_head head;
include/drm/drm_modeset_lock.h
114
WARN_ON(!list_empty(&lock->head));
include/drm/drm_modeset_lock.h
95
struct list_head head;
include/drm/drm_plane.h
1006
list_for_each_entry(plane, &(dev)->mode_config.plane_list, head)
include/drm/drm_plane.h
652
struct list_head head;
include/drm/drm_plane.h
982
list_for_each_entry((plane), &(dev)->mode_config.plane_list, head) \
include/drm/drm_plane.h
995
list_for_each_entry(plane, &(dev)->mode_config.plane_list, head) \
include/drm/drm_property.h
50
struct list_head head;
include/drm/drm_property.h
84
struct list_head head;
include/drm/spsc_queue.h
104
WRITE_ONCE(queue->head, next);
include/drm/spsc_queue.h
110
(long)&node->next, (long) &queue->head) != (long)&node->next) {
include/drm/spsc_queue.h
114
} while (unlikely(!(queue->head = READ_ONCE(node->next))));
include/drm/spsc_queue.h
40
struct spsc_node *head;
include/drm/spsc_queue.h
50
queue->head = NULL;
include/drm/spsc_queue.h
51
atomic_long_set(&queue->tail, (long)&queue->head);
include/drm/spsc_queue.h
57
return queue->head;
include/drm/spsc_queue.h
87
return tail == &queue->head;
include/drm/spsc_queue.h
98
node = READ_ONCE(queue->head);
include/drm/ttm/ttm_execbuf_util.h
49
struct list_head head;
include/drm/ttm/ttm_resource.h
491
ttm_lru_first_res_or_null(struct list_head *head);
include/linux/acpi_iort.h
41
struct list_head *head);
include/linux/acpi_iort.h
43
struct list_head *head);
include/linux/acpi_iort.h
47
void iort_iommu_get_resv_regions(struct device *dev, struct list_head *head);
include/linux/acpi_iort.h
63
void iort_get_rmr_sids(struct fwnode_handle *iommu_fwnode, struct list_head *head) { }
include/linux/acpi_iort.h
65
void iort_put_rmr_sids(struct fwnode_handle *iommu_fwnode, struct list_head *head) { }
include/linux/acpi_iort.h
72
void iort_iommu_get_resv_regions(struct device *dev, struct list_head *head)
include/linux/bio.h
537
struct bio *head;
include/linux/bio.h
543
return bl->head == NULL;
include/linux/bio.h
548
bl->head = bl->tail = NULL;
include/linux/bio.h
554
for (bio = (bl)->head; bio; bio = bio->bi_next)
include/linux/bio.h
574
bl->head = bio;
include/linux/bio.h
581
bio->bi_next = bl->head;
include/linux/bio.h
583
bl->head = bio;
include/linux/bio.h
591
if (!bl2->head)
include/linux/bio.h
595
bl->tail->bi_next = bl2->head;
include/linux/bio.h
597
bl->head = bl2->head;
include/linux/bio.h
612
if (!bl2->head)
include/linux/bio.h
615
if (bl->head)
include/linux/bio.h
616
bl2->tail->bi_next = bl->head;
include/linux/bio.h
620
bl->head = bl2->head;
include/linux/bio.h
625
return bl->head;
include/linux/bio.h
630
struct bio *bio = bl->head;
include/linux/bio.h
633
bl->head = bl->head->bi_next;
include/linux/bio.h
634
if (!bl->head)
include/linux/bio.h
645
struct bio *bio = bl->head;
include/linux/bio.h
647
bl->head = bl->tail = NULL;
include/linux/blk-mq.h
253
return rl->head == NULL;
include/linux/blk-mq.h
258
rl->head = NULL;
include/linux/blk-mq.h
268
rl->head = rq;
include/linux/blk-mq.h
275
rl->head = rq;
include/linux/blk-mq.h
285
rl->head = rl->head->rq_next;
include/linux/blk-mq.h
286
if (!rl->head)
include/linux/blkdev.h
1156
struct request *head;
include/linux/bpf-cgroup.h
136
int __cgroup_bpf_run_filter_sysctl(struct ctl_table_header *head,
include/linux/bpf-cgroup.h
369
#define BPF_CGROUP_RUN_PROG_SYSCTL(head, table, write, buf, count, pos) \
include/linux/bpf-cgroup.h
373
__ret = __cgroup_bpf_run_filter_sysctl(head, table, write, \
include/linux/bpf-cgroup.h
505
#define BPF_CGROUP_RUN_PROG_SYSCTL(head,table,write,buf,count,pos) ({ 0; })
include/linux/bpf_verifier.h
754
struct bpf_verifier_stack_elem *head; /* stack of verifier states to be processed */
include/linux/btree-128.h
102
return btree_grim_visitor(&head->h, &btree_geo128, opaque,
include/linux/btree-128.h
106
#define btree_for_each_safe128(head, k1, k2, val) \
include/linux/btree-128.h
107
for (val = btree_last128(head, &k1, &k2); \
include/linux/btree-128.h
109
val = btree_get_prev128(head, &k1, &k2))
include/linux/btree-128.h
12
static inline int btree_init128(struct btree_head128 *head)
include/linux/btree-128.h
14
return btree_init(&head->h);
include/linux/btree-128.h
17
static inline void btree_destroy128(struct btree_head128 *head)
include/linux/btree-128.h
19
btree_destroy(&head->h);
include/linux/btree-128.h
22
static inline void *btree_lookup128(struct btree_head128 *head, u64 k1, u64 k2)
include/linux/btree-128.h
25
return btree_lookup(&head->h, &btree_geo128, (unsigned long *)&key);
include/linux/btree-128.h
28
static inline void *btree_get_prev128(struct btree_head128 *head,
include/linux/btree-128.h
34
val = btree_get_prev(&head->h, &btree_geo128,
include/linux/btree-128.h
41
static inline int btree_insert128(struct btree_head128 *head, u64 k1, u64 k2,
include/linux/btree-128.h
45
return btree_insert(&head->h, &btree_geo128,
include/linux/btree-128.h
49
static inline int btree_update128(struct btree_head128 *head, u64 k1, u64 k2,
include/linux/btree-128.h
53
return btree_update(&head->h, &btree_geo128,
include/linux/btree-128.h
57
static inline void *btree_remove128(struct btree_head128 *head, u64 k1, u64 k2)
include/linux/btree-128.h
6
static inline void btree_init_mempool128(struct btree_head128 *head,
include/linux/btree-128.h
60
return btree_remove(&head->h, &btree_geo128, (unsigned long *)&key);
include/linux/btree-128.h
63
static inline void *btree_last128(struct btree_head128 *head, u64 *k1, u64 *k2)
include/linux/btree-128.h
68
val = btree_last(&head->h, &btree_geo128, (unsigned long *)&key[0]);
include/linux/btree-128.h
9
btree_init_mempool(&head->h, mempool);
include/linux/btree-128.h
90
static inline size_t btree_visitor128(struct btree_head128 *head,
include/linux/btree-128.h
94
return btree_visitor(&head->h, &btree_geo128, opaque,
include/linux/btree-128.h
98
static inline size_t btree_grim_visitor128(struct btree_head128 *head,
include/linux/btree-type.h
100
static inline void *BTREE_FN(remove)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE key)
include/linux/btree-type.h
102
return btree_remove(&head->h, BTREE_TYPE_GEO, (unsigned long *)&key);
include/linux/btree-type.h
105
static inline void *BTREE_FN(last)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE *key)
include/linux/btree-type.h
107
return btree_last(&head->h, BTREE_TYPE_GEO, (unsigned long *)key);
include/linux/btree-type.h
110
static inline void *BTREE_FN(get_prev)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE *key)
include/linux/btree-type.h
112
return btree_get_prev(&head->h, BTREE_TYPE_GEO, (unsigned long *)key);
include/linux/btree-type.h
122
static inline size_t BTREE_FN(visitor)(BTREE_TYPE_HEAD *head,
include/linux/btree-type.h
126
return btree_visitor(&head->h, BTREE_TYPE_GEO, opaque,
include/linux/btree-type.h
130
static inline size_t BTREE_FN(grim_visitor)(BTREE_TYPE_HEAD *head,
include/linux/btree-type.h
134
return btree_grim_visitor(&head->h, BTREE_TYPE_GEO, opaque,
include/linux/btree-type.h
14
static inline void BTREE_FN(init_mempool)(BTREE_TYPE_HEAD *head,
include/linux/btree-type.h
17
btree_init_mempool(&head->h, mempool);
include/linux/btree-type.h
20
static inline int BTREE_FN(init)(BTREE_TYPE_HEAD *head)
include/linux/btree-type.h
22
return btree_init(&head->h);
include/linux/btree-type.h
25
static inline void BTREE_FN(destroy)(BTREE_TYPE_HEAD *head)
include/linux/btree-type.h
27
btree_destroy(&head->h);
include/linux/btree-type.h
38
static inline void *BTREE_FN(lookup)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE key)
include/linux/btree-type.h
41
return btree_lookup(&head->h, BTREE_TYPE_GEO, &_key);
include/linux/btree-type.h
44
static inline int BTREE_FN(insert)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE key,
include/linux/btree-type.h
48
return btree_insert(&head->h, BTREE_TYPE_GEO, &_key, val, gfp);
include/linux/btree-type.h
51
static inline int BTREE_FN(update)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE key,
include/linux/btree-type.h
55
return btree_update(&head->h, BTREE_TYPE_GEO, &_key, val);
include/linux/btree-type.h
58
static inline void *BTREE_FN(remove)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE key)
include/linux/btree-type.h
61
return btree_remove(&head->h, BTREE_TYPE_GEO, &_key);
include/linux/btree-type.h
64
static inline void *BTREE_FN(last)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE *key)
include/linux/btree-type.h
67
void *val = btree_last(&head->h, BTREE_TYPE_GEO, &_key);
include/linux/btree-type.h
73
static inline void *BTREE_FN(get_prev)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE *key)
include/linux/btree-type.h
76
void *val = btree_get_prev(&head->h, BTREE_TYPE_GEO, &_key);
include/linux/btree-type.h
82
static inline void *BTREE_FN(lookup)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE key)
include/linux/btree-type.h
84
return btree_lookup(&head->h, BTREE_TYPE_GEO, (unsigned long *)&key);
include/linux/btree-type.h
87
static inline int BTREE_FN(insert)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE key,
include/linux/btree-type.h
90
return btree_insert(&head->h, BTREE_TYPE_GEO, (unsigned long *)&key,
include/linux/btree-type.h
94
static inline int BTREE_FN(update)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE key,
include/linux/btree-type.h
97
return btree_update(&head->h, BTREE_TYPE_GEO, (unsigned long *)&key, val);
include/linux/btree.h
100
void *btree_lookup(struct btree_head *head, struct btree_geo *geo,
include/linux/btree.h
115
int __must_check btree_insert(struct btree_head *head, struct btree_geo *geo,
include/linux/btree.h
128
int btree_update(struct btree_head *head, struct btree_geo *geo,
include/linux/btree.h
140
void *btree_remove(struct btree_head *head, struct btree_geo *geo,
include/linux/btree.h
172
void *btree_last(struct btree_head *head, struct btree_geo *geo,
include/linux/btree.h
186
void *btree_get_prev(struct btree_head *head, struct btree_geo *geo,
include/linux/btree.h
191
size_t btree_visitor(struct btree_head *head, struct btree_geo *geo,
include/linux/btree.h
199
size_t btree_grim_visitor(struct btree_head *head, struct btree_geo *geo,
include/linux/btree.h
216
#define btree_for_each_safel(head, key, val) \
include/linux/btree.h
217
for (val = btree_lastl(head, &key); \
include/linux/btree.h
219
val = btree_get_prevl(head, &key))
include/linux/btree.h
227
#define btree_for_each_safe32(head, key, val) \
include/linux/btree.h
228
for (val = btree_last32(head, &key); \
include/linux/btree.h
230
val = btree_get_prev32(head, &key))
include/linux/btree.h
239
#define btree_for_each_safe64(head, key, val) \
include/linux/btree.h
240
for (val = btree_last64(head, &key); \
include/linux/btree.h
242
val = btree_get_prev64(head, &key))
include/linux/btree.h
67
void btree_init_mempool(struct btree_head *head, mempool_t *mempool);
include/linux/btree.h
79
int __must_check btree_init(struct btree_head *head);
include/linux/btree.h
89
void btree_destroy(struct btree_head *head);
include/linux/ceph/pagelist.h
11
struct list_head head;
include/linux/circ_buf.h
11
int head;
include/linux/circ_buf.h
16
#define CIRC_CNT(head,tail,size) (((head) - (tail)) & ((size)-1))
include/linux/circ_buf.h
21
#define CIRC_SPACE(head,tail,size) CIRC_CNT((tail),((head)+1),(size))
include/linux/circ_buf.h
26
#define CIRC_CNT_TO_END(head,tail,size) \
include/linux/circ_buf.h
28
int n = ((head) + end) & ((size)-1); \
include/linux/circ_buf.h
32
#define CIRC_SPACE_TO_END(head,tail,size) \
include/linux/circ_buf.h
33
({int end = (size) - 1 - (head); \
include/linux/compat.h
675
compat_sys_set_robust_list(struct compat_robust_list_head __user *head,
include/linux/dma-fence-array.h
78
#define dma_fence_array_for_each(fence, index, head) \
include/linux/dma-fence-array.h
79
for (index = 0, fence = dma_fence_array_first(head); fence; \
include/linux/dma-fence-array.h
80
++(index), fence = dma_fence_array_next(head, index))
include/linux/dma-fence-array.h
95
struct dma_fence *dma_fence_array_first(struct dma_fence *head);
include/linux/dma-fence-array.h
96
struct dma_fence *dma_fence_array_next(struct dma_fence *head,
include/linux/dma-fence-chain.h
120
#define dma_fence_chain_for_each(iter, head) \
include/linux/dma-fence-chain.h
121
for (iter = dma_fence_get(head); iter; \
include/linux/dma-fence-unwrap.h
33
struct dma_fence *dma_fence_unwrap_first(struct dma_fence *head,
include/linux/dma-fence-unwrap.h
47
#define dma_fence_unwrap_for_each(fence, cursor, head) \
include/linux/dma-fence-unwrap.h
48
for (fence = dma_fence_unwrap_first(head, cursor); fence; \
include/linux/etherdevice.h
67
struct sk_buff *eth_gro_receive(struct list_head *head, struct sk_buff *skb);
include/linux/fb.h
821
struct list_head *head);
include/linux/fb.h
823
struct list_head *head);
include/linux/fb.h
825
struct list_head *head);
include/linux/fb.h
827
struct list_head *head);
include/linux/fb.h
829
struct list_head *head);
include/linux/fb.h
830
extern void fb_destroy_modelist(struct list_head *head);
include/linux/fb.h
832
struct list_head *head);
include/linux/fb.h
834
struct list_head *head);
include/linux/fbcon.h
21
void fbcon_delete_modelist(struct list_head *head);
include/linux/fbcon.h
42
static inline void fbcon_delete_modelist(struct list_head *head) {}
include/linux/fd.h
13
compat_uint_t head;
include/linux/filter.h
1559
struct ctl_table_header *head;
include/linux/fs_context.h
184
u8 head; /* Insertion index in buffer[] */
include/linux/hidraw.h
29
int head;
include/linux/if_macvlan.h
68
extern void macvlan_dellink(struct net_device *dev, struct list_head *head);
include/linux/if_team.h
252
struct hlist_head *head = team_port_index_hash(team, port_index);
include/linux/if_team.h
254
hlist_for_each_entry(port, head, hlist)
include/linux/if_team.h
273
struct hlist_head *head = team_port_index_hash(team, port_index);
include/linux/if_team.h
275
hlist_for_each_entry_rcu(port, head, hlist)
include/linux/iio/buffer-dma.h
55
struct list_head head;
include/linux/io_uring_types.h
140
u32 head;
include/linux/io_uring_types.h
245
struct io_kiocb *head;
include/linux/iommu.h
1299
struct list_head *head)
include/linux/iommu.h
942
struct list_head *head);
include/linux/jbd2.h
1409
static inline void jbd2_file_log_bh(struct list_head *head, struct buffer_head *bh)
include/linux/jbd2.h
1411
list_add_tail(&bh->b_assoc_buffers, head);
include/linux/kexec.h
340
kimage_entry_t head;
include/linux/list.h
1148
#define hlist_for_each(pos, head) \
include/linux/list.h
1149
for (pos = (head)->first; pos ; pos = pos->next)
include/linux/list.h
1151
#define hlist_for_each_safe(pos, n, head) \
include/linux/list.h
1152
for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \
include/linux/list.h
1166
#define hlist_for_each_entry(pos, head, member) \
include/linux/list.h
1167
for (pos = hlist_entry_safe((head)->first, typeof(*(pos)), member);\
include/linux/list.h
1197
#define hlist_for_each_entry_safe(pos, n, head, member) \
include/linux/list.h
1198
for (pos = hlist_entry_safe((head)->first, typeof(*pos), member);\
include/linux/list.h
1206
static inline size_t hlist_count_nodes(struct hlist_head *head)
include/linux/list.h
1211
hlist_for_each(pos, head)
include/linux/list.h
175
static inline void list_add(struct list_head *new, struct list_head *head)
include/linux/list.h
177
__list_add(new, head, head->next);
include/linux/list.h
189
static inline void list_add_tail(struct list_head *new, struct list_head *head)
include/linux/list.h
191
__list_add(new, head->prev, head);
include/linux/list.h
304
static inline void list_move(struct list_head *list, struct list_head *head)
include/linux/list.h
307
list_add(list, head);
include/linux/list.h
316
struct list_head *head)
include/linux/list.h
319
list_add_tail(list, head);
include/linux/list.h
331
static inline void list_bulk_move_tail(struct list_head *head,
include/linux/list.h
338
head->prev->next = first;
include/linux/list.h
339
first->prev = head->prev;
include/linux/list.h
341
last->next = head;
include/linux/list.h
342
head->prev = last;
include/linux/list.h
350
static inline int list_is_first(const struct list_head *list, const struct list_head *head)
include/linux/list.h
352
return list->prev == head;
include/linux/list.h
360
static inline int list_is_last(const struct list_head *list, const struct list_head *head)
include/linux/list.h
362
return list->next == head;
include/linux/list.h
370
static inline int list_is_head(const struct list_head *list, const struct list_head *head)
include/linux/list.h
372
return list == head;
include/linux/list.h
379
static inline int list_empty(const struct list_head *head)
include/linux/list.h
381
return READ_ONCE(head->next) == head;
include/linux/list.h
415
static inline int list_empty_careful(const struct list_head *head)
include/linux/list.h
417
struct list_head *next = smp_load_acquire(&head->next);
include/linux/list.h
418
return list_is_head(next, head) && (next == READ_ONCE(head->prev));
include/linux/list.h
425
static inline void list_rotate_left(struct list_head *head)
include/linux/list.h
429
if (!list_empty(head)) {
include/linux/list.h
430
first = head->next;
include/linux/list.h
431
list_move_tail(first, head);
include/linux/list.h
443
struct list_head *head)
include/linux/list.h
450
list_move_tail(head, list);
include/linux/list.h
457
static inline int list_is_singular(const struct list_head *head)
include/linux/list.h
459
return !list_empty(head) && (head->next == head->prev);
include/linux/list.h
463
struct list_head *head, struct list_head *entry)
include/linux/list.h
466
list->next = head->next;
include/linux/list.h
470
head->next = new_first;
include/linux/list.h
471
new_first->prev = head;
include/linux/list.h
489
struct list_head *head, struct list_head *entry)
include/linux/list.h
491
if (list_empty(head))
include/linux/list.h
493
if (list_is_singular(head) && !list_is_head(entry, head) && (entry != head->next))
include/linux/list.h
495
if (list_is_head(entry, head))
include/linux/list.h
498
__list_cut_position(list, head, entry);
include/linux/list.h
516
struct list_head *head,
include/linux/list.h
519
if (head->next == entry) {
include/linux/list.h
523
list->next = head->next;
include/linux/list.h
527
head->next = entry;
include/linux/list.h
528
entry->prev = head;
include/linux/list.h
551
struct list_head *head)
include/linux/list.h
554
__list_splice(list, head, head->next);
include/linux/list.h
563
struct list_head *head)
include/linux/list.h
566
__list_splice(list, head->prev, head);
include/linux/list.h
577
struct list_head *head)
include/linux/list.h
580
__list_splice(list, head, head->next);
include/linux/list.h
594
struct list_head *head)
include/linux/list.h
597
__list_splice(list, head->prev, head);
include/linux/list.h
678
#define list_next_entry_circular(pos, head, member) \
include/linux/list.h
679
(list_is_last(&(pos)->member, head) ? \
include/linux/list.h
680
list_first_entry(head, typeof(*(pos)), member) : list_next_entry(pos, member))
include/linux/list.h
699
#define list_prev_entry_circular(pos, head, member) \
include/linux/list.h
700
(list_is_first(&(pos)->member, head) ? \
include/linux/list.h
701
list_last_entry(head, typeof(*(pos)), member) : list_prev_entry(pos, member))
include/linux/list.h
708
#define list_for_each(pos, head) \
include/linux/list.h
709
for (pos = (head)->next; !list_is_head(pos, (head)); pos = pos->next)
include/linux/list.h
718
#define list_for_each_continue(pos, head) \
include/linux/list.h
719
for (pos = pos->next; !list_is_head(pos, (head)); pos = pos->next)
include/linux/list.h
726
#define list_for_each_prev(pos, head) \
include/linux/list.h
727
for (pos = (head)->prev; !list_is_head(pos, (head)); pos = pos->prev)
include/linux/list.h
735
#define list_for_each_safe(pos, n, head) \
include/linux/list.h
736
for (pos = (head)->next, n = pos->next; \
include/linux/list.h
737
!list_is_head(pos, (head)); \
include/linux/list.h
746
#define list_for_each_prev_safe(pos, n, head) \
include/linux/list.h
747
for (pos = (head)->prev, n = pos->prev; \
include/linux/list.h
748
!list_is_head(pos, (head)); \
include/linux/list.h
755
static inline size_t list_count_nodes(struct list_head *head)
include/linux/list.h
760
list_for_each(pos, head)
include/linux/list.h
772
#define list_entry_is_head(pos, head, member) \
include/linux/list.h
773
list_is_head(&pos->member, (head))
include/linux/list.h
781
#define list_for_each_entry(pos, head, member) \
include/linux/list.h
782
for (pos = list_first_entry(head, typeof(*pos), member); \
include/linux/list.h
783
!list_entry_is_head(pos, head, member); \
include/linux/list.h
792
#define list_for_each_entry_reverse(pos, head, member) \
include/linux/list.h
793
for (pos = list_last_entry(head, typeof(*pos), member); \
include/linux/list.h
794
!list_entry_is_head(pos, head, member); \
include/linux/list.h
805
#define list_prepare_entry(pos, head, member) \
include/linux/list.h
806
((pos) ? : list_entry(head, typeof(*pos), member))
include/linux/list.h
817
#define list_for_each_entry_continue(pos, head, member) \
include/linux/list.h
819
!list_entry_is_head(pos, head, member); \
include/linux/list.h
831
#define list_for_each_entry_continue_reverse(pos, head, member) \
include/linux/list.h
833
!list_entry_is_head(pos, head, member); \
include/linux/list.h
844
#define list_for_each_entry_from(pos, head, member) \
include/linux/list.h
845
for (; !list_entry_is_head(pos, head, member); \
include/linux/list.h
857
#define list_for_each_entry_from_reverse(pos, head, member) \
include/linux/list.h
858
for (; !list_entry_is_head(pos, head, member); \
include/linux/list.h
868
#define list_for_each_entry_safe(pos, n, head, member) \
include/linux/list.h
869
for (pos = list_first_entry(head, typeof(*pos), member), \
include/linux/list.h
871
!list_entry_is_head(pos, head, member); \
include/linux/list.h
884
#define list_for_each_entry_safe_continue(pos, n, head, member) \
include/linux/list.h
887
!list_entry_is_head(pos, head, member); \
include/linux/list.h
900
#define list_for_each_entry_safe_from(pos, n, head, member) \
include/linux/list.h
902
!list_entry_is_head(pos, head, member); \
include/linux/list.h
915
#define list_for_each_entry_safe_reverse(pos, n, head, member) \
include/linux/list.h
916
for (pos = list_last_entry(head, typeof(*pos), member), \
include/linux/list.h
918
!list_entry_is_head(pos, head, member); \
include/linux/list_bl.h
171
#define hlist_bl_for_each_entry(tpos, pos, head, member) \
include/linux/list_bl.h
172
for (pos = hlist_bl_first(head); \
include/linux/list_bl.h
185
#define hlist_bl_for_each_entry_safe(tpos, pos, n, head, member) \
include/linux/list_bl.h
186
for (pos = hlist_bl_first(head); \
include/linux/list_lru.h
206
struct list_head *head);
include/linux/list_nulls.h
128
#define hlist_nulls_for_each_entry(tpos, pos, head, member) \
include/linux/list_nulls.h
129
for (pos = (head)->first; \
include/linux/list_private.h
103
#define list_private_entry_is_head(pos, head, member) \
include/linux/list_private.h
104
list_is_head(&ACCESS_PRIVATE(pos, member), (head))
include/linux/list_private.h
112
#define list_private_for_each_entry(pos, head, member) \
include/linux/list_private.h
113
for (pos = list_private_first_entry(head, typeof(*pos), member); \
include/linux/list_private.h
114
!list_private_entry_is_head(pos, head, member); \
include/linux/list_private.h
123
#define list_private_for_each_entry_reverse(pos, head, member) \
include/linux/list_private.h
124
for (pos = list_private_last_entry(head, typeof(*pos), member); \
include/linux/list_private.h
125
!list_private_entry_is_head(pos, head, member); \
include/linux/list_private.h
137
#define list_private_for_each_entry_continue(pos, head, member) \
include/linux/list_private.h
139
!list_private_entry_is_head(pos, head, member); \
include/linux/list_private.h
151
#define list_private_for_each_entry_continue_reverse(pos, head, member) \
include/linux/list_private.h
153
!list_private_entry_is_head(pos, head, member); \
include/linux/list_private.h
164
#define list_private_for_each_entry_from(pos, head, member) \
include/linux/list_private.h
165
for (; !list_private_entry_is_head(pos, head, member); \
include/linux/list_private.h
177
#define list_private_for_each_entry_from_reverse(pos, head, member) \
include/linux/list_private.h
178
for (; !list_private_entry_is_head(pos, head, member); \
include/linux/list_private.h
188
#define list_private_for_each_entry_safe(pos, n, head, member) \
include/linux/list_private.h
189
for (pos = list_private_first_entry(head, typeof(*pos), member), \
include/linux/list_private.h
191
!list_private_entry_is_head(pos, head, member); \
include/linux/list_private.h
204
#define list_private_for_each_entry_safe_continue(pos, n, head, member) \
include/linux/list_private.h
207
!list_private_entry_is_head(pos, head, member); \
include/linux/list_private.h
220
#define list_private_for_each_entry_safe_from(pos, n, head, member) \
include/linux/list_private.h
222
!list_private_entry_is_head(pos, head, member); \
include/linux/list_private.h
235
#define list_private_for_each_entry_safe_reverse(pos, n, head, member) \
include/linux/list_private.h
236
for (pos = list_private_last_entry(head, typeof(*pos), member), \
include/linux/list_private.h
238
!list_private_entry_is_head(pos, head, member); \
include/linux/list_private.h
70
#define list_private_next_entry_circular(pos, head, member) \
include/linux/list_private.h
71
(list_is_last(&ACCESS_PRIVATE(pos, member), head) ? \
include/linux/list_private.h
72
list_private_first_entry(head, typeof(*(pos)), member) : \
include/linux/list_private.h
92
#define list_private_prev_entry_circular(pos, head, member) \
include/linux/list_private.h
93
(list_is_first(&ACCESS_PRIVATE(pos, member), head) ? \
include/linux/list_private.h
94
list_private_last_entry(head, typeof(*(pos)), member) : \
include/linux/list_sort.h
13
void list_sort(void *priv, struct list_head *head, list_cmp_func_t cmp);
include/linux/llist.h
216
static inline bool llist_empty(const struct llist_head *head)
include/linux/llist.h
218
return READ_ONCE(head->first) == NULL;
include/linux/llist.h
236
struct llist_head *head)
include/linux/llist.h
238
struct llist_node *first = READ_ONCE(head->first);
include/linux/llist.h
242
} while (!try_cmpxchg(&head->first, &first, new_first));
include/linux/llist.h
249
struct llist_head *head)
include/linux/llist.h
251
new_last->next = head->first;
include/linux/llist.h
252
head->first = new_first;
include/linux/llist.h
263
static inline bool llist_add(struct llist_node *new, struct llist_head *head)
include/linux/llist.h
265
return llist_add_batch(new, new, head);
include/linux/llist.h
268
static inline bool __llist_add(struct llist_node *new, struct llist_head *head)
include/linux/llist.h
270
return __llist_add_batch(new, new, head);
include/linux/llist.h
281
static inline struct llist_node *llist_del_all(struct llist_head *head)
include/linux/llist.h
283
return xchg(&head->first, NULL);
include/linux/llist.h
286
static inline struct llist_node *__llist_del_all(struct llist_head *head)
include/linux/llist.h
288
struct llist_node *first = head->first;
include/linux/llist.h
290
head->first = NULL;
include/linux/llist.h
294
extern struct llist_node *llist_del_first(struct llist_head *head);
include/linux/llist.h
303
static inline struct llist_node *llist_del_first_init(struct llist_head *head)
include/linux/llist.h
305
struct llist_node *n = llist_del_first(head);
include/linux/llist.h
312
extern bool llist_del_first_this(struct llist_head *head,
include/linux/llist.h
315
struct llist_node *llist_reverse_order(struct llist_node *head);
include/linux/mISDNif.h
480
struct hlist_head head;
include/linux/maple_tree.h
313
struct maple_enode *head;
include/linux/maple_tree.h
508
.head = NULL, \
include/linux/memremap.h
108
void (*folio_split)(struct folio *head, struct folio *tail);
include/linux/mlx5/driver.h
270
struct list_head head;
include/linux/mm.h
249
static inline struct folio *lru_to_folio(struct list_head *head)
include/linux/mm.h
251
return list_entry((head)->prev, struct folio, lru);
include/linux/mmzone.h
535
struct list_head *head;
include/linux/mroute_base.h
159
void (*free)(struct rcu_head *head);
include/linux/msdos_partition.h
9
u8 head; /* starting head */
include/linux/mtd/inftl.h
42
int head,sect,cyl;
include/linux/mtd/nftl.h
33
int head,sect,cyl;
include/linux/net/intel/libie/fwlog.h
58
u16 head;
include/linux/netdevice.h
2957
struct sk_buff *(*gro_receive)(struct list_head *head,
include/linux/netdevice.h
3373
void netif_close_many(struct list_head *head, bool unlink);
include/linux/netdevice.h
3405
void unregister_netdevice_queue(struct net_device *dev, struct list_head *head);
include/linux/netdevice.h
3406
void unregister_netdevice_many(struct list_head *head);
include/linux/netdevice.h
4247
void netif_receive_skb_list_internal(struct list_head *head);
include/linux/netdevice.h
4248
void netif_receive_skb_list(struct list_head *head);
include/linux/netfilter.h
119
struct rcu_head head;
include/linux/netfilter.h
218
void nf_hook_slow_list(struct list_head *head, struct nf_hook_state *state,
include/linux/netfilter.h
324
struct list_head *head, struct net_device *in, struct net_device *out,
include/linux/netfilter.h
354
nf_hook_slow_list(head, &state, hook_head);
include/linux/netfilter.h
427
struct list_head *head, struct net_device *in, struct net_device *out,
include/linux/netfilter/ipset/ip_set.h
178
int (*head)(struct ip_set *set, struct sk_buff *skb);
include/linux/netlink.h
260
if (is_vmalloc_addr(skb->head))
include/linux/nfs_page.h
155
extern void nfs_join_page_group(struct nfs_page *head,
include/linux/nfs_page.h
241
nfs_list_add_request(struct nfs_page *req, struct list_head *head)
include/linux/nfs_page.h
243
list_add_tail(&req->wb_list, head);
include/linux/nfs_page.h
252
nfs_list_move_request(struct nfs_page *req, struct list_head *head)
include/linux/nfs_page.h
254
list_move_tail(&req->wb_list, head);
include/linux/nfs_page.h
270
nfs_list_entry(struct list_head *head)
include/linux/nfs_page.h
272
return list_entry(head, struct nfs_page, wb_list);
include/linux/nfs_xdr.h
1733
void (*error_cleanup)(struct list_head *head, int);
include/linux/notifier.h
100
.head = NULL }
include/linux/notifier.h
103
.head = NULL }
include/linux/notifier.h
105
.head = NULL }
include/linux/notifier.h
110
.head = NULL, \
include/linux/notifier.h
62
struct notifier_block __rcu *head;
include/linux/notifier.h
67
struct notifier_block __rcu *head;
include/linux/notifier.h
71
struct notifier_block __rcu *head;
include/linux/notifier.h
78
struct notifier_block __rcu *head;
include/linux/notifier.h
83
(name)->head = NULL; \
include/linux/notifier.h
87
(name)->head = NULL; \
include/linux/notifier.h
90
(name)->head = NULL; \
include/linux/objpool.h
130
uint32_t head = smp_load_acquire(&slot->head);
include/linux/objpool.h
132
while (head != READ_ONCE(slot->last)) {
include/linux/objpool.h
147
if (READ_ONCE(slot->last) - head - 1 >= pool->nr_objs) {
include/linux/objpool.h
148
head = READ_ONCE(slot->head);
include/linux/objpool.h
153
obj = READ_ONCE(slot->entries[head & slot->mask]);
include/linux/objpool.h
156
if (try_cmpxchg_release(&slot->head, &head, head + 1))
include/linux/objpool.h
194
uint32_t head, tail;
include/linux/objpool.h
200
head = READ_ONCE(slot->head);
include/linux/objpool.h
202
WARN_ON_ONCE(tail - head > pool->nr_objs);
include/linux/objpool.h
54
uint32_t head;
include/linux/objpool.h
70
typedef int (*objpool_fini_cb)(struct objpool_head *head, void *context);
include/linux/page-flags.h
226
unsigned long head = READ_ONCE(page[1].compound_head);
include/linux/page-flags.h
228
if (likely(head & 1))
include/linux/page-flags.h
229
return (const struct page *)(head - 1);
include/linux/page-flags.h
284
unsigned long head = READ_ONCE(page->compound_head);
include/linux/page-flags.h
286
if (unlikely(head & 1))
include/linux/page-flags.h
287
return head - 1;
include/linux/page-flags.h
853
__SETPAGEFLAG(Head, head, PF_ANY)
include/linux/page-flags.h
854
__CLEARPAGEFLAG(Head, head, PF_ANY)
include/linux/page-flags.h
855
CLEARPAGEFLAG(Head, head, PF_ANY)
include/linux/page-flags.h
868
static __always_inline void set_compound_head(struct page *page, struct page *head)
include/linux/page-flags.h
870
WRITE_ONCE(page->compound_head, (unsigned long)head + 1);
include/linux/pagemap.h
168
void *head;
include/linux/pagemap.h
179
head = rcu_access_pointer(mapping->i_pages.xa_head);
include/linux/pagemap.h
180
if (!head)
include/linux/pagemap.h
189
if (!xa_is_node(head) && xa_is_value(head))
include/linux/perf_event.h
1165
unsigned long head;
include/linux/perf_event.h
1816
struct hlist_head *head, int rctx,
include/linux/phy.h
820
list_for_each_entry(port, &(phydev)->ports, head)
include/linux/phy_port.h
55
struct list_head head;
include/linux/pipe_fs_i.h
175
static inline unsigned int pipe_occupancy(unsigned int head, unsigned int tail)
include/linux/pipe_fs_i.h
177
return (pipe_index_t)(head - tail);
include/linux/pipe_fs_i.h
185
static inline bool pipe_empty(unsigned int head, unsigned int tail)
include/linux/pipe_fs_i.h
187
return !pipe_occupancy(head, tail);
include/linux/pipe_fs_i.h
196
static inline bool pipe_full(unsigned int head, unsigned int tail,
include/linux/pipe_fs_i.h
199
return pipe_occupancy(head, tail) >= limit;
include/linux/pipe_fs_i.h
208
return pipe_full(pipe->head, pipe->tail, pipe->max_usage);
include/linux/pipe_fs_i.h
217
return pipe_empty(pipe->head, pipe->tail);
include/linux/pipe_fs_i.h
226
return pipe_occupancy(pipe->head, pipe->tail);
include/linux/pipe_fs_i.h
246
return pipe_buf(pipe, pipe->head);
include/linux/pipe_fs_i.h
56
pipe_index_t head;
include/linux/plist.h
115
plist_head_init(struct plist_head *head)
include/linux/plist.h
117
INIT_LIST_HEAD(&head->node_list);
include/linux/plist.h
132
extern void plist_add(struct plist_node *node, struct plist_head *head);
include/linux/plist.h
133
extern void plist_del(struct plist_node *node, struct plist_head *head);
include/linux/plist.h
135
extern void plist_requeue(struct plist_node *node, struct plist_head *head);
include/linux/plist.h
142
#define plist_for_each(pos, head) \
include/linux/plist.h
143
list_for_each_entry(pos, &(head)->node_list, node_list)
include/linux/plist.h
152
#define plist_for_each_continue(pos, head) \
include/linux/plist.h
153
list_for_each_entry_continue(pos, &(head)->node_list, node_list)
include/linux/plist.h
163
#define plist_for_each_safe(pos, n, head) \
include/linux/plist.h
164
list_for_each_entry_safe(pos, n, &(head)->node_list, node_list)
include/linux/plist.h
172
#define plist_for_each_entry(pos, head, mem) \
include/linux/plist.h
173
list_for_each_entry(pos, &(head)->node_list, mem.node_list)
include/linux/plist.h
184
#define plist_for_each_entry_continue(pos, head, m) \
include/linux/plist.h
185
list_for_each_entry_continue(pos, &(head)->node_list, m.node_list)
include/linux/plist.h
196
#define plist_for_each_entry_safe(pos, n, head, m) \
include/linux/plist.h
197
list_for_each_entry_safe(pos, n, &(head)->node_list, m.node_list)
include/linux/plist.h
203
static inline int plist_head_empty(const struct plist_head *head)
include/linux/plist.h
205
return list_empty(&head->node_list);
include/linux/plist.h
226
# define plist_first_entry(head, type, member) \
include/linux/plist.h
228
WARN_ON(plist_head_empty(head)); \
include/linux/plist.h
229
container_of(plist_first(head), type, member); \
include/linux/plist.h
232
# define plist_first_entry(head, type, member) \
include/linux/plist.h
233
container_of(plist_first(head), type, member)
include/linux/plist.h
243
# define plist_last_entry(head, type, member) \
include/linux/plist.h
245
WARN_ON(plist_head_empty(head)); \
include/linux/plist.h
246
container_of(plist_last(head), type, member); \
include/linux/plist.h
249
# define plist_last_entry(head, type, member) \
include/linux/plist.h
250
container_of(plist_last(head), type, member)
include/linux/plist.h
273
static inline struct plist_node *plist_first(const struct plist_head *head)
include/linux/plist.h
275
return list_entry(head->node_list.next,
include/linux/plist.h
285
static inline struct plist_node *plist_last(const struct plist_head *head)
include/linux/plist.h
287
return list_entry(head->node_list.prev,
include/linux/plist.h
86
#define PLIST_HEAD_INIT(head) \
include/linux/plist.h
88
.node_list = LIST_HEAD_INIT((head).node_list) \
include/linux/plist.h
95
#define PLIST_HEAD(head) \
include/linux/plist.h
96
struct plist_head head = PLIST_HEAD_INIT(head)
include/linux/posix-timers.h
61
struct timerqueue_head *head;
include/linux/posix-timers.h
69
static inline bool cpu_timer_enqueue(struct timerqueue_head *head,
include/linux/posix-timers.h
72
ctmr->head = head;
include/linux/posix-timers.h
73
return timerqueue_add(head, &ctmr->node);
include/linux/posix-timers.h
78
return !!ctmr->head;
include/linux/posix-timers.h
84
timerqueue_del(ctmr->head, &ctmr->node);
include/linux/posix-timers.h
85
ctmr->head = NULL;
include/linux/ptr_ring.h
251
int head = consumer_head;
include/linux/ptr_ring.h
258
while (likely(head > r->consumer_tail))
include/linux/ptr_ring.h
259
r->queue[--head] = NULL;
include/linux/ptr_ring.h
544
int head = r->consumer_head - 1;
include/linux/ptr_ring.h
545
if (head < 0)
include/linux/ptr_ring.h
546
head = r->size - 1;
include/linux/ptr_ring.h
547
if (r->queue[head]) {
include/linux/ptr_ring.h
551
r->queue[head] = batch[--n];
include/linux/ptr_ring.h
552
r->consumer_tail = head;
include/linux/ptr_ring.h
554
WRITE_ONCE(r->consumer_head, head);
include/linux/rcu_segcblist.h
191
struct rcu_head *head;
include/linux/rcu_segcblist.h
205
.head = NULL, \
include/linux/rcu_segcblist.h
206
.tails[RCU_DONE_TAIL] = &n.head, \
include/linux/rcu_segcblist.h
207
.tails[RCU_WAIT_TAIL] = &n.head, \
include/linux/rcu_segcblist.h
208
.tails[RCU_NEXT_READY_TAIL] = &n.head, \
include/linux/rcu_segcblist.h
209
.tails[RCU_NEXT_TAIL] = &n.head, \
include/linux/rcu_segcblist.h
22
struct rcu_head *head;
include/linux/rcu_segcblist.h
27
#define RCU_CBLIST_INITIALIZER(n) { .head = NULL, .tail = &n.head }
include/linux/rculist.h
125
static inline void list_add_rcu(struct list_head *new, struct list_head *head)
include/linux/rculist.h
127
__list_add_rcu(new, head, head->next);
include/linux/rculist.h
147
struct list_head *head)
include/linux/rculist.h
149
__list_add_rcu(new, head->prev, head);
include/linux/rculist.h
332
struct list_head *head,
include/linux/rculist.h
336
__list_splice_init_rcu(list, head, head->next, sync);
include/linux/rculist.h
347
struct list_head *head,
include/linux/rculist.h
351
__list_splice_init_rcu(list, head->prev, head, sync);
include/linux/rculist.h
426
#define list_next_or_null_rcu(head, ptr, type, member) \
include/linux/rculist.h
428
struct list_head *__head = (head); \
include/linux/rculist.h
446
#define list_for_each_entry_rcu(pos, head, member, cond...) \
include/linux/rculist.h
448
pos = list_entry_rcu((head)->next, typeof(*pos), member); \
include/linux/rculist.h
449
&pos->member != (head); \
include/linux/rculist.h
465
#define list_for_each_entry_srcu(pos, head, member, cond) \
include/linux/rculist.h
467
pos = list_entry_rcu((head)->next, typeof(*pos), member); \
include/linux/rculist.h
468
&pos->member != (head); \
include/linux/rculist.h
50
#define list_for_each_rcu(pos, head) \
include/linux/rculist.h
500
#define list_for_each_entry_lockless(pos, head, member) \
include/linux/rculist.h
501
for (pos = list_entry_lockless((head)->next, typeof(*pos), member); \
include/linux/rculist.h
502
&pos->member != (head); \
include/linux/rculist.h
51
for (pos = rcu_dereference((head)->next); \
include/linux/rculist.h
52
!list_is_head(pos, (head)); \
include/linux/rculist.h
523
#define list_for_each_entry_continue_rcu(pos, head, member) \
include/linux/rculist.h
525
&pos->member != (head); \
include/linux/rculist.h
545
#define list_for_each_entry_from_rcu(pos, head, member) \
include/linux/rculist.h
546
for (; &(pos)->member != (head); \
include/linux/rculist.h
620
#define hlist_first_rcu(head) (*((struct hlist_node __rcu **)(&(head)->first)))
include/linux/rculist.h
63
#define list_tail_rcu(head) (*((struct list_head __rcu **)(&(head)->prev)))
include/linux/rculist.h
747
#define __hlist_for_each_rcu(pos, head) \
include/linux/rculist.h
748
for (pos = rcu_dereference(hlist_first_rcu(head)); \
include/linux/rculist.h
763
#define hlist_for_each_entry_rcu(pos, head, member, cond...) \
include/linux/rculist.h
765
pos = hlist_entry_safe(rcu_dereference_raw(hlist_first_rcu(head)),\
include/linux/rculist.h
784
#define hlist_for_each_entry_srcu(pos, head, member, cond) \
include/linux/rculist.h
786
pos = hlist_entry_safe(rcu_dereference_raw(hlist_first_rcu(head)),\
include/linux/rculist.h
805
#define hlist_for_each_entry_rcu_notrace(pos, head, member) \
include/linux/rculist.h
806
for (pos = hlist_entry_safe(rcu_dereference_raw_check(hlist_first_rcu(head)),\
include/linux/rculist.h
822
#define hlist_for_each_entry_rcu_bh(pos, head, member) \
include/linux/rculist.h
823
for (pos = hlist_entry_safe(rcu_dereference_bh(hlist_first_rcu(head)),\
include/linux/rculist_bl.h
95
#define hlist_bl_for_each_entry_rcu(tpos, pos, head, member) \
include/linux/rculist_bl.h
96
for (pos = hlist_bl_first_rcu(head); \
include/linux/rculist_nulls.h
226
#define hlist_nulls_for_each_entry_rcu(tpos, pos, head, member) \
include/linux/rculist_nulls.h
228
pos = rcu_dereference_raw(hlist_nulls_first_rcu(head)); \
include/linux/rculist_nulls.h
241
#define hlist_nulls_for_each_entry_safe(tpos, pos, head, member) \
include/linux/rculist_nulls.h
243
pos = rcu_dereference_raw(hlist_nulls_first_rcu(head)); \
include/linux/rculist_nulls.h
45
#define hlist_nulls_first_rcu(head) \
include/linux/rculist_nulls.h
46
(*((struct hlist_nulls_node __rcu __force **)&(head)->first))
include/linux/rcupdate.h
1111
void kvfree_call_rcu(struct rcu_head *head, void *ptr);
include/linux/rcupdate.h
121
void call_rcu_hurry(struct rcu_head *head, rcu_callback_t func);
include/linux/rcupdate.h
123
static inline void call_rcu_hurry(struct rcu_head *head, rcu_callback_t func)
include/linux/rcupdate.h
125
call_rcu(head, func);
include/linux/rcupdate.h
179
void call_rcu_tasks(struct rcu_head *head, rcu_callback_t func);
include/linux/rcupdate.h
286
void init_rcu_head(struct rcu_head *head);
include/linux/rcupdate.h
287
void destroy_rcu_head(struct rcu_head *head);
include/linux/rcupdate.h
288
void init_rcu_head_on_stack(struct rcu_head *head);
include/linux/rcupdate.h
289
void destroy_rcu_head_on_stack(struct rcu_head *head);
include/linux/rcupdate.h
291
static inline void init_rcu_head(struct rcu_head *head) { }
include/linux/rcupdate.h
292
static inline void destroy_rcu_head(struct rcu_head *head) { }
include/linux/rcupdate.h
293
static inline void init_rcu_head_on_stack(struct rcu_head *head) { }
include/linux/rcupdate.h
294
static inline void destroy_rcu_head_on_stack(struct rcu_head *head) { }
include/linux/rcupdate.h
51
void call_rcu(struct rcu_head *head, rcu_callback_t func);
include/linux/rcupdate_wait.h
17
struct rcu_head head;
include/linux/rcupdate_wait.h
23
void wakeme_after_rcu(struct rcu_head *head);
include/linux/reboot-mode.h
7
struct list_head head;
include/linux/resource_ext.h
32
extern void resource_list_free(struct list_head *head);
include/linux/resource_ext.h
35
struct list_head *head)
include/linux/resource_ext.h
37
list_add(&entry->node, head);
include/linux/resource_ext.h
41
struct list_head *head)
include/linux/resource_ext.h
43
list_add_tail(&entry->node, head);
include/linux/rhashtable.h
449
#define rht_for_each_from(pos, head, tbl, hash) \
include/linux/rhashtable.h
450
for (pos = head; \
include/linux/rhashtable.h
473
#define rht_for_each_entry_from(tpos, pos, head, tbl, hash, member) \
include/linux/rhashtable.h
474
for (pos = head; \
include/linux/rhashtable.h
523
#define rht_for_each_rcu_from(pos, head, tbl, hash) \
include/linux/rhashtable.h
525
pos = head; \
include/linux/rhashtable.h
558
#define rht_for_each_entry_rcu_from(tpos, pos, head, tbl, hash, member) \
include/linux/rhashtable.h
560
pos = head; \
include/linux/rhashtable.h
772
struct rhash_head *head;
include/linux/rhashtable.h
797
rht_for_each_from(head, rht_ptr(bkt, tbl, hash), tbl, hash) {
include/linux/rhashtable.h
804
params.obj_cmpfn(&arg, rht_obj(ht, head)) :
include/linux/rhashtable.h
805
rhashtable_compare(&arg, rht_obj(ht, head)))) {
include/linux/rhashtable.h
806
pprev = &head->next;
include/linux/rhashtable.h
810
data = rht_obj(ht, head);
include/linux/rhashtable.h
817
plist = container_of(head, struct rhlist_head, rhead);
include/linux/rhashtable.h
820
head = rht_dereference_bucket(head->next, tbl, hash);
include/linux/rhashtable.h
821
RCU_INIT_POINTER(list->rhead.next, head);
include/linux/rhashtable.h
842
head = rht_ptr(bkt, tbl, hash);
include/linux/rhashtable.h
844
RCU_INIT_POINTER(obj->next, head);
include/linux/rolling_buffer.h
25
struct folio_queue *head; /* Producer's insertion point */
include/linux/rseq_entry.h
271
u64 start_ip, abort_ip, offset, cs_end, head, tasksize = TASK_SIZE;
include/linux/rseq_entry.h
311
unsafe_get_user(head, uc_head, efault);
include/linux/rseq_entry.h
312
if (unlikely(head))
include/linux/rtnetlink.h
186
void rtnl_kfree_skbs(struct sk_buff *head, struct sk_buff *tail);
include/linux/sched/wake_q.h
51
static inline void wake_q_init(struct wake_q_head *head)
include/linux/sched/wake_q.h
53
head->first = WAKE_Q_TAIL;
include/linux/sched/wake_q.h
54
head->lastp = &head->first;
include/linux/sched/wake_q.h
57
static inline bool wake_q_empty(struct wake_q_head *head)
include/linux/sched/wake_q.h
59
return head->first == WAKE_Q_TAIL;
include/linux/sched/wake_q.h
62
extern void wake_q_add(struct wake_q_head *head, struct task_struct *task);
include/linux/sched/wake_q.h
63
extern void wake_q_add_safe(struct wake_q_head *head, struct task_struct *task);
include/linux/sched/wake_q.h
64
extern void wake_up_q(struct wake_q_head *head);
include/linux/seq_file.h
299
extern struct list_head *seq_list_start(struct list_head *head,
include/linux/seq_file.h
301
extern struct list_head *seq_list_start_head(struct list_head *head,
include/linux/seq_file.h
303
extern struct list_head *seq_list_next(void *v, struct list_head *head,
include/linux/seq_file.h
306
extern struct list_head *seq_list_start_rcu(struct list_head *head, loff_t pos);
include/linux/seq_file.h
307
extern struct list_head *seq_list_start_head_rcu(struct list_head *head, loff_t pos);
include/linux/seq_file.h
308
extern struct list_head *seq_list_next_rcu(void *v, struct list_head *head, loff_t *ppos);
include/linux/seq_file.h
314
extern struct hlist_node *seq_hlist_start(struct hlist_head *head,
include/linux/seq_file.h
316
extern struct hlist_node *seq_hlist_start_head(struct hlist_head *head,
include/linux/seq_file.h
318
extern struct hlist_node *seq_hlist_next(void *v, struct hlist_head *head,
include/linux/seq_file.h
321
extern struct hlist_node *seq_hlist_start_rcu(struct hlist_head *head,
include/linux/seq_file.h
323
extern struct hlist_node *seq_hlist_start_head_rcu(struct hlist_head *head,
include/linux/seq_file.h
326
struct hlist_head *head,
include/linux/seq_file.h
330
extern struct hlist_node *seq_hlist_start_percpu(struct hlist_head __percpu *head, int *cpu, loff_t pos);
include/linux/seq_file.h
332
extern struct hlist_node *seq_hlist_next_percpu(void *v, struct hlist_head __percpu *head, int *cpu, loff_t *pos);
include/linux/skbuff.h
1094
unsigned char *head,
include/linux/skbuff.h
1724
return skb->head + skb->end;
include/linux/skbuff.h
1744
return skb->end - skb->head;
include/linux/skbuff.h
1749
skb->end = skb->head + offset;
include/linux/skbuff.h
2355
struct sk_buff_head *head)
include/linux/skbuff.h
2358
__skb_queue_splice(list, (struct sk_buff *) head, head->next);
include/linux/skbuff.h
2359
head->qlen += list->qlen;
include/linux/skbuff.h
2371
struct sk_buff_head *head)
include/linux/skbuff.h
2374
__skb_queue_splice(list, (struct sk_buff *) head, head->next);
include/linux/skbuff.h
2375
head->qlen += list->qlen;
include/linux/skbuff.h
2386
struct sk_buff_head *head)
include/linux/skbuff.h
2389
__skb_queue_splice(list, head->prev, (struct sk_buff *) head);
include/linux/skbuff.h
2390
head->qlen += list->qlen;
include/linux/skbuff.h
2403
struct sk_buff_head *head)
include/linux/skbuff.h
2406
__skb_queue_splice(list, head->prev, (struct sk_buff *) head);
include/linux/skbuff.h
2407
head->qlen += list->qlen;
include/linux/skbuff.h
2704
return skb->head + skb->tail;
include/linux/skbuff.h
2709
skb->tail = skb->data - skb->head;
include/linux/skbuff.h
2816
DEBUG_NET_WARN_ON_ONCE(skb->data < skb->head);
include/linux/skbuff.h
2889
return skb->data - skb->head;
include/linux/skbuff.h
2998
return skb->head + skb->inner_transport_header;
include/linux/skbuff.h
3008
long offset = skb->data - skb->head;
include/linux/skbuff.h
3023
return skb->head + skb->inner_network_header;
include/linux/skbuff.h
3028
long offset = skb->data - skb->head;
include/linux/skbuff.h
3048
return skb->head + skb->inner_mac_header;
include/linux/skbuff.h
3053
long offset = skb->data - skb->head;
include/linux/skbuff.h
3073
return skb->head + skb->transport_header;
include/linux/skbuff.h
3078
long offset = skb->data - skb->head;
include/linux/skbuff.h
3095
long offset = skb->data - skb->head;
include/linux/skbuff.h
3116
return skb->head + skb->network_header;
include/linux/skbuff.h
3121
long offset = skb->data - skb->head;
include/linux/skbuff.h
3136
return skb->head + skb->mac_header;
include/linux/skbuff.h
3157
long offset = skb->data - skb->head;
include/linux/skbuff.h
3218
return skb->head + skb->csum_start;
include/linux/skbuff.h
4929
skb->csum_start = ((unsigned char *)ptr + start) - skb->head;
include/linux/srcu.h
90
void call_srcu(struct srcu_struct *ssp, struct rcu_head *head,
include/linux/srcu.h
91
void (*func)(struct rcu_head *head));
include/linux/sunrpc/svc.h
507
struct kvec *argv = buf->head;
include/linux/sunrpc/svc.h
509
WARN_ON(buf->len != buf->head->iov_len + buf->page_len + buf->tail->iov_len);
include/linux/sunrpc/svc.h
510
buf->len = buf->head->iov_len + buf->page_len + buf->tail->iov_len;
include/linux/sunrpc/svc.h
525
struct kvec *resv = buf->head;
include/linux/sunrpc/svc.h
572
struct kvec *resv = buf->head;
include/linux/sunrpc/svc_rdma.h
289
struct svc_rdma_recv_ctxt *head);
include/linux/sunrpc/xdr.h
57
struct kvec head[1], /* RPC header + non-page data */
include/linux/sunrpc/xdr.h
76
buf->head[0].iov_base = start;
include/linux/sunrpc/xdr.h
77
buf->head[0].iov_len = len;
include/linux/sunrpc/xprt.h
122
#define rq_svec rq_snd_buf.head
include/linux/syscalls.h
578
asmlinkage long sys_set_robust_list(struct robust_list_head __user *head,
include/linux/sysctl.h
236
void (*set_ownership)(struct ctl_table_header *head,
include/linux/sysctl.h
238
int (*permissions)(struct ctl_table_header *head, const struct ctl_table *table);
include/linux/thunderbolt.h
544
int head;
include/linux/timerqueue.h
10
extern bool timerqueue_del(struct timerqueue_head *head,
include/linux/timerqueue.h
23
struct timerqueue_node *timerqueue_getnext(struct timerqueue_head *head)
include/linux/timerqueue.h
25
struct rb_node *leftmost = rb_first_cached(&head->rb_root);
include/linux/timerqueue.h
40
static inline void timerqueue_init_head(struct timerqueue_head *head)
include/linux/timerqueue.h
42
head->rb_root = RB_ROOT_CACHED;
include/linux/timerqueue.h
8
extern bool timerqueue_add(struct timerqueue_head *head,
include/linux/trace_events.h
931
struct pt_regs *regs, struct hlist_head *head,
include/linux/trace_events.h
936
u64 count, struct pt_regs *regs, void *head,
include/linux/trace_events.h
939
perf_tp_event(type, count, raw_data, size, regs, head, rctx, task);
include/linux/tty_buffer.h
36
struct tty_buffer *head; /* Queue head */
include/linux/types.h
248
void (*func)(struct callback_head *head);
include/linux/types.h
252
typedef void (*rcu_callback_t)(struct rcu_head *head);
include/linux/types.h
253
typedef void (*call_rcu_func_t)(struct rcu_head *head, rcu_callback_t func);
include/linux/udp.h
93
struct list_head *head,
include/linux/usb/hcd.h
57
struct list_head head;
include/linux/usb/phy.h
123
struct list_head head;
include/linux/virtio_vsock.h
130
return (size_t)(skb_end_pointer(skb) - skb->head);
include/linux/virtio_vsock.h
22
return (struct virtio_vsock_hdr *)skb->head;
include/linux/vmw_vmci_defs.h
916
u64 head;
include/linux/vmw_vmci_defs.h
920
head = vmci_q_header_consumer_head(consume_q_header);
include/linux/vmw_vmci_defs.h
922
if (tail >= produce_q_size || head >= produce_q_size)
include/linux/vmw_vmci_defs.h
930
if (tail >= head)
include/linux/vmw_vmci_defs.h
931
free_space = produce_q_size - (tail - head) - 1;
include/linux/vmw_vmci_defs.h
933
free_space = head - tail - 1;
include/linux/vringh.h
163
u16 *head);
include/linux/vringh.h
173
int vringh_complete_user(struct vringh *vrh, u16 head, u32 len);
include/linux/vringh.h
232
u16 *head,
include/linux/vringh.h
235
int vringh_complete_kern(struct vringh *vrh, u16 head, u32 len);
include/linux/vringh.h
305
u16 *head,
include/linux/vringh.h
315
int vringh_complete_iotlb(struct vringh *vrh, u16 head, u32 len);
include/linux/wait.h
127
return !list_empty(&wq_head->head);
include/linux/wait.h
140
return list_is_singular(&wq_head->head);
include/linux/wait.h
173
struct list_head *head = &wq_head->head;
include/linux/wait.h
176
list_for_each_entry(wq, &wq_head->head, entry) {
include/linux/wait.h
179
head = &wq->entry;
include/linux/wait.h
181
list_add(&wq_entry->entry, head);
include/linux/wait.h
196
list_add_tail(&wq_entry->entry, &wq_head->head);
include/linux/wait.h
37
struct list_head head;
include/linux/wait.h
57
.head = LIST_HEAD_INIT(name.head) }
include/linux/wwan.h
188
struct list_head *head);
include/media/v4l2-fwnode.h
122
struct list_head head;
include/media/v4l2-fwnode.h
403
struct v4l2_connector_link, head)
include/media/v4l2-fwnode.h
412
list_last_entry(&(v4l2c)->links, struct v4l2_connector_link, head)
include/net/act_api.h
115
struct list_head head;
include/net/bluetooth/bluetooth.h
415
struct hlist_head head;
include/net/bluetooth/coredump.h
48
char *head;
include/net/bluetooth/l2cap.h
505
__u16 head;
include/net/bond_alb.h
116
u32 head; /* Index to the head of the bi-directional clients
include/net/cfg80211.h
1400
const u8 *head, *tail;
include/net/fq_impl.h
101
head = &tin->new_flows;
include/net/fq_impl.h
102
if (list_empty(head)) {
include/net/fq_impl.h
103
head = &tin->old_flows;
include/net/fq_impl.h
104
if (list_empty(head))
include/net/fq_impl.h
108
flow = list_first_entry(head, struct fq_flow, flowchain);
include/net/fq_impl.h
120
if ((head == &tin->new_flows) &&
include/net/fq_impl.h
315
struct list_head *head;
include/net/fq_impl.h
319
head = &tin->new_flows;
include/net/fq_impl.h
320
if (list_empty(head)) {
include/net/fq_impl.h
321
head = &tin->old_flows;
include/net/fq_impl.h
322
if (list_empty(head))
include/net/fq_impl.h
326
flow = list_first_entry(head, struct fq_flow, flowchain);
include/net/fq_impl.h
95
struct list_head *head;
include/net/gre.h
140
skb->csum_start = skb_transport_header(skb) - skb->head;
include/net/gro.h
110
struct list_head *head,
include/net/gro.h
118
return cb(head, skb);
include/net/gro.h
125
struct list_head *head,
include/net/gro.h
133
return cb(sk, head, skb);
include/net/gro.h
411
#define indirect_call_gro_receive_inet(cb, f2, f1, head, skb) \
include/net/gro.h
415
INDIRECT_CALL_INET(cb, f2, f1, head, skb); \
include/net/gro.h
418
struct sk_buff *udp_gro_receive(struct list_head *head, struct sk_buff *skb,
include/net/gso.h
28
return (skb_mac_header(inner_skb) - inner_skb->head) -
include/net/gso.h
54
SKB_GSO_CB(skb)->csum_start = skb_checksum_start(skb) - skb->head;
include/net/gso.h
68
int plen = (skb->head + SKB_GSO_CB(skb)->csum_start) - csum_start;
include/net/gso.h
72
SKB_GSO_CB(skb)->csum_start = csum_start - skb->head;
include/net/inet_common.h
65
struct sk_buff *inet_gro_receive(struct list_head *head, struct sk_buff *skb);
include/net/inet_common.h
76
#define indirect_call_gro_receive(f2, f1, cb, head, skb) \
include/net/inet_common.h
80
INDIRECT_CALL_2(cb, f2, f1, head, skb); \
include/net/inet_frag.h
178
void inet_frag_reasm_finish(struct inet_frag_queue *q, struct sk_buff *head,
include/net/inet_hashtables.h
125
#define inet_bind_bucket_for_each(tb, head) \
include/net/inet_hashtables.h
126
hlist_for_each_entry(tb, head, node)
include/net/inet_hashtables.h
219
struct inet_bind_hashbucket *head,
include/net/inet_hashtables.h
229
struct inet_bind_hashbucket *head,
include/net/inet_hashtables.h
237
inet_bind2_bucket_find(const struct inet_bind_hashbucket *head,
include/net/ioam6.h
19
struct rhash_head head;
include/net/ioam6.h
30
struct rhash_head head;
include/net/ip.h
166
void ip_list_rcv(struct list_head *head, struct packet_type *pt,
include/net/ip6_checksum.h
61
skb->csum_start = skb_transport_header(skb) - skb->head;
include/net/ip6_fib.h
328
void fib6_info_destroy_rcu(struct rcu_head *head);
include/net/ip_tunnels.h
395
void ip_tunnel_dellink(struct net_device *dev, struct list_head *head);
include/net/ip_vs.h
1642
void ip_vs_dest_dst_rcu_free(struct rcu_head *head);
include/net/ipv6.h
1004
void ipv6_list_rcv(struct list_head *head, struct packet_type *pt,
include/net/ipv6_frag.h
101
head = inet_frag_pull_head(&fq->q);
include/net/ipv6_frag.h
102
if (!head)
include/net/ipv6_frag.h
105
head->dev = dev;
include/net/ipv6_frag.h
108
icmpv6_send(head, ICMPV6_TIME_EXCEED, ICMPV6_EXC_FRAGTIME, 0);
include/net/ipv6_frag.h
109
kfree_skb_reason(head, SKB_DROP_REASON_FRAG_REASM_TIMEOUT);
include/net/ipv6_frag.h
68
struct sk_buff *head;
include/net/ipv6_stubs.h
65
struct list_head *head,
include/net/iucv/af_iucv.h
159
struct hlist_head head;
include/net/libeth/xdp.h
941
u32 head, nr_frags, i, ret = LIBETH_XDP_TX;
include/net/libeth/xdp.h
949
head = libeth_xdp_xmit_queue_head(bq, xdpf, dev);
include/net/libeth/xdp.h
950
if (head == LIBETH_XDP_PASS)
include/net/libeth/xdp.h
952
else if (head == LIBETH_XDP_DROP)
include/net/libeth/xsk.h
333
struct libeth_xdp_buff *libeth_xsk_buff_add_frag(struct libeth_xdp_buff *head,
include/net/libeth/xsk.h
351
libeth_xsk_process_buff(struct libeth_xdp_buff *head,
include/net/libeth/xsk.h
356
return head;
include/net/libeth/xsk.h
362
if (head)
include/net/libeth/xsk.h
363
return libeth_xsk_buff_add_frag(head, xdp);
include/net/libeth/xsk.h
80
for (const struct libeth_xdp_buff *head = xdp; ; ) {
include/net/libeth/xsk.h
81
xdp = container_of(xsk_buff_get_frag(&head->base),
include/net/mac80211.h
4895
struct sk_buff *head,
include/net/mana/gdma.h
326
u32 head;
include/net/neighbour.h
283
#define neigh_for_each_in_bucket(pos, head) hlist_for_each_entry(pos, head, hash)
include/net/neighbour.h
284
#define neigh_for_each_in_bucket_rcu(pos, head) \
include/net/neighbour.h
285
hlist_for_each_entry_rcu(pos, head, hash)
include/net/neighbour.h
286
#define neigh_for_each_in_bucket_safe(pos, tmp, head) \
include/net/neighbour.h
287
hlist_for_each_entry_safe(pos, tmp, head, hash)
include/net/netfilter/nf_conntrack_count.h
14
struct list_head head; /* connections with the same filtering key */
include/net/netfilter/nf_conntrack_helper.h
147
struct list_head head;
include/net/netlink.h
1212
#define nlmsg_for_each_msg(pos, head, len, rem) \
include/net/netlink.h
1213
for (pos = head, rem = len; \
include/net/netlink.h
2397
#define nla_for_each_attr(pos, head, len, rem) \
include/net/netlink.h
2398
for (pos = head, rem = len; \
include/net/netlink.h
2410
#define nla_for_each_attr_type(pos, type, head, len, rem) \
include/net/netlink.h
2411
nla_for_each_attr(pos, head, len, rem) \
include/net/netlink.h
541
int __nla_validate(const struct nlattr *head, int len, int maxtype,
include/net/netlink.h
544
int __nla_parse(struct nlattr **tb, int maxtype, const struct nlattr *head,
include/net/netlink.h
548
struct nlattr *nla_find(const struct nlattr *head, int len, int attrtype);
include/net/netlink.h
707
const struct nlattr *head, int len,
include/net/netlink.h
711
return __nla_parse(tb, maxtype, head, len, policy,
include/net/netlink.h
732
const struct nlattr *head, int len,
include/net/netlink.h
736
return __nla_parse(tb, maxtype, head, len, policy,
include/net/netlink.h
757
const struct nlattr *head,
include/net/netlink.h
762
return __nla_parse(tb, maxtype, head, len, policy,
include/net/netlink.h
884
static inline int nla_validate_deprecated(const struct nlattr *head, int len,
include/net/netlink.h
889
return __nla_validate(head, len, maxtype, policy, NL_VALIDATE_LIBERAL,
include/net/netlink.h
907
static inline int nla_validate(const struct nlattr *head, int len, int maxtype,
include/net/netlink.h
911
return __nla_validate(head, len, maxtype, policy, NL_VALIDATE_STRICT,
include/net/netns/ipv6.h
123
struct hlist_head head;
include/net/nexthop.h
262
void nexthop_free_rcu(struct rcu_head *head);
include/net/pkt_cls.h
551
ptr >= skb->head &&
include/net/rtnetlink.h
170
struct list_head *head);
include/net/sch_generic.h
1061
qh->head = skb;
include/net/sch_generic.h
1076
skb->next = qh->head;
include/net/sch_generic.h
1078
if (!qh->head)
include/net/sch_generic.h
1080
qh->head = skb;
include/net/sch_generic.h
1086
struct sk_buff *skb = qh->head;
include/net/sch_generic.h
1089
qh->head = skb->next;
include/net/sch_generic.h
1091
if (qh->head == NULL)
include/net/sch_generic.h
1221
return qh->head;
include/net/sch_generic.h
1299
rtnl_kfree_skbs(qh->head, qh->tail);
include/net/sch_generic.h
1301
qh->head = NULL;
include/net/sch_generic.h
355
struct list_head head;
include/net/sch_generic.h
60
struct sk_buff *head;
include/net/sch_generic.h
711
void dev_deactivate_many(struct list_head *head);
include/net/sctp/sctp.h
367
#define sctp_skb_for_each(pos, head, tmp) \
include/net/sctp/sctp.h
368
skb_queue_walk_safe(head, pos, tmp)
include/net/sctp/sctp.h
408
static inline int sctp_list_single_entry(struct list_head *head)
include/net/sctp/sctp.h
410
return list_is_singular(head);
include/net/sctp/sctp.h
504
#define sctp_for_each_hentry(ep, head) \
include/net/sctp/sctp.h
505
hlist_for_each_entry(ep, head, node)
include/net/sock.h
1135
WRITE_ONCE(sk->sk_backlog.head, skb);
include/net/sock.h
418
struct sk_buff *head;
include/net/sock.h
762
static inline struct sock *__sk_head(const struct hlist_head *head)
include/net/sock.h
764
return hlist_entry(head->first, struct sock, sk_node);
include/net/sock.h
767
static inline struct sock *sk_head(const struct hlist_head *head)
include/net/sock.h
769
return hlist_empty(head) ? NULL : __sk_head(head);
include/net/sock.h
772
static inline struct sock *__sk_nulls_head(const struct hlist_nulls_head *head)
include/net/sock.h
774
return hlist_nulls_entry(head->first, struct sock, sk_nulls_node);
include/net/sock.h
777
static inline struct sock *sk_nulls_head(const struct hlist_nulls_head *head)
include/net/sock.h
779
return hlist_nulls_empty(head) ? NULL : __sk_nulls_head(head);
include/net/sock.h
970
#define sk_for_each_entry_offset_rcu(tpos, pos, head, offset) \
include/net/sock.h
971
for (pos = rcu_dereference(hlist_first_rcu(head)); \
include/net/tcp.h
1953
struct hlist_head head;
include/net/tcp.h
2377
struct sk_buff *tcp_gro_lookup(struct list_head *head, struct tcphdr *th);
include/net/tcp.h
2378
struct sk_buff *tcp_gro_receive(struct list_head *head, struct sk_buff *skb,
include/net/tcp.h
2381
INDIRECT_CALLABLE_DECLARE(struct sk_buff *tcp4_gro_receive(struct list_head *head, struct sk_buff *skb));
include/net/tcp_ao.h
93
struct hlist_head head;
include/net/udp.h
62
struct hlist_head head;
include/net/xdp_sock_drv.h
160
static inline bool xsk_buff_add_frag(struct xdp_buff *head,
include/net/xdp_sock_drv.h
166
if (!__xdp_buff_add_frag(head, virt_to_netmem(data),
include/net/xdp_sock_drv.h
426
static inline bool xsk_buff_add_frag(struct xdp_buff *head,
include/net/xfrm.h
1844
struct sk_buff *xfrm4_gro_udp_encap_rcv(struct sock *sk, struct list_head *head,
include/net/xfrm.h
1846
struct sk_buff *xfrm6_gro_udp_encap_rcv(struct sock *sk, struct list_head *head,
include/rdma/rdmavt_cq.h
38
u32 head; /* index of next entry to fill */
include/rdma/rdmavt_qp.h
198
u32 head; /* new work requests posted to the head */
include/rdma/rdmavt_qp.h
277
static inline u32 rvt_get_rq_count(struct rvt_rq *rq, u32 head, u32 tail)
include/rdma/rdmavt_qp.h
279
u32 count = head - tail;
include/rdma/rdmavt_qp.h
944
RDMA_READ_UAPI_ATOMIC(cq->queue->head) :
include/rdma/rdmavt_qp.h
945
ibcq_to_rvtcq(send_cq)->kqueue->head;
include/trace/events/icmp.h
38
if (proto_4 != IPPROTO_UDP || (u8 *)uh < skb->head ||
include/trace/events/jbd2.h
101
__field( tid_t, head )
include/trace/events/jbd2.h
108
__entry->head = journal->j_tail_sequence;
include/trace/events/jbd2.h
113
__entry->transaction, __entry->sync_commit, __entry->head)
include/trace/events/mmflags.h
158
DEF_PAGEFLAG_NAME(head), \
include/trace/events/rpcrdma.h
1391
__entry->headlen = rqst->rq_rcv_buf.head[0].iov_len;
include/trace/events/rpcrdma.h
862
__entry->headlen = rqst->rq_snd_buf.head[0].iov_len;
include/trace/events/sunrpc.h
1588
__be32 *p = (__be32 *)xdr->head[0].iov_base;
include/trace/events/sunrpc.h
1592
__entry->head_len = xdr->head[0].iov_len;
include/trace/events/sunrpc.h
1638
__entry->head_base = xdr->head[0].iov_base;
include/trace/events/sunrpc.h
1639
__entry->head_len = xdr->head[0].iov_len;
include/trace/events/sunrpc.h
711
__entry->head_base = xdr->buf->head[0].iov_base,
include/trace/events/sunrpc.h
712
__entry->head_len = xdr->buf->head[0].iov_len,
include/trace/events/sunrpc.h
74
__entry->head_base = xdr->head[0].iov_base;
include/trace/events/sunrpc.h
75
__entry->head_len = xdr->head[0].iov_len;
include/trace/events/sunrpc.h
769
__entry->head_base = xdr->buf->head[0].iov_base,
include/trace/events/sunrpc.h
770
__entry->head_len = xdr->buf->head[0].iov_len,
include/trace/perf.h
26
struct hlist_head *head; \
include/trace/perf.h
33
head = this_cpu_ptr(event_call->perf_events); \
include/trace/perf.h
36
hlist_empty(head)) \
include/trace/perf.h
55
head, __task); \
include/uapi/linux/fd.h
17
head, /* nr of heads */
include/uapi/linux/fd.h
74
unsigned int device,head,track;
include/uapi/linux/fsmap.h
82
struct fsmap_head *head)
include/uapi/linux/fsmap.h
84
head->fmh_keys[0] = head->fmh_recs[head->fmh_entries - 1];
include/uapi/linux/io_uring.h
1070
__u32 head;
include/uapi/linux/io_uring.h
555
__u32 head;
include/uapi/linux/io_uring.h
574
__u32 head;
include/uapi/linux/io_uring.h
907
__u32 head; /* output */
include/uapi/linux/virtio_iommu.h
100
struct virtio_iommu_req_head head;
include/uapi/linux/virtio_iommu.h
110
struct virtio_iommu_req_head head;
include/uapi/linux/virtio_iommu.h
132
struct virtio_iommu_probe_property head;
include/uapi/linux/virtio_iommu.h
140
struct virtio_iommu_req_head head;
include/uapi/linux/virtio_iommu.h
75
struct virtio_iommu_req_head head;
include/uapi/linux/virtio_iommu.h
84
struct virtio_iommu_req_head head;
include/uapi/linux/virtio_rtc.h
104
struct virtio_rtc_resp_head head;
include/uapi/linux/virtio_rtc.h
113
struct virtio_rtc_req_head head;
include/uapi/linux/virtio_rtc.h
119
struct virtio_rtc_resp_head head;
include/uapi/linux/virtio_rtc.h
138
struct virtio_rtc_req_head head;
include/uapi/linux/virtio_rtc.h
145
struct virtio_rtc_resp_head head;
include/uapi/linux/virtio_rtc.h
154
struct virtio_rtc_req_head head;
include/uapi/linux/virtio_rtc.h
160
struct virtio_rtc_resp_head head;
include/uapi/linux/virtio_rtc.h
170
struct virtio_rtc_req_head head;
include/uapi/linux/virtio_rtc.h
179
struct virtio_rtc_resp_head head;
include/uapi/linux/virtio_rtc.h
186
struct virtio_rtc_req_head head;
include/uapi/linux/virtio_rtc.h
194
struct virtio_rtc_resp_head head;
include/uapi/linux/virtio_rtc.h
227
struct virtio_rtc_notif_head head;
include/uapi/linux/virtio_rtc.h
63
struct virtio_rtc_req_head head;
include/uapi/linux/virtio_rtc.h
69
struct virtio_rtc_resp_head head;
include/uapi/linux/virtio_rtc.h
76
struct virtio_rtc_req_head head;
include/uapi/linux/virtio_rtc.h
89
struct virtio_rtc_resp_head head;
include/uapi/linux/virtio_rtc.h
99
struct virtio_rtc_req_head head;
include/uapi/rdma/rvt-abi.h
30
RDMA_ATOMIC_UAPI(__u32, head);
include/uapi/rdma/rvt-abi.h
61
RDMA_ATOMIC_UAPI(__u32, head);
include/xen/grant_table.h
134
void gnttab_free_grant_references(grant_ref_t head);
include/xen/grant_table.h
136
void gnttab_free_grant_reference_seq(grant_ref_t head, unsigned int count);
include/xen/interface/event_channel.h
276
event_word_t head[EVTCHN_FIFO_MAX_QUEUES];
init/initramfs.c
122
for (p = head; hardlink_seen && p < head + 32; p++) {
init/initramfs.c
80
} *head[32];
init/initramfs.c
94
for (p = head + hash(major, minor, ino); *p; p = &(*p)->next) {
io_uring/bpf_filter.c
103
filters = container_of(head, struct io_bpf_filters, rcu_head);
io_uring/bpf_filter.c
97
static void io_free_bpf_filters(struct rcu_head *head)
io_uring/cancel.c
392
static bool io_match_linked(struct io_kiocb *head)
io_uring/cancel.c
396
io_for_each_link(req, head) {
io_uring/cancel.c
407
bool io_match_task_safe(struct io_kiocb *head, struct io_uring_task *tctx,
io_uring/cancel.c
412
if (tctx && head->tctx != tctx)
io_uring/cancel.c
417
if (head->flags & REQ_F_LINK_TIMEOUT) {
io_uring/cancel.c
418
struct io_ring_ctx *ctx = head->ctx;
io_uring/cancel.c
422
matched = io_match_linked(head);
io_uring/cancel.c
425
matched = io_match_linked(head);
io_uring/cancel.h
24
bool io_match_task_safe(struct io_kiocb *head, struct io_uring_task *tctx,
io_uring/fdinfo.c
65
unsigned int sq_head = READ_ONCE(r->sq.head);
io_uring/fdinfo.c
67
unsigned int cq_head = READ_ONCE(r->cq.head);
io_uring/io_uring.c
1692
struct io_kiocb *head = ctx->submit_state.link.head;
io_uring/io_uring.c
1695
if (head) {
io_uring/io_uring.c
1703
head->flags |= REQ_F_IO_DRAIN | REQ_F_FORCE_ASYNC;
io_uring/io_uring.c
1784
if (unlikely(ctx->drain_next) && !ctx->submit_state.link.head) {
io_uring/io_uring.c
1836
struct io_kiocb *head = link->head;
io_uring/io_uring.c
1847
if (head && !(head->flags & REQ_F_FAIL))
io_uring/io_uring.c
1848
req_fail_link_node(head, -ECANCELED);
io_uring/io_uring.c
1851
if (head) {
io_uring/io_uring.c
1853
link->head = NULL;
io_uring/io_uring.c
1854
req = head;
io_uring/io_uring.c
1860
if (head)
io_uring/io_uring.c
1863
link->head = req;
io_uring/io_uring.c
1894
if (unlikely(link->head)) {
io_uring/io_uring.c
1903
req = link->head;
io_uring/io_uring.c
1904
link->head = NULL;
io_uring/io_uring.c
1911
link->head = req;
io_uring/io_uring.c
1931
if (unlikely(state->link.head))
io_uring/io_uring.c
1932
io_queue_sqe_fallback(state->link.head);
io_uring/io_uring.c
1949
state->link.head = NULL;
io_uring/io_uring.c
1964
smp_store_release(&rings->sq.head, ctx->cached_sq_head);
io_uring/io_uring.c
1979
unsigned head = ctx->cached_sq_head++ & mask;
io_uring/io_uring.c
1983
head = READ_ONCE(ctx->sq_array[head]);
io_uring/io_uring.c
1984
if (unlikely(head >= ctx->sq_entries)) {
io_uring/io_uring.c
1989
head = array_index_nospec(head, ctx->sq_entries);
io_uring/io_uring.c
2003
head <<= 1;
io_uring/io_uring.c
2004
*sqe = &ctx->sq_sqes[head];
io_uring/io_uring.c
2882
p->sq_off.head = offsetof(struct io_rings, sq.head);
io_uring/io_uring.c
2892
p->cq_off.head = offsetof(struct io_rings, cq.head);
io_uring/io_uring.h
253
#define io_for_each_link(pos, head) \
io_uring/io_uring.h
254
for (pos = (head); pos; pos = pos->link)
io_uring/io_uring.h
457
return READ_ONCE(r->sq.tail) - READ_ONCE(r->sq.head) == ctx->sq_entries;
io_uring/kbuf.c
197
__u16 tail, head = bl->head;
io_uring/kbuf.c
203
if (unlikely(tail == head))
io_uring/kbuf.c
206
if (head + 1 == tail)
io_uring/kbuf.c
209
buf = io_ring_head_to_buf(br, head, bl->mask);
io_uring/kbuf.c
24
#define io_ring_head_to_buf(br, head, mask) &(br)->bufs[(head) & (mask)]
io_uring/kbuf.c
255
__u16 nr_avail, tail, head;
io_uring/kbuf.c
259
head = bl->head;
io_uring/kbuf.c
260
nr_avail = min_t(__u16, tail - head, UIO_MAXIOV);
io_uring/kbuf.c
264
buf = io_ring_head_to_buf(br, head, bl->mask);
io_uring/kbuf.c
321
buf = io_ring_head_to_buf(br, ++head, bl->mask);
io_uring/kbuf.c
324
if (head == tail)
io_uring/kbuf.c
45
buf = io_ring_head_to_buf(bl->buf_ring, bl->head, bl->mask);
io_uring/kbuf.c
56
bl->head++;
io_uring/kbuf.c
739
buf_status.head = bl->head;
io_uring/kbuf.c
74
bl->head += nr;
io_uring/kbuf.h
31
__u16 head;
io_uring/msg_ring.c
208
static void io_msg_tw_fd_complete(struct callback_head *head)
io_uring/msg_ring.c
210
struct io_msg *msg = container_of(head, struct io_msg, tw);
io_uring/notif.c
102
nd->head = prev_nd->head;
io_uring/notif.c
105
net_zcopy_get(&nd->head->uarg);
io_uring/notif.c
135
nd->head = nd;
io_uring/notif.c
60
if (nd->head != nd) {
io_uring/notif.c
61
io_tx_ubuf_complete(skb, &nd->head->uarg, success);
io_uring/notif.c
88
if (unlikely(nd->head != nd || nd->next))
io_uring/notif.h
18
struct io_notif_data *head;
io_uring/poll.c
133
poll->head = NULL;
io_uring/poll.c
157
smp_store_release(&poll->head, NULL);
io_uring/poll.c
162
struct wait_queue_head *head = smp_load_acquire(&poll->head);
io_uring/poll.c
164
if (head) {
io_uring/poll.c
165
spin_lock_irq(&head->lock);
io_uring/poll.c
167
spin_unlock_irq(&head->lock);
io_uring/poll.c
437
struct wait_queue_head *head;
io_uring/poll.c
442
head = smp_load_acquire(&poll->head);
io_uring/poll.c
449
if (head) {
io_uring/poll.c
450
spin_lock_irq(&head->lock);
io_uring/poll.c
454
spin_unlock_irq(&head->lock);
io_uring/poll.c
457
return !!head;
io_uring/poll.c
461
struct wait_queue_head *head,
io_uring/poll.c
476
if (first->head == head)
io_uring/poll.c
480
if ((*poll_ptr)->head == head)
io_uring/poll.c
507
poll->head = head;
io_uring/poll.c
511
add_wait_queue_exclusive(head, &poll->wait);
io_uring/poll.c
513
add_wait_queue(head, &poll->wait);
io_uring/poll.c
517
static void io_poll_queue_proc(struct file *file, struct wait_queue_head *head,
io_uring/poll.c
523
__io_queue_proc(poll, pt, head,
io_uring/poll.c
633
static void io_async_queue_proc(struct file *file, struct wait_queue_head *head,
io_uring/poll.c
639
__io_queue_proc(&apoll->poll, pt, head, &apoll->double_poll);
io_uring/poll.h
15
struct wait_queue_head *head;
io_uring/register.c
598
old_head = READ_ONCE(o.rings->sq.head);
io_uring/register.c
607
WRITE_ONCE(n.rings->sq.head, old_head);
io_uring/register.c
611
old_head = READ_ONCE(o.rings->cq.head);
io_uring/register.c
627
WRITE_ONCE(n.rings->cq.head, old_head);
io_uring/slist.h
10
#define wq_list_for_each(pos, prv, head) \
io_uring/slist.h
11
for (pos = (head)->first, prv = NULL; pos; prv = pos, pos = (pos)->next)
io_uring/slist.h
7
#define __wq_list_for_each(pos, head) \
io_uring/slist.h
8
for (pos = (head)->first; pos; pos = (pos)->next)
io_uring/timeout.c
23
struct io_kiocb *head;
io_uring/timeout.c
248
timeout->head = NULL;
io_uring/timeout.c
360
prev = timeout->head;
io_uring/timeout.c
361
timeout->head = NULL;
io_uring/timeout.c
571
if (!link->head)
io_uring/timeout.c
575
timeout->head = link->last;
io_uring/timeout.c
655
if (timeout->head) {
io_uring/timeout.c
667
static bool io_match_task(struct io_kiocb *head, struct io_uring_task *tctx,
io_uring/timeout.c
669
__must_hold(&head->ctx->timeout_lock)
io_uring/timeout.c
673
if (tctx && head->tctx != tctx)
io_uring/timeout.c
678
io_for_each_link(req, head) {
io_uring/tw.c
174
struct llist_node *head;
io_uring/tw.c
188
head = READ_ONCE(ctx->work_llist.first);
io_uring/tw.c
191
if (head) {
io_uring/tw.c
192
struct io_kiocb *first_req = container_of(head,
io_uring/tw.c
211
req->io_task_work.node.next = head;
io_uring/tw.c
212
} while (!try_cmpxchg(&ctx->work_llist.first, &head,
io_uring/tw.c
223
if (!head) {
io_uring/wait.c
219
iowq.cq_tail = READ_ONCE(rings->cq.head) + min_events;
io_uring/wait.c
319
return READ_ONCE(io_get_rings(ctx)->cq.head) == READ_ONCE(io_get_rings(ctx)->cq.tail) ? ret : 0;
io_uring/wait.h
32
return ctx->cached_cq_tail - READ_ONCE(rings->cq.head);
io_uring/wait.h
39
return READ_ONCE(rings->cq.tail) - READ_ONCE(rings->cq.head);
io_uring/waitid.c
115
struct wait_queue_head *head;
io_uring/waitid.c
117
head = smp_load_acquire(&iw->head);
io_uring/waitid.c
118
if (head) {
io_uring/waitid.c
121
smp_store_release(&iw->head, NULL);
io_uring/waitid.c
122
spin_lock_irq(&head->lock);
io_uring/waitid.c
124
spin_unlock_irq(&head->lock);
io_uring/waitid.c
220
iw->head = &current->signal->wait_chldexit;
io_uring/waitid.c
221
add_wait_queue(iw->head, &iwa->wo.child_wait);
io_uring/waitid.c
249
smp_store_release(&iw->head, NULL);
io_uring/waitid.c
276
iw->head = NULL;
io_uring/waitid.c
30
struct wait_queue_head *head;
io_uring/waitid.c
312
iw->head = &current->signal->wait_chldexit;
io_uring/waitid.c
317
add_wait_queue(iw->head, &iwa->wo.child_wait);
io_uring/zcrx.c
1118
smp_store_release(&zcrx->rq_ring->head, zcrx->cached_rq_head);
io_uring/zcrx.c
370
offsets->head = offsetof(struct io_uring, head);
io_uring/zcrx.c
985
smp_store_release(&ifq->rq_ring->head, ifq->cached_rq_head);
ipc/ipc_sysctl.c
193
static void ipc_set_ownership(struct ctl_table_header *head,
ipc/ipc_sysctl.c
197
container_of(head->set, struct ipc_namespace, ipc_set);
ipc/ipc_sysctl.c
206
static int ipc_permissions(struct ctl_table_header *head, const struct ctl_table *table)
ipc/ipc_sysctl.c
212
container_of(head->set, struct ipc_namespace, ipc_set);
ipc/ipc_sysctl.c
225
ipc_set_ownership(head, &ns_root_uid, &ns_root_gid);
ipc/mq_sysctl.c
79
static void mq_set_ownership(struct ctl_table_header *head,
ipc/mq_sysctl.c
83
container_of(head->set, struct ipc_namespace, mq_set);
ipc/mq_sysctl.c
92
static int mq_permissions(struct ctl_table_header *head, const struct ctl_table *table)
ipc/mq_sysctl.c
98
mq_set_ownership(head, &ns_root_uid, &ns_root_gid);
ipc/msg.c
128
static void msg_rcu_free(struct rcu_head *head)
ipc/msg.c
130
struct kern_ipc_perm *p = container_of(head, struct kern_ipc_perm, rcu);
ipc/sem.c
323
static void sem_rcu_free(struct rcu_head *head)
ipc/sem.c
325
struct kern_ipc_perm *p = container_of(head, struct kern_ipc_perm, rcu);
ipc/shm.c
231
static void shm_rcu_free(struct rcu_head *head)
ipc/shm.c
233
struct kern_ipc_perm *ptr = container_of(head, struct kern_ipc_perm,
ipc/util.c
534
void (*func)(struct rcu_head *head))
ipc/util.h
174
void (*func)(struct rcu_head *head));
kernel/audit.h
256
extern void audit_free_rule_rcu(struct rcu_head *head);
kernel/audit_tree.c
121
kfree_rcu(tree, head);
kernel/audit_tree.c
149
struct audit_chunk *chunk = container_of(rcu, struct audit_chunk, head);
kernel/audit_tree.c
160
call_rcu(&chunk->head, __put_chunk);
kernel/audit_tree.c
21
struct rcu_head head;
kernel/audit_tree.c
32
struct rcu_head head;
kernel/auditfilter.c
101
struct audit_entry *e = container_of(head, struct audit_entry, rcu);
kernel/auditfilter.c
99
void audit_free_rule_rcu(struct rcu_head *head)
kernel/bpf/arena.c
319
struct list_head head;
kernel/bpf/arena.c
333
list_add(&vml->head, &arena->vma_list);
kernel/bpf/arena.c
354
list_del(&vml->head);
kernel/bpf/arena.c
658
list_for_each_entry(vml, &arena->vma_list, head)
kernel/bpf/btf.c
1311
return show->obj.head + show->obj.size - data;
kernel/bpf/btf.c
1386
show->obj.head = data;
kernel/bpf/btf.c
439
void *head;
kernel/bpf/cgroup.c
1084
struct hlist_head *head;
kernel/bpf/cgroup.c
1100
head = &cg->bpf.progs[atype];
kernel/bpf/cgroup.c
1101
hlist_for_each_entry(pl, head, node) {
kernel/bpf/cgroup.c
1889
int __cgroup_bpf_run_filter_sysctl(struct ctl_table_header *head,
kernel/bpf/cgroup.c
1895
.head = head,
kernel/bpf/cgroup.c
2252
if (!ctx->head)
kernel/bpf/cgroup.c
2254
tmp_ret = sysctl_cpy_dir(ctx->head->parent, &buf, &buf_len);
kernel/bpf/cgroup.c
384
static u32 prog_list_length(struct hlist_head *head, int *preorder_cnt)
kernel/bpf/cgroup.c
389
hlist_for_each_entry(pl, head, node) {
kernel/bpf/cgroup.c
932
struct hlist_head *head;
kernel/bpf/cgroup.c
947
head = &cg->bpf.progs[atype];
kernel/bpf/cgroup.c
948
hlist_for_each_entry(pl, head, node) {
kernel/bpf/core.c
89
if (ptr >= skb->head && ptr + size <= skb_tail_pointer(skb))
kernel/bpf/crypto.c
223
static void crypto_free_cb(struct rcu_head *head)
kernel/bpf/crypto.c
227
ctx = container_of(head, struct bpf_crypto_ctx, rcu);
kernel/bpf/devmap.c
1108
struct hlist_head *head;
kernel/bpf/devmap.c
1111
head = dev_map_index_hash(dtab, i);
kernel/bpf/devmap.c
1113
hlist_for_each_entry_safe(dev, next, head, index_hlist) {
kernel/bpf/devmap.c
221
struct hlist_head *head;
kernel/bpf/devmap.c
224
head = dev_map_index_hash(dtab, i);
kernel/bpf/devmap.c
226
hlist_for_each_entry_safe(dev, next, head, index_hlist) {
kernel/bpf/devmap.c
280
struct hlist_head *head = dev_map_index_hash(dtab, key);
kernel/bpf/devmap.c
283
hlist_for_each_entry_rcu(dev, head, index_hlist,
kernel/bpf/devmap.c
297
struct hlist_head *head;
kernel/bpf/devmap.c
322
head = dev_map_index_hash(dtab, i);
kernel/bpf/devmap.c
324
next_dev = hlist_entry_safe(rcu_dereference_raw(hlist_first_rcu(head)),
kernel/bpf/devmap.c
628
struct hlist_head *head;
kernel/bpf/devmap.c
666
head = dev_map_index_hash(dtab, i);
kernel/bpf/devmap.c
667
hlist_for_each_entry_rcu(dst, head, index_hlist,
kernel/bpf/devmap.c
749
struct hlist_head *head;
kernel/bpf/devmap.c
789
head = dev_map_index_hash(dtab, i);
kernel/bpf/devmap.c
790
hlist_for_each_entry_safe(dst, next, head, index_hlist) {
kernel/bpf/hashtab.c
1172
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
1189
head = &b->head;
kernel/bpf/hashtab.c
1195
l_old = lookup_nulls_elem_raw(head, hash, key, key_size,
kernel/bpf/hashtab.c
1217
l_old = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1248
hlist_nulls_add_head_rcu(&l_new->hash_node, head);
kernel/bpf/hashtab.c
1279
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
1296
head = &b->head;
kernel/bpf/hashtab.c
1312
l_old = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1321
hlist_nulls_add_head_rcu(&l_new->hash_node, head);
kernel/bpf/hashtab.c
1355
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
1373
head = &b->head;
kernel/bpf/hashtab.c
1379
l_old = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1403
hlist_nulls_add_head_rcu(&l_new->hash_node, head);
kernel/bpf/hashtab.c
1418
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
143
INIT_HLIST_NULLS_HEAD(&htab->buckets[i].head, i);
kernel/bpf/hashtab.c
1435
head = &b->head;
kernel/bpf/hashtab.c
1452
l_old = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1467
hlist_nulls_add_head_rcu(&l_new->hash_node, head);
kernel/bpf/hashtab.c
1498
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
1511
head = &b->head;
kernel/bpf/hashtab.c
1517
l = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1533
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
1546
head = &b->head;
kernel/bpf/hashtab.c
1552
l = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1573
struct hlist_nulls_head *head = select_bucket(htab, i);
kernel/bpf/hashtab.c
1577
hlist_nulls_for_each_entry_safe(l, n, head, hash_node) {
kernel/bpf/hashtab.c
1591
struct hlist_nulls_head *head = select_bucket(htab, i);
kernel/bpf/hashtab.c
1595
hlist_nulls_for_each_entry(l, n, head, hash_node) {
kernel/bpf/hashtab.c
1676
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
1687
head = &b->head;
kernel/bpf/hashtab.c
1693
l = lookup_elem_raw(head, hash, key, key_size);
kernel/bpf/hashtab.c
1781
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
1844
head = &b->head;
kernel/bpf/hashtab.c
1856
hlist_nulls_for_each_entry_rcu(l, n, head, hash_node)
kernel/bpf/hashtab.c
1893
hlist_nulls_for_each_entry_safe(l, n, head, hash_node) {
kernel/bpf/hashtab.c
2090
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
2120
head = &b->head;
kernel/bpf/hashtab.c
2121
hlist_nulls_for_each_entry_rcu(elem, n, head, hash_node) {
kernel/bpf/hashtab.c
2263
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
2287
head = &b->head;
kernel/bpf/hashtab.c
2288
hlist_nulls_for_each_entry_safe(elem, n, head, hash_node) {
kernel/bpf/hashtab.c
2608
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
2613
head = select_bucket(htab, i);
kernel/bpf/hashtab.c
2615
hlist_nulls_for_each_entry_safe(l, n, head, hash_node) {
kernel/bpf/hashtab.c
688
return &__select_bucket(htab, hash)->head;
kernel/bpf/hashtab.c
692
static struct htab_elem *lookup_elem_raw(struct hlist_nulls_head *head, u32 hash,
kernel/bpf/hashtab.c
698
hlist_nulls_for_each_entry_rcu(l, n, head, hash_node)
kernel/bpf/hashtab.c
709
static struct htab_elem *lookup_nulls_elem_raw(struct hlist_nulls_head *head,
kernel/bpf/hashtab.c
717
hlist_nulls_for_each_entry_rcu(l, n, head, hash_node)
kernel/bpf/hashtab.c
735
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
745
head = select_bucket(htab, hash);
kernel/bpf/hashtab.c
747
l = lookup_nulls_elem_raw(head, hash, key, key_size, htab->n_buckets);
kernel/bpf/hashtab.c
81
struct hlist_nulls_head head;
kernel/bpf/hashtab.c
863
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
871
head = &b->head;
kernel/bpf/hashtab.c
877
hlist_nulls_for_each_entry_rcu(l, n, head, hash_node)
kernel/bpf/hashtab.c
895
struct hlist_nulls_head *head;
kernel/bpf/hashtab.c
909
head = select_bucket(htab, hash);
kernel/bpf/hashtab.c
912
l = lookup_nulls_elem_raw(head, hash, key, key_size, htab->n_buckets);
kernel/bpf/hashtab.c
934
head = select_bucket(htab, i);
kernel/bpf/hashtab.c
937
next_l = hlist_nulls_entry_safe(rcu_dereference_raw(hlist_nulls_first_rcu(head)),
kernel/bpf/helpers.c
2235
struct list_head *head = list_head, *orig_head = list_head;
kernel/bpf/helpers.c
2247
if (!head->next || list_empty(head))
kernel/bpf/helpers.c
2249
head = head->next;
kernel/bpf/helpers.c
2254
while (head != orig_head) {
kernel/bpf/helpers.c
2255
void *obj = head;
kernel/bpf/helpers.c
2258
head = head->next;
kernel/bpf/helpers.c
2383
struct bpf_list_head *head,
kernel/bpf/helpers.c
2386
struct list_head *n = &node->list_head, *h = (void *)head;
kernel/bpf/helpers.c
2404
WRITE_ONCE(node->owner, head);
kernel/bpf/helpers.c
2409
__bpf_kfunc int bpf_list_push_front_impl(struct bpf_list_head *head,
kernel/bpf/helpers.c
2416
return __bpf_list_add(n, head, false, meta ? meta->record : NULL, off);
kernel/bpf/helpers.c
2419
__bpf_kfunc int bpf_list_push_back_impl(struct bpf_list_head *head,
kernel/bpf/helpers.c
2426
return __bpf_list_add(n, head, true, meta ? meta->record : NULL, off);
kernel/bpf/helpers.c
2429
static struct bpf_list_node *__bpf_list_del(struct bpf_list_head *head, bool tail)
kernel/bpf/helpers.c
2431
struct list_head *n, *h = (void *)head;
kernel/bpf/helpers.c
2444
if (WARN_ON_ONCE(READ_ONCE(node->owner) != head))
kernel/bpf/helpers.c
2452
__bpf_kfunc struct bpf_list_node *bpf_list_pop_front(struct bpf_list_head *head)
kernel/bpf/helpers.c
2454
return __bpf_list_del(head, false);
kernel/bpf/helpers.c
2457
__bpf_kfunc struct bpf_list_node *bpf_list_pop_back(struct bpf_list_head *head)
kernel/bpf/helpers.c
2459
return __bpf_list_del(head, true);
kernel/bpf/helpers.c
2462
__bpf_kfunc struct bpf_list_node *bpf_list_front(struct bpf_list_head *head)
kernel/bpf/helpers.c
2464
struct list_head *h = (struct list_head *)head;
kernel/bpf/helpers.c
2472
__bpf_kfunc struct bpf_list_node *bpf_list_back(struct bpf_list_head *head)
kernel/bpf/helpers.c
2474
struct list_head *h = (struct list_head *)head;
kernel/bpf/memalloc.c
130
static struct llist_node notrace *__llist_del_first(struct llist_head *head)
kernel/bpf/memalloc.c
134
entry = head->first;
kernel/bpf/memalloc.c
138
head->first = next;
kernel/bpf/memalloc.c
279
static void __free_rcu(struct rcu_head *head)
kernel/bpf/memalloc.c
281
struct bpf_mem_cache *c = container_of(head, struct bpf_mem_cache, rcu_ttrace);
kernel/bpf/memalloc.c
287
static void __free_rcu_tasks_trace(struct rcu_head *head)
kernel/bpf/memalloc.c
293
__free_rcu(head);
kernel/bpf/memalloc.c
295
call_rcu(head, __free_rcu);
kernel/bpf/memalloc.c
365
static void __free_by_rcu(struct rcu_head *head)
kernel/bpf/memalloc.c
367
struct bpf_mem_cache *c = container_of(head, struct bpf_mem_cache, rcu);
kernel/bpf/percpu_freelist.c
103
struct pcpu_freelist_head *head;
kernel/bpf/percpu_freelist.c
107
head = per_cpu_ptr(s->freelist, cpu);
kernel/bpf/percpu_freelist.c
108
if (!READ_ONCE(head->first))
kernel/bpf/percpu_freelist.c
110
if (raw_res_spin_lock(&head->lock))
kernel/bpf/percpu_freelist.c
112
node = head->first;
kernel/bpf/percpu_freelist.c
114
WRITE_ONCE(head->first, node->next);
kernel/bpf/percpu_freelist.c
115
raw_res_spin_unlock(&head->lock);
kernel/bpf/percpu_freelist.c
118
raw_res_spin_unlock(&head->lock);
kernel/bpf/percpu_freelist.c
15
struct pcpu_freelist_head *head = per_cpu_ptr(s->freelist, cpu);
kernel/bpf/percpu_freelist.c
17
raw_res_spin_lock_init(&head->lock);
kernel/bpf/percpu_freelist.c
18
head->first = NULL;
kernel/bpf/percpu_freelist.c
28
static inline void pcpu_freelist_push_node(struct pcpu_freelist_head *head,
kernel/bpf/percpu_freelist.c
31
node->next = head->first;
kernel/bpf/percpu_freelist.c
32
WRITE_ONCE(head->first, node);
kernel/bpf/percpu_freelist.c
35
static inline bool ___pcpu_freelist_push(struct pcpu_freelist_head *head,
kernel/bpf/percpu_freelist.c
38
if (raw_res_spin_lock(&head->lock))
kernel/bpf/percpu_freelist.c
40
pcpu_freelist_push_node(head, node);
kernel/bpf/percpu_freelist.c
41
raw_res_spin_unlock(&head->lock);
kernel/bpf/percpu_freelist.c
48
struct pcpu_freelist_head *head;
kernel/bpf/percpu_freelist.c
58
head = per_cpu_ptr(s->freelist, cpu);
kernel/bpf/percpu_freelist.c
59
if (raw_res_spin_lock(&head->lock))
kernel/bpf/percpu_freelist.c
61
pcpu_freelist_push_node(head, node);
kernel/bpf/percpu_freelist.c
62
raw_res_spin_unlock(&head->lock);
kernel/bpf/percpu_freelist.c
81
struct pcpu_freelist_head *head;
kernel/bpf/percpu_freelist.c
89
head = per_cpu_ptr(s->freelist, cpu);
kernel/bpf/percpu_freelist.c
93
pcpu_freelist_push_node(head, buf);
kernel/bpf/queue_stack_maps.c
142
index = qs->head - 1;
kernel/bpf/queue_stack_maps.c
150
qs->head = index;
kernel/bpf/queue_stack_maps.c
20
u32 head, tail;
kernel/bpf/queue_stack_maps.c
212
dst = &qs->elements[qs->head * qs->map.value_size];
kernel/bpf/queue_stack_maps.c
215
if (unlikely(++qs->head >= qs->size))
kernel/bpf/queue_stack_maps.c
216
qs->head = 0;
kernel/bpf/queue_stack_maps.c
33
return qs->head == qs->tail;
kernel/bpf/queue_stack_maps.c
38
u32 head = qs->head + 1;
kernel/bpf/queue_stack_maps.c
40
if (unlikely(head >= qs->size))
kernel/bpf/queue_stack_maps.c
41
head = 0;
kernel/bpf/queue_stack_maps.c
43
return head == qs->tail;
kernel/bpf/stream.c
124
struct llist_node *head, *tail;
kernel/bpf/stream.c
131
head = llist_reverse_order(tail);
kernel/bpf/stream.c
134
stream->backlog_head = head;
kernel/bpf/stream.c
137
stream->backlog_tail->next = head;
kernel/bpf/stream.c
332
struct llist_node *list, *head, *tail;
kernel/bpf/stream.c
345
head = tail = list;
kernel/bpf/stream.c
353
llist_add_batch(head, tail, &stream->log);
kernel/bpf/trampoline.c
334
struct hlist_head *head;
kernel/bpf/trampoline.c
338
head = &trampoline_key_table[hash_64(key, TRAMPOLINE_HASH_BITS)];
kernel/bpf/trampoline.c
339
hlist_for_each_entry(tr, head, hlist_key) {
kernel/bpf/trampoline.c
358
hlist_add_head(&tr->hlist_key, head);
kernel/bpf/trampoline.c
359
head = &trampoline_ip_table[hash_64(tr->ip, TRAMPOLINE_HASH_BITS)];
kernel/bpf/trampoline.c
360
hlist_add_head(&tr->hlist_ip, head);
kernel/bpf/verifier.c
19781
struct list_head *pos, *head;
kernel/bpf/verifier.c
19783
head = explored_state(env, insn);
kernel/bpf/verifier.c
19784
list_for_each(pos, head) {
kernel/bpf/verifier.c
20483
struct list_head *pos, *tmp, *head;
kernel/bpf/verifier.c
20505
head = explored_state(env, insn_idx);
kernel/bpf/verifier.c
20506
list_for_each_safe(pos, tmp, head) {
kernel/bpf/verifier.c
20832
list_add(&new_sl->node, head);
kernel/bpf/verifier.c
2092
struct bpf_verifier_stack_elem *elem, *head = env->head;
kernel/bpf/verifier.c
2095
if (env->head == NULL)
kernel/bpf/verifier.c
2099
err = copy_verifier_state(cur, &head->st);
kernel/bpf/verifier.c
2104
bpf_vlog_reset(&env->log, head->log_pos);
kernel/bpf/verifier.c
2106
*insn_idx = head->insn_idx;
kernel/bpf/verifier.c
2108
*prev_insn_idx = head->prev_insn_idx;
kernel/bpf/verifier.c
2109
elem = head->next;
kernel/bpf/verifier.c
2110
free_verifier_state(&head->st, false);
kernel/bpf/verifier.c
2111
kfree(head);
kernel/bpf/verifier.c
2112
env->head = elem;
kernel/bpf/verifier.c
2143
elem->next = env->head;
kernel/bpf/verifier.c
2145
env->head = elem;
kernel/bpf/verifier.c
24532
struct list_head *head, *pos, *tmp;
kernel/bpf/verifier.c
24561
head = &env->explored_states[i];
kernel/bpf/verifier.c
24563
list_for_each_safe(pos, tmp, head) {
kernel/bpf/verifier.c
3019
elem->next = env->head;
kernel/bpf/verifier.c
3021
env->head = elem;
kernel/bpf/verifier.c
9063
struct list_head *pos, *head;
kernel/bpf/verifier.c
9066
head = explored_state(env, insn_idx);
kernel/bpf/verifier.c
9067
list_for_each(pos, head) {
kernel/cgroup/cpuset.c
2557
static void flush_migrate_mm_task_workfn(struct callback_head *head)
kernel/cgroup/cpuset.c
2560
kfree(head);
kernel/cgroup/rstat.c
199
struct cgroup_subsys_state *head,
kernel/cgroup/rstat.c
214
lockdep_assert_held(ss_rstat_lock(head->ss));
kernel/cgroup/rstat.c
240
child->rstat_flush_next = head;
kernel/cgroup/rstat.c
241
head = child;
kernel/cgroup/rstat.c
260
return head;
kernel/cgroup/rstat.c
285
struct cgroup_subsys_state *head = NULL, *parent, *child;
kernel/cgroup/rstat.c
317
head = root;
kernel/cgroup/rstat.c
322
head = css_rstat_push_children(head, child, cpu);
kernel/cgroup/rstat.c
324
return head;
kernel/events/callchain.c
50
static void release_callchain_buffers_rcu(struct rcu_head *head)
kernel/events/callchain.c
55
entries = container_of(head, struct callchain_cpus_entries, rcu_head);
kernel/events/core.c
10996
struct hlist_head *head;
kernel/events/core.c
10999
head = find_swevent_head_rcu(swhash, type, event_id);
kernel/events/core.c
11000
if (!head)
kernel/events/core.c
11003
hlist_for_each_entry_rcu(event, head, hlist_entry) {
kernel/events/core.c
11059
struct hlist_head *head;
kernel/events/core.c
11068
head = find_swevent_head(swhash, event);
kernel/events/core.c
11069
if (WARN_ON_ONCE(!head))
kernel/events/core.c
11072
hlist_add_head_rcu(&event->hlist_entry, head);
kernel/events/core.c
11324
struct pt_regs *regs, struct hlist_head *head,
kernel/events/core.c
11329
if (!trace_call_bpf(call, raw_data) || hlist_empty(head)) {
kernel/events/core.c
11334
perf_tp_event(call->event.type, count, raw_data, size, regs, head,
kernel/events/core.c
11383
struct pt_regs *regs, struct hlist_head *head, int rctx,
kernel/events/core.c
11403
hlist_for_each_entry_rcu(event, head, hlist_entry) {
kernel/events/core.c
11772
struct list_head *head)
kernel/events/core.c
11787
if (head)
kernel/events/core.c
11788
list_splice(head, &event->addr_filters.list);
kernel/events/core.c
1375
static void free_ctx(struct rcu_head *head)
kernel/events/core.c
1379
ctx = container_of(head, struct perf_event_context, rcu_head);
kernel/events/core.c
5248
static void free_cpc_rcu(struct rcu_head *head)
kernel/events/core.c
5251
container_of(head, typeof(*cpc), epc.rcu_head);
kernel/events/core.c
5256
static void free_epc_rcu(struct rcu_head *head)
kernel/events/core.c
5258
struct perf_event_pmu_context *epc = container_of(head, typeof(*epc), rcu_head);
kernel/events/core.c
5299
static void free_event_rcu(struct rcu_head *head)
kernel/events/core.c
5301
struct perf_event *event = container_of(head, typeof(*event), rcu_head);
kernel/events/core.c
7645
static void perf_pending_task(struct callback_head *head)
kernel/events/core.c
7647
struct perf_event *event = container_of(head, struct perf_event, pending_task);
kernel/events/core.c
9924
void perf_event_aux_event(struct perf_event *event, unsigned long head,
kernel/events/core.c
9940
.offset = head,
kernel/events/hw_breakpoint.c
325
struct rhlist_head *head, *pos;
kernel/events/hw_breakpoint.c
335
head = rhltable_lookup(&task_bps_ht, &bp->hw.target, task_bps_ht_params);
kernel/events/hw_breakpoint.c
336
if (!head)
kernel/events/hw_breakpoint.c
339
rhl_for_each_entry_rcu(iter, pos, head, hw.bp_list) {
kernel/events/internal.h
26
local_t head; /* write position */
kernel/events/internal.h
95
void perf_event_aux_event(struct perf_event *event, unsigned long head,
kernel/events/ring_buffer.c
114
WRITE_ONCE(rb->user_page->data_head, head);
kernel/events/ring_buffer.c
129
if (unlikely(head != local_read(&rb->head))) {
kernel/events/ring_buffer.c
142
ring_buffer_has_space(unsigned long head, unsigned long tail,
kernel/events/ring_buffer.c
147
return CIRC_SPACE(head, tail, data_size) >= size;
kernel/events/ring_buffer.c
149
return CIRC_SPACE(tail, head, data_size) >= size;
kernel/events/ring_buffer.c
159
unsigned long tail, offset, head;
kernel/events/ring_buffer.c
199
offset = local_read(&rb->head);
kernel/events/ring_buffer.c
201
head = offset;
kernel/events/ring_buffer.c
204
if (unlikely(!ring_buffer_has_space(head, tail,
kernel/events/ring_buffer.c
223
head += size;
kernel/events/ring_buffer.c
225
head -= size;
kernel/events/ring_buffer.c
226
} while (!local_try_cmpxchg(&rb->head, &offset, head));
kernel/events/ring_buffer.c
229
offset = head;
kernel/events/ring_buffer.c
230
head = (u64)(-head);
kernel/events/ring_buffer.c
238
if (unlikely(head - local_read(&rb->wakeup) > rb->watermark))
kernel/events/ring_buffer.c
423
handle->head = aux_head;
kernel/events/ring_buffer.c
498
aux_head = handle->head;
kernel/events/ring_buffer.c
561
handle->head = rb->aux_head;
kernel/events/ring_buffer.c
57
unsigned long head;
kernel/events/ring_buffer.c
80
head = local_read(&rb->head);
kernel/events/uprobes.c
1529
struct list_head *head)
kernel/events/uprobes.c
1535
INIT_LIST_HEAD(head);
kernel/events/uprobes.c
1548
list_add(&u->pending_list, head);
kernel/events/uprobes.c
1556
list_add(&u->pending_list, head);
kernel/events/uprobes.c
2039
#define for_each_ret_instance_rcu(pos, head) \
kernel/events/uprobes.c
2040
for (pos = rcu_dereference_raw(head); pos; pos = rcu_dereference_raw(pos->next))
kernel/futex/core.c
1134
struct robust_list __user * __user *head,
kernel/futex/core.c
1139
if (get_user(uentry, (unsigned long __user *)head))
kernel/futex/core.c
1156
struct robust_list_head __user *head = curr->robust_list;
kernel/futex/core.c
1167
if (fetch_robust_entry(&entry, &head->list.next, &pi))
kernel/futex/core.c
1172
if (get_user(futex_offset, &head->futex_offset))
kernel/futex/core.c
1178
if (fetch_robust_entry(&pending, &head->list_op_pending, &pip))
kernel/futex/core.c
1182
while (entry != &head->list) {
kernel/futex/core.c
1231
compat_uptr_t __user *head, unsigned int *pi)
kernel/futex/core.c
1233
if (get_user(*uentry, head))
kernel/futex/core.c
1250
struct compat_robust_list_head __user *head = curr->compat_robust_list;
kernel/futex/core.c
1262
if (compat_fetch_robust_entry(&uentry, &entry, &head->list.next, &pi))
kernel/futex/core.c
1267
if (get_user(futex_offset, &head->futex_offset))
kernel/futex/core.c
1274
&head->list_op_pending, &pip))
kernel/futex/core.c
1278
while (entry != (struct robust_list __user *) &head->list) {
kernel/futex/core.c
1326
struct list_head *next, *head = &curr->pi_state_list;
kernel/futex/core.c
1347
while (!list_empty(head)) {
kernel/futex/core.c
1348
next = head->next;
kernel/futex/core.c
1379
if (head->next != next) {
kernel/futex/core.c
1542
static void futex_ref_rcu(struct rcu_head *head);
kernel/futex/core.c
1608
static void futex_ref_rcu(struct rcu_head *head)
kernel/futex/core.c
1610
struct mm_struct *mm = container_of(head, struct mm_struct, futex_rcu);
kernel/futex/syscalls.c
102
struct robust_list_head __user *head = futex_get_robust_list_common(pid, false);
kernel/futex/syscalls.c
104
if (IS_ERR(head))
kernel/futex/syscalls.c
105
return PTR_ERR(head);
kernel/futex/syscalls.c
107
if (put_user(sizeof(*head), len_ptr))
kernel/futex/syscalls.c
109
return put_user(head, head_ptr);
kernel/futex/syscalls.c
28
SYSCALL_DEFINE2(set_robust_list, struct robust_list_head __user *, head,
kernel/futex/syscalls.c
34
if (unlikely(len != sizeof(*head)))
kernel/futex/syscalls.c
37
current->robust_list = head;
kernel/futex/syscalls.c
479
struct compat_robust_list_head __user *, head,
kernel/futex/syscalls.c
482
if (unlikely(len != sizeof(*head)))
kernel/futex/syscalls.c
485
current->compat_robust_list = head;
kernel/futex/syscalls.c
494
struct compat_robust_list_head __user *head = futex_get_robust_list_common(pid, true);
kernel/futex/syscalls.c
496
if (IS_ERR(head))
kernel/futex/syscalls.c
497
return PTR_ERR(head);
kernel/futex/syscalls.c
499
if (put_user(sizeof(*head), len_ptr))
kernel/futex/syscalls.c
501
return put_user(ptr_to_compat(head), head_ptr);
kernel/futex/syscalls.c
54
void __user *head;
kernel/futex/syscalls.c
78
head = futex_task_robust_list(p, compat);
kernel/futex/syscalls.c
83
return head;
kernel/gcov/clang.c
120
INIT_LIST_HEAD(&info->head);
kernel/gcov/clang.c
124
list_add_tail(&info->head, &current_info->functions);
kernel/gcov/clang.c
131
struct gcov_fn_info, head);
kernel/gcov/clang.c
177
struct gcov_info, head);
kernel/gcov/clang.c
178
if (list_is_last(&info->head, &clang_gcov_list))
kernel/gcov/clang.c
180
return list_next_entry(info, head);
kernel/gcov/clang.c
189
list_add_tail(&info->head, &clang_gcov_list);
kernel/gcov/clang.c
200
__list_del_entry(&info->head);
kernel/gcov/clang.c
229
list_for_each_entry(fn, &info->functions, head)
kernel/gcov/clang.c
244
&info1->functions, struct gcov_fn_info, head);
kernel/gcov/clang.c
246
&info2->functions, struct gcov_fn_info, head);
kernel/gcov/clang.c
252
while (!list_is_last(&fn_ptr1->head, &info1->functions) &&
kernel/gcov/clang.c
253
!list_is_last(&fn_ptr2->head, &info2->functions)) {
kernel/gcov/clang.c
258
fn_ptr1 = list_next_entry(fn_ptr1, head);
kernel/gcov/clang.c
259
fn_ptr2 = list_next_entry(fn_ptr2, head);
kernel/gcov/clang.c
261
return list_is_last(&fn_ptr1->head, &info1->functions) &&
kernel/gcov/clang.c
262
list_is_last(&fn_ptr2->head, &info2->functions);
kernel/gcov/clang.c
276
struct gcov_fn_info, head);
kernel/gcov/clang.c
278
list_for_each_entry(dfn_ptr, &dst->functions, head) {
kernel/gcov/clang.c
284
sfn_ptr = list_next_entry(sfn_ptr, head);
kernel/gcov/clang.c
295
INIT_LIST_HEAD(&fn_dup->head);
kernel/gcov/clang.c
323
INIT_LIST_HEAD(&dup->head);
kernel/gcov/clang.c
329
list_for_each_entry(fn, &info->functions, head) {
kernel/gcov/clang.c
334
list_add_tail(&fn_dup->head, &dup->functions);
kernel/gcov/clang.c
352
list_for_each_entry_safe(fn, tmp, &info->functions, head) {
kernel/gcov/clang.c
354
list_del(&fn->head);
kernel/gcov/clang.c
378
list_for_each_entry(fi_ptr, &info->functions, head) {
kernel/gcov/clang.c
58
struct list_head head;
kernel/gcov/clang.c
68
struct list_head head;
kernel/gcov/clang.c
89
INIT_LIST_HEAD(&info->head);
kernel/gcov/clang.c
94
list_add_tail(&info->head, &clang_gcov_list);
kernel/kexec_core.c
238
image->entry = &image->head;
kernel/kexec_core.c
239
image->last_entry = &image->head;
kernel/kexec_core.c
546
for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE); \
kernel/kexec_file.c
453
image->type, image->start, image->head, flags);
kernel/kprobes.c
1947
static int kretprobe_fini_pool(struct objpool_head *head, void *context)
kernel/kprobes.c
1953
static void free_rp_inst_rcu(struct rcu_head *head)
kernel/kprobes.c
1955
struct kretprobe_instance *ri = container_of(head, struct kretprobe_instance, rcu);
kernel/kprobes.c
2667
struct hlist_head *head;
kernel/kprobes.c
2687
head = &kprobe_table[i];
kernel/kprobes.c
2688
hlist_for_each_entry(p, head, hlist)
kernel/kprobes.c
2731
struct hlist_head *head;
kernel/kprobes.c
2739
head = &kprobe_table[i];
kernel/kprobes.c
2740
hlist_for_each_entry(p, head, hlist) {
kernel/kprobes.c
2857
struct hlist_head *head;
kernel/kprobes.c
2864
head = &kprobe_table[i];
kernel/kprobes.c
2866
hlist_for_each_entry_rcu(p, head, hlist) {
kernel/kprobes.c
2933
struct hlist_head *head;
kernel/kprobes.c
2952
head = &kprobe_table[i];
kernel/kprobes.c
2954
hlist_for_each_entry(p, head, hlist) {
kernel/kprobes.c
2977
struct hlist_head *head;
kernel/kprobes.c
2991
head = &kprobe_table[i];
kernel/kprobes.c
2993
hlist_for_each_entry(p, head, hlist) {
kernel/kprobes.c
382
struct hlist_head *head;
kernel/kprobes.c
385
head = &kprobe_table[hash_ptr(addr, KPROBE_HASH_BITS)];
kernel/kprobes.c
386
hlist_for_each_entry_rcu(p, head, hlist,
kernel/kprobes.c
952
struct hlist_head *head;
kernel/kprobes.c
964
head = &kprobe_table[i];
kernel/kprobes.c
965
hlist_for_each_entry(p, head, hlist)
kernel/kprobes.c
976
struct hlist_head *head;
kernel/kprobes.c
988
head = &kprobe_table[i];
kernel/kprobes.c
989
hlist_for_each_entry(p, head, hlist) {
kernel/locking/lockdep.c
1088
struct hlist_head *head;
kernel/locking/lockdep.c
1115
head = chainhash_table + i;
kernel/locking/lockdep.c
1116
hlist_for_each_entry_rcu(chain, head, entry) {
kernel/locking/lockdep.c
1425
struct lock_class *links_to, struct list_head *head,
kernel/locking/lockdep.c
1448
list_add_tail_rcu(&entry->entry, head);
kernel/locking/lockdep.c
1743
struct list_head *head;
kernel/locking/lockdep.c
1814
head = get_dep_list(lock, offset);
kernel/locking/lockdep.c
1815
list_for_each_entry_rcu(entry, head, entry) {
kernel/locking/lockdep.c
6230
struct hlist_head *head;
kernel/locking/lockdep.c
6234
head = chainhash_table + i;
kernel/locking/lockdep.c
6235
hlist_for_each_entry_rcu(chain, head, entry) {
kernel/locking/lockdep.c
6397
struct hlist_head *head;
kernel/locking/lockdep.c
6402
head = classhash_table + i;
kernel/locking/lockdep.c
6403
hlist_for_each_entry_rcu(class, head, hash_entry) {
kernel/locking/lockdep.c
6481
struct hlist_head *head;
kernel/locking/lockdep.c
6485
head = classhash_table + i;
kernel/locking/lockdep.c
6486
hlist_for_each_entry_rcu(class, head, hash_entry) {
kernel/locking/rtmutex.c
553
wake_q_add(&wqh->head, task);
kernel/locking/rtmutex.c
571
if (!wake_q_empty(&wqh->head))
kernel/locking/rtmutex.c
572
wake_up_q(&wqh->head);
kernel/locking/rtmutex_common.h
68
struct wake_q_head head;
kernel/locking/rtmutex_common.h
74
.head = WAKE_Q_HEAD_INITIALIZER(name.head), \
kernel/module/main.c
3607
struct hlist_head *head = idem_hash + hash;
kernel/module/main.c
3617
hlist_for_each_entry(existing, head, entry) {
kernel/module/main.c
3639
struct hlist_head *head = idem_hash + hash;
kernel/module/main.c
3644
hlist_for_each_entry_safe(pos, next, head, entry) {
kernel/notifier.c
148
ret = notifier_chain_register(&nh->head, n, false);
kernel/notifier.c
171
ret = notifier_chain_register(&nh->head, n, true);
kernel/notifier.c
193
ret = notifier_chain_unregister(&nh->head, n);
kernel/notifier.c
223
ret = notifier_call_chain(&nh->head, val, v, -1, NULL);
kernel/notifier.c
241
return !rcu_access_pointer(nh->head);
kernel/notifier.c
261
return notifier_chain_register(&nh->head, n, unique_priority);
kernel/notifier.c
264
ret = notifier_chain_register(&nh->head, n, unique_priority);
kernel/notifier.c
324
return notifier_chain_unregister(&nh->head, n);
kernel/notifier.c
327
ret = notifier_chain_unregister(&nh->head, n);
kernel/notifier.c
343
if (rcu_access_pointer(nh->head)) {
kernel/notifier.c
345
ret = notifier_call_chain_robust(&nh->head, val_up, val_down, v);
kernel/notifier.c
378
if (rcu_access_pointer(nh->head)) {
kernel/notifier.c
380
ret = notifier_call_chain(&nh->head, val, v, -1, NULL);
kernel/notifier.c
405
return notifier_chain_register(&nh->head, n, false);
kernel/notifier.c
422
return notifier_chain_unregister(&nh->head, n);
kernel/notifier.c
429
return notifier_call_chain_robust(&nh->head, val_up, val_down, v);
kernel/notifier.c
453
return notifier_call_chain(&nh->head, val, v, -1, NULL);
kernel/notifier.c
483
return notifier_chain_register(&nh->head, n, false);
kernel/notifier.c
486
ret = notifier_chain_register(&nh->head, n, false);
kernel/notifier.c
513
return notifier_chain_unregister(&nh->head, n);
kernel/notifier.c
516
ret = notifier_chain_unregister(&nh->head, n);
kernel/notifier.c
546
ret = notifier_call_chain(&nh->head, val, v, -1, NULL);
kernel/notifier.c
569
nh->head = NULL;
kernel/nstree.c
556
const struct list_head *head;
kernel/nstree.c
581
head = &to_ns_common(kls->user_ns)->ns_owner_root.ns_list_head;
kernel/nstree.c
587
first_ns = list_entry_rcu(head->next, typeof(*first_ns), ns_owner_node.ns_list_entry);
kernel/nstree.c
590
list_for_each_entry_from_rcu(ns, head, ns_owner_node.ns_list_entry) {
kernel/nstree.c
670
static inline struct ns_common *first_ns_common(const struct list_head *head,
kernel/nstree.c
674
return list_entry_rcu(head->next, struct ns_common, ns_tree_node.ns_list_entry);
kernel/nstree.c
675
return list_entry_rcu(head->next, struct ns_common, ns_unified_node.ns_list_entry);
kernel/nstree.c
687
const struct list_head *head,
kernel/nstree.c
691
return &ns->ns_tree_node.ns_list_entry == head;
kernel/nstree.c
692
return &ns->ns_unified_node.ns_list_entry == head;
kernel/nstree.c
701
const struct list_head *head;
kernel/nstree.c
725
head = &ns_tree->ns_list_head;
kernel/nstree.c
727
head = &ns_unified_root.ns_list_head;
kernel/nstree.c
732
first_ns = first_ns_common(head, ns_tree);
kernel/nstree.c
734
for (ns = first_ns; !ns_common_is_head(ns, head, ns_tree) && nr_ns_ids;
kernel/padata.c
110
struct list_head *head)
kernel/padata.c
122
list_add(&pw->pw_list, head);
kernel/pid.c
720
static int pid_table_root_permissions(struct ctl_table_header *head,
kernel/pid.c
724
container_of(head->set, struct pid_namespace, set);
kernel/pid.c
737
static void pid_table_root_set_ownership(struct ctl_table_header *head,
kernel/pid.c
741
container_of(head->set, struct pid_namespace, set);
kernel/power/console.c
124
list_for_each_entry(entry, &pm_vt_switch_list, head) {
kernel/power/console.c
24
struct list_head head;
kernel/power/console.c
53
list_for_each_entry(tmp, &pm_vt_switch_list, head) {
kernel/power/console.c
70
list_add(&entry->head, &pm_vt_switch_list);
kernel/power/console.c
88
list_for_each_entry(tmp, &pm_vt_switch_list, head) {
kernel/power/console.c
90
list_del(&tmp->head);
kernel/rcu/rcu.h
232
static inline int debug_rcu_head_queue(struct rcu_head *head)
kernel/rcu/rcu.h
236
r1 = debug_object_activate(head, &rcuhead_debug_descr);
kernel/rcu/rcu.h
237
debug_object_active_state(head, &rcuhead_debug_descr,
kernel/rcu/rcu.h
243
static inline void debug_rcu_head_unqueue(struct rcu_head *head)
kernel/rcu/rcu.h
245
debug_object_active_state(head, &rcuhead_debug_descr,
kernel/rcu/rcu.h
248
debug_object_deactivate(head, &rcuhead_debug_descr);
kernel/rcu/rcu.h
251
static inline int debug_rcu_head_queue(struct rcu_head *head)
kernel/rcu/rcu.h
256
static inline void debug_rcu_head_unqueue(struct rcu_head *head)
kernel/rcu/rcu_segcblist.c
20
rclp->head = NULL;
kernel/rcu/rcu_segcblist.c
21
rclp->tail = &rclp->head;
kernel/rcu/rcu_segcblist.c
243
rsclp->head = NULL;
kernel/rcu/rcu_segcblist.c
245
rsclp->tails[i] = &rsclp->head;
kernel/rcu/rcu_segcblist.c
270
&rsclp->head != READ_ONCE(rsclp->tails[RCU_DONE_TAIL]);
kernel/rcu/rcu_segcblist.c
290
return rsclp->head;
kernel/rcu/rcu_segcblist.c
382
*rclp->tail = rsclp->head;
kernel/rcu/rcu_segcblist.c
383
WRITE_ONCE(rsclp->head, *rsclp->tails[RCU_DONE_TAIL]);
kernel/rcu/rcu_segcblist.c
388
WRITE_ONCE(rsclp->tails[i], &rsclp->head);
kernel/rcu/rcu_segcblist.c
436
if (!rclp->head)
kernel/rcu/rcu_segcblist.c
439
*rclp->tail = rsclp->head;
kernel/rcu/rcu_segcblist.c
440
WRITE_ONCE(rsclp->head, rclp->head);
kernel/rcu/rcu_segcblist.c
442
if (&rsclp->head == rsclp->tails[i])
kernel/rcu/rcu_segcblist.c
446
rclp->head = NULL;
kernel/rcu/rcu_segcblist.c
447
rclp->tail = &rclp->head;
kernel/rcu/rcu_segcblist.c
457
if (!rclp->head)
kernel/rcu/rcu_segcblist.c
461
WRITE_ONCE(*rsclp->tails[RCU_NEXT_TAIL], rclp->head);
kernel/rcu/rcu_segcblist.c
47
drclp->head = srclp->head;
kernel/rcu/rcu_segcblist.c
48
if (drclp->head)
kernel/rcu/rcu_segcblist.c
51
drclp->tail = &drclp->head;
kernel/rcu/rcu_segcblist.c
57
srclp->head = rhp;
kernel/rcu/rcu_segcblist.c
71
rhp = rclp->head;
kernel/rcu/rcu_segcblist.c
75
rclp->head = rhp->next;
kernel/rcu/rcu_segcblist.c
76
if (!rclp->head)
kernel/rcu/rcu_segcblist.c
77
rclp->tail = &rclp->head;
kernel/rcu/rcu_segcblist.h
115
return &rsclp->head == rsclp->tails[RCU_DONE_TAIL];
kernel/rcu/rcu_segcblist.h
45
return !READ_ONCE(rsclp->head);
kernel/rcu/rcuscale.c
157
void (*async)(struct rcu_head *head, rcu_callback_t func);
kernel/rcu/rcuscale.c
229
static void srcu_call_rcu(struct rcu_head *head, rcu_callback_t func)
kernel/rcu/rcuscale.c
231
call_srcu(srcu_ctlp, head, func);
kernel/rcu/rcutorture.c
665
call_rcu_busted(struct rcu_head *head, rcu_callback_t func)
kernel/rcu/rcutorture.c
668
func(head);
kernel/rcu/rcutorture.c
868
static void srcu_torture_call(struct rcu_head *head,
kernel/rcu/rcutorture.c
871
call_srcu(srcu_ctlp, head, func);
kernel/rcu/srcutiny.c
255
init_rcu_head_on_stack(&rs.head);
kernel/rcu/srcutiny.c
257
call_srcu(ssp, &rs.head, wakeme_after_rcu);
kernel/rcu/srcutiny.c
259
destroy_rcu_head_on_stack(&rs.head);
kernel/rcu/srcutree.c
1494
init_rcu_head_on_stack(&rcu.head);
kernel/rcu/srcutree.c
1495
__call_srcu(ssp, &rcu.head, wakeme_after_rcu, do_norm);
kernel/rcu/srcutree.c
1497
destroy_rcu_head_on_stack(&rcu.head);
kernel/rcu/tiny.c
158
void call_rcu(struct rcu_head *head, rcu_callback_t func)
kernel/rcu/tiny.c
163
if (debug_rcu_head_queue(head)) {
kernel/rcu/tiny.c
165
pr_err("%s(): Double-freed CB %p->%pS()!!! ", __func__, head, head->func);
kernel/rcu/tiny.c
166
mem_dump_obj(head);
kernel/rcu/tiny.c
171
head->func = func;
kernel/rcu/tiny.c
172
head->next = NULL;
kernel/rcu/tiny.c
175
*rcu_ctrlblk.curtail = head;
kernel/rcu/tiny.c
176
rcu_ctrlblk.curtail = &head->next;
kernel/rcu/tiny.c
83
static inline bool rcu_reclaim_tiny(struct rcu_head *head)
kernel/rcu/tiny.c
89
trace_rcu_invoke_callback("", head);
kernel/rcu/tiny.c
90
f = head->func;
kernel/rcu/tiny.c
91
debug_rcu_head_callback(head);
kernel/rcu/tiny.c
92
WRITE_ONCE(head->func, (rcu_callback_t)0L);
kernel/rcu/tiny.c
93
f(head);
kernel/rcu/tree.c
1645
(struct rcu_head *) node, struct rcu_synchronize, head);
kernel/rcu/tree.c
1657
struct llist_node *done, *rcu, *next, *head;
kernel/rcu/tree.c
1675
head = done->next;
kernel/rcu/tree.c
1687
llist_for_each_safe(rcu, next, head) {
kernel/rcu/tree.c
1798
llist_add((struct llist_node *) &rs->head, &rcu_state.srs_next);
kernel/rcu/tree.c
2653
trace_rcu_batch_end(rcu_state.name, count, !!rcl.head, need_resched(),
kernel/rcu/tree.c
2998
static void rcutree_enqueue(struct rcu_data *rdp, struct rcu_head *head, rcu_callback_t func)
kernel/rcu/tree.c
3000
rcu_segcblist_enqueue(&rdp->cblist, head);
kernel/rcu/tree.c
3001
trace_rcu_callback(rcu_state.name, head,
kernel/rcu/tree.c
3009
static void call_rcu_core(struct rcu_data *rdp, struct rcu_head *head,
kernel/rcu/tree.c
3012
rcutree_enqueue(rdp, head, func);
kernel/rcu/tree.c
3044
rcu_segcblist_first_pend_cb(&rdp->cblist) != head)
kernel/rcu/tree.c
3102
__call_rcu_common(struct rcu_head *head, rcu_callback_t func, bool lazy_in)
kernel/rcu/tree.c
3110
WARN_ON_ONCE((unsigned long)head & (sizeof(void *) - 1));
kernel/rcu/tree.c
3116
if (debug_rcu_head_queue(head)) {
kernel/rcu/tree.c
3123
pr_err("%s(): Double-freed CB %p->%pS()!!! ", __func__, head, head->func);
kernel/rcu/tree.c
3124
mem_dump_obj(head);
kernel/rcu/tree.c
3126
WRITE_ONCE(head->func, rcu_leak_callback);
kernel/rcu/tree.c
3129
head->func = func;
kernel/rcu/tree.c
3130
head->next = NULL;
kernel/rcu/tree.c
3131
kasan_record_aux_stack(head);
kernel/rcu/tree.c
3153
call_rcu_nocb(rdp, head, func, flags, lazy);
kernel/rcu/tree.c
3155
call_rcu_core(rdp, head, func, flags);
kernel/rcu/tree.c
3183
void call_rcu_hurry(struct rcu_head *head, rcu_callback_t func)
kernel/rcu/tree.c
3185
__call_rcu_common(head, func, false);
kernel/rcu/tree.c
3249
void call_rcu(struct rcu_head *head, rcu_callback_t func)
kernel/rcu/tree.c
3251
__call_rcu_common(head, func, enable_rcu_lazy);
kernel/rcu/tree.c
3281
trace_rcu_sr_normal(rcu_state.name, &rs.head, TPS("request"));
kernel/rcu/tree.c
3288
init_rcu_head_on_stack(&rs.head);
kernel/rcu/tree.c
3305
destroy_rcu_head_on_stack(&rs.head);
kernel/rcu/tree.c
3308
trace_rcu_sr_normal(rcu_state.name, &rs.head, TPS("complete"));
kernel/rcu/tree.h
505
static void call_rcu_nocb(struct rcu_data *rdp, struct rcu_head *head,
kernel/rcu/tree_nocb.h
1645
static void call_rcu_nocb(struct rcu_data *rdp, struct rcu_head *head,
kernel/rcu/tree_nocb.h
575
static void call_rcu_nocb(struct rcu_data *rdp, struct rcu_head *head,
kernel/rcu/tree_nocb.h
580
if (!rcu_nocb_try_bypass(rdp, head, &was_alldone, flags, lazy)) {
kernel/rcu/tree_nocb.h
582
rcutree_enqueue(rdp, head, func);
kernel/rcu/update.c
402
void wakeme_after_rcu(struct rcu_head *head)
kernel/rcu/update.c
406
rcu = container_of(head, struct rcu_synchronize, head);
kernel/rcu/update.c
428
init_rcu_head_on_stack(&rs_array[i].head);
kernel/rcu/update.c
430
(crcu_array[i])(&rs_array[i].head, wakeme_after_rcu);
kernel/rcu/update.c
444
destroy_rcu_head_on_stack(&rs_array[i].head);
kernel/rcu/update.c
458
void init_rcu_head(struct rcu_head *head)
kernel/rcu/update.c
460
debug_object_init(head, &rcuhead_debug_descr);
kernel/rcu/update.c
464
void destroy_rcu_head(struct rcu_head *head)
kernel/rcu/update.c
466
debug_object_free(head, &rcuhead_debug_descr);
kernel/rcu/update.c
485
void init_rcu_head_on_stack(struct rcu_head *head)
kernel/rcu/update.c
487
debug_object_init_on_stack(head, &rcuhead_debug_descr);
kernel/rcu/update.c
502
void destroy_rcu_head_on_stack(struct rcu_head *head)
kernel/rcu/update.c
504
debug_object_free(head, &rcuhead_debug_descr);
kernel/rcu/update.c
607
static struct rcu_head head;
kernel/rcu/update.c
614
call_rcu(&head, test_callback);
kernel/resource.c
1951
void resource_list_free(struct list_head *head)
kernel/resource.c
1955
list_for_each_entry_safe(entry, tmp, head, node)
kernel/sched/core.c
1018
static bool __wake_q_add(struct wake_q_head *head, struct task_struct *task)
kernel/sched/core.c
1037
*head->lastp = node;
kernel/sched/core.c
1038
head->lastp = &node->next;
kernel/sched/core.c
1054
void wake_q_add(struct wake_q_head *head, struct task_struct *task)
kernel/sched/core.c
1056
if (__wake_q_add(head, task))
kernel/sched/core.c
1077
void wake_q_add_safe(struct wake_q_head *head, struct task_struct *task)
kernel/sched/core.c
1079
if (!__wake_q_add(head, task))
kernel/sched/core.c
1083
void wake_up_q(struct wake_q_head *head)
kernel/sched/core.c
1085
struct wake_q_node *node = head->first;
kernel/sched/core.c
4916
static void do_balance_callbacks(struct rq *rq, struct balance_callback *head)
kernel/sched/core.c
4923
while (head) {
kernel/sched/core.c
4924
func = (void (*)(struct rq *))head->func;
kernel/sched/core.c
4925
next = head->next;
kernel/sched/core.c
4926
head->next = NULL;
kernel/sched/core.c
4927
head = next;
kernel/sched/core.c
4954
struct balance_callback *head = rq->balance_callback;
kernel/sched/core.c
4956
if (likely(!head))
kernel/sched/core.c
4968
if (split && head == &balance_push_callback)
kernel/sched/core.c
4969
head = NULL;
kernel/sched/core.c
4973
return head;
kernel/sched/core.c
4990
void balance_callbacks(struct rq *rq, struct balance_callback *head)
kernel/sched/core.c
4994
if (unlikely(head)) {
kernel/sched/core.c
4996
do_balance_callbacks(rq, head);
kernel/sched/rt.c
1467
requeue_rt_entity(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int head)
kernel/sched/rt.c
1473
if (head)
kernel/sched/rt.c
1480
static void requeue_task_rt(struct rq *rq, struct task_struct *p, int head)
kernel/sched/rt.c
1487
requeue_rt_entity(rt_rq, rt_se, head);
kernel/sched/rt.c
1747
struct plist_head *head = &rq->rt.pushable_tasks;
kernel/sched/rt.c
1753
plist_for_each_entry(p, head, pushable_tasks) {
kernel/sched/sched.h
2068
struct balance_callback *head,
kernel/sched/sched.h
2078
if (unlikely(head->next || rq->balance_callback == &balance_push_callback))
kernel/sched/sched.h
2081
head->func = func;
kernel/sched/sched.h
2082
head->next = rq->balance_callback;
kernel/sched/sched.h
2083
rq->balance_callback = head;
kernel/sched/sched.h
4073
extern void balance_callbacks(struct rq *rq, struct balance_callback *head);
kernel/sched/syscalls.c
500
struct balance_callback *head;
kernel/sched/syscalls.c
703
head = splice_balance_callbacks(rq);
kernel/sched/syscalls.c
713
balance_callbacks(rq, head);
kernel/sched/topology.c
1428
static void free_asym_cap_entry(struct rcu_head *head)
kernel/sched/topology.c
1430
struct asym_cap_data *entry = container_of(head, struct asym_cap_data, rcu);
kernel/sched/wait.c
101
if (&curr->entry == &wq_head->head)
kernel/sched/wait.c
104
list_for_each_entry_safe_from(curr, next, &wq_head->head, entry) {
kernel/sched/wait.c
13
INIT_LIST_HEAD(&wq_head->head);
kernel/sched/wait.c
271
was_empty = list_empty(&wq_head->head);
kernel/sched/wait.c
54
struct list_head *head = &wq_head->head;
kernel/sched/wait.c
60
if (!list_empty(head) &&
kernel/sched/wait.c
61
(list_first_entry(head, typeof(*wq_entry), entry)->flags & WQ_FLAG_PRIORITY))
kernel/sched/wait.c
64
list_add(&wq_entry->entry, head);
kernel/sched/wait.c
99
curr = list_first_entry(&wq_head->head, wait_queue_entry_t, entry);
kernel/signal.c
2104
struct hlist_head *head = &tsk->signal->ignored_posix_timers;
kernel/signal.c
2108
if (likely(hlist_empty(head)))
kernel/signal.c
2118
hlist_for_each_entry_safe(tmr, tmp , head, ignored_list) {
kernel/smp.c
484
struct llist_head *head;
kernel/smp.c
494
head = this_cpu_ptr(&call_single_queue);
kernel/smp.c
495
entry = llist_del_all(head);
kernel/softirq.c
1041
&per_cpu(tasklet_vec, cpu).head;
kernel/softirq.c
1043
&per_cpu(tasklet_hi_vec, cpu).head;
kernel/softirq.c
1080
if (&per_cpu(tasklet_vec, cpu).head != per_cpu(tasklet_vec, cpu).tail) {
kernel/softirq.c
1081
*__this_cpu_read(tasklet_vec.tail) = per_cpu(tasklet_vec, cpu).head;
kernel/softirq.c
1083
per_cpu(tasklet_vec, cpu).head = NULL;
kernel/softirq.c
1084
per_cpu(tasklet_vec, cpu).tail = &per_cpu(tasklet_vec, cpu).head;
kernel/softirq.c
1088
if (&per_cpu(tasklet_hi_vec, cpu).head != per_cpu(tasklet_hi_vec, cpu).tail) {
kernel/softirq.c
1089
*__this_cpu_read(tasklet_hi_vec.tail) = per_cpu(tasklet_hi_vec, cpu).head;
kernel/softirq.c
1091
per_cpu(tasklet_hi_vec, cpu).head = NULL;
kernel/softirq.c
1092
per_cpu(tasklet_hi_vec, cpu).tail = &per_cpu(tasklet_hi_vec, cpu).head;
kernel/softirq.c
802
struct tasklet_struct *head;
kernel/softirq.c
813
struct tasklet_head *head;
kernel/softirq.c
817
head = this_cpu_ptr(headp);
kernel/softirq.c
819
*head->tail = t;
kernel/softirq.c
820
head->tail = &(t->next);
kernel/softirq.c
909
list = tl_head->head;
kernel/softirq.c
910
tl_head->head = NULL;
kernel/softirq.c
911
tl_head->tail = &tl_head->head;
kernel/task_work.c
203
struct callback_head *work, *head, *next;
kernel/task_work.c
212
head = NULL;
kernel/task_work.c
215
head = &work_exited;
kernel/task_work.c
219
} while (!try_cmpxchg(&task->task_works, &work, head));
kernel/task_work.c
62
struct callback_head *head;
kernel/task_work.c
73
head = READ_ONCE(task->task_works);
kernel/task_work.c
75
if (unlikely(head == &work_exited))
kernel/task_work.c
77
work->next = head;
kernel/task_work.c
78
} while (!try_cmpxchg(&task->task_works, &head, work));
kernel/time/posix-cpu-timers.c
494
WARN_ON_ONCE(ctmr->head || timerqueue_node_queued(&ctmr->node));
kernel/time/posix-cpu-timers.c
521
static void cleanup_timerqueue(struct timerqueue_head *head)
kernel/time/posix-cpu-timers.c
526
while ((node = timerqueue_getnext(head))) {
kernel/time/posix-cpu-timers.c
527
timerqueue_del(head, node);
kernel/time/posix-cpu-timers.c
529
ctmr->head = NULL;
kernel/time/posix-cpu-timers.c
800
static u64 collect_timerqueue(struct timerqueue_head *head,
kernel/time/posix-cpu-timers.c
806
while ((next = timerqueue_getnext(head))) {
kernel/time/posix-timers.c
117
struct hlist_head *head = &bucket->head;
kernel/time/posix-timers.c
120
hlist_for_each_entry_rcu(timer, head, t_hash, lockdep_is_held(&bucket->lock)) {
kernel/time/posix-timers.c
150
hlist_add_head_rcu(&timer->t_hash, &bucket->head);
kernel/time/posix-timers.c
1563
INIT_HLIST_HEAD(&timer_buckets[i].head);
kernel/time/posix-timers.c
44
struct hlist_head head;
kernel/time/posix-timers.c
95
hlist_for_each_entry_rcu(timer, &bucket->head, t_hash) {
kernel/time/timer.c
1766
static void expire_timers(struct timer_base *base, struct hlist_head *head)
kernel/time/timer.c
1775
while (!hlist_empty(head)) {
kernel/time/timer.c
1779
timer = hlist_entry(head->first, struct timer_list, entry);
kernel/time/timer.c
2485
static void migrate_timer_list(struct timer_base *new_base, struct hlist_head *head)
kernel/time/timer.c
2490
while (!hlist_empty(head)) {
kernel/time/timer.c
2491
timer = hlist_entry(head->first, struct timer_list, entry);
kernel/trace/fprobe.c
112
struct hlist_head *head;
kernel/trace/fprobe.c
115
head = &fprobe_table[hash_ptr(fp, FPROBE_HASH_BITS)];
kernel/trace/fprobe.c
116
hlist_for_each_entry_rcu(fph, head, hlist,
kernel/trace/fprobe.c
128
struct hlist_head *head;
kernel/trace/fprobe.c
138
head = &fprobe_table[hash_ptr(fp, FPROBE_HASH_BITS)];
kernel/trace/fprobe.c
139
hlist_add_head_rcu(&fp->hlist_array->hlist, head);
kernel/trace/fprobe.c
264
struct rhlist_head *head, *pos;
kernel/trace/fprobe.c
278
head = rhltable_lookup(&fprobe_ip_table, &ip, fprobe_rht_params);
kernel/trace/fprobe.c
280
rhl_for_each_entry_rcu(node, pos, head, hlist) {
kernel/trace/fprobe.c
379
struct rhlist_head *head, *pos;
kernel/trace/fprobe.c
389
head = rhltable_lookup(&fprobe_ip_table, &func, fprobe_rht_params);
kernel/trace/fprobe.c
391
rhl_for_each_entry_rcu(node, pos, head, hlist) {
kernel/trace/fprobe.c
407
rhl_for_each_entry_rcu(node, pos, head, hlist) {
kernel/trace/fprobe.c
424
rhl_for_each_entry_rcu(node, pos, head, hlist) {
kernel/trace/ftrace.c
1279
static void clear_ftrace_mod_list(struct list_head *head)
kernel/trace/ftrace.c
1284
if (!head)
kernel/trace/ftrace.c
1288
list_for_each_entry_safe(p, n, head, list)
kernel/trace/ftrace.c
5070
struct list_head *head = enable ? &tr->mod_trace : &tr->mod_notrace;
kernel/trace/ftrace.c
5081
list_for_each_entry_safe(ftrace_mod, n, head, list) {
kernel/trace/ftrace.c
5106
static void process_mod_list(struct list_head *head, struct ftrace_ops *ops,
kernel/trace/ftrace.c
5128
list_for_each_entry_safe(ftrace_mod, n, head, list) {
kernel/trace/ftrace.c
5159
if (enable && list_empty(head))
kernel/trace/ftrace.c
7073
struct hlist_head *head;
kernel/trace/ftrace.c
7089
head = &fgd->hash->buckets[i];
kernel/trace/ftrace.c
7090
hlist_for_each_entry(entry, head, hlist) {
kernel/trace/rethook.c
127
static void free_rethook_node_rcu(struct rcu_head *head)
kernel/trace/rethook.c
129
struct rethook_node *node = container_of(head, struct rethook_node, rcu);
kernel/trace/rethook.c
35
static void rethook_free_rcu(struct rcu_head *head)
kernel/trace/rethook.c
37
struct rethook *rh = container_of(head, struct rethook, rcu);
kernel/trace/rethook.c
79
static int rethook_fini_pool(struct objpool_head *head, void *context)
kernel/trace/ring_buffer.c
1301
struct buffer_page *head;
kernel/trace/ring_buffer.c
1303
head = cpu_buffer->head_page;
kernel/trace/ring_buffer.c
1304
if (!head)
kernel/trace/ring_buffer.c
1310
rb_set_list_to_head(head->list.prev);
kernel/trace/ring_buffer.c
1314
meta->head_buffer = (unsigned long)head->page;
kernel/trace/ring_buffer.c
1341
struct buffer_page *head,
kernel/trace/ring_buffer.c
1346
unsigned long val = (unsigned long)&head->list;
kernel/trace/ring_buffer.c
1364
struct buffer_page *head,
kernel/trace/ring_buffer.c
1368
return rb_head_page_set(cpu_buffer, head, prev,
kernel/trace/ring_buffer.c
1373
struct buffer_page *head,
kernel/trace/ring_buffer.c
1377
return rb_head_page_set(cpu_buffer, head, prev,
kernel/trace/ring_buffer.c
1382
struct buffer_page *head,
kernel/trace/ring_buffer.c
1386
return rb_head_page_set(cpu_buffer, head, prev,
kernel/trace/ring_buffer.c
1407
struct buffer_page *head;
kernel/trace/ring_buffer.c
1420
page = head = cpu_buffer->head_page;
kernel/trace/ring_buffer.c
1434
} while (page != head);
kernel/trace/ring_buffer.c
1549
struct list_head *head, *tmp;
kernel/trace/ring_buffer.c
1578
head = rb_list_head(cpu_buffer->pages);
kernel/trace/ring_buffer.c
1579
if (!rb_check_links(cpu_buffer, head))
kernel/trace/ring_buffer.c
1582
tmp = head;
kernel/trace/ring_buffer.c
1595
if (tmp == head)
kernel/trace/ring_buffer.c
2451
struct list_head *head = cpu_buffer->pages;
kernel/trace/ring_buffer.c
2458
if (head) {
kernel/trace/ring_buffer.c
2461
list_for_each_entry_safe(bpage, tmp, head, list) {
kernel/trace/ring_buffer.c
2465
bpage = list_entry(head, struct buffer_page, list);
kernel/trace/ring_buffer.c
3188
if (iter->head != iter->next_event)
kernel/trace/ring_buffer.c
3200
if (iter->head > commit - 8)
kernel/trace/ring_buffer.c
3203
event = __rb_page_index(iter_head_page, iter->head);
kernel/trace/ring_buffer.c
3212
if ((iter->head + length) > commit || length > iter->event_size)
kernel/trace/ring_buffer.c
3228
iter->next_event = iter->head + length;
kernel/trace/ring_buffer.c
3233
iter->head = 0;
kernel/trace/ring_buffer.c
3277
iter->head = 0;
kernel/trace/ring_buffer.c
5283
iter->head = cpu_buffer->reader_page->read;
kernel/trace/ring_buffer.c
5284
iter->next_event = iter->head;
kernel/trace/ring_buffer.c
5290
if (iter->head) {
kernel/trace/ring_buffer.c
5363
return ((iter->head_page == commit_page && iter->head >= commit) ||
kernel/trace/ring_buffer.c
5366
iter->head == rb_page_size(cpu_buffer->reader_page)));
kernel/trace/ring_buffer.c
5632
if (iter->head == iter->next_event) {
kernel/trace/ring_buffer.c
5638
iter->head = iter->next_event;
kernel/trace/ring_buffer.c
5774
if (iter->head >= rb_page_size(iter->head_page)) {
kernel/trace/ring_buffer.c
602
unsigned long head;
kernel/trace/rv/rv.c
502
struct list_head *head;
kernel/trace/rv/rv.c
510
head = &rv_monitors_list;
kernel/trace/rv/rv.c
513
head = enabled_monitors_next(m, head, &l);
kernel/trace/rv/rv.c
514
if (!head)
kernel/trace/rv/rv.c
518
return head;
kernel/trace/trace.c
10114
if (map->head.mod == mod)
kernel/trace/trace.c
207
struct trace_eval_map_head head;
kernel/trace/trace.c
2999
struct list_head *head;
kernel/trace/trace.c
3016
head = trace_get_fields(event);
kernel/trace/trace.c
3017
if (!head) {
kernel/trace/trace.c
3026
list_for_each_entry(field, head, link) {
kernel/trace/trace.c
5115
return ptr + ptr->head.length + 1;
kernel/trace/trace.c
5155
map_array->head.mod = mod;
kernel/trace/trace.c
5156
map_array->head.length = len;
kernel/trace/trace_event_perf.c
440
struct hlist_head head;
kernel/trace/trace_event_perf.c
463
head.first = &event->hlist_entry;
kernel/trace/trace_event_perf.c
480
1, &regs, &head, NULL);
kernel/trace/trace_events.c
101
struct list_head *head;
kernel/trace/trace_events.c
103
head = trace_get_fields(call);
kernel/trace/trace_events.c
104
field = __find_event_field(head, name);
kernel/trace/trace_events.c
115
static int __trace_define_field(struct list_head *head, const char *type,
kernel/trace/trace_events.c
140
list_add(&field->link, head);
kernel/trace/trace_events.c
149
struct list_head *head;
kernel/trace/trace_events.c
154
head = trace_get_fields(call);
kernel/trace/trace_events.c
155
return __trace_define_field(head, type, name, offset, size,
kernel/trace/trace_events.c
164
struct list_head *head;
kernel/trace/trace_events.c
169
head = trace_get_fields(call);
kernel/trace/trace_events.c
170
return __trace_define_field(head, type, name, offset, size,
kernel/trace/trace_events.c
2052
struct list_head *head = trace_get_fields(call);
kernel/trace/trace_events.c
2063
node = head;
kernel/trace/trace_events.c
2074
else if (node == head)
kernel/trace/trace_events.c
223
struct list_head *head;
kernel/trace/trace_events.c
225
head = trace_get_fields(call);
kernel/trace/trace_events.c
226
list_for_each_entry_safe(field, next, head, link) {
kernel/trace/trace_events.c
239
struct list_head *head;
kernel/trace/trace_events.c
241
head = trace_get_fields(call);
kernel/trace/trace_events.c
246
tail = list_first_entry(head, struct ftrace_event_field, link);
kernel/trace/trace_events.c
3012
struct list_head *head;
kernel/trace/trace_events.c
3019
head = trace_get_fields(call);
kernel/trace/trace_events.c
3020
if (list_empty(head)) {
kernel/trace/trace_events.c
3500
struct list_head *head;
kernel/trace/trace_events.c
3514
head = trace_get_fields(call);
kernel/trace/trace_events.c
3515
list_for_each_entry(field, head, link) {
kernel/trace/trace_events.c
85
__find_event_field(struct list_head *head, const char *name)
kernel/trace/trace_events.c
89
list_for_each_entry(field, head, link) {
kernel/trace/trace_events_filter.c
1387
static void delay_free_filter(struct filter_head *head)
kernel/trace/trace_events_filter.c
1389
call_rcu_tasks_trace(&head->rcu, free_filter_list_tasks);
kernel/trace/trace_events_filter.c
1394
struct filter_head *head;
kernel/trace/trace_events_filter.c
1397
head = kmalloc_obj(*head);
kernel/trace/trace_events_filter.c
1398
if (!head)
kernel/trace/trace_events_filter.c
1401
INIT_LIST_HEAD(&head->list);
kernel/trace/trace_events_filter.c
1405
kfree(head);
kernel/trace/trace_events_filter.c
1410
list_add_tail(&item->list, &head->list);
kernel/trace/trace_events_filter.c
1411
delay_free_filter(head);
kernel/trace/trace_events_filter.c
1442
struct filter_head *head;
kernel/trace/trace_events_filter.c
1445
head = kmalloc_obj(*head);
kernel/trace/trace_events_filter.c
1446
if (!head)
kernel/trace/trace_events_filter.c
1449
INIT_LIST_HEAD(&head->list);
kernel/trace/trace_events_filter.c
1458
list_add_tail(&item->list, &head->list);
kernel/trace/trace_events_filter.c
1467
list_add_tail(&item->list, &head->list);
kernel/trace/trace_events_filter.c
1469
delay_free_filter(head);
kernel/trace/trace_events_filter.c
1474
if (head)
kernel/trace/trace_events_filter.c
1475
free_filter_list(head);
kernel/trace/trace_events_inject.c
141
struct list_head *head;
kernel/trace/trace_events_inject.c
144
head = trace_get_fields(call);
kernel/trace/trace_events_inject.c
145
list_for_each_entry(field, head, link) {
kernel/trace/trace_events_inject.c
157
struct list_head *head;
kernel/trace/trace_events_inject.c
165
head = trace_get_fields(call);
kernel/trace/trace_events_inject.c
166
list_for_each_entry(field, head, link) {
kernel/trace/trace_events_user.c
1089
struct list_head *head = &user->validators;
kernel/trace/trace_events_user.c
1091
list_for_each_entry_safe(validator, next, head, user_event_link) {
kernel/trace/trace_events_user.c
1100
struct list_head *head = &user->fields;
kernel/trace/trace_events_user.c
1102
list_for_each_entry_safe(field, next, head, link) {
kernel/trace/trace_events_user.c
1389
struct list_head *head = &user->fields;
kernel/trace/trace_events_user.c
1395
list_for_each_entry_reverse(field, head, link) {
kernel/trace/trace_events_user.c
1407
list_for_each_entry_reverse(field, head, link) {
kernel/trace/trace_events_user.c
1546
struct list_head *head = &user->validators;
kernel/trace/trace_events_user.c
1551
list_for_each_entry(validator, head, user_event_link) {
kernel/trace/trace_events_user.c
1807
struct list_head *head;
kernel/trace/trace_events_user.c
1812
head = trace_get_fields(&user->call);
kernel/trace/trace_events_user.c
1814
list_for_each_entry_reverse(field, head, link) {
kernel/trace/trace_events_user.c
1894
struct list_head *head = &user->fields;
kernel/trace/trace_events_user.c
1898
return list_empty(head);
kernel/trace/trace_events_user.c
1900
list_for_each_entry_reverse(field, head, link) {
kernel/trace/trace_fprobe.c
456
struct hlist_head *head;
kernel/trace/trace_fprobe.c
461
head = this_cpu_ptr(call->perf_events);
kernel/trace/trace_fprobe.c
462
if (hlist_empty(head))
kernel/trace/trace_fprobe.c
480
head, NULL);
kernel/trace/trace_fprobe.c
492
struct hlist_head *head;
kernel/trace/trace_fprobe.c
497
head = this_cpu_ptr(call->perf_events);
kernel/trace/trace_fprobe.c
498
if (hlist_empty(head))
kernel/trace/trace_fprobe.c
516
head, NULL);
kernel/trace/trace_kprobe.c
1678
struct hlist_head *head;
kernel/trace/trace_kprobe.c
1699
head = this_cpu_ptr(call->perf_events);
kernel/trace/trace_kprobe.c
1700
if (hlist_empty(head))
kernel/trace/trace_kprobe.c
1716
head, NULL);
kernel/trace/trace_kprobe.c
1728
struct hlist_head *head;
kernel/trace/trace_kprobe.c
1735
head = this_cpu_ptr(call->perf_events);
kernel/trace/trace_kprobe.c
1736
if (hlist_empty(head))
kernel/trace/trace_kprobe.c
1752
head, NULL);
kernel/trace/trace_output.c
1121
struct list_head *head;
kernel/trace/trace_output.c
1145
head = trace_get_fields(call);
kernel/trace/trace_output.c
1149
if (head && !list_empty(head))
kernel/trace/trace_output.c
1150
print_fields(iter, call, head);
kernel/trace/trace_output.c
968
struct list_head *head)
kernel/trace/trace_output.c
981
list_for_each_entry_reverse(field, head, link) {
kernel/trace/trace_probe.c
322
struct list_head *head;
kernel/trace/trace_probe.c
324
head = trace_get_fields(ctx->event);
kernel/trace/trace_probe.c
325
list_for_each_entry(field, head, link) {
kernel/trace/trace_syscalls.c
1402
struct hlist_head *head;
kernel/trace/trace_syscalls.c
1442
head = this_cpu_ptr(sys_data->enter_event->perf_events);
kernel/trace/trace_syscalls.c
1444
if (!valid_prog_array && hlist_empty(head))
kernel/trace/trace_syscalls.c
1464
hlist_empty(head)) {
kernel/trace/trace_syscalls.c
1471
head, NULL);
kernel/trace/trace_syscalls.c
1540
struct hlist_head *head;
kernel/trace/trace_syscalls.c
1563
head = this_cpu_ptr(sys_data->exit_event->perf_events);
kernel/trace/trace_syscalls.c
1565
if (!valid_prog_array && hlist_empty(head))
kernel/trace/trace_syscalls.c
1581
hlist_empty(head)) {
kernel/trace/trace_syscalls.c
1587
1, regs, head, NULL);
kernel/trace/trace_uprobe.c
1393
struct hlist_head *head;
kernel/trace/trace_uprobe.c
1421
head = this_cpu_ptr(call->perf_events);
kernel/trace/trace_uprobe.c
1422
if (hlist_empty(head))
kernel/trace/trace_uprobe.c
1444
head, NULL);
kernel/tracepoint.c
110
static void rcu_free_old_probes(struct rcu_head *head)
kernel/tracepoint.c
112
kfree(container_of(head, struct tp_probes, rcu));
kernel/ucount.c
42
static int set_permissions(struct ctl_table_header *head,
kernel/ucount.c
46
container_of(head->set, struct user_namespace, set);
kernel/unwind/deferred.c
187
static void unwind_deferred_task_work(struct callback_head *head)
kernel/user-return-notifier.c
39
struct hlist_head *head;
kernel/user-return-notifier.c
41
head = &get_cpu_var(return_notifier_list);
kernel/user-return-notifier.c
42
hlist_for_each_entry_safe(urn, tmp2, head, link)
kernel/watch_queue.c
104
unsigned int head, tail, note, offset, len;
kernel/watch_queue.c
109
head = pipe->head;
kernel/watch_queue.c
111
if (pipe_full(head, tail, pipe->ring_size))
kernel/watch_queue.c
126
buf = pipe_buf(pipe, head);
kernel/watch_queue.c
133
smp_store_release(&pipe->head, head + 1); /* vs pipe_read() */
kernel/watch_queue.c
149
buf = pipe_buf(pipe, head - 1);
kernel/workqueue.c
1146
static void move_linked_works(struct work_struct *work, struct list_head *head,
kernel/workqueue.c
1156
list_move_tail(&work->entry, head);
kernel/workqueue.c
2194
struct list_head *head, unsigned int extra_flags)
kernel/workqueue.c
2203
list_add_tail(&work->entry, head);
kernel/workqueue.c
3851
struct list_head *head;
kernel/workqueue.c
3879
head = worker->scheduled.next;
kernel/workqueue.c
3884
head = target->entry.next;
kernel/workqueue.c
3894
insert_work(pwq, &barr->work, head, work_flags);
lib/assoc_array.c
1298
static void assoc_array_rcu_cleanup(struct rcu_head *head)
lib/assoc_array.c
1301
container_of(head, struct assoc_array_edit, rcu);
lib/bootconfig.c
454
static struct xbc_node * __init __xbc_add_sibling(char *data, uint32_t flag, bool head)
lib/bootconfig.c
466
if (!last_parent->child || head) {
lib/btree.c
175
static inline void __btree_init(struct btree_head *head)
lib/btree.c
177
head->node = NULL;
lib/btree.c
178
head->height = 0;
lib/btree.c
181
void btree_init_mempool(struct btree_head *head, mempool_t *mempool)
lib/btree.c
183
__btree_init(head);
lib/btree.c
184
head->mempool = mempool;
lib/btree.c
188
int btree_init(struct btree_head *head)
lib/btree.c
190
__btree_init(head);
lib/btree.c
191
head->mempool = mempool_create(0, btree_alloc, btree_free, NULL);
lib/btree.c
192
if (!head->mempool)
lib/btree.c
198
void btree_destroy(struct btree_head *head)
lib/btree.c
200
mempool_free(head->node, head->mempool);
lib/btree.c
201
mempool_destroy(head->mempool);
lib/btree.c
202
head->mempool = NULL;
lib/btree.c
206
void *btree_last(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
209
int height = head->height;
lib/btree.c
210
unsigned long *node = head->node;
lib/btree.c
240
static void *btree_lookup_node(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
243
int i, height = head->height;
lib/btree.c
244
unsigned long *node = head->node;
lib/btree.c
262
void *btree_lookup(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
268
node = btree_lookup_node(head, geo, key);
lib/btree.c
279
int btree_update(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
285
node = btree_lookup_node(head, geo, key);
lib/btree.c
306
void *btree_get_prev(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
316
if (head->height == 0)
lib/btree.c
322
node = head->node;
lib/btree.c
323
for (height = head->height ; height > 1; height--) {
lib/btree.c
383
static unsigned long *find_level(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
386
unsigned long *node = head->node;
lib/btree.c
389
for (height = head->height; height > level; height--) {
lib/btree.c
408
static int btree_grow(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
414
node = btree_node_alloc(head, gfp);
lib/btree.c
417
if (head->node) {
lib/btree.c
418
fill = getfill(geo, head->node, 0);
lib/btree.c
419
setkey(geo, node, 0, bkey(geo, head->node, fill - 1));
lib/btree.c
420
setval(geo, node, 0, head->node);
lib/btree.c
422
head->node = node;
lib/btree.c
423
head->height++;
lib/btree.c
427
static void btree_shrink(struct btree_head *head, struct btree_geo *geo)
lib/btree.c
432
if (head->height <= 1)
lib/btree.c
435
node = head->node;
lib/btree.c
438
head->node = bval(geo, node, 0);
lib/btree.c
439
head->height--;
lib/btree.c
440
mempool_free(node, head->mempool);
lib/btree.c
443
static int btree_insert_level(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
451
if (head->height < level) {
lib/btree.c
452
err = btree_grow(head, geo, gfp);
lib/btree.c
458
node = find_level(head, geo, key, level);
lib/btree.c
468
new = btree_node_alloc(head, gfp);
lib/btree.c
471
err = btree_insert_level(head, geo,
lib/btree.c
475
mempool_free(new, head->mempool);
lib/btree.c
505
int btree_insert(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
509
return btree_insert_level(head, geo, key, val, 1, gfp);
lib/btree.c
513
static void *btree_remove_level(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
515
static void merge(struct btree_head *head, struct btree_geo *geo, int level,
lib/btree.c
531
btree_remove_level(head, geo, bkey(geo, parent, lpos), level + 1);
lib/btree.c
532
mempool_free(right, head->mempool);
lib/btree.c
535
static void rebalance(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
546
btree_remove_level(head, geo, key, level + 1);
lib/btree.c
547
mempool_free(child, head->mempool);
lib/btree.c
551
parent = find_level(head, geo, key, level + 1);
lib/btree.c
559
merge(head, geo, level,
lib/btree.c
570
merge(head, geo, level,
lib/btree.c
586
static void *btree_remove_level(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
593
if (level > head->height) {
lib/btree.c
595
head->height = 0;
lib/btree.c
596
head->node = NULL;
lib/btree.c
600
node = find_level(head, geo, key, level);
lib/btree.c
615
if (level < head->height)
lib/btree.c
616
rebalance(head, geo, key, level, node, fill - 1);
lib/btree.c
618
btree_shrink(head, geo);
lib/btree.c
624
void *btree_remove(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
627
if (head->height == 0)
lib/btree.c
630
return btree_remove_level(head, geo, key, 1);
lib/btree.c
671
static size_t __btree_for_each(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
686
count = __btree_for_each(head, geo, child, opaque,
lib/btree.c
693
mempool_free(node, head->mempool);
lib/btree.c
741
size_t btree_visitor(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
752
if (head->node)
lib/btree.c
753
count = __btree_for_each(head, geo, head->node, opaque, func,
lib/btree.c
754
func2, 0, head->height, 0);
lib/btree.c
759
size_t btree_grim_visitor(struct btree_head *head, struct btree_geo *geo,
lib/btree.c
770
if (head->node)
lib/btree.c
771
count = __btree_for_each(head, geo, head->node, opaque, func,
lib/btree.c
772
func2, 1, head->height, 0);
lib/btree.c
773
__btree_init(head);
lib/btree.c
92
static unsigned long *btree_node_alloc(struct btree_head *head, gfp_t gfp)
lib/btree.c
96
node = mempool_alloc(head->mempool, gfp);
lib/debugobjects.c
182
static bool pool_push_batch(struct obj_pool *dst, struct hlist_head *head)
lib/debugobjects.c
190
obj = hlist_entry(head->first, typeof(*obj), node);
lib/debugobjects.c
193
hlist_splice_init(head, last, &dst->objects);
lib/debugobjects.c
198
static bool pool_pop_batch(struct hlist_head *head, struct obj_pool *src)
lib/debugobjects.c
207
hlist_move_list(&src->objects, head);
lib/debugobjects.c
209
obj = hlist_entry(head->first, typeof(*obj), node);
lib/debugobjects.c
318
static void free_object_list(struct hlist_head *head)
lib/debugobjects.c
324
hlist_for_each_entry_safe(obj, tmp, head, node) {
lib/debugobjects.c
365
static bool kmem_alloc_batch(struct hlist_head *head, struct kmem_cache *cache, gfp_t gfp)
lib/debugobjects.c
373
free_object_list(head);
lib/debugobjects.c
382
hlist_add_head(&obj->node, head);
lib/debugobjects.c
402
HLIST_HEAD(head);
lib/debugobjects.c
420
if (!kmem_alloc_batch(&head, obj_cache, gfp))
lib/debugobjects.c
424
if (!pool_push_batch(&pool_global, &head))
lib/debugobjects.c
425
pool_push_batch(&pool_to_free, &head);
lib/digsig.c
157
head = len - l;
lib/digsig.c
158
memcpy(out1 + head, p, l);
lib/digsig.c
71
int head, i;
lib/iov_iter.c
340
struct page *head;
lib/iov_iter.c
353
head = compound_head(page);
lib/iov_iter.c
354
v += (page - head) << PAGE_SHIFT;
lib/iov_iter.c
356
if (WARN_ON(n > v || v > page_size(head)))
lib/list_sort.c
16
struct list_head *head, **tail = &head;
lib/list_sort.c
189
void list_sort(void *priv, struct list_head *head, list_cmp_func_t cmp)
lib/list_sort.c
191
struct list_head *list = head->next, *pending = NULL;
lib/list_sort.c
194
if (list == head->prev) /* Zero or one elements */
lib/list_sort.c
198
head->prev->next = NULL;
lib/list_sort.c
255
merge_final(priv, cmp, head, pending, list);
lib/list_sort.c
38
return head;
lib/list_sort.c
49
static void merge_final(void *priv, list_cmp_func_t cmp, struct list_head *head,
lib/list_sort.c
52
struct list_head *tail = head;
lib/list_sort.c
94
tail->next = head;
lib/list_sort.c
95
head->prev = tail;
lib/llist.c
31
struct llist_node *llist_del_first(struct llist_head *head)
lib/llist.c
35
entry = smp_load_acquire(&head->first);
lib/llist.c
40
} while (!try_cmpxchg(&head->first, &entry, next));
lib/llist.c
57
bool llist_del_first_this(struct llist_head *head,
lib/llist.c
63
entry = smp_load_acquire(&head->first);
lib/llist.c
68
} while (!try_cmpxchg(&head->first, &entry, next));
lib/llist.c
81
struct llist_node *llist_reverse_order(struct llist_node *head)
lib/llist.c
85
while (head) {
lib/llist.c
86
struct llist_node *tmp = head;
lib/llist.c
87
head = head->next;
lib/maple_tree.c
4907
static void mt_free_walk(struct rcu_head *head)
lib/maple_tree.c
4915
node = container_of(head, struct maple_node, rcu);
lib/maple_tree.c
913
mat->tail = mat->head = dead_enode;
lib/maple_tree.c
921
static void mt_free_walk(struct rcu_head *head);
lib/maple_tree.c
937
while (mat->head) {
lib/maple_tree.c
938
next = mte_to_mat(mat->head)->next;
lib/maple_tree.c
939
node = mte_to_node(mat->head);
lib/maple_tree.c
940
mt_destroy_walk(mat->head, mas->tree, !in_rcu);
lib/maple_tree.c
943
mat->head = next;
lib/nlattr.c
60
static int __nla_validate_parse(const struct nlattr *head, int len, int maxtype,
lib/nlattr.c
604
static int __nla_validate_parse(const struct nlattr *head, int len, int maxtype,
lib/nlattr.c
622
nla_for_each_attr(nla, head, len, rem) {
lib/nlattr.c
673
int __nla_validate(const struct nlattr *head, int len, int maxtype,
lib/nlattr.c
677
return __nla_validate_parse(head, len, maxtype, policy, validate,
lib/nlattr.c
728
const struct nlattr *head, int len,
lib/nlattr.c
732
return __nla_validate_parse(head, len, maxtype, policy, validate,
lib/nlattr.c
745
struct nlattr *nla_find(const struct nlattr *head, int len, int attrtype)
lib/nlattr.c
750
nla_for_each_attr(nla, head, len, rem)
lib/nlattr.c
89
static int nla_validate_array(const struct nlattr *head, int len, int maxtype,
lib/nlattr.c
97
nla_for_each_attr(entry, head, len, rem) {
lib/plist.c
113
plist_check_head(head);
lib/plist.c
122
void plist_del(struct plist_node *node, struct plist_head *head)
lib/plist.c
124
plist_check_head(head);
lib/plist.c
127
if (node->node_list.next != &head->node_list) {
lib/plist.c
142
plist_check_head(head);
lib/plist.c
155
void plist_requeue(struct plist_node *node, struct plist_head *head)
lib/plist.c
158
struct list_head *node_next = &head->node_list;
lib/plist.c
160
plist_check_head(head);
lib/plist.c
161
BUG_ON(plist_head_empty(head));
lib/plist.c
164
if (node == plist_last(head))
lib/plist.c
172
plist_del(node, head);
lib/plist.c
185
plist_for_each_continue(iter, head) {
lib/plist.c
194
plist_check_head(head);
lib/plist.c
56
static void plist_check_head(struct plist_head *head)
lib/plist.c
58
if (!plist_head_empty(head))
lib/plist.c
59
plist_check_list(&plist_first(head)->prio_list);
lib/plist.c
60
plist_check_list(&head->node_list);
lib/plist.c
73
void plist_add(struct plist_node *node, struct plist_head *head)
lib/plist.c
76
struct list_head *node_next = &head->node_list;
lib/plist.c
78
plist_check_head(head);
lib/plist.c
82
if (plist_head_empty(head))
lib/plist.c
85
first = iter = plist_first(head);
lib/radix-tree.c
290
void radix_tree_node_rcu_free(struct rcu_head *head)
lib/radix-tree.c
293
container_of(head, struct radix_tree_node, rcu_head);
lib/radix-tree.h
8
extern void radix_tree_node_rcu_free(struct rcu_head *head);
lib/ref_tracker.c
17
struct list_head head; /* anchor into dir->list or dir->quarantine */
lib/ref_tracker.c
232
list_for_each_entry_safe(tracker, n, &dir->quarantine, head) {
lib/ref_tracker.c
233
list_del(&tracker->head);
lib/ref_tracker.c
240
list_for_each_entry_safe(tracker, n, &dir->list, head) {
lib/ref_tracker.c
241
list_del(&tracker->head);
lib/ref_tracker.c
280
list_add(&tracker->head, &dir->list);
lib/ref_tracker.c
329
list_move_tail(&tracker->head, &dir->quarantine);
lib/ref_tracker.c
331
tracker = list_first_entry(&dir->quarantine, struct ref_tracker, head);
lib/ref_tracker.c
332
list_del(&tracker->head);
lib/ref_tracker.c
83
list_for_each_entry(tracker, &dir->list, head) {
lib/rhashtable.c
117
static void bucket_table_free_rcu(struct rcu_head *head)
lib/rhashtable.c
119
bucket_table_free(container_of(head, struct bucket_table, rcu));
lib/rhashtable.c
235
struct rhash_head *head, *next, *entry;
lib/rhashtable.c
264
head = rht_ptr(new_tbl->buckets + new_hash, new_tbl, new_hash);
lib/rhashtable.c
266
RCU_INIT_POINTER(entry->next, head);
lib/rhashtable.c
506
struct rhash_head *head;
lib/rhashtable.c
510
rht_for_each_from(head, rht_ptr(bkt, tbl, hash), tbl, hash) {
lib/rhashtable.c
517
ht->p.obj_cmpfn(&arg, rht_obj(ht, head)) :
lib/rhashtable.c
518
rhashtable_compare(&arg, rht_obj(ht, head)))) {
lib/rhashtable.c
519
pprev = &head->next;
lib/rhashtable.c
524
return rht_obj(ht, head);
lib/rhashtable.c
527
plist = container_of(head, struct rhlist_head, rhead);
lib/rhashtable.c
530
head = rht_dereference_bucket(head->next, tbl, hash);
lib/rhashtable.c
531
RCU_INIT_POINTER(list->rhead.next, head);
lib/rhashtable.c
553
struct rhash_head *head;
lib/rhashtable.c
574
head = rht_ptr(bkt, tbl, hash);
lib/rhashtable.c
576
RCU_INIT_POINTER(obj->next, head);
lib/test_hmm.c
1689
static void dmirror_devmem_folio_split(struct folio *head, struct folio *tail)
lib/test_hmm.c
1691
struct page *rpage = BACKING_PAGE(folio_page(head, 0));
lib/test_hmm.c
1710
offset = folio_pfn(tail) - folio_pfn(head);
lib/test_hmm.c
1718
folio_page(tail, 0)->mapping = folio_page(head, 0)->mapping;
lib/test_hmm.c
1719
tail->pgmap = head->pgmap;
lib/test_objpool.c
437
static int ot_objpool_release(struct objpool_head *head, void *context)
lib/test_objpool.c
441
WARN_ON(!head || !sop || head != &sop->pool);
lib/tests/list-private-test.c
32
LIST_HEAD(head);
lib/tests/list-private-test.c
35
list_add(&ACCESS_PRIVATE(&entry, list), &head);
lib/tests/list-private-test.c
39
pos = list_private_first_entry(&head, struct list_test_struct, list);
lib/tests/list-private-test.c
40
pos = list_private_last_entry(&head, struct list_test_struct, list);
lib/tests/list-private-test.c
43
pos = list_private_next_entry_circular(pos, &head, list);
lib/tests/list-private-test.c
44
pos = list_private_prev_entry_circular(pos, &head, list);
lib/tests/list-private-test.c
46
if (list_private_entry_is_head(pos, &head, list))
lib/tests/list-private-test.c
49
list_private_for_each_entry(pos, &head, list) { }
lib/tests/list-private-test.c
50
list_private_for_each_entry_reverse(pos, &head, list) { }
lib/tests/list-private-test.c
51
list_private_for_each_entry_continue(pos, &head, list) { }
lib/tests/list-private-test.c
52
list_private_for_each_entry_continue_reverse(pos, &head, list) { }
lib/tests/list-private-test.c
53
list_private_for_each_entry_from(pos, &head, list) { }
lib/tests/list-private-test.c
54
list_private_for_each_entry_from_reverse(pos, &head, list) { }
lib/tests/list-private-test.c
56
list_private_for_each_entry_safe(pos, n, &head, list)
lib/tests/list-private-test.c
58
list_private_for_each_entry_safe_continue(pos, n, &head, list) { }
lib/tests/list-private-test.c
59
list_private_for_each_entry_safe_from(pos, n, &head, list) { }
lib/tests/list-private-test.c
60
list_private_for_each_entry_safe_reverse(pos, n, &head, list) { }
lib/tests/test_list_sort.c
104
KUNIT_EXPECT_PTR_EQ_MSG(test, head.prev, cur, "list is corrupted");
lib/tests/test_list_sort.c
63
LIST_HEAD(head);
lib/tests/test_list_sort.c
79
list_add_tail(&el->list, &head);
lib/tests/test_list_sort.c
82
list_sort(test, &head, cmp);
lib/tests/test_list_sort.c
84
for (cur = head.next; cur->next != &head; cur = cur->next) {
lib/timerqueue.c
35
bool timerqueue_add(struct timerqueue_head *head, struct timerqueue_node *node)
lib/timerqueue.c
40
return rb_add_cached(&node->node, &head->rb_root, __timerqueue_less);
lib/timerqueue.c
53
bool timerqueue_del(struct timerqueue_head *head, struct timerqueue_node *node)
lib/timerqueue.c
57
rb_erase_cached(&node->node, &head->rb_root);
lib/timerqueue.c
60
return !RB_EMPTY_ROOT(&head->rb_root.rb_root);
lib/xarray.c
567
static int xas_expand(struct xa_state *xas, void *head)
lib/xarray.c
574
if (!head) {
lib/xarray.c
580
} else if (xa_is_node(head)) {
lib/xarray.c
581
node = xa_to_node(head);
lib/xarray.c
586
while (max > max_index(head)) {
lib/xarray.c
595
if (xa_is_value(head))
lib/xarray.c
597
RCU_INIT_POINTER(node->slots[0], head);
lib/xarray.c
619
if (xa_is_node(head)) {
lib/xarray.c
620
xa_to_node(head)->offset = 0;
lib/xarray.c
621
rcu_assign_pointer(xa_to_node(head)->parent, node);
lib/xarray.c
623
head = xa_mk_node(node);
lib/xarray.c
624
rcu_assign_pointer(xa->xa_head, head);
lib/zlib_deflate/deflate.c
173
s->prev[(str) & s->w_mask] = match_head = s->head[s->ins_h], \
lib/zlib_deflate/deflate.c
174
s->head[s->ins_h] = (Pos)(str))
lib/zlib_deflate/deflate.c
181
s->head[s->hash_size-1] = NIL; \
lib/zlib_deflate/deflate.c
182
memset((char *)s->head, 0, (unsigned)(s->hash_size-1)*sizeof(*s->head));
lib/zlib_deflate/deflate.c
260
s->head = (Pos *) mem->head_memory;
lib/zlib_deflate/deflate.c
780
p = &s->head[n];
lib/zlib_deflate/defutil.h
112
Pos *head; /* Heads of the hash chains or NIL. */
lib/zstd/compress/zstd_lazy.c
1000
if (nbChunks == 1) return ZSTD_rotateRight_U16((U16)matches[0], head);
lib/zstd/compress/zstd_lazy.c
1001
if (nbChunks == 2) return ZSTD_rotateRight_U32((U32)matches[1] << 16 | (U32)matches[0], head);
lib/zstd/compress/zstd_lazy.c
1003
return ZSTD_rotateRight_U64((U64)matches[3] << 48 | (U64)matches[2] << 32 | (U64)matches[1] << 16 | (U64)matches[0], head);
lib/zstd/compress/zstd_lazy.c
989
ZSTD_row_getSSEMask(int nbChunks, const BYTE* const src, const BYTE tag, const U32 head)
mm/filemap.c
3025
pipe->head++;
mm/hugetlb.c
1141
struct list_head *head = &resv_map->region_cache;
mm/hugetlb.c
1148
list_for_each_entry_safe(rg, trg, head, link) {
mm/hugetlb.c
601
struct list_head *head = &resv->regions;
mm/hugetlb.c
613
list_for_each_entry_safe(iter, trg, head, link) {
mm/hugetlb.c
648
rg = head->prev;
mm/hugetlb.c
857
struct list_head *head = &resv->regions;
mm/hugetlb.c
864
list_for_each_entry_safe(rg, trg, head, link) {
mm/hugetlb.c
984
struct list_head *head = &resv->regions;
mm/hugetlb.c
990
list_for_each_entry(rg, head, link) {
mm/hugetlb_vmemmap.c
102
struct page *head;
mm/hugetlb_vmemmap.c
110
head = pmd_leaf(*pmd) ? pmd_page(*pmd) : NULL;
mm/hugetlb_vmemmap.c
126
struct page *page = head ? head + pte_index(addr) :
mm/hugetlb_vmemmap.c
133
if (!head || ret)
mm/hugetlb_vmemmap.c
136
return vmemmap_split_pmd(pmd, head, addr & PMD_MASK, vmemmap_walk);
mm/hugetlb_vmemmap.c
52
static int vmemmap_split_pmd(pmd_t *pmd, struct page *head, unsigned long start,
mm/hugetlb_vmemmap.c
70
entry = mk_pte(head + i, pgprot);
mm/hugetlb_vmemmap.c
82
if (!PageReserved(head))
mm/hugetlb_vmemmap.c
83
split_page(head, get_order(PMD_SIZE));
mm/internal.h
881
static inline void prep_compound_tail(struct page *head, int tail_idx)
mm/internal.h
883
struct page *p = head + tail_idx;
mm/internal.h
886
set_compound_head(p, head);
mm/kasan/quarantine.c
173
qlink = q->head;
mm/kasan/quarantine.c
299
curr = from->head;
mm/kasan/quarantine.c
36
struct qlist_node *head;
mm/kasan/quarantine.c
46
return !q->head;
mm/kasan/quarantine.c
51
q->head = q->tail = NULL;
mm/kasan/quarantine.c
59
q->head = qlink;
mm/kasan/quarantine.c
78
to->tail->next = from->head;
mm/kmsan/report.c
122
head = entries[1];
mm/kmsan/report.c
126
stack_depot_fetch(head, &chained_entries);
mm/kmsan/report.c
91
depot_stack_handle_t head;
mm/ksm.c
1032
stable_node = rmap_item->head;
mm/ksm.c
1052
rmap_item->head = NULL;
mm/ksm.c
163
struct list_head *head;
mm/ksm.c
1837
if (page_node && page_node->head != &migrate_nodes) {
mm/ksm.c
1879
VM_BUG_ON(page_node->head != &migrate_nodes);
mm/ksm.c
1967
VM_BUG_ON(page_node->head != &migrate_nodes);
mm/ksm.c
1985
VM_BUG_ON(page_node->head != &migrate_nodes);
mm/ksm.c
1997
stable_node_dup->head = &migrate_nodes;
mm/ksm.c
1998
list_add(&stable_node_dup->list, stable_node_dup->head);
mm/ksm.c
2025
VM_BUG_ON(page_node->head != &migrate_nodes);
mm/ksm.c
217
struct ksm_stable_node *head;
mm/ksm.c
2227
rmap_item->head = stable_node;
mm/ksm.c
2261
if (stable_node->head != &migrate_nodes &&
mm/ksm.c
2265
stable_node->head = &migrate_nodes;
mm/ksm.c
2266
list_add(&stable_node->list, stable_node->head);
mm/ksm.c
2268
if (stable_node->head != &migrate_nodes &&
mm/ksm.c
2269
rmap_item->head == stable_node)
mm/ksm.c
2298
if (kfolio == folio && rmap_item->head == stable_node) {
mm/ksm.c
529
return dup->head == STABLE_NODE_DUP_HEAD;
mm/ksm.c
536
dup->head = STABLE_NODE_DUP_HEAD;
mm/ksm.c
557
dup->head = NULL;
mm/ksm.c
914
if (stable_node->head == &migrate_nodes)
mm/list_lru.c
245
struct list_head *head)
mm/list_lru.c
247
list_move(item, head);
mm/memory-failure.c
1223
{ head, head, MF_MSG_HUGE, me_huge_page },
mm/memory-failure.c
1874
struct llist_node *head;
mm/memory-failure.c
1878
head = llist_del_all(raw_hwp_list_head(folio));
mm/memory-failure.c
1879
llist_for_each_entry_safe(p, next, head, node) {
mm/memory-failure.c
1902
struct llist_head *head;
mm/memory-failure.c
1914
head = raw_hwp_list_head(folio);
mm/memory-failure.c
1915
llist_for_each_entry(p, head->first, node) {
mm/memory-failure.c
1923
llist_add(&raw_hwp->node, head);
mm/migrate.c
1000
} while (bh != head);
mm/migrate.c
902
static bool buffer_migrate_lock_buffers(struct buffer_head *head,
mm/migrate.c
905
struct buffer_head *bh = head;
mm/migrate.c
918
} while (bh != head);
mm/migrate.c
925
bh = head;
mm/migrate.c
938
struct buffer_head *bh, *head;
mm/migrate.c
942
head = folio_buffers(src);
mm/migrate.c
943
if (!head)
mm/migrate.c
951
if (!buffer_migrate_lock_buffers(head, mode))
mm/migrate.c
958
migrating = test_and_set_bit_lock(BH_Migrate, &head->b_state);
mm/migrate.c
963
bh = head;
mm/migrate.c
970
} while (bh != head);
mm/migrate.c
987
bh = head;
mm/migrate.c
991
} while (bh != head);
mm/migrate.c
995
clear_bit_unlock(BH_Migrate, &head->b_state);
mm/migrate.c
996
bh = head;
mm/mm_init.c
1082
static void __ref memmap_init_compound(struct page *head,
mm/mm_init.c
1097
__SetPageHead(head);
mm/mm_init.c
1102
prep_compound_tail(head, pfn - head_pfn);
mm/mm_init.c
1105
prep_compound_head(head, order);
mm/mmu_gather.c
289
static void tlb_remove_table_rcu(struct rcu_head *head)
mm/mmu_gather.c
291
__tlb_remove_table_free(container_of(head, struct mmu_table_batch, rcu));
mm/mmu_gather.c
324
static inline void __tlb_remove_table_one_rcu(struct rcu_head *head)
mm/mmu_gather.c
328
ptdesc = container_of(head, struct ptdesc, pt_rcu_head);
mm/page_alloc.c
5326
int head = PageHead(page);
mm/page_alloc.c
5332
else if (!head) {
mm/page_alloc.c
7116
struct page *head = pfn_to_page(start);
mm/page_alloc.c
7118
check_new_pages(head, order);
mm/page_alloc.c
7119
prep_new_page(head, order, gfp_mask, 0);
mm/page_alloc.c
7485
const struct page *head = page - (pfn & ((1 << order) - 1));
mm/page_alloc.c
7487
if (PageBuddy(head) &&
mm/page_alloc.c
7488
buddy_order_unsafe(head) >= order)
mm/page_isolation.c
417
struct page *head = compound_head(page);
mm/page_isolation.c
418
unsigned long head_pfn = page_to_pfn(head);
mm/page_isolation.c
419
unsigned long nr_pages = compound_nr(head);
mm/pgtable-generic.c
240
static void pte_free_now(struct rcu_head *head)
mm/pgtable-generic.c
244
page = container_of(head, struct page, rcu_head);
mm/shmem.c
3547
pipe->head++;
mm/shmem.c
3733
WARN_ON_ONCE(!list_empty(&shmem_falloc_waitq.head));
mm/shrinker.c
761
static void shrinker_free_rcu_cb(struct rcu_head *head)
mm/shrinker.c
763
struct shrinker *shrinker = container_of(head, struct shrinker, rcu);
mm/slab.h
134
unsigned long head;
mm/slab.h
136
head = READ_ONCE(page->compound_head);
mm/slab.h
137
if (head & 1)
mm/slab.h
138
page = (struct page *)(head - 1);
mm/slab.h
640
void kvfree_rcu_cb(struct rcu_head *head);
mm/slab_common.c
1268
void kvfree_call_rcu(struct rcu_head *head, void *ptr)
mm/slab_common.c
1270
if (head) {
mm/slab_common.c
1272
call_rcu(head, kvfree_rcu_cb);
mm/slab_common.c
1385
struct rcu_head *head;
mm/slab_common.c
1526
kvfree_rcu_list(struct rcu_head *head)
mm/slab_common.c
1530
for (; head; head = next) {
mm/slab_common.c
1531
void *ptr = (void *) head->func;
mm/slab_common.c
1532
unsigned long offset = (void *) head - ptr;
mm/slab_common.c
1534
next = head->next;
mm/slab_common.c
1537
trace_rcu_invoke_kvfree_callback("slab", head, offset);
mm/slab_common.c
1555
struct rcu_head *head;
mm/slab_common.c
1571
head = krwp->head_free;
mm/slab_common.c
1590
if (head && !WARN_ON_ONCE(!poll_state_synchronize_rcu_full(&head_gp_snap)))
mm/slab_common.c
1591
kvfree_rcu_list(head);
mm/slab_common.c
1622
return !!READ_ONCE(krcp->head);
mm/slab_common.c
1695
if (krcp->head && poll_state_synchronize_rcu(krcp->head_gp_snap)) {
mm/slab_common.c
1696
head_ready = krcp->head;
mm/slab_common.c
1698
WRITE_ONCE(krcp->head, NULL);
mm/slab_common.c
1748
krwp->head_free = krcp->head;
mm/slab_common.c
1751
WRITE_ONCE(krcp->head, NULL);
mm/slab_common.c
1941
void kvfree_call_rcu(struct rcu_head *head, void *ptr)
mm/slab_common.c
1954
if (!head)
mm/slab_common.c
1964
__func__, head);
mm/slab_common.c
1971
success = add_ptr_to_bulk_krc_lock(&krcp, &flags, ptr, !head);
mm/slab_common.c
1975
if (head == NULL)
mm/slab_common.c
1979
head->func = ptr;
mm/slab_common.c
1980
head->next = krcp->head;
mm/slab_common.c
1981
WRITE_ONCE(krcp->head, head);
mm/slub.c
2003
struct slab *slab, void *head, void *tail, int *bulk_cnt,
mm/slub.c
2572
struct rcu_head head;
mm/slub.c
2649
call_rcu(&delayed_free->head, slab_free_after_rcu_debug);
mm/slub.c
2689
bool slab_free_freelist_hook(struct kmem_cache *s, void **head, void **tail,
mm/slub.c
2694
void *next = *head;
mm/slub.c
2704
*head = NULL;
mm/slub.c
2716
set_freepointer(s, object, *head);
mm/slub.c
2717
*head = object;
mm/slub.c
2729
return *head != NULL;
mm/slub.c
2968
static void rcu_free_sheaf_nobarn(struct rcu_head *head)
mm/slub.c
2973
sheaf = container_of(head, struct slab_sheaf, rcu_head);
mm/slub.c
4115
struct slab *slab, void *head, void *tail, int *bulk_cnt,
mm/slub.c
4119
void *object = head;
mm/slub.c
5399
void *head, void *tail, int bulk_cnt,
mm/slub.c
5417
if (free_debug_processing(s, slab, head, tail, &cnt, addr, handle)) {
mm/slub.c
5423
slab->freelist = head;
mm/slub.c
5471
void *head, void *tail, int cnt,
mm/slub.c
5482
free_to_partial_list(s, slab, head, tail, cnt, addr);
mm/slub.c
5499
new.freelist = head;
mm/slub.c
5789
static void rcu_free_sheaf(struct rcu_head *head)
mm/slub.c
5796
sheaf = container_of(head, struct slab_sheaf, rcu_head);
mm/slub.c
6137
static void defer_free(struct kmem_cache *s, void *head)
mm/slub.c
6143
head = kasan_reset_tag(head);
mm/slub.c
6146
if (llist_add(head + s->offset, &df->objects))
mm/slub.c
6193
void slab_free_bulk(struct kmem_cache *s, struct slab *slab, void *head,
mm/slub.c
6202
if (likely(slab_free_freelist_hook(s, &head, &tail, &cnt))) {
mm/slub.c
6203
__slab_free(s, slab, head, tail, cnt, addr);
mm/slub.c
6212
container_of(rcu_head, struct rcu_delayed_free, head);
mm/slub.c
6415
void kvfree_rcu_cb(struct rcu_head *head)
mm/slub.c
6417
void *obj = head;
mm/slub.c
7050
void *head = object;
mm/slub.c
7059
__slab_free(s, slab, head, tail, cnt, _RET_IP_);
mm/swapfile.c
3705
struct page *head;
mm/swapfile.c
3747
head = vmalloc_to_page(si->swap_map + offset);
mm/swapfile.c
3755
if (!page_private(head)) {
mm/swapfile.c
3757
INIT_LIST_HEAD(&head->lru);
mm/swapfile.c
3758
set_page_private(head, SWP_CONTINUED);
mm/swapfile.c
3762
list_for_each_entry(list_page, &head->lru, lru) {
mm/swapfile.c
3784
list_add_tail(&page->lru, &head->lru);
mm/swapfile.c
3809
struct page *head;
mm/swapfile.c
3814
head = vmalloc_to_page(si->swap_map + offset);
mm/swapfile.c
3815
if (page_private(head) != SWP_CONTINUED) {
mm/swapfile.c
3822
page = list_next_entry(head, lru);
mm/swapfile.c
3835
BUG_ON(page == head);
mm/swapfile.c
3841
if (page == head) {
mm/swapfile.c
3850
while ((page = list_prev_entry(page, lru)) != head) {
mm/swapfile.c
3865
BUG_ON(page == head);
mm/swapfile.c
3873
while ((page = list_prev_entry(page, lru)) != head) {
mm/swapfile.c
3895
struct page *head;
mm/swapfile.c
3896
head = vmalloc_to_page(si->swap_map + offset);
mm/swapfile.c
3897
if (page_private(head)) {
mm/swapfile.c
3900
list_for_each_entry_safe(page, next, &head->lru, lru) {
mm/swapfile.c
430
static void swap_table_free_folio_rcu_cb(struct rcu_head *head)
mm/swapfile.c
434
folio = page_folio(container_of(head, struct page, rcu_head));
mm/util.c
1240
unsigned long head, nr_pages = 1;
mm/util.c
1250
head = ps->page_snapshot.compound_head;
mm/util.c
1251
if ((head & 1) == 0) {
mm/util.c
1262
foliop = (struct folio *)(head - 1);
mm/vmalloc.c
1249
struct list_head *head, bool augment)
mm/vmalloc.c
1256
head = &rb_entry(parent, struct vmap_area, rb_node)->list;
mm/vmalloc.c
1258
head = head->prev;
mm/vmalloc.c
1283
list_add(&va->list, head);
mm/vmalloc.c
1289
struct list_head *head)
mm/vmalloc.c
1291
__link_va(va, root, parent, link, head, false);
mm/vmalloc.c
1297
struct list_head *head)
mm/vmalloc.c
1299
__link_va(va, root, parent, link, head, true);
mm/vmalloc.c
1401
struct rb_root *root, struct list_head *head)
mm/vmalloc.c
1408
link_va(va, root, parent, link, head);
mm/vmalloc.c
1414
struct list_head *head)
mm/vmalloc.c
1425
link_va_augment(va, root, parent, link, head);
mm/vmalloc.c
1443
struct rb_root *root, struct list_head *head, bool augment)
mm/vmalloc.c
1473
if (next != head) {
mm/vmalloc.c
1494
if (next->prev != head) {
mm/vmalloc.c
1520
__link_va(va, root, parent, link, head, augment);
mm/vmalloc.c
1527
struct rb_root *root, struct list_head *head)
mm/vmalloc.c
1529
return __merge_or_add_vmap_area(va, root, head, false);
mm/vmalloc.c
1534
struct rb_root *root, struct list_head *head)
mm/vmalloc.c
1536
va = __merge_or_add_vmap_area(va, root, head, true);
mm/vmalloc.c
1637
find_vmap_lowest_linear_match(struct list_head *head, unsigned long size,
mm/vmalloc.c
1642
list_for_each_entry(va, head, list) {
mm/vmalloc.c
1653
find_vmap_lowest_match_check(struct rb_root *root, struct list_head *head,
mm/vmalloc.c
1664
va_2 = find_vmap_lowest_linear_match(head, size, align, vstart);
mm/vmalloc.c
1707
va_clip(struct rb_root *root, struct list_head *head,
mm/vmalloc.c
1800
insert_vmap_area_augment(lva, &va->rb_node, root, head);
mm/vmalloc.c
1808
struct rb_root *root, struct list_head *head,
mm/vmalloc.c
1825
ret = va_clip(root, head, va, nva_start_addr, size);
mm/vmalloc.c
1837
__alloc_vmap_area(struct rb_root *root, struct list_head *head,
mm/vmalloc.c
1861
nva_start_addr = va_alloc(va, root, head, size, align, vstart, vend);
mm/vmalloc.c
1865
find_vmap_lowest_match_check(root, head, size, align);
mm/vmalloc.c
1938
list_add(&va->list, &vp->head);
mm/vmalloc.c
1955
if (!vp || list_empty(&vp->head))
mm/vmalloc.c
1959
if (!list_empty(&vp->head)) {
mm/vmalloc.c
1960
va = list_first_entry(&vp->head, struct vmap_area, list);
mm/vmalloc.c
1978
list_move_tail(&va->list, &vp->head);
mm/vmalloc.c
2122
insert_vmap_area(va, &vn->busy.root, &vn->busy.head);
mm/vmalloc.c
2209
reclaim_list_global(struct list_head *head)
mm/vmalloc.c
2213
if (list_empty(head))
mm/vmalloc.c
2217
list_for_each_entry_safe(va, n, head, list)
mm/vmalloc.c
2235
if (list_empty(&vn->pool[i].head))
mm/vmalloc.c
2240
list_replace_init(&vn->pool[i].head, &tmp_list);
mm/vmalloc.c
2267
list_replace_init(&tmp_list, &vn->pool[i].head);
mm/vmalloc.c
2368
list_replace_init(&vn->lazy.head, &vn->purge_list);
mm/vmalloc.c
2467
insert_vmap_area(va, &vn->lazy.root, &vn->lazy.head);
mm/vmalloc.c
3766
struct llist_node *head;
mm/vmalloc.c
3768
head = llist_del_all(&pending_vm_area_cleanup);
mm/vmalloc.c
3769
if (!head)
mm/vmalloc.c
3772
llist_for_each_entry_safe(area, tmp, head, llnode) {
mm/vmalloc.c
5029
insert_vmap_area(vas[area], &vn->busy.root, &vn->busy.head);
mm/vmalloc.c
5207
list_for_each_entry(va, &vn->lazy.head, list) {
mm/vmalloc.c
5228
list_for_each_entry(va, &vn->busy.head, list) {
mm/vmalloc.c
5383
INIT_LIST_HEAD(&vn->busy.head);
mm/vmalloc.c
5387
INIT_LIST_HEAD(&vn->lazy.head);
mm/vmalloc.c
5391
INIT_LIST_HEAD(&vn->pool[i].head);
mm/vmalloc.c
5467
insert_vmap_area(va, &vn->busy.root, &vn->busy.head);
mm/vmalloc.c
915
struct list_head head;
mm/vmalloc.c
927
struct list_head head;
mm/vmscan.c
2892
mm = list_entry(mm_state->head, struct mm_struct, lru_gen.list);
mm/vmscan.c
2951
if (mm_state->head == &mm->lru_gen.list)
mm/vmscan.c
2952
mm_state->head = mm_state->head->prev;
mm/vmscan.c
3070
if (!mm_state->head)
mm/vmscan.c
3071
mm_state->head = &mm_list->fifo;
mm/vmscan.c
3073
if (mm_state->head == &mm_list->fifo)
mm/vmscan.c
3077
mm_state->head = mm_state->head->next;
mm/vmscan.c
3078
if (mm_state->head == &mm_list->fifo) {
mm/vmscan.c
3085
if (!mm_state->tail || mm_state->tail == mm_state->head) {
mm/vmscan.c
3086
mm_state->tail = mm_state->head->next;
mm/vmscan.c
3119
mm_state->head = NULL;
mm/vmscan.c
3864
struct list_head *head = &lrugen->folios[old_gen][type][zone];
mm/vmscan.c
3866
while (!list_empty(head)) {
mm/vmscan.c
3867
struct folio *folio = lru_to_folio(head);
mm/vmscan.c
4567
struct list_head *head = &lrugen->folios[gen][type][zone];
mm/vmscan.c
4569
while (!list_empty(head)) {
mm/vmscan.c
4570
struct folio *folio = lru_to_folio(head);
mm/vmscan.c
4595
list_splice(&moved, head);
mm/vmscan.c
5115
struct list_head *head = &lruvec->lists[lru];
mm/vmscan.c
5117
while (!list_empty(head)) {
mm/vmscan.c
5119
struct folio *folio = lru_to_folio(head);
mm/vmscan.c
5144
struct list_head *head = &lruvec->lrugen.folios[gen][type][zone];
mm/vmscan.c
5146
while (!list_empty(head)) {
mm/vmscan.c
5148
struct folio *folio = lru_to_folio(head);
net/6lowpan/iphc.c
1136
u8 head[LOWPAN_IPHC_MAX_HC_BUF_LEN] = {};
net/6lowpan/iphc.c
1144
hc_ptr = head + 2;
net/6lowpan/iphc.c
1299
head[0] = iphc0;
net/6lowpan/iphc.c
1300
head[1] = iphc1;
net/6lowpan/iphc.c
1304
memcpy(skb_push(skb, hc_ptr - head), head, hc_ptr - head);
net/6lowpan/iphc.c
1307
pr_debug("header len %d skb %u\n", (int)(hc_ptr - head), skb->len);
net/8021q/vlan.c
117
unregister_netdevice_queue(dev, head);
net/8021q/vlan.c
88
void unregister_vlan_dev(struct net_device *dev, struct list_head *head)
net/8021q/vlan.h
143
void unregister_vlan_dev(struct net_device *dev, struct list_head *head);
net/8021q/vlan_core.c
464
static struct sk_buff *vlan_gro_receive(struct list_head *head,
net/8021q/vlan_core.c
491
list_for_each_entry(p, head, list) {
net/8021q/vlan_core.c
507
head, skb);
net/atm/addr.c
108
struct list_head *head;
net/atm/addr.c
116
head = &dev->lecs;
net/atm/addr.c
118
head = &dev->local;
net/atm/addr.c
119
list_for_each_entry(this, head, entry) {
net/atm/addr.c
124
if (head == &dev->local)
net/atm/addr.c
138
struct list_head *head;
net/atm/addr.c
144
head = &dev->lecs;
net/atm/addr.c
146
head = &dev->local;
net/atm/addr.c
147
list_for_each_entry(this, head, entry)
net/atm/addr.c
154
list_for_each_entry(this, head, entry)
net/atm/addr.c
52
struct list_head *head;
net/atm/addr.c
56
head = &dev->lecs;
net/atm/addr.c
58
head = &dev->local;
net/atm/addr.c
59
list_for_each_entry_safe(this, p, head, entry) {
net/atm/addr.c
64
if (head == &dev->local)
net/atm/addr.c
73
struct list_head *head;
net/atm/addr.c
81
head = &dev->lecs;
net/atm/addr.c
83
head = &dev->local;
net/atm/addr.c
84
list_for_each_entry(this, head, entry) {
net/atm/addr.c
96
list_add(&this->entry, head);
net/atm/addr.c
98
if (head == &dev->local)
net/atm/common.c
268
struct hlist_head *head = &vcc_hash[i];
net/atm/common.c
273
sk_for_each_safe(s, tmp, head) {
net/atm/common.c
315
struct hlist_head *head = &vcc_hash[vci & (VCC_HTABLE_SIZE - 1)];
net/atm/common.c
319
sk_for_each(s, head) {
net/atm/common.c
46
struct hlist_head *head = &vcc_hash[vcc->vci & (VCC_HTABLE_SIZE - 1)];
net/atm/common.c
48
sk_add_node(sk, head);
net/atm/lec.c
1552
struct hlist_head *head;
net/atm/lec.c
1557
head = &priv->lec_arp_tables[HASH(mac_addr[ETH_ALEN - 1])];
net/atm/lec.c
1558
hlist_for_each_entry(entry, head, next) {
net/atm/lec.c
237
(long)skb->head, (long)skb->data, (long)skb_tail_pointer(skb),
net/atm/proc.c
84
struct hlist_head *head = &vcc_hash[*bucket];
net/atm/proc.c
86
sk = hlist_empty(head) ? NULL : __sk_head(head);
net/atm/signaling.c
262
struct hlist_head *head = &vcc_hash[i];
net/atm/signaling.c
264
sk_for_each(s, head) {
net/batman-adv/bat_iv_ogm.c
1940
struct hlist_head *head, int *idx_s, int *sub)
net/batman-adv/bat_iv_ogm.c
1946
hlist_for_each_entry_rcu(orig_node, head, hash_entry) {
net/batman-adv/bat_iv_ogm.c
1978
struct hlist_head *head;
net/batman-adv/bat_iv_ogm.c
1985
head = &hash->table[bucket];
net/batman-adv/bat_iv_ogm.c
1989
bat_priv, if_outgoing, head,
net/batman-adv/bat_iv_ogm.c
754
struct hlist_head *head;
net/batman-adv/bat_iv_ogm.c
762
head = &hash->table[i];
net/batman-adv/bat_iv_ogm.c
765
hlist_for_each_entry_rcu(orig_node, head, hash_entry) {
net/batman-adv/bat_v.c
387
struct hlist_head *head, int *idx_s, int *sub)
net/batman-adv/bat_v.c
393
hlist_for_each_entry_rcu(orig_node, head, hash_entry) {
net/batman-adv/bat_v.c
424
struct hlist_head *head;
net/batman-adv/bat_v.c
431
head = &hash->table[bucket];
net/batman-adv/bat_v.c
435
bat_priv, if_outgoing, head, &idx,
net/batman-adv/bridge_loop_avoidance.c
1223
struct hlist_head *head;
net/batman-adv/bridge_loop_avoidance.c
1233
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
1238
head, hash_entry) {
net/batman-adv/bridge_loop_avoidance.c
1278
struct hlist_head *head;
net/batman-adv/bridge_loop_avoidance.c
1287
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
1290
hlist_for_each_entry_rcu(claim, head, hash_entry) {
net/batman-adv/bridge_loop_avoidance.c
1333
struct hlist_head *head;
net/batman-adv/bridge_loop_avoidance.c
1357
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
1360
hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) {
net/batman-adv/bridge_loop_avoidance.c
1430
struct hlist_head *head;
net/batman-adv/bridge_loop_avoidance.c
1471
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
1474
hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) {
net/batman-adv/bridge_loop_avoidance.c
1735
struct hlist_head *head;
net/batman-adv/bridge_loop_avoidance.c
1746
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
1749
hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) {
net/batman-adv/bridge_loop_avoidance.c
222
struct hlist_head *head;
net/batman-adv/bridge_loop_avoidance.c
231
head = &hash->table[index];
net/batman-adv/bridge_loop_avoidance.c
234
hlist_for_each_entry_rcu(claim, head, hash_entry) {
net/batman-adv/bridge_loop_avoidance.c
262
struct hlist_head *head;
net/batman-adv/bridge_loop_avoidance.c
274
head = &hash->table[index];
net/batman-adv/bridge_loop_avoidance.c
277
hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) {
net/batman-adv/bridge_loop_avoidance.c
302
struct hlist_head *head;
net/batman-adv/bridge_loop_avoidance.c
312
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
317
head, hash_entry) {
net/batman-adv/bridge_loop_avoidance.c
594
struct hlist_head *head;
net/batman-adv/bridge_loop_avoidance.c
612
head = &hash->table[i];
net/batman-adv/bridge_loop_avoidance.c
615
hlist_for_each_entry_rcu(claim, head, hash_entry) {
net/batman-adv/distributed-arp-table.c
163
struct hlist_head *head;
net/batman-adv/distributed-arp-table.c
170
head = &bat_priv->dat.hash->table[i];
net/batman-adv/distributed-arp-table.c
174
hlist_for_each_entry_safe(dat_entry, node_tmp, head,
net/batman-adv/distributed-arp-table.c
328
struct hlist_head *head;
net/batman-adv/distributed-arp-table.c
340
head = &hash->table[index];
net/batman-adv/distributed-arp-table.c
343
hlist_for_each_entry_rcu(dat_entry, head, hash_entry) {
net/batman-adv/distributed-arp-table.c
567
struct hlist_head *head;
net/batman-adv/distributed-arp-table.c
579
head = &hash->table[i];
net/batman-adv/distributed-arp-table.c
582
hlist_for_each_entry_rcu(orig_node, head, hash_entry) {
net/batman-adv/fragmentation.c
314
struct hlist_head head = HLIST_HEAD_INIT;
net/batman-adv/fragmentation.c
318
if (!batadv_frag_insert_packet(orig_node_src, *skb, &head))
net/batman-adv/fragmentation.c
322
if (hlist_empty(&head))
net/batman-adv/fragmentation.c
325
skb_out = batadv_frag_merge_packets(&head);
net/batman-adv/fragmentation.c
37
static void batadv_frag_clear_chain(struct hlist_head *head, bool dropped)
net/batman-adv/fragmentation.c
42
hlist_for_each_entry_safe(entry, node, head, list) {
net/batman-adv/hash.h
105
hlist_add_head_rcu(data_node, head);
net/batman-adv/hash.h
136
struct hlist_head *head;
net/batman-adv/hash.h
140
head = &hash->table[index];
net/batman-adv/hash.h
143
hlist_for_each(node, head) {
net/batman-adv/hash.h
83
struct hlist_head *head;
net/batman-adv/hash.h
91
head = &hash->table[index];
net/batman-adv/hash.h
96
hlist_for_each(node, head) {
net/batman-adv/mesh-interface.c
1085
struct list_head *head)
net/batman-adv/mesh-interface.c
1103
unregister_netdevice_queue(mesh_iface, head);
net/batman-adv/multicast.c
1606
struct hlist_head *head = &bat_priv->mcast.want_all_unsnoopables_list;
net/batman-adv/multicast.c
1619
hlist_add_head_rcu(node, head);
net/batman-adv/multicast.c
1651
struct hlist_head *head = &bat_priv->mcast.want_all_ipv4_list;
net/batman-adv/multicast.c
1664
hlist_add_head_rcu(node, head);
net/batman-adv/multicast.c
1696
struct hlist_head *head = &bat_priv->mcast.want_all_ipv6_list;
net/batman-adv/multicast.c
1709
hlist_add_head_rcu(node, head);
net/batman-adv/multicast.c
1741
struct hlist_head *head = &bat_priv->mcast.want_all_rtr4_list;
net/batman-adv/multicast.c
1754
hlist_add_head_rcu(node, head);
net/batman-adv/multicast.c
1786
struct hlist_head *head = &bat_priv->mcast.want_all_rtr6_list;
net/batman-adv/multicast.c
1799
hlist_add_head_rcu(node, head);
net/batman-adv/multicast_forw.c
193
struct hlist_head *head,
net/batman-adv/multicast_forw.c
202
__hlist_for_each_rcu(node, head) {
net/batman-adv/multicast_forw.c
282
struct hlist_head *head = NULL;
net/batman-adv/multicast_forw.c
288
head = &bat_priv->mcast.want_all_ipv4_list;
net/batman-adv/multicast_forw.c
293
head = &bat_priv->mcast.want_all_ipv6_list;
net/batman-adv/multicast_forw.c
301
ret = batadv_mcast_forw_push_dests_list(bat_priv, skb, vid, head,
net/batman-adv/multicast_forw.c
328
struct hlist_head *head = NULL;
net/batman-adv/multicast_forw.c
334
head = &bat_priv->mcast.want_all_rtr4_list;
net/batman-adv/multicast_forw.c
339
head = &bat_priv->mcast.want_all_rtr6_list;
net/batman-adv/multicast_forw.c
347
ret = batadv_mcast_forw_push_dests_list(bat_priv, skb, vid, head,
net/batman-adv/originator.c
1263
struct hlist_head *head;
net/batman-adv/originator.c
1273
head = &hash->table[i];
net/batman-adv/originator.c
1274
if (hlist_empty(head))
net/batman-adv/originator.c
1280
head, hash_entry) {
net/batman-adv/originator.c
57
struct hlist_head *head;
net/batman-adv/originator.c
65
head = &hash->table[index];
net/batman-adv/originator.c
68
hlist_for_each_entry_rcu(orig_node, head, hash_entry) {
net/batman-adv/originator.c
901
struct hlist_head *head;
net/batman-adv/originator.c
914
head = &hash->table[i];
net/batman-adv/originator.c
919
head, hash_entry) {
net/batman-adv/send.c
1096
struct hlist_head head = HLIST_HEAD_INIT;
net/batman-adv/send.c
1108
batadv_forw_packet_list_steal(&bat_priv->forw_bcast_list, &head,
net/batman-adv/send.c
1114
batadv_forw_packet_list_steal(&bat_priv->forw_bat_list, &head,
net/batman-adv/send.c
1119
batadv_forw_packet_list_free(&head);
net/batman-adv/send.c
625
static void batadv_forw_packet_list_free(struct hlist_head *head)
net/batman-adv/send.c
630
hlist_for_each_entry_safe(forw_packet, safe_tmp_node, head,
net/batman-adv/send.c
657
spinlock_t *lock, struct hlist_head *head,
net/batman-adv/send.c
675
hlist_add_head(&forw_packet->list, head);
net/batman-adv/translation-table.c
1273
struct hlist_head *head,
net/batman-adv/translation-table.c
1280
hlist_for_each_entry_safe(tt_common_entry, node_tmp, head,
net/batman-adv/translation-table.c
130
struct hlist_head *head;
net/batman-adv/translation-table.c
1310
struct hlist_head *head;
net/batman-adv/translation-table.c
1315
head = &hash->table[i];
net/batman-adv/translation-table.c
1319
batadv_tt_local_purge_list(bat_priv, head, timeout);
net/batman-adv/translation-table.c
1331
struct hlist_head *head;
net/batman-adv/translation-table.c
1340
head = &hash->table[i];
net/batman-adv/translation-table.c
1345
head, hash_entry) {
net/batman-adv/translation-table.c
1409
const struct hlist_head *head;
net/batman-adv/translation-table.c
141
head = &hash->table[index];
net/batman-adv/translation-table.c
1412
head = &entry->orig_list;
net/batman-adv/translation-table.c
1413
hlist_for_each_entry_rcu(tmp_orig_entry, head, list) {
net/batman-adv/translation-table.c
144
hlist_for_each_entry_rcu(tt, head, hash_entry) {
net/batman-adv/translation-table.c
1471
const struct hlist_head *head;
net/batman-adv/translation-table.c
1475
head = &tt_global->orig_list;
net/batman-adv/translation-table.c
1476
hlist_for_each_entry_rcu(orig_entry, head, list)
net/batman-adv/translation-table.c
1722
struct hlist_head *head;
net/batman-adv/translation-table.c
1725
head = &tt_global_entry->orig_list;
net/batman-adv/translation-table.c
1726
hlist_for_each_entry_rcu(orig_entry, head, list) {
net/batman-adv/translation-table.c
1832
struct hlist_head *head;
net/batman-adv/translation-table.c
1838
head = &global->orig_list;
net/batman-adv/translation-table.c
1840
hlist_for_each_entry_rcu(orig_entry, head, list) {
net/batman-adv/translation-table.c
1872
struct hlist_head *head, int *idx_s, int *sub)
net/batman-adv/translation-table.c
1878
hlist_for_each_entry_rcu(common, head, hash_entry) {
net/batman-adv/translation-table.c
1909
struct hlist_head *head;
net/batman-adv/translation-table.c
1931
head = &hash->table[bucket];
net/batman-adv/translation-table.c
1935
head, &idx, &sub))
net/batman-adv/translation-table.c
1985
struct hlist_head *head;
net/batman-adv/translation-table.c
1990
head = &tt_global_entry->orig_list;
net/batman-adv/translation-table.c
1991
hlist_for_each_entry_safe(orig_entry, safe, head, list)
net/batman-adv/translation-table.c
2012
struct hlist_head *head;
net/batman-adv/translation-table.c
2018
head = &tt_global_entry->orig_list;
net/batman-adv/translation-table.c
2019
hlist_for_each_entry_safe(orig_entry, safe, head, list) {
net/batman-adv/translation-table.c
2045
struct hlist_head *head;
net/batman-adv/translation-table.c
2053
head = &tt_global_entry->orig_list;
net/batman-adv/translation-table.c
2054
hlist_for_each_entry_rcu(orig_entry, head, list) {
net/batman-adv/translation-table.c
2158
struct hlist_head *head;
net/batman-adv/translation-table.c
2166
head = &hash->table[i];
net/batman-adv/translation-table.c
2171
head, hash_entry) {
net/batman-adv/translation-table.c
2223
struct hlist_head *head;
net/batman-adv/translation-table.c
2232
head = &hash->table[i];
net/batman-adv/translation-table.c
2236
hlist_for_each_entry_safe(tt_common, node_tmp, head,
net/batman-adv/translation-table.c
2266
struct hlist_head *head;
net/batman-adv/translation-table.c
2275
head = &hash->table[i];
net/batman-adv/translation-table.c
2280
head, hash_entry) {
net/batman-adv/translation-table.c
2401
struct hlist_head *head;
net/batman-adv/translation-table.c
2407
head = &hash->table[i];
net/batman-adv/translation-table.c
2410
hlist_for_each_entry_rcu(tt_common, head, hash_entry) {
net/batman-adv/translation-table.c
2479
struct hlist_head *head;
net/batman-adv/translation-table.c
2485
head = &hash->table[i];
net/batman-adv/translation-table.c
2488
hlist_for_each_entry_rcu(tt_common, head, hash_entry) {
net/batman-adv/translation-table.c
2719
struct hlist_head *head;
net/batman-adv/translation-table.c
2733
head = &hash->table[i];
net/batman-adv/translation-table.c
2736
head, hash_entry) {
net/batman-adv/translation-table.c
3558
struct hlist_head *head;
net/batman-adv/translation-table.c
3565
head = &hash->table[i];
net/batman-adv/translation-table.c
3569
head, hash_entry) {
net/batman-adv/translation-table.c
3597
struct hlist_head *head;
net/batman-adv/translation-table.c
3605
head = &hash->table[i];
net/batman-adv/translation-table.c
3609
hlist_for_each_entry_safe(tt_common, node_tmp, head,
net/batman-adv/translation-table.c
599
struct hlist_head *head;
net/batman-adv/translation-table.c
720
head = &tt_global->orig_list;
net/batman-adv/translation-table.c
722
hlist_for_each_entry_rcu(orig_entry, head, list) {
net/bluetooth/af_bluetooth.c
178
sk_add_node(sk, &l->head);
net/bluetooth/af_bluetooth.c
200
sk_for_each(sk, &l->head) {
net/bluetooth/af_bluetooth.c
784
return seq_hlist_start_head(&l->head, *pos);
net/bluetooth/af_bluetooth.c
791
return seq_hlist_next(v, &l->head, pos);
net/bluetooth/coredump.c
103
vfree(hdev->dump.head);
net/bluetooth/coredump.c
111
hdev->dump.head = vmalloc(size);
net/bluetooth/coredump.c
112
if (!hdev->dump.head)
net/bluetooth/coredump.c
116
hdev->dump.tail = hdev->dump.head;
net/bluetooth/coredump.c
117
hdev->dump.end = hdev->dump.head + size;
net/bluetooth/coredump.c
250
size = hdev->dump.tail - hdev->dump.head;
net/bluetooth/coredump.c
255
skb_put_data(skb, hdev->dump.head, size);
net/bluetooth/coredump.c
260
dev_coredumpv(&hdev->dev, hdev->dump.head, size, GFP_KERNEL);
net/bluetooth/coredump.c
274
dump_size = hdev->dump.tail - hdev->dump.head;
net/bluetooth/coredump.c
293
dump_size = hdev->dump.tail - hdev->dump.head;
net/bluetooth/coredump.c
409
dump_size = hdev->dump.tail - hdev->dump.head;
net/bluetooth/coredump.c
58
return hci_devcd_update_hdr_state(hdev->dump.head,
net/bluetooth/coredump.c
90
hdev->dump.head = NULL;
net/bluetooth/hci_conn.c
2637
struct list_head *head = &hdev->conn_hash.list;
net/bluetooth/hci_conn.c
2646
while ((conn = list_first_entry_or_null(head,
net/bluetooth/hci_sock.c
212
sk_for_each(sk, &hci_sk_list.head) {
net/bluetooth/hci_sock.c
323
sk_for_each(sk, &hci_sk_list.head) {
net/bluetooth/hci_sock.c
440
sk_for_each(sk, &hci_sk_list.head) {
net/bluetooth/hci_sock.c
770
sk_for_each(sk, &hci_sk_list.head) {
net/bluetooth/hci_sock.c
841
sk_for_each(sk, &hci_sk_list.head) {
net/bluetooth/hci_sync.c
5773
struct list_head *head = &hdev->conn_hash.list;
net/bluetooth/hci_sync.c
5777
while ((conn = list_first_or_null_rcu(head, struct hci_conn, list))) {
net/bluetooth/iso.c
2626
sk_for_each(sk, &iso_sk_list.head) {
net/bluetooth/iso.c
598
sk_for_each(sk, &iso_sk_list.head) {
net/bluetooth/iso.c
617
sk_for_each(sk, &iso_sk_list.head) {
net/bluetooth/iso.c
649
sk_for_each(sk, &iso_sk_list.head) {
net/bluetooth/iso.c
709
sk_for_each(sk, &iso_sk_list.head) {
net/bluetooth/l2cap_core.c
2051
while (chan->retrans_list.head != L2CAP_SEQ_LIST_CLEAR) {
net/bluetooth/l2cap_core.c
2725
initial_head = chan->srej_list.head;
net/bluetooth/l2cap_core.c
2735
} while (chan->srej_list.head != initial_head);
net/bluetooth/l2cap_core.c
298
static struct sk_buff *l2cap_ertm_seq_in_queue(struct sk_buff_head *head,
net/bluetooth/l2cap_core.c
303
skb_queue_walk(head, skb) {
net/bluetooth/l2cap_core.c
337
seq_list->head = L2CAP_SEQ_LIST_CLEAR;
net/bluetooth/l2cap_core.c
359
u16 seq = seq_list->head;
net/bluetooth/l2cap_core.c
362
seq_list->head = seq_list->list[seq & mask];
net/bluetooth/l2cap_core.c
365
if (seq_list->head == L2CAP_SEQ_LIST_TAIL) {
net/bluetooth/l2cap_core.c
366
seq_list->head = L2CAP_SEQ_LIST_CLEAR;
net/bluetooth/l2cap_core.c
377
if (seq_list->head == L2CAP_SEQ_LIST_CLEAR)
net/bluetooth/l2cap_core.c
383
seq_list->head = L2CAP_SEQ_LIST_CLEAR;
net/bluetooth/l2cap_core.c
397
seq_list->head = seq;
net/bluetooth/l2cap_core.c
5998
if (chan->srej_list.head == txseq) {
net/bluetooth/msft.c
308
struct list_head head;
net/bluetooth/msft.c
311
INIT_LIST_HEAD(&head);
net/bluetooth/msft.c
339
list_add_tail(&address_filter->list, &head);
net/bluetooth/msft.c
344
list_for_each_entry_safe(address_filter, n, &head, list) {
net/bluetooth/rfcomm/sock.c
110
sk_for_each(sk, &rfcomm_sk_list.head) {
net/bluetooth/rfcomm/sock.c
133
sk_for_each(sk, &rfcomm_sk_list.head) {
net/bluetooth/rfcomm/sock.c
989
sk_for_each(sk, &rfcomm_sk_list.head) {
net/bluetooth/rfcomm/tty.c
221
struct list_head *head = &rfcomm_dev_list;
net/bluetooth/rfcomm/tty.c
238
head = &entry->list;
net/bluetooth/rfcomm/tty.c
252
head = &entry->list;
net/bluetooth/rfcomm/tty.c
263
list_add(&dev->list, head);
net/bluetooth/sco.c
1435
sk_for_each(sk, &sco_sk_list.head) {
net/bluetooth/sco.c
1529
sk_for_each(sk, &sco_sk_list.head) {
net/bluetooth/sco.c
442
sk_for_each(sk, &sco_sk_list.head) {
net/bluetooth/sco.c
462
sk_for_each(sk, &sco_sk_list.head) {
net/bpf/test_run.c
121
struct xdp_page_head *head =
net/bpf/test_run.c
135
new_ctx = &head->ctx;
net/bpf/test_run.c
136
frm = head->frame;
net/bpf/test_run.c
137
data = head->data;
net/bpf/test_run.c
147
memcpy(&head->orig_ctx, new_ctx, sizeof(head->orig_ctx));
net/bpf/test_run.c
212
static bool frame_was_changed(const struct xdp_page_head *head)
net/bpf/test_run.c
218
return head->frame->data != head->orig_ctx.data ||
net/bpf/test_run.c
219
head->frame->flags != head->orig_ctx.flags;
net/bpf/test_run.c
222
static bool ctx_was_changed(struct xdp_page_head *head)
net/bpf/test_run.c
224
return head->orig_ctx.data != head->ctx.data ||
net/bpf/test_run.c
225
head->orig_ctx.data_meta != head->ctx.data_meta ||
net/bpf/test_run.c
226
head->orig_ctx.data_end != head->ctx.data_end;
net/bpf/test_run.c
229
static void reset_ctx(struct xdp_page_head *head)
net/bpf/test_run.c
231
if (likely(!frame_was_changed(head) && !ctx_was_changed(head)))
net/bpf/test_run.c
234
head->ctx.data = head->orig_ctx.data;
net/bpf/test_run.c
235
head->ctx.data_meta = head->orig_ctx.data_meta;
net/bpf/test_run.c
236
head->ctx.data_end = head->orig_ctx.data_end;
net/bpf/test_run.c
237
xdp_update_frame_from_buff(&head->ctx, head->frame);
net/bpf/test_run.c
238
head->frame->mem_type = head->orig_ctx.rxq->mem.type;
net/bpf/test_run.c
281
struct xdp_page_head *head;
net/bpf/test_run.c
301
head = phys_to_virt(page_to_phys(page));
net/bpf/test_run.c
302
reset_ctx(head);
net/bpf/test_run.c
303
ctx = &head->ctx;
net/bpf/test_run.c
304
frm = head->frame;
net/bpf/test_run.c
310
if (unlikely(ctx_was_changed(head))) {
net/bridge/br_cfm.c
11
hlist_for_each_entry(mep, &br->mep_list, head)
net/bridge/br_cfm.c
23
hlist_for_each_entry_rcu(mep, &br->mep_list, head,
net/bridge/br_cfm.c
36
hlist_for_each_entry_rcu(peer_mep, &mep->peer_mep_list, head,
net/bridge/br_cfm.c
564
hlist_add_tail_rcu(&mep->head, &br->mep_list);
net/bridge/br_cfm.c
578
hlist_for_each_entry_safe(peer_mep, n_store, &mep->peer_mep_list, head) {
net/bridge/br_cfm.c
580
hlist_del_rcu(&peer_mep->head);
net/bridge/br_cfm.c
587
hlist_del_rcu(&mep->head);
net/bridge/br_cfm.c
658
hlist_for_each_entry(peer_mep, &mep->peer_mep_list, head)
net/bridge/br_cfm.c
663
hlist_for_each_entry(peer_mep, &mep->peer_mep_list, head)
net/bridge/br_cfm.c
707
hlist_add_tail_rcu(&peer_mep->head, &mep->peer_mep_list);
net/bridge/br_cfm.c
737
hlist_del_rcu(&peer_mep->head);
net/bridge/br_cfm.c
827
hlist_for_each_entry_rcu(mep, &br->mep_list, head)
net/bridge/br_cfm.c
842
hlist_for_each_entry_rcu(mep, &br->mep_list, head)
net/bridge/br_cfm.c
843
hlist_for_each_entry_rcu(peer_mep, &mep->peer_mep_list, head)
net/bridge/br_cfm.c
864
hlist_for_each_entry_safe(mep, n_store, &br->mep_list, head)
net/bridge/br_cfm_netlink.c
461
hlist_for_each_entry_rcu(mep, &br->mep_list, head) {
net/bridge/br_cfm_netlink.c
587
hlist_for_each_entry_rcu(peer_mep, &mep->peer_mep_list, head) {
net/bridge/br_cfm_netlink.c
624
hlist_for_each_entry_rcu(mep, &br->mep_list, head) {
net/bridge/br_cfm_netlink.c
658
hlist_for_each_entry_rcu(peer_mep, &mep->peer_mep_list, head) {
net/bridge/br_if.c
292
static void destroy_nbp_rcu(struct rcu_head *head)
net/bridge/br_if.c
295
container_of(head, struct net_bridge_port, rcu);
net/bridge/br_if.c
380
void br_dev_delete(struct net_device *dev, struct list_head *head)
net/bridge/br_if.c
397
unregister_netdevice_queue(br->dev, head);
net/bridge/br_multicast.c
893
static void br_multicast_gc(struct hlist_head *head)
net/bridge/br_multicast.c
898
hlist_for_each_entry_safe(gcent, tmp, head, gc_node) {
net/bridge/br_private_cfm.h
119
struct hlist_node head;
net/bridge/br_private_cfm.h
138
struct hlist_node head;
net/caif/cfrfml.c
217
u8 head[6];
net/caif/cfrfml.c
234
err = cfpkt_peek_head(pkt, head, 6);
net/caif/cfrfml.c
268
if (cfpkt_add_head(frontpkt, head, 6) < 0)
net/can/bcm.c
358
static void bcm_send_to_user(struct bcm_op *op, struct bcm_msg_head *head,
net/can/bcm.c
365
unsigned int datalen = head->nframes * op->cfsiz;
net/can/bcm.c
370
skb = alloc_skb(sizeof(*head) + datalen, gfp_any());
net/can/bcm.c
374
skb_put_data(skb, head, sizeof(*head));
net/can/bcm.c
384
if (head->nframes) {
net/can/bcm.c
396
if (head->nframes == 1) {
net/can/bcm.c
490
struct bcm_msg_head head;
net/can/bcm.c
502
memset(&head, 0, sizeof(head));
net/can/bcm.c
503
head.opcode = RX_CHANGED;
net/can/bcm.c
504
head.flags = op->flags;
net/can/bcm.c
505
head.count = op->count;
net/can/bcm.c
506
head.ival1 = op->ival1;
net/can/bcm.c
507
head.ival2 = op->ival2;
net/can/bcm.c
508
head.can_id = op->can_id;
net/can/bcm.c
509
head.nframes = 1;
net/can/bcm.c
511
bcm_send_to_user(op, &head, data, 1);
net/ceph/auth_x.c
540
struct ceph_x_request_header *head = buf;
net/ceph/auth_x.c
554
struct ceph_x_authenticate *auth = (void *)(head + 1);
net/ceph/auth_x.c
564
head->op = cpu_to_le16(CEPHX_GET_AUTH_SESSION_KEY);
net/ceph/messenger.c
546
static void ceph_msg_remove_list(struct list_head *head)
net/ceph/messenger.c
548
while (!list_empty(head)) {
net/ceph/messenger.c
549
struct ceph_msg *msg = list_first_entry(head, struct ceph_msg,
net/ceph/messenger.c
912
BUG_ON(list_empty(&pagelist->head));
net/ceph/messenger.c
913
page = list_first_entry(&pagelist->head, struct page, lru);
net/ceph/messenger.c
968
BUG_ON(list_is_last(&cursor->page->lru, &pagelist->head));
net/ceph/osd_client.c
2243
} __packed head;
net/ceph/osd_client.c
2255
memcpy(&head, p, sizeof(head));
net/ceph/osd_client.c
2256
p += sizeof(head);
net/ceph/osd_client.c
2275
ceph_encode_copy(&p, &head.client_inc, sizeof(head.client_inc));
net/ceph/osd_client.c
2276
ceph_encode_copy(&p, &head.epoch, sizeof(head.epoch));
net/ceph/osd_client.c
2277
ceph_encode_copy(&p, &head.flags, sizeof(head.flags));
net/ceph/osd_client.c
2278
ceph_encode_copy(&p, &head.mtime, sizeof(head.mtime));
net/ceph/osd_client.c
2288
pgid.seed = le32_to_cpu(head.hash);
net/ceph/pagelist.c
17
INIT_LIST_HEAD(&pl->head);
net/ceph/pagelist.c
32
struct page *page = list_entry(pl->head.prev, struct page, lru);
net/ceph/pagelist.c
43
while (!list_empty(&pl->head)) {
net/ceph/pagelist.c
44
struct page *page = list_first_entry(&pl->head, struct page,
net/ceph/pagelist.c
69
list_add_tail(&page->lru, &pl->head);
net/core/datagram.c
644
struct page *head, *last_head = NULL;
net/core/datagram.c
664
head = compound_head(pages[n]);
net/core/datagram.c
665
order = compound_order(head);
net/core/datagram.c
670
if (pages[n] - head > (1UL << order) - 1) {
net/core/datagram.c
671
head = compound_head(pages[n]);
net/core/datagram.c
672
order = compound_order(head);
net/core/datagram.c
675
start += (pages[n] - head) << PAGE_SHIFT;
net/core/datagram.c
681
if (head == skb_frag_page(last) &&
net/core/datagram.c
689
last_head = head;
net/core/datagram.c
698
skb_fill_page_desc_noacc(skb, frag++, head, start, size);
net/core/dev.c
12298
void unregister_netdevice_queue(struct net_device *dev, struct list_head *head)
net/core/dev.c
12302
if (head) {
net/core/dev.c
12303
list_move_tail(&dev->unreg_list, head);
net/core/dev.c
12355
void unregister_netdevice_many_notify(struct list_head *head,
net/core/dev.c
12365
if (list_empty(head))
net/core/dev.c
12368
list_for_each_entry_safe(dev, tmp, head, unreg_list) {
net/core/dev.c
12386
list_for_each_entry(dev, head, unreg_list) {
net/core/dev.c
12397
list_for_each_entry(dev, head, unreg_list) {
net/core/dev.c
12403
list_for_each_entry(dev, head, unreg_list) {
net/core/dev.c
12414
list_for_each_entry(dev, head, unreg_list) {
net/core/dev.c
12475
list_for_each_entry(dev, head, unreg_list) {
net/core/dev.c
12482
list_del(head);
net/core/dev.c
12492
void unregister_netdevice_many(struct list_head *head)
net/core/dev.c
12494
unregister_netdevice_many_notify(head, 0, NULL);
net/core/dev.c
1735
static void __dev_close_many(struct list_head *head)
net/core/dev.c
1742
list_for_each_entry(dev, head, close_list) {
net/core/dev.c
1759
dev_deactivate_many(head);
net/core/dev.c
1761
list_for_each_entry(dev, head, close_list) {
net/core/dev.c
1791
void netif_close_many(struct list_head *head, bool unlink)
net/core/dev.c
1796
list_for_each_entry_safe(dev, tmp, head, close_list)
net/core/dev.c
1800
__dev_close_many(head);
net/core/dev.c
1802
list_for_each_entry_safe(dev, tmp, head, close_list) {
net/core/dev.c
317
struct hlist_head *head = dev_name_hash(net, name);
net/core/dev.c
320
hlist_for_each_entry(name_node, head, hlist)
net/core/dev.c
329
struct hlist_head *head = dev_name_hash(net, name);
net/core/dev.c
332
hlist_for_each_entry_rcu(name_node, head, hlist)
net/core/dev.c
362
static void netdev_name_node_alt_free(struct rcu_head *head)
net/core/dev.c
365
container_of(head, struct netdev_name_node, rcu);
net/core/dev.c
4080
struct sk_buff *next, *head = NULL, *tail;
net/core/dev.c
4093
if (!head)
net/core/dev.c
4094
head = skb;
net/core/dev.c
4102
return head;
net/core/dev.c
5025
u32 head;
net/core/dev.c
5076
head = READ_ONCE(per_cpu(softnet_data, next_cpu).input_queue_head);
net/core/dev.c
5077
rps_input_queue_tail_save(&rflow->last_qtail, head);
net/core/dev.c
5785
struct Qdisc *head;
net/core/dev.c
5788
head = sd->output_queue;
net/core/dev.c
5795
while (head) {
net/core/dev.c
5798
struct Qdisc *q = head;
net/core/dev.c
5800
head = head->next_sched;
net/core/dev.c
6213
static inline void __netif_receive_skb_list_ptype(struct list_head *head,
net/core/dev.c
6221
if (list_empty(head))
net/core/dev.c
6225
ip_list_rcv, head, pt_prev, orig_dev);
net/core/dev.c
6227
list_for_each_entry_safe(skb, next, head, list) {
net/core/dev.c
6233
static void __netif_receive_skb_list_core(struct list_head *head, bool pfmemalloc)
net/core/dev.c
6251
list_for_each_entry_safe(skb, next, head, list) {
net/core/dev.c
626
struct list_head *head = ptype_head(pt);
net/core/dev.c
628
if (WARN_ON_ONCE(!head))
net/core/dev.c
6299
static void __netif_receive_skb_list(struct list_head *head)
net/core/dev.c
6305
list_for_each_entry_safe(skb, next, head, list) {
net/core/dev.c
6310
list_cut_before(&sublist, head, &skb->list);
net/core/dev.c
632
list_add_rcu(&pt->list, head);
net/core/dev.c
6322
if (!list_empty(head))
net/core/dev.c
6323
__netif_receive_skb_list_core(head, pfmemalloc);
net/core/dev.c
6385
void netif_receive_skb_list_internal(struct list_head *head)
net/core/dev.c
6390
list_for_each_entry_safe(skb, next, head, list) {
net/core/dev.c
6397
list_splice_init(&sublist, head);
net/core/dev.c
6402
list_for_each_entry_safe(skb, next, head, list) {
net/core/dev.c
6414
__netif_receive_skb_list(head);
net/core/dev.c
6456
void netif_receive_skb_list(struct list_head *head)
net/core/dev.c
6460
if (list_empty(head))
net/core/dev.c
6463
list_for_each_entry(skb, head, list)
net/core/dev.c
6466
netif_receive_skb_list_internal(head);
net/core/dev.c
652
struct list_head *head = ptype_head(pt);
net/core/dev.c
655
if (!head)
net/core/dev.c
660
list_for_each_entry(pt1, head, list) {
net/core/dev.c
961
struct hlist_head *head = dev_index_hash(net, ifindex);
net/core/dev.c
963
hlist_for_each_entry(dev, head, index_hlist)
net/core/dev.c
985
struct hlist_head *head = dev_index_hash(net, ifindex);
net/core/dev.c
987
hlist_for_each_entry_rcu(dev, head, index_hlist)
net/core/dev.h
153
void unregister_netdevice_many_notify(struct list_head *head,
net/core/dst.c
129
static void dst_destroy_rcu(struct rcu_head *head)
net/core/dst.c
131
struct dst_entry *dst = container_of(head, struct dst_entry, rcu_head);
net/core/filter.c
9674
*insn++ = BPF_LDX_MEM(BPF_FIELD_SIZEOF(struct sk_buff, head),
net/core/filter.c
9676
offsetof(struct sk_buff, head));
net/core/gro.c
161
struct page *page = virt_to_head_page(skb->head);
net/core/gro.c
254
struct list_head *head = &net_hotdata.offload_base;
net/core/gro.c
269
list_for_each_entry_rcu(ptype, head, list) {
net/core/gro.c
281
WARN_ON(&ptype->list == head);
net/core/gro.c
292
struct list_head *head = &gro->hash[index].list;
net/core/gro.c
295
list_for_each_entry_safe_reverse(skb, p, head, list) {
net/core/gro.c
343
static void gro_list_prepare(const struct list_head *head,
net/core/gro.c
350
list_for_each_entry(p, head, list) {
net/core/gro.c
443
static void gro_flush_oldest(struct gro_node *gro, struct list_head *head)
net/core/gro.c
447
oldest = list_last_entry(head, struct sk_buff, list);
net/core/gro.c
466
struct list_head *head = &net_hotdata.offload_base;
net/core/gro.c
480
list_for_each_entry_rcu(ptype, head, list) {
net/core/gro.c
54
struct list_head *head = &net_hotdata.offload_base;
net/core/gro.c
59
list_for_each_entry(po1, head, list) {
net/core/gro_cells.c
106
static void percpu_free_defer_callback(struct rcu_head *head)
net/core/gro_cells.c
110
defer = container_of(head, struct percpu_free_defer, rcu);
net/core/neighbour.c
1800
static void neigh_rcu_free_parms(struct rcu_head *head)
net/core/neighbour.c
1803
container_of(head, struct neigh_parms, rcu_head);
net/core/neighbour.c
581
static void neigh_hash_free_rcu(struct rcu_head *head)
net/core/neighbour.c
583
struct neigh_hash_table *nht = container_of(head,
net/core/neighbour.c
875
LIST_HEAD(head);
net/core/neighbour.c
887
list_add(&n->free_node, &head);
net/core/neighbour.c
897
while (!list_empty(&head)) {
net/core/neighbour.c
898
n = list_first_entry(&head, typeof(*n), free_node);
net/core/pktgen.c
3169
skb->csum_start = skb_transport_header(skb) - skb->head;
net/core/rtnetlink.c
106
struct sk_buff *head = defer_kfree_skb_list;
net/core/rtnetlink.c
145
while (head) {
net/core/rtnetlink.c
146
struct sk_buff *next = head->next;
net/core/rtnetlink.c
148
kfree_skb(head);
net/core/rtnetlink.c
150
head = next;
net/core/rtnetlink.c
95
void rtnl_kfree_skbs(struct sk_buff *head, struct sk_buff *tail)
net/core/rtnetlink.c
97
if (head && tail) {
net/core/rtnetlink.c
99
defer_kfree_skb_list = head;
net/core/selftests.c
131
skb->csum_start = skb_transport_header(skb) - skb->head;
net/core/skbuff.c
1084
static void skb_kfree_head(void *head, unsigned int end_offset)
net/core/skbuff.c
1086
kfree(head);
net/core/skbuff.c
1091
unsigned char *head = skb->head;
net/core/skbuff.c
1094
if (skb_pp_recycle(skb, head))
net/core/skbuff.c
1096
skb_free_frag(head);
net/core/skbuff.c
1098
skb_kfree_head(head, skb_end_offset(skb));
net/core/skbuff.c
1199
if (likely(skb->head))
net/core/skbuff.c
1367
16, 1, skb->head, headroom, false);
net/core/skbuff.c
1628
C(head);
net/core/skbuff.c
2001
struct page *page, *head = NULL;
net/core/skbuff.c
2026
while (head) {
net/core/skbuff.c
2027
struct page *next = (struct page *)page_private(head);
net/core/skbuff.c
2028
put_page(head);
net/core/skbuff.c
2029
head = next;
net/core/skbuff.c
2033
set_page_private(page, (unsigned long)head);
net/core/skbuff.c
2034
head = page;
net/core/skbuff.c
2037
page = head;
net/core/skbuff.c
2071
__skb_fill_netmem_desc(skb, i, page_to_netmem(head), 0, psize);
net/core/skbuff.c
2072
head = (struct page *)page_private(head);
net/core/skbuff.c
2074
__skb_fill_netmem_desc(skb, new_frags - 1, page_to_netmem(head), 0,
net/core/skbuff.c
211
msg, addr, skb->len, sz, skb->head, skb->data,
net/core/skbuff.c
2202
BUG_ON(skb_copy_bits(skb, -headerlen, n->head, headerlen + skb->len));
net/core/skbuff.c
2320
memcpy(data + nhead, skb->head, skb_tail_pointer(skb) - skb->head);
net/core/skbuff.c
2346
off = (data + nhead) - skb->head;
net/core/skbuff.c
2348
skb->head = data;
net/core/skbuff.c
2438
memmove(skb->head + saved_end_offset,
net/core/skbuff.c
2553
BUG_ON(skb_copy_bits(skb, -head_copy_len, n->head + head_copy_off,
net/core/skbuff.c
2670
if (unlikely(skb->data < skb->head))
net/core/skbuff.c
3307
struct sk_buff *head = skb;
net/core/skbuff.c
3388
if (skb == head) {
net/core/skbuff.c
3889
page = virt_to_head_page(from->head);
net/core/skbuff.c
398
skb->head = data;
net/core/skbuff.c
4656
page = virt_to_head_page(frag_skb->head);
net/core/skbuff.c
6209
page = virt_to_head_page(from->head);
net/core/skbuff.c
6842
skb->head = data;
net/core/skbuff.c
6975
skb->head = data;
net/core/sock.c
2343
static void __sk_destruct(struct rcu_head *head)
net/core/sock.c
2345
struct sock *sk = container_of(head, struct sock, sk_rcu);
net/core/sock.c
2510
newsk->sk_backlog.head = newsk->sk_backlog.tail = NULL;
net/core/sock.c
3203
while ((skb = sk->sk_backlog.head) != NULL) {
net/core/sock.c
3204
sk->sk_backlog.head = sk->sk_backlog.tail = NULL;
net/core/sock_map.c
1021
elem = sock_hash_lookup_elem_raw(&bucket->head, hash, key, key_size);
net/core/sock_map.c
1040
hlist_add_head_rcu(&elem_new->node, &bucket->head);
net/core/sock_map.c
1062
struct hlist_head *head;
net/core/sock_map.c
1068
head = &sock_hash_select_bucket(htab, hash)->head;
net/core/sock_map.c
1069
elem = sock_hash_lookup_elem_raw(head, hash, key, key_size);
net/core/sock_map.c
1084
head = &sock_hash_select_bucket(htab, i)->head;
net/core/sock_map.c
1085
elem_next = hlist_entry_safe(rcu_dereference(hlist_first_rcu(head)),
net/core/sock_map.c
1134
INIT_HLIST_HEAD(&htab->buckets[i].head);
net/core/sock_map.c
1168
hlist_for_each_entry(elem, &bucket->head, node)
net/core/sock_map.c
1170
hlist_move_list(&bucket->head, &unlink_list);
net/core/sock_map.c
1337
node = rcu_dereference(hlist_first_rcu(&bucket->head));
net/core/sock_map.c
854
struct hlist_head head;
net/core/sock_map.c
879
sock_hash_lookup_elem_raw(struct hlist_head *head, u32 hash, void *key,
net/core/sock_map.c
884
hlist_for_each_entry_rcu(elem, head, node) {
net/core/sock_map.c
904
elem = sock_hash_lookup_elem_raw(&bucket->head, hash, key, key_size);
net/core/sock_map.c
931
elem_probe = sock_hash_lookup_elem_raw(&bucket->head, elem->hash,
net/core/sock_map.c
953
elem = sock_hash_lookup_elem_raw(&bucket->head, hash, key, key_size);
net/core/sock_reuseport.c
302
static void reuseport_free_rcu(struct rcu_head *head)
net/core/sock_reuseport.c
306
reuse = container_of(head, struct sock_reuseport, rcu);
net/ethernet/eth.c
390
struct sk_buff *eth_gro_receive(struct list_head *head, struct sk_buff *skb)
net/ethernet/eth.c
408
list_for_each_entry(p, head, list) {
net/ethernet/eth.c
432
head, skb);
net/ethtool/cmis_fw_update.c
106
__struct_group(cmis_cdb_start_fw_download_pl_h, head, /* no attrs */,
net/handshake/tlshd.c
64
struct nlattr *head = nlmsg_attrdata(info->nlhdr, GENL_HDRLEN);
net/handshake/tlshd.c
70
nla_for_each_attr(nla, head, len, rem) {
net/handshake/tlshd.c
80
nla_for_each_attr(nla, head, len, rem) {
net/hsr/hsr_netlink.c
132
static void hsr_dellink(struct net_device *dev, struct list_head *head)
net/hsr/hsr_netlink.c
148
unregister_netdevice_queue(dev, head);
net/ieee802154/6lowpan/core.c
188
static void lowpan_dellink(struct net_device *ldev, struct list_head *head)
net/ipv4/af_inet.c
1419
skb->network_header = skb_mac_header(skb) + nhoff - skb->head;
net/ipv4/af_inet.c
1445
skb->head - (unsigned char *)iph;
net/ipv4/af_inet.c
1457
skb->network_header = (u8 *)iph - skb->head;
net/ipv4/af_inet.c
1474
struct sk_buff *inet_gro_receive(struct list_head *head, struct sk_buff *skb)
net/ipv4/af_inet.c
1509
list_for_each_entry(p, head, list) {
net/ipv4/af_inet.c
1539
ops->callbacks.gro_receive, head, skb);
net/ipv4/af_inet.c
1547
static struct sk_buff *ipip_gro_receive(struct list_head *head,
net/ipv4/af_inet.c
1557
return inet_gro_receive(head, skb);
net/ipv4/devinet.c
224
static void inet_rcu_free_ifa(struct rcu_head *head)
net/ipv4/devinet.c
226
struct in_ifaddr *ifa = container_of(head, struct in_ifaddr, rcu_head);
net/ipv4/devinet.c
241
static void in_dev_free_rcu(struct rcu_head *head)
net/ipv4/devinet.c
243
struct in_device *idev = container_of(head, struct in_device, rcu_head);
net/ipv4/devinet.c
722
struct hlist_head *head = &net->ipv4.inet_addr_lst[i];
net/ipv4/devinet.c
726
hlist_for_each_entry_rcu(ifa, head, addr_lst) {
net/ipv4/devinet.c
765
hlist_for_each_entry_safe(ifa, n, head, addr_lst) {
net/ipv4/esp4_offload.c
30
static struct sk_buff *esp4_gro_receive(struct list_head *head,
net/ipv4/fib_frontend.c
1030
struct hlist_head *head;
net/ipv4/fib_frontend.c
1068
head = &net->ipv4.fib_table_hash[h];
net/ipv4/fib_frontend.c
1069
hlist_for_each_entry_rcu(tb, head, tb_hlist) {
net/ipv4/fib_frontend.c
116
struct hlist_head *head;
net/ipv4/fib_frontend.c
123
head = &net->ipv4.fib_table_hash[h];
net/ipv4/fib_frontend.c
124
hlist_for_each_entry_rcu(tb, head, tb_hlist,
net/ipv4/fib_frontend.c
1611
struct hlist_head *head = &net->ipv4.fib_table_hash[i];
net/ipv4/fib_frontend.c
1615
hlist_for_each_entry_safe(tb, tmp, head, tb_hlist) {
net/ipv4/fib_frontend.c
191
struct hlist_head *head = &net->ipv4.fib_table_hash[h];
net/ipv4/fib_frontend.c
195
hlist_for_each_entry_safe(tb, tmp, head, tb_hlist)
net/ipv4/fib_semantics.c
1547
struct hlist_head *head;
net/ipv4/fib_semantics.c
1549
head = fib_info_laddrhash_bucket(net, fi->fib_prefsrc);
net/ipv4/fib_semantics.c
1550
hlist_add_head(&fi->fib_lhash, head);
net/ipv4/fib_semantics.c
1556
struct hlist_head *head;
net/ipv4/fib_semantics.c
1560
head = fib_nh_head(nexthop_nh->fib_nh_dev);
net/ipv4/fib_semantics.c
1561
hlist_add_head_rcu(&nexthop_nh->nh_hash, head);
net/ipv4/fib_semantics.c
1822
struct hlist_head *head;
net/ipv4/fib_semantics.c
1829
head = fib_info_laddrhash_bucket(net, local);
net/ipv4/fib_semantics.c
1830
hlist_for_each_entry(fi, head, fib_lhash) {
net/ipv4/fib_semantics.c
1912
struct hlist_head *head = fib_nh_head(dev);
net/ipv4/fib_semantics.c
1915
hlist_for_each_entry(nh, head, nh_hash) {
net/ipv4/fib_semantics.c
1931
struct hlist_head *head = fib_nh_head(dev);
net/ipv4/fib_semantics.c
1940
hlist_for_each_entry(nh, head, nh_hash) {
net/ipv4/fib_semantics.c
2081
struct hlist_head *head;
net/ipv4/fib_semantics.c
2096
head = fib_nh_head(dev);
net/ipv4/fib_semantics.c
2099
hlist_for_each_entry(nh, head, nh_hash) {
net/ipv4/fib_semantics.c
222
static void free_fib_info_rcu(struct rcu_head *head)
net/ipv4/fib_semantics.c
224
struct fib_info *fi = container_of(head, struct fib_info, rcu);
net/ipv4/fib_semantics.c
371
static void fib_info_hash_free(struct hlist_head *head)
net/ipv4/fib_semantics.c
373
kvfree(head);
net/ipv4/fib_semantics.c
394
struct hlist_head *head = &old_info_hash[i];
net/ipv4/fib_semantics.c
398
hlist_for_each_entry_safe(fi, n, head, fib_hash)
net/ipv4/fib_semantics.c
420
struct hlist_head *head;
net/ipv4/fib_semantics.c
429
head = &net->ipv4.fib_info_hash[hash];
net/ipv4/fib_semantics.c
431
hlist_for_each_entry(fi, head, fib_hash) {
net/ipv4/fib_semantics.c
450
struct hlist_head *head = fib_info_hash_bucket(nfi);
net/ipv4/fib_semantics.c
453
hlist_for_each_entry(fi, head, fib_hash) {
net/ipv4/fib_semantics.c
478
struct hlist_head *head;
net/ipv4/fib_semantics.c
481
head = fib_nh_head(dev);
net/ipv4/fib_semantics.c
483
hlist_for_each_entry_rcu(nh, head, nh_hash) {
net/ipv4/fib_trie.c
2143
struct hlist_head *head = &net->ipv4.fib_table_hash[h];
net/ipv4/fib_trie.c
2146
hlist_for_each_entry_rcu(tb, head, tb_hlist,
net/ipv4/fib_trie.c
2213
struct hlist_head *head = &net->ipv4.fib_table_hash[h];
net/ipv4/fib_trie.c
2216
hlist_for_each_entry_rcu(tb, head, tb_hlist) {
net/ipv4/fib_trie.c
2225
static void __trie_free_rcu(struct rcu_head *head)
net/ipv4/fib_trie.c
2227
struct fib_table *tb = container_of(head, struct fib_table, rcu);
net/ipv4/fib_trie.c
2622
struct hlist_head *head = &net->ipv4.fib_table_hash[h];
net/ipv4/fib_trie.c
2625
hlist_for_each_entry_rcu(tb, head, tb_hlist) {
net/ipv4/fib_trie.c
2655
struct hlist_head *head = &net->ipv4.fib_table_hash[h];
net/ipv4/fib_trie.c
2658
hlist_for_each_entry_rcu(tb, head, tb_hlist) {
net/ipv4/fib_trie.c
2707
struct hlist_head *head = &net->ipv4.fib_table_hash[h];
net/ipv4/fib_trie.c
2708
hlist_for_each_entry_rcu(tb, head, tb_hlist) {
net/ipv4/fib_trie.c
303
static void __node_free_rcu(struct rcu_head *head)
net/ipv4/fib_trie.c
305
struct tnode *n = container_of(head, struct tnode, rcu);
net/ipv4/fib_trie.c
486
struct callback_head *head = &tn_info(tn)->rcu;
net/ipv4/fib_trie.c
488
while (head) {
net/ipv4/fib_trie.c
489
head = head->next;
net/ipv4/fib_trie.c
493
tn = container_of(head, struct tnode, rcu)->kv;
net/ipv4/fou_core.c
245
struct list_head *head,
net/ipv4/fou_core.c
270
pp = call_gro_receive(ops->callbacks.gro_receive, head, skb);
net/ipv4/fou_core.c
326
struct list_head *head,
net/ipv4/fou_core.c
416
list_for_each_entry(p, head, list) {
net/ipv4/fou_core.c
459
pp = call_gro_receive(ops->callbacks.gro_receive, head, skb);
net/ipv4/gre_offload.c
114
skb->csum_start = skb_transport_header(skb) - skb->head;
net/ipv4/gre_offload.c
122
static struct sk_buff *gre_gro_receive(struct list_head *head,
net/ipv4/gre_offload.c
192
list_for_each_entry(p, head, list) {
net/ipv4/gre_offload.c
227
pp = call_gro_receive(ptype->callbacks.gro_receive, head, skb);
net/ipv4/icmp.c
1036
if (unlikely(!skb_in || skb_network_header(skb_in) < skb_in->head ||
net/ipv4/icmp.c
839
if ((u8 *)iph < skb_in->head ||
net/ipv4/inet_connection_sock.c
322
struct inet_bind_hashbucket *head, *head2;
net/ipv4/inet_connection_sock.c
362
head = &hinfo->bhash[inet_bhashfn(net, port,
net/ipv4/inet_connection_sock.c
364
spin_lock_bh(&head->lock);
net/ipv4/inet_connection_sock.c
373
inet_bind_bucket_for_each(tb, &head->chain)
net/ipv4/inet_connection_sock.c
384
spin_unlock_bh(&head->lock);
net/ipv4/inet_connection_sock.c
409
return head;
net/ipv4/inet_connection_sock.c
508
struct inet_bind_hashbucket *head, *head2;
net/ipv4/inet_connection_sock.c
517
head = inet_csk_find_open_port(sk, &tb, &tb2, &head2, &port);
net/ipv4/inet_connection_sock.c
518
if (!head)
net/ipv4/inet_connection_sock.c
527
head = &hinfo->bhash[inet_bhashfn(net, port,
net/ipv4/inet_connection_sock.c
529
spin_lock_bh(&head->lock);
net/ipv4/inet_connection_sock.c
530
inet_bind_bucket_for_each(tb, &head->chain)
net/ipv4/inet_connection_sock.c
537
head, port, l3mdev);
net/ipv4/inet_connection_sock.c
593
spin_unlock_bh(&head->lock);
net/ipv4/inet_fragment.c
290
static void inet_frag_destroy_rcu(struct rcu_head *head)
net/ipv4/inet_fragment.c
292
struct inet_frag_queue *q = container_of(head, struct inet_frag_queue,
net/ipv4/inet_fragment.c
501
struct sk_buff *fp, *head = skb_rb_first(&q->rb_fragments);
net/ipv4/inet_fragment.c
518
if (head != skb) {
net/ipv4/inet_fragment.c
521
head = skb;
net/ipv4/inet_fragment.c
538
skb_morph(skb, head);
net/ipv4/inet_fragment.c
539
FRAG_CB(skb)->next_frag = FRAG_CB(head)->next_frag;
net/ipv4/inet_fragment.c
540
rb_replace_node(&head->rbnode, &skb->rbnode,
net/ipv4/inet_fragment.c
542
consume_skb(head);
net/ipv4/inet_fragment.c
543
head = skb;
net/ipv4/inet_fragment.c
545
WARN_ON(FRAG_CB(head)->ip_defrag_offset != 0);
net/ipv4/inet_fragment.c
547
delta = -head->truesize;
net/ipv4/inet_fragment.c
550
if (skb_unclone(head, GFP_ATOMIC))
net/ipv4/inet_fragment.c
553
delta += head->truesize;
net/ipv4/inet_fragment.c
561
if (skb_has_frag_list(head)) {
net/ipv4/inet_fragment.c
568
skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list;
net/ipv4/inet_fragment.c
569
skb_frag_list_init(head);
net/ipv4/inet_fragment.c
570
for (i = 0; i < skb_shinfo(head)->nr_frags; i++)
net/ipv4/inet_fragment.c
571
plen += skb_frag_size(&skb_shinfo(head)->frags[i]);
net/ipv4/inet_fragment.c
572
clone->data_len = head->data_len - plen;
net/ipv4/inet_fragment.c
574
head->truesize += clone->truesize;
net/ipv4/inet_fragment.c
576
clone->ip_summed = head->ip_summed;
net/ipv4/inet_fragment.c
578
skb_shinfo(head)->frag_list = clone;
net/ipv4/inet_fragment.c
581
nextp = &skb_shinfo(head)->frag_list;
net/ipv4/inet_fragment.c
586
int ts_delta = head->truesize - orig_truesize;
net/ipv4/inet_fragment.c
594
head->sk = sk;
net/ipv4/inet_fragment.c
595
head->destructor = destructor;
net/ipv4/inet_fragment.c
603
void inet_frag_reasm_finish(struct inet_frag_queue *q, struct sk_buff *head,
net/ipv4/inet_fragment.c
606
struct sock *sk = is_skb_wmem(head) ? head->sk : NULL;
net/ipv4/inet_fragment.c
607
const unsigned int head_truesize = head->truesize;
net/ipv4/inet_fragment.c
613
skb_push(head, head->data - skb_network_header(head));
net/ipv4/inet_fragment.c
616
fp = FRAG_CB(head)->next_frag;
net/ipv4/inet_fragment.c
617
rbn = rb_next(&head->rbnode);
net/ipv4/inet_fragment.c
618
rb_erase(&head->rbnode, &q->rb_fragments);
net/ipv4/inet_fragment.c
620
sum_truesize = head->truesize;
net/ipv4/inet_fragment.c
632
if (head->ip_summed != fp->ip_summed)
net/ipv4/inet_fragment.c
633
head->ip_summed = CHECKSUM_NONE;
net/ipv4/inet_fragment.c
634
else if (head->ip_summed == CHECKSUM_COMPLETE)
net/ipv4/inet_fragment.c
635
head->csum = csum_add(head->csum, fp->csum);
net/ipv4/inet_fragment.c
637
if (try_coalesce && skb_try_coalesce(head, fp, &stolen,
net/ipv4/inet_fragment.c
645
head->data_len += fp->len;
net/ipv4/inet_fragment.c
646
head->len += fp->len;
net/ipv4/inet_fragment.c
647
head->truesize += fp->truesize;
net/ipv4/inet_fragment.c
667
skb_mark_not_on_list(head);
net/ipv4/inet_fragment.c
668
head->prev = NULL;
net/ipv4/inet_fragment.c
669
head->tstamp = q->stamp;
net/ipv4/inet_fragment.c
670
head->tstamp_type = q->tstamp_type;
net/ipv4/inet_fragment.c
679
struct sk_buff *head, *skb;
net/ipv4/inet_fragment.c
681
head = skb_rb_first(&q->rb_fragments);
net/ipv4/inet_fragment.c
682
if (!head)
net/ipv4/inet_fragment.c
684
skb = FRAG_CB(head)->next_frag;
net/ipv4/inet_fragment.c
686
rb_replace_node(&head->rbnode, &skb->rbnode,
net/ipv4/inet_fragment.c
689
rb_erase(&head->rbnode, &q->rb_fragments);
net/ipv4/inet_fragment.c
690
memset(&head->rbnode, 0, sizeof(head->rbnode));
net/ipv4/inet_fragment.c
693
if (head == q->fragments_tail)
net/ipv4/inet_fragment.c
696
sub_frag_mem_limit(q->fqdir, head->truesize);
net/ipv4/inet_fragment.c
698
return head;
net/ipv4/inet_hashtables.c
1001
spin_unlock_bh(&head->lock);
net/ipv4/inet_hashtables.c
1043
struct inet_bind_hashbucket *head, *head2;
net/ipv4/inet_hashtables.c
1093
head = &hinfo->bhash[inet_bhashfn(net, port,
net/ipv4/inet_hashtables.c
1096
hlist_for_each_entry_rcu(tb, &head->chain, node) {
net/ipv4/inet_hashtables.c
1111
spin_lock_bh(&head->lock);
net/ipv4/inet_hashtables.c
1116
inet_bind_bucket_for_each(tb, &head->chain) {
net/ipv4/inet_hashtables.c
1131
net, head, port, l3mdev);
net/ipv4/inet_hashtables.c
1133
spin_unlock_bh(&head->lock);
net/ipv4/inet_hashtables.c
1141
spin_unlock_bh(&head->lock);
net/ipv4/inet_hashtables.c
1190
spin_unlock(&head->lock);
net/ipv4/inet_hashtables.c
1218
spin_unlock(&head->lock);
net/ipv4/inet_hashtables.c
125
struct inet_bind_hashbucket *head,
net/ipv4/inet_hashtables.c
147
hlist_add_head(&tb2->node, &head->chain);
net/ipv4/inet_hashtables.c
153
struct inet_bind_hashbucket *head,
net/ipv4/inet_hashtables.c
160
inet_bind2_bucket_init(tb2, net, head, tb, sk);
net/ipv4/inet_hashtables.c
215
struct inet_bind_hashbucket *head, *head2;
net/ipv4/inet_hashtables.c
221
head = &hashinfo->bhash[bhash];
net/ipv4/inet_hashtables.c
224
spin_lock(&head->lock);
net/ipv4/inet_hashtables.c
241
spin_unlock(&head->lock);
net/ipv4/inet_hashtables.c
256
struct inet_bind_hashbucket *head, *head2;
net/ipv4/inet_hashtables.c
265
head = &table->bhash[bhash];
net/ipv4/inet_hashtables.c
268
spin_lock(&head->lock);
net/ipv4/inet_hashtables.c
274
spin_unlock(&head->lock);
net/ipv4/inet_hashtables.c
285
inet_bind_bucket_for_each(tb, &head->chain) {
net/ipv4/inet_hashtables.c
291
net, head, port, l3mdev);
net/ipv4/inet_hashtables.c
294
spin_unlock(&head->lock);
net/ipv4/inet_hashtables.c
318
spin_unlock(&head->lock);
net/ipv4/inet_hashtables.c
326
spin_unlock(&head->lock);
net/ipv4/inet_hashtables.c
535
struct inet_ehash_bucket *head;
net/ipv4/inet_hashtables.c
543
head = &hashinfo->ehash[slot];
net/ipv4/inet_hashtables.c
546
sk_nulls_for_each_rcu(sk, node, &head->chain) {
net/ipv4/inet_hashtables.c
590
struct inet_ehash_bucket *head = inet_ehash_bucket(hinfo, hash);
net/ipv4/inet_hashtables.c
597
sk_nulls_for_each(sk2, node, &head->chain) {
net/ipv4/inet_hashtables.c
611
sk_nulls_for_each(sk2, node, &head->chain) {
net/ipv4/inet_hashtables.c
632
__sk_nulls_add_node_rcu(sk, &head->chain);
net/ipv4/inet_hashtables.c
708
struct inet_ehash_bucket *head;
net/ipv4/inet_hashtables.c
716
head = inet_ehash_bucket(hashinfo, sk->sk_hash);
net/ipv4/inet_hashtables.c
717
list = &head->chain;
net/ipv4/inet_hashtables.c
75
struct inet_bind_hashbucket *head,
net/ipv4/inet_hashtables.c
88
hlist_add_head_rcu(&tb->node, &head->chain);
net/ipv4/inet_hashtables.c
886
inet_bind2_bucket_find(const struct inet_bind_hashbucket *head, const struct net *net,
net/ipv4/inet_hashtables.c
891
inet_bind_bucket_for_each(bhash2, &head->chain)
net/ipv4/inet_hashtables.c
930
struct inet_bind_hashbucket *head, *head2;
net/ipv4/inet_hashtables.c
967
head = &hinfo->bhash[bhash];
net/ipv4/inet_hashtables.c
973
spin_lock_bh(&head->lock);
net/ipv4/inet_timewait_sock.c
307
struct inet_ehash_bucket *head = &hashinfo->ehash[0];
net/ipv4/inet_timewait_sock.c
313
for (slot = 0; slot <= ehash_mask; slot++, head++) {
net/ipv4/inet_timewait_sock.c
314
if (hlist_nulls_empty(&head->chain))
net/ipv4/inet_timewait_sock.c
321
sk_nulls_for_each_rcu(sk, node, &head->chain) {
net/ipv4/ip_fragment.c
128
struct sk_buff *head = NULL;
net/ipv4/ip_fragment.c
161
head = inet_frag_pull_head(&qp->q);
net/ipv4/ip_fragment.c
162
if (!head)
net/ipv4/ip_fragment.c
164
head->dev = dev_get_by_index_rcu(net, qp->iif);
net/ipv4/ip_fragment.c
165
if (!head->dev)
net/ipv4/ip_fragment.c
170
iph = ip_hdr(head);
net/ipv4/ip_fragment.c
171
reason = ip_route_input_noref(head, iph->daddr, iph->saddr,
net/ipv4/ip_fragment.c
172
ip4h_dscp(iph), head->dev);
net/ipv4/ip_fragment.c
181
(skb_rtable(head)->rt_type != RTN_LOCAL))
net/ipv4/ip_fragment.c
185
icmp_send(head, ICMP_TIME_EXCEEDED, ICMP_EXC_FRAGTIME, 0);
net/ipv4/ip_fragment.c
192
kfree_skb_reason(head, reason);
net/ipv4/ip_input.c
578
static void ip_sublist_rcv_finish(struct list_head *head)
net/ipv4/ip_input.c
582
list_for_each_entry_safe(skb, next, head, list) {
net/ipv4/ip_input.c
602
static void ip_list_rcv_finish(struct net *net, struct list_head *head)
net/ipv4/ip_input.c
608
list_for_each_entry_safe(skb, next, head, list) {
net/ipv4/ip_input.c
639
static void ip_sublist_rcv(struct list_head *head, struct net_device *dev,
net/ipv4/ip_input.c
643
head, dev, NULL, ip_rcv_finish);
net/ipv4/ip_input.c
644
ip_list_rcv_finish(net, head);
net/ipv4/ip_input.c
648
void ip_list_rcv(struct list_head *head, struct packet_type *pt,
net/ipv4/ip_input.c
656
list_for_each_entry_safe(skb, next, head, list) {
net/ipv4/ip_sockglue.c
335
static void ip_ra_destroy_rcu(struct rcu_head *head)
net/ipv4/ip_sockglue.c
337
struct ip_ra_chain *ra = container_of(head, struct ip_ra_chain, rcu);
net/ipv4/ip_tunnel.c
108
hlist_for_each_entry_rcu(t, head, hash_node) {
net/ipv4/ip_tunnel.c
1092
void ip_tunnel_dellink(struct net_device *dev, struct list_head *head)
net/ipv4/ip_tunnel.c
1101
unregister_netdevice_queue(dev, head);
net/ipv4/ip_tunnel.c
1165
struct list_head *head)
net/ipv4/ip_tunnel.c
1175
unregister_netdevice_queue(dev, head);
net/ipv4/ip_tunnel.c
1187
unregister_netdevice_queue(t->dev, head);
net/ipv4/ip_tunnel.c
124
head = &itn->tunnels[hash];
net/ipv4/ip_tunnel.c
126
hlist_for_each_entry_rcu(t, head, hash_node) {
net/ipv4/ip_tunnel.c
143
hlist_for_each_entry_rcu(t, head, hash_node) {
net/ipv4/ip_tunnel.c
194
struct hlist_head *head = ip_bucket(itn, &t->parms);
net/ipv4/ip_tunnel.c
198
hlist_add_head_rcu(&t->hash_node, head);
net/ipv4/ip_tunnel.c
218
struct hlist_head *head = ip_bucket(itn, parms);
net/ipv4/ip_tunnel.c
222
hlist_for_each_entry_rcu(t, head, hash_node, lockdep_rtnl_is_held()) {
net/ipv4/ip_tunnel.c
412
nh = skb_network_header(skb) - skb->head;
net/ipv4/ip_tunnel.c
421
iph = (struct iphdr *)(skb->head + nh);
net/ipv4/ip_tunnel.c
87
struct hlist_head *head;
net/ipv4/ip_tunnel.c
92
head = &itn->tunnels[hash];
net/ipv4/ip_tunnel.c
94
hlist_for_each_entry_rcu(t, head, hash_node) {
net/ipv4/ip_tunnel_core.c
1107
if (skb_network_header(skb) >= skb->head &&
net/ipv4/ip_tunnel_core.c
1111
if (skb_network_header(skb) >= skb->head &&
net/ipv4/ipmr.c
668
struct list_head *head)
net/ipv4/ipmr.c
717
unregister_netdevice_queue(dev, head);
net/ipv4/ipmr.c
723
static void ipmr_cache_free_rcu(struct rcu_head *head)
net/ipv4/ipmr.c
725
struct mr_mfc *c = container_of(head, struct mr_mfc, rcu);
net/ipv4/netfilter/nf_reject_ipv4.c
251
nskb->csum_start = (unsigned char *)tcph - nskb->head;
net/ipv4/nexthop.c
2698
struct hlist_head *head = &net->nexthop.devhash[hash];
net/ipv4/nexthop.c
2702
hlist_for_each_entry_safe(nhi, n, head, dev_hash) {
net/ipv4/nexthop.c
3924
struct hlist_head *head = &net->nexthop.devhash[hash];
net/ipv4/nexthop.c
3928
hlist_for_each_entry_safe(nhi, n, head, dev_hash) {
net/ipv4/nexthop.c
470
struct hlist_head *head;
net/ipv4/nexthop.c
476
head = &net->nexthop.devhash[hash];
net/ipv4/nexthop.c
477
hlist_add_head(&nhi->dev_hash, head);
net/ipv4/nexthop.c
519
void nexthop_free_rcu(struct rcu_head *head)
net/ipv4/nexthop.c
521
struct nexthop *nh = container_of(head, struct nexthop, rcu);
net/ipv4/nexthop.c
93
return !net->nexthop.notifier_chain.head;
net/ipv4/route.c
1522
struct list_head head;
net/ipv4/route.c
1534
list_add_tail(&rt->dst.rt_uncached, &ul->head);
net/ipv4/route.c
1563
if (list_empty(&ul->head))
net/ipv4/route.c
1567
list_for_each_entry_safe(rt, safe, &ul->head, dst.rt_uncached) {
net/ipv4/route.c
3747
INIT_LIST_HEAD(&ul->head);
net/ipv4/tcp_ao.c
1121
hlist_for_each_entry_safe(key, next, &ao_info->head, node) {
net/ipv4/tcp_ao.c
1167
hlist_for_each_entry_rcu(key, &ao->head, node, lockdep_sock_is_held(sk))
net/ipv4/tcp_ao.c
1186
hlist_for_each_entry_rcu(key, &ao->head, node, lockdep_sock_is_held(sk))
net/ipv4/tcp_ao.c
119
hlist_for_each_entry_rcu(key, &ao->head, node, lockdep_sock_is_held(sk)) {
net/ipv4/tcp_ao.c
1230
hlist_for_each_entry_rcu(key, &ao->head, node) {
net/ipv4/tcp_ao.c
1259
key_head = rcu_dereference(hlist_first_rcu(&new_ao->head));
net/ipv4/tcp_ao.c
1281
hlist_for_each_entry_safe(key, key_head, &new_ao->head, node) {
net/ipv4/tcp_ao.c
1809
hlist_add_head_rcu(&key->node, &ao_info->head);
net/ipv4/tcp_ao.c
1908
hlist_for_each_entry_rcu(key, &ao_info->head, node,
net/ipv4/tcp_ao.c
1953
if (rcu_dereference_check(hlist_first_rcu(&md5sig->head),
net/ipv4/tcp_ao.c
210
hlist_for_each_entry_rcu(key, &ao->head, node, lockdep_sock_is_held(sk)) {
net/ipv4/tcp_ao.c
2237
hlist_for_each_entry_rcu(key, &ao_info->head, node,
net/ipv4/tcp_ao.c
234
INIT_HLIST_HEAD(&ao->head);
net/ipv4/tcp_ao.c
2405
hlist_for_each_entry_rcu(key, &ao->head, node, lockdep_sock_is_held(sk))
net/ipv4/tcp_ao.c
242
hlist_add_head_rcu(&mkt->node, &ao->head);
net/ipv4/tcp_ao.c
264
static void tcp_ao_key_free_rcu(struct rcu_head *head)
net/ipv4/tcp_ao.c
266
struct tcp_ao_key *key = container_of(head, struct tcp_ao_key, rcu);
net/ipv4/tcp_ao.c
277
hlist_for_each_entry_safe(key, n, &ao->head, node) {
net/ipv4/tcp_ao.c
291
hlist_for_each_entry(key, &ao->head, node)
net/ipv4/tcp_ao.c
325
hlist_for_each_entry_safe(key, n, &ao_info->head, node) {
net/ipv4/tcp_diag.c
160
hlist_for_each_entry_rcu(key, &md5sig->head, node)
net/ipv4/tcp_diag.c
481
struct inet_ehash_bucket *head = &hashinfo->ehash[i];
net/ipv4/tcp_diag.c
488
if (hlist_nulls_empty(&head->chain))
net/ipv4/tcp_diag.c
498
sk_nulls_for_each(sk, node, &head->chain) {
net/ipv4/tcp_diag.c
66
hlist_for_each_entry_rcu(key, &md5sig->head, node)
net/ipv4/tcp_diag.c
78
hlist_for_each_entry_rcu(key, &md5sig->head, node) {
net/ipv4/tcp_fastopen.c
117
static void tcp_fastopen_ctx_free(struct rcu_head *head)
net/ipv4/tcp_fastopen.c
120
container_of(head, struct tcp_fastopen_context, rcu);
net/ipv4/tcp_input.c
2526
struct sk_buff *skb, *head;
net/ipv4/tcp_input.c
2529
head = tcp_rtx_queue_head(sk);
net/ipv4/tcp_input.c
2530
is_reneg = head && (TCP_SKB_CB(head)->sacked & TCPCB_SACKED_ACKED);
net/ipv4/tcp_input.c
2540
skb = head;
net/ipv4/tcp_input.c
2544
else if (skb != head && tcp_rack_skb_timeout(tp, skb, 0) > 0)
net/ipv4/tcp_input.c
3810
struct sk_buff *head = tcp_send_head(sk);
net/ipv4/tcp_input.c
3814
if (!head)
net/ipv4/tcp_input.c
3816
if (!after(TCP_SKB_CB(head)->end_seq, tcp_wnd_end(tp))) {
net/ipv4/tcp_input.c
5793
struct sk_buff *head, struct sk_buff *tail, u32 start, u32 end)
net/ipv4/tcp_input.c
5795
struct sk_buff *skb = head, *n;
net/ipv4/tcp_input.c
5902
struct sk_buff *skb, *head;
net/ipv4/tcp_input.c
5915
for (head = skb;;) {
net/ipv4/tcp_input.c
5925
if (range_truesize != head->truesize ||
net/ipv4/tcp_input.c
5928
head, skb, start, end);
net/ipv4/tcp_ipv4.c
1277
hlist_for_each_entry_rcu(key, &md5sig->head, node,
net/ipv4/tcp_ipv4.c
1323
hlist_for_each_entry_rcu(key, &md5sig->head, node,
net/ipv4/tcp_ipv4.c
1361
INIT_HLIST_HEAD(&md5sig->head);
net/ipv4/tcp_ipv4.c
1412
hlist_add_head_rcu(&key->node, &md5sig->head);
net/ipv4/tcp_ipv4.c
1498
hlist_for_each_entry_safe(key, n, &md5sig->head, node) {
net/ipv4/tcp_ipv4.c
668
skb->csum_start = skb_transport_header(skb) - skb->head;
net/ipv4/tcp_offload.c
265
struct sk_buff *tcp_gro_lookup(struct list_head *head, struct tcphdr *th)
net/ipv4/tcp_offload.c
270
list_for_each_entry(p, head, list) {
net/ipv4/tcp_offload.c
286
struct sk_buff *tcp_gro_receive(struct list_head *head, struct sk_buff *skb,
net/ipv4/tcp_offload.c
302
p = tcp_gro_lookup(head, th);
net/ipv4/tcp_offload.c
377
skb->csum_start = (unsigned char *)th - skb->head;
net/ipv4/tcp_offload.c
389
static void tcp4_check_fraglist_gro(struct list_head *head, struct sk_buff *skb,
net/ipv4/tcp_offload.c
401
p = tcp_gro_lookup(head, th);
net/ipv4/tcp_offload.c
419
struct sk_buff *tcp4_gro_receive(struct list_head *head, struct sk_buff *skb)
net/ipv4/tcp_offload.c
433
tcp4_check_fraglist_gro(head, skb, th);
net/ipv4/tcp_offload.c
435
return tcp_gro_receive(head, skb, th);
net/ipv4/tcp_output.c
1254
struct list_head head; /* queue of tcp sockets */
net/ipv4/tcp_output.c
1301
list_splice_init(&tsq->head, &list);
net/ipv4/tcp_output.c
1370
INIT_LIST_HEAD(&tsq->head);
net/ipv4/tcp_output.c
1414
empty = list_empty(&tsq->head);
net/ipv4/tcp_output.c
1415
list_add(&tp->tsq_node, &tsq->head);
net/ipv4/tcp_output.c
2440
struct sk_buff *head;
net/ipv4/tcp_output.c
2497
head = tcp_rtx_queue_head(sk);
net/ipv4/tcp_output.c
2498
if (!head)
net/ipv4/tcp_output.c
2503
expected_ack = head->tstamp + srtt_in_ns;
net/ipv4/tcp_sigpool.c
42
static void free_old_scratches(struct rcu_head *head)
net/ipv4/tcp_sigpool.c
46
stf = container_of(head, struct scratches_to_free, rcu);
net/ipv4/tunnel4.c
116
struct xfrm_tunnel __rcu *head;
net/ipv4/tunnel4.c
120
head = (proto == IPPROTO_IPIP) ? tunnel4_handlers : tunnel64_handlers;
net/ipv4/tunnel4.c
122
for_each_tunnel_rcu(head, handler) {
net/ipv4/tunnel4.c
90
#define for_each_tunnel_rcu(head, handler) \
net/ipv4/tunnel4.c
91
for (handler = rcu_dereference(head); \
net/ipv4/udp.c
1064
skb->csum_start = skb_transport_header(skb) - skb->head;
net/ipv4/udp.c
1110
skb->csum_start = skb_transport_header(skb) - skb->head;
net/ipv4/udp.c
149
sk_for_each(sk2, &hslot->head) {
net/ipv4/udp.c
186
udp_portaddr_for_each_entry(sk2, &hslot2->head) {
net/ipv4/udp.c
214
sk_for_each(sk2, &hslot->head) {
net/ipv4/udp.c
2306
&nhslot2->head);
net/ipv4/udp.c
2580
sk_for_each_entry_offset_rcu(sk, node, &hslot->head, offset) {
net/ipv4/udp.c
2823
sk_for_each_rcu(sk, &hslot->head) {
net/ipv4/udp.c
2856
udp_portaddr_for_each_entry_rcu(sk, &hslot2->head) {
net/ipv4/udp.c
331
sk_add_node_rcu(sk, &hslot->head);
net/ipv4/udp.c
3344
if (hlist_empty(&hslot->head))
net/ipv4/udp.c
3348
sk_for_each(sk, &hslot->head) {
net/ipv4/udp.c
340
&hslot2->head);
net/ipv4/udp.c
343
&hslot2->head);
net/ipv4/udp.c
3541
if (hlist_empty(&hslot2->head))
net/ipv4/udp.c
3545
sk = hlist_entry_safe(hslot2->head.first, struct sock,
net/ipv4/udp.c
3829
INIT_HLIST_HEAD(&table->hash[i].head);
net/ipv4/udp.c
3834
INIT_HLIST_HEAD(&table->hash2[i].hslot.head);
net/ipv4/udp.c
3883
INIT_HLIST_HEAD(&udptable->hash[i].head);
net/ipv4/udp.c
3887
INIT_HLIST_HEAD(&udptable->hash2[i].hslot.head);
net/ipv4/udp.c
453
sk_for_each_rcu(sk, &hslot->head) {
net/ipv4/udp.c
479
udp_portaddr_for_each_entry_rcu(sk, &hslot2->head) {
net/ipv4/udp_diag.c
106
if (hlist_empty(&hslot->head))
net/ipv4/udp_diag.c
110
sk_for_each(sk, &hslot->head) {
net/ipv4/udp_offload.c
148
struct list_head *head,
net/ipv4/udp_offload.c
156
return static_call(udp_tunnel_gro_rcv)(sk, head, skb);
net/ipv4/udp_offload.c
158
return call_gro_receive_sk(udp_sk(sk)->gro_receive, sk, head, skb);
net/ipv4/udp_offload.c
164
struct list_head *head,
net/ipv4/udp_offload.c
167
return call_gro_receive_sk(udp_sk(sk)->gro_receive, sk, head, skb);
net/ipv4/udp_offload.c
24
struct list_head *head,
net/ipv4/udp_offload.c
284
skb->head - (unsigned char *)uh);
net/ipv4/udp_offload.c
301
skb->csum_start = skb_transport_header(skb) - skb->head;
net/ipv4/udp_offload.c
526
gso_skb->csum_start = skb_transport_header(gso_skb) - gso_skb->head;
net/ipv4/udp_offload.c
702
static struct sk_buff *udp_gro_receive_segment(struct list_head *head,
net/ipv4/udp_offload.c
728
list_for_each_entry(p, head, list) {
net/ipv4/udp_offload.c
787
struct sk_buff *udp_gro_receive(struct list_head *head, struct sk_buff *skb,
net/ipv4/udp_offload.c
814
return call_gro_receive(udp_gro_receive_segment, head, skb);
net/ipv4/udp_offload.c
831
list_for_each_entry(p, head, list) {
net/ipv4/udp_offload.c
849
pp = udp_tunnel_gro_rcv(sk, head, skb);
net/ipv4/udp_offload.c
877
struct sk_buff *udp4_gro_receive(struct list_head *head, struct sk_buff *skb)
net/ipv4/udp_offload.c
900
pp = udp_gro_receive(head, skb, uh, sk);
net/ipv4/udp_offload.c
912
skb->csum_start = (unsigned char *)uh - skb->head;
net/ipv4/xfrm4_input.c
182
struct sk_buff *xfrm4_gro_udp_encap_rcv(struct sock *sk, struct list_head *head,
net/ipv4/xfrm4_input.c
213
pp = call_gro_receive(ops->callbacks.gro_receive, head, skb);
net/ipv4/xfrm4_protocol.c
40
#define for_each_protocol_rcu(head, handler) \
net/ipv4/xfrm4_protocol.c
41
for (handler = rcu_dereference(head); \
net/ipv4/xfrm4_protocol.c
49
struct xfrm4_protocol __rcu **head = proto_handlers(protocol);
net/ipv4/xfrm4_protocol.c
51
if (!head)
net/ipv4/xfrm4_protocol.c
54
for_each_protocol_rcu(*head, handler)
net/ipv4/xfrm4_protocol.c
66
struct xfrm4_protocol __rcu **head = proto_handlers(nexthdr);
net/ipv4/xfrm4_protocol.c
72
if (!head)
net/ipv4/xfrm4_protocol.c
83
for_each_protocol_rcu(*head, handler)
net/ipv6/addrconf_core.c
252
static void in6_dev_finish_destroy_rcu(struct rcu_head *head)
net/ipv6/addrconf_core.c
254
struct inet6_dev *idev = container_of(head, struct inet6_dev, rcu);
net/ipv6/addrlabel.c
127
hlist_for_each_entry_rcu(p, &net->ipv6.ip6addrlbl_table.head, list) {
net/ipv6/addrlabel.c
206
hlist_for_each_entry_safe(p, n, &net->ipv6.ip6addrlbl_table.head, list) {
net/ipv6/addrlabel.c
227
hlist_add_head_rcu(&newp->list, &net->ipv6.ip6addrlbl_table.head);
net/ipv6/addrlabel.c
269
hlist_for_each_entry_safe(p, n, &net->ipv6.ip6addrlbl_table.head, list) {
net/ipv6/addrlabel.c
308
INIT_HLIST_HEAD(&net->ipv6.ip6addrlbl_table.head);
net/ipv6/addrlabel.c
322
hlist_for_each_entry_safe(p, n, &net->ipv6.ip6addrlbl_table.head, list) {
net/ipv6/addrlabel.c
336
hlist_for_each_entry_safe(p, n, &net->ipv6.ip6addrlbl_table.head, list) {
net/ipv6/addrlabel.c
503
hlist_for_each_entry_rcu(p, &net->ipv6.ip6addrlbl_table.head, list) {
net/ipv6/esp6_offload.c
53
static struct sk_buff *esp6_gro_receive(struct list_head *head,
net/ipv6/icmp.c
659
if ((u8 *)hdr < skb->head ||
net/ipv6/ila/ila_xlat.c
203
struct ila_map *ila, *head;
net/ipv6/ila/ila_xlat.c
235
head = rhashtable_lookup_fast(&ilan->xlat.rhash_table,
net/ipv6/ila/ila_xlat.c
238
if (!head) {
net/ipv6/ila/ila_xlat.c
243
struct ila_map *tila = head, *prev = NULL;
net/ipv6/ila/ila_xlat.c
265
RCU_INIT_POINTER(ila->next, head);
net/ipv6/ila/ila_xlat.c
267
&head->node,
net/ipv6/ila/ila_xlat.c
286
struct ila_map *ila, *head, *prev;
net/ipv6/ila/ila_xlat.c
292
head = rhashtable_lookup_fast(&ilan->xlat.rhash_table,
net/ipv6/ila/ila_xlat.c
294
ila = head;
net/ipv6/ila/ila_xlat.c
315
head = rcu_dereference_protected(ila->next,
net/ipv6/ila/ila_xlat.c
317
if (head) {
net/ipv6/ila/ila_xlat.c
323
&head->node, rht_params);
net/ipv6/inet6_hashtables.c
279
struct inet_ehash_bucket *head = inet_ehash_bucket(hinfo, hash);
net/ipv6/inet6_hashtables.c
286
sk_nulls_for_each(sk2, node, &head->chain) {
net/ipv6/inet6_hashtables.c
301
sk_nulls_for_each(sk2, node, &head->chain) {
net/ipv6/inet6_hashtables.c
323
__sk_nulls_add_node_rcu(sk, &head->chain);
net/ipv6/inet6_hashtables.c
58
struct inet_ehash_bucket *head;
net/ipv6/inet6_hashtables.c
66
head = &hashinfo->ehash[slot];
net/ipv6/inet6_hashtables.c
68
sk_nulls_for_each_rcu(sk, node, &head->chain) {
net/ipv6/ioam6.c
147
err = rhashtable_lookup_insert_fast(&nsdata->namespaces, &ns->head,
net/ipv6/ioam6.c
182
err = rhashtable_remove_fast(&nsdata->namespaces, &ns->head,
net/ipv6/ioam6.c
343
err = rhashtable_lookup_insert_fast(&nsdata->schemas, &sc->head,
net/ipv6/ioam6.c
380
err = rhashtable_remove_fast(&nsdata->schemas, &sc->head,
net/ipv6/ioam6.c
66
.head_offset = offsetof(struct ioam6_namespace, head),
net/ipv6/ioam6.c
74
.head_offset = offsetof(struct ioam6_schema, head),
net/ipv6/ip6_checksum.c
132
skb->csum_start = skb_transport_header(skb) - skb->head;
net/ipv6/ip6_fib.c
168
void fib6_info_destroy_rcu(struct rcu_head *head)
net/ipv6/ip6_fib.c
170
struct fib6_info *f6i = container_of(head, struct fib6_info, rcu);
net/ipv6/ip6_fib.c
2320
struct hlist_head *head;
net/ipv6/ip6_fib.c
2325
head = &net->ipv6.fib_table_hash[h];
net/ipv6/ip6_fib.c
2326
hlist_for_each_entry_rcu(table, head, tb6_hlist) {
net/ipv6/ip6_fib.c
2416
struct hlist_head *head;
net/ipv6/ip6_fib.c
2421
head = &net->ipv6.fib_table_hash[h];
net/ipv6/ip6_fib.c
2422
hlist_for_each_entry_rcu(table, head, tb6_hlist) {
net/ipv6/ip6_fib.c
2547
struct hlist_head *head = &net->ipv6.fib_table_hash[i];
net/ipv6/ip6_fib.c
2551
hlist_for_each_entry_safe(tb, tmp, head, tb6_hlist) {
net/ipv6/ip6_fib.c
284
struct hlist_head *head;
net/ipv6/ip6_fib.c
290
head = &net->ipv6.fib_table_hash[id & (FIB6_TABLE_HASHSZ - 1)];
net/ipv6/ip6_fib.c
295
hlist_for_each_entry_rcu(tb, head, tb6_hlist, true)
net/ipv6/ip6_fib.c
359
const struct hlist_head *head = &net->ipv6.fib_table_hash[h];
net/ipv6/ip6_fib.c
362
hlist_for_each_entry_rcu(tb, head, tb6_hlist)
net/ipv6/ip6_fib.c
508
struct hlist_head *head = &net->ipv6.fib_table_hash[h];
net/ipv6/ip6_fib.c
511
hlist_for_each_entry_rcu(tb, head, tb6_hlist) {
net/ipv6/ip6_fib.c
638
struct hlist_head *head;
net/ipv6/ip6_fib.c
706
head = &net->ipv6.fib_table_hash[h];
net/ipv6/ip6_fib.c
707
hlist_for_each_entry_rcu(tb, head, tb6_hlist) {
net/ipv6/ip6_flowlabel.c
104
static void fl_free_rcu(struct rcu_head *head)
net/ipv6/ip6_flowlabel.c
106
struct ip6_flowlabel *fl = container_of(head, struct ip6_flowlabel, rcu);
net/ipv6/ip6_gre.c
1552
static void __net_exit ip6gre_exit_rtnl_net(struct net *net, struct list_head *head)
net/ipv6/ip6_gre.c
1562
unregister_netdevice_queue(dev, head);
net/ipv6/ip6_gre.c
1576
unregister_netdevice_queue(t->dev, head);
net/ipv6/ip6_gre.c
2061
static void ip6gre_dellink(struct net_device *dev, struct list_head *head)
net/ipv6/ip6_gre.c
2067
unregister_netdevice_queue(dev, head);
net/ipv6/ip6_icmp.c
70
if (unlikely(!skb_in || skb_network_header(skb_in) < skb_in->head ||
net/ipv6/ip6_input.c
110
struct list_head *head)
net/ipv6/ip6_input.c
116
list_for_each_entry_safe(skb, next, head, list) {
net/ipv6/ip6_input.c
316
static void ip6_sublist_rcv(struct list_head *head, struct net_device *dev,
net/ipv6/ip6_input.c
320
head, dev, NULL, ip6_rcv_finish);
net/ipv6/ip6_input.c
321
ip6_list_rcv_finish(net, NULL, head);
net/ipv6/ip6_input.c
325
void ipv6_list_rcv(struct list_head *head, struct packet_type *pt,
net/ipv6/ip6_input.c
333
list_for_each_entry_safe(skb, next, head, list) {
net/ipv6/ip6_input.c
82
static void ip6_sublist_rcv_finish(struct list_head *head)
net/ipv6/ip6_input.c
86
list_for_each_entry_safe(skb, next, head, list) {
net/ipv6/ip6_offload.c
137
skb->network_header = skb_mac_header(skb) + nhoff - skb->head;
net/ipv6/ip6_offload.c
150
skb->head - (unsigned char *)(ipv6h + 1);
net/ipv6/ip6_offload.c
154
skb->network_header = (u8 *)ipv6h - skb->head;
net/ipv6/ip6_offload.c
203
INDIRECT_CALLABLE_SCOPE struct sk_buff *ipv6_gro_receive(struct list_head *head,
net/ipv6/ip6_offload.c
247
list_for_each_entry(p, head, list) {
net/ipv6/ip6_offload.c
288
pp = tcp6_gro_receive(head, skb);
net/ipv6/ip6_offload.c
291
pp = udp6_gro_receive(head, skb);
net/ipv6/ip6_offload.c
294
pp = ops->callbacks.gro_receive(head, skb);
net/ipv6/ip6_offload.c
301
static struct sk_buff *sit_ip6ip6_gro_receive(struct list_head *head,
net/ipv6/ip6_offload.c
313
return ipv6_gro_receive(head, skb);
net/ipv6/ip6_offload.c
316
static struct sk_buff *ip4ip6_gro_receive(struct list_head *head,
net/ipv6/ip6_offload.c
328
return inet_gro_receive(head, skb);
net/ipv6/ip6_tunnel.c
2142
static void ip6_tnl_dellink(struct net_device *dev, struct list_head *head)
net/ipv6/ip6_tunnel.c
2148
unregister_netdevice_queue(dev, head);
net/ipv6/ip6_tunnel.c
848
nh = skb_network_header(skb) - skb->head;
net/ipv6/ip6_tunnel.c
859
ipv6h = (struct ipv6hdr *)(skb->head + nh);
net/ipv6/ip6_vti.c
1021
static void vti6_dellink(struct net_device *dev, struct list_head *head)
net/ipv6/ip6_vti.c
1027
unregister_netdevice_queue(dev, head);
net/ipv6/ip6mr.c
718
struct list_head *head)
net/ipv6/ip6mr.c
768
unregister_netdevice_queue(dev, head);
net/ipv6/ip6mr.c
774
static inline void ip6mr_cache_free_rcu(struct rcu_head *head)
net/ipv6/ip6mr.c
776
struct mr_mfc *c = container_of(head, struct mr_mfc, rcu);
net/ipv6/netfilter/ip6t_eui64.c
24
if (!(skb_mac_header(skb) >= skb->head &&
net/ipv6/netfilter/nf_conntrack_reasm.c
349
memmove(skb->head + sizeof(struct frag_hdr), skb->head,
net/ipv6/netfilter/nf_conntrack_reasm.c
350
(skb->data - skb->head) - sizeof(struct frag_hdr));
net/ipv6/reassembly.c
287
memmove(skb->head + sizeof(struct frag_hdr), skb->head,
net/ipv6/reassembly.c
288
(skb->data - skb->head) - sizeof(struct frag_hdr));
net/ipv6/route.c
133
struct list_head head;
net/ipv6/route.c
145
list_add_tail(&rt->dst.rt_uncached, &ul->head);
net/ipv6/route.c
168
if (list_empty(&ul->head))
net/ipv6/route.c
172
list_for_each_entry_safe(rt, safe, &ul->head, dst.rt_uncached) {
net/ipv6/route.c
4524
struct hlist_head *head;
net/ipv6/route.c
4530
head = &net->ipv6.fib_table_hash[h];
net/ipv6/route.c
4531
hlist_for_each_entry_rcu(table, head, tb6_hlist) {
net/ipv6/route.c
6935
INIT_LIST_HEAD(&ul->head);
net/ipv6/seg6_iptunnel.c
33
int head = 0;
net/ipv6/seg6_iptunnel.c
40
head = sizeof(struct ipv6hdr);
net/ipv6/seg6_iptunnel.c
47
return ((tuninfo->srh->hdrlen + 1) << 3) + head;
net/ipv6/sit.c
1755
static void ipip6_dellink(struct net_device *dev, struct list_head *head)
net/ipv6/sit.c
1761
unregister_netdevice_queue(dev, head);
net/ipv6/sit.c
1799
static void __net_exit sit_exit_rtnl_net(struct net *net, struct list_head *head)
net/ipv6/sit.c
1807
unregister_netdevice_queue(dev, head);
net/ipv6/sit.c
1820
unregister_netdevice_queue(t->dev, head);
net/ipv6/sit.c
412
static void prl_list_destroy_rcu(struct rcu_head *head)
net/ipv6/sit.c
416
p = container_of(head, struct ip_tunnel_prl_entry, rcu_head);
net/ipv6/tcpv6_offload.c
17
static void tcp6_check_fraglist_gro(struct list_head *head, struct sk_buff *skb,
net/ipv6/tcpv6_offload.c
27
p = tcp_gro_lookup(head, th);
net/ipv6/tcpv6_offload.c
45
static __always_inline struct sk_buff *tcp6_gro_receive(struct list_head *head,
net/ipv6/tcpv6_offload.c
61
tcp6_check_fraglist_gro(head, skb, th);
net/ipv6/tcpv6_offload.c
63
return tcp_gro_receive(head, skb, th);
net/ipv6/tunnel6.c
117
#define for_each_tunnel_rcu(head, handler) \
net/ipv6/tunnel6.c
118
for (handler = rcu_dereference(head); \
net/ipv6/tunnel6.c
161
struct xfrm6_tunnel __rcu *head;
net/ipv6/tunnel6.c
165
head = (proto == IPPROTO_IPV6) ? tunnel6_handlers : tunnel46_handlers;
net/ipv6/tunnel6.c
167
for_each_tunnel_rcu(head, handler) {
net/ipv6/udp.c
1002
sk_for_each_entry_offset_rcu(sk, node, &hslot->head, offset) {
net/ipv6/udp.c
1213
udp_portaddr_for_each_entry_rcu(sk, &hslot2->head) {
net/ipv6/udp.c
1340
skb->csum_start = skb_transport_header(skb) - skb->head;
net/ipv6/udp.c
206
sk_for_each_rcu(sk, &hslot->head) {
net/ipv6/udp.c
231
udp_portaddr_for_each_entry_rcu(sk, &hslot2->head) {
net/ipv6/udp_offload.c
135
struct sk_buff *udp6_gro_receive(struct list_head *head, struct sk_buff *skb)
net/ipv6/udp_offload.c
159
pp = udp_gro_receive(head, skb, uh, sk);
net/ipv6/udp_offload.c
94
packet_start = (u8 *) skb->head + SKB_GSO_CB(skb)->mac_offset;
net/ipv6/xfrm6_input.c
179
struct sk_buff *xfrm6_gro_udp_encap_rcv(struct sock *sk, struct list_head *head,
net/ipv6/xfrm6_input.c
190
return xfrm4_gro_udp_encap_rcv(sk, head, skb);
net/ipv6/xfrm6_input.c
213
pp = call_gro_receive(ops->callbacks.gro_receive, head, skb);
net/ipv6/xfrm6_protocol.c
41
#define for_each_protocol_rcu(head, handler) \
net/ipv6/xfrm6_protocol.c
42
for (handler = rcu_dereference(head); \
net/ipv6/xfrm6_protocol.c
50
struct xfrm6_protocol __rcu **head = proto_handlers(protocol);
net/ipv6/xfrm6_protocol.c
52
if (!head)
net/ipv6/xfrm6_protocol.c
67
struct xfrm6_protocol __rcu **head = proto_handlers(nexthdr);
net/ipv6/xfrm6_protocol.c
73
if (!head)
net/ipv6/xfrm6_protocol.c
96
for_each_protocol_rcu(*head, handler)
net/ipv6/xfrm6_tunnel.c
181
static void x6spi_destroy_rcu(struct rcu_head *head)
net/ipv6/xfrm6_tunnel.c
184
container_of(head, struct xfrm6_tunnel_spi, rcu_head));
net/iucv/af_iucv.c
1060
struct iucv_array *iba = (struct iucv_array *)skb->head;
net/iucv/af_iucv.c
1162
struct iucv_array *iba = (struct iucv_array *)skb->head;
net/iucv/af_iucv.c
1601
sk_for_each(sk, &iucv_sk_list.head)
net/iucv/af_iucv.c
2058
sk_for_each(sk, &iucv_sk_list.head) {
net/iucv/af_iucv.c
2189
sk_for_each(sk, &iucv_sk_list.head) {
net/iucv/af_iucv.c
269
sk_for_each(sk, &iucv_sk_list.head)
net/iucv/af_iucv.c
309
sk_add_node(sk, &l->head);
net/kcm/kcmsock.c
1204
struct list_head *head;
net/kcm/kcmsock.c
1217
head = &mux->kcm_socks;
net/kcm/kcmsock.c
1221
head = &tkcm->kcm_sock_list;
net/kcm/kcmsock.c
1225
list_add(&kcm->kcm_sock_list, head);
net/kcm/kcmsock.c
1246
struct list_head *head;
net/kcm/kcmsock.c
1317
head = &mux->psocks;
net/kcm/kcmsock.c
1321
head = &tpsock->psock_list;
net/kcm/kcmsock.c
1325
list_add(&psock->psock_list, head);
net/kcm/kcmsock.c
222
static void requeue_rx_msgs(struct kcm_mux *mux, struct sk_buff_head *head)
net/kcm/kcmsock.c
227
while ((skb = skb_dequeue(head))) {
net/kcm/kcmsock.c
256
struct sk_buff *head)
net/kcm/kcmsock.c
276
psock->ready_rx_msg = head;
net/kcm/kcmsock.c
588
struct sk_buff *head;
net/kcm/kcmsock.c
606
while ((head = skb_peek(&sk->sk_write_queue))) {
net/kcm/kcmsock.c
610
struct kcm_tx_msg *txm = kcm_tx_msg(head);
net/kcm/kcmsock.c
619
skb = head;
net/kcm/kcmsock.c
679
if (skb == head) {
net/kcm/kcmsock.c
695
kfree_skb(head);
net/kcm/kcmsock.c
699
if (!head) {
net/kcm/kcmsock.c
752
struct sk_buff *skb = NULL, *head = NULL, *frag_prev = NULL;
net/kcm/kcmsock.c
770
head = kcm->seq_skb;
net/kcm/kcmsock.c
771
skb = kcm_tx_msg(head)->last_skb;
net/kcm/kcmsock.c
786
head = alloc_skb(0, sk->sk_allocation);
net/kcm/kcmsock.c
787
while (!head) {
net/kcm/kcmsock.c
793
head = alloc_skb(0, sk->sk_allocation);
net/kcm/kcmsock.c
796
skb = head;
net/kcm/kcmsock.c
822
if (head == skb)
net/kcm/kcmsock.c
823
skb_shinfo(head)->frag_list = tskb;
net/kcm/kcmsock.c
852
if (head != skb)
net/kcm/kcmsock.c
853
head->truesize += copy;
net/kcm/kcmsock.c
881
if (head != skb) {
net/kcm/kcmsock.c
882
head->len += copy;
net/kcm/kcmsock.c
883
head->data_len += copy;
net/kcm/kcmsock.c
898
if (head) {
net/kcm/kcmsock.c
900
__skb_queue_tail(&sk->sk_write_queue, head);
net/kcm/kcmsock.c
922
if (head) {
net/kcm/kcmsock.c
923
kcm->seq_skb = head;
net/kcm/kcmsock.c
924
kcm_tx_msg(head)->last_skb = skb;
net/kcm/kcmsock.c
944
if (head == frag_prev)
net/kcm/kcmsock.c
945
skb_shinfo(head)->frag_list = NULL;
net/kcm/kcmsock.c
959
if (head != kcm->seq_skb)
net/kcm/kcmsock.c
960
kfree_skb(head);
net/kcm/kcmsock.c
962
kfree_skb(head);
net/llc/llc_proc.c
43
struct hlist_nulls_head *head = &sap->sk_laddr_hash[i];
net/llc/llc_proc.c
46
sk_nulls_for_each(sk, node, head) {
net/mac80211/cfg.c
1338
if (!params->head && !old)
net/mac80211/cfg.c
1342
if (params->head)
net/mac80211/cfg.c
1377
new->head = ((u8 *) new) + sizeof(*new);
net/mac80211/cfg.c
1378
new->tail = new->head + new_head_len;
net/mac80211/cfg.c
1413
if (params->head)
net/mac80211/cfg.c
1414
memcpy(new->head, params->head, new_head_len);
net/mac80211/cfg.c
1416
memcpy(new->head, old->head, new_head_len);
net/mac80211/cfg.c
3965
new_beacon->head = pos;
net/mac80211/cfg.c
3966
memcpy(pos, beacon->head, beacon->head_len);
net/mac80211/driver-ops.h
1405
struct sk_buff *head,
net/mac80211/driver-ops.h
1411
return local->ops->can_aggregate_in_amsdu(&local->hw, head, skb);
net/mac80211/ibss.c
142
presp->cntdwn_counter_offsets[0] = (pos - presp->head);
net/mac80211/ibss.c
1537
skb_put_data(skb, presp->head, presp->head_len);
net/mac80211/ibss.c
201
presp->head_len = pos - presp->head;
net/mac80211/ibss.c
314
mgmt = (void *)presp->head;
net/mac80211/ibss.c
73
presp->head = (void *)(presp + 1);
net/mac80211/ibss.c
75
mgmt = (void *) presp->head;
net/mac80211/ieee80211_i.h
278
u8 *head, *tail;
net/mac80211/mesh.c
1012
bcn->head = ((u8 *) bcn) + sizeof(*bcn);
net/mac80211/mesh.c
1103
memcpy(bcn->head, skb->data, bcn->head_len);
net/mac80211/mesh.c
1107
bcn->tail = bcn->head + bcn->head_len;
net/mac80211/mesh.c
1449
skb_put_data(presp, bcn->head, bcn->head_len);
net/mac80211/tx.c
3416
struct sk_buff **frag_tail, *head;
net/mac80211/tx.c
3467
head = skb_peek_tail(&flow->queue);
net/mac80211/tx.c
3468
if (!head || skb_is_gso(head))
net/mac80211/tx.c
3471
orig_truesize = head->truesize;
net/mac80211/tx.c
3472
orig_len = head->len;
net/mac80211/tx.c
3474
if (skb->len + head->len > max_amsdu_len)
net/mac80211/tx.c
3478
nfrags += 1 + skb_shinfo(head)->nr_frags;
net/mac80211/tx.c
3479
frag_tail = &skb_shinfo(head)->frag_list;
net/mac80211/tx.c
3492
if (!drv_can_aggregate_in_amsdu(local, head, skb))
net/mac80211/tx.c
3495
if (!ieee80211_amsdu_prepare_head(sdata, fast_tx, head))
net/mac80211/tx.c
3504
frag_tail = &skb_shinfo(head)->frag_list;
net/mac80211/tx.c
3514
if ((head->len - hdrlen) & 3)
net/mac80211/tx.c
3515
pad = 4 - ((head->len - hdrlen) & 3);
net/mac80211/tx.c
3533
head->len += skb->len;
net/mac80211/tx.c
3534
head->data_len += skb->len;
net/mac80211/tx.c
3538
fq->memory_usage += head->truesize - orig_truesize;
net/mac80211/tx.c
3539
if (head->len != orig_len) {
net/mac80211/tx.c
3540
flow->backlog += head->len - orig_len;
net/mac80211/tx.c
3541
tin->backlog_bytes += head->len - orig_len;
net/mac80211/tx.c
4058
struct txq_info *txqi = NULL, *head = NULL;
net/mac80211/tx.c
4073
if (txqi == head) {
net/mac80211/tx.c
4080
if (!head)
net/mac80211/tx.c
4081
head = txqi;
net/mac80211/tx.c
5101
beacon_data = beacon->head;
net/mac80211/tx.c
5105
beacon_data = beacon->head;
net/mac80211/tx.c
5251
beacon_data = beacon->head;
net/mac80211/tx.c
5260
beacon_data = beacon->head;
net/mac80211/tx.c
5444
skb_put_data(skb, beacon->head, beacon->head_len);
net/mac80211/tx.c
5663
skb_put_data(skb, beacon->head, beacon->head_len);
net/mac80211/tx.c
5701
skb_put_data(skb, beacon->head, beacon->head_len);
net/mac80211/tx.c
6096
if (!beacon || !beacon->head)
net/mpls/af_mpls.c
1520
static void mpls_dev_destroy_rcu(struct rcu_head *head)
net/mpls/af_mpls.c
1522
struct mpls_dev *mdev = container_of(head, struct mpls_dev, rcu);
net/mpls/af_mpls.c
557
static void mpls_rt_free_rcu(struct rcu_head *head)
net/mpls/af_mpls.c
561
rt = container_of(head, struct mpls_route, rt_rcu);
net/mptcp/protocol.c
4602
INIT_LIST_HEAD(&delegated->head);
net/mptcp/protocol.h
495
struct list_head head;
net/mptcp/protocol.h
727
schedule = list_empty(&delegated->head);
net/mptcp/protocol.h
728
list_add_tail(&subflow->delegated_node, &delegated->head);
net/mptcp/protocol.h
742
if (list_empty(&delegated->head)) {
net/mptcp/protocol.h
747
ret = list_first_entry(&delegated->head, struct mptcp_subflow_context, delegated_node);
net/mptcp/subflow.c
1896
struct request_sock *req, *head, *tail;
net/mptcp/subflow.c
1905
head = queue->rskq_accept_head;
net/mptcp/subflow.c
1911
if (!head)
net/mptcp/subflow.c
1919
for (req = head; req; req = req->dl_next) {
net/mptcp/subflow.c
1959
queue->rskq_accept_head = head;
net/netfilter/core.c
651
void nf_hook_slow_list(struct list_head *head, struct nf_hook_state *state,
net/netfilter/core.c
658
list_for_each_entry_safe(skb, next, head, list) {
net/netfilter/core.c
66
struct nf_hook_entries_rcu_head *head;
net/netfilter/core.c
665
list_splice(&sublist, head);
net/netfilter/core.c
68
head = container_of(h, struct nf_hook_entries_rcu_head, head);
net/netfilter/core.c
69
kvfree(head->allocation);
net/netfilter/core.c
74
struct nf_hook_entries_rcu_head *head;
net/netfilter/core.c
83
head = (void *)&ops[num];
net/netfilter/core.c
84
head->allocation = e;
net/netfilter/core.c
85
call_rcu(&head->head, __nf_hook_entries_free);
net/netfilter/ipset/ip_set_bitmap_gen.h
310
.head = mtype_head,
net/netfilter/ipset/ip_set_bitmap_ipmac.c
223
if (skb_mac_header(skb) < skb->head ||
net/netfilter/ipset/ip_set_core.c
1188
ip_set_destroy_set_rcu(struct rcu_head *head)
net/netfilter/ipset/ip_set_core.c
1190
struct ip_set *set = container_of(head, struct ip_set, rcu);
net/netfilter/ipset/ip_set_core.c
1651
ret = set->variant->head(set, skb);
net/netfilter/ipset/ip_set_hash_gen.h
1446
.head = mtype_head,
net/netfilter/ipset/ip_set_hash_ipmac.c
208
if (skb_mac_header(skb) < skb->head ||
net/netfilter/ipset/ip_set_hash_ipmac.c
92
if (skb_mac_header(skb) < skb->head ||
net/netfilter/ipset/ip_set_hash_mac.c
80
if (skb_mac_header(skb) < skb->head ||
net/netfilter/ipset/ip_set_list_set.c
565
.head = list_set_head,
net/netfilter/ipvs/ip_vs_app.c
61
static void ip_vs_app_inc_rcu_free(struct rcu_head *head)
net/netfilter/ipvs/ip_vs_app.c
63
struct ip_vs_app *inc = container_of(head, struct ip_vs_app, rcu_head);
net/netfilter/ipvs/ip_vs_conn.c
811
static void ip_vs_conn_rcu_free(struct rcu_head *head)
net/netfilter/ipvs/ip_vs_conn.c
813
struct ip_vs_conn *cp = container_of(head, struct ip_vs_conn,
net/netfilter/ipvs/ip_vs_core.c
837
skb->csum_start = skb_network_header(skb) - skb->head + icmp_offset;
net/netfilter/ipvs/ip_vs_ctl.c
517
static void ip_vs_service_rcu_free(struct rcu_head *head)
net/netfilter/ipvs/ip_vs_ctl.c
521
svc = container_of(head, struct ip_vs_service, rcu_head);
net/netfilter/ipvs/ip_vs_ctl.c
747
void ip_vs_dest_dst_rcu_free(struct rcu_head *head)
net/netfilter/ipvs/ip_vs_ctl.c
749
struct ip_vs_dest_dst *dest_dst = container_of(head,
net/netfilter/ipvs/ip_vs_ctl.c
819
static void ip_vs_dest_rcu_free(struct rcu_head *head)
net/netfilter/ipvs/ip_vs_ctl.c
823
dest = container_of(head, struct ip_vs_dest, rcu_head);
net/netfilter/ipvs/ip_vs_ctl.c
858
static void ip_vs_stats_rcu_free(struct rcu_head *head)
net/netfilter/ipvs/ip_vs_ctl.c
860
struct ip_vs_stats_rcu *rs = container_of(head,
net/netfilter/ipvs/ip_vs_lblc.c
128
static void ip_vs_lblc_rcu_free(struct rcu_head *head)
net/netfilter/ipvs/ip_vs_lblc.c
130
struct ip_vs_lblc_entry *en = container_of(head,
net/netfilter/ipvs/ip_vs_lblcr.c
123
static void ip_vs_lblcr_elem_rcu_free(struct rcu_head *head)
net/netfilter/ipvs/ip_vs_lblcr.c
127
e = container_of(head, struct ip_vs_dest_set_elem, rcu_head);
net/netfilter/ipvs/ip_vs_mh.c
370
static void ip_vs_mh_state_free(struct rcu_head *head)
net/netfilter/ipvs/ip_vs_mh.c
374
s = container_of(head, struct ip_vs_mh_state, rcu_head);
net/netfilter/ipvs/ip_vs_sync.c
275
unsigned char *head;
net/netfilter/ipvs/ip_vs_sync.c
348
sb->head = (unsigned char *)sb->mesg + sizeof(struct ip_vs_sync_mesg);
net/netfilter/ipvs/ip_vs_sync.c
434
sb->head = (unsigned char *)mesg + sizeof(struct ip_vs_sync_mesg_v0);
net/netfilter/ipvs/ip_vs_sync.c
572
if (buff->head + len > buff->end || !m->nr_conns) {
net/netfilter/ipvs/ip_vs_sync.c
589
s = (struct ip_vs_sync_conn_v0 *) buff->head;
net/netfilter/ipvs/ip_vs_sync.c
610
buff->head += len;
net/netfilter/ipvs/ip_vs_sync.c
690
pad = (4 - (size_t) buff->head) & 3;
net/netfilter/ipvs/ip_vs_sync.c
692
if (buff->head + len + pad > buff->end || m->reserved) {
net/netfilter/ipvs/ip_vs_sync.c
711
p = buff->head;
net/netfilter/ipvs/ip_vs_sync.c
712
buff->head += pad + len;
net/netfilter/nf_conncount.c
202
list_for_each_entry_safe(conn, conn_n, &list->head, node) {
net/netfilter/nf_conncount.c
264
list_add_tail(&conn->node, &list->head);
net/netfilter/nf_conncount.c
292
INIT_LIST_HEAD(&list->head);
net/netfilter/nf_conncount.c
313
list_for_each_entry_safe(conn, conn_n, &list->head, node) {
net/netfilter/nf_conncount.c
473
list_add(&conn->node, &rbconn->list.head);
net/netfilter/nf_conncount.c
654
list_for_each_entry_safe(conn, conn_n, &list->head, node)
net/netfilter/nf_conntrack_core.c
1422
struct hlist_nulls_head *head)
net/netfilter/nf_conntrack_core.c
1429
hlist_nulls_for_each_entry_rcu(h, n, head, hnnode) {
net/netfilter/nf_conntrack_expect.c
392
static void nf_ct_expect_free_rcu(struct rcu_head *head)
net/netfilter/nf_conntrack_expect.c
396
exp = container_of(head, struct nf_conntrack_expect, rcu);
net/netfilter/nf_conntrack_expect.c
609
struct hlist_node *head)
net/netfilter/nf_conntrack_expect.c
613
head = rcu_dereference(hlist_next_rcu(head));
net/netfilter/nf_conntrack_expect.c
614
while (head == NULL) {
net/netfilter/nf_conntrack_expect.c
617
head = rcu_dereference(hlist_first_rcu(&nf_ct_expect_hash[st->bucket]));
net/netfilter/nf_conntrack_expect.c
619
return head;
net/netfilter/nf_conntrack_expect.c
624
struct hlist_node *head = ct_expect_get_first(seq);
net/netfilter/nf_conntrack_expect.c
626
if (head)
net/netfilter/nf_conntrack_expect.c
627
while (pos && (head = ct_expect_get_next(seq, head)))
net/netfilter/nf_conntrack_expect.c
629
return pos ? NULL : head;
net/netfilter/nf_conntrack_helper.c
273
list_add_rcu(&n->head, &nf_ct_helper_expectfn_list);
net/netfilter/nf_conntrack_helper.c
281
list_del_rcu(&n->head);
net/netfilter/nf_conntrack_helper.c
293
list_for_each_entry_rcu(cur, &nf_ct_helper_expectfn_list, head) {
net/netfilter/nf_conntrack_helper.c
310
list_for_each_entry_rcu(cur, &nf_ct_helper_expectfn_list, head) {
net/netfilter/nf_flow_table_xdp.c
11
struct list_head head;
net/netfilter/nf_flow_table_xdp.c
112
list_for_each_entry_safe(ft_elem, ft_next, &elem->head, head) {
net/netfilter/nf_flow_table_xdp.c
114
list_del_rcu(&ft_elem->head);
net/netfilter/nf_flow_table_xdp.c
119
if (list_empty(&elem->head))
net/netfilter/nf_flow_table_xdp.c
19
struct list_head head;
net/netfilter/nf_flow_table_xdp.c
40
ft_elem = list_first_or_null_rcu(&iter->head,
net/netfilter/nf_flow_table_xdp.c
42
head);
net/netfilter/nf_flow_table_xdp.c
78
INIT_LIST_HEAD(&elem->head);
net/netfilter/nf_flow_table_xdp.c
81
list_add_tail_rcu(&ft_elem->head, &elem->head);
net/netfilter/nf_log_syslog.c
812
if (p < skb->head)
net/netfilter/nf_synproxy_core.c
427
nskb->csum_start = (unsigned char *)nth - nskb->head;
net/netfilter/nf_synproxy_core.c
825
nskb->csum_start = (unsigned char *)nth - nskb->head;
net/netfilter/nf_tables_api.c
10154
LIST_HEAD(head);
net/netfilter/nf_tables_api.c
10157
list_splice_init(&nft_net->destroy_list, &head);
net/netfilter/nf_tables_api.c
10160
if (list_empty(&head))
net/netfilter/nf_tables_api.c
10165
list_for_each_entry_safe(trans, next, &head, list) {
net/netfilter/nfnetlink.c
317
struct list_head head;
net/netfilter/nfnetlink.c
335
list_add_tail(&nfnl_err->head, list);
net/netfilter/nfnetlink.c
342
list_del(&nfnl_err->head);
net/netfilter/nfnetlink.c
350
list_for_each_entry_safe(nfnl_err, next, err_list, head)
net/netfilter/nfnetlink.c
358
list_for_each_entry_safe(nfnl_err, next, err_list, head) {
net/netfilter/nfnetlink_acct.c
137
list_add_tail_rcu(&nfacct->head, &nfnl_acct_net->nfnl_acct_list);
net/netfilter/nfnetlink_acct.c
209
list_for_each_entry_rcu(cur, &nfnl_acct_net->nfnl_acct_list, head) {
net/netfilter/nfnetlink_acct.c
297
list_for_each_entry(cur, &nfnl_acct_net->nfnl_acct_list, head) {
net/netfilter/nfnetlink_acct.c
33
struct list_head head;
net/netfilter/nfnetlink_acct.c
335
list_del_rcu(&cur->head);
net/netfilter/nfnetlink_acct.c
352
list_for_each_entry_safe(cur, tmp, &nfnl_acct_net->nfnl_acct_list, head)
net/netfilter/nfnetlink_acct.c
359
list_for_each_entry(cur, &nfnl_acct_net->nfnl_acct_list, head) {
net/netfilter/nfnetlink_acct.c
423
list_for_each_entry_rcu(cur, &nfnl_acct_net->nfnl_acct_list, head) {
net/netfilter/nfnetlink_acct.c
515
list_for_each_entry_safe(cur, tmp, &nfnl_acct_net->nfnl_acct_list, head) {
net/netfilter/nfnetlink_acct.c
516
list_del_rcu(&cur->head);
net/netfilter/nfnetlink_acct.c
75
list_for_each_entry(nfacct, &nfnl_acct_net->nfnl_acct_list, head) {
net/netfilter/nfnetlink_cttimeout.c
119
list_for_each_entry(timeout, &pernet->nfct_timeout_list, head) {
net/netfilter/nfnetlink_cttimeout.c
173
list_add_tail_rcu(&timeout->head, &pernet->nfct_timeout_list);
net/netfilter/nfnetlink_cttimeout.c
241
list_for_each_entry_rcu(cur, &pernet->nfct_timeout_list, head) {
net/netfilter/nfnetlink_cttimeout.c
282
list_for_each_entry(cur, &pernet->nfct_timeout_list, head) {
net/netfilter/nfnetlink_cttimeout.c
320
list_del_rcu(&timeout->head);
net/netfilter/nfnetlink_cttimeout.c
340
head)
net/netfilter/nfnetlink_cttimeout.c
347
list_for_each_entry(cur, &pernet->nfct_timeout_list, head) {
net/netfilter/nfnetlink_cttimeout.c
37
struct list_head head;
net/netfilter/nfnetlink_cttimeout.c
511
list_for_each_entry_rcu(timeout, &pernet->nfct_timeout_list, head) {
net/netfilter/nfnetlink_cttimeout.c
592
list_for_each_entry_safe(cur, tmp, &pernet->nfct_timeout_list, head) {
net/netfilter/nfnetlink_cttimeout.c
593
list_del_rcu(&cur->head);
net/netfilter/nfnetlink_hook.c
389
const void *head;
net/netfilter/nfnetlink_hook.c
405
head = nfnl_hook_entries_head(family, hooknum, net, name);
net/netfilter/nfnetlink_hook.c
408
if (head && IS_ERR(head))
net/netfilter/nfnetlink_hook.c
409
return PTR_ERR(head);
net/netfilter/nfnetlink_hook.c
416
ctx->headv = (unsigned long)head;
net/netfilter/nfnetlink_log.c
1034
struct hlist_head *head = &log->instance_table[st->bucket];
net/netfilter/nfnetlink_log.c
1036
if (!hlist_empty(head))
net/netfilter/nfnetlink_log.c
1037
return rcu_dereference(hlist_first_rcu(head));
net/netfilter/nfnetlink_log.c
1048
struct hlist_head *head;
net/netfilter/nfnetlink_log.c
1054
head = &log->instance_table[st->bucket];
net/netfilter/nfnetlink_log.c
1055
h = rcu_dereference(hlist_first_rcu(head));
net/netfilter/nfnetlink_log.c
1063
struct hlist_node *head;
net/netfilter/nfnetlink_log.c
1064
head = get_first(net, st);
net/netfilter/nfnetlink_log.c
1066
if (head)
net/netfilter/nfnetlink_log.c
1067
while (pos && (head = get_next(net, st, head)))
net/netfilter/nfnetlink_log.c
1069
return pos ? NULL : head;
net/netfilter/nfnetlink_log.c
108
const struct hlist_head *head;
net/netfilter/nfnetlink_log.c
111
head = &log->instance_table[instance_hashfn(group_num)];
net/netfilter/nfnetlink_log.c
112
hlist_for_each_entry_rcu(inst, head, hlist) {
net/netfilter/nfnetlink_log.c
149
static void nfulnl_instance_free_rcu(struct rcu_head *head)
net/netfilter/nfnetlink_log.c
152
container_of(head, struct nfulnl_instance, rcu);
net/netfilter/nfnetlink_log.c
595
if (hwhdrp >= skb->head &&
net/netfilter/nfnetlink_log.c
840
struct hlist_head *head = &log->instance_table[i];
net/netfilter/nfnetlink_log.c
842
hlist_for_each_entry_safe(inst, t2, head, hlist) {
net/netfilter/nfnetlink_queue.c
121
struct hlist_head *head;
net/netfilter/nfnetlink_queue.c
1223
struct hlist_head *head = &q->instance_table[i];
net/netfilter/nfnetlink_queue.c
1225
hlist_for_each_entry_rcu(inst, head, hlist)
net/netfilter/nfnetlink_queue.c
124
head = &q->instance_table[instance_hashfn(queue_num)];
net/netfilter/nfnetlink_queue.c
125
hlist_for_each_entry_rcu(inst, head, hlist) {
net/netfilter/nfnetlink_queue.c
1265
struct hlist_head *head = &q->instance_table[i];
net/netfilter/nfnetlink_queue.c
1267
hlist_for_each_entry_rcu(inst, head, hlist)
net/netfilter/nfnetlink_queue.c
1287
struct hlist_head *head = &q->instance_table[i];
net/netfilter/nfnetlink_queue.c
1289
hlist_for_each_entry_safe(inst, t2, head, hlist) {
net/netfilter/nfnetlink_queue.c
1756
struct hlist_node *head;
net/netfilter/nfnetlink_queue.c
1757
head = get_first(seq);
net/netfilter/nfnetlink_queue.c
1759
if (head)
net/netfilter/nfnetlink_queue.c
1760
while (pos && (head = get_next(seq, head)))
net/netfilter/nfnetlink_queue.c
1762
return pos ? NULL : head;
net/netfilter/nft_set_bitmap.c
103
list_for_each_entry_rcu(be, &priv->list, head,
net/netfilter/nft_set_bitmap.c
123
list_for_each_entry_rcu(be, &priv->list, head) {
net/netfilter/nft_set_bitmap.c
151
list_add_tail_rcu(&new->head, &priv->list);
net/netfilter/nft_set_bitmap.c
167
list_del_rcu(&be->head);
net/netfilter/nft_set_bitmap.c
17
struct list_head head;
net/netfilter/nft_set_bitmap.c
229
list_for_each_entry_rcu(be, &priv->list, head,
net/netfilter/nft_set_bitmap.c
285
list_for_each_entry_safe(be, n, &priv->list, head)
net/netfilter/x_tables.c
1595
struct list_head *head, *curr;
net/netfilter/x_tables.c
1623
trav->head = trav->curr = is_target ?
net/netfilter/x_tables.c
1628
if (trav->curr != trav->head)
net/netfilter/x_tables.c
1632
trav->head = trav->curr = is_target ?
net/netfilter/x_tables.c
1638
if (trav->curr != trav->head)
net/netfilter/x_tables.c
1693
if (trav->curr == trav->head)
net/netfilter/x_tables.c
1727
if (trav->curr == trav->head)
net/netfilter/xt_hashlimit.c
258
static void dsthash_free_rcu(struct rcu_head *head)
net/netfilter/xt_hashlimit.c
260
struct dsthash_ent *ent = container_of(head, struct dsthash_ent, rcu);
net/netfilter/xt_hashlimit.c
364
struct hlist_head *head = &ht->hash[i];
net/netfilter/xt_hashlimit.c
368
if (hlist_empty(head))
net/netfilter/xt_hashlimit.c
372
hlist_for_each_entry_safe(dh, n, head, node) {
net/netfilter/xt_mac.c
32
if (skb_mac_header(skb) < skb->head)
net/netfilter/xt_recent.c
501
const struct list_head *head = e->list.next;
net/netfilter/xt_recent.c
504
while (head == &t->iphash[st->bucket]) {
net/netfilter/xt_recent.c
507
head = t->iphash[st->bucket].next;
net/netfilter/xt_recent.c
509
return list_entry(head, struct recent_entry, list);
net/netlabel/netlabel_addrlist.c
101
list_for_each_entry_rcu(iter, head, list)
net/netlabel/netlabel_addrlist.c
123
struct list_head *head)
net/netlabel/netlabel_addrlist.c
127
list_for_each_entry_rcu(iter, head, list)
net/netlabel/netlabel_addrlist.c
148
int netlbl_af4list_add(struct netlbl_af4list *entry, struct list_head *head)
net/netlabel/netlabel_addrlist.c
152
iter = netlbl_af4list_search(entry->addr, head);
net/netlabel/netlabel_addrlist.c
161
list_for_each_entry_rcu(iter, head, list)
net/netlabel/netlabel_addrlist.c
169
list_add_tail_rcu(&entry->list, head);
net/netlabel/netlabel_addrlist.c
185
int netlbl_af6list_add(struct netlbl_af6list *entry, struct list_head *head)
net/netlabel/netlabel_addrlist.c
189
iter = netlbl_af6list_search(&entry->addr, head);
net/netlabel/netlabel_addrlist.c
199
list_for_each_entry_rcu(iter, head, list)
net/netlabel/netlabel_addrlist.c
207
list_add_tail_rcu(&entry->list, head);
net/netlabel/netlabel_addrlist.c
240
struct list_head *head)
net/netlabel/netlabel_addrlist.c
244
entry = netlbl_af4list_search_exact(addr, mask, head);
net/netlabel/netlabel_addrlist.c
281
struct list_head *head)
net/netlabel/netlabel_addrlist.c
285
entry = netlbl_af6list_search_exact(addr, mask, head);
net/netlabel/netlabel_addrlist.c
47
struct list_head *head)
net/netlabel/netlabel_addrlist.c
51
list_for_each_entry_rcu(iter, head, list)
net/netlabel/netlabel_addrlist.c
72
struct list_head *head)
net/netlabel/netlabel_addrlist.c
76
list_for_each_entry_rcu(iter, head, list)
net/netlabel/netlabel_addrlist.c
97
struct list_head *head)
net/netlabel/netlabel_addrlist.h
101
struct list_head *head);
net/netlabel/netlabel_addrlist.h
104
struct list_head *head);
net/netlabel/netlabel_addrlist.h
107
struct list_head *head);
net/netlabel/netlabel_addrlist.h
149
#define netlbl_af6list_foreach(iter, head) \
net/netlabel/netlabel_addrlist.h
150
for (iter = __af6list_valid((head)->next, head); \
net/netlabel/netlabel_addrlist.h
151
&iter->list != (head); \
net/netlabel/netlabel_addrlist.h
152
iter = __af6list_valid(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
154
#define netlbl_af6list_foreach_rcu(iter, head) \
net/netlabel/netlabel_addrlist.h
155
for (iter = __af6list_valid_rcu((head)->next, head); \
net/netlabel/netlabel_addrlist.h
156
&iter->list != (head); \
net/netlabel/netlabel_addrlist.h
157
iter = __af6list_valid_rcu(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
159
#define netlbl_af6list_foreach_safe(iter, tmp, head) \
net/netlabel/netlabel_addrlist.h
160
for (iter = __af6list_valid((head)->next, head), \
net/netlabel/netlabel_addrlist.h
161
tmp = __af6list_valid(iter->list.next, head); \
net/netlabel/netlabel_addrlist.h
162
&iter->list != (head); \
net/netlabel/netlabel_addrlist.h
163
iter = tmp, tmp = __af6list_valid(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
166
struct list_head *head);
net/netlabel/netlabel_addrlist.h
169
struct list_head *head);
net/netlabel/netlabel_addrlist.h
172
struct list_head *head);
net/netlabel/netlabel_addrlist.h
175
struct list_head *head);
net/netlabel/netlabel_addrlist.h
82
#define netlbl_af4list_foreach(iter, head) \
net/netlabel/netlabel_addrlist.h
83
for (iter = __af4list_valid((head)->next, head); \
net/netlabel/netlabel_addrlist.h
84
&iter->list != (head); \
net/netlabel/netlabel_addrlist.h
85
iter = __af4list_valid(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
87
#define netlbl_af4list_foreach_rcu(iter, head) \
net/netlabel/netlabel_addrlist.h
88
for (iter = __af4list_valid_rcu((head)->next, head); \
net/netlabel/netlabel_addrlist.h
89
&iter->list != (head); \
net/netlabel/netlabel_addrlist.h
90
iter = __af4list_valid_rcu(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
92
#define netlbl_af4list_foreach_safe(iter, tmp, head) \
net/netlabel/netlabel_addrlist.h
93
for (iter = __af4list_valid((head)->next, head), \
net/netlabel/netlabel_addrlist.h
94
tmp = __af4list_valid(iter->list.next, head); \
net/netlabel/netlabel_addrlist.h
95
&iter->list != (head); \
net/netlabel/netlabel_addrlist.h
96
iter = tmp, tmp = __af4list_valid(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
99
struct list_head *head);
net/netlink/af_netlink.c
1288
if (is_vmalloc_addr(skb->head) || delta * 2 < skb->truesize)
net/netlink/af_netlink.c
297
if (is_vmalloc_addr(skb->head))
net/netlink/af_netlink.c
374
if (is_vmalloc_addr(skb->head)) {
net/netlink/af_netlink.c
377
vfree_atomic(skb->head);
net/netlink/af_netlink.c
379
skb->head = NULL;
net/netlink/af_netlink.c
703
static void deferred_put_nlk_sk(struct rcu_head *head)
net/netlink/af_netlink.c
705
struct netlink_sock *nlk = container_of(head, struct netlink_sock, rcu);
net/nfc/llcp.h
33
struct hlist_head head;
net/nfc/llcp_commands.c
180
void nfc_llcp_free_sdp_tlv_list(struct hlist_head *head)
net/nfc/llcp_commands.c
185
hlist_for_each_entry_safe(sdp, n, head, node) {
net/nfc/llcp_core.c
126
sk_for_each_safe(sk, tmp, &local->raw_sockets.head) {
net/nfc/llcp_core.c
216
sk_for_each(sk, &local->sockets.head) {
net/nfc/llcp_core.c
28
sk_add_node(sk, &l->head);
net/nfc/llcp_core.c
379
sk_for_each(sk, &local->sockets.head) {
net/nfc/llcp_core.c
730
sk_for_each(sk, &local->raw_sockets.head) {
net/nfc/llcp_core.c
77
sk_for_each_safe(sk, tmp, &local->sockets.head) {
net/nfc/llcp_core.c
826
sk_for_each(sk, &local->connecting_sockets.head) {
net/nfc/nfc.h
36
struct hlist_head head;
net/nfc/nfc.h
59
void nfc_llcp_free_sdp_tlv_list(struct hlist_head *head);
net/nfc/rawsock.c
26
sk_add_node(sk, &l->head);
net/nfc/rawsock.c
379
sk_for_each(sk, &raw_sk_list.head) {
net/openvswitch/actions.c
64
fifo->head = 0;
net/openvswitch/actions.c
70
return (fifo->head == fifo->tail);
net/openvswitch/actions.c
83
if (fifo->head >= DEFERRED_ACTION_FIFO_SIZE - 1)
net/openvswitch/actions.c
86
return &fifo->fifo[fifo->head++];
net/openvswitch/conntrack.c
1623
struct hlist_head *head = &info->limits[i];
net/openvswitch/conntrack.c
1627
hlist_for_each_entry_safe(ct_limit, next, head, hlist_node)
net/openvswitch/conntrack.c
1823
struct hlist_head *head;
net/openvswitch/conntrack.c
1832
head = &info->limits[i];
net/openvswitch/conntrack.c
1833
hlist_for_each_entry_rcu(ct_limit, head, hlist_node) {
net/openvswitch/conntrack.c
883
struct hlist_head *head;
net/openvswitch/conntrack.c
885
head = ct_limit_hash_bucket(info, new_ct_limit->zone);
net/openvswitch/conntrack.c
886
hlist_for_each_entry_rcu(ct_limit, head, hlist_node) {
net/openvswitch/conntrack.c
895
hlist_add_head_rcu(&new_ct_limit->hlist_node, head);
net/openvswitch/conntrack.c
902
struct hlist_head *head;
net/openvswitch/conntrack.c
905
head = ct_limit_hash_bucket(info, zone);
net/openvswitch/conntrack.c
906
hlist_for_each_entry_safe(ct_limit, n, head, hlist_node) {
net/openvswitch/conntrack.c
919
struct hlist_head *head;
net/openvswitch/conntrack.c
921
head = ct_limit_hash_bucket(info, zone);
net/openvswitch/conntrack.c
922
hlist_for_each_entry_rcu(ct_limit, head, hlist_node) {
net/openvswitch/datapath.c
187
struct hlist_head *head;
net/openvswitch/datapath.c
189
head = vport_hash_bucket(dp, port_no);
net/openvswitch/datapath.c
190
hlist_for_each_entry_rcu(vport, head, dp_hash_node,
net/openvswitch/datapath.c
206
struct hlist_head *head = vport_hash_bucket(dp, vport->port_no);
net/openvswitch/datapath.c
208
hlist_add_head_rcu(&vport->dp_hash_node, head);
net/openvswitch/datapath.c
2679
struct list_head *head)
net/openvswitch/datapath.c
2695
list_add(&vport->detach_list, head);
net/openvswitch/datapath.c
2707
LIST_HEAD(head);
net/openvswitch/datapath.c
2718
list_vports_from_net(net, dnet, &head);
net/openvswitch/datapath.c
2722
list_for_each_entry_safe(vport, vport_next, &head, detach_list) {
net/openvswitch/datapath.h
208
int head;
net/openvswitch/flow_netlink.c
2415
static void __ovs_nla_free_flow_actions(struct rcu_head *head)
net/openvswitch/flow_netlink.c
2417
ovs_nla_free_flow_actions(container_of(head, struct sw_flow_actions, rcu));
net/openvswitch/flow_table.c
481
struct hlist_head *head = &ti->buckets[i];
net/openvswitch/flow_table.c
485
hlist_for_each_entry_safe(flow, n, head,
net/openvswitch/flow_table.c
527
struct hlist_head *head;
net/openvswitch/flow_table.c
534
head = &ti->buckets[*bucket];
net/openvswitch/flow_table.c
535
hlist_for_each_entry_rcu(flow, head, flow_table.node[ver]) {
net/openvswitch/flow_table.c
559
struct hlist_head *head;
net/openvswitch/flow_table.c
561
head = find_bucket(ti, flow->flow_table.hash);
net/openvswitch/flow_table.c
562
hlist_add_head_rcu(&flow->flow_table.node[ti->node_ver], head);
net/openvswitch/flow_table.c
568
struct hlist_head *head;
net/openvswitch/flow_table.c
570
head = find_bucket(ti, flow->ufid_table.hash);
net/openvswitch/flow_table.c
571
hlist_add_head_rcu(&flow->ufid_table.node[ti->node_ver], head);
net/openvswitch/flow_table.c
586
struct hlist_head *head = &old->buckets[i];
net/openvswitch/flow_table.c
589
hlist_for_each_entry_rcu(flow, head,
net/openvswitch/flow_table.c
594
hlist_for_each_entry_rcu(flow, head,
net/openvswitch/flow_table.c
702
struct hlist_head *head;
net/openvswitch/flow_table.c
708
head = find_bucket(ti, hash);
net/openvswitch/flow_table.c
711
hlist_for_each_entry_rcu(flow, head, flow_table.node[ti->node_ver],
net/openvswitch/flow_table.c
920
struct hlist_head *head;
net/openvswitch/flow_table.c
924
head = find_bucket(ti, hash);
net/openvswitch/flow_table.c
925
hlist_for_each_entry_rcu(flow, head, ufid_table.node[ti->node_ver],
net/packet/af_packet.c
1116
po->rx_ring.head, status);
net/packet/af_packet.c
1188
unsigned int previous = rb->head ? rb->head - 1 : rb->frame_max;
net/packet/af_packet.c
1194
buff->head = buff->head != buff->frame_max ? buff->head+1 : 0;
net/packet/af_packet.c
1248
idx = READ_ONCE(po->rx_ring.head);
net/packet/af_packet.c
2355
slot_id = po->rx_ring.head;
net/packet/af_packet.c
248
struct sk_buff *next, *head = NULL, *tail;
net/packet/af_packet.c
259
if (!head)
net/packet/af_packet.c
260
head = skb;
net/packet/af_packet.c
268
return head;
net/packet/af_packet.c
4541
rb->head = 0;
net/packet/af_packet.c
537
return packet_lookup_frame(po, rb, rb->head, status);
net/packet/internal.h
57
unsigned int head;
net/psp/psp_sock.c
106
static void psp_assoc_free_queue(struct rcu_head *head)
net/psp/psp_sock.c
108
struct psp_assoc *pas = container_of(head, struct psp_assoc, rcu);
net/rds/connection.c
170
struct hlist_head *head = rds_conn_bucket(laddr, faddr);
net/rds/connection.c
178
conn = rds_conn_lookup(net, head, laddr, faddr, trans, tos, dev_if);
net/rds/connection.c
321
found = rds_conn_lookup(net, head, laddr, faddr, trans,
net/rds/connection.c
342
hlist_add_head_rcu(&conn->c_hash_node, head);
net/rds/connection.c
571
struct hlist_head *head;
net/rds/connection.c
587
for (i = 0, head = rds_conn_hash; i < ARRAY_SIZE(rds_conn_hash);
net/rds/connection.c
588
i++, head++) {
net/rds/connection.c
589
hlist_for_each_entry_rcu(conn, head, c_hash_node) {
net/rds/connection.c
691
struct hlist_head *head;
net/rds/connection.c
700
for (i = 0, head = rds_conn_hash; i < ARRAY_SIZE(rds_conn_hash);
net/rds/connection.c
701
i++, head++) {
net/rds/connection.c
702
hlist_for_each_entry_rcu(conn, head, c_hash_node) {
net/rds/connection.c
729
struct hlist_head *head;
net/rds/connection.c
738
for (i = 0, head = rds_conn_hash; i < ARRAY_SIZE(rds_conn_hash);
net/rds/connection.c
739
i++, head++) {
net/rds/connection.c
740
hlist_for_each_entry_rcu(conn, head, c_hash_node) {
net/rds/connection.c
84
struct hlist_head *head,
net/rds/connection.c
92
hlist_for_each_entry_rcu(conn, head, c_hash_node) {
net/rds/ib_recv.c
104
struct rds_ib_cache_head *head;
net/rds/ib_recv.c
112
head = per_cpu_ptr(cache->percpu, cpu);
net/rds/ib_recv.c
113
head->first = NULL;
net/rds/ib_recv.c
114
head->count = 0;
net/rds/ib_recv.c
139
struct rds_ib_cache_head *head;
net/rds/ib_recv.c
143
head = per_cpu_ptr(cache->percpu, cpu);
net/rds/ib_recv.c
144
if (head->first) {
net/rds/ib_recv.c
145
list_splice_entire_tail(head->first, caller_list);
net/rds/ib_recv.c
146
head->first = NULL;
net/rds/ib_recv.c
521
struct list_head *head = cache->ready;
net/rds/ib_recv.c
523
if (head) {
net/rds/ib_recv.c
524
if (!list_empty(head)) {
net/rds/ib_recv.c
525
cache->ready = head->next;
net/rds/ib_recv.c
526
list_del_init(head);
net/rds/ib_recv.c
531
return head;
net/rds/message.c
101
struct list_head *head;
net/rds/message.c
107
head = &q->zcookie_head;
net/rds/message.c
108
if (!list_empty(head)) {
net/rds/message.c
109
info = list_first_entry(head, struct rds_msg_zcopy_info,
net/rxrpc/call_accept.c
183
unsigned int size = RXRPC_BACKLOG_MAX, head, tail;
net/rxrpc/call_accept.c
195
head = b->peer_backlog_head;
net/rxrpc/call_accept.c
197
while (CIRC_CNT(head, tail, size) > 0) {
net/rxrpc/call_accept.c
204
head = b->conn_backlog_head;
net/rxrpc/call_accept.c
206
while (CIRC_CNT(head, tail, size) > 0) {
net/rxrpc/call_accept.c
218
head = b->call_backlog_head;
net/rxrpc/call_accept.c
220
while (CIRC_CNT(head, tail, size) > 0) {
net/rxrpc/call_accept.c
45
unsigned int head, tail, call_head, call_tail;
net/rxrpc/call_accept.c
68
head = b->peer_backlog_head;
net/rxrpc/call_accept.c
70
if (CIRC_CNT(head, tail, size) < max) {
net/rxrpc/call_accept.c
76
b->peer_backlog[head] = peer;
net/rxrpc/call_accept.c
78
(head + 1) & (size - 1));
net/rxrpc/call_accept.c
81
head = b->conn_backlog_head;
net/rxrpc/call_accept.c
83
if (CIRC_CNT(head, tail, size) < max) {
net/rxrpc/call_accept.c
89
b->conn_backlog[head] = conn;
net/rxrpc/call_accept.c
91
(head + 1) & (size - 1));
net/sched/act_api.c
1041
list_for_each_entry(a, &act_base, head) {
net/sched/act_api.c
1047
list_add_tail(&act->head, &act_base);
net/sched/act_api.c
1069
list_for_each_entry(a, &act_base, head) {
net/sched/act_api.c
1071
list_del(&act->head);
net/sched/act_api.c
1093
list_for_each_entry(a, &act_base, head) {
net/sched/act_api.c
1112
list_for_each_entry(a, &act_base, head) {
net/sched/act_ct.c
892
static void tcf_ct_params_free_rcu(struct rcu_head *head)
net/sched/act_ct.c
896
params = container_of(head, struct tcf_ct_params, rcu);
net/sched/act_gate.c
35
static void tcf_gate_params_free_rcu(struct rcu_head *head);
net/sched/act_gate.c
555
static void tcf_gate_params_free_rcu(struct rcu_head *head)
net/sched/act_gate.c
557
struct tcf_gate_params *p = container_of(head, struct tcf_gate_params, rcu);
net/sched/act_ife.c
418
static void tcf_ife_cleanup_params(struct rcu_head *head)
net/sched/act_ife.c
420
struct tcf_ife_params *p = container_of(head, struct tcf_ife_params,
net/sched/act_pedit.c
145
static void tcf_pedit_cleanup_rcu(struct rcu_head *head)
net/sched/act_pedit.c
148
container_of(head, struct tcf_pedit_parms, rcu);
net/sched/act_tunnel_key.c
221
const struct nlattr *attr, *head = nla_data(nla);
net/sched/act_tunnel_key.c
223
err = nla_validate_deprecated(head, len, TCA_TUNNEL_KEY_ENC_OPTS_MAX,
net/sched/act_tunnel_key.c
228
nla_for_each_attr(attr, head, len, rem) {
net/sched/cls_api.c
236
list_for_each_entry(t, &tcf_proto_base, head) {
net/sched/cls_api.c
286
list_for_each_entry(t, &tcf_proto_base, head)
net/sched/cls_api.c
290
list_add_tail(&ops->head, &tcf_proto_base);
net/sched/cls_api.c
312
list_for_each_entry(t, &tcf_proto_base, head) {
net/sched/cls_api.c
314
list_del(&t->head);
net/sched/cls_basic.c
111
struct basic_head *head = rtnl_dereference(tp->root);
net/sched/cls_basic.c
114
list_for_each_entry_safe(f, n, &head->flist, link) {
net/sched/cls_basic.c
117
idr_remove(&head->handle_idr, f->handle);
net/sched/cls_basic.c
123
idr_destroy(&head->handle_idr);
net/sched/cls_basic.c
124
kfree_rcu(head, rcu);
net/sched/cls_basic.c
130
struct basic_head *head = rtnl_dereference(tp->root);
net/sched/cls_basic.c
135
idr_remove(&head->handle_idr, f->handle);
net/sched/cls_basic.c
138
*last = list_empty(&head->flist);
net/sched/cls_basic.c
178
struct basic_head *head = rtnl_dereference(tp->root);
net/sched/cls_basic.c
206
err = idr_alloc_u32(&head->handle_idr, fnew, &handle,
net/sched/cls_basic.c
209
err = idr_alloc_u32(&head->handle_idr, fnew, &handle,
net/sched/cls_basic.c
225
idr_remove(&head->handle_idr, fnew->handle);
net/sched/cls_basic.c
232
idr_replace(&head->handle_idr, fnew, fnew->handle);
net/sched/cls_basic.c
238
list_add_rcu(&fnew->link, &head->flist);
net/sched/cls_basic.c
252
struct basic_head *head = rtnl_dereference(tp->root);
net/sched/cls_basic.c
255
list_for_each_entry(f, &head->flist, link) {
net/sched/cls_basic.c
45
struct basic_head *head = rcu_dereference_bh(tp->root);
net/sched/cls_basic.c
48
list_for_each_entry_rcu(f, &head->flist, link) {
net/sched/cls_basic.c
64
struct basic_head *head = rtnl_dereference(tp->root);
net/sched/cls_basic.c
67
list_for_each_entry(f, &head->flist, link) {
net/sched/cls_basic.c
78
struct basic_head *head;
net/sched/cls_basic.c
80
head = kzalloc_obj(*head);
net/sched/cls_basic.c
81
if (head == NULL)
net/sched/cls_basic.c
83
INIT_LIST_HEAD(&head->flist);
net/sched/cls_basic.c
84
idr_init(&head->handle_idr);
net/sched/cls_basic.c
85
rcu_assign_pointer(tp->root, head);
net/sched/cls_bpf.c
243
struct cls_bpf_head *head;
net/sched/cls_bpf.c
245
head = kzalloc_obj(*head);
net/sched/cls_bpf.c
246
if (head == NULL)
net/sched/cls_bpf.c
249
INIT_LIST_HEAD_RCU(&head->plist);
net/sched/cls_bpf.c
250
idr_init(&head->handle_idr);
net/sched/cls_bpf.c
251
rcu_assign_pointer(tp->root, head);
net/sched/cls_bpf.c
289
struct cls_bpf_head *head = rtnl_dereference(tp->root);
net/sched/cls_bpf.c
291
idr_remove(&head->handle_idr, prog->handle);
net/sched/cls_bpf.c
304
struct cls_bpf_head *head = rtnl_dereference(tp->root);
net/sched/cls_bpf.c
307
*last = list_empty(&head->plist);
net/sched/cls_bpf.c
314
struct cls_bpf_head *head = rtnl_dereference(tp->root);
net/sched/cls_bpf.c
317
list_for_each_entry_safe(prog, tmp, &head->plist, link)
net/sched/cls_bpf.c
320
idr_destroy(&head->handle_idr);
net/sched/cls_bpf.c
321
kfree_rcu(head, rcu);
net/sched/cls_bpf.c
326
struct cls_bpf_head *head = rtnl_dereference(tp->root);
net/sched/cls_bpf.c
329
list_for_each_entry(prog, &head->plist, link) {
net/sched/cls_bpf.c
413
struct cls_bpf_head *head = rtnl_dereference(tp->root);
net/sched/cls_bpf.c
447
ret = idr_alloc_u32(&head->handle_idr, prog, &handle,
net/sched/cls_bpf.c
450
ret = idr_alloc_u32(&head->handle_idr, prog, &handle,
net/sched/cls_bpf.c
513
idr_replace(&head->handle_idr, prog, handle);
net/sched/cls_bpf.c
519
list_add_rcu(&prog->link, &head->plist);
net/sched/cls_bpf.c
531
idr_remove(&head->handle_idr, prog->handle);
net/sched/cls_bpf.c
641
struct cls_bpf_head *head = rtnl_dereference(tp->root);
net/sched/cls_bpf.c
644
list_for_each_entry(prog, &head->plist, link) {
net/sched/cls_bpf.c
653
struct cls_bpf_head *head = rtnl_dereference(tp->root);
net/sched/cls_bpf.c
659
list_for_each_entry(prog, &head->plist, link) {
net/sched/cls_bpf.c
85
struct cls_bpf_head *head = rcu_dereference_bh(tp->root);
net/sched/cls_bpf.c
90
list_for_each_entry_rcu(prog, &head->plist, link) {
net/sched/cls_cgroup.c
123
if (head) {
net/sched/cls_cgroup.c
124
tcf_exts_get_net(&head->exts);
net/sched/cls_cgroup.c
125
tcf_queue_work(&head->rwork, cls_cgroup_destroy_work);
net/sched/cls_cgroup.c
137
struct cls_cgroup_head *head = rtnl_dereference(tp->root);
net/sched/cls_cgroup.c
140
if (head) {
net/sched/cls_cgroup.c
141
if (tcf_exts_get_net(&head->exts))
net/sched/cls_cgroup.c
142
tcf_queue_work(&head->rwork, cls_cgroup_destroy_work);
net/sched/cls_cgroup.c
144
__cls_cgroup_destroy(head);
net/sched/cls_cgroup.c
157
struct cls_cgroup_head *head = rtnl_dereference(tp->root);
net/sched/cls_cgroup.c
162
if (!head)
net/sched/cls_cgroup.c
164
if (arg->fn(tp, head, arg) < 0) {
net/sched/cls_cgroup.c
175
struct cls_cgroup_head *head = rtnl_dereference(tp->root);
net/sched/cls_cgroup.c
178
t->tcm_handle = head->handle;
net/sched/cls_cgroup.c
184
if (tcf_exts_dump(skb, &head->exts) < 0 ||
net/sched/cls_cgroup.c
185
tcf_em_tree_dump(skb, &head->ematches, TCA_CGROUP_EMATCHES) < 0)
net/sched/cls_cgroup.c
190
if (tcf_exts_dump_stats(skb, &head->exts) < 0)
net/sched/cls_cgroup.c
30
struct cls_cgroup_head *head = rcu_dereference_bh(tp->root);
net/sched/cls_cgroup.c
33
if (unlikely(!head))
net/sched/cls_cgroup.c
37
if (!tcf_em_tree_match(skb, &head->ematches, NULL))
net/sched/cls_cgroup.c
43
return tcf_exts_exec(skb, &head->exts, res);
net/sched/cls_cgroup.c
60
static void __cls_cgroup_destroy(struct cls_cgroup_head *head)
net/sched/cls_cgroup.c
62
tcf_exts_destroy(&head->exts);
net/sched/cls_cgroup.c
63
tcf_em_tree_destroy(&head->ematches);
net/sched/cls_cgroup.c
64
tcf_exts_put_net(&head->exts);
net/sched/cls_cgroup.c
65
kfree(head);
net/sched/cls_cgroup.c
70
struct cls_cgroup_head *head = container_of(to_rcu_work(work),
net/sched/cls_cgroup.c
74
__cls_cgroup_destroy(head);
net/sched/cls_cgroup.c
85
struct cls_cgroup_head *head = rtnl_dereference(tp->root);
net/sched/cls_cgroup.c
92
if (!head && !handle)
net/sched/cls_cgroup.c
95
if (head && handle != head->handle)
net/sched/cls_cgroup.c
98
new = kzalloc_obj(*head);
net/sched/cls_flow.c
300
struct flow_head *head = rcu_dereference_bh(tp->root);
net/sched/cls_flow.c
307
list_for_each_entry_rcu(f, &head->filters, list) {
net/sched/cls_flow.c
396
struct flow_head *head = rtnl_dereference(tp->root);
net/sched/cls_flow.c
557
list_add_tail_rcu(&fnew->list, &head->filters);
net/sched/cls_flow.c
580
struct flow_head *head = rtnl_dereference(tp->root);
net/sched/cls_flow.c
586
*last = list_empty(&head->filters);
net/sched/cls_flow.c
592
struct flow_head *head;
net/sched/cls_flow.c
594
head = kzalloc_obj(*head);
net/sched/cls_flow.c
595
if (head == NULL)
net/sched/cls_flow.c
597
INIT_LIST_HEAD(&head->filters);
net/sched/cls_flow.c
598
rcu_assign_pointer(tp->root, head);
net/sched/cls_flow.c
605
struct flow_head *head = rtnl_dereference(tp->root);
net/sched/cls_flow.c
608
list_for_each_entry_safe(f, next, &head->filters, list) {
net/sched/cls_flow.c
615
kfree_rcu(head, rcu);
net/sched/cls_flow.c
620
struct flow_head *head = rtnl_dereference(tp->root);
net/sched/cls_flow.c
623
list_for_each_entry(f, &head->filters, list)
net/sched/cls_flow.c
693
struct flow_head *head = rtnl_dereference(tp->root);
net/sched/cls_flow.c
696
list_for_each_entry(f, &head->filters, list) {
net/sched/cls_flower.c
2234
static struct fl_flow_mask *fl_create_new_mask(struct cls_fl_head *head,
net/sched/cls_flower.c
2261
err = rhashtable_replace_fast(&head->ht, &mask->ht_node,
net/sched/cls_flower.c
2266
spin_lock(&head->masks_lock);
net/sched/cls_flower.c
2267
list_add_tail_rcu(&newmask->list, &head->masks);
net/sched/cls_flower.c
2268
spin_unlock(&head->masks_lock);
net/sched/cls_flower.c
2280
static int fl_check_assign_mask(struct cls_fl_head *head,
net/sched/cls_flower.c
2294
fnew->mask = rhashtable_lookup_get_insert_fast(&head->ht,
net/sched/cls_flower.c
2305
newmask = fl_create_new_mask(head, mask);
net/sched/cls_flower.c
2325
rhashtable_remove_fast(&head->ht, &mask->ht_node,
net/sched/cls_flower.c
2363
struct cls_fl_head *head = fl_head_dereference(tp);
net/sched/cls_flower.c
2423
err = idr_alloc_u32(&head->handle_idr, NULL, &handle,
net/sched/cls_flower.c
2426
err = idr_alloc_u32(&head->handle_idr, NULL, &handle,
net/sched/cls_flower.c
2489
err = fl_check_assign_mask(head, fnew, fold, mask);
net/sched/cls_flower.c
2543
idr_replace(&head->handle_idr, fnew, fnew->handle);
net/sched/cls_flower.c
2549
fl_mask_put(head, fold->mask);
net/sched/cls_flower.c
2559
idr_replace(&head->handle_idr, fnew, fnew->handle);
net/sched/cls_flower.c
2583
fl_mask_put(head, fnew->mask);
net/sched/cls_flower.c
2597
idr_remove(&head->handle_idr, fnew->handle);
net/sched/cls_flower.c
2614
struct cls_fl_head *head = fl_head_dereference(tp);
net/sched/cls_flower.c
2620
*last = list_empty(&head->masks);
net/sched/cls_flower.c
2629
struct cls_fl_head *head = fl_head_dereference(tp);
net/sched/cls_flower.c
2636
idr_for_each_entry_continue_ul(&head->handle_idr, f, tmp, id) {
net/sched/cls_flower.c
2659
struct cls_fl_head *head = fl_head_dereference(tp);
net/sched/cls_flower.c
2662
if (list_empty(&head->hw_filters)) {
net/sched/cls_flower.c
2668
f = list_entry(&head->hw_filters, struct cls_fl_filter,
net/sched/cls_flower.c
2670
list_for_each_entry_continue(f, &head->hw_filters, hw_list) {
net/sched/cls_flower.c
2748
struct cls_fl_head *head = fl_head_dereference(tp);
net/sched/cls_flower.c
2751
list_add(&f->hw_list, &head->hw_filters);
net/sched/cls_flower.c
328
struct cls_fl_head *head = rcu_dereference_bh(tp->root);
net/sched/cls_flower.c
335
list_for_each_entry_rcu(mask, &head->masks, list) {
net/sched/cls_flower.c
364
struct cls_fl_head *head;
net/sched/cls_flower.c
366
head = kzalloc_obj(*head);
net/sched/cls_flower.c
367
if (!head)
net/sched/cls_flower.c
370
spin_lock_init(&head->masks_lock);
net/sched/cls_flower.c
371
INIT_LIST_HEAD_RCU(&head->masks);
net/sched/cls_flower.c
372
INIT_LIST_HEAD(&head->hw_filters);
net/sched/cls_flower.c
373
rcu_assign_pointer(tp->root, head);
net/sched/cls_flower.c
374
idr_init(&head->handle_idr);
net/sched/cls_flower.c
376
return rhashtable_init(&head->ht, &mask_ht_params);
net/sched/cls_flower.c
3823
struct cls_fl_head *head = fl_head_dereference(tp);
net/sched/cls_flower.c
3826
tp->deleting = idr_is_empty(&head->handle_idr);
net/sched/cls_flower.c
405
static bool fl_mask_put(struct cls_fl_head *head, struct fl_flow_mask *mask)
net/sched/cls_flower.c
410
rhashtable_remove_fast(&head->ht, &mask->ht_node, mask_ht_params);
net/sched/cls_flower.c
412
spin_lock(&head->masks_lock);
net/sched/cls_flower.c
414
spin_unlock(&head->masks_lock);
net/sched/cls_flower.c
536
static struct cls_fl_filter *__fl_get(struct cls_fl_head *head, u32 handle)
net/sched/cls_flower.c
541
f = idr_find(&head->handle_idr, handle);
net/sched/cls_flower.c
551
struct cls_fl_head *head = rcu_dereference_bh(tp->root);
net/sched/cls_flower.c
554
f = idr_find(&head->handle_idr, handle);
net/sched/cls_flower.c
562
struct cls_fl_head *head = fl_head_dereference(tp);
net/sched/cls_flower.c
575
idr_remove(&head->handle_idr, f->handle);
net/sched/cls_flower.c
579
*last = fl_mask_put(head, f->mask);
net/sched/cls_flower.c
590
struct cls_fl_head *head = container_of(to_rcu_work(work),
net/sched/cls_flower.c
594
rhashtable_destroy(&head->ht);
net/sched/cls_flower.c
595
kfree(head);
net/sched/cls_flower.c
602
struct cls_fl_head *head = fl_head_dereference(tp);
net/sched/cls_flower.c
607
list_for_each_entry_safe(mask, next_mask, &head->masks, list) {
net/sched/cls_flower.c
614
idr_destroy(&head->handle_idr);
net/sched/cls_flower.c
617
tcf_queue_work(&head->rwork, fl_destroy_sleepable);
net/sched/cls_flower.c
629
struct cls_fl_head *head = fl_head_dereference(tp);
net/sched/cls_flower.c
631
return __fl_get(head, handle);
net/sched/cls_fw.c
135
struct fw_head *head = rtnl_dereference(tp->root);
net/sched/cls_fw.c
139
if (head == NULL)
net/sched/cls_fw.c
143
while ((f = rtnl_dereference(head->ht[h])) != NULL) {
net/sched/cls_fw.c
144
RCU_INIT_POINTER(head->ht[h],
net/sched/cls_fw.c
153
kfree_rcu(head, rcu);
net/sched/cls_fw.c
159
struct fw_head *head = rtnl_dereference(tp->root);
net/sched/cls_fw.c
166
if (head == NULL || f == NULL)
net/sched/cls_fw.c
169
fp = &head->ht[fw_hash(f->id)];
net/sched/cls_fw.c
185
if (rcu_access_pointer(head->ht[h])) {
net/sched/cls_fw.c
206
struct fw_head *head = rtnl_dereference(tp->root);
net/sched/cls_fw.c
226
if (mask != head->mask)
net/sched/cls_fw.c
228
} else if (head->mask != 0xFFFFFFFF)
net/sched/cls_fw.c
244
struct fw_head *head = rtnl_dereference(tp->root);
net/sched/cls_fw.c
297
fp = &head->ht[fw_hash(fnew->id)];
net/sched/cls_fw.c
316
if (!head) {
net/sched/cls_fw.c
321
head = kzalloc_obj(*head);
net/sched/cls_fw.c
322
if (!head)
net/sched/cls_fw.c
324
head->mask = mask;
net/sched/cls_fw.c
326
rcu_assign_pointer(tp->root, head);
net/sched/cls_fw.c
343
RCU_INIT_POINTER(f->next, head->ht[fw_hash(handle)]);
net/sched/cls_fw.c
344
rcu_assign_pointer(head->ht[fw_hash(handle)], f);
net/sched/cls_fw.c
358
struct fw_head *head = rtnl_dereference(tp->root);
net/sched/cls_fw.c
361
if (head == NULL)
net/sched/cls_fw.c
370
for (f = rtnl_dereference(head->ht[h]); f;
net/sched/cls_fw.c
381
struct fw_head *head = rtnl_dereference(tp->root);
net/sched/cls_fw.c
406
if (head->mask != 0xFFFFFFFF &&
net/sched/cls_fw.c
407
nla_put_u32(skb, TCA_FW_MASK, head->mask))
net/sched/cls_fw.c
55
struct fw_head *head = rcu_dereference_bh(tp->root);
net/sched/cls_fw.c
60
if (head != NULL) {
net/sched/cls_fw.c
61
id &= head->mask;
net/sched/cls_fw.c
63
for (f = rcu_dereference_bh(head->ht[fw_hash(id)]); f;
net/sched/cls_fw.c
93
struct fw_head *head = rtnl_dereference(tp->root);
net/sched/cls_fw.c
96
if (head == NULL)
net/sched/cls_fw.c
99
f = rtnl_dereference(head->ht[fw_hash(handle)]);
net/sched/cls_matchall.c
102
err = tc_setup_offload_action(&cls_mall.rule->action, &head->exts,
net/sched/cls_matchall.c
106
mall_destroy_hw_filter(tp, head, cookie, NULL);
net/sched/cls_matchall.c
112
skip_sw, &head->flags, &head->in_hw_count, true);
net/sched/cls_matchall.c
117
mall_destroy_hw_filter(tp, head, cookie, NULL);
net/sched/cls_matchall.c
121
if (skip_sw && !(head->flags & TCA_CLS_FLAGS_IN_HW))
net/sched/cls_matchall.c
130
struct cls_mall_head *head = rtnl_dereference(tp->root);
net/sched/cls_matchall.c
132
if (!head)
net/sched/cls_matchall.c
135
tcf_unbind_filter(tp, &head->res);
net/sched/cls_matchall.c
137
if (!tc_skip_hw(head->flags))
net/sched/cls_matchall.c
138
mall_destroy_hw_filter(tp, head, (unsigned long) head, extack);
net/sched/cls_matchall.c
140
if (tcf_exts_get_net(&head->exts))
net/sched/cls_matchall.c
141
tcf_queue_work(&head->rwork, mall_destroy_work);
net/sched/cls_matchall.c
143
__mall_destroy(head);
net/sched/cls_matchall.c
148
struct cls_mall_head *head = rtnl_dereference(tp->root);
net/sched/cls_matchall.c
150
if (head && head->handle == handle)
net/sched/cls_matchall.c
151
return head;
net/sched/cls_matchall.c
168
struct cls_mall_head *head = rtnl_dereference(tp->root);
net/sched/cls_matchall.c
178
if (head)
net/sched/cls_matchall.c
233
*arg = head;
net/sched/cls_matchall.c
252
struct cls_mall_head *head = rtnl_dereference(tp->root);
net/sched/cls_matchall.c
254
head->deleting = true;
net/sched/cls_matchall.c
262
struct cls_mall_head *head = rtnl_dereference(tp->root);
net/sched/cls_matchall.c
267
if (!head || head->deleting)
net/sched/cls_matchall.c
269
if (arg->fn(tp, head, arg) < 0)
net/sched/cls_matchall.c
278
struct cls_mall_head *head = rtnl_dereference(tp->root);
net/sched/cls_matchall.c
283
if (tc_skip_hw(head->flags))
net/sched/cls_matchall.c
286
cls_mall.rule = flow_rule_alloc(tcf_exts_num_actions(&head->exts));
net/sched/cls_matchall.c
290
tc_cls_common_offload_init(&cls_mall.common, tp, head->flags, extack);
net/sched/cls_matchall.c
293
cls_mall.cookie = (unsigned long)head;
net/sched/cls_matchall.c
295
err = tc_setup_offload_action(&cls_mall.rule->action, &head->exts,
net/sched/cls_matchall.c
300
return add && tc_skip_sw(head->flags) ? err : 0;
net/sched/cls_matchall.c
304
&cls_mall, cb_priv, &head->flags,
net/sched/cls_matchall.c
305
&head->in_hw_count);
net/sched/cls_matchall.c
313
struct cls_mall_head *head,
net/sched/cls_matchall.c
319
tc_cls_common_offload_init(&cls_mall.common, tp, head->flags, NULL);
net/sched/cls_matchall.c
32
struct cls_mall_head *head = rcu_dereference_bh(tp->root);
net/sched/cls_matchall.c
325
tcf_exts_hw_stats_update(&head->exts, &cls_mall.stats, cls_mall.use_act_stats);
net/sched/cls_matchall.c
332
struct cls_mall_head *head = fh;
net/sched/cls_matchall.c
336
if (!head)
net/sched/cls_matchall.c
339
if (!tc_skip_hw(head->flags))
net/sched/cls_matchall.c
34
if (unlikely(!head))
net/sched/cls_matchall.c
340
mall_stats_hw_filter(tp, head, (unsigned long)head);
net/sched/cls_matchall.c
342
t->tcm_handle = head->handle;
net/sched/cls_matchall.c
348
if (head->res.classid &&
net/sched/cls_matchall.c
349
nla_put_u32(skb, TCA_MATCHALL_CLASSID, head->res.classid))
net/sched/cls_matchall.c
352
if (head->flags && nla_put_u32(skb, TCA_MATCHALL_FLAGS, head->flags))
net/sched/cls_matchall.c
356
struct tc_matchall_pcnt *pf = per_cpu_ptr(head->pf, cpu);
net/sched/cls_matchall.c
366
if (tcf_exts_dump(skb, &head->exts))
net/sched/cls_matchall.c
37
if (tc_skip_sw(head->flags))
net/sched/cls_matchall.c
371
if (tcf_exts_dump_stats(skb, &head->exts) < 0)
net/sched/cls_matchall.c
384
struct cls_mall_head *head = fh;
net/sched/cls_matchall.c
386
tc_cls_bind_class(classid, cl, q, &head->res, base);
net/sched/cls_matchall.c
40
*res = head->res;
net/sched/cls_matchall.c
41
__this_cpu_inc(head->pf->rhit);
net/sched/cls_matchall.c
42
return tcf_exts_exec(skb, &head->exts, res);
net/sched/cls_matchall.c
50
static void __mall_destroy(struct cls_mall_head *head)
net/sched/cls_matchall.c
52
tcf_exts_destroy(&head->exts);
net/sched/cls_matchall.c
53
tcf_exts_put_net(&head->exts);
net/sched/cls_matchall.c
54
free_percpu(head->pf);
net/sched/cls_matchall.c
55
kfree(head);
net/sched/cls_matchall.c
60
struct cls_mall_head *head = container_of(to_rcu_work(work),
net/sched/cls_matchall.c
64
__mall_destroy(head);
net/sched/cls_matchall.c
69
struct cls_mall_head *head,
net/sched/cls_matchall.c
76
tc_cls_common_offload_init(&cls_mall.common, tp, head->flags, extack);
net/sched/cls_matchall.c
81
&head->flags, &head->in_hw_count, true);
net/sched/cls_matchall.c
85
struct cls_mall_head *head,
net/sched/cls_matchall.c
91
bool skip_sw = tc_skip_sw(head->flags);
net/sched/cls_matchall.c
94
cls_mall.rule = flow_rule_alloc(tcf_exts_num_actions(&head->exts));
net/sched/cls_matchall.c
98
tc_cls_common_offload_init(&cls_mall.common, tp, head->flags, extack);
net/sched/cls_route.c
121
route4_set_fastmap(head, id, iif, f); \
net/sched/cls_route.c
129
struct route4_head *head = rcu_dereference_bh(tp->root);
net/sched/cls_route.c
147
if (id == head->fastmap[h].id &&
net/sched/cls_route.c
148
iif == head->fastmap[h].iif &&
net/sched/cls_route.c
149
(f = head->fastmap[h].filter) != NULL) {
net/sched/cls_route.c
164
b = rcu_dereference_bh(head->table[h]);
net/sched/cls_route.c
190
route4_set_fastmap(head, id, iif, ROUTE4_FAILURE);
net/sched/cls_route.c
219
struct route4_head *head = rtnl_dereference(tp->root);
net/sched/cls_route.c
232
b = rtnl_dereference(head->table[h1]);
net/sched/cls_route.c
245
struct route4_head *head;
net/sched/cls_route.c
247
head = kzalloc_obj(struct route4_head);
net/sched/cls_route.c
248
if (head == NULL)
net/sched/cls_route.c
251
rcu_assign_pointer(tp->root, head);
net/sched/cls_route.c
280
struct route4_head *head = rtnl_dereference(tp->root);
net/sched/cls_route.c
283
if (head == NULL)
net/sched/cls_route.c
289
b = rtnl_dereference(head->table[h1]);
net/sched/cls_route.c
306
RCU_INIT_POINTER(head->table[h1], NULL);
net/sched/cls_route.c
310
kfree_rcu(head, rcu);
net/sched/cls_route.c
316
struct route4_head *head = rtnl_dereference(tp->root);
net/sched/cls_route.c
324
if (!head || !f)
net/sched/cls_route.c
341
route4_reset_fastmap(head);
net/sched/cls_route.c
358
RCU_INIT_POINTER(head->table[to_hash(h)], NULL);
net/sched/cls_route.c
367
if (rcu_access_pointer(head->table[h1])) {
net/sched/cls_route.c
385
u32 handle, struct route4_head *head,
net/sched/cls_route.c
439
b = rtnl_dereference(head->table[h1]);
net/sched/cls_route.c
445
rcu_assign_pointer(head->table[h1], b);
net/sched/cls_route.c
481
struct route4_head *head = rtnl_dereference(tp->root);
net/sched/cls_route.c
528
err = route4_set_parms(net, tp, base, f, handle, head, tb,
net/sched/cls_route.c
548
b = rtnl_dereference(head->table[th]);
net/sched/cls_route.c
561
route4_reset_fastmap(head);
net/sched/cls_route.c
580
struct route4_head *head = rtnl_dereference(tp->root);
net/sched/cls_route.c
583
if (head == NULL || arg->stop)
net/sched/cls_route.c
587
struct route4_bucket *b = rtnl_dereference(head->table[h]);
net/sched/cls_route.c
69
route4_reset_fastmap(struct route4_head *head)
net/sched/cls_route.c
72
memset(head->fastmap, 0, sizeof(head->fastmap));
net/sched/cls_route.c
77
route4_set_fastmap(struct route4_head *head, u32 id, int iif,
net/sched/cls_route.c
84
head->fastmap[h].id = id;
net/sched/cls_route.c
85
head->fastmap[h].iif = iif;
net/sched/cls_route.c
86
head->fastmap[h].filter = f;
net/sched/sch_cake.c
1229
if (flow->head == flow->tail)
net/sched/sch_cake.c
1252
for (skb_check = flow->head;
net/sched/sch_cake.c
129
struct sk_buff *head;
net/sched/sch_cake.c
1360
flow->head = elig_ack->next;
net/sched/sch_cake.c
1976
if (flow->head) {
net/sched/sch_cake.c
2010
struct list_head *head;
net/sched/sch_cake.c
2128
head = &b->decaying_flows;
net/sched/sch_cake.c
2129
if (!first_flow || list_empty(head)) {
net/sched/sch_cake.c
2130
head = &b->new_flows;
net/sched/sch_cake.c
2131
if (list_empty(head)) {
net/sched/sch_cake.c
2132
head = &b->old_flows;
net/sched/sch_cake.c
2133
if (unlikely(list_empty(head))) {
net/sched/sch_cake.c
2134
head = &b->decaying_flows;
net/sched/sch_cake.c
2135
if (unlikely(list_empty(head)))
net/sched/sch_cake.c
2140
flow = list_first_entry(head, struct cake_flow, flowchain);
net/sched/sch_cake.c
2151
if (flow->head) {
net/sched/sch_cake.c
2226
if (reason == SKB_NOT_DROPPED_YET || !flow->head)
net/sched/sch_cake.c
3131
if (flow->head) {
net/sched/sch_cake.c
3133
skb = flow->head;
net/sched/sch_cake.c
914
struct sk_buff *skb = flow->head;
net/sched/sch_cake.c
917
flow->head = skb->next;
net/sched/sch_cake.c
928
if (!flow->head)
net/sched/sch_cake.c
929
flow->head = skb;
net/sched/sch_choke.c
107
if (q->head == q->tail)
net/sched/sch_choke.c
121
if (idx == q->head)
net/sched/sch_choke.c
186
*pidx = (q->head + get_random_u32_below(choke_len(q))) & q->tab_mask;
net/sched/sch_choke.c
192
return q->tab[*pidx = q->head];
net/sched/sch_choke.c
205
if (q->head == q->tail)
net/sched/sch_choke.c
288
if (q->head == q->tail) {
net/sched/sch_choke.c
294
skb = q->tab[q->head];
net/sched/sch_choke.c
295
q->tab[q->head] = NULL;
net/sched/sch_choke.c
308
while (q->head != q->tail) {
net/sched/sch_choke.c
309
struct sk_buff *skb = q->tab[q->head];
net/sched/sch_choke.c
311
q->head = (q->head + 1) & q->tab_mask;
net/sched/sch_choke.c
319
q->head = q->tail = 0;
net/sched/sch_choke.c
383
while (q->head != q->tail) {
net/sched/sch_choke.c
384
struct sk_buff *skb = q->tab[q->head];
net/sched/sch_choke.c
386
q->head = (q->head + 1) & q->tab_mask;
net/sched/sch_choke.c
399
q->head = 0;
net/sched/sch_choke.c
417
if (q->head == q->tail)
net/sched/sch_choke.c
484
return (q->head != q->tail) ? q->tab[q->head] : NULL;
net/sched/sch_choke.c
66
unsigned int head;
net/sched/sch_choke.c
77
return (q->tail - q->head) & q->tab_mask;
net/sched/sch_choke.c
96
q->head = (q->head + 1) & q->tab_mask;
net/sched/sch_choke.c
97
if (q->head == q->tail)
net/sched/sch_choke.c
99
} while (q->tab[q->head] == NULL);
net/sched/sch_etf.c
113
p = rb_first_cached(&q->head);
net/sched/sch_etf.c
166
struct rb_node **p = &q->head.rb_root.rb_node, *parent = NULL;
net/sched/sch_etf.c
189
rb_insert_color_cached(&nskb->rbnode, &q->head, leftmost);
net/sched/sch_etf.c
211
rb_erase_cached(&skb->rbnode, &q->head);
net/sched/sch_etf.c
235
rb_erase_cached(&skb->rbnode, &q->head);
net/sched/sch_etf.c
35
struct rb_root_cached head;
net/sched/sch_etf.c
420
struct rb_node *p = rb_first_cached(&q->head);
net/sched/sch_etf.c
427
rb_erase_cached(&skb->rbnode, &q->head);
net/sched/sch_fq.c
201
struct fq_flow_head *head = (list_sel == NEW_FLOW) ?
net/sched/sch_fq.c
205
if (head->first)
net/sched/sch_fq.c
206
head->last->next = flow;
net/sched/sch_fq.c
208
head->first = flow;
net/sched/sch_fq.c
209
head->last = flow;
net/sched/sch_fq.c
464
struct sk_buff *head = flow->head;
net/sched/sch_fq.c
467
return head;
net/sched/sch_fq.c
469
if (!head)
net/sched/sch_fq.c
472
if (fq_skb_cb(skb)->time_to_send < fq_skb_cb(head)->time_to_send)
net/sched/sch_fq.c
474
return head;
net/sched/sch_fq.c
480
if (skb == flow->head) {
net/sched/sch_fq.c
484
flow->head = next;
net/sched/sch_fq.c
507
struct sk_buff *head, *aux;
net/sched/sch_fq.c
509
head = flow->head;
net/sched/sch_fq.c
510
if (!head ||
net/sched/sch_fq.c
512
if (!head)
net/sched/sch_fq.c
513
flow->head = skb;
net/sched/sch_fq.c
654
struct fq_flow_head *head;
net/sched/sch_fq.c
677
head = fq_pband_head_select(pband);
net/sched/sch_fq.c
678
if (!head) {
net/sched/sch_fq.c
695
f = head->first;
net/sched/sch_fq.c
699
head->first = f->next;
net/sched/sch_fq.c
710
head->first = f->next;
net/sched/sch_fq.c
72
struct sk_buff *head; /* list of skbs for this flow : first skb */
net/sched/sch_fq.c
725
head->first = f->next;
net/sched/sch_fq.c
727
if (head == &pband->new_flows) {
net/sched/sch_fq.c
795
rtnl_kfree_skbs(flow->head, flow->tail);
net/sched/sch_fq.c
796
flow->head = NULL;
net/sched/sch_fq_codel.c
118
struct sk_buff *skb = flow->head;
net/sched/sch_fq_codel.c
120
flow->head = skb->next;
net/sched/sch_fq_codel.c
129
if (flow->head == NULL)
net/sched/sch_fq_codel.c
130
flow->head = skb;
net/sched/sch_fq_codel.c
264
if (flow->head) {
net/sched/sch_fq_codel.c
287
struct list_head *head;
net/sched/sch_fq_codel.c
290
head = &q->new_flows;
net/sched/sch_fq_codel.c
291
if (list_empty(head)) {
net/sched/sch_fq_codel.c
292
head = &q->old_flows;
net/sched/sch_fq_codel.c
293
if (list_empty(head))
net/sched/sch_fq_codel.c
296
flow = list_first_entry(head, struct fq_codel_flow, flowchain);
net/sched/sch_fq_codel.c
310
if ((head == &q->new_flows) && !list_empty(&q->old_flows))
net/sched/sch_fq_codel.c
330
rtnl_kfree_skbs(flow->head, flow->tail);
net/sched/sch_fq_codel.c
331
flow->head = NULL;
net/sched/sch_fq_codel.c
43
struct sk_buff *head;
net/sched/sch_fq_codel.c
672
if (flow->head) {
net/sched/sch_fq_codel.c
674
skb = flow->head;
net/sched/sch_fq_pie.c
122
if (!flow->head)
net/sched/sch_fq_pie.c
123
flow->head = skb;
net/sched/sch_fq_pie.c
231
struct sk_buff *skb = flow->head;
net/sched/sch_fq_pie.c
233
flow->head = skb->next;
net/sched/sch_fq_pie.c
243
struct list_head *head;
net/sched/sch_fq_pie.c
247
head = &q->new_flows;
net/sched/sch_fq_pie.c
248
if (list_empty(head)) {
net/sched/sch_fq_pie.c
249
head = &q->old_flows;
net/sched/sch_fq_pie.c
250
if (list_empty(head))
net/sched/sch_fq_pie.c
254
flow = list_first_entry(head, struct fq_pie_flow, flowchain);
net/sched/sch_fq_pie.c
262
if (flow->head) {
net/sched/sch_fq_pie.c
272
if (head == &q->new_flows && !list_empty(&q->old_flows))
net/sched/sch_fq_pie.c
51
struct sk_buff *head;
net/sched/sch_fq_pie.c
545
rtnl_kfree_skbs(flow->head, flow->tail);
net/sched/sch_fq_pie.c
546
flow->head = NULL;
net/sched/sch_generic.c
1052
static void qdisc_free_cb(struct rcu_head *head)
net/sched/sch_generic.c
1054
struct Qdisc *q = container_of(head, struct Qdisc, rcu);
net/sched/sch_generic.c
1325
void dev_deactivate_many(struct list_head *head)
net/sched/sch_generic.c
1330
list_for_each_entry(dev, head, close_list) {
net/sched/sch_generic.c
1344
list_for_each_entry(dev, head, close_list) {
net/sched/sch_generic.c
1352
list_for_each_entry(dev, head, close_list) {
net/sched/sch_hfsc.c
1133
struct hfsc_class *head, *cl;
net/sched/sch_hfsc.c
1144
head = &q->root;
net/sched/sch_hfsc.c
1163
if (cl->level >= head->level)
net/sched/sch_hfsc.c
1172
head = cl;
net/sched/sch_hhf.c
121
struct sk_buff *head;
net/sched/sch_hhf.c
181
struct list_head *head,
net/sched/sch_hhf.c
187
if (list_empty(head))
net/sched/sch_hhf.c
190
list_for_each_entry_safe(flow, next, head, flowchain) {
net/sched/sch_hhf.c
197
if (list_is_last(&flow->flowchain, head))
net/sched/sch_hhf.c
212
static struct hh_flow_state *alloc_new_hh(struct list_head *head,
net/sched/sch_hhf.c
218
if (!list_empty(head)) {
net/sched/sch_hhf.c
220
list_for_each_entry(flow, head, flowchain) {
net/sched/sch_hhf.c
239
list_add_tail(&flow->flowchain, head);
net/sched/sch_hhf.c
331
struct sk_buff *skb = bucket->head;
net/sched/sch_hhf.c
333
bucket->head = skb->next;
net/sched/sch_hhf.c
341
if (bucket->head == NULL)
net/sched/sch_hhf.c
342
bucket->head = skb;
net/sched/sch_hhf.c
356
if (!bucket->head)
net/sched/sch_hhf.c
359
if (bucket->head) {
net/sched/sch_hhf.c
423
struct list_head *head;
net/sched/sch_hhf.c
426
head = &q->new_buckets;
net/sched/sch_hhf.c
427
if (list_empty(head)) {
net/sched/sch_hhf.c
428
head = &q->old_buckets;
net/sched/sch_hhf.c
429
if (list_empty(head))
net/sched/sch_hhf.c
432
bucket = list_first_entry(head, struct wdrr_bucket, bucketchain);
net/sched/sch_hhf.c
443
if (bucket->head) {
net/sched/sch_hhf.c
451
if ((head == &q->new_buckets) && !list_empty(&q->old_buckets))
net/sched/sch_hhf.c
486
struct list_head *head = &q->hh_flows[i];
net/sched/sch_hhf.c
488
if (list_empty(head))
net/sched/sch_hhf.c
490
list_for_each_entry_safe(flow, next, head, flowchain) {
net/sched/sch_sfq.c
349
struct sk_buff *head;
net/sched/sch_sfq.c
426
head = slot_dequeue_head(slot);
net/sched/sch_sfq.c
427
delta = qdisc_pkt_len(head) - qdisc_pkt_len(skb);
net/sched/sch_sfq.c
430
qdisc_drop(head, sch, to_free);
net/sched/sch_taprio.c
196
static void taprio_free_sched_cb(struct rcu_head *head)
net/sched/sch_taprio.c
198
struct sched_gate_list *sched = container_of(head, struct sched_gate_list, rcu);
net/sctp/endpointola.c
191
static void sctp_endpoint_destroy_rcu(struct rcu_head *head)
net/sctp/endpointola.c
193
struct sctp_endpoint *ep = container_of(head, struct sctp_endpoint, rcu);
net/sctp/input.c
737
struct sctp_hashbucket *head;
net/sctp/input.c
741
head = &sctp_ep_hashtable[ep->hashent];
net/sctp/input.c
743
write_lock(&head->lock);
net/sctp/input.c
755
sctp_for_each_hentry(ep2, &head->chain) {
net/sctp/input.c
782
hlist_add_head(&ep->node, &head->chain);
net/sctp/input.c
784
write_unlock(&head->lock);
net/sctp/input.c
804
struct sctp_hashbucket *head;
net/sctp/input.c
808
head = &sctp_ep_hashtable[ep->hashent];
net/sctp/input.c
810
write_lock(&head->lock);
net/sctp/input.c
814
write_unlock(&head->lock);
net/sctp/input.c
846
struct sctp_hashbucket *head;
net/sctp/input.c
854
head = &sctp_ep_hashtable[hash];
net/sctp/input.c
855
read_lock(&head->lock);
net/sctp/input.c
856
sctp_for_each_hentry(ep, &head->chain) {
net/sctp/input.c
874
read_unlock(&head->lock);
net/sctp/output.c
420
static void sctp_packet_gso_append(struct sk_buff *head, struct sk_buff *skb)
net/sctp/output.c
422
if (SCTP_OUTPUT_CB(head)->last == head)
net/sctp/output.c
423
skb_shinfo(head)->frag_list = skb;
net/sctp/output.c
425
SCTP_OUTPUT_CB(head)->last->next = skb;
net/sctp/output.c
426
SCTP_OUTPUT_CB(head)->last = skb;
net/sctp/output.c
428
head->truesize += skb->truesize;
net/sctp/output.c
429
head->data_len += skb->len;
net/sctp/output.c
430
head->len += skb->len;
net/sctp/output.c
431
refcount_add(skb->truesize, &head->sk->sk_wmem_alloc);
net/sctp/output.c
437
struct sk_buff *head, int gso, gfp_t gfp)
net/sctp/output.c
443
struct sock *sk = head->sk;
net/sctp/output.c
448
skb_shinfo(head)->gso_type = sk->sk_gso_type;
net/sctp/output.c
449
SCTP_OUTPUT_CB(head)->last = head;
net/sctp/output.c
451
nskb = head;
net/sctp/output.c
531
sctp_packet_gso_append(head, nskb);
net/sctp/output.c
537
memset(head->cb, 0, max(sizeof(struct inet_skb_parm),
net/sctp/output.c
539
skb_shinfo(head)->gso_segs = pkt_count;
net/sctp/output.c
540
skb_shinfo(head)->gso_size = GSO_BY_FRAGS;
net/sctp/output.c
550
(struct sctphdr *)skb_transport_header(head);
net/sctp/output.c
552
sh->checksum = sctp_compute_cksum(head, 0);
net/sctp/output.c
555
head->ip_summed = CHECKSUM_PARTIAL;
net/sctp/output.c
556
head->csum_not_inet = 1;
net/sctp/output.c
557
head->csum_start = skb_transport_header(head) - head->head;
net/sctp/output.c
558
head->csum_offset = offsetof(struct sctphdr, checksum);
net/sctp/output.c
575
struct sk_buff *head;
net/sctp/output.c
598
head = alloc_skb((gso ? packet->overhead : packet->size) +
net/sctp/output.c
600
if (!head)
net/sctp/output.c
602
skb_reserve(head, packet->overhead + MAX_HEADER);
net/sctp/output.c
603
skb_set_owner_w(head, sk);
net/sctp/output.c
606
sh = skb_push(head, sizeof(struct sctphdr));
net/sctp/output.c
607
skb_reset_transport_header(head);
net/sctp/output.c
616
kfree_skb(head);
net/sctp/output.c
621
pkt_count = sctp_packet_pack(packet, head, gso, gfp);
net/sctp/output.c
623
kfree_skb(head);
net/sctp/output.c
626
pr_debug("***sctp_transmit_packet*** skb->len:%d\n", head->len);
net/sctp/output.c
647
head->ignore_df = packet->ipfragok;
net/sctp/output.c
649
skb_set_dst_pending_confirm(head, 1);
net/sctp/output.c
653
if (tp->af_specific->sctp_xmit(head, tp) >= 0 &&
net/sctp/outqueue.c
1107
chunk->skb ? chunk->skb->head : NULL, chunk->skb ?
net/sctp/outqueue.c
319
static void sctp_insert_list(struct list_head *head, struct list_head *new)
net/sctp/outqueue.c
329
list_for_each(pos, head) {
net/sctp/outqueue.c
339
list_add_tail(new, head);
net/sctp/proc.c
163
struct sctp_hashbucket *head;
net/sctp/proc.c
171
head = &sctp_ep_hashtable[hash];
net/sctp/proc.c
172
read_lock_bh(&head->lock);
net/sctp/proc.c
173
sctp_for_each_hentry(ep, &head->chain) {
net/sctp/proc.c
186
read_unlock_bh(&head->lock);
net/sctp/socket.c
5366
struct sctp_hashbucket *head;
net/sctp/socket.c
5368
for (head = sctp_ep_hashtable; hash < sctp_ep_hashsize;
net/sctp/socket.c
5369
hash++, head++) {
net/sctp/socket.c
5370
read_lock_bh(&head->lock);
net/sctp/socket.c
5371
sctp_for_each_hentry(ep, &head->chain) {
net/sctp/socket.c
5376
read_unlock_bh(&head->lock);
net/sctp/socket.c
8380
struct sctp_bind_hashbucket *head, struct net *, unsigned short snum);
net/sctp/socket.c
8386
struct sctp_bind_hashbucket *head; /* hash list */
net/sctp/socket.c
8413
head = &sctp_port_hashtable[index];
net/sctp/socket.c
8414
spin_lock_bh(&head->lock);
net/sctp/socket.c
8415
sctp_for_each_hentry(pp, &head->chain)
net/sctp/socket.c
8421
spin_unlock_bh(&head->lock);
net/sctp/socket.c
8442
head = &sctp_port_hashtable[sctp_phashfn(net, snum)];
net/sctp/socket.c
8443
spin_lock_bh(&head->lock);
net/sctp/socket.c
8444
sctp_for_each_hentry(pp, &head->chain) {
net/sctp/socket.c
8504
if (!pp && !(pp = sctp_bucket_create(head, net, snum)))
net/sctp/socket.c
8546
spin_unlock_bh(&head->lock);
net/sctp/socket.c
8746
struct sctp_bind_hashbucket *head, struct net *net, unsigned short snum)
net/sctp/socket.c
8757
hlist_add_head(&pp->node, &head->chain);
net/sctp/socket.c
8775
struct sctp_bind_hashbucket *head =
net/sctp/socket.c
8780
spin_lock(&head->lock);
net/sctp/socket.c
8786
spin_unlock(&head->lock);
net/sctp/socket.c
9510
struct sctp_bind_hashbucket *head;
net/sctp/socket.c
9519
head = &sctp_port_hashtable[sctp_phashfn(sock_net(oldsk),
net/sctp/socket.c
9521
spin_lock_bh(&head->lock);
net/sctp/socket.c
9526
spin_unlock_bh(&head->lock);
net/sctp/transport.c
138
static void sctp_transport_destroy_rcu(struct rcu_head *head)
net/sctp/transport.c
142
transport = container_of(head, struct sctp_transport, rcu);
net/smc/af_smc.c
200
struct hlist_head *head;
net/smc/af_smc.c
202
head = &h->ht;
net/smc/af_smc.c
205
sk_add_node(sk, head);
net/smc/smc_diag.c
203
struct hlist_head *head;
net/smc/smc_diag.c
208
head = &prot->h.smc_hash->ht;
net/smc/smc_diag.c
209
if (hlist_empty(head))
net/smc/smc_diag.c
212
sk_for_each(sk, head) {
net/strparser/strparser.c
104
head = strp->skb_head;
net/strparser/strparser.c
105
if (head) {
net/strparser/strparser.c
133
err = skb_unclone(head, GFP_ATOMIC);
net/strparser/strparser.c
140
if (unlikely(skb_shinfo(head)->frag_list)) {
net/strparser/strparser.c
147
if (WARN_ON(head->next)) {
net/strparser/strparser.c
152
skb = alloc_skb_for_msg(head);
net/strparser/strparser.c
159
strp->skb_nextp = &head->next;
net/strparser/strparser.c
161
head = skb;
net/strparser/strparser.c
164
&skb_shinfo(head)->frag_list;
net/strparser/strparser.c
180
head = strp->skb_head;
net/strparser/strparser.c
181
if (!head) {
net/strparser/strparser.c
182
head = skb;
net/strparser/strparser.c
183
strp->skb_head = head;
net/strparser/strparser.c
186
stm = _strp_msg(head);
net/strparser/strparser.c
202
stm = _strp_msg(head);
net/strparser/strparser.c
205
head->data_len += skb->len;
net/strparser/strparser.c
206
head->len += skb->len;
net/strparser/strparser.c
207
head->truesize += skb->truesize;
net/strparser/strparser.c
213
len = (*strp->cb.parse_msg)(strp, head);
net/strparser/strparser.c
240
} else if (len <= (ssize_t)head->len -
net/strparser/strparser.c
301
strp->cb.rcv_msg(strp, head);
net/strparser/strparser.c
95
struct sk_buff *head, *skb;
net/sunrpc/auth.c
324
void rpcauth_destroy_credlist(struct list_head *head)
net/sunrpc/auth.c
328
while (!list_empty(head)) {
net/sunrpc/auth.c
329
cred = list_entry(head->next, struct rpc_cred, cr_lru);
net/sunrpc/auth.c
381
struct hlist_head *head;
net/sunrpc/auth.c
389
head = &cache->hashtable[i];
net/sunrpc/auth.c
390
while (!hlist_empty(head)) {
net/sunrpc/auth.c
391
cred = hlist_entry(head->first, struct rpc_cred, cr_hash);
net/sunrpc/auth_gss/auth_gss.c
1312
gss_free_ctx_callback(struct rcu_head *head)
net/sunrpc/auth_gss/auth_gss.c
1314
struct gss_cl_ctx *ctx = container_of(head, struct gss_cl_ctx, gc_rcu);
net/sunrpc/auth_gss/auth_gss.c
1331
gss_free_cred_callback(struct rcu_head *head)
net/sunrpc/auth_gss/auth_gss.c
1333
struct gss_cred *gss_cred = container_of(head, struct gss_cred, gc_base.cr_rcu);
net/sunrpc/auth_gss/auth_gss.c
1572
iov.iov_base = req->rq_snd_buf.head[0].iov_base;
net/sunrpc/auth_gss/auth_gss.c
1765
offset = (u8 *)p - (u8 *)snd_buf->head[0].iov_base;
net/sunrpc/auth_gss/auth_gss.c
1880
offset = (u8 *)p - (u8 *)snd_buf->head[0].iov_base;
net/sunrpc/auth_gss/auth_gss.c
1899
iov = snd_buf->head;
net/sunrpc/auth_gss/auth_gss.c
2066
struct kvec *head = rqstp->rq_rcv_buf.head;
net/sunrpc/auth_gss/auth_gss.c
2074
offset = (u8 *)(p) - (u8 *)head->iov_base;
net/sunrpc/auth_gss/gss_krb5_crypto.c
252
page_pos = desc->pos - outbuf->head[0].iov_len;
net/sunrpc/auth_gss/gss_krb5_crypto.c
382
p = buf->head[0].iov_base + base;
net/sunrpc/auth_gss/gss_krb5_crypto.c
384
memmove(p + shiftlen, p, buf->head[0].iov_len - base);
net/sunrpc/auth_gss/gss_krb5_crypto.c
386
buf->head[0].iov_len += shiftlen;
net/sunrpc/auth_gss/gss_krb5_crypto.c
608
krb5_make_confounder(buf->head[0].iov_base + offset, conflen);
net/sunrpc/auth_gss/gss_krb5_crypto.c
614
buf->tail[0].iov_base = buf->head[0].iov_base
net/sunrpc/auth_gss/gss_krb5_crypto.c
615
+ buf->head[0].iov_len;
net/sunrpc/auth_gss/gss_krb5_crypto.c
621
memcpy(ecptr, buf->head[0].iov_base + offset, GSS_KRB5_TOK_HDR_LEN);
net/sunrpc/auth_gss/gss_krb5_crypto.c
837
krb5_make_confounder(buf->head[0].iov_base + offset, conflen);
net/sunrpc/auth_gss/gss_krb5_crypto.c
843
buf->tail[0].iov_base = buf->head[0].iov_base
net/sunrpc/auth_gss/gss_krb5_crypto.c
844
+ buf->head[0].iov_len;
net/sunrpc/auth_gss/gss_krb5_crypto.c
849
memcpy(ecptr, buf->head[0].iov_base + offset, GSS_KRB5_TOK_HDR_LEN);
net/sunrpc/auth_gss/gss_krb5_test.c
101
buf.head[0].iov_base = kunit_kzalloc(test, buf.head[0].iov_len, GFP_KERNEL);
net/sunrpc/auth_gss/gss_krb5_test.c
102
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, buf.head[0].iov_base);
net/sunrpc/auth_gss/gss_krb5_test.c
103
memcpy(buf.head[0].iov_base, param->plaintext->data, buf.head[0].iov_len);
net/sunrpc/auth_gss/gss_krb5_test.c
1166
buf.head[0].iov_base = text;
net/sunrpc/auth_gss/gss_krb5_test.c
1167
buf.head[0].iov_len = param->confounder->len + param->plaintext->len;
net/sunrpc/auth_gss/gss_krb5_test.c
1168
buf.len = buf.head[0].iov_len;
net/sunrpc/auth_gss/gss_krb5_test.c
1199
buf.head[0].iov_base, buf.len), 0,
net/sunrpc/auth_gss/gss_krb5_test.c
1659
buf.head[0].iov_base = text;
net/sunrpc/auth_gss/gss_krb5_test.c
1660
buf.head[0].iov_len = param->confounder->len + param->plaintext->len;
net/sunrpc/auth_gss/gss_krb5_test.c
1661
buf.len = buf.head[0].iov_len;
net/sunrpc/auth_gss/gss_krb5_test.c
1692
buf.head[0].iov_base,
net/sunrpc/auth_gss/gss_krb5_test.c
1815
buf.head[0].iov_base = text;
net/sunrpc/auth_gss/gss_krb5_test.c
1816
buf.head[0].iov_len = param->plaintext->len;
net/sunrpc/auth_gss/gss_krb5_test.c
1817
buf.len = buf.head[0].iov_len;
net/sunrpc/auth_gss/gss_krb5_test.c
1831
buf.head[0].iov_base, buf.len), 0,
net/sunrpc/auth_gss/gss_krb5_test.c
559
buf.head[0].iov_base = text;
net/sunrpc/auth_gss/gss_krb5_test.c
560
buf.head[0].iov_len = param->plaintext->len;
net/sunrpc/auth_gss/gss_krb5_test.c
561
buf.len = buf.head[0].iov_len;
net/sunrpc/auth_gss/gss_krb5_test.c
76
.head[0].iov_len = param->plaintext->len,
net/sunrpc/auth_gss/gss_krb5_wrap.c
111
ptr = buf->head[0].iov_base + offset;
net/sunrpc/auth_gss/gss_krb5_wrap.c
158
ptr = buf->head[0].iov_base + offset;
net/sunrpc/auth_gss/gss_krb5_wrap.c
223
movelen = min_t(unsigned int, buf->head[0].iov_len, len);
net/sunrpc/auth_gss/gss_krb5_wrap.c
226
buf->head[0].iov_len);
net/sunrpc/auth_gss/gss_krb5_wrap.c
228
buf->head[0].iov_len -= GSS_KRB5_TOK_HDR_LEN + headskip;
net/sunrpc/auth_gss/gss_krb5_wrap.c
57
char head[LOCAL_BUF_LEN];
net/sunrpc/auth_gss/gss_krb5_wrap.c
63
read_bytes_from_xdr_buf(buf, 0, head, shift);
net/sunrpc/auth_gss/gss_krb5_wrap.c
69
write_bytes_to_xdr_buf(buf, buf->len - shift, head, shift);
net/sunrpc/auth_gss/svcauth_gss.c
117
static void rsi_free_rcu(struct rcu_head *head)
net/sunrpc/auth_gss/svcauth_gss.c
119
struct rsi *rsii = container_of(head, struct rsi, rcu_head);
net/sunrpc/auth_gss/svcauth_gss.c
1852
struct kvec *head = buf->head;
net/sunrpc/auth_gss/svcauth_gss.c
1877
if (tail->iov_base >= head->iov_base + PAGE_SIZE)
net/sunrpc/auth_gss/svcauth_gss.c
1879
if (tail->iov_base < head->iov_base)
net/sunrpc/auth_gss/svcauth_gss.c
1881
if (tail->iov_len + head->iov_len
net/sunrpc/auth_gss/svcauth_gss.c
1896
if (head->iov_len + 2 * RPC_MAX_AUTH_SIZE > PAGE_SIZE)
net/sunrpc/auth_gss/svcauth_gss.c
1898
tail->iov_base = head->iov_base
net/sunrpc/auth_gss/svcauth_gss.c
1899
+ head->iov_len + RPC_MAX_AUTH_SIZE;
net/sunrpc/auth_gss/svcauth_gss.c
1988
svcauth_gss_domain_release_rcu(struct rcu_head *head)
net/sunrpc/auth_gss/svcauth_gss.c
1990
struct auth_domain *dom = container_of(head, struct auth_domain, rcu_head);
net/sunrpc/auth_gss/svcauth_gss.c
388
static void rsc_free_rcu(struct rcu_head *head)
net/sunrpc/auth_gss/svcauth_gss.c
390
struct rsc *rsci = container_of(head, struct rsc, rcu_head);
net/sunrpc/auth_unix.c
62
unx_free_cred_callback(struct rcu_head *head)
net/sunrpc/auth_unix.c
64
struct rpc_cred *rpc_cred = container_of(head, struct rpc_cred, cr_rcu);
net/sunrpc/backchannel_rqst.c
63
free_page((unsigned long)xbufp->head[0].iov_base);
net/sunrpc/backchannel_rqst.c
65
free_page((unsigned long)xbufp->head[0].iov_base);
net/sunrpc/backchannel_rqst.c
71
buf->head[0].iov_len = PAGE_SIZE;
net/sunrpc/cache.c
105
struct hlist_head *head = &detail->hash_table[hash];
net/sunrpc/cache.c
120
hlist_for_each_entry_rcu(tmp, head, cache_list,
net/sunrpc/cache.c
137
hlist_add_head_rcu(&new->cache_list, head);
net/sunrpc/cache.c
164
static void cache_fresh_locked(struct cache_head *head, time64_t expiry,
net/sunrpc/cache.c
171
head->expiry_time = expiry;
net/sunrpc/cache.c
172
head->last_refresh = now;
net/sunrpc/cache.c
174
set_bit(CACHE_VALID, &head->flags);
net/sunrpc/cache.c
177
static void cache_fresh_unlocked(struct cache_head *head,
net/sunrpc/cache.c
180
if (test_and_clear_bit(CACHE_PENDING, &head->flags)) {
net/sunrpc/cache.c
181
cache_revisit_request(head);
net/sunrpc/cache.c
182
cache_dequeue(detail, head);
net/sunrpc/cache.c
479
struct hlist_head *head;
net/sunrpc/cache.c
483
head = &current_detail->hash_table[current_index];
net/sunrpc/cache.c
484
hlist_for_each_entry_safe(ch, tmp, head, cache_list) {
net/sunrpc/cache.c
547
struct hlist_head *head = NULL;
net/sunrpc/cache.c
558
head = &detail->hash_table[i];
net/sunrpc/cache.c
559
while (!hlist_empty(head)) {
net/sunrpc/cache.c
560
ch = hlist_entry(head->first, struct cache_head,
net/sunrpc/cache.c
59
static void cache_fresh_unlocked(struct cache_head *head,
net/sunrpc/cache.c
66
struct hlist_head *head = &detail->hash_table[hash];
net/sunrpc/cache.c
70
hlist_for_each_entry_rcu(tmp, head, cache_list) {
net/sunrpc/clnt.c
1950
req->rq_snd_buf.head[0].iov_len = 0;
net/sunrpc/clnt.c
1952
req->rq_snd_buf.head[0].iov_base, req);
net/sunrpc/clnt.c
2623
req->rq_rcv_buf.head[0].iov_base, req);
net/sunrpc/rpc_pipe.c
64
static void rpc_purge_list(wait_queue_head_t *waitq, struct list_head *head,
net/sunrpc/rpc_pipe.c
69
if (list_empty(head))
net/sunrpc/rpc_pipe.c
72
msg = list_entry(head->next, struct rpc_pipe_msg, list);
net/sunrpc/rpc_pipe.c
76
} while (!list_empty(head));
net/sunrpc/socklib.c
233
want = xdr->head[0].iov_len + rmsize;
net/sunrpc/socklib.c
242
&xdr->head[0], base);
net/sunrpc/socklib.c
244
err = xprt_send_kvec(sock, msg, &xdr->head[0], base);
net/sunrpc/socklib.c
73
if (xdr->head[0].iov_len == 0)
net/sunrpc/socklib.c
76
ret = xdr_skb_read_bits(desc, xdr->head[0].iov_base,
net/sunrpc/socklib.c
77
xdr->head[0].iov_len);
net/sunrpc/socklib.c
78
if (ret != xdr->head[0].iov_len || !desc->count)
net/sunrpc/svc.c
1586
struct kvec *resv = &rqstp->rq_res.head[0];
net/sunrpc/svc.c
1662
if (rqstp->rq_arg.len <= rqstp->rq_arg.head[0].iov_len) {
net/sunrpc/svc.c
1663
rqstp->rq_arg.head[0].iov_len = rqstp->rq_arg.len;
net/sunrpc/svc.c
1665
} else if (rqstp->rq_arg.len <= rqstp->rq_arg.head[0].iov_len +
net/sunrpc/svc.c
1668
rqstp->rq_arg.head[0].iov_len;
net/sunrpc/svc.c
1670
rqstp->rq_arg.len = rqstp->rq_arg.head[0].iov_len +
net/sunrpc/svc.c
1674
rqstp->rq_res.head[0].iov_len = 0;
net/sunrpc/svc_xprt.c
1244
skip = rqstp->rq_arg.len - rqstp->rq_arg.head[0].iov_len;
net/sunrpc/svc_xprt.c
1245
memcpy(dr->args, rqstp->rq_arg.head[0].iov_base - skip,
net/sunrpc/svc_xprt.c
1269
rqstp->rq_arg.head[0].iov_base = dr->args;
net/sunrpc/svc_xprt.c
1271
rqstp->rq_arg.head[0].iov_len = dr->argslen << 2;
net/sunrpc/svc_xprt.c
525
space += rqstp->rq_res.head[0].iov_len;
net/sunrpc/svc_xprt.c
569
rqstp->rq_res.head[0].iov_len = 0;
net/sunrpc/svc_xprt.c
677
arg->head[0].iov_base = page_address(rqstp->rq_pages[0]);
net/sunrpc/svc_xprt.c
678
arg->head[0].iov_len = PAGE_SIZE;
net/sunrpc/svc_xprt.c
941
xb->len = xb->head[0].iov_len +
net/sunrpc/svcauth.c
241
struct hlist_head *head;
net/sunrpc/svcauth.c
243
head = &auth_domain_table[hash_str(name, DN_HASHBITS)];
net/sunrpc/svcauth.c
247
hlist_for_each_entry(hp, head, hash) {
net/sunrpc/svcauth.c
255
hlist_add_head_rcu(&new->hash, head);
net/sunrpc/svcauth.c
264
struct hlist_head *head;
net/sunrpc/svcauth.c
266
head = &auth_domain_table[hash_str(name, DN_HASHBITS)];
net/sunrpc/svcauth.c
269
hlist_for_each_entry_rcu(hp, head, hash) {
net/sunrpc/svcauth_unix.c
43
static void svcauth_unix_domain_release_rcu(struct rcu_head *head)
net/sunrpc/svcauth_unix.c
45
struct auth_domain *dom = container_of(head, struct auth_domain, rcu_head);
net/sunrpc/svcsock.c
1002
rqstp->rq_arg.head[0].iov_base = page_address(rqstp->rq_pages[0]);
net/sunrpc/svcsock.c
1090
__be32 *p = (__be32 *)rqstp->rq_arg.head[0].iov_base;
net/sunrpc/svcsock.c
1106
dst = &req->rq_private_buf.head[0];
net/sunrpc/svcsock.c
1107
src = &rqstp->rq_arg.head[0];
net/sunrpc/svcsock.c
1175
if (rqstp->rq_arg.len <= rqstp->rq_arg.head[0].iov_len) {
net/sunrpc/svcsock.c
1176
rqstp->rq_arg.head[0].iov_len = rqstp->rq_arg.len;
net/sunrpc/svcsock.c
1179
rqstp->rq_arg.page_len = rqstp->rq_arg.len - rqstp->rq_arg.head[0].iov_len;
net/sunrpc/svcsock.c
1188
p = (__be32 *)rqstp->rq_arg.head[0].iov_base;
net/sunrpc/svcsock.c
669
rqstp->rq_arg.head[0].iov_base = skb->data;
net/sunrpc/svcsock.c
670
rqstp->rq_arg.head[0].iov_len = len;
net/sunrpc/svcsock.c
677
if (len <= rqstp->rq_arg.head[0].iov_len) {
net/sunrpc/svcsock.c
678
rqstp->rq_arg.head[0].iov_len = len;
net/sunrpc/svcsock.c
682
rqstp->rq_arg.page_len = len - rqstp->rq_arg.head[0].iov_len;
net/sunrpc/sysfs.c
17
static void free_xprt_addr(struct rcu_head *head)
net/sunrpc/sysfs.c
19
struct xprt_addr *addr = container_of(head, struct xprt_addr, rcu);
net/sunrpc/xdr.c
1121
if (xdr->iov == xdr->buf->head) {
net/sunrpc/xdr.c
1164
struct kvec *head = buf->head;
net/sunrpc/xdr.c
1201
xdr->end = head->iov_base + head->iov_len;
net/sunrpc/xdr.c
1204
head->iov_len = len;
net/sunrpc/xdr.c
1206
xdr->p = head->iov_base + head->iov_len;
net/sunrpc/xdr.c
1207
xdr->iov = buf->head;
net/sunrpc/xdr.c
1313
xdr_stream_set_pos(xdr, base + buf->page_len + buf->head->iov_len);
net/sunrpc/xdr.c
1391
else if (xdr->iov == xdr->buf->head)
net/sunrpc/xdr.c
1410
if (xdr_set_iov(xdr, buf->head, 0, buf->len) == 0 &&
net/sunrpc/xdr.c
150
const struct kvec *head = xdr->head;
net/sunrpc/xdr.c
1520
struct kvec *iov = buf->head;
net/sunrpc/xdr.c
154
if (head->iov_len) {
net/sunrpc/xdr.c
155
bvec_set_virt(bvec++, head->iov_base, head->iov_len);
net/sunrpc/xdr.c
1641
buf->head[0] = *iov;
net/sunrpc/xdr.c
1666
if (base < buf->head[0].iov_len) {
net/sunrpc/xdr.c
1667
subbuf->head[0].iov_base = buf->head[0].iov_base + base;
net/sunrpc/xdr.c
1668
subbuf->head[0].iov_len = min_t(unsigned int, len,
net/sunrpc/xdr.c
1669
buf->head[0].iov_len - base);
net/sunrpc/xdr.c
1670
len -= subbuf->head[0].iov_len;
net/sunrpc/xdr.c
1673
base -= buf->head[0].iov_len;
net/sunrpc/xdr.c
1674
subbuf->head[0].iov_base = buf->head[0].iov_base;
net/sunrpc/xdr.c
1675
subbuf->head[0].iov_len = 0;
net/sunrpc/xdr.c
1801
if (buf.head[0].iov_len)
net/sunrpc/xdr.c
1802
xdr_buf_iov_zero(buf.head, 0, buf.head[0].iov_len);
net/sunrpc/xdr.c
1842
if (buf->head[0].iov_len) {
net/sunrpc/xdr.c
1843
cur = min_t(size_t, buf->head[0].iov_len, trim);
net/sunrpc/xdr.c
1844
buf->head[0].iov_len -= cur;
net/sunrpc/xdr.c
1857
this_len = min_t(unsigned int, len, subbuf->head[0].iov_len);
net/sunrpc/xdr.c
1858
memcpy(obj, subbuf->head[0].iov_base, this_len);
net/sunrpc/xdr.c
1889
this_len = min_t(unsigned int, len, subbuf->head[0].iov_len);
net/sunrpc/xdr.c
1890
memcpy(subbuf->head[0].iov_base, obj, this_len);
net/sunrpc/xdr.c
1964
if (todo && base < buf->head->iov_len) {
net/sunrpc/xdr.c
1965
c = buf->head->iov_base + base;
net/sunrpc/xdr.c
1967
buf->head->iov_len - base);
net/sunrpc/xdr.c
1993
base = buf->head->iov_len; /* align to start of pages */
net/sunrpc/xdr.c
1997
base -= buf->head->iov_len;
net/sunrpc/xdr.c
203
struct kvec *head = xdr->head;
net/sunrpc/xdr.c
205
char *buf = (char *)head->iov_base;
net/sunrpc/xdr.c
206
unsigned int buflen = head->iov_len;
net/sunrpc/xdr.c
208
head->iov_len = offset;
net/sunrpc/xdr.c
2145
buf->head->iov_len + buf->page_len + buf->tail->iov_len)
net/sunrpc/xdr.c
2162
if (offset >= buf->head[0].iov_len) {
net/sunrpc/xdr.c
2163
offset -= buf->head[0].iov_len;
net/sunrpc/xdr.c
2165
thislen = buf->head[0].iov_len - offset;
net/sunrpc/xdr.c
2168
sg_set_buf(sg, buf->head[0].iov_base + offset, thislen);
net/sunrpc/xdr.c
510
if (buflen <= buf->head->iov_len)
net/sunrpc/xdr.c
512
pagelen = buflen - buf->head->iov_len;
net/sunrpc/xdr.c
533
struct kvec *head = buf->head;
net/sunrpc/xdr.c
535
unsigned int sum = head->iov_len + buf->page_len + tail->iov_len;
net/sunrpc/xdr.c
608
const struct kvec *head = buf->head;
net/sunrpc/xdr.c
614
if (base >= head->iov_len)
net/sunrpc/xdr.c
616
if (len > head->iov_len - base)
net/sunrpc/xdr.c
617
len = head->iov_len - base;
net/sunrpc/xdr.c
618
if (to >= buf->page_len + head->iov_len) {
net/sunrpc/xdr.c
619
tato = to - buf->page_len - head->iov_len;
net/sunrpc/xdr.c
621
} else if (to >= head->iov_len) {
net/sunrpc/xdr.c
622
pgto = to - head->iov_len;
net/sunrpc/xdr.c
640
memcpy(tail->iov_base + tato, head->iov_base + base, talen);
net/sunrpc/xdr.c
644
_copy_to_pages(buf->pages, buf->page_base + pgto, head->iov_base + base,
net/sunrpc/xdr.c
648
memmove(head->iov_base + to, head->iov_base + base, len);
net/sunrpc/xdr.c
682
const struct kvec *head = buf->head;
net/sunrpc/xdr.c
686
if (base >= head->iov_len) {
net/sunrpc/xdr.c
687
xdr_buf_pages_shift_right(buf, head->iov_len - base, len,
net/sunrpc/xdr.c
691
if (base + len > head->iov_len)
net/sunrpc/xdr.c
692
xdr_buf_pages_shift_right(buf, 0, base + len - head->iov_len,
net/sunrpc/xdr.c
708
const struct kvec *head = buf->head;
net/sunrpc/xdr.c
710
head->iov_len + buf->page_len + base - shift;
net/sunrpc/xdr.c
713
if (WARN_ONCE(shift > head->iov_len + buf->page_len + base,
net/sunrpc/xdr.c
716
if (hdto + hdlen > head->iov_len)
net/sunrpc/xdr.c
717
hdlen = head->iov_len - hdto;
net/sunrpc/xdr.c
718
memcpy(head->iov_base + hdto, tail->iov_base + base, hdlen);
net/sunrpc/xdr.c
753
const struct kvec *head = buf->head;
net/sunrpc/xdr.c
754
unsigned int hdto = head->iov_len + base - shift;
net/sunrpc/xdr.c
757
if (WARN_ONCE(shift > head->iov_len + base,
net/sunrpc/xdr.c
760
if (hdto + hdlen > head->iov_len)
net/sunrpc/xdr.c
761
hdlen = head->iov_len - hdto;
net/sunrpc/xdr.c
762
_copy_from_pages(head->iov_base + hdto, buf->pages,
net/sunrpc/xdr.c
804
const struct kvec *head = buf->head;
net/sunrpc/xdr.c
818
if (base < head->iov_len) {
net/sunrpc/xdr.c
819
bytes = min_t(unsigned int, len, head->iov_len - base);
net/sunrpc/xdr.c
820
memmove(head->iov_base + (base - shift),
net/sunrpc/xdr.c
821
head->iov_base + base, bytes);
net/sunrpc/xdr.c
825
xdr_buf_pages_shift_left(buf, base - head->iov_len, len, shift);
net/sunrpc/xdr.c
839
struct kvec *head = buf->head;
net/sunrpc/xdr.c
842
WARN_ON_ONCE(len > head->iov_len);
net/sunrpc/xdr.c
843
if (head->iov_len > buflen) {
net/sunrpc/xdr.c
844
buf->buflen -= head->iov_len - buflen;
net/sunrpc/xdr.c
845
head->iov_len = buflen;
net/sunrpc/xdr.c
847
if (len >= head->iov_len)
net/sunrpc/xdr.c
849
shift = head->iov_len - len;
net/sunrpc/xdr.c
852
head->iov_len = len;
net/sunrpc/xdr.c
868
unsigned int shift, buflen = buf->len - buf->head->iov_len;
net/sunrpc/xdr.c
871
if (buf->head->iov_len >= buf->len || len > buflen)
net/sunrpc/xdr.c
907
xdr_stream_set_pos(xdr, pos + xdr->buf->head[0].iov_len);
net/sunrpc/xdr.c
918
WARN_ON(pos < xdr->buf->head[0].iov_len);
net/sunrpc/xdr.c
919
return pos - xdr->buf->head[0].iov_len;
net/sunrpc/xdr.c
940
struct kvec *iov = buf->head;
net/sunrpc/xprt.c
2011
xbufp->len = xbufp->head[0].iov_len + xbufp->page_len +
net/sunrpc/xprtmultipath.c
263
struct rpc_xprt *xprt_switch_find_first_entry(struct list_head *head)
net/sunrpc/xprtmultipath.c
267
list_for_each_entry_rcu(pos, head, xprt_switch) {
net/sunrpc/xprtmultipath.c
275
struct rpc_xprt *xprt_switch_find_first_entry_offline(struct list_head *head)
net/sunrpc/xprtmultipath.c
279
list_for_each_entry_rcu(pos, head, xprt_switch) {
net/sunrpc/xprtmultipath.c
297
struct rpc_xprt *_xprt_switch_find_current_entry(struct list_head *head,
net/sunrpc/xprtmultipath.c
304
list_for_each_entry_rcu(pos, head, xprt_switch) {
net/sunrpc/xprtmultipath.c
315
struct rpc_xprt *xprt_switch_find_current_entry(struct list_head *head,
net/sunrpc/xprtmultipath.c
318
return _xprt_switch_find_current_entry(head, cur, true);
net/sunrpc/xprtmultipath.c
323
struct rpc_xprt *first_entry(struct list_head *head),
net/sunrpc/xprtmultipath.c
324
struct rpc_xprt *current_entry(struct list_head *head,
net/sunrpc/xprtmultipath.c
328
struct list_head *head;
net/sunrpc/xprtmultipath.c
332
head = &xps->xps_xprt_list;
net/sunrpc/xprtmultipath.c
334
return first_entry(head);
net/sunrpc/xprtmultipath.c
335
return current_entry(head, xpi->xpi_cursor);
net/sunrpc/xprtmultipath.c
346
struct rpc_xprt *xprt_switch_find_current_entry_offline(struct list_head *head,
net/sunrpc/xprtmultipath.c
349
return _xprt_switch_find_current_entry(head, cur, false);
net/sunrpc/xprtmultipath.c
364
struct list_head *head;
net/sunrpc/xprtmultipath.c
370
head = &xps->xps_xprt_list;
net/sunrpc/xprtmultipath.c
371
list_for_each_entry_rcu(pos, head, xprt_switch) {
net/sunrpc/xprtmultipath.c
394
struct rpc_xprt *xprt_switch_find_next_entry(struct list_head *head,
net/sunrpc/xprtmultipath.c
400
list_for_each_entry_rcu(pos, head, xprt_switch) {
net/sunrpc/xprtmultipath.c
440
struct rpc_xprt *__xprt_switch_find_next_entry_roundrobin(struct list_head *head,
net/sunrpc/xprtmultipath.c
445
ret = xprt_switch_find_next_entry(head, cur, true);
net/sunrpc/xprtmultipath.c
448
return xprt_switch_find_first_entry(head);
net/sunrpc/xprtmultipath.c
455
struct list_head *head = &xps->xps_xprt_list;
net/sunrpc/xprtmultipath.c
462
xprt = __xprt_switch_find_next_entry_roundrobin(head, cur);
net/sunrpc/xprtrdma/backchannel.c
250
buf->head[0].iov_base = p;
net/sunrpc/xprtrdma/backchannel.c
251
buf->head[0].iov_len = size;
net/sunrpc/xprtrdma/rpc_rdma.c
1051
rqst->rq_rcv_buf.head[0].iov_base = srcp;
net/sunrpc/xprtrdma/rpc_rdma.c
1052
rqst->rq_private_buf.head[0].iov_base = srcp;
net/sunrpc/xprtrdma/rpc_rdma.c
1057
curlen = rqst->rq_rcv_buf.head[0].iov_len;
net/sunrpc/xprtrdma/rpc_rdma.c
1445
rep->rr_hdrbuf.head[0].iov_base, NULL);
net/sunrpc/xprtrdma/rpc_rdma.c
174
return (buf->head[0].iov_len + buf->tail[0].iov_len) <
net/sunrpc/xprtrdma/rpc_rdma.c
238
seg = rpcrdma_convert_kvec(&xdrbuf->head[0], seg, &n);
net/sunrpc/xprtrdma/rpc_rdma.c
347
pos = rqst->rq_snd_buf.head[0].iov_len;
net/sunrpc/xprtrdma/rpc_rdma.c
407
rqst->rq_rcv_buf.head[0].iov_len,
net/sunrpc/xprtrdma/rpc_rdma.c
668
dst = (unsigned char *)xdr->head[0].iov_base;
net/sunrpc/xprtrdma/rpc_rdma.c
669
dst += xdr->head[0].iov_len + xdr->page_len;
net/sunrpc/xprtrdma/rpc_rdma.c
684
dst = (unsigned char *)xdr->head[0].iov_base;
net/sunrpc/xprtrdma/rpc_rdma.c
685
dst += xdr->head[0].iov_len;
net/sunrpc/xprtrdma/rpc_rdma.c
732
if (!rpcrdma_prepare_head_iov(r_xprt, req, xdr->head[0].iov_len))
net/sunrpc/xprtrdma/rpc_rdma.c
752
if (!rpcrdma_prepare_head_iov(r_xprt, req, xdr->head[0].iov_len))
net/sunrpc/xprtrdma/svc_rdma_backchannel.c
26
struct kvec *dst, *src = &rcvbuf->head[0];
net/sunrpc/xprtrdma/svc_rdma_backchannel.c
36
dst = &req->rq_private_buf.head[0];
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
406
arg->head[0].iov_base = ctxt->rc_recv_buf;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
407
arg->head[0].iov_len = ctxt->rc_byte_len;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
678
rdma_argp = rq_arg->head[0].iov_base;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
710
rq_arg->head[0].iov_base = rctxt->rc_stream.p;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
712
rq_arg->head[0].iov_len -= hdr_len;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
800
buf->tail[0].iov_base = buf->head[0].iov_base + chunk->ch_position;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
801
buf->tail[0].iov_len = buf->head[0].iov_len - chunk->ch_position;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
802
buf->head[0].iov_len = chunk->ch_position;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
833
buf->head[0].iov_base = page_address(rqstp->rq_pages[0]);
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
834
buf->head[0].iov_len = min_t(size_t, PAGE_SIZE, ctxt->rc_readbytes);
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
836
buf->page_len = ctxt->rc_readbytes - buf->head[0].iov_len;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
852
buf->head[0].iov_base = page_address(rqstp->rq_pages[0]);
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
853
buf->head[0].iov_len = min_t(size_t, PAGE_SIZE, ctxt->rc_readbytes);
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
855
buf->page_len = ctxt->rc_readbytes - buf->head[0].iov_len;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1003
head->rc_readbytes += dummy.rs_length;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1023
struct svc_rdma_recv_ctxt *head)
net/sunrpc/xprtrdma/svc_rdma_rw.c
1026
pcl_first_chunk(&head->rc_call_pcl);
net/sunrpc/xprtrdma/svc_rdma_rw.c
1027
const struct svc_rdma_pcl *pcl = &head->rc_read_pcl;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1033
return svc_rdma_build_read_chunk(rqstp, head, call_chunk);
net/sunrpc/xprtrdma/svc_rdma_rw.c
1038
ret = svc_rdma_read_chunk_range(rqstp, head, call_chunk,
net/sunrpc/xprtrdma/svc_rdma_rw.c
1044
ret = svc_rdma_build_read_chunk(rqstp, head, chunk);
net/sunrpc/xprtrdma/svc_rdma_rw.c
1053
length = next->ch_position - head->rc_readbytes;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1054
ret = svc_rdma_read_chunk_range(rqstp, head, call_chunk,
net/sunrpc/xprtrdma/svc_rdma_rw.c
1062
return svc_rdma_read_chunk_range(rqstp, head, call_chunk,
net/sunrpc/xprtrdma/svc_rdma_rw.c
1086
struct svc_rdma_recv_ctxt *head)
net/sunrpc/xprtrdma/svc_rdma_rw.c
1088
return svc_rdma_read_call_chunk(rqstp, head);
net/sunrpc/xprtrdma/svc_rdma_rw.c
1102
struct svc_rdma_recv_ctxt *head)
net/sunrpc/xprtrdma/svc_rdma_rw.c
1106
for (i = 0; i < head->rc_page_count; i++) {
net/sunrpc/xprtrdma/svc_rdma_rw.c
1107
head->rc_pages[i] = rqstp->rq_pages[i];
net/sunrpc/xprtrdma/svc_rdma_rw.c
1137
struct svc_rdma_recv_ctxt *head)
net/sunrpc/xprtrdma/svc_rdma_rw.c
1139
struct svc_rdma_chunk_ctxt *cc = &head->rc_cc;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1144
head->rc_pageoff = 0;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1145
head->rc_curpage = 0;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1146
head->rc_readbytes = 0;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1148
if (pcl_is_empty(&head->rc_call_pcl)) {
net/sunrpc/xprtrdma/svc_rdma_rw.c
1149
if (head->rc_read_pcl.cl_count == 1)
net/sunrpc/xprtrdma/svc_rdma_rw.c
1150
ret = svc_rdma_read_data_item(rqstp, head);
net/sunrpc/xprtrdma/svc_rdma_rw.c
1152
ret = svc_rdma_read_multiple_chunks(rqstp, head);
net/sunrpc/xprtrdma/svc_rdma_rw.c
1154
ret = svc_rdma_read_special(rqstp, head);
net/sunrpc/xprtrdma/svc_rdma_rw.c
1155
svc_rdma_clear_rqst_pages(rqstp, head);
net/sunrpc/xprtrdma/svc_rdma_rw.c
577
info->wi_next_off = offset - xdr->head[0].iov_len;
net/sunrpc/xprtrdma/svc_rdma_rw.c
598
if (xdr->head[0].iov_len) {
net/sunrpc/xprtrdma/svc_rdma_rw.c
599
ret = svc_rdma_iov_write(info, &xdr->head[0]);
net/sunrpc/xprtrdma/svc_rdma_rw.c
605
ret = svc_rdma_pages_write(info, xdr, xdr->head[0].iov_len,
net/sunrpc/xprtrdma/svc_rdma_rw.c
748
struct svc_rdma_recv_ctxt *head,
net/sunrpc/xprtrdma/svc_rdma_rw.c
752
struct svc_rdma_chunk_ctxt *cc = &head->rc_cc;
net/sunrpc/xprtrdma/svc_rdma_rw.c
758
if (check_add_overflow(head->rc_pageoff, len, &total))
net/sunrpc/xprtrdma/svc_rdma_rw.c
768
PAGE_SIZE - head->rc_pageoff);
net/sunrpc/xprtrdma/svc_rdma_rw.c
770
if (!head->rc_pageoff)
net/sunrpc/xprtrdma/svc_rdma_rw.c
771
head->rc_page_count++;
net/sunrpc/xprtrdma/svc_rdma_rw.c
774
rqstp->rq_pages[head->rc_curpage],
net/sunrpc/xprtrdma/svc_rdma_rw.c
775
seg_len, head->rc_pageoff);
net/sunrpc/xprtrdma/svc_rdma_rw.c
777
head->rc_pageoff += seg_len;
net/sunrpc/xprtrdma/svc_rdma_rw.c
778
if (head->rc_pageoff == PAGE_SIZE) {
net/sunrpc/xprtrdma/svc_rdma_rw.c
779
head->rc_curpage++;
net/sunrpc/xprtrdma/svc_rdma_rw.c
780
head->rc_pageoff = 0;
net/sunrpc/xprtrdma/svc_rdma_rw.c
784
if (len && ((head->rc_curpage + 1) > rqstp->rq_maxpages))
net/sunrpc/xprtrdma/svc_rdma_rw.c
800
trace_svcrdma_page_overrun_err(&cc->cc_cid, head->rc_curpage);
net/sunrpc/xprtrdma/svc_rdma_rw.c
817
struct svc_rdma_recv_ctxt *head,
net/sunrpc/xprtrdma/svc_rdma_rw.c
825
ret = svc_rdma_build_read_segment(rqstp, head, segment);
net/sunrpc/xprtrdma/svc_rdma_rw.c
828
head->rc_readbytes += segment->rs_length;
net/sunrpc/xprtrdma/svc_rdma_rw.c
850
struct svc_rdma_recv_ctxt *head,
net/sunrpc/xprtrdma/svc_rdma_rw.c
854
unsigned char *dst, *src = head->rc_recv_buf;
net/sunrpc/xprtrdma/svc_rdma_rw.c
857
numpages = PAGE_ALIGN(head->rc_pageoff + remaining) >> PAGE_SHIFT;
net/sunrpc/xprtrdma/svc_rdma_rw.c
861
if (head->rc_curpage >= rqstp->rq_maxpages)
net/sunrpc/xprtrdma/svc_rdma_rw.c
865
PAGE_SIZE - head->rc_pageoff);
net/sunrpc/xprtrdma/svc_rdma_rw.c
867
if (!head->rc_pageoff)
net/sunrpc/xprtrdma/svc_rdma_rw.c
868
head->rc_page_count++;
net/sunrpc/xprtrdma/svc_rdma_rw.c
870
dst = page_address(rqstp->rq_pages[head->rc_curpage]);
net/sunrpc/xprtrdma/svc_rdma_rw.c
871
memcpy((unsigned char *)dst + head->rc_pageoff, src + offset, page_len);
net/sunrpc/xprtrdma/svc_rdma_rw.c
873
head->rc_readbytes += page_len;
net/sunrpc/xprtrdma/svc_rdma_rw.c
874
head->rc_pageoff += page_len;
net/sunrpc/xprtrdma/svc_rdma_rw.c
875
if (head->rc_pageoff == PAGE_SIZE) {
net/sunrpc/xprtrdma/svc_rdma_rw.c
876
head->rc_curpage++;
net/sunrpc/xprtrdma/svc_rdma_rw.c
877
head->rc_pageoff = 0;
net/sunrpc/xprtrdma/svc_rdma_rw.c
903
struct svc_rdma_recv_ctxt *head)
net/sunrpc/xprtrdma/svc_rdma_rw.c
905
const struct svc_rdma_pcl *pcl = &head->rc_read_pcl;
net/sunrpc/xprtrdma/svc_rdma_rw.c
913
ret = svc_rdma_copy_inline_range(rqstp, head, start, length);
net/sunrpc/xprtrdma/svc_rdma_rw.c
918
ret = svc_rdma_build_read_chunk(rqstp, head, chunk);
net/sunrpc/xprtrdma/svc_rdma_rw.c
927
length = next->ch_position - head->rc_readbytes;
net/sunrpc/xprtrdma/svc_rdma_rw.c
928
ret = svc_rdma_copy_inline_range(rqstp, head, start, length);
net/sunrpc/xprtrdma/svc_rdma_rw.c
934
length = head->rc_byte_len - start;
net/sunrpc/xprtrdma/svc_rdma_rw.c
935
return svc_rdma_copy_inline_range(rqstp, head, start, length);
net/sunrpc/xprtrdma/svc_rdma_rw.c
957
struct svc_rdma_recv_ctxt *head)
net/sunrpc/xprtrdma/svc_rdma_rw.c
959
return svc_rdma_build_read_chunk(rqstp, head,
net/sunrpc/xprtrdma/svc_rdma_rw.c
960
pcl_first_chunk(&head->rc_read_pcl));
net/sunrpc/xprtrdma/svc_rdma_rw.c
979
struct svc_rdma_recv_ctxt *head,
net/sunrpc/xprtrdma/svc_rdma_rw.c
999
ret = svc_rdma_build_read_segment(rqstp, head, &dummy);
net/sunrpc/xprtrdma/svc_rdma_sendto.c
638
ret = svc_rdma_iov_dma_map(data, &xdr->head[0]);
net/sunrpc/xprtrdma/svc_rdma_sendto.c
684
if (xdr->head[0].iov_len)
net/sunrpc/xprtrdma/svc_rdma_sendto.c
751
if (xdr->head[0].iov_len) {
net/sunrpc/xprtrdma/svc_rdma_sendto.c
752
memcpy(args->pd_dest, xdr->head[0].iov_base, xdr->head[0].iov_len);
net/sunrpc/xprtrdma/svc_rdma_sendto.c
753
args->pd_dest += xdr->head[0].iov_len;
net/sunrpc/xprtrdma/xprt_rdma.h
576
xdr->head[0].iov_len = len;
net/sunrpc/xprtsock.c
487
want = min_t(size_t, count, buf->head[0].iov_len);
net/sunrpc/xprtsock.c
489
ret = xs_read_kvec(sock, msg, flags, &buf->head[0], want, seek);
net/sunrpc/xprtsock.c
551
if (buf->head[0].iov_len >= transport->recv.offset)
net/sunrpc/xprtsock.c
552
memcpy(buf->head[0].iov_base,
net/sysctl_net.c
42
static int net_ctl_permissions(struct ctl_table_header *head,
net/sysctl_net.c
45
struct net *net = container_of(head->set, struct net, sysctls);
net/sysctl_net.c
56
static void net_ctl_set_ownership(struct ctl_table_header *head,
net/sysctl_net.c
59
struct net *net = container_of(head->set, struct net, sysctls);
net/tipc/crypto.c
409
struct tipc_tfm *tfm_entry, *head, *tmp;
net/tipc/crypto.c
414
head = *get_cpu_ptr(aead->tfm_entry);
net/tipc/crypto.c
416
list_for_each_entry_safe(tfm_entry, tmp, &head->list, list) {
net/tipc/crypto.c
422
crypto_free_aead(head->tfm);
net/tipc/crypto.c
423
list_del(&head->list);
net/tipc/crypto.c
424
kfree(head);
net/tipc/crypto.c
517
struct tipc_tfm *tfm_entry, *head;
net/tipc/crypto.c
574
head = tfm_entry;
net/tipc/crypto.c
576
*per_cpu_ptr(tmp->tfm_entry, cpu) = head;
net/tipc/crypto.c
579
list_add_tail(&tfm_entry->list, &head->list);
net/tipc/link.c
2556
u16 head = hskb ? msg_seqno(buf_msg(hskb)) : l->snd_nxt - 1;
net/tipc/link.c
2561
skb_queue_len(&l->transmq), head, tail,
net/tipc/monitor.c
324
static void mon_assign_roles(struct tipc_monitor *mon, struct tipc_peer *head)
net/tipc/monitor.c
326
struct tipc_peer *peer = peer_nxt(head);
net/tipc/monitor.c
334
if (i++ < head->applied) {
net/tipc/monitor.c
336
if (head == self)
net/tipc/monitor.c
345
head = peer;
net/tipc/monitor.c
346
head->is_head = true;
net/tipc/monitor.c
356
struct tipc_peer *peer, *prev, *head;
net/tipc/monitor.c
372
head = peer_head(prev);
net/tipc/monitor.c
373
if (head == self)
net/tipc/monitor.c
385
mon_assign_roles(mon, head);
net/tipc/monitor.c
426
struct tipc_peer *peer, *head;
net/tipc/monitor.c
433
head = peer_head(peer);
net/tipc/monitor.c
434
if (head == self)
net/tipc/monitor.c
436
mon_assign_roles(mon, head);
net/tipc/monitor.c
445
struct tipc_peer *peer, *head;
net/tipc/monitor.c
470
head = peer_head(peer);
net/tipc/monitor.c
471
if (head == self)
net/tipc/monitor.c
473
mon_assign_roles(mon, head);
net/tipc/msg.c
126
struct sk_buff *head = *headbuf;
net/tipc/msg.c
143
if (unlikely(head))
net/tipc/msg.c
151
head = *headbuf = frag;
net/tipc/msg.c
152
TIPC_SKB_CB(head)->tail = NULL;
net/tipc/msg.c
156
if (!head)
net/tipc/msg.c
164
if (skb_try_coalesce(head, frag, &headstolen, &delta)) {
net/tipc/msg.c
167
tail = TIPC_SKB_CB(head)->tail;
net/tipc/msg.c
168
if (!skb_has_frag_list(head))
net/tipc/msg.c
169
skb_shinfo(head)->frag_list = frag;
net/tipc/msg.c
172
head->truesize += frag->truesize;
net/tipc/msg.c
173
head->data_len += frag->len;
net/tipc/msg.c
174
head->len += frag->len;
net/tipc/msg.c
175
TIPC_SKB_CB(head)->tail = frag;
net/tipc/msg.c
179
TIPC_SKB_CB(head)->validated = 0;
net/tipc/msg.c
180
if (unlikely(!tipc_msg_validate(&head)))
net/tipc/msg.c
182
*buf = head;
net/tipc/msg.c
183
TIPC_SKB_CB(head)->tail = NULL;
net/tipc/msg.c
768
struct sk_buff *head = NULL;
net/tipc/msg.c
788
if (tipc_buf_append(&head, &frag))
net/tipc/msg.c
790
if (!head)
net/tipc/msg.c
797
kfree_skb(head);
net/tipc/msg.h
1265
struct sk_buff_head *head)
net/tipc/msg.h
1267
spin_lock_bh(&head->lock);
net/tipc/msg.h
1268
skb_queue_splice_tail(list, head);
net/tipc/msg.h
1269
spin_unlock_bh(&head->lock);
net/tipc/msg.h
1277
struct sk_buff_head *head)
net/tipc/msg.h
1286
tipc_skb_queue_splice_tail(&tmp, head);
net/tipc/name_distr.c
210
struct sk_buff_head head;
net/tipc/name_distr.c
213
__skb_queue_head_init(&head);
net/tipc/name_distr.c
221
named_distribute(net, &head, dnode, &nt->cluster_scope, seqno);
net/tipc/name_distr.c
222
tipc_node_xmit(net, &head, dnode, 0);
net/tipc/name_table.c
1057
struct hlist_head *head;
net/tipc/name_table.c
1068
head = &tn->nametbl->services[i];
net/tipc/name_table.c
1078
hlist_for_each_entry_rcu(service, head, service_list)
net/tipc/node.c
1762
struct sk_buff_head head;
net/tipc/node.c
1764
__skb_queue_head_init(&head);
net/tipc/node.c
1765
__skb_queue_tail(&head, skb);
net/tipc/node.c
1766
tipc_node_xmit(net, &head, dnode, selector);
net/tipc/socket.c
4001
i += tipc_skb_dump(sk->sk_backlog.head, false, buf + i);
net/tipc/socket.c
4002
if (sk->sk_backlog.tail != sk->sk_backlog.head) {
net/tipc/socket.c
540
static void tipc_sk_callback(struct rcu_head *head)
net/tipc/socket.c
542
struct tipc_sock *tsk = container_of(head, struct tipc_sock, rcu);
net/tls/tls_device_fallback.c
169
skb->csum_start = skb_transport_header(skb) - skb->head;
net/vmw_vsock/diag.c
103
struct list_head *head = &vsock_connected_table[bucket];
net/vmw_vsock/diag.c
106
list_for_each_entry(vsk, head, connected_table) {
net/vmw_vsock/diag.c
73
struct list_head *head = &vsock_bind_table[bucket];
net/vmw_vsock/diag.c
76
list_for_each_entry(vsk, head, bound_table) {
net/vmw_vsock/virtio_transport.c
326
memset(skb->head, 0, VIRTIO_VSOCK_SKB_HEADROOM);
net/vmw_vsock/vmci_transport_notify.c
186
u64 head;
net/vmw_vsock/vmci_transport_notify.c
201
vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair, &tail, &head);
net/vmw_vsock/vmci_transport_notify.c
202
room_left = vmci_trans(vsk)->consume_size - head;
net/vmw_vsock/vmci_transport_notify.c
208
waiting_info.offset = head + room_needed;
net/vmw_vsock/vmci_transport_notify.c
228
u64 head;
net/vmw_vsock/vmci_transport_notify.c
237
vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair, &tail, &head);
net/wireless/nl80211.c
6197
bcn->head = nla_data(attrs[NL80211_ATTR_BEACON_HEAD]);
net/wireless/trace.h
738
__dynamic_array(u8, head, info->beacon.head_len)
net/wireless/trace.h
749
if (info->beacon.head)
net/wireless/trace.h
750
memcpy(__get_dynamic_array(head),
net/wireless/trace.h
751
info->beacon.head,
net/wireless/util.c
677
frag_page = virt_to_head_page(skb->head);
net/xfrm/trace_iptfs.h
26
__field(void *, head)
net/xfrm/trace_iptfs.h
41
__entry->head = skb->head;
net/xfrm/trace_iptfs.h
52
__entry->head_pg_addr = page_address(virt_to_head_page(skb->head));
net/xfrm/trace_iptfs.h
63
__entry->tailroom, __entry->tail, __entry->end, __entry->head,
net/xfrm/xfrm_interface_core.c
859
static void xfrmi_dellink(struct net_device *dev, struct list_head *head)
net/xfrm/xfrm_interface_core.c
861
unregister_netdevice_queue(dev, head);
net/xfrm/xfrm_policy.c
1152
if (rhashtable_remove_fast(&xfrm_policy_inexact_table, &b->head,
net/xfrm/xfrm_policy.c
129
struct rhash_head head;
net/xfrm/xfrm_policy.c
1534
.head_offset = offsetof(struct xfrm_pol_inexact_bin, head),
net/xfrm/xfrm_policy.c
451
static void xfrm_policy_destroy_rcu(struct rcu_head *head)
net/xfrm/xfrm_policy.c
453
struct xfrm_policy *policy = container_of(head, struct xfrm_policy, rcu);
net/xfrm/xfrm_policy.c
779
&bin->k, &bin->head,
samples/bpf/test_lru_dist.c
55
static inline void list_add(struct list_head *new, struct list_head *head)
samples/bpf/test_lru_dist.c
57
__list_add(new, head, head->next);
samples/bpf/test_lru_dist.c
71
static inline void list_move(struct list_head *list, struct list_head *head)
samples/bpf/test_lru_dist.c
74
list_add(list, head);
samples/vfio-mdev/mtty.c
114
u8 head, tail;
samples/vfio-mdev/mtty.c
384
mdev_state->s[index].rxtx.head] = data;
samples/vfio-mdev/mtty.c
386
CIRCULAR_BUF_INC_IDX(mdev_state->s[index].rxtx.head);
samples/vfio-mdev/mtty.c
429
(mdev_state->s[index].rxtx.head ==
samples/vfio-mdev/mtty.c
450
mdev_state->s[index].rxtx.head = 0;
samples/vfio-mdev/mtty.c
546
if (mdev_state->s[index].rxtx.head !=
samples/vfio-mdev/mtty.c
554
if (mdev_state->s[index].rxtx.head ==
samples/vfio-mdev/mtty.c
597
(mdev_state->s[index].rxtx.head ==
samples/vfio-mdev/mtty.c
628
if (mdev_state->s[index].rxtx.head !=
samples/vfio-mdev/mtty.c
637
if (mdev_state->s[index].rxtx.head ==
scripts/dtc/livetree.c
76
struct property *head = NULL;
scripts/dtc/livetree.c
81
p->next = head;
scripts/dtc/livetree.c
82
head = p;
scripts/dtc/livetree.c
85
return head;
scripts/include/list.h
165
static inline void list_move(struct list_head *list, struct list_head *head)
scripts/include/list.h
168
list_add(list, head);
scripts/include/list.h
177
struct list_head *head)
scripts/include/list.h
180
list_add_tail(list, head);
scripts/include/list.h
188
static inline int list_is_first(const struct list_head *list, const struct list_head *head)
scripts/include/list.h
190
return list->prev == head;
scripts/include/list.h
198
static inline int list_is_last(const struct list_head *list, const struct list_head *head)
scripts/include/list.h
200
return list->next == head;
scripts/include/list.h
208
static inline int list_is_head(const struct list_head *list, const struct list_head *head)
scripts/include/list.h
210
return list == head;
scripts/include/list.h
217
static inline int list_empty(const struct list_head *head)
scripts/include/list.h
219
return head->next == head;
scripts/include/list.h
275
#define list_entry_is_head(pos, head, member) \
scripts/include/list.h
276
(&pos->member == (head))
scripts/include/list.h
284
#define list_for_each_entry(pos, head, member) \
scripts/include/list.h
285
for (pos = list_first_entry(head, typeof(*pos), member); \
scripts/include/list.h
286
!list_entry_is_head(pos, head, member); \
scripts/include/list.h
295
#define list_for_each_entry_reverse(pos, head, member) \
scripts/include/list.h
296
for (pos = list_last_entry(head, typeof(*pos), member); \
scripts/include/list.h
297
!list_entry_is_head(pos, head, member); \
scripts/include/list.h
307
#define list_for_each_entry_safe(pos, n, head, member) \
scripts/include/list.h
308
for (pos = list_first_entry(head, typeof(*pos), member), \
scripts/include/list.h
310
!list_entry_is_head(pos, head, member); \
scripts/include/list.h
411
#define hlist_for_each_entry(pos, head, member) \
scripts/include/list.h
412
for (pos = hlist_entry_safe((head)->first, typeof(*(pos)), member);\
scripts/include/list.h
423
#define hlist_for_each_entry_safe(pos, n, head, member) \
scripts/include/list.h
424
for (pos = hlist_entry_safe((head)->first, typeof(*pos), member);\
scripts/include/list.h
81
static inline void list_add(struct list_head *new, struct list_head *head)
scripts/include/list.h
83
__list_add(new, head, head->next);
scripts/include/list.h
94
static inline void list_add_tail(struct list_head *new, struct list_head *head)
scripts/include/list.h
96
__list_add(new, head->prev, head);
scripts/kconfig/lkc.h
103
struct gstr get_relations_str(struct symbol **sym_arr, struct list_head *head);
scripts/kconfig/mconf.c
427
LIST_HEAD(head);
scripts/kconfig/mconf.c
429
.head = &head,
scripts/kconfig/mconf.c
434
res = get_relations_str(sym_arr, &head);
scripts/kconfig/mconf.c
446
list_for_each_entry_safe(pos, tmp, &head, entries)
scripts/kconfig/menu.c
642
struct list_head *head)
scripts/kconfig/menu.c
668
if (head && location) {
scripts/kconfig/menu.c
671
list_add_tail(&jump->entries, head);
scripts/kconfig/menu.c
722
struct list_head *head)
scripts/kconfig/menu.c
745
get_prompt_str(r, menu->prompt, head);
scripts/kconfig/menu.c
773
struct gstr get_relations_str(struct symbol **sym_arr, struct list_head *head)
scripts/kconfig/menu.c
780
get_symbol_str(&res, sym, head);
scripts/kconfig/mnconf-common.c
30
list_for_each_entry(pos, data->head, entries) {
scripts/kconfig/mnconf-common.h
10
struct list_head *head;
scripts/kconfig/nconf.c
738
LIST_HEAD(head);
scripts/kconfig/nconf.c
740
.head = &head,
scripts/kconfig/nconf.c
744
res = get_relations_str(sym_arr, &head);
scripts/mod/modpost.c
288
static bool contains_namespace(struct list_head *head, const char *namespace)
scripts/mod/modpost.c
299
list_for_each_entry(list, head, list) {
scripts/mod/modpost.c
307
static void add_namespace(struct list_head *head, const char *namespace)
scripts/mod/modpost.c
311
if (!contains_namespace(head, namespace)) {
scripts/mod/modpost.c
314
list_add_tail(&ns_entry->list, head);
security/apparmor/domain.c
390
struct aa_ns *ns, struct list_head *head,
security/apparmor/domain.c
398
AA_BUG(!head);
security/apparmor/domain.c
402
list_for_each_entry_rcu(profile, head, base.list) {
security/apparmor/include/lib.h
233
static inline struct aa_policy *__policy_find(struct list_head *head,
security/apparmor/include/lib.h
238
list_for_each_entry_rcu(policy, head, list) {
security/apparmor/include/lib.h
258
static inline struct aa_policy *__policy_strn_find(struct list_head *head,
security/apparmor/include/lib.h
263
list_for_each_entry_rcu(policy, head, list) {
security/apparmor/include/policy.h
178
struct rhash_head head;
security/apparmor/include/policy.h
311
void __aa_profile_list_release(struct list_head *head);
security/apparmor/include/policy_ns.h
141
static inline struct aa_ns *__aa_findn_ns(struct list_head *head,
security/apparmor/include/policy_ns.h
144
return (struct aa_ns *)__policy_strn_find(head, name, n);
security/apparmor/include/policy_ns.h
147
static inline struct aa_ns *__aa_find_ns(struct list_head *head,
security/apparmor/include/policy_ns.h
150
return __aa_findn_ns(head, name, strlen(name));
security/apparmor/label.c
369
static void label_free_rcu(struct rcu_head *head)
security/apparmor/label.c
371
struct aa_label *label = container_of(head, struct aa_label, rcu);
security/apparmor/lsm.c
2154
if (!list_empty(&cache->head)) {
security/apparmor/lsm.c
2155
aa_buf = list_first_entry(&cache->head, union aa_buffer, list);
security/apparmor/lsm.c
2239
list_add(&aa_buf->list, &cache->head);
security/apparmor/lsm.c
2290
INIT_LIST_HEAD(&per_cpu(aa_local_buffers, i).head);
security/apparmor/lsm.c
58
struct list_head head;
security/apparmor/policy.c
243
void __aa_profile_list_release(struct list_head *head)
security/apparmor/policy.c
246
list_for_each_entry_safe(profile, tmp, head, base.list)
security/apparmor/policy.c
461
static struct aa_profile *__strn_find_child(struct list_head *head,
security/apparmor/policy.c
464
return (struct aa_profile *)__policy_strn_find(head, name, len);
security/apparmor/policy.c
476
static struct aa_profile *__find_child(struct list_head *head, const char *name)
security/apparmor/policy.c
478
return __strn_find_child(head, name, strlen(name));
security/apparmor/policy_ns.c
300
static void __ns_list_release(struct list_head *head);
security/apparmor/policy_ns.c
350
static void __ns_list_release(struct list_head *head)
security/apparmor/policy_ns.c
354
list_for_each_entry_safe(ns, tmp, head, base.list)
security/apparmor/policy_unpack.c
1356
params.head_offset = offsetof(struct aa_data, head);
security/apparmor/policy_unpack.c
1383
if (rhashtable_insert_fast(profile->data, &data->head,
security/keys/trusted-keys/trusted_tpm2.c
484
struct tpm_header *head;
security/keys/trusted-keys/trusted_tpm2.c
528
head = (struct tpm_header *)buf.data;
security/keys/trusted-keys/trusted_tpm2.c
530
head->tag = cpu_to_be16(TPM2_ST_NO_SESSIONS);
security/keys/user_defined.c
89
static void user_free_payload_rcu(struct rcu_head *head)
security/keys/user_defined.c
93
payload = container_of(head, struct user_key_payload, rcu);
security/security.c
1200
static void inode_free_by_rcu(struct rcu_head *head)
security/security.c
1203
call_void_hook(inode_free_security_rcu, head);
security/security.c
1204
kmem_cache_free(lsm_inode_cache, head);
security/selinux/avc.c
148
struct hlist_head *head;
security/selinux/avc.c
155
head = &selinux_avc.avc_cache.slots[i];
security/selinux/avc.c
156
if (!hlist_empty(head)) {
security/selinux/avc.c
159
hlist_for_each_entry_rcu(node, head, list)
security/selinux/avc.c
464
struct hlist_head *head;
security/selinux/avc.c
470
head = &selinux_avc.avc_cache.slots[hvalue];
security/selinux/avc.c
477
hlist_for_each_entry(node, head, list) {
security/selinux/avc.c
525
struct hlist_head *head;
security/selinux/avc.c
528
head = &selinux_avc.avc_cache.slots[hvalue];
security/selinux/avc.c
529
hlist_for_each_entry_rcu(node, head, list) {
security/selinux/avc.c
613
struct hlist_head *head;
security/selinux/avc.c
629
head = &selinux_avc.avc_cache.slots[hvalue];
security/selinux/avc.c
632
hlist_for_each_entry(pos, head, list) {
security/selinux/avc.c
640
hlist_add_head_rcu(&node->list, head);
security/selinux/avc.c
838
struct hlist_head *head;
security/selinux/avc.c
850
head = &selinux_avc.avc_cache.slots[hvalue];
security/selinux/avc.c
855
hlist_for_each_entry(pos, head, list) {
security/selinux/avc.c
927
struct hlist_head *head;
security/selinux/avc.c
934
head = &selinux_avc.avc_cache.slots[i];
security/selinux/avc.c
943
hlist_for_each_entry(node, head, list)
security/selinux/ss/mls.c
108
head = -2;
security/selinux/ss/mls.c
115
if (prev != head) {
security/selinux/ss/mls.c
116
if (prev - head > 1)
security/selinux/ss/mls.c
131
head = i;
security/selinux/ss/mls.c
136
if (prev != head) {
security/selinux/ss/mls.c
137
if (prev - head > 1)
security/selinux/ss/mls.c
34
int i, l, len, head, prev;
security/selinux/ss/mls.c
48
head = -2;
security/selinux/ss/mls.c
55
if (head != prev) {
security/selinux/ss/mls.c
61
head = i;
security/selinux/ss/mls.c
65
if (prev != head) {
security/selinux/ss/mls.c
90
int i, l, head, prev;
security/selinux/ss/policydb.c
2212
for (l = NULL, c = genfs->head; c; l = c, c = c->next) {
security/selinux/ss/policydb.c
2231
genfs->head = newc;
security/selinux/ss/policydb.c
3468
for (c = genfs->head; c; c = c->next)
security/selinux/ss/policydb.c
3474
for (c = genfs->head; c; c = c->next) {
security/selinux/ss/policydb.c
821
c = g->head;
security/selinux/ss/policydb.c
869
struct ocontext *head, *c;
security/selinux/ss/policydb.c
882
head = p->ocontexts[OCON_ISID];
security/selinux/ss/policydb.c
883
for (c = head; c; c = c->next) {
security/selinux/ss/policydb.h
208
struct ocontext *head;
security/selinux/ss/services.c
2920
for (c = genfs->head; c; c = c->next) {
security/smack/smack_access.c
413
struct hlist_head *head;
security/smack/smack_access.c
416
head = &smack_known_hash[hash & (SMACK_HASH_SLOTS - 1)];
security/smack/smack_access.c
418
hlist_add_head_rcu(&skp->smk_hashed, head);
security/smack/smack_access.c
432
struct hlist_head *head;
security/smack/smack_access.c
436
head = &smack_known_hash[hash & (SMACK_HASH_SLOTS - 1)];
security/smack/smack_access.c
438
hlist_for_each_entry_rcu(skp, head, smk_hashed)
security/smack/smackfs.c
534
struct list_head *head)
security/smack/smackfs.c
540
for (list = rcu_dereference(list_next_rcu(head));
security/smack/smackfs.c
541
list != head;
security/smack/smackfs.c
551
struct list_head *head)
security/smack/smackfs.c
558
return (list == head) ? NULL : list;
security/tomoyo/audit.c
440
void tomoyo_read_log(struct tomoyo_io_buffer *head)
security/tomoyo/audit.c
444
if (head->r.w_pos)
security/tomoyo/audit.c
446
kfree(head->read_buf);
security/tomoyo/audit.c
447
head->read_buf = NULL;
security/tomoyo/audit.c
457
head->read_buf = ptr->log;
security/tomoyo/audit.c
458
head->r.w[head->r.w_pos++] = head->read_buf;
security/tomoyo/common.c
1024
head->w.domain = domain;
security/tomoyo/common.c
1026
if (!head->read_buf)
security/tomoyo/common.c
1028
memset(&head->r, 0, sizeof(head->r));
security/tomoyo/common.c
1029
head->r.print_this_domain_only = true;
security/tomoyo/common.c
1031
head->r.domain = &domain->list;
security/tomoyo/common.c
1033
head->r.eof = true;
security/tomoyo/common.c
1034
tomoyo_io_printf(head, "# select %s\n", data);
security/tomoyo/common.c
1036
tomoyo_io_printf(head, "# This is a deleted domain.\n");
security/tomoyo/common.c
1051
const struct tomoyo_task_acl *p1 = container_of(a, typeof(*p1), head);
security/tomoyo/common.c
1052
const struct tomoyo_task_acl *p2 = container_of(b, typeof(*p2), head);
security/tomoyo/common.c
1073
.head.type = TOMOYO_TYPE_MANUAL_TASK_ACL,
security/tomoyo/common.c
1078
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/common.c
1180
static int tomoyo_write_domain(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
1182
__must_hold(&head->io_sem)
security/tomoyo/common.c
1184
char *data = head->write_buf;
security/tomoyo/common.c
1186
struct tomoyo_domain_info *domain = head->w.domain;
security/tomoyo/common.c
1187
const bool is_delete = head->w.is_delete;
security/tomoyo/common.c
1201
head->w.domain = domain;
security/tomoyo/common.c
1242
static bool tomoyo_print_condition(struct tomoyo_io_buffer *head,
security/tomoyo/common.c
1244
__must_hold(&head->io_sem)
security/tomoyo/common.c
1246
switch (head->r.cond_step) {
security/tomoyo/common.c
1248
head->r.cond_index = 0;
security/tomoyo/common.c
1249
head->r.cond_step++;
security/tomoyo/common.c
1251
tomoyo_set_space(head);
security/tomoyo/common.c
1252
tomoyo_set_string(head, cond->transit->name);
security/tomoyo/common.c
1271
for (skip = 0; skip < head->r.cond_index; skip++) {
security/tomoyo/common.c
1296
while (head->r.cond_index < condc) {
security/tomoyo/common.c
1301
if (!tomoyo_flush(head))
security/tomoyo/common.c
1304
head->r.cond_index++;
security/tomoyo/common.c
1305
tomoyo_set_space(head);
security/tomoyo/common.c
1308
tomoyo_io_printf(head,
security/tomoyo/common.c
1311
tomoyo_set_string(head,
security/tomoyo/common.c
1313
tomoyo_set_string(head, "\"");
security/tomoyo/common.c
1317
tomoyo_set_string(head,
security/tomoyo/common.c
1319
tomoyo_set_string(head,
security/tomoyo/common.c
1321
tomoyo_io_printf(head, "\"]%s=", envp->is_not ? "!" : "");
security/tomoyo/common.c
1323
tomoyo_set_string(head, "\"");
security/tomoyo/common.c
1324
tomoyo_set_string(head, envp->value->name);
security/tomoyo/common.c
1325
tomoyo_set_string(head, "\"");
security/tomoyo/common.c
1327
tomoyo_set_string(head,
security/tomoyo/common.c
1334
(head, numbers_p++);
security/tomoyo/common.c
1337
tomoyo_set_string(head,
security/tomoyo/common.c
1341
tomoyo_set_string(head, match ? "=" : "!=");
security/tomoyo/common.c
1345
(head, names_p++);
security/tomoyo/common.c
1349
(head, numbers_p++);
security/tomoyo/common.c
1352
tomoyo_set_string(head,
security/tomoyo/common.c
1358
head->r.cond_step++;
security/tomoyo/common.c
1361
if (!tomoyo_flush(head))
security/tomoyo/common.c
1363
head->r.cond_step++;
security/tomoyo/common.c
1367
tomoyo_io_printf(head, " grant_log=%s",
security/tomoyo/common.c
1370
tomoyo_set_lf(head);
security/tomoyo/common.c
1384
static void tomoyo_set_group(struct tomoyo_io_buffer *head,
security/tomoyo/common.c
1386
__must_hold(&head->io_sem)
security/tomoyo/common.c
1388
if (head->type == TOMOYO_EXCEPTIONPOLICY) {
security/tomoyo/common.c
1389
tomoyo_print_namespace(head);
security/tomoyo/common.c
1390
tomoyo_io_printf(head, "acl_group %u ",
security/tomoyo/common.c
1391
head->r.acl_group_index);
security/tomoyo/common.c
1393
tomoyo_set_string(head, category);
security/tomoyo/common.c
1404
static bool tomoyo_print_entry(struct tomoyo_io_buffer *head,
security/tomoyo/common.c
1406
__must_hold(&head->io_sem)
security/tomoyo/common.c
1412
if (head->r.print_cond_part)
security/tomoyo/common.c
1416
if (!tomoyo_flush(head))
security/tomoyo/common.c
1420
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1426
if (head->r.print_transition_related_only &&
security/tomoyo/common.c
1430
tomoyo_set_group(head, "file ");
security/tomoyo/common.c
1433
tomoyo_set_slash(head);
security/tomoyo/common.c
1435
tomoyo_set_string(head, tomoyo_path_keyword[bit]);
security/tomoyo/common.c
1439
tomoyo_print_name_union(head, &ptr->name);
security/tomoyo/common.c
1442
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1444
tomoyo_set_group(head, "task ");
security/tomoyo/common.c
1445
tomoyo_set_string(head, "manual_domain_transition ");
security/tomoyo/common.c
1446
tomoyo_set_string(head, ptr->domainname->name);
security/tomoyo/common.c
1447
} else if (head->r.print_transition_related_only) {
security/tomoyo/common.c
1451
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1458
tomoyo_set_group(head, "file ");
security/tomoyo/common.c
1461
tomoyo_set_slash(head);
security/tomoyo/common.c
1463
tomoyo_set_string(head, tomoyo_mac_keywords
security/tomoyo/common.c
1468
tomoyo_print_name_union(head, &ptr->name1);
security/tomoyo/common.c
1469
tomoyo_print_name_union(head, &ptr->name2);
security/tomoyo/common.c
1472
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1479
tomoyo_set_group(head, "file ");
security/tomoyo/common.c
1482
tomoyo_set_slash(head);
security/tomoyo/common.c
1484
tomoyo_set_string(head, tomoyo_mac_keywords
security/tomoyo/common.c
1489
tomoyo_print_name_union(head, &ptr->name);
security/tomoyo/common.c
1490
tomoyo_print_number_union(head, &ptr->number);
security/tomoyo/common.c
1493
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1500
tomoyo_set_group(head, "file ");
security/tomoyo/common.c
1503
tomoyo_set_slash(head);
security/tomoyo/common.c
1505
tomoyo_set_string(head, tomoyo_mac_keywords
security/tomoyo/common.c
1510
tomoyo_print_name_union(head, &ptr->name);
security/tomoyo/common.c
1511
tomoyo_print_number_union(head, &ptr->mode);
security/tomoyo/common.c
1512
tomoyo_print_number_union(head, &ptr->major);
security/tomoyo/common.c
1513
tomoyo_print_number_union(head, &ptr->minor);
security/tomoyo/common.c
1516
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1523
tomoyo_set_group(head, "network inet ");
security/tomoyo/common.c
1524
tomoyo_set_string(head, tomoyo_proto_keyword
security/tomoyo/common.c
1526
tomoyo_set_space(head);
security/tomoyo/common.c
1529
tomoyo_set_slash(head);
security/tomoyo/common.c
1531
tomoyo_set_string(head, tomoyo_socket_keyword[bit]);
security/tomoyo/common.c
1535
tomoyo_set_space(head);
security/tomoyo/common.c
1537
tomoyo_set_string(head, "@");
security/tomoyo/common.c
1538
tomoyo_set_string(head, ptr->address.group->group_name
security/tomoyo/common.c
1544
tomoyo_io_printf(head, "%s", buf);
security/tomoyo/common.c
1546
tomoyo_print_number_union(head, &ptr->port);
security/tomoyo/common.c
1549
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1556
tomoyo_set_group(head, "network unix ");
security/tomoyo/common.c
1557
tomoyo_set_string(head, tomoyo_proto_keyword
security/tomoyo/common.c
1559
tomoyo_set_space(head);
security/tomoyo/common.c
1562
tomoyo_set_slash(head);
security/tomoyo/common.c
1564
tomoyo_set_string(head, tomoyo_socket_keyword[bit]);
security/tomoyo/common.c
1568
tomoyo_print_name_union(head, &ptr->name);
security/tomoyo/common.c
1571
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1573
tomoyo_set_group(head, "file mount");
security/tomoyo/common.c
1574
tomoyo_print_name_union(head, &ptr->dev_name);
security/tomoyo/common.c
1575
tomoyo_print_name_union(head, &ptr->dir_name);
security/tomoyo/common.c
1576
tomoyo_print_name_union(head, &ptr->fs_type);
security/tomoyo/common.c
1577
tomoyo_print_number_union(head, &ptr->flags);
security/tomoyo/common.c
1580
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1582
tomoyo_set_group(head, "misc env ");
security/tomoyo/common.c
1583
tomoyo_set_string(head, ptr->env->name);
security/tomoyo/common.c
1586
head->r.print_cond_part = true;
security/tomoyo/common.c
1587
head->r.cond_step = 0;
security/tomoyo/common.c
1588
if (!tomoyo_flush(head))
security/tomoyo/common.c
1591
if (!tomoyo_print_condition(head, acl->cond))
security/tomoyo/common.c
1593
head->r.print_cond_part = false;
security/tomoyo/common.c
1595
tomoyo_set_lf(head);
security/tomoyo/common.c
1610
static bool tomoyo_read_domain2(struct tomoyo_io_buffer *head,
security/tomoyo/common.c
1613
__must_hold(&head->io_sem)
security/tomoyo/common.c
1615
list_for_each_cookie(head->r.acl, list) {
security/tomoyo/common.c
1617
list_entry(head->r.acl, typeof(*ptr), list);
security/tomoyo/common.c
1619
if (!tomoyo_print_entry(head, ptr))
security/tomoyo/common.c
1622
head->r.acl = NULL;
security/tomoyo/common.c
1633
static void tomoyo_read_domain(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
1635
__must_hold(&head->io_sem)
security/tomoyo/common.c
1637
if (head->r.eof)
security/tomoyo/common.c
1639
list_for_each_cookie(head->r.domain, &tomoyo_domain_list) {
security/tomoyo/common.c
1641
list_entry(head->r.domain, typeof(*domain), list);
security/tomoyo/common.c
1644
switch (head->r.step) {
security/tomoyo/common.c
1647
!head->r.print_this_domain_only)
security/tomoyo/common.c
1650
tomoyo_set_string(head, domain->domainname->name);
security/tomoyo/common.c
1651
tomoyo_set_lf(head);
security/tomoyo/common.c
1652
tomoyo_io_printf(head, "use_profile %u\n",
security/tomoyo/common.c
1656
tomoyo_set_string(head, tomoyo_dif[i]);
security/tomoyo/common.c
1657
head->r.index = 0;
security/tomoyo/common.c
1658
head->r.step++;
security/tomoyo/common.c
1661
while (head->r.index < TOMOYO_MAX_ACL_GROUPS) {
security/tomoyo/common.c
1662
i = head->r.index++;
security/tomoyo/common.c
1665
tomoyo_io_printf(head, "use_group %u\n", i);
security/tomoyo/common.c
1666
if (!tomoyo_flush(head))
security/tomoyo/common.c
1669
head->r.index = 0;
security/tomoyo/common.c
1670
head->r.step++;
security/tomoyo/common.c
1671
tomoyo_set_lf(head);
security/tomoyo/common.c
1674
if (!tomoyo_read_domain2(head, &domain->acl_info_list))
security/tomoyo/common.c
1676
head->r.step++;
security/tomoyo/common.c
1677
if (!tomoyo_set_lf(head))
security/tomoyo/common.c
1681
head->r.step = 0;
security/tomoyo/common.c
1682
if (head->r.print_this_domain_only)
security/tomoyo/common.c
1687
head->r.eof = true;
security/tomoyo/common.c
1697
static int tomoyo_write_pid(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
1699
head->r.eof = false;
security/tomoyo/common.c
1713
static void tomoyo_read_pid(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
1714
__must_hold(&head->io_sem)
security/tomoyo/common.c
1716
char *buf = head->write_buf;
security/tomoyo/common.c
1724
head->r.eof = true;
security/tomoyo/common.c
1727
if (head->r.w_pos || head->r.eof)
security/tomoyo/common.c
1729
head->r.eof = true;
security/tomoyo/common.c
1744
tomoyo_io_printf(head, "%u %u ", pid, domain->profile);
security/tomoyo/common.c
1745
tomoyo_set_string(head, domain->domainname->name);
security/tomoyo/common.c
1774
static int tomoyo_write_exception(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
1776
__must_hold(&head->io_sem)
security/tomoyo/common.c
1778
const bool is_delete = head->w.is_delete;
security/tomoyo/common.c
1780
.ns = head->w.ns,
security/tomoyo/common.c
1782
.data = head->write_buf,
security/tomoyo/common.c
1801
(head->w.ns, &head->w.ns->acl_group[group],
security/tomoyo/common.c
1817
static bool tomoyo_read_group(struct tomoyo_io_buffer *head, const int idx)
security/tomoyo/common.c
1819
__must_hold(&head->io_sem)
security/tomoyo/common.c
1822
container_of(head->r.ns, typeof(*ns), namespace_list);
security/tomoyo/common.c
1825
list_for_each_cookie(head->r.group, list) {
security/tomoyo/common.c
1827
list_entry(head->r.group, typeof(*group), head.list);
security/tomoyo/common.c
1829
list_for_each_cookie(head->r.acl, &group->member_list) {
security/tomoyo/common.c
1831
list_entry(head->r.acl, typeof(*ptr), list);
security/tomoyo/common.c
1835
if (!tomoyo_flush(head))
security/tomoyo/common.c
1837
tomoyo_print_namespace(head);
security/tomoyo/common.c
1838
tomoyo_set_string(head, tomoyo_group_name[idx]);
security/tomoyo/common.c
1839
tomoyo_set_string(head, group->group_name->name);
security/tomoyo/common.c
1841
tomoyo_set_space(head);
security/tomoyo/common.c
1842
tomoyo_set_string(head, container_of
security/tomoyo/common.c
1844
head)->member_name->name);
security/tomoyo/common.c
1846
tomoyo_print_number_union(head, &container_of
security/tomoyo/common.c
1849
head)->number);
security/tomoyo/common.c
1854
head);
security/tomoyo/common.c
1858
tomoyo_io_printf(head, " %s", buffer);
security/tomoyo/common.c
1860
tomoyo_set_lf(head);
security/tomoyo/common.c
1862
head->r.acl = NULL;
security/tomoyo/common.c
1864
head->r.group = NULL;
security/tomoyo/common.c
1878
static bool tomoyo_read_policy(struct tomoyo_io_buffer *head, const int idx)
security/tomoyo/common.c
1882
container_of(head->r.ns, typeof(*ns), namespace_list);
security/tomoyo/common.c
1885
list_for_each_cookie(head->r.acl, list) {
security/tomoyo/common.c
1887
container_of(head->r.acl, typeof(*acl), list);
security/tomoyo/common.c
1890
if (!tomoyo_flush(head))
security/tomoyo/common.c
1896
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1898
tomoyo_print_namespace(head);
security/tomoyo/common.c
1899
tomoyo_set_string(head, tomoyo_transition_type
security/tomoyo/common.c
1901
tomoyo_set_string(head, ptr->program ?
security/tomoyo/common.c
1903
tomoyo_set_string(head, " from ");
security/tomoyo/common.c
1904
tomoyo_set_string(head, ptr->domainname ?
security/tomoyo/common.c
1912
container_of(acl, typeof(*ptr), head);
security/tomoyo/common.c
1914
tomoyo_print_namespace(head);
security/tomoyo/common.c
1915
tomoyo_set_string(head, "aggregator ");
security/tomoyo/common.c
1916
tomoyo_set_string(head,
security/tomoyo/common.c
1918
tomoyo_set_space(head);
security/tomoyo/common.c
1919
tomoyo_set_string(head,
security/tomoyo/common.c
1926
tomoyo_set_lf(head);
security/tomoyo/common.c
1928
head->r.acl = NULL;
security/tomoyo/common.c
1939
static void tomoyo_read_exception(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
1941
__must_hold(&head->io_sem)
security/tomoyo/common.c
1944
container_of(head->r.ns, typeof(*ns), namespace_list);
security/tomoyo/common.c
1946
if (head->r.eof)
security/tomoyo/common.c
1948
while (head->r.step < TOMOYO_MAX_POLICY &&
security/tomoyo/common.c
1949
tomoyo_read_policy(head, head->r.step))
security/tomoyo/common.c
1950
head->r.step++;
security/tomoyo/common.c
1951
if (head->r.step < TOMOYO_MAX_POLICY)
security/tomoyo/common.c
1953
while (head->r.step < TOMOYO_MAX_POLICY + TOMOYO_MAX_GROUP &&
security/tomoyo/common.c
1954
tomoyo_read_group(head, head->r.step - TOMOYO_MAX_POLICY))
security/tomoyo/common.c
1955
head->r.step++;
security/tomoyo/common.c
1956
if (head->r.step < TOMOYO_MAX_POLICY + TOMOYO_MAX_GROUP)
security/tomoyo/common.c
1958
while (head->r.step < TOMOYO_MAX_POLICY + TOMOYO_MAX_GROUP
security/tomoyo/common.c
1960
head->r.acl_group_index = head->r.step - TOMOYO_MAX_POLICY
security/tomoyo/common.c
1962
if (!tomoyo_read_domain2(head, &ns->acl_group
security/tomoyo/common.c
1963
[head->r.acl_group_index]))
security/tomoyo/common.c
1965
head->r.step++;
security/tomoyo/common.c
1967
head->r.eof = true;
security/tomoyo/common.c
2026
static bool tomoyo_numscan(const char *str, const char *head, int *width, const char tail)
security/tomoyo/common.c
2029
const int n = strlen(head);
security/tomoyo/common.c
2031
if (!strncmp(str, head, n)) {
security/tomoyo/common.c
205
static bool tomoyo_flush(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
207
while (head->r.w_pos) {
security/tomoyo/common.c
208
const char *w = head->r.w[0];
security/tomoyo/common.c
212
if (len > head->read_user_buf_avail)
security/tomoyo/common.c
213
len = head->read_user_buf_avail;
security/tomoyo/common.c
216
if (copy_to_user(head->read_user_buf, w, len))
security/tomoyo/common.c
218
head->read_user_buf_avail -= len;
security/tomoyo/common.c
219
head->read_user_buf += len;
security/tomoyo/common.c
222
head->r.w[0] = w;
security/tomoyo/common.c
226
if (head->poll) {
security/tomoyo/common.c
227
if (!head->read_user_buf_avail ||
security/tomoyo/common.c
228
copy_to_user(head->read_user_buf, "", 1))
security/tomoyo/common.c
230
head->read_user_buf_avail--;
security/tomoyo/common.c
231
head->read_user_buf++;
security/tomoyo/common.c
233
head->r.w_pos--;
security/tomoyo/common.c
2337
static void tomoyo_read_query(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
2338
__must_hold(&head->io_sem)
security/tomoyo/common.c
234
for (len = 0; len < head->r.w_pos; len++)
security/tomoyo/common.c
2345
if (head->r.w_pos)
security/tomoyo/common.c
2347
kfree(head->read_buf);
security/tomoyo/common.c
2348
head->read_buf = NULL;
security/tomoyo/common.c
235
head->r.w[len] = head->r.w[len + 1];
security/tomoyo/common.c
2353
if (pos++ != head->r.query_index)
security/tomoyo/common.c
2360
head->r.query_index = 0;
security/tomoyo/common.c
237
head->r.avail = 0;
security/tomoyo/common.c
2371
if (pos++ != head->r.query_index)
security/tomoyo/common.c
2384
head->read_buf = buf;
security/tomoyo/common.c
2385
head->r.w[head->r.w_pos++] = buf;
security/tomoyo/common.c
2386
head->r.query_index++;
security/tomoyo/common.c
2399
static int tomoyo_write_answer(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
2400
__must_hold(&head->io_sem)
security/tomoyo/common.c
2402
char *data = head->write_buf;
security/tomoyo/common.c
2439
static void tomoyo_read_version(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
2440
__must_hold(&head->io_sem)
security/tomoyo/common.c
2442
if (!head->r.eof) {
security/tomoyo/common.c
2443
tomoyo_io_printf(head, "2.6.0");
security/tomoyo/common.c
2444
head->r.eof = true;
security/tomoyo/common.c
2488
static void tomoyo_read_stat(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
2489
__must_hold(&head->io_sem)
security/tomoyo/common.c
2494
if (head->r.eof)
security/tomoyo/common.c
2497
tomoyo_io_printf(head, "Policy %-30s %10u",
security/tomoyo/common.c
2504
tomoyo_io_printf(head, " (Last: %04u/%02u/%02u %02u:%02u:%02u)",
security/tomoyo/common.c
2508
tomoyo_set_lf(head);
security/tomoyo/common.c
251
static void tomoyo_set_string(struct tomoyo_io_buffer *head, const char *string)
security/tomoyo/common.c
2514
tomoyo_io_printf(head, "Memory used by %-22s %10u",
security/tomoyo/common.c
2518
tomoyo_io_printf(head, " (Quota: %10u)", used);
security/tomoyo/common.c
2519
tomoyo_set_lf(head);
security/tomoyo/common.c
2521
tomoyo_io_printf(head, "Total memory used: %10u\n",
security/tomoyo/common.c
2523
head->r.eof = true;
security/tomoyo/common.c
253
if (head->r.w_pos < TOMOYO_MAX_IO_READ_QUEUE) {
security/tomoyo/common.c
2533
static int tomoyo_write_stat(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
2534
__must_hold(&head->io_sem)
security/tomoyo/common.c
2536
char *data = head->write_buf;
security/tomoyo/common.c
254
head->r.w[head->r.w_pos++] = string;
security/tomoyo/common.c
255
tomoyo_flush(head);
security/tomoyo/common.c
2556
struct tomoyo_io_buffer *head = kzalloc_obj(*head, GFP_NOFS);
security/tomoyo/common.c
2558
if (!head)
security/tomoyo/common.c
2560
guard(mutex_init)(&head->io_sem);
security/tomoyo/common.c
2561
head->type = type;
security/tomoyo/common.c
2565
head->write = tomoyo_write_domain;
security/tomoyo/common.c
2566
head->read = tomoyo_read_domain;
security/tomoyo/common.c
2570
head->write = tomoyo_write_exception;
security/tomoyo/common.c
2571
head->read = tomoyo_read_exception;
security/tomoyo/common.c
2575
head->poll = tomoyo_poll_log;
security/tomoyo/common.c
2576
head->read = tomoyo_read_log;
security/tomoyo/common.c
2580
head->write = tomoyo_write_pid;
security/tomoyo/common.c
2581
head->read = tomoyo_read_pid;
security/tomoyo/common.c
2585
head->read = tomoyo_read_version;
security/tomoyo/common.c
2586
head->readbuf_size = 128;
security/tomoyo/common.c
2590
head->write = tomoyo_write_stat;
security/tomoyo/common.c
2591
head->read = tomoyo_read_stat;
security/tomoyo/common.c
2592
head->readbuf_size = 1024;
security/tomoyo/common.c
2596
head->write = tomoyo_write_profile;
security/tomoyo/common.c
2597
head->read = tomoyo_read_profile;
security/tomoyo/common.c
260
static void tomoyo_io_printf(struct tomoyo_io_buffer *head, const char *fmt,
security/tomoyo/common.c
2600
head->poll = tomoyo_poll_query;
security/tomoyo/common.c
2601
head->write = tomoyo_write_answer;
security/tomoyo/common.c
2602
head->read = tomoyo_read_query;
security/tomoyo/common.c
2606
head->write = tomoyo_write_manager;
security/tomoyo/common.c
2607
head->read = tomoyo_read_manager;
security/tomoyo/common.c
2615
head->read = NULL;
security/tomoyo/common.c
2616
head->poll = NULL;
security/tomoyo/common.c
2617
} else if (!head->poll) {
security/tomoyo/common.c
2619
if (!head->readbuf_size)
security/tomoyo/common.c
2620
head->readbuf_size = 4096 * 2;
security/tomoyo/common.c
2621
head->read_buf = kzalloc(head->readbuf_size, GFP_NOFS);
security/tomoyo/common.c
2622
if (!head->read_buf) {
security/tomoyo/common.c
2623
kfree(head);
security/tomoyo/common.c
2632
head->write = NULL;
security/tomoyo/common.c
2633
} else if (head->write) {
security/tomoyo/common.c
2634
head->writebuf_size = 4096 * 2;
security/tomoyo/common.c
2635
head->write_buf = kzalloc(head->writebuf_size, GFP_NOFS);
security/tomoyo/common.c
2636
if (!head->write_buf) {
security/tomoyo/common.c
2637
kfree(head->read_buf);
security/tomoyo/common.c
2638
kfree(head);
security/tomoyo/common.c
2650
file->private_data = head;
security/tomoyo/common.c
2651
tomoyo_notify_gc(head, true);
security/tomoyo/common.c
2666
struct tomoyo_io_buffer *head = file->private_data;
security/tomoyo/common.c
2668
if (head->poll)
security/tomoyo/common.c
2669
return head->poll(file, wait) | EPOLLOUT | EPOLLWRNORM;
security/tomoyo/common.c
2680
static inline void tomoyo_set_namespace_cursor(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
2684
if (head->type != TOMOYO_EXCEPTIONPOLICY &&
security/tomoyo/common.c
2685
head->type != TOMOYO_PROFILE)
security/tomoyo/common.c
269
static void tomoyo_io_printf(struct tomoyo_io_buffer *head, const char *fmt,
security/tomoyo/common.c
2691
ns = head->r.ns;
security/tomoyo/common.c
2692
if (!ns || (head->r.eof && ns->next != &tomoyo_namespace_list)) {
security/tomoyo/common.c
2694
memset(&head->r, 0, sizeof(head->r));
security/tomoyo/common.c
2695
head->r.ns = ns ? ns->next : tomoyo_namespace_list.next;
security/tomoyo/common.c
2706
static inline bool tomoyo_has_more_namespace(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
2708
return (head->type == TOMOYO_EXCEPTIONPOLICY ||
security/tomoyo/common.c
2709
head->type == TOMOYO_PROFILE) && head->r.eof &&
security/tomoyo/common.c
271
__must_hold(&head->io_sem)
security/tomoyo/common.c
2710
head->r.ns->next != &tomoyo_namespace_list;
security/tomoyo/common.c
2722
ssize_t tomoyo_read_control(struct tomoyo_io_buffer *head, char __user *buffer,
security/tomoyo/common.c
2728
if (!head->read)
security/tomoyo/common.c
2730
if (mutex_lock_interruptible(&head->io_sem))
security/tomoyo/common.c
2732
head->read_user_buf = buffer;
security/tomoyo/common.c
2733
head->read_user_buf_avail = buffer_len;
security/tomoyo/common.c
2735
if (tomoyo_flush(head))
security/tomoyo/common.c
2738
tomoyo_set_namespace_cursor(head);
security/tomoyo/common.c
2739
head->read(head);
security/tomoyo/common.c
2740
} while (tomoyo_flush(head) &&
security/tomoyo/common.c
2741
tomoyo_has_more_namespace(head));
security/tomoyo/common.c
2743
len = head->read_user_buf - buffer;
security/tomoyo/common.c
2744
mutex_unlock(&head->io_sem);
security/tomoyo/common.c
275
size_t pos = head->r.avail;
security/tomoyo/common.c
2758
static int tomoyo_parse_policy(struct tomoyo_io_buffer *head, char *line)
security/tomoyo/common.c
276
int size = head->readbuf_size - pos;
security/tomoyo/common.c
2760
__must_hold(&head->io_sem)
security/tomoyo/common.c
2763
head->w.is_delete = !strncmp(line, "delete ", 7);
security/tomoyo/common.c
2764
if (head->w.is_delete)
security/tomoyo/common.c
2767
if (head->type == TOMOYO_EXCEPTIONPOLICY ||
security/tomoyo/common.c
2768
head->type == TOMOYO_PROFILE) {
security/tomoyo/common.c
2774
head->w.ns = tomoyo_assign_namespace(line);
security/tomoyo/common.c
2777
head->w.ns = NULL;
security/tomoyo/common.c
2779
head->w.ns = &tomoyo_kernel_namespace;
security/tomoyo/common.c
2781
if (!head->w.ns)
security/tomoyo/common.c
2785
return head->write(head);
security/tomoyo/common.c
2797
ssize_t tomoyo_write_control(struct tomoyo_io_buffer *head,
security/tomoyo/common.c
2805
if (!head->write)
security/tomoyo/common.c
2807
if (mutex_lock_interruptible(&head->io_sem))
security/tomoyo/common.c
2809
cp0 = head->write_buf;
security/tomoyo/common.c
281
len = vsnprintf(head->read_buf + pos, size, fmt, args) + 1;
security/tomoyo/common.c
2810
head->read_user_buf_avail = 0;
security/tomoyo/common.c
2816
if (head->w.avail >= head->writebuf_size - 1) {
security/tomoyo/common.c
2817
const int len = head->writebuf_size * 2;
security/tomoyo/common.c
2824
memmove(cp, cp0, head->w.avail);
security/tomoyo/common.c
2826
head->write_buf = cp;
security/tomoyo/common.c
2828
head->writebuf_size = len;
security/tomoyo/common.c
283
if (pos + len >= head->readbuf_size) {
security/tomoyo/common.c
2836
cp0[head->w.avail++] = c;
security/tomoyo/common.c
2839
cp0[head->w.avail - 1] = '\0';
security/tomoyo/common.c
2840
head->w.avail = 0;
security/tomoyo/common.c
2843
head->w.ns = &tomoyo_kernel_namespace;
security/tomoyo/common.c
2844
head->w.domain = NULL;
security/tomoyo/common.c
2845
memset(&head->r, 0, sizeof(head->r));
security/tomoyo/common.c
2849
switch (head->type) {
security/tomoyo/common.c
2854
if (tomoyo_select_domain(head, cp0))
security/tomoyo/common.c
2859
head->r.print_transition_related_only = true;
security/tomoyo/common.c
2869
switch (tomoyo_parse_policy(head, cp0)) {
security/tomoyo/common.c
287
head->r.avail += len;
security/tomoyo/common.c
2874
switch (head->type) {
security/tomoyo/common.c
288
tomoyo_set_string(head, head->read_buf + pos);
security/tomoyo/common.c
2890
mutex_unlock(&head->io_sem);
security/tomoyo/common.c
2899
void tomoyo_close_control(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
2905
if (head->type == TOMOYO_QUERY &&
security/tomoyo/common.c
2908
tomoyo_notify_gc(head, false);
security/tomoyo/common.c
2976
struct tomoyo_io_buffer head = { };
security/tomoyo/common.c
298
static void tomoyo_set_space(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
2982
head.type = TOMOYO_PROFILE;
security/tomoyo/common.c
2983
head.write = tomoyo_write_profile;
security/tomoyo/common.c
2987
head.type = TOMOYO_EXCEPTIONPOLICY;
security/tomoyo/common.c
2988
head.write = tomoyo_write_exception;
security/tomoyo/common.c
2992
head.type = TOMOYO_DOMAINPOLICY;
security/tomoyo/common.c
2993
head.write = tomoyo_write_domain;
security/tomoyo/common.c
2997
head.type = TOMOYO_MANAGER;
security/tomoyo/common.c
2998
head.write = tomoyo_write_manager;
security/tomoyo/common.c
300
tomoyo_set_string(head, " ");
security/tomoyo/common.c
3002
head.type = TOMOYO_STAT;
security/tomoyo/common.c
3003
head.write = tomoyo_write_stat;
security/tomoyo/common.c
3015
head.write_buf = start;
security/tomoyo/common.c
3016
tomoyo_parse_policy(&head, start);
security/tomoyo/common.c
310
static bool tomoyo_set_lf(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
312
tomoyo_set_string(head, "\n");
security/tomoyo/common.c
313
return !head->r.w_pos;
security/tomoyo/common.c
323
static void tomoyo_set_slash(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
325
tomoyo_set_string(head, "/");
security/tomoyo/common.c
362
static void tomoyo_print_namespace(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
366
tomoyo_set_string(head,
security/tomoyo/common.c
367
container_of(head->r.ns,
security/tomoyo/common.c
370
tomoyo_set_space(head);
security/tomoyo/common.c
379
static void tomoyo_print_name_union(struct tomoyo_io_buffer *head,
security/tomoyo/common.c
382
tomoyo_set_space(head);
security/tomoyo/common.c
384
tomoyo_set_string(head, "@");
security/tomoyo/common.c
385
tomoyo_set_string(head, ptr->group->group_name->name);
security/tomoyo/common.c
387
tomoyo_set_string(head, ptr->filename->name);
security/tomoyo/common.c
399
static void tomoyo_print_name_union_quoted(struct tomoyo_io_buffer *head,
security/tomoyo/common.c
403
tomoyo_set_string(head, "@");
security/tomoyo/common.c
404
tomoyo_set_string(head, ptr->group->group_name->name);
security/tomoyo/common.c
406
tomoyo_set_string(head, "\"");
security/tomoyo/common.c
407
tomoyo_set_string(head, ptr->filename->name);
security/tomoyo/common.c
408
tomoyo_set_string(head, "\"");
security/tomoyo/common.c
421
tomoyo_print_number_union_nospace(struct tomoyo_io_buffer *head, const struct tomoyo_number_union *ptr)
security/tomoyo/common.c
422
__must_hold(&head->io_sem)
security/tomoyo/common.c
425
tomoyo_set_string(head, "@");
security/tomoyo/common.c
426
tomoyo_set_string(head, ptr->group->group_name->name);
security/tomoyo/common.c
457
tomoyo_io_printf(head, "%s", buffer);
security/tomoyo/common.c
469
static void tomoyo_print_number_union(struct tomoyo_io_buffer *head,
security/tomoyo/common.c
471
__must_hold(&head->io_sem)
security/tomoyo/common.c
473
tomoyo_set_space(head);
security/tomoyo/common.c
474
tomoyo_print_number_union_nospace(head, ptr);
security/tomoyo/common.c
669
static int tomoyo_write_profile(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
670
__must_hold(&head->io_sem)
security/tomoyo/common.c
672
char *data = head->write_buf;
security/tomoyo/common.c
677
if (sscanf(data, "PROFILE_VERSION=%u", &head->w.ns->profile_version)
security/tomoyo/common.c
684
profile = tomoyo_assign_profile(head->w.ns, i);
security/tomoyo/common.c
725
static void tomoyo_print_config(struct tomoyo_io_buffer *head, const u8 config)
security/tomoyo/common.c
726
__must_hold(&head->io_sem)
security/tomoyo/common.c
728
tomoyo_io_printf(head, "={ mode=%s grant_log=%s reject_log=%s }\n",
security/tomoyo/common.c
741
static void tomoyo_read_profile(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
742
__must_hold(&head->io_sem)
security/tomoyo/common.c
746
container_of(head->r.ns, typeof(*ns), namespace_list);
security/tomoyo/common.c
749
if (head->r.eof)
security/tomoyo/common.c
752
index = head->r.index;
security/tomoyo/common.c
754
switch (head->r.step) {
security/tomoyo/common.c
756
tomoyo_print_namespace(head);
security/tomoyo/common.c
757
tomoyo_io_printf(head, "PROFILE_VERSION=%u\n",
security/tomoyo/common.c
759
head->r.step++;
security/tomoyo/common.c
762
for ( ; head->r.index < TOMOYO_MAX_PROFILES;
security/tomoyo/common.c
763
head->r.index++)
security/tomoyo/common.c
764
if (ns->profile_ptr[head->r.index])
security/tomoyo/common.c
766
if (head->r.index == TOMOYO_MAX_PROFILES) {
security/tomoyo/common.c
767
head->r.eof = true;
security/tomoyo/common.c
770
head->r.step++;
security/tomoyo/common.c
778
tomoyo_print_namespace(head);
security/tomoyo/common.c
779
tomoyo_io_printf(head, "%u-COMMENT=", index);
security/tomoyo/common.c
780
tomoyo_set_string(head, comment ? comment->name : "");
security/tomoyo/common.c
781
tomoyo_set_lf(head);
security/tomoyo/common.c
782
tomoyo_print_namespace(head);
security/tomoyo/common.c
783
tomoyo_io_printf(head, "%u-PREFERENCE={ ", index);
security/tomoyo/common.c
785
tomoyo_io_printf(head, "%s=%u ",
security/tomoyo/common.c
788
tomoyo_set_string(head, "}\n");
security/tomoyo/common.c
789
head->r.step++;
security/tomoyo/common.c
794
tomoyo_print_namespace(head);
security/tomoyo/common.c
795
tomoyo_io_printf(head, "%u-%s", index, "CONFIG");
security/tomoyo/common.c
796
tomoyo_print_config(head, profile->default_config);
security/tomoyo/common.c
797
head->r.bit = 0;
security/tomoyo/common.c
798
head->r.step++;
security/tomoyo/common.c
802
for ( ; head->r.bit < TOMOYO_MAX_MAC_INDEX
security/tomoyo/common.c
803
+ TOMOYO_MAX_MAC_CATEGORY_INDEX; head->r.bit++) {
security/tomoyo/common.c
804
const u8 i = head->r.bit;
security/tomoyo/common.c
809
tomoyo_print_namespace(head);
security/tomoyo/common.c
811
tomoyo_io_printf(head, "%u-CONFIG::%s::%s",
security/tomoyo/common.c
817
tomoyo_io_printf(head, "%u-CONFIG::%s", index,
security/tomoyo/common.c
819
tomoyo_print_config(head, config);
security/tomoyo/common.c
820
head->r.bit++;
security/tomoyo/common.c
823
if (head->r.bit == TOMOYO_MAX_MAC_INDEX
security/tomoyo/common.c
825
head->r.index++;
security/tomoyo/common.c
826
head->r.step = 1;
security/tomoyo/common.c
830
if (tomoyo_flush(head))
security/tomoyo/common.c
845
return container_of(a, struct tomoyo_manager, head)->manager ==
security/tomoyo/common.c
846
container_of(b, struct tomoyo_manager, head)->manager;
security/tomoyo/common.c
876
error = tomoyo_update_policy(&e.head, sizeof(e), &param,
security/tomoyo/common.c
892
static int tomoyo_write_manager(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
894
__must_hold(&head->io_sem)
security/tomoyo/common.c
896
char *data = head->write_buf;
security/tomoyo/common.c
899
tomoyo_manage_by_non_root = !head->w.is_delete;
security/tomoyo/common.c
902
return tomoyo_update_manager_entry(data, head->w.is_delete);
security/tomoyo/common.c
912
static void tomoyo_read_manager(struct tomoyo_io_buffer *head)
security/tomoyo/common.c
915
if (head->r.eof)
security/tomoyo/common.c
917
list_for_each_cookie(head->r.acl, &tomoyo_kernel_namespace.policy_list[TOMOYO_ID_MANAGER]) {
security/tomoyo/common.c
919
list_entry(head->r.acl, typeof(*ptr), head.list);
security/tomoyo/common.c
921
if (ptr->head.is_deleted)
security/tomoyo/common.c
923
if (!tomoyo_flush(head))
security/tomoyo/common.c
925
tomoyo_set_string(head, ptr->manager->name);
security/tomoyo/common.c
926
tomoyo_set_lf(head);
security/tomoyo/common.c
928
head->r.eof = true;
security/tomoyo/common.c
957
list_for_each_entry_rcu(ptr, &tomoyo_kernel_namespace.policy_list[TOMOYO_ID_MANAGER], head.list,
security/tomoyo/common.c
959
if (!ptr->head.is_deleted &&
security/tomoyo/common.c
993
static bool tomoyo_select_domain(struct tomoyo_io_buffer *head,
security/tomoyo/common.c
996
__must_hold(&head->io_sem)
security/tomoyo/common.h
1002
void tomoyo_close_control(struct tomoyo_io_buffer *head);
security/tomoyo/common.h
1057
ssize_t tomoyo_read_control(struct tomoyo_io_buffer *head, char __user *buffer,
security/tomoyo/common.h
1059
ssize_t tomoyo_write_control(struct tomoyo_io_buffer *head,
security/tomoyo/common.h
1087
void tomoyo_notify_gc(struct tomoyo_io_buffer *head, const bool is_register);
security/tomoyo/common.h
1094
void tomoyo_read_log(struct tomoyo_io_buffer *head) __must_hold(&head->io_sem);
security/tomoyo/common.h
1185
atomic_dec(&ptr->head.users);
security/tomoyo/common.h
1199
atomic_dec(&cond->head.users);
security/tomoyo/common.h
1212
atomic_dec(&group->head.users);
security/tomoyo/common.h
1287
#define list_for_each_cookie(pos, head) \
security/tomoyo/common.h
1289
pos = srcu_dereference((head)->next, &tomoyo_ss); \
security/tomoyo/common.h
1290
for ( ; pos != (head); pos = srcu_dereference(pos->next, &tomoyo_ss))
security/tomoyo/common.h
514
struct tomoyo_shared_acl_head head;
security/tomoyo/common.h
542
struct tomoyo_shared_acl_head head;
security/tomoyo/common.h
549
struct tomoyo_acl_head head;
security/tomoyo/common.h
555
struct tomoyo_acl_head head;
security/tomoyo/common.h
561
struct tomoyo_acl_head head;
security/tomoyo/common.h
653
struct tomoyo_shared_acl_head head;
security/tomoyo/common.h
699
struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_MANUAL_TASK_ACL */
security/tomoyo/common.h
710
struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_PATH_ACL */
security/tomoyo/common.h
720
struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_PATH_NUMBER_ACL */
security/tomoyo/common.h
729
struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_MKDEV_ACL */
security/tomoyo/common.h
741
struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_PATH2_ACL */
security/tomoyo/common.h
749
struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_MOUNT_ACL */
security/tomoyo/common.h
758
struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_ENV_ACL */
security/tomoyo/common.h
764
struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_INET_ACL */
security/tomoyo/common.h
773
struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_UNIX_ACL */
security/tomoyo/common.h
794
void (*read)(struct tomoyo_io_buffer *head);
security/tomoyo/common.h
795
int (*write)(struct tomoyo_io_buffer *head);
security/tomoyo/common.h
850
struct tomoyo_acl_head head;
security/tomoyo/common.h
860
struct tomoyo_acl_head head;
security/tomoyo/common.h
867
struct tomoyo_acl_head head;
security/tomoyo/condition.c
418
list_for_each_entry(ptr, &tomoyo_condition_list, head.list) {
security/tomoyo/condition.c
420
atomic_read(&ptr->head.users) == TOMOYO_GC_IN_PROGRESS)
security/tomoyo/condition.c
423
atomic_inc(&ptr->head.users);
security/tomoyo/condition.c
429
atomic_set(&entry->head.users, 1);
security/tomoyo/condition.c
430
list_add(&entry->head.list, &tomoyo_condition_list);
security/tomoyo/condition.c
439
tomoyo_del_condition(&entry->head.list);
security/tomoyo/condition.c
698
tomoyo_del_condition(&entry->head.list);
security/tomoyo/domain.c
117
container_of(new_entry, struct tomoyo_path_acl, head)
security/tomoyo/domain.c
224
head);
security/tomoyo/domain.c
227
head);
security/tomoyo/domain.c
276
error = tomoyo_update_policy(&e.head, sizeof(e), param,
security/tomoyo/domain.c
304
list_for_each_entry_rcu(ptr, list, head.list,
security/tomoyo/domain.c
306
if (ptr->head.is_deleted || ptr->type != type)
security/tomoyo/domain.c
386
head);
security/tomoyo/domain.c
388
head);
security/tomoyo/domain.c
419
error = tomoyo_update_policy(&e.head, sizeof(e), param,
security/tomoyo/domain.c
751
list_for_each_entry_rcu(ptr, list, head.list,
security/tomoyo/domain.c
753
if (ptr->head.is_deleted ||
security/tomoyo/environ.c
106
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/environ.c
22
container_of(ptr, typeof(*acl), head);
security/tomoyo/environ.c
80
const struct tomoyo_env_acl *p1 = container_of(a, typeof(*p1), head);
security/tomoyo/environ.c
81
const struct tomoyo_env_acl *p2 = container_of(b, typeof(*p2), head);
security/tomoyo/environ.c
97
struct tomoyo_env_acl e = { .head.type = TOMOYO_TYPE_ENV_ACL };
security/tomoyo/file.c
260
head);
security/tomoyo/file.c
283
container_of(ptr, typeof(*acl), head);
security/tomoyo/file.c
304
container_of(ptr, typeof(*acl), head);
security/tomoyo/file.c
324
container_of(ptr, typeof(*acl), head);
security/tomoyo/file.c
348
const struct tomoyo_path_acl *p1 = container_of(a, typeof(*p1), head);
security/tomoyo/file.c
349
const struct tomoyo_path_acl *p2 = container_of(b, typeof(*p2), head);
security/tomoyo/file.c
367
u16 * const a_perm = &container_of(a, struct tomoyo_path_acl, head)
security/tomoyo/file.c
370
const u16 b_perm = container_of(b, struct tomoyo_path_acl, head)->perm;
security/tomoyo/file.c
394
.head.type = TOMOYO_TYPE_PATH_ACL,
security/tomoyo/file.c
402
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/file.c
420
const struct tomoyo_mkdev_acl *p1 = container_of(a, typeof(*p1), head);
security/tomoyo/file.c
421
const struct tomoyo_mkdev_acl *p2 = container_of(b, typeof(*p2), head);
security/tomoyo/file.c
443
head)->perm;
security/tomoyo/file.c
445
const u8 b_perm = container_of(b, struct tomoyo_mkdev_acl, head)
security/tomoyo/file.c
470
.head.type = TOMOYO_TYPE_MKDEV_ACL,
security/tomoyo/file.c
481
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/file.c
502
const struct tomoyo_path2_acl *p1 = container_of(a, typeof(*p1), head);
security/tomoyo/file.c
503
const struct tomoyo_path2_acl *p2 = container_of(b, typeof(*p2), head);
security/tomoyo/file.c
522
u8 * const a_perm = &container_of(a, struct tomoyo_path2_acl, head)
security/tomoyo/file.c
525
const u8 b_perm = container_of(b, struct tomoyo_path2_acl, head)->perm;
security/tomoyo/file.c
549
.head.type = TOMOYO_TYPE_PATH2_ACL,
security/tomoyo/file.c
558
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/file.c
640
head);
security/tomoyo/file.c
642
head);
security/tomoyo/file.c
662
head)->perm;
security/tomoyo/file.c
664
const u8 b_perm = container_of(b, struct tomoyo_path_number_acl, head)
security/tomoyo/file.c
687
.head.type = TOMOYO_TYPE_PATH_NUMBER_ACL,
security/tomoyo/file.c
696
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/file.c
970
const struct tomoyo_mount_acl *p1 = container_of(a, typeof(*p1), head);
security/tomoyo/file.c
971
const struct tomoyo_mount_acl *p2 = container_of(b, typeof(*p2), head);
security/tomoyo/file.c
990
struct tomoyo_mount_acl e = { .head.type = TOMOYO_TYPE_MOUNT_ACL };
security/tomoyo/file.c
999
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/gc.c
111
container_of(element, typeof(*ptr), head.list);
security/tomoyo/gc.c
127
container_of(element, typeof(*ptr), head.list);
security/tomoyo/gc.c
143
container_of(element, typeof(*ptr), head.list);
security/tomoyo/gc.c
165
= container_of(acl, typeof(*entry), head);
security/tomoyo/gc.c
172
= container_of(acl, typeof(*entry), head);
security/tomoyo/gc.c
180
= container_of(acl, typeof(*entry), head);
security/tomoyo/gc.c
188
= container_of(acl, typeof(*entry), head);
security/tomoyo/gc.c
198
= container_of(acl, typeof(*entry), head);
security/tomoyo/gc.c
208
container_of(acl, typeof(*entry), head);
security/tomoyo/gc.c
216
container_of(acl, typeof(*entry), head);
security/tomoyo/gc.c
225
container_of(acl, typeof(*entry), head);
security/tomoyo/gc.c
233
container_of(acl, typeof(*entry), head);
security/tomoyo/gc.c
279
head.list);
security/tomoyo/gc.c
331
container_of(element, typeof(*member), head.list);
security/tomoyo/gc.c
346
container_of(element, typeof(*group), head.list);
security/tomoyo/gc.c
41
struct tomoyo_io_buffer *head;
security/tomoyo/gc.c
439
head.list)->entry.name))
security/tomoyo/gc.c
45
list_for_each_entry(head, &tomoyo_io_buffer_list, list) {
security/tomoyo/gc.c
46
head->users++;
security/tomoyo/gc.c
48
mutex_lock(&head->io_sem);
security/tomoyo/gc.c
49
if (head->r.domain == element || head->r.group == element ||
security/tomoyo/gc.c
50
head->r.acl == element || &head->w.domain->list == element)
security/tomoyo/gc.c
52
mutex_unlock(&head->io_sem);
security/tomoyo/gc.c
54
head->users--;
security/tomoyo/gc.c
580
list_for_each_entry_safe(group, tmp, list, head.list) {
security/tomoyo/gc.c
583
atomic_read(&group->head.users) > 0)
security/tomoyo/gc.c
585
atomic_set(&group->head.users,
security/tomoyo/gc.c
588
&group->head.list);
security/tomoyo/gc.c
623
struct tomoyo_io_buffer *head;
security/tomoyo/gc.c
627
list_for_each_entry_safe(head, tmp, &tomoyo_io_buffer_list,
security/tomoyo/gc.c
629
if (head->users)
security/tomoyo/gc.c
631
list_del(&head->list);
security/tomoyo/gc.c
634
kfree(head->read_buf);
security/tomoyo/gc.c
635
kfree(head->write_buf);
security/tomoyo/gc.c
637
kfree(head);
security/tomoyo/gc.c
655
void tomoyo_notify_gc(struct tomoyo_io_buffer *head, const bool is_register)
security/tomoyo/gc.c
661
head->users = 1;
security/tomoyo/gc.c
662
list_add(&head->list, &tomoyo_io_buffer_list);
security/tomoyo/gc.c
668
is_write = data_race(head->write_buf != NULL);
security/tomoyo/gc.c
669
if (!--head->users) {
security/tomoyo/gc.c
670
list_del(&head->list);
security/tomoyo/gc.c
673
kfree(head->read_buf);
security/tomoyo/gc.c
674
kfree(head->write_buf);
security/tomoyo/gc.c
676
kfree(head);
security/tomoyo/gc.c
72
struct tomoyo_io_buffer *head;
security/tomoyo/gc.c
77
list_for_each_entry(head, &tomoyo_io_buffer_list, list) {
security/tomoyo/gc.c
80
head->users++;
security/tomoyo/gc.c
82
mutex_lock(&head->io_sem);
security/tomoyo/gc.c
84
const char *w = head->r.w[i];
security/tomoyo/gc.c
91
mutex_unlock(&head->io_sem);
security/tomoyo/gc.c
93
head->users--;
security/tomoyo/group.c
111
error = tomoyo_update_policy(&e.head, sizeof(e), param,
security/tomoyo/group.c
136
list_for_each_entry_rcu(member, &group->member_list, head.list,
security/tomoyo/group.c
138
if (member->head.is_deleted)
security/tomoyo/group.c
165
list_for_each_entry_rcu(member, &group->member_list, head.list,
security/tomoyo/group.c
167
if (member->head.is_deleted)
security/tomoyo/group.c
196
list_for_each_entry_rcu(member, &group->member_list, head.list,
security/tomoyo/group.c
198
if (member->head.is_deleted)
security/tomoyo/group.c
24
return container_of(a, struct tomoyo_path_group, head)->member_name ==
security/tomoyo/group.c
25
container_of(b, struct tomoyo_path_group, head)->member_name;
security/tomoyo/group.c
39
return !memcmp(&container_of(a, struct tomoyo_number_group, head)
security/tomoyo/group.c
41
&container_of(b, struct tomoyo_number_group, head)
security/tomoyo/group.c
43
sizeof(container_of(a, struct tomoyo_number_group, head)
security/tomoyo/group.c
59
head);
security/tomoyo/group.c
61
head);
security/tomoyo/group.c
90
error = tomoyo_update_policy(&e.head, sizeof(e), param,
security/tomoyo/group.c
99
error = tomoyo_update_policy(&e.head, sizeof(e), param,
security/tomoyo/memory.c
112
list_for_each_entry(group, list, head.list) {
security/tomoyo/memory.c
114
atomic_read(&group->head.users) == TOMOYO_GC_IN_PROGRESS)
security/tomoyo/memory.c
116
atomic_inc(&group->head.users);
security/tomoyo/memory.c
125
atomic_set(&entry->head.users, 1);
security/tomoyo/memory.c
126
list_add_tail_rcu(&entry->head.list, list);
security/tomoyo/memory.c
157
struct list_head *head;
security/tomoyo/memory.c
163
head = &tomoyo_name_list[hash_long(hash, TOMOYO_HASH_BITS)];
security/tomoyo/memory.c
166
list_for_each_entry(ptr, head, head.list) {
security/tomoyo/memory.c
168
atomic_read(&ptr->head.users) == TOMOYO_GC_IN_PROGRESS)
security/tomoyo/memory.c
170
atomic_inc(&ptr->head.users);
security/tomoyo/memory.c
177
atomic_set(&ptr->head.users, 1);
security/tomoyo/memory.c
179
list_add_tail(&ptr->head.list, head);
security/tomoyo/mount.c
52
container_of(ptr, typeof(*acl), head);
security/tomoyo/network.c
195
const struct tomoyo_inet_acl *p1 = container_of(a, typeof(*p1), head);
security/tomoyo/network.c
196
const struct tomoyo_inet_acl *p2 = container_of(b, typeof(*p2), head);
security/tomoyo/network.c
214
const struct tomoyo_unix_acl *p1 = container_of(a, typeof(*p1), head);
security/tomoyo/network.c
215
const struct tomoyo_unix_acl *p2 = container_of(b, typeof(*p2), head);
security/tomoyo/network.c
235
&container_of(a, struct tomoyo_inet_acl, head)->perm;
security/tomoyo/network.c
237
const u8 b_perm = container_of(b, struct tomoyo_inet_acl, head)->perm;
security/tomoyo/network.c
261
&container_of(a, struct tomoyo_unix_acl, head)->perm;
security/tomoyo/network.c
263
const u8 b_perm = container_of(b, struct tomoyo_unix_acl, head)->perm;
security/tomoyo/network.c
284
struct tomoyo_inet_acl e = { .head.type = TOMOYO_TYPE_INET_ACL };
security/tomoyo/network.c
311
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/network.c
329
struct tomoyo_unix_acl e = { .head.type = TOMOYO_TYPE_UNIX_ACL };
security/tomoyo/network.c
345
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/network.c
426
container_of(ptr, typeof(*acl), head);
security/tomoyo/network.c
456
container_of(ptr, typeof(*acl), head);
security/tomoyo/securityfs_if.c
23
head);
security/tomoyo/util.c
1069
perm = data_race(container_of(ptr, struct tomoyo_path_acl, head)->perm);
security/tomoyo/util.c
1072
perm = data_race(container_of(ptr, struct tomoyo_path2_acl, head)->perm);
security/tomoyo/util.c
1075
perm = data_race(container_of(ptr, struct tomoyo_path_number_acl, head)
security/tomoyo/util.c
1079
perm = data_race(container_of(ptr, struct tomoyo_mkdev_acl, head)->perm);
security/tomoyo/util.c
1082
perm = data_race(container_of(ptr, struct tomoyo_inet_acl, head)->perm);
security/tomoyo/util.c
1085
perm = data_race(container_of(ptr, struct tomoyo_unix_acl, head)->perm);
sound/core/seq/oss/seq_oss_readq.c
167
memcpy(rec, &q->q[q->head], sizeof(*rec));
sound/core/seq/oss/seq_oss_readq.c
178
(q->qlen > 0 || q->head == q->tail),
sound/core/seq/oss/seq_oss_readq.c
190
q->head = (q->head + 1) % q->maxlen;
sound/core/seq/oss/seq_oss_readq.c
49
q->head = q->tail = 0;
sound/core/seq/oss/seq_oss_readq.c
78
q->head = q->tail = 0;
sound/core/seq/oss/seq_oss_readq.h
22
int head, tail;
sound/core/seq/seq_fifo.c
122
if (f->head == NULL)
sound/core/seq/seq_fifo.c
123
f->head = cell;
sound/core/seq/seq_fifo.c
141
cell = f->head;
sound/core/seq/seq_fifo.c
143
f->head = cell->next;
sound/core/seq/seq_fifo.c
199
cell->next = f->head;
sound/core/seq/seq_fifo.c
200
f->head = cell;
sound/core/seq/seq_fifo.c
238
oldhead = f->head;
sound/core/seq/seq_fifo.c
241
f->head = NULL;
sound/core/seq/seq_fifo.c
42
f->head = NULL;
sound/core/seq/seq_fifo.h
17
struct snd_seq_event_cell *head; /* pointer to head of fifo */
sound/core/seq/seq_prioq.c
163
cur = f->head; /* cursor */
sound/core/seq/seq_prioq.c
190
if (f->head == cur) /* this is the first cell, set head to it */
sound/core/seq/seq_prioq.c
191
f->head = cell;
sound/core/seq/seq_prioq.c
219
cell = f->head;
sound/core/seq/seq_prioq.c
223
f->head = cell->next;
sound/core/seq/seq_prioq.c
258
for (cell = f->head; cell; cell = next) {
sound/core/seq/seq_prioq.c
266
if (cell == f->head)
sound/core/seq/seq_prioq.c
267
f->head = cell->next;
sound/core/seq/seq_prioq.c
51
f->head = NULL;
sound/core/seq/seq_prioq.h
15
struct snd_seq_event_cell *head; /* pointer to head of prioq */
sound/core/timer.c
766
struct list_head *head)
sound/core/timer.c
771
while (!list_empty(head)) {
sound/core/timer.c
772
ti = list_first_entry(head, struct snd_timer_instance,
sound/core/timer.c
794
struct list_head *head)
sound/core/timer.c
797
while (!list_empty(head))
sound/core/timer.c
798
list_del_init(head->next);
sound/firewire/amdtp-stream.c
494
static unsigned int calculate_cached_cycle_count(struct amdtp_stream *s, unsigned int head)
sound/firewire/amdtp-stream.c
499
if (cycles < head)
sound/firewire/amdtp-stream.c
501
cycles -= head;
sound/firewire/motu/amdtp-motu.c
378
unsigned int cache_head = cache->head;
sound/firewire/motu/amdtp-motu.c
391
cache->head = cache_head;
sound/firewire/motu/motu-stream.c
287
motu->cache.head = 0;
sound/firewire/motu/motu.h
47
unsigned int head;
sound/isa/msnd/msnd_pinnacle.c
156
u16 head, tail, size;
sound/isa/msnd/msnd_pinnacle.c
162
head = readw(chip->DSPQ + JQS_wHead);
sound/isa/msnd/msnd_pinnacle.c
165
if (head > size || tail > size)
sound/isa/msnd/msnd_pinnacle.c
167
while (head != tail) {
sound/isa/msnd/msnd_pinnacle.c
168
snd_msnd_eval_dsp_msg(chip, readw(pwDSPQData + 2 * head));
sound/isa/msnd/msnd_pinnacle.c
169
if (++head > size)
sound/isa/msnd/msnd_pinnacle.c
170
head = 0;
sound/isa/msnd/msnd_pinnacle.c
171
writew(head, chip->DSPQ + JQS_wHead);
sound/pci/ctxfi/ctimap.c
101
while (!list_empty(head)) {
sound/pci/ctxfi/ctimap.c
102
pos = head->next;
sound/pci/ctxfi/ctimap.c
21
struct list_head *pos, *pre, *head;
sound/pci/ctxfi/ctimap.c
24
head = mappers;
sound/pci/ctxfi/ctimap.c
26
if (list_empty(head)) {
sound/pci/ctxfi/ctimap.c
29
list_add(&entry->list, head);
sound/pci/ctxfi/ctimap.c
33
list_for_each(pos, head) {
sound/pci/ctxfi/ctimap.c
41
if (pos != head) {
sound/pci/ctxfi/ctimap.c
43
if (pre == head)
sound/pci/ctxfi/ctimap.c
44
pre = head->prev;
sound/pci/ctxfi/ctimap.c
48
pre = head->prev;
sound/pci/ctxfi/ctimap.c
49
pos = head->next;
sound/pci/ctxfi/ctimap.c
50
list_add_tail(&entry->list, head);
sound/pci/ctxfi/ctimap.c
67
struct list_head *next, *pre, *head;
sound/pci/ctxfi/ctimap.c
70
head = mappers;
sound/pci/ctxfi/ctimap.c
72
if (list_empty(head))
sound/pci/ctxfi/ctimap.c
75
pre = (entry->list.prev == head) ? head->prev : entry->list.prev;
sound/pci/ctxfi/ctimap.c
76
next = (entry->list.next == head) ? head->next : entry->list.next;
sound/pci/ctxfi/ctimap.c
96
void free_input_mapper_list(struct list_head *head)
sound/pci/echoaudio/echoaudio_dsp.c
1115
int head = pipe->sglist_head;
sound/pci/echoaudio/echoaudio_dsp.c
1118
if (head < MAX_SGLIST_ENTRIES - 1) {
sound/pci/echoaudio/echoaudio_dsp.c
1119
list[head].addr = cpu_to_le32(address);
sound/pci/echoaudio/echoaudio_dsp.c
1120
list[head].size = cpu_to_le32(length);
sound/soc/codecs/hdac_hdmi.c
1014
list_for_each_entry(cvt, &hdmi->cvt_list, head) {
sound/soc/codecs/hdac_hdmi.c
1100
list_for_each_entry(cvt, &hdmi->cvt_list, head) {
sound/soc/codecs/hdac_hdmi.c
1112
list_for_each_entry(pin, &hdmi->pin_list, head) {
sound/soc/codecs/hdac_hdmi.c
1130
list_for_each_entry(pin, &hdmi->pin_list, head) {
sound/soc/codecs/hdac_hdmi.c
1156
list_for_each_entry(pin, &hdmi->pin_list, head) {
sound/soc/codecs/hdac_hdmi.c
1191
list_for_each_entry(cvt, &hdmi->cvt_list, head) {
sound/soc/codecs/hdac_hdmi.c
1224
list_add_tail(&cvt->head, &hdmi->cvt_list);
sound/soc/codecs/hdac_hdmi.c
1388
list_add_tail(&pin->head, &hdmi->pin_list);
sound/soc/codecs/hdac_hdmi.c
1470
list_for_each_entry(cvt, &hdmi->cvt_list, head) {
sound/soc/codecs/hdac_hdmi.c
148
list_for_each_entry(pcm, &hdmi->pcm_list, head) {
sound/soc/codecs/hdac_hdmi.c
1617
list_for_each_entry(pin, &hdmi->pin_list, head) {
sound/soc/codecs/hdac_hdmi.c
1653
list_for_each_entry(pin, &hdmi->pin_list, head) {
sound/soc/codecs/hdac_hdmi.c
1799
list_for_each_entry(port, &pcm->port_list, head)
sound/soc/codecs/hdac_hdmi.c
1831
port = list_first_entry(&pcm->port_list, struct hdac_hdmi_port, head);
sound/soc/codecs/hdac_hdmi.c
1922
list_for_each_entry(pin, &hdmi->pin_list, head)
sound/soc/codecs/hdac_hdmi.c
286
list_for_each_entry(pcm, &hdmi->pcm_list, head) {
sound/soc/codecs/hdac_hdmi.c
534
list_for_each_entry(pcm, &hdmi->pcm_list, head) {
sound/soc/codecs/hdac_hdmi.c
539
list_for_each_entry(port, &pcm->port_list, head) {
sound/soc/codecs/hdac_hdmi.c
55
struct list_head head;
sound/soc/codecs/hdac_hdmi.c
571
list_for_each_entry(cvt, &hdmi->cvt_list, head) {
sound/soc/codecs/hdac_hdmi.c
725
list_for_each_entry(pcm, &hdmi->pcm_list, head) {
sound/soc/codecs/hdac_hdmi.c
729
list_for_each_entry(p, &pcm->port_list, head) {
sound/soc/codecs/hdac_hdmi.c
75
struct list_head head;
sound/soc/codecs/hdac_hdmi.c
84
struct list_head head;
sound/soc/codecs/hdac_hdmi.c
924
list_for_each_entry(pcm, &hdmi->pcm_list, head) {
sound/soc/codecs/hdac_hdmi.c
928
list_for_each_entry_safe(p, p_next, &pcm->port_list, head) {
sound/soc/codecs/hdac_hdmi.c
932
list_del(&p->head);
sound/soc/codecs/hdac_hdmi.c
941
list_for_each_entry(pcm, &hdmi->pcm_list, head) {
sound/soc/codecs/hdac_hdmi.c
943
list_add_tail(&port->head, &pcm->port_list);
sound/soc/codecs/hdac_hdmi.c
98
struct list_head head;
sound/soc/codecs/sigmadsp.c
250
list_add_tail(&ctrl->head, &sigmadsp->ctrl_list);
sound/soc/codecs/sigmadsp.c
281
list_add_tail(&data->head, &sigmadsp->data_list);
sound/soc/codecs/sigmadsp.c
30
struct list_head head;
sound/soc/codecs/sigmadsp.c
42
struct list_head head;
sound/soc/codecs/sigmadsp.c
423
list_add_tail(&data->head, &sigmadsp->data_list);
sound/soc/codecs/sigmadsp.c
470
list_for_each_entry_safe(ctrl, _ctrl, &sigmadsp->ctrl_list, head) {
sound/soc/codecs/sigmadsp.c
475
list_for_each_entry_safe(data, _data, &sigmadsp->data_list, head)
sound/soc/codecs/sigmadsp.c
710
list_for_each_entry(ctrl, &sigmadsp->ctrl_list, head) {
sound/soc/codecs/sigmadsp.c
745
list_for_each_entry(data, &sigmadsp->data_list, head) {
sound/soc/codecs/sigmadsp.c
755
list_for_each_entry(ctrl, &sigmadsp->ctrl_list, head)
sound/soc/codecs/sigmadsp.c
779
list_for_each_entry(ctrl, &sigmadsp->ctrl_list, head)
sound/soc/intel/boards/ehl_rt5660.c
104
pcm = list_first_entry(&ctx->hdmi_pcm_list, struct sof_hdmi_pcm, head);
sound/soc/intel/boards/ehl_rt5660.c
69
struct list_head head;
sound/soc/intel/boards/ehl_rt5660.c
88
list_add_tail(&pcm->head, &ctx->hdmi_pcm_list);
sound/soc/intel/boards/sof_es8336.c
266
list_add_tail(&pcm->head, &priv->hdmi_pcm_list);
sound/soc/intel/boards/sof_es8336.c
415
pcm = list_first_entry(&priv->hdmi_pcm_list, struct sof_hdmi_pcm, head);
sound/soc/intel/boards/sof_es8336.c
70
struct list_head head;
sound/soc/intel/boards/sof_pcm512x.c
149
pcm = list_first_entry(&ctx->hdmi_pcm_list, struct sof_hdmi_pcm, head);
sound/soc/intel/boards/sof_pcm512x.c
43
struct list_head head;
sound/soc/intel/boards/sof_pcm512x.c
85
list_add_tail(&pcm->head, &ctx->hdmi_pcm_list);
sound/soc/sof/ipc3-loader.c
114
const struct sof_ext_man_header *head;
sound/soc/sof/ipc3-loader.c
116
head = (struct sof_ext_man_header *)fw->data;
sound/soc/sof/ipc3-loader.c
123
if (fw->size < sizeof(*head))
sound/soc/sof/ipc3-loader.c
130
if (head->magic == SOF_EXT_MAN_MAGIC_NUMBER)
sound/soc/sof/ipc3-loader.c
131
return head->full_size;
sound/soc/sof/ipc3-loader.c
135
head->magic);
sound/soc/sof/ipc3-loader.c
143
const struct sof_ext_man_header *head;
sound/soc/sof/ipc3-loader.c
149
head = (struct sof_ext_man_header *)fw->data;
sound/soc/sof/ipc3-loader.c
150
remaining = head->full_size - head->header_size;
sound/soc/sof/ipc3-loader.c
161
head->header_version)) {
sound/soc/sof/ipc3-loader.c
164
head->header_version, SOF_EXT_MAN_VERSION);
sound/soc/sof/ipc3-loader.c
169
iptr = (uintptr_t)fw->data + head->header_size;
sound/usb/fcp.c
340
elem->head.mixer = mixer;
sound/usb/fcp.c
342
elem->head.id = 0;
sound/usb/fcp.c
355
err = snd_usb_mixer_add_control(&elem->head, kctl);
sound/usb/fcp.c
384
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/fcp.c
429
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer.c
1077
struct snd_usb_audio *chip = cval->head.mixer->chip;
sound/usb/mixer.c
1220
if (!cval->head.mixer->ignore_ctl_error)
sound/usb/mixer.c
1221
usb_audio_warn(cval->head.mixer->chip,
sound/usb/mixer.c
1223
cval->head.id, mixer_ctrl_intf(cval->head.mixer),
sound/usb/mixer.c
1256
usb_audio_err(cval->head.mixer->chip,
sound/usb/mixer.c
1258
cval->head.id, mixer_ctrl_intf(cval->head.mixer),
sound/usb/mixer.c
1259
cval->control, cval->head.id);
sound/usb/mixer.c
1266
} else if (cval->head.mixer->protocol == UAC_VERSION_1) {
sound/usb/mixer.c
130
((cval)->head.mixer->ignore_ctl_error ? 0 : (err))
sound/usb/mixer.c
1340
usb_audio_info(cval->head.mixer->chip,
sound/usb/mixer.c
1342
cval->head.id, mixer_ctrl_intf(cval->head.mixer),
sound/usb/mixer.c
1399
snd_ctl_notify(cval->head.mixer->chip->card,
sound/usb/mixer.c
1505
struct snd_usb_audio *chip = cval->head.mixer->chip;
sound/usb/mixer.c
1516
idx = mixer_ctrl_intf(cval->head.mixer) | (cval->head.id << 8);
sound/usb/mixer.c
1517
if (cval->head.mixer->protocol == UAC_VERSION_2) {
sound/usb/mixer.c
1690
snd_usb_mixer_elem_init_std(&cval->head, mixer, unitid);
sound/usb/mixer.c
1796
cval->head.id, kctl->id.name);
sound/usb/mixer.c
1825
cval->head.id, kctl->id.name, cval->channels,
sound/usb/mixer.c
1830
cval->head.id, kctl->id.name, cval->channels,
sound/usb/mixer.c
1832
snd_usb_mixer_add_control(&cval->head, kctl);
sound/usb/mixer.c
1900
snd_usb_mixer_elem_init_std(&cval->head, mixer, term->id);
sound/usb/mixer.c
1903
cval->head.resume = connector_mixer_resume;
sound/usb/mixer.c
1937
snd_usb_mixer_add_control(&cval->head, kctl);
sound/usb/mixer.c
1963
snd_usb_mixer_elem_init_std(&cval->head, state->mixer, hdr->bClockID);
sound/usb/mixer.c
1989
return snd_usb_mixer_add_control(&cval->head, kctl);
sound/usb/mixer.c
2184
snd_usb_mixer_elem_init_std(&cval->head, state->mixer, unitid);
sound/usb/mixer.c
2217
cval->head.id, kctl->id.name, cval->channels, cval->min, cval->max);
sound/usb/mixer.c
2218
snd_usb_mixer_add_control(&cval->head, kctl);
sound/usb/mixer.c
2531
snd_usb_mixer_elem_init_std(&cval->head, state->mixer, unitid);
sound/usb/mixer.c
2623
cval->head.id, kctl->id.name, cval->channels,
sound/usb/mixer.c
2626
err = snd_usb_mixer_add_control(&cval->head, kctl);
sound/usb/mixer.c
2777
snd_usb_mixer_elem_init_std(&cval->head, state->mixer, unitid);
sound/usb/mixer.c
2870
cval->head.id, kctl->id.name, desc->bNrInPins);
sound/usb/mixer.c
2871
return snd_usb_mixer_add_control(&cval->head, kctl);
sound/usb/mixer.c
310
struct snd_usb_audio *chip = cval->head.mixer->chip;
sound/usb/mixer.c
321
idx = mixer_ctrl_intf(cval->head.mixer) | (cval->head.id << 8);
sound/usb/mixer.c
3402
"channels=%i, type=\"%s\"\n", cval->head.id,
sound/usb/mixer.c
341
struct snd_usb_audio *chip = cval->head.mixer->chip;
sound/usb/mixer.c
3483
&info->head.kctl->id);
sound/usb/mixer.c
365
idx = mixer_ctrl_intf(cval->head.mixer) | (cval->head.id << 8);
sound/usb/mixer.c
408
return (cval->head.mixer->protocol == UAC_VERSION_1) ?
sound/usb/mixer.c
439
if (!cval->head.mixer->ignore_ctl_error)
sound/usb/mixer.c
440
usb_audio_dbg(cval->head.mixer->chip,
sound/usb/mixer.c
457
struct snd_usb_audio *chip = cval->head.mixer->chip;
sound/usb/mixer.c
464
if (cval->head.mixer->protocol == UAC_VERSION_1) {
sound/usb/mixer.c
489
idx = mixer_ctrl_intf(cval->head.mixer) | (cval->head.id << 8);
sound/usb/mixer.c
519
usb_audio_dbg(cval->head.mixer->chip,
sound/usb/mixer.h
79
container_of(list, struct usb_mixer_elem_info, head)
sound/usb/mixer.h
82
struct usb_mixer_elem_list head;
sound/usb/mixer_quirks.c
109
return snd_usb_mixer_add_control(&cval->head, kctl);
sound/usb/mixer_quirks.c
2221
struct snd_usb_audio *chip = cval->head.mixer->chip;
sound/usb/mixer_quirks.c
2284
snd_usb_mixer_elem_init_std(&cval->head, mixer, unitid);
sound/usb/mixer_quirks.c
2285
cval->head.resume = realtek_resume_jack;
sound/usb/mixer_quirks.c
2298
return snd_usb_mixer_add_control(&cval->head, kctl);
sound/usb/mixer_quirks.c
2385
struct snd_usb_audio *chip = cval->head.mixer->chip;
sound/usb/mixer_quirks.c
555
me = &mei->info.head;
sound/usb/mixer_quirks.c
575
snd_dev = mei->info.head.mixer->chip->dev;
sound/usb/mixer_quirks.c
658
me = &mei->info.head;
sound/usb/mixer_quirks.c
720
snd_usb_mixer_elem_init_std(&mei->info.head, mixer,
sound/usb/mixer_quirks.c
724
mei->info.head.resume = snd_dualsense_resume_jack;
sound/usb/mixer_quirks.c
739
err = snd_usb_mixer_add_control(&mei->info.head, kctl);
sound/usb/mixer_quirks.c
74
snd_usb_mixer_elem_init_std(&cval->head, mixer, unitid);
sound/usb/mixer_s1810c.c
541
elem->head.mixer = mixer;
sound/usb/mixer_s1810c.c
543
elem->head.id = 0;
sound/usb/mixer_s1810c.c
553
return snd_usb_mixer_add_control(&elem->head, kctl);
sound/usb/mixer_scarlett.c
258
struct snd_usb_audio *chip = elem->head.mixer->chip;
sound/usb/mixer_scarlett.c
261
int idx = snd_usb_ctrl_intf(elem->head.mixer->hostif) | (elem->head.id << 8);
sound/usb/mixer_scarlett.c
702
struct snd_usb_audio *chip = elem->head.mixer->chip;
sound/usb/mixer_scarlett.c
705
int idx = snd_usb_ctrl_intf(elem->head.mixer->hostif) | (elem->head.id << 8);
sound/usb/mixer_scarlett.c
826
elem->head.mixer = mixer;
sound/usb/mixer_scarlett.c
827
elem->head.resume = resume;
sound/usb/mixer_scarlett.c
830
elem->head.id = index;
sound/usb/mixer_scarlett.c
847
err = snd_usb_mixer_add_control(&elem->head, kctl);
sound/usb/mixer_scarlett2.c
3206
elem->head.mixer = mixer;
sound/usb/mixer_scarlett2.c
3208
elem->head.id = 0;
sound/usb/mixer_scarlett2.c
3221
err = snd_usb_mixer_add_control(&elem->head, kctl);
sound/usb/mixer_scarlett2.c
3238
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
3277
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
3335
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3595
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3612
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3634
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3656
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3695
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
3725
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3751
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3774
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3854
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3876
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3915
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3970
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
3990
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4013
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4070
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4096
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4119
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4180
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4203
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4262
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4263
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
4281
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4388
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4411
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4446
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4469
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4565
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4588
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4649
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
4725
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4780
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4800
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4825
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4888
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4911
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4965
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
4987
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5023
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5077
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5099
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5175
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
5186
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5258
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
5269
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
5281
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5310
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5418
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
5434
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
5450
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5498
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5597
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5620
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5754
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5776
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5826
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
5836
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
5947
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6026
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6104
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6126
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6170
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6183
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6255
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6277
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6327
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
6339
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6786
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6808
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6914
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6936
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
6995
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
7006
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
7106
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
7151
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
7174
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
7265
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
7324
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
7334
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
7388
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
7398
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
7478
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
7550
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
7572
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_scarlett2.c
7661
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
7676
struct scarlett2_data *private = elem->head.mixer->private_data;
sound/usb/mixer_scarlett2.c
7686
struct usb_mixer_interface *mixer = elem->head.mixer;
sound/usb/mixer_us16x08.c
1066
elem->head.mixer = mixer;
sound/usb/mixer_us16x08.c
1067
elem->head.resume = NULL;
sound/usb/mixer_us16x08.c
1070
elem->head.id = index;
sound/usb/mixer_us16x08.c
1088
err = snd_usb_mixer_add_control(&elem->head, kctl);
sound/usb/mixer_us16x08.c
195
struct snd_usb_audio *chip = elem->head.mixer->chip;
sound/usb/mixer_us16x08.c
263
struct snd_usb_audio *chip = elem->head.mixer->chip;
sound/usb/mixer_us16x08.c
280
buf[6] = elem->head.id;
sound/usb/mixer_us16x08.c
300
struct snd_usb_audio *chip = elem->head.mixer->chip;
sound/usb/mixer_us16x08.c
307
switch (elem->head.id) {
sound/usb/mixer_us16x08.c
321
buf[6] = elem->head.id;
sound/usb/mixer_us16x08.c
342
switch (elem->head.id) {
sound/usb/mixer_us16x08.c
373
struct snd_usb_audio *chip = elem->head.mixer->chip;
sound/usb/mixer_us16x08.c
390
buf[6] = elem->head.id;
sound/usb/mixer_us16x08.c
422
int val_idx = COMP_STORE_IDX(elem->head.id);
sound/usb/mixer_us16x08.c
433
struct snd_usb_audio *chip = elem->head.mixer->chip;
sound/usb/mixer_us16x08.c
448
val_idx = elem->head.id - SND_US16X08_ID_COMP_BASE;
sound/usb/mixer_us16x08.c
492
val = store->val[EQ_STORE_BAND_IDX(elem->head.id)]
sound/usb/mixer_us16x08.c
493
[EQ_STORE_PARAM_IDX(elem->head.id)][index];
sound/usb/mixer_us16x08.c
503
struct snd_usb_audio *chip = elem->head.mixer->chip;
sound/usb/mixer_us16x08.c
549
int b_idx = EQ_STORE_BAND_IDX(elem->head.id) - 1;
sound/usb/mixer_us16x08.c
550
int p_idx = EQ_STORE_PARAM_IDX(elem->head.id);
sound/usb/mixer_us16x08.c
563
struct snd_usb_audio *chip = elem->head.mixer->chip;
sound/usb/mixer_us16x08.c
568
int b_idx = EQ_STORE_BAND_IDX(elem->head.id) - 1;
sound/usb/mixer_us16x08.c
569
int p_idx = EQ_STORE_PARAM_IDX(elem->head.id);
sound/usb/mixer_us16x08.c
695
struct snd_usb_audio *chip = elem->head.mixer->chip;
sound/usb/qcom/qc_audio_offload.c
394
struct list_head *head, size_t size)
sound/usb/qcom/qc_audio_offload.c
410
list_for_each_entry(info, head, list) {
sound/usb/qcom/qc_audio_offload.c
453
list_add_tail(&info->list, head);
sound/usb/qcom/qc_audio_offload.c
460
*head, size_t *curr_iova_size)
sound/usb/qcom/qc_audio_offload.c
466
list_for_each_entry(info, head, list) {
sound/usb/usx2y/usx2yhwdeppcm.c
145
int pack, head;
sound/usb/usx2y/usx2yhwdeppcm.c
151
head = shm->captured_iso_head + 1;
sound/usb/usx2y/usx2yhwdeppcm.c
152
if (head >= ARRAY_SIZE(shm->captured_iso))
sound/usb/usx2y/usx2yhwdeppcm.c
153
head = 0;
sound/usb/usx2y/usx2yhwdeppcm.c
154
shm->captured_iso[head].frame = urb->start_frame + pack;
sound/usb/usx2y/usx2yhwdeppcm.c
155
shm->captured_iso[head].offset = desc->offset;
sound/usb/usx2y/usx2yhwdeppcm.c
156
shm->captured_iso[head].length = desc->actual_length;
sound/usb/usx2y/usx2yhwdeppcm.c
157
shm->captured_iso_head = head;
sound/usb/usx2y/usx2yhwdeppcm.c
55
int head;
sound/usb/usx2y/usx2yhwdeppcm.c
58
head = usx2y->hwdep_pcm_shm->captured_iso_head + 1;
sound/usb/usx2y/usx2yhwdeppcm.c
59
if (head >= ARRAY_SIZE(usx2y->hwdep_pcm_shm->captured_iso))
sound/usb/usx2y/usx2yhwdeppcm.c
60
head = 0;
sound/usb/usx2y/usx2yhwdeppcm.c
61
usx2y->hwdep_pcm_shm->capture_iso_start = head;
sound/usb/usx2y/usx2yhwdeppcm.c
62
dev_dbg(&usx2y->dev->dev, "cap start %i\n", head);
tools/bpf/bpftool/cfg.c
106
new_bb->head = insn;
tools/bpf/bpftool/cfg.c
208
bb->tail = bb_next(bb)->head - 1;
tools/bpf/bpftool/cfg.c
249
if (bb->head == insn)
tools/bpf/bpftool/cfg.c
30
struct bpf_insn *head;
tools/bpf/bpftool/cfg.c
405
start_idx = bb->head - func->start;
tools/bpf/bpftool/cfg.c
406
dump_xlated_for_graph(dd, bb->head, bb->tail, start_idx,
tools/bpf/bpftool/cfg.c
94
if (bb->head == insn)
tools/bpf/bpftool/cfg.c
96
else if (bb->head > insn)
tools/include/io_uring/mini_liburing.h
116
cq->khead = ptr + p->cq_off.head;
tools/include/io_uring/mini_liburing.h
12
unsigned int *head;
tools/include/io_uring/mini_liburing.h
187
unsigned int head = *cq->khead;
tools/include/io_uring/mini_liburing.h
193
if (head != *cq->ktail) {
tools/include/io_uring/mini_liburing.h
194
*cqe_ptr = &cq->cqes[head & mask];
tools/include/io_uring/mini_liburing.h
21
unsigned int *head;
tools/include/io_uring/mini_liburing.h
89
sq->khead = ptr + p->sq_off.head;
tools/include/linux/list.h
154
static inline void list_move(struct list_head *list, struct list_head *head)
tools/include/linux/list.h
157
list_add(list, head);
tools/include/linux/list.h
166
struct list_head *head)
tools/include/linux/list.h
169
list_add_tail(list, head);
tools/include/linux/list.h
177
static inline int list_is_first(const struct list_head *list, const struct list_head *head)
tools/include/linux/list.h
179
return list->prev == head;
tools/include/linux/list.h
188
const struct list_head *head)
tools/include/linux/list.h
190
return list->next == head;
tools/include/linux/list.h
197
static inline int list_empty(const struct list_head *head)
tools/include/linux/list.h
199
return head->next == head;
tools/include/linux/list.h
215
static inline int list_empty_careful(const struct list_head *head)
tools/include/linux/list.h
217
struct list_head *next = head->next;
tools/include/linux/list.h
218
return (next == head) && (next == head->prev);
tools/include/linux/list.h
225
static inline void list_rotate_left(struct list_head *head)
tools/include/linux/list.h
229
if (!list_empty(head)) {
tools/include/linux/list.h
230
first = head->next;
tools/include/linux/list.h
231
list_move_tail(first, head);
tools/include/linux/list.h
239
static inline int list_is_singular(const struct list_head *head)
tools/include/linux/list.h
241
return !list_empty(head) && (head->next == head->prev);
tools/include/linux/list.h
245
struct list_head *head, struct list_head *entry)
tools/include/linux/list.h
248
list->next = head->next;
tools/include/linux/list.h
252
head->next = new_first;
tools/include/linux/list.h
253
new_first->prev = head;
tools/include/linux/list.h
271
struct list_head *head, struct list_head *entry)
tools/include/linux/list.h
273
if (list_empty(head))
tools/include/linux/list.h
275
if (list_is_singular(head) &&
tools/include/linux/list.h
276
(head->next != entry && head != entry))
tools/include/linux/list.h
278
if (entry == head)
tools/include/linux/list.h
281
__list_cut_position(list, head, entry);
tools/include/linux/list.h
304
struct list_head *head)
tools/include/linux/list.h
307
__list_splice(list, head, head->next);
tools/include/linux/list.h
316
struct list_head *head)
tools/include/linux/list.h
319
__list_splice(list, head->prev, head);
tools/include/linux/list.h
330
struct list_head *head)
tools/include/linux/list.h
333
__list_splice(list, head, head->next);
tools/include/linux/list.h
347
struct list_head *head)
tools/include/linux/list.h
350
__list_splice(list, head->prev, head);
tools/include/linux/list.h
429
#define list_for_each(pos, head) \
tools/include/linux/list.h
430
for (pos = (head)->next; pos != (head); pos = pos->next)
tools/include/linux/list.h
437
#define list_for_each_prev(pos, head) \
tools/include/linux/list.h
438
for (pos = (head)->prev; pos != (head); pos = pos->prev)
tools/include/linux/list.h
446
#define list_for_each_safe(pos, n, head) \
tools/include/linux/list.h
447
for (pos = (head)->next, n = pos->next; pos != (head); \
tools/include/linux/list.h
456
#define list_for_each_prev_safe(pos, n, head) \
tools/include/linux/list.h
457
for (pos = (head)->prev, n = pos->prev; \
tools/include/linux/list.h
458
pos != (head); \
tools/include/linux/list.h
467
#define list_for_each_entry(pos, head, member) \
tools/include/linux/list.h
468
for (pos = list_first_entry(head, typeof(*pos), member); \
tools/include/linux/list.h
469
&pos->member != (head); \
tools/include/linux/list.h
478
#define list_for_each_entry_reverse(pos, head, member) \
tools/include/linux/list.h
479
for (pos = list_last_entry(head, typeof(*pos), member); \
tools/include/linux/list.h
480
&pos->member != (head); \
tools/include/linux/list.h
491
#define list_prepare_entry(pos, head, member) \
tools/include/linux/list.h
492
((pos) ? : list_entry(head, typeof(*pos), member))
tools/include/linux/list.h
503
#define list_for_each_entry_continue(pos, head, member) \
tools/include/linux/list.h
505
&pos->member != (head); \
tools/include/linux/list.h
517
#define list_for_each_entry_continue_reverse(pos, head, member) \
tools/include/linux/list.h
519
&pos->member != (head); \
tools/include/linux/list.h
530
#define list_for_each_entry_from(pos, head, member) \
tools/include/linux/list.h
531
for (; &pos->member != (head); \
tools/include/linux/list.h
541
#define list_for_each_entry_safe(pos, n, head, member) \
tools/include/linux/list.h
542
for (pos = list_first_entry(head, typeof(*pos), member), \
tools/include/linux/list.h
544
&pos->member != (head); \
tools/include/linux/list.h
557
#define list_for_each_entry_safe_continue(pos, n, head, member) \
tools/include/linux/list.h
560
&pos->member != (head); \
tools/include/linux/list.h
573
#define list_for_each_entry_safe_from(pos, n, head, member) \
tools/include/linux/list.h
575
&pos->member != (head); \
tools/include/linux/list.h
588
#define list_for_each_entry_safe_reverse(pos, n, head, member) \
tools/include/linux/list.h
589
for (pos = list_last_entry(head, typeof(*pos), member), \
tools/include/linux/list.h
591
&pos->member != (head); \
tools/include/linux/list.h
61
static inline void list_add(struct list_head *new, struct list_head *head)
tools/include/linux/list.h
63
__list_add(new, head, head->next);
tools/include/linux/list.h
717
#define hlist_for_each(pos, head) \
tools/include/linux/list.h
718
for (pos = (head)->first; pos ; pos = pos->next)
tools/include/linux/list.h
720
#define hlist_for_each_safe(pos, n, head) \
tools/include/linux/list.h
721
for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \
tools/include/linux/list.h
735
#define hlist_for_each_entry(pos, head, member) \
tools/include/linux/list.h
736
for (pos = hlist_entry_safe((head)->first, typeof(*(pos)), member);\
tools/include/linux/list.h
75
static inline void list_add_tail(struct list_head *new, struct list_head *head)
tools/include/linux/list.h
766
#define hlist_for_each_entry_safe(pos, n, head, member) \
tools/include/linux/list.h
767
for (pos = hlist_entry_safe((head)->first, typeof(*pos), member);\
tools/include/linux/list.h
77
__list_add(new, head->prev, head);
tools/include/linux/list.h
790
#define list_for_each_from(pos, head) \
tools/include/linux/list.h
791
for (; pos != (head); pos = pos->next)
tools/include/linux/list_sort.h
13
void list_sort(void *priv, struct list_head *head, list_cmp_func_t cmp);
tools/include/linux/ring_buffer.h
61
u64 head = READ_ONCE(base->data_head);
tools/include/linux/ring_buffer.h
64
return head;
tools/include/uapi/linux/io_uring.h
436
__u32 head;
tools/include/uapi/linux/io_uring.h
455
__u32 head;
tools/lib/bpf/nlattr.c
104
int libbpf_nla_parse(struct nlattr *tb[], int maxtype, struct nlattr *head,
tools/lib/bpf/nlattr.c
112
libbpf_nla_for_each_attr(nla, head, len, rem) {
tools/lib/bpf/nlattr.h
123
int libbpf_nla_parse(struct nlattr *tb[], int maxtype, struct nlattr *head,
tools/lib/bpf/nlattr.h
75
#define libbpf_nla_for_each_attr(pos, head, len, rem) \
tools/lib/bpf/nlattr.h
76
for (pos = head, rem = len; \
tools/lib/list_sort.c
16
struct list_head *head, **tail = &head;
tools/lib/list_sort.c
172
void list_sort(void *priv, struct list_head *head, list_cmp_func_t cmp)
tools/lib/list_sort.c
174
struct list_head *list = head->next, *pending = NULL;
tools/lib/list_sort.c
177
if (list == head->prev) /* Zero or one elements */
tools/lib/list_sort.c
181
head->prev->next = NULL;
tools/lib/list_sort.c
238
merge_final(priv, cmp, head, pending, list);
tools/lib/list_sort.c
38
return head;
tools/lib/list_sort.c
49
static void merge_final(void *priv, list_cmp_func_t cmp, struct list_head *head,
tools/lib/list_sort.c
52
struct list_head *tail = head;
tools/lib/list_sort.c
84
tail->next = head;
tools/lib/list_sort.c
85
head->prev = tail;
tools/lib/perf/evsel.c
569
struct hlist_head *head;
tools/lib/perf/evsel.c
577
head = &sid->periods[hash];
tools/lib/perf/evsel.c
579
hlist_for_each_entry(res, head, hnode)
tools/lib/perf/mmap.c
149
u64 head = perf_mmap__read_head(md);
tools/lib/perf/mmap.c
154
md->start = md->overwrite ? head : old;
tools/lib/perf/mmap.c
155
md->end = md->overwrite ? old : head;
tools/lib/perf/mmap.c
165
md->prev = head;
tools/objtool/check.c
364
struct hlist_head *head = &cfi_hash[hash_min(cfi_key(cfi), cfi_bits)];
tools/objtool/check.c
367
hlist_for_each_entry(obj, head, hash) {
tools/objtool/check.c
376
hlist_add_head(&obj->hash, head);
tools/objtool/check.c
383
struct hlist_head *head = &cfi_hash[hash_min(cfi_key(cfi), cfi_bits)];
tools/objtool/check.c
385
hlist_add_head(&cfi->hash, head);
tools/objtool/elf.c
47
struct elf_hash_node **head)
tools/objtool/elf.c
51
if (node == *head) {
tools/objtool/elf.c
52
*head = node->next;
tools/objtool/elf.c
56
for (prev = NULL, cur = *head; cur; prev = cur, cur = cur->next) {
tools/perf/arch/arm64/util/arm-spe.c
534
size_t buffer_size, u64 head)
tools/perf/arch/arm64/util/arm-spe.c
545
if (head >= buffer_size)
tools/perf/arch/arm64/util/arm-spe.c
564
if (head > watermark)
tools/perf/arch/arm64/util/arm-spe.c
565
watermark = head;
tools/perf/arch/arm64/util/arm-spe.c
586
u64 *head, u64 *old)
tools/perf/arch/arm64/util/arm-spe.c
610
if (!wrapped && arm_spe_buffer_has_wrapped(data, mm->len, *head)) {
tools/perf/arch/arm64/util/arm-spe.c
616
__func__, idx, (size_t)*old, (size_t)*head, mm->len);
tools/perf/arch/arm64/util/arm-spe.c
628
if (*head >= mm->len) {
tools/perf/arch/arm64/util/arm-spe.c
629
*old = *head - mm->len;
tools/perf/arch/arm64/util/arm-spe.c
631
*head += mm->len;
tools/perf/arch/arm64/util/arm-spe.c
632
*old = *head - mm->len;
tools/perf/arch/x86/util/intel-bts.c
367
u64 *head, u64 *old)
tools/perf/arch/x86/util/intel-bts.c
375
__func__, idx, (size_t)*old, (size_t)*head);
tools/perf/arch/x86/util/intel-bts.c
396
*old = *head;
tools/perf/arch/x86/util/intel-bts.c
397
*head += mm->len;
tools/perf/arch/x86/util/intel-bts.c
403
if (*old > *head)
tools/perf/arch/x86/util/intel-bts.c
404
*head += mm->len;
tools/perf/arch/x86/util/intel-bts.c
408
__func__, wrapped ? "" : "not ", (size_t)*old, (size_t)*head);
tools/perf/arch/x86/util/intel-pt.c
1051
void *data, size_t head)
tools/perf/arch/x86/util/intel-pt.c
1056
if (head > ref_offset || head < ref_end - buf_size)
tools/perf/arch/x86/util/intel-pt.c
1058
} else if (head > ref_offset && head < ref_end) {
tools/perf/arch/x86/util/intel-pt.c
1067
void *data, size_t head)
tools/perf/arch/x86/util/intel-pt.c
1069
if (head >= ref_size) {
tools/perf/arch/x86/util/intel-pt.c
1070
memcpy(ref_buf, data + head - ref_size, ref_size);
tools/perf/arch/x86/util/intel-pt.c
1072
memcpy(ref_buf, data, head);
tools/perf/arch/x86/util/intel-pt.c
1073
ref_size -= head;
tools/perf/arch/x86/util/intel-pt.c
1074
memcpy(ref_buf + head, data + buf_size - ref_size, ref_size);
tools/perf/arch/x86/util/intel-pt.c
1080
u64 head)
tools/perf/arch/x86/util/intel-pt.c
1087
data, head);
tools/perf/arch/x86/util/intel-pt.c
1090
data, head);
tools/perf/arch/x86/util/intel-pt.c
1114
u64 *head, u64 *old)
tools/perf/arch/x86/util/intel-pt.c
1122
__func__, idx, (size_t)*old, (size_t)*head);
tools/perf/arch/x86/util/intel-pt.c
1140
wrapped = intel_pt_wrapped(ptr, idx, mm, data, *head);
tools/perf/arch/x86/util/intel-pt.c
1156
*old = *head;
tools/perf/arch/x86/util/intel-pt.c
1157
*head += mm->len;
tools/perf/arch/x86/util/intel-pt.c
1163
if (*old > *head)
tools/perf/arch/x86/util/intel-pt.c
1164
*head += mm->len;
tools/perf/arch/x86/util/intel-pt.c
1168
__func__, wrapped ? "" : "not ", (size_t)*old, (size_t)*head);
tools/perf/builtin-annotate.c
373
static void print_annotate_item_stat(struct list_head *head, const char *title)
tools/perf/builtin-annotate.c
381
list_splice_init(head, &tmp);
tools/perf/builtin-annotate.c
389
list_for_each_entry(iter, head, list) {
tools/perf/builtin-annotate.c
404
list_for_each_entry(istat, head, list)
tools/perf/builtin-diff.c
510
list_for_each_entry(pair, &he->pairs.head, pairs.node)
tools/perf/builtin-ftrace.c
1588
struct list_head *head = opt->value;
tools/perf/builtin-ftrace.c
1596
list_add_tail(&entry->list, head);
tools/perf/builtin-ftrace.c
1601
static void delete_filter_func(struct list_head *head)
tools/perf/builtin-ftrace.c
1605
list_for_each_entry_safe(pos, tmp, head, list) {
tools/perf/builtin-ftrace.c
1614
struct list_head *head = opt->value;
tools/perf/builtin-ftrace.c
1629
list_add_tail(&entry->list, head);
tools/perf/builtin-inject.c
1276
struct hlist_head *head;
tools/perf/builtin-inject.c
1281
head = &gs->heads[hash];
tools/perf/builtin-inject.c
1283
hlist_for_each_entry(guest_id, head, node)
tools/perf/builtin-inject.c
1383
struct hlist_head *head;
tools/perf/builtin-inject.c
1388
head = &gs->tids[hash];
tools/perf/builtin-inject.c
1390
hlist_for_each_entry(guest_tid, head, node)
tools/perf/builtin-kwork.c
1513
static unsigned long long nr_list_entry(struct list_head *head)
tools/perf/builtin-kwork.c
1518
list_for_each(pos, head)
tools/perf/builtin-sched.c
4171
static int get_all_cpu_stats(struct list_head *head)
tools/perf/builtin-sched.c
4173
struct schedstat_cpu *cptr = list_first_entry(head, struct schedstat_cpu, cpu_list);
tools/perf/builtin-sched.c
4206
list_for_each_entry(cptr, head, cpu_list) {
tools/perf/builtin-sched.c
4207
if (list_is_first(&cptr->cpu_list, head))
tools/perf/builtin-sched.c
4210
if (list_is_last(&cptr->cpu_list, head))
tools/perf/builtin-sched.c
4224
list_add(&summary_head->cpu_list, head);
tools/perf/builtin-sched.c
4468
static void free_schedstat(struct list_head *head)
tools/perf/builtin-sched.c
4473
list_for_each_entry_safe(cptr, n2, head, cpu_list) {
tools/perf/builtin-script.c
2253
list_splice_init(&metric_me->head, &script_me->head);
tools/perf/builtin-script.c
2254
list_for_each_entry(expr, &script_me->head, nd) {
tools/perf/check-header_ignore_hunks/lib/list_sort.c
4
struct list_head *tail = head;
tools/perf/tests/parse-metric.c
58
list_for_each_entry (mexp, &me->head, nd) {
tools/perf/tests/pmu-events.c
883
list_for_each_entry (mexp, &me->head, nd) {
tools/perf/ui/browser.c
105
struct list_head *head = browser->entries;
tools/perf/ui/browser.c
113
pos = ui_browser__list_head_filter_entries(browser, head->next);
tools/perf/ui/browser.c
119
pos = ui_browser__list_head_filter_prev_entries(browser, head->prev);
tools/perf/ui/browser.c
513
struct list_head *head = browser->entries;
tools/perf/ui/browser.c
517
browser->top = ui_browser__list_head_filter_entries(browser, head->next);
tools/perf/ui/browser.c
521
list_for_each_from(pos, head) {
tools/perf/ui/browsers/annotate.c
387
struct list_head *head = browser->b.entries;
tools/perf/ui/browsers/annotate.c
390
list_for_each_entry(al, head, node) {
tools/perf/ui/hist.c
218
list_for_each_entry(pair, &a->pairs.head, pairs.node) {
tools/perf/ui/hist.c
223
list_for_each_entry(pair, &b->pairs.head, pairs.node) {
tools/perf/ui/hist.c
89
list_for_each_entry(pair, &he->pairs.head, pairs.node) {
tools/perf/util/annotate-data.c
1833
list_for_each_entry(pair, &he->pairs.head, pairs.node)
tools/perf/util/annotate.c
1487
size_t disasm__fprintf(struct list_head *head, FILE *fp)
tools/perf/util/annotate.c
1492
list_for_each_entry(pos, head, al.node)
tools/perf/util/annotate.c
2653
static struct annotated_item_stat *annotate_data_stat(struct list_head *head,
tools/perf/util/annotate.c
2658
list_for_each_entry(istat, head, list) {
tools/perf/util/annotate.c
2673
list_add_tail(&istat->list, head);
tools/perf/util/annotate.c
3002
static bool basic_block_has_offset(struct list_head *head, s64 offset)
tools/perf/util/annotate.c
3006
list_for_each_entry(link, head, node) {
tools/perf/util/annotate.c
3133
struct list_head *head)
tools/perf/util/annotate.c
3138
list_move(&link->bb->list, head);
tools/perf/util/annotate.c
3176
struct list_head *head)
tools/perf/util/annotate.c
3198
link_found_basic_blocks(link, head);
tools/perf/util/annotate.c
656
void annotation_line__add(struct annotation_line *al, struct list_head *head)
tools/perf/util/annotate.c
658
list_add_tail(&al->node, head);
tools/perf/util/annotate.c
662
annotation_line__next(struct annotation_line *pos, struct list_head *head)
tools/perf/util/annotate.c
664
list_for_each_entry_continue(pos, head, node)
tools/perf/util/annotate.h
145
void annotation_line__add(struct annotation_line *al, struct list_head *head);
tools/perf/util/annotate.h
189
annotation_line__next(struct annotation_line *pos, struct list_head *head);
tools/perf/util/annotate.h
222
size_t disasm__fprintf(struct list_head *head, FILE *fp);
tools/perf/util/annotate.h
581
struct list_head *head);
tools/perf/util/arm-spe.c
1210
if (list_empty(&queue->head) || speq)
tools/perf/util/auxtrace.c
1001
list_for_each_entry(auxtrace_index, head, list) {
tools/perf/util/auxtrace.c
1010
static int auxtrace_index__process_entry(int fd, struct list_head *head,
tools/perf/util/auxtrace.c
1020
auxtrace_index = auxtrace_index__last(head);
tools/perf/util/auxtrace.c
1042
struct list_head *head = &session->auxtrace_index;
tools/perf/util/auxtrace.c
1057
err = auxtrace_index__process_entry(fd, head, needs_swap);
tools/perf/util/auxtrace.c
1101
if (list_is_last(&buffer->list, &queue->head))
tools/perf/util/auxtrace.c
1106
if (list_empty(&queue->head))
tools/perf/util/auxtrace.c
1108
return list_entry(queue->head.next, struct auxtrace_buffer,
tools/perf/util/auxtrace.c
1918
u64 head, old = mm->prev, offset, ref;
tools/perf/util/auxtrace.c
1925
head = auxtrace_mmap__read_head(mm, kernel_is_64_bit);
tools/perf/util/auxtrace.c
1928
auxtrace_record__find_snapshot(itr, mm->idx, mm, data, &head, &old))
tools/perf/util/auxtrace.c
1931
if (old == head)
tools/perf/util/auxtrace.c
1935
mm->idx, old, head, head - old);
tools/perf/util/auxtrace.c
1938
head_off = head & mm->mask;
tools/perf/util/auxtrace.c
1941
head_off = head % mm->len;
tools/perf/util/auxtrace.c
1955
if (head > old || size <= head || mm->mask) {
tools/perf/util/auxtrace.c
1956
offset = head - size;
tools/perf/util/auxtrace.c
1965
offset = head - size - rem;
tools/perf/util/auxtrace.c
2005
mm->prev = head;
tools/perf/util/auxtrace.c
2010
err = auxtrace_mmap__write_tail(mm, head, kernel_is_64_bit);
tools/perf/util/auxtrace.c
2217
list_add_tail(&filt->list, &filts->head);
tools/perf/util/auxtrace.c
2230
INIT_LIST_HEAD(&filts->head);
tools/perf/util/auxtrace.c
2238
list_for_each_entry_safe(filt, n, &filts->head, list) {
tools/perf/util/auxtrace.c
228
INIT_LIST_HEAD(&queue_array[i].head);
tools/perf/util/auxtrace.c
270
list_splice_tail(&queues->queue_array[i].head,
tools/perf/util/auxtrace.c
271
&queue_array[i].head);
tools/perf/util/auxtrace.c
2846
list_for_each_entry(filt, &filts.head, list) {
tools/perf/util/auxtrace.c
329
list_add_tail(&buffer->list, &queue->head);
tools/perf/util/auxtrace.c
475
while (!list_empty(&queues->queue_array[i].head)) {
tools/perf/util/auxtrace.c
478
buffer = list_entry(queues->queue_array[i].head.next,
tools/perf/util/auxtrace.c
620
unsigned char *data, u64 *head, u64 *old)
tools/perf/util/auxtrace.c
623
return itr->find_snapshot(itr, idx, mm, data, head, old);
tools/perf/util/auxtrace.c
906
static int auxtrace_index__alloc(struct list_head *head)
tools/perf/util/auxtrace.c
917
list_add_tail(&auxtrace_index->list, head);
tools/perf/util/auxtrace.c
922
void auxtrace_index__free(struct list_head *head)
tools/perf/util/auxtrace.c
926
list_for_each_entry_safe(auxtrace_index, n, head, list) {
tools/perf/util/auxtrace.c
932
static struct auxtrace_index *auxtrace_index__last(struct list_head *head)
tools/perf/util/auxtrace.c
937
if (list_empty(head)) {
tools/perf/util/auxtrace.c
938
err = auxtrace_index__alloc(head);
tools/perf/util/auxtrace.c
943
auxtrace_index = list_entry(head->prev, struct auxtrace_index, list);
tools/perf/util/auxtrace.c
946
err = auxtrace_index__alloc(head);
tools/perf/util/auxtrace.c
949
auxtrace_index = list_entry(head->prev, struct auxtrace_index,
tools/perf/util/auxtrace.c
956
int auxtrace_index__auxtrace_event(struct list_head *head,
tools/perf/util/auxtrace.c
962
auxtrace_index = auxtrace_index__last(head);
tools/perf/util/auxtrace.c
989
int auxtrace_index__write(int fd, struct list_head *head)
tools/perf/util/auxtrace.c
995
list_for_each_entry(auxtrace_index, head, list)
tools/perf/util/auxtrace.h
275
struct list_head head;
tools/perf/util/auxtrace.h
402
u64 *head, u64 *old);
tools/perf/util/auxtrace.h
451
struct list_head head;
tools/perf/util/auxtrace.h
464
u64 head;
tools/perf/util/auxtrace.h
470
head = READ_ONCE(pc->aux_head);
tools/perf/util/auxtrace.h
474
return head;
tools/perf/util/auxtrace.h
592
unsigned char *data, u64 *head, u64 *old);
tools/perf/util/auxtrace.h
596
int auxtrace_index__auxtrace_event(struct list_head *head, union perf_event *event,
tools/perf/util/auxtrace.h
598
int auxtrace_index__write(int fd, struct list_head *head);
tools/perf/util/auxtrace.h
601
void auxtrace_index__free(struct list_head *head);
tools/perf/util/block-range.c
116
struct block_range *head = malloc(sizeof(struct block_range));
tools/perf/util/block-range.c
117
if (!head)
tools/perf/util/block-range.c
120
*head = (struct block_range){
tools/perf/util/block-range.c
127
rb_link_left_of_node(&head->node, &next->node);
tools/perf/util/block-range.c
128
rb_insert_color(&head->node, &block_ranges.root);
tools/perf/util/block-range.c
131
iter.start = head;
tools/perf/util/block-range.c
163
struct block_range *head = malloc(sizeof(struct block_range));
tools/perf/util/block-range.c
164
if (!head)
tools/perf/util/block-range.c
167
*head = (struct block_range){
tools/perf/util/block-range.c
181
rb_link_left_of_node(&head->node, &entry->node);
tools/perf/util/block-range.c
182
rb_insert_color(&head->node, &block_ranges.root);
tools/perf/util/callchain.c
1566
LIST_HEAD(head);
tools/perf/util/callchain.c
1576
list_add_tail(&new->list, &head);
tools/perf/util/callchain.c
1581
list_for_each_entry_safe_reverse(chain, new, &head, list)
tools/perf/util/callchain.c
1594
list_for_each_entry_safe(chain, new, &head, list) {
tools/perf/util/cs-etm-decoder/cs-etm-decoder.c
114
packet_queue->head = (packet_queue->head + 1) &
tools/perf/util/cs-etm-decoder/cs-etm-decoder.c
117
*packet = packet_queue->packet_buffer[packet_queue->head];
tools/perf/util/cs-etm.c
2873
list_for_each_entry(buf, &etm->queues.queue_array[i].head, list)
tools/perf/util/cs-etm.c
3331
bool empty = list_empty(&queues->queue_array[i].head);
tools/perf/util/cs-etm.c
571
queue->head = 0;
tools/perf/util/cs-etm.h
203
u32 head;
tools/perf/util/evlist.c
660
struct hlist_head *head;
tools/perf/util/evlist.c
665
head = &evlist->core.heads[hash];
tools/perf/util/evlist.c
667
hlist_for_each_entry(sid, head, node)
tools/perf/util/evlist.c
728
struct hlist_head *head;
tools/perf/util/evlist.c
748
head = &evlist->core.heads[hash];
tools/perf/util/evlist.c
750
hlist_for_each_entry(sid, head, node) {
tools/perf/util/hisi-ptt.c
40
uint32_t head = *(uint32_t *)buf;
tools/perf/util/hisi-ptt.c
42
if ((HISI_PTT_8DW_CHECK_MASK & head) == HISI_PTT_IS_8DW_PKT)
tools/perf/util/hist.c
2738
list_for_each_entry(leader, &pos->pairs.head, pairs.node) {
tools/perf/util/hist.h
243
struct list_head head;
tools/perf/util/hist.h
341
list_add_tail(&pair->pairs.node, &he->pairs.head);
tools/perf/util/intel-bts.c
172
if (list_empty(&queue->head))
tools/perf/util/intel-bts.c
254
if (b->list.prev == &queue->head)
tools/perf/util/intel-pt-decoder/intel-pt-log.c
105
size_t space = b->buf_sz - b->head;
tools/perf/util/intel-pt-decoder/intel-pt-log.c
108
memcpy(b->buf + b->head, buf, n);
tools/perf/util/intel-pt-decoder/intel-pt-log.c
111
b->head += n;
tools/perf/util/intel-pt-decoder/intel-pt-log.c
112
if (sz && b->head >= b->buf_sz) {
tools/perf/util/intel-pt-decoder/intel-pt-log.c
113
b->head = 0;
tools/perf/util/intel-pt-decoder/intel-pt-log.c
176
write_lines(b->buf + b->head, b->buf_sz - b->head, b->backend, &remove_first);
tools/perf/util/intel-pt-decoder/intel-pt-log.c
178
write_lines(b->buf, b->head, b->backend, &remove_first);
tools/perf/util/intel-pt-decoder/intel-pt-log.c
181
b->head = 0;
tools/perf/util/intel-pt-decoder/intel-pt-log.c
30
size_t head;
tools/perf/util/intel-pt.c
1581
if (list_empty(&queue->head))
tools/perf/util/intel-pt.c
955
list_for_each_entry(filt, &pt->filts.head, list) {
tools/perf/util/metricgroup.c
1534
list_add(&expr->nd, &me->head);
tools/perf/util/metricgroup.c
1673
list_for_each_entry(old_expr, &old_me->head, nd) {
tools/perf/util/metricgroup.c
1729
list_add(&new_expr->nd, &new_me->head);
tools/perf/util/metricgroup.c
86
INIT_LIST_HEAD(&me->head);
tools/perf/util/metricgroup.c
96
list_for_each_entry_safe(expr, tmp, &me->head, nd) {
tools/perf/util/metricgroup.h
26
struct list_head head; /* list of metric_expr */
tools/perf/util/mmap.c
323
u64 head = perf_mmap__read_head(&md->core);
tools/perf/util/mmap.c
355
md->core.prev = head;
tools/perf/util/ordered-events.c
226
struct list_head *head = &oe->events;
tools/perf/util/ordered-events.c
239
list_for_each_entry_safe(iter, tmp, head, list) {
tools/perf/util/ordered-events.c
256
if (list_empty(head))
tools/perf/util/ordered-events.c
259
oe->last = list_entry(head->prev, struct ordered_event, list);
tools/perf/util/ordered-events.c
295
struct list_head *head = &oe->events;
tools/perf/util/ordered-events.c
297
first = list_entry(head->next, struct ordered_event, list);
tools/perf/util/ordered-events.c
301
if (WARN_ONCE(!last || list_empty(head), "empty queue"))
tools/perf/util/parse-events.c
1105
const struct parse_events_terms *head,
tools/perf/util/parse-events.c
1111
list_for_each_entry(term, &head->terms, list)
tools/perf/util/parse-events.c
1981
const struct list_head *head)
tools/perf/util/parse-events.c
2025
list_for_each_entry(pos, head, core.node) {
tools/perf/util/parse-events.c
351
const struct parse_events_terms *head,
tools/perf/util/parse-events.y
488
struct parse_events_terms *head = $1;
tools/perf/util/parse-events.y
491
if (!head) {
tools/perf/util/parse-events.y
495
list_add_tail(&term->list, &head->terms);
tools/perf/util/parse-events.y
501
struct parse_events_terms *head = malloc(sizeof(*head));
tools/perf/util/parse-events.y
504
if (!head)
tools/perf/util/parse-events.y
506
parse_events_terms__init(head);
tools/perf/util/parse-events.y
507
list_add_tail(&term->list, &head->terms);
tools/perf/util/parse-events.y
508
$$ = head;
tools/perf/util/powerpc-vpadtl.c
453
if (list_empty(&queue->head) || vpaq)
tools/perf/util/python.c
1320
list_for_each(pos, &me->head) {
tools/perf/util/python.c
1427
list_for_each(pos, &me->head) {
tools/perf/util/python.c
2049
list_for_each(mpos, &me->head) {
tools/perf/util/s390-cpumsf.c
807
if (list_empty(&queue->head))
tools/perf/util/session.c
2041
u64 head;
tools/perf/util/session.c
2057
head = 0;
tools/perf/util/session.c
2112
if ((skip = perf_session__process_event(session, event, head, "pipe")) < 0) {
tools/perf/util/session.c
2114
head, event->header.size, event->header.type);
tools/perf/util/session.c
2119
head += size;
tools/perf/util/session.c
2122
head += skip;
tools/perf/util/session.c
2157
prefetch_event(char *buf, u64 head, size_t mmap_size,
tools/perf/util/session.c
2167
if (head + sizeof(event->header) > mmap_size)
tools/perf/util/session.c
2170
event = (union perf_event *)(buf + head);
tools/perf/util/session.c
2175
if (head + event_size <= mmap_size)
tools/perf/util/session.c
2183
if (event_size <= mmap_size - head % page_size) {
tools/perf/util/session.c
2190
" fuzzed or compressed perf.data?\n", __func__, head, event_size, mmap_size);
tools/perf/util/session.c
2196
fetch_mmaped_event(u64 head, size_t mmap_size, char *buf, bool needs_swap)
tools/perf/util/session.c
2198
return prefetch_event(buf, head, mmap_size, needs_swap, ERR_PTR(-EINVAL));
tools/perf/util/session.c
2202
fetch_decomp_event(u64 head, size_t mmap_size, char *buf, bool needs_swap)
tools/perf/util/session.c
2204
return prefetch_event(buf, head, mmap_size, needs_swap, NULL);
tools/perf/util/session.c
2216
while (decomp->head < decomp->size && !session_done()) {
tools/perf/util/session.c
2217
union perf_event *event = fetch_decomp_event(decomp->head, decomp->size, decomp->data,
tools/perf/util/session.c
2229
decomp->file_pos + decomp->head, event->header.size, event->header.type);
tools/perf/util/session.c
2236
decomp->head += size;
tools/perf/util/session.c
2274
u64 head;
tools/perf/util/session.c
2287
rd->head = rd->data_offset;
tools/perf/util/session.c
2335
page_offset = page_size * (rd->head / page_size);
tools/perf/util/session.c
2337
rd->head -= page_offset;
tools/perf/util/session.c
2347
rd->file_pos = rd->file_offset + rd->head;
tools/perf/util/session.c
2370
event = fetch_mmaped_event(rd->head, rd->mmap_size, rd->mmap_cur,
tools/perf/util/session.c
2386
rd->file_offset + rd->head, event->header.size,
tools/perf/util/session.c
2396
rd->head += size;
tools/perf/util/session.h
101
u64 head;
tools/perf/util/stat-shadow.c
290
mexp = list_first_entry(&me->head, typeof(*mexp), nd);
tools/perf/util/stat-shadow.c
292
list_for_each_entry_from(mexp, &me->head, nd) {
tools/perf/util/symbol-elf.c
2872
int get_sdt_note_list(struct list_head *head, const char *target)
tools/perf/util/symbol-elf.c
2886
ret = construct_sdt_notes_list(elf, head);
tools/perf/util/symbol.h
254
int get_sdt_note_list(struct list_head *head, const char *target);
tools/perf/util/tool.c
28
decomp_last_rem = decomp_last->size - decomp_last->head;
tools/perf/util/tool.c
43
decomp->head = 0;
tools/perf/util/tool.c
46
memcpy(decomp->data, &(decomp_last->data[decomp_last->head]), decomp_last_rem);
tools/power/acpi/tools/acpidbg/acpidbg.c
128
p = &crc->buf[crc->head];
tools/power/acpi/tools/acpidbg/acpidbg.c
134
crc->head = (crc->head + len) & (ACPI_AML_BUF_SIZE - 1);
tools/power/acpi/tools/acpidbg/acpidbg.c
144
p = &crc->buf[crc->head];
tools/power/acpi/tools/acpidbg/acpidbg.c
155
crc->head = (crc->head + len) & (ACPI_AML_BUF_SIZE - 1);
tools/power/acpi/tools/acpidbg/acpidbg.c
165
p = &crc->buf[crc->head];
tools/power/acpi/tools/acpidbg/acpidbg.c
171
crc->head = (crc->head + 1) & (ACPI_AML_BUF_SIZE - 1);
tools/power/acpi/tools/acpidbg/acpidbg.c
187
crc->head = (crc->head + 1) & (ACPI_AML_BUF_SIZE - 1);
tools/power/acpi/tools/acpidbg/acpidbg.c
198
crc->head = (crc->head + 1) & (ACPI_AML_BUF_SIZE - 1);
tools/power/acpi/tools/acpidbg/acpidbg.c
211
crc->head = (crc->head + 1) & (ACPI_AML_BUF_SIZE - 1);
tools/power/acpi/tools/acpidbg/acpidbg.c
42
(CIRC_CNT((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE))
tools/power/acpi/tools/acpidbg/acpidbg.c
44
(CIRC_CNT_TO_END((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE))
tools/power/acpi/tools/acpidbg/acpidbg.c
46
(CIRC_SPACE((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE))
tools/power/acpi/tools/acpidbg/acpidbg.c
48
(CIRC_SPACE_TO_END((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE))
tools/power/acpi/tools/acpidbg/acpidbg.c
77
.head = 0,
tools/power/acpi/tools/acpidbg/acpidbg.c
82
.head = 0,
tools/power/x86/turbostat/turbostat.c
10082
struct pmt_mmio *head = NULL, *last = NULL;
tools/power/x86/turbostat/turbostat.c
10160
head = new_pmt;
tools/power/x86/turbostat/turbostat.c
10176
if (head)
tools/power/x86/turbostat/turbostat.c
10177
pmt_mmios = head;
tools/power/x86/turbostat/turbostat.c
10179
return head;
tools/power/x86/turbostat/turbostat.c
10561
struct msr_counter *find_msrp_by_name(struct msr_counter *head, char *name)
tools/power/x86/turbostat/turbostat.c
10565
for (mp = head; mp; mp = mp->next) {
tools/sched_ext/include/scx/common.bpf.h
308
int bpf_list_push_front_impl(struct bpf_list_head *head,
tools/sched_ext/include/scx/common.bpf.h
311
#define bpf_list_push_front(head, node) bpf_list_push_front_impl(head, node, NULL, 0)
tools/sched_ext/include/scx/common.bpf.h
313
int bpf_list_push_back_impl(struct bpf_list_head *head,
tools/sched_ext/include/scx/common.bpf.h
316
#define bpf_list_push_back(head, node) bpf_list_push_back_impl(head, node, NULL, 0)
tools/sched_ext/include/scx/common.bpf.h
318
struct bpf_list_node *bpf_list_pop_front(struct bpf_list_head *head) __ksym;
tools/sched_ext/include/scx/common.bpf.h
319
struct bpf_list_node *bpf_list_pop_back(struct bpf_list_head *head) __ksym;
tools/sched_ext/include/scx/common.bpf.h
325
#define bpf_rbtree_add(head, node, less) bpf_rbtree_add_impl(head, node, less, NULL, 0)
tools/testing/radix-tree/test.c
72
static void item_free_rcu(struct rcu_head *head)
tools/testing/radix-tree/test.c
74
struct item *item = container_of(head, struct item, rcu_head);
tools/testing/selftests/arm64/signal/test_signals_utils.c
182
struct _aarch64_ctx *head;
tools/testing/selftests/arm64/signal/test_signals_utils.c
195
head = (struct _aarch64_ctx *)uc->uc_mcontext.__reserved;
tools/testing/selftests/arm64/signal/test_signals_utils.c
196
head = get_header(head, EXTRA_MAGIC, td->live_sz, &offset);
tools/testing/selftests/arm64/signal/test_signals_utils.c
197
if (head) {
tools/testing/selftests/arm64/signal/test_signals_utils.c
198
extra = (struct extra_context *)head;
tools/testing/selftests/arm64/signal/test_signals_utils.c
233
copied_extra->head.size = sizeof(*copied_extra) + 16;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_magic.c
21
struct _aarch64_ctx *shead = GET_SF_RESV_HEAD(sf), *head;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_magic.c
28
head = get_starting_head(shead, HDR_SZ * 2, GET_SF_RESV_SIZE(sf), NULL);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_magic.c
29
if (!head)
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_magic.c
36
head->magic = KSFT_BAD_MAGIC;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_magic.c
37
head->size = HDR_SZ;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_magic.c
38
write_terminator_record(GET_RESV_NEXT_HEAD(head));
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size.c
24
struct _aarch64_ctx *shead = GET_SF_RESV_HEAD(sf), *head;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size.c
33
head = get_starting_head(shead, need_sz, resv_sz, &offset);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size.c
34
if (!head)
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size.c
45
head->magic = ESR_MAGIC;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size.c
46
head->size = sizeof(struct esr_context);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size.c
48
write_terminator_record(GET_RESV_NEXT_HEAD(head));
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size.c
60
head->size = (resv_sz - offset - need_sz + MIN_SZ_ALIGN) & ~0xfUL;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size.c
62
head->size += MIN_SZ_ALIGN;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size.c
64
write_terminator_record(GET_RESV_NEXT_HEAD(head));
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size_for_magic0.c
21
struct _aarch64_ctx *shead = GET_SF_RESV_HEAD(sf), *head;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size_for_magic0.c
28
head = get_starting_head(shead, HDR_SZ, GET_SF_RESV_SIZE(sf), NULL);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size_for_magic0.c
29
if (!head)
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size_for_magic0.c
32
head->magic = 0;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_bad_size_for_magic0.c
33
head->size = HDR_SZ;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_duplicated_fpsimd.c
21
struct _aarch64_ctx *shead = GET_SF_RESV_HEAD(sf), *head;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_duplicated_fpsimd.c
27
head = get_starting_head(shead, sizeof(struct fpsimd_context) + HDR_SZ,
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_duplicated_fpsimd.c
29
if (!head)
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_duplicated_fpsimd.c
33
head->magic = FPSIMD_MAGIC;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_duplicated_fpsimd.c
34
head->size = sizeof(struct fpsimd_context);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_duplicated_fpsimd.c
36
write_terminator_record(GET_RESV_NEXT_HEAD(head));
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_missing_fpsimd.c
23
struct _aarch64_ctx *head = GET_SF_RESV_HEAD(sf);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_missing_fpsimd.c
30
head = get_header(head, FPSIMD_MAGIC, resv_sz, &offset);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_missing_fpsimd.c
31
if (head && resv_sz - offset >= HDR_SZ) {
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_missing_fpsimd.c
35
write_terminator_record(head);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_sme_change_vl.c
37
struct _aarch64_ctx *head = GET_SF_RESV_HEAD(sf);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_sme_change_vl.c
45
head = get_header(head, ZA_MAGIC, resv_sz, &offset);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_sme_change_vl.c
46
if (!head) {
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_sme_change_vl.c
51
if (head->size != sizeof(struct za_context)) {
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_sme_change_vl.c
56
za = (struct za_context *)head;
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_sve_change_vl.c
37
struct _aarch64_ctx *head = GET_SF_RESV_HEAD(sf);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_sve_change_vl.c
45
head = get_header(head, SVE_MAGIC, resv_sz, &offset);
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_sve_change_vl.c
46
if (!head) {
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_sve_change_vl.c
51
if (head->size != sizeof(struct sve_context)) {
tools/testing/selftests/arm64/signal/testcases/fake_sigreturn_sve_change_vl.c
56
sve = (struct sve_context *)head;
tools/testing/selftests/arm64/signal/testcases/fpmr_siginfo.c
41
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/fpmr_siginfo.c
56
get_header(head, FPMR_MAGIC, td->live_sz, &offset);
tools/testing/selftests/arm64/signal/testcases/gcs_frame.c
21
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/gcs_frame.c
53
head = get_header(head, GCS_MAGIC, GET_BUF_RESV_SIZE(context),
tools/testing/selftests/arm64/signal/testcases/gcs_frame.c
55
if (!head) {
tools/testing/selftests/arm64/signal/testcases/gcs_frame.c
60
gcs = (struct gcs_context *)head;
tools/testing/selftests/arm64/signal/testcases/poe_siginfo.c
41
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/poe_siginfo.c
56
get_header(head, POE_MAGIC, td->live_sz, &offset);
tools/testing/selftests/arm64/signal/testcases/sme_vl.c
33
struct _aarch64_ctx *head = GET_SF_RESV_HEAD(sf);
tools/testing/selftests/arm64/signal/testcases/sme_vl.c
41
head = get_header(head, ZA_MAGIC, resv_sz, &offset);
tools/testing/selftests/arm64/signal/testcases/sme_vl.c
42
if (!head) {
tools/testing/selftests/arm64/signal/testcases/sme_vl.c
46
za = (struct za_context *)head;
tools/testing/selftests/arm64/signal/testcases/ssve_regs.c
46
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/ssve_regs.c
66
head = get_header(head, SVE_MAGIC, GET_BUF_RESV_SIZE(context),
tools/testing/selftests/arm64/signal/testcases/ssve_regs.c
68
if (!head) {
tools/testing/selftests/arm64/signal/testcases/ssve_regs.c
73
ssve = (struct sve_context *)head;
tools/testing/selftests/arm64/signal/testcases/ssve_regs.c
86
head->size, ssve->vl);
tools/testing/selftests/arm64/signal/testcases/ssve_za_regs.c
51
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/ssve_za_regs.c
73
regs = get_header(head, SVE_MAGIC, GET_BUF_RESV_SIZE(context),
tools/testing/selftests/arm64/signal/testcases/ssve_za_regs.c
94
regs = get_header(head, ZA_MAGIC, GET_BUF_RESV_SIZE(context),
tools/testing/selftests/arm64/signal/testcases/sve_regs.c
46
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/sve_regs.c
64
head = get_header(head, SVE_MAGIC, GET_BUF_RESV_SIZE(context),
tools/testing/selftests/arm64/signal/testcases/sve_regs.c
66
if (!head) {
tools/testing/selftests/arm64/signal/testcases/sve_regs.c
71
sve = (struct sve_context *)head;
tools/testing/selftests/arm64/signal/testcases/sve_regs.c
79
head->size, sve->vl);
tools/testing/selftests/arm64/signal/testcases/sve_vl.c
33
struct _aarch64_ctx *head = GET_SF_RESV_HEAD(sf);
tools/testing/selftests/arm64/signal/testcases/sve_vl.c
41
head = get_header(head, SVE_MAGIC, resv_sz, &offset);
tools/testing/selftests/arm64/signal/testcases/sve_vl.c
42
if (!head) {
tools/testing/selftests/arm64/signal/testcases/sve_vl.c
46
sve = (struct sve_context *)head;
tools/testing/selftests/arm64/signal/testcases/testcases.c
100
if (zt->head.size != ZT_SIG_CONTEXT_SIZE(zt->nregs)) {
tools/testing/selftests/arm64/signal/testcases/testcases.c
118
struct _aarch64_ctx *head =
tools/testing/selftests/arm64/signal/testcases/testcases.c
127
while (head && !terminated && offs < resv_sz) {
tools/testing/selftests/arm64/signal/testcases/testcases.c
128
if ((uint64_t)head & 0x0fUL) {
tools/testing/selftests/arm64/signal/testcases/testcases.c
135
switch (head->magic) {
tools/testing/selftests/arm64/signal/testcases/testcases.c
137
if (head->size) {
tools/testing/selftests/arm64/signal/testcases/testcases.c
141
head = extra_data;
tools/testing/selftests/arm64/signal/testcases/testcases.c
155
else if (head->size !=
tools/testing/selftests/arm64/signal/testcases/testcases.c
161
if (head->size != sizeof(struct esr_context))
tools/testing/selftests/arm64/signal/testcases/testcases.c
165
if (head->size != sizeof(struct poe_context))
tools/testing/selftests/arm64/signal/testcases/testcases.c
169
if (head->size != sizeof(struct tpidr2_context))
tools/testing/selftests/arm64/signal/testcases/testcases.c
176
sve = (struct sve_context *)head;
tools/testing/selftests/arm64/signal/testcases/testcases.c
18
term = GET_RESV_NEXT_HEAD(&extra->head);
tools/testing/selftests/arm64/signal/testcases/testcases.c
183
za = (struct za_context *)head;
tools/testing/selftests/arm64/signal/testcases/testcases.c
190
zt = (struct zt_context *)head;
tools/testing/selftests/arm64/signal/testcases/testcases.c
196
else if (head->size !=
tools/testing/selftests/arm64/signal/testcases/testcases.c
204
if (head->size != sizeof(struct gcs_context))
tools/testing/selftests/arm64/signal/testcases/testcases.c
211
else if (head->size !=
tools/testing/selftests/arm64/signal/testcases/testcases.c
215
extra = (struct extra_context *)head;
tools/testing/selftests/arm64/signal/testcases/testcases.c
234
memcpy(magic, &head->magic, sizeof(magic));
tools/testing/selftests/arm64/signal/testcases/testcases.c
241
head->magic,
tools/testing/selftests/arm64/signal/testcases/testcases.c
249
offs += head->size;
tools/testing/selftests/arm64/signal/testcases/testcases.c
250
if (resv_sz < offs + sizeof(*head)) {
tools/testing/selftests/arm64/signal/testcases/testcases.c
271
head = GET_RESV_NEXT_HEAD(head);
tools/testing/selftests/arm64/signal/testcases/testcases.c
310
struct _aarch64_ctx *head;
tools/testing/selftests/arm64/signal/testcases/testcases.c
312
head = get_terminator(shead, resv_sz, &offs);
tools/testing/selftests/arm64/signal/testcases/testcases.c
314
if (!head)
tools/testing/selftests/arm64/signal/testcases/testcases.c
315
return head;
tools/testing/selftests/arm64/signal/testcases/testcases.c
319
head = get_header(shead, EXTRA_MAGIC, resv_sz, &offs);
tools/testing/selftests/arm64/signal/testcases/testcases.c
320
if (!head || resv_sz - offs < need_sz) {
tools/testing/selftests/arm64/signal/testcases/testcases.c
330
return head;
tools/testing/selftests/arm64/signal/testcases/testcases.c
48
if ((sve->head.size != sizeof(struct sve_context)) &&
tools/testing/selftests/arm64/signal/testcases/testcases.c
49
(sve->head.size != regs_size)) {
tools/testing/selftests/arm64/signal/testcases/testcases.c
73
if ((za->head.size != sizeof(struct za_context)) &&
tools/testing/selftests/arm64/signal/testcases/testcases.c
74
(za->head.size != regs_size)) {
tools/testing/selftests/arm64/signal/testcases/testcases.h
101
if (!head || resv_sz < HDR_SZ)
tools/testing/selftests/arm64/signal/testcases/testcases.h
105
head->magic != magic && head->magic) {
tools/testing/selftests/arm64/signal/testcases/testcases.h
106
offs += head->size;
tools/testing/selftests/arm64/signal/testcases/testcases.h
107
head = GET_RESV_NEXT_HEAD(head);
tools/testing/selftests/arm64/signal/testcases/testcases.h
109
if (head->magic == magic) {
tools/testing/selftests/arm64/signal/testcases/testcases.h
110
found = head;
tools/testing/selftests/arm64/signal/testcases/testcases.h
119
static inline struct _aarch64_ctx *get_terminator(struct _aarch64_ctx *head,
tools/testing/selftests/arm64/signal/testcases/testcases.h
123
return get_header(head, 0, resv_sz, offset);
tools/testing/selftests/arm64/signal/testcases/testcases.h
95
static inline struct _aarch64_ctx *get_header(struct _aarch64_ctx *head, uint32_t magic,
tools/testing/selftests/arm64/signal/testcases/tpidr2_siginfo.c
41
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/tpidr2_siginfo.c
56
get_header(head, TPIDR2_MAGIC, td->live_sz, &offset);
tools/testing/selftests/arm64/signal/testcases/za_no_regs.c
40
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/za_no_regs.c
57
head = get_header(head, ZA_MAGIC, GET_BUF_RESV_SIZE(context), &offset);
tools/testing/selftests/arm64/signal/testcases/za_no_regs.c
58
if (!head) {
tools/testing/selftests/arm64/signal/testcases/za_no_regs.c
63
za = (struct za_context *)head;
tools/testing/selftests/arm64/signal/testcases/za_no_regs.c
69
if (head->size != ZA_SIG_REGS_OFFSET) {
tools/testing/selftests/arm64/signal/testcases/za_no_regs.c
71
head->size, ZA_SIG_REGS_OFFSET);
tools/testing/selftests/arm64/signal/testcases/za_no_regs.c
77
head->size, za->vl);
tools/testing/selftests/arm64/signal/testcases/za_regs.c
48
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/za_regs.c
66
head = get_header(head, ZA_MAGIC, GET_BUF_RESV_SIZE(context), &offset);
tools/testing/selftests/arm64/signal/testcases/za_regs.c
67
if (!head) {
tools/testing/selftests/arm64/signal/testcases/za_regs.c
72
za = (struct za_context *)head;
tools/testing/selftests/arm64/signal/testcases/za_regs.c
78
if (head->size != ZA_SIG_CONTEXT_SIZE(sve_vq_from_vl(vl))) {
tools/testing/selftests/arm64/signal/testcases/za_regs.c
80
head->size, ZA_SIG_CONTEXT_SIZE(sve_vq_from_vl(vl)));
tools/testing/selftests/arm64/signal/testcases/za_regs.c
85
head->size, za->vl);
tools/testing/selftests/arm64/signal/testcases/zt_no_regs.c
24
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/zt_no_regs.c
33
head = get_header(head, ZT_MAGIC, GET_BUF_RESV_SIZE(context), &offset);
tools/testing/selftests/arm64/signal/testcases/zt_no_regs.c
34
if (head) {
tools/testing/selftests/arm64/signal/testcases/zt_regs.c
30
struct _aarch64_ctx *head = GET_BUF_RESV_HEAD(context);
tools/testing/selftests/arm64/signal/testcases/zt_regs.c
42
head = get_header(head, ZT_MAGIC, GET_BUF_RESV_SIZE(context), &offset);
tools/testing/selftests/arm64/signal/testcases/zt_regs.c
43
if (!head) {
tools/testing/selftests/arm64/signal/testcases/zt_regs.c
48
zt = (struct zt_context *)head;
tools/testing/selftests/arm64/signal/testcases/zt_regs.c
55
head->size, zt->nregs);
tools/testing/selftests/bpf/bpf_arena_htab.h
29
return &__select_bucket(htab, hash)->head;
tools/testing/selftests/bpf/bpf_arena_htab.h
40
static hashtab_elem_t *lookup_elem_raw(arena_list_head_t *head, __u32 hash, int key)
tools/testing/selftests/bpf/bpf_arena_htab.h
44
list_for_each_entry(l, head, hash_node)
tools/testing/selftests/bpf/bpf_arena_htab.h
59
arena_list_head_t *head;
tools/testing/selftests/bpf/bpf_arena_htab.h
62
head = select_bucket(htab, key);
tools/testing/selftests/bpf/bpf_arena_htab.h
63
l_old = lookup_elem_raw(head, htab_hash(key), key);
tools/testing/selftests/bpf/bpf_arena_htab.h
72
arena_list_head_t *head;
tools/testing/selftests/bpf/bpf_arena_htab.h
75
head = select_bucket(htab, key);
tools/testing/selftests/bpf/bpf_arena_htab.h
76
l_old = lookup_elem_raw(head, htab_hash(key), key);
tools/testing/selftests/bpf/bpf_arena_htab.h
85
list_add_head(&l_new->hash_node, head);
tools/testing/selftests/bpf/bpf_arena_htab.h
9
struct arena_list_head head;
tools/testing/selftests/bpf/bpf_arena_list.h
36
#define list_for_each_entry(pos, head, member) \
tools/testing/selftests/bpf/bpf_arena_list.h
37
for (void * ___tmp = (pos = list_entry_safe((head)->first, \
tools/testing/selftests/bpf/bpf_experimental.h
100
extern struct bpf_list_node *bpf_list_pop_back(struct bpf_list_head *head) __ksym;
tools/testing/selftests/bpf/bpf_experimental.h
124
#define bpf_rbtree_add(head, node, less) bpf_rbtree_add_impl(head, node, less, NULL, 0)
tools/testing/selftests/bpf/bpf_experimental.h
65
extern int bpf_list_push_front_impl(struct bpf_list_head *head,
tools/testing/selftests/bpf/bpf_experimental.h
70
#define bpf_list_push_front(head, node) bpf_list_push_front_impl(head, node, NULL, 0)
tools/testing/selftests/bpf/bpf_experimental.h
81
extern int bpf_list_push_back_impl(struct bpf_list_head *head,
tools/testing/selftests/bpf/bpf_experimental.h
86
#define bpf_list_push_back(head, node) bpf_list_push_back_impl(head, node, NULL, 0)
tools/testing/selftests/bpf/bpf_experimental.h
93
extern struct bpf_list_node *bpf_list_pop_front(struct bpf_list_head *head) __ksym;
tools/testing/selftests/bpf/prog_tests/arena_list.c
20
static int list_sum(struct arena_list_head *head)
tools/testing/selftests/bpf/prog_tests/arena_list.c
25
list_for_each_entry(n, head, node)
tools/testing/selftests/bpf/prog_tests/core_read_macros.c
18
void (*func)(struct callback_head *head);
tools/testing/selftests/bpf/prog_tests/core_read_macros.c
8
void (*func)(struct callback_head *head);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
167
fq_flows_add_head(struct bpf_list_head *head, struct bpf_spin_lock *lock,
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
171
bpf_list_push_front(head, &flow->list_node);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
177
fq_flows_add_tail(struct bpf_list_head *head, struct bpf_spin_lock *lock,
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
181
bpf_list_push_back(head, &flow->list_node);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
187
fq_flows_remove_front(struct bpf_list_head *head, struct bpf_spin_lock *lock,
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
191
*node = bpf_list_pop_front(head);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
197
fq_flows_is_empty(struct bpf_list_head *head, struct bpf_spin_lock *lock)
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
202
node = bpf_list_pop_front(head);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
204
bpf_list_push_front(head, node);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
464
struct bpf_list_head *head;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
473
head = &fq_new_flows;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
477
head = &fq_old_flows;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
486
fq_flows_remove_front(head, lock, &node, cnt);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
502
if (head == &fq_new_flows && !is_empty) {
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
540
fq_flows_add_head(head, lock, flow, cnt);
tools/testing/selftests/bpf/progs/decap_sanity.c
23
return skb->data - skb->head;
tools/testing/selftests/bpf/progs/kfree_skb.c
21
void (*func)(struct callback_head *head);
tools/testing/selftests/bpf/progs/linked_list.c
109
int list_push_pop_multiple(struct bpf_spin_lock *lock, struct bpf_list_head *head, bool leave_in_map)
tools/testing/selftests/bpf/progs/linked_list.c
13
struct bpf_list_head head __contains(foo, node2);
tools/testing/selftests/bpf/progs/linked_list.c
135
bpf_list_push_front(head, &f[i]->node2);
tools/testing/selftests/bpf/progs/linked_list.c
136
bpf_list_push_front(head, &f[i + 1]->node2);
tools/testing/selftests/bpf/progs/linked_list.c
142
n = bpf_list_pop_front(head);
tools/testing/selftests/bpf/progs/linked_list.c
152
bpf_list_push_back(head, &pf->node2);
tools/testing/selftests/bpf/progs/linked_list.c
161
n = bpf_list_pop_back(head);
tools/testing/selftests/bpf/progs/linked_list.c
173
n = bpf_list_pop_back(head);
tools/testing/selftests/bpf/progs/linked_list.c
181
n = bpf_list_pop_front(head);
tools/testing/selftests/bpf/progs/linked_list.c
191
int list_in_list(struct bpf_spin_lock *lock, struct bpf_list_head *head, bool leave_in_map)
tools/testing/selftests/bpf/progs/linked_list.c
209
bpf_list_push_back(&f->head, &b->node);
tools/testing/selftests/bpf/progs/linked_list.c
215
bpf_list_push_front(head, &f->node2);
tools/testing/selftests/bpf/progs/linked_list.c
222
n = bpf_list_pop_front(head);
tools/testing/selftests/bpf/progs/linked_list.c
234
n = bpf_list_pop_front(&f->head);
tools/testing/selftests/bpf/progs/linked_list.c
249
n = bpf_list_pop_front(&f->head);
tools/testing/selftests/bpf/progs/linked_list.c
261
int test_list_push_pop(struct bpf_spin_lock *lock, struct bpf_list_head *head)
tools/testing/selftests/bpf/progs/linked_list.c
265
ret = list_push_pop(lock, head, false);
tools/testing/selftests/bpf/progs/linked_list.c
268
return list_push_pop(lock, head, true);
tools/testing/selftests/bpf/progs/linked_list.c
272
int test_list_push_pop_multiple(struct bpf_spin_lock *lock, struct bpf_list_head *head)
tools/testing/selftests/bpf/progs/linked_list.c
276
ret = list_push_pop_multiple(lock, head, false);
tools/testing/selftests/bpf/progs/linked_list.c
279
return list_push_pop_multiple(lock, head, true);
tools/testing/selftests/bpf/progs/linked_list.c
28
int list_push_pop(struct bpf_spin_lock *lock, struct bpf_list_head *head, bool leave_in_map)
tools/testing/selftests/bpf/progs/linked_list.c
283
int test_list_in_list(struct bpf_spin_lock *lock, struct bpf_list_head *head)
tools/testing/selftests/bpf/progs/linked_list.c
287
ret = list_in_list(lock, head, false);
tools/testing/selftests/bpf/progs/linked_list.c
290
return list_in_list(lock, head, true);
tools/testing/selftests/bpf/progs/linked_list.c
301
return test_list_push_pop(&v->lock, &v->head);
tools/testing/selftests/bpf/progs/linked_list.c
316
return test_list_push_pop(&v->lock, &v->head);
tools/testing/selftests/bpf/progs/linked_list.c
328
return test_list_push_pop(&ghead_nested.inner.lock, &ghead_nested.inner.head);
tools/testing/selftests/bpf/progs/linked_list.c
359
return test_list_push_pop_multiple(&v->lock, &v->head);
tools/testing/selftests/bpf/progs/linked_list.c
374
return test_list_push_pop_multiple(&v->lock, &v->head);
tools/testing/selftests/bpf/progs/linked_list.c
38
n = bpf_list_pop_front(head);
tools/testing/selftests/bpf/progs/linked_list.c
396
return test_list_in_list(&v->lock, &v->head);
tools/testing/selftests/bpf/progs/linked_list.c
411
return test_list_in_list(&v->lock, &v->head);
tools/testing/selftests/bpf/progs/linked_list.c
47
n = bpf_list_pop_back(head);
tools/testing/selftests/bpf/progs/linked_list.c
58
bpf_list_push_front(head, &f->node2);
tools/testing/selftests/bpf/progs/linked_list.c
63
n = bpf_list_pop_back(head);
tools/testing/selftests/bpf/progs/linked_list.c
75
bpf_list_push_front(head, &f->node2);
tools/testing/selftests/bpf/progs/linked_list.c
78
n = bpf_list_pop_front(head);
tools/testing/selftests/bpf/progs/linked_list.c
90
n = bpf_list_pop_front(head);
tools/testing/selftests/bpf/progs/linked_list.c
98
n = bpf_list_pop_back(head);
tools/testing/selftests/bpf/progs/linked_list.h
16
struct bpf_list_head head __contains(bar, node);
tools/testing/selftests/bpf/progs/linked_list.h
25
struct bpf_list_head head __contains(foo, node2);
tools/testing/selftests/bpf/progs/linked_list_fail.c
106
CHECK(kptr_kptr, op, &f1->lock, &f2->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
108
CHECK(kptr_map, op, &f1->lock, &v->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
109
CHECK(kptr_inner_map, op, &f1->lock, &iv->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
112
CHECK(global_kptr, op, &glock, &f1->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
113
CHECK(global_map, op, &glock, &v->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
114
CHECK(global_inner_map, op, &glock, &iv->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
116
CHECK(map_map, op, &v->lock, &v2->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
117
CHECK(map_kptr, op, &v->lock, &f2->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
119
CHECK(map_inner_map, op, &v->lock, &iv->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
121
CHECK(inner_map_inner_map, op, &iv->lock, &iv2->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
122
CHECK(inner_map_kptr, op, &iv->lock, &f2->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
124
CHECK(inner_map_map, op, &iv->lock, &v->head);
tools/testing/selftests/bpf/progs/linked_list_fail.c
143
CHECK(kptr_kptr, op, &f1->lock, &f2->head, &b->node); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
145
CHECK(kptr_map, op, &f1->lock, &v->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
146
CHECK(kptr_inner_map, op, &f1->lock, &iv->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
149
CHECK(global_kptr, op, &glock, &f1->head, &b->node); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
150
CHECK(global_map, op, &glock, &v->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
151
CHECK(global_inner_map, op, &glock, &iv->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
153
CHECK(map_map, op, &v->lock, &v2->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
154
CHECK(map_kptr, op, &v->lock, &f2->head, &b->node); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
156
CHECK(map_inner_map, op, &v->lock, &iv->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
158
CHECK(inner_map_inner_map, op, &iv->lock, &iv2->head, &f->node2);\
tools/testing/selftests/bpf/progs/linked_list_fail.c
159
CHECK(inner_map_kptr, op, &iv->lock, &f2->head, &b->node); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
161
CHECK(inner_map_map, op, &iv->lock, &v->head, &f->node2);
tools/testing/selftests/bpf/progs/linked_list_fail.c
317
return *(int *)&f->head;
tools/testing/selftests/bpf/progs/linked_list_fail.c
328
*(int *)&f->head = 0;
tools/testing/selftests/bpf/progs/linked_list_fail.c
533
bpf_list_push_front((void *)&f->head + ctx->protocol, &f->node2);
tools/testing/selftests/bpf/progs/linked_list_fail.c
555
bpf_list_push_front((void *)&f->head + 1, &b->node);
tools/testing/selftests/bpf/progs/linked_list_fail.c
57
CHECK(kptr, pop_front, &f->head);
tools/testing/selftests/bpf/progs/linked_list_fail.c
578
int pop_ptr_off(void *(*op)(void *head))
tools/testing/selftests/bpf/progs/linked_list_fail.c
58
CHECK(kptr, pop_back, &f->head);
tools/testing/selftests/bpf/progs/linked_list_fail.c
581
struct bpf_list_head head __contains(foo, node2);
tools/testing/selftests/bpf/progs/linked_list_fail.c
590
n = op(&p->head);
tools/testing/selftests/bpf/progs/linked_list_fail.c
63
CHECK(map, pop_front, &v->head);
tools/testing/selftests/bpf/progs/linked_list_fail.c
64
CHECK(map, pop_back, &v->head);
tools/testing/selftests/bpf/progs/linked_list_fail.c
66
CHECK(inner_map, pop_front, &iv->head);
tools/testing/selftests/bpf/progs/linked_list_fail.c
67
CHECK(inner_map, pop_back, &iv->head);
tools/testing/selftests/bpf/progs/linked_list_fail.c
80
CHECK(kptr, push_front, &f->head, &b->node);
tools/testing/selftests/bpf/progs/linked_list_fail.c
81
CHECK(kptr, push_back, &f->head, &b->node);
tools/testing/selftests/bpf/progs/linked_list_fail.c
86
CHECK(map, push_front, &v->head, &f->node2);
tools/testing/selftests/bpf/progs/linked_list_fail.c
87
CHECK(map, push_back, &v->head, &f->node2);
tools/testing/selftests/bpf/progs/linked_list_fail.c
89
CHECK(inner_map, push_front, &iv->head, &f->node2);
tools/testing/selftests/bpf/progs/linked_list_fail.c
90
CHECK(inner_map, push_back, &iv->head, &f->node2);
tools/testing/selftests/bpf/progs/map_in_map_btf.c
16
struct bpf_list_head head __contains(node_data, node);
tools/testing/selftests/bpf/progs/map_in_map_btf.c
68
bpf_list_push_back(&value->head, &new->node);
tools/testing/selftests/bpf/progs/mptcp_bpf.h
11
return list == head;
tools/testing/selftests/bpf/progs/mptcp_bpf.h
23
#define list_entry_is_head(pos, head, member) \
tools/testing/selftests/bpf/progs/mptcp_bpf.h
24
list_is_head(&pos->member, (head))
tools/testing/selftests/bpf/progs/mptcp_bpf.h
27
#define list_for_each_entry(pos, head, member) \
tools/testing/selftests/bpf/progs/mptcp_bpf.h
28
for (pos = list_first_entry(head, typeof(*pos), member); \
tools/testing/selftests/bpf/progs/mptcp_bpf.h
29
!list_entry_is_head(pos, head, member) && can_loop; \
tools/testing/selftests/bpf/progs/mptcp_bpf.h
9
const struct list_head *head)
tools/testing/selftests/bpf/progs/net_timestamping.c
134
shinfo = bpf_core_cast(skb->head + skb->end, struct skb_shared_info);
tools/testing/selftests/bpf/progs/normal_map_btf.c
16
struct bpf_list_head head __contains(node_data, node);
tools/testing/selftests/bpf/progs/normal_map_btf.c
51
bpf_list_push_back(&value->head, &new->node);
tools/testing/selftests/bpf/progs/refcounted_kptr.c
173
static long __read_from_list(struct bpf_list_head *head,
tools/testing/selftests/bpf/progs/refcounted_kptr.c
183
l = bpf_list_pop_front(head);
tools/testing/selftests/bpf/progs/refcounted_kptr.c
193
if (bpf_list_push_back(head, &n->l)) {
tools/testing/selftests/bpf/progs/refcounted_kptr.c
233
err = __insert_in_tree_and_list(&head, &root, &lock); \
tools/testing/selftests/bpf/progs/refcounted_kptr.c
243
err = __read_from_list(&head, &lock, rem_list); \
tools/testing/selftests/bpf/progs/refcounted_kptr.c
272
err = __insert_in_tree_and_list(&head, &root, &lock); \
tools/testing/selftests/bpf/progs/refcounted_kptr.c
276
err = __read_from_list(&head, &lock, rem_list); \
tools/testing/selftests/bpf/progs/refcounted_kptr.c
310
err = __insert_in_tree_and_list(&head, &root, &lock); \
tools/testing/selftests/bpf/progs/refcounted_kptr.c
333
INSERT_DOUBLE_READ_AND_DEL(__read_from_list, head, "insert_double_del: 2x read-and-del from list");
tools/testing/selftests/bpf/progs/refcounted_kptr.c
43
private(A) struct bpf_list_head head __contains(node_data, l);
tools/testing/selftests/bpf/progs/refcounted_kptr.c
492
bpf_list_push_back(&head, &m->l);
tools/testing/selftests/bpf/progs/refcounted_kptr.c
581
static int __insert_in_list(struct bpf_list_head *head, struct bpf_spin_lock *lock,
tools/testing/selftests/bpf/progs/refcounted_kptr.c
599
bpf_list_push_front(head, &node_ref->l);
tools/testing/selftests/bpf/progs/refcounted_kptr.c
622
return __insert_in_list(&head, &lock, &v->node);
tools/testing/selftests/bpf/progs/refcounted_kptr.c
73
static long __insert_in_tree_and_list(struct bpf_list_head *head,
tools/testing/selftests/bpf/progs/refcounted_kptr.c
97
if (bpf_list_push_front(head, &m->l)) {
tools/testing/selftests/bpf/progs/test_cls_redirect.c
129
net_ptr head;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
151
size_t off = (size_t)buf->head;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
162
buf->head += len;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
173
buf->head += len;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
186
if (buf->head + len > buf->tail) {
tools/testing/selftests/bpf/progs/test_cls_redirect.c
194
void *ptr = buf->head;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
195
buf->head += len;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
505
.head = (uint8_t *)(long)skb->data,
tools/testing/selftests/bpf/progs/test_cls_redirect.c
960
.head = (uint8_t *)(long)skb->data,
tools/testing/selftests/bpf/progs/test_core_read_macros.c
12
void (*func)(struct callback_head___shuffled *head);
tools/testing/selftests/bpf/progs/type_cast.c
49
shared_info = bpf_core_cast(kskb->head + kskb->end, struct skb_shared_info);
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
289
static void testmod_free_cb(struct rcu_head *head)
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
293
ctx = container_of(head, struct bpf_testmod_ctx, rcu);
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
212
rq_ring.khead = (unsigned int *)((char *)ring_ptr + reg.offsets.head);
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
317
unsigned int head;
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
322
io_uring_for_each_cqe(ring, head, cqe) {
tools/testing/selftests/filesystems/utils.c
101
static inline void list_add_tail(struct list *head, struct list *list)
tools/testing/selftests/filesystems/utils.c
103
__list_add(list, head->prev, head);
tools/testing/selftests/filesystems/utils.c
309
struct list head, uid_mapl, gid_mapl;
tools/testing/selftests/filesystems/utils.c
323
list_init(&head);
tools/testing/selftests/filesystems/utils.c
326
list_add_tail(&head, &uid_mapl);
tools/testing/selftests/filesystems/utils.c
327
list_add_tail(&head, &gid_mapl);
tools/testing/selftests/filesystems/utils.c
329
return get_userns_fd_from_idmap(&head);
tools/testing/selftests/futex/functional/robust_list.c
100
head->futex_offset = (size_t) offsetof(struct lock_struct, futex) -
tools/testing/selftests/futex/functional/robust_list.c
102
head->list.next = &head->list;
tools/testing/selftests/futex/functional/robust_list.c
103
head->list_op_pending = NULL;
tools/testing/selftests/futex/functional/robust_list.c
111
static int mutex_lock(struct lock_struct *lock, struct robust_list_head *head, bool error_inject)
tools/testing/selftests/futex/functional/robust_list.c
122
head->list_op_pending = &lock->list;
tools/testing/selftests/futex/functional/robust_list.c
128
struct robust_list *list = &head->list;
tools/testing/selftests/futex/functional/robust_list.c
134
while (list->next != &head->list)
tools/testing/selftests/futex/functional/robust_list.c
138
lock->list.next = &head->list;
tools/testing/selftests/futex/functional/robust_list.c
163
head->list_op_pending = NULL;
tools/testing/selftests/futex/functional/robust_list.c
174
struct robust_list_head head;
tools/testing/selftests/futex/functional/robust_list.c
177
ret = set_list(&head);
tools/testing/selftests/futex/functional/robust_list.c
183
ret = mutex_lock(lock, &head, false);
tools/testing/selftests/futex/functional/robust_list.c
211
struct robust_list_head head;
tools/testing/selftests/futex/functional/robust_list.c
214
ret = set_list(&head);
tools/testing/selftests/futex/functional/robust_list.c
228
ret = mutex_lock(&lock, &head, false);
tools/testing/selftests/futex/functional/robust_list.c
251
struct robust_list_head head;
tools/testing/selftests/futex/functional/robust_list.c
252
size_t head_size = sizeof(head);
tools/testing/selftests/futex/functional/robust_list.c
255
ret = set_robust_list(&head, head_size);
tools/testing/selftests/futex/functional/robust_list.c
258
ret = set_robust_list(&head, head_size * 2);
tools/testing/selftests/futex/functional/robust_list.c
262
ret = set_robust_list(&head, head_size - 1);
tools/testing/selftests/futex/functional/robust_list.c
266
ret = set_robust_list(&head, 0);
tools/testing/selftests/futex/functional/robust_list.c
278
struct robust_list_head head, head2, *get_head;
tools/testing/selftests/futex/functional/robust_list.c
279
size_t head_size = sizeof(head), len_ptr;
tools/testing/selftests/futex/functional/robust_list.c
282
ret = set_robust_list(&head, head_size);
tools/testing/selftests/futex/functional/robust_list.c
287
ASSERT_EQ(get_head, &head);
tools/testing/selftests/futex/functional/robust_list.c
303
struct robust_list_head *head = arg;
tools/testing/selftests/futex/functional/robust_list.c
306
ret = set_robust_list(head, sizeof(*head));
tools/testing/selftests/futex/functional/robust_list.c
331
struct robust_list_head head, *get_head;
tools/testing/selftests/futex/functional/robust_list.c
340
tid = create_child(&child_list, &head);
tools/testing/selftests/futex/functional/robust_list.c
347
ASSERT_EQ(&head, get_head);
tools/testing/selftests/futex/functional/robust_list.c
363
struct robust_list_head head;
tools/testing/selftests/futex/functional/robust_list.c
366
ret = set_list(&head);
tools/testing/selftests/futex/functional/robust_list.c
372
ret = mutex_lock(lock, &head, true);
tools/testing/selftests/futex/functional/robust_list.c
395
struct robust_list_head head;
tools/testing/selftests/futex/functional/robust_list.c
398
ret = set_list(&head);
tools/testing/selftests/futex/functional/robust_list.c
408
ret = mutex_lock(&lock, &head, false);
tools/testing/selftests/futex/functional/robust_list.c
429
struct robust_list_head head;
tools/testing/selftests/futex/functional/robust_list.c
432
set_list(&head);
tools/testing/selftests/futex/functional/robust_list.c
436
mutex_lock(&locks[i], &head, false);
tools/testing/selftests/futex/functional/robust_list.c
451
struct robust_list_head head;
tools/testing/selftests/futex/functional/robust_list.c
455
ret = mutex_lock(lock, &head, false);
tools/testing/selftests/futex/functional/robust_list.c
47
static int set_robust_list(struct robust_list_head *head, size_t len)
tools/testing/selftests/futex/functional/robust_list.c
49
return syscall(SYS_set_robust_list, head, len);
tools/testing/selftests/futex/functional/robust_list.c
512
static struct robust_list_head head;
tools/testing/selftests/futex/functional/robust_list.c
516
ret = set_list(&head);
tools/testing/selftests/futex/functional/robust_list.c
52
static int get_robust_list(int pid, struct robust_list_head **head, size_t *len_ptr)
tools/testing/selftests/futex/functional/robust_list.c
522
head.list.next = &a.list;
tools/testing/selftests/futex/functional/robust_list.c
54
return syscall(SYS_get_robust_list, pid, head, len_ptr);
tools/testing/selftests/futex/functional/robust_list.c
92
static int set_list(struct robust_list_head *head)
tools/testing/selftests/futex/functional/robust_list.c
96
ret = set_robust_list(head, sizeof(*head));
tools/testing/selftests/kselftest_harness.h
822
#define __LIST_APPEND(head, item) \
tools/testing/selftests/kselftest_harness.h
825
if (head == NULL) { \
tools/testing/selftests/kselftest_harness.h
826
head = item; \
tools/testing/selftests/kselftest_harness.h
833
item->prev = head->prev; \
tools/testing/selftests/kselftest_harness.h
835
head->prev = item; \
tools/testing/selftests/kselftest_harness.h
837
item->next = head; \
tools/testing/selftests/kselftest_harness.h
840
head = item; \
tools/testing/selftests/rseq/basic_percpu_ops_test.c
182
expect = (intptr_t)RSEQ_READ_ONCE(list->c[cpu].head);
tools/testing/selftests/rseq/basic_percpu_ops_test.c
184
targetptr = (intptr_t *)&list->c[cpu].head;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
205
struct percpu_list_node *head;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
211
targetptr = (intptr_t *)&list->c[cpu].head;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
214
load = (intptr_t *)&head;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
221
return head;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
237
node = list->c[cpu].head;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
240
list->c[cpu].head = node->next;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
297
node->next = list.c[i].head;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
298
list.c[i].head = node;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
75
struct percpu_list_node *head;
tools/testing/selftests/rseq/param_test.c
1302
list->c[i].head = node;
tools/testing/selftests/rseq/param_test.c
1311
free(list->c[i].head);
tools/testing/selftests/rseq/param_test.c
1345
if (expect_b != __atomic_load_n(&list_b.c[cpu_b].head->data, __ATOMIC_ACQUIRE)) {
tools/testing/selftests/rseq/param_test.c
1361
expect_a = __atomic_load_n(&list_a.c[cpu_a].head->data, __ATOMIC_ACQUIRE);
tools/testing/selftests/rseq/param_test.c
1368
if (expect_a != __atomic_load_n(&list_a.c[cpu_a].head->data, __ATOMIC_ACQUIRE)) {
tools/testing/selftests/rseq/param_test.c
1381
expect_b = __atomic_load_n(&list_b.c[cpu_b].head->data, __ATOMIC_ACQUIRE);
tools/testing/selftests/rseq/param_test.c
398
struct percpu_list_node *head;
tools/testing/selftests/rseq/param_test.c
643
expect = (intptr_t)RSEQ_READ_ONCE(list->c[cpu].head);
tools/testing/selftests/rseq/param_test.c
645
targetptr = (intptr_t *)&list->c[cpu].head;
tools/testing/selftests/rseq/param_test.c
669
struct percpu_list_node *head;
tools/testing/selftests/rseq/param_test.c
675
targetptr = (intptr_t *)&list->c[cpu].head;
tools/testing/selftests/rseq/param_test.c
678
load = (intptr_t *)&head;
tools/testing/selftests/rseq/param_test.c
683
node = head;
tools/testing/selftests/rseq/param_test.c
703
node = list->c[cpu].head;
tools/testing/selftests/rseq/param_test.c
706
list->c[cpu].head = node->next;
tools/testing/selftests/rseq/param_test.c
762
node->next = list.c[i].head;
tools/testing/selftests/rseq/param_test.c
763
list.c[i].head = node;
tools/testing/selftests/rseq/param_test.c
844
struct percpu_buffer_node *head;
tools/testing/selftests/rseq/param_test.c
856
head = NULL;
tools/testing/selftests/rseq/param_test.c
859
head = RSEQ_READ_ONCE(buffer->c[cpu].array[offset - 1]);
tools/testing/selftests/rseq/param_test.c
865
(intptr_t)head, newval, cpu);
tools/testing/selftests/rseq/param_test.c
872
return head;
tools/testing/selftests/rseq/param_test.c
882
struct percpu_buffer_node *head;
tools/testing/selftests/rseq/param_test.c
888
head = buffer->c[cpu].array[offset - 1];
tools/testing/selftests/rseq/param_test.c
890
return head;
tools/testing/selftests/timers/clocksource-switch.c
45
char *head, *tmp;
tools/testing/selftests/timers/clocksource-switch.c
56
head = buf;
tools/testing/selftests/timers/clocksource-switch.c
58
while (head - buf < size) {
tools/testing/selftests/timers/clocksource-switch.c
60
for (tmp = head; *tmp != ' '; tmp++) {
tools/testing/selftests/timers/clocksource-switch.c
67
strcpy(list[i], head);
tools/testing/selftests/timers/clocksource-switch.c
68
head = tmp + 1;
tools/testing/selftests/ublk/kublk.c
913
unsigned head;
tools/testing/selftests/ublk/kublk.c
916
io_uring_for_each_cqe(&t->ring, head, cqe) {
tools/testing/selftests/vfio/lib/drivers/dsa/registers.h
578
u32 head:16;
tools/testing/selftests/x86/lam.c
521
sring->head = sq_ptr + p.sq_off.head;
tools/testing/selftests/x86/lam.c
541
cring->head = cq_ptr + p.cq_off.head;
tools/testing/selftests/x86/lam.c
572
unsigned int head;
tools/testing/selftests/x86/lam.c
575
head = *cring->head;
tools/testing/selftests/x86/lam.c
579
if (head == *cring->tail)
tools/testing/selftests/x86/lam.c
582
cqe = &cring->queue.cqes[head & *s->cq_ring.ring_mask];
tools/testing/selftests/x86/lam.c
592
head++;
tools/testing/selftests/x86/lam.c
595
*cring->head = head;
tools/testing/selftests/x86/lam.c
98
unsigned int *head;
tools/testing/selftests/x86/test_shadow_stack.c
563
struct node *head = NULL, *cur;
tools/testing/selftests/x86/test_shadow_stack.c
581
cur->next = head;
tools/testing/selftests/x86/test_shadow_stack.c
582
head = cur;
tools/testing/selftests/x86/test_shadow_stack.c
585
while (head) {
tools/testing/selftests/x86/test_shadow_stack.c
586
cur = head;
tools/testing/selftests/x86/test_shadow_stack.c
587
head = cur->next;
tools/testing/selftests/x86/test_shadow_stack.c
606
struct node *head = NULL, *cur;
tools/testing/selftests/x86/test_shadow_stack.c
630
cur->next = head;
tools/testing/selftests/x86/test_shadow_stack.c
631
head = cur;
tools/testing/selftests/x86/test_shadow_stack.c
640
while (head) {
tools/testing/selftests/x86/test_shadow_stack.c
641
cur = head;
tools/testing/selftests/x86/test_shadow_stack.c
642
head = cur->next;
tools/testing/shared/linux/radix-tree.h
10
static inline void trace_call_rcu(struct rcu_head *head,
tools/testing/shared/linux/radix-tree.h
11
void (*func)(struct rcu_head *head))
tools/testing/shared/linux/radix-tree.h
14
printf("Delaying free of %p to slab\n", (char *)head -
tools/testing/shared/linux/radix-tree.h
16
call_rcu(head, func);
tools/testing/shared/maple-shared.h
13
void maple_rcu_cb(struct rcu_head *head);
tools/testing/shared/maple-shim.c
10
void maple_rcu_cb(struct rcu_head *head) {
tools/testing/shared/maple-shim.c
11
struct maple_node *node = container_of(head, struct maple_node, rcu);
tools/testing/vsock/vsock_diag_test.c
106
static void print_vsock_stats(FILE *fp, struct list_head *head)
tools/testing/vsock/vsock_diag_test.c
110
list_for_each_entry(st, head, list)
tools/testing/vsock/vsock_diag_test.c
114
static struct vsock_stat *find_vsock_stat(struct list_head *head, int fd)
tools/testing/vsock/vsock_diag_test.c
124
list_for_each_entry(st, head, list)
tools/testing/vsock/vsock_diag_test.c
132
static void check_no_sockets(struct list_head *head)
tools/testing/vsock/vsock_diag_test.c
134
if (!list_empty(head)) {
tools/testing/vsock/vsock_diag_test.c
136
print_vsock_stats(stderr, head);
tools/testing/vsock/vsock_diag_test.c
141
static void check_num_sockets(struct list_head *head, int expected)
tools/testing/vsock/vsock_diag_test.c
146
list_for_each(node, head)
tools/testing/vsock/vsock_diag_test.c
152
print_vsock_stats(stderr, head);
tools/usb/usbip/libsrc/list.h
111
#define list_for_each(pos, head) \
tools/usb/usbip/libsrc/list.h
112
for (pos = (head)->next; pos != (head); pos = pos->next)
tools/usb/usbip/libsrc/list.h
120
#define list_for_each_safe(pos, n, head) \
tools/usb/usbip/libsrc/list.h
121
for (pos = (head)->next, n = pos->next; pos != (head); \
tools/usb/usbip/libsrc/list.h
58
static inline void list_add(struct list_head *new, struct list_head *head)
tools/usb/usbip/libsrc/list.h
60
__list_add(new, head, head->next);
tools/virtio/ringtest/ring.c
114
unsigned head, index;
tools/virtio/ringtest/ring.c
120
head = (ring_size - 1) & (guest.avail_idx++);
tools/virtio/ringtest/ring.c
125
ring[head].addr = (unsigned long)(void*)buf;
tools/virtio/ringtest/ring.c
126
ring[head].len = len;
tools/virtio/ringtest/ring.c
133
index = ring[head].index;
tools/virtio/ringtest/ring.c
138
ring[head].flags = DESC_HW;
tools/virtio/ringtest/ring.c
145
unsigned head = (ring_size - 1) & guest.last_used_idx;
tools/virtio/ringtest/ring.c
149
if (ring[head].flags & DESC_HW)
tools/virtio/ringtest/ring.c
153
*lenp = ring[head].len;
tools/virtio/ringtest/ring.c
154
index = ring[head].index & (ring_size - 1);
tools/virtio/ringtest/ring.c
166
unsigned head = (ring_size - 1) & guest.last_used_idx;
tools/virtio/ringtest/ring.c
168
return (ring[head].flags & DESC_HW);
tools/virtio/ringtest/ring.c
221
unsigned head = (ring_size - 1) & host.used_idx;
tools/virtio/ringtest/ring.c
223
return !(ring[head].flags & DESC_HW);
tools/virtio/ringtest/ring.c
228
unsigned head = (ring_size - 1) & host.used_idx;
tools/virtio/ringtest/ring.c
230
if (!(ring[head].flags & DESC_HW))
tools/virtio/ringtest/ring.c
241
ring[head].len--;
tools/virtio/ringtest/ring.c
249
ring[head].flags = 0;
tools/virtio/ringtest/virtio_ring_0_9.c
101
unsigned head;
tools/virtio/ringtest/virtio_ring_0_9.c
111
head = (ring_size - 1) & (guest.avail_idx++);
tools/virtio/ringtest/virtio_ring_0_9.c
113
head = guest.free_head;
tools/virtio/ringtest/virtio_ring_0_9.c
118
desc[head].flags = VRING_DESC_F_NEXT;
tools/virtio/ringtest/virtio_ring_0_9.c
119
desc[head].addr = (unsigned long)(void *)buf;
tools/virtio/ringtest/virtio_ring_0_9.c
120
desc[head].len = len;
tools/virtio/ringtest/virtio_ring_0_9.c
125
desc[head].flags &= ~VRING_DESC_F_NEXT;
tools/virtio/ringtest/virtio_ring_0_9.c
127
guest.free_head = desc[head].next;
tools/virtio/ringtest/virtio_ring_0_9.c
130
data[head].data = datap;
tools/virtio/ringtest/virtio_ring_0_9.c
137
(head | (avail & ~(ring_size - 1))) ^ 0x8000;
tools/virtio/ringtest/virtio_ring_0_9.c
143
ring.avail->ring[avail] = head;
tools/virtio/ringtest/virtio_ring_0_9.c
154
unsigned head;
tools/virtio/ringtest/virtio_ring_0_9.c
159
head = (ring_size - 1) & guest.last_used_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
160
index = ring.used->ring[head].id;
tools/virtio/ringtest/virtio_ring_0_9.c
172
head = (ring_size - 1) & guest.last_used_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
173
index = head;
tools/virtio/ringtest/virtio_ring_0_9.c
175
head = (ring_size - 1) & guest.last_used_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
176
index = ring.used->ring[head].id;
tools/virtio/ringtest/virtio_ring_0_9.c
183
*lenp = ring.used->ring[head].len;
tools/virtio/ringtest/virtio_ring_0_9.c
201
unsigned short head = last_used_idx & (ring_size - 1);
tools/virtio/ringtest/virtio_ring_0_9.c
202
unsigned index = ring.used->ring[head].id;
tools/virtio/ringtest/virtio_ring_0_9.c
260
unsigned head = host.used_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
262
unsigned index = ring.avail->ring[head & (ring_size - 1)];
tools/virtio/ringtest/virtio_ring_0_9.c
264
return ((index ^ head ^ 0x8000) & ~(ring_size - 1));
tools/virtio/ringtest/virtio_ring_0_9.c
266
return head == ring.avail->idx;
tools/virtio/ringtest/virtio_ring_0_9.c
274
unsigned head;
tools/virtio/ringtest/virtio_ring_0_9.c
277
head = ring.avail->ring[used_idx & (ring_size - 1)];
tools/virtio/ringtest/virtio_ring_0_9.c
278
if ((used_idx ^ head ^ 0x8000) & ~(ring_size - 1))
tools/virtio/ringtest/virtio_ring_0_9.c
284
desc = &ring.desc[head & (ring_size - 1)];
tools/virtio/ringtest/virtio_ring_0_9.c
294
head = used_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
296
head = ring.avail->ring[used_idx];
tools/virtio/ringtest/virtio_ring_0_9.c
298
desc = &ring.desc[head];
tools/virtio/ringtest/virtio_ring_0_9.c
308
ring.used->ring[used_idx].id = head;
tools/virtio/vringh_test.c
111
static inline int vringh_get_head(struct vringh *vrh, u16 *head)
tools/virtio/vringh_test.c
128
err = get_user(*head, &vrh->vring.avail->ring[i]);
tools/virtio/vringh_test.c
201
u16 head, written;
tools/virtio/vringh_test.c
205
err = vringh_get_head(&vrh, &head);
tools/virtio/vringh_test.c
230
getrange, &head);
tools/virtio/vringh_test.c
276
err = vringh_complete_user(&vrh, head, written);
tools/virtio/vringh_test.c
452
u16 head;
tools/virtio/vringh_test.c
502
err = vringh_getdesc_user(&vrh, &riov, &wiov, getrange, &head);
tools/virtio/vringh_test.c
526
err = vringh_getdesc_user(&vrh, &riov, &wiov, getrange, &head);
tools/virtio/vringh_test.c
561
err = vringh_complete_user(&vrh, head, err);
tools/virtio/vringh_test.c
595
err = vringh_getdesc_user(&vrh, &riov, &wiov, getrange, &head);
tools/virtio/vringh_test.c
623
used[0].id = head;
tools/virtio/vringh_test.c
648
err = vringh_getdesc_user(&vrh, &riov, &wiov, getrange, &head);
tools/virtio/vringh_test.c
651
used[i].id = head;
tools/virtio/vringh_test.c
732
err = vringh_getdesc_user(&vrh, &riov, &wiov, getrange, &head);
tools/virtio/vringh_test.c
736
if (head != 0)
tools/virtio/vringh_test.c
737
errx(1, "vringh_getdesc_user: head %i not 0", head);