Symbol: next
arch/alpha/include/asm/core_marvel.h
310
struct io7 *next;
arch/alpha/include/asm/gct.h
24
gct6_handle next;
arch/alpha/include/asm/mmu_context.h
122
unsigned long next = asn + 1;
arch/alpha/include/asm/mmu_context.h
127
next = (asn & ~HARDWARE_ASN_MASK) + ASN_FIRST_VERSION;
arch/alpha/include/asm/mmu_context.h
129
cpu_last_asn(cpu) = next;
arch/alpha/include/asm/mmu_context.h
130
return next;
arch/alpha/include/asm/mmu_context.h
135
struct task_struct *next)
arch/alpha/include/asm/mmu_context.h
160
task_thread_info(next)->pcb.asn = mmc & HARDWARE_ASN_MASK;
arch/alpha/include/asm/pci.h
22
struct pci_controller *next;
arch/alpha/kernel/console.c
85
for (hose = hose_head; hose; hose = hose->next) {
arch/alpha/kernel/core_marvel.c
104
for (io7 = io7_head; io7 && io7->pe != pe; io7 = io7->next)
arch/alpha/kernel/core_marvel.c
138
io7->next = io7_head;
arch/alpha/kernel/core_marvel.c
141
for (insp = io7_head; insp; insp = insp->next) {
arch/alpha/kernel/core_marvel.c
148
if (NULL == insp->next ||
arch/alpha/kernel/core_marvel.c
149
insp->next->pe > io7->pe) { /* insert here */
arch/alpha/kernel/core_marvel.c
150
io7->next = insp->next;
arch/alpha/kernel/core_marvel.c
151
insp->next = io7;
arch/alpha/kernel/core_marvel.c
159
io7->next = io7_head;
arch/alpha/kernel/core_marvel.c
698
for (hose = hose_head; hose; hose = hose->next) {
arch/alpha/kernel/core_marvel.c
96
return (prev ? prev->next : io7_head);
arch/alpha/kernel/core_mcpcia.c
440
for (hose = hose_head; hose; hose = hose->next)
arch/alpha/kernel/core_mcpcia.c
551
for (hose = hose_head; hose; hose = hose->next, iodpp++) {
arch/alpha/kernel/core_mcpcia.c
595
for (hose = hose_head; hose; hose = hose->next)
arch/alpha/kernel/core_titan.c
479
for (hose = hose_head; hose; hose = hose->next)
arch/alpha/kernel/core_titan.c
753
for (hose = hose_head; hose; hose = hose->next)
arch/alpha/kernel/err_common.c
159
struct el_subpacket *next = NULL;
arch/alpha/kernel/err_common.c
162
for (; h && h->class != header->class; h = h->next);
arch/alpha/kernel/err_common.c
163
if (h) next = h->handler(header);
arch/alpha/kernel/err_common.c
165
return next;
arch/alpha/kernel/err_common.c
197
struct el_subpacket *next = NULL;
arch/alpha/kernel/err_common.c
204
next = el_process_header_subpacket(header);
arch/alpha/kernel/err_common.c
207
if (NULL == (next = el_process_subpacket_reg(header))) {
arch/alpha/kernel/err_common.c
216
return next;
arch/alpha/kernel/err_common.c
225
for (a = subpacket_annotation_list; a; a = a->next) {
arch/alpha/kernel/err_common.c
286
for (; a->next != NULL; a = a->next) {
arch/alpha/kernel/err_common.c
294
a->next = new;
arch/alpha/kernel/err_common.c
296
new->next = NULL;
arch/alpha/kernel/err_common.c
308
for (; h->next != NULL; h = h->next) {
arch/alpha/kernel/err_common.c
315
h->next = new;
arch/alpha/kernel/err_common.c
317
new->next = NULL;
arch/alpha/kernel/err_impl.h
18
struct el_subpacket_annotation *next;
arch/alpha/kernel/err_impl.h
28
struct el_subpacket_handler *next;
arch/alpha/kernel/gct.c
40
if (node->next)
arch/alpha/kernel/gct.c
41
status |= gct6_find_nodes(GCT_NODE_PTR(node->next), search);
arch/alpha/kernel/module.c
122
for (g = chains[i].next; g ; g = n) {
arch/alpha/kernel/module.c
123
n = g->next;
arch/alpha/kernel/module.c
23
struct got_entry *next;
arch/alpha/kernel/module.c
40
for (g = chains + r_sym; g ; g = g->next)
arch/alpha/kernel/module.c
50
g->next = chains[r_sym].next;
arch/alpha/kernel/module.c
54
chains[r_sym].next = g;
arch/alpha/kernel/pci.c
204
struct pdev_srm_saved_conf *next;
arch/alpha/kernel/pci.c
228
tmp->next = srm_saved_configs;
arch/alpha/kernel/pci.c
246
for (tmp = srm_saved_configs; tmp; tmp = tmp->next) {
arch/alpha/kernel/pci.c
334
for (next_busno = 0, hose = hose_head; hose; hose = hose->next) {
arch/alpha/kernel/pci.c
382
for (hose = hose_head; hose; hose = hose->next) {
arch/alpha/kernel/pci.c
397
hose_tail = &hose->next;
arch/alpha/kernel/pci.c
420
for(hose = hose_head; hose; hose = hose->next)
arch/alpha/kernel/setup.c
1340
.next = c_next,
arch/alpha/kernel/setup.c
228
for (hose = hose_head; hose; hose = hose->next)
arch/alpha/kernel/smc37c669.c
2344
int __init smcc669_open( struct FILE *fp, char *info, char *next, char *mode )
arch/alpha/kernel/sys_rawhide.c
172
for (hose = hose_head; hose; hose = hose->next) {
arch/arc/include/asm/dsp-impl.h
101
long unsigned int *readfrom = &next->thread.dsp.ACC0_GLO;
arch/arc/include/asm/dsp-impl.h
98
struct task_struct *next)
arch/arc/include/asm/mmu_context.h
132
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/arc/include/asm/mmu_context.h
148
cpumask_set_cpu(cpu, mm_cpumask(next));
arch/arc/include/asm/mmu_context.h
150
mmu_setup_pgd(next, next->pgd);
arch/arc/include/asm/mmu_context.h
152
get_new_mmu_context(next);
arch/arc/include/asm/switch_to.h
17
#define switch_to(prev, next, last) \
arch/arc/include/asm/switch_to.h
19
dsp_save_restore(prev, next); \
arch/arc/include/asm/switch_to.h
20
fpu_save_restore(prev, next); \
arch/arc/include/asm/switch_to.h
21
last = __switch_to(prev, next);\
arch/arc/kernel/fpu.c
32
void fpu_save_restore(struct task_struct *prev, struct task_struct *next)
arch/arc/kernel/fpu.c
35
unsigned int *readfrom = &next->thread.fpu.aux_dpfp[0].l;
arch/arc/kernel/fpu.c
69
void fpu_save_restore(struct task_struct *prev, struct task_struct *next)
arch/arc/kernel/fpu.c
72
struct arc_fpu *restore = &next->thread.fpu;
arch/arc/kernel/setup.c
637
.next = c_next,
arch/arm/include/asm/ecard.h
140
struct expansion_card *next;
arch/arm/include/asm/fiq.h
23
struct fiq_handler *next;
arch/arm/include/asm/mmu_context.h
108
#define activate_mm(prev,next) switch_mm(prev, next, NULL)
arch/arm/include/asm/mmu_context.h
117
switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/arm/include/asm/mmu_context.h
129
!cpumask_empty(mm_cpumask(next)) &&
arch/arm/include/asm/mmu_context.h
130
!cpumask_test_cpu(cpu, mm_cpumask(next)))
arch/arm/include/asm/mmu_context.h
133
if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) {
arch/arm/include/asm/mmu_context.h
134
check_and_switch_context(next, tsk);
arch/arm/include/asm/spinlock.h
116
return lock.tickets.owner == lock.tickets.next;
arch/arm/include/asm/spinlock.h
127
return (tickets.next - tickets.owner) > 1;
arch/arm/include/asm/spinlock.h
73
while (lockval.tickets.next != lockval.tickets.owner) {
arch/arm/include/asm/spinlock_types.h
16
u16 next;
arch/arm/include/asm/spinlock_types.h
20
u16 next;
arch/arm/include/asm/switch_to.h
27
#define switch_to(prev,next,last) \
arch/arm/include/asm/switch_to.h
31
__this_cpu_write(__entry_task, next); \
arch/arm/include/asm/switch_to.h
32
last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \
arch/arm/kernel/fiq.c
119
f->next = current_fiq;
arch/arm/kernel/fiq.c
136
current_fiq = current_fiq->next;
arch/arm/kernel/hibernate.c
86
for (pbe = restore_pblist; pbe; pbe = pbe->next)
arch/arm/kernel/setup.c
1342
.next = c_next,
arch/arm/mach-omap2/powerdomain.c
164
int prev, next, state, trace_state = 0;
arch/arm/mach-omap2/powerdomain.c
185
next = pwrdm_read_next_pwrst(pwrdm);
arch/arm/mach-omap2/powerdomain.c
186
if (next != prev) {
arch/arm/mach-omap2/powerdomain.c
188
((next & OMAP_POWERSTATE_MASK) << 8) |
arch/arm/mach-rpc/ecard.c
499
for (ec = cards; ec; ec = ec->next) {
arch/arm/mach-rpc/ecard.c
562
for (ec = cards; ec; ec = ec->next) {
arch/arm/mach-rpc/ecard.c
656
ec = ec->next;
arch/arm/mach-rpc/ecard.c
966
for (ecp = &cards; *ecp; ecp = &(*ecp)->next);
arch/arm/mm/idmap.c
105
next = pgd_addr_end(addr, end);
arch/arm/mm/idmap.c
106
idmap_add_pud(pgd, addr, next, prot);
arch/arm/mm/idmap.c
107
} while (pgd++, addr = next, addr != end);
arch/arm/mm/idmap.c
28
unsigned long next;
arch/arm/mm/idmap.c
49
next = pmd_addr_end(addr, end);
arch/arm/mm/idmap.c
52
} while (pmd++, addr = next, addr != end);
arch/arm/mm/idmap.c
73
unsigned long next;
arch/arm/mm/idmap.c
76
next = pud_addr_end(addr, end);
arch/arm/mm/idmap.c
77
idmap_add_pmd(pud, addr, next, prot);
arch/arm/mm/idmap.c
78
} while (pud++, addr = next, addr != end);
arch/arm/mm/idmap.c
85
unsigned long next;
arch/arm/mm/kasan_init.c
118
next = pmd_addr_end(addr, end);
arch/arm/mm/kasan_init.c
119
kasan_pte_populate(pmdp, addr, next, early);
arch/arm/mm/kasan_init.c
120
} while (pmdp++, addr = next, addr != end);
arch/arm/mm/kasan_init.c
126
unsigned long next;
arch/arm/mm/kasan_init.c
149
next = pgd_addr_end(addr, end);
arch/arm/mm/kasan_init.c
158
kasan_pmd_populate(pudp, addr, next, early);
arch/arm/mm/kasan_init.c
159
} while (pgdp++, addr = next, addr != end);
arch/arm/mm/kasan_init.c
46
unsigned long next;
arch/arm/mm/kasan_init.c
53
next = addr + PAGE_SIZE;
arch/arm/mm/kasan_init.c
87
} while (ptep++, addr = next, addr != end);
arch/arm/mm/kasan_init.c
96
unsigned long next;
arch/arm/mm/mmu.c
1081
unsigned long addr, next = 0;
arch/arm/mm/mmu.c
1087
if (addr < next)
arch/arm/mm/mmu.c
1114
next = (addr + PMD_SIZE - 1) & PMD_MASK;
arch/arm/mm/mmu.c
808
unsigned long next;
arch/arm/mm/mmu.c
815
next = pmd_addr_end(addr, end);
arch/arm/mm/mmu.c
822
((addr | next | phys) & ~SECTION_MASK) == 0) {
arch/arm/mm/mmu.c
823
__map_init_section(pmd, addr, next, phys, type, ng);
arch/arm/mm/mmu.c
825
alloc_init_pte(pmd, addr, next,
arch/arm/mm/mmu.c
829
phys += next - addr;
arch/arm/mm/mmu.c
831
} while (pmd++, addr = next, addr != end);
arch/arm/mm/mmu.c
840
unsigned long next;
arch/arm/mm/mmu.c
843
next = pud_addr_end(addr, end);
arch/arm/mm/mmu.c
844
alloc_init_pmd(pud, addr, next, phys, type, alloc, ng);
arch/arm/mm/mmu.c
845
phys += next - addr;
arch/arm/mm/mmu.c
846
} while (pud++, addr = next, addr != end);
arch/arm/mm/mmu.c
855
unsigned long next;
arch/arm/mm/mmu.c
858
next = p4d_addr_end(addr, end);
arch/arm/mm/mmu.c
859
alloc_init_pud(p4d, addr, next, phys, type, alloc, ng);
arch/arm/mm/mmu.c
860
phys += next - addr;
arch/arm/mm/mmu.c
861
} while (p4d++, addr = next, addr != end);
arch/arm/mm/mmu.c
961
unsigned long next = pgd_addr_end(addr, end);
arch/arm/mm/mmu.c
963
alloc_init_p4d(pgd, addr, next, phys, type, alloc, ng);
arch/arm/mm/mmu.c
965
phys += next - addr;
arch/arm/mm/mmu.c
966
addr = next;
arch/arm/probes/decode.c
417
const struct decode_header *next;
arch/arm/probes/decode.c
445
for (;; h = next) {
arch/arm/probes/decode.c
452
next = (struct decode_header *)
arch/arm/probes/decode.c
465
next = (struct decode_header *)d->table.table;
arch/arm64/include/asm/fpsimd.h
77
extern void fpsimd_thread_switch(struct task_struct *next);
arch/arm64/include/asm/hw_breakpoint.h
127
extern void hw_breakpoint_thread_switch(struct task_struct *next);
arch/arm64/include/asm/hw_breakpoint.h
130
static inline void hw_breakpoint_thread_switch(struct task_struct *next)
arch/arm64/include/asm/mmu_context.h
235
static inline void __switch_mm(struct mm_struct *next)
arch/arm64/include/asm/mmu_context.h
241
if (next == &init_mm) {
arch/arm64/include/asm/mmu_context.h
246
check_and_switch_context(next);
arch/arm64/include/asm/mmu_context.h
250
switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/arm64/include/asm/mmu_context.h
253
if (prev != next)
arch/arm64/include/asm/mmu_context.h
254
__switch_mm(next);
arch/arm64/include/asm/mmu_context.h
262
update_saved_ttbr0(tsk, next);
arch/arm64/include/asm/mmu_context.h
31
static inline void contextidr_thread_switch(struct task_struct *next)
arch/arm64/include/asm/mmu_context.h
36
write_sysreg(task_pid_nr(next), contextidr_el1);
arch/arm64/include/asm/mte.h
102
void mte_thread_switch(struct task_struct *next);
arch/arm64/include/asm/mte.h
140
static inline void mte_thread_switch(struct task_struct *next)
arch/arm64/include/asm/processor.h
381
struct task_struct *next);
arch/arm64/kernel/asm-offsets.c
152
DEFINE(HIBERN_PBE_NEXT, offsetof(struct pbe, next));
arch/arm64/kernel/cpuinfo.c
297
.next = c_next,
arch/arm64/kernel/fpsimd.c
1509
void fpsimd_thread_switch(struct task_struct *next)
arch/arm64/kernel/fpsimd.c
1524
if (test_tsk_thread_flag(next, TIF_KERNEL_FPSTATE)) {
arch/arm64/kernel/fpsimd.c
1526
fpsimd_load_kernel_state(next);
arch/arm64/kernel/fpsimd.c
1535
&next->thread.uw.fpsimd_state;
arch/arm64/kernel/fpsimd.c
1536
wrong_cpu = next->thread.fpsimd_cpu != smp_processor_id();
arch/arm64/kernel/fpsimd.c
1538
update_tsk_thread_flag(next, TIF_FOREIGN_FPSTATE,
arch/arm64/kernel/hw_breakpoint.c
904
void hw_breakpoint_thread_switch(struct task_struct *next)
arch/arm64/kernel/hw_breakpoint.c
918
next_debug_info = &next->thread.debug;
arch/arm64/kernel/mte.c
283
void mte_thread_switch(struct task_struct *next)
arch/arm64/kernel/mte.c
288
mte_update_sctlr_user(next);
arch/arm64/kernel/mte.c
289
mte_update_gcr_excl(next);
arch/arm64/kernel/mte.c
292
mte_disable_tco_entry(next);
arch/arm64/kernel/pi/map_range.c
52
u64 next = min((start | lmask) + 1, PAGE_ALIGN(end));
arch/arm64/kernel/pi/map_range.c
54
if (level < 2 || (level == 2 && (start | next | pa) & lmask)) {
arch/arm64/kernel/pi/map_range.c
64
map_range(pte, start, next, pa, prot, level + 1,
arch/arm64/kernel/pi/map_range.c
85
pa += next - start;
arch/arm64/kernel/pi/map_range.c
86
start = next;
arch/arm64/kernel/process.c
528
static void tls_thread_switch(struct task_struct *next)
arch/arm64/kernel/process.c
532
if (is_compat_thread(task_thread_info(next)))
arch/arm64/kernel/process.c
533
write_sysreg(next->thread.uw.tp_value, tpidrro_el0);
arch/arm64/kernel/process.c
537
write_sysreg(*task_user_tls(next), tpidr_el0);
arch/arm64/kernel/process.c
539
write_sysreg_s(next->thread.tpidr2_el0, SYS_TPIDR2_EL0);
arch/arm64/kernel/process.c
546
static void ssbs_thread_switch(struct task_struct *next)
arch/arm64/kernel/process.c
552
if (unlikely(next->flags & PF_KTHREAD))
arch/arm64/kernel/process.c
562
spectre_v4_enable_task_mitigation(next);
arch/arm64/kernel/process.c
574
static void entry_task_switch(struct task_struct *next)
arch/arm64/kernel/process.c
576
__this_cpu_write(__entry_task, next);
arch/arm64/kernel/process.c
586
static void gcs_thread_switch(struct task_struct *next)
arch/arm64/kernel/process.c
593
write_sysreg_s(next->thread.gcspr_el0, SYS_GCSPR_EL0);
arch/arm64/kernel/process.c
595
if (current->thread.gcs_el0_mode != next->thread.gcs_el0_mode)
arch/arm64/kernel/process.c
596
gcs_set_el0_mode(next);
arch/arm64/kernel/process.c
607
if (task_gcs_el0_enabled(current) || task_gcs_el0_enabled(next))
arch/arm64/kernel/process.c
613
static void gcs_thread_switch(struct task_struct *next)
arch/arm64/kernel/process.c
625
static void update_cntkctl_el1(struct task_struct *next)
arch/arm64/kernel/process.c
627
struct thread_info *ti = task_thread_info(next);
arch/arm64/kernel/process.c
640
struct task_struct *next)
arch/arm64/kernel/process.c
644
(read_ti_thread_flags(task_thread_info(next)) &
arch/arm64/kernel/process.c
646
update_cntkctl_el1(next);
arch/arm64/kernel/process.c
668
static void permission_overlay_switch(struct task_struct *next)
arch/arm64/kernel/process.c
674
if (current->thread.por_el0 != next->thread.por_el0) {
arch/arm64/kernel/process.c
675
write_sysreg_s(next->thread.por_el0, SYS_POR_EL0);
arch/arm64/kernel/process.c
707
struct task_struct *next)
arch/arm64/kernel/process.c
711
fpsimd_thread_switch(next);
arch/arm64/kernel/process.c
712
tls_thread_switch(next);
arch/arm64/kernel/process.c
713
hw_breakpoint_thread_switch(next);
arch/arm64/kernel/process.c
714
contextidr_thread_switch(next);
arch/arm64/kernel/process.c
715
entry_task_switch(next);
arch/arm64/kernel/process.c
716
ssbs_thread_switch(next);
arch/arm64/kernel/process.c
717
cntkctl_thread_switch(prev, next);
arch/arm64/kernel/process.c
718
ptrauth_thread_switch_user(next);
arch/arm64/kernel/process.c
719
permission_overlay_switch(next);
arch/arm64/kernel/process.c
720
gcs_thread_switch(next);
arch/arm64/kernel/process.c
736
mte_thread_switch(next);
arch/arm64/kernel/process.c
738
if (prev->thread.sctlr_user != next->thread.sctlr_user)
arch/arm64/kernel/process.c
739
update_sctlr_el1(next->thread.sctlr_user);
arch/arm64/kernel/process.c
742
last = cpu_switch_to(prev, next);
arch/arm64/kernel/smp.c
689
goto next;
arch/arm64/kernel/smp.c
694
goto next;
arch/arm64/kernel/smp.c
707
goto next;
arch/arm64/kernel/smp.c
723
goto next;
arch/arm64/kernel/smp.c
729
next:
arch/arm64/kvm/hyp/include/nvhe/spinlock.h
105
return lockval.owner != lockval.next;
arch/arm64/kvm/hyp/include/nvhe/spinlock.h
24
u16 next, owner;
arch/arm64/kvm/hyp/include/nvhe/spinlock.h
26
u16 owner, next;
arch/arm64/kvm/hyp/nvhe/list_debug.c
31
struct list_head *next)
arch/arm64/kvm/hyp/nvhe/list_debug.c
33
if (NVHE_CHECK_DATA_CORRUPTION(next->prev != prev) ||
arch/arm64/kvm/hyp/nvhe/list_debug.c
34
NVHE_CHECK_DATA_CORRUPTION(prev->next != next) ||
arch/arm64/kvm/hyp/nvhe/list_debug.c
35
NVHE_CHECK_DATA_CORRUPTION(new == prev || new == next))
arch/arm64/kvm/hyp/nvhe/list_debug.c
44
struct list_head *prev, *next;
arch/arm64/kvm/hyp/nvhe/list_debug.c
47
next = entry->next;
arch/arm64/kvm/hyp/nvhe/list_debug.c
49
if (NVHE_CHECK_DATA_CORRUPTION(next == LIST_POISON1) ||
arch/arm64/kvm/hyp/nvhe/list_debug.c
51
NVHE_CHECK_DATA_CORRUPTION(prev->next != entry) ||
arch/arm64/kvm/hyp/nvhe/list_debug.c
52
NVHE_CHECK_DATA_CORRUPTION(next->prev != entry))
arch/arm64/kvm/hyp/nvhe/page_alloc.c
214
p = node_to_page(pool->free_area[i].next);
arch/arm64/kvm/mmu.c
126
u64 next, chunk_size;
arch/arm64/kvm/mmu.c
155
next = __stage2_range_addr_end(addr, end, chunk_size);
arch/arm64/kvm/mmu.c
156
ret = KVM_PGT_FN(kvm_pgtable_stage2_split)(pgt, addr, next - addr, cache);
arch/arm64/kvm/mmu.c
159
} while (addr = next, addr != end);
arch/arm64/kvm/mmu.c
69
u64 next;
arch/arm64/kvm/mmu.c
76
next = stage2_range_addr_end(addr, end);
arch/arm64/kvm/mmu.c
77
ret = fn(pgt, addr, next - addr);
arch/arm64/kvm/mmu.c
81
if (resched && next != end)
arch/arm64/kvm/mmu.c
83
} while (addr = next, addr != end);
arch/arm64/kvm/mmu.c
920
u64 next;
arch/arm64/kvm/mmu.c
923
next = stage2_range_addr_end(addr, end);
arch/arm64/kvm/mmu.c
925
next - addr);
arch/arm64/kvm/mmu.c
926
if (next != end)
arch/arm64/kvm/mmu.c
928
} while (addr = next, addr != end);
arch/arm64/kvm/sys_regs.c
5055
.next = idregs_debug_next,
arch/arm64/kvm/sys_regs.c
5121
.next = sr_resx_next,
arch/arm64/kvm/vgic/vgic-debug.c
271
.next = vgic_debug_next,
arch/arm64/kvm/vgic/vgic-debug.c
480
.next = vgic_its_debug_next,
arch/arm64/kvm/vgic/vgic-init.c
434
struct vgic_redist_region *rdreg, *next;
arch/arm64/kvm/vgic/vgic-init.c
445
list_for_each_entry_safe(rdreg, next, &dist->rd_regions, list)
arch/arm64/kvm/vgic/vgic-its.c
2014
struct its_device *next;
arch/arm64/kvm/vgic/vgic-its.c
2019
next = list_next_entry(dev, dev_list);
arch/arm64/kvm/vgic/vgic-its.c
2020
next_offset = next->device_id - dev->device_id;
arch/arm64/kvm/vgic/vgic-its.c
2027
struct its_ite *next;
arch/arm64/kvm/vgic/vgic-its.c
2032
next = list_next_entry(ite, ite_list);
arch/arm64/kvm/vgic/vgic-its.c
2033
next_offset = next->event_id - ite->event_id;
arch/arm64/mm/contpte.c
454
unsigned long next;
arch/arm64/mm/contpte.c
475
next = pte_cont_addr_end(addr, end);
arch/arm64/mm/contpte.c
476
nr = (next - addr) >> PAGE_SHIFT;
arch/arm64/mm/contpte.c
479
if (((addr | next | (pfn << PAGE_SHIFT)) & ~CONT_PTE_MASK) == 0)
arch/arm64/mm/contpte.c
486
addr = next;
arch/arm64/mm/fixmap.c
58
unsigned long next;
arch/arm64/mm/fixmap.c
68
next = pmd_addr_end(addr, end);
arch/arm64/mm/fixmap.c
70
} while (pmdp++, addr = next, addr != end);
arch/arm64/mm/kasan_init.c
118
unsigned long next;
arch/arm64/mm/kasan_init.c
127
next = addr + PAGE_SIZE;
arch/arm64/mm/kasan_init.c
129
} while (ptep++, addr = next, addr != end && pte_none(__ptep_get(ptep)));
arch/arm64/mm/kasan_init.c
135
unsigned long next;
arch/arm64/mm/kasan_init.c
139
next = pmd_addr_end(addr, end);
arch/arm64/mm/kasan_init.c
140
kasan_pte_populate(pmdp, addr, next, node, early);
arch/arm64/mm/kasan_init.c
141
} while (pmdp++, addr = next, addr != end && pmd_none(READ_ONCE(*pmdp)));
arch/arm64/mm/kasan_init.c
147
unsigned long next;
arch/arm64/mm/kasan_init.c
151
next = pud_addr_end(addr, end);
arch/arm64/mm/kasan_init.c
152
kasan_pmd_populate(pudp, addr, next, node, early);
arch/arm64/mm/kasan_init.c
153
} while (pudp++, addr = next, addr != end && pud_none(READ_ONCE(*pudp)));
arch/arm64/mm/kasan_init.c
159
unsigned long next;
arch/arm64/mm/kasan_init.c
163
next = p4d_addr_end(addr, end);
arch/arm64/mm/kasan_init.c
164
kasan_pud_populate(p4dp, addr, next, node, early);
arch/arm64/mm/kasan_init.c
165
} while (p4dp++, addr = next, addr != end && p4d_none(READ_ONCE(*p4dp)));
arch/arm64/mm/kasan_init.c
171
unsigned long next;
arch/arm64/mm/kasan_init.c
176
next = pgd_addr_end(addr, end);
arch/arm64/mm/kasan_init.c
177
kasan_p4d_populate(pgdp, addr, next, node, early);
arch/arm64/mm/kasan_init.c
178
} while (pgdp++, addr = next, addr != end);
arch/arm64/mm/mmu.c
1472
unsigned long next;
arch/arm64/mm/mmu.c
1476
next = pmd_addr_end(addr, end);
arch/arm64/mm/mmu.c
1497
unmap_hotplug_pte_range(pmdp, addr, next, free_mapped, altmap);
arch/arm64/mm/mmu.c
1498
} while (addr = next, addr < end);
arch/arm64/mm/mmu.c
1505
unsigned long next;
arch/arm64/mm/mmu.c
1509
next = pud_addr_end(addr, end);
arch/arm64/mm/mmu.c
1530
unmap_hotplug_pmd_range(pudp, addr, next, free_mapped, altmap);
arch/arm64/mm/mmu.c
1531
} while (addr = next, addr < end);
arch/arm64/mm/mmu.c
1538
unsigned long next;
arch/arm64/mm/mmu.c
1542
next = p4d_addr_end(addr, end);
arch/arm64/mm/mmu.c
1549
unmap_hotplug_pud_range(p4dp, addr, next, free_mapped, altmap);
arch/arm64/mm/mmu.c
1550
} while (addr = next, addr < end);
arch/arm64/mm/mmu.c
1556
unsigned long next;
arch/arm64/mm/mmu.c
1568
next = pgd_addr_end(addr, end);
arch/arm64/mm/mmu.c
1575
unmap_hotplug_p4d_range(pgdp, addr, next, free_mapped, altmap);
arch/arm64/mm/mmu.c
1576
} while (addr = next, addr < end);
arch/arm64/mm/mmu.c
1621
unsigned long i, next, start = addr;
arch/arm64/mm/mmu.c
1624
next = pmd_addr_end(addr, end);
arch/arm64/mm/mmu.c
1631
free_empty_pte_table(pmdp, addr, next, floor, ceiling);
arch/arm64/mm/mmu.c
1632
} while (addr = next, addr < end);
arch/arm64/mm/mmu.c
1661
unsigned long i, next, start = addr;
arch/arm64/mm/mmu.c
1664
next = pud_addr_end(addr, end);
arch/arm64/mm/mmu.c
1671
free_empty_pmd_table(pudp, addr, next, floor, ceiling);
arch/arm64/mm/mmu.c
1672
} while (addr = next, addr < end);
arch/arm64/mm/mmu.c
1701
unsigned long i, next, start = addr;
arch/arm64/mm/mmu.c
1704
next = p4d_addr_end(addr, end);
arch/arm64/mm/mmu.c
1711
free_empty_pud_table(p4dp, addr, next, floor, ceiling);
arch/arm64/mm/mmu.c
1712
} while (addr = next, addr < end);
arch/arm64/mm/mmu.c
1739
unsigned long next;
arch/arm64/mm/mmu.c
1743
next = pgd_addr_end(addr, end);
arch/arm64/mm/mmu.c
1750
free_empty_p4d_table(pgdp, addr, next, floor, ceiling);
arch/arm64/mm/mmu.c
1751
} while (addr = next, addr < end);
arch/arm64/mm/mmu.c
1756
unsigned long addr, unsigned long next)
arch/arm64/mm/mmu.c
1762
unsigned long addr, unsigned long next)
arch/arm64/mm/mmu.c
1764
vmemmap_verify((pte_t *)pmdp, node, addr, next);
arch/arm64/mm/mmu.c
1881
unsigned long next, end;
arch/arm64/mm/mmu.c
1910
next = addr;
arch/arm64/mm/mmu.c
1918
__pmd_free_pte_page(pmdp, next, /* acquire_mmap_lock = */ false);
arch/arm64/mm/mmu.c
1919
} while (pmdp++, next += PMD_SIZE, next != end);
arch/arm64/mm/mmu.c
203
unsigned long next;
arch/arm64/mm/mmu.c
230
next = pte_cont_addr_end(addr, end);
arch/arm64/mm/mmu.c
233
if ((((addr | next | phys) & ~CONT_PTE_MASK) == 0) &&
arch/arm64/mm/mmu.c
237
init_pte(ptep, addr, next, phys, __prot);
arch/arm64/mm/mmu.c
239
ptep += pte_index(next) - pte_index(addr);
arch/arm64/mm/mmu.c
240
phys += next - addr;
arch/arm64/mm/mmu.c
241
} while (addr = next, addr != end);
arch/arm64/mm/mmu.c
257
unsigned long next;
arch/arm64/mm/mmu.c
262
next = pmd_addr_end(addr, end);
arch/arm64/mm/mmu.c
265
if (((addr | next | phys) & ~PMD_MASK) == 0 &&
arch/arm64/mm/mmu.c
278
ret = alloc_init_cont_pte(pmdp, addr, next, phys, prot,
arch/arm64/mm/mmu.c
286
phys += next - addr;
arch/arm64/mm/mmu.c
287
} while (pmdp++, addr = next, addr != end);
arch/arm64/mm/mmu.c
299
unsigned long next;
arch/arm64/mm/mmu.c
329
next = pmd_cont_addr_end(addr, end);
arch/arm64/mm/mmu.c
332
if ((((addr | next | phys) & ~CONT_PMD_MASK) == 0) &&
arch/arm64/mm/mmu.c
336
ret = init_pmd(pmdp, addr, next, phys, __prot, pgtable_alloc, flags);
arch/arm64/mm/mmu.c
340
pmdp += pmd_index(next) - pmd_index(addr);
arch/arm64/mm/mmu.c
341
phys += next - addr;
arch/arm64/mm/mmu.c
342
} while (addr = next, addr != end);
arch/arm64/mm/mmu.c
356
unsigned long next;
arch/arm64/mm/mmu.c
382
next = pud_addr_end(addr, end);
arch/arm64/mm/mmu.c
388
((addr | next | phys) & ~PUD_MASK) == 0 &&
arch/arm64/mm/mmu.c
399
ret = alloc_init_cont_pmd(pudp, addr, next, phys, prot,
arch/arm64/mm/mmu.c
407
phys += next - addr;
arch/arm64/mm/mmu.c
408
} while (pudp++, addr = next, addr != end);
arch/arm64/mm/mmu.c
422
unsigned long next;
arch/arm64/mm/mmu.c
448
next = p4d_addr_end(addr, end);
arch/arm64/mm/mmu.c
450
ret = alloc_init_pud(p4dp, addr, next, phys, prot,
arch/arm64/mm/mmu.c
458
phys += next - addr;
arch/arm64/mm/mmu.c
459
} while (p4dp++, addr = next, addr != end);
arch/arm64/mm/mmu.c
474
unsigned long addr, end, next;
arch/arm64/mm/mmu.c
489
next = pgd_addr_end(addr, end);
arch/arm64/mm/mmu.c
490
ret = alloc_init_p4d(pgdp, addr, next, phys, prot, pgtable_alloc,
arch/arm64/mm/mmu.c
494
phys += next - addr;
arch/arm64/mm/mmu.c
495
} while (pgdp++, addr = next, addr != end);
arch/arm64/mm/mmu.c
831
unsigned long next, struct mm_walk *walk)
arch/arm64/mm/mmu.c
844
unsigned long next, struct mm_walk *walk)
arch/arm64/mm/mmu.c
866
unsigned long next, struct mm_walk *walk)
arch/arm64/mm/pageattr.c
35
unsigned long next, struct mm_walk *walk)
arch/arm64/mm/pageattr.c
40
if (WARN_ON_ONCE((next - addr) != PUD_SIZE))
arch/arm64/mm/pageattr.c
51
unsigned long next, struct mm_walk *walk)
arch/arm64/mm/pageattr.c
56
if (WARN_ON_ONCE((next - addr) != PMD_SIZE))
arch/arm64/mm/pageattr.c
67
unsigned long next, struct mm_walk *walk)
arch/arm64/mm/trans_pgd.c
105
next = pmd_addr_end(addr, end);
arch/arm64/mm/trans_pgd.c
109
if (copy_pte(info, dst_pmdp, src_pmdp, addr, next))
arch/arm64/mm/trans_pgd.c
115
} while (dst_pmdp++, src_pmdp++, addr = next, addr != end);
arch/arm64/mm/trans_pgd.c
126
unsigned long next;
arch/arm64/mm/trans_pgd.c
141
next = pud_addr_end(addr, end);
arch/arm64/mm/trans_pgd.c
145
if (copy_pmd(info, dst_pudp, src_pudp, addr, next))
arch/arm64/mm/trans_pgd.c
151
} while (dst_pudp++, src_pudp++, addr = next, addr != end);
arch/arm64/mm/trans_pgd.c
162
unsigned long next;
arch/arm64/mm/trans_pgd.c
175
next = p4d_addr_end(addr, end);
arch/arm64/mm/trans_pgd.c
178
if (copy_pud(info, dst_p4dp, src_p4dp, addr, next))
arch/arm64/mm/trans_pgd.c
180
} while (dst_p4dp++, src_p4dp++, addr = next, addr != end);
arch/arm64/mm/trans_pgd.c
188
unsigned long next;
arch/arm64/mm/trans_pgd.c
194
next = pgd_addr_end(addr, end);
arch/arm64/mm/trans_pgd.c
197
if (copy_p4d(info, dst_pgdp, src_pgdp, addr, next))
arch/arm64/mm/trans_pgd.c
199
} while (dst_pgdp++, src_pgdp++, addr = next, addr != end);
arch/arm64/mm/trans_pgd.c
90
unsigned long next;
arch/csky/include/asm/mmu_context.h
24
switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/csky/include/asm/mmu_context.h
29
if (prev != next)
arch/csky/include/asm/mmu_context.h
30
check_and_switch_context(next, cpu);
arch/csky/include/asm/mmu_context.h
32
setup_pgd(next->pgd, next->context.asid.counter);
arch/csky/include/asm/mmu_context.h
34
flush_icache_deferred(next);
arch/csky/include/asm/switch_to.h
10
struct task_struct *next)
arch/csky/include/asm/switch_to.h
13
restore_from_user_fp(&next->thread.user_fp);
arch/csky/include/asm/switch_to.h
17
struct task_struct *next)
arch/csky/include/asm/switch_to.h
27
#define switch_to(prev, next, last) \
arch/csky/include/asm/switch_to.h
30
struct task_struct *__next = (next); \
arch/csky/include/asm/switch_to.h
32
((last) = __switch_to((prev), (next))); \
arch/csky/kernel/cpu-probe.c
14
unsigned int cur, next, i;
arch/csky/kernel/cpu-probe.c
24
next = mfcr("cr13");
arch/csky/kernel/cpu-probe.c
27
if (cur == next)
arch/csky/kernel/cpu-probe.c
30
cur = next;
arch/csky/kernel/cpu-probe.c
33
if (!(next >> 28)) {
arch/csky/kernel/cpu-probe.c
76
.next = c_next,
arch/csky/lib/string.c
46
unsigned long last, next;
arch/csky/lib/string.c
56
next = s.as_ulong[0];
arch/csky/lib/string.c
58
last = next;
arch/csky/lib/string.c
59
next = s.as_ulong[1];
arch/csky/lib/string.c
62
next << ((BYTES_LONG - distance) * 8);
arch/hexagon/include/asm/mmu_context.h
29
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/hexagon/include/asm/mmu_context.h
38
if (next->context.generation < prev->context.generation) {
arch/hexagon/include/asm/mmu_context.h
40
next->pgd[l1] = init_mm.pgd[l1];
arch/hexagon/include/asm/mmu_context.h
42
next->context.generation = prev->context.generation;
arch/hexagon/include/asm/mmu_context.h
45
__vmnewmap((void *)next->context.ptbase);
arch/hexagon/include/asm/mmu_context.h
52
static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
arch/hexagon/include/asm/mmu_context.h
57
switch_mm(prev, next, current_thread_info()->task);
arch/hexagon/kernel/setup.c
134
.next = &c_next,
arch/loongarch/include/asm/hw_breakpoint.h
124
extern void hw_breakpoint_thread_switch(struct task_struct *next);
arch/loongarch/include/asm/hw_breakpoint.h
129
static inline void hw_breakpoint_thread_switch(struct task_struct *next)
arch/loongarch/include/asm/mmu_context.h
101
atomic_update_pgd_asid(cpu_asid(cpu, next), (unsigned long)invalid_pg_dir);
arch/loongarch/include/asm/mmu_context.h
110
cpumask_set_cpu(cpu, mm_cpumask(next));
arch/loongarch/include/asm/mmu_context.h
115
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/loongarch/include/asm/mmu_context.h
121
switch_mm_irqs_off(prev, next, tsk);
arch/loongarch/include/asm/mmu_context.h
133
#define activate_mm(prev, next) switch_mm(prev, next, current)
arch/loongarch/include/asm/mmu_context.h
88
static inline void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
arch/loongarch/include/asm/mmu_context.h
95
if (!asid_valid(next, cpu))
arch/loongarch/include/asm/mmu_context.h
96
get_new_mmu_context(next, cpu, &need_flush);
arch/loongarch/include/asm/mmu_context.h
98
if (next != &init_mm)
arch/loongarch/include/asm/mmu_context.h
99
atomic_update_pgd_asid(cpu_asid(cpu, next), (unsigned long)next->pgd);
arch/loongarch/include/asm/switch_to.h
26
struct task_struct *next, struct thread_info *next_ti,
arch/loongarch/include/asm/switch_to.h
35
#define switch_to(prev, next, last) \
arch/loongarch/include/asm/switch_to.h
39
hw_breakpoint_thread_switch(next); \
arch/loongarch/include/asm/switch_to.h
40
(last) = __switch_to(prev, next, task_thread_info(next), \
arch/loongarch/kernel/asm-offsets.c
281
OFFSET(PBE_NEXT, pbe, next);
arch/loongarch/kernel/hw_breakpoint.c
545
void hw_breakpoint_thread_switch(struct task_struct *next)
arch/loongarch/kernel/hw_breakpoint.c
548
struct pt_regs *regs = task_pt_regs(next);
arch/loongarch/kernel/hw_breakpoint.c
550
if (test_tsk_thread_flag(next, TIF_SINGLESTEP)) {
arch/loongarch/kernel/proc.c
127
.next = c_next,
arch/loongarch/kvm/mmu.c
187
phys_addr_t next, start, size;
arch/loongarch/kvm/mmu.c
196
next = addr + (0x1UL << ctx->pgtable_shift);
arch/loongarch/kvm/mmu.c
201
} while (entry++, addr = next, addr < end);
arch/loongarch/kvm/mmu.c
221
phys_addr_t next, start, size;
arch/loongarch/kvm/mmu.c
230
next = kvm_pgtable_addr_end(ctx, addr, end);
arch/loongarch/kvm/mmu.c
241
ret |= kvm_ptw_leaf(entry, addr, next, ctx);
arch/loongarch/kvm/mmu.c
243
ret |= kvm_ptw_dir(entry, addr, next, ctx);
arch/loongarch/kvm/mmu.c
245
} while (entry++, addr = next, addr < end);
arch/loongarch/kvm/mmu.c
265
phys_addr_t next;
arch/loongarch/kvm/mmu.c
271
next = kvm_pgtable_addr_end(ctx, addr, end);
arch/loongarch/kvm/mmu.c
276
ret |= kvm_ptw_dir(entry, addr, next, ctx);
arch/loongarch/kvm/mmu.c
278
} while (entry++, addr = next, addr < end);
arch/loongarch/mm/init.c
107
unsigned long addr, unsigned long next)
arch/loongarch/mm/init.c
117
unsigned long addr, unsigned long next)
arch/loongarch/mm/init.c
122
vmemmap_verify((pte_t *)pmd, node, addr, next);
arch/loongarch/mm/kasan_init.c
175
unsigned long next;
arch/loongarch/mm/kasan_init.c
182
next = addr + PAGE_SIZE;
arch/loongarch/mm/kasan_init.c
184
} while (ptep++, addr = next, addr != end && __pte_none(early, ptep_get(ptep)));
arch/loongarch/mm/kasan_init.c
190
unsigned long next;
arch/loongarch/mm/kasan_init.c
194
next = pmd_addr_end(addr, end);
arch/loongarch/mm/kasan_init.c
195
kasan_pte_populate(pmdp, addr, next, node, early);
arch/loongarch/mm/kasan_init.c
196
} while (pmdp++, addr = next, addr != end && __pmd_none(early, pmdp_get(pmdp)));
arch/loongarch/mm/kasan_init.c
202
unsigned long next;
arch/loongarch/mm/kasan_init.c
206
next = pud_addr_end(addr, end);
arch/loongarch/mm/kasan_init.c
207
kasan_pmd_populate(pudp, addr, next, node, early);
arch/loongarch/mm/kasan_init.c
208
} while (pudp++, addr = next, addr != end && __pud_none(early, READ_ONCE(*pudp)));
arch/loongarch/mm/kasan_init.c
214
unsigned long next;
arch/loongarch/mm/kasan_init.c
218
next = p4d_addr_end(addr, end);
arch/loongarch/mm/kasan_init.c
219
kasan_pud_populate(p4dp, addr, next, node, early);
arch/loongarch/mm/kasan_init.c
220
} while (p4dp++, addr = next, addr != end && __p4d_none(early, READ_ONCE(*p4dp)));
arch/loongarch/mm/kasan_init.c
226
unsigned long next;
arch/loongarch/mm/kasan_init.c
232
next = pgd_addr_end(addr, end);
arch/loongarch/mm/kasan_init.c
233
kasan_p4d_populate(pgdp, addr, next, node, early);
arch/loongarch/mm/kasan_init.c
234
} while (pgdp++, addr = next, addr != end);
arch/loongarch/mm/pageattr.c
29
unsigned long next, struct mm_walk *walk)
arch/loongarch/mm/pageattr.c
42
unsigned long next, struct mm_walk *walk)
arch/loongarch/mm/pageattr.c
55
unsigned long next, struct mm_walk *walk)
arch/loongarch/mm/pageattr.c
68
unsigned long next, struct mm_walk *walk)
arch/loongarch/mm/pageattr.c
81
unsigned long next, struct mm_walk *walk)
arch/loongarch/mm/pageattr.c
91
static int pageattr_pte_hole(unsigned long addr, unsigned long next,
arch/m68k/emu/nfblock.c
183
struct nfhd_device *dev, *next;
arch/m68k/emu/nfblock.c
185
list_for_each_entry_safe(dev, next, &nfhd_list, list) {
arch/m68k/include/asm/dvma.h
117
struct Linux_SBus_DMA *next;
arch/m68k/include/asm/mac_iop.h
140
struct iop_msg *next; /* next message in queue or NULL */
arch/m68k/include/asm/mmu_context.h
207
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/m68k/include/asm/mmu_context.h
291
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk)
arch/m68k/include/asm/mmu_context.h
293
if (prev != next) {
arch/m68k/include/asm/mmu_context.h
295
switch_mm_0230(next);
arch/m68k/include/asm/mmu_context.h
297
switch_mm_0460(next);
arch/m68k/include/asm/mmu_context.h
72
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/m68k/include/asm/mmu_context.h
76
set_context(tsk->mm->context, next->pgd);
arch/m68k/include/asm/mmu_context.h
91
#define prepare_arch_switch(next) load_ksp_mmu(next)
arch/m68k/include/asm/switch_to.h
31
#define switch_to(prev,next,last) do { \
arch/m68k/include/asm/switch_to.h
33
register void *_next __asm__ ("a1") = (next); \
arch/m68k/include/asm/tlbflush.h
199
goto next;
arch/m68k/include/asm/tlbflush.h
206
next:
arch/m68k/kernel/setup_mm.c
472
.next = c_next,
arch/m68k/kernel/setup_no.c
215
.next = c_next,
arch/m68k/mac/iop.c
398
msg = msg->next;
arch/m68k/mac/iop.c
463
msg->next = NULL;
arch/m68k/mac/iop.c
475
while (q->next) q = q->next;
arch/m68k/mac/iop.c
476
q->next = msg;
arch/m68k/mm/kmap.c
117
for (p = &iolist; (tmp = *p) ; p = &tmp->next) {
arch/m68k/mm/kmap.c
128
area->next = *p;
arch/m68k/mm/kmap.c
140
for (p = &iolist ; (tmp = *p) ; p = &tmp->next) {
arch/m68k/mm/kmap.c
142
*p = tmp->next;
arch/m68k/mm/motorola.c
144
ptable_desc *dp = ptable_list[type].next;
arch/m68k/mm/motorola.c
219
} else if (ptable_list[type].next != dp) {
arch/m68k/sun3/sun3dvma.c
132
ret = list_entry(hole_cache.next, struct hole, list);
arch/microblaze/include/asm/mmu_context_mm.h
114
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/microblaze/include/asm/mmu_context_mm.h
117
tsk->thread.pgdir = next->pgd;
arch/microblaze/include/asm/mmu_context_mm.h
118
get_mmu_context(next);
arch/microblaze/include/asm/mmu_context_mm.h
119
set_context(next->context, next->pgd);
arch/microblaze/include/asm/switch_to.h
13
struct thread_info *next);
arch/microblaze/include/asm/switch_to.h
15
#define switch_to(prev, next, last) \
arch/microblaze/include/asm/switch_to.h
18
task_thread_info(next)); \
arch/microblaze/kernel/cpu/mb.c
155
.next = c_next,
arch/mips/alchemy/common/clock.c
1008
goto next;
arch/mips/alchemy/common/clock.c
1032
next:
arch/mips/bcm63xx/irq.c
102
unsigned int to_call = *next; \
arch/mips/bcm63xx/irq.c
104
*next = (*next + 1) & (width - 1); \
arch/mips/bcm63xx/irq.c
81
unsigned int *next = &i[cpu]; \
arch/mips/cavium-octeon/executive/cvmx-bootmem.c
103
static void cvmx_bootmem_phy_set_next(uint64_t addr, uint64_t next)
arch/mips/cavium-octeon/executive/cvmx-bootmem.c
105
cvmx_write64_uint64((addr + NEXT_OFFSET) | (1ull << 63), next);
arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h
114
# This is the variable where the next core to boot is stored
arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h
116
# Get the core id of the next to be booted
arch/mips/include/asm/mmu_context.h
155
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/mips/include/asm/mmu_context.h
163
check_switch_mmu_context(next);
arch/mips/include/asm/mmu_context.h
170
cpumask_set_cpu(cpu, mm_cpumask(next));
arch/mips/include/asm/switch_to.h
100
# define __sanitize_fcr31(next) do { (void) (next); } while (0)
arch/mips/include/asm/switch_to.h
109
#define switch_to(prev, next, last) \
arch/mips/include/asm/switch_to.h
113
if (tsk_used_math(next)) \
arch/mips/include/asm/switch_to.h
114
__sanitize_fcr31(next); \
arch/mips/include/asm/switch_to.h
117
__restore_dsp(next); \
arch/mips/include/asm/switch_to.h
128
if (KSTK_STATUS(next) & ST0_CU2 && \
arch/mips/include/asm/switch_to.h
130
cop2_restore(next); \
arch/mips/include/asm/switch_to.h
137
write_c0_userlocal(task_thread_info(next)->tp_value); \
arch/mips/include/asm/switch_to.h
138
__restore_watch(next); \
arch/mips/include/asm/switch_to.h
139
(last) = resume(prev, next, task_thread_info(next)); \
arch/mips/include/asm/switch_to.h
33
struct task_struct *next, struct thread_info *next_ti);
arch/mips/include/asm/switch_to.h
62
next->thread.emulated_fp = 0; \
arch/mips/include/asm/switch_to.h
88
# define __sanitize_fcr31(next) \
arch/mips/include/asm/switch_to.h
90
unsigned long fcr31 = mask_fcr31_x(next->thread.fpu.fcr31); \
arch/mips/include/asm/switch_to.h
94
pc = (void __user *)task_pt_regs(next)->cp0_epc; \
arch/mips/include/asm/switch_to.h
95
next->thread.fpu.fcr31 &= ~fcr31; \
arch/mips/include/asm/switch_to.h
96
force_fcr31_sig(fcr31, pc, next); \
arch/mips/kernel/asm-offsets.c
343
OFFSET(PBE_NEXT, pbe, next);
arch/mips/kernel/module.c
116
struct mips_hi16 *next;
arch/mips/kernel/module.c
144
next = l->next;
arch/mips/kernel/module.c
146
l = next;
arch/mips/kernel/module.c
25
struct mips_hi16 *next;
arch/mips/kernel/module.c
81
n->next = me->arch.r_mips_hi16_list;
arch/mips/kernel/module.c
89
struct mips_hi16 *next;
arch/mips/kernel/module.c
92
next = l->next;
arch/mips/kernel/module.c
94
l = next;
arch/mips/kernel/proc.c
328
.next = c_next,
arch/mips/kernel/vpe-mt.c
213
t = list_entry(v->tc.next, struct tc, tc);
arch/mips/kernel/vpe-mt.c
232
t = list_entry(v->tc.next, struct tc, tc);
arch/mips/kernel/vpe.c
220
struct mips_hi16 *next;
arch/mips/kernel/vpe.c
327
n->next = mips_hi16_list;
arch/mips/kernel/vpe.c
338
struct mips_hi16 *l, *next;
arch/mips/kernel/vpe.c
376
next = l->next;
arch/mips/kernel/vpe.c
378
l = next;
arch/mips/kernel/vpe.c
395
next = l->next;
arch/mips/kernel/vpe.c
397
l = next;
arch/mips/sibyte/common/sb_tbprof.c
145
u64 next = (1ULL << 40) - tb_period;
arch/mips/sibyte/common/sb_tbprof.c
178
__raw_writeq(next, IOADDR(A_SCD_PERF_CNT_1));
arch/nios2/include/asm/mmu_context.h
43
void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/nios2/include/asm/mmu_context.h
51
void activate_mm(struct mm_struct *prev, struct mm_struct *next);
arch/nios2/include/asm/switch_to.h
17
#define switch_to(prev, next, last) \
arch/nios2/include/asm/switch_to.h
26
: "r" (prev), "r" (next) \
arch/nios2/kernel/cpuinfo.c
187
.next = cpuinfo_next,
arch/nios2/mm/mmu_context.c
106
void activate_mm(struct mm_struct *prev, struct mm_struct *next)
arch/nios2/mm/mmu_context.c
108
next->context = get_new_context();
arch/nios2/mm/mmu_context.c
109
set_context(next->context);
arch/nios2/mm/mmu_context.c
110
pgd_current = next->pgd;
arch/nios2/mm/mmu_context.c
80
void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/nios2/mm/mmu_context.c
89
if (unlikely(CTX_VERSION(next->context) !=
arch/nios2/mm/mmu_context.c
91
next->context = get_new_context();
arch/nios2/mm/mmu_context.c
94
pgd_current = next->pgd;
arch/nios2/mm/mmu_context.c
97
set_context(next->context);
arch/openrisc/include/asm/mmu_context.h
24
extern void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/openrisc/include/asm/mmu_context.h
27
#define activate_mm(prev, next) switch_mm((prev), (next), NULL)
arch/openrisc/kernel/dma.c
26
unsigned long next, struct mm_walk *walk)
arch/openrisc/kernel/dma.c
37
local_dcache_range_flush(__pa(addr), __pa(next));
arch/openrisc/kernel/dma.c
48
unsigned long next, struct mm_walk *walk)
arch/openrisc/kernel/setup.c
337
.next = c_next,
arch/openrisc/mm/tlb.c
137
void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/openrisc/mm/tlb.c
142
if (unlikely(prev == next))
arch/openrisc/mm/tlb.c
148
cpumask_set_cpu(cpu, mm_cpumask(next));
arch/openrisc/mm/tlb.c
156
current_pgd[cpu] = next->pgd;
arch/parisc/include/asm/mmu_context.h
51
struct mm_struct *next, struct task_struct *tsk)
arch/parisc/include/asm/mmu_context.h
53
if (prev != next) {
arch/parisc/include/asm/mmu_context.h
57
spinlock_t *pgd_lock = &next->page_table_lock;
arch/parisc/include/asm/mmu_context.h
60
mtctl(__pa(next->pgd), 25);
arch/parisc/include/asm/mmu_context.h
61
load_context(next->context);
arch/parisc/include/asm/mmu_context.h
66
struct mm_struct *next, struct task_struct *tsk)
arch/parisc/include/asm/mmu_context.h
70
if (prev == next)
arch/parisc/include/asm/mmu_context.h
74
switch_mm_irqs_off(prev, next, tsk);
arch/parisc/include/asm/mmu_context.h
80
static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
arch/parisc/include/asm/mmu_context.h
90
BUG_ON(next == &init_mm); /* Should never happen */
arch/parisc/include/asm/mmu_context.h
92
if (next->context.space_id == 0)
arch/parisc/include/asm/mmu_context.h
93
next->context.space_id = alloc_sid();
arch/parisc/include/asm/mmu_context.h
95
switch_mm(prev,next,current);
arch/parisc/include/asm/parisc-device.h
34
struct parisc_driver *next;
arch/parisc/include/asm/ropes.h
74
struct sba_device *next; /* list of SBA's in system */
arch/parisc/include/asm/switch_to.h
10
(last) = _switch_to(prev, next); \
arch/parisc/include/asm/switch_to.h
9
#define switch_to(prev, next, last) do { \
arch/parisc/kernel/irq.c
207
while ((action = action->next))
arch/parisc/kernel/irq.c
210
for ( ;action; action = action->next) {
arch/parisc/kernel/setup.c
170
.next = c_next,
arch/powerpc/include/asm/bootx.h
29
u32 next;
arch/powerpc/include/asm/bootx.h
45
u32 next;
arch/powerpc/include/asm/iommu.h
192
struct list_head next;
arch/powerpc/include/asm/kvm_book3s_64.h
32
struct kvm_nested_guest *next;
arch/powerpc/include/asm/kvm_book3s_64.h
95
(pos)->list.next)), true); \
arch/powerpc/include/asm/kvm_host.h
182
struct list_head next;
arch/powerpc/include/asm/membarrier.h
16
likely(!(atomic_read(&next->membarrier_state) &
arch/powerpc/include/asm/membarrier.h
5
struct mm_struct *next,
arch/powerpc/include/asm/mmu_context.h
100
extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next,
arch/powerpc/include/asm/mmu_context.h
223
extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
arch/powerpc/include/asm/mmu_context.h
226
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/powerpc/include/asm/mmu_context.h
232
switch_mm_irqs_off(prev, next, tsk);
arch/powerpc/include/asm/mmu_context.h
242
static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
arch/powerpc/include/asm/mmu_context.h
244
switch_mm_irqs_off(prev, next, current);
arch/powerpc/include/asm/mmu_context.h
56
struct mm_struct *next);
arch/powerpc/include/asm/mmu_context.h
58
struct mm_struct *next,
arch/powerpc/include/asm/mmu_context.h
62
return radix__switch_mmu_context(prev, next);
arch/powerpc/include/asm/mmu_context.h
63
return switch_slb(tsk, next);
arch/powerpc/include/asm/mpic.h
332
struct mpic *next;
arch/powerpc/include/asm/opal-api.h
985
__be64 next;
arch/powerpc/include/asm/switch_to.h
17
#define switch_to(prev, next, last) ((last) = __switch_to((prev), (next)))
arch/powerpc/include/asm/switch_to.h
20
struct thread_struct *next);
arch/powerpc/kernel/asm-offsets.c
330
OFFSET(pbe_next, pbe, next);
arch/powerpc/kernel/cacheinfo.c
196
struct cache *next = cache;
arch/powerpc/kernel/cacheinfo.c
198
while (next) {
arch/powerpc/kernel/cacheinfo.c
199
WARN_ONCE(cpumask_test_cpu(cpu, &next->shared_cpu_map),
arch/powerpc/kernel/cacheinfo.c
201
cpu, next->ofnode,
arch/powerpc/kernel/cacheinfo.c
202
cache_type_string(next));
arch/powerpc/kernel/cacheinfo.c
203
cpumask_set_cpu(cpu, &next->shared_cpu_map);
arch/powerpc/kernel/cacheinfo.c
204
next = next->next_local;
arch/powerpc/kernel/cacheinfo.c
42
struct cache_index_dir *next; /* next index in parent directory */
arch/powerpc/kernel/cacheinfo.c
804
index_dir->next = cache_dir->index;
arch/powerpc/kernel/cacheinfo.c
870
struct cache_index_dir *next;
arch/powerpc/kernel/cacheinfo.c
872
next = index->next;
arch/powerpc/kernel/cacheinfo.c
874
index = next;
arch/powerpc/kernel/cacheinfo.c
893
struct cache *next = cache->next_local;
arch/powerpc/kernel/cacheinfo.c
907
cache = next;
arch/powerpc/kernel/eeh_event.c
52
event = list_entry(eeh_eventlist.next,
arch/powerpc/kernel/eeh_pe.c
185
struct list_head *next = pe->child_list.next;
arch/powerpc/kernel/eeh_pe.c
187
if (next == &pe->child_list) {
arch/powerpc/kernel/eeh_pe.c
191
next = pe->child.next;
arch/powerpc/kernel/eeh_pe.c
192
if (next != &pe->parent->child_list)
arch/powerpc/kernel/eeh_pe.c
198
return list_entry(next, struct eeh_pe, child);
arch/powerpc/kernel/rtas_flash.c
109
struct flash_block_list *next;
arch/powerpc/kernel/rtas_flash.c
171
for (f = flist; f; f = f->next) {
arch/powerpc/kernel/rtas_flash.c
196
struct flash_block_list *next;
arch/powerpc/kernel/rtas_flash.c
202
next = f->next;
arch/powerpc/kernel/rtas_flash.c
204
f = next;
arch/powerpc/kernel/rtas_flash.c
334
while (fl->next)
arch/powerpc/kernel/rtas_flash.c
335
fl = fl->next; /* seek to last block_list for append */
arch/powerpc/kernel/rtas_flash.c
339
fl->next = kmem_cache_zalloc(flash_block_cache, GFP_KERNEL);
arch/powerpc/kernel/rtas_flash.c
340
if (!fl->next)
arch/powerpc/kernel/rtas_flash.c
342
fl = fl->next;
arch/powerpc/kernel/rtas_flash.c
536
struct flash_block_list *f, *next, *flist;
arch/powerpc/kernel/rtas_flash.c
571
flist->next = rtas_firmware_flash_list;
arch/powerpc/kernel/rtas_flash.c
583
for (f = flist; f; f = next) {
arch/powerpc/kernel/rtas_flash.c
590
next = f->next;
arch/powerpc/kernel/rtas_flash.c
592
if (f->next)
arch/powerpc/kernel/rtas_flash.c
593
f->next = (struct flash_block_list *)cpu_to_be64(__pa(f->next));
arch/powerpc/kernel/rtas_flash.c
595
f->next = NULL;
arch/powerpc/kernel/setup-common.c
355
.next = c_next,
arch/powerpc/kvm/book3s_64_vio.c
162
list_for_each_entry_rcu(stit, &stt->iommu_tables, next) {
arch/powerpc/kvm/book3s_64_vio.c
190
list_add_rcu(&stit->next, &stt->iommu_tables);
arch/powerpc/kvm/book3s_64_vio.c
266
list_for_each_entry_safe(stit, tmp, &stt->iommu_tables, next) {
arch/powerpc/kvm/book3s_64_vio.c
387
list_for_each_entry_rcu(stit, &stt->iommu_tables, next) {
arch/powerpc/kvm/book3s_64_vio.c
605
list_for_each_entry_lockless(stit, &stt->iommu_tables, next) {
arch/powerpc/kvm/book3s_64_vio.c
700
list_for_each_entry_lockless(stit, &stt->iommu_tables, next) {
arch/powerpc/kvm/book3s_64_vio.c
742
list_for_each_entry_lockless(stit, &stt->iommu_tables, next) {
arch/powerpc/kvm/book3s_64_vio.c
76
list_del_rcu(&stit->next);
arch/powerpc/kvm/book3s_64_vio.c
96
list_for_each_entry_safe(stit, tmp, &stt->iommu_tables, next) {
arch/powerpc/kvm/book3s_hv.h
47
void accumulate_time(struct kvm_vcpu *vcpu, struct kvmhv_tb_accumulator *next);
arch/powerpc/kvm/book3s_hv.h
48
#define start_timing(vcpu, next) accumulate_time(vcpu, next)
arch/powerpc/kvm/book3s_hv.h
51
#define accumulate_time(vcpu, next) do {} while (0)
arch/powerpc/kvm/book3s_hv.h
52
#define start_timing(vcpu, next) do {} while (0)
arch/powerpc/kvm/book3s_hv_nested.c
812
gp->next = freelist;
arch/powerpc/kvm/book3s_hv_nested.c
820
freelist = gp->next;
arch/powerpc/kvm/book3s_hv_nested.c
948
(*n_rmap)->list.next = (struct llist_node *) rmap;
arch/powerpc/kvm/book3s_hv_p9_entry.c
220
void accumulate_time(struct kvm_vcpu *vcpu, struct kvmhv_tb_accumulator *next)
arch/powerpc/kvm/book3s_hv_p9_entry.c
230
vcpu->arch.cur_activity = next;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
151
struct revmap_entry *next, *prev;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
166
next = real_vmalloc_addr(&kvm->arch.hpt.rev[rev->forw]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
168
next->back = rev->back;
arch/powerpc/kvm/book3s_hv_uvmem.c
275
struct kvmppc_uvmem_slot *p, *next;
arch/powerpc/kvm/book3s_hv_uvmem.c
278
list_for_each_entry_safe(p, next, &kvm->arch.uvmem_pfns, list) {
arch/powerpc/kvm/book3s_xive.c
422
goto next;
arch/powerpc/kvm/book3s_xive.c
425
goto next;
arch/powerpc/kvm/book3s_xive.c
430
goto next;
arch/powerpc/kvm/book3s_xive.c
448
next:
arch/powerpc/kvm/mpic.c
1066
__func__, idx, dst->raised.next);
arch/powerpc/kvm/mpic.c
142
int next;
arch/powerpc/kvm/mpic.c
294
int next = -1;
arch/powerpc/kvm/mpic.c
306
next = irq;
arch/powerpc/kvm/mpic.c
311
q->next = next;
arch/powerpc/kvm/mpic.c
320
return q->next;
arch/powerpc/kvm/mpic.c
386
__func__, n_IRQ, dst->servicing.next, n_CPU);
arch/powerpc/kvm/mpic.c
389
__func__, n_CPU, n_IRQ, dst->raised.next);
arch/powerpc/kvm/mpic.c
397
__func__, n_IRQ, dst->raised.next,
arch/powerpc/kvm/mpic.c
548
opp->dst[i].raised.next = -1;
arch/powerpc/kvm/mpic.c
550
opp->dst[i].servicing.next = -1;
arch/powerpc/lib/qspinlock.c
17
struct qnode *next;
arch/powerpc/lib/qspinlock.c
356
struct qnode *next;
arch/powerpc/lib/qspinlock.c
364
next = READ_ONCE(node->next);
arch/powerpc/lib/qspinlock.c
365
if (!next)
arch/powerpc/lib/qspinlock.c
368
if (next->sleepy)
arch/powerpc/lib/qspinlock.c
373
next->sleepy = 1;
arch/powerpc/lib/qspinlock.c
396
if (node->next && !node->next->sleepy) {
arch/powerpc/lib/qspinlock.c
405
node->next->sleepy = 1;
arch/powerpc/lib/qspinlock.c
530
struct qnode *next, *node;
arch/powerpc/lib/qspinlock.c
556
node->next = NULL;
arch/powerpc/lib/qspinlock.c
580
WRITE_ONCE(prev->next, node);
arch/powerpc/lib/qspinlock.c
601
next = READ_ONCE(node->next);
arch/powerpc/lib/qspinlock.c
602
if (next)
arch/powerpc/lib/qspinlock.c
603
prefetchw(next);
arch/powerpc/lib/qspinlock.c
671
next = READ_ONCE(node->next);
arch/powerpc/lib/qspinlock.c
672
if (!next) {
arch/powerpc/lib/qspinlock.c
674
while (!(next = READ_ONCE(node->next)))
arch/powerpc/lib/qspinlock.c
688
int next_cpu = next->cpu;
arch/powerpc/lib/qspinlock.c
689
WRITE_ONCE(next->locked, 1);
arch/powerpc/lib/qspinlock.c
695
WRITE_ONCE(next->locked, 1);
arch/powerpc/lib/rheap.c
133
blk = list_entry(info->empty_list.next, rh_block_t, list);
arch/powerpc/lib/rheap.c
156
rh_block_t *next;
arch/powerpc/lib/rheap.c
170
next = NULL;
arch/powerpc/lib/rheap.c
178
if (next == NULL && s >= bs)
arch/powerpc/lib/rheap.c
179
next = blk;
arch/powerpc/lib/rheap.c
202
if (next != NULL)
arch/powerpc/lib/rheap.c
203
list_add(&blkn->list, &next->list);
arch/powerpc/lib/rheap.c
35
pp = (unsigned long *)&l->next;
arch/powerpc/mm/book3s32/mmu_context.c
114
void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk)
arch/powerpc/mm/book3s32/mmu_context.c
116
long id = next->context.id;
arch/powerpc/mm/book3s32/mmu_context.c
119
panic("mm_struct %p has no context ID", next);
arch/powerpc/mm/book3s32/mmu_context.c
123
update_user_segments(next->context.sr0);
arch/powerpc/mm/book3s32/mmu_context.c
126
abatron_pteptrs[1] = next->pgd;
arch/powerpc/mm/book3s32/mmu_context.c
129
mtspr(SPRN_SDR1, rol32(__pa(next->pgd), 4) & 0xffff01ff);
arch/powerpc/mm/book3s64/iommu_api.c
132
list_for_each_entry_rcu(mem2, &mm->context.iommu_group_mem_list, next,
arch/powerpc/mm/book3s64/iommu_api.c
165
list_add_rcu(&mem->next, &mm->context.iommu_group_mem_list);
arch/powerpc/mm/book3s64/iommu_api.c
245
list_del_rcu(&mem->next);
arch/powerpc/mm/book3s64/iommu_api.c
28
struct list_head next;
arch/powerpc/mm/book3s64/iommu_api.c
294
list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) {
arch/powerpc/mm/book3s64/iommu_api.c
315
list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next,
arch/powerpc/mm/book3s64/iommu_api.c
361
list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) {
arch/powerpc/mm/book3s64/mmu_context.c
321
void radix__switch_mmu_context(struct mm_struct *prev, struct mm_struct *next)
arch/powerpc/mm/book3s64/mmu_context.c
323
mtspr(SPRN_PID, next->context.id);
arch/powerpc/mm/book3s64/radix_pgtable.c
1002
unsigned long addr, unsigned long next)
arch/powerpc/mm/book3s64/radix_pgtable.c
1012
vmemmap_verify(ptep, node, addr, next);
arch/powerpc/mm/book3s64/radix_pgtable.c
1118
unsigned long next;
arch/powerpc/mm/book3s64/radix_pgtable.c
1145
for (addr = start; addr < end; addr = next) {
arch/powerpc/mm/book3s64/radix_pgtable.c
1146
next = pmd_addr_end(addr, end);
arch/powerpc/mm/book3s64/radix_pgtable.c
1181
vmemmap_set_pmd(pmd, p, node, addr, next);
arch/powerpc/mm/book3s64/radix_pgtable.c
1193
} else if (vmemmap_check_pmd(pmd, node, addr, next)) {
arch/powerpc/mm/book3s64/radix_pgtable.c
1215
next = addr + PAGE_SIZE;
arch/powerpc/mm/book3s64/radix_pgtable.c
1318
unsigned long next;
arch/powerpc/mm/book3s64/radix_pgtable.c
1325
for (addr = start; addr < end; addr = next) {
arch/powerpc/mm/book3s64/radix_pgtable.c
1339
next = pmd_addr_end(addr, end);
arch/powerpc/mm/book3s64/radix_pgtable.c
1352
next = addr + PAGE_SIZE;
arch/powerpc/mm/book3s64/radix_pgtable.c
1380
next = addr + 2 * PAGE_SIZE;
arch/powerpc/mm/book3s64/radix_pgtable.c
1396
next = addr + PAGE_SIZE;
arch/powerpc/mm/book3s64/radix_pgtable.c
1406
next = addr + PAGE_SIZE;
arch/powerpc/mm/book3s64/radix_pgtable.c
1520
pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next;
arch/powerpc/mm/book3s64/radix_pgtable.c
791
unsigned long next, pages = 0;
arch/powerpc/mm/book3s64/radix_pgtable.c
795
for (; addr < end; addr = next, pte++) {
arch/powerpc/mm/book3s64/radix_pgtable.c
796
next = (addr + PAGE_SIZE) & PAGE_MASK;
arch/powerpc/mm/book3s64/radix_pgtable.c
797
if (next > end)
arch/powerpc/mm/book3s64/radix_pgtable.c
798
next = end;
arch/powerpc/mm/book3s64/radix_pgtable.c
803
if (PAGE_ALIGNED(addr) && PAGE_ALIGNED(next)) {
arch/powerpc/mm/book3s64/radix_pgtable.c
810
else if (!direct && vmemmap_page_is_unused(addr, next)) {
arch/powerpc/mm/book3s64/radix_pgtable.c
824
unsigned long next, pages = 0;
arch/powerpc/mm/book3s64/radix_pgtable.c
829
for (; addr < end; addr = next, pmd++) {
arch/powerpc/mm/book3s64/radix_pgtable.c
830
next = pmd_addr_end(addr, end);
arch/powerpc/mm/book3s64/radix_pgtable.c
837
IS_ALIGNED(next, PMD_SIZE)) {
arch/powerpc/mm/book3s64/radix_pgtable.c
844
else if (!direct && vmemmap_pmd_is_unused(addr, next)) {
arch/powerpc/mm/book3s64/radix_pgtable.c
853
remove_pte_table(pte_base, addr, next, direct, altmap);
arch/powerpc/mm/book3s64/radix_pgtable.c
864
unsigned long next, pages = 0;
arch/powerpc/mm/book3s64/radix_pgtable.c
869
for (; addr < end; addr = next, pud++) {
arch/powerpc/mm/book3s64/radix_pgtable.c
870
next = pud_addr_end(addr, end);
arch/powerpc/mm/book3s64/radix_pgtable.c
877
!IS_ALIGNED(next, PUD_SIZE)) {
arch/powerpc/mm/book3s64/radix_pgtable.c
887
remove_pmd_table(pmd_base, addr, next, direct, altmap);
arch/powerpc/mm/book3s64/radix_pgtable.c
898
unsigned long addr, next;
arch/powerpc/mm/book3s64/radix_pgtable.c
905
for (addr = start; addr < end; addr = next) {
arch/powerpc/mm/book3s64/radix_pgtable.c
906
next = pgd_addr_end(addr, end);
arch/powerpc/mm/book3s64/radix_pgtable.c
915
!IS_ALIGNED(next, P4D_SIZE)) {
arch/powerpc/mm/book3s64/radix_pgtable.c
925
remove_pud_table(pud_base, addr, next, direct, altmap);
arch/powerpc/mm/book3s64/radix_pgtable.c
991
unsigned long addr, unsigned long next)
arch/powerpc/mm/book3s64/radix_pgtable.c
996
vmemmap_verify(pmdp_ptep(pmdp), node, addr, next);
arch/powerpc/mm/book3s64/subpage_prot.c
108
for (; addr < limit; addr = next) {
arch/powerpc/mm/book3s64/subpage_prot.c
109
next = pmd_addr_end(addr, limit);
arch/powerpc/mm/book3s64/subpage_prot.c
124
if (addr + (nw << PAGE_SHIFT) > next)
arch/powerpc/mm/book3s64/subpage_prot.c
125
nw = (next - addr) >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/subpage_prot.c
192
unsigned long next, limit;
arch/powerpc/mm/book3s64/subpage_prot.c
233
for (limit = addr + len; addr < limit; addr = next) {
arch/powerpc/mm/book3s64/subpage_prot.c
234
next = pmd_addr_end(addr, limit);
arch/powerpc/mm/book3s64/subpage_prot.c
263
if (addr + (nw << PAGE_SHIFT) > next)
arch/powerpc/mm/book3s64/subpage_prot.c
264
nw = (next - addr) >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/subpage_prot.c
97
unsigned long next, limit;
arch/powerpc/mm/init_64.c
129
static struct vmemmap_backing *next;
arch/powerpc/mm/init_64.c
147
vmem_back = next;
arch/powerpc/mm/init_64.c
148
next = next->list;
arch/powerpc/mm/init_64.c
155
next = vmemmap_alloc_block(PAGE_SIZE, node);
arch/powerpc/mm/init_64.c
156
if (unlikely(!next)) {
arch/powerpc/mm/init_64.c
165
return next++;
arch/powerpc/mm/init_64.c
313
vmem_back->list = next;
arch/powerpc/mm/init_64.c
314
next = vmem_back;
arch/powerpc/mm/kasan/init_32.c
181
unsigned long next;
arch/powerpc/mm/kasan/init_32.c
189
next = pgd_addr_end(addr, end);
arch/powerpc/mm/kasan/init_32.c
191
} while (pmd++, addr = next, addr != end);
arch/powerpc/mm/mmu_context.c
104
switch_mmu_context(prev, next, tsk);
arch/powerpc/mm/mmu_context.c
43
void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
arch/powerpc/mm/mmu_context.c
50
if (!cpumask_test_cpu(cpu, mm_cpumask(next))) {
arch/powerpc/mm/mmu_context.c
51
VM_WARN_ON_ONCE(next == &init_mm);
arch/powerpc/mm/mmu_context.c
52
cpumask_set_cpu(cpu, mm_cpumask(next));
arch/powerpc/mm/mmu_context.c
53
inc_mm_active_cpus(next);
arch/powerpc/mm/mmu_context.c
84
switch_mm_pgdir(tsk, next);
arch/powerpc/mm/mmu_context.c
87
if (prev == next)
arch/powerpc/mm/mmu_context.c
98
membarrier_arch_switch_mm(prev, next, tsk);
arch/powerpc/mm/nohash/e500_hugetlbpage.c
21
int this, next;
arch/powerpc/mm/nohash/e500_hugetlbpage.c
26
next = this + 1;
arch/powerpc/mm/nohash/e500_hugetlbpage.c
27
if (next >= tcd->esel_max)
arch/powerpc/mm/nohash/e500_hugetlbpage.c
28
next = tcd->esel_first;
arch/powerpc/mm/nohash/e500_hugetlbpage.c
30
tcd->esel_next = next;
arch/powerpc/mm/nohash/mmu_context.c
219
void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next,
arch/powerpc/mm/nohash/mmu_context.c
231
next->context.active++;
arch/powerpc/mm/nohash/mmu_context.c
241
id = next->context.id;
arch/powerpc/mm/nohash/mmu_context.c
275
context_mm[id] = next;
arch/powerpc/mm/nohash/mmu_context.c
276
next->context.id = id;
arch/powerpc/mm/nohash/mmu_context.c
284
local_flush_tlb_mm(next);
arch/powerpc/mm/nohash/mmu_context.c
296
abatron_pteptrs[1] = next->pgd;
arch/powerpc/mm/nohash/mmu_context.c
297
set_context(id, next->pgd);
arch/powerpc/perf/hv-24x7.c
1276
struct hv_24x7_result **next)
arch/powerpc/perf/hv-24x7.c
1292
if (next)
arch/powerpc/perf/hv-24x7.c
1293
*next = (struct hv_24x7_result *) res->elements;
arch/powerpc/perf/hv-24x7.c
1334
if (next)
arch/powerpc/perf/hv-24x7.c
1335
*next = element_data - data_offset;
arch/powerpc/platforms/83xx/mpc832x_rdb.c
110
goto next;
arch/powerpc/platforms/83xx/mpc832x_rdb.c
115
next:
arch/powerpc/platforms/cell/spufs/inode.c
283
aff_supp = !list_empty(&(list_entry(cbe_spu_info[0].spus.next,
arch/powerpc/platforms/cell/spufs/inode.c
310
!list_entry(neighbor->aff_list.next, struct spu_context,
arch/powerpc/platforms/cell/spufs/inode.c
361
|| list_entry(neighbor->aff_list.next, struct spu_context,
arch/powerpc/platforms/cell/spufs/sched.c
377
list_for_each_entry_reverse(spu, ref->aff_list.next, aff_list) {
arch/powerpc/platforms/cell/spufs/switch.c
1951
static void restore_lscsa(struct spu_state *next, struct spu *spu)
arch/powerpc/platforms/cell/spufs/switch.c
1959
set_watchdog_timer(next, spu); /* Step 26. */
arch/powerpc/platforms/cell/spufs/switch.c
1960
setup_spu_status_part1(next, spu); /* Step 27. */
arch/powerpc/platforms/cell/spufs/switch.c
1961
setup_spu_status_part2(next, spu); /* Step 28. */
arch/powerpc/platforms/cell/spufs/switch.c
1962
restore_mfc_rag(next, spu); /* Step 29. */
arch/powerpc/platforms/cell/spufs/switch.c
1964
setup_mfc_slbs(next, spu, spu_restore_code, sizeof(spu_restore_code));
arch/powerpc/platforms/cell/spufs/switch.c
1965
set_spu_npc(next, spu); /* Step 31. */
arch/powerpc/platforms/cell/spufs/switch.c
1966
set_signot1(next, spu); /* Step 32. */
arch/powerpc/platforms/cell/spufs/switch.c
1967
set_signot2(next, spu); /* Step 33. */
arch/powerpc/platforms/cell/spufs/switch.c
1968
setup_decr(next, spu); /* Step 34. */
arch/powerpc/platforms/cell/spufs/switch.c
1969
setup_ppu_mb(next, spu); /* Step 35. */
arch/powerpc/platforms/cell/spufs/switch.c
1970
setup_ppuint_mb(next, spu); /* Step 36. */
arch/powerpc/platforms/cell/spufs/switch.c
1971
send_restore_code(next, spu); /* Step 37. */
arch/powerpc/platforms/cell/spufs/switch.c
1972
set_ppu_querymask(next, spu); /* Step 38. */
arch/powerpc/platforms/cell/spufs/switch.c
1973
wait_tag_complete(next, spu); /* Step 39. */
arch/powerpc/platforms/cell/spufs/switch.c
1974
wait_spu_stopped(next, spu); /* Step 40. */
arch/powerpc/platforms/cell/spufs/switch.c
1977
static void restore_csa(struct spu_state *next, struct spu *spu)
arch/powerpc/platforms/cell/spufs/switch.c
1984
restore_spu_privcntl(next, spu); /* Step 41. */
arch/powerpc/platforms/cell/spufs/switch.c
1985
restore_status_part1(next, spu); /* Step 42. */
arch/powerpc/platforms/cell/spufs/switch.c
1986
restore_status_part2(next, spu); /* Step 43. */
arch/powerpc/platforms/cell/spufs/switch.c
1987
restore_ls_16kb(next, spu); /* Step 44. */
arch/powerpc/platforms/cell/spufs/switch.c
1988
wait_tag_complete(next, spu); /* Step 45. */
arch/powerpc/platforms/cell/spufs/switch.c
1989
suspend_mfc(next, spu); /* Step 46. */
arch/powerpc/platforms/cell/spufs/switch.c
1990
wait_suspend_mfc_complete(next, spu); /* Step 47. */
arch/powerpc/platforms/cell/spufs/switch.c
1991
issue_mfc_tlbie(next, spu); /* Step 48. */
arch/powerpc/platforms/cell/spufs/switch.c
1992
clear_interrupts(next, spu); /* Step 49. */
arch/powerpc/platforms/cell/spufs/switch.c
1993
restore_mfc_queues(next, spu); /* Step 50. */
arch/powerpc/platforms/cell/spufs/switch.c
1994
restore_ppu_querymask(next, spu); /* Step 51. */
arch/powerpc/platforms/cell/spufs/switch.c
1995
restore_ppu_querytype(next, spu); /* Step 52. */
arch/powerpc/platforms/cell/spufs/switch.c
1996
restore_mfc_csr_tsq(next, spu); /* Step 53. */
arch/powerpc/platforms/cell/spufs/switch.c
1997
restore_mfc_csr_cmd(next, spu); /* Step 54. */
arch/powerpc/platforms/cell/spufs/switch.c
1998
restore_mfc_csr_ato(next, spu); /* Step 55. */
arch/powerpc/platforms/cell/spufs/switch.c
1999
restore_mfc_tclass_id(next, spu); /* Step 56. */
arch/powerpc/platforms/cell/spufs/switch.c
2000
set_llr_event(next, spu); /* Step 57. */
arch/powerpc/platforms/cell/spufs/switch.c
2001
restore_decr_wrapped(next, spu); /* Step 58. */
arch/powerpc/platforms/cell/spufs/switch.c
2002
restore_ch_part1(next, spu); /* Step 59. */
arch/powerpc/platforms/cell/spufs/switch.c
2003
restore_ch_part2(next, spu); /* Step 60. */
arch/powerpc/platforms/cell/spufs/switch.c
2004
restore_spu_lslr(next, spu); /* Step 61. */
arch/powerpc/platforms/cell/spufs/switch.c
2005
restore_spu_cfg(next, spu); /* Step 62. */
arch/powerpc/platforms/cell/spufs/switch.c
2006
restore_pm_trace(next, spu); /* Step 63. */
arch/powerpc/platforms/cell/spufs/switch.c
2007
restore_spu_npc(next, spu); /* Step 64. */
arch/powerpc/platforms/cell/spufs/switch.c
2008
restore_spu_mb(next, spu); /* Step 65. */
arch/powerpc/platforms/cell/spufs/switch.c
2009
check_ppu_mb_stat(next, spu); /* Step 66. */
arch/powerpc/platforms/cell/spufs/switch.c
2010
check_ppuint_mb_stat(next, spu); /* Step 67. */
arch/powerpc/platforms/cell/spufs/switch.c
2012
restore_mfc_sr1(next, spu); /* Step 69. */
arch/powerpc/platforms/cell/spufs/switch.c
2013
set_int_route(next, spu); /* NEW */
arch/powerpc/platforms/cell/spufs/switch.c
2014
restore_other_spu_access(next, spu); /* Step 70. */
arch/powerpc/platforms/cell/spufs/switch.c
2015
restore_spu_runcntl(next, spu); /* Step 71. */
arch/powerpc/platforms/cell/spufs/switch.c
2016
restore_mfc_cntl(next, spu); /* Step 72. */
arch/powerpc/platforms/cell/spufs/switch.c
2017
enable_user_access(next, spu); /* Step 73. */
arch/powerpc/platforms/cell/spufs/switch.c
2018
reset_switch_active(next, spu); /* Step 74. */
arch/powerpc/platforms/cell/spufs/switch.c
2019
reenable_interrupts(next, spu); /* Step 75. */
arch/powerpc/platforms/cell/spufs/switch.c
2054
static int __do_spu_restore(struct spu_state *next, struct spu *spu)
arch/powerpc/platforms/cell/spufs/switch.c
2069
restore_lscsa(next, spu); /* Steps 24-39. */
arch/powerpc/platforms/cell/spufs/switch.c
2070
rc = check_restore_status(next, spu); /* Step 40. */
arch/powerpc/platforms/cell/spufs/switch.c
2080
restore_csa(next, spu);
arch/powerpc/platforms/powermac/bootx_init.c
105
ppp = &pp->next;
arch/powerpc/platforms/powermac/bootx_init.c
257
goto next;
arch/powerpc/platforms/powermac/bootx_init.c
262
next:
arch/powerpc/platforms/powermac/bootx_init.c
263
ppp = &pp->next;
arch/powerpc/platforms/powermac/bootx_init.c
317
goto next;
arch/powerpc/platforms/powermac/bootx_init.c
320
goto next;
arch/powerpc/platforms/powermac/bootx_init.c
326
next:
arch/powerpc/platforms/powermac/bootx_init.c
327
ppp = &pp->next;
arch/powerpc/platforms/powermac/pci.c
667
unsigned long base, end, next = -1;
arch/powerpc/platforms/powermac/pci.c
683
if (base != next) {
arch/powerpc/platforms/powermac/pci.c
698
next = end + 1;
arch/powerpc/platforms/powermac/pfunc_core.c
798
list_entry(dev->functions.next, typeof(*func), link);
arch/powerpc/platforms/powernv/opal-async.c
242
.next = NULL,
arch/powerpc/platforms/powernv/opal-hmi.c
281
msg_node = list_entry(opal_hmi_evt_list.next,
arch/powerpc/platforms/powernv/opal-hmi.c
362
.next = NULL,
arch/powerpc/platforms/powernv/opal-memory-errors.c
114
.next = NULL,
arch/powerpc/platforms/powernv/opal-memory-errors.c
62
msg_node = list_entry(opal_memory_err_list.next,
arch/powerpc/platforms/powernv/opal-power.c
119
.next = NULL,
arch/powerpc/platforms/powernv/opal-power.c
126
.next = NULL,
arch/powerpc/platforms/powernv/opal-power.c
133
.next = NULL,
arch/powerpc/platforms/powernv/opal-prd.c
377
.next = NULL,
arch/powerpc/platforms/powernv/opal-prd.c
383
.next = NULL,
arch/powerpc/platforms/powernv/opal.c
1152
struct opal_sg_list *next;
arch/powerpc/platforms/powernv/opal.c
1154
next = kzalloc(PAGE_SIZE, GFP_KERNEL);
arch/powerpc/platforms/powernv/opal.c
1155
if (!next)
arch/powerpc/platforms/powernv/opal.c
1161
sg->next = cpu_to_be64(__pa(next));
arch/powerpc/platforms/powernv/opal.c
1162
sg = next;
arch/powerpc/platforms/powernv/opal.c
1182
uint64_t next = be64_to_cpu(sg->next);
arch/powerpc/platforms/powernv/opal.c
1186
if (next)
arch/powerpc/platforms/powernv/opal.c
1187
sg = __va(next);
arch/powerpc/platforms/powernv/pci-ioda-tce.c
384
list_for_each_entry_rcu(tgl, &tbl->it_group_list, next) {
arch/powerpc/platforms/powernv/pci-ioda-tce.c
386
list_del_rcu(&tgl->next);
arch/powerpc/platforms/powernv/pci-ioda-tce.c
425
list_add_rcu(&tgl->next, &tbl->it_group_list);
arch/powerpc/platforms/powernv/pci-ioda.c
1224
list_for_each_entry_lockless(tgl, &tbl->it_group_list, next) {
arch/powerpc/platforms/ps3/mm.c
579
last = list_entry(r->chunk_list.head.next,
arch/powerpc/platforms/ps3/os-area.c
419
next:
arch/powerpc/platforms/ps3/os-area.c
439
goto next;
arch/powerpc/platforms/ps3/os-area.c
442
goto next;
arch/powerpc/platforms/pseries/dlpar.c
106
dn->properties = prop->next;
arch/powerpc/platforms/pseries/dlpar.c
203
last_property->next = property;
arch/powerpc/platforms/pseries/hvCall_inst.c
75
.next = hc_next,
arch/powerpc/platforms/pseries/hvcserver.c
71
element = head->next;
arch/powerpc/platforms/pseries/iommu.c
443
u64 tce_size, num_tce, dma_offset, next;
arch/powerpc/platforms/pseries/iommu.c
449
next = start_pfn << PAGE_SHIFT;
arch/powerpc/platforms/pseries/iommu.c
453
num_tce += next & (tce_size - 1);
arch/powerpc/platforms/pseries/iommu.c
454
next &= ~(tce_size - 1);
arch/powerpc/platforms/pseries/iommu.c
466
dma_offset = next + be64_to_cpu(maprange->dma_base);
arch/powerpc/platforms/pseries/iommu.c
471
next += limit * tce_size;
arch/powerpc/platforms/pseries/iommu.c
482
u64 tce_size, num_tce, dma_offset, next, proto_tce, liobn;
arch/powerpc/platforms/pseries/iommu.c
518
next = start_pfn << PAGE_SHIFT;
arch/powerpc/platforms/pseries/iommu.c
522
num_tce += next & (tce_size - 1);
arch/powerpc/platforms/pseries/iommu.c
523
next &= ~(tce_size - 1);
arch/powerpc/platforms/pseries/iommu.c
536
dma_offset = next + be64_to_cpu(maprange->dma_base);
arch/powerpc/platforms/pseries/iommu.c
539
tcep[l] = cpu_to_be64(proto_tce | next);
arch/powerpc/platforms/pseries/iommu.c
540
next += tce_size;
arch/powerpc/platforms/pseries/papr-indices.c
101
params->next = rets;
arch/powerpc/platforms/pseries/papr-indices.c
105
params->next = 0;
arch/powerpc/platforms/pseries/papr-indices.c
143
param->next = 1;
arch/powerpc/platforms/pseries/papr-indices.c
180
init_state = (p->next == 1) ? true : false;
arch/powerpc/platforms/pseries/papr-indices.c
43
u32 next;
arch/powerpc/platforms/pseries/papr-indices.c
85
params->next);
arch/powerpc/platforms/pseries/papr-indices.c
98
params->next = 1;
arch/powerpc/platforms/pseries/reconfig.c
184
new->next = last;
arch/powerpc/platforms/pseries/reconfig.c
92
struct property *next;
arch/powerpc/platforms/pseries/reconfig.c
93
for (; prop; prop = next) {
arch/powerpc/platforms/pseries/reconfig.c
94
next = prop->next;
arch/powerpc/sysdev/mpic.c
1519
mpic->next = mpics;
arch/powerpc/sysdev/mpic.c
1951
mpic = mpic->next;
arch/powerpc/sysdev/mpic.c
1993
mpic = mpic->next;
arch/powerpc/sysdev/mpic.c
576
goto next;
arch/powerpc/sysdev/mpic.c
580
goto next;
arch/powerpc/sysdev/mpic.c
585
next:
arch/riscv/include/asm/membarrier.h
17
likely(!(atomic_read(&next->membarrier_state) &
arch/riscv/include/asm/membarrier.h
6
struct mm_struct *next,
arch/riscv/include/asm/mmu_context.h
16
void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/riscv/include/asm/mmu_context.h
21
struct mm_struct *next)
arch/riscv/include/asm/mmu_context.h
24
next->context.pmlen = 0;
arch/riscv/include/asm/mmu_context.h
26
switch_mm(prev, next, NULL);
arch/riscv/include/asm/switch_to.h
113
#define switch_to(prev, next, last) \
arch/riscv/include/asm/switch_to.h
116
struct task_struct *__next = (next); \
arch/riscv/include/asm/switch_to.h
52
struct task_struct *next)
arch/riscv/include/asm/switch_to.h
58
fstate_restore(next, task_pt_regs(next));
arch/riscv/include/asm/switch_to.h
84
static inline void __switch_to_envcfg(struct task_struct *next)
arch/riscv/include/asm/switch_to.h
88
:: "r" (next->thread.envcfg) : "memory");
arch/riscv/include/asm/vector.h
370
struct task_struct *next)
arch/riscv/include/asm/vector.h
390
if (riscv_preempt_v_started(next)) {
arch/riscv/include/asm/vector.h
391
if (next->thread.riscv_v_flags & RISCV_PREEMPT_V_IN_SCHEDULE) {
arch/riscv/include/asm/vector.h
392
next->thread.riscv_v_flags &= ~RISCV_PREEMPT_V_IN_SCHEDULE;
arch/riscv/include/asm/vector.h
395
riscv_preempt_v_set_restore(next);
arch/riscv/include/asm/vector.h
398
riscv_v_vstate_set_restore(next, task_pt_regs(next));
arch/riscv/kernel/asm-offsets.c
138
OFFSET(HIBERN_PBE_NEXT, pbe, next);
arch/riscv/kernel/cpu.c
376
.next = c_next,
arch/riscv/kernel/hibernate.c
198
unsigned long next;
arch/riscv/kernel/hibernate.c
217
next = pmd_addr_end(start, end);
arch/riscv/kernel/hibernate.c
225
ret = temp_pgtable_map_pte(dst_pmdp, src_pmdp, start, next, prot);
arch/riscv/kernel/hibernate.c
229
} while (dst_pmdp++, src_pmdp++, start = next, start != end);
arch/riscv/kernel/hibernate.c
237
unsigned long next;
arch/riscv/kernel/hibernate.c
256
next = pud_addr_end(start, end);
arch/riscv/kernel/hibernate.c
264
ret = temp_pgtable_map_pmd(dst_pudp, src_pudp, start, next, prot);
arch/riscv/kernel/hibernate.c
268
} while (dst_pudp++, src_pudp++, start = next, start != end);
arch/riscv/kernel/hibernate.c
276
unsigned long next;
arch/riscv/kernel/hibernate.c
295
next = p4d_addr_end(start, end);
arch/riscv/kernel/hibernate.c
303
ret = temp_pgtable_map_pud(dst_p4dp, src_p4dp, start, next, prot);
arch/riscv/kernel/hibernate.c
307
} while (dst_p4dp++, src_p4dp++, start = next, start != end);
arch/riscv/kernel/hibernate.c
316
unsigned long next;
arch/riscv/kernel/hibernate.c
322
next = pgd_addr_end(start, end);
arch/riscv/kernel/hibernate.c
330
ret = temp_pgtable_map_p4d(dst_pgdp, src_pgdp, start, next, prot);
arch/riscv/kernel/hibernate.c
334
} while (dst_pgdp++, src_pgdp++, start = next, start != end);
arch/riscv/kernel/vector.c
233
int cur, next;
arch/riscv/kernel/vector.c
238
next = riscv_v_ctrl_get_next(tsk);
arch/riscv/kernel/vector.c
239
if (!next) {
arch/riscv/kernel/vector.c
245
cur = next;
arch/riscv/kernel/vector.c
250
next = PR_RISCV_V_VSTATE_CTRL_DEFAULT;
arch/riscv/kernel/vector.c
252
riscv_v_ctrl_set(tsk, cur, next, inherit);
arch/riscv/kernel/vector.c
266
int cur, next;
arch/riscv/kernel/vector.c
291
next = VSTATE_CTRL_GET_NEXT(arg);
arch/riscv/kernel/vector.c
293
switch (next) {
arch/riscv/kernel/vector.c
297
riscv_v_ctrl_set(current, cur, next, inherit);
arch/riscv/kvm/gstage.c
271
goto next;
arch/riscv/kvm/gstage.c
277
next:
arch/riscv/kvm/gstage.c
305
goto next;
arch/riscv/kvm/gstage.c
311
next:
arch/riscv/mm/context.c
208
struct mm_struct *next, unsigned int cpu)
arch/riscv/mm/context.c
218
cpumask_set_cpu(cpu, mm_cpumask(next));
arch/riscv/mm/context.c
220
set_mm_asid(next, cpu);
arch/riscv/mm/context.c
223
set_mm_noasid(next);
arch/riscv/mm/context.c
278
struct mm_struct *next, unsigned int cpu)
arch/riscv/mm/context.c
318
void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/riscv/mm/context.c
323
if (unlikely(prev == next))
arch/riscv/mm/context.c
326
membarrier_arch_switch_mm(prev, next, task);
arch/riscv/mm/context.c
335
set_mm(prev, next, cpu);
arch/riscv/mm/context.c
337
flush_icache_deferred(next, cpu, task);
arch/riscv/mm/init.c
1441
unsigned long addr, unsigned long next)
arch/riscv/mm/init.c
1447
unsigned long addr, unsigned long next)
arch/riscv/mm/init.c
1449
vmemmap_verify((pte_t *)pmdp, node, addr, next);
arch/riscv/mm/init.c
1653
unsigned long next;
arch/riscv/mm/init.c
1656
for (; addr < end; addr = next) {
arch/riscv/mm/init.c
1657
next = (addr + PAGE_SIZE) & PAGE_MASK;
arch/riscv/mm/init.c
1658
if (next > end)
arch/riscv/mm/init.c
1659
next = end;
arch/riscv/mm/init.c
1675
unsigned long next;
arch/riscv/mm/init.c
1679
for (; addr < end; addr = next) {
arch/riscv/mm/init.c
1680
next = pmd_addr_end(addr, end);
arch/riscv/mm/init.c
1694
remove_pte_mapping(pte_base, addr, next, is_vmemmap, altmap);
arch/riscv/mm/init.c
1702
unsigned long next;
arch/riscv/mm/init.c
1706
for (; addr < end; addr = next) {
arch/riscv/mm/init.c
1707
next = pud_addr_end(addr, end);
arch/riscv/mm/init.c
1723
remove_pmd_mapping(pmd_base, addr, next, is_vmemmap, altmap);
arch/riscv/mm/init.c
1733
unsigned long next;
arch/riscv/mm/init.c
1737
for (; addr < end; addr = next) {
arch/riscv/mm/init.c
1738
next = p4d_addr_end(addr, end);
arch/riscv/mm/init.c
1754
remove_pud_mapping(pud_base, addr, next, is_vmemmap, altmap);
arch/riscv/mm/init.c
1764
unsigned long addr, next;
arch/riscv/mm/init.c
1768
for (addr = va; addr < end; addr = next) {
arch/riscv/mm/init.c
1769
next = pgd_addr_end(addr, end);
arch/riscv/mm/init.c
1779
remove_p4d_mapping(p4d_base, addr, next, is_vmemmap, altmap);
arch/riscv/mm/kasan_init.c
107
kasan_populate_pmd(pudp, vaddr, next);
arch/riscv/mm/kasan_init.c
108
} while (pudp++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
116
unsigned long next;
arch/riscv/mm/kasan_init.c
126
next = p4d_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
129
(next - vaddr) >= P4D_SIZE) {
arch/riscv/mm/kasan_init.c
138
kasan_populate_pud(p4dp, vaddr, next);
arch/riscv/mm/kasan_init.c
139
} while (p4dp++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
146
unsigned long next;
arch/riscv/mm/kasan_init.c
149
next = pgd_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
152
(next - vaddr) >= PGDIR_SIZE) {
arch/riscv/mm/kasan_init.c
161
kasan_populate_p4d(pgdp, vaddr, next);
arch/riscv/mm/kasan_init.c
162
} while (pgdp++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
169
unsigned long next;
arch/riscv/mm/kasan_init.c
179
next = pud_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
181
if (IS_ALIGNED(vaddr, PUD_SIZE) && (next - vaddr) >= PUD_SIZE) {
arch/riscv/mm/kasan_init.c
187
} while (pudp++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
194
unsigned long next;
arch/riscv/mm/kasan_init.c
204
next = p4d_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
207
(next - vaddr) >= P4D_SIZE) {
arch/riscv/mm/kasan_init.c
212
kasan_early_clear_pud(p4dp, vaddr, next);
arch/riscv/mm/kasan_init.c
213
} while (p4dp++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
219
unsigned long next;
arch/riscv/mm/kasan_init.c
222
next = pgd_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
225
(next - vaddr) >= PGDIR_SIZE) {
arch/riscv/mm/kasan_init.c
230
kasan_early_clear_p4d(pgdp, vaddr, next);
arch/riscv/mm/kasan_init.c
231
} while (pgdp++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
240
unsigned long next;
arch/riscv/mm/kasan_init.c
250
next = pud_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
253
(next - vaddr) >= PUD_SIZE) {
arch/riscv/mm/kasan_init.c
260
} while (pudp++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
269
unsigned long next;
arch/riscv/mm/kasan_init.c
288
next = p4d_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
291
(next - vaddr) >= P4D_SIZE) {
arch/riscv/mm/kasan_init.c
297
kasan_early_populate_pud(p4dp, vaddr, next);
arch/riscv/mm/kasan_init.c
298
} while (p4dp++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
306
unsigned long next;
arch/riscv/mm/kasan_init.c
309
next = pgd_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
312
(next - vaddr) >= PGDIR_SIZE) {
arch/riscv/mm/kasan_init.c
318
kasan_early_populate_p4d(pgdp, vaddr, next);
arch/riscv/mm/kasan_init.c
319
} while (pgdp++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
380
unsigned long next;
arch/riscv/mm/kasan_init.c
385
next = pud_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
394
} while (pud_k++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
400
unsigned long next;
arch/riscv/mm/kasan_init.c
405
next = p4d_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
414
} while (p4d_k++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
419
unsigned long next;
arch/riscv/mm/kasan_init.c
424
next = pgd_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
432
kasan_shallow_populate_p4d(pgd_k, vaddr, next);
arch/riscv/mm/kasan_init.c
433
} while (pgd_k++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
54
unsigned long next;
arch/riscv/mm/kasan_init.c
64
next = pmd_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
67
(next - vaddr) >= PMD_SIZE) {
arch/riscv/mm/kasan_init.c
76
kasan_populate_pte(pmdp, vaddr, next);
arch/riscv/mm/kasan_init.c
77
} while (pmdp++, vaddr = next, vaddr != end);
arch/riscv/mm/kasan_init.c
85
unsigned long next;
arch/riscv/mm/kasan_init.c
95
next = pud_addr_end(vaddr, end);
arch/riscv/mm/kasan_init.c
98
(next - vaddr) >= PUD_SIZE) {
arch/riscv/mm/pageattr.c
100
unsigned long next;
arch/riscv/mm/pageattr.c
105
next = pmd_addr_end(vaddr, end);
arch/riscv/mm/pageattr.c
107
if (next - vaddr >= PMD_SIZE &&
arch/riscv/mm/pageattr.c
108
vaddr <= (vaddr & PMD_MASK) && end >= next)
arch/riscv/mm/pageattr.c
130
} while (pmdp++, vaddr = next, vaddr != end);
arch/riscv/mm/pageattr.c
139
unsigned long next;
arch/riscv/mm/pageattr.c
145
next = pud_addr_end(vaddr, end);
arch/riscv/mm/pageattr.c
147
if (next - vaddr >= PUD_SIZE &&
arch/riscv/mm/pageattr.c
148
vaddr <= (vaddr & PUD_MASK) && end >= next)
arch/riscv/mm/pageattr.c
172
ret = __split_linear_mapping_pmd(pudp, vaddr, next);
arch/riscv/mm/pageattr.c
175
} while (pudp++, vaddr = next, vaddr != end);
arch/riscv/mm/pageattr.c
184
unsigned long next;
arch/riscv/mm/pageattr.c
190
next = p4d_addr_end(vaddr, end);
arch/riscv/mm/pageattr.c
196
if (next - vaddr >= P4D_SIZE &&
arch/riscv/mm/pageattr.c
197
vaddr <= (vaddr & P4D_MASK) && end >= next)
arch/riscv/mm/pageattr.c
230
ret = __split_linear_mapping_pud(p4dp, vaddr, next);
arch/riscv/mm/pageattr.c
233
} while (p4dp++, vaddr = next, vaddr != end);
arch/riscv/mm/pageattr.c
242
unsigned long next;
arch/riscv/mm/pageattr.c
246
next = pgd_addr_end(vaddr, end);
arch/riscv/mm/pageattr.c
248
ret = __split_linear_mapping_p4d(pgdp, vaddr, next);
arch/riscv/mm/pageattr.c
251
} while (pgdp++, vaddr = next, vaddr != end);
arch/riscv/mm/pageattr.c
30
unsigned long next, struct mm_walk *walk)
arch/riscv/mm/pageattr.c
43
unsigned long next, struct mm_walk *walk)
arch/riscv/mm/pageattr.c
56
unsigned long next, struct mm_walk *walk)
arch/riscv/mm/pageattr.c
69
unsigned long next, struct mm_walk *walk)
arch/riscv/mm/pageattr.c
79
static int pageattr_pte_hole(unsigned long addr, unsigned long next,
arch/s390/boot/vmem.c
357
unsigned long pa, next, pages = 0;
arch/s390/boot/vmem.c
362
for (; addr < end; addr = next, pmd++) {
arch/s390/boot/vmem.c
363
next = pmd_addr_end(addr, end);
arch/s390/boot/vmem.c
365
if (kasan_pmd_populate_zero_shadow(pmd, addr, next, mode))
arch/s390/boot/vmem.c
367
pa = try_get_large_pmd_pa(pmd, addr, next, mode);
arch/s390/boot/vmem.c
380
pgtable_pte_populate(pmd, addr, next, mode);
arch/s390/boot/vmem.c
389
unsigned long pa, next, pages = 0;
arch/s390/boot/vmem.c
394
for (; addr < end; addr = next, pud++) {
arch/s390/boot/vmem.c
395
next = pud_addr_end(addr, end);
arch/s390/boot/vmem.c
397
if (kasan_pud_populate_zero_shadow(pud, addr, next, mode))
arch/s390/boot/vmem.c
399
pa = try_get_large_pud_pa(pud, addr, next, mode);
arch/s390/boot/vmem.c
412
pgtable_pmd_populate(pud, addr, next, mode);
arch/s390/boot/vmem.c
421
unsigned long next;
arch/s390/boot/vmem.c
426
for (; addr < end; addr = next, p4d++) {
arch/s390/boot/vmem.c
427
next = p4d_addr_end(addr, end);
arch/s390/boot/vmem.c
429
if (kasan_p4d_populate_zero_shadow(p4d, addr, next, mode))
arch/s390/boot/vmem.c
434
pgtable_pud_populate(p4d, addr, next, mode);
arch/s390/boot/vmem.c
440
unsigned long next;
arch/s390/boot/vmem.c
452
for (; addr < end; addr = next, pgd++) {
arch/s390/boot/vmem.c
453
next = pgd_addr_end(addr, end);
arch/s390/boot/vmem.c
455
if (kasan_pgd_populate_zero_shadow(pgd, addr, next, mode))
arch/s390/boot/vmem.c
464
pgtable_p4d_populate(pgd, addr, next, mode);
arch/s390/include/asm/debug.h
440
.next = NULL, \
arch/s390/include/asm/debug.h
45
struct debug_info *next;
arch/s390/include/asm/kvm_host.h
608
int next;
arch/s390/include/asm/mmu_context.h
121
struct mm_struct *next)
arch/s390/include/asm/mmu_context.h
123
switch_mm_irqs_off(prev, next, current);
arch/s390/include/asm/mmu_context.h
124
cpumask_set_cpu(smp_processor_id(), mm_cpumask(next));
arch/s390/include/asm/mmu_context.h
67
static inline void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
arch/s390/include/asm/mmu_context.h
72
if (next == &init_mm)
arch/s390/include/asm/mmu_context.h
75
get_lowcore()->user_asce.val = next->context.asce;
arch/s390/include/asm/mmu_context.h
76
cpumask_set_cpu(cpu, &next->context.cpu_attach_mask);
arch/s390/include/asm/mmu_context.h
80
if (prev != next)
arch/s390/include/asm/mmu_context.h
85
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/s390/include/asm/mmu_context.h
91
switch_mm_irqs_off(prev, next, tsk);
arch/s390/kernel/debug.c
774
debug_area_last->next = id;
arch/s390/kernel/debug.c
778
id->next = NULL;
arch/s390/kernel/debug.c
919
debug_area_first = id->next;
arch/s390/kernel/debug.c
923
id->prev->next = id->next;
arch/s390/kernel/debug.c
924
if (id->next)
arch/s390/kernel/debug.c
925
id->next->prev = id->prev;
arch/s390/kernel/diag/diag.c
123
.next = show_diag_stat_next,
arch/s390/kernel/entry.h
22
struct task_struct *__switch_to_asm(struct task_struct *prev, struct task_struct *next);
arch/s390/kernel/process.c
186
struct task_struct *__switch_to(struct task_struct *prev, struct task_struct *next)
arch/s390/kernel/process.c
193
update_cr_regs(next);
arch/s390/kernel/process.c
194
restore_kernel_fpu_regs(&next->thread);
arch/s390/kernel/process.c
195
restore_access_regs(&next->thread.acrs[0]);
arch/s390/kernel/process.c
196
restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb);
arch/s390/kernel/process.c
197
restore_gs_cb(next->thread.gs_cb);
arch/s390/kernel/process.c
198
return __switch_to_asm(prev, next);
arch/s390/kernel/processor.c
383
.next = c_next,
arch/s390/kernel/sysinfo.c
311
struct service_level *ptr, *next;
arch/s390/kernel/sysinfo.c
315
list_for_each_entry_safe(ptr, next, &service_level_list, list) {
arch/s390/kernel/sysinfo.c
354
.next = service_level_next,
arch/s390/kernel/topology.c
162
info = info->next;
arch/s390/kernel/topology.c
167
info = info->next;
arch/s390/kernel/topology.c
172
info = info->next;
arch/s390/kernel/topology.c
196
drawer = drawer->next;
arch/s390/kernel/topology.c
200
book = book->next;
arch/s390/kernel/topology.c
204
socket = socket->next;
arch/s390/kernel/topology.c
43
struct mask_info *next;
arch/s390/kernel/topology.c
545
mask->next = memblock_alloc_or_panic(sizeof(*mask->next), 8);
arch/s390/kernel/topology.c
546
mask = mask->next;
arch/s390/kernel/topology.c
82
info = info->next;
arch/s390/kvm/dat.c
1000
static long dat_set_pn_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1018
static long dat_set_pn_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1144
static long dat_reset_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1154
return next;
arch/s390/kvm/dat.c
1176
static long __dat_peek_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1184
state->end = next;
arch/s390/kvm/dat.c
1189
static long __dat_peek_cmma_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1194
state->end = min(walk->end, next);
arch/s390/kvm/dat.c
1216
static long __dat_get_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1238
state->end = next;
arch/s390/kvm/dat.c
1271
static long __dat_set_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
539
gfn_t cur, next;
arch/s390/kvm/dat.c
546
for (cur = ALIGN_DOWN(start, cur_size); cur < end; idx++, cur = next) {
arch/s390/kvm/dat.c
547
next = cur + cur_size;
arch/s390/kvm/dat.c
560
rc = the_op(walk->last, cur, next, walk);
arch/s390/kvm/dat.c
567
rc = dat_crste_walk_range(max(start, cur), min(end, next),
arch/s390/kvm/dat.c
570
rc = dat_pte_walk_range(max(start, cur), min(end, next),
arch/s390/kvm/dat.c
788
static long dat_reset_skeys_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
802
return next;
arch/s390/kvm/dat.c
806
static long dat_reset_skeys_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
814
end = ((min(next, walk->end) - gfn) << PAGE_SHIFT) + origin;
arch/s390/kvm/dat.c
821
return next;
arch/s390/kvm/dat.c
841
static long _dat_slot_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
857
static long _dat_slot_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
870
if (walk->start <= gfn && walk->end >= next) {
arch/s390/kvm/dat.h
348
typedef long (*dat_walk_op)(union crste *crste, gfn_t gfn, gfn_t next, struct dat_walk *w);
arch/s390/kvm/dat.h
360
long (*pte_entry)(union pte *pte, gfn_t gfn, gfn_t next, struct dat_walk *w);
arch/s390/kvm/dat.h
463
struct vsie_rmap *next;
arch/s390/kvm/gmap.c
1040
static long __set_cmma_dirty_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
1044
return next;
arch/s390/kvm/gmap.c
1091
struct gmap_cache *gmap_cache, *next;
arch/s390/kvm/gmap.c
1101
list_for_each_entry_safe(gmap_cache, next, &sg->scb_users, list) {
arch/s390/kvm/gmap.c
1112
struct gmap *sg, *next;
arch/s390/kvm/gmap.c
1115
list_for_each_entry_safe(sg, next, &parent->children, list) {
arch/s390/kvm/gmap.c
371
static long _gmap_unmap_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *w)
arch/s390/kvm/gmap.c
393
static long _gmap_unmap_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
851
static long _gmap_split_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
861
gmap_unmap_prefix(gmap, gfn, next);
arch/s390/kvm/gmap.c
870
return next;
arch/s390/kvm/gmap.c
924
static long _destroy_pages_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
930
return next;
arch/s390/kvm/gmap.c
934
static long _destroy_pages_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
943
end = ((min(next, walk->end) - gfn) << PAGE_SHIFT) + origin;
arch/s390/kvm/gmap.c
947
return next;
arch/s390/kvm/gmap.c
988
rmap->next = radix_tree_deref_slot_protected(slot, &sg->host_to_rmap_lock);
arch/s390/kvm/gmap.c
989
for (temp = rmap->next; temp; temp = temp->next) {
arch/s390/kvm/gmap.c
995
rmap->next = NULL;
arch/s390/kvm/gmap.h
82
for (pos = (head); n = pos ? pos->next : NULL, pos; pos = n)
arch/s390/kvm/vsie.c
1454
vsie_page = kvm->arch.vsie.pages[kvm->arch.vsie.next];
arch/s390/kvm/vsie.c
1457
kvm->arch.vsie.next++;
arch/s390/kvm/vsie.c
1458
kvm->arch.vsie.next %= nr_vcpus;
arch/s390/kvm/vsie.c
601
struct vsie_page *cur, *next;
arch/s390/kvm/vsie.c
609
list_for_each_entry_safe(cur, next, &gmap->scb_users, gmap_cache.list) {
arch/s390/lib/spinlock.c
158
struct spin_wait *node, *next;
arch/s390/lib/spinlock.c
165
node->prev = node->next = NULL;
arch/s390/lib/spinlock.c
196
WRITE_ONCE(node->prev->next, node);
arch/s390/lib/spinlock.c
241
while ((next = READ_ONCE(node->next)) == NULL)
arch/s390/lib/spinlock.c
243
next->prev = NULL;
arch/s390/lib/spinlock.c
60
struct spin_wait *next, *prev;
arch/s390/mm/cmm.c
115
pa = pa->next;
arch/s390/mm/cmm.c
39
struct cmm_page_array *next;
arch/s390/mm/cmm.c
86
npa->next = pa;
arch/s390/mm/pageattr.c
167
unsigned long next;
arch/s390/mm/pageattr.c
176
next = pmd_addr_end(addr, end);
arch/s390/mm/pageattr.c
180
need_split |= !!(addr + PMD_SIZE > next);
arch/s390/mm/pageattr.c
189
rc = walk_pte_level(pmdp, addr, next, flags);
arch/s390/mm/pageattr.c
194
addr = next;
arch/s390/mm/pageattr.c
254
unsigned long next;
arch/s390/mm/pageattr.c
263
next = pud_addr_end(addr, end);
arch/s390/mm/pageattr.c
267
need_split |= !!(addr + PUD_SIZE > next);
arch/s390/mm/pageattr.c
276
rc = walk_pmd_level(pudp, addr, next, flags);
arch/s390/mm/pageattr.c
279
addr = next;
arch/s390/mm/pageattr.c
288
unsigned long next;
arch/s390/mm/pageattr.c
296
next = p4d_addr_end(addr, end);
arch/s390/mm/pageattr.c
297
rc = walk_pud_level(p4dp, addr, next, flags);
arch/s390/mm/pageattr.c
299
addr = next;
arch/s390/mm/pageattr.c
310
unsigned long next;
arch/s390/mm/pageattr.c
318
next = pgd_addr_end(addr, end);
arch/s390/mm/pageattr.c
319
rc = walk_p4d_level(pgdp, addr, next, flags);
arch/s390/mm/pageattr.c
323
} while (pgdp++, addr = next, addr < end && !rc);
arch/s390/mm/pfault.c
212
struct thread_struct *thread, *next;
arch/s390/mm/pfault.c
216
list_for_each_entry_safe(thread, next, &pfault_list, list) {
arch/s390/mm/pgalloc.c
210
unsigned long next = (addr + (SIZE)) & ~((SIZE) - 1); \
arch/s390/mm/pgalloc.c
212
return (next - 1) < (end - 1) ? next : end; \
arch/s390/mm/pgalloc.c
234
unsigned long *pte, next;
arch/s390/mm/pgalloc.c
241
next = base_page_addr_end(addr, end);
arch/s390/mm/pgalloc.c
243
} while (pte++, addr = next, addr < end);
arch/s390/mm/pgalloc.c
250
unsigned long *ste, next, *table;
arch/s390/mm/pgalloc.c
256
next = base_segment_addr_end(addr, end);
arch/s390/mm/pgalloc.c
266
rc = base_page_walk(table, addr, next, alloc);
arch/s390/mm/pgalloc.c
272
} while (ste++, addr = next, addr < end);
arch/s390/mm/pgalloc.c
279
unsigned long *rtte, next, *table;
arch/s390/mm/pgalloc.c
285
next = base_region3_addr_end(addr, end);
arch/s390/mm/pgalloc.c
295
rc = base_segment_walk(table, addr, next, alloc);
arch/s390/mm/pgalloc.c
300
} while (rtte++, addr = next, addr < end);
arch/s390/mm/pgalloc.c
307
unsigned long *rste, next, *table;
arch/s390/mm/pgalloc.c
313
next = base_region2_addr_end(addr, end);
arch/s390/mm/pgalloc.c
323
rc = base_region3_walk(table, addr, next, alloc);
arch/s390/mm/pgalloc.c
328
} while (rste++, addr = next, addr < end);
arch/s390/mm/pgalloc.c
335
unsigned long *rfte, next, *table;
arch/s390/mm/pgalloc.c
341
next = base_region1_addr_end(addr, end);
arch/s390/mm/pgalloc.c
351
rc = base_region2_walk(table, addr, next, alloc);
arch/s390/mm/pgalloc.c
356
} while (rfte++, addr = next, addr < end);
arch/s390/mm/pgtable.c
346
pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next;
arch/s390/mm/vmem.c
227
unsigned long next, prot, pages = 0;
arch/s390/mm/vmem.c
234
for (; addr < end; addr = next, pmd++) {
arch/s390/mm/vmem.c
235
next = pmd_addr_end(addr, end);
arch/s390/mm/vmem.c
241
IS_ALIGNED(next, PMD_SIZE)) {
arch/s390/mm/vmem.c
246
} else if (!direct && vmemmap_unuse_sub_pmd(addr, next)) {
arch/s390/mm/vmem.c
254
IS_ALIGNED(next, PMD_SIZE) &&
arch/s390/mm/vmem.c
274
!IS_ALIGNED(next, PMD_SIZE)) {
arch/s390/mm/vmem.c
275
vmemmap_use_new_sub_pmd(addr, next);
arch/s390/mm/vmem.c
286
vmemmap_use_sub_pmd(addr, next);
arch/s390/mm/vmem.c
289
ret = modify_pte_table(pmd, addr, next, add, direct, altmap);
arch/s390/mm/vmem.c
318
unsigned long next, prot, pages = 0;
arch/s390/mm/vmem.c
325
for (; addr < end; addr = next, pud++) {
arch/s390/mm/vmem.c
326
next = pud_addr_end(addr, end);
arch/s390/mm/vmem.c
332
IS_ALIGNED(next, PUD_SIZE)) {
arch/s390/mm/vmem.c
344
IS_ALIGNED(next, PUD_SIZE) &&
arch/s390/mm/vmem.c
358
ret = modify_pmd_table(pud, addr, next, add, direct, altmap);
arch/s390/mm/vmem.c
388
unsigned long next;
arch/s390/mm/vmem.c
394
for (; addr < end; addr = next, p4d++) {
arch/s390/mm/vmem.c
395
next = p4d_addr_end(addr, end);
arch/s390/mm/vmem.c
405
ret = modify_pud_table(p4d, addr, next, add, direct, altmap);
arch/s390/mm/vmem.c
433
unsigned long addr, next;
arch/s390/mm/vmem.c
449
for (addr = start; addr < end; addr = next) {
arch/s390/mm/vmem.c
450
next = pgd_addr_end(addr, end);
arch/s390/mm/vmem.c
462
ret = modify_p4d_table(pgd, addr, next, add, direct, altmap);
arch/s390/pci/pci_bus.c
292
struct zpci_bus *curp = *pos, *next = NULL;
arch/s390/pci/pci_bus.c
296
next = list_next_entry(curp, bus_next);
arch/s390/pci/pci_bus.c
298
next = list_first_entry(&zbus_list, typeof(*curp), bus_next);
arch/s390/pci/pci_bus.c
300
if (list_entry_is_head(next, &zbus_list, bus_next))
arch/s390/pci/pci_bus.c
301
next = NULL;
arch/s390/pci/pci_bus.c
303
if (next)
arch/s390/pci/pci_bus.c
304
__zpci_bus_get(next);
arch/s390/pci/pci_bus.c
306
*pos = next;
arch/sh/drivers/pci/pci.c
116
hose_tail = &hose->next;
arch/sh/drivers/pci/pci.c
155
for (hose = hose_head; hose; hose = hose->next)
arch/sh/drivers/pci/pci.c
262
for (hose = hose_head; hose; hose = hose->next) {
arch/sh/drivers/platform_early.c
129
if (!dev->devres_head.next) {
arch/sh/drivers/platform_early.c
72
if (!epdrv->list.next) {
arch/sh/include/asm/dwarf.h
273
struct dwarf_frame *prev, *next;
arch/sh/include/asm/mmu_context.h
110
struct mm_struct *next,
arch/sh/include/asm/mmu_context.h
115
if (likely(prev != next)) {
arch/sh/include/asm/mmu_context.h
116
cpumask_set_cpu(cpu, mm_cpumask(next));
arch/sh/include/asm/mmu_context.h
117
set_TTB(next->pgd);
arch/sh/include/asm/mmu_context.h
118
activate_context(next, cpu);
arch/sh/include/asm/mmu_context.h
120
if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)))
arch/sh/include/asm/mmu_context.h
121
activate_context(next, cpu);
arch/sh/include/asm/pci.h
16
struct pci_channel *next;
arch/sh/include/asm/processor.h
61
unsigned long long next;
arch/sh/include/asm/switch_to_32.h
65
struct task_struct *next);
arch/sh/include/asm/switch_to_32.h
70
#define switch_to(prev, next, last) \
arch/sh/include/asm/switch_to_32.h
82
if (is_dsp_enabled(next)) \
arch/sh/include/asm/switch_to_32.h
83
__restore_dsp(next); \
arch/sh/include/asm/switch_to_32.h
88
__ts5 = (u32 *)next; \
arch/sh/include/asm/switch_to_32.h
89
__ts6 = (u32 *)&next->thread.sp; \
arch/sh/include/asm/switch_to_32.h
90
__ts7 = next->thread.pc; \
arch/sh/kernel/asm-offsets.c
34
DEFINE(PBE_NEXT, offsetof(struct pbe, next));
arch/sh/kernel/cpu/proc.c
147
.next = c_next,
arch/sh/kernel/cpu/sh4/sq.c
212
for (p = &sq_mapping_list; (map = *p); p = &map->next)
arch/sh/kernel/cpu/sh4/sq.c
294
for (list = &sq_mapping_list; (entry = *list); list = &entry->next)
arch/sh/kernel/cpu/sh4/sq.c
34
struct sq_mapping *next;
arch/sh/kernel/cpu/sh4/sq.c
78
p = &tmp->next;
arch/sh/kernel/cpu/sh4/sq.c
80
map->next = tmp;
arch/sh/kernel/cpu/sh4/sq.c
92
for (p = &sq_mapping_list; (tmp = *p); p = &tmp->next)
arch/sh/kernel/cpu/sh4/sq.c
94
*p = tmp->next;
arch/sh/kernel/process_32.c
145
__switch_to(struct task_struct *prev, struct task_struct *next)
arch/sh/kernel/process_32.c
147
struct thread_struct *next_t = &next->thread;
arch/sh/kernel/process_32.c
150
__stack_chk_guard = next->stack_canary;
arch/sh/kernel/process_32.c
156
if (next->thread.fpu_counter > 5)
arch/sh/kernel/process_32.c
166
: "r" (task_thread_info(next)));
arch/sh/kernel/process_32.c
174
if (next->thread.fpu_counter > 5)
arch/sh/kernel/unwinder.c
34
.next = &unwinder_list,
arch/sh/kernel/unwinder.c
52
.next = &stack_reader.list,
arch/sh/kernel/unwinder.c
73
best = list_entry(unwinder_list.next, struct unwinder, list);
arch/sparc/include/asm/cpudata_32.h
25
int next;
arch/sparc/include/asm/leon_amba.h
161
struct amba_confarea_type *next;/* next bus in chain */
arch/sparc/include/asm/mdesc.h
72
struct mdesc_notifier_client *next;
arch/sparc/include/asm/switch_to_32.h
28
#define SWITCH_DO_LAZY_FPU(next) /* */
arch/sparc/include/asm/switch_to_32.h
38
#define prepare_arch_switch(next) do { \
arch/sparc/include/asm/switch_to_32.h
58
#define switch_to(prev, next, last) do { \
arch/sparc/include/asm/switch_to_32.h
60
SWITCH_DO_LAZY_FPU(next); \
arch/sparc/include/asm/switch_to_32.h
61
cpumask_set_cpu(smp_processor_id(), mm_cpumask(next->active_mm)); \
arch/sparc/include/asm/switch_to_32.h
93
"r" (task_thread_info(next)), \
arch/sparc/include/asm/switch_to_64.h
21
#define switch_to(prev, next, last) \
arch/sparc/include/asm/switch_to_64.h
26
task_thread_info(next); \
arch/sparc/include/asm/switch_to_64.h
57
: "0" (task_thread_info(next)), \
arch/sparc/include/asm/switch_to_64.h
7
#define prepare_arch_switch(next) \
arch/sparc/kernel/cpu.c
434
.next = c_next,
arch/sparc/kernel/ds.c
1220
dp->next = ds_info_list;
arch/sparc/kernel/ds.c
216
struct ds_info *next;
arch/sparc/kernel/ds.c
752
for (dp = ds_info_list; dp; dp = dp->next) {
arch/sparc/kernel/ds.c
762
for (dp = ds_info_list; dp; dp = dp->next) {
arch/sparc/kernel/irq.h
7
struct irq_bucket *next;
arch/sparc/kernel/irq_32.c
170
p->next = irq_map[pil];
arch/sparc/kernel/irq_32.c
189
pnext = &(*pnext)->next;
arch/sparc/kernel/irq_32.c
190
*pnext = p->next;
arch/sparc/kernel/irq_32.c
229
struct irq_bucket *next = p->next;
arch/sparc/kernel/irq_32.c
232
p = next;
arch/sparc/kernel/leon_pci_grpci2.c
164
unsigned int next; /* 0x0C Next Data Descriptor in chan */
arch/sparc/kernel/leon_smp.c
240
prev = &cpu_data(i).next;
arch/sparc/kernel/mdesc.c
310
client->next = client_list;
arch/sparc/kernel/mdesc.c
497
p = p->next;
arch/sparc/kernel/of_device_64.c
446
goto next;
arch/sparc/kernel/of_device_64.c
454
next:
arch/sparc/kernel/pci_fire.c
454
pbm->next = pci_pbm_root;
arch/sparc/kernel/pci_impl.h
61
struct pci_pbm_info *next;
arch/sparc/kernel/pci_psycho.c
501
for (pbm = pci_pbm_root; pbm; pbm = pbm->next) {
arch/sparc/kernel/pci_psycho.c
578
pbm->next = pci_pbm_root;
arch/sparc/kernel/pci_sabre.c
572
pbm->next = pci_pbm_root;
arch/sparc/kernel/pci_schizo.c
1351
pbm->next = pci_pbm_root;
arch/sparc/kernel/pci_schizo.c
1411
for (pbm = pci_pbm_root; pbm; pbm = pbm->next) {
arch/sparc/kernel/pci_sun4v.c
1224
pbm->next = pci_pbm_root;
arch/sparc/kernel/prom_common.c
89
prevp = &(*prevp)->next;
arch/sparc/kernel/starfire.c
41
struct starfire_irqinfo *next;
arch/sparc/kernel/starfire.c
73
p->next = sflist;
arch/sparc/kernel/starfire.c
85
for (p = sflist; p != NULL; p = p->next)
arch/sparc/kernel/sun4d_irq.c
136
struct irq_bucket *next;
arch/sparc/kernel/sun4d_irq.c
138
next = p->next;
arch/sparc/kernel/sun4d_irq.c
140
p = next;
arch/sparc/kernel/sun4d_irq.c
175
struct irq_bucket *next;
arch/sparc/kernel/sun4d_irq.c
177
next = p->next;
arch/sparc/kernel/sun4d_irq.c
179
p = next;
arch/sparc/kernel/sun4d_smp.c
173
prev = &cpu_data(i).next;
arch/sparc/kernel/sun4m_smp.c
131
prev = &cpu_data(i).next;
arch/sparc/kernel/time_32.c
205
unsigned int next = (unsigned int)delta;
arch/sparc/kernel/time_32.c
207
sparc_config.load_profile_irq(cpu, next);
arch/sparc/mm/io-unit.c
115
next: j = (i & 15);
arch/sparc/mm/io-unit.c
130
goto next;
arch/sparc/mm/srmmu.c
379
struct ctx_list *next;
arch/sparc/mm/srmmu.c
394
entry->next->prev = entry->prev;
arch/sparc/mm/srmmu.c
395
entry->prev->next = entry->next;
arch/sparc/mm/srmmu.c
400
entry->next = head;
arch/sparc/mm/srmmu.c
401
(entry->prev = head->prev)->next = entry;
arch/sparc/mm/srmmu.c
412
ctxp = ctx_free.next;
arch/sparc/mm/srmmu.c
420
ctxp = ctx_used.next;
arch/sparc/mm/srmmu.c
422
ctxp = ctxp->next;
arch/sparc/mm/srmmu.c
458
ctx_free.next = ctx_free.prev = &ctx_free;
arch/sparc/mm/srmmu.c
459
ctx_used.next = ctx_used.prev = &ctx_used;
arch/sparc/mm/tlb.c
306
pmd_huge_pte(mm, pmdp) = (pgtable_t) lh->next;
arch/um/drivers/chan_kern.c
368
struct list_head *ele, *next;
arch/um/drivers/chan_kern.c
371
list_for_each_safe(ele, next, chans) {
arch/um/drivers/line.c
714
struct list_head *ele, *next;
arch/um/drivers/line.c
720
list_for_each_safe(ele, next, &winch_handlers) {
arch/um/drivers/mconsole_kern.c
330
unplugged = list_entry(unplugged_pages.next,
arch/um/drivers/mconsole_kern.c
356
struct list_head *entry = unplugged_pages.next;
arch/um/drivers/mconsole_kern.c
68
req = list_entry(mc_requests.next, struct mconsole_entry, list);
arch/um/drivers/port_kern.c
240
conn = list_entry(port->connections.next, struct connection,
arch/um/include/asm/mmu_context.h
16
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/um/kernel/time.c
531
static void time_travel_update_time(unsigned long long next, bool idle)
arch/um/kernel/time.c
540
__time_travel_add_event(&ne, next);
arch/um/kernel/time.c
627
unsigned long long next = S64_MAX;
arch/um/kernel/time.c
633
time_travel_update_time(next, true);
arch/um/kernel/time.c
644
os_timer_one_shot(cpu, time_travel_timer_event.time - next);
arch/um/kernel/tlb.c
102
next = pmd_addr_end(addr, end);
arch/um/kernel/tlb.c
106
next - addr);
arch/um/kernel/tlb.c
110
else ret = update_pte_range(pmd, addr, next, ops);
arch/um/kernel/tlb.c
111
} while (pmd++, addr = next, ((addr < end) && !ret));
arch/um/kernel/tlb.c
120
unsigned long next;
arch/um/kernel/tlb.c
125
next = pud_addr_end(addr, end);
arch/um/kernel/tlb.c
129
next - addr);
arch/um/kernel/tlb.c
133
else ret = update_pmd_range(pud, addr, next, ops);
arch/um/kernel/tlb.c
134
} while (pud++, addr = next, ((addr < end) && !ret));
arch/um/kernel/tlb.c
143
unsigned long next;
arch/um/kernel/tlb.c
148
next = p4d_addr_end(addr, end);
arch/um/kernel/tlb.c
152
next - addr);
arch/um/kernel/tlb.c
156
ret = update_pud_range(p4d, addr, next, ops);
arch/um/kernel/tlb.c
157
} while (p4d++, addr = next, ((addr < end) && !ret));
arch/um/kernel/tlb.c
165
unsigned long addr, next;
arch/um/kernel/tlb.c
185
next = pgd_addr_end(addr, mm->context.sync_tlb_range_to);
arch/um/kernel/tlb.c
189
next - addr);
arch/um/kernel/tlb.c
193
ret = update_p4d_range(pgd, addr, next, &ops);
arch/um/kernel/tlb.c
194
} while (pgd++, addr = next,
arch/um/kernel/tlb.c
97
unsigned long next;
arch/um/kernel/um_arch.c
119
.next = c_next,
arch/x86/boot/compressed/efi.c
103
pa_data = data->next;
arch/x86/boot/compressed/ident_map_64.c
177
sd = (struct setup_data *)sd->next;
arch/x86/boot/compressed/kaslr.c
446
ptr = (struct setup_data *)(unsigned long)ptr->next;
arch/x86/boot/compressed/kaslr.c
786
ptr = (struct setup_data *)(unsigned long)ptr->next;
arch/x86/boot/startup/sev-shared.c
542
hdr = (struct setup_data *)hdr->next;
arch/x86/crypto/aesni-intel_glue.c
1302
goto next;
arch/x86/crypto/aesni-intel_glue.c
1316
next:
arch/x86/events/amd/uncore.c
571
struct amd_uncore_ctx *curr, *next;
arch/x86/events/amd/uncore.c
586
next = *per_cpu_ptr(pmu->ctx, j);
arch/x86/events/amd/uncore.c
587
if (!next || cpu == j)
arch/x86/events/amd/uncore.c
590
if (curr == next) {
arch/x86/events/amd/uncore.c
594
next->cpu = j;
arch/x86/events/core.c
2172
for (quirk = x86_pmu.quirks; quirk; quirk = quirk->next)
arch/x86/events/intel/pt.c
858
buf->cur = list_entry(buf->cur->list.next, struct topa,
arch/x86/events/intel/uncore_discovery.c
439
struct intel_uncore_discovery_type *type, *next;
arch/x86/events/intel/uncore_discovery.c
443
rbtree_postorder_for_each_entry_safe(type, next, &discovery_tables, node) {
arch/x86/events/perf_event.h
1094
__quirk.next = x86_pmu.quirks; \
arch/x86/events/perf_event.h
670
struct x86_pmu_quirk *next;
arch/x86/include/asm/mce.h
185
unsigned next;
arch/x86/include/asm/mmu_context.h
181
extern void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/x86/include/asm/mmu_context.h
184
extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
arch/x86/include/asm/mmu_context.h
188
#define activate_mm(prev, next) \
arch/x86/include/asm/mmu_context.h
190
paravirt_enter_mmap(next); \
arch/x86/include/asm/mmu_context.h
191
switch_mm_irqs_off((prev), (next), NULL); \
arch/x86/include/asm/mmu_context.h
76
extern void switch_ldt(struct mm_struct *prev, struct mm_struct *next);
arch/x86/include/asm/mmu_context.h
82
static inline void switch_ldt(struct mm_struct *prev, struct mm_struct *next)
arch/x86/include/asm/paravirt.h
282
static inline void paravirt_enter_mmap(struct mm_struct *next)
arch/x86/include/asm/paravirt.h
284
PVOP_VCALL1(pv_ops, mmu.enter_mmap, next);
arch/x86/include/asm/paravirt.h
490
static inline void arch_end_context_switch(struct task_struct *next)
arch/x86/include/asm/paravirt.h
492
PVOP_VCALL1(pv_ops, cpu.end_context_switch, next);
arch/x86/include/asm/paravirt_types.h
86
void (*end_context_switch)(struct task_struct *next);
arch/x86/include/asm/switch_to.h
10
struct task_struct *next);
arch/x86/include/asm/switch_to.h
13
struct task_struct *next);
arch/x86/include/asm/switch_to.h
49
#define switch_to(prev, next, last) \
arch/x86/include/asm/switch_to.h
51
((last) = __switch_to_asm((prev), (next))); \
arch/x86/include/uapi/asm/setup_data.h
28
__u64 next;
arch/x86/kernel/alternative.c
1276
char *next = strchr(str, ',');
arch/x86/kernel/alternative.c
1277
if (next) {
arch/x86/kernel/alternative.c
1278
*next = 0;
arch/x86/kernel/alternative.c
1279
next++;
arch/x86/kernel/alternative.c
1318
str = next;
arch/x86/kernel/alternative.c
2158
struct list_head next;
arch/x86/kernel/alternative.c
2193
list_add_tail(&smp->next, &smp_alt_modules);
arch/x86/kernel/alternative.c
2205
list_for_each_entry(item, &smp_alt_modules, next) {
arch/x86/kernel/alternative.c
2208
list_del(&item->next);
arch/x86/kernel/alternative.c
2229
list_for_each_entry(mod, &smp_alt_modules, next)
arch/x86/kernel/alternative.c
2250
list_for_each_entry(mod, &smp_alt_modules, next) {
arch/x86/kernel/alternative.c
374
for (int next, i = 0; i < len; i = next) {
arch/x86/kernel/alternative.c
380
next = i + insn.length;
arch/x86/kernel/alternative.c
389
next = skip_nops(buf, next, len);
arch/x86/kernel/alternative.c
391
add_nop(buf + nop, next - nop);
arch/x86/kernel/alternative.c
392
DUMP_BYTES(ALT, buf, len, "%px: [%d:%d) optimized NOPs: ", instr, nop, next);
arch/x86/kernel/alternative.c
469
for (int next, i = 0; i < instrlen; i = next) {
arch/x86/kernel/alternative.c
475
next = i + insn.length;
arch/x86/kernel/alternative.c
488
if (need_reloc(next + insn.immediate.value, repl, repl_len)) {
arch/x86/kernel/alternative.c
512
if (need_reloc(next + insn.displacement.value, repl, repl_len)) {
arch/x86/kernel/apm_32.c
1154
for (as = user_list; as != NULL; as = as->next) {
arch/x86/kernel/apm_32.c
1242
for (as = user_list; as != NULL; as = as->next) {
arch/x86/kernel/apm_32.c
1557
user_list = as->next;
arch/x86/kernel/apm_32.c
1562
(as1 != NULL) && (as1->next != as);
arch/x86/kernel/apm_32.c
1563
as1 = as1->next)
arch/x86/kernel/apm_32.c
1568
as1->next = as->next;
arch/x86/kernel/apm_32.c
1598
as->next = user_list;
arch/x86/kernel/apm_32.c
327
struct apm_user *next;
arch/x86/kernel/asm-offsets.c
53
OFFSET(pbe_next, pbe, next);
arch/x86/kernel/cpu/bugs.c
1187
char *next = strchr(str, ',');
arch/x86/kernel/cpu/bugs.c
1188
if (next) {
arch/x86/kernel/cpu/bugs.c
1189
*next = 0;
arch/x86/kernel/cpu/bugs.c
1190
next++;
arch/x86/kernel/cpu/bugs.c
1211
str = next;
arch/x86/kernel/cpu/mce/dev-mcelog.c
206
unsigned next;
arch/x86/kernel/cpu/mce/dev-mcelog.c
222
next = mcelog->next;
arch/x86/kernel/cpu/mce/dev-mcelog.c
225
for (i = 0; i < next; i++) {
arch/x86/kernel/cpu/mce/dev-mcelog.c
232
memset(mcelog->entry, 0, next * sizeof(struct mce));
arch/x86/kernel/cpu/mce/dev-mcelog.c
233
mcelog->next = 0;
arch/x86/kernel/cpu/mce/dev-mcelog.c
247
if (READ_ONCE(mcelog->next))
arch/x86/kernel/cpu/mce/dev-mcelog.c
47
entry = mcelog->next;
arch/x86/kernel/cpu/mce/dev-mcelog.c
58
mcelog->next = entry + 1;
arch/x86/kernel/cpu/mce/severity.c
450
.next = s_next,
arch/x86/kernel/cpu/microcode/amd.c
910
__list_del(p->plist.prev, p->plist.next);
arch/x86/kernel/cpu/mtrr/generic.c
199
struct cache_map *next = cache_map + idx;
arch/x86/kernel/cpu/mtrr/generic.c
201
if (!next->fixed && end == next->start && type == next->type)
arch/x86/kernel/cpu/proc.c
179
.next = c_next,
arch/x86/kernel/dumpstack.c
264
goto next;
arch/x86/kernel/dumpstack.c
287
next:
arch/x86/kernel/hpet.c
106
char *next = strchr(str, ',');
arch/x86/kernel/hpet.c
108
if (next)
arch/x86/kernel/hpet.c
109
*next++ = 0;
arch/x86/kernel/hpet.c
116
str = next;
arch/x86/kernel/jailhouse.c
231
pa_data = header.next;
arch/x86/kernel/kdebugfs.c
117
pa_next = data->next;
arch/x86/kernel/kexec-bzimage64.c
138
sd->next = params->hdr.setup_data;
arch/x86/kernel/kexec-bzimage64.c
182
sd->next = params->hdr.setup_data;
arch/x86/kernel/kexec-bzimage64.c
242
sd->next = params->hdr.setup_data;
arch/x86/kernel/kexec-bzimage64.c
269
sd->next = params->hdr.setup_data;
arch/x86/kernel/kexec-bzimage64.c
296
sd->next = params->hdr.setup_data;
arch/x86/kernel/ksysfs.c
104
pa_next = data->next;
arch/x86/kernel/ksysfs.c
319
pa_data = data->next;
arch/x86/kernel/ksysfs.c
85
pa_data = data->next;
arch/x86/kernel/kvm.c
195
struct hlist_node *p, *next;
arch/x86/kernel/kvm.c
198
hlist_for_each_safe(p, next, &b->list) {
arch/x86/kernel/ldt.c
110
(unsigned long)next->context.ldt))
arch/x86/kernel/ldt.c
111
load_mm_ldt(next);
arch/x86/kernel/ldt.c
90
void switch_ldt(struct mm_struct *prev, struct mm_struct *next)
arch/x86/kernel/process.h
14
struct task_struct *next)
arch/x86/kernel/process.h
16
unsigned long next_tif = read_task_thread_flags(next);
arch/x86/kernel/process.h
38
__switch_to_xtra(prev, next);
arch/x86/kernel/process_32.c
158
*next = &next_p->thread;
arch/x86/kernel/process_32.c
180
load_TLS(next, cpu);
arch/x86/kernel/process_32.c
197
refresh_sysenter_cs(next);
arch/x86/kernel/process_32.c
205
if (prev->gs | next->gs)
arch/x86/kernel/process_32.c
206
loadsegment(gs, next->gs);
arch/x86/kernel/process_64.c
375
struct thread_struct *next)
arch/x86/kernel/process_64.c
387
if (prev->pkru != next->pkru)
arch/x86/kernel/process_64.c
388
wrpkru(next->pkru);
arch/x86/kernel/process_64.c
392
struct thread_struct *next)
arch/x86/kernel/process_64.c
396
if (unlikely(prev->fsindex || next->fsindex))
arch/x86/kernel/process_64.c
397
loadseg(FS, next->fsindex);
arch/x86/kernel/process_64.c
398
if (unlikely(prev->gsindex || next->gsindex))
arch/x86/kernel/process_64.c
399
loadseg(GS, next->gsindex);
arch/x86/kernel/process_64.c
402
wrfsbase(next->fsbase);
arch/x86/kernel/process_64.c
403
__wrgsbase_inactive(next->gsbase);
arch/x86/kernel/process_64.c
406
next->fsindex, next->fsbase, FS);
arch/x86/kernel/process_64.c
408
next->gsindex, next->gsbase, GS);
arch/x86/kernel/process_64.c
613
struct thread_struct *next = &next_p->thread;
arch/x86/kernel/process_64.c
632
load_TLS(next, cpu);
arch/x86/kernel/process_64.c
656
if (unlikely(next->es | prev->es))
arch/x86/kernel/process_64.c
657
loadsegment(es, next->es);
arch/x86/kernel/process_64.c
660
if (unlikely(next->ds | prev->ds))
arch/x86/kernel/process_64.c
661
loadsegment(ds, next->ds);
arch/x86/kernel/process_64.c
663
x86_fsgsbase_load(prev, next);
arch/x86/kernel/process_64.c
665
x86_pkru_load(prev, next);
arch/x86/kernel/setup.c
491
pa_next = data->next;
arch/x86/kernel/setup.c
584
pa_next = data->next;
arch/x86/kvm/mmu/mmu.c
1134
struct pte_list_desc *desc, *next;
arch/x86/kvm/mmu/mmu.c
1149
for (; desc; desc = next) {
arch/x86/kvm/mmu/mmu.c
1152
next = desc->more;
arch/x86/mm/ident_map.c
104
next = pud_addr_end(addr, end);
arch/x86/mm/ident_map.c
117
use_gbpage &= ((next & ~PUD_MASK) == 0);
arch/x86/mm/ident_map.c
132
ident_pmd_init(info, pmd, addr, next);
arch/x86/mm/ident_map.c
138
ident_pmd_init(info, pmd, addr, next);
arch/x86/mm/ident_map.c
148
unsigned long next;
arch/x86/mm/ident_map.c
151
for (; addr < end; addr = next) {
arch/x86/mm/ident_map.c
155
next = p4d_addr_end(addr, end);
arch/x86/mm/ident_map.c
158
result = ident_pud_init(info, pud, addr, next);
arch/x86/mm/ident_map.c
168
result = ident_pud_init(info, pud, addr, next);
arch/x86/mm/ident_map.c
183
unsigned long next;
arch/x86/mm/ident_map.c
193
for (; addr < end; addr = next) {
arch/x86/mm/ident_map.c
197
next = pgd_addr_end(addr, end);
arch/x86/mm/ident_map.c
200
result = ident_p4d_init(info, p4d, addr, next);
arch/x86/mm/ident_map.c
209
result = ident_p4d_init(info, p4d, addr, next);
arch/x86/mm/ident_map.c
97
unsigned long next;
arch/x86/mm/ident_map.c
99
for (; addr < end; addr = next) {
arch/x86/mm/init.c
700
unsigned long next, start;
arch/x86/mm/init.c
716
next = round_up(start + 1, step_size);
arch/x86/mm/init.c
717
if (next > map_end)
arch/x86/mm/init.c
718
next = map_end;
arch/x86/mm/init.c
720
next = map_end;
arch/x86/mm/init.c
723
mapped_ram_size += init_range_memory_mapping(start, next);
arch/x86/mm/init.c
724
start = next;
arch/x86/mm/init_64.c
1098
unsigned long next, pages = 0;
arch/x86/mm/init_64.c
1103
for (; addr < end; addr = next, pte++) {
arch/x86/mm/init_64.c
1104
next = (addr + PAGE_SIZE) & PAGE_MASK;
arch/x86/mm/init_64.c
1105
if (next > end)
arch/x86/mm/init_64.c
1106
next = end;
arch/x86/mm/init_64.c
1141
unsigned long next, pages = 0;
arch/x86/mm/init_64.c
1146
for (; addr < end; addr = next, pmd++) {
arch/x86/mm/init_64.c
1147
next = pmd_addr_end(addr, end);
arch/x86/mm/init_64.c
1154
IS_ALIGNED(next, PMD_SIZE)) {
arch/x86/mm/init_64.c
1163
} else if (vmemmap_pmd_is_unused(addr, next)) {
arch/x86/mm/init_64.c
1174
remove_pte_table(pte_base, addr, next, direct);
arch/x86/mm/init_64.c
1187
unsigned long next, pages = 0;
arch/x86/mm/init_64.c
1192
for (; addr < end; addr = next, pud++) {
arch/x86/mm/init_64.c
1193
next = pud_addr_end(addr, end);
arch/x86/mm/init_64.c
1200
IS_ALIGNED(next, PUD_SIZE)) {
arch/x86/mm/init_64.c
1209
remove_pmd_table(pmd_base, addr, next, direct, altmap);
arch/x86/mm/init_64.c
1221
unsigned long next, pages = 0;
arch/x86/mm/init_64.c
1226
for (; addr < end; addr = next, p4d++) {
arch/x86/mm/init_64.c
1227
next = p4d_addr_end(addr, end);
arch/x86/mm/init_64.c
1235
remove_pud_table(pud_base, addr, next, altmap, direct);
arch/x86/mm/init_64.c
1254
unsigned long next;
arch/x86/mm/init_64.c
1259
for (addr = start; addr < end; addr = next) {
arch/x86/mm/init_64.c
1260
next = pgd_addr_end(addr, end);
arch/x86/mm/init_64.c
1267
remove_p4d_table(p4d, addr, next, altmap, direct);
arch/x86/mm/init_64.c
1519
unsigned long addr, unsigned long next)
arch/x86/mm/init_64.c
1541
!IS_ALIGNED(next, PMD_SIZE))
arch/x86/mm/init_64.c
1542
vmemmap_use_new_sub_pmd(addr, next);
arch/x86/mm/init_64.c
1546
unsigned long addr, unsigned long next)
arch/x86/mm/init_64.c
1551
vmemmap_verify((pte_t *)pmd, node, addr, next);
arch/x86/mm/init_64.c
1552
vmemmap_use_sub_pmd(addr, next);
arch/x86/mm/init_64.c
1587
unsigned long next;
arch/x86/mm/init_64.c
1595
for (; addr < end; addr = next) {
arch/x86/mm/init_64.c
1600
next = (addr + PAGE_SIZE) & PAGE_MASK;
arch/x86/mm/init_64.c
1607
next = (addr + PAGE_SIZE) & PAGE_MASK;
arch/x86/mm/init_64.c
1614
next = (addr + PAGE_SIZE) & PAGE_MASK;
arch/x86/mm/init_64.c
1621
next = (addr + PAGE_SIZE) & PAGE_MASK;
arch/x86/mm/init_64.c
1626
next = (addr + PAGE_SIZE) & PAGE_MASK;
arch/x86/mm/init_64.c
1636
next = pmd_addr_end(addr, end);
arch/x86/mm/init_64.c
1637
nr_pmd_pages = (next - addr) >> PAGE_SHIFT;
arch/x86/mm/ioremap.c
666
paddr_next = data->next;
arch/x86/mm/kasan_init_64.c
106
unsigned long next;
arch/x86/mm/kasan_init_64.c
116
next = pud_addr_end(addr, end);
arch/x86/mm/kasan_init_64.c
118
kasan_populate_pud(pud, addr, next, nid);
arch/x86/mm/kasan_init_64.c
119
} while (pud++, addr = next, addr != end);
arch/x86/mm/kasan_init_64.c
127
unsigned long next;
arch/x86/mm/kasan_init_64.c
136
next = p4d_addr_end(addr, end);
arch/x86/mm/kasan_init_64.c
137
kasan_populate_p4d(p4d, addr, next, nid);
arch/x86/mm/kasan_init_64.c
138
} while (p4d++, addr = next, addr != end);
arch/x86/mm/kasan_init_64.c
145
unsigned long next;
arch/x86/mm/kasan_init_64.c
151
next = pgd_addr_end(addr, end);
arch/x86/mm/kasan_init_64.c
152
kasan_populate_pgd(pgd, addr, next, nid);
arch/x86/mm/kasan_init_64.c
153
} while (pgd++, addr = next, addr != end);
arch/x86/mm/kasan_init_64.c
209
unsigned long next;
arch/x86/mm/kasan_init_64.c
219
next = p4d_addr_end(addr, end);
arch/x86/mm/kasan_init_64.c
227
} while (p4d++, addr = next, addr != end && p4d_none(*p4d));
arch/x86/mm/kasan_init_64.c
235
unsigned long next;
arch/x86/mm/kasan_init_64.c
239
next = pgd_addr_end(addr, end);
arch/x86/mm/kasan_init_64.c
240
kasan_early_p4d_populate(pgd, addr, next);
arch/x86/mm/kasan_init_64.c
241
} while (pgd++, addr = next, addr != end);
arch/x86/mm/kasan_init_64.c
249
unsigned long next;
arch/x86/mm/kasan_init_64.c
254
next = p4d_addr_end(addr, end);
arch/x86/mm/kasan_init_64.c
260
} while (p4d++, addr = next, addr != end);
arch/x86/mm/kasan_init_64.c
265
unsigned long addr, next;
arch/x86/mm/kasan_init_64.c
272
next = pgd_addr_end(addr, (unsigned long)end);
arch/x86/mm/kasan_init_64.c
283
kasan_shallow_populate_p4ds(pgd, addr, next);
arch/x86/mm/kasan_init_64.c
284
} while (pgd++, addr = next, addr != (unsigned long)end);
arch/x86/mm/kasan_init_64.c
76
unsigned long next;
arch/x86/mm/kasan_init_64.c
96
next = pmd_addr_end(addr, end);
arch/x86/mm/kasan_init_64.c
98
kasan_populate_pmd(pmd, addr, next, nid);
arch/x86/mm/kasan_init_64.c
99
} while (pmd++, addr = next, addr != end);
arch/x86/mm/kmmio.c
487
struct kmmio_fault_page *next = f->release_next;
arch/x86/mm/kmmio.c
490
f = next;
arch/x86/mm/pat/memtype.c
1022
.next = memtype_seq_next,
arch/x86/mm/tlb.c
225
static struct new_asid choose_new_asid(struct mm_struct *next, u64 next_tlb_gen)
arch/x86/mm/tlb.c
241
u16 global_asid = mm_global_asid(next);
arch/x86/mm/tlb.c
255
next->context.ctx_id)
arch/x86/mm/tlb.c
607
void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/x86/mm/tlb.c
613
switch_mm_irqs_off(NULL, next, tsk);
arch/x86/mm/tlb.c
629
struct task_struct *next)
arch/x86/mm/tlb.c
646
clear_ti_thread_flag(&next->thread_info, TIF_SPEC_L1D_FLUSH);
arch/x86/mm/tlb.c
647
next->l1d_flush_kill.func = l1d_flush_force_sigbus;
arch/x86/mm/tlb.c
648
task_work_add(next, &next->l1d_flush_kill, TWA_RESUME);
arch/x86/mm/tlb.c
652
static unsigned long mm_mangle_tif_spec_bits(struct task_struct *next)
arch/x86/mm/tlb.c
654
unsigned long next_tif = read_task_thread_flags(next);
arch/x86/mm/tlb.c
663
return (unsigned long)next->mm | spec_bits;
arch/x86/mm/tlb.c
666
static void cond_mitigation(struct task_struct *next)
arch/x86/mm/tlb.c
670
if (!next || !next->mm)
arch/x86/mm/tlb.c
673
next_mm = mm_mangle_tif_spec_bits(next);
arch/x86/mm/tlb.c
735
if ((prev_mm & ~LAST_USER_MM_SPEC_MASK) != (unsigned long)next->mm)
arch/x86/mm/tlb.c
746
l1d_flush_evaluate(prev_mm, next_mm, next);
arch/x86/mm/tlb.c
783
void switch_mm_irqs_off(struct mm_struct *unused, struct mm_struct *next,
arch/x86/mm/tlb.c
841
if (prev == next) {
arch/x86/mm/tlb.c
845
next->context.ctx_id);
arch/x86/mm/tlb.c
859
!cpumask_test_cpu(cpu, mm_cpumask(next))))
arch/x86/mm/tlb.c
860
cpumask_set_cpu(cpu, mm_cpumask(next));
arch/x86/mm/tlb.c
863
if (mm_needs_global_asid(next, prev_asid)) {
arch/x86/mm/tlb.c
864
next_tlb_gen = atomic64_read(&next->context.tlb_gen);
arch/x86/mm/tlb.c
865
ns = choose_new_asid(next, next_tlb_gen);
arch/x86/mm/tlb.c
891
next_tlb_gen = atomic64_read(&next->context.tlb_gen);
arch/x86/mm/tlb.c
935
if (next != &init_mm && !cpumask_test_cpu(cpu, mm_cpumask(next)))
arch/x86/mm/tlb.c
936
cpumask_set_cpu(cpu, mm_cpumask(next));
arch/x86/mm/tlb.c
940
next_tlb_gen = atomic64_read(&next->context.tlb_gen);
arch/x86/mm/tlb.c
942
ns = choose_new_asid(next, next_tlb_gen);
arch/x86/mm/tlb.c
946
new_lam = mm_lam_cr3_mask(next);
arch/x86/mm/tlb.c
949
this_cpu_write(cpu_tlbstate.ctxs[ns.asid].ctx_id, next->context.ctx_id);
arch/x86/mm/tlb.c
951
load_new_mm_cr3(next->pgd, ns.asid, new_lam, true);
arch/x86/mm/tlb.c
956
load_new_mm_cr3(next->pgd, ns.asid, new_lam, false);
arch/x86/mm/tlb.c
964
this_cpu_write(cpu_tlbstate.loaded_mm, next);
arch/x86/mm/tlb.c
966
cpu_tlbstate_update_lam(new_lam, mm_untag_mask(next));
arch/x86/mm/tlb.c
968
if (next != prev) {
arch/x86/mm/tlb.c
969
cr4_update_pce_mm(next);
arch/x86/mm/tlb.c
970
switch_ldt(prev, next);
arch/x86/pci/common.c
670
pa_data = data->next;
arch/x86/pci/i386.c
120
struct pcibios_fwaddrmap *entry, *next;
arch/x86/pci/i386.c
123
list_for_each_entry_safe(entry, next, &pcibios_fwaddrmappings, list) {
arch/x86/pci/mmconfig-shared.c
335
if (cfg->list.next == &pci_mmcfg_list)
arch/x86/pci/mmconfig-shared.c
338
cfgx = list_entry(cfg->list.next, typeof(*cfg), list);
arch/x86/virt/vmx/tdx/tdx.c
1031
u64 next;
arch/x86/virt/vmx/tdx/tdx.c
1052
next = args.rdx;
arch/x86/virt/vmx/tdx/tdx.c
1055
} while (next < tdmr->base + tdmr->size);
arch/x86/xen/enlighten_pv.c
433
static void xen_end_context_switch(struct task_struct *next)
arch/x86/xen/enlighten_pv.c
439
if (__task_lazy_mmu_mode_active(next))
arch/x86/xen/enlighten_pvh.c
101
struct boot_e820_entry *next;
arch/x86/xen/enlighten_pvh.c
109
next = e + 1;
arch/x86/xen/enlighten_pvh.c
110
memmove(next, e,
arch/x86/xen/enlighten_pvh.c
113
next->addr = PAGE_ALIGN(e->addr) + PFN_PHYS(pages);
arch/x86/xen/enlighten_pvh.c
114
e->size = next->addr - e->addr;
arch/x86/xen/enlighten_pvh.c
115
next->size -= e->size;
arch/xtensa/include/asm/mmu_context.h
126
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
arch/xtensa/include/asm/mmu_context.h
130
int migrated = next->context.cpu != cpu;
arch/xtensa/include/asm/mmu_context.h
134
next->context.cpu = cpu;
arch/xtensa/include/asm/mmu_context.h
136
if (migrated || prev != next)
arch/xtensa/include/asm/mmu_context.h
137
activate_context(next, cpu);
arch/xtensa/include/asm/pci-bridge.h
35
struct pci_controller *next;
arch/xtensa/include/asm/switch_to.h
15
extern void *_switch_to(void *last, void *next);
arch/xtensa/include/asm/switch_to.h
17
#define switch_to(prev,next,last) \
arch/xtensa/include/asm/switch_to.h
19
(last) = _switch_to(prev, next); \
arch/xtensa/kernel/asm-offsets.c
160
DEFINE(PBE_NEXT, offsetof(struct pbe, next));
arch/xtensa/kernel/setup.c
717
.next = c_next,
arch/xtensa/kernel/time.c
64
unsigned long flags, next;
arch/xtensa/kernel/time.c
68
next = get_ccount() + delta;
arch/xtensa/kernel/time.c
69
set_linux_timer(next);
arch/xtensa/kernel/time.c
70
if (next - get_ccount() > delta)
arch/xtensa/platforms/iss/network.c
630
struct list_head *ele, *next;
arch/xtensa/platforms/iss/network.c
634
list_for_each_safe(ele, next, &eth_cmd_line) {
block/badblocks.c
858
sector_t next = s + sectors;
block/badblocks.c
861
roundup(next, 1 << bb->shift);
block/badblocks.c
862
sectors = next - s;
block/bfq-iosched.c
1008
rq = rq_entry_fifo(bfqq->fifo.next);
block/bfq-iosched.c
1023
struct request *next, *prev = NULL;
block/bfq-iosched.c
1026
next = bfq_check_fifo(bfqq, last);
block/bfq-iosched.c
1027
if (next)
block/bfq-iosched.c
1028
return next;
block/bfq-iosched.c
1034
next = rb_entry_rq(rbnext);
block/bfq-iosched.c
1038
next = rb_entry_rq(rbnext);
block/bfq-iosched.c
1041
return bfq_choose_req(bfqd, next, prev, blk_rq_pos(last));
block/bfq-iosched.c
2556
struct request *next)
block/bfq-iosched.c
2559
*next_bfqq = RQ_BFQQ(next);
block/bfq-iosched.c
2574
!list_empty(&rq->queuelist) && !list_empty(&next->queuelist) &&
block/bfq-iosched.c
2575
next->fifo_time < rq->fifo_time) {
block/bfq-iosched.c
2577
list_replace_init(&next->queuelist, &rq->queuelist);
block/bfq-iosched.c
2578
rq->fifo_time = next->fifo_time;
block/bfq-iosched.c
2581
if (bfqq->next_rq == next)
block/bfq-iosched.c
2584
bfqg_stats_update_io_merged(bfqq_group(bfqq), next->cmd_flags);
block/bfq-iosched.c
2587
if (!RB_EMPTY_NODE(&next->rb_node)) {
block/bfq-iosched.c
2588
bfq_remove_request(next->q, next);
block/bfq-iosched.c
2591
next->cmd_flags);
block/bfq-iosched.c
5419
struct bfq_queue *__bfqq, *next;
block/bfq-iosched.c
5428
next = __bfqq->new_bfqq;
block/bfq-iosched.c
5430
__bfqq = next;
block/bfq-wf2q.c
353
struct rb_node *next;
block/bfq-wf2q.c
356
next = rb_next(&entity->rb_node);
block/bfq-wf2q.c
357
st->first_idle = bfq_entity_of(next);
block/bfq-wf2q.c
361
next = rb_prev(&entity->rb_node);
block/bfq-wf2q.c
362
st->last_idle = bfq_entity_of(next);
block/bio-integrity.c
284
size_t next = min_t(size_t, PAGE_SIZE, bytes);
block/bio-integrity.c
290
size += next;
block/bio-integrity.c
291
bytes -= next;
block/bio.c
1680
struct bio *bio, *next;
block/bio.c
1683
next = bio_dirty_list;
block/bio.c
1687
while ((bio = next) != NULL) {
block/bio.c
1688
next = bio->bi_private;
block/blk-crypto-internal.h
125
struct request *next)
block/blk-crypto-internal.h
52
struct request *next)
block/blk-crypto-internal.h
55
next->crypt_ctx);
block/blk-integrity.c
141
struct request *next)
block/blk-integrity.c
145
if (blk_integrity_rq(req) == 0 && blk_integrity_rq(next) == 0)
block/blk-integrity.c
148
if (blk_integrity_rq(req) == 0 || blk_integrity_rq(next) == 0)
block/blk-integrity.c
152
bip_next = bio_integrity(next->bio);
block/blk-integrity.c
160
if (req->nr_integrity_segments + next->nr_integrity_segments >
block/blk-integrity.c
164
if (integrity_req_gap_back_merge(req, next->bio))
block/blk-iolatency.c
676
goto next;
block/blk-iolatency.c
682
goto next;
block/blk-iolatency.c
707
next:
block/blk-merge.c
52
struct request *prev_rq, struct bio *prev, struct bio *next)
block/blk-merge.c
613
struct request *next)
block/blk-merge.c
619
if (blk_rq_sectors(req) + bio_sectors(next->bio) >
block/blk-merge.c
623
req->nr_phys_segments = segments + blk_rq_nr_discard_segments(next);
block/blk-merge.c
631
struct request *next)
block/blk-merge.c
635
if (req_gap_back_merge(req, next->bio))
block/blk-merge.c
641
if ((blk_rq_sectors(req) + blk_rq_sectors(next)) >
block/blk-merge.c
645
total_phys_segments = req->nr_phys_segments + next->nr_phys_segments;
block/blk-merge.c
649
if (!blk_cgroup_mergeable(req, next->bio))
block/blk-merge.c
652
if (blk_integrity_merge_rq(q, req, next) == false)
block/blk-merge.c
655
if (!bio_crypt_ctx_merge_rq(req, next))
block/blk-merge.c
660
req->nr_integrity_segments += next->nr_integrity_segments;
block/blk-merge.c
733
struct request *next)
block/blk-merge.c
737
else if (blk_rq_pos(req) + blk_rq_sectors(req) == blk_rq_pos(next))
block/blk-merge.c
750
struct request *next)
block/blk-merge.c
752
return (rq->cmd_flags & REQ_ATOMIC) == (next->cmd_flags & REQ_ATOMIC);
block/blk-merge.c
755
u8 bio_seg_gap(struct request_queue *q, struct bio *prev, struct bio *next,
block/blk-merge.c
764
gaps_bit = min_not_zero(gaps_bit, next->bi_bvec_gap_bit);
block/blk-merge.c
767
bio_get_first_bvec(next, &nb);
block/blk-merge.c
778
struct request *req, struct request *next)
block/blk-merge.c
780
if (!rq_mergeable(req) || !rq_mergeable(next))
block/blk-merge.c
783
if (req_op(req) != req_op(next))
block/blk-merge.c
786
if (req->bio->bi_write_hint != next->bio->bi_write_hint)
block/blk-merge.c
788
if (req->bio->bi_write_stream != next->bio->bi_write_stream)
block/blk-merge.c
790
if (req->bio->bi_ioprio != next->bio->bi_ioprio)
block/blk-merge.c
792
if (!blk_atomic_write_mergeable_rqs(req, next))
block/blk-merge.c
803
switch (blk_try_req_merge(req, next)) {
block/blk-merge.c
805
if (!req_attempt_discard_merge(q, req, next))
block/blk-merge.c
809
if (!ll_merge_requests_fn(q, req, next))
block/blk-merge.c
81
bio_get_first_bvec(next, &nb);
block/blk-merge.c
822
if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) ||
block/blk-merge.c
824
(next->cmd_flags & REQ_FAILFAST_MASK)) {
block/blk-merge.c
826
blk_rq_set_mixed_merge(next);
block/blk-merge.c
834
if (next->start_time_ns < req->start_time_ns)
block/blk-merge.c
835
req->start_time_ns = next->start_time_ns;
block/blk-merge.c
837
req->phys_gap_bit = bio_seg_gap(req->q, req->biotail, next->bio,
block/blk-merge.c
838
min_not_zero(next->phys_gap_bit,
block/blk-merge.c
840
req->biotail->bi_next = next->bio;
block/blk-merge.c
841
req->biotail = next->biotail;
block/blk-merge.c
843
req->__data_len += blk_rq_bytes(next);
block/blk-merge.c
846
elv_merge_requests(q, req, next);
block/blk-merge.c
848
blk_crypto_rq_put_keyslot(next);
block/blk-merge.c
853
blk_account_io_merge_request(next);
block/blk-merge.c
855
trace_block_rq_merge(next);
block/blk-merge.c
861
next->bio = NULL;
block/blk-merge.c
862
return next;
block/blk-merge.c
868
struct request *next = elv_latter_request(q, rq);
block/blk-merge.c
870
if (next)
block/blk-merge.c
871
return attempt_merge(q, rq, next);
block/blk-merge.c
893
struct request *next)
block/blk-merge.c
895
return attempt_merge(q, rq, next);
block/blk-mq-debugfs.c
325
.next = hctx_dispatch_next,
block/blk-mq-debugfs.c
48
.next = queue_requeue_list_next,
block/blk-mq-debugfs.c
514
.next = ctx_##name##_rq_list_next, \
block/blk-mq-debugfs.c
779
rqos = rqos->next;
block/blk-mq-dma.c
48
struct bio_vec next;
block/blk-mq-dma.c
53
next = mp_bvec_iter_bvec(iter->bvecs, iter->iter);
block/blk-mq-dma.c
54
if (bv.bv_len + next.bv_len > max_size ||
block/blk-mq-dma.c
55
!biovec_phys_mergeable(req->q, &bv, &next))
block/blk-mq-dma.c
58
bv.bv_len += next.bv_len;
block/blk-mq-dma.c
59
bvec_iter_advance_single(iter->bvecs, &iter->iter, next.bv_len);
block/blk-mq.c
1250
struct request *rq, *next;
block/blk-mq.c
1252
llist_for_each_entry_safe(rq, next, entry, ipi_list)
block/blk-mq.c
1582
rq = list_entry(rq_list.next, struct request, queuelist);
block/blk-mq.c
1597
rq = list_entry(flush_list.next, struct request, queuelist);
block/blk-mq.c
1674
unsigned long next;
block/blk-mq.c
1691
if (expired->next == 0)
block/blk-mq.c
1692
expired->next = deadline;
block/blk-mq.c
1693
else if (time_after(expired->next, deadline))
block/blk-mq.c
1694
expired->next = deadline;
block/blk-mq.c
1772
expired.next = 0;
block/blk-mq.c
1776
if (expired.next != 0) {
block/blk-mq.c
1777
mod_timer(&q->timeout, expired.next);
block/blk-mq.c
1842
dispatch_data->rq = list_entry_rq(ctx->rq_lists[type].next);
block/blk-mq.c
4395
struct blk_mq_hw_ctx *hctx, *next;
block/blk-mq.c
4402
list_for_each_entry_safe(hctx, next, &q->unused_hctx_list, hctx_list) {
block/blk-mq.c
908
struct bio *next = bio->bi_next;
block/blk-mq.c
918
bio = next;
block/blk-mq.h
392
struct request *rq = list_entry_rq(list->next);
block/blk-rq-qos.c
103
rqos = rqos->next;
block/blk-rq-qos.c
31
rqos = rqos->next;
block/blk-rq-qos.c
318
q->rq_qos = rqos->next;
block/blk-rq-qos.c
345
rqos->next = q->rq_qos;
block/blk-rq-qos.c
365
for (cur = &q->rq_qos; *cur; cur = &(*cur)->next) {
block/blk-rq-qos.c
367
*cur = rqos->next;
block/blk-rq-qos.c
40
rqos = rqos->next;
block/blk-rq-qos.c
49
rqos = rqos->next;
block/blk-rq-qos.c
58
rqos = rqos->next;
block/blk-rq-qos.c
67
rqos = rqos->next;
block/blk-rq-qos.c
76
rqos = rqos->next;
block/blk-rq-qos.c
85
rqos = rqos->next;
block/blk-rq-qos.c
94
rqos = rqos->next;
block/blk-rq-qos.h
31
struct rq_qos *next;
block/blk-rq-qos.h
65
for (rqos = q->rq_qos; rqos; rqos = rqos->next) {
block/blk.h
254
struct bio *next)
block/blk.h
257
struct bio_integrity_payload *bip_next = bio_integrity(next);
block/blk.h
288
struct bio *next)
block/blk.h
453
struct request *next);
block/elevator.c
197
struct hlist_node *next;
block/elevator.c
200
hash_for_each_possible_safe(e->hash, rq, next, hash, offset) {
block/elevator.c
377
struct request *next)
block/elevator.c
382
e->type->ops.requests_merged(q, rq, next);
block/genhd.c
1418
.next = disk_seqf_next,
block/genhd.c
179
struct blk_major_name *next;
block/genhd.c
201
for (dp = major_names[major_to_index(offset)]; dp; dp = dp->next)
block/genhd.c
278
p->next = NULL;
block/genhd.c
282
for (n = &major_names[index]; *n; n = &(*n)->next) {
block/genhd.c
311
for (n = &major_names[index]; *n; n = &(*n)->next)
block/genhd.c
318
*n = p->next;
block/genhd.c
882
for (n = &major_names[major_to_index(major)]; *n; n = &(*n)->next) {
block/genhd.c
989
.next = disk_seqf_next,
block/kyber-iosched.c
581
struct request *rq, *next;
block/kyber-iosched.c
583
list_for_each_entry_safe(rq, next, rq_list, queuelist) {
block/kyber-iosched.c
926
.next = kyber_##name##_rqs_next, \
block/mq-deadline.c
205
struct request *next)
block/mq-deadline.c
208
const u8 ioprio_class = dd_rq_ioclass(next);
block/mq-deadline.c
219
if (!list_empty(&req->queuelist) && !list_empty(&next->queuelist)) {
block/mq-deadline.c
220
if (time_before((unsigned long)next->fifo_time,
block/mq-deadline.c
222
list_move(&req->queuelist, &next->queuelist);
block/mq-deadline.c
223
req->fifo_time = next->fifo_time;
block/mq-deadline.c
230
deadline_remove_request(q, &dd->per_prio[prio], next);
block/mq-deadline.c
263
struct request *rq = rq_entry_fifo(per_prio->fifo_list[data_dir].next);
block/mq-deadline.c
279
return rq_entry_fifo(per_prio->fifo_list[data_dir].next);
block/mq-deadline.c
832
.next = deadline_##name##_fifo_next, \
block/mq-deadline.c
952
.next = deadline_dispatch_next,
block/partitions/acorn.c
529
sector_t next;
block/partitions/acorn.c
534
next = le32_to_cpu(p->start);
block/partitions/acorn.c
536
put_partition(state, slot++, start, next - start);
block/partitions/acorn.c
537
start = next;
block/partitions/check.h
20
int next;
block/partitions/cmdline.c
119
char *next;
block/partitions/cmdline.c
129
next = strsep(&bdevdef, ":");
block/partitions/cmdline.c
130
if (!next) {
block/partitions/cmdline.c
135
strscpy(newparts->name, next, sizeof(newparts->name));
block/partitions/cmdline.c
140
while ((next = strsep(&bdevdef, ","))) {
block/partitions/cmdline.c
141
ret = parse_subpart(next_subpart, next);
block/partitions/cmdline.c
182
char *next;
block/partitions/cmdline.c
193
while ((next = strsep(&pbuf, ";"))) {
block/partitions/cmdline.c
194
ret = parse_parts(next_parts, next);
block/partitions/cmdline.c
74
char *next = strsep(&partdef, ")");
block/partitions/cmdline.c
76
if (!next) {
block/partitions/cmdline.c
82
strscpy(new_subpart->name, next, sizeof(new_subpart->name));
block/partitions/msdos.c
150
if (state->next == state->limit)
block/partitions/msdos.c
174
sector_t offs, size, next;
block/partitions/msdos.c
183
next = this_sector + offs;
block/partitions/msdos.c
187
if (next < first_sector)
block/partitions/msdos.c
189
if (next + size > first_sector + first_size)
block/partitions/msdos.c
193
put_partition(state, state->next, next, size);
block/partitions/msdos.c
194
set_info(state, state->next, disksig);
block/partitions/msdos.c
196
state->parts[state->next].flags = ADDPART_FLAG_RAID;
block/partitions/msdos.c
198
if (++state->next == state->limit)
block/partitions/msdos.c
283
for (i = 0; i < max_nparts && state->next < state->limit; i++) {
block/partitions/msdos.c
293
put_partition(state, state->next++,
block/partitions/msdos.c
380
if (state->next == state->limit)
block/partitions/msdos.c
397
put_partition(state, state->next++, bsd_start, bsd_size);
block/partitions/msdos.c
508
if (state->next == state->limit)
block/partitions/msdos.c
512
put_partition(state, state->next++,
block/partitions/msdos.c
554
if (state->next == state->limit)
block/partitions/msdos.c
558
put_partition(state, state->next++,
block/partitions/msdos.c
662
state->next = 5;
block/sed-opal.c
2489
struct opal_suspend_data *suspend, *next;
block/sed-opal.c
2492
list_for_each_entry_safe(suspend, next, &dev->unlk_lst, node) {
block/t10-pi.c
106
next:
block/t10-pi.c
262
goto next;
block/t10-pi.c
274
goto next;
block/t10-pi.c
290
next:
block/t10-pi.c
76
goto next;
block/t10-pi.c
89
goto next;
crypto/algapi.c
225
if (spawns->next == NULL)
crypto/algapi.c
664
struct crypto_spawn *next;
crypto/algapi.c
669
next = spawn->next;
crypto/algapi.c
677
spawn = next;
crypto/algapi.c
744
spawn->next = inst->spawns;
crypto/algapi.c
992
queue->backlog = queue->backlog->next;
crypto/algapi.c
994
request = queue->list.next;
crypto/asymmetric_keys/asymmetric_type.c
509
char *next;
crypto/asymmetric_keys/asymmetric_type.c
524
next = parse_buf;
crypto/asymmetric_keys/asymmetric_type.c
525
restrict_method = strsep(&next, ":");
crypto/asymmetric_keys/asymmetric_type.c
527
if ((strcmp(restrict_method, "key_or_keyring") == 0) && next) {
crypto/asymmetric_keys/asymmetric_type.c
535
key_text = strsep(&next, ":");
crypto/asymmetric_keys/asymmetric_type.c
537
if (next) {
crypto/asymmetric_keys/asymmetric_type.c
538
if (strcmp(next, "chain") != 0)
crypto/asymmetric_keys/pkcs7_parser.c
110
for (sinfo = sinfo->next; sinfo; sinfo = sinfo->next) {
crypto/asymmetric_keys/pkcs7_parser.c
172
ctx->certs = cert->next;
crypto/asymmetric_keys/pkcs7_parser.c
466
ctx->ppcerts = &x509->next;
crypto/asymmetric_keys/pkcs7_parser.c
65
pkcs7->certs = cert->next;
crypto/asymmetric_keys/pkcs7_parser.c
70
pkcs7->crl = cert->next;
crypto/asymmetric_keys/pkcs7_parser.c
730
ctx->ppsinfo = &sinfo->next;
crypto/asymmetric_keys/pkcs7_parser.c
75
pkcs7->signed_infos = sinfo->next;
crypto/asymmetric_keys/pkcs7_parser.h
18
struct pkcs7_signed_info *next;
crypto/asymmetric_keys/pkcs7_trust.c
166
for (p = pkcs7->certs; p; p = p->next)
crypto/asymmetric_keys/pkcs7_trust.c
169
for (sinfo = pkcs7->signed_infos; sinfo; sinfo = sinfo->next) {
crypto/asymmetric_keys/pkcs7_verify.c
147
if (sinfo == NULL || sinfo->next != NULL)
crypto/asymmetric_keys/pkcs7_verify.c
184
for (x509 = pkcs7->certs; x509; x509 = x509->next, certix++) {
crypto/asymmetric_keys/pkcs7_verify.c
221
for (p = pkcs7->certs; p; p = p->next)
crypto/asymmetric_keys/pkcs7_verify.c
269
for (p = pkcs7->certs; p; p = p->next) {
crypto/asymmetric_keys/pkcs7_verify.c
278
for (p = pkcs7->certs; p; p = p->next) {
crypto/asymmetric_keys/pkcs7_verify.c
466
for (sinfo = pkcs7->signed_infos; sinfo; sinfo = sinfo->next) {
crypto/asymmetric_keys/x509_parser.h
15
struct x509_certificate *next;
crypto/jitterentropy.c
300
static inline __u64 jent_delta(__u64 prev, __u64 next)
crypto/jitterentropy.c
303
return (prev < next) ? (next - prev) :
crypto/jitterentropy.c
304
(JENT_UINT64_MAX - prev + 1 + next);
crypto/proc.c
90
.next = c_next,
crypto/rsa-pkcs1pad.c
64
struct scatterlist *next)
crypto/rsa-pkcs1pad.c
66
int nsegs = next ? 2 : 1;
crypto/rsa-pkcs1pad.c
71
if (next)
crypto/rsa-pkcs1pad.c
72
sg_chain(sg, nsegs, next);
crypto/xor.c
148
for (f = fastest; f; f = f->next)
crypto/xor.c
60
tmpl->next = template_list;
crypto/xor.c
89
tmpl->next = template_list;
drivers/accel/habanalabs/common/command_submission.c
508
struct hl_cs *next = NULL, *iter, *first_cs;
drivers/accel/habanalabs/common/command_submission.c
545
next = iter;
drivers/accel/habanalabs/common/command_submission.c
549
if (next && !next->tdr_active) {
drivers/accel/habanalabs/common/command_submission.c
550
next->tdr_active = true;
drivers/accel/habanalabs/common/command_submission.c
551
schedule_delayed_work(&next->work_tdr, next->timeout_jiffies);
drivers/accel/habanalabs/common/memory.c
449
struct hl_vm_va_block *prev, *next;
drivers/accel/habanalabs/common/memory.c
460
next = list_next_entry(va_block, node);
drivers/accel/habanalabs/common/memory.c
461
if (&next->node != va_list && va_block->end + 1 == next->start) {
drivers/accel/habanalabs/common/memory.c
462
next->start = va_block->start;
drivers/accel/habanalabs/common/memory.c
463
next->size = next->end - next->start + 1;
drivers/accel/ivpu/ivpu_fw_log.c
129
u32 next = 0;
drivers/accel/ivpu/ivpu_fw_log.c
131
while (fw_log_from_bo(vdev, bo, &next, &log) == 0)
drivers/accel/ivpu/ivpu_fw_log.c
144
u32 next;
drivers/accel/ivpu/ivpu_fw_log.c
146
next = 0;
drivers/accel/ivpu/ivpu_fw_log.c
147
while (fw_log_from_bo(vdev, vdev->fw->mem_log_crit, &next, &log) == 0) {
drivers/accel/ivpu/ivpu_fw_log.c
152
next = 0;
drivers/accel/ivpu/ivpu_fw_log.c
153
while (fw_log_from_bo(vdev, vdev->fw->mem_log_verb, &next, &log) == 0) {
drivers/accel/ivpu/ivpu_fw_log.c
162
u32 next;
drivers/accel/ivpu/ivpu_fw_log.c
164
next = 0;
drivers/accel/ivpu/ivpu_fw_log.c
165
while (fw_log_from_bo(vdev, vdev->fw->mem_log_crit, &next, &log) == 0) {
drivers/accel/ivpu/ivpu_fw_log.c
170
next = 0;
drivers/accel/ivpu/ivpu_fw_log.c
171
while (fw_log_from_bo(vdev, vdev->fw->mem_log_verb, &next, &log) == 0) {
drivers/accessibility/speakup/makemapdata.c
121
this = this->next;
drivers/accessibility/speakup/utils.h
20
struct st_key *next;
drivers/accessibility/speakup/utils.h
78
this = this->next;
drivers/accessibility/speakup/utils.h
90
while (this->next) {
drivers/accessibility/speakup/utils.h
93
this = this->next;
drivers/accessibility/speakup/utils.h
95
this->next = extra_keys++;
drivers/accessibility/speakup/utils.h
96
this = this->next;
drivers/acpi/acpi_video.c
1967
struct acpi_video_device *dev, *next;
drivers/acpi/acpi_video.c
1970
list_for_each_entry_safe(dev, next, &video->video_device_list, entry) {
drivers/acpi/acpica/acdebug.h
251
char **next, acpi_object_type *return_type);
drivers/acpi/acpica/aclocal.h
1063
struct acpi_interface_info *next;
drivers/acpi/acpica/aclocal.h
1163
struct acpi_external_list *next;
drivers/acpi/acpica/aclocal.h
1181
struct acpi_external_file *next;
drivers/acpi/acpica/aclocal.h
1186
struct acpi_parse_object_list *next;
drivers/acpi/acpica/aclocal.h
1257
struct acpi_debug_mem_block *next; \
drivers/acpi/acpica/aclocal.h
193
struct acpi_new_table_desc *next;
drivers/acpi/acpica/aclocal.h
412
struct acpi_sci_handler_info *next;
drivers/acpi/acpica/aclocal.h
431
struct acpi_gpe_notify_info *next;
drivers/acpi/acpica/aclocal.h
483
struct acpi_gpe_block_info *next;
drivers/acpi/acpica/aclocal.h
499
struct acpi_gpe_xrupt_info *next;
drivers/acpi/acpica/aclocal.h
549
struct acpi_ged_handler_info *next;
drivers/acpi/acpica/aclocal.h
567
void *next; \
drivers/acpi/acpica/aclocal.h
703
struct acpi_address_range *next;
drivers/acpi/acpica/aclocal.h
759
union acpi_parse_object *next; /* Next op */\
drivers/acpi/acpica/aclocal.h
818
struct acpi_comment_node *next;
drivers/acpi/acpica/aclocal.h
823
struct acpi_comment_addr_node *next;
drivers/acpi/acpica/aclocal.h
835
struct acpi_file_node *next;
drivers/acpi/acpica/acobject.h
136
union acpi_operand_object *next; /* Link for list of acquired mutexes */
drivers/acpi/acpica/acobject.h
146
union acpi_operand_object *next;
drivers/acpi/acpica/acobject.h
309
union acpi_operand_object *next[2]; /* Device and System handler lists */
drivers/acpi/acpica/acobject.h
322
union acpi_operand_object *next;
drivers/acpi/acpica/acobject.h
395
union acpi_operand_object *next; /* Link for object cache and internal lists */
drivers/acpi/acpica/acstruct.h
38
struct acpi_walk_state *next; /* Next walk_state in list */
drivers/acpi/acpica/amlresrc.h
90
struct asl_resource_node *next;
drivers/acpi/acpica/dbcmds.c
1067
ged_info = ged_info->next;
drivers/acpi/acpica/dbcmds.c
429
next_interface = next_interface->next;
drivers/acpi/acpica/dbconvert.c
168
char *next;
drivers/acpi/acpica/dbconvert.c
182
this = acpi_db_get_next_token(this, &next, &type);
drivers/acpi/acpica/dbconvert.c
196
this = next;
drivers/acpi/acpica/dbdisply.c
1091
handler_obj = handler_obj->address_space.next;
drivers/acpi/acpica/dbdisply.c
124
acpi_os_printf("%20.20s : %p\n", "NextOp", op->common.next);
drivers/acpi/acpica/dbdisply.c
538
frame = frame->results.next;
drivers/acpi/acpica/dbdisply.c
575
walk_state = walk_state->next;
drivers/acpi/acpica/dbdisply.c
883
notify = notify->next;
drivers/acpi/acpica/dbdisply.c
910
gpe_block = gpe_block->next;
drivers/acpi/acpica/dbdisply.c
913
gpe_xrupt_info = gpe_xrupt_info->next;
drivers/acpi/acpica/dbdisply.c
992
handler_obj = handler_obj->address_space.next;
drivers/acpi/acpica/dbfileio.c
114
table_list_head = table_list_head->next;
drivers/acpi/acpica/dbinput.c
399
const struct acpi_db_command_help *next;
drivers/acpi/acpica/dbinput.c
402
next = acpi_gbl_db_command_help;
drivers/acpi/acpica/dbinput.c
403
while (next->invocation) {
drivers/acpi/acpica/dbinput.c
404
matched = acpi_db_match_command_help(command, next);
drivers/acpi/acpica/dbinput.c
409
next++;
drivers/acpi/acpica/dbinput.c
429
const struct acpi_db_command_help *next = acpi_gbl_db_command_help;
drivers/acpi/acpica/dbinput.c
437
while (next->invocation) {
drivers/acpi/acpica/dbinput.c
438
acpi_os_printf("%-38s%s", next->invocation,
drivers/acpi/acpica/dbinput.c
439
next->description);
drivers/acpi/acpica/dbinput.c
440
next++;
drivers/acpi/acpica/dbinput.c
465
char **next, acpi_object_type *return_type)
drivers/acpi/acpica/dbinput.c
584
*next = NULL;
drivers/acpi/acpica/dbinput.c
587
*next = string + 1;
drivers/acpi/acpica/dbinput.c
611
char *next;
drivers/acpi/acpica/dbinput.c
625
acpi_gbl_db_args[i] = acpi_db_get_next_token(this, &next,
drivers/acpi/acpica/dbinput.c
632
this = next;
drivers/acpi/acpica/dbutils.c
370
start_op = (method->common.value.arg)->common.next;
drivers/acpi/acpica/dbxface.c
205
union acpi_parse_object *next;
drivers/acpi/acpica/dbxface.c
291
next = op->common.next;
drivers/acpi/acpica/dbxface.c
292
op->common.next = NULL;
drivers/acpi/acpica/dbxface.c
325
op->common.next = next;
drivers/acpi/acpica/dsdebug.c
161
next = op->common.next;
drivers/acpi/acpica/dsdebug.c
162
op->common.next = NULL;
drivers/acpi/acpica/dsdebug.c
175
op->common.next = next;
drivers/acpi/acpica/dsdebug.c
189
next_walk_state = next_walk_state->next;
drivers/acpi/acpica/dsdebug.c
93
union acpi_parse_object *next;
drivers/acpi/acpica/dsfield.c
452
arg = arg->common.next;
drivers/acpi/acpica/dsfield.c
510
arg = arg->common.next;
drivers/acpi/acpica/dsfield.c
519
status = acpi_ds_get_field_names(&info, walk_state, arg->common.next);
drivers/acpi/acpica/dsfield.c
652
arg = arg->common.next;
drivers/acpi/acpica/dsfield.c
707
arg = arg->common.next;
drivers/acpi/acpica/dsfield.c
724
arg = arg->common.next;
drivers/acpi/acpica/dsfield.c
728
arg = arg->common.next;
drivers/acpi/acpica/dsfield.c
747
status = acpi_ds_get_field_names(&info, walk_state, arg->common.next);
drivers/acpi/acpica/dsfield.c
792
arg = arg->common.next;
drivers/acpi/acpica/dsfield.c
806
arg = arg->common.next;
drivers/acpi/acpica/dsfield.c
814
status = acpi_ds_get_field_names(&info, walk_state, arg->common.next);
drivers/acpi/acpica/dsobject.c
187
byte_list = arg->named.next;
drivers/acpi/acpica/dsopcode.c
378
next_op = next_op->common.next;
drivers/acpi/acpica/dsopcode.c
704
next_op = next_op->common.next;
drivers/acpi/acpica/dsopcode.c
708
next_op = next_op->common.next;
drivers/acpi/acpica/dsopcode.c
757
arg = arg->common.next;
drivers/acpi/acpica/dspkginit.c
130
arg = arg->common.next;
drivers/acpi/acpica/dspkginit.c
246
arg = arg->common.next;
drivers/acpi/acpica/dspkginit.c
280
arg = arg->common.next;
drivers/acpi/acpica/dsutils.c
690
arg = arg->common.next;
drivers/acpi/acpica/dswexec.c
500
next_op = next_op->common.next;
drivers/acpi/acpica/dswload.c
517
(u8)op->common.value.arg->common.next->common.value.integer;
drivers/acpi/acpica/dswscope.c
40
walk_state->scope_info = scope_info->scope.next;
drivers/acpi/acpica/dswstate.c
446
walk_state->next = thread->walk_state_list;
drivers/acpi/acpica/dswstate.c
478
thread->walk_state_list = walk_state->next;
drivers/acpi/acpica/dswstate.c
703
walk_state->control_state = state->common.next;
drivers/acpi/acpica/dswstate.c
712
walk_state->scope_info = state->common.next;
drivers/acpi/acpica/dswstate.c
721
walk_state->results = state->common.next;
drivers/acpi/acpica/evgpe.c
432
gpe_block = gpe_block->next;
drivers/acpi/acpica/evgpe.c
485
notify = notify->next;
drivers/acpi/acpica/evgpeblk.c
117
if (!gpe_block->previous && !gpe_block->next) {
drivers/acpi/acpica/evgpeblk.c
130
gpe_block->previous->next = gpe_block->next;
drivers/acpi/acpica/evgpeblk.c
133
gpe_block->next;
drivers/acpi/acpica/evgpeblk.c
136
if (gpe_block->next) {
drivers/acpi/acpica/evgpeblk.c
137
gpe_block->next->previous = gpe_block->previous;
drivers/acpi/acpica/evgpeblk.c
67
while (next_gpe_block->next) {
drivers/acpi/acpica/evgpeblk.c
68
next_gpe_block = next_gpe_block->next;
drivers/acpi/acpica/evgpeblk.c
71
next_gpe_block->next = gpe_block;
drivers/acpi/acpica/evgpeinit.c
251
gpe_block = gpe_block->next;
drivers/acpi/acpica/evgpeinit.c
254
gpe_xrupt_info = gpe_xrupt_info->next;
drivers/acpi/acpica/evgpeutil.c
150
next_gpe_xrupt = next_gpe_xrupt->next;
drivers/acpi/acpica/evgpeutil.c
167
while (next_gpe_xrupt->next) {
drivers/acpi/acpica/evgpeutil.c
168
next_gpe_xrupt = next_gpe_xrupt->next;
drivers/acpi/acpica/evgpeutil.c
171
next_gpe_xrupt->next = gpe_xrupt;
drivers/acpi/acpica/evgpeutil.c
237
gpe_xrupt->previous->next = gpe_xrupt->next;
drivers/acpi/acpica/evgpeutil.c
241
acpi_gbl_gpe_xrupt_list_head = gpe_xrupt->next;
drivers/acpi/acpica/evgpeutil.c
244
if (gpe_xrupt->next) {
drivers/acpi/acpica/evgpeutil.c
245
gpe_xrupt->next->previous = gpe_xrupt->previous;
drivers/acpi/acpica/evgpeutil.c
276
struct acpi_gpe_notify_info *next;
drivers/acpi/acpica/evgpeutil.c
311
next = notify->next;
drivers/acpi/acpica/evgpeutil.c
313
notify = next;
drivers/acpi/acpica/evgpeutil.c
64
gpe_block = gpe_block->next;
drivers/acpi/acpica/evgpeutil.c
67
gpe_xrupt_info = gpe_xrupt_info->next;
drivers/acpi/acpica/evhandler.c
140
handler_obj = handler_obj->address_space.next;
drivers/acpi/acpica/evhandler.c
304
handler_obj = handler_obj->address_space.next;
drivers/acpi/acpica/evhandler.c
509
handler_obj->address_space.next = obj_desc->common_notify.handler;
drivers/acpi/acpica/evmisc.c
185
handler_obj->notify.next[info->notify.handler_list_id];
drivers/acpi/acpica/evregion.c
393
*last_obj_ptr = obj_desc->region.next;
drivers/acpi/acpica/evregion.c
394
obj_desc->region.next = NULL; /* Must clear field */
drivers/acpi/acpica/evregion.c
475
last_obj_ptr = &obj_desc->region.next;
drivers/acpi/acpica/evregion.c
476
obj_desc = obj_desc->region.next;
drivers/acpi/acpica/evregion.c
535
region_obj->region.next = handler_obj->address_space.region_list;
drivers/acpi/acpica/evsci.c
206
acpi_gbl_sci_handler_list = sci_handler->next;
drivers/acpi/acpica/evsci.c
56
sci_handler = sci_handler->next;
drivers/acpi/acpica/evxface.c
152
handler_obj = handler_obj->notify.next[i];
drivers/acpi/acpica/evxface.c
174
handler_obj->notify.next[i] =
drivers/acpi/acpica/evxface.c
296
handler_obj = handler_obj->notify.next[i];
drivers/acpi/acpica/evxface.c
307
previous_handler_obj->notify.next[i] =
drivers/acpi/acpica/evxface.c
308
handler_obj->notify.next[i];
drivers/acpi/acpica/evxface.c
312
handler_obj->notify.next[i];
drivers/acpi/acpica/evxface.c
430
sci_handler = sci_handler->next;
drivers/acpi/acpica/evxface.c
435
new_sci_handler->next = acpi_gbl_sci_handler_list;
drivers/acpi/acpica/evxface.c
493
prev_sci_handler->next = next_sci_handler->next;
drivers/acpi/acpica/evxface.c
496
next_sci_handler->next;
drivers/acpi/acpica/evxface.c
505
next_sci_handler = next_sci_handler->next;
drivers/acpi/acpica/evxfgpe.c
448
notify = notify->next;
drivers/acpi/acpica/evxfgpe.c
454
new_notify->next = gpe_event_info->dispatch.notify_list;
drivers/acpi/acpica/evxfregn.c
231
*last_obj_ptr = handler_obj->address_space.next;
drivers/acpi/acpica/evxfregn.c
241
last_obj_ptr = &handler_obj->address_space.next;
drivers/acpi/acpica/evxfregn.c
242
handler_obj = handler_obj->address_space.next;
drivers/acpi/acpica/exdump.c
122
{ACPI_EXD_POINTER, ACPI_EXD_OFFSET(region.next), "Next"}
drivers/acpi/acpica/exdump.c
210
{ACPI_EXD_HDLR_LIST, ACPI_EXD_OFFSET(address_space.next), "Next"},
drivers/acpi/acpica/exdump.c
223
{ACPI_EXD_POINTER, ACPI_EXD_OFFSET(notify.next[0]),
drivers/acpi/acpica/exdump.c
225
{ACPI_EXD_POINTER, ACPI_EXD_OFFSET(notify.next[1]), "Next Device Notify"}
drivers/acpi/acpica/exdump.c
338
union acpi_operand_object *next;
drivers/acpi/acpica/exdump.c
440
next = start;
drivers/acpi/acpica/exdump.c
442
acpi_os_printf("%20s : %p ", name, next);
drivers/acpi/acpica/exdump.c
443
if (next) {
drivers/acpi/acpica/exdump.c
446
(next), next->common.type);
drivers/acpi/acpica/exdump.c
448
while (next->common.next_object) {
drivers/acpi/acpica/exdump.c
449
if ((next->common.type ==
drivers/acpi/acpica/exdump.c
451
data = next;
drivers/acpi/acpica/exdump.c
454
next = next->common.next_object;
drivers/acpi/acpica/exdump.c
455
acpi_os_printf("->%p(%s %2.2X)", next,
drivers/acpi/acpica/exdump.c
457
(next),
drivers/acpi/acpica/exdump.c
458
next->common.type);
drivers/acpi/acpica/exdump.c
460
if ((next == start) || (next == data)) {
drivers/acpi/acpica/exdump.c
476
next = start;
drivers/acpi/acpica/exdump.c
478
acpi_os_printf("%20s : %p", name, next);
drivers/acpi/acpica/exdump.c
479
if (next) {
drivers/acpi/acpica/exdump.c
482
(next),
drivers/acpi/acpica/exdump.c
483
next->address_space.space_id);
drivers/acpi/acpica/exdump.c
485
while (next->address_space.next) {
drivers/acpi/acpica/exdump.c
486
if ((next->common.type ==
drivers/acpi/acpica/exdump.c
488
data = next;
drivers/acpi/acpica/exdump.c
491
next = next->address_space.next;
drivers/acpi/acpica/exdump.c
492
acpi_os_printf("->%p(%s %2.2X)", next,
drivers/acpi/acpica/exdump.c
494
(next),
drivers/acpi/acpica/exdump.c
495
next->address_space.
drivers/acpi/acpica/exdump.c
498
if ((next == start) || (next == data)) {
drivers/acpi/acpica/exdump.c
512
next = start;
drivers/acpi/acpica/exdump.c
514
acpi_os_printf("%20s : %p", name, next);
drivers/acpi/acpica/exdump.c
515
if (next) {
drivers/acpi/acpica/exdump.c
518
(next), next->common.type);
drivers/acpi/acpica/exdump.c
520
while (next->region.next) {
drivers/acpi/acpica/exdump.c
521
if ((next->common.type ==
drivers/acpi/acpica/exdump.c
523
data = next;
drivers/acpi/acpica/exdump.c
526
next = next->region.next;
drivers/acpi/acpica/exdump.c
527
acpi_os_printf("->%p(%s %2.2X)", next,
drivers/acpi/acpica/exdump.c
529
(next),
drivers/acpi/acpica/exdump.c
530
next->common.type);
drivers/acpi/acpica/exdump.c
532
if ((next == start) || (next == data)) {
drivers/acpi/acpica/exmutex.c
45
if (obj_desc->mutex.next) {
drivers/acpi/acpica/exmutex.c
46
(obj_desc->mutex.next)->mutex.prev = obj_desc->mutex.prev;
drivers/acpi/acpica/exmutex.c
463
union acpi_operand_object *next = thread->acquired_mutex_list;
drivers/acpi/acpica/exmutex.c
470
while (next) {
drivers/acpi/acpica/exmutex.c
471
obj_desc = next;
drivers/acpi/acpica/exmutex.c
496
next = obj_desc->mutex.next;
drivers/acpi/acpica/exmutex.c
499
obj_desc->mutex.next = NULL;
drivers/acpi/acpica/exmutex.c
50
(obj_desc->mutex.prev)->mutex.next = obj_desc->mutex.next;
drivers/acpi/acpica/exmutex.c
61
thread->acquired_mutex_list = obj_desc->mutex.next;
drivers/acpi/acpica/exmutex.c
89
obj_desc->mutex.next = list_head;
drivers/acpi/acpica/hwpci.c
179
list_element->next = *return_list_head;
drivers/acpi/acpica/hwpci.c
237
info = info->next;
drivers/acpi/acpica/hwpci.c
24
struct acpi_pci_device *next;
drivers/acpi/acpica/hwpci.c
264
struct acpi_pci_device *next;
drivers/acpi/acpica/hwpci.c
267
next = list_head;
drivers/acpi/acpica/hwpci.c
268
while (next) {
drivers/acpi/acpica/hwpci.c
269
previous = next;
drivers/acpi/acpica/hwpci.c
270
next = previous->next;
drivers/acpi/acpica/psargs.c
703
union acpi_parse_object *next;
drivers/acpi/acpica/psargs.c
707
next = cur->common.next;
drivers/acpi/acpica/psargs.c
717
cur = next;
drivers/acpi/acpica/psargs.c
797
prev->common.next = field;
drivers/acpi/acpica/psparse.c
104
union acpi_parse_object *next;
drivers/acpi/acpica/psparse.c
235
replacement_op->common.next = op->common.next;
drivers/acpi/acpica/psparse.c
238
op->common.next;
drivers/acpi/acpica/psparse.c
249
next = prev->common.next;
drivers/acpi/acpica/psparse.c
250
if (next == op) {
drivers/acpi/acpica/psparse.c
258
prev->common.next =
drivers/acpi/acpica/psparse.c
260
replacement_op->common.next =
drivers/acpi/acpica/psparse.c
261
op->common.next;
drivers/acpi/acpica/psparse.c
262
next = NULL;
drivers/acpi/acpica/psparse.c
264
prev->common.next =
drivers/acpi/acpica/psparse.c
265
op->common.next;
drivers/acpi/acpica/psparse.c
266
next = NULL;
drivers/acpi/acpica/psparse.c
269
prev = next;
drivers/acpi/acpica/psscope.c
174
if (scope->common.next) {
drivers/acpi/acpica/pstree.c
133
while (prev_arg->common.next) {
drivers/acpi/acpica/pstree.c
134
prev_arg = prev_arg->common.next;
drivers/acpi/acpica/pstree.c
136
prev_arg->common.next = arg;
drivers/acpi/acpica/pstree.c
147
arg = arg->common.next;
drivers/acpi/acpica/pstree.c
172
union acpi_parse_object *next = NULL;
drivers/acpi/acpica/pstree.c
184
next = acpi_ps_get_arg(op, 0);
drivers/acpi/acpica/pstree.c
185
if (next) {
drivers/acpi/acpica/pstree.c
186
ASL_CV_LABEL_FILENODE(next);
drivers/acpi/acpica/pstree.c
187
return (next);
drivers/acpi/acpica/pstree.c
192
next = op->common.next;
drivers/acpi/acpica/pstree.c
193
if (next) {
drivers/acpi/acpica/pstree.c
194
ASL_CV_LABEL_FILENODE(next);
drivers/acpi/acpica/pstree.c
195
return (next);
drivers/acpi/acpica/pstree.c
207
arg = arg->common.next;
drivers/acpi/acpica/pstree.c
217
if (parent->common.next) {
drivers/acpi/acpica/pstree.c
221
ASL_CV_LABEL_FILENODE(parent->common.next);
drivers/acpi/acpica/pstree.c
222
return (parent->common.next);
drivers/acpi/acpica/pstree.c
229
ASL_CV_LABEL_FILENODE(next);
drivers/acpi/acpica/pstree.c
230
return (next);
drivers/acpi/acpica/pstree.c
74
arg = arg->common.next;
drivers/acpi/acpica/pswalk.c
32
union acpi_parse_object *next = NULL;
drivers/acpi/acpica/pswalk.c
72
next = acpi_ps_get_arg(op, 0);
drivers/acpi/acpica/pswalk.c
73
if (next) {
drivers/acpi/acpica/pswalk.c
77
op = next;
drivers/acpi/acpica/pswalk.c
85
next = op->common.next;
drivers/acpi/acpica/pswalk.c
96
if (next) {
drivers/acpi/acpica/pswalk.c
97
op = next;
drivers/acpi/acpica/utaddress.c
115
range_info->next;
drivers/acpi/acpica/utaddress.c
117
prev->next = range_info->next;
drivers/acpi/acpica/utaddress.c
134
range_info = range_info->next;
drivers/acpi/acpica/utaddress.c
218
range_info = range_info->next;
drivers/acpi/acpica/utaddress.c
239
struct acpi_address_range *next;
drivers/acpi/acpica/utaddress.c
246
next = acpi_gbl_address_range_list[i];
drivers/acpi/acpica/utaddress.c
248
while (next) {
drivers/acpi/acpica/utaddress.c
249
range_info = next;
drivers/acpi/acpica/utaddress.c
250
next = range_info->next;
drivers/acpi/acpica/utaddress.c
65
range_info->next = acpi_gbl_address_range_list[space_id];
drivers/acpi/acpica/utcache.c
76
void *next;
drivers/acpi/acpica/utcache.c
96
next = ACPI_GET_DESCRIPTOR_PTR(cache->list_head);
drivers/acpi/acpica/utcache.c
99
cache->list_head = next;
drivers/acpi/acpica/utdelete.c
124
next_desc = handler_desc->address_space.next;
drivers/acpi/acpica/utdelete.c
216
next_desc->region.next;
drivers/acpi/acpica/utdelete.c
222
last_obj_ptr = &next_desc->region.next;
drivers/acpi/acpica/utdelete.c
223
next_desc = next_desc->region.next;
drivers/acpi/acpica/utdelete.c
521
prev_object->notify.next[i];
drivers/acpi/acpica/utinit.c
57
next_gpe_block = gpe_block->next;
drivers/acpi/acpica/utinit.c
64
next_gpe_xrupt_info = gpe_xrupt_info->next;
drivers/acpi/acpica/utosi.c
131
acpi_default_supported_interfaces[i].next =
drivers/acpi/acpica/utosi.c
164
acpi_gbl_supported_interfaces = next_interface->next;
drivers/acpi/acpica/utosi.c
224
interface_info->next = acpi_gbl_supported_interfaces;
drivers/acpi/acpica/utosi.c
261
next_interface->next;
drivers/acpi/acpica/utosi.c
263
previous_interface->next =
drivers/acpi/acpica/utosi.c
264
next_interface->next;
drivers/acpi/acpica/utosi.c
285
next_interface = next_interface->next;
drivers/acpi/acpica/utosi.c
330
next_interface = next_interface->next;
drivers/acpi/acpica/utosi.c
359
next_interface = next_interface->next;
drivers/acpi/acpica/utstate.c
34
state->common.next = *list_head;
drivers/acpi/acpica/utstate.c
65
*list_head = state->common.next;
drivers/acpi/acpica/uttrack.c
299
if (!element->next) {
drivers/acpi/acpica/uttrack.c
303
element = element->next;
drivers/acpi/acpica/uttrack.c
383
allocation->next = mem_list->list_head;
drivers/acpi/acpica/uttrack.c
390
allocation->next = element->next;
drivers/acpi/acpica/uttrack.c
393
if (element->next) {
drivers/acpi/acpica/uttrack.c
394
(element->next)->previous = allocation;
drivers/acpi/acpica/uttrack.c
397
element->next = allocation;
drivers/acpi/acpica/uttrack.c
452
(allocation->previous)->next = allocation->next;
drivers/acpi/acpica/uttrack.c
454
mem_list->list_head = allocation->next;
drivers/acpi/acpica/uttrack.c
457
if (allocation->next) {
drivers/acpi/acpica/uttrack.c
458
(allocation->next)->previous = allocation->previous;
drivers/acpi/acpica/uttrack.c
687
element = element->next;
drivers/acpi/apei/apei-base.c
334
res1 = list_entry(res_list1->next, struct apei_res, list);
drivers/acpi/apei/apei-base.c
335
resn1 = list_entry(res1->list.next, struct apei_res, list);
drivers/acpi/apei/apei-base.c
364
resn1 = list_entry(resn1->list.next, struct apei_res, list);
drivers/acpi/apei/ghes.c
1262
struct llist_node *llnode, *next;
drivers/acpi/apei/ghes.c
1275
next = llnode->next;
drivers/acpi/apei/ghes.c
1292
llnode = next;
drivers/acpi/apei/ghes.c
1315
llnode = llnode->next;
drivers/acpi/arm64/iort.c
1316
struct iommu_resv_region *entry, *next;
drivers/acpi/arm64/iort.c
1318
list_for_each_entry_safe(entry, next, head, list)
drivers/acpi/evged.c
166
struct acpi_ged_event *event, *next;
drivers/acpi/evged.c
168
list_for_each_entry_safe(event, next, &geddev->event_list, node) {
drivers/acpi/nvs.c
120
list_for_each_entry_safe(entry, next, &nvs_list, node) {
drivers/acpi/nvs.c
97
struct nvs_page *entry, *next;
drivers/acpi/pci_root.c
818
goto next;
drivers/acpi/pci_root.c
826
goto next;
drivers/acpi/pci_root.c
848
goto next;
drivers/acpi/pci_root.c
852
next:
drivers/acpi/platform_profile.c
500
enum platform_profile_option next = PLATFORM_PROFILE_LAST;
drivers/acpi/platform_profile.c
524
next = find_next_bit_wrap(data.aggregate,
drivers/acpi/platform_profile.c
528
err = class_for_each_device(&platform_profile_class, NULL, &next,
drivers/acpi/pptt.c
485
struct acpi_pptt_processor *next;
drivers/acpi/pptt.c
493
next = fetch_pptt_node(table_hdr, cpu->parent);
drivers/acpi/pptt.c
494
if (!(next && next->flags & ACPI_PPTT_ACPI_IDENTICAL))
drivers/acpi/proc.c
52
dev->physical_node_list.next)
drivers/acpi/property.c
1356
struct list_head *next;
drivers/acpi/property.c
1379
next = dn->sibling.next;
drivers/acpi/property.c
1380
if (next == head)
drivers/acpi/property.c
1383
dn = list_entry(next, struct acpi_data_node, sibling);
drivers/acpi/property.c
658
struct acpi_data_node *dn, *next;
drivers/acpi/property.c
663
list_for_each_entry_safe_reverse(dn, next, list, sibling) {
drivers/acpi/sleep.c
59
.next = NULL,
drivers/android/binder.c
1690
struct binder_transaction *next;
drivers/android/binder.c
1725
next = t->from_parent;
drivers/android/binder.c
1732
if (next == NULL) {
drivers/android/binder.c
1737
t = next;
drivers/android/binder.c
447
BUG_ON(work->entry.next && !list_empty(&work->entry));
drivers/android/binder_alloc.c
53
return list_entry(buffer->entry.next, struct binder_buffer, entry);
drivers/android/binder_alloc.c
653
struct binder_buffer *buffer, *next;
drivers/android/binder_alloc.c
675
next = kzalloc_obj(*next);
drivers/android/binder_alloc.c
676
if (!next)
drivers/android/binder_alloc.c
680
buffer = binder_alloc_new_buf_locked(alloc, next, size, is_async);
drivers/android/binder_alloc.c
715
struct binder_buffer *prev, *next;
drivers/android/binder_alloc.c
720
BUG_ON(alloc->buffers.next == &buffer->entry);
drivers/android/binder_alloc.c
727
next = binder_buffer_next(buffer);
drivers/android/binder_alloc.c
728
if (buffer_start_page(next) == buffer_start_page(buffer))
drivers/android/binder_alloc.c
773
struct binder_buffer *next = binder_buffer_next(buffer);
drivers/android/binder_alloc.c
775
if (next->free) {
drivers/android/binder_alloc.c
776
rb_erase(&next->rb_node, &alloc->free_buffers);
drivers/android/binder_alloc.c
777
binder_delete_free_buffer(alloc, next);
drivers/android/binder_alloc.c
780
if (alloc->buffers.next != &buffer->entry) {
drivers/ata/libata-core.c
256
next:
drivers/ata/libata-core.c
274
goto next;
drivers/ata/libata-core.c
6713
char *p, *cur, *next;
drivers/ata/libata-core.c
6728
for (cur = ata_force_param_buf; *cur != '\0'; cur = next) {
drivers/ata/libata-core.c
6732
next = cur;
drivers/ata/libata-core.c
6733
if (ata_parse_force_one(&next, &te, &reason)) {
drivers/atm/adummy.c
189
struct adummy_dev *adummy_dev, *next;
drivers/atm/adummy.c
191
list_for_each_entry_safe(adummy_dev, next, &adummy_devs, entry) {
drivers/atm/eni.c
660
eni_dev->fast = ENI_VCC(curr)->next;
drivers/atm/eni.c
661
ENI_VCC(curr)->next = ENI_VCC_NOS;
drivers/atm/eni.c
668
eni_dev->slow = ENI_VCC(curr)->next;
drivers/atm/eni.c
669
ENI_VCC(curr)->next = ENI_VCC_NOS;
drivers/atm/eni.c
695
if (ENI_VCC(vcc)->next != ENI_VCC_NOS) {
drivers/atm/eni.c
701
ENI_VCC(vcc)->next = NULL;
drivers/atm/eni.c
704
ENI_VCC(eni_dev->last_fast)->next = vcc;
drivers/atm/eni.c
710
ENI_VCC(eni_dev->last_slow)->next = vcc;
drivers/atm/eni.c
799
eni_vcc->next = ENI_VCC_NOS;
drivers/atm/eni.h
64
struct atm_vcc *next; /* next pending RX */
drivers/atm/fore200e.c
1076
buffer->next = bsq->freebuf;
drivers/atm/fore200e.c
2032
buffer[ i ].next = bsq->freebuf;
drivers/atm/fore200e.c
890
buffer = buffer->next;
drivers/atm/fore200e.c
935
bsq->freebuf = buffer->next;
drivers/atm/fore200e.h
579
struct buffer* next; /* next receive buffer */
drivers/atm/he.c
1523
struct he_buff *heb, *next;
drivers/atm/he.c
1567
list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry)
drivers/atm/he.c
1646
struct he_buff *heb, *next;
drivers/atm/he.c
1772
list_for_each_entry_safe(heb, next, &he_vcc->buffers, entry)
drivers/atm/he.c
393
he_dev->next = NULL;
drivers/atm/he.c
395
he_dev->next = he_devs;
drivers/atm/he.c
771
struct he_buff *heb, *next;
drivers/atm/he.c
888
list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry)
drivers/atm/he.h
314
struct he_dev *next;
drivers/atm/idt77105.c
120
for (walk = idt77105_all; walk; walk = walk->next) {
drivers/atm/idt77105.c
269
PRIV(dev)->next = idt77105_all;
drivers/atm/idt77105.c
333
prev = walk, walk = walk->next) {
drivers/atm/idt77105.c
336
prev->next = walk->next;
drivers/atm/idt77105.c
338
idt77105_all = walk->next;
drivers/atm/idt77105.c
37
struct idt77105_priv *next;
drivers/atm/idt77105.c
91
for (walk = idt77105_all; walk; walk = walk->next) {
drivers/atm/idt77252.c
1215
if (card->rsq.next == card->rsq.last)
drivers/atm/idt77252.c
1218
rsqe = card->rsq.next + 1;
drivers/atm/idt77252.c
1228
card->rsq.next = rsqe;
drivers/atm/idt77252.c
1229
if (card->rsq.next == card->rsq.last)
drivers/atm/idt77252.c
1232
rsqe = card->rsq.next + 1;
drivers/atm/idt77252.c
1235
writel((unsigned long) card->rsq.next - (unsigned long) card->rsq.base,
drivers/atm/idt77252.c
1341
struct sk_buff *next;
drivers/atm/idt77252.c
1347
next = sb_pool_skb(card, handle);
drivers/atm/idt77252.c
1350
if (next) {
drivers/atm/idt77252.c
1351
card->raw_cell_head = next;
drivers/atm/idt77252.c
1388
card->tsq.next = card->tsq.last;
drivers/atm/idt77252.c
1393
writel((unsigned long) card->tsq.next - (unsigned long) card->tsq.base,
drivers/atm/idt77252.c
1414
if (card->tsq.next == card->tsq.last)
drivers/atm/idt77252.c
1417
tsqe = card->tsq.next + 1;
drivers/atm/idt77252.c
1420
card->tsq.base, card->tsq.next, card->tsq.last);
drivers/atm/idt77252.c
1511
card->tsq.next = tsqe;
drivers/atm/idt77252.c
1512
if (card->tsq.next == card->tsq.last)
drivers/atm/idt77252.c
1515
tsqe = card->tsq.next + 1;
drivers/atm/idt77252.c
1518
card->tsq.base, card->tsq.next, card->tsq.last);
drivers/atm/idt77252.c
1524
writel((unsigned long)card->tsq.next - (unsigned long)card->tsq.base,
drivers/atm/idt77252.c
1529
readl(SAR_REG_TSQT), card->tsq.next);
drivers/atm/idt77252.c
3715
last = &card->next;
drivers/atm/idt77252.c
3771
idt77252_chain = card->next;
drivers/atm/idt77252.c
651
scq->next = scq->base;
drivers/atm/idt77252.c
662
scq->base, scq->next, scq->last, (unsigned long long)scq->paddr);
drivers/atm/idt77252.c
710
TXPRINTK("%s: SCQ: next 0x%p\n", card->name, scq->next);
drivers/atm/idt77252.c
745
scq->next->word_1 = cpu_to_le32(tbd->word_1 |
drivers/atm/idt77252.c
747
scq->next->word_2 = cpu_to_le32(tbd->word_2);
drivers/atm/idt77252.c
748
scq->next->word_3 = cpu_to_le32(tbd->word_3);
drivers/atm/idt77252.c
749
scq->next->word_4 = cpu_to_le32(tbd->word_4);
drivers/atm/idt77252.c
751
if (scq->next == scq->last)
drivers/atm/idt77252.c
752
scq->next = scq->base;
drivers/atm/idt77252.c
754
scq->next++;
drivers/atm/idt77252.c
758
(u32)((unsigned long)scq->next - (unsigned long)scq->base));
drivers/atm/idt77252.c
772
read_sram(card, scq->scd + 1), scq->next);
drivers/atm/idt77252.c
798
card->name, atomic_read(&scq->used), scq->next);
drivers/atm/idt77252.c
985
card->rsq.next = card->rsq.last;
drivers/atm/idt77252.h
163
struct scqe *next;
drivers/atm/idt77252.h
272
struct rsq_entry *next;
drivers/atm/idt77252.h
307
struct tsq_entry *next;
drivers/atm/idt77252.h
413
struct idt77252_dev *next;
drivers/atm/iphase.c
100
que->next = NULL;
drivers/atm/iphase.c
106
data->next = NULL;
drivers/atm/iphase.c
107
if (que->next == NULL)
drivers/atm/iphase.c
108
que->next = que->tail = data;
drivers/atm/iphase.c
110
data->next = que->next;
drivers/atm/iphase.c
111
que->next = data;
drivers/atm/iphase.c
121
entry->next = NULL;
drivers/atm/iphase.c
122
if (que->next == NULL)
drivers/atm/iphase.c
123
que->next = que->tail = entry;
drivers/atm/iphase.c
125
que->tail->next = entry;
drivers/atm/iphase.c
126
que->tail = que->tail->next;
drivers/atm/iphase.c
133
if (que->next == NULL)
drivers/atm/iphase.c
135
tmpdata = que->next;
drivers/atm/iphase.c
136
if ( que->next == que->tail)
drivers/atm/iphase.c
137
que->next = que->tail = NULL;
drivers/atm/iphase.c
139
que->next = que->next->next;
drivers/atm/iphase.h
300
struct free_desc_q *next;
drivers/atm/iphase.h
887
struct ia_rtn_q *next, *tail;
drivers/atm/nicstar.c
1472
scqep = scq->next;
drivers/atm/nicstar.c
1483
if (!ns_scqe_is_tsr(scqep) && scq->tail != scq->next) {
drivers/atm/nicstar.c
1491
scqi = scq->next - scq->base;
drivers/atm/nicstar.c
1495
*scq->next = tsr;
drivers/atm/nicstar.c
1498
if (scq->next == scq->last)
drivers/atm/nicstar.c
1499
scq->next = scq->base;
drivers/atm/nicstar.c
1501
scq->next++;
drivers/atm/nicstar.c
1502
data = scq_virt_to_bus(scq, scq->next);
drivers/atm/nicstar.c
1555
card->tsq.base, card->tsq.next,
drivers/atm/nicstar.c
1559
card->rsq.base, card->rsq.next,
drivers/atm/nicstar.c
1739
while (scq->tail == scq->next) {
drivers/atm/nicstar.c
1748
scq->tail != scq->next,
drivers/atm/nicstar.c
1759
*scq->next = *tbd;
drivers/atm/nicstar.c
1760
index = (int)(scq->next - scq->base);
drivers/atm/nicstar.c
1767
scq->next);
drivers/atm/nicstar.c
1768
if (scq->next == scq->last)
drivers/atm/nicstar.c
1769
scq->next = scq->base;
drivers/atm/nicstar.c
1771
scq->next++;
drivers/atm/nicstar.c
1784
while (scq->tail == scq->next) {
drivers/atm/nicstar.c
1786
data = scq_virt_to_bus(scq, scq->next);
drivers/atm/nicstar.c
1798
scq->tail != scq->next,
drivers/atm/nicstar.c
1809
scqi = scq->next - scq->base;
drivers/atm/nicstar.c
1814
*scq->next = tsr;
drivers/atm/nicstar.c
1821
le32_to_cpu(tsr.word_4), scq->next);
drivers/atm/nicstar.c
1822
if (scq->next == scq->last)
drivers/atm/nicstar.c
1823
scq->next = scq->base;
drivers/atm/nicstar.c
1825
scq->next++;
drivers/atm/nicstar.c
1832
data = scq_virt_to_bus(scq, scq->next);
drivers/atm/nicstar.c
1849
if (card->tsq.next == card->tsq.last)
drivers/atm/nicstar.c
1852
one_ahead = card->tsq.next + 1;
drivers/atm/nicstar.c
1859
while (!ns_tsi_isempty(card->tsq.next) || !ns_tsi_isempty(one_ahead) ||
drivers/atm/nicstar.c
1866
while (ns_tsi_isempty(card->tsq.next)) {
drivers/atm/nicstar.c
1867
if (card->tsq.next == card->tsq.last)
drivers/atm/nicstar.c
1868
card->tsq.next = card->tsq.base;
drivers/atm/nicstar.c
1870
card->tsq.next++;
drivers/atm/nicstar.c
1873
if (!ns_tsi_tmrof(card->tsq.next)) {
drivers/atm/nicstar.c
1874
scdi = ns_tsi_getscdindex(card->tsq.next);
drivers/atm/nicstar.c
1882
ns_tsi_init(card->tsq.next);
drivers/atm/nicstar.c
1887
drain_scq(card, scq, ns_tsi_getscqpos(card->tsq.next));
drivers/atm/nicstar.c
1892
ns_tsi_init(card->tsq.next);
drivers/atm/nicstar.c
1893
previous = card->tsq.next;
drivers/atm/nicstar.c
1894
if (card->tsq.next == card->tsq.last)
drivers/atm/nicstar.c
1895
card->tsq.next = card->tsq.base;
drivers/atm/nicstar.c
1897
card->tsq.next++;
drivers/atm/nicstar.c
1899
if (card->tsq.next == card->tsq.last)
drivers/atm/nicstar.c
1902
one_ahead = card->tsq.next + 1;
drivers/atm/nicstar.c
1961
if (!ns_rsqe_valid(card->rsq.next))
drivers/atm/nicstar.c
1964
dequeue_rx(card, card->rsq.next);
drivers/atm/nicstar.c
1965
ns_rsqe_init(card->rsq.next);
drivers/atm/nicstar.c
1966
previous = card->rsq.next;
drivers/atm/nicstar.c
1967
if (card->rsq.next == card->rsq.last)
drivers/atm/nicstar.c
1968
card->rsq.next = card->rsq.base;
drivers/atm/nicstar.c
1970
card->rsq.next++;
drivers/atm/nicstar.c
1971
} while (ns_rsqe_valid(card->rsq.next));
drivers/atm/nicstar.c
552
card->tsq.next = card->tsq.base;
drivers/atm/nicstar.c
571
card->rsq.next = card->rsq.base;
drivers/atm/nicstar.c
888
scq->next = scq->base;
drivers/atm/nicstar.h
658
ns_tsi *next;
drivers/atm/nicstar.h
667
ns_scqe *next;
drivers/atm/nicstar.h
684
ns_rsqe *next;
drivers/atm/suni.c
314
PRIV(dev)->next = sunis;
drivers/atm/suni.c
332
(unsigned long) poll_timer.list.next);
drivers/atm/suni.c
348
walk = &PRIV((*walk)->dev)->next);
drivers/atm/suni.c
349
*walk = PRIV((*walk)->dev)->next;
drivers/atm/suni.c
62
for (walk = sunis; walk; walk = walk->next) {
drivers/atm/suni.h
236
struct suni_priv *next; /* next SUNI */
drivers/base/cacheinfo.c
295
struct device_node *next __free(device_node) = of_find_next_cache_node(np);
drivers/base/cacheinfo.c
296
if (next) {
drivers/base/devres.c
457
first = first->next;
drivers/base/devres.c
524
if (WARN_ON(dev->devres_head.next == NULL))
drivers/base/devres.c
532
cnt = remove_nodes(dev, dev->devres_head.next, &dev->devres_head, &todo);
drivers/base/devres.c
687
end = grp->node[1].entry.next;
drivers/base/devtmpfs.c
116
req->next = requests;
drivers/base/devtmpfs.c
398
struct req *next = req->next;
drivers/base/devtmpfs.c
402
req = next;
drivers/base/devtmpfs.c
47
struct req *next;
drivers/base/firmware_loader/fallback.c
52
struct fw_priv *next;
drivers/base/firmware_loader/fallback.c
55
list_for_each_entry_safe(fw_priv, next, &pending_fw_head,
drivers/base/firmware_loader/main.c
1469
fce = list_entry(fwc->fw_names.next,
drivers/base/isa.c
105
struct device *tmp = to_isa_dev(dev)->next;
drivers/base/isa.c
152
isa_dev->next = isa_driver->devices;
drivers/base/isa.c
20
struct device *next;
drivers/base/map.c
103
for (p = domain->probes[MAJOR(dev) % 255]; p; p = p->next) {
drivers/base/map.c
21
struct probe *next;
drivers/base/map.c
60
s = &(*s)->next;
drivers/base/map.c
61
p->next = *s;
drivers/base/map.c
81
for (s = &domain->probes[index % 255]; *s; s = &(*s)->next) {
drivers/base/map.c
84
*s = p->next;
drivers/base/node.c
419
struct node_cache_info *info, *next;
drivers/base/node.c
424
list_for_each_entry_safe(info, next, &node->cache_attrs, node) {
drivers/base/power/main.c
1177
dev = to_device(dpm_suspended_list.next);
drivers/base/power/main.c
2231
struct device *dev = to_device(dpm_list.next);
drivers/base/power/main.c
848
dev = to_device(dpm_noirq_list.next);
drivers/base/power/main.c
998
dev = to_device(dpm_late_early_list.next);
drivers/base/power/wakeup.c
1153
.next = wakeup_sources_stats_seq_next,
drivers/base/property.c
800
struct fwnode_handle *next;
drivers/base/property.c
806
next = fwnode_call_ptr_op(fwnode, get_next_child_node, child);
drivers/base/property.c
807
if (next)
drivers/base/property.c
808
return next;
drivers/base/regmap/regcache-rbtree.c
201
struct rb_node *next;
drivers/base/regmap/regcache-rbtree.c
211
next = rb_first(&rbtree_ctx->root);
drivers/base/regmap/regcache-rbtree.c
212
while (next) {
drivers/base/regmap/regcache-rbtree.c
213
rbtree_node = rb_entry(next, struct regcache_rbtree_node, node);
drivers/base/regmap/regcache-rbtree.c
214
next = rb_next(&rbtree_node->node);
drivers/base/regmap/regmap-debugfs.c
538
struct rb_node *next;
drivers/base/regmap/regmap-debugfs.c
640
next = rb_first(&map->range_tree);
drivers/base/regmap/regmap-debugfs.c
641
while (next) {
drivers/base/regmap/regmap-debugfs.c
642
range_node = rb_entry(next, struct regmap_range_node, node);
drivers/base/regmap/regmap-debugfs.c
649
next = rb_next(&range_node->node);
drivers/base/regmap/regmap.c
546
struct rb_node *next;
drivers/base/regmap/regmap.c
549
next = rb_first(&map->range_tree);
drivers/base/regmap/regmap.c
550
while (next) {
drivers/base/regmap/regmap.c
551
range_node = rb_entry(next, struct regmap_range_node, node);
drivers/base/regmap/regmap.c
552
next = rb_next(&range_node->node);
drivers/block/aoe/aoe.h
158
struct aoedev *next;
drivers/block/aoe/aoeblk.c
135
list_empty(&(*t)->ffree) ? NULL : (*t)->ffree.next);
drivers/block/aoe/aoecmd.c
1207
pos = iocq[id].head.next;
drivers/block/aoe/aoecmd.c
1670
pos = flist.next;
drivers/block/aoe/aoecmd.c
220
pos = t->ffree.next;
drivers/block/aoe/aoecmd.c
776
pos = flist.next;
drivers/block/aoe/aoedev.c
340
for (d = devlist; d; d = d->next) {
drivers/block/aoe/aoedev.c
371
for (d = devlist; d; d = d->next) {
drivers/block/aoe/aoedev.c
389
*dd = d->next;
drivers/block/aoe/aoedev.c
392
dd = &d->next;
drivers/block/aoe/aoedev.c
460
for (d=devlist; d; d=d->next)
drivers/block/aoe/aoedev.c
502
d->next = devlist;
drivers/block/brd.c
360
struct brd_device *brd, *next;
drivers/block/brd.c
364
list_for_each_entry_safe(brd, next, &brd_devices, brd_list) {
drivers/block/drbd/drbd_actlog.c
736
int next = (device->rs_last_mark + 1) % DRBD_SYNC_MARKS;
drivers/block/drbd/drbd_actlog.c
741
device->rs_mark_time[next] = now;
drivers/block/drbd/drbd_actlog.c
742
device->rs_mark_left[next] = still_to_go;
drivers/block/drbd/drbd_actlog.c
743
device->rs_last_mark = next;
drivers/block/drbd/drbd_nl.c
3714
if (list_empty(connections) || connections->next->next != connections)
drivers/block/drbd/drbd_nl.c
3917
pos = list_entry_rcu(resource->resources.next,
drivers/block/drbd/drbd_nl.c
4889
goto next;
drivers/block/drbd/drbd_nl.c
4895
goto next;
drivers/block/drbd/drbd_nl.c
4901
goto next;
drivers/block/drbd/drbd_nl.c
4907
goto next;
drivers/block/drbd/drbd_nl.c
4910
next:
drivers/block/drbd/drbd_nl.c
4913
list_entry(state_change->list.next,
drivers/block/drbd/drbd_nl.c
4965
list_entry(head.next, struct drbd_state_change, list);
drivers/block/drbd/drbd_receiver.c
1213
next_epoch = list_entry(epoch->list.next, struct drbd_epoch, list);
drivers/block/floppy.c
3015
if (raw_cmd->next &&
drivers/block/floppy.c
3020
raw_cmd = raw_cmd->next;
drivers/block/floppy.c
3041
cmd.next = NULL;
drivers/block/floppy.c
3057
ptr = ptr->next;
drivers/block/floppy.c
3065
struct floppy_raw_cmd *next;
drivers/block/floppy.c
3076
next = this->next;
drivers/block/floppy.c
3078
this = next;
drivers/block/floppy.c
3097
ptr->next = NULL;
drivers/block/floppy.c
3123
rcmd = &(ptr->next);
drivers/block/loop.c
1956
cmd_list->next, struct loop_cmd, list_entry);
drivers/block/loop.c
1957
list_del(cmd_list->next);
drivers/block/nbd.c
742
struct bio *next = bio->bi_next;
drivers/block/nbd.c
747
bool is_last = !next && bio_iter_last(bvec, iter);
drivers/block/nbd.c
781
bio = next;
drivers/block/null_blk/main.c
2188
nullb = list_entry(nullb_list.next, struct nullb, list);
drivers/block/null_blk/main.c
2207
nullb = list_entry(nullb_list.next, struct nullb, list);
drivers/block/ps3vram.c
538
struct bio *next;
drivers/block/ps3vram.c
571
next = bio_list_peek(&priv->list);
drivers/block/ps3vram.c
576
return next;
drivers/block/ublk_drv.c
2012
struct request *next;
drivers/block/ublk_drv.c
2015
next = rq->rq_next;
drivers/block/ublk_drv.c
2018
rq = next;
drivers/block/xen-blkback/blkback.c
421
req = list_entry(ring->pending_free.next, struct pending_req,
drivers/block/xen-blkback/blkback.c
837
goto next;
drivers/block/xen-blkback/blkback.c
856
goto next;
drivers/block/xen-blkback/blkback.c
865
goto next;
drivers/block/xen-blkback/blkback.c
871
goto next;
drivers/block/xen-blkback/blkback.c
882
next:
drivers/block/zram/zram_drv.c
1086
goto next;
drivers/block/zram/zram_drv.c
1092
goto next;
drivers/block/zram/zram_drv.c
1114
next:
drivers/block/zram/zram_drv.c
1210
goto next;
drivers/block/zram/zram_drv.c
1214
goto next;
drivers/block/zram/zram_drv.c
1218
goto next;
drivers/block/zram/zram_drv.c
1221
goto next;
drivers/block/zram/zram_drv.c
1224
goto next;
drivers/block/zram/zram_drv.c
1227
next:
drivers/block/zram/zram_drv.c
1563
goto next;
drivers/block/zram/zram_drv.c
1582
next:
drivers/block/zram/zram_drv.c
2348
goto next;
drivers/block/zram/zram_drv.c
2352
goto next;
drivers/block/zram/zram_drv.c
2356
goto next;
drivers/block/zram/zram_drv.c
2361
goto next;
drivers/block/zram/zram_drv.c
2365
goto next;
drivers/block/zram/zram_drv.c
2368
next:
drivers/block/zram/zram_drv.c
2656
goto next;
drivers/block/zram/zram_drv.c
2661
next:
drivers/bluetooth/btrtl.c
1107
goto next;
drivers/bluetooth/btrtl.c
1135
next:
drivers/bluetooth/btrtl.c
469
struct list_head *next;
drivers/bluetooth/btrtl.c
472
list_for_each_safe(pos, next, &btrtl_dev->patch_subsecs) {
drivers/bluetooth/btsdio.c
284
tuple = tuple->next;
drivers/bus/brcmstb_gisb.c
146
struct list_head next;
drivers/bus/brcmstb_gisb.c
296
list_for_each_entry(gdev, &brcmstb_gisb_arb_device_list, next) {
drivers/bus/brcmstb_gisb.c
385
list_for_each_entry(gdev, &brcmstb_gisb_arb_device_list, next)
drivers/bus/brcmstb_gisb.c
430
INIT_LIST_HEAD(&gdev->next);
drivers/bus/brcmstb_gisb.c
492
list_add_tail(&gdev->next, &brcmstb_gisb_arb_device_list);
drivers/cdrom/cdrom.c
3246
long next = 0;
drivers/cdrom/cdrom.c
3248
ret = cdrom_get_next_writable(cdi, &next);
drivers/cdrom/cdrom.c
3251
if (copy_to_user((long __user *)arg, &next, sizeof(next)))
drivers/char/agp/isoch.c
299
for (pos=head->next; cdev<ndevs; cdev++, pos=pos->next) {
drivers/char/agp/isoch.c
40
struct list_head *pos, *tmp, *head = &list->list, *start = head->next;
drivers/char/agp/isoch.c
455
for (pos=head->next; pos!=head; ) {
drivers/char/agp/isoch.c
458
pos = pos->next;
drivers/char/agp/isoch.c
53
pos = pos->next;
drivers/char/ipmi/ipmi_devintf.c
125
struct ipmi_recv_msg *msg, *next;
drivers/char/ipmi/ipmi_devintf.c
129
list_for_each_entry_safe(msg, next, &priv->recv_msgs, link)
drivers/char/ipmi/ipmi_devintf.c
223
entry = priv->recv_msgs.next;
drivers/char/ipmi/ipmi_dmi.c
110
info = info->next;
drivers/char/ipmi/ipmi_dmi.c
28
struct ipmi_dmi_info *next;
drivers/char/ipmi/ipmi_dmi.c
85
info->next = ipmi_dmi_infos;
drivers/char/ipmi/ipmi_msghandler.c
1389
rcvr->next = rcvrs;
drivers/char/ipmi/ipmi_msghandler.c
1396
rcvrs = rcvr->next;
drivers/char/ipmi/ipmi_msghandler.c
1761
rcvr->next = rcvrs;
drivers/char/ipmi/ipmi_msghandler.c
1772
rcvrs = rcvr->next;
drivers/char/ipmi/ipmi_msghandler.c
219
struct cmd_rcvr *next;
drivers/char/ipmi/ipmi_msghandler.c
3808
entry = tmplist.next;
drivers/char/ipmi/ipmi_msghandler.c
4825
smi_msg = list_entry(intf->waiting_rcv_msgs.next,
drivers/char/ipmi/ipmi_msghandler.c
4878
entry = intf->hp_xmit_msgs.next;
drivers/char/ipmi/ipmi_msghandler.c
4880
entry = intf->xmit_msgs.next;
drivers/char/ipmi/ipmi_msghandler.c
5608
.next = NULL,
drivers/char/ipmi/ipmi_si_hotmod.c
189
char *str, *curr, *next;
drivers/char/ipmi/ipmi_si_hotmod.c
196
for (curr = strstrip(str); curr; curr = next) {
drivers/char/ipmi/ipmi_si_hotmod.c
199
next = strchr(curr, ':');
drivers/char/ipmi/ipmi_si_hotmod.c
200
if (next) {
drivers/char/ipmi/ipmi_si_hotmod.c
201
*next = '\0';
drivers/char/ipmi/ipmi_si_hotmod.c
202
next++;
drivers/char/ipmi/ipmi_watchdog.c
1129
.next = NULL,
drivers/char/misc.c
112
.next = misc_seq_next,
drivers/char/tpm/eventlog/tpm1.c
283
.next = tpm1_bios_measurements_next,
drivers/char/tpm/eventlog/tpm1.c
290
.next = tpm1_bios_measurements_next,
drivers/char/tpm/eventlog/tpm2.c
156
.next = tpm2_bios_measurements_next,
drivers/char/xillybus/xillyusb.c
469
struct list_head *this, *next;
drivers/char/xillybus/xillyusb.c
476
list_for_each_safe(this, next, &ep->buffers) {
drivers/clk/clk.c
5536
struct clock_provider *clk_provider, *next;
drivers/clk/clk.c
5553
list_for_each_entry_safe(clk_provider, next,
drivers/clk/clk.c
5570
list_for_each_entry_safe(clk_provider, next,
drivers/clocksource/em_sti.c
116
static u64 em_sti_set_next(struct em_sti_priv *p, u64 next)
drivers/clocksource/em_sti.c
126
em_sti_write(p, STI_COMPA_H, next >> 32);
drivers/clocksource/em_sti.c
127
em_sti_write(p, STI_COMPA_L, next & 0xffffffff);
drivers/clocksource/em_sti.c
137
return next;
drivers/clocksource/em_sti.c
253
u64 next;
drivers/clocksource/em_sti.c
256
next = em_sti_set_next(p, em_sti_count(p) + delta);
drivers/clocksource/em_sti.c
257
safe = em_sti_count(p) < (next - 1);
drivers/clocksource/ingenic-sysost.c
248
static int ingenic_ost_cevt_set_next(unsigned long next,
drivers/clocksource/ingenic-sysost.c
254
writel(next, ost->base + OST_REG_OST1DFR);
drivers/clocksource/ingenic-timer.c
88
static int ingenic_tcu_cevt_set_next(unsigned long next,
drivers/clocksource/ingenic-timer.c
94
if (next > 0xffff)
drivers/clocksource/ingenic-timer.c
97
regmap_write(tcu->map, TCU_REG_TDFRc(timer->channel), next);
drivers/clocksource/mps2-timer.c
62
static int mps2_timer_set_next_event(unsigned long next, struct clock_event_device *ce)
drivers/clocksource/mps2-timer.c
64
clockevent_mps2_writel(next, ce, TIMER_VALUE);
drivers/clocksource/timer-cadence-ttc.c
353
ttccs->ttc.clk_rate_change_nb.next = NULL;
drivers/clocksource/timer-cadence-ttc.c
432
ttcce->ttc.clk_rate_change_nb.next = NULL;
drivers/clocksource/timer-econet-en751221.c
83
u32 next;
drivers/clocksource/timer-econet-en751221.c
87
next = ioread32(reg_count(cpu)) + delta;
drivers/clocksource/timer-econet-en751221.c
88
iowrite32(next, reg_compare(cpu));
drivers/clocksource/timer-econet-en751221.c
90
if ((s32)(next - ioread32(reg_count(cpu))) < ECONET_MIN_DELTA / 2)
drivers/clocksource/timer-ep93xx.c
87
static int ep93xx_clkevt_set_next_event(unsigned long next,
drivers/clocksource/timer-ep93xx.c
99
writel(next, tcu->base + EP93XX_TIMER3_LOAD);
drivers/clocksource/timer-imx-sysctr.c
74
u64 next;
drivers/clocksource/timer-imx-sysctr.c
78
next = sysctr_read_counter(evt);
drivers/clocksource/timer-imx-sysctr.c
80
next += delta;
drivers/clocksource/timer-imx-sysctr.c
82
cmp_hi = (next >> 32) & 0x00fffff;
drivers/clocksource/timer-imx-sysctr.c
83
cmp_lo = next & 0xffffffff;
drivers/clocksource/timer-imx-tpm.c
86
unsigned long next, prev, now;
drivers/clocksource/timer-imx-tpm.c
89
next = prev + delta;
drivers/clocksource/timer-imx-tpm.c
90
writel(next, timer_base + TPM_C0V);
drivers/clocksource/timer-imx-tpm.c
97
if ((next & 0xffffffff) != readl(timer_base + TPM_C0V))
drivers/clocksource/timer-integrator-ap.c
103
static int clkevt_set_next_event(unsigned long next, struct clock_event_device *evt)
drivers/clocksource/timer-integrator-ap.c
108
writel(next, clkevt_base + TIMER_LOAD);
drivers/clocksource/timer-pxa.c
81
unsigned long next, oscr;
drivers/clocksource/timer-pxa.c
84
next = timer_readl(OSCR) + delta;
drivers/clocksource/timer-pxa.c
85
timer_writel(next, OSMR0);
drivers/clocksource/timer-pxa.c
88
return (signed)(next - oscr) <= MIN_OSCR_DELTA ? -ETIME : 0;
drivers/clocksource/timer-sp804.c
207
static int sp804_set_next_event(unsigned long next,
drivers/clocksource/timer-sp804.c
213
writel(next, common_clkevt->load);
drivers/clocksource/timer-stm32.c
130
unsigned long now, next;
drivers/clocksource/timer-stm32.c
132
next = readl_relaxed(timer_of_base(to) + TIM_CNT) + evt;
drivers/clocksource/timer-stm32.c
133
writel_relaxed(next, timer_of_base(to) + TIM_CCR1);
drivers/clocksource/timer-stm32.c
136
if ((next - now) > evt)
drivers/clocksource/timer-sun5i.c
280
st->clk_rate_cb.next = NULL;
drivers/comedi/drivers.c
1033
for (driv = comedi_drivers; driv; driv = driv->next) {
drivers/comedi/drivers.c
1048
for (driv = comedi_drivers; driv; driv = driv->next) {
drivers/comedi/drivers.c
1210
driver->next = comedi_drivers;
drivers/comedi/drivers.c
1235
comedi_drivers = driver->next;
drivers/comedi/drivers.c
1237
for (prev = comedi_drivers; prev->next; prev = prev->next) {
drivers/comedi/drivers.c
1238
if (prev->next == driver) {
drivers/comedi/drivers.c
1239
prev->next = driver->next;
drivers/comedi/drivers/cb_pcidas64.c
1519
devpriv->ai_dma_desc[i].next =
drivers/comedi/drivers/cb_pcidas64.c
1534
devpriv->ao_dma_desc[i].next =
drivers/comedi/drivers/cb_pcidas64.c
2944
next_bits = le32_to_cpu(devpriv->ao_dma_desc[buffer_index].next);
drivers/comedi/drivers/cb_pcidas64.c
2946
devpriv->ao_dma_desc[buffer_index].next = cpu_to_le32(next_bits);
drivers/comedi/drivers/cb_pcidas64.c
2951
next_bits = le32_to_cpu(devpriv->ao_dma_desc[prev_buffer_index].next);
drivers/comedi/drivers/cb_pcidas64.c
2953
devpriv->ao_dma_desc[prev_buffer_index].next = cpu_to_le32(next_bits);
drivers/comedi/drivers/gsc_hpdi.c
149
unsigned int next;
drivers/comedi/drivers/gsc_hpdi.c
151
next = readl(devpriv->plx9080_mmio + PLX_REG_DMAPADR(channel));
drivers/comedi/drivers/gsc_hpdi.c
156
for (desc = 0; (next < start || next >= start + devpriv->block_size) &&
drivers/comedi/drivers/gsc_hpdi.c
432
devpriv->dma_desc[i].next = cpu_to_le32((phys_addr +
drivers/comedi/drivers/gsc_hpdi.c
446
devpriv->dma_desc[i - 1].next = cpu_to_le32(phys_addr | next_bits);
drivers/comedi/drivers/mite.c
673
desc->next = cpu_to_le32(ring->dma_addr +
drivers/comedi/drivers/mite.c
686
desc->next = cpu_to_le32(ring->dma_addr);
drivers/comedi/drivers/mite.h
25
__le32 next;
drivers/comedi/drivers/plx9080.h
46
__le32 next;
drivers/comedi/kcomedilib/kcomedilib_main.c
66
unsigned int next = 1 - cur;
drivers/comedi/kcomedilib/kcomedilib_main.c
75
bitmap_zero(destinations[next], COMEDI_NUM_BOARD_MINORS);
drivers/comedi/kcomedilib/kcomedilib_main.c
83
set_bit(f, destinations[next]);
drivers/comedi/kcomedilib/kcomedilib_main.c
87
cur = next;
drivers/comedi/proc.c
50
for (driv = comedi_drivers; driv; driv = driv->next) {
drivers/cpufreq/cpufreq.c
448
for (nb = cpufreq_transition_notifier_list.head; nb; nb = nb->next)
drivers/cpufreq/powernv-cpufreq.c
1013
.next = NULL,
drivers/cpufreq/powernv-cpufreq.c
269
goto next;
drivers/cpufreq/powernv-cpufreq.c
275
goto next;
drivers/cpufreq/powernv-cpufreq.c
283
next:
drivers/cpufreq/powernv-cpufreq.c
570
goto next;
drivers/cpufreq/powernv-cpufreq.c
591
next:
drivers/cpufreq/pxa3xx-cpufreq.c
167
struct pxa3xx_freq_info *next;
drivers/cpufreq/pxa3xx-cpufreq.c
173
next = &pxa3xx_freqs[index];
drivers/cpufreq/pxa3xx-cpufreq.c
176
__update_core_freq(next);
drivers/cpufreq/pxa3xx-cpufreq.c
177
__update_bus_freq(next);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce.h
183
__le32 next;
drivers/crypto/axis/artpec6_crypto.c
1064
struct artpec6_crypto_bounce_buffer *next;
drivers/crypto/axis/artpec6_crypto.c
1066
list_for_each_entry_safe(b, next, &dma->bounce_buffers, list) {
drivers/crypto/axis/artpec6_crypto.c
433
struct artpec6_crypto_bounce_buffer *next;
drivers/crypto/axis/artpec6_crypto.c
435
list_for_each_entry_safe(b, next, &dma->bounce_buffers, list) {
drivers/crypto/caam/caamalg_qi2.c
5468
if (hash_list.next) {
drivers/crypto/caam/caamhash.c
1890
if (!hash_list.next)
drivers/crypto/ccp/ccp-crypto-main.c
120
req_queue.backlog = crypto_cmd->entry.next;
drivers/crypto/ccp/ccp-crypto-main.c
124
req_queue.backlog = req_queue.backlog->next;
drivers/crypto/ccp/ccp-crypto-main.c
128
req_queue.backlog = crypto_cmd->entry.next;
drivers/crypto/ccp/ccp-crypto-main.c
143
struct ccp_crypto_cmd *held, *next, *backlog;
drivers/crypto/ccp/ccp-crypto-main.c
194
next = ccp_crypto_cmd_complete(held, &backlog);
drivers/crypto/ccp/ccp-crypto-main.c
201
held = next;
drivers/crypto/hisilicon/sec/sec_algs.c
164
sgl_next = sgl_current->next;
drivers/crypto/hisilicon/sec/sec_algs.c
206
sgl_current->next = sgl_next;
drivers/crypto/hisilicon/sec/sec_drv.h
388
struct sec_hw_sgl *next;
drivers/crypto/hisilicon/sec2/sec.h
45
struct sec_hw_sgl *next;
drivers/crypto/hisilicon/sgl.c
31
struct hisi_acc_hw_sgl *next;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
1032
req_ctx->dst = dst_hook.next;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
1042
req_ctx->src = src_hook.next;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
106
struct buffer_desc *next;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
1142
req_ctx->src = src_hook.next;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
1162
req_ctx->dst = dst_hook.next;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
340
buf1 = buf->next;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
728
buf->next = NULL;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
891
buf->next = next_buf;
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
899
buf->next = NULL;
drivers/crypto/intel/keembay/ocs-aes.c
1481
ll[i].next = dll_desc->dma_addr + (sizeof(*ll) * (i + 1));
drivers/crypto/intel/keembay/ocs-aes.c
1485
ll[i - 1].next = 0;
drivers/crypto/intel/keembay/ocs-aes.c
185
u32 next;
drivers/crypto/intel/qat/qat_common/adf_cfg.c
50
.next = qat_dev_cfg_next,
drivers/crypto/intel/qat/qat_common/adf_cfg_user.h
13
struct adf_user_cfg_key_val *next;
drivers/crypto/intel/qat/qat_common/adf_cfg_user.h
26
struct adf_user_cfg_section *next;
drivers/crypto/intel/qat/qat_common/adf_cnv_dbgfs.c
155
.next = qat_cnv_errors_seq_next,
drivers/crypto/intel/qat/qat_common/adf_ctl_drv.c
166
params_head = key_val.next;
drivers/crypto/intel/qat/qat_common/adf_ctl_drv.c
168
section_head = section.next;
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
191
struct vf_id_map *next;
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
197
next = list_next_entry(map, list);
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
198
while (next && &next->list != &vfs_table) {
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
199
next->fake_id++;
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
200
next = list_next_entry(next, list);
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
257
struct vf_id_map *map, *next;
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
266
next = list_next_entry(map, list);
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
267
while (next && &next->list != &vfs_table) {
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
268
next->fake_id--;
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
269
next = list_next_entry(next, list);
drivers/crypto/intel/qat/qat_common/adf_fw_counters.c
188
.next = qat_fw_counters_seq_next,
drivers/crypto/intel/qat/qat_common/adf_transport_debug.c
191
.next = adf_bank_next,
drivers/crypto/intel/qat/qat_common/adf_transport_debug.c
90
.next = adf_ring_next,
drivers/crypto/intel/qat/qat_common/icp_qat_uclo.h
408
struct icp_qat_uof_batch_init *next;
drivers/crypto/intel/qat/qat_common/qat_hal.c
1300
plm_init = lm_init_header->next;
drivers/crypto/intel/qat/qat_common/qat_hal.c
1319
plm_init = plm_init->next;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
171
umem_init = umem_init_header->next;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
180
umem_init = umem_init->next;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
195
umem_init = umem_init->next;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
269
while (tail_old->next)
drivers/crypto/intel/qat/qat_common/qat_uclo.c
270
tail_old = tail_old->next;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
280
mem_init->next = NULL;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
281
tail->next = mem_init;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
289
tail_old = tail_old->next;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
296
mem_init = tail_old->next;
drivers/crypto/marvell/cesa/cesa.h
313
struct mv_cesa_tdma_desc *next;
drivers/crypto/marvell/cesa/tdma.c
117
last->next = dreq->chain.first;
drivers/crypto/marvell/cesa/tdma.c
135
struct mv_cesa_tdma_desc *tdma = NULL, *next = NULL;
drivers/crypto/marvell/cesa/tdma.c
141
for (tdma = engine->chain_hw.first; tdma; tdma = next) {
drivers/crypto/marvell/cesa/tdma.c
143
next = tdma->next;
drivers/crypto/marvell/cesa/tdma.c
163
engine->chain_hw.first = tdma->next;
drivers/crypto/marvell/cesa/tdma.c
164
tdma->next = NULL;
drivers/crypto/marvell/cesa/tdma.c
216
chain->last->next = new_tdma;
drivers/crypto/marvell/cesa/tdma.c
240
for (op_desc = chain->first; op_desc; op_desc = op_desc->next) {
drivers/crypto/marvell/cesa/tdma.c
79
tdma = tdma->next;
drivers/crypto/marvell/cesa/tdma.c
93
for (tdma = dreq->chain.first; tdma; tdma = tdma->next) {
drivers/crypto/sahara.c
130
u32 next;
drivers/crypto/sahara.c
136
u32 next;
drivers/crypto/sahara.c
409
dev->hw_desc[i]->next);
drivers/crypto/sahara.c
427
dev->hw_link[i]->next);
drivers/crypto/sahara.c
452
dev->hw_desc[idx]->next = dev->hw_phys_desc[1];
drivers/crypto/sahara.c
496
dev->hw_link[i]->next = 0;
drivers/crypto/sahara.c
499
dev->hw_link[i]->next = dev->hw_phys_link[i + 1];
drivers/crypto/sahara.c
512
dev->hw_link[j]->next = 0;
drivers/crypto/sahara.c
515
dev->hw_link[j]->next = dev->hw_phys_link[j + 1];
drivers/crypto/sahara.c
524
dev->hw_desc[idx]->next = 0;
drivers/crypto/sahara.c
788
dev->hw_link[i]->next = 0;
drivers/crypto/sahara.c
791
dev->hw_link[i]->next = dev->hw_phys_link[i + 1];
drivers/crypto/sahara.c
838
dev->hw_link[i]->next = 0;
drivers/crypto/sahara.c
866
dev->hw_link[index]->next = 0;
drivers/crypto/sahara.c
945
dev->hw_desc[0]->next = 0;
drivers/crypto/sahara.c
951
dev->hw_desc[0]->next = dev->hw_phys_desc[1];
drivers/crypto/sahara.c
956
dev->hw_desc[1]->next = 0;
drivers/crypto/talitos.c
1108
goto next;
drivers/crypto/talitos.c
1132
next:
drivers/cxl/acpi.c
787
struct resource *res, *next, *cxl = data;
drivers/cxl/acpi.c
789
for (res = cxl->child; res; res = next) {
drivers/cxl/acpi.c
792
next = res->sibling;
drivers/cxl/acpi.c
825
struct resource *res, *new, *next;
drivers/cxl/acpi.c
827
for (res = cxl_res->child; res; res = next) {
drivers/cxl/acpi.c
845
next = res->sibling;
drivers/cxl/acpi.c
846
while (next && resource_overlaps(new, next)) {
drivers/cxl/acpi.c
847
if (resource_contains(new, next)) {
drivers/cxl/acpi.c
848
struct resource *_next = next->sibling;
drivers/cxl/acpi.c
850
remove_resource(next);
drivers/cxl/acpi.c
851
del_cxl_resource(next);
drivers/cxl/acpi.c
852
next = _next;
drivers/cxl/acpi.c
854
next->start = new->end + 1;
drivers/cxl/core/port.c
158
struct cxl_dport *next = NULL;
drivers/cxl/core/port.c
164
next = cxlsd->target[i + 1];
drivers/cxl/core/port.c
166
next ? "," : "");
drivers/cxl/core/region.c
1164
if (ep_iter->next == ep->next) {
drivers/cxl/core/region.c
1174
if (!found || !ep->next) {
drivers/cxl/core/region.c
1239
ep ? ep->next ? dev_name(ep->next->uport_dev) :
drivers/cxl/core/region.c
1281
if (ep_iter->next == ep->next) {
drivers/cxl/core/region.c
1636
iter = ep->next, ep = cxl_ep_load(iter, cxlmd))
drivers/cxl/core/region.c
1673
iter = ep->next, ep = cxl_ep_load(iter, cxlmd)) {
drivers/cxl/core/region.c
275
iter = ep->next, ep = cxl_ep_load(iter, cxlmd)) {
drivers/cxl/core/region.c
337
iter = ep->next, ep = cxl_ep_load(iter, cxlmd)) {
drivers/cxl/cxl.h
736
struct cxl_port *next;
drivers/cxl/port.c
302
ep->next = down;
drivers/dax/bus.c
1058
struct resource *next = res->sibling;
drivers/dax/bus.c
1069
if (next && next->start > res->end + 1)
drivers/dax/bus.c
1070
alloc = min(next->start - (res->end + 1), to_alloc);
drivers/dax/bus.c
1073
if (!alloc && !next && res->end < region_res->end)
drivers/dma-buf/sw_sync.c
212
struct sync_pt *pt, *next;
drivers/dma-buf/sw_sync.c
220
list_for_each_entry_safe(pt, next, &obj->pt_list, link) {
drivers/dma-buf/sw_sync.c
234
list_for_each_entry_safe(pt, next, &signalled, link) {
drivers/dma-buf/sw_sync.c
329
struct sync_pt *pt, *next;
drivers/dma-buf/sw_sync.c
333
list_for_each_entry_safe(pt, next, &obj->pt_list, link) {
drivers/dma/altera-msgdma.c
233
struct msgdma_sw_desc *child, *next;
drivers/dma/altera-msgdma.c
237
list_for_each_entry_safe(child, next, &desc->tx_list, node) {
drivers/dma/altera-msgdma.c
251
struct msgdma_sw_desc *desc, *next;
drivers/dma/altera-msgdma.c
253
list_for_each_entry_safe(desc, next, list, node)
drivers/dma/altera-msgdma.c
535
struct msgdma_sw_desc *sdesc, *next;
drivers/dma/altera-msgdma.c
539
list_for_each_entry_safe(sdesc, next, &desc->tx_list, node)
drivers/dma/altera-msgdma.c
583
struct msgdma_sw_desc *desc, *next;
drivers/dma/altera-msgdma.c
588
list_for_each_entry_safe(desc, next, &mdev->done_list, node) {
drivers/dma/amba-pl08x.c
2422
struct pl08x_dma_chan *next;
drivers/dma/amba-pl08x.c
2425
next, &dmadev->channels, vc.chan.device_node) {
drivers/dma/amba-pl08x.c
906
struct pl08x_dma_chan *p, *next;
drivers/dma/amba-pl08x.c
909
next = NULL;
drivers/dma/amba-pl08x.c
920
next = p;
drivers/dma/amba-pl08x.c
924
if (!next && pl08x->has_slave) {
drivers/dma/amba-pl08x.c
928
next = p;
drivers/dma/amba-pl08x.c
936
if (next) {
drivers/dma/amba-pl08x.c
943
spin_lock(&next->vc.lock);
drivers/dma/amba-pl08x.c
945
success = next->state == PL08X_CHAN_WAITING;
drivers/dma/amba-pl08x.c
947
pl08x_phy_reassign_start(plchan->phychan, next);
drivers/dma/amba-pl08x.c
948
spin_unlock(&next->vc.lock);
drivers/dma/bcm-sba-raid.c
109
struct list_head next;
drivers/dma/bcm-sba-raid.c
1501
INIT_LIST_HEAD(&req->next);
drivers/dma/bcm-sba-raid.c
224
INIT_LIST_HEAD(&req->next);
drivers/dma/bcm-sba-raid.c
295
list_for_each_entry(nreq, &req->next, next)
drivers/dma/bcm-sba-raid.c
309
list_add_tail(&req->next, &first->next);
drivers/dma/bcm-sba-raid.c
434
list_for_each_entry(nreq, &first->next, next)
drivers/dma/bcm-sba-raid.c
436
INIT_LIST_HEAD(&first->next);
drivers/dma/bcm-sba-raid.c
532
list_for_each_entry(nreq, &req->next, next)
drivers/dma/bcm2835-dma.c
325
control_block->next = 0;
drivers/dma/bcm2835-dma.c
340
d->cb_list[frame - 1].cb->next = cb_entry->paddr;
drivers/dma/bcm2835-dma.c
59
uint32_t next;
drivers/dma/bcm2835-dma.c
756
d->cb_list[d->frames - 1].cb->next = d->cb_list[0].paddr;
drivers/dma/bcm2835-dma.c
827
struct bcm2835_chan *c, *next;
drivers/dma/bcm2835-dma.c
829
list_for_each_entry_safe(c, next, &od->ddev.channels,
drivers/dma/bcm2835-dma.c
864
struct bcm2835_chan *c, *next;
drivers/dma/bcm2835-dma.c
866
list_for_each_entry_safe(c, next, &od->ddev.channels,
drivers/dma/dw/core.c
168
dwc->tx_node_active = dwc->tx_node_active->next;
drivers/dma/dw/core.c
224
list_move(dwc->queue.next, &dwc->active_list);
drivers/dma/dw/core.c
326
if (active == head->next)
drivers/dma/dw/core.c
438
list_move(dwc->queue.next, dwc->active_list.prev);
drivers/dma/dw/core.c
51
return to_dw_desc(dwc->active_list.next);
drivers/dma/fsldma.c
121
struct fsl_dma_ld_hw *hw, dma_addr_t next)
drivers/dma/fsldma.c
127
hw->next_ln_addr = CPU_TO_DMA(chan, snoop_bits | next, 64);
drivers/dma/img-mdc-dma.c
1023
struct mdc_chan *mchan, *next;
drivers/dma/img-mdc-dma.c
1028
list_for_each_entry_safe(mchan, next, &mdma->dma_dev.channels,
drivers/dma/img-mdc-dma.c
262
struct mdc_hw_list_desc *curr, *next;
drivers/dma/img-mdc-dma.c
268
next = curr->next_desc;
drivers/dma/img-mdc-dma.c
271
curr = next;
drivers/dma/imx-dma.c
626
list_move_tail(imxdmac->ld_active.next, &imxdmac->ld_free);
drivers/dma/imx-dma.c
631
list_move_tail(imxdmac->ld_queue.next, &imxdmac->ld_active);
drivers/dma/imx-dma.c
731
list_move_tail(imxdmac->ld_free.next, &imxdmac->ld_queue);
drivers/dma/imx-dma.c
997
list_move_tail(imxdmac->ld_queue.next,
drivers/dma/ioat/dma.c
437
struct ioat_ring_ent *next = ring[i+1];
drivers/dma/ioat/dma.c
440
hw->next = next->txd.phys;
drivers/dma/ioat/dma.c
442
ring[i]->hw->next = ring[0]->txd.phys;
drivers/dma/ioat/dma.h
229
(unsigned long long) hw->next, tx->cookie, tx->flags,
drivers/dma/ioat/hw.h
124
uint64_t next;
drivers/dma/ioat/hw.h
135
uint64_t next;
drivers/dma/ioat/hw.h
178
uint64_t next;
drivers/dma/ioat/hw.h
192
uint64_t next;
drivers/dma/ioat/hw.h
222
uint64_t next;
drivers/dma/ioat/hw.h
91
uint64_t next;
drivers/dma/ioat/init.c
330
dma_chan = container_of(dma->channels.next, struct dma_chan,
drivers/dma/ioat/init.c
830
dma_chan = container_of(dma->channels.next, struct dma_chan,
drivers/dma/ioat/prep.c
290
(unsigned long long) (pq_ex ? pq_ex->next : pq->next),
drivers/dma/ioat/prep.c
300
dev_dbg(dev, "\tNEXT: %#llx\n", pq->next);
drivers/dma/ioat/prep.c
323
(unsigned long long) pq->next,
drivers/dma/mediatek/mtk-cqdma.c
527
tx->next = NULL;
drivers/dma/mediatek/mtk-cqdma.c
532
prev_tx->next = tx;
drivers/dma/mediatek/mtk-hsdma.c
542
u16 next;
drivers/dma/mediatek/mtk-hsdma.c
559
next = MTK_HSDMA_NEXT_DESP_IDX(pc->ring.cur_rptr,
drivers/dma/mediatek/mtk-hsdma.c
561
rxd = &pc->ring.rxd[next];
drivers/dma/mediatek/mtk-hsdma.c
571
cb = &pc->ring.cb[next];
drivers/dma/mediatek/mtk-hsdma.c
611
pc->ring.cur_rptr = next;
drivers/dma/mv_xor.c
311
iter = list_entry(mv_chan->chain.next,
drivers/dma/mv_xor.c
321
iter = list_entry(iter->node.next,
drivers/dma/mxs-dma.c
507
ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * idx;
drivers/dma/mxs-dma.c
543
ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * idx;
drivers/dma/mxs-dma.c
608
ccw->next = mxs_chan->ccw_phys;
drivers/dma/mxs-dma.c
610
ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * (i + 1);
drivers/dma/mxs-dma.c
96
u32 next;
drivers/dma/nbpfaxi.c
127
u32 next;
drivers/dma/nbpfaxi.c
836
prev->hwdesc->next = (u32)ldesc->hwdesc_dma_addr;
drivers/dma/nbpfaxi.c
845
prev->hwdesc->next = 0;
drivers/dma/owl-dma.c
1053
struct owl_dma_vchan *next;
drivers/dma/owl-dma.c
1056
next, &od->dma.channels, vc.chan.device_node) {
drivers/dma/owl-dma.c
366
struct owl_dma_lli *next,
drivers/dma/owl-dma.c
370
list_add_tail(&next->node, &txd->lli_list);
drivers/dma/owl-dma.c
373
prev->hw[OWL_DMADESC_NEXT_LLI] = next->phys;
drivers/dma/owl-dma.c
378
return next;
drivers/dma/pch_dma.c
329
pd_chan->chan.chan_id, desc->regs.next);
drivers/dma/pch_dma.c
335
channel_writel(pd_chan, NEXT, desc->regs.next);
drivers/dma/pch_dma.c
599
desc->regs.next = DMA_DESC_FOLLOW_WITHOUT_IRQ;
drivers/dma/pch_dma.c
624
prev->regs.next |= desc->txd.phys;
drivers/dma/pch_dma.c
632
desc->regs.next = DMA_DESC_END_WITH_IRQ;
drivers/dma/pch_dma.c
634
desc->regs.next = DMA_DESC_END_WITHOUT_IRQ;
drivers/dma/pch_dma.c
68
u32 next;
drivers/dma/pch_dma.c
750
pd->ch_regs[i].next = channel_readl(pd_chan, NEXT);
drivers/dma/pch_dma.c
773
channel_writel(pd_chan, NEXT, pd->ch_regs[i].next);
drivers/dma/pl330.c
2512
desc = list_entry(last->node.next, struct dma_pl330_desc, node);
drivers/dma/pl330.c
2575
desc = list_entry(pool->next,
drivers/dma/pl330.c
2717
desc = list_entry(first->node.next,
drivers/dma/pl330.c
2830
desc = list_entry(first->node.next,
drivers/dma/ppc4xx/adma.c
1451
slot = list_entry(slot->slot_node.next,
drivers/dma/ppc4xx/adma.c
1740
iter = list_entry(iter->slot_node.next,
drivers/dma/ppc4xx/adma.c
1809
list_entry(ppc440spe_chan->all_slots.next,
drivers/dma/ppc4xx/adma.c
2115
iter->hw_next = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
2198
iter->hw_next = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
2221
iter->hw_next = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
2466
list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
2639
iter->hw_next = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
2665
iter = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
2671
iter->hw_next = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
2992
iter = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
3004
iter = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
3073
end = list_entry(end->chain_node.next,
drivers/dma/ppc4xx/adma.c
3247
desc = list_entry(hdesc->chain_node.next,
drivers/dma/ppc4xx/adma.c
3368
desc = list_entry(desc->chain_node.next,
drivers/dma/ppc4xx/adma.c
3409
desc = list_entry(desc->chain_node.next,
drivers/dma/ppc4xx/adma.c
343
iter->hw_next = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
447
iter = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
451
iter = list_entry(iter->chain_node.next,
drivers/dma/ppc4xx/adma.c
478
iter->hw_next = list_entry(iter->chain_node.next,
drivers/dma/qcom/hidma.c
114
struct hidma_desc *next;
drivers/dma/qcom/hidma.c
126
list_for_each_entry_safe(mdesc, next, &list, node) {
drivers/dma/qcom/hidma.c
174
if (mdesc->node.next) {
drivers/dma/qcom/hidma.c
232
struct hidma_desc *qdesc, *next;
drivers/dma/qcom/hidma.c
236
list_for_each_entry_safe(qdesc, next, &mchan->queued, node) {
drivers/dma/sh/rz-dmac.c
516
list_move_tail(channel->ld_free.next, &channel->ld_queue);
drivers/dma/sh/rz-dmac.c
554
list_move_tail(channel->ld_free.next, &channel->ld_queue);
drivers/dma/sh/rz-dmac.c
600
list_move_tail(channel->ld_queue.next,
drivers/dma/sh/rz-dmac.c
745
list_move_tail(channel->ld_active.next, &channel->ld_free);
drivers/dma/sh/rz-dmac.c
751
list_move_tail(channel->ld_queue.next, &channel->ld_active);
drivers/dma/st_fdma.c
332
hw_node->next = 0;
drivers/dma/st_fdma.c
487
hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
drivers/dma/st_fdma.c
536
hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
drivers/dma/st_fdma.h
50
u32 next;
drivers/dma/ste_dma40_ll.c
224
dma_addr_t next = lli_phys;
drivers/dma/ste_dma40_ll.c
244
next = cyclic ? first_phys : 0;
drivers/dma/ste_dma40_ll.c
246
next = ALIGN(next + sizeof(struct d40_phy_lli),
drivers/dma/ste_dma40_ll.c
249
err = d40_phy_fill_lli(lli, addr, size_seg, next,
drivers/dma/ste_dma40_ll.c
313
int next, unsigned int flags)
drivers/dma/ste_dma40_ll.c
319
if (next != -EINVAL) {
drivers/dma/ste_dma40_ll.c
320
slos = next * 2;
drivers/dma/ste_dma40_ll.c
321
dlos = next * 2 + 1;
drivers/dma/ste_dma40_ll.c
339
int next, unsigned int flags)
drivers/dma/ste_dma40_ll.c
341
d40_log_lli_link(lli_dst, lli_src, next, flags);
drivers/dma/ste_dma40_ll.c
352
int next, unsigned int flags)
drivers/dma/ste_dma40_ll.c
354
d40_log_lli_link(lli_dst, lli_src, next, flags);
drivers/dma/ste_dma40_ll.h
463
int next, unsigned int flags);
drivers/dma/ste_dma40_ll.h
468
int next, unsigned int flags);
drivers/dma/stm32/stm32-dma3.c
506
u32 next = curr + 1;
drivers/dma/stm32/stm32-dma3.c
521
next_lli = swdesc->lli[next].hwdesc_addr;
drivers/dma/stm32/stm32-mdma.c
696
u32 next = count + 1;
drivers/dma/stm32/stm32-mdma.c
718
hwdesc->clar = desc->node[next].hwdesc_phys;
drivers/dma/sun6i-dma.c
375
struct sun6i_dma_lli *next,
drivers/dma/sun6i-dma.c
379
if ((!prev && !txd) || !next)
drivers/dma/sun6i-dma.c
384
txd->v_lli = next;
drivers/dma/sun6i-dma.c
387
prev->v_lli_next = next;
drivers/dma/sun6i-dma.c
390
next->p_lli_next = LLI_LAST_ITEM;
drivers/dma/sun6i-dma.c
391
next->v_lli_next = NULL;
drivers/dma/sun6i-dma.c
393
return next;
drivers/dma/timb_dma.c
188
td_desc = list_entry(td_chan->active_list.next, struct timb_dma_desc,
drivers/dma/timb_dma.c
229
td_desc = list_entry(td_chan->active_list.next, struct timb_dma_desc,
drivers/dma/timb_dma.c
267
list_entry(td_chan->active_list.next,
drivers/dma/timb_dma.c
284
td_desc = list_entry(td_chan->queue.next, struct timb_dma_desc,
drivers/dma/txx9dmac.c
165
return list_entry(dc->active_list.next,
drivers/dma/txx9dmac.c
177
return list_entry(dc->queue.next, struct txx9dmac_desc, desc_node);
drivers/dma/txx9dmac.c
948
desc = list_entry(list.next, struct txx9dmac_desc, desc_node);
drivers/dma/xilinx/xilinx_dma.c
1069
struct xilinx_dma_tx_descriptor *desc, *next;
drivers/dma/xilinx/xilinx_dma.c
1074
list_for_each_entry_safe(desc, next, &chan->done_list, node) {
drivers/dma/xilinx/xilinx_dma.c
1750
struct xilinx_dma_tx_descriptor *desc, *next;
drivers/dma/xilinx/xilinx_dma.c
1756
list_for_each_entry_safe(desc, next, &chan->active_list, node) {
drivers/dma/xilinx/xilinx_dma.c
857
struct xilinx_vdma_tx_segment *segment, *next;
drivers/dma/xilinx/xilinx_dma.c
866
list_for_each_entry_safe(segment, next, &desc->segments, node) {
drivers/dma/xilinx/xilinx_dma.c
903
struct xilinx_dma_tx_descriptor *desc, *next;
drivers/dma/xilinx/xilinx_dma.c
905
list_for_each_entry_safe(desc, next, list, node) {
drivers/dma/xilinx/xilinx_dpdma.c
657
struct xilinx_dpdma_sw_desc *sw_desc, *next;
drivers/dma/xilinx/xilinx_dpdma.c
665
list_for_each_entry_safe(sw_desc, next, &desc->descriptors, node) {
drivers/dma/xilinx/zynqmp_dma.c
445
struct zynqmp_dma_desc_sw *child, *next;
drivers/dma/xilinx/zynqmp_dma.c
449
list_for_each_entry_safe(child, next, &sdesc->tx_list, node) {
drivers/dma/xilinx/zynqmp_dma.c
463
struct zynqmp_dma_desc_sw *desc, *next;
drivers/dma/xilinx/zynqmp_dma.c
465
list_for_each_entry_safe(desc, next, list, node)
drivers/dma/xilinx/zynqmp_dma.c
616
struct zynqmp_dma_desc_sw *desc, *next;
drivers/dma/xilinx/zynqmp_dma.c
621
list_for_each_entry_safe(desc, next, &chan->done_list, node) {
drivers/dpll/zl3073x/core.c
925
struct zl3073x_dpll *zldpll, *next;
drivers/dpll/zl3073x/core.c
938
list_for_each_entry_safe(zldpll, next, &zldev->dplls, list) {
drivers/dpll/zl3073x/dpll.c
1515
struct zl3073x_dpll_pin *pin, *next;
drivers/dpll/zl3073x/dpll.c
1517
list_for_each_entry_safe(pin, next, &zldpll->pins, list) {
drivers/edac/altera_edac.c
1606
list_add(&altdev->next, &altdev->edac->a10_ecc_devices);
drivers/edac/altera_edac.c
2010
list_add(&altdev->next, &edac->a10_ecc_devices);
drivers/edac/altera_edac.c
2087
next);
drivers/edac/altera_edac.h
376
struct list_head next;
drivers/edac/sb_edac.c
1725
goto next;
drivers/edac/sb_edac.c
1740
next:
drivers/edac/sb_edac.c
757
sbridge_dev = list_entry(prev ? prev->list.next
drivers/edac/sb_edac.c
758
: sbridge_edac_list.next, struct sbridge_dev, list);
drivers/edac/xgene_edac.c
1018
struct list_head next;
drivers/edac/xgene_edac.c
1246
list_add(&ctx->next, &edac->l3s);
drivers/edac/xgene_edac.c
133
struct list_head next;
drivers/edac/xgene_edac.c
1783
list_add(&ctx->next, &edac->socs);
drivers/edac/xgene_edac.c
1825
list_for_each_entry(mcu, &ctx->mcus, next)
drivers/edac/xgene_edac.c
1829
list_for_each_entry(pmd, &ctx->pmds, next) {
drivers/edac/xgene_edac.c
1834
list_for_each_entry(node, &ctx->l3s, next)
drivers/edac/xgene_edac.c
1837
list_for_each_entry(node, &ctx->socs, next)
drivers/edac/xgene_edac.c
1972
list_for_each_entry_safe(mcu, temp_mcu, &edac->mcus, next)
drivers/edac/xgene_edac.c
1975
list_for_each_entry_safe(pmd, temp_pmd, &edac->pmds, next)
drivers/edac/xgene_edac.c
1978
list_for_each_entry_safe(node, temp_node, &edac->l3s, next)
drivers/edac/xgene_edac.c
1981
list_for_each_entry_safe(node, temp_node, &edac->socs, next)
drivers/edac/xgene_edac.c
422
list_add(&ctx->next, &edac->mcus);
drivers/edac/xgene_edac.c
510
struct list_head next;
drivers/edac/xgene_edac.c
964
list_add(&ctx->next, &edac->pmds);
drivers/firewire/core-cdev.c
1118
struct fw_cdev_iso_packet __user *p, *end, *next;
drivers/firewire/core-cdev.c
1182
next = (struct fw_cdev_iso_packet __user *)
drivers/firewire/core-cdev.c
1184
if (next > end)
drivers/firewire/core-cdev.c
1198
p = next;
drivers/firewire/core-topology.c
215
child = fw_node(child->link.next);
drivers/firewire/core-topology.c
267
struct fw_node *node, *next, *child, *parent;
drivers/firewire/core-topology.c
293
list_for_each_entry_safe(node, next, &list, link)
drivers/firewire/core-topology.c
370
node0 = fw_node(list0.next);
drivers/firewire/core-topology.c
371
node1 = fw_node(list1.next);
drivers/firewire/core-topology.c
433
node0 = fw_node(node0->link.next);
drivers/firewire/core-topology.c
434
next1 = fw_node(node1->link.next);
drivers/firewire/net.c
301
fi2 = list_entry(fi->fi_link.next,
drivers/firewire/net.c
437
fi = list_entry(pd->fi_list.next, struct fwnet_fragment_info, fi_link);
drivers/firewire/nosy.c
43
__le32 next;
drivers/firewire/nosy.c
586
lynx->rcv_start_pcl->next = cpu_to_le32(lynx->rcv_pcl_bus);
drivers/firewire/nosy.c
587
lynx->rcv_pcl->next = cpu_to_le32(PCL_NEXT_INVALID);
drivers/firewire/ohci.c
1050
ctx->buffer_tail = list_entry(ctx->buffer_list.next,
drivers/firewire/ohci.c
1096
if (desc->list.next == &ctx->buffer_list) {
drivers/firewire/ohci.c
1102
desc = list_entry(desc->list.next,
drivers/firewire/ohci.c
780
void *next;
drivers/firewire/ohci.c
783
next = handle_ar_packet(ctx, p);
drivers/firewire/ohci.c
784
if (!next)
drivers/firewire/ohci.c
786
p = next;
drivers/firewire/ohci.c
953
desc = list_entry(ctx->buffer_list.next,
drivers/firewire/ohci.c
967
desc = list_entry(desc->list.next,
drivers/firewire/sbp2.c
1216
struct sbp2_logical_unit *lu, *next;
drivers/firewire/sbp2.c
1224
list_for_each_entry_safe(lu, next, &tgt->lu_list, link) {
drivers/firewire/sbp2.c
1460
orb->request.next.high = cpu_to_be32(SBP2_ORB_NULL);
drivers/firewire/sbp2.c
308
struct sbp2_pointer next;
drivers/firewire/sbp2.c
516
struct sbp2_orb *orb, *next;
drivers/firewire/sbp2.c
525
list_for_each_entry_safe(orb, next, &list, link) {
drivers/firmware/dmi-sysfs.c
640
struct dmi_sysfs_entry *entry, *next;
drivers/firmware/dmi-sysfs.c
643
list_for_each_entry_safe(entry, next, &entry_list, list) {
drivers/firmware/dmi_scan.c
1003
for (d = head->next; d != &dmi_devices; d = d->next) {
drivers/firmware/efi/apple-properties.c
195
pa_data = data->next;
drivers/firmware/efi/efi.c
1122
for (prsv = efi_memreserve_root->next; prsv; ) {
drivers/firmware/efi/efi.c
1134
prsv = rsv->next;
drivers/firmware/efi/efi.c
1161
rsv->next = efi_memreserve_root->next;
drivers/firmware/efi/efi.c
1162
efi_memreserve_root->next = __pa(rsv);
drivers/firmware/efi/efi.c
799
prsv = rsv->next;
drivers/firmware/efi/esrt.c
361
struct esre_entry *entry, *next;
drivers/firmware/efi/esrt.c
363
list_for_each_entry_safe(entry, next, &entry_list, list) {
drivers/firmware/efi/libstub/efi-stub.c
93
rsv->next = 0;
drivers/firmware/efi/libstub/gop.c
101
*next = option;
drivers/firmware/efi/libstub/gop.c
105
static bool parse_list(char *option, char **next)
drivers/firmware/efi/libstub/gop.c
114
*next = option;
drivers/firmware/efi/libstub/gop.c
39
static bool parse_modenum(char *option, char **next)
drivers/firmware/efi/libstub/gop.c
52
*next = option;
drivers/firmware/efi/libstub/gop.c
56
static bool parse_res(char *option, char **next)
drivers/firmware/efi/libstub/gop.c
88
*next = option;
drivers/firmware/efi/libstub/gop.c
92
static bool parse_auto(char *option, char **next)
drivers/firmware/efi/libstub/x86-stub.c
131
while (data && data->next)
drivers/firmware/efi/libstub/x86-stub.c
132
data = (struct setup_data *)(unsigned long)data->next;
drivers/firmware/efi/libstub/x86-stub.c
148
data->next = (unsigned long)rom;
drivers/firmware/efi/libstub/x86-stub.c
194
new->next = 0;
drivers/firmware/efi/libstub/x86-stub.c
200
while (data->next)
drivers/firmware/efi/libstub/x86-stub.c
201
data = (struct setup_data *)(unsigned long)data->next;
drivers/firmware/efi/libstub/x86-stub.c
202
data->next = (unsigned long)new;
drivers/firmware/efi/libstub/x86-stub.c
262
const u8 *next;
drivers/firmware/efi/libstub/x86-stub.c
272
next = p + hdr->length;
drivers/firmware/efi/libstub/x86-stub.c
275
while ((next[0] != 0 || next[1] != 0) && next + 1 < end)
drivers/firmware/efi/libstub/x86-stub.c
276
next++;
drivers/firmware/efi/libstub/x86-stub.c
278
next += 2;
drivers/firmware/efi/libstub/x86-stub.c
279
p = next;
drivers/firmware/efi/libstub/x86-stub.c
372
unsigned long end, next;
drivers/firmware/efi/libstub/x86-stub.c
407
for (end = start + size; start < end; start = next) {
drivers/firmware/efi/libstub/x86-stub.c
414
next = desc.base_address + desc.length;
drivers/firmware/efi/libstub/x86-stub.c
428
unprotect_size = min(rounded_end, next) - unprotect_start;
drivers/firmware/efi/libstub/x86-stub.c
558
e820ext->next = 0;
drivers/firmware/efi/libstub/x86-stub.c
562
while (data && data->next)
drivers/firmware/efi/libstub/x86-stub.c
563
data = (struct setup_data *)(unsigned long)data->next;
drivers/firmware/efi/libstub/x86-stub.c
566
data->next = (unsigned long)e820ext;
drivers/firmware/efi/libstub/x86-stub.c
76
rom->data.next = 0;
drivers/firmware/efi/mokvar-table.c
145
next:
drivers/firmware/efi/mokvar-table.c
170
goto next;
drivers/firmware/qcom/qcom_tzmem.c
303
struct qcom_tzmem_area *area, *next;
drivers/firmware/qcom/qcom_tzmem.c
326
list_for_each_entry_safe(area, next, &pool->areas, list) {
drivers/firmware/qemu_fw_cfg.c
388
struct fw_cfg_sysfs_entry *entry, *next;
drivers/firmware/qemu_fw_cfg.c
390
list_for_each_entry_safe(entry, next, &fw_cfg_entry_cache, list) {
drivers/firmware/qemu_fw_cfg.c
564
struct kobject *k, *next;
drivers/firmware/qemu_fw_cfg.c
566
list_for_each_entry_safe(k, next, &kset->list, entry)
drivers/firmware/stratix10-svc.c
1328
hash_add(actrl->trx_list, &handle->next,
drivers/firmware/stratix10-svc.c
1357
hash_del(&handle->next);
drivers/firmware/stratix10-svc.c
1462
if (!hash_hashed(&handle->next)) {
drivers/firmware/stratix10-svc.c
1532
if (!hash_hashed(&handle->next)) {
drivers/firmware/stratix10-svc.c
1536
hash_del(&handle->next);
drivers/firmware/stratix10-svc.c
1655
hash_for_each_safe(actrl->trx_list, i, tmp, handler, next) {
drivers/firmware/stratix10-svc.c
1658
hash_del(&handler->next);
drivers/firmware/stratix10-svc.c
197
struct hlist_node next;
drivers/fpga/dfl.c
1121
u64 v, next;
drivers/fpga/dfl.c
1130
next = FIELD_GET(DFHv1_PARAM_HDR_NEXT_OFFSET, v);
drivers/fpga/dfl.c
1131
if (!next)
drivers/fpga/dfl.c
1134
size += next * sizeof(u64);
drivers/fpga/dfl.c
984
u64 v, next;
drivers/fpga/dfl.c
994
next = FIELD_GET(DFHv1_PARAM_HDR_NEXT_OFFSET, v);
drivers/fpga/dfl.c
995
params += next;
drivers/fpga/fpga-bridge.c
215
struct fpga_bridge *bridge, *next;
drivers/fpga/fpga-bridge.c
218
list_for_each_entry_safe(bridge, next, bridge_list, node) {
drivers/gpib/common/gpib_os.c
1163
for (list_ptr = head->next; list_ptr != head; list_ptr = list_ptr->next) {
drivers/gpib/common/gpib_os.c
1195
for (list_ptr = head->next; list_ptr != head; list_ptr = list_ptr->next) {
drivers/gpib/common/gpib_os.c
1971
struct list_head *front = head->next;
drivers/gpib/common/gpib_os.c
2122
for (list_ptr = registered_drivers.next; list_ptr != &registered_drivers;) {
drivers/gpib/common/gpib_os.c
2126
list_ptr = list_ptr->next;
drivers/gpib/common/gpib_os.c
223
struct list_head *front = head->next;
drivers/gpib/common/gpib_os.c
257
for (list_ptr = head->next; list_ptr != head; list_ptr = list_ptr->next) {
drivers/gpib/common/gpib_os.c
441
if (head->next == head)
drivers/gpib/common/gpib_os.c
448
for (cur = head->next; cur != head; cur = cur->next) {
drivers/gpib/common/gpib_os.c
836
for (list_ptr = registered_drivers.next; list_ptr != &registered_drivers;
drivers/gpib/common/gpib_os.c
837
list_ptr = list_ptr->next) {
drivers/gpib/tnt4882/mite.c
105
struct mite_struct *mite, *next;
drivers/gpib/tnt4882/mite.c
107
for (mite = mite_devices; mite; mite = next) {
drivers/gpib/tnt4882/mite.c
108
next = mite->next;
drivers/gpib/tnt4882/mite.c
66
mite->next = mite_devices;
drivers/gpib/tnt4882/mite.h
28
u32 next;
drivers/gpib/tnt4882/mite.h
32
struct mite_struct *next;
drivers/gpib/tnt4882/tnt4882_gpib.c
931
for (mite = mite_devices; mite; mite = mite->next) {
drivers/gpio/gpiolib.c
497
struct gpio_device *prev, *next;
drivers/gpio/gpiolib.c
507
next = list_first_entry(&gpio_devices, struct gpio_device, list);
drivers/gpio/gpiolib.c
508
if (gdev->base + gdev->ngpio <= next->base) {
drivers/gpio/gpiolib.c
521
list_for_each_entry_safe(prev, next, &gpio_devices, list) {
drivers/gpio/gpiolib.c
523
if (&next->list == &gpio_devices)
drivers/gpio/gpiolib.c
528
&& gdev->base + gdev->ngpio <= next->base) {
drivers/gpio/gpiolib.c
5394
struct gpio_device *gdev = v, *next;
drivers/gpio/gpiolib.c
5396
next = list_entry_rcu(gdev->list.next, struct gpio_device, list);
drivers/gpio/gpiolib.c
5397
gdev = &next->list == &gpio_devices ? NULL : next;
drivers/gpio/gpiolib.c
5456
.next = gpiolib_seq_next,
drivers/gpu/drm/adp/adp-mipi.c
177
struct drm_bridge *next;
drivers/gpu/drm/adp/adp-mipi.c
180
next = devm_drm_of_get_bridge(adp->dsi.dev, adp->dsi.dev->of_node, 1, 0);
drivers/gpu/drm/adp/adp-mipi.c
181
if (IS_ERR(next))
drivers/gpu/drm/adp/adp-mipi.c
182
return PTR_ERR(next);
drivers/gpu/drm/adp/adp-mipi.c
184
adp->next_bridge = next;
drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
491
for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1067
for (base = robj->vm_bo; base; base = base->next)
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
124
char *next;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
140
next = strchr(p, ',');
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
141
if (!next)
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
143
p = next + 1;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
459
fault = &gmc->fault_ring[fault->next];
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
472
fault->next = gmc->fault_hash[hash].idx;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
528
fault = &gmc->fault_ring[fault->next];
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h
102
uint64_t next:AMDGPU_GMC_FAULT_RING_ORDER;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
130
struct list_head *next;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
148
next = block->link.next;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
149
block = list_entry(next, struct drm_buddy_block, link);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
59
struct list_head *head, *next;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
82
next = block->link.next;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
83
if (next != head)
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
84
block = list_entry(next, struct drm_buddy_block, link);
drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
457
TP_PROTO(void *prev, void *next),
drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
458
TP_ARGS(prev, next),
drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
461
__field(void *, next)
drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
466
__entry->next = next;
drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
470
__entry->next)
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2512
struct dma_fence *next = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2530
&next, true,
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2536
*fence = next;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2568
struct dma_fence *next;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2581
&next, true, k_job_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2586
fence = next;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
320
struct dma_fence *next;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
360
&next, true, copy_flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
365
fence = next;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2026
struct amdgpu_bo_va_mapping *before, *after, *tmp, *next;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2082
list_for_each_entry_safe(tmp, next, &removed, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2190
struct amdgpu_bo_va_mapping *mapping, *next;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2203
base = &(*base)->next) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2208
*base = bo_va->base.next;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2217
list_for_each_entry_safe(mapping, next, &bo_va->valids, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2224
list_for_each_entry_safe(mapping, next, &bo_va->invalids, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2285
for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2320
for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
336
for (base = bo->vm_bo; base; base = base->next)
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
414
base->next = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
419
base->next = bo->vm_bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
485
while (!list_is_head(prev->next, &vm->done)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
486
bo_va = list_entry(prev->next, typeof(*bo_va), base.vm_status);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
500
prev = prev->next;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
918
for (base = bo->vm_bo; base; base = base->next) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
206
struct amdgpu_vm_bo_base *next;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
564
struct amdgpu_vm_bo_base *entry, *next;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
575
list_for_each_entry_safe(entry, next, &params->tlb_flush_waitlist, vm_status)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
70
while (head != block->link.next) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
74
block = list_entry(block->link.next, struct drm_buddy_block, link);
drivers/gpu/drm/amd/amdkfd/kfd_debugfs.c
175
struct debugfs_proc_entry *entry, *next;
drivers/gpu/drm/amd/amdkfd/kfd_debugfs.c
178
list_for_each_entry_safe(entry, next, &procs, list) {
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1439
struct device_process_node *cur, *next;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1446
list_for_each_entry_safe(cur, next, &dqm->queues, list) {
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
135
struct dma_fence *next;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
162
NULL, &next, true, 0);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
169
*mfence = next;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
540
unsigned long next;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
546
next = min(vma->vm_end, end);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
547
r = svm_migrate_vma_to_vram(node, prange, vma, addr, next, trigger, ttm_res_offset);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
554
ttm_res_offset += next - addr;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
555
addr = next;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
833
unsigned long next;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
842
next = min(vma->vm_end, end);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
843
r = svm_migrate_vma_to_ram(node, prange, vma, addr, next, trigger,
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
851
addr = next;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
110
struct temp_sdma_queue_list *sdma_q, *next;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
221
list_for_each_entry_safe(sdma_q, next, &sdma_q_list.list, list) {
drivers/gpu/drm/amd/amdkfd/kfd_process.c
238
list_for_each_entry_safe(sdma_q, next, &sdma_q_list.list, list) {
drivers/gpu/drm/amd/amdkfd/kfd_process.c
247
list_for_each_entry_safe(sdma_q, next, &sdma_q_list.list, list) {
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
220
struct process_queue_node *pqn, *next;
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
222
list_for_each_entry_safe(pqn, next, &pqm->queues, process_queue_list) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1756
unsigned long next = 0;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1765
next = min(vma->vm_end, end);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1766
npages = (next - addr) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1786
addr = next;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1842
if (!r && next == end)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1847
addr = next;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2213
struct interval_tree_node *next;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2220
next = interval_tree_iter_next(node, start, last);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2273
node = next;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3322
struct svm_range *next;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3340
list_for_each_entry_safe(prange, next, &p->svms.list, list) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3707
struct svm_range *next;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3743
list_for_each_entry_safe(prange, next, &insert_list, list) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3753
list_for_each_entry_safe(prange, next, &remove_list, update_list) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3922
struct interval_tree_node *next;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3925
next = interval_tree_iter_next(node, start, last);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3963
node = next;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
4012
struct criu_svm_metadata *next = NULL;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
4098
list_for_each_entry_safe(criu_svm_md, next, &svms->criu_svm_metadata_list, list) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
122
while (dev->mem_props.next != &dev->mem_props) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
123
mem = container_of(dev->mem_props.next,
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
129
while (dev->cache_props.next != &dev->cache_props) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
130
cache = container_of(dev->cache_props.next,
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
136
while (dev->io_link_props.next != &dev->io_link_props) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
137
iolink = container_of(dev->io_link_props.next,
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
143
while (dev->p2p_link_props.next != &dev->p2p_link_props) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
144
p2plink = container_of(dev->p2p_link_props.next,
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
150
while (dev->perf_props.next != &dev->perf_props) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
151
perf = container_of(dev->perf_props.next,
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1595
goto next;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1605
next:
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
920
list_move_tail(temp_list->next, master_list);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
1630
colorop = old_colorop->next;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
1790
colorop = old_colorop->next;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
1891
colorop = colorop->next;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
1902
colorop = colorop->next;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
1914
colorop = colorop->next;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
1925
colorop = colorop->next;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
1930
colorop = colorop->next;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
1942
colorop = colorop->next;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
1953
colorop = colorop->next;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_helpers.c
254
list_for_each_entry(pos, &mst_state->payloads, next) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c
594
handler_data = container_of(handler_list->next, struct amdgpu_dm_irq_handler_data, list);
drivers/gpu/drm/amd/pm/powerplay/inc/power_state.h
41
struct pp_power_state *next;
drivers/gpu/drm/amd/pm/powerplay/inc/smu_ucode_xfer_cz.h
116
uint16_t next;
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu8_smumgr.c
338
task->next = is_last ? END_OF_TASK_LIST : smu8_smu->toc_entry_used_count;
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu8_smumgr.c
375
task->next = is_last ? END_OF_TASK_LIST : smu8_smu->toc_entry_used_count;
drivers/gpu/drm/arm/display/komeda/komeda_private_obj.c
428
struct drm_private_obj *obj, *next;
drivers/gpu/drm/arm/display/komeda/komeda_private_obj.c
430
list_for_each_entry_safe(obj, next, &config->privobj_list, head)
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_dc.h
243
dma_addr_t next;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
1129
dscr->next = dscr_dma;
drivers/gpu/drm/bridge/ti-tdp158.c
14
struct drm_bridge *next;
drivers/gpu/drm/bridge/ti-tdp158.c
54
return drm_bridge_attach(encoder, tdp158->next, bridge, flags);
drivers/gpu/drm/bridge/ti-tdp158.c
76
tdp158->next = devm_drm_of_get_bridge(dev, dev->of_node, 1, 0);
drivers/gpu/drm/bridge/ti-tdp158.c
77
if (IS_ERR(tdp158->next))
drivers/gpu/drm/bridge/ti-tdp158.c
78
return dev_err_probe(dev, PTR_ERR(tdp158->next), "missing bridge");
drivers/gpu/drm/display/drm_dp_mst_topology.c
1317
list_del(&txmsg->next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
1710
list_for_each_entry(payload, &state->payloads, next)
drivers/gpu/drm/display/drm_dp_mst_topology.c
1848
list_add(&port->next, &mgr->destroy_port_list);
drivers/gpu/drm/display/drm_dp_mst_topology.c
1948
list_for_each_entry(port, &mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
1983
list_for_each_entry(port, &mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
2020
list_for_each_entry(port, &mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
2310
list_del(&port->next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
2412
list_add(&port->next, &mstb->ports);
drivers/gpu/drm/display/drm_dp_mst_topology.c
2559
list_for_each_entry(port, &mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
2594
list_for_each_entry(port, &mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
2640
list_for_each_entry(port, &mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
2865
struct drm_dp_sideband_msg_tx, next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
2869
list_del(&txmsg->next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
2879
list_add_tail(&txmsg->next, &mgr->tx_msg_downq);
drivers/gpu/drm/display/drm_dp_mst_topology.c
2977
list_for_each_entry_safe(port, tmp, &mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
2983
list_del(&port->next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
3412
list_for_each_entry(pos, &mst_state->payloads, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
3720
list_for_each_entry(port, &mstb->ports, next)
drivers/gpu/drm/display/drm_dp_mst_topology.c
3986
struct drm_dp_sideband_msg_tx, next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
4020
list_del(&txmsg->next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
4109
next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
4111
list_del(&up_req->next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
4144
INIT_LIST_HEAD(&up_req->next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
4192
list_add_tail(&up_req->next, &mgr->up_req_list);
drivers/gpu/drm/display/drm_dp_mst_topology.c
4279
struct drm_dp_sideband_msg_tx, next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
4490
list_add(&payload->next, &topology_state->payloads);
drivers/gpu/drm/display/drm_dp_mst_topology.c
4662
list_for_each_entry(old_payload, &old_mst_state->payloads, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
4855
list_for_each_entry(port, &mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
4940
list_for_each_entry(payload, &state->payloads, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
5041
list_for_each_entry_safe(port, port_tmp, &mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
5042
list_del(&port->next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
5049
list_for_each_entry_safe(txmsg, txmsg_tmp, &mgr->tx_msg_downq, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
5054
list_del(&txmsg->next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
5104
next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
5106
list_del(&port->next);
drivers/gpu/drm/display/drm_dp_mst_topology.c
5140
list_for_each_entry(pos, &old_state->payloads, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
5150
list_add(&payload->next, &state->payloads);
drivers/gpu/drm/display/drm_dp_mst_topology.c
5156
list_for_each_entry_safe(pos, payload, &state->payloads, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
5173
list_for_each_entry_safe(pos, tmp, &mst_state->payloads, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
5272
list_for_each_entry(payload, &state->payloads, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
5290
list_for_each_entry(port, &mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
5359
list_for_each_entry(payload, &mst_state->payloads, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
5431
list_for_each_entry(pos, &mst_state->payloads, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
6035
list_for_each_entry(downstream_port, &port->mstb->ports, next) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
63
struct list_head next;
drivers/gpu/drm/drm_atomic.c
851
drm_printf(p, "\tnext=%d\n", colorop->next ? colorop->next->base.id : 0);
drivers/gpu/drm/drm_bridge.c
830
struct drm_bridge *next, *limit;
drivers/gpu/drm/drm_bridge.c
841
next = list_next_entry(bridge, chain_node);
drivers/gpu/drm/drm_bridge.c
843
if (next->pre_enable_prev_first) {
drivers/gpu/drm/drm_bridge.c
847
limit = next;
drivers/gpu/drm/drm_bridge.c
852
list_for_each_entry_from(next, &encoder->bridge_chain,
drivers/gpu/drm/drm_bridge.c
854
if (!next->pre_enable_prev_first) {
drivers/gpu/drm/drm_bridge.c
855
next = list_prev_entry(next, chain_node);
drivers/gpu/drm/drm_bridge.c
856
limit = next;
drivers/gpu/drm/drm_bridge.c
860
if (list_is_last(&next->chain_node,
drivers/gpu/drm/drm_bridge.c
862
limit = next;
drivers/gpu/drm/drm_bridge.c
868
list_for_each_entry_from_reverse(next, &encoder->bridge_chain,
drivers/gpu/drm/drm_bridge.c
870
if (next == bridge)
drivers/gpu/drm/drm_bridge.c
873
drm_atomic_bridge_call_post_disable(next,
drivers/gpu/drm/drm_bridge.c
924
struct drm_bridge *iter, *next, *limit;
drivers/gpu/drm/drm_bridge.c
933
next = iter;
drivers/gpu/drm/drm_bridge.c
935
list_for_each_entry_from_reverse(next,
drivers/gpu/drm/drm_bridge.c
938
if (next == bridge)
drivers/gpu/drm/drm_bridge.c
941
if (!next->pre_enable_prev_first) {
drivers/gpu/drm/drm_bridge.c
945
limit = next;
drivers/gpu/drm/drm_bridge.c
950
list_for_each_entry_from(next, &encoder->bridge_chain, chain_node) {
drivers/gpu/drm/drm_bridge.c
954
if (next == iter)
drivers/gpu/drm/drm_bridge.c
960
drm_atomic_bridge_call_pre_enable(next, state);
drivers/gpu/drm/drm_colorop.c
111
colorop->next = NULL;
drivers/gpu/drm/drm_colorop.c
191
struct drm_colorop *colorop, *next;
drivers/gpu/drm/drm_colorop.c
193
list_for_each_entry_safe(colorop, next, &config->colorop_list, head) {
drivers/gpu/drm/drm_colorop.c
592
void drm_colorop_set_next_property(struct drm_colorop *colorop, struct drm_colorop *next)
drivers/gpu/drm/drm_colorop.c
596
next ? next->base.id : 0);
drivers/gpu/drm/drm_colorop.c
597
colorop->next = next;
drivers/gpu/drm/drm_connector.c
1096
if (lhead->next == &config->connector_list) {
drivers/gpu/drm/drm_connector.c
1101
lhead = lhead->next;
drivers/gpu/drm/drm_encoder.c
190
struct drm_bridge *bridge, *next;
drivers/gpu/drm/drm_encoder.c
197
list_for_each_entry_safe(bridge, next, &encoder->bridge_chain,
drivers/gpu/drm/drm_gpusvm.c
486
struct drm_gpusvm_notifier *notifier, *next;
drivers/gpu/drm/drm_gpusvm.c
488
drm_gpusvm_for_each_notifier_safe(notifier, next, gpusvm, 0, LONG_MAX) {
drivers/gpu/drm/drm_gpuvm.c
1489
struct drm_gpuvm_bo *vm_bo, *next;
drivers/gpu/drm/drm_gpuvm.c
1494
list_for_each_entry_safe(vm_bo, next, &gpuvm->evict.list,
drivers/gpu/drm/drm_gpuvm.c
1794
bo_defer = bo_defer->next;
drivers/gpu/drm/drm_gpuvm.c
2317
struct drm_gpuva *next,
drivers/gpu/drm/drm_gpuvm.c
2330
if (op->next) {
drivers/gpu/drm/drm_gpuvm.c
2331
drm_gpuva_init_from_op(next, op->next);
drivers/gpu/drm/drm_gpuvm.c
2332
drm_gpuva_insert(gpuvm, next);
drivers/gpu/drm/drm_gpuvm.c
2372
struct drm_gpuva_op_map *next,
drivers/gpu/drm/drm_gpuvm.c
2381
r->next = next;
drivers/gpu/drm/drm_gpuvm.c
2411
struct drm_gpuva *va, *next;
drivers/gpu/drm/drm_gpuvm.c
2421
drm_gpuvm_for_each_va_range_safe(va, next, gpuvm, req_addr, req_end) {
drivers/gpu/drm/drm_gpuvm.c
2587
struct drm_gpuva *va, *next;
drivers/gpu/drm/drm_gpuvm.c
2594
drm_gpuvm_for_each_va_range_safe(va, next, gpuvm, req_addr, req_end) {
drivers/gpu/drm/drm_gpuvm.c
2595
struct drm_gpuva_op_map prev = {}, next = {};
drivers/gpu/drm/drm_gpuvm.c
2613
next.va.addr = req_end;
drivers/gpu/drm/drm_gpuvm.c
2614
next.va.range = end - req_end;
drivers/gpu/drm/drm_gpuvm.c
2615
next.gem.obj = obj;
drivers/gpu/drm/drm_gpuvm.c
2616
next.gem.offset = offset + (req_end - addr);
drivers/gpu/drm/drm_gpuvm.c
2626
next_split ? &next : NULL,
drivers/gpu/drm/drm_gpuvm.c
2909
if (__r->next) {
drivers/gpu/drm/drm_gpuvm.c
2910
r->next = kmemdup(__r->next, sizeof(*r->next),
drivers/gpu/drm/drm_gpuvm.c
2912
if (unlikely(!r->next))
drivers/gpu/drm/drm_gpuvm.c
3225
struct drm_gpuva_op *op, *next;
drivers/gpu/drm/drm_gpuvm.c
3227
drm_gpuva_for_each_op_safe(op, next, ops) {
drivers/gpu/drm/drm_gpuvm.c
3232
kfree(op->remap.next);
drivers/gpu/drm/drm_pagemap.c
241
goto next;
drivers/gpu/drm/drm_pagemap.c
253
goto next;
drivers/gpu/drm/drm_pagemap.c
276
next:
drivers/gpu/drm/drm_pagemap.c
315
goto next;
drivers/gpu/drm/drm_pagemap.c
327
next:
drivers/gpu/drm/drm_pagemap.c
711
goto next;
drivers/gpu/drm/drm_pagemap.c
715
goto next;
drivers/gpu/drm/drm_pagemap.c
720
goto next;
drivers/gpu/drm/drm_pagemap.c
737
next:
drivers/gpu/drm/drm_pagemap.c
830
struct drm_pagemap_dev_hold *dev_hold, *next;
drivers/gpu/drm/drm_pagemap.c
837
llist_for_each_entry_safe(dev_hold, next, node, link) {
drivers/gpu/drm/drm_suballoc.c
133
if (sa_manager->hole->next == &sa_manager->olist)
drivers/gpu/drm/drm_suballoc.c
136
sa = list_entry(sa_manager->hole->next, struct drm_suballoc, olist);
drivers/gpu/drm/drm_suballoc.c
159
if (hole->next != &sa_manager->olist)
drivers/gpu/drm/drm_suballoc.c
160
return list_entry(hole->next, struct drm_suballoc, olist)->soffset;
drivers/gpu/drm/drm_suballoc.c
235
if (sa_manager->hole->next == &sa_manager->olist) {
drivers/gpu/drm/drm_syncobj.c
294
if (!wait->node.next)
drivers/gpu/drm/drm_syncobj.c
510
if (wait.node.next)
drivers/gpu/drm/drm_vblank_work.c
50
struct drm_vblank_work *work, *next;
drivers/gpu/drm/drm_vblank_work.c
56
list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
drivers/gpu/drm/drm_vblank_work.c
74
struct drm_vblank_work *work, *next;
drivers/gpu/drm/drm_vblank_work.c
81
list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
drivers/gpu/drm/gma500/mmu.c
487
unsigned long next;
drivers/gpu/drm/gma500/mmu.c
510
next = psb_pd_addr_end(addr, end);
drivers/gpu/drm/gma500/mmu.c
517
(addr & clflush_mask) < next);
drivers/gpu/drm/gma500/mmu.c
520
} while (addr = next, next != end);
drivers/gpu/drm/gma500/mmu.c
532
unsigned long next;
drivers/gpu/drm/gma500/mmu.c
541
next = psb_pd_addr_end(addr, end);
drivers/gpu/drm/gma500/mmu.c
548
} while (addr += PAGE_SIZE, addr < next);
drivers/gpu/drm/gma500/mmu.c
551
} while (addr = next, next != end);
drivers/gpu/drm/gma500/mmu.c
574
unsigned long next;
drivers/gpu/drm/gma500/mmu.c
597
next = psb_pd_addr_end(addr, end);
drivers/gpu/drm/gma500/mmu.c
605
} while (addr += PAGE_SIZE, addr < next);
drivers/gpu/drm/gma500/mmu.c
608
} while (addr = next, next != end);
drivers/gpu/drm/gma500/mmu.c
629
unsigned long next;
drivers/gpu/drm/gma500/mmu.c
639
next = psb_pd_addr_end(addr, end);
drivers/gpu/drm/gma500/mmu.c
649
} while (addr += PAGE_SIZE, addr < next);
drivers/gpu/drm/gma500/mmu.c
652
} while (addr = next, next != end);
drivers/gpu/drm/gma500/mmu.c
678
unsigned long next;
drivers/gpu/drm/gma500/mmu.c
703
next = psb_pd_addr_end(addr, end);
drivers/gpu/drm/gma500/mmu.c
712
} while (addr += PAGE_SIZE, addr < next);
drivers/gpu/drm/gma500/mmu.c
715
} while (addr = next, next != end);
drivers/gpu/drm/i915/display/intel_bios.c
358
static void next_lfp_data_ptr(struct lfp_data_ptr_table *next,
drivers/gpu/drm/i915/display/intel_bios.c
362
next->table_size = prev->table_size;
drivers/gpu/drm/i915/display/intel_bios.c
363
next->offset = prev->offset + size;
drivers/gpu/drm/i915/display/intel_dsi_vbt.c
660
const u8 *next = data + operation_size;
drivers/gpu/drm/i915/display/intel_dsi_vbt.c
665
if (operation_size && data != next) {
drivers/gpu/drm/i915/display/intel_global_state.c
129
struct intel_global_obj *obj, *next;
drivers/gpu/drm/i915/display/intel_global_state.c
131
list_for_each_entry_safe(obj, next, &display->global.obj_list, head) {
drivers/gpu/drm/i915/display/intel_panel.c
255
struct drm_display_mode *mode, *next;
drivers/gpu/drm/i915/display/intel_panel.c
257
list_for_each_entry_safe(mode, next, &connector->base.probed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
304
struct drm_display_mode *mode, *next;
drivers/gpu/drm/i915/display/intel_panel.c
306
list_for_each_entry_safe(mode, next, &connector->base.probed_modes, head) {
drivers/gpu/drm/i915/display/intel_panel.c
451
struct drm_display_mode *fixed_mode, *next;
drivers/gpu/drm/i915/display/intel_panel.c
460
list_for_each_entry_safe(fixed_mode, next, &panel->fixed_modes, head) {
drivers/gpu/drm/i915/display/intel_plane.c
388
iter_colorop = iter_colorop->next;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1433
struct i915_gem_engines *pos, *next;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1437
list_for_each_entry_safe(pos, next, &ctx->stale.engines, link) {
drivers/gpu/drm/i915/gem/i915_gem_context.c
1450
list_safe_reset_next(pos, next, link);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2015
capture->next = eb->capture_lists[j];
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
392
struct i915_vma *vma, *next;
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
407
list_for_each_entry_safe(vma, next,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
663
struct drm_mm_node *hole, *next;
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
677
list_for_each_entry_safe(hole, next, &mm->hole_stack, hole_stack) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
766
drm_mm_for_each_node_safe(hole, next, mm) {
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
168
node->next = head;
drivers/gpu/drm/i915/gt/intel_context.c
440
struct intel_context *child, *next;
drivers/gpu/drm/i915/gt/intel_context.c
448
for_each_child_safe(ce, child, next)
drivers/gpu/drm/i915/gt/intel_engine_pm.c
247
struct llist_node *node, *next;
drivers/gpu/drm/i915/gt/intel_engine_pm.c
249
llist_for_each_safe(node, next, llist_del_all(&engine->barrier_tasks)) {
drivers/gpu/drm/i915/gt/intel_engine_user.c
210
struct list_head *it, *next;
drivers/gpu/drm/i915/gt/intel_engine_user.c
218
list_for_each_safe(it, next, &engines) {
drivers/gpu/drm/i915/gt/intel_engine_user.c
86
struct llist_node *pos, *next;
drivers/gpu/drm/i915/gt/intel_engine_user.c
88
llist_for_each_safe(pos, next, get_engines(i915)) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
357
const struct i915_request *next)
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
369
return rq_prio(prev) >= rq_prio(next);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
948
const struct intel_context *next)
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
950
if (prev != next)
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
965
const struct i915_request *next)
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
967
GEM_BUG_ON(prev == next);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
968
GEM_BUG_ON(!assert_priority_queue(prev, next));
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
978
if (__i915_request_is_complete(next))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
981
if (unlikely((i915_request_flags(prev) | i915_request_flags(next)) &
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
986
if (!can_merge_ctx(prev->context, next->context))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
989
GEM_BUG_ON(i915_seqno_passed(prev->fence.seqno, next->fence.seqno));
drivers/gpu/drm/i915/gt/intel_gt_requests.c
67
struct intel_timeline *next = xchg(&tl->retire, NULL);
drivers/gpu/drm/i915/gt/intel_gt_requests.c
83
GEM_BUG_ON(!next);
drivers/gpu/drm/i915/gt/intel_gt_requests.c
84
tl = ptr_mask_bits(next, 1);
drivers/gpu/drm/i915/gt/intel_ring.h
60
u32 next, u32 prev)
drivers/gpu/drm/i915/gt/intel_ring.h
62
typecheck(typeof(ring->size), next);
drivers/gpu/drm/i915/gt/intel_ring.h
64
return (next - prev) << ring->wrap;
drivers/gpu/drm/i915/gt/selftest_execlists.c
2785
rq->mock.link.next = &(*prev)->mock.link;
drivers/gpu/drm/i915/gt/selftest_ring.c
32
u32 next, u32 prev,
drivers/gpu/drm/i915/gt/selftest_ring.c
37
result = intel_ring_direction(ring, next, prev);
drivers/gpu/drm/i915/gt/selftest_ring.c
45
next, prev, result, expected);
drivers/gpu/drm/i915/gt/selftest_ring.c
54
u32 prev = x, next = intel_ring_wrap(ring, x + step);
drivers/gpu/drm/i915/gt/selftest_ring.c
57
err |= check_ring_direction(ring, next, next, 0);
drivers/gpu/drm/i915/gt/selftest_ring.c
59
err |= check_ring_direction(ring, next, prev, 1);
drivers/gpu/drm/i915/gt/selftest_ring.c
60
err |= check_ring_direction(ring, prev, next, -1);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5410
goto next;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5414
goto next;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5417
goto next;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5440
next:
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5469
goto next;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5473
goto next;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5476
goto next;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5484
next:
drivers/gpu/drm/i915/gvt/cfg_space.c
328
u8 next;
drivers/gpu/drm/i915/gvt/cfg_space.c
368
next = vgpu_cfg_space(vgpu)[PCI_CAPABILITY_LIST];
drivers/gpu/drm/i915/gvt/cfg_space.c
370
if (vgpu_cfg_space(vgpu)[next + PCI_CAP_LIST_ID] == PCI_CAP_ID_PM) {
drivers/gpu/drm/i915/gvt/cfg_space.c
371
vgpu->cfg_space.pmcsr_off = next + PCI_PM_CTRL;
drivers/gpu/drm/i915/gvt/cfg_space.c
374
next = vgpu_cfg_space(vgpu)[next + PCI_CAP_LIST_NEXT];
drivers/gpu/drm/i915/gvt/cfg_space.c
375
} while (next);
drivers/gpu/drm/i915/gvt/debugfs.c
115
list_for_each_entry_safe(node, next, &param.diff_mmio_list, node) {
drivers/gpu/drm/i915/gvt/debugfs.c
95
struct diff_mmio *node, *next;
drivers/gpu/drm/i915/gvt/execlist.c
399
struct list_head *next = workload_q_head(vgpu, workload->engine)->next;
drivers/gpu/drm/i915/gvt/execlist.c
412
next_workload = container_of(next,
drivers/gpu/drm/i915/gvt/gtt.c
1450
oos_page = container_of(gtt->oos_page_use_list_head.next,
drivers/gpu/drm/i915/gvt/gtt.c
1459
oos_page = container_of(gtt->oos_page_free_list_head.next,
drivers/gpu/drm/i915/gvt/gtt.c
2447
struct intel_gvt_partial_pte *pos, *next;
drivers/gpu/drm/i915/gvt/gtt.c
2449
list_for_each_entry_safe(pos, next,
drivers/gpu/drm/i915/gvt/handlers.c
903
u32 next = i915_mmio_reg_offset(_next);
drivers/gpu/drm/i915/gvt/handlers.c
905
u32 stride = next - start;
drivers/gpu/drm/i915/gvt/mmio_context.c
411
static void switch_mocs(struct intel_vgpu *pre, struct intel_vgpu *next,
drivers/gpu/drm/i915/gvt/mmio_context.c
441
if (next)
drivers/gpu/drm/i915/gvt/mmio_context.c
442
new_v = vgpu_vreg_t(next, offset);
drivers/gpu/drm/i915/gvt/mmio_context.c
459
if (next)
drivers/gpu/drm/i915/gvt/mmio_context.c
460
new_v = vgpu_vreg_t(next, l3_offset);
drivers/gpu/drm/i915/gvt/mmio_context.c
486
struct intel_vgpu *next,
drivers/gpu/drm/i915/gvt/mmio_context.c
495
switch_mocs(pre, next, engine);
drivers/gpu/drm/i915/gvt/mmio_context.c
523
if (next) {
drivers/gpu/drm/i915/gvt/mmio_context.c
524
s = &next->submission;
drivers/gpu/drm/i915/gvt/mmio_context.c
535
new_v = vgpu_vreg_t(next, mmio->reg) |
drivers/gpu/drm/i915/gvt/mmio_context.c
538
new_v = vgpu_vreg_t(next, mmio->reg);
drivers/gpu/drm/i915/gvt/mmio_context.c
551
next ? next->id : 0,
drivers/gpu/drm/i915/gvt/mmio_context.c
557
if (next)
drivers/gpu/drm/i915/gvt/mmio_context.c
558
handle_tlb_pending_event(next, engine);
drivers/gpu/drm/i915/gvt/mmio_context.c
571
struct intel_vgpu *next,
drivers/gpu/drm/i915/gvt/mmio_context.c
574
if (WARN(!pre && !next, "switch ring %s from host to HOST\n",
drivers/gpu/drm/i915/gvt/mmio_context.c
579
pre ? "vGPU" : "host", next ? "vGPU" : "HOST");
drivers/gpu/drm/i915/gvt/mmio_context.c
587
switch_mmio(pre, next, engine);
drivers/gpu/drm/i915/gvt/mmio_context.h
48
struct intel_vgpu *next,
drivers/gpu/drm/i915/i915_active.c
370
struct llist_node *pos, *next;
drivers/gpu/drm/i915/i915_active.c
389
llist_for_each_safe(pos, next, llist_del_all(&engine->barrier_tasks)) {
drivers/gpu/drm/i915/i915_active.c
395
pos->next = head;
drivers/gpu/drm/i915/i915_active.c
895
first->next = prev;
drivers/gpu/drm/i915/i915_active.c
910
first = first->next;
drivers/gpu/drm/i915/i915_active.c
922
struct llist_node *pos, *next;
drivers/gpu/drm/i915/i915_active.c
933
llist_for_each_safe(pos, next, take_preallocated_barriers(ref)) {
drivers/gpu/drm/i915/i915_active.c
971
struct llist_node *node, *next;
drivers/gpu/drm/i915/i915_active.c
987
llist_for_each_safe(node, next, node) {
drivers/gpu/drm/i915/i915_gem_evict.c
157
struct i915_vma *vma, *next;
drivers/gpu/drm/i915/i915_gem_evict.c
199
list_for_each_entry_safe(vma, next, &vm->bound_list, vm_link) {
drivers/gpu/drm/i915/i915_gem_evict.c
235
list_for_each_entry_safe(vma, next, &eviction_list, evict_link) {
drivers/gpu/drm/i915/i915_gem_evict.c
282
list_for_each_entry_safe(vma, next, &eviction_list, evict_link) {
drivers/gpu/drm/i915/i915_gem_evict.c
293
list_for_each_entry_safe(vma, next, &eviction_list, evict_link) {
drivers/gpu/drm/i915/i915_gem_evict.c
337
struct i915_vma *vma, *next;
drivers/gpu/drm/i915/i915_gem_evict.c
428
list_for_each_entry_safe(vma, next, &eviction_list, evict_link) {
drivers/gpu/drm/i915/i915_gpu_error.c
1071
struct i915_vma_coredump *next = vma->next;
drivers/gpu/drm/i915/i915_gpu_error.c
1080
vma = next;
drivers/gpu/drm/i915/i915_gpu_error.c
1107
gt->engine = ee->next;
drivers/gpu/drm/i915/i915_gpu_error.c
1128
error->gt = gt->next;
drivers/gpu/drm/i915/i915_gpu_error.c
1169
dst->next = NULL;
drivers/gpu/drm/i915/i915_gpu_error.c
1489
struct intel_engine_capture_vma *next;
drivers/gpu/drm/i915/i915_gpu_error.c
1496
capture_vma_snapshot(struct intel_engine_capture_vma *next,
drivers/gpu/drm/i915/i915_gpu_error.c
1503
return next;
drivers/gpu/drm/i915/i915_gpu_error.c
1507
return next;
drivers/gpu/drm/i915/i915_gpu_error.c
1511
return next;
drivers/gpu/drm/i915/i915_gpu_error.c
1517
c->next = next;
drivers/gpu/drm/i915/i915_gpu_error.c
1522
capture_vma(struct intel_engine_capture_vma *next,
drivers/gpu/drm/i915/i915_gpu_error.c
1528
return next;
drivers/gpu/drm/i915/i915_gpu_error.c
1536
return next;
drivers/gpu/drm/i915/i915_gpu_error.c
1538
next = capture_vma_snapshot(next, vma->resource, gfp, name);
drivers/gpu/drm/i915/i915_gpu_error.c
1540
return next;
drivers/gpu/drm/i915/i915_gpu_error.c
1550
for (c = rq->capture_list; c; c = c->next)
drivers/gpu/drm/i915/i915_gpu_error.c
1561
vma->next = ee->vma;
drivers/gpu/drm/i915/i915_gpu_error.c
1681
capture = this->next;
drivers/gpu/drm/i915/i915_gpu_error.c
1772
ee->next = gt->engine;
drivers/gpu/drm/i915/i915_gpu_error.c
2035
for (gt = error->gt; gt; gt = gt->next) {
drivers/gpu/drm/i915/i915_gpu_error.c
2038
for (cs = gt->engine; cs; cs = cs->next) {
drivers/gpu/drm/i915/i915_gpu_error.c
2327
char *buf, *ptr, *next;
drivers/gpu/drm/i915/i915_gpu_error.c
2383
next = strnchr(ptr, got, '\n');
drivers/gpu/drm/i915/i915_gpu_error.c
2384
if (next) {
drivers/gpu/drm/i915/i915_gpu_error.c
2385
count = next - ptr;
drivers/gpu/drm/i915/i915_gpu_error.c
2386
*next = 0;
drivers/gpu/drm/i915/i915_gpu_error.c
2427
ptr = next;
drivers/gpu/drm/i915/i915_gpu_error.c
2429
if (next) {
drivers/gpu/drm/i915/i915_gpu_error.c
518
vma = vma->next;
drivers/gpu/drm/i915/i915_gpu_error.c
871
for (ee = gt->engine; ee; ee = ee->next) {
drivers/gpu/drm/i915/i915_gpu_error.c
888
for (vma = ee->vma; vma; vma = vma->next)
drivers/gpu/drm/i915/i915_gpu_error.c
918
for (ee = error->gt ? error->gt->engine : NULL; ee; ee = ee->next)
drivers/gpu/drm/i915/i915_gpu_error.h
127
struct intel_engine_coredump *next;
drivers/gpu/drm/i915/i915_gpu_error.h
187
struct intel_gt_coredump *next;
drivers/gpu/drm/i915/i915_gpu_error.h
35
struct i915_vma_coredump *next;
drivers/gpu/drm/i915/i915_list_util.h
14
WRITE_ONCE(head->next, first);
drivers/gpu/drm/i915/i915_list_util.h
20
return READ_ONCE(list->next) == head;
drivers/gpu/drm/i915/i915_request.c
1105
if (unlikely(READ_ONCE(prev->link.next) != &signal->link)) {
drivers/gpu/drm/i915/i915_request.c
336
struct i915_capture_list *next = capture->next;
drivers/gpu/drm/i915/i915_request.c
340
capture = next;
drivers/gpu/drm/i915/i915_request.h
57
struct i915_capture_list *next;
drivers/gpu/drm/i915/i915_scheduler.c
227
if (stack.dfs_link.next == stack.dfs_link.prev)
drivers/gpu/drm/i915/i915_sw_fence.c
145
wait_queue_entry_t *pos, *next;
drivers/gpu/drm/i915/i915_sw_fence.c
160
list_for_each_entry_safe(pos, next, &x->head, entry) {
drivers/gpu/drm/i915/i915_sw_fence.c
170
list_for_each_entry_safe(pos, next, &x->head, entry) {
drivers/gpu/drm/i915/i915_syncmap.c
276
struct i915_syncmap *next;
drivers/gpu/drm/i915/i915_syncmap.c
282
next = kzalloc_flex(*next, child, KSYNCMAP);
drivers/gpu/drm/i915/i915_syncmap.c
283
if (unlikely(!next))
drivers/gpu/drm/i915/i915_syncmap.c
289
next->height = above + p->height;
drivers/gpu/drm/i915/i915_syncmap.c
290
next->prefix = __sync_branch_prefix(next, id);
drivers/gpu/drm/i915/i915_syncmap.c
295
__sync_child(p->parent)[idx] = next;
drivers/gpu/drm/i915/i915_syncmap.c
298
next->parent = p->parent;
drivers/gpu/drm/i915/i915_syncmap.c
302
__sync_set_child(next, idx, p);
drivers/gpu/drm/i915/i915_syncmap.c
303
p->parent = next;
drivers/gpu/drm/i915/i915_syncmap.c
306
p = next;
drivers/gpu/drm/i915/i915_syncmap.c
315
next = __sync_child(p)[idx];
drivers/gpu/drm/i915/i915_syncmap.c
316
if (!next) {
drivers/gpu/drm/i915/i915_syncmap.c
317
next = __sync_alloc_leaf(p, id);
drivers/gpu/drm/i915/i915_syncmap.c
318
if (unlikely(!next))
drivers/gpu/drm/i915/i915_syncmap.c
321
__sync_set_child(p, idx, next);
drivers/gpu/drm/i915/i915_syncmap.c
322
p = next;
drivers/gpu/drm/i915/i915_syncmap.c
326
p = next;
drivers/gpu/drm/i915/i915_user_extensions.c
26
u64 next;
drivers/gpu/drm/i915/i915_user_extensions.c
53
if (get_user(next, &ext->next_extension) ||
drivers/gpu/drm/i915/i915_user_extensions.c
54
overflows_type(next, uintptr_t))
drivers/gpu/drm/i915/i915_user_extensions.c
57
ext = u64_to_user_ptr(next);
drivers/gpu/drm/i915/i915_vma.c
1872
struct i915_vma *vma, *next;
drivers/gpu/drm/i915/i915_vma.c
1876
list_for_each_entry_safe(vma, next, &gt->closed_vma, closed_link) {
drivers/gpu/drm/i915/i915_vma.c
1895
list_for_each_entry_safe(vma, next, &closed, closed_link) {
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
385
struct reserved *next;
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
439
r->next = reserved;
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
522
struct reserved *next = reserved->next;
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
527
reserved = next;
drivers/gpu/drm/imagination/pvr_fw_meta.c
324
if (l1_data->next == 0xFFFFFFFF)
drivers/gpu/drm/imagination/pvr_fw_meta.c
328
l1_data->next);
drivers/gpu/drm/imagination/pvr_fw_trace.c
445
.next = fw_trace_seq_next,
drivers/gpu/drm/imagination/pvr_rogue_meta.h
151
u32 next;
drivers/gpu/drm/imagination/pvr_vm.c
441
if (op->remap.next) {
drivers/gpu/drm/ingenic/ingenic-drm-drv.c
1059
desc->next = dma_hwdesc_addr(priv, next_hwdesc);
drivers/gpu/drm/ingenic/ingenic-drm-drv.c
50
u32 next;
drivers/gpu/drm/ingenic/ingenic-drm-drv.c
693
hwdesc->next = dma_hwdesc_addr(priv, next_id);
drivers/gpu/drm/logicvc/logicvc_layer.c
587
struct logicvc_layer *next;
drivers/gpu/drm/logicvc/logicvc_layer.c
626
list_for_each_entry_safe(layer, next, &logicvc->layers_list, list)
drivers/gpu/drm/mediatek/mtk_ddp_comp.h
325
unsigned int next)
drivers/gpu/drm/mediatek/mtk_ddp_comp.h
328
comp->funcs->connect(comp->dev, mmsys_dev, next);
drivers/gpu/drm/mediatek/mtk_ddp_comp.h
335
unsigned int next)
drivers/gpu/drm/mediatek/mtk_ddp_comp.h
338
comp->funcs->disconnect(comp->dev, mmsys_dev, next);
drivers/gpu/drm/mediatek/mtk_ddp_comp.h
87
void (*connect)(struct device *dev, struct device *mmsys_dev, unsigned int next);
drivers/gpu/drm/mediatek/mtk_ddp_comp.h
88
void (*disconnect)(struct device *dev, struct device *mmsys_dev, unsigned int next);
drivers/gpu/drm/mediatek/mtk_disp_drv.h
115
unsigned int next);
drivers/gpu/drm/mediatek/mtk_disp_drv.h
117
unsigned int next);
drivers/gpu/drm/mediatek/mtk_disp_ovl_adaptor.c
448
void mtk_ovl_adaptor_connect(struct device *dev, struct device *mmsys_dev, unsigned int next)
drivers/gpu/drm/mediatek/mtk_disp_ovl_adaptor.c
450
mtk_mmsys_ddp_connect(mmsys_dev, DDP_COMPONENT_ETHDR_MIXER, next);
drivers/gpu/drm/mediatek/mtk_disp_ovl_adaptor.c
460
void mtk_ovl_adaptor_disconnect(struct device *dev, struct device *mmsys_dev, unsigned int next)
drivers/gpu/drm/mediatek/mtk_disp_ovl_adaptor.c
462
mtk_mmsys_ddp_disconnect(mmsys_dev, DDP_COMPONENT_ETHDR_MIXER, next);
drivers/gpu/drm/mediatek/mtk_drm_drv.c
856
struct device_node **next, unsigned int *cid)
drivers/gpu/drm/mediatek/mtk_drm_drv.c
876
*next = ep_dev_node;
drivers/gpu/drm/mediatek/mtk_drm_drv.c
922
struct device_node *next = NULL, *prev, *vdo = dev->parent->of_node;
drivers/gpu/drm/mediatek/mtk_drm_drv.c
930
ret = mtk_drm_of_get_ddp_ep_cid(vdo, 0, cpath, &next, &temp_path[idx]);
drivers/gpu/drm/mediatek/mtk_drm_drv.c
932
if (next && temp_path[idx] == DDP_COMPONENT_DRM_OVL_ADAPTOR) {
drivers/gpu/drm/mediatek/mtk_drm_drv.c
933
dev_dbg(dev, "Adding OVL Adaptor for %pOF\n", next);
drivers/gpu/drm/mediatek/mtk_drm_drv.c
936
if (next)
drivers/gpu/drm/mediatek/mtk_drm_drv.c
937
dev_err(dev, "Invalid component %pOF\n", next);
drivers/gpu/drm/mediatek/mtk_drm_drv.c
951
prev = next;
drivers/gpu/drm/mediatek/mtk_drm_drv.c
952
ret = mtk_drm_of_get_ddp_ep_cid(next, 1, cpath, &next, &temp_path[idx]);
drivers/gpu/drm/mediatek/mtk_drm_drv.c
955
of_node_put(next);
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
51
ring->cur = ring->next;
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
172
ring->cur = ring->next;
drivers/gpu/drm/msm/adreno/a8xx_gpu.c
168
ring->cur = ring->next;
drivers/gpu/drm/msm/adreno/adreno_gpu.c
1077
uint32_t wptr = ring->next - ring->start;
drivers/gpu/drm/msm/adreno/adreno_gpu.c
668
ring->next = ring->start;
drivers/gpu/drm/msm/adreno/adreno_gpu.c
721
ring->cur = ring->next;
drivers/gpu/drm/msm/msm_gem_vma.c
1154
ops_are_same_pte(struct msm_vm_bind_op *first, struct msm_vm_bind_op *next)
drivers/gpu/drm/msm/msm_gem_vma.c
1162
return ((first->iova + first->range) & pte_mask) == (next->iova & pte_mask);
drivers/gpu/drm/msm/msm_gem_vma.c
607
if (op->remap.next) {
drivers/gpu/drm/msm/msm_gem_vma.c
608
next_vma = vma_from_op(arg, op->remap.next);
drivers/gpu/drm/msm/msm_ringbuffer.c
105
ring->next = ring->start;
drivers/gpu/drm/msm/msm_ringbuffer.h
135
if (ring->next == ring->end)
drivers/gpu/drm/msm/msm_ringbuffer.h
136
ring->next = ring->start;
drivers/gpu/drm/msm/msm_ringbuffer.h
137
*(ring->next++) = data;
drivers/gpu/drm/msm/msm_ringbuffer.h
51
uint32_t *start, *end, *cur, *next;
drivers/gpu/drm/nouveau/include/nvkm/core/mm.h
58
return !node->next;
drivers/gpu/drm/nouveau/include/nvkm/core/mm.h
75
} while ((node = node->next));
drivers/gpu/drm/nouveau/include/nvkm/core/mm.h
9
struct nvkm_mm_node *next;
drivers/gpu/drm/nouveau/include/nvkm/subdev/fb.h
158
struct nvkm_ram_data *next;
drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
104
u64 next;
drivers/gpu/drm/nouveau/nouveau_gem.c
419
nvbo = list_entry(op->list.next, struct nouveau_bo, entry);
drivers/gpu/drm/nouveau/nouveau_svm.c
177
unsigned long next;
drivers/gpu/drm/nouveau/nouveau_svm.c
184
next = min(vma->vm_end, end);
drivers/gpu/drm/nouveau/nouveau_svm.c
187
next);
drivers/gpu/drm/nouveau/nouveau_svm.c
188
addr = next;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1037
struct bind_job_op *op, *next;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1039
list_for_each_op_safe(op, next, &job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1205
if (op->remap.next)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1206
drm_gpuva_link(&new->next->va, va->vm_bo);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1651
struct bind_job_op *op, *next;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1653
list_for_each_op_safe(op, next, ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1967
struct drm_gpuva *va, *next;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1970
drm_gpuvm_for_each_va_safe(va, next, &uvmm->base) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
534
if (r->next)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
535
op_map_prepare_unwind(new->next);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
56
struct nouveau_uvma *next;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
578
if (r->next)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
581
if (r->prev && r->next)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
733
if (r->next) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
734
ret = op_map_prepare(uvmm, &new->next, r->next,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
746
if (args && (r->prev && r->next))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
887
if (r->next)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
888
end = r->next->va.addr;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
949
struct drm_gpuva_op_map *n = r->next;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
114
struct nvkm_mm_node *prev, *this, *next;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
133
next = node(this, next);
drivers/gpu/drm/nouveau/nvkm/core/mm.c
134
if (next && next->type != type)
drivers/gpu/drm/nouveau/nvkm/core/mm.c
150
this->next = NULL;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
189
struct nvkm_mm_node *prev, *this, *next;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
207
next = node(this, next);
drivers/gpu/drm/nouveau/nvkm/core/mm.c
208
if (next && next->type != type) {
drivers/gpu/drm/nouveau/nvkm/core/mm.c
210
c = next->offset - e;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
229
this->next = NULL;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
243
u32 next;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
247
next = prev->offset + prev->length;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
248
if (next != offset) {
drivers/gpu/drm/nouveau/nvkm/core/mm.c
249
BUG_ON(next > offset);
drivers/gpu/drm/nouveau/nvkm/core/mm.c
253
node->offset = next;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
254
node->length = offset - next;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
54
struct nvkm_mm_node *next = node(this, next);
drivers/gpu/drm/nouveau/nvkm/core/mm.c
62
if (next && next->type == NVKM_MM_TYPE_NONE) {
drivers/gpu/drm/nouveau/nvkm/core/mm.c
63
next->offset = this->offset;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
64
next->length += this->length;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
460
u32 next = nvkm_rd32(device, list->data[i].addr + base + 0);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
466
if (prev != next)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
467
snprintf(mods, sizeof(mods), "-> %08x", next);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
200
} prev, next, *chan;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
210
status->next.tsg = !!(stat & 0x10000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
211
status->next.id = (stat & 0x0fff0000) >> 16;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
222
status->chan = &status->next;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
227
status->chan = &status->next;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
240
status->next.tsg ? "tsg" : "ch", status->next.id,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
241
status->chan == &status->next ? "*" : " ");
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1086
u32 next = init->addr + init->count * init->pitch;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1088
while (addr < next) {
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1106
u32 next = init->addr + init->count * init->pitch;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1116
while (addr < next) {
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1145
u32 next = init->addr + init->count * init->pitch;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1153
while (addr < next) {
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1118
struct nv04_gr_chan *next = NULL;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1130
next = gr->chan[chid];
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1131
if (next)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1132
nv04_gr_load_context(next, chid);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
936
struct nv10_gr_chan *next = NULL;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
948
next = gr->chan[chid];
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
949
if (next)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
950
nv10_gr_load_context(next, chid);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
75
switch (ram->next->bios.timing_ver) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
77
CWL = ram->next->bios.timing_10_CWL;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
78
CL = ram->next->bios.timing_10_CL;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
79
WR = ram->next->bios.timing_10_WR;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
80
DLL = !ram->next->bios.ramcfg_DLLoff;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
81
ODT = ram->next->bios.timing_10_ODT;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
82
RON = ram->next->bios.ramcfg_RON;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
85
CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
86
CL = (ram->next->bios.timing[1] & 0x0000001f) >> 0;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
87
WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
96
if (ram->next->bios.timing_ver == 0x20 ||
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr3.c
97
ram->next->bios.ramcfg_timing == 0xff) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
41
xd = !ram->next->bios.ramcfg_DLLoff;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
43
switch (ram->next->bios.ramcfg_ver) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
45
pd = ram->next->bios.ramcfg_11_01_80;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
46
lf = ram->next->bios.ramcfg_11_01_40;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
47
vh = ram->next->bios.ramcfg_11_02_10;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
48
vr = ram->next->bios.ramcfg_11_02_04;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
49
vo = ram->next->bios.ramcfg_11_06;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
50
l3 = !ram->next->bios.ramcfg_11_07_02;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
56
switch (ram->next->bios.timing_ver) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
58
WL = (ram->next->bios.timing[1] & 0x00000f80) >> 7;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
59
CL = (ram->next->bios.timing[1] & 0x0000001f);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
60
WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
61
at[0] = ram->next->bios.timing_20_2e_c0;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
62
at[1] = ram->next->bios.timing_20_2e_30;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
63
dt = ram->next->bios.timing_20_2e_03;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c
64
ds = ram->next->bios.timing_20_2f_03;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
187
node = &r->next;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
89
struct nvkm_mm_node *next = vram->mn;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
92
if (next) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
93
if (likely(next->nl_entry.next)){
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
95
while ((node = next)) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
96
next = node->next;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1041
gk104_ram_calc_xits(struct gk104_ram *ram, struct nvkm_ram_data *next)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1052
ram->mode = (next->freq > fuc->refpll.vco1.max_freq) ? 2 : 1;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1063
refclk = next->freq;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1065
ret = gk104_pll_calc_hiclk(next->freq, subdev->device->crystal,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1074
" (refclock: %i kHz)\n", next->freq, ret);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1090
ram->base.freq = next->freq;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1096
ret = gk104_ram_calc_sddr3(ram, next->freq);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1101
ret = gk104_ram_calc_gddr5(ram, next->freq);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1120
if (ram->base.next == NULL) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1143
ram->base.next = &ram->base.target;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1145
ram->base.next = &ram->base.xition;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1147
BUG_ON(ram->base.next != &ram->base.xition);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1148
ram->base.next = &ram->base.target;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1151
return gk104_ram_calc_xits(ram, ram->base.next);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1236
struct nvkm_ram_data *next = ram->base.next;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1240
return (ram->base.next == &ram->base.xition);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1245
gk104_ram_prog_0(ram, next->freq);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1247
return (ram->base.next == &ram->base.xition);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1254
ram->base.next = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
210
struct nvkm_ram_data *next = ram->base.next;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
211
u8 v0 = next->bios.ramcfg_11_03_c0;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
212
u8 v1 = next->bios.ramcfg_11_03_30;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
243
u32 next = (prev & ~mask) | data;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
244
nvkm_memx_wr32(fuc->memx, addr, next);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
255
struct nvkm_ram_data *next = ram->base.next;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
256
int vc = !next->bios.ramcfg_11_02_08;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
257
int mv = !next->bios.ramcfg_11_02_04;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
311
switch (next->bios.ramcfg_11_03_c0) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
318
switch (next->bios.ramcfg_11_03_30) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
326
if (next->bios.ramcfg_11_02_80)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
328
if (next->bios.ramcfg_11_02_40)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
330
if (next->bios.ramcfg_11_07_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
332
if (next->bios.ramcfg_11_07_08)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
367
if (next->bios.ramcfg_11_02_40 ||
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
368
next->bios.ramcfg_11_07_10) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
391
if (next->bios.ramcfg_11_07_40)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
395
ram_wr32(fuc, 0x10f65c, 0x00000011 * next->bios.rammap_11_11_0c);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
396
ram_wr32(fuc, 0x10f6b8, 0x01010101 * next->bios.ramcfg_11_09);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
397
ram_wr32(fuc, 0x10f6bc, 0x01010101 * next->bios.ramcfg_11_09);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
399
if (!next->bios.ramcfg_11_07_08 && !next->bios.ramcfg_11_07_04) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
400
ram_wr32(fuc, 0x10f698, 0x01010101 * next->bios.ramcfg_11_04);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
401
ram_wr32(fuc, 0x10f69c, 0x01010101 * next->bios.ramcfg_11_04);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
403
if (!next->bios.ramcfg_11_07_08) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
409
u32 data = 0x01000100 * next->bios.ramcfg_11_04;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
414
if (ram->mode == 2 && next->bios.ramcfg_11_08_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
422
if (!next->bios.ramcfg_11_02_80)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
424
if (!next->bios.ramcfg_11_02_40)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
426
if (!next->bios.ramcfg_11_07_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
428
if (!next->bios.ramcfg_11_07_08)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
434
if (next->bios.ramcfg_11_01_08)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
445
if (next->bios.ramcfg_11_08_01)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
452
if (next->bios.ramcfg_11_08_08)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
454
if (next->bios.ramcfg_11_08_04)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
456
if (next->bios.ramcfg_11_08_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
461
ram_mask(fuc, 0x10f248, 0xffffffff, next->bios.timing[10]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
462
ram_mask(fuc, 0x10f290, 0xffffffff, next->bios.timing[0]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
463
ram_mask(fuc, 0x10f294, 0xffffffff, next->bios.timing[1]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
464
ram_mask(fuc, 0x10f298, 0xffffffff, next->bios.timing[2]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
465
ram_mask(fuc, 0x10f29c, 0xffffffff, next->bios.timing[3]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
466
ram_mask(fuc, 0x10f2a0, 0xffffffff, next->bios.timing[4]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
467
ram_mask(fuc, 0x10f2a4, 0xffffffff, next->bios.timing[5]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
468
ram_mask(fuc, 0x10f2a8, 0xffffffff, next->bios.timing[6]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
469
ram_mask(fuc, 0x10f2ac, 0xffffffff, next->bios.timing[7]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
470
ram_mask(fuc, 0x10f2cc, 0xffffffff, next->bios.timing[8]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
471
ram_mask(fuc, 0x10f2e8, 0xffffffff, next->bios.timing[9]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
475
if (next->bios.ramcfg_11_08_20)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
483
data |= next->bios.ramcfg_11_02_03 << 8;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
487
if (next->bios.ramcfg_11_01_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
495
data |= next->bios.timing_20_30_07 << 28;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
499
if (next->bios.ramcfg_11_01_01)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
507
data |= next->bios.timing_20_30_07 << 28;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
511
if (next->bios.ramcfg_11_01_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
519
if (!next->bios.ramcfg_11_01_04)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
521
if (!next->bios.ramcfg_11_07_80)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
526
if (next->bios.ramcfg_11_03_f0) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
527
if (next->bios.rammap_11_08_0c) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
528
if (!next->bios.ramcfg_11_07_80)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
541
ram_wr32(fuc, 0x10f870, 0x11111111 * next->bios.ramcfg_11_03_0f);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
545
data |= next->bios.ramcfg_11_02_03;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
549
if (next->bios.ramcfg_11_01_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
560
data = next->bios.timing_20_30_07 << 8;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
561
if (next->bios.ramcfg_11_01_01)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
565
ram_mask(fuc, 0x10f250, 0x000003f0, next->bios.timing_20_2c_003f << 4);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
566
data = (next->bios.timing[10] & 0x7f000000) >> 24;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
567
if (data < next->bios.timing_20_2c_1fc0)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
568
data = next->bios.timing_20_2c_1fc0;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
570
ram_mask(fuc, 0x10f224, 0x001f0000, next->bios.timing_20_30_f8 << 16);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
572
ram_mask(fuc, 0x10fec4, 0x041e0f07, next->bios.timing_20_31_0800 << 26 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
573
next->bios.timing_20_31_0780 << 17 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
574
next->bios.timing_20_31_0078 << 8 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
575
next->bios.timing_20_31_0007);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
576
ram_mask(fuc, 0x10fec8, 0x00000027, next->bios.timing_20_31_8000 << 5 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
577
next->bios.timing_20_31_7000);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
585
if (next->bios.ramcfg_11_08_10 && (ram->mode == 2) /*XXX*/) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
618
if (!next->bios.ramcfg_11_07_08) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
619
if (!next->bios.ramcfg_11_07_04)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
633
if (!next->bios.ramcfg_11_07_08) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
635
if ( next->bios.ramcfg_11_07_04)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
637
if (!next->bios.rammap_11_08_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
659
if (next->bios.ramcfg_11_07_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
667
if (next->bios.rammap_11_08_01)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
706
struct nvkm_ram_data *next = ram->base.next;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
707
int vc = !next->bios.ramcfg_11_02_08;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
708
int mv = !next->bios.ramcfg_11_02_04;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
726
if (next->bios.ramcfg_11_03_f0)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
731
if (next->bios.ramcfg_DLLoff)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
755
switch (next->bios.ramcfg_11_03_c0) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
762
switch (next->bios.ramcfg_11_03_30) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
770
if (next->bios.ramcfg_11_02_80)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
772
if (next->bios.ramcfg_11_02_40)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
774
if (next->bios.ramcfg_11_07_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
776
if (next->bios.ramcfg_11_07_08)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
786
data |= next->bios.ramcfg_11_03_30 << 16;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
818
if (next->bios.ramcfg_11_02_40 ||
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
819
next->bios.ramcfg_11_07_10) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
825
if (next->bios.ramcfg_11_07_40)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
829
ram_wr32(fuc, 0x10f65c, 0x00000011 * next->bios.rammap_11_11_0c);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
830
ram_wr32(fuc, 0x10f6b8, 0x01010101 * next->bios.ramcfg_11_09);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
831
ram_wr32(fuc, 0x10f6bc, 0x01010101 * next->bios.ramcfg_11_09);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
835
if (!next->bios.ramcfg_11_02_80)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
837
if (!next->bios.ramcfg_11_02_40)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
839
if (!next->bios.ramcfg_11_07_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
841
if (!next->bios.ramcfg_11_07_08)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
848
if (next->bios.ramcfg_11_08_01)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
855
ram_mask(fuc, 0x10f248, 0xffffffff, next->bios.timing[10]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
856
ram_mask(fuc, 0x10f290, 0xffffffff, next->bios.timing[0]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
857
ram_mask(fuc, 0x10f294, 0xffffffff, next->bios.timing[1]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
858
ram_mask(fuc, 0x10f298, 0xffffffff, next->bios.timing[2]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
859
ram_mask(fuc, 0x10f29c, 0xffffffff, next->bios.timing[3]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
860
ram_mask(fuc, 0x10f2a0, 0xffffffff, next->bios.timing[4]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
861
ram_mask(fuc, 0x10f2a4, 0xffffffff, next->bios.timing[5]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
862
ram_mask(fuc, 0x10f2a8, 0xffffffff, next->bios.timing[6]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
863
ram_mask(fuc, 0x10f2ac, 0xffffffff, next->bios.timing[7]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
864
ram_mask(fuc, 0x10f2cc, 0xffffffff, next->bios.timing[8]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
865
ram_mask(fuc, 0x10f2e8, 0xffffffff, next->bios.timing[9]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
869
if (!next->bios.ramcfg_11_01_04)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
871
if (!next->bios.ramcfg_11_07_80)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
876
if (next->bios.ramcfg_11_03_f0) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
877
if (next->bios.rammap_11_08_0c) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
878
if (!next->bios.ramcfg_11_07_80)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
892
ram_wr32(fuc, 0x10f870, 0x11111111 * next->bios.ramcfg_11_03_0f);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
894
ram_mask(fuc, 0x10f250, 0x000003f0, next->bios.timing_20_2c_003f << 4);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
896
data = (next->bios.timing[10] & 0x7f000000) >> 24;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
897
if (data < next->bios.timing_20_2c_1fc0)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
898
data = next->bios.timing_20_2c_1fc0;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
901
ram_mask(fuc, 0x10f224, 0x001f0000, next->bios.timing_20_30_f8 << 16);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
911
if (!next->bios.ramcfg_DLLoff) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
921
if (!next->bios.ramcfg_DLLoff) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
949
if (next->bios.rammap_11_08_01)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
502
struct nvkm_ram_data *next;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
511
next = &ram->base.target;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
512
next->freq = freq;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
513
ram->base.next = next;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
520
&next->bios);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
534
&ver, &hdr, &next->bios);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
541
if (next->bios.ramcfg_timing != 0xff) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
542
data = nvbios_timingEp(bios, next->bios.ramcfg_timing,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
544
&next->bios);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
597
if (!next->bios.ramcfg_DLLoff)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
605
if (next->bios.ramcfg_10_02_10) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
629
if (!next->bios.ramcfg_10_02_10) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
637
switch (next->bios.ramcfg_DLLoff * ram->base.type) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
646
if (next->bios.timing_10_ODT)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
663
next->bios.ramcfg_FBVDDQ) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
671
gt215_ram_gpio(fuc, 0x18, !next->bios.ramcfg_FBVDDQ);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
707
if (next->bios.rammap_10_04_08) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
708
ram_wr32(fuc, 0x1005a0, next->bios.ramcfg_10_06 << 16 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
709
next->bios.ramcfg_10_05 << 8 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
710
next->bios.ramcfg_10_05);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
711
ram_wr32(fuc, 0x1005a4, next->bios.ramcfg_10_08 << 8 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
712
next->bios.ramcfg_10_07);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
713
ram_wr32(fuc, 0x10f804, next->bios.ramcfg_10_09_f0 << 20 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
714
next->bios.ramcfg_10_03_0f << 16 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
715
next->bios.ramcfg_10_09_0f |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
766
ram_mask(fuc, 0x100200, 0x00001000, !next->bios.ramcfg_10_02_08 << 12);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
777
if (!next->bios.ramcfg_10_02_04)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
780
if (next->bios.ramcfg_10_02_04) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
785
if (next->bios.ramcfg_10_02_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
800
if (next->bios.ramcfg_10_02_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
814
unk714 |= (next->bios.ramcfg_10_04_01) << 8;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
816
if (next->bios.ramcfg_10_02_20)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
818
if (next->bios.ramcfg_10_02_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
820
if (next->bios.ramcfg_10_02_01)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
822
if (next->bios.timing_10_24 != 0xff) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
824
unk718 |= next->bios.timing_10_24 << 28;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
826
if (next->bios.ramcfg_10_02_10)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
834
if (!next->bios.timing_10_ODT)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
838
if (!next->bios.ramcfg_DLLoff)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
861
if (next->bios.rammap_10_04_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
863
if (next->bios.ramcfg_10_02_10) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
226
struct nvkm_ram_data *next;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
234
next = &ram->base.target;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
235
next->freq = freq;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
236
ram->base.next = next;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
250
nvbios_rammapEp_from_perf(bios, data, hdr, &next->bios);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
260
&next->bios);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
267
if (next->bios.ramcfg_timing != 0xff) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
268
data = nvbios_timingEp(bios, next->bios.ramcfg_timing,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
269
&ver, &hdr, &cnt, &len, &next->bios);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
304
if (subdev->device->chipset <= 0x96 && !next->bios.ramcfg_00_03_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
318
if (next->bios.timing_10_ODT)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
354
next->bios.rammap_00_16_40 << 14);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
362
nv50_ram_gpio(hwsq, 0x18, !next->bios.ramcfg_FBVDDQ);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
399
if (!next->bios.ramcfg_00_03_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
401
ram_mask(hwsq, 0x100200, 0x00001000, !next->bios.ramcfg_00_04_02 << 12);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
412
if (!next->bios.ramcfg_00_03_08)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
414
if (!next->bios.ramcfg_FBVDDQ)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
416
if ( next->bios.ramcfg_00_04_04)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
421
if (!next->bios.ramcfg_00_03_08)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
425
if ( next->bios.ramcfg_00_03_01)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
427
if ( next->bios.ramcfg_00_03_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
429
if (!next->bios.ramcfg_00_03_08)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
431
if ( next->bios.ramcfg_00_04_04)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
433
if ( next->bios.ramcfg_00_04_20)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
443
if (next->bios.rammap_00_16_20) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
444
ram_wr32(hwsq, 0x1005a0, next->bios.ramcfg_00_07 << 16 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
445
next->bios.ramcfg_00_06 << 8 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
446
next->bios.ramcfg_00_05);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
447
ram_wr32(hwsq, 0x1005a4, next->bios.ramcfg_00_09 << 8 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
448
next->bios.ramcfg_00_08);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
455
if (!next->bios.timing_10_ODT)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
459
if (!next->bios.ramcfg_DLLoff)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
467
if (next->bios.rammap_00_17_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
469
if (!next->bios.rammap_00_16_40)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
471
if (next->bios.ramcfg_00_03_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
473
if (subdev->device->chipset <= 0x96 && next->bios.ramcfg_00_03_02)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
97
unkt3b = 0x19 + ram->base.next->bios.rammap_00_16_40;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
99
ram->base.next->bios.rammap_00_16_40) << 16 |
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c
65
switch (ram->next->bios.timing_ver) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c
67
CL = ram->next->bios.timing_10_CL;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c
68
WR = ram->next->bios.timing_10_WR;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c
69
DLL = !ram->next->bios.ramcfg_DLLoff;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c
70
ODT = ram->next->bios.timing_10_ODT & 3;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c
73
CL = (ram->next->bios.timing[1] & 0x0000001f);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c
74
WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c
80
if (ram->next->bios.timing_ver == 0x20 ||
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr2.c
81
ram->next->bios.ramcfg_timing == 0xff) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c
74
DLL = !ram->next->bios.ramcfg_DLLoff;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c
76
switch (ram->next->bios.timing_ver) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c
78
if (ram->next->bios.timing_hdr < 0x17) {
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c
82
CWL = ram->next->bios.timing_10_CWL;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c
83
CL = ram->next->bios.timing_10_CL;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c
84
WR = ram->next->bios.timing_10_WR;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c
85
ODT = ram->next->bios.timing_10_ODT;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c
88
CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c
89
CL = (ram->next->bios.timing[1] & 0x0000001f) >> 0;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/sddr3.c
90
WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
656
char *start, *next = p, *equal;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
666
while ((start = strsep(&next, ";"))) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
646
void *next;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
649
next = r535_gsp_rpc_get(gsp, fn, max_payload_size);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
650
if (IS_ERR(next)) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
651
repv = next;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
655
memcpy(next, payload, max_payload_size);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
657
repv = r535_gsp_rpc_send(gsp, next, NVKM_GSP_RPC_REPLY_NOWAIT, 0);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
669
next = r535_gsp_rpc_get(gsp, NV_VGPU_MSG_FUNCTION_CONTINUATION_RECORD, size);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
670
if (IS_ERR(next)) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
671
repv = next;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
675
memcpy(next, payload, size);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
677
repv = r535_gsp_rpc_send(gsp, next, NVKM_GSP_RPC_REPLY_NOWAIT, 0);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
205
if (likely(iobj->lru.next) && iobj->map) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
249
if (likely(iobj->lru.next))
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
275
if (likely(iobj->lru.next)) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
277
iobj->lru.next = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
303
iobj->lru.next = NULL; /* Exclude from eviction. */
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
325
if (likely(iobj->lru.next))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1239
struct nvkm_vma *next = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1246
if (vma->addr + vma->size == addr + size && (next = node(vma, next))) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1247
if (!next->part ||
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1248
next->memory || next->mapped != map)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1249
next = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1252
if (prev || next)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1253
return nvkm_vmm_node_merge(vmm, prev, vma, next, size);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1261
struct nvkm_vma *next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1277
next = nvkm_vmm_pfn_split_merge(vmm, vma, start, size, 0, false);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1278
if (!WARN_ON(!next)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1279
vma = next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1283
} while ((vma = node(vma, next)) && (start = vma->addr) < limit);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1343
goto next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1363
goto next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1394
next:
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1397
vma = node(vma, next);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1420
struct nvkm_vma *next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1428
if ((next = node(vma, next)) && (!next->part || next->mapped))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1429
next = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1430
nvkm_vmm_node_merge(vmm, prev, vma, next, vma->size);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1553
for (; map->off; map->mem = map->mem->next) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1614
struct nvkm_vma *prev, *next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1622
if ((next = node(vma, next)) && !next->used) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1623
vma->size += next->size;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1624
nvkm_vmm_free_delete(vmm, next);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1634
struct nvkm_vma *next = vma;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1640
const bool mem = next->memory != NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1641
const bool map = next->mapped;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1642
const u8 refd = next->refd;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1643
const u64 addr = next->addr;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1644
u64 size = next->size;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1647
while ((next = node(next, next)) && next->part &&
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1648
(next->mapped == map) &&
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1649
(next->memory != NULL) == mem &&
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1650
(next->refd == refd))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1651
size += next->size;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1666
} while (next && next->part);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1673
next = vma;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1675
if (next->mapped)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1676
nvkm_vmm_unmap_region(vmm, next);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1677
} while ((next = node(vma, next)) && next->part);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1796
struct nvkm_vma *next = node(this, next);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1805
if (vmm->func->page_block && next && next->page != p)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
474
bool next = !!pgt->pte[ptei].s.sptes;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
475
if (spte != next)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
674
u64 next = 1ULL << page[i - 1].shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
675
u64 part = ALIGN(addr, next) - addr;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
676
if (size - part >= next)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
959
struct nvkm_vma *vma, struct nvkm_vma *next, u64 size)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
961
if (next) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
963
vma->size += next->size;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
964
nvkm_vmm_node_delete(vmm, next);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
974
nvkm_vmm_node_remove(vmm, next);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
976
next->addr -= size;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
977
next->size += size;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
978
nvkm_vmm_node_insert(vmm, next);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
979
return next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
353
(MAP->mem = MAP->mem->next))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
253
map->next = (1 << page->shift) >> 8;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
305
map->next |= 1ULL << 44;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
38
if (map->ctag && !(map->next & (1ULL << 44))) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
45
base += map->next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
52
data += map->next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
101
data += map->next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
20
data += map->next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
233
map->next = 1ULL << page->shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
142
data += map->next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
225
data += map->next;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
436
map->next = (1ULL << page->shift) >> 4;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
482
map->next |= map->ctag;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
242
map->next = 1 << page->shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
313
map->next |= map->ctag;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
35
u64 next = addr + map->type, data;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
48
data = next | (log2blk << 7);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
49
next += pten * map->next;
drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c
38
u32 next, addr;
drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c
41
next = init->addr + init->count * 8;
drivers/gpu/drm/nouveau/nvkm/subdev/therm/gf100.c
46
while (addr < next) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
324
engine = list_entry(dmm->idle_head.next, struct refill_engine,
drivers/gpu/drm/panthor/panthor_heap.c
178
hdr->next = (prev_gpuva & GENMASK_ULL(63, 12)) |
drivers/gpu/drm/panthor/panthor_heap.c
33
u64 next;
drivers/gpu/drm/panthor/panthor_mmu.c
2140
if (op->next && aligned_unmap_end > unmap_end &&
drivers/gpu/drm/panthor/panthor_mmu.c
2141
op->next->va.addr + op->next->va.range >= aligned_unmap_end &&
drivers/gpu/drm/panthor/panthor_mmu.c
2142
iova_mapped_as_huge_page(op->next, unmap_end - 1)) {
drivers/gpu/drm/panthor/panthor_mmu.c
2189
if (op->remap.next) {
drivers/gpu/drm/panthor/panthor_mmu.c
2190
struct panthor_gem_object *bo = to_panthor_bo(op->remap.next->gem.obj);
drivers/gpu/drm/panthor/panthor_mmu.c
2191
u64 addr = op->remap.next->va.addr;
drivers/gpu/drm/panthor/panthor_mmu.c
2192
u64 size = unmap_start + unmap_range - op->remap.next->va.addr;
drivers/gpu/drm/panthor/panthor_mmu.c
2195
bo->base.sgt, op->remap.next->gem.offset, size);
drivers/gpu/drm/qxl/qxl_cmd.c
228
next_id = info->next;
drivers/gpu/drm/qxl/qxl_dev.h
373
uint64_t next; /* out */
drivers/gpu/drm/qxl/qxl_release.c
127
entry = container_of(release->bos.next,
drivers/gpu/drm/radeon/mkregtable.c
102
for (pos = list_entry((head)->next, typeof(*pos), member); \
drivers/gpu/drm/radeon/mkregtable.c
104
pos = list_entry(pos->member.next, typeof(*pos), member))
drivers/gpu/drm/radeon/mkregtable.c
43
struct list_head *next, *prev;
drivers/gpu/drm/radeon/mkregtable.c
49
list->next = list;
drivers/gpu/drm/radeon/mkregtable.c
61
struct list_head *prev, struct list_head *next)
drivers/gpu/drm/radeon/mkregtable.c
63
next->prev = new;
drivers/gpu/drm/radeon/mkregtable.c
64
new->next = next;
drivers/gpu/drm/radeon/mkregtable.c
66
prev->next = new;
drivers/gpu/drm/radeon/mkregtable.c
70
struct list_head *prev, struct list_head *next);
drivers/gpu/drm/radeon/rv6xx_dpm.c
197
struct rv6xx_sclk_stepping next;
drivers/gpu/drm/radeon/rv6xx_dpm.c
199
next.post_divider = cur->post_divider;
drivers/gpu/drm/radeon/rv6xx_dpm.c
202
next.vco_frequency = (cur->vco_frequency * (100 + step_size)) / 100;
drivers/gpu/drm/radeon/rv6xx_dpm.c
204
next.vco_frequency = (cur->vco_frequency * 100 + 99 + step_size) / (100 + step_size);
drivers/gpu/drm/radeon/rv6xx_dpm.c
206
return next;
drivers/gpu/drm/radeon/rv6xx_dpm.c
222
struct rv6xx_sclk_stepping next = *cur;
drivers/gpu/drm/radeon/rv6xx_dpm.c
224
while (rv6xx_can_step_post_div(rdev, &next, target))
drivers/gpu/drm/radeon/rv6xx_dpm.c
225
next.post_divider--;
drivers/gpu/drm/radeon/rv6xx_dpm.c
227
return next;
drivers/gpu/drm/radeon/rv6xx_dpm.c
259
struct rv6xx_sclk_stepping next;
drivers/gpu/drm/radeon/rv6xx_dpm.c
262
next = rv6xx_next_post_div_step(rdev, &cur, &target);
drivers/gpu/drm/radeon/rv6xx_dpm.c
264
next = rv6xx_next_vco_step(rdev, &cur, increasing_vco, R600_VCOSTEPPCT_DFLT);
drivers/gpu/drm/radeon/rv6xx_dpm.c
266
if (rv6xx_reached_stepping_target(rdev, &next, &target, increasing_vco)) {
drivers/gpu/drm/radeon/rv6xx_dpm.c
269
tiny.post_divider = next.post_divider;
drivers/gpu/drm/radeon/rv6xx_dpm.c
274
if ((next.post_divider != target.post_divider) &&
drivers/gpu/drm/radeon/rv6xx_dpm.c
275
(next.vco_frequency != target.vco_frequency)) {
drivers/gpu/drm/radeon/rv6xx_dpm.c
279
final_vco.post_divider = next.post_divider;
drivers/gpu/drm/radeon/rv6xx_dpm.c
287
rv6xx_output_stepping(rdev, step_index++, &next);
drivers/gpu/drm/radeon/rv6xx_dpm.c
289
cur = next;
drivers/gpu/drm/scheduler/sched_entity.c
500
struct drm_sched_job *next;
drivers/gpu/drm/scheduler/sched_entity.c
502
next = drm_sched_entity_queue_peek(entity);
drivers/gpu/drm/scheduler/sched_entity.c
503
if (next) {
drivers/gpu/drm/scheduler/sched_entity.c
510
next->submit_ts);
drivers/gpu/drm/scheduler/sched_main.c
1131
struct drm_sched_job *job, *next;
drivers/gpu/drm/scheduler/sched_main.c
1145
next = list_first_entry_or_null(&sched->pending_list,
drivers/gpu/drm/scheduler/sched_main.c
1146
typeof(*next), list);
drivers/gpu/drm/scheduler/sched_main.c
1147
if (next) {
drivers/gpu/drm/scheduler/sched_main.c
1150
&next->s_fence->scheduled.flags))
drivers/gpu/drm/scheduler/sched_main.c
1151
next->s_fence->scheduled.timestamp =
drivers/gpu/drm/scheduler/sched_main.c
1154
*have_more = dma_fence_is_signaled(&next->s_fence->finished);
drivers/gpu/drm/scheduler/tests/mock_scheduler.c
346
struct drm_mock_sched_job *job, *next;
drivers/gpu/drm/scheduler/tests/mock_scheduler.c
356
list_for_each_entry_safe(job, next, &sched->job_list, link) {
drivers/gpu/drm/scheduler/tests/mock_scheduler.c
78
struct drm_mock_sched_job *next;
drivers/gpu/drm/scheduler/tests/mock_scheduler.c
84
list_for_each_entry_safe(job, next, &sched->job_list, link) {
drivers/gpu/drm/tegra/gem.c
29
dma_addr_t next = ~(dma_addr_t)0;
drivers/gpu/drm/tegra/gem.c
38
if (sg_dma_address(s) != next) {
drivers/gpu/drm/tegra/gem.c
39
next = sg_dma_address(s) + sg_dma_len(s);
drivers/gpu/drm/tests/drm_connector_test.c
268
KUNIT_ASSERT_PTR_EQ(test, connector->head.next, &connector->head);
drivers/gpu/drm/tests/drm_connector_test.c
423
KUNIT_ASSERT_PTR_EQ(test, connector->head.next, &priv->drm.mode_config.connector_list);
drivers/gpu/drm/tests/drm_connector_test.c
530
KUNIT_ASSERT_PTR_EQ(test, priv->connector.head.next, &priv->drm.mode_config.connector_list);
drivers/gpu/drm/tests/drm_framebuffer_test.c
606
KUNIT_EXPECT_PTR_EQ(test, dev->mode_config.fb_list.next, &fb1.head);
drivers/gpu/drm/tests/drm_mm_test.c
245
struct drm_mm_node *node, *next;
drivers/gpu/drm/tests/drm_mm_test.c
272
drm_mm_for_each_node_safe(node, next, &mm) {
drivers/gpu/drm/ttm/ttm_resource.c
624
struct ttm_resource *next = ttm_lru_item_to_res(next_lru);
drivers/gpu/drm/ttm/ttm_resource.c
629
bulk = next->bo->bulk_move;
drivers/gpu/drm/ttm/ttm_resource.c
634
cursor->mem_type = next->mem_type;
drivers/gpu/drm/ttm/ttm_resource.c
72
struct ttm_resource_cursor *cursor, *next;
drivers/gpu/drm/ttm/ttm_resource.c
74
list_for_each_entry_safe(cursor, next, &bulk->cursor_list, bulk_link)
drivers/gpu/drm/ttm/ttm_resource.c
81
struct ttm_resource_cursor *cursor, *next;
drivers/gpu/drm/ttm/ttm_resource.c
83
list_for_each_entry_safe(cursor, next, &bulk->cursor_list, bulk_link)
drivers/gpu/drm/v3d/v3d_submit.c
846
user_ext = u64_to_user_ptr(ext.next);
drivers/gpu/drm/vboxvideo/vbva_base.c
168
u32 next;
drivers/gpu/drm/vboxvideo/vbva_base.c
176
next = (vbva_ctx->vbva->record_free_index + 1) % VBVA_MAX_RECORDS;
drivers/gpu/drm/vboxvideo/vbva_base.c
179
if (next == vbva_ctx->vbva->record_first_index)
drivers/gpu/drm/vboxvideo/vbva_base.c
183
if (next == vbva_ctx->vbva->record_first_index)
drivers/gpu/drm/vboxvideo/vbva_base.c
188
vbva_ctx->vbva->record_free_index = next;
drivers/gpu/drm/vc4/vc4_gem.c
503
struct vc4_exec_info *next;
drivers/gpu/drm/vc4/vc4_gem.c
506
next = vc4_first_bin_job(vc4);
drivers/gpu/drm/vc4/vc4_gem.c
513
if (next && next->perfmon == exec->perfmon)
drivers/gpu/drm/vc4/vc4_hvs.c
1758
struct drm_mm_node *node, *next;
drivers/gpu/drm/vc4/vc4_hvs.c
1763
drm_mm_for_each_node_safe(node, next, &vc4->hvs->dlist_mm)
drivers/gpu/drm/vc4/vc4_hvs.c
1768
drm_mm_for_each_node_safe(node, next, &vc4->hvs->lbm_mm)
drivers/gpu/drm/vc4/vc4_irq.c
121
struct vc4_exec_info *next, *exec = vc4_first_bin_job(vc4);
drivers/gpu/drm/vc4/vc4_irq.c
129
next = vc4_first_bin_job(vc4);
drivers/gpu/drm/vc4/vc4_irq.c
135
if (next && next->perfmon == exec->perfmon)
drivers/gpu/drm/virtio/virtgpu_drv.h
124
struct list_head next;
drivers/gpu/drm/virtio/virtgpu_gem.c
277
list_add_tail(&objs->next, &vgdev->obj_free_list);
drivers/gpu/drm/virtio/virtgpu_gem.c
291
struct virtio_gpu_object_array, next);
drivers/gpu/drm/virtio/virtgpu_gem.c
292
list_del(&objs->next);
drivers/gpu/drm/vkms/vkms_composer.c
227
colorop = colorop->next;
drivers/gpu/drm/vmwgfx/ttm_object.c
395
list = tfile->ref_list.next;
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1408
struct vmw_ctx_bindinfo *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1410
list_for_each_entry_safe(entry, next, &cbs->list, ctx_list)
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
430
struct vmw_ctx_bindinfo *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
433
list_for_each_entry_safe(entry, next, &cbs->list, ctx_list)
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
472
struct vmw_ctx_bindinfo *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
475
list_for_each_entry_safe(entry, next, head, res_list)
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
523
struct vmw_ctx_bindinfo *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
525
list_for_each_entry_safe(entry, next, &from->list, ctx_list) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
385
struct vmw_cmdbuf_header *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
389
list_for_each_entry_safe(entry, next, &ctx->hw_submitted, list) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
520
struct vmw_cmdbuf_header *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
533
list_for_each_entry_safe(entry, next, &man->error, list) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
126
struct vmw_cmdbuf_res *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
128
list_for_each_entry_safe(entry, next, list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
162
struct vmw_cmdbuf_res *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
164
list_for_each_entry_safe(entry, next, list, head) {
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
309
struct vmw_cmdbuf_res *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c
311
list_for_each_entry_safe(entry, next, &man->list, head)
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
330
bool (*next)(struct vmw_piter *);
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
951
return viter->next(viter);
drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c
109
entry = list_entry(lds->active.next, typeof(*entry), active);
drivers/gpu/drm/vmwgfx/vmwgfx_shader.c
548
struct vmw_dx_shader *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_shader.c
552
list_for_each_entry_safe(entry, next, list, cotable_head) {
drivers/gpu/drm/vmwgfx/vmwgfx_so.c
425
struct vmw_view *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_so.c
429
list_for_each_entry_safe(entry, next, list, cotable_head)
drivers/gpu/drm/vmwgfx/vmwgfx_so.c
445
struct vmw_view *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_so.c
449
list_for_each_entry_safe(entry, next, list, srf_head)
drivers/gpu/drm/vmwgfx/vmwgfx_streamoutput.c
359
struct vmw_dx_streamoutput *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_streamoutput.c
363
list_for_each_entry_safe(entry, next, list, cotable_head) {
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c
111
viter->next = &__vmw_piter_non_sg_next;
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c
117
viter->next = &__vmw_piter_sg_next;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
137
struct page *entry, *next;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
139
list_for_each_entry_safe(entry, next, &ctx->page_list, lru) {
drivers/gpu/drm/xe/tests/xe_bo.c
445
struct xe_bo_link *link, *next;
drivers/gpu/drm/xe/tests/xe_bo.c
555
list_for_each_entry_safe(link, next, &bos, link) {
drivers/gpu/drm/xe/tests/xe_gt_sriov_pf_config_kunit.c
24
unsigned long next = 1 + (unsigned long)prev;
drivers/gpu/drm/xe/tests/xe_gt_sriov_pf_config_kunit.c
26
if (next > TEST_MAX_VFS)
drivers/gpu/drm/xe/tests/xe_gt_sriov_pf_config_kunit.c
29
next, str_plural(next));
drivers/gpu/drm/xe/tests/xe_gt_sriov_pf_config_kunit.c
30
return (void *)next;
drivers/gpu/drm/xe/tests/xe_pci.c
258
const void *next = pre_gmdid_graphics_ip_gen_params(test, prev, desc);
drivers/gpu/drm/xe/tests/xe_pci.c
260
if (next)
drivers/gpu/drm/xe/tests/xe_pci.c
261
return next;
drivers/gpu/drm/xe/tests/xe_pci.c
283
const void *next = pre_gmdid_media_ip_gen_params(test, prev, desc);
drivers/gpu/drm/xe/tests/xe_pci.c
285
if (next)
drivers/gpu/drm/xe/tests/xe_pci.c
286
return next;
drivers/gpu/drm/xe/tests/xe_pci.c
401
struct device *next;
drivers/gpu/drm/xe/tests/xe_pci.c
403
next = driver_find_next_device(&xe_pci_driver.driver, dev);
drivers/gpu/drm/xe/tests/xe_pci.c
406
if (!next)
drivers/gpu/drm/xe/tests/xe_pci.c
409
snprintf(desc, KUNIT_PARAM_DESC_SIZE, "%s", dev_name(next));
drivers/gpu/drm/xe/tests/xe_pci.c
410
return pdev_to_xe_device(to_pci_dev(next));
drivers/gpu/drm/xe/xe_bo.c
3561
struct xe_bo *bo, *next;
drivers/gpu/drm/xe/xe_bo.c
3570
llist_for_each_entry_safe(bo, next, freed, freed)
drivers/gpu/drm/xe/xe_exec_queue.c
1395
struct xe_exec_queue *eq = q, *next;
drivers/gpu/drm/xe/xe_exec_queue.c
1397
list_for_each_entry_safe(eq, next, &eq->multi_gt_list,
drivers/gpu/drm/xe/xe_exec_queue.c
474
struct xe_exec_queue *eq, *next;
drivers/gpu/drm/xe/xe_exec_queue.c
490
list_for_each_entry_safe(eq, next, &q->multi_gt_list,
drivers/gpu/drm/xe/xe_hw_fence.c
53
struct xe_hw_fence *fence, *next;
drivers/gpu/drm/xe/xe_hw_fence.c
59
list_for_each_entry_safe(fence, next, &irq->pending, irq_link) {
drivers/gpu/drm/xe/xe_hw_fence.c
84
struct xe_hw_fence *fence, *next;
drivers/gpu/drm/xe/xe_hw_fence.c
91
list_for_each_entry_safe(fence, next, &irq->pending, irq_link) {
drivers/gpu/drm/xe/xe_lmtt.c
372
u64 next, pde, pt_addr;
drivers/gpu/drm/xe/xe_lmtt.c
388
next = min(end, round_up(offset + 1, pte_addr_shift));
drivers/gpu/drm/xe/xe_lmtt.c
391
err = __lmtt_alloc_range(lmtt, pt, offset, next);
drivers/gpu/drm/xe/xe_lmtt.c
396
offset = next;
drivers/gpu/drm/xe/xe_pm.c
385
struct xe_vm *vm, *next;
drivers/gpu/drm/xe/xe_pm.c
388
list_for_each_entry_safe(vm, next, &xe->rebind_resume_list,
drivers/gpu/drm/xe/xe_pt.c
1264
if (!err && op->remap.next)
drivers/gpu/drm/xe/xe_pt.c
1265
err = vma_add_deps(op->remap.next, job);
drivers/gpu/drm/xe/xe_pt.c
1447
if (!err && op->remap.next && !op->remap.skip_next)
drivers/gpu/drm/xe/xe_pt.c
1448
err = vma_check_userptr(vm, op->remap.next, pt_update);
drivers/gpu/drm/xe/xe_pt.c
1551
static bool xe_pt_check_kill(u64 addr, u64 next, unsigned int level,
drivers/gpu/drm/xe/xe_pt.c
1561
if (IS_ALIGNED(addr, size) && IS_ALIGNED(next, size) &&
drivers/gpu/drm/xe/xe_pt.c
1562
((next - addr) >> shift) == child->num_live) {
drivers/gpu/drm/xe/xe_pt.c
1569
if (xe_walk->modified_end <= next)
drivers/gpu/drm/xe/xe_pt.c
1570
xe_walk->modified_end = round_up(next, size);
drivers/gpu/drm/xe/xe_pt.c
1642
unsigned int level, u64 addr, u64 next,
drivers/gpu/drm/xe/xe_pt.c
1660
pgoff_t count = xe_pt_num_entries(addr, next, xe_child->level, walk);
drivers/gpu/drm/xe/xe_pt.c
1673
(i == count - 1 && !IS_ALIGNED(next, pt_size));
drivers/gpu/drm/xe/xe_pt.c
1721
killed = xe_pt_check_kill(addr, next, level - 1, xe_child, action, walk);
drivers/gpu/drm/xe/xe_pt.c
1729
bool covered = xe_pt_covers(addr, next, xe_child->level, &xe_walk->base);
drivers/gpu/drm/xe/xe_pt.c
1738
level, addr, next, xe_child->num_live);
drivers/gpu/drm/xe/xe_pt.c
1746
unsigned int level, u64 addr, u64 next,
drivers/gpu/drm/xe/xe_pt.c
1760
if (!IS_ALIGNED(next, size))
drivers/gpu/drm/xe/xe_pt.c
1761
next = xe_walk->modified_end;
drivers/gpu/drm/xe/xe_pt.c
1765
xe_pt_check_kill(addr, next, level, xe_child, action, walk))
drivers/gpu/drm/xe/xe_pt.c
1768
if (!xe_pt_nonshared_offsets(addr, next, level, walk, action, &offset,
drivers/gpu/drm/xe/xe_pt.c
2206
if (!err && op->remap.next && !op->remap.skip_next) {
drivers/gpu/drm/xe/xe_pt.c
2208
op->remap.next, false);
drivers/gpu/drm/xe/xe_pt.c
2434
if (op->remap.next && !op->remap.skip_next)
drivers/gpu/drm/xe/xe_pt.c
2435
bind_op_commit(vm, tile, pt_update_ops, op->remap.next,
drivers/gpu/drm/xe/xe_pt.c
432
static bool xe_pt_hugepte_possible(u64 addr, u64 next, unsigned int level,
drivers/gpu/drm/xe/xe_pt.c
441
if (!xe_pt_covers(addr, next, level, &xe_walk->base))
drivers/gpu/drm/xe/xe_pt.c
445
if (next - xe_walk->va_curs_start > xe_walk->curs->size)
drivers/gpu/drm/xe/xe_pt.c
457
size = next - addr;
drivers/gpu/drm/xe/xe_pt.c
468
xe_pt_scan_64K(u64 addr, u64 next, struct xe_pt_stage_bind_walk *xe_walk)
drivers/gpu/drm/xe/xe_pt.c
475
if (next > xe_walk->l0_end_addr)
drivers/gpu/drm/xe/xe_pt.c
483
for (; addr < next; addr += SZ_64K) {
drivers/gpu/drm/xe/xe_pt.c
490
return addr == next;
drivers/gpu/drm/xe/xe_pt.c
502
xe_pt_is_pte_ps64K(u64 addr, u64 next, struct xe_pt_stage_bind_walk *xe_walk)
drivers/gpu/drm/xe/xe_pt.c
516
unsigned int level, u64 addr, u64 next,
drivers/gpu/drm/xe/xe_pt.c
532
if (level == 0 || xe_pt_hugepte_possible(addr, next, level, xe_walk)) {
drivers/gpu/drm/xe/xe_pt.c
556
if (xe_pt_is_pte_ps64K(addr, next, xe_walk)) {
drivers/gpu/drm/xe/xe_pt.c
572
xe_res_next(curs, next - addr);
drivers/gpu/drm/xe/xe_pt.c
573
xe_walk->va_curs_start = next;
drivers/gpu/drm/xe/xe_pt.c
587
xe_walk->l0_end_addr = next;
drivers/gpu/drm/xe/xe_pt.c
590
covers = xe_pt_covers(addr, next, level, &xe_walk->base);
drivers/gpu/drm/xe/xe_pt.c
615
covers && xe_pt_scan_64K(addr, next, xe_walk)) {
drivers/gpu/drm/xe/xe_pt.c
854
unsigned int level, u64 addr, u64 next,
drivers/gpu/drm/xe/xe_pt.c
872
if (xe_pt_nonshared_offsets(addr, next, --level, walk, action, &offset,
drivers/gpu/drm/xe/xe_pt_walk.c
105
err = xe_pt_walk_range(child, level - 1, addr, next, walk);
drivers/gpu/drm/xe/xe_pt_walk.c
109
next, &child, &action, walk);
drivers/gpu/drm/xe/xe_pt_walk.c
113
} while (xe_pt_next(&offset, &addr, next, end, level, walk));
drivers/gpu/drm/xe/xe_pt_walk.c
36
static bool xe_pt_next(pgoff_t *offset, u64 *addr, u64 next, u64 end,
drivers/gpu/drm/xe/xe_pt_walk.c
46
if (skip_to > next) {
drivers/gpu/drm/xe/xe_pt_walk.c
47
step += (skip_to - next) >> shift;
drivers/gpu/drm/xe/xe_pt_walk.c
48
next = skip_to;
drivers/gpu/drm/xe/xe_pt_walk.c
52
*addr = next;
drivers/gpu/drm/xe/xe_pt_walk.c
55
return next != end;
drivers/gpu/drm/xe/xe_pt_walk.c
83
u64 next;
drivers/gpu/drm/xe/xe_pt_walk.c
86
next = xe_pt_addr_end(addr, end, level, walk);
drivers/gpu/drm/xe/xe_pt_walk.c
87
if (walk->shared_pt_mode && xe_pt_covers(addr, next, level,
drivers/gpu/drm/xe/xe_pt_walk.c
93
err = ops->pt_entry(parent, offset, level, addr, next,
drivers/gpu/drm/xe/xe_pt_walk.h
64
unsigned int level, u64 addr, u64 next,
drivers/gpu/drm/xe/xe_range_fence.c
32
struct xe_range_fence *rfence, *next;
drivers/gpu/drm/xe/xe_range_fence.c
34
llist_for_each_entry_safe(rfence, next, node, link) {
drivers/gpu/drm/xe/xe_res_cursor.h
108
struct list_head *head, *next;
drivers/gpu/drm/xe/xe_res_cursor.h
122
next = block->link.next;
drivers/gpu/drm/xe/xe_res_cursor.h
123
if (next != head)
drivers/gpu/drm/xe/xe_res_cursor.h
124
block = list_entry(next, struct drm_buddy_block,
drivers/gpu/drm/xe/xe_res_cursor.h
263
struct list_head *next;
drivers/gpu/drm/xe/xe_res_cursor.h
297
next = block->link.next;
drivers/gpu/drm/xe/xe_res_cursor.h
298
block = list_entry(next, struct drm_buddy_block, link);
drivers/gpu/drm/xe/xe_res_cursor.h
304
next = block->link.next;
drivers/gpu/drm/xe/xe_res_cursor.h
305
block = list_entry(next, struct drm_buddy_block, link);
drivers/gpu/drm/xe/xe_svm.c
1409
struct drm_gpusvm_notifier *notifier, *next;
drivers/gpu/drm/xe/xe_svm.c
1413
drm_gpusvm_for_each_notifier_safe(notifier, next, &vm->svm.gpusvm, start, end) {
drivers/gpu/drm/xe/xe_tlb_inval.c
164
struct xe_tlb_inval_fence *fence, *next;
drivers/gpu/drm/xe/xe_tlb_inval.c
196
list_for_each_entry_safe(fence, next,
drivers/gpu/drm/xe/xe_tlb_inval.c
360
struct xe_tlb_inval_fence *fence, *next;
drivers/gpu/drm/xe/xe_tlb_inval.c
392
list_for_each_entry_safe(fence, next,
drivers/gpu/drm/xe/xe_tlb_inval.c
71
struct xe_tlb_inval_fence *fence, *next;
drivers/gpu/drm/xe/xe_tlb_inval.c
77
list_for_each_entry_safe(fence, next,
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
35
while (head != block->link.next) {
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
39
block = list_entry(block->link.next, struct drm_buddy_block,
drivers/gpu/drm/xe/xe_userptr.c
188
struct xe_userptr_vma *uvma, *next;
drivers/gpu/drm/xe/xe_userptr.c
197
list_for_each_entry_safe(uvma, next, &vm->userptr.invalidated,
drivers/gpu/drm/xe/xe_userptr.c
206
list_for_each_entry_safe(uvma, next, &vm->userptr.repin_list,
drivers/gpu/drm/xe/xe_userptr.c
249
list_for_each_entry_safe(uvma, next, &vm->userptr.repin_list,
drivers/gpu/drm/xe/xe_vm.c
163
link = list->next;
drivers/gpu/drm/xe/xe_vm.c
1738
struct drm_gpuva *gpuva, *next;
drivers/gpu/drm/xe/xe_vm.c
1775
drm_gpuvm_for_each_va_safe(gpuva, next, &vm->gpuvm) {
drivers/gpu/drm/xe/xe_vm.c
2193
if (op->remap.next)
drivers/gpu/drm/xe/xe_vm.c
2196
(ULL)op->remap.next->va.addr,
drivers/gpu/drm/xe/xe_vm.c
2197
(ULL)op->remap.next->va.range);
drivers/gpu/drm/xe/xe_vm.c
2559
if (op->remap.next) {
drivers/gpu/drm/xe/xe_vm.c
2560
err |= xe_vm_insert_vma(vm, op->remap.next);
drivers/gpu/drm/xe/xe_vm.c
2564
op->remap.next->tile_present =
drivers/gpu/drm/xe/xe_vm.c
2678
if (op->base.remap.next)
drivers/gpu/drm/xe/xe_vm.c
2679
end = op->base.remap.next->va.addr;
drivers/gpu/drm/xe/xe_vm.c
2725
if (op->base.remap.next) {
drivers/gpu/drm/xe/xe_vm.c
2726
vma = new_vma(vm, op->base.remap.next,
drivers/gpu/drm/xe/xe_vm.c
2731
op->remap.next = vma;
drivers/gpu/drm/xe/xe_vm.c
2833
if (op->remap.next) {
drivers/gpu/drm/xe/xe_vm.c
2834
prep_vma_destroy(vm, op->remap.next, next_post_commit);
drivers/gpu/drm/xe/xe_vm.c
2835
xe_vma_destroy_unlocked(op->remap.next);
drivers/gpu/drm/xe/xe_vm.c
3010
if (!err && op->remap.next)
drivers/gpu/drm/xe/xe_vm.c
3011
err = vma_lock_and_validate(exec, op->remap.next,
drivers/gpu/drm/xe/xe_vm.c
3106
if (op->remap.next)
drivers/gpu/drm/xe/xe_vm.c
3107
trace_xe_vma_bind(op->remap.next);
drivers/gpu/drm/xe/xe_vm.c
3275
if (op->remap.next)
drivers/gpu/drm/xe/xe_vm.c
3276
vma_add_ufence(op->remap.next, ufence);
drivers/gpu/drm/xe/xe_vm.c
4546
struct xe_vma *prev, *next;
drivers/gpu/drm/xe/xe_vm.c
4557
next = xe_vm_find_vma_by_addr(vm, *end + 1);
drivers/gpu/drm/xe/xe_vm.c
4558
if (is_cpu_addr_vma_with_default_attr(next))
drivers/gpu/drm/xe/xe_vm.c
4559
*end = xe_vma_end(next);
drivers/gpu/drm/xe/xe_vm.c
690
struct xe_vma *vma, *next;
drivers/gpu/drm/xe/xe_vm.c
728
list_for_each_entry_safe(vma, next, &vm->rebind_list,
drivers/gpu/drm/xe/xe_vm.c
83
struct list_head *link, *next;
drivers/gpu/drm/xe/xe_vm.c
85
list_for_each_safe(link, next, list)
drivers/gpu/drm/xe/xe_vm_types.h
371
struct xe_vma *next;
drivers/gpu/ipu-v3/ipu-image-convert.c
1542
run = list_entry(chan->done_q.next,
drivers/greybus/interface.c
1184
struct gb_bundle *next;
drivers/greybus/interface.c
1197
list_for_each_entry_safe(bundle, next, &intf->bundles, links)
drivers/greybus/manifest.c
232
struct manifest_desc *desc, *next, *tmp;
drivers/greybus/manifest.c
240
list_for_each_entry_safe(desc, next, &intf->manifest_descs, links) {
drivers/greybus/manifest.c
285
list_for_each_entry_safe(desc, next, &list, links) {
drivers/greybus/manifest.c
54
struct manifest_desc *next;
drivers/greybus/manifest.c
56
list_for_each_entry_safe(descriptor, next, &intf->manifest_descs, links)
drivers/greybus/manifest.c
79
struct manifest_desc *next;
drivers/greybus/manifest.c
81
list_for_each_entry_safe(descriptor, next, &intf->manifest_descs, links)
drivers/hid/hid-core.c
1266
const __u8 *next;
drivers/hid/hid-core.c
1333
while ((next = fetch_item(start, end, &item)) != NULL) {
drivers/hid/hid-core.c
1334
start = next;
drivers/hid/hid-core.c
1756
struct hid_field_entry *next;
drivers/hid/hid-core.c
1763
list_for_each_entry(next,
drivers/hid/hid-core.c
1770
if (entry->priority > next->priority) {
drivers/hid/hid-core.c
1771
list_add_tail(&entry->list, &next->list);
drivers/hid/hid-cougar.c
197
struct hid_input *next, *hidinput = NULL;
drivers/hid/hid-cougar.c
233
list_for_each_entry_safe(hidinput, next, &hdev->inputs, list) {
drivers/hid/hid-debug.c
3062
list = report_enum->report_list.next;
drivers/hid/hid-debug.c
3076
list = list->next;
drivers/hid/hid-gaff.c
78
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-gaff.c
86
report_ptr = report_ptr->next;
drivers/hid/hid-google-stadiaff.c
72
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-holtekff.c
137
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-holtekff.c
145
report = list_entry(report_list->next, struct hid_report, list);
drivers/hid/hid-input.c
2303
struct hid_input *next, *hidinput = NULL;
drivers/hid/hid-input.c
2367
list_for_each_entry_safe(hidinput, next, &hid->inputs, list) {
drivers/hid/hid-input.c
2404
struct hid_input *hidinput, *next;
drivers/hid/hid-input.c
2408
list_for_each_entry_safe(hidinput, next, &hid->inputs, list) {
drivers/hid/hid-lg2ff.c
61
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-lg3ff.c
126
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-lg3ff.c
49
struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
drivers/hid/hid-lg3ff.c
87
struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
drivers/hid/hid-lg4ff.c
1265
struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
drivers/hid/hid-lg4ff.c
1279
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-lgff.c
103
struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
drivers/hid/hid-lgff.c
128
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-lgff.c
61
struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
drivers/hid/hid-logitech-hidpp.c
2762
struct hid_input *hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-logitech-hidpp.c
2772
struct hid_input *hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-logitech-hidpp.c
2823
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-megaworld.c
51
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-mf.c
88
input_ptr = input_ptr->next;
drivers/hid/hid-microsoft.c
336
hidinput = list_entry(hdev->inputs.next, struct hid_input, list);
drivers/hid/hid-pl.c
103
report_ptr = report_ptr->next;
drivers/hid/hid-sjoy.c
68
report_ptr = report_ptr->next;
drivers/hid/hid-sony.c
1337
struct hid_report *report = list_entry(report_list->next,
drivers/hid/hid-sony.c
1714
hidinput = list_entry(sc->hdev->inputs.next, struct hid_input, list);
drivers/hid/hid-sony.c
1857
if (sc->list_node.next) {
drivers/hid/hid-steelseries.c
134
struct hid_report *report = list_entry(report_list->next, struct hid_report, list);
drivers/hid/hid-tmff.c
132
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/hid-zpff.c
65
hidinput = list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/intel-ish-hid/ishtp/client-buffers.c
105
rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb,
drivers/hid/intel-ish-hid/ishtp/client-buffers.c
115
rb = list_entry(cl->in_process_list.list.next,
drivers/hid/intel-ish-hid/ishtp/client-buffers.c
138
tx_buf = list_entry(cl->tx_free_list.list.next,
drivers/hid/intel-ish-hid/ishtp/client-buffers.c
150
tx_buf = list_entry(cl->tx_list.list.next,
drivers/hid/intel-ish-hid/ishtp/client.c
1057
new_rb = list_entry(cl->free_rb_list.list.next,
drivers/hid/intel-ish-hid/ishtp/client.c
1186
new_rb = list_entry(cl->free_rb_list.list.next,
drivers/hid/intel-ish-hid/ishtp/client.c
26
struct ishtp_cl_rb *next;
drivers/hid/intel-ish-hid/ishtp/client.c
30
list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list)
drivers/hid/intel-ish-hid/ishtp/client.c
627
rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, list);
drivers/hid/intel-ish-hid/ishtp/client.c
823
cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring,
drivers/hid/intel-ish-hid/ishtp/client.c
910
cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring,
drivers/hid/usbhid/hid-pidff.c
1521
list_entry(hid->inputs.next, struct hid_input, list);
drivers/hid/usbhid/hiddev.c
79
list = report_enum->report_list.next;
drivers/hid/usbhid/hiddev.c
89
list = report->list.next;
drivers/hid/wacom_sys.c
1531
int group, next;
drivers/hid/wacom_sys.c
1537
next = cur->id;
drivers/hid/wacom_sys.c
1540
next_led = wacom_led_find(wacom, group, ++next);
drivers/hsi/clients/cmt_speech.c
1322
struct list_head *cursor, *next;
drivers/hsi/clients/cmt_speech.c
1325
list_for_each_safe(cursor, next, head) {
drivers/hsi/clients/cmt_speech.c
167
entry = list_entry(head->next, struct char_queue, list);
drivers/hv/ring_buffer.c
77
u32 next = ring_info->ring_buffer->write_index;
drivers/hv/ring_buffer.c
79
return next;
drivers/hwmon/ibmpex.c
250
struct ibmpex_bmc_data *p, *next;
drivers/hwmon/ibmpex.c
252
list_for_each_entry_safe(p, next, &driver_data.bmc_data, list)
drivers/hwmon/ibmpex.c
576
struct ibmpex_bmc_data *p, *next;
drivers/hwmon/ibmpex.c
579
list_for_each_entry_safe(p, next, &driver_data.bmc_data, list)
drivers/hwmon/pmbus/pmbus_core.c
1457
sensor->next = data->sensors;
drivers/hwmon/pmbus/pmbus_core.c
165
for (sensor = data->sensors; sensor; sensor = sensor->next)
drivers/hwmon/pmbus/pmbus_core.c
175
for (sensor = data->sensors; sensor; sensor = sensor->next)
drivers/hwmon/pmbus/pmbus_core.c
47
struct pmbus_sensor *next;
drivers/hwmon/pmbus/pmbus_core.c
540
for (sensor = data->sensors; sensor; sensor = sensor->next) {
drivers/hwtracing/coresight/coresight-core.c
868
struct coresight_node *nd, *next;
drivers/hwtracing/coresight/coresight-core.c
870
list_for_each_entry_safe(nd, next, &path->path_list, link) {
drivers/hwtracing/intel_th/msu.c
315
return win->entry.next == &win->msc->win_list;
drivers/hwtracing/stm/policy.c
501
next:
drivers/hwtracing/stm/policy.c
521
goto next;
drivers/i2c/busses/scx200_acb.c
461
iface->next = scx200_acb_list;
drivers/i2c/busses/scx200_acb.c
585
scx200_acb_list = iface->next;
drivers/i2c/busses/scx200_acb.c
64
struct scx200_acb_iface *next;
drivers/i2c/i2c-core-base.c
1336
struct i2c_client *client, *next;
drivers/i2c/i2c-core-base.c
1356
list_for_each_entry_safe(client, next, &adap->userspace_clients,
drivers/i2c/i2c-core-base.c
1758
struct i2c_client *client, *next;
drivers/i2c/i2c-core-base.c
1779
list_for_each_entry_safe(client, next, &adap->userspace_clients,
drivers/iio/adc/ina2xx-adc.c
780
struct timespec64 next, now, delta;
drivers/iio/adc/ina2xx-adc.c
790
ktime_get_ts64(&next);
drivers/iio/adc/ina2xx-adc.c
803
ktime_get_ts64(&next);
drivers/iio/adc/ina2xx-adc.c
821
timespec64_add_ns(&next, 1000 * sampling_us);
drivers/iio/adc/ina2xx-adc.c
822
delta = timespec64_sub(next, now);
drivers/iio/imu/adis16480.c
1437
u16 next = be16_to_cpu(buffer[offset + 1]);
drivers/iio/imu/adis16480.c
1439
if (curr == st->burst_id && next != st->burst_id) {
drivers/iio/inkern.c
35
struct iio_map_internal *mapi, *next;
drivers/iio/inkern.c
37
list_for_each_entry_safe(mapi, next, &iio_map_list, l) {
drivers/infiniband/core/cm.c
893
work = list_entry(cm_id_priv->work_list.next, struct cm_work, list);
drivers/infiniband/core/counters.c
303
goto next;
drivers/infiniband/core/counters.c
307
next:
drivers/infiniband/core/counters.c
422
goto next;
drivers/infiniband/core/counters.c
426
next:
drivers/infiniband/core/mad.c
1301
next_send_buf = send_buf->next;
drivers/infiniband/core/mad.c
2001
mad_send_wr = list_entry(mad_agent_priv->backlog_list.next,
drivers/infiniband/core/mad.c
2414
mad_send_wr = list_entry(mad_agent_priv->wait_list.next,
drivers/infiniband/core/mad.c
2440
if (mad_agent_priv->wait_list.next == &mad_send_wr->agent_list)
drivers/infiniband/core/mad.c
2542
mad_list = container_of(qp_info->overflow_list.next,
drivers/infiniband/core/mad.c
2767
local = list_entry(mad_agent_priv->local_list.next,
drivers/infiniband/core/mad.c
2891
mad_send_wr = list_entry(mad_agent_priv->wait_list.next,
drivers/infiniband/core/mad.c
2940
recv_wr.next = NULL;
drivers/infiniband/core/mad.c
3014
mad_list = list_entry(qp_info->recv_queue.list.next,
drivers/infiniband/core/mad.c
848
send_wr->cur_seg = container_of(send_wr->rmpp_list.next,
drivers/infiniband/core/mad_rmpp.c
388
if (seg->list.next == rmpp_list)
drivers/infiniband/core/mad_rmpp.c
391
return container_of(seg->list.next, struct ib_mad_recv_buf, list);
drivers/infiniband/core/multicast.c
402
member = list_entry(group->active_list.next,
drivers/infiniband/core/multicast.c
444
member = list_entry(group->pending_list.next,
drivers/infiniband/core/multicast.c
497
member = list_entry(group->pending_list.next,
drivers/infiniband/core/nldev.c
1634
goto next;
drivers/infiniband/core/nldev.c
1637
goto next;
drivers/infiniband/core/nldev.c
1664
next: idx++;
drivers/infiniband/core/rw.c
131
prev->wr.wr.next = &reg->inv_wr;
drivers/infiniband/core/rw.c
133
prev->wr.wr.next = &reg->reg_wr.wr;
drivers/infiniband/core/rw.c
136
reg->reg_wr.wr.next = &reg->wr.wr;
drivers/infiniband/core/rw.c
191
prev->wr.wr.next = NULL;
drivers/infiniband/core/rw.c
271
prev->wr.wr.next = NULL;
drivers/infiniband/core/rw.c
330
rdma_wr->wr.next = i + 1 < ctx->nr_ops ?
drivers/infiniband/core/rw.c
462
rdma_wr->wr.next = i + 1 < nr_ops ?
drivers/infiniband/core/rw.c
82
reg->inv_wr.next = &reg->reg_wr.wr;
drivers/infiniband/core/rw.c
823
ctx->reg->reg_wr.wr.next = &rdma_wr->wr;
drivers/infiniband/core/rw.c
85
reg->inv_wr.next = NULL;
drivers/infiniband/core/rw.c
884
if (ctx->reg[0].inv_wr.next)
drivers/infiniband/core/rw.c
907
last_wr->next = chain_wr;
drivers/infiniband/core/sa_query.c
1020
query = list_entry(ib_nl_request_list.next,
drivers/infiniband/core/sa_query.c
896
if (ib_nl_request_list.next == &query->list)
drivers/infiniband/core/user_mad.c
346
recv_buf = container_of(recv_buf->list.next,
drivers/infiniband/core/user_mad.c
417
packet = list_entry(file->recv_list.next, struct ib_umad_packet, list);
drivers/infiniband/core/uverbs_cmd.c
2025
struct ib_send_wr *wr = NULL, *last, *next;
drivers/infiniband/core/uverbs_cmd.c
2106
next = &ud->wr;
drivers/infiniband/core/uverbs_cmd.c
2122
next = &rdma->wr;
drivers/infiniband/core/uverbs_cmd.c
2139
next = &atomic->wr;
drivers/infiniband/core/uverbs_cmd.c
2143
next_size = sizeof(*next);
drivers/infiniband/core/uverbs_cmd.c
2144
next = alloc_wr(next_size, user_wr->num_sge);
drivers/infiniband/core/uverbs_cmd.c
2145
if (!next) {
drivers/infiniband/core/uverbs_cmd.c
2156
next->ex.imm_data =
drivers/infiniband/core/uverbs_cmd.c
2159
next->ex.invalidate_rkey = user_wr->ex.invalidate_rkey;
drivers/infiniband/core/uverbs_cmd.c
2163
wr = next;
drivers/infiniband/core/uverbs_cmd.c
2165
last->next = next;
drivers/infiniband/core/uverbs_cmd.c
2166
last = next;
drivers/infiniband/core/uverbs_cmd.c
2168
next->next = NULL;
drivers/infiniband/core/uverbs_cmd.c
2169
next->wr_id = user_wr->wr_id;
drivers/infiniband/core/uverbs_cmd.c
2170
next->num_sge = user_wr->num_sge;
drivers/infiniband/core/uverbs_cmd.c
2171
next->opcode = user_wr->opcode;
drivers/infiniband/core/uverbs_cmd.c
2172
next->send_flags = user_wr->send_flags;
drivers/infiniband/core/uverbs_cmd.c
2174
if (next->num_sge) {
drivers/infiniband/core/uverbs_cmd.c
2175
next->sg_list = (void *) next +
drivers/infiniband/core/uverbs_cmd.c
2177
if (copy_from_user(next->sg_list, sgls + sg_ind,
drivers/infiniband/core/uverbs_cmd.c
2178
next->num_sge *
drivers/infiniband/core/uverbs_cmd.c
2183
sg_ind += next->num_sge;
drivers/infiniband/core/uverbs_cmd.c
2185
next->sg_list = NULL;
drivers/infiniband/core/uverbs_cmd.c
2191
for (next = wr; next; next = next->next) {
drivers/infiniband/core/uverbs_cmd.c
2193
if (next == bad_wr)
drivers/infiniband/core/uverbs_cmd.c
2208
next = wr->next;
drivers/infiniband/core/uverbs_cmd.c
2210
wr = next;
drivers/infiniband/core/uverbs_cmd.c
2224
struct ib_recv_wr *wr = NULL, *last, *next;
drivers/infiniband/core/uverbs_cmd.c
2264
(U32_MAX - ALIGN(sizeof(*next), sizeof(struct ib_sge))) /
drivers/infiniband/core/uverbs_cmd.c
2270
next = kmalloc(ALIGN(sizeof(*next), sizeof(struct ib_sge)) +
drivers/infiniband/core/uverbs_cmd.c
2273
if (!next) {
drivers/infiniband/core/uverbs_cmd.c
2279
wr = next;
drivers/infiniband/core/uverbs_cmd.c
2281
last->next = next;
drivers/infiniband/core/uverbs_cmd.c
2282
last = next;
drivers/infiniband/core/uverbs_cmd.c
2284
next->next = NULL;
drivers/infiniband/core/uverbs_cmd.c
2285
next->wr_id = user_wr->wr_id;
drivers/infiniband/core/uverbs_cmd.c
2286
next->num_sge = user_wr->num_sge;
drivers/infiniband/core/uverbs_cmd.c
2288
if (next->num_sge) {
drivers/infiniband/core/uverbs_cmd.c
2289
next->sg_list = (void *)next +
drivers/infiniband/core/uverbs_cmd.c
2290
ALIGN(sizeof(*next), sizeof(struct ib_sge));
drivers/infiniband/core/uverbs_cmd.c
2291
if (copy_from_user(next->sg_list, sgls + sg_ind,
drivers/infiniband/core/uverbs_cmd.c
2292
next->num_sge *
drivers/infiniband/core/uverbs_cmd.c
2297
sg_ind += next->num_sge;
drivers/infiniband/core/uverbs_cmd.c
2299
next->sg_list = NULL;
drivers/infiniband/core/uverbs_cmd.c
2309
next = wr->next;
drivers/infiniband/core/uverbs_cmd.c
2311
wr = next;
drivers/infiniband/core/uverbs_cmd.c
2321
struct ib_recv_wr *wr, *next;
drivers/infiniband/core/uverbs_cmd.c
2348
for (next = wr; next; next = next->next) {
drivers/infiniband/core/uverbs_cmd.c
2350
if (next == bad_wr)
drivers/infiniband/core/uverbs_cmd.c
2360
next = wr->next;
drivers/infiniband/core/uverbs_cmd.c
2362
wr = next;
drivers/infiniband/core/uverbs_cmd.c
2372
struct ib_recv_wr *wr, *next;
drivers/infiniband/core/uverbs_cmd.c
2400
for (next = wr; next; next = next->next) {
drivers/infiniband/core/uverbs_cmd.c
2402
if (next == bad_wr)
drivers/infiniband/core/uverbs_cmd.c
2412
next = wr->next;
drivers/infiniband/core/uverbs_cmd.c
2414
wr = next;
drivers/infiniband/core/uverbs_ioctl.c
126
buf->next = pbundle->allocated_mem;
drivers/infiniband/core/uverbs_ioctl.c
40
struct bundle_alloc_head *next;
drivers/infiniband/core/uverbs_ioctl.c
553
memblock = memblock->next;
drivers/infiniband/core/uverbs_ioctl.c
588
pbundle->alloc_head.next = NULL;
drivers/infiniband/core/uverbs_main.c
256
event = list_entry(ev_queue->event_list.next, struct ib_uverbs_event, list);
drivers/infiniband/core/uverbs_main.c
262
list_del(ev_queue->event_list.next);
drivers/infiniband/core/verbs.c
2897
.next = NULL,
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2059
wr = wr->next;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2887
wr = wr->next;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2995
wr = wr->next;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3032
wr = wr->next;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3085
wr = wr->next;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3663
rwr.next = NULL;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3676
swr->next = NULL;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
758
int i, next;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
767
next = srq->start_idx;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
768
srq->start_idx = srq->swq[next].next_idx;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
784
srqe->wr_id[0] = cpu_to_le32((u32)next);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
785
srq->swq[next].wr_id = wqe->wr_id;
drivers/infiniband/hw/cxgb4/cm.c
4287
tmp = timeout_list.next;
drivers/infiniband/hw/cxgb4/cm.c
4289
tmp->next = NULL;
drivers/infiniband/hw/cxgb4/cm.c
4338
if (!ep->entry.next) {
drivers/infiniband/hw/cxgb4/qp.c
1027
wr = wr->next;
drivers/infiniband/hw/cxgb4/qp.c
1070
wr = wr->next;
drivers/infiniband/hw/cxgb4/qp.c
1121
wr && wr->next && !wr->next->next &&
drivers/infiniband/hw/cxgb4/qp.c
1124
(wr->next->opcode == IB_WR_SEND ||
drivers/infiniband/hw/cxgb4/qp.c
1125
wr->next->opcode == IB_WR_SEND_WITH_INV) &&
drivers/infiniband/hw/cxgb4/qp.c
1126
wr->next->sg_list[0].length == T4_WRITE_CMPL_MAX_CQE &&
drivers/infiniband/hw/cxgb4/qp.c
1127
wr->next->num_sge == 1 && num_wrs >= 2) {
drivers/infiniband/hw/cxgb4/qp.c
1243
wr = wr->next;
drivers/infiniband/hw/cxgb4/qp.c
1325
wr = wr->next;
drivers/infiniband/hw/cxgb4/qp.c
1412
wr = wr->next;
drivers/infiniband/hw/cxgb4/qp.c
633
if (wr->next->opcode == IB_WR_SEND)
drivers/infiniband/hw/cxgb4/qp.c
636
wcwr->stag_inv = cpu_to_be32(wr->next->ex.invalidate_rkey);
drivers/infiniband/hw/cxgb4/qp.c
641
if (wr->next->send_flags & IB_SEND_INLINE)
drivers/infiniband/hw/cxgb4/qp.c
642
build_immd_cmpl(sq, &wcwr->u_cmpl.immd_src, wr->next);
drivers/infiniband/hw/cxgb4/qp.c
645
&wcwr->u_cmpl.isgl_src, wr->next->sg_list, 1, NULL);
drivers/infiniband/hw/cxgb4/qp.c
689
bool send_signaled = (wr->next->send_flags & IB_SEND_SIGNALED) ||
drivers/infiniband/hw/cxgb4/qp.c
730
if (wr->next->opcode == IB_WR_SEND)
drivers/infiniband/hw/cxgb4/qp.c
738
swsqe->wr_id = wr->next->wr_id;
drivers/infiniband/hw/cxgb4/resource.c
116
entry = list_entry(uctx->cqids.next, struct c4iw_qid_list,
drivers/infiniband/hw/cxgb4/resource.c
186
entry = list_entry(uctx->qpids.next, struct c4iw_qid_list,
drivers/infiniband/hw/erdma/erdma_cm.c
206
work = list_entry(cep->work_freelist.next, struct erdma_cm_work,
drivers/infiniband/hw/erdma/erdma_qp.c
690
wr = wr->next;
drivers/infiniband/hw/erdma/erdma_qp.c
747
wr = wr->next;
drivers/infiniband/hw/hfi1/debugfs.h
14
.next = _##name##_seq_next, \
drivers/infiniband/hw/hfi1/driver.c
1858
goto next;
drivers/infiniband/hw/hfi1/driver.c
1861
goto next;
drivers/infiniband/hw/hfi1/driver.c
1873
goto next; /* just in case */
drivers/infiniband/hw/hfi1/driver.c
1881
next:
drivers/infiniband/hw/hfi1/driver.c
592
goto next;
drivers/infiniband/hw/hfi1/driver.c
595
goto next;
drivers/infiniband/hw/hfi1/driver.c
608
goto next; /* just in case */
drivers/infiniband/hw/hfi1/driver.c
612
goto next;
drivers/infiniband/hw/hfi1/driver.c
621
goto next;
drivers/infiniband/hw/hfi1/driver.c
630
next:
drivers/infiniband/hw/hfi1/pio.c
1407
u32 head, next;
drivers/infiniband/hw/hfi1/pio.c
1467
next = head + 1;
drivers/infiniband/hw/hfi1/pio.c
1468
if (next >= sc->sr_size)
drivers/infiniband/hw/hfi1/pio.c
1469
next = 0;
drivers/infiniband/hw/hfi1/pio.c
1475
sc->sr_head = next;
drivers/infiniband/hw/hfi1/rc.c
109
if (++next > rvt_size_atomic(&dev->rdi))
drivers/infiniband/hw/hfi1/rc.c
110
next = 0;
drivers/infiniband/hw/hfi1/rc.c
118
qp->s_acked_ack_queue = next;
drivers/infiniband/hw/hfi1/rc.c
119
qp->s_tail_ack_queue = next;
drivers/infiniband/hw/hfi1/rc.c
3018
u8 next;
drivers/infiniband/hw/hfi1/rc.c
3022
next = qp->r_head_ack_queue + 1;
drivers/infiniband/hw/hfi1/rc.c
3024
if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device)))
drivers/infiniband/hw/hfi1/rc.c
3025
next = 0;
drivers/infiniband/hw/hfi1/rc.c
3027
if (unlikely(next == qp->s_acked_ack_queue)) {
drivers/infiniband/hw/hfi1/rc.c
3028
if (!qp->s_ack_queue[next].sent)
drivers/infiniband/hw/hfi1/rc.c
3030
update_ack_queue(qp, next);
drivers/infiniband/hw/hfi1/rc.c
3070
qp->r_head_ack_queue = next;
drivers/infiniband/hw/hfi1/rc.c
3093
u8 next;
drivers/infiniband/hw/hfi1/rc.c
3098
next = qp->r_head_ack_queue + 1;
drivers/infiniband/hw/hfi1/rc.c
3099
if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device)))
drivers/infiniband/hw/hfi1/rc.c
3100
next = 0;
drivers/infiniband/hw/hfi1/rc.c
3102
if (unlikely(next == qp->s_acked_ack_queue)) {
drivers/infiniband/hw/hfi1/rc.c
3103
if (!qp->s_ack_queue[next].sent)
drivers/infiniband/hw/hfi1/rc.c
3105
update_ack_queue(qp, next);
drivers/infiniband/hw/hfi1/rc.c
3141
qp->r_head_ack_queue = next;
drivers/infiniband/hw/hfi1/rc.c
81
u8 next = qp->s_tail_ack_queue;
drivers/infiniband/hw/hfi1/rc.h
15
unsigned int next;
drivers/infiniband/hw/hfi1/rc.h
17
next = n + 1;
drivers/infiniband/hw/hfi1/rc.h
18
if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device)))
drivers/infiniband/hw/hfi1/rc.h
19
next = 0;
drivers/infiniband/hw/hfi1/rc.h
20
qp->s_tail_ack_queue = next;
drivers/infiniband/hw/hfi1/rc.h
21
qp->s_acked_ack_queue = next;
drivers/infiniband/hw/hfi1/tid_rdma.c
2240
u8 next;
drivers/infiniband/hw/hfi1/tid_rdma.c
2273
next = qp->r_head_ack_queue + 1;
drivers/infiniband/hw/hfi1/tid_rdma.c
2274
if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device)))
drivers/infiniband/hw/hfi1/tid_rdma.c
2275
next = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
2277
if (unlikely(next == qp->s_tail_ack_queue)) {
drivers/infiniband/hw/hfi1/tid_rdma.c
2278
if (!qp->s_ack_queue[next].sent) {
drivers/infiniband/hw/hfi1/tid_rdma.c
2282
update_ack_queue(qp, next);
drivers/infiniband/hw/hfi1/tid_rdma.c
2309
qp->r_head_ack_queue = next;
drivers/infiniband/hw/hfi1/tid_rdma.c
3670
u8 next;
drivers/infiniband/hw/hfi1/tid_rdma.c
3710
next = qp->r_head_ack_queue + 1;
drivers/infiniband/hw/hfi1/tid_rdma.c
3711
if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device)))
drivers/infiniband/hw/hfi1/tid_rdma.c
3712
next = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
3714
if (unlikely(next == qp->s_acked_ack_queue)) {
drivers/infiniband/hw/hfi1/tid_rdma.c
3715
if (!qp->s_ack_queue[next].sent)
drivers/infiniband/hw/hfi1/tid_rdma.c
3717
update_ack_queue(qp, next);
drivers/infiniband/hw/hfi1/tid_rdma.c
3804
qp->r_head_ack_queue = next;
drivers/infiniband/hw/hfi1/tid_rdma.c
4280
u32 psn, next;
drivers/infiniband/hw/hfi1/tid_rdma.c
4368
for (next = priv->r_tid_tail + 1; ; next++) {
drivers/infiniband/hw/hfi1/tid_rdma.c
4369
if (next > rvt_size_atomic(&dev->rdi))
drivers/infiniband/hw/hfi1/tid_rdma.c
4370
next = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
4371
if (next == priv->r_tid_head)
drivers/infiniband/hw/hfi1/tid_rdma.c
4373
e = &qp->s_ack_queue[next];
drivers/infiniband/hw/hfi1/tid_rdma.c
4377
priv->r_tid_tail = next;
drivers/infiniband/hw/hfi1/tid_rdma.c
4946
u32 next;
drivers/infiniband/hw/hfi1/tid_rdma.c
4951
next = flow->flow_state.r_next_psn;
drivers/infiniband/hw/hfi1/tid_rdma.c
4952
flow->npkts = delta_psn(lpsn, next - 1);
drivers/infiniband/hw/hfi1/tid_rdma.c
5194
u32 hwords, next;
drivers/infiniband/hw/hfi1/tid_rdma.c
5267
next = qpriv->r_tid_ack + 1;
drivers/infiniband/hw/hfi1/tid_rdma.c
5268
if (next > rvt_size_atomic(&dev->rdi))
drivers/infiniband/hw/hfi1/tid_rdma.c
5269
next = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
5270
qpriv->r_tid_ack = next;
drivers/infiniband/hw/hfi1/tid_rdma.c
5271
if (qp->s_ack_queue[next].opcode != TID_OP(WRITE_REQ))
drivers/infiniband/hw/hfi1/tid_rdma.c
5273
nreq = ack_to_tid_req(&qp->s_ack_queue[next]);
drivers/infiniband/hw/hfi1/user_sdma.c
251
static u8 next;
drivers/infiniband/hw/hfi1/user_sdma.c
261
mapping[hash] = next;
drivers/infiniband/hw/hfi1/user_sdma.c
262
next = (next + 1) & 0x7F;
drivers/infiniband/hw/hns/hns_roce_cmd.c
121
cmd->free_head = context->next;
drivers/infiniband/hw/hns/hns_roce_cmd.c
230
hr_cmd->context[i].next = i + 1;
drivers/infiniband/hw/hns/hns_roce_cmd.c
233
hr_cmd->context[hr_cmd->max_cmds - 1].next = 0;
drivers/infiniband/hw/hns/hns_roce_device.h
488
u32 next; /* Next ID to allocate. */
drivers/infiniband/hw/hns/hns_roce_device.h
540
int next;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1011
.next = NULL,
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1148
for (nreq = 0; wr; ++nreq, wr = wr->next) {
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
719
for (nreq = 0; wr; ++nreq, wr = wr->next) {
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
847
for (nreq = 0; wr; ++nreq, wr = wr->next) {
drivers/infiniband/hw/hns/hns_roce_qp.c
1621
hr_dev->qp_table.bank[i].next = hr_dev->qp_table.bank[i].min;
drivers/infiniband/hw/hns/hns_roce_qp.c
227
id = ida_alloc_range(&bank->ida, bank->next, bank->max, GFP_KERNEL);
drivers/infiniband/hw/hns/hns_roce_qp.c
236
bank->next = (id + 1) > bank->max ? bank->min : id + 1;
drivers/infiniband/hw/ionic/ionic_controlpath.c
2093
qp->rq_meta[i].next = &qp->rq_meta[i + 1];
drivers/infiniband/hw/ionic/ionic_controlpath.c
2094
qp->rq_meta[i].next = IONIC_META_LAST;
drivers/infiniband/hw/ionic/ionic_controlpath.c
2513
qp->rq_meta[i].next = &qp->rq_meta[i + 1];
drivers/infiniband/hw/ionic/ionic_controlpath.c
2514
qp->rq_meta[i].next = IONIC_META_LAST;
drivers/infiniband/hw/ionic/ionic_datapath.c
1216
qp->rq_meta_head = meta->next;
drivers/infiniband/hw/ionic/ionic_datapath.c
1217
meta->next = IONIC_META_POSTED;
drivers/infiniband/hw/ionic/ionic_datapath.c
1263
wr = wr->next;
drivers/infiniband/hw/ionic/ionic_datapath.c
1336
wr = wr->next;
drivers/infiniband/hw/ionic/ionic_datapath.c
173
if (unlikely(meta->next != IONIC_META_POSTED)) {
drivers/infiniband/hw/ionic/ionic_datapath.c
180
meta->next = qp->rq_meta_head;
drivers/infiniband/hw/ionic/ionic_datapath.c
54
if (unlikely(meta->next != IONIC_META_POSTED)) {
drivers/infiniband/hw/ionic/ionic_datapath.c
69
meta->next = qp->rq_meta_head;
drivers/infiniband/hw/ionic/ionic_ibdev.h
239
struct ionic_rq_meta *next;
drivers/infiniband/hw/irdma/ctrl.c
27
entry = head->next;
drivers/infiniband/hw/irdma/ctrl.c
30
entry = lastentry->next;
drivers/infiniband/hw/irdma/ctrl.c
3912
list_entry(cqp->ooo_avail.next,
drivers/infiniband/hw/irdma/ctrl.c
4502
int next;
drivers/infiniband/hw/irdma/ctrl.c
4505
next = IRDMA_RING_GET_NEXT_TAIL(ceq->ceq_ring, 0);
drivers/infiniband/hw/irdma/ctrl.c
4508
ceqe = IRDMA_GET_CEQ_ELEM_AT_POS(ceq, next);
drivers/infiniband/hw/irdma/ctrl.c
4520
next = IRDMA_RING_GET_NEXT_TAIL(ceq->ceq_ring, i);
drivers/infiniband/hw/irdma/ctrl.c
4521
if (!next)
drivers/infiniband/hw/irdma/main.h
444
u32 *req_rsrc_num, u32 *next)
drivers/infiniband/hw/irdma/main.h
450
rsrc_num = find_next_zero_bit(rsrc_array, max_rsrc, *next);
drivers/infiniband/hw/irdma/main.h
462
*next = rsrc_num + 1;
drivers/infiniband/hw/irdma/main.h
463
if (*next == max_rsrc)
drivers/infiniband/hw/irdma/main.h
464
*next = 0;
drivers/infiniband/hw/irdma/pble.c
22
chunk = (struct irdma_chunk *) pinfo->clist.next;
drivers/infiniband/hw/irdma/puda.c
25
buf = (struct irdma_puda_buf *)list->next;
drivers/infiniband/hw/irdma/puda.c
949
nextbuf = buf->next;
drivers/infiniband/hw/irdma/puda.c
982
buf->next = nextbuf;
drivers/infiniband/hw/irdma/puda.h
59
struct irdma_puda_buf *next; /* for alloclist in rsrc struct */
drivers/infiniband/hw/irdma/utils.c
2136
struct list_head *chunk_entry = pprm->clist.next;
drivers/infiniband/hw/irdma/utils.c
2154
chunk_entry = pchunk->list.next;
drivers/infiniband/hw/irdma/utils.c
2294
u32 next = 1;
drivers/infiniband/hw/irdma/utils.c
2298
&node_id, &next))
drivers/infiniband/hw/irdma/utils.c
871
entry = list->next;
drivers/infiniband/hw/irdma/verbs.c
4135
ib_wr = ib_wr->next;
drivers/infiniband/hw/irdma/verbs.c
4183
ib_wr = ib_wr->next;
drivers/infiniband/hw/irdma/verbs.c
4231
ib_wr = ib_wr->next;
drivers/infiniband/hw/mana/wr.c
150
for (; wr; wr = wr->next) {
drivers/infiniband/hw/mana/wr.c
55
for (; wr; wr = wr->next) {
drivers/infiniband/hw/mlx4/alias_GUID.c
810
cb_ctx = list_entry(det->cb_list.next,
drivers/infiniband/hw/mlx4/mad.c
1330
recv_wr.next = NULL;
drivers/infiniband/hw/mlx4/mad.c
1440
wr.wr.next = NULL;
drivers/infiniband/hw/mlx4/mad.c
662
wr.wr.next = NULL;
drivers/infiniband/hw/mlx4/qp.c
3579
for (nreq = 0; wr; ++nreq, wr = wr->next) {
drivers/infiniband/hw/mlx4/qp.c
3816
if (wr->next)
drivers/infiniband/hw/mlx4/qp.c
3876
for (nreq = 0; wr; ++nreq, wr = wr->next) {
drivers/infiniband/hw/mlx4/qp.c
4461
.next = NULL,
drivers/infiniband/hw/mlx4/srq.c
153
next = get_wqe(srq, i);
drivers/infiniband/hw/mlx4/srq.c
154
next->next_wqe_index =
drivers/infiniband/hw/mlx4/srq.c
157
for (scatter = (void *) (next + 1);
drivers/infiniband/hw/mlx4/srq.c
158
(void *) scatter < (void *) next + desc_size;
drivers/infiniband/hw/mlx4/srq.c
295
struct mlx4_wqe_srq_next_seg *next;
drivers/infiniband/hw/mlx4/srq.c
300
next = get_wqe(srq, srq->tail);
drivers/infiniband/hw/mlx4/srq.c
301
next->next_wqe_index = cpu_to_be16(wqe_index);
drivers/infiniband/hw/mlx4/srq.c
311
struct mlx4_wqe_srq_next_seg *next;
drivers/infiniband/hw/mlx4/srq.c
326
for (nreq = 0; wr; ++nreq, wr = wr->next) {
drivers/infiniband/hw/mlx4/srq.c
341
next = get_wqe(srq, srq->head);
drivers/infiniband/hw/mlx4/srq.c
342
srq->head = be16_to_cpu(next->next_wqe_index);
drivers/infiniband/hw/mlx4/srq.c
343
scat = (struct mlx4_wqe_data_seg *) (next + 1);
drivers/infiniband/hw/mlx4/srq.c
80
struct mlx4_wqe_srq_next_seg *next;
drivers/infiniband/hw/mlx5/cq.c
354
tail = qp->sq.w_list[idx].next;
drivers/infiniband/hw/mlx5/cq.c
356
tail = qp->sq.w_list[idx].next;
drivers/infiniband/hw/mlx5/cq.c
422
wq->last_poll = wq->w_list[idx].next;
drivers/infiniband/hw/mlx5/cq.c
588
struct mlx5_ib_wc *soft_wc, *next;
drivers/infiniband/hw/mlx5/cq.c
591
list_for_each_entry_safe(soft_wc, next, &cq->wc_list, list) {
drivers/infiniband/hw/mlx5/devx.c
2774
event = list_entry(ev_queue->event_list.next,
drivers/infiniband/hw/mlx5/devx.c
2784
list_del(ev_queue->event_list.next);
drivers/infiniband/hw/mlx5/gsi.c
440
for (; wr; wr = wr->next) {
drivers/infiniband/hw/mlx5/gsi.c
443
cur_wr.wr.next = NULL;
drivers/infiniband/hw/mlx5/mlx5_ib.h
353
u16 next;
drivers/infiniband/hw/mlx5/mr.c
2087
struct mlx5_ib_mr *mr, *next;
drivers/infiniband/hw/mlx5/mr.c
2091
list_for_each_entry_safe(mr, next, &dev->data_direct_mr_list, dd_node) {
drivers/infiniband/hw/mlx5/odp.c
1106
frame->next = head;
drivers/infiniband/hw/mlx5/odp.c
1122
head = frame->next;
drivers/infiniband/hw/mlx5/odp.c
1139
head = frame->next;
drivers/infiniband/hw/mlx5/odp.c
952
struct pf_frame *next;
drivers/infiniband/hw/mlx5/qp.c
5785
.next = NULL,
drivers/infiniband/hw/mlx5/srq.c
110
struct mlx5_wqe_srq_next_seg *next;
drivers/infiniband/hw/mlx5/srq.c
133
next = get_wqe(srq, i);
drivers/infiniband/hw/mlx5/srq.c
134
next->next_wqe_index =
drivers/infiniband/hw/mlx5/srq.c
397
struct mlx5_wqe_srq_next_seg *next;
drivers/infiniband/hw/mlx5/srq.c
402
next = get_wqe(srq, srq->tail);
drivers/infiniband/hw/mlx5/srq.c
403
next->next_wqe_index = cpu_to_be16(wqe_index);
drivers/infiniband/hw/mlx5/srq.c
413
struct mlx5_wqe_srq_next_seg *next;
drivers/infiniband/hw/mlx5/srq.c
430
for (nreq = 0; wr; nreq++, wr = wr->next) {
drivers/infiniband/hw/mlx5/srq.c
445
next = get_wqe(srq, srq->head);
drivers/infiniband/hw/mlx5/srq.c
446
srq->head = be16_to_cpu(next->next_wqe_index);
drivers/infiniband/hw/mlx5/srq.c
447
scat = (struct mlx5_wqe_data_seg *)(next + 1);
drivers/infiniband/hw/mlx5/wr.c
1082
for (nreq = 0; wr; nreq++, wr = wr->next) {
drivers/infiniband/hw/mlx5/wr.c
1233
for (nreq = 0; wr; nreq++, wr = wr->next) {
drivers/infiniband/hw/mlx5/wr.c
775
qp->sq.w_list[idx].next = qp->sq.cur_post;
drivers/infiniband/hw/mthca/mthca_cmd.c
1727
enum ib_qp_state next, u32 num, int is_ee,
drivers/infiniband/hw/mthca/mthca_cmd.c
1774
if (op[cur][next] == CMD_ERR2RST_QPEE) {
drivers/infiniband/hw/mthca/mthca_cmd.c
1789
op[cur][next], CMD_TIME_CLASS_C);
drivers/infiniband/hw/mthca/mthca_cmd.c
181
int next;
drivers/infiniband/hw/mthca/mthca_cmd.c
1823
op_mod, op[cur][next], CMD_TIME_CLASS_C);
drivers/infiniband/hw/mthca/mthca_cmd.c
424
dev->cmd.free_head = context->next;
drivers/infiniband/hw/mthca/mthca_cmd.c
460
context->next = dev->cmd.free_head;
drivers/infiniband/hw/mthca/mthca_cmd.c
569
dev->cmd.context[i].next = i + 1;
drivers/infiniband/hw/mthca/mthca_cmd.c
572
dev->cmd.context[dev->cmd.max_cmds - 1].next = -1;
drivers/infiniband/hw/mthca/mthca_cmd.h
309
enum ib_qp_state next, u32 num, int is_ee,
drivers/infiniband/hw/mthca/mthca_memfree.h
104
NULL : list_entry(icm->chunk_list.next,
drivers/infiniband/hw/mthca/mthca_memfree.h
117
if (iter->chunk->list.next == &iter->icm->chunk_list) {
drivers/infiniband/hw/mthca/mthca_memfree.h
122
iter->chunk = list_entry(iter->chunk->list.next,
drivers/infiniband/hw/mthca/mthca_qp.c
1172
struct mthca_next_seg *next;
drivers/infiniband/hw/mthca/mthca_qp.c
1220
next = get_recv_wqe(qp, i);
drivers/infiniband/hw/mthca/mthca_qp.c
1221
next->nda_op = cpu_to_be32(((i + 1) & (qp->rq.max - 1)) <<
drivers/infiniband/hw/mthca/mthca_qp.c
1223
next->ee_nds = cpu_to_be32(size);
drivers/infiniband/hw/mthca/mthca_qp.c
1225
for (scatter = (void *) (next + 1);
drivers/infiniband/hw/mthca/mthca_qp.c
1226
(void *) scatter < (void *) next + (1 << qp->rq.wqe_shift);
drivers/infiniband/hw/mthca/mthca_qp.c
1232
next = get_send_wqe(qp, i);
drivers/infiniband/hw/mthca/mthca_qp.c
1233
next->nda_op = cpu_to_be32((((i + 1) & (qp->sq.max - 1)) <<
drivers/infiniband/hw/mthca/mthca_qp.c
1239
next = get_recv_wqe(qp, i);
drivers/infiniband/hw/mthca/mthca_qp.c
1240
next->nda_op = htonl((((i + 1) % qp->rq.max) <<
drivers/infiniband/hw/mthca/mthca_qp.c
1655
for (nreq = 0; wr; ++nreq, wr = wr->next) {
drivers/infiniband/hw/mthca/mthca_qp.c
1851
for (nreq = 0; wr; wr = wr->next) {
drivers/infiniband/hw/mthca/mthca_qp.c
1959
for (nreq = 0; wr; ++nreq, wr = wr->next) {
drivers/infiniband/hw/mthca/mthca_qp.c
2180
for (nreq = 0; wr; ++nreq, wr = wr->next) {
drivers/infiniband/hw/mthca/mthca_qp.c
2236
struct mthca_next_seg *next;
drivers/infiniband/hw/mthca/mthca_qp.c
2248
next = get_send_wqe(qp, index);
drivers/infiniband/hw/mthca/mthca_qp.c
2250
next = get_recv_wqe(qp, index);
drivers/infiniband/hw/mthca/mthca_qp.c
2252
*dbd = !!(next->ee_nds & cpu_to_be32(MTHCA_NEXT_DBD));
drivers/infiniband/hw/mthca/mthca_qp.c
2253
if (next->ee_nds & cpu_to_be32(0x3f))
drivers/infiniband/hw/mthca/mthca_qp.c
2254
*new_wqe = (next->nda_op & cpu_to_be32(~0x3f)) |
drivers/infiniband/hw/mthca/mthca_qp.c
2255
(next->ee_nds & cpu_to_be32(0x3f));
drivers/infiniband/hw/mthca/mthca_srq.c
183
struct mthca_next_seg *next;
drivers/infiniband/hw/mthca/mthca_srq.c
185
next = wqe = get_wqe(srq, i);
drivers/infiniband/hw/mthca/mthca_srq.c
189
next->nda_op = htonl(((i + 1) << srq->wqe_shift) | 1);
drivers/infiniband/hw/mthca/mthca_srq.c
192
next->nda_op = 0;
drivers/infiniband/hw/mthca/mthca_srq.c
502
for (nreq = 0; wr; wr = wr->next) {
drivers/infiniband/hw/mthca/mthca_srq.c
592
for (nreq = 0; wr; ++nreq, wr = wr->next) {
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2187
wr = wr->next;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2254
wr = wr->next;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2321
wr = wr->next;
drivers/infiniband/hw/qedr/qedr_roce_cm.c
595
if (wr->next) {
drivers/infiniband/hw/qedr/qedr_roce_cm.c
598
*bad_wr = wr->next;
drivers/infiniband/hw/qedr/qedr_roce_cm.c
656
wr = wr->next;
drivers/infiniband/hw/qedr/verbs.c
3797
wr = wr->next;
drivers/infiniband/hw/qedr/verbs.c
3905
wr = wr->next;
drivers/infiniband/hw/qedr/verbs.c
4002
wr = wr->next;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
851
wr = wr->next;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
942
wr = wr->next;
drivers/infiniband/sw/rdmavt/cq.c
34
u32 next;
drivers/infiniband/sw/rdmavt/cq.c
57
next = 0;
drivers/infiniband/sw/rdmavt/cq.c
59
next = head + 1;
drivers/infiniband/sw/rdmavt/cq.c
62
if (unlikely(next == tail || cq->cq_full)) {
drivers/infiniband/sw/rdmavt/cq.c
96
RDMA_WRITE_UAPI_ATOMIC(u_wc->head, next);
drivers/infiniband/sw/rdmavt/cq.c
99
k_wc->head = next;
drivers/infiniband/sw/rdmavt/mr.c
162
r = rkt->next;
drivers/infiniband/sw/rdmavt/mr.c
171
rkt->next = (r + 1) & (rkt->max - 1);
drivers/infiniband/sw/rdmavt/qp.c
1097
RCU_INIT_POINTER(qp->next, NULL);
drivers/infiniband/sw/rdmavt/qp.c
1383
qp->next = rdi->qp_dev->qp_table[n];
drivers/infiniband/sw/rdmavt/qp.c
1794
for (; wr; wr = wr->next) {
drivers/infiniband/sw/rdmavt/qp.c
1796
u32 next;
drivers/infiniband/sw/rdmavt/qp.c
1805
next = wq->head + 1;
drivers/infiniband/sw/rdmavt/qp.c
1806
if (next >= qp->r_rq.size)
drivers/infiniband/sw/rdmavt/qp.c
1807
next = 0;
drivers/infiniband/sw/rdmavt/qp.c
1808
if (next == READ_ONCE(wq->tail)) {
drivers/infiniband/sw/rdmavt/qp.c
1835
smp_store_release(&wq->head, next);
drivers/infiniband/sw/rdmavt/qp.c
1963
u32 next;
drivers/infiniband/sw/rdmavt/qp.c
2030
next = qp->s_head + 1;
drivers/infiniband/sw/rdmavt/qp.c
2031
if (next >= qp->s_size)
drivers/infiniband/sw/rdmavt/qp.c
2032
next = 0;
drivers/infiniband/sw/rdmavt/qp.c
2114
qp->s_head = next;
drivers/infiniband/sw/rdmavt/qp.c
2167
call_send = qp->s_head == READ_ONCE(qp->s_last) && !wr->next;
drivers/infiniband/sw/rdmavt/qp.c
2169
for (; wr; wr = wr->next) {
drivers/infiniband/sw/rdmavt/qp.c
2209
for (; wr; wr = wr->next) {
drivers/infiniband/sw/rdmavt/qp.c
2211
u32 next;
drivers/infiniband/sw/rdmavt/qp.c
2221
next = wq->head + 1;
drivers/infiniband/sw/rdmavt/qp.c
2222
if (next >= srq->rq.size)
drivers/infiniband/sw/rdmavt/qp.c
2223
next = 0;
drivers/infiniband/sw/rdmavt/qp.c
2224
if (next == READ_ONCE(wq->tail)) {
drivers/infiniband/sw/rdmavt/qp.c
2239
smp_store_release(&wq->head, next);
drivers/infiniband/sw/rdmavt/qp.c
2704
qp = rcu_dereference(pqp->next);
drivers/infiniband/sw/rdmavt/qp.c
751
qpp = &q->next) {
drivers/infiniband/sw/rdmavt/qp.c
754
rcu_dereference_protected(qp->next,
drivers/infiniband/sw/rxe/rxe_icrc.c
21
static __be32 rxe_crc32(struct rxe_dev *rxe, __be32 crc, void *next, size_t len)
drivers/infiniband/sw/rxe/rxe_icrc.c
23
return (__force __be32)crc32_le((__force u32)crc, next, len);
drivers/infiniband/sw/rxe/rxe_pool.c
142
&pool->next, gfp_flags);
drivers/infiniband/sw/rxe/rxe_pool.h
44
u32 next;
drivers/infiniband/sw/rxe/rxe_recv.c
233
if (mca->qp_list.next != &mcg->qp_list) {
drivers/infiniband/sw/rxe/rxe_verbs.c
1061
wr = wr->next;
drivers/infiniband/sw/rxe/rxe_verbs.c
522
wr = wr->next;
drivers/infiniband/sw/rxe/rxe_verbs.c
923
ibwr = ibwr->next;
drivers/infiniband/sw/siw/siw_cm.c
324
work = list_entry(cep->work_freelist.next, struct siw_cm_work,
drivers/infiniband/sw/siw/siw_mem.c
63
u32 id, next;
drivers/infiniband/sw/siw/siw_mem.c
77
get_random_bytes(&next, 4);
drivers/infiniband/sw/siw/siw_mem.c
78
next &= SIW_STAG_MAX_INDEX;
drivers/infiniband/sw/siw/siw_mem.c
80
if (xa_alloc_cyclic(&sdev->mem_xa, &id, mem, limit, &next,
drivers/infiniband/sw/siw/siw_qp_tx.c
1270
qp->tx_list.next = NULL;
drivers/infiniband/sw/siw/siw_qp_tx.c
1278
qp->tx_list.next = NULL;
drivers/infiniband/sw/siw/siw_verbs.c
1098
wr = wr->next;
drivers/infiniband/sw/siw/siw_verbs.c
1816
wr = wr->next;
drivers/infiniband/sw/siw/siw_verbs.c
731
wr = wr->next;
drivers/infiniband/sw/siw/siw_verbs.c
751
wr = wr->next;
drivers/infiniband/sw/siw/siw_verbs.c
951
wr = wr->next;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1355
p = list_entry(priv->cm.start_list.next, typeof(*p), list);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1409
p = list_entry(priv->cm.reap_list.next, typeof(*p), list);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
225
p = list_entry(priv->cm.rx_flush_list.next, typeof(*p), list);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
342
wr->next = NULL;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
942
p = list_entry(priv->cm.passive_ids.next, typeof(*p), list);
drivers/infiniband/ulp/ipoib/ipoib_fs.c
129
.next = ipoib_mcg_seq_next,
drivers/infiniband/ulp/ipoib/ipoib_fs.c
213
.next = ipoib_path_seq_next,
drivers/infiniband/ulp/ipoib/ipoib_verbs.c
233
priv->rx_wr.next = NULL;
drivers/infiniband/ulp/iser/iscsi_iser.c
211
tx_desc->inv_wr.next = NULL;
drivers/infiniband/ulp/iser/iscsi_iser.c
212
tx_desc->reg_wr.wr.next = NULL;
drivers/infiniband/ulp/iser/iser_memory.c
244
inv_wr->next = next_wr;
drivers/infiniband/ulp/iser/iser_memory.c
281
wr->wr.next = &tx_desc->send_wr;
drivers/infiniband/ulp/iser/iser_memory.c
328
wr->wr.next = &tx_desc->send_wr;
drivers/infiniband/ulp/iser/iser_verbs.c
821
wr.next = NULL;
drivers/infiniband/ulp/iser/iser_verbs.c
840
wr.next = NULL;
drivers/infiniband/ulp/iser/iser_verbs.c
867
wr->next = NULL;
drivers/infiniband/ulp/iser/iser_verbs.c
874
if (tx_desc->inv_wr.next)
drivers/infiniband/ulp/iser/iser_verbs.c
876
else if (tx_desc->reg_wr.wr.next)
drivers/infiniband/ulp/isert/ib_isert.c
737
rx_wr->next = rx_wr + 1;
drivers/infiniband/ulp/isert/ib_isert.c
741
rx_wr->next = NULL; /* mark end of work requests list */
drivers/infiniband/ulp/isert/ib_isert.c
768
rx_wr.next = NULL;
drivers/infiniband/ulp/isert/ib_isert.c
789
send_wr.next = NULL;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2243
struct rtrs_clt_path *next;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2289
next = rtrs_clt_get_next_path_or_null(&clt->paths_list, clt_path);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2315
next))
drivers/infiniband/ulp/rtrs/rtrs-clt.c
595
wr->next = &wr_arr[i - 1];
drivers/infiniband/ulp/rtrs/rtrs-clt.c
766
READ_ONCE((&clt_path->s.entry)->next),
drivers/infiniband/ulp/rtrs/rtrs-srv.c
1224
id = list_entry(con->rsp_wr_wait_list.next,
drivers/infiniband/ulp/rtrs/rtrs-srv.c
247
wr->wr.next = &rwr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
248
rwr.wr.next = &inv_wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
249
inv_wr.next = &imm_wr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
251
wr->wr.next = &rwr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
252
rwr.wr.next = &imm_wr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
254
wr->wr.next = &inv_wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
255
inv_wr.next = &imm_wr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
257
wr->wr.next = &imm_wr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
275
imm_wr.wr.next = NULL;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
377
inv_wr.next = &rwr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
378
rwr.wr.next = &imm_wr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
381
rwr.wr.next = &imm_wr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
384
inv_wr.next = &imm_wr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
395
imm_wr.wr.next = NULL;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
400
rwr.wr.next = &imm_wr.wr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
865
rwr[mri].wr.next = mri ? &rwr[mri - 1].wr : NULL;
drivers/infiniband/ulp/rtrs/rtrs.c
113
struct ib_send_wr *next = head;
drivers/infiniband/ulp/rtrs/rtrs.c
115
while (next->next)
drivers/infiniband/ulp/rtrs/rtrs.c
116
next = next->next;
drivers/infiniband/ulp/rtrs/rtrs.c
117
next->next = wr;
drivers/infiniband/ulp/rtrs/rtrs.c
123
wr->next = tail;
drivers/infiniband/ulp/srp/ib_srp.c
1166
.next = NULL,
drivers/infiniband/ulp/srp/ib_srp.c
1433
if (state->fr.next >= state->fr.end) {
drivers/infiniband/ulp/srp/ib_srp.c
1474
wr.wr.next = NULL;
drivers/infiniband/ulp/srp/ib_srp.c
1485
*state->fr.next++ = desc;
drivers/infiniband/ulp/srp/ib_srp.c
1506
state->fr.next = req->fr_list;
drivers/infiniband/ulp/srp/ib_srp.c
1564
state.gen.next = next_mr;
drivers/infiniband/ulp/srp/ib_srp.c
1770
ret = srp_map_idb(ch, req, state.gen.next, state.gen.end,
drivers/infiniband/ulp/srp/ib_srp.c
1904
wr.next = NULL;
drivers/infiniband/ulp/srp/ib_srp.c
1926
wr.next = NULL;
drivers/infiniband/ulp/srp/ib_srp.h
329
struct srp_fr_desc **next;
drivers/infiniband/ulp/srp/ib_srp.h
333
void **next;
drivers/infiniband/ulp/srpt/ib_srpt.c
2938
send_wr.next = NULL;
drivers/infiniband/ulp/srpt/ib_srpt.c
901
wr.next = NULL;
drivers/infiniband/ulp/srpt/ib_srpt.c
924
.next = NULL,
drivers/input/gameport/gameport.c
295
struct gameport_event *e, *next;
drivers/input/gameport/gameport.c
300
list_for_each_entry_safe(e, next, &gameport_event_list, node) {
drivers/input/gameport/gameport.c
410
struct gameport_event *event, *next;
drivers/input/gameport/gameport.c
415
list_for_each_entry_safe(event, next, &gameport_event_list, node) {
drivers/input/input.c
1136
.next = input_devices_seq_next,
drivers/input/input.c
1197
.next = input_handlers_seq_next,
drivers/input/input.c
2183
struct input_handle *handle, *next;
drivers/input/input.c
2188
list_for_each_entry_safe(handle, next, &dev->h_list, d_node)
drivers/input/input.c
2485
struct input_handle *handle, *next;
drivers/input/input.c
2489
list_for_each_entry_safe(handle, next, &handler->h_list, h_node)
drivers/input/serio/i8042-acpipnpio.h
1487
id = id->next;
drivers/input/serio/serio.c
166
struct serio_event *e, *next;
drivers/input/serio/serio.c
170
list_for_each_entry_safe(e, next, &serio_event_list, node) {
drivers/input/serio/serio.c
276
struct serio_event *event, *next;
drivers/input/serio/serio.c
280
list_for_each_entry_safe(event, next, &serio_event_list, node) {
drivers/input/serio/serio.c
600
s = list_entry(s->child_node.next,
drivers/input/serio/serio.c
689
struct serio *s, *next;
drivers/input/serio/serio.c
693
list_for_each_entry_safe(s, next, &serio->children, child_node) {
drivers/input/touchscreen/sur40.c
547
new_buf = list_entry(sur40->buf_list.next, struct sur40_buffer, list);
drivers/interconnect/core.c
306
struct icc_node *next, *prev = NULL;
drivers/interconnect/core.c
312
next = path->reqs[i].node;
drivers/interconnect/core.c
313
p = next->provider;
drivers/interconnect/core.c
317
prev = next;
drivers/interconnect/core.c
322
ret = p->set(prev, next);
drivers/interconnect/core.c
326
prev = next;
drivers/iommu/amd/amd_iommu_types.h
450
#define for_each_pci_segment_safe(pci_seg, next) \
drivers/iommu/amd/amd_iommu_types.h
451
list_for_each_entry_safe((pci_seg), (next), &amd_iommu_pci_seg_list, list)
drivers/iommu/amd/amd_iommu_types.h
457
#define for_each_iommu_safe(iommu, next) \
drivers/iommu/amd/amd_iommu_types.h
458
list_for_each_entry_safe((iommu), (next), &amd_iommu_list, list)
drivers/iommu/amd/amd_iommu_types.h
462
#define for_each_pdom_dev_data_safe(pdom_dev_data, next, pdom) \
drivers/iommu/amd/amd_iommu_types.h
463
list_for_each_entry_safe((pdom_dev_data), (next), &pdom->dev_data_list, list)
drivers/iommu/amd/init.c
1770
struct amd_iommu_pci_seg *pci_seg, *next;
drivers/iommu/amd/init.c
1772
for_each_pci_segment_safe(pci_seg, next) {
drivers/iommu/amd/init.c
1802
struct amd_iommu *iommu, *next;
drivers/iommu/amd/init.c
1804
for_each_iommu_safe(iommu, next) {
drivers/iommu/amd/init.c
2619
struct unity_map_entry *entry, *next;
drivers/iommu/amd/init.c
2623
list_for_each_entry_safe(entry, next, &pci_seg->unity_map, list) {
drivers/iommu/amd/pasid.c
82
struct pdom_dev_data *pdom_dev_data, *next;
drivers/iommu/amd/pasid.c
91
for_each_pdom_dev_data_safe(pdom_dev_data, next, sva_pdom)
drivers/iommu/dma-iommu.c
542
if (window->node.next == &bridge->dma_ranges &&
drivers/iommu/exynos-iommu.c
1012
list_for_each_entry_safe(data, next, &domain->clients, domain_node) {
drivers/iommu/exynos-iommu.c
949
struct sysmmu_drvdata *data, *next;
drivers/iommu/exynos-iommu.c
957
list_for_each_entry_safe(data, next, &domain->clients, domain_node) {
drivers/iommu/exynos-iommu.c
993
struct sysmmu_drvdata *data, *next;
drivers/iommu/intel/dmar.c
586
struct acpi_dmar_header *iter, *next;
drivers/iommu/intel/dmar.c
589
for (iter = start; iter < end; iter = next) {
drivers/iommu/intel/dmar.c
590
next = (void *)iter + iter->length;
drivers/iommu/intel/dmar.c
595
} else if (next > end) {
drivers/iommu/io-pgfault.c
105
list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) {
drivers/iommu/io-pgfault.c
357
struct iopf_fault *iopf, *next;
drivers/iommu/io-pgfault.c
366
list_for_each_entry_safe(iopf, next, &iopf_param->partial,
drivers/iommu/io-pgfault.c
44
struct iopf_fault *iopf, *next;
drivers/iommu/io-pgfault.c
453
struct iopf_fault *next;
drivers/iommu/io-pgfault.c
46
list_for_each_entry_safe(iopf, next, &group->faults, list) {
drivers/iommu/io-pgfault.c
468
list_for_each_entry_safe(partial_iopf, next, &fault_param->partial, list)
drivers/iommu/io-pgfault.c
538
struct iommu_fault_param *iopf_param, *next;
drivers/iommu/io-pgfault.c
543
list_for_each_entry_safe(iopf_param, next, &queue->devices, queue_list)
drivers/iommu/io-pgfault.c
85
struct iopf_fault *iopf, *next;
drivers/iommu/iommu-sva.c
118
list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) {
drivers/iommu/iommu-sva.c
145
list_add(&domain->next, &iommu_mm->sva_domains);
drivers/iommu/iommu-sva.c
184
list_del(&domain->next);
drivers/iommu/iommu.c
1935
struct iommu_group *group, *next;
drivers/iommu/iommu.c
1943
list_for_each_entry_safe(group, next, &group_list, entry) {
drivers/iommu/iommu.c
2814
goto next;
drivers/iommu/iommu.c
2823
next:
drivers/iommu/iommu.c
2934
struct iommu_resv_region *entry, *next;
drivers/iommu/iommu.c
2936
list_for_each_entry_safe(entry, next, list, list) {
drivers/iommu/iommu.c
959
struct iommu_resv_region *region, *next;
drivers/iommu/iommu.c
966
list_for_each_entry_safe(region, next, &group_resv_regions, list) {
drivers/iommu/iommufd/eventq.c
227
struct iommufd_vevent *cur, *next;
drivers/iommu/iommufd/eventq.c
23
struct iopf_group *group, *next;
drivers/iommu/iommufd/eventq.c
231
list_for_each_entry_safe(cur, next, &eventq->deliver, node) {
drivers/iommu/iommufd/eventq.c
260
struct iommufd_vevent *next;
drivers/iommu/iommufd/eventq.c
262
next = list_first_entry(list, struct iommufd_vevent, node);
drivers/iommu/iommufd/eventq.c
264
if (next == &veventq->lost_events_header) {
drivers/iommu/iommufd/eventq.c
269
list_del(&next->node);
drivers/iommu/iommufd/eventq.c
271
memcpy(vevent, next, sizeof(*vevent));
drivers/iommu/iommufd/eventq.c
273
vevent = next;
drivers/iommu/iommufd/eventq.c
33
list_for_each_entry_safe(group, next, &fault->common.deliver, node) {
drivers/iommu/iommufd/eventq.c
40
list_for_each_entry_safe(group, next, &free_list, node) {
drivers/iommu/iommufd/eventq.c
61
struct iopf_group *group, *next;
drivers/iommu/iommufd/eventq.c
70
list_for_each_entry_safe(group, next, &fault->common.deliver, node) {
drivers/iommu/iommufd/io_pagetable.c
266
list_for_each_entry(elm, pages_list, next) {
drivers/iommu/iommufd/io_pagetable.c
28
struct list_head next;
drivers/iommu/iommufd/io_pagetable.c
281
next);
drivers/iommu/iommufd/io_pagetable.c
312
list_for_each_entry(elm, pages_list, next) {
drivers/iommu/iommufd/io_pagetable.c
342
struct iopt_pages_list, next))) {
drivers/iommu/iommufd/io_pagetable.c
347
list_del(&elm->next);
drivers/iommu/iommufd/io_pagetable.c
358
list_for_each_entry(elm, pages_list, next) {
drivers/iommu/iommufd/io_pagetable.c
366
list_for_each_entry(undo_elm, pages_list, next) {
drivers/iommu/iommufd/io_pagetable.c
392
list_for_each_entry(elm, pages_list, next) {
drivers/iommu/iommufd/io_pagetable.c
423
list_add(&elm.next, &pages_list);
drivers/iommu/iommufd/io_pagetable.c
727
list_add_tail(&elm->next, pages_list);
drivers/iommu/iommufd/io_pagetable.c
903
struct iopt_reserved *reserved, *next;
drivers/iommu/iommufd/io_pagetable.c
908
reserved = next) {
drivers/iommu/iommufd/io_pagetable.c
909
next = iopt_reserved_iter_next(reserved, 0, ULONG_MAX);
drivers/iommu/iommufd/iommufd_private.h
682
struct iommufd_veventq *veventq, *next;
drivers/iommu/iommufd/iommufd_private.h
686
list_for_each_entry_safe(veventq, next, &viommu->veventqs, node) {
drivers/iommu/iommufd/main.c
328
struct iommufd_sw_msi_map *next;
drivers/iommu/iommufd/main.c
382
list_for_each_entry_safe(cur, next, &ictx->sw_msi_list, sw_msi_item)
drivers/iommu/iommufd/vfio_compat.c
484
put_user(total_cap_size, &last_cap->next)) {
drivers/iommu/iova.c
101
struct rb_node *node, *next;
drivers/iommu/iova.c
124
next = node->rb_left;
drivers/iommu/iova.c
125
while (next->rb_right) {
drivers/iommu/iova.c
126
next = next->rb_right;
drivers/iommu/iova.c
127
if (to_iova(next)->pfn_lo >= limit_pfn) {
drivers/iommu/iova.c
128
node = next;
drivers/iommu/iova.c
573
struct iova_magazine *next;
drivers/iommu/iova.c
681
kmemleak_transient_leak(mag->next);
drivers/iommu/iova.c
682
rcache->depot = mag->next;
drivers/iommu/iova.c
690
mag->next = rcache->depot;
drivers/iommu/virtio-iommu.c
367
struct interval_tree_node *node, *next;
drivers/iommu/virtio-iommu.c
370
next = interval_tree_iter_first(&vdomain->mappings, iova, end);
drivers/iommu/virtio-iommu.c
371
while (next) {
drivers/iommu/virtio-iommu.c
372
node = next;
drivers/iommu/virtio-iommu.c
374
next = interval_tree_iter_next(node, iova, end);
drivers/iommu/virtio-iommu.c
489
struct iommu_resv_region *region = NULL, *next;
drivers/iommu/virtio-iommu.c
523
list_for_each_entry(next, &vdev->resv_regions, list) {
drivers/iommu/virtio-iommu.c
524
if (next->start > region->start)
drivers/iommu/virtio-iommu.c
527
list_add_tail(&region->list, &next->list);
drivers/isdn/capi/kcapi_proc.c
106
.next = controller_next,
drivers/isdn/capi/kcapi_proc.c
181
.next = applications_next,
drivers/isdn/capi/kcapi_proc.c
188
.next = applications_next,
drivers/isdn/capi/kcapi_proc.c
99
.next = controller_next,
drivers/isdn/hardware/mISDN/hfcmulti.c
1087
struct hfc_multi *pos, *next, *plx_last_hc;
drivers/isdn/hardware/mISDN/hfcmulti.c
1189
list_for_each_entry_safe(pos, next, &HFClist, list) {
drivers/isdn/hardware/mISDN/hfcmulti.c
5440
struct hfc_multi *card, *next;
drivers/isdn/hardware/mISDN/hfcmulti.c
5443
list_for_each_entry_safe(card, next, &HFClist, list)
drivers/isdn/hardware/mISDN/hfcmulti.c
884
struct hfc_multi *hc, *next, *pcmmaster = NULL;
drivers/isdn/hardware/mISDN/hfcmulti.c
901
list_for_each_entry_safe(hc, next, &HFClist, list) {
drivers/isdn/hardware/mISDN/hfcmulti.c
912
list_for_each_entry_safe(hc, next, &HFClist, list) {
drivers/isdn/hardware/mISDN/hfcsusb.c
2133
struct hfcsusb *next;
drivers/isdn/hardware/mISDN/hfcsusb.c
2141
list_for_each_entry_safe(hw, next, &HFClist, list)
drivers/isdn/mISDN/dsp_cmx.c
1110
member = list_entry(conf->mlist.next, struct dsp_conf_member,
drivers/isdn/mISDN/dsp_cmx.c
1449
other = (list_entry(conf->mlist.next,
drivers/isdn/mISDN/dsp_cmx.c
1694
member = list_entry(conf->mlist.next,
drivers/isdn/mISDN/dsp_cmx.c
502
member = list_entry(conf->mlist.next, struct dsp_conf_member, list);
drivers/isdn/mISDN/dsp_cmx.c
661
member = list_entry(conf->mlist.next, struct dsp_conf_member,
drivers/isdn/mISDN/dsp_cmx.c
675
member = list_entry(conf->mlist.next, struct dsp_conf_member,
drivers/isdn/mISDN/dsp_cmx.c
677
nextm = list_entry(member->list.next, struct dsp_conf_member,
drivers/isdn/mISDN/dsp_hwec.c
54
char *dup, *next, *tok, *name, *val;
drivers/isdn/mISDN/dsp_hwec.c
57
dup = next = kstrdup(arg, GFP_ATOMIC);
drivers/isdn/mISDN/dsp_hwec.c
61
while ((tok = strsep(&next, ","))) {
drivers/isdn/mISDN/dsp_pipeline.c
200
char *dup, *next, *tok, *name, *args;
drivers/isdn/mISDN/dsp_pipeline.c
211
dup = next = kstrdup(cfg, GFP_ATOMIC);
drivers/isdn/mISDN/dsp_pipeline.c
214
while ((tok = strsep(&next, "|"))) {
drivers/isdn/mISDN/l1oip_core.c
1274
struct l1oip *hc, *next;
drivers/isdn/mISDN/l1oip_core.c
1276
list_for_each_entry_safe(hc, next, &l1oip_ilist, list)
drivers/isdn/mISDN/socket.c
283
struct mISDNchannel *bchan, *next;
drivers/isdn/mISDN/socket.c
297
list_for_each_entry_safe(bchan, next,
drivers/isdn/mISDN/timerdev.c
68
struct mISDNtimer *timer, *next;
drivers/isdn/mISDN/timerdev.c
85
list_for_each_entry_safe(timer, next, &dev->expired, list) {
drivers/leds/trigger/ledtrig-pattern.c
115
if (data->curr->brightness == data->next->brightness) {
drivers/leds/trigger/ledtrig-pattern.c
120
if (!data->next->delta_t) {
drivers/leds/trigger/ledtrig-pattern.c
188
data->next = data->patterns + 1;
drivers/leds/trigger/ledtrig-pattern.c
35
struct led_pattern *next;
drivers/leds/trigger/ledtrig-pattern.c
49
data->curr = data->next;
drivers/leds/trigger/ledtrig-pattern.c
53
if (data->next == data->patterns + data->npatterns - 1)
drivers/leds/trigger/ledtrig-pattern.c
54
data->next = data->patterns;
drivers/leds/trigger/ledtrig-pattern.c
56
data->next++;
drivers/leds/trigger/ledtrig-pattern.c
73
step_brightness = abs(data->next->brightness - data->curr->brightness);
drivers/leds/trigger/ledtrig-pattern.c
76
if (data->next->brightness > data->curr->brightness)
drivers/macintosh/adb-iop.c
226
req->next = NULL;
drivers/macintosh/adb-iop.c
234
last_req->next = req;
drivers/macintosh/adb-iop.c
68
current_req = req->next;
drivers/macintosh/adb.c
637
ap = &(*ap)->next;
drivers/macintosh/adb.c
638
req->next = NULL;
drivers/macintosh/adb.c
738
state->completed = req->next;
drivers/macintosh/macio-adb.c
181
req->next = NULL;
drivers/macintosh/macio-adb.c
188
last_req->next = req;
drivers/macintosh/macio-adb.c
228
current_req = req->next;
drivers/macintosh/macio-adb.c
248
current_req = req->next;
drivers/macintosh/smu.c
120
cmd = list_entry(smu->cmd_list.next, struct smu_cmd, link);
drivers/macintosh/smu.c
733
newcmd = list_entry(smu->cmd_i2c_list.next,
drivers/macintosh/smu.c
984
prop->next = NULL;
drivers/macintosh/via-cuda.c
511
req->next = NULL;
drivers/macintosh/via-cuda.c
518
last_req->next = req;
drivers/macintosh/via-cuda.c
642
current_req = req->next;
drivers/macintosh/via-cuda.c
696
current_req = req->next;
drivers/macintosh/via-macii.c
221
req.next = current_req;
drivers/macintosh/via-macii.c
257
req->next = NULL;
drivers/macintosh/via-macii.c
265
last_req->next = req;
drivers/macintosh/via-macii.c
471
current_req = req->next;
drivers/macintosh/via-macii.c
478
current_req = req->next;
drivers/macintosh/via-macii.c
518
current_req = req->next;
drivers/macintosh/via-pmu.c
1161
req->next = NULL;
drivers/macintosh/via-pmu.c
1167
last_req->next = req;
drivers/macintosh/via-pmu.c
1371
next:
drivers/macintosh/via-pmu.c
1463
goto next;
drivers/macintosh/via-pmu.c
1504
current_req = req->next;
drivers/macintosh/via-pmu.c
1554
current_req = req->next;
drivers/macintosh/via-pmu.c
2126
for (list = &all_pmu_pvt; (list = list->next) != &all_pmu_pvt; ) {
drivers/macintosh/windfarm_core.c
112
next += HZ;
drivers/macintosh/windfarm_core.c
115
delay = next - jiffies;
drivers/macintosh/windfarm_core.c
87
unsigned long next, delay;
drivers/macintosh/windfarm_core.c
89
next = jiffies;
drivers/macintosh/windfarm_core.c
97
if (time_after_eq(jiffies, next)) {
drivers/macintosh/windfarm_smu_controls.c
313
fct = list_entry(smu_fans.next, struct smu_fan_control, link);
drivers/macintosh/windfarm_smu_sensors.c
467
ads = list_entry(smu_ads.next, struct smu_ad_sensor, link);
drivers/mailbox/bcm-flexrm-mailbox.c
1041
next = flexrm_write_descs(msg, nhcnt, reqid,
drivers/mailbox/bcm-flexrm-mailbox.c
1045
if (IS_ERR(next)) {
drivers/mailbox/bcm-flexrm-mailbox.c
1046
ret = PTR_ERR(next);
drivers/mailbox/bcm-flexrm-mailbox.c
1052
ring->bd_write_offset = (unsigned long)(next - ring->bd_base);
drivers/mailbox/bcm-flexrm-mailbox.c
978
void *next;
drivers/md/bcache/alloc.c
87
unsigned long next = c->nbuckets * c->cache->sb.bucket_size / 1024;
drivers/md/bcache/alloc.c
97
} while (atomic_cmpxchg(&c->rescale, r, r + next) != r);
drivers/md/bcache/bset.c
113
struct bkey *k = iter->data->k, *next = bkey_next(k);
drivers/md/bcache/bset.c
115
if (next < iter->data->end &&
drivers/md/bcache/bset.c
117
&START_KEY(next) : next) > 0) {
drivers/md/bcache/bset.c
23
struct bkey *k, *next;
drivers/md/bcache/bset.c
25
for (k = i->start; k < bset_bkey_last(i); k = next) {
drivers/md/bcache/bset.c
26
next = bkey_next(k);
drivers/md/bcache/bset.c
36
if (next < bset_bkey_last(i) &&
drivers/md/bcache/bset.c
38
&START_KEY(next) : next) > 0)
drivers/md/bcache/sysfs.c
239
next_io = wb ? div64_s64(dc->writeback_rate.next-local_clock(),
drivers/md/bcache/util.c
205
d->next += div_u64(done * NSEC_PER_SEC, atomic_long_read(&d->rate));
drivers/md/bcache/util.c
213
if (time_before64(now + NSEC_PER_SEC * 5LLU / 2LLU, d->next))
drivers/md/bcache/util.c
214
d->next = now + NSEC_PER_SEC * 5LLU / 2LLU;
drivers/md/bcache/util.c
216
if (time_after64(now - NSEC_PER_SEC * 2, d->next))
drivers/md/bcache/util.c
217
d->next = now - NSEC_PER_SEC * 2;
drivers/md/bcache/util.c
219
return time_after64(d->next, now)
drivers/md/bcache/util.c
220
? div_u64(d->next - now, NSEC_PER_SEC / HZ)
drivers/md/bcache/util.h
419
uint64_t next;
drivers/md/bcache/util.h
430
d->next = local_clock();
drivers/md/bcache/writeback.c
131
dc->writeback_rate.next + NSEC_PER_MSEC))) {
drivers/md/bcache/writeback.c
477
struct keybuf_key *next, *keys[MAX_WRITEBACKS_IN_PASS], *w;
drivers/md/bcache/writeback.c
493
next = bch_keybuf_next(&dc->writeback_keys);
drivers/md/bcache/writeback.c
497
next) {
drivers/md/bcache/writeback.c
502
BUG_ON(ptr_stale(dc->disk.c, &next->key, 0));
drivers/md/bcache/writeback.c
528
&START_KEY(&next->key)))
drivers/md/bcache/writeback.c
531
size += KEY_SIZE(&next->key);
drivers/md/bcache/writeback.c
532
keys[nk++] = next;
drivers/md/bcache/writeback.c
533
} while ((next = bch_keybuf_next(&dc->writeback_keys)));
drivers/md/bcache/writeback.c
670
goto next;
drivers/md/bcache/writeback.c
687
next:
drivers/md/dm-bufio.c
1479
list_entry(write_list->next, struct dm_buffer, write_list);
drivers/md/dm-bufio.c
1640
b = list_to_buffer(c->reserved_buffers.next);
drivers/md/dm-bufio.c
175
it->e = to_le(it->e->list.next);
drivers/md/dm-bufio.c
195
it->e = to_le(it->e->list.next);
drivers/md/dm-bufio.c
221
lru->cursor = lru->cursor->next;
drivers/md/dm-bufio.c
2615
struct dm_buffer *b = list_to_buffer(c->reserved_buffers.next);
drivers/md/dm-bufio.c
2657
struct dm_buffer *b = list_to_buffer(c->reserved_buffers.next);
drivers/md/dm-bufio.c
2709
h = dm_bufio_all_clients.next;
drivers/md/dm-bufio.c
2721
struct list_head *h = dm_bufio_all_clients.next;
drivers/md/dm-bufio.c
2727
h = h->next;
drivers/md/dm-cache-policy-smq.c
135
return to_entry(es, e->next);
drivers/md/dm-cache-policy-smq.c
152
e->next = l->head;
drivers/md/dm-cache-policy-smq.c
168
e->next = INDEXER_NULL;
drivers/md/dm-cache-policy-smq.c
172
tail->next = l->tail = to_index(es, e);
drivers/md/dm-cache-policy-smq.c
190
e->next = to_index(es, old);
drivers/md/dm-cache-policy-smq.c
191
prev->next = old->prev = to_index(es, e);
drivers/md/dm-cache-policy-smq.c
201
struct entry *next = l_next(es, e);
drivers/md/dm-cache-policy-smq.c
204
prev->next = e->next;
drivers/md/dm-cache-policy-smq.c
206
l->head = e->next;
drivers/md/dm-cache-policy-smq.c
208
if (next)
drivers/md/dm-cache-policy-smq.c
209
next->prev = e->prev;
drivers/md/dm-cache-policy-smq.c
42
unsigned int next:28;
drivers/md/dm-cache-policy-smq.c
722
e->next = INDEXER_NULL;
drivers/md/dm-core.h
290
struct dm_io *next;
drivers/md/dm-delay.c
88
struct dm_delay_info *delayed, *next;
drivers/md/dm-delay.c
99
list_for_each_entry_safe(delayed, next, &local_list, list) {
drivers/md/dm-init.c
118
char *next;
drivers/md/dm-init.c
128
next = str_field_delimit(&field[i], ',');
drivers/md/dm-init.c
153
return next;
drivers/md/dm-init.c
197
char *next;
drivers/md/dm-init.c
207
next = str_field_delimit(&field[i], ';');
drivers/md/dm-init.c
230
return next;
drivers/md/dm-integrity.c
1363
struct rb_node *next;
drivers/md/dm-integrity.c
1373
next = rb_next(&node->node);
drivers/md/dm-integrity.c
1374
if (unlikely(!next))
drivers/md/dm-integrity.c
1377
next_node = container_of(next, struct journal_node, node);
drivers/md/dm-integrity.c
1383
struct rb_node *next;
drivers/md/dm-integrity.c
1389
next = rb_next(&node->node);
drivers/md/dm-integrity.c
1390
if (unlikely(!next))
drivers/md/dm-integrity.c
1393
next_node = container_of(next, struct journal_node, node);
drivers/md/dm-integrity.c
2889
struct bio *next = flushes->bi_next;
drivers/md/dm-integrity.c
2893
flushes = next;
drivers/md/dm-integrity.c
3912
ic->reboot_notifier.next = NULL;
drivers/md/dm-integrity.c
4258
pl[i - 1].next = &pl[i];
drivers/md/dm-integrity.c
4261
pl[i].next = NULL;
drivers/md/dm-io.c
201
dp->context_ptr = pl->next;
drivers/md/dm-ioctl.c
1344
spec->next = outptr - outbuf;
drivers/md/dm-ioctl.c
1416
static int next_target(struct dm_target_spec *last, uint32_t next, const char *end,
drivers/md/dm-ioctl.c
1433
if (remaining - sizeof(struct dm_target_spec) <= next) {
drivers/md/dm-ioctl.c
1438
if (next % __alignof__(struct dm_target_spec)) {
drivers/md/dm-ioctl.c
1440
next, __alignof__(struct dm_target_spec));
drivers/md/dm-ioctl.c
1444
*spec = (struct dm_target_spec *) ((unsigned char *) last + next);
drivers/md/dm-ioctl.c
1456
uint32_t next = param->data_start;
drivers/md/dm-ioctl.c
1469
if (next < min_size) {
drivers/md/dm-ioctl.c
1471
__func__, next, i ? "previous target" : "'struct dm_ioctl'");
drivers/md/dm-ioctl.c
1475
r = next_target(spec, next, end, &spec, &target_params);
drivers/md/dm-ioctl.c
1499
next = spec->next;
drivers/md/dm-ioctl.c
638
old_nl->next = (uint32_t) ((void *) nl -
drivers/md/dm-ioctl.c
642
nl->next = 0;
drivers/md/dm-ioctl.c
692
info->old_vers->next = (uint32_t) ((void *)info->vers - (void *)info->old_vers);
drivers/md/dm-ioctl.c
697
info->vers->next = 0;
drivers/md/dm-kcopyd.c
247
struct page_list *next;
drivers/md/dm-kcopyd.c
250
next = pl->next;
drivers/md/dm-kcopyd.c
255
pl->next = kc->pages;
drivers/md/dm-kcopyd.c
260
pl = next;
drivers/md/dm-kcopyd.c
278
kc->pages = pl->next;
drivers/md/dm-kcopyd.c
281
pl->next = *pages;
drivers/md/dm-kcopyd.c
298
struct page_list *next;
drivers/md/dm-kcopyd.c
301
next = pl->next;
drivers/md/dm-kcopyd.c
303
pl = next;
drivers/md/dm-kcopyd.c
313
struct page_list *pl = NULL, *next;
drivers/md/dm-kcopyd.c
316
next = alloc_pl(GFP_KERNEL);
drivers/md/dm-kcopyd.c
317
if (!next) {
drivers/md/dm-kcopyd.c
322
next->next = pl;
drivers/md/dm-kcopyd.c
323
pl = next;
drivers/md/dm-kcopyd.c
401
zero_page_list.next = &zero_page_list;
drivers/md/dm-kcopyd.c
454
job = list_entry(jobs->next, struct kcopyd_job, list);
drivers/md/dm-log-writes.c
448
goto next;
drivers/md/dm-log-writes.c
465
goto next;
drivers/md/dm-log-writes.c
474
next:
drivers/md/dm-pcache/cache_key.c
263
goto next;
drivers/md/dm-pcache/cache_key.c
307
goto next;
drivers/md/dm-pcache/cache_key.c
334
next:
drivers/md/dm-pcache/cache_req.c
688
next:
drivers/md/dm-pcache/cache_req.c
725
goto next;
drivers/md/dm-ps-historical-service-time.c
122
static u64 fixed_ema(u64 last, u64 next, u64 weight)
drivers/md/dm-ps-historical-service-time.c
125
last += next * (HST_FIXED_1 - weight);
drivers/md/dm-ps-historical-service-time.c
219
struct path_info *pi, *next;
drivers/md/dm-ps-historical-service-time.c
221
list_for_each_entry_safe(pi, next, paths, list) {
drivers/md/dm-ps-queue-length.c
69
struct path_info *pi, *next;
drivers/md/dm-ps-queue-length.c
71
list_for_each_entry_safe(pi, next, paths, list) {
drivers/md/dm-ps-round-robin.c
197
pi = list_entry(s->valid_paths.next, struct path_info, list);
drivers/md/dm-ps-round-robin.c
37
struct path_info *pi, *next;
drivers/md/dm-ps-round-robin.c
39
list_for_each_entry_safe(pi, next, paths, list) {
drivers/md/dm-ps-service-time.c
65
struct path_info *pi, *next;
drivers/md/dm-ps-service-time.c
67
list_for_each_entry_safe(pi, next, paths, list) {
drivers/md/dm-region-hash.c
445
struct dm_region *reg, *next;
drivers/md/dm-region-hash.c
486
list_for_each_entry_safe(reg, next, &recovered, list) {
drivers/md/dm-region-hash.c
492
list_for_each_entry_safe(reg, next, &failed_recovered, list) {
drivers/md/dm-region-hash.c
497
list_for_each_entry_safe(reg, next, &clean, list) {
drivers/md/dm-region-hash.c
654
reg = list_entry(rh->quiesced_regions.next,
drivers/md/dm-snap.c
1743
struct rb_node *next;
drivers/md/dm-snap.c
1748
next = rb_first(&s->out_of_order_tree);
drivers/md/dm-snap.c
1749
while (next) {
drivers/md/dm-snap.c
1750
pe = rb_entry(next, struct dm_snap_pending_exception,
drivers/md/dm-snap.c
1754
next = rb_next(next);
drivers/md/dm-stats.c
222
s = container_of(stats->list.next, struct dm_stat, list_entry);
drivers/md/dm-table.c
168
struct list_head *tmp, *next;
drivers/md/dm-table.c
170
list_for_each_safe(tmp, next, devices) {
drivers/md/dm-uevent.c
124
struct dm_uevent *event, *next;
drivers/md/dm-uevent.c
126
list_for_each_entry_safe(event, next, events, elist) {
drivers/md/dm-vdo/action-manager.c
124
manager->actions[0].next = &manager->actions[1];
drivers/md/dm-vdo/action-manager.c
125
manager->current_action = manager->actions[1].next =
drivers/md/dm-vdo/action-manager.c
258
manager->current_action = manager->current_action->next;
drivers/md/dm-vdo/action-manager.c
36
struct action *next;
drivers/md/dm-vdo/action-manager.c
364
} else if (!manager->current_action->next->in_use) {
drivers/md/dm-vdo/action-manager.c
365
current_action = manager->current_action->next;
drivers/md/dm-vdo/action-manager.c
381
.next = current_action->next,
drivers/md/dm-vdo/dm-vdo-target.c
1575
instances.next);
drivers/md/dm-vdo/dm-vdo-target.c
1587
instances.next = instance + 1;
drivers/md/dm-vdo/dm-vdo-target.c
169
unsigned int next;
drivers/md/dm-vdo/encodings.c
808
partition->next = layout->head;
drivers/md/dm-vdo/encodings.c
929
layout->head = part->next;
drivers/md/dm-vdo/encodings.c
949
for (partition = layout->head; partition != NULL; partition = partition->next) {
drivers/md/dm-vdo/encodings.c
997
for (partition = layout->head; partition != NULL; partition = partition->next) {
drivers/md/dm-vdo/encodings.h
584
struct partition *next; /* A pointer to the next partition in the layout */
drivers/md/dm-vdo/flush.c
234
if (zone->next == NULL) {
drivers/md/dm-vdo/flush.c
240
flusher->logical_zone_to_notify = zone->next;
drivers/md/dm-vdo/flush.c
321
for (zone = &flusher->vdo->logical_zones->zones[0]; zone != NULL; zone = zone->next)
drivers/md/dm-vdo/funnel-queue.c
116
queue->oldest = READ_ONCE(oldest->next);
drivers/md/dm-vdo/funnel-queue.c
128
WRITE_ONCE(oldest->next, NULL);
drivers/md/dm-vdo/funnel-queue.c
25
queue->stub.next = NULL;
drivers/md/dm-vdo/funnel-queue.c
46
struct funnel_queue_entry *next = READ_ONCE(oldest->next);
drivers/md/dm-vdo/funnel-queue.c
53
if (next == NULL)
drivers/md/dm-vdo/funnel-queue.c
59
oldest = next;
drivers/md/dm-vdo/funnel-queue.c
61
next = READ_ONCE(oldest->next);
drivers/md/dm-vdo/funnel-queue.c
68
if (next == NULL) {
drivers/md/dm-vdo/funnel-queue.c
86
next = READ_ONCE(oldest->next);
drivers/md/dm-vdo/funnel-queue.c
87
if (next == NULL) {
drivers/md/dm-vdo/funnel-queue.h
101
WRITE_ONCE(previous->next, entry);
drivers/md/dm-vdo/funnel-queue.h
48
struct funnel_queue_entry *next;
drivers/md/dm-vdo/funnel-queue.h
95
WRITE_ONCE(entry->next, NULL);
drivers/md/dm-vdo/indexer/radix-sort.c
72
static inline void insert_key(const struct task task, sort_key_t *next)
drivers/md/dm-vdo/indexer/radix-sort.c
75
sort_key_t unsorted = *next;
drivers/md/dm-vdo/indexer/radix-sort.c
78
while ((--next >= task.first_key) &&
drivers/md/dm-vdo/indexer/radix-sort.c
79
(compare(unsorted, next[0], task.offset, task.length) < 0))
drivers/md/dm-vdo/indexer/radix-sort.c
80
next[1] = next[0];
drivers/md/dm-vdo/indexer/radix-sort.c
83
next[1] = unsorted;
drivers/md/dm-vdo/indexer/radix-sort.c
92
sort_key_t *next;
drivers/md/dm-vdo/indexer/radix-sort.c
94
for (next = task.first_key + 1; next <= task.last_key; next++)
drivers/md/dm-vdo/indexer/radix-sort.c
95
insert_key(task, next);
drivers/md/dm-vdo/indexer/volume.c
605
struct uds_request *next;
drivers/md/dm-vdo/indexer/volume.c
607
for (request = entry->first_request; request != NULL; request = next) {
drivers/md/dm-vdo/indexer/volume.c
608
next = request->next_request;
drivers/md/dm-vdo/io-submitter.c
168
struct bio *bio, *next;
drivers/md/dm-vdo/io-submitter.c
172
for (bio = get_bio_list(vio); bio != NULL; bio = next) {
drivers/md/dm-vdo/io-submitter.c
173
next = bio->bi_next;
drivers/md/dm-vdo/logical-zone.c
338
zone->allocation_zone = zone->allocation_zone->next;
drivers/md/dm-vdo/logical-zone.c
63
zone->next = &zones->zones[zone_number + 1];
drivers/md/dm-vdo/logical-zone.h
52
struct logical_zone *next;
drivers/md/dm-vdo/memory-alloc.c
123
block->next = memory_stats.vmalloc_list;
drivers/md/dm-vdo/memory-alloc.c
140
block_ptr = &block->next) {
drivers/md/dm-vdo/memory-alloc.c
142
*block_ptr = block->next;
drivers/md/dm-vdo/memory-alloc.c
76
struct vmalloc_block_info *next;
drivers/md/dm-vdo/packer.c
278
struct data_vio *client, *next;
drivers/md/dm-vdo/packer.c
286
for (client = agent->compression.next_in_batch; client != NULL; client = next) {
drivers/md/dm-vdo/packer.c
287
next = client->compression.next_in_batch;
drivers/md/dm-vdo/packer.c
299
struct data_vio *client, *next;
drivers/md/dm-vdo/packer.c
308
for (client = agent->compression.next_in_batch; client != NULL; client = next) {
drivers/md/dm-vdo/packer.c
309
next = client->compression.next_in_batch;
drivers/md/dm-vdo/physical-zone.c
345
zone->next = &zones->zones[(zone_number + 1) % vdo->thread_config.physical_zone_count];
drivers/md/dm-vdo/physical-zone.c
547
bool tried_all = (allocation->first_allocation_zone == zone->next->zone_number);
drivers/md/dm-vdo/physical-zone.c
575
allocation->zone = zone->next;
drivers/md/dm-vdo/physical-zone.h
68
struct physical_zone *next;
drivers/md/dm-vdo/priority-table.c
174
entry = bucket->queue.next;
drivers/md/dm-vdo/priority-table.c
204
next_entry = entry->next;
drivers/md/dm-vdo/slab-depot.c
161
list_move_tail(&journal->dirty_entry, dirty_journal->dirty_entry.next);
drivers/md/dm-vdo/slab-depot.c
3111
.next = (((slabs == NULL) || (start < end)) ? NULL : slabs[start]),
drivers/md/dm-vdo/slab-depot.c
3132
struct vdo_slab *slab = iterator->next;
drivers/md/dm-vdo/slab-depot.c
3135
iterator->next = NULL;
drivers/md/dm-vdo/slab-depot.c
3137
iterator->next = iterator->slabs[slab->slab_number - iterator->stride];
drivers/md/dm-vdo/slab-depot.c
3173
while (iterator.next != NULL) {
drivers/md/dm-vdo/slab-depot.c
3441
while (iterator.next != NULL) {
drivers/md/dm-vdo/slab-depot.c
3500
if (allocator->slabs_to_erase.next == NULL) {
drivers/md/dm-vdo/slab-depot.c
3573
while (iterator.next != NULL) {
drivers/md/dm-vdo/slab-depot.c
3662
while (iterator.next != NULL) {
drivers/md/dm-vdo/slab-depot.c
4385
while (iterator.next != NULL) {
drivers/md/dm-vdo/slab-depot.c
721
pooled = container_of(journal->uncommitted_blocks.next,
drivers/md/dm-vdo/slab-depot.h
316
struct vdo_slab *next;
drivers/md/dm-vdo/vdo.c
1008
.next = thread->listeners,
drivers/md/dm-vdo/vdo.c
1155
listener = listener->next;
drivers/md/dm-vdo/vdo.c
643
struct read_only_listener *listener, *next;
drivers/md/dm-vdo/vdo.c
645
for (listener = vdo_forget(thread->listeners); listener != NULL; listener = next) {
drivers/md/dm-vdo/vdo.c
646
next = vdo_forget(listener->next);
drivers/md/dm-vdo/vdo.h
53
struct read_only_listener *next;
drivers/md/dm-writecache.c
1400
struct rb_node *next = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
1402
if (!next)
drivers/md/dm-writecache.c
1404
f = container_of(next, struct wc_entry, rb_node);
drivers/md/dm-writecache.c
1689
wb = list_entry(list->next, struct writeback_struct, endio_entry);
drivers/md/dm-writecache.c
1726
c = list_entry(list->next, struct copy_struct, endio_entry);
drivers/md/dm-writecache.c
1772
list.next->prev = list.prev->next = &list;
drivers/md/dm-writecache.c
721
struct rb_node *next;
drivers/md/dm-writecache.c
728
next = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
730
if (unlikely(!next))
drivers/md/dm-writecache.c
731
next = rb_first(&wc->freetree);
drivers/md/dm-writecache.c
732
wc->current_free = next ? container_of(next, struct wc_entry, rb_node) : NULL;
drivers/md/dm-writecache.c
736
e = container_of(wc->freelist.next, struct wc_entry, lru);
drivers/md/dm-writecache.c
775
wc->lru.next = LIST_POISON1;
drivers/md/dm-writecache.c
777
wc->freelist.next = LIST_POISON1;
drivers/md/dm-writecache.c
804
e = container_of(wc->lru.next, struct wc_entry, lru);
drivers/md/dm-writecache.c
815
if (unlikely(e->lru.next == &wc->lru))
drivers/md/dm-writecache.c
817
e2 = container_of(e->lru.next, struct wc_entry, lru);
drivers/md/dm-zoned-metadata.c
1709
goto next;
drivers/md/dm-zoned-metadata.c
1737
goto next;
drivers/md/dm-zoned-metadata.c
1767
next:
drivers/md/dm-zoned-metadata.c
2778
struct dmz_mblock *mblk, *next;
drivers/md/dm-zoned-metadata.c
2818
rbtree_postorder_for_each_entry_safe(mblk, next, root, node) {
drivers/md/dm.c
1006
struct dm_io *next = io->next;
drivers/md/dm.c
1010
io->next = NULL;
drivers/md/dm.c
1012
io = next;
drivers/md/dm.c
1705
io->next = NULL;
drivers/md/dm.c
1712
io->next = *head;
drivers/md/dm.c
2017
if (devices->next == devices->prev)
drivers/md/dm.c
2114
struct dm_io *curr, *next;
drivers/md/dm.c
2133
for (curr = list, next = curr->next; curr; curr = next, next =
drivers/md/dm.c
2134
curr ? curr->next : NULL) {
drivers/md/dm.c
2142
curr->next = tmp;
drivers/md/dm.c
873
struct dm_io *next = md->requeue_list;
drivers/md/dm.c
876
io->next = next;
drivers/md/md-bitmap.c
540
struct buffer_head *next = bh->b_this_page;
drivers/md/md-bitmap.c
542
bh = next;
drivers/md/md.c
10483
.next = NULL,
drivers/md/md.c
10776
node_detected_dev = list_entry(all_detected_devices.next,
drivers/md/md.c
4832
= list_entry(mddev->disks.next,
drivers/md/md.c
7096
rdev0 = list_entry(pending_raid_disks.next,
drivers/md/md.c
7333
= list_entry(mddev->disks.next,
drivers/md/md.c
8941
.next = md_seq_next,
drivers/md/md.c
9735
int next = (last_mark+1) % SYNC_MARKS;
drivers/md/md.c
9737
mddev->resync_mark = mark[next];
drivers/md/md.c
9738
mddev->resync_mark_cnt = mark_cnt[next];
drivers/md/md.c
9739
mark[next] = jiffies;
drivers/md/md.c
9740
mark_cnt[next] = io_sectors - atomic_read(&mddev->recovery_active);
drivers/md/md.c
9741
last_mark = next;
drivers/md/persistent-data/dm-space-map-metadata.c
125
unsigned int next = brb_next(brb, brb->end);
drivers/md/persistent-data/dm-space-map-metadata.c
131
if (next == brb->begin)
drivers/md/persistent-data/dm-space-map-metadata.c
139
brb->end = next;
drivers/md/raid1.c
906
struct bio *next = bio->bi_next;
drivers/md/raid1.c
909
bio = next;
drivers/md/raid10.c
1107
struct bio *next = bio->bi_next;
drivers/md/raid10.c
1110
bio = next;
drivers/md/raid10.c
4600
sector_t next, safe, last;
drivers/md/raid10.c
4636
next = first_dev_address(conf->reshape_progress - 1,
drivers/md/raid10.c
4645
if (next + conf->offset_diff < safe)
drivers/md/raid10.c
4657
next = last_dev_address(conf->reshape_progress, &conf->geo);
drivers/md/raid10.c
4667
if (next > safe + conf->offset_diff)
drivers/md/raid10.c
891
struct bio *next = bio->bi_next;
drivers/md/raid10.c
894
bio = next;
drivers/md/raid5-cache.c
1197
struct r5l_io_unit *io, *next;
drivers/md/raid5-cache.c
1202
list_for_each_entry_safe(io, next, &log->finished_ios, log_sibling) {
drivers/md/raid5-cache.c
1397
struct stripe_head *sh, *next;
drivers/md/raid5-cache.c
1404
list_for_each_entry_safe(sh, next, &conf->r5c_full_stripe_list, lru) {
drivers/md/raid5-cache.c
1411
list_for_each_entry_safe(sh, next,
drivers/md/raid5-cache.c
1943
struct stripe_head *sh, *next;
drivers/md/raid5-cache.c
1945
list_for_each_entry_safe(sh, next, cached_stripe_list, lru) {
drivers/md/raid5-cache.c
1956
struct stripe_head *sh, *next;
drivers/md/raid5-cache.c
1958
list_for_each_entry_safe(sh, next, cached_stripe_list, lru)
drivers/md/raid5-cache.c
2415
struct stripe_head *sh, *next;
drivers/md/raid5-cache.c
2427
list_for_each_entry_safe(sh, next, &ctx->cached_list, lru) {
drivers/md/raid5-cache.c
514
struct stripe_head *sh, *next;
drivers/md/raid5-cache.c
516
list_for_each_entry_safe(sh, next, &io->stripe_list, log_list) {
drivers/md/raid5-cache.c
528
struct r5l_io_unit *io, *next;
drivers/md/raid5-cache.c
532
list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) {
drivers/md/raid5-cache.c
544
struct r5l_io_unit *io, *next;
drivers/md/raid5-cache.c
548
list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) {
drivers/md/raid5-ppl.c
403
struct stripe_head *sh, *next;
drivers/md/raid5-ppl.c
410
list_for_each_entry_safe(sh, next, &io->stripe_list, log_list) {
drivers/md/raid5.c
1059
struct list_head *first, *next = NULL;
drivers/md/raid5.c
1067
first = conf->pending_list.next;
drivers/md/raid5.c
1078
first = data->sibling.next;
drivers/md/raid5.c
1079
next = data->sibling.next;
drivers/md/raid5.c
1090
if (next != &conf->pending_list)
drivers/md/raid5.c
1091
conf->next_pending_data = list_entry(next,
drivers/md/raid5.c
2563
nsh = list_entry(newstripes.next, struct stripe_head, lru);
drivers/md/raid5.c
2641
nsh = list_entry(newstripes.next, struct stripe_head, lru);
drivers/md/raid5.c
454
first = (conf->inactive_list + hash)->next;
drivers/md/raid5.c
4850
struct stripe_head *sh, *next;
drivers/md/raid5.c
4853
list_for_each_entry_safe(sh, next, &head_sh->batch_list, batch_list) {
drivers/md/raid5.c
5306
struct list_head *l = conf->delayed_list.next;
drivers/md/raid5.c
5327
struct stripe_head *sh = list_entry(head.next, struct stripe_head, lru);
drivers/md/raid5.c
5562
sh = list_entry(handle_list->next, typeof(*sh), lru);
drivers/md/raid5.c
5567
if (conf->hold_list.next == conf->last_hold)
drivers/md/raid5.c
5570
conf->last_hold = conf->hold_list.next;
drivers/md/raid5.c
5632
if (cb->list.next && !list_empty(&cb->list)) {
drivers/md/raid5.c
5676
if (cb->list.next == NULL) {
drivers/md/raid5.c
6460
sh = list_entry(stripes.next, struct stripe_head, lru);
drivers/media/common/saa7146/saa7146_fops.c
104
struct saa7146_buf *buf,*next = NULL;
drivers/media/common/saa7146/saa7146_fops.c
114
buf = list_entry(q->queue.next, struct saa7146_buf, list);
drivers/media/common/saa7146/saa7146_fops.c
117
next = list_entry(q->queue.next, struct saa7146_buf, list);
drivers/media/common/saa7146/saa7146_fops.c
120
buf, q->queue.prev, q->queue.next);
drivers/media/common/saa7146/saa7146_fops.c
121
buf->activate(dev,buf,next);
drivers/media/common/saa7146/saa7146_hlp.c
721
void saa7146_set_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next)
drivers/media/common/saa7146/saa7146_hlp.c
728
DEB_CAP("buf:%p, next:%p\n", buf, next);
drivers/media/common/saa7146/saa7146_vbi.c
135
static void saa7146_set_vbi_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next)
drivers/media/common/saa7146/saa7146_vbi.c
207
struct saa7146_buf *next)
drivers/media/common/saa7146/saa7146_vbi.c
211
DEB_VBI("dev:%p, buf:%p, next:%p\n", dev, buf, next);
drivers/media/common/saa7146/saa7146_vbi.c
212
saa7146_set_vbi_capture(dev,buf,next);
drivers/media/common/saa7146/saa7146_vbi.c
302
buf = list_entry(dq->queue.next, struct saa7146_buf, list);
drivers/media/common/saa7146/saa7146_video.c
539
struct saa7146_buf *next)
drivers/media/common/saa7146/saa7146_video.c
543
saa7146_set_capture(dev,buf,next);
drivers/media/common/saa7146/saa7146_video.c
647
buf = list_entry(dq->queue.next, struct saa7146_buf, list);
drivers/media/common/siano/smscoreapi.c
1205
cb = (struct smscore_buffer_t *) coredev->buffers.next;
drivers/media/common/siano/smscoreapi.c
1630
cb = (struct smscore_buffer_t *) coredev->buffers.next;
drivers/media/common/siano/smscoreapi.c
1756
(struct smscore_idlist_t *) client->idlist.next;
drivers/media/common/siano/smscoreapi.c
2128
g_smscore_notifyees.next;
drivers/media/common/siano/smscoreapi.c
2139
g_smscore_registry.next;
drivers/media/common/siano/smscoreapi.c
430
struct list_head *next;
drivers/media/common/siano/smscoreapi.c
433
for (next = g_smscore_registry.next;
drivers/media/common/siano/smscoreapi.c
434
next != &g_smscore_registry;
drivers/media/common/siano/smscoreapi.c
435
next = next->next) {
drivers/media/common/siano/smscoreapi.c
436
entry = (struct smscore_registry_entry_t *) next;
drivers/media/common/siano/smscoreapi.c
527
struct list_head *next, *first;
drivers/media/common/siano/smscoreapi.c
535
for (next = first->next;
drivers/media/common/siano/smscoreapi.c
536
next != first && !rc;
drivers/media/common/siano/smscoreapi.c
537
next = next->next) {
drivers/media/common/siano/smscoreapi.c
539
(struct smscore_device_t *) next;
drivers/media/common/siano/smscoreapi.c
565
struct list_head *next, *first;
drivers/media/common/siano/smscoreapi.c
571
for (next = first->next; next != first;) {
drivers/media/common/siano/smscoreapi.c
573
(struct smscore_device_notifyee_t *) next;
drivers/media/common/siano/smscoreapi.c
574
next = next->next;
drivers/media/common/siano/smscoreapi.c
592
client = (struct smscore_client_t *) coredev->clients.next;
drivers/media/common/siano/smsdvb-main.c
1261
smsdvb_unregister_client((struct smsdvb_client_t *)g_smsdvb_clients.next);
drivers/media/dvb-core/dmxdev.c
493
list_for_each_entry(feed, &dmxdevfilter->feed.ts, next)
drivers/media/dvb-core/dmxdev.c
514
list_for_each_entry(feed, &filter->feed.ts, next) {
drivers/media/dvb-core/dmxdev.c
574
list_for_each_entry(feed, &dmxdevfilter->feed.ts, next) {
drivers/media/dvb-core/dmxdev.c
594
list_for_each_entry_safe(feed, tmp, &dmxdevfilter->feed.ts, next) {
drivers/media/dvb-core/dmxdev.c
595
list_del(&feed->next);
drivers/media/dvb-core/dmxdev.c
768
list_for_each_entry(feed, &filter->feed.ts, next) {
drivers/media/dvb-core/dmxdev.c
902
list_add(&feed->next, &filter->feed.ts);
drivers/media/dvb-core/dmxdev.c
919
list_for_each_entry_safe(feed, tmp, &filter->feed.ts, next) {
drivers/media/dvb-core/dmxdev.c
924
list_del(&feed->next);
drivers/media/dvb-core/dvb_demux.c
1042
dvbdmxfeed->filter = dvbdmxfilter->next;
drivers/media/dvb-core/dvb_demux.c
1044
while (f->next != dvbdmxfilter)
drivers/media/dvb-core/dvb_demux.c
1045
f = f->next;
drivers/media/dvb-core/dvb_demux.c
1046
f->next = f->next->next;
drivers/media/dvb-core/dvb_demux.c
181
} while ((f = f->next) && sec->is_filtering);
drivers/media/dvb-core/dvb_demux.c
892
dvbdmxfilter->next = dvbdmxfeed->filter;
drivers/media/dvb-core/dvb_demux.c
942
} while ((f = f->next));
drivers/media/dvb-core/dvb_vb2.c
262
ctx->buf = list_entry(ctx->dvb_q.next,
drivers/media/dvb-core/dvb_vb2.c
98
buf = list_entry(ctx->dvb_q.next,
drivers/media/dvb-core/dvbdev.c
1119
struct dvbdevfops_node *node, *next;
drivers/media/dvb-core/dvbdev.c
1125
list_for_each_entry_safe(node, next, &dvbdevfops_list, list_head) {
drivers/media/dvb-frontends/drx39xyj/drx_driver.h
1216
struct drx_version_list *next;
drivers/media/dvb-frontends/rtl2832_sdr.c
163
buf = list_entry(dev->queued_bufs.next,
drivers/media/dvb-frontends/rtl2832_sdr.c
413
buf = list_entry(dev->queued_bufs.next,
drivers/media/i2c/et8ek8/et8ek8_driver.c
268
const struct et8ek8_reg *next;
drivers/media/i2c/et8ek8/et8ek8_driver.c
277
next = regs;
drivers/media/i2c/et8ek8/et8ek8_driver.c
284
while (next->type != ET8EK8_REG_TERM &&
drivers/media/i2c/et8ek8/et8ek8_driver.c
285
next->type != ET8EK8_REG_DELAY) {
drivers/media/i2c/et8ek8/et8ek8_driver.c
290
if (WARN(next->type != ET8EK8_REG_8BIT &&
drivers/media/i2c/et8ek8/et8ek8_driver.c
291
next->type != ET8EK8_REG_16BIT,
drivers/media/i2c/et8ek8/et8ek8_driver.c
292
"Invalid type = %d", next->type)) {
drivers/media/i2c/et8ek8/et8ek8_driver.c
300
next++;
drivers/media/i2c/et8ek8/et8ek8_driver.c
316
if (next->type == ET8EK8_REG_DELAY) {
drivers/media/i2c/et8ek8/et8ek8_driver.c
317
msleep(next->val);
drivers/media/i2c/et8ek8/et8ek8_driver.c
322
next++;
drivers/media/i2c/et8ek8/et8ek8_driver.c
323
regs = next;
drivers/media/i2c/et8ek8/et8ek8_driver.c
326
} while (next->type != ET8EK8_REG_TERM);
drivers/media/i2c/et8ek8/et8ek8_driver.c
363
struct et8ek8_reglist **next = &meta->reglist[0].ptr;
drivers/media/i2c/et8ek8/et8ek8_driver.c
365
while (*next) {
drivers/media/i2c/et8ek8/et8ek8_driver.c
366
if ((*next)->type == type)
drivers/media/i2c/et8ek8/et8ek8_driver.c
367
return *next;
drivers/media/i2c/et8ek8/et8ek8_driver.c
369
next++;
drivers/media/i2c/imx274.c
633
const struct reg_8 *next;
drivers/media/i2c/imx274.c
641
for (next = table;; next++) {
drivers/media/i2c/imx274.c
642
if ((next->addr != range_start + range_count) ||
drivers/media/i2c/imx274.c
643
(next->addr == IMX274_TABLE_END) ||
drivers/media/i2c/imx274.c
644
(next->addr == IMX274_TABLE_WAIT_MS) ||
drivers/media/i2c/imx274.c
663
if (next->addr == IMX274_TABLE_END)
drivers/media/i2c/imx274.c
666
if (next->addr == IMX274_TABLE_WAIT_MS) {
drivers/media/i2c/imx274.c
667
msleep_range(next->val);
drivers/media/i2c/imx274.c
672
val = next->val;
drivers/media/i2c/imx274.c
675
range_start = next->addr;
drivers/media/i2c/tvp514x.c
358
const struct tvp514x_reg *next = reglist;
drivers/media/i2c/tvp514x.c
360
for (; next->token != TOK_TERM; next++) {
drivers/media/i2c/tvp514x.c
361
if (next->token == TOK_DELAY) {
drivers/media/i2c/tvp514x.c
362
msleep(next->val);
drivers/media/i2c/tvp514x.c
366
if (next->token == TOK_SKIP)
drivers/media/i2c/tvp514x.c
369
err = tvp514x_write_reg(sd, next->reg, (u8) next->val);
drivers/media/mc/mc-device.c
608
struct media_entity_notify *notify, *next;
drivers/media/mc/mc-device.c
642
list_for_each_entry_safe(notify, next, &mdev->entity_notify, list)
drivers/media/mc/mc-device.c
814
struct media_entity *next;
drivers/media/mc/mc-device.c
833
list_for_each_entry_safe(entity, next, &mdev->entities, graph_obj.list)
drivers/media/mc/mc-device.c
84
int next = id & MEDIA_ENT_ID_FLAG_NEXT;
drivers/media/mc/mc-device.c
89
if (((media_entity_id(entity) == id) && !next) ||
drivers/media/mc/mc-device.c
90
((media_entity_id(entity) > id) && next)) {
drivers/media/mc/mc-entity.c
1007
iter->cursor = pipe->pads.next;
drivers/media/mc/mc-entity.c
1013
iter->cursor = iter->cursor->next;
drivers/media/mc/mc-entity.c
1038
iter->cursor = pipe->pads.next;
drivers/media/mc/mc-entity.c
1046
iter->cursor = iter->cursor->next;
drivers/media/mc/mc-entity.c
300
graph->stack[graph->top].link = entity->links.next;
drivers/media/mc/mc-entity.c
363
struct media_entity *next;
drivers/media/mc/mc-entity.c
369
link_top(graph) = link_top(graph)->next;
drivers/media/mc/mc-entity.c
375
link_top(graph) = link_top(graph)->next;
drivers/media/mc/mc-entity.c
384
next = media_entity_other(entity, link);
drivers/media/mc/mc-entity.c
387
if (media_entity_enum_test_and_set(&graph->ent_enum, next)) {
drivers/media/mc/mc-entity.c
388
link_top(graph) = link_top(graph)->next;
drivers/media/mc/mc-entity.c
391
next->name);
drivers/media/mc/mc-entity.c
396
link_top(graph) = link_top(graph)->next;
drivers/media/mc/mc-entity.c
397
stack_push(graph, next);
drivers/media/mc/mc-entity.c
399
next->name);
drivers/media/mc/mc-entity.c
527
entry->links = pad->entity->links.next;
drivers/media/mc/mc-entity.c
550
if (entry->links->next == &entry->pad->entity->links) {
drivers/media/mc/mc-entity.c
559
entry->links = entry->links->next;
drivers/media/pci/bt8xx/bttv-driver.c
1547
buf = list_entry(btv->capture.next,
drivers/media/pci/bt8xx/bttv-driver.c
2663
item = list_entry(btv->capture.next, struct bttv_buffer, list);
drivers/media/pci/bt8xx/bttv-driver.c
2672
item->list.next != &btv->capture) {
drivers/media/pci/bt8xx/bttv-driver.c
2673
item = list_entry(item->list.next,
drivers/media/pci/bt8xx/bttv-driver.c
2837
item = list_entry(btv->capture.next, struct bttv_buffer, list);
drivers/media/pci/bt8xx/bttv-driver.c
2844
item = list_entry(btv->vcapture.next, struct bttv_buffer, list);
drivers/media/pci/bt8xx/bttv-driver.c
2933
new = list_entry(btv->vcapture.next, struct bttv_buffer, list);
drivers/media/pci/bt8xx/bttv-risc.c
485
unsigned long next = btv->main.dma + ((slot+2) << 2);
drivers/media/pci/bt8xx/bttv-risc.c
489
btv->main.cpu[slot+1] = cpu_to_le32(next);
drivers/media/pci/bt8xx/bttv-risc.c
501
risc->jmp[1] = cpu_to_le32(next);
drivers/media/pci/bt8xx/bttv-risc.c
611
if (set->top->list.next)
drivers/media/pci/bt8xx/bttv-risc.c
614
if (set->top->list.next)
drivers/media/pci/bt8xx/bttv-risc.c
616
if (set->bottom->list.next)
drivers/media/pci/bt8xx/bttv-risc.c
630
if (set->top->list.next)
drivers/media/pci/bt8xx/bttv-risc.c
640
if (set->bottom->list.next)
drivers/media/pci/bt8xx/bttv-vbi.c
135
buf = list_entry(btv->vcapture.next,
drivers/media/pci/cobalt/cobalt-omnitek.c
155
dma_addr_t next = desc->bus;
drivers/media/pci/cobalt/cobalt-omnitek.c
164
WARN_ON(next & 3);
drivers/media/pci/cobalt/cobalt-omnitek.c
209
next += sizeof(struct sg_dma_descriptor);
drivers/media/pci/cobalt/cobalt-omnitek.c
210
d->next_h = (u32)((u64)next >> 32);
drivers/media/pci/cobalt/cobalt-omnitek.c
211
d->next_l = (u32)next |
drivers/media/pci/cobalt/cobalt-omnitek.c
253
next += sizeof(struct sg_dma_descriptor);
drivers/media/pci/cobalt/cobalt-omnitek.c
263
d->next_h = (u32)((u64)next >> 32);
drivers/media/pci/cobalt/cobalt-omnitek.c
264
d->next_l = (u32)next | (to_pci ? WRITE_TO_PCI : 0);
drivers/media/pci/cobalt/cobalt-omnitek.c
272
struct sg_dma_desc_info *next)
drivers/media/pci/cobalt/cobalt-omnitek.c
277
if (next == NULL) {
drivers/media/pci/cobalt/cobalt-omnitek.c
281
d->next_h = (u32)((u64)next->bus >> 32);
drivers/media/pci/cobalt/cobalt-omnitek.c
282
d->next_l = (u32)next->bus | direction | INTERRUPT_ENABLE;
drivers/media/pci/cobalt/cobalt-omnitek.h
40
struct sg_dma_desc_info *next);
drivers/media/pci/cx23885/cx23885-417.c
1176
struct cx23885_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx23885/cx23885-417.c
1184
struct cx23885_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx23885/cx23885-core.c
1655
buf = list_entry(q->active.next, struct cx23885_buffer,
drivers/media/pci/cx23885/cx23885-core.c
434
buf = list_entry(q->active.next,
drivers/media/pci/cx23885/cx23885-dvb.c
154
struct cx23885_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx23885/cx23885-vbi.c
221
struct cx23885_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx23885/cx23885-vbi.c
237
struct cx23885_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx23885/cx23885-video.c
493
struct cx23885_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx23885/cx23885-video.c
509
struct cx23885_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx23885/cx23885-video.c
95
buf = list_entry(q->active.next,
drivers/media/pci/cx25821/cx25821-video.c
113
buf = list_entry(dmaq->active.next,
drivers/media/pci/cx25821/cx25821-video.c
266
struct cx25821_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx25821/cx25821-video.c
284
struct cx25821_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx88/cx88-blackbird.c
734
buf = list_entry(dmaq->active.next, struct cx88_buffer, list);
drivers/media/pci/cx88/cx88-blackbird.c
742
struct cx88_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx88/cx88-blackbird.c
770
struct cx88_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx88/cx88-core.c
525
buf = list_entry(q->active.next,
drivers/media/pci/cx88/cx88-dvb.c
126
buf = list_entry(dmaq->active.next, struct cx88_buffer, list);
drivers/media/pci/cx88/cx88-dvb.c
141
struct cx88_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx88/cx88-mpeg.c
207
buf = list_entry(q->active.next, struct cx88_buffer, list);
drivers/media/pci/cx88/cx88-mpeg.c
280
buf = list_entry(q->active.next, struct cx88_buffer, list);
drivers/media/pci/cx88/cx88-vbi.c
106
buf = list_entry(q->active.next, struct cx88_buffer, list);
drivers/media/pci/cx88/cx88-vbi.c
198
struct cx88_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx88/cx88-vbi.c
217
struct cx88_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx88/cx88-video.c
410
buf = list_entry(q->active.next, struct cx88_buffer, list);
drivers/media/pci/cx88/cx88-video.c
533
struct cx88_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/cx88/cx88-video.c
551
struct cx88_buffer *buf = list_entry(dmaq->active.next,
drivers/media/pci/intel/ipu3/ipu3-cio2.c
911
unsigned int i, j, next = q->bufs_next;
drivers/media/pci/intel/ipu3/ipu3-cio2.c
943
if (bufs_queued <= 1 || fbpt_rp == next)
drivers/media/pci/intel/ipu3/ipu3-cio2.c
945
next = (fbpt_rp + 1) % CIO2_MAX_BUFFERS;
drivers/media/pci/intel/ipu3/ipu3-cio2.c
954
if (!q->bufs[next]) {
drivers/media/pci/intel/ipu3/ipu3-cio2.c
955
q->bufs[next] = b;
drivers/media/pci/intel/ipu3/ipu3-cio2.c
956
entry = &q->fbpt[next * CIO2_MAX_LOPS];
drivers/media/pci/intel/ipu3/ipu3-cio2.c
959
q->bufs_next = (next + 1) % CIO2_MAX_BUFFERS;
drivers/media/pci/intel/ipu3/ipu3-cio2.c
966
dev_dbg(dev, "entry %i was full!\n", next);
drivers/media/pci/intel/ipu3/ipu3-cio2.c
967
next = (next + 1) % CIO2_MAX_BUFFERS;
drivers/media/pci/intel/ipu6/ipu6.c
69
u32 next;
drivers/media/pci/ivtv/ivtv-irq.c
217
buf = list_entry(s->q_predma.list.next, struct ivtv_buffer, list);
drivers/media/pci/ivtv/ivtv-queue.c
66
buf = list_entry(q->list.next, struct ivtv_buffer, list);
drivers/media/pci/ivtv/ivtv-queue.c
67
list_del_init(q->list.next);
drivers/media/pci/ivtv/ivtv-queue.c
79
struct ivtv_buffer *buf = list_entry(from->list.next, struct ivtv_buffer, list);
drivers/media/pci/ivtv/ivtv-queue.c
81
list_move_tail(from->list.next, &to->list);
drivers/media/pci/mgb4/mgb4_vin.c
739
list_del_init(vindev->buf_list.next);
drivers/media/pci/mgb4/mgb4_vout.c
219
list_del_init(voutdev->buf_list.next);
drivers/media/pci/mgb4/mgb4_vout.c
607
list_del_init(voutdev->buf_list.next);
drivers/media/pci/saa7134/saa7134-core.c
1166
list_for_each_entry(mops, &mops_list, next)
drivers/media/pci/saa7134/saa7134-core.c
1315
list_for_each_entry(mops, &mops_list, next)
drivers/media/pci/saa7134/saa7134-core.c
1350
struct saa7134_buf *buf, *next;
drivers/media/pci/saa7134/saa7134-core.c
1355
next = buf;
drivers/media/pci/saa7134/saa7134-core.c
1364
next = list_entry(q->queue.next, struct saa7134_buf,
drivers/media/pci/saa7134/saa7134-core.c
1366
buf->activate(dev, buf, next);
drivers/media/pci/saa7134/saa7134-core.c
1466
list_add_tail(&ops->next,&mops_list);
drivers/media/pci/saa7134/saa7134-core.c
1476
list_del(&ops->next);
drivers/media/pci/saa7134/saa7134-core.c
264
struct saa7134_buf *next = NULL;
drivers/media/pci/saa7134/saa7134-core.c
276
next = list_entry(q->queue.next, struct saa7134_buf,
drivers/media/pci/saa7134/saa7134-core.c
279
buf->activate(dev, buf, next);
drivers/media/pci/saa7134/saa7134-core.c
304
struct saa7134_buf *buf,*next = NULL;
drivers/media/pci/saa7134/saa7134-core.c
311
buf = list_entry(q->queue.next, struct saa7134_buf, entry);
drivers/media/pci/saa7134/saa7134-core.c
313
buf, q->queue.prev, q->queue.next);
drivers/media/pci/saa7134/saa7134-core.c
316
next = list_entry(q->queue.next, struct saa7134_buf, entry);
drivers/media/pci/saa7134/saa7134-core.c
318
buf->activate(dev, buf, next);
drivers/media/pci/saa7134/saa7134-core.c
320
q->queue.prev, q->queue.next);
drivers/media/pci/saa7134/saa7134-ts.c
33
struct saa7134_buf *next)
drivers/media/pci/saa7134/saa7134-ts.c
42
if (NULL == next)
drivers/media/pci/saa7134/saa7134-ts.c
43
next = buf;
drivers/media/pci/saa7134/saa7134-ts.c
45
ts_dbg("- [top] buf=%p next=%p\n", buf, next);
drivers/media/pci/saa7134/saa7134-ts.c
47
saa_writel(SAA7134_RS_BA2(5),saa7134_buffer_base(next));
drivers/media/pci/saa7134/saa7134-ts.c
50
ts_dbg("- [bottom] buf=%p next=%p\n", buf, next);
drivers/media/pci/saa7134/saa7134-ts.c
51
saa_writel(SAA7134_RS_BA1(5),saa7134_buffer_base(next));
drivers/media/pci/saa7134/saa7134-vbi.c
71
struct saa7134_buf *next)
drivers/media/pci/saa7134/saa7134-video.c
635
struct saa7134_buf *next)
drivers/media/pci/saa7134/saa7134.h
471
struct saa7134_buf *next);
drivers/media/pci/saa7134/saa7134.h
530
struct list_head next;
drivers/media/pci/saa7164/saa7164-core.c
1142
.next = saa7164_seq_next,
drivers/media/pci/solo6x10/solo6x10-g723.c
311
ss; ss = ss->next, i++)
drivers/media/pci/solo6x10/solo6x10-g723.c
83
for (ss = pstr->substream; ss != NULL; ss = ss->next) {
drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c
724
solo_enc->vidq_active.next,
drivers/media/pci/solo6x10/solo6x10-v4l2.c
339
solo_dev->vidq_active.next,
drivers/media/pci/tw5864/tw5864-core.c
209
goto next;
drivers/media/pci/tw5864/tw5864-core.c
236
next:
drivers/media/pci/tw5864/tw5864-video.c
438
struct tw5864_buf *buf = list_entry(input->active.next,
drivers/media/pci/tw5864/tw5864-video.c
460
struct tw5864_buf *buf = list_entry(input->active.next,
drivers/media/pci/tw68/tw68-core.c
397
buf = container_of(dev->active.next, struct tw68_buf, list);
drivers/media/pci/tw68/tw68-video.c
498
container_of(dev->active.next, struct tw68_buf, list);
drivers/media/pci/tw68/tw68-video.c
513
container_of(dev->active.next, struct tw68_buf, list);
drivers/media/pci/tw68/tw68-video.c
985
buf = list_entry(dev->active.next, struct tw68_buf, list);
drivers/media/pci/tw686x/tw686x-audio.c
284
ss; ss = ss->next, i++)
drivers/media/pci/tw686x/tw686x-audio.c
40
struct tw686x_audio_buf *next = NULL;
drivers/media/pci/tw686x/tw686x-audio.c
54
next = list_first_entry(&ac->buf_list,
drivers/media/pci/tw686x/tw686x-audio.c
56
list_move_tail(&next->list, &ac->buf_list);
drivers/media/pci/tw686x/tw686x-audio.c
58
ac->curr_bufs[pb] = next;
drivers/media/pci/tw686x/tw686x-audio.c
74
reg_write(dev, reg, next->dma);
drivers/media/pci/zoran/videocodec.c
146
h->list = a->next;
drivers/media/pci/zoran/videocodec.c
149
prev->next = a->next;
drivers/media/pci/zoran/videocodec.c
158
a = a->next;
drivers/media/pci/zoran/videocodec.c
160
h = h->next;
drivers/media/pci/zoran/videocodec.c
192
while (h->next)
drivers/media/pci/zoran/videocodec.c
193
h = h->next; // find the end
drivers/media/pci/zoran/videocodec.c
194
h->next = ptr;
drivers/media/pci/zoran/videocodec.c
21
struct attached_list *next;
drivers/media/pci/zoran/videocodec.c
233
codeclist_top = h->next;
drivers/media/pci/zoran/videocodec.c
237
prev->next = h->next;
drivers/media/pci/zoran/videocodec.c
245
h = h->next;
drivers/media/pci/zoran/videocodec.c
272
a = a->next;
drivers/media/pci/zoran/videocodec.c
274
h = h->next;
drivers/media/pci/zoran/videocodec.c
28
struct codec_list *next;
drivers/media/pci/zoran/videocodec.c
86
while (a->next)
drivers/media/pci/zoran/videocodec.c
87
a = a->next; // find end
drivers/media/pci/zoran/videocodec.c
88
a->next = ptr;
drivers/media/pci/zoran/videocodec.c
98
h = h->next;
drivers/media/pci/zoran/zoran_driver.c
935
buf = list_entry(zr->queued_bufs.next, struct zr_buffer, queue);
drivers/media/platform/amphion/vpu_helpers.h
35
void vpu_helper_get_kmp_next(const u8 *pattern, int *next, int size);
drivers/media/platform/amphion/vpu_helpers.h
36
int vpu_helper_kmp_search(u8 *s, int s_len, const u8 *p, int p_len, int *next);
drivers/media/platform/amphion/vpu_helpers.h
39
const u8 *p, int p_len, int *next);
drivers/media/platform/atmel/atmel-isi.c
191
isi->active = list_entry(isi->video_buffer_list.next,
drivers/media/platform/atmel/atmel-isi.c
330
desc = list_entry(isi->dma_desc_head.next,
drivers/media/platform/intel/pxa_camera.c
962
pcdev->active = list_entry(pcdev->capture.next,
drivers/media/platform/nxp/imx7-media-csi.c
1582
buf = list_entry(csi->ready_q.next, struct imx7_csi_vb2_buffer,
drivers/media/platform/nxp/imx7-media-csi.c
736
struct imx7_csi_vb2_buffer *done, *next;
drivers/media/platform/nxp/imx7-media-csi.c
751
next = imx7_csi_video_next_buf(csi);
drivers/media/platform/nxp/imx7-media-csi.c
752
if (next) {
drivers/media/platform/nxp/imx7-media-csi.c
753
dma_addr = vb2_dma_contig_plane_dma_addr(&next->vbuf.vb2_buf, 0);
drivers/media/platform/nxp/imx7-media-csi.c
754
csi->active_vb2_buf[csi->buf_num] = next;
drivers/media/platform/qcom/iris/iris_buffer.c
422
struct iris_buffer *buffer, *next;
drivers/media/platform/qcom/iris/iris_buffer.c
427
list_for_each_entry_safe(buffer, next, &buffers->list, list) {
drivers/media/platform/qcom/iris/iris_buffer.c
446
struct iris_buffer *buffer, *next;
drivers/media/platform/qcom/iris/iris_buffer.c
472
list_for_each_entry_safe(buffer, next, &buffers->list, list) {
drivers/media/platform/qcom/iris/iris_buffer.c
505
struct iris_buffer *buf, *next;
drivers/media/platform/qcom/iris/iris_buffer.c
531
list_for_each_entry_safe(buf, next, &buffers->list, list) {
drivers/media/platform/qcom/iris/iris_buffer.c
552
list_for_each_entry_safe(buf, next, &buffers->list, list) {
drivers/media/platform/qcom/iris/iris_buffer.c
577
struct iris_buffer *buffer, *next;
drivers/media/platform/qcom/iris/iris_buffer.c
580
list_for_each_entry_safe(buffer, next, &buffers->list, list) {
drivers/media/platform/qcom/iris/iris_buffer.c
621
struct iris_buffer *buffer, *next;
drivers/media/platform/qcom/iris/iris_buffer.c
636
list_for_each_entry_safe(buffer, next, &buffers->list, list) {
drivers/media/platform/qcom/iris/iris_vidc.c
247
struct iris_buffer *buf, *next;
drivers/media/platform/qcom/iris/iris_vidc.c
264
list_for_each_entry_safe(buf, next, &buffers->list, list)
drivers/media/platform/qcom/iris/iris_vidc.c
277
list_for_each_entry_safe(buf, next, &buffers->list, list)
drivers/media/platform/qcom/venus/helpers.c
105
struct intbuf *buf, *next;
drivers/media/platform/qcom/venus/helpers.c
114
list_for_each_entry_safe(buf, next, &inst->dpbbufs, list) {
drivers/media/platform/renesas/rcar-vin/rcar-dma.c
978
buf = list_entry(vin->buf_list.next, struct rvin_buffer, list);
drivers/media/platform/renesas/rcar_fdp1.c
1233
job->next = fdp1_peek_queued_field(ctx);
drivers/media/platform/renesas/rcar_fdp1.c
548
struct fdp1_field_buffer *next;
drivers/media/platform/renesas/rcar_fdp1.c
934
if (job->next)
drivers/media/platform/renesas/rcar_fdp1.c
935
fdp1_write(fdp1, job->next->addrs[0], FD1_RPF2_ADDR_Y);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-video.c
217
buf = list_entry(cru->buf_list.next,
drivers/media/platform/renesas/sh_vou.c
1065
vou_dev->active = list_entry(vou_dev->buf_list.next,
drivers/media/platform/renesas/sh_vou.c
1072
struct sh_vou_buffer *new = list_entry(vou_dev->active->list.next,
drivers/media/platform/renesas/sh_vou.c
1349
struct v4l2_subdev *sd = list_entry(v4l2_dev->subdevs.next,
drivers/media/platform/renesas/sh_vou.c
306
buf = list_entry(vou_dev->buf_list.next, struct sh_vou_buffer, list);
drivers/media/platform/renesas/sh_vou.c
316
buf = list_entry(buf->list.next, struct sh_vou_buffer, list);
drivers/media/platform/renesas/vsp1/vsp1_dl.c
1187
struct vsp1_dl_list *dl, *next;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
1192
list_for_each_entry_safe(dl, next, &dlm->free, list) {
drivers/media/platform/renesas/vsp1/vsp1_dl.c
822
struct vsp1_dl_list *next = list_next_entry(dl, chain);
drivers/media/platform/renesas/vsp1/vsp1_dl.c
824
dl->header->next_header = next->dma;
drivers/media/platform/renesas/vsp1/vsp1_drm.c
109
uif->sink = next;
drivers/media/platform/renesas/vsp1/vsp1_drm.c
549
struct vsp1_entity *next;
drivers/media/platform/renesas/vsp1/vsp1_drm.c
565
list_for_each_entry_safe(entity, next, &pipe->entities, list_pipe) {
drivers/media/platform/renesas/vsp1/vsp1_drm.c
68
struct vsp1_entity *next, unsigned int next_pad)
drivers/media/platform/renesas/vsp1/vsp1_drm.c
80
prev->sink = next;
drivers/media/platform/renesas/vsp1/vsp1_video.c
210
struct vsp1_vb2_buffer *next = NULL;
drivers/media/platform/renesas/vsp1/vsp1_video.c
228
next = list_first_entry(&video->irqqueue,
drivers/media/platform/renesas/vsp1/vsp1_video.c
240
return next;
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
737
cap->buf.curr = cap->buf.next;
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
738
cap->buf.next = NULL;
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
743
cap->buf.next = list_first_entry(&cap->buf.queue, struct rkisp1_buffer, queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
744
list_del(&cap->buf.next->queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
746
buff_addr = cap->buf.next->buff_addr;
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
966
if (cap->buf.next) {
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
967
vb2_buffer_done(&cap->buf.next->vb.vb2_buf, state);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
968
cap->buf.next = NULL;
drivers/media/platform/rockchip/rkisp1/rkisp1-common.h
341
struct rkisp1_buffer *next;
drivers/media/platform/samsung/exynos4-is/fimc-capture.c
1445
buf = list_entry(fimc->vid_cap.active_buf_q.next,
drivers/media/platform/samsung/exynos4-is/fimc-core.h
693
buf = list_entry(vid_cap->active_buf_q.next,
drivers/media/platform/samsung/exynos4-is/fimc-core.h
721
buf = list_entry(vid_cap->pending_buf_q.next,
drivers/media/platform/samsung/exynos4-is/fimc-lite.h
200
struct flite_buffer *buf = list_entry(dev->active_buf_q.next,
drivers/media/platform/samsung/exynos4-is/fimc-lite.h
215
struct flite_buffer *buf = list_entry(dev->pending_buf_q.next,
drivers/media/platform/samsung/exynos4-is/media-dev.c
391
p = list_entry(fmd->pipelines.next, typeof(*p), list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
136
b = list_entry(lh->next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
233
dst_buf = list_entry(ctx->dst_queue.next,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
273
src_buf = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
435
src_buf = list_entry(ctx->src_queue.next, struct s5p_mfc_buf,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
553
src_buf = list_entry(ctx->src_queue.next,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
592
src_buf = list_entry(ctx->src_queue.next,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
629
mb_entry = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1134
mb_entry = list_entry((&ctx->ref_queue)->next,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1154
dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1170
dst_mb = list_entry(ctx->dst_queue.next,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1209
src_mb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1221
dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1295
mb_entry = list_entry(ctx->src_queue.next, struct s5p_mfc_buf,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1307
mb_entry = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v5.c
1181
temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v5.c
1218
src_mb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v5.c
1238
dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v5.c
1257
temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v5.c
1276
dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v5.c
1302
temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v6.c
1973
temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v6.c
2017
src_mb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v6.c
2041
dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v6.c
2061
temp_vb = list_entry(ctx->src_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v6.c
2077
dst_mb = list_entry(ctx->dst_queue.next, struct s5p_mfc_buf, list);
drivers/media/platform/st/sti/delta/delta-v4l2.c
208
struct delta_dts *next;
drivers/media/platform/st/sti/delta/delta-v4l2.c
216
list_for_each_entry_safe(dts, next, &ctx->dts, list)
drivers/media/platform/st/stm32/stm32-dcmi.c
246
buf = list_entry(dcmi->buffers.next, struct dcmi_buf, list);
drivers/media/platform/st/stm32/stm32-dcmi.c
799
buf = list_entry(dcmi->buffers.next, struct dcmi_buf, list);
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
146
struct dcmipp_buf *active, *next;
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
438
vcap->next = list_first_entry(&vcap->buffers, typeof(*buf), list);
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
440
vcap->next->vb.vb2_buf.index, vcap->next, &vcap->next->addr);
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
442
dcmipp_start_capture(vcap, vcap->next);
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
588
vcap->next = buf;
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
689
if (!vcap->next && list_is_singular(&vcap->buffers)) {
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
699
vcap->next = NULL;
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
705
if (!vcap->next)
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
706
vcap->next = list_next_entry(vcap->active, list);
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
713
reg_write(vcap, DCMIPP_P0PPM0AR1, vcap->next->addr);
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
715
vcap->next->vb.vb2_buf.index, vcap->next, &vcap->next->addr);
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
787
swap(vcap->active, vcap->next);
drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c
181
unsigned int next;
drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c
184
next = !(reg & CSI_BUF_CTRL_DBS);
drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c
187
sun4i_csi_buffer_mark_done(csi, next, sequence);
drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c
190
return sun4i_csi_buffer_fill_slot(csi, next);
drivers/media/platform/ti/am437x/am437x-vpfe.c
1119
vpfe->next_frm = list_entry(vpfe->dma_queue.next,
drivers/media/platform/ti/am437x/am437x-vpfe.c
1891
vpfe->next_frm = list_entry(vpfe->dma_queue.next,
drivers/media/platform/ti/davinci/vpif_capture.c
207
common->cur_frm = common->next_frm = list_entry(common->dma_queue.next,
drivers/media/platform/ti/davinci/vpif_capture.c
298
common->next_frm = list_entry(common->dma_queue.next,
drivers/media/platform/ti/davinci/vpif_capture.c
344
common->next_frm = list_entry(common->dma_queue.next,
drivers/media/platform/ti/davinci/vpif_display.c
200
list_entry(common->dma_queue.next,
drivers/media/platform/ti/davinci/vpif_display.c
285
common->next_frm = list_entry(common->dma_queue.next,
drivers/media/platform/ti/davinci/vpif_display.c
308
common->next_frm = list_entry(common->dma_queue.next,
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
651
buf = list_entry(dma->queue.next, struct ti_csi2rx_buffer, list);
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
861
buf = list_entry(dma->queue.next, struct ti_csi2rx_buffer, list);
drivers/media/platform/ti/omap/omap_vout.c
1005
vout->next_frm = vout->cur_frm = list_entry(vout->dma_queue.next,
drivers/media/platform/ti/omap/omap_vout.c
543
vout->next_frm = list_entry(vout->dma_queue.next,
drivers/media/platform/ti/omap3isp/isp.c
1063
struct isp_reg *next = reg_list;
drivers/media/platform/ti/omap3isp/isp.c
1065
for (; next->reg != ISP_TOK_TERM; next++)
drivers/media/platform/ti/omap3isp/isp.c
1066
next->val = isp_reg_readl(isp, next->mmio_range, next->reg);
drivers/media/platform/ti/omap3isp/isp.c
1078
struct isp_reg *next = reg_list;
drivers/media/platform/ti/omap3isp/isp.c
1080
for (; next->reg != ISP_TOK_TERM; next++)
drivers/media/platform/ti/omap3isp/isp.c
1081
isp_reg_writel(isp, next->val, next->mmio_range, next->reg);
drivers/media/platform/ti/vpe/vip.c
2208
buf = list_entry(stream->post_bufs.next,
drivers/media/platform/ti/vpe/vip.c
2217
buf = list_entry(stream->vidq.next, struct vip_buffer, list);
drivers/media/platform/ti/vpe/vip.c
2367
buf = list_entry(stream->vidq.next,
drivers/media/platform/ti/vpe/vip.c
884
buf = list_entry(stream->dropq.next,
drivers/media/platform/ti/vpe/vip.c
891
buf = list_entry(stream->vidq.next,
drivers/media/platform/ti/vpe/vpdma.c
489
list->next = list->buf.addr;
drivers/media/platform/ti/vpe/vpdma.c
503
list->next = list->buf.addr;
drivers/media/platform/ti/vpe/vpdma.c
515
list->next = NULL;
drivers/media/platform/ti/vpe/vpdma.c
538
list_size = (list->next - list->buf.addr) >> 4;
drivers/media/platform/ti/vpe/vpdma.c
657
cfd = list->next;
drivers/media/platform/ti/vpe/vpdma.c
666
list->next = cfd + 1;
drivers/media/platform/ti/vpe/vpdma.c
686
cfd = list->next;
drivers/media/platform/ti/vpe/vpdma.c
695
list->next = cfd + 1;
drivers/media/platform/ti/vpe/vpdma.c
724
ctd = list->next;
drivers/media/platform/ti/vpe/vpdma.c
733
list->next = ctd + 1;
drivers/media/platform/ti/vpe/vpdma.c
749
ctd = list->next;
drivers/media/platform/ti/vpe/vpdma.c
758
list->next = ctd + 1;
drivers/media/platform/ti/vpe/vpdma.c
858
dtd = list->next;
drivers/media/platform/ti/vpe/vpdma.c
877
list->next = dtd + 1;
drivers/media/platform/ti/vpe/vpdma.c
927
dtd = list->next;
drivers/media/platform/ti/vpe/vpdma.c
949
list->next = dtd + 1;
drivers/media/platform/ti/vpe/vpdma.h
27
void *next;
drivers/media/platform/via/via-camera.c
325
return list_entry(cam->buffer_queue.next, struct via_buffer, queue);
drivers/media/test-drivers/vidtv/vidtv_channel.c
132
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_channel.c
133
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
135
head->next = s302m;
drivers/media/test-drivers/vidtv/vidtv_channel.c
194
curr = curr->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
197
cur_chnl = cur_chnl->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
244
curr = curr->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
247
cur_chnl = cur_chnl->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
295
curr = curr->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
298
cur_chnl = cur_chnl->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
352
s = s->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
360
cur_chnl = cur_chnl->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
371
e = e->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
404
prev_e->next = curr_e;
drivers/media/test-drivers/vidtv/vidtv_channel.c
409
desc = desc->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
411
s = s->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
42
curr = curr->next;
drivers/media/test-drivers/vidtv/vidtv_channel.c
546
curr = curr->next;
drivers/media/test-drivers/vidtv/vidtv_channel.h
62
struct vidtv_channel *next;
drivers/media/test-drivers/vidtv/vidtv_encoder.h
162
struct vidtv_encoder *next;
drivers/media/test-drivers/vidtv/vidtv_encoder.h
28
struct vidtv_access_unit *next;
drivers/media/test-drivers/vidtv/vidtv_mux.c
105
p = p->next;
drivers/media/test-drivers/vidtv/vidtv_mux.c
309
au = au->next;
drivers/media/test-drivers/vidtv/vidtv_mux.c
339
e = e->next;
drivers/media/test-drivers/vidtv/vidtv_mux.c
343
cur_chnl = cur_chnl->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1039
p = p->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1083
stream->next = NULL;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1086
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_psi.c
1087
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1089
head->next = stream;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1102
curr_stream = curr_stream->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1139
program = program->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1246
table_descriptor = table_descriptor->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1269
stream_descriptor = stream_descriptor->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1272
stream = stream->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1411
service_desc = service_desc->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1414
service = service->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1459
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_psi.c
1460
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1462
head->next = service;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1476
curr = curr->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1523
program = program->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1530
for (program = pat->program; program; program = program->next) {
drivers/media/test-drivers/vidtv/vidtv_psi.c
1601
t = t->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1734
table_descriptor = table_descriptor->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1765
transport_descriptor = transport_descriptor->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1768
transport = transport->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1788
curr_t = curr_t->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1824
e = e->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1957
event_descriptor = event_descriptor->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1960
event = event->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
2025
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_psi.c
2026
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
2028
head->next = e;
drivers/media/test-drivers/vidtv/vidtv_psi.c
2041
curr_e = curr_e->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
272
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_psi.c
273
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
275
head->next = desc;
drivers/media/test-drivers/vidtv/vidtv_psi.c
404
head_e = head_e->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
420
prev_e->next = curr_e;
drivers/media/test-drivers/vidtv/vidtv_psi.c
423
entry = entry->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
545
curr->next = NULL;
drivers/media/test-drivers/vidtv/vidtv_psi.c
549
prev->next = curr;
drivers/media/test-drivers/vidtv/vidtv_psi.c
552
desc = desc->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
567
curr = curr->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
587
sl_entry = sl_entry->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
620
desc = desc->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
731
serv_list_entry = serv_list_entry->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
850
s = s->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
883
s = s->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
906
program->next = NULL;
drivers/media/test-drivers/vidtv/vidtv_psi.c
909
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_psi.c
910
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
912
head->next = program;
drivers/media/test-drivers/vidtv/vidtv_psi.c
926
curr = curr->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
948
program = program->next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
101
struct vidtv_psi_desc *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
110
struct vidtv_psi_desc_service_list_entry *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
118
struct vidtv_psi_desc *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
129
struct vidtv_psi_desc *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
168
struct vidtv_psi_table_pat_program *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
193
struct vidtv_psi_table_sdt_service *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
235
struct vidtv_psi_table_pmt_stream *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
52
struct vidtv_psi_desc *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
63
struct vidtv_psi_desc *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
657
struct vidtv_psi_table_transport *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
727
struct vidtv_psi_table_eit_event *next;
drivers/media/test-drivers/vidtv/vidtv_psi.h
79
struct vidtv_psi_desc *next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
156
while (head->next)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
157
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
159
head->next = au;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
172
head = head->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
192
sync_au = sync_au->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
216
au = au->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
217
sync_au = sync_au->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
229
au = au->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
230
sync_au = sync_au->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
386
au = au->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
431
au = au->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
493
while (args.head->next)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
494
args.head = args.head->next;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
496
args.head->next = e;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
499
e->next = NULL;
drivers/media/test-drivers/vivid/vivid-kthread-cap.c
252
vid_out_buf = list_entry(out_dev->vid_out_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-cap.c
601
vid_cap_buf = list_entry(dev->vid_cap_active.next, struct vivid_buffer, list);
drivers/media/test-drivers/vivid/vivid-kthread-cap.c
607
vbi_cap_buf = list_entry(dev->vbi_cap_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-cap.c
613
meta_cap_buf = list_entry(dev->meta_cap_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-cap.c
866
buf = list_entry(dev->vid_cap_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-cap.c
881
buf = list_entry(dev->vbi_cap_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-cap.c
896
buf = list_entry(dev->meta_cap_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-out.c
313
buf = list_entry(dev->vid_out_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-out.c
328
buf = list_entry(dev->vbi_out_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-out.c
343
buf = list_entry(dev->meta_out_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-out.c
64
vid_out_buf = list_entry(dev->vid_out_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-out.c
71
vbi_out_buf = list_entry(dev->vbi_out_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-out.c
76
meta_out_buf = list_entry(dev->meta_out_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-touch.c
184
buf = list_entry(dev->touch_cap_active.next,
drivers/media/test-drivers/vivid/vivid-kthread-touch.c
20
tch_cap_buf = list_entry(dev->touch_cap_active.next,
drivers/media/test-drivers/vivid/vivid-sdr-cap.c
318
buf = list_entry(dev->sdr_cap_active.next,
drivers/media/test-drivers/vivid/vivid-sdr-cap.c
98
sdr_cap_buf = list_entry(dev->sdr_cap_active.next,
drivers/media/usb/airspy/airspy.c
223
buf = list_entry(s->queued_bufs.next,
drivers/media/usb/airspy/airspy.c
451
buf = list_entry(s->queued_bufs.next,
drivers/media/usb/au0828/au0828-video.c
402
*buf = list_entry(dma_q->active.next, struct au0828_buffer, list);
drivers/media/usb/au0828/au0828-video.c
477
*buf = list_entry(dma_q->active.next, struct au0828_buffer, list);
drivers/media/usb/au0828/au0828-video.c
871
buf = list_entry(vidq->active.next, struct au0828_buffer, list);
drivers/media/usb/au0828/au0828-video.c
901
buf = list_entry(vbiq->active.next, struct au0828_buffer, list);
drivers/media/usb/cx231xx/cx231xx-417.c
1249
buf = list_entry(dma_q->active.next,
drivers/media/usb/cx231xx/cx231xx-417.c
1304
buf = list_entry(dma_q->active.next, struct cx231xx_buffer, list);
drivers/media/usb/cx231xx/cx231xx-core.c
108
list_del(&ops->next);
drivers/media/usb/cx231xx/cx231xx-core.c
118
list_for_each_entry(ops, &cx231xx_extension_devlist, next) {
drivers/media/usb/cx231xx/cx231xx-core.c
130
list_for_each_entry(ops, &cx231xx_extension_devlist, next) {
drivers/media/usb/cx231xx/cx231xx-core.c
88
list_add_tail(&ops->next, &cx231xx_extension_devlist);
drivers/media/usb/cx231xx/cx231xx-vbi.c
577
*buf = list_entry(dma_q->active.next, struct cx231xx_buffer, list);
drivers/media/usb/cx231xx/cx231xx-video.c
245
*buf = list_entry(dma_q->active.next, struct cx231xx_buffer, list);
drivers/media/usb/cx231xx/cx231xx.h
728
struct list_head next;
drivers/media/usb/em28xx/em28xx-core.c
1077
list_add_tail(&ops->next, &em28xx_extension_devlist);
drivers/media/usb/em28xx/em28xx-core.c
1103
list_del(&ops->next);
drivers/media/usb/em28xx/em28xx-core.c
1115
list_for_each_entry(ops, &em28xx_extension_devlist, next) {
drivers/media/usb/em28xx/em28xx-core.c
1130
list_for_each_entry(ops, &em28xx_extension_devlist, next) {
drivers/media/usb/em28xx/em28xx-core.c
1147
list_for_each_entry(ops, &em28xx_extension_devlist, next) {
drivers/media/usb/em28xx/em28xx-core.c
1164
list_for_each_entry(ops, &em28xx_extension_devlist, next) {
drivers/media/usb/em28xx/em28xx-video.c
1165
buf = list_entry(vidq->active.next, struct em28xx_buffer, list);
drivers/media/usb/em28xx/em28xx-video.c
1200
buf = list_entry(vbiq->active.next, struct em28xx_buffer, list);
drivers/media/usb/em28xx/em28xx-video.c
598
buf = list_entry(dma_q->active.next, struct em28xx_buffer, list);
drivers/media/usb/em28xx/em28xx.h
780
struct list_head next;
drivers/media/usb/hackrf/hackrf.c
447
buffer = list_entry(buffer_list->next, struct hackrf_buffer, list);
drivers/media/usb/hdpvr/hdpvr-video.c
109
for (p = q->next; p != q;) {
drivers/media/usb/hdpvr/hdpvr-video.c
116
tmp = p->next;
drivers/media/usb/hdpvr/hdpvr-video.c
201
buf = list_entry(dev->free_buff_list.next, struct hdpvr_buffer,
drivers/media/usb/hdpvr/hdpvr-video.c
242
buf = list_entry(dev->rec_buff_list.next, struct hdpvr_buffer,
drivers/media/usb/msi2500/msi2500.c
153
buf = list_entry(dev->queued_bufs.next, struct msi2500_frame_buf, list);
drivers/media/usb/msi2500/msi2500.c
554
buf = list_entry(dev->queued_bufs.next,
drivers/media/usb/pvrusb2/pvrusb2-io.c
382
while ((lp = sp->queued_list.next) != &sp->queued_list) {
drivers/media/usb/pvrusb2/pvrusb2-io.c
549
struct list_head *lp = sp->idle_list.next;
drivers/media/usb/pvrusb2/pvrusb2-io.c
556
struct list_head *lp = sp->ready_list.next;
drivers/media/usb/pwc/pwc-if.c
195
buf = list_entry(pdev->queued_bufs.next, struct pwc_frame_buf, list);
drivers/media/usb/pwc/pwc-if.c
562
buf = list_entry(pdev->queued_bufs.next, struct pwc_frame_buf,
drivers/media/usb/s2255/s2255drv.c
555
buf = list_entry(vc->buf_list.next,
drivers/media/usb/ttusb-dec/ttusb_dec.c
1597
while ((item = dec->urb_frame_list.next) != &dec->urb_frame_list) {
drivers/media/usb/ttusb-dec/ttusb_dec.c
1615
while ((item = dec->filter_info_list.next) != &dec->filter_info_list) {
drivers/media/usb/ttusb-dec/ttusb_dec.c
569
for (item = dec->filter_info_list.next; item != &dec->filter_info_list;
drivers/media/usb/ttusb-dec/ttusb_dec.c
570
item = item->next) {
drivers/media/usb/ttusb-dec/ttusb_dec.c
779
if ((item = dec->urb_frame_list.next) != &dec->urb_frame_list) {
drivers/media/usb/uvc/uvc_ctrl.c
1175
int next, int next_compound)
drivers/media/usb/uvc/uvc_ctrl.c
1190
if (map->id == v4l2_id && !next && !next_compound) {
drivers/media/usb/uvc/uvc_ctrl.c
1199
next_compound : next)) {
drivers/media/usb/uvc/uvc_ctrl.c
1212
int next = v4l2_id & V4L2_CTRL_FLAG_NEXT_CTRL;
drivers/media/usb/uvc/uvc_ctrl.c
1222
__uvc_find_control(entity, v4l2_id, mapping, &ctrl, next,
drivers/media/usb/uvc/uvc_ctrl.c
1224
if (ctrl && !next && !next_compound)
drivers/media/usb/uvc/uvc_ctrl.c
1228
if (!ctrl && !next && !next_compound)
drivers/media/usb/uvc/uvc_driver.c
1528
if (forward->chain.next || forward->chain.prev) {
drivers/media/usb/uvc/uvc_driver.c
1643
if (term->chain.next || term->chain.prev) {
drivers/media/usb/uvc/uvc_driver.c
1700
if (entity->chain.next || entity->chain.prev) {
drivers/media/usb/uvc/uvc_driver.c
1887
if (term->chain.next || term->chain.prev)
drivers/media/v4l2-core/v4l2-async.c
674
if (!notifier || !notifier->waiting_list.next)
drivers/media/v4l2-core/v4l2-async.c
888
if (!sd->async_list.next)
drivers/media/v4l2-core/v4l2-async.c
900
if (sd->asc_list.next) {
drivers/media/v4l2-core/v4l2-async.c
908
sd->async_list.next = NULL;
drivers/media/v4l2-core/v4l2-ctrls-api.c
31
u32 next;
drivers/media/v4l2-core/v4l2-ctrls-api.c
341
h->next = 0;
drivers/media/v4l2-core/v4l2-ctrls-api.c
375
mref->helper->next = i;
drivers/media/v4l2-core/v4l2-ctrls-api.c
505
idx = helpers[idx].next;
drivers/media/v4l2-core/v4l2-ctrls-api.c
665
tmp_idx = helpers[tmp_idx].next;
drivers/media/v4l2-core/v4l2-ctrls-api.c
690
idx = helpers[idx].next;
drivers/media/v4l2-core/v4l2-ctrls-api.c
711
idx = helpers[idx].next;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1823
ref = ref->next;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1918
new_ref->next = hdl->buckets[bucket];
drivers/media/v4l2-core/v4l2-device.c
90
struct v4l2_subdev *sd, *next;
drivers/media/v4l2-core/v4l2-device.c
99
list_for_each_entry_safe(sd, next, &v4l2_dev->subdevs, list) {
drivers/media/v4l2-core/v4l2-subdev.c
1722
if (!sd->async_subdev_endpoint_list.next)
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1014
ccfifo_writel(emc, (next->burst_regs[EMC_MRW6_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1018
ccfifo_writel(emc, (next->burst_regs[EMC_MRW14_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
102
next->ptfv_list[dqs] = \
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1025
(next->burst_regs[EMC_MRW7_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1030
(next->burst_regs[EMC_MRW15_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
104
(next->ptfv_list[dqs] * \
drivers/memory/tegra/tegra210-emc-cc-r21021.c
105
next->ptfv_list[w])) / \
drivers/memory/tegra/tegra210-emc-cc-r21021.c
106
(next->ptfv_list[w] + 1); \
drivers/memory/tegra/tegra210-emc-cc-r21021.c
109
__stringify(dev), nval, next->ptfv_list[dqs]); \
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1158
div_o3(1000 * next->dram_timings[T_PDEX],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1166
next->dram_timings[T_PDEX]);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1171
delay = div_o3(1000 * next->dram_timings[T_PDEX],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1251
ccfifo_writel(emc, next->emc_mrw2, EMC_MRW2, 0);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1252
ccfifo_writel(emc, next->emc_mrw, EMC_MRW, 0);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1254
ccfifo_writel(emc, next->emc_mrw4, EMC_MRW4, 0);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1257
ccfifo_writel(emc, next->emc_emrs &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1259
ccfifo_writel(emc, next->emc_emrs2 &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1261
ccfifo_writel(emc, next->emc_mrs |
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1320
next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1349
ccfifo_writel(emc, next->burst_regs[EMC_ZCAL_INTERVAL_INDEX],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1353
ccfifo_writel(emc, next->burst_regs[EMC_CFG_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1370
next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1375
next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1407
if (next->rate > last->rate) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1408
for (i = 0; i < next->num_up_down; i++)
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1409
mc_writel(emc->mc, next->la_scale_regs[i],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1423
emc_writel(emc, next->burst_regs[EMC_ZCAL_WAIT_CNT_INDEX],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1425
emc_writel(emc, next->burst_regs[EMC_ZCAL_INTERVAL_INDEX],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1436
emc_writel(emc, next->burst_regs[EMC_MRS_WAIT_CNT_INDEX],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1439
emc_writel(emc, next->burst_regs[EMC_ZCAL_WAIT_CNT_INDEX],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1451
emc_writel(emc, next->burst_regs[EMC_CFG_INDEX], EMC_CFG);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1453
emc_writel(emc, next->emc_fdpd_ctrl_cmd_no_ramp,
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1455
emc_writel(emc, next->emc_sel_dpd_ctrl, EMC_SEL_DPD_CTRL);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1465
next->burst_regs[EMC_PMACRO_AUTOCAL_CFG_COMMON_INDEX],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1496
if (next->burst_regs[EMC_CFG_DIG_DLL_INDEX] & EMC_CFG_DIG_DLL_CFG_DLL_EN) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1508
emc_writel(emc, next->emc_auto_cal_config, EMC_AUTO_CAL_CONFIG);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
184
struct tegra210_emc_timing *next)
drivers/memory/tegra/tegra210-emc-cc-r21021.c
190
u32 i, samples = next->ptfv_list[PTFV_DVFS_SAMPLES_INDEX];
drivers/memory/tegra/tegra210-emc-cc-r21021.c
194
if (!next->periodic_training)
drivers/memory/tegra/tegra210-emc-cc-r21021.c
199
(next->ptfv_list[PTFV_CONFIG_CTRL_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
207
__COPY_EMA(next, last, idx);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
211
__MOVAVG(next, idx) = 0;
drivers/memory/tegra/tegra210-emc-cc-r21021.c
225
over |= tegra210_emc_compare_update_delay(next,
drivers/memory/tegra/tegra210-emc-cc-r21021.c
226
__MOVAVG_AC(next, idx), idx);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
235
over |= tegra210_emc_compare_update_delay(next,
drivers/memory/tegra/tegra210-emc-cc-r21021.c
236
__MOVAVG_AC(next, idx), idx);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
359
struct tegra210_emc_timing *fake, *last = emc->last, *next = emc->next;
drivers/memory/tegra/tegra210-emc-cc-r21021.c
364
u32 tFC_lpddr4 = 1000 * next->dram_timings[T_FC_LPDDR4];
drivers/memory/tegra/tegra210-emc-cc-r21021.c
388
if ((next->burst_regs[EMC_ZCAL_INTERVAL_INDEX] != 0 &&
drivers/memory/tegra/tegra210-emc-cc-r21021.c
394
opt_dll_mode = tegra210_emc_get_dll_state(next);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
396
if ((next->burst_regs[EMC_FBIO_CFG5_INDEX] & BIT(25)) &&
drivers/memory/tegra/tegra210-emc-cc-r21021.c
404
dst_clk_period = 1000000000 / next->rate;
drivers/memory/tegra/tegra210-emc-cc-r21021.c
417
emc_cfg = next->burst_regs[EMC_CFG_INDEX];
drivers/memory/tegra/tegra210-emc-cc-r21021.c
420
emc_sel_dpd_ctrl = next->emc_sel_dpd_ctrl;
drivers/memory/tegra/tegra210-emc-cc-r21021.c
432
emc_dbg(emc, INFO, "DLL clksrc: 0x%08x\n", next->dll_clk_src);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
434
next->rate);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
460
emc_auto_cal_config = next->emc_auto_cal_config;
drivers/memory/tegra/tegra210-emc-cc-r21021.c
476
if (next->periodic_training) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
477
tegra210_emc_reset_dram_clktree_values(next);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
489
if (periodic_compensation_handler(emc, DVFS_SEQUENCE, fake, next))
drivers/memory/tegra/tegra210-emc-cc-r21021.c
499
emc_writel(emc, next->emc_fdpd_ctrl_cmd_no_ramp &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
504
((next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
508
((next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
513
(next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
516
(next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
536
((next->burst_regs[EMC_PMACRO_DATA_PAD_TX_CTRL_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
540
((next->burst_regs[EMC_PMACRO_DATA_PAD_TX_CTRL_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
543
next->burst_regs[EMC_PMACRO_DATA_PAD_TX_CTRL_INDEX];
drivers/memory/tegra/tegra210-emc-cc-r21021.c
570
if (next->burst_regs[EMC_CFG_DIG_DLL_INDEX] &
drivers/memory/tegra/tegra210-emc-cc-r21021.c
587
emc_writel(emc, next->emc_auto_cal_config2, EMC_AUTO_CAL_CONFIG2);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
588
emc_writel(emc, next->emc_auto_cal_config3, EMC_AUTO_CAL_CONFIG3);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
589
emc_writel(emc, next->emc_auto_cal_config4, EMC_AUTO_CAL_CONFIG4);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
590
emc_writel(emc, next->emc_auto_cal_config5, EMC_AUTO_CAL_CONFIG5);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
591
emc_writel(emc, next->emc_auto_cal_config6, EMC_AUTO_CAL_CONFIG6);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
592
emc_writel(emc, next->emc_auto_cal_config7, EMC_AUTO_CAL_CONFIG7);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
593
emc_writel(emc, next->emc_auto_cal_config8, EMC_AUTO_CAL_CONFIG8);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
609
emc_writel(emc, next->emc_cfg_2, EMC_CFG_2);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
620
zq_wait_long = max(next->min_mrs_wait,
drivers/memory/tegra/tegra210-emc-cc-r21021.c
673
next->burst_regs[EMC_RP_INDEX]);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
728
mr13_flip_fspwr = (next->emc_mrw3 & 0xffffff3f) | 0x80;
drivers/memory/tegra/tegra210-emc-cc-r21021.c
729
mr13_flip_fspop = (next->emc_mrw3 & 0xffffff3f) | 0x00;
drivers/memory/tegra/tegra210-emc-cc-r21021.c
731
mr13_flip_fspwr = (next->emc_mrw3 & 0xffffff3f) | 0x40;
drivers/memory/tegra/tegra210-emc-cc-r21021.c
732
mr13_flip_fspop = (next->emc_mrw3 & 0xffffff3f) | 0xc0;
drivers/memory/tegra/tegra210-emc-cc-r21021.c
737
emc_writel(emc, next->emc_mrw, EMC_MRW);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
738
emc_writel(emc, next->emc_mrw2, EMC_MRW2);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
748
for (i = 0; i < next->num_burst; i++) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
755
value = next->burst_regs[i];
drivers/memory/tegra/tegra210-emc-cc-r21021.c
78
({ next->ptfv_list[(dev)] = \
drivers/memory/tegra/tegra210-emc-cc-r21021.c
79
next->ptfv_list[(dev)] / \
drivers/memory/tegra/tegra210-emc-cc-r21021.c
80
next->ptfv_list[PTFV_DVFS_SAMPLES_INDEX]; })
drivers/memory/tegra/tegra210-emc-cc-r21021.c
816
tegra210_emc_adjust_timing(emc, next);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
820
(next->run_clocks & EMC_MRW_MRW_OP_MASK);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
827
for (i = 0; i < next->num_burst_per_ch; i++) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
852
next->burst_reg_per_ch[i], burst[i].offset);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
854
next->burst_reg_per_ch[i],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
86
({ next->ptfv_list[(dev)] += \
drivers/memory/tegra/tegra210-emc-cc-r21021.c
861
for (i = 0; i < next->vref_num; i++) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
872
next->vref_perch_regs[i], vref[i].offset);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
873
emc_channel_writel(emc, vref[i].bank, next->vref_perch_regs[i],
drivers/memory/tegra/tegra210-emc-cc-r21021.c
880
for (i = 0; i < next->num_trim; i++) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
897
value = tegra210_emc_compensate(next, offsets[i]);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
905
next->trim_regs[i], offsets[i]);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
906
emc_writel(emc, next->trim_regs[i], offsets[i]);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
913
for (i = 0; i < next->num_trim_per_ch; i++) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
937
value = tegra210_emc_compensate(next, offset);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
945
next->trim_perch_regs[i], offset);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
947
next->trim_perch_regs[i], offset);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
953
for (i = 0; i < next->num_mc_regs; i++) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
955
u32 *values = next->burst_mc_regs;
drivers/memory/tegra/tegra210-emc-cc-r21021.c
963
if (next->rate < last->rate) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
968
for (i = 0; i < next->num_up_down; i++) {
drivers/memory/tegra/tegra210-emc-cc-r21021.c
970
next->la_scale_regs[i], la[i]);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
971
mc_writel(emc->mc, next->la_scale_regs[i], la[i]);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
984
value = next->burst_regs[EMC_ZCAL_WAIT_CNT_INDEX];
drivers/memory/tegra/tegra210-emc-core.c
1011
delta[0] = 128 * (next->current_dram_clktree[C0D1U0] -
drivers/memory/tegra/tegra210-emc-core.c
1012
next->trained_dram_clktree[C0D1U0]);
drivers/memory/tegra/tegra210-emc-core.c
1013
delta[1] = 128 * (next->current_dram_clktree[C0D1U1] -
drivers/memory/tegra/tegra210-emc-core.c
1014
next->trained_dram_clktree[C0D1U1]);
drivers/memory/tegra/tegra210-emc-core.c
1015
delta[2] = 128 * (next->current_dram_clktree[C1D1U0] -
drivers/memory/tegra/tegra210-emc-core.c
1016
next->trained_dram_clktree[C1D1U0]);
drivers/memory/tegra/tegra210-emc-core.c
1017
delta[3] = 128 * (next->current_dram_clktree[C1D1U1] -
drivers/memory/tegra/tegra210-emc-core.c
1018
next->trained_dram_clktree[C1D1U1]);
drivers/memory/tegra/tegra210-emc-core.c
1026
if ((delta_taps[i] > next->tree_margin) ||
drivers/memory/tegra/tegra210-emc-core.c
1027
(delta_taps[i] < (-1 * next->tree_margin))) {
drivers/memory/tegra/tegra210-emc-core.c
1170
value = emc->next->burst_regs[EMC_DLL_CFG_0_INDEX];
drivers/memory/tegra/tegra210-emc-core.c
1176
if (emc->next->rate >= 400000 && emc->next->rate < 600000)
drivers/memory/tegra/tegra210-emc-core.c
1178
else if (emc->next->rate >= 600000 && emc->next->rate < 800000)
drivers/memory/tegra/tegra210-emc-core.c
1180
else if (emc->next->rate >= 800000 && emc->next->rate < 1000000)
drivers/memory/tegra/tegra210-emc-core.c
1182
else if (emc->next->rate >= 1000000 && emc->next->rate < 1200000)
drivers/memory/tegra/tegra210-emc-core.c
1231
timing = emc->next;
drivers/memory/tegra/tegra210-emc-core.c
1345
entry = emc->next;
drivers/memory/tegra/tegra210-emc-core.c
1555
emc->next = timing;
drivers/memory/tegra/tegra210-emc-core.c
62
next->trim_regs[EMC_PMACRO_OB_DDLL_LONG_DQ_RANK ## \
drivers/memory/tegra/tegra210-emc-core.c
69
next->trim_perch_regs[EMC ## chan ## \
drivers/memory/tegra/tegra210-emc-core.c
725
if (emc->next->periodic_training)
drivers/memory/tegra/tegra210-emc-core.c
734
u32 dll_setting = emc->next->dll_clk_src;
drivers/memory/tegra/tegra210-emc-core.c
761
if (emc->next->clk_out_enb_x_0_clk_enb_emc_dll)
drivers/memory/tegra/tegra210-emc-core.c
798
emc->next = &timings[index];
drivers/memory/tegra/tegra210-emc-core.c
895
u32 tegra210_emc_get_dll_state(struct tegra210_emc_timing *next)
drivers/memory/tegra/tegra210-emc-core.c
897
if (next->emc_emrs & 0x1)
drivers/memory/tegra/tegra210-emc-core.c
940
u32 tegra210_emc_compensate(struct tegra210_emc_timing *next, u32 offset)
drivers/memory/tegra/tegra210-emc-core.c
942
u32 temp = 0, rate = next->rate / 1000;
drivers/memory/tegra/tegra210-emc-core.c
973
delta[0] = 128 * (next->current_dram_clktree[C0D0U0] -
drivers/memory/tegra/tegra210-emc-core.c
974
next->trained_dram_clktree[C0D0U0]);
drivers/memory/tegra/tegra210-emc-core.c
975
delta[1] = 128 * (next->current_dram_clktree[C0D0U1] -
drivers/memory/tegra/tegra210-emc-core.c
976
next->trained_dram_clktree[C0D0U1]);
drivers/memory/tegra/tegra210-emc-core.c
977
delta[2] = 128 * (next->current_dram_clktree[C1D0U0] -
drivers/memory/tegra/tegra210-emc-core.c
978
next->trained_dram_clktree[C1D0U0]);
drivers/memory/tegra/tegra210-emc-core.c
979
delta[3] = 128 * (next->current_dram_clktree[C1D0U1] -
drivers/memory/tegra/tegra210-emc-core.c
980
next->trained_dram_clktree[C1D0U1]);
drivers/memory/tegra/tegra210-emc-core.c
988
if ((delta_taps[i] > next->tree_margin) ||
drivers/memory/tegra/tegra210-emc-core.c
989
(delta_taps[i] < (-1 * next->tree_margin))) {
drivers/memory/tegra/tegra210-emc.h
1005
u32 tegra210_emc_compensate(struct tegra210_emc_timing *next, u32 offset);
drivers/memory/tegra/tegra210-emc.h
911
struct tegra210_emc_timing *next;
drivers/memory/tegra/tegra210-emc.h
997
u32 tegra210_emc_get_dll_state(struct tegra210_emc_timing *next);
drivers/message/fusion/mptbase.c
1126
mpt_add_chain(void *pAddr, u8 next, u16 length, dma_addr_t dma_addr)
drivers/message/fusion/mptbase.c
1132
pChain->NextChainOffset = next;
drivers/message/fusion/mptbase.c
1146
mpt_add_chain_64bit(void *pAddr, u8 next, u16 length, dma_addr_t dma_addr)
drivers/message/fusion/mptbase.c
1155
pChain->NextChainOffset = next;
drivers/message/fusion/mptbase.c
907
mf = list_entry(ioc->FreeQ.next, MPT_FRAME_HDR,
drivers/message/fusion/mptbase.h
595
typedef void (*MPT_ADD_CHAIN)(void *pAddr, u8 next, u16 length,
drivers/message/fusion/mptbase.h
808
void *next;
drivers/message/fusion/mptsas.c
1196
target_reset_list = list_entry(head->next,
drivers/message/fusion/mptsas.c
1271
target_reset_list = list_entry(head->next,
drivers/message/fusion/mptsas.c
586
struct mptsas_device_info *sas_info, *next;
drivers/message/fusion/mptsas.c
595
list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list,
drivers/message/fusion/mptsas.c
688
struct mptsas_device_info *sas_info, *next;
drivers/message/fusion/mptsas.c
750
list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list,
drivers/message/fusion/mptsas.c
818
struct mptsas_device_info *sas_info, *next;
drivers/message/fusion/mptsas.c
823
list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list,
drivers/message/fusion/mptsas.c
838
struct mptsas_device_info *sas_info, *next;
drivers/message/fusion/mptsas.c
841
list_for_each_entry_safe(sas_info, next, &ioc->sas_device_info_list,
drivers/message/fusion/mptscsih.c
145
chainBuf = list_entry(ioc->FreeChainQ.next, MPT_FRAME_HDR,
drivers/message/fusion/mptscsih.c
1453
int next;
drivers/message/fusion/mptscsih.c
1464
next = ioc->ChainToChain[chain_idx];
drivers/message/fusion/mptscsih.c
1482
chain_idx = next;
drivers/mfd/ipaq-micro.c
101
micro->msg = list_entry(micro->queue.next,
drivers/mfd/twl4030-power.c
200
u8 delay, u8 next)
drivers/mfd/twl4030-power.c
214
err = twl4030_write_script_byte(address++, next);
drivers/misc/altera-stapl/altera.c
116
struct altera_procinfo *next;
drivers/misc/altera-stapl/altera.c
2351
procptr->next = NULL;
drivers/misc/altera-stapl/altera.c
2358
while (tmpptr->next != NULL)
drivers/misc/altera-stapl/altera.c
2359
tmpptr = tmpptr->next;
drivers/misc/altera-stapl/altera.c
2360
tmpptr->next = procptr;
drivers/misc/altera-stapl/altera.c
2457
proc_list = procptr->next;
drivers/misc/genwqe/card_dev.c
192
struct list_head *node, *next;
drivers/misc/genwqe/card_dev.c
197
list_for_each_safe(node, next, &cfile->map_list) {
drivers/misc/genwqe/card_dev.c
230
struct list_head *node, *next;
drivers/misc/genwqe/card_dev.c
234
list_for_each_safe(node, next, &cfile->pin_list) {
drivers/misc/ibmasm/command.c
71
struct list_head *next;
drivers/misc/ibmasm/command.c
76
next = sp->command_queue.next;
drivers/misc/ibmasm/command.c
77
list_del_init(next);
drivers/misc/ibmasm/command.c
78
cmd = list_entry(next, struct command, queue_node);
drivers/misc/lkdtm/bugs.c
610
test_head.next = redirection;
drivers/misc/lkdtm/bugs.c
637
item.node.next = redirection;
drivers/misc/mei/bus.c
1487
struct mei_cl_device *cldev, *next;
drivers/misc/mei/bus.c
1490
list_for_each_entry_safe(cldev, next, &bus->device_list, bus_list)
drivers/misc/mei/client.c
283
struct mei_me_client *me_cl, *next;
drivers/misc/mei/client.c
286
list_for_each_entry_safe(me_cl, next, &dev->me_clients, list)
drivers/misc/mei/client.c
397
struct mei_cl_cb *cb, *next;
drivers/misc/mei/client.c
399
list_for_each_entry_safe(cb, next, head, list) {
drivers/misc/mei/client.c
419
struct mei_cl_cb *cb, *next;
drivers/misc/mei/client.c
421
list_for_each_entry_safe(cb, next, head, list) {
drivers/misc/mei/client.c
435
struct mei_cl_cb *cb, *next;
drivers/misc/mei/client.c
437
list_for_each_entry_safe(cb, next, head, list)
drivers/misc/mei/hbm.c
1002
struct mei_cl_cb *cb, *next;
drivers/misc/mei/hbm.c
1005
list_for_each_entry_safe(cb, next, &dev->ctrl_rd_list, list) {
drivers/misc/mei/hbm.c
661
struct mei_cl_cb *cb, *next;
drivers/misc/mei/hbm.c
664
list_for_each_entry_safe(cb, next, &dev->ctrl_rd_list, list) {
drivers/misc/mei/hbm.c
691
struct mei_cl_cb *cb, *next;
drivers/misc/mei/hbm.c
694
list_for_each_entry_safe(cb, next, &dev->ctrl_rd_list, list) {
drivers/misc/mei/interrupt.c
31
struct mei_cl_cb *cb, *next;
drivers/misc/mei/interrupt.c
34
list_for_each_entry_safe(cb, next, cmpl_list, list) {
drivers/misc/mei/interrupt.c
509
struct mei_cl_cb *cb, *next;
drivers/misc/mei/interrupt.c
527
list_for_each_entry_safe(cb, next, &dev->write_waiting_list, list) {
drivers/misc/mei/interrupt.c
538
list_for_each_entry_safe(cb, next, &dev->ctrl_wr_list, list) {
drivers/misc/mei/interrupt.c
592
list_for_each_entry_safe(cb, next, &dev->write_list, list) {
drivers/misc/mei/main.c
97
struct mei_cl_vtag *vtag_l, *next;
drivers/misc/mei/main.c
99
list_for_each_entry_safe(vtag_l, next, &cl->vtag_map, list) {
drivers/misc/sgi-gru/grufile.c
65
struct list_head *entry, *next;
drivers/misc/sgi-gru/grufile.c
74
list_for_each_safe(entry, next, &vdata->vd_head) {
drivers/misc/sgi-gru/grukservices.c
119
void *next __gru_cacheline_aligned__;/* CL 1 */
drivers/misc/sgi-gru/grukservices.c
553
mq->next = &mq->data;
drivers/misc/sgi-gru/grukservices.c
838
struct message_header *mhdr = mq->next;
drivers/misc/sgi-gru/grukservices.c
839
void *next, *pnext;
drivers/misc/sgi-gru/grukservices.c
847
pnext = mq->next;
drivers/misc/sgi-gru/grukservices.c
848
next = pnext + GRU_CACHE_LINE_BYTES * lines;
drivers/misc/sgi-gru/grukservices.c
849
if (next == mq->limit) {
drivers/misc/sgi-gru/grukservices.c
850
next = mq->start;
drivers/misc/sgi-gru/grukservices.c
852
} else if (pnext < mq->start2 && next >= mq->start2) {
drivers/misc/sgi-gru/grukservices.c
858
mq->next = next;
drivers/misc/sgi-gru/grukservices.c
870
struct message_header *mhdr = mq->next;
drivers/misc/sgi-gru/grukservices.c
876
mhdr = mq->next;
drivers/misc/sgi-gru/gruprocfs.c
230
.next = seq_next,
drivers/misc/sgi-gru/gruprocfs.c
237
.next = seq_next,
drivers/misc/sgi-xp/xpc.h
245
struct xpc_fifo_entry_uv *next;
drivers/misc/sgi-xp/xpc.h
268
struct xpc_fifo_entry_uv next; /* FOR XPC INTERNAL USE ONLY */
drivers/misc/sgi-xp/xpc.h
290
struct xpc_fifo_entry_uv next;
drivers/misc/sgi-xp/xpc_uv.c
1019
&msg_slot->next);
drivers/misc/sgi-xp/xpc_uv.c
1263
msg_slot = container_of(entry, struct xpc_send_msg_slot_uv, next);
drivers/misc/sgi-xp/xpc_uv.c
1272
xpc_put_fifo_entry_uv(&ch->sn.uv.msg_slot_free_list, &msg_slot->next);
drivers/misc/sgi-xp/xpc_uv.c
1373
xpc_put_fifo_entry_uv(&ch_uv->recv_msg_list, &msg_slot->hdr.u.next);
drivers/misc/sgi-xp/xpc_uv.c
1568
hdr.u.next);
drivers/misc/sgi-xp/xpc_uv.c
889
head->first = first->next;
drivers/misc/sgi-xp/xpc_uv.c
896
first->next = NULL;
drivers/misc/sgi-xp/xpc_uv.c
908
last->next = NULL;
drivers/misc/sgi-xp/xpc_uv.c
911
head->last->next = last;
drivers/misc/vmw_vmci/vmci_context.c
505
list_item = context->datagram_queue.next;
drivers/misc/vmw_vmci/vmci_context.c
531
list_item = context->datagram_queue.next;
drivers/mmc/core/mmc_test.c
1936
unsigned int next;
drivers/mmc/core/mmc_test.c
1947
next = rnd_next;
drivers/mmc/core/mmc_test.c
1951
rnd_next = next;
drivers/mmc/core/mmc_test.c
1959
next = rnd_next;
drivers/mmc/core/mmc_test.c
1963
rnd_next = next;
drivers/mmc/core/sdio_cis.c
326
this->next = NULL;
drivers/mmc/core/sdio_cis.c
330
prev = &this->next;
drivers/mmc/core/sdio_cis.c
391
tuple = tuple->next;
drivers/mmc/core/sdio_cis.c
426
tuple = tuple->next;
drivers/mmc/core/sdio_uart.c
1039
for (tpl = func->tuples; tpl; tpl = tpl->next) {
drivers/mmc/host/atmel-mci.c
1611
slot = list_entry(host->queue.next,
drivers/mmc/host/dw_mmc.c
1906
slot = list_entry(host->queue.next,
drivers/mmc/host/mmci.c
1087
bool next)
drivers/mmc/host/mmci.c
1095
if (next)
drivers/mmc/host/mmci.c
1127
struct mmci_dmae_next *next = &dmae->next_data;
drivers/mmc/host/mmci.c
1132
WARN_ON(!data->host_cookie && (next->desc || next->chan));
drivers/mmc/host/mmci.c
1134
dmae->desc_current = next->desc;
drivers/mmc/host/mmci.c
1135
dmae->cur = next->chan;
drivers/mmc/host/mmci.c
1136
next->desc = NULL;
drivers/mmc/host/mmci.c
1137
next->chan = NULL;
drivers/mmc/host/mmci.c
1152
struct mmci_dmae_next *next = &dmae->next_data;
drivers/mmc/host/mmci.c
1160
if (dmae->desc_current == next->desc)
drivers/mmc/host/mmci.c
1163
if (dmae->cur == next->chan) {
drivers/mmc/host/mmci.c
1168
next->desc = NULL;
drivers/mmc/host/mmci.c
1169
next->chan = NULL;
drivers/mmc/host/mmci.c
543
static int mmci_prep_data(struct mmci_host *host, struct mmc_data *data, bool next)
drivers/mmc/host/mmci.c
550
err = host->ops->prep_data(host, data, next);
drivers/mmc/host/mmci.c
552
if (next && !err)
drivers/mmc/host/mmci.h
392
bool next);
drivers/mmc/host/mmci.h
477
bool next);
drivers/mmc/host/mmci_stm32_sdmmc.c
161
struct mmc_data *data, bool next)
drivers/mmc/host/mmci_stm32_sdmmc.c
164
if (!next && data->host_cookie == host->next_cookie)
drivers/mmc/host/mtk-sd.c
2101
gpd->next = lower_32_bits(dma_addr);
drivers/mmc/host/mtk-sd.c
2113
bd[i].next = lower_32_bits(dma_addr);
drivers/mmc/host/mtk-sd.c
382
u32 next;
drivers/mmc/host/mtk-sd.c
400
u32 next;
drivers/mmc/host/omap_hsmmc.c
1192
struct omap_hsmmc_next *next,
drivers/mmc/host/omap_hsmmc.c
1197
if (!next && data->host_cookie &&
drivers/mmc/host/omap_hsmmc.c
1206
if (next || data->host_cookie != host->next_data.cookie) {
drivers/mmc/host/omap_hsmmc.c
1219
if (next) {
drivers/mmc/host/omap_hsmmc.c
1220
next->dma_len = dma_len;
drivers/mmc/host/omap_hsmmc.c
1221
data->host_cookie = ++next->cookie < 0 ? 1 : next->cookie;
drivers/mmc/host/usdhi6rol0.c
487
struct scatterlist *next = sg_next(host->sg);
drivers/mmc/host/usdhi6rol0.c
491
if (!next)
drivers/mmc/host/usdhi6rol0.c
493
host->sg = next;
drivers/mmc/host/usdhi6rol0.c
495
if (WARN(next && sg_dma_len(next) % data->blksz,
drivers/mmc/host/usdhi6rol0.c
497
sg_dma_len(next), data->blksz))
drivers/mtd/devices/block2mtd.c
511
struct list_head *pos, *next;
drivers/mtd/devices/block2mtd.c
514
list_for_each_safe(pos, next, &blkmtd_device_list) {
drivers/mtd/devices/ms02-nv.c
220
mp->next = root_ms02nv_mtd;
drivers/mtd/devices/ms02-nv.c
251
root_ms02nv_mtd = mp->next;
drivers/mtd/devices/ms02-nv.h
91
struct mtd_info *next;
drivers/mtd/devices/slram.c
135
curmtd = &(*curmtd)->next;
drivers/mtd/devices/slram.c
144
(*curmtd)->next = NULL;
drivers/mtd/devices/slram.c
204
nextitem = slram_mtdlist->next;
drivers/mtd/devices/slram.c
62
struct slram_mtd_list *next;
drivers/mtd/mtd_blkdevs.c
448
struct mtd_blktrans_dev *dev, *next;
drivers/mtd/mtd_blkdevs.c
451
list_for_each_entry_safe(dev, next, &tr->devs, list)
drivers/mtd/mtd_blkdevs.c
480
if (!blktrans_notifier.list.next)
drivers/mtd/mtd_blkdevs.c
508
struct mtd_blktrans_dev *dev, *next;
drivers/mtd/mtd_blkdevs.c
515
list_for_each_entry_safe(dev, next, &tr->devs, list)
drivers/mtd/mtd_blkdevs.c
529
if (blktrans_notifier.list.next)
drivers/mtd/mtdpart.c
307
struct mtd_info *child, *next;
drivers/mtd/mtdpart.c
310
list_for_each_entry_safe(child, next, &mtd->partitions, part.node) {
drivers/mtd/mtdpart.c
332
struct mtd_info *child, *next;
drivers/mtd/mtdpart.c
335
list_for_each_entry_safe(child, next, &mtd->partitions, part.node) {
drivers/mtd/nand/raw/brcmnand/brcmnand.c
1313
u32 next;
drivers/mtd/nand/raw/brcmnand/brcmnand.c
1318
next = (section * sas);
drivers/mtd/nand/raw/brcmnand/brcmnand.c
1320
next += 6;
drivers/mtd/nand/raw/brcmnand/brcmnand.c
1331
next--;
drivers/mtd/nand/raw/brcmnand/brcmnand.c
1335
oobregion->length = next - oobregion->offset;
drivers/mtd/parsers/cmdlinepart.c
296
this_mtd->next = partitions;
drivers/mtd/parsers/cmdlinepart.c
345
for (part = partitions; part; part = part->next) {
drivers/mtd/parsers/cmdlinepart.c
52
struct cmdline_mtd_partition *next;
drivers/mtd/parsers/redboot.c
223
prev = &(*prev)->next;
drivers/mtd/parsers/redboot.c
224
new_fl->next = *prev;
drivers/mtd/parsers/redboot.c
235
for (tmp_fl = fl; tmp_fl->next; tmp_fl = tmp_fl->next) {
drivers/mtd/parsers/redboot.c
236
if (tmp_fl->img->flash_base + tmp_fl->img->size + master->erasesize <= tmp_fl->next->img->flash_base) {
drivers/mtd/parsers/redboot.c
282
if (fl->next && fl->img->flash_base + fl->img->size + master->erasesize <= fl->next->img->flash_base) {
drivers/mtd/parsers/redboot.c
285
parts[i].size = fl->next->img->flash_base - parts[i].offset;
drivers/mtd/parsers/redboot.c
290
fl = fl->next;
drivers/mtd/parsers/redboot.c
299
fl = fl->next;
drivers/mtd/parsers/redboot.c
33
struct fis_list *next;
drivers/mtd/spi-nor/core.c
1578
struct spi_nor_erase_command *cmd, *next;
drivers/mtd/spi-nor/core.c
1580
list_for_each_entry_safe(cmd, next, erase_list, list) {
drivers/mtd/spi-nor/core.c
1661
struct spi_nor_erase_command *cmd, *next;
drivers/mtd/spi-nor/core.c
1668
list_for_each_entry_safe(cmd, next, &erase_list, list) {
drivers/mtd/ubi/attach.c
825
aeb = list_entry(ai->free.next, struct ubi_ainf_peb, u.list);
drivers/mtd/ubi/block.c
639
struct ubiblock *next;
drivers/mtd/ubi/block.c
643
list_for_each_entry_safe(dev, next, &ubiblock_devices, list) {
drivers/mtd/ubi/debug.c
542
.next = eraseblk_count_seq_next,
drivers/mtd/ubi/wl.c
1743
wrk = list_entry(ubi->works.next, struct ubi_work, list);
drivers/mtd/ubi/wl.c
215
wrk = list_entry(ubi->works.next, struct ubi_work, list);
drivers/net/bareudp.c
57
struct list_head next; /* bareudp node on namespace list */
drivers/net/bareudp.c
638
list_for_each_entry(bareudp, &bn->bareudp_list, next) {
drivers/net/bareudp.c
677
list_add(&bareudp->next, &bn->bareudp_list);
drivers/net/bareudp.c
698
list_del(&bareudp->next);
drivers/net/bareudp.c
785
struct bareudp_dev *bareudp, *next;
drivers/net/bareudp.c
787
list_for_each_entry_safe(bareudp, next, &bn->bareudp_list, next)
drivers/net/bonding/bond_alb.c
106
u32 next_index = tx_hash_table[index].next;
drivers/net/bonding/bond_alb.c
209
hash_table[hash_index].next = next_index;
drivers/net/bonding/bond_alb.c
821
u32 next;
drivers/net/bonding/bond_alb.c
824
next = bond_info->rx_hashtbl[ip_src_hash].src_first;
drivers/net/bonding/bond_alb.c
825
bond_info->rx_hashtbl[ip_dst_hash].src_next = next;
drivers/net/bonding/bond_alb.c
826
if (next != RLB_NULL_INDEX)
drivers/net/bonding/bond_alb.c
827
bond_info->rx_hashtbl[next].src_prev = ip_dst_hash;
drivers/net/bonding/bond_alb.c
83
entry->next = TLB_NULL_INDEX;
drivers/net/bonding/bond_main.c
4463
struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
drivers/net/bonding/bond_main.c
4471
next = NULL;
drivers/net/bonding/bond_main.c
4477
next = ldev;
drivers/net/bonding/bond_main.c
4486
if (!next) {
drivers/net/bonding/bond_main.c
4489
next = dev_stack[--cur];
drivers/net/bonding/bond_main.c
4493
now = next;
drivers/net/bonding/bond_procfs.c
278
.next = bond_info_seq_next,
drivers/net/dsa/bcm_sf2_cfp.c
1045
list_del(&rule->next);
drivers/net/dsa/bcm_sf2_cfp.c
1220
list_for_each_entry_safe_reverse(rule, n, &priv->cfp.rules_list, next)
drivers/net/dsa/bcm_sf2_cfp.c
1242
list_for_each_entry(rule, &priv->cfp.rules_list, next) {
drivers/net/dsa/bcm_sf2_cfp.c
25
struct list_head next;
drivers/net/dsa/bcm_sf2_cfp.c
572
list_for_each_entry(rule, &priv->cfp.rules_list, next) {
drivers/net/dsa/bcm_sf2_cfp.c
590
list_for_each_entry(rule, &priv->cfp.rules_list, next) {
drivers/net/dsa/bcm_sf2_cfp.c
965
list_add_tail(&rule->next, &priv->cfp.rules_list);
drivers/net/dsa/hirschmann/hellcreek.c
1582
const struct tc_taprio_sched_entry *cur, *initial, *next;
drivers/net/dsa/hirschmann/hellcreek.c
1586
next = cur + 1;
drivers/net/dsa/hirschmann/hellcreek.c
1596
gates = next->gate_mask ^
drivers/net/dsa/hirschmann/hellcreek.c
1622
next++;
drivers/net/dsa/microchip/ksz9477_acl.c
701
struct ksz9477_acl_entry *curr, *next;
drivers/net/dsa/microchip/ksz9477_acl.c
713
next = &acles->entries[j];
drivers/net/dsa/microchip/ksz9477_acl.c
715
if (curr->prio > next->prio) {
drivers/net/dsa/microchip/ksz_ptp.c
637
struct timespec64 next;
drivers/net/dsa/microchip/ksz_ptp.c
666
next = ns_to_timespec64(next_ns);
drivers/net/dsa/microchip/ksz_ptp.c
667
request.start.sec = next.tv_sec;
drivers/net/dsa/microchip/ksz_ptp.c
668
request.start.nsec = next.tv_nsec;
drivers/net/dsa/ocelot/felix_vsc9959.c
2063
return vsc9959_psfp_sfi_list_add(ocelot, sfi2, last->next);
drivers/net/ethernet/3com/3c515.c
1016
vp->tx_ring[entry].next = 0;
drivers/net/ethernet/3com/3c515.c
1028
prev_entry->next = isa_virt_to_bus(&vp->tx_ring[entry]);
drivers/net/ethernet/3com/3c515.c
1557
vp = list_entry(root_corkscrew_dev.next,
drivers/net/ethernet/3com/3c515.c
274
u32 next;
drivers/net/ethernet/3com/3c515.c
287
u32 next;
drivers/net/ethernet/3com/3c515.c
332
next:8; /* The media type to try next. */
drivers/net/ethernet/3com/3c515.c
719
dev->if_port = media_tbl[dev->if_port].next;
drivers/net/ethernet/3com/3c515.c
810
vp->rx_ring[i].next =
drivers/net/ethernet/3com/3c515.c
813
vp->rx_ring[i].next = 0;
drivers/net/ethernet/3com/3c515.c
824
vp->rx_ring[i - 1].next =
drivers/net/ethernet/3com/3c515.c
910
media_tbl[dev->if_port].next;
drivers/net/ethernet/3com/3c59x.c
1584
dev->if_port = media_tbl[dev->if_port].next;
drivers/net/ethernet/3com/3c59x.c
1738
vp->rx_ring[i].next = cpu_to_le32(vp->rx_ring_dma + sizeof(struct boom_rx_desc) * (i+1));
drivers/net/ethernet/3com/3c59x.c
1761
vp->rx_ring[i-1].next = cpu_to_le32(vp->rx_ring_dma);
drivers/net/ethernet/3com/3c59x.c
1841
dev->if_port = media_tbl[dev->if_port].next;
drivers/net/ethernet/3com/3c59x.c
2147
vp->tx_ring[entry].next = 0;
drivers/net/ethernet/3com/3c59x.c
2216
prev_entry->next = cpu_to_le32(vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc));
drivers/net/ethernet/3com/3c59x.c
544
__le32 next; /* Last entry points to 0. */
drivers/net/ethernet/3com/3c59x.c
564
__le32 next; /* Last entry points to 0. */
drivers/net/ethernet/3com/3c59x.c
724
next:8; /* The media type to try next. */
drivers/net/ethernet/8390/8390.h
24
unsigned char next; /* pointer to next packet. */
drivers/net/ethernet/8390/axnet_cs.c
1387
rx_frame.next);
drivers/net/ethernet/8390/axnet_cs.c
1421
rx_frame.status, rx_frame.next,
drivers/net/ethernet/8390/axnet_cs.c
1428
next_frame = rx_frame.next;
drivers/net/ethernet/8390/lib8390.c
709
if (rx_frame.next != next_frame &&
drivers/net/ethernet/8390/lib8390.c
710
rx_frame.next != next_frame + 1 &&
drivers/net/ethernet/8390/lib8390.c
711
rx_frame.next != next_frame - num_rx_pages &&
drivers/net/ethernet/8390/lib8390.c
712
rx_frame.next != next_frame + 1 - num_rx_pages) {
drivers/net/ethernet/8390/lib8390.c
723
rx_frame.next);
drivers/net/ethernet/8390/lib8390.c
751
rx_frame.status, rx_frame.next,
drivers/net/ethernet/8390/lib8390.c
758
next_frame = rx_frame.next;
drivers/net/ethernet/8390/stnic.c
198
hdr->next = buf[0] >> 8;
drivers/net/ethernet/8390/stnic.c
207
ring_page, hdr->status, hdr->next, hdr->count);
drivers/net/ethernet/agere/et131x.c
1760
tcb->next = tcb + 1;
drivers/net/ethernet/agere/et131x.c
1766
tcb->next = NULL;
drivers/net/ethernet/agere/et131x.c
2016
rfd = list_entry(rx_ring->recv_list.next,
drivers/net/ethernet/agere/et131x.c
2138
struct fbr_desc *next;
drivers/net/ethernet/agere/et131x.c
2145
next = (struct fbr_desc *)(fbr->ring_virtaddr) +
drivers/net/ethernet/agere/et131x.c
2152
next->addr_hi = fbr->bus_high[buff_index];
drivers/net/ethernet/agere/et131x.c
2153
next->addr_lo = fbr->bus_low[buff_index];
drivers/net/ethernet/agere/et131x.c
2154
next->word2 = buff_index;
drivers/net/ethernet/agere/et131x.c
2244
element = rx_local->recv_list.next;
drivers/net/ethernet/agere/et131x.c
2573
tx_ring->send_tail->next = tcb;
drivers/net/ethernet/agere/et131x.c
2579
WARN_ON(tcb->next != NULL);
drivers/net/ethernet/agere/et131x.c
2639
tx_ring->tcb_qhead = tcb->next;
drivers/net/ethernet/agere/et131x.c
2647
tcb->next = NULL;
drivers/net/ethernet/agere/et131x.c
2655
tx_ring->tcb_qtail->next = tcb;
drivers/net/ethernet/agere/et131x.c
2715
tx_ring->tcb_qtail->next = tcb;
drivers/net/ethernet/agere/et131x.c
2739
struct tcb *next = tcb->next;
drivers/net/ethernet/agere/et131x.c
2741
tx_ring->send_head = next;
drivers/net/ethernet/agere/et131x.c
2743
if (next == NULL)
drivers/net/ethernet/agere/et131x.c
2792
tx_ring->send_head = tcb->next;
drivers/net/ethernet/agere/et131x.c
2793
if (tcb->next == NULL)
drivers/net/ethernet/agere/et131x.c
2807
tx_ring->send_head = tcb->next;
drivers/net/ethernet/agere/et131x.c
2808
if (tcb->next == NULL)
drivers/net/ethernet/agere/et131x.c
350
struct tcb *next; /* Next entry in ring */
drivers/net/ethernet/alteon/acenic.h
684
struct net_device *next;
drivers/net/ethernet/altera/altera_sgdma.c
320
csrwr32(lower_32_bits(ndesc_phys), desc, sgdma_descroffs(next));
drivers/net/ethernet/altera/altera_sgdma.c
427
entry = list_entry((list)->next, type, member); \
drivers/net/ethernet/altera/altera_sgdma.c
436
entry = list_entry((list)->next, type, member); \
drivers/net/ethernet/altera/altera_sgdmahw.h
15
u32 next;
drivers/net/ethernet/amd/amd8111e.h
767
struct net_device *next;
drivers/net/ethernet/amd/declance.c
1049
dev = lp->next;
drivers/net/ethernet/amd/declance.c
1268
lp->next = root_lance_dev;
drivers/net/ethernet/amd/declance.c
1311
root_lance_dev = lp->next;
drivers/net/ethernet/amd/declance.c
251
struct net_device *next;
drivers/net/ethernet/amd/pcnet32.c
1993
lp->next = pcnet32_dev;
drivers/net/ethernet/amd/pcnet32.c
294
struct net_device *next;
drivers/net/ethernet/amd/pcnet32.c
3029
next_dev = lp->next;
drivers/net/ethernet/apple/mace.c
538
int fill, next, len;
drivers/net/ethernet/apple/mace.c
543
next = fill + 1;
drivers/net/ethernet/apple/mace.c
544
if (next >= N_TX_RING)
drivers/net/ethernet/apple/mace.c
545
next = 0;
drivers/net/ethernet/apple/mace.c
546
if (next == mp->tx_empty) {
drivers/net/ethernet/apple/mace.c
565
np = mp->tx_cmds + NCMDS_TX * next;
drivers/net/ethernet/apple/mace.c
570
mp->tx_fill = next;
drivers/net/ethernet/apple/mace.c
578
if (++next >= N_TX_RING)
drivers/net/ethernet/apple/mace.c
579
next = 0;
drivers/net/ethernet/apple/mace.c
580
if (next == mp->tx_empty)
drivers/net/ethernet/apple/mace.c
883
int i, nb, stat, next;
drivers/net/ethernet/apple/mace.c
895
next = i + 1;
drivers/net/ethernet/apple/mace.c
896
if (next >= N_RX_RING)
drivers/net/ethernet/apple/mace.c
897
next = 0;
drivers/net/ethernet/apple/mace.c
898
np = mp->rx_cmds + next;
drivers/net/ethernet/apple/mace.c
899
if (next != mp->rx_fill &&
drivers/net/ethernet/apple/mace.c
952
next = i + 1;
drivers/net/ethernet/apple/mace.c
953
if (next >= N_RX_RING)
drivers/net/ethernet/apple/mace.c
954
next = 0;
drivers/net/ethernet/apple/mace.c
955
if (next == mp->rx_empty)
drivers/net/ethernet/apple/mace.c
978
i = next;
drivers/net/ethernet/aquantia/atlantic/aq_macsec.c
1243
rx_sc = rcu_dereference_bh(rx_sc->next)) {
drivers/net/ethernet/aquantia/atlantic/aq_macsec.c
1269
rx_sc = rcu_dereference_bh(rx_sc->next)) {
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
500
buff_ = &ring->buff_ring[buff_->next];
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
559
if (buff_->next >= self->size) {
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
565
next_ = buff_->next;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
587
if (buff_->next >= self->size) {
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
591
next_ = buff_->next;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
650
next_ = buff_->next;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
734
if (buff_->next >= rx_ring->size) {
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
738
next_ = buff_->next;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
760
if (buff_->next >= rx_ring->size) {
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
764
next_ = buff_->next;
drivers/net/ethernet/aquantia/atlantic/aq_ring.h
48
u16 next;
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c
712
buff->next = 0U;
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c
716
buff->next = aq_ring_next_dx(ring,
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c
982
buff->next = 0U;
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c
991
buff->next = rxd_wb->next_desc_ptr;
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c
995
buff->next =
drivers/net/ethernet/atheros/ag71xx.c
1160
desc->next = (u32)(ring->descs_dma +
drivers/net/ethernet/atheros/ag71xx.c
1238
desc->next = (u32)(ring->descs_dma +
drivers/net/ethernet/atheros/ag71xx.c
1242
desc, desc->next);
drivers/net/ethernet/atheros/ag71xx.c
1650
goto next;
drivers/net/ethernet/atheros/ag71xx.c
1661
next:
drivers/net/ethernet/atheros/ag71xx.c
1897
ag->stop_desc->next = (u32)ag->stop_desc_dma;
drivers/net/ethernet/atheros/ag71xx.c
297
u32 next;
drivers/net/ethernet/atheros/alx/main.c
124
cur = next;
drivers/net/ethernet/atheros/alx/main.c
125
if (++next == alx->rx_ringsz)
drivers/net/ethernet/atheros/alx/main.c
126
next = 0;
drivers/net/ethernet/atheros/alx/main.c
77
u16 cur, next, count = 0;
drivers/net/ethernet/atheros/alx/main.c
79
next = cur = rxq->write_idx;
drivers/net/ethernet/atheros/alx/main.c
80
if (++next == alx->rx_ringsz)
drivers/net/ethernet/atheros/alx/main.c
81
next = 0;
drivers/net/ethernet/atheros/alx/main.c
84
while (!cur_buf->skb && next != rxq->read_idx) {
drivers/net/ethernet/atheros/atlx/atl1.c
1851
goto next;
drivers/net/ethernet/atheros/atlx/atl1.c
1883
next:
drivers/net/ethernet/broadcom/asp2/bcmasp_intf.c
519
goto next;
drivers/net/ethernet/broadcom/asp2/bcmasp_intf.c
534
goto next;
drivers/net/ethernet/broadcom/asp2/bcmasp_intf.c
560
next:
drivers/net/ethernet/broadcom/bcmsysport.c
768
goto next;
drivers/net/ethernet/broadcom/bcmsysport.c
787
goto next;
drivers/net/ethernet/broadcom/bcmsysport.c
795
goto next;
drivers/net/ethernet/broadcom/bcmsysport.c
805
goto next;
drivers/net/ethernet/broadcom/bcmsysport.c
837
next:
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1136
u16 next, max = rxr->rx_agg_bmap_size;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1138
next = find_next_zero_bit(rxr->rx_agg_bmap, max, idx);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1139
if (next >= max)
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1140
next = find_first_zero_bit(rxr->rx_agg_bmap, max);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1141
return next;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
533
u64 next;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
923
t2[i].next = (u64)(t2_mapping +
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
612
u8 *next = base;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
624
memcpy(next, &pos->u, size);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
627
counter, next);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
628
next += stride + size;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1022
u16 next, max = rxr->rx_agg_bmap_size;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1024
next = find_next_zero_bit(rxr->rx_agg_bmap, max, idx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1025
if (next >= max)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1026
next = find_first_zero_bit(rxr->rx_agg_bmap, max);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1027
return next;
drivers/net/ethernet/broadcom/cnic.c
1010
struct bnx2x_bd_chain_next *next =
drivers/net/ethernet/broadcom/cnic.c
1016
next->addr_hi = (u64) info->dma.pg_map_arr[j] >> 32;
drivers/net/ethernet/broadcom/cnic.c
1017
next->addr_lo = info->dma.pg_map_arr[j] & 0xffffffff;
drivers/net/ethernet/broadcom/cnic.c
5760
udev = list_entry(cnic_udev_list.next, struct cnic_uio_dev,
drivers/net/ethernet/broadcom/cnic.c
658
u32 next)
drivers/net/ethernet/broadcom/cnic.c
662
id_tbl->next = next;
drivers/net/ethernet/broadcom/cnic.c
700
id = find_next_zero_bit(id_tbl->table, id_tbl->max, id_tbl->next);
drivers/net/ethernet/broadcom/cnic.c
703
if (id_tbl->next != 0) {
drivers/net/ethernet/broadcom/cnic.c
704
id = find_first_zero_bit(id_tbl->table, id_tbl->next);
drivers/net/ethernet/broadcom/cnic.c
705
if (id >= id_tbl->next)
drivers/net/ethernet/broadcom/cnic.c
712
id_tbl->next = (id + 1) & (id_tbl->max - 1);
drivers/net/ethernet/broadcom/cnic.h
144
u32 next;
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2349
goto next;
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2377
goto next;
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2385
goto next;
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2413
goto next;
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2445
next:
drivers/net/ethernet/broadcom/tg3.c
7926
struct sk_buff *segs, *seg, *next;
drivers/net/ethernet/broadcom/tg3.c
7951
skb_list_walk_safe(segs, seg, next) {
drivers/net/ethernet/brocade/bna/bnad.c
687
goto next;
drivers/net/ethernet/brocade/bna/bnad.c
720
next:
drivers/net/ethernet/cadence/macb_main.c
678
unsigned int i, cycles, shift, curr, next;
drivers/net/ethernet/cadence/macb_main.c
712
next = (curr + shift) % ring_size;
drivers/net/ethernet/cadence/macb_main.c
714
while (next != i) {
drivers/net/ethernet/cadence/macb_main.c
716
desc_next = macb_tx_desc(queue, next);
drivers/net/ethernet/cadence/macb_main.c
720
if (next == ring_size - 1)
drivers/net/ethernet/cadence/macb_main.c
726
skb_next = macb_tx_skb(queue, next);
drivers/net/ethernet/cadence/macb_main.c
729
curr = next;
drivers/net/ethernet/cadence/macb_main.c
730
next = (curr + shift) % ring_size;
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1056
while (dispatch->next != dispatch) {
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1057
temp = dispatch->next;
drivers/net/ethernet/cavium/liquidio/octeon_network.h
618
node = root->next;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
167
next = &rbdr->pgcache[rbdr->pgidx];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
168
page = next->page;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
595
goto next;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
599
goto next;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
617
next:
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
98
struct pgcache *pgcache, *next;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
102
if ((void *)e->next >= (void *)t->tid_tab &&
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
103
(void *)e->next < (void *)&t->atid_tab[t->natids])
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
83
if ((void *)e->next >= (void *)t->tid_tab &&
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
84
(void *)e->next < (void *)&t->atid_tab[t->natids])
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1148
t->stid_tab[nstids - 1].next = &t->stid_tab[nstids];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1153
t->atid_tab[natids - 1].next = &t->atid_tab[natids];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
508
p->next = t->afree;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
646
t->afree = p->next;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h
141
union listen_entry *next;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.h
146
union active_open_entry *next;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
206
for (p = &d->l2tab[hash].first; *p; p = &(*p)->next)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
208
*p = e->next;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
297
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
310
e->next = d->l2tab[hash].first;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
372
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb3/l2t.h
66
struct l2t_entry *next; /* next l2t_entry on chain */
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1073
.next = devlog_next,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1217
.next = mboxlog_next,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1913
.next = mps_tcam_next,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2544
.next = dcb_info_next,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3228
.next = sge_queue_next,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
91
.next = seq_tab_next,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1602
t->afree = p->next;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1619
p->next = t->afree;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1892
t->atid_tab[natids - 1].next = &t->atid_tab[natids];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
228
const struct cxgb4_next_header *next;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
241
next = is_ipv6 ? cxgb4_ipv6_jumps : cxgb4_ipv4_jumps;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
244
for (i = 0; next[i].jump; i++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
245
if (next[i].sel.offoff != cls->knode.sel->offoff ||
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
246
next[i].sel.offshift != cls->knode.sel->offshift ||
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
247
next[i].sel.offmask != cls->knode.sel->offmask ||
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
248
next[i].sel.off != cls->knode.sel->off)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
260
if (next[i].key.off == off &&
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
261
next[i].key.val == val &&
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
262
next[i].key.mask == mask) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
282
link->match_field = next[i].jump;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
94
union aopen_entry *next;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
287
for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
289
*p = e->next;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
290
e->next = NULL;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
329
for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
331
*p = e->next;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
332
e->next = NULL;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
440
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
464
e->next = d->l2tab[hash].first;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
520
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
719
.next = l2t_seq_next,
drivers/net/ethernet/chelsio/cxgb4/l2t.h
80
struct l2t_entry *next; /* next l2t_entry on chain */
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2030
.next = mboxlog_next,
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2165
.next = sge_queue_next,
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2294
.next = sge_qstats_next,
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2375
.next = interfaces_next,
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls.h
160
struct listen_info *next; /* Link to next entry */
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls.h
320
struct sk_buff *next;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
332
int next = (int)new_state[sk->sk_state];
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
334
tcp_set_state(sk, next & TCP_STATE_MASK);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
335
return next & TCP_ACTION_FIN;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
559
p->next = cdev->listen_hash_tab[key];
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
576
for (p = cdev->listen_hash_tab[key]; p; p = p->next)
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
596
for (p = *prev; p; prev = &p->next, p = p->next)
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
599
*prev = p->next;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
134
pool->next, count, 0);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
136
pool->next = 0;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
141
pool->next = i + count;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
142
if (pool->next >= ppm->pool_index_max)
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
143
pool->next = 0;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
149
pool->next);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
164
ppm->next, count, 0);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
166
ppm->next = 0;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
173
ppm->next = i + count;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
174
if (ppm->max_index_in_edram && (ppm->next >= ppm->max_index_in_edram))
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
175
ppm->next = 0;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
176
else if (ppm->next >= ppm->bmap_index_max)
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
177
ppm->next = 0;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
182
__func__, i, count, i + ppm->pool_rsvd, ppm->next,
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
206
if (i < pool->next)
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
207
pool->next = i;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
211
__func__, cpu, i, pool->next);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
218
if (i < ppm->next)
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
219
ppm->next = i;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
222
pr_debug("%s: idx %d, next %u.\n", __func__, i, ppm->next);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
270
nr_pages, npods, ppm->next, caller_data);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
478
ppm->next = 0;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h
123
unsigned int next; /* next possible free index */
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h
145
unsigned int next;
drivers/net/ethernet/cisco/enic/enic_main.c
821
while (!buf->os_buf && (buf->next != wq->to_clean)) {
drivers/net/ethernet/cisco/enic/enic_main.c
826
wq->to_use = buf->next;
drivers/net/ethernet/cisco/enic/enic_rq.c
379
buf = buf->next;
drivers/net/ethernet/cisco/enic/enic_rq.c
404
vrq->to_clean = vrq_buf->next;
drivers/net/ethernet/cisco/enic/enic_rq.c
405
vrq_buf = vrq_buf->next;
drivers/net/ethernet/cisco/enic/vnic_rq.c
183
buf = buf->next;
drivers/net/ethernet/cisco/enic/vnic_rq.c
37
buf->next = rq->bufs[0];
drivers/net/ethernet/cisco/enic/vnic_rq.c
40
buf->next = rq->bufs[i + 1];
drivers/net/ethernet/cisco/enic/vnic_rq.c
42
buf->next = buf + 1;
drivers/net/ethernet/cisco/enic/vnic_rq.h
121
buf = buf->next;
drivers/net/ethernet/cisco/enic/vnic_rq.h
172
rq->to_clean = buf->next;
drivers/net/ethernet/cisco/enic/vnic_rq.h
56
struct vnic_rq_buf *next;
drivers/net/ethernet/cisco/enic/vnic_wq.c
193
buf = wq->to_clean = buf->next;
drivers/net/ethernet/cisco/enic/vnic_wq.c
37
buf->next = wq->bufs[0];
drivers/net/ethernet/cisco/enic/vnic_wq.c
38
buf->next->prev = buf;
drivers/net/ethernet/cisco/enic/vnic_wq.c
41
buf->next = wq->bufs[i + 1];
drivers/net/ethernet/cisco/enic/vnic_wq.c
42
buf->next->prev = buf;
drivers/net/ethernet/cisco/enic/vnic_wq.c
44
buf->next = buf + 1;
drivers/net/ethernet/cisco/enic/vnic_wq.c
45
buf->next->prev = buf;
drivers/net/ethernet/cisco/enic/vnic_wq.h
135
buf = buf->next;
drivers/net/ethernet/cisco/enic/vnic_wq.h
156
wq->to_clean = buf->next;
drivers/net/ethernet/cisco/enic/vnic_wq.h
41
struct vnic_wq_buf *next;
drivers/net/ethernet/dec/tulip/de2104x.c
556
goto next;
drivers/net/ethernet/dec/tulip/de2104x.c
563
goto next;
drivers/net/ethernet/dec/tulip/de2104x.c
592
next:
drivers/net/ethernet/dec/tulip/tulip.h
389
struct mediainfo *next;
drivers/net/ethernet/dlink/dl2k.c
345
int cid, next;
drivers/net/ethernet/dlink/dl2k.c
377
next = psib[i++];
drivers/net/ethernet/dlink/dl2k.c
378
if ((cid == 0 && next == 0) || (cid == 0xff && next == 0xff)) {
drivers/net/ethernet/dlink/dl2k.c
395
j = (next - i > 255) ? 255 : next - i;
drivers/net/ethernet/dlink/dl2k.c
407
i = next;
drivers/net/ethernet/ec_bhf.c
379
u32 next;
drivers/net/ethernet/ec_bhf.c
382
next = (u8 *)(desc + 1) - priv->rx_buf.buf;
drivers/net/ethernet/ec_bhf.c
384
next = 0;
drivers/net/ethernet/ec_bhf.c
385
next |= RXHDR_NEXT_VALID;
drivers/net/ethernet/ec_bhf.c
386
desc->header.next = cpu_to_le32(next);
drivers/net/ethernet/ec_bhf.c
76
__le32 next;
drivers/net/ethernet/engleder/tsnep_hw.h
198
__le64 next;
drivers/net/ethernet/engleder/tsnep_hw.h
228
__le64 next;
drivers/net/ethernet/engleder/tsnep_main.c
1028
entry->desc->next = __cpu_to_le64(next_entry->desc_dma);
drivers/net/ethernet/engleder/tsnep_main.c
324
entry->desc->next = __cpu_to_le64(next_entry->desc_dma);
drivers/net/ethernet/engleder/tsnep_selftests.c
194
static int get_operation(struct tsnep_gcl *gcl, u64 system_time, u64 *next)
drivers/net/ethernet/engleder/tsnep_selftests.c
200
*next = cycle_start;
drivers/net/ethernet/engleder/tsnep_selftests.c
202
*next += gcl->operation[i].interval;
drivers/net/ethernet/engleder/tsnep_selftests.c
203
if (*next > system_time)
drivers/net/ethernet/fealnx.c
1316
struct fealnx_desc *next;
drivers/net/ethernet/fealnx.c
1327
next = np->cur_tx_copy->next_desc_logical;
drivers/net/ethernet/fealnx.c
1328
next->skbuff = skb;
drivers/net/ethernet/fealnx.c
1329
next->control = TXIC | TXLD | CRCEnable | PADEnable;
drivers/net/ethernet/fealnx.c
1330
next->control |= (skb->len << PKTSShift); /* pkt size */
drivers/net/ethernet/fealnx.c
1331
next->control |= ((skb->len - BPT) << TBSShift); /* buf size */
drivers/net/ethernet/fealnx.c
1335
next->buffer = dma_map_single(&ep->pci_dev->dev,
drivers/net/ethernet/fealnx.c
1339
next->status = TXOWN;
drivers/net/ethernet/fealnx.c
1342
np->cur_tx_copy = next->next_desc_logical;
drivers/net/ethernet/fealnx.c
1492
struct fealnx_desc *next;
drivers/net/ethernet/fealnx.c
1494
next = np->cur_tx->next_desc_logical;
drivers/net/ethernet/fealnx.c
1495
tx_status = next->status;
drivers/net/ethernet/fealnx.c
1496
tx_control = next->control;
drivers/net/ethernet/freescale/fman/fman_mac.h
190
hash_entry = ETH_HASH_ENTRY_OBJ(addr_lst->next);
drivers/net/ethernet/freescale/gianfar.c
2108
struct txbd8 *bdp, *next = NULL;
drivers/net/ethernet/freescale/gianfar.c
2152
next = next_txbd(bdp, base, tx_ring_size);
drivers/net/ethernet/freescale/gianfar.c
2153
buflen = be16_to_cpu(next->length) +
drivers/net/ethernet/freescale/gianfar.c
2172
bdp = next;
drivers/net/ethernet/freescale/ucc_geth.c
215
struct list_head *node = lh->next;
drivers/net/ethernet/google/gve/gve.h
205
s16 next;
drivers/net/ethernet/google/gve/gve.h
448
s16 next;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
28
rx->dqo.free_buf_states = buf_state->next;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
31
buf_state->next = buffer_id;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
45
return buf_state->next == buffer_id;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
53
buf_state->next = rx->dqo.free_buf_states;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
70
list->head = buf_state->next;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
71
if (buf_state->next == -1)
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
75
buf_state->next = buffer_id;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
85
buf_state->next = -1;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
93
rx->dqo.buf_states[tail].next = buffer_id;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
57
rx->dqo.buf_states[i].next = i + 1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
58
rx->dqo.buf_states[rx->dqo.num_buf_states - 1].next = -1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
588
rx->ctx.skb_tail->next = skb;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1138
tx->dqo.pending_packets[old_tail].next = index;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1140
pending_packet->next = -1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1151
next_index = pkt->next;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1157
tx->dqo.pending_packets[prev_index].next = next_index;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1297
next_index = pending_packet->next;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1347
next_index = pending_packet->next;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
146
tx->dqo_tx.free_pending_packets = pending_packet->next;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
162
pending_packet->next = old_head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
356
tx->dqo.pending_packets[i].next = i + 1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
358
tx->dqo.pending_packets[tx->dqo.num_pending_packets - 1].next = -1;
drivers/net/ethernet/hisilicon/hisi_femac.c
280
goto next;
drivers/net/ethernet/hisilicon/hisi_femac.c
287
next:
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
548
goto next;
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
555
next:
drivers/net/ethernet/hisilicon/hns/hns_dsaf_ppe.c
122
ppe_cb->next = NULL;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_ppe.h
74
struct hns_ppe_cb *next; /* pointer to next ppe device */
drivers/net/ethernet/hisilicon/hns3/hnae3.h
383
struct hnae3_ring_chain_node *next;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4165
ring->tail_skb->next = new_skb;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4581
while (cur_chain->next)
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4582
cur_chain = cur_chain->next;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4590
cur_chain->next = chain;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4602
ring = ring->next;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4625
chain = cur_chain->next;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4642
chain_tmp = chain->next;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4651
ring->next = group->ring;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
780
ring = ring->next;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.h
469
struct hns3_enet_ring *next;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.h
697
for ((pos) = (head).ring; (pos); (pos) = (pos)->next)
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
5010
for (node = ring_chain; node; node = node->next) {
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
5231
new_rule->rule_node.next = old_rule->rule_node.next;
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c
161
chain = head->next;
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c
164
chain_tmp = chain->next;
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c
230
cur_chain->next = new_chain;
drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c
671
for (node = ring_chain; node; node = node->next) {
drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c
682
if (i == HCLGE_MBX_MAX_RING_CHAIN_PARAM_NUM || !node->next) {
drivers/net/ethernet/huawei/hinic/hinic_rx.c
302
curr_skb->next = skb;
drivers/net/ethernet/i825xx/82596.c
1079
tbd->next = I596_NULL;
drivers/net/ethernet/i825xx/82596.c
211
struct i596_tbd *next;
drivers/net/ethernet/i825xx/ether1.c
446
int i, status, addr, next, next2;
drivers/net/ethernet/i825xx/ether1.c
479
next = addr + RFD_SIZE + RBD_SIZE + ETH_FRAME_LEN + 10;
drivers/net/ethernet/i825xx/ether1.c
480
next2 = next + RFD_SIZE + RBD_SIZE + ETH_FRAME_LEN + 10;
drivers/net/ethernet/i825xx/ether1.c
483
next = RX_AREA_START;
drivers/net/ethernet/i825xx/ether1.c
492
init_rfd.rfd_link = next;
drivers/net/ethernet/i825xx/ether1.c
493
init_rbd.rbd_link = next + RFD_SIZE;
drivers/net/ethernet/i825xx/ether1.c
498
addr = next;
drivers/net/ethernet/i825xx/lib82596.c
1029
tbd->next = I596_NULL;
drivers/net/ethernet/i825xx/lib82596.c
177
u32 next;
drivers/net/ethernet/i825xx/sun3_82586.c
595
p->xmit_buffs[i]->next = 0xffff;
drivers/net/ethernet/i825xx/sun3_82586.c
647
rfd[i].next = make16(rfd + (i+1) % (p->num_recv_buffs+rfdadd) );
drivers/net/ethernet/i825xx/sun3_82586.c
662
rbd[i].next = make16((rbd + (i+1) % p->num_recv_buffs));
drivers/net/ethernet/i825xx/sun3_82586.c
805
rbd = (struct rbd_struct *) make32(rbd->next);
drivers/net/ethernet/i825xx/sun3_82586.c
823
p->rfd_top = (struct rfd_struct *) make32(p->rfd_top->next); /* step to next RFD */
drivers/net/ethernet/i825xx/sun3_82586.c
863
rfds = (struct rfd_struct *) make32(rfds->next);
drivers/net/ethernet/i825xx/sun3_82586.h
135
unsigned short next; /* linkoffset to next RFD */
drivers/net/ethernet/i825xx/sun3_82586.h
166
unsigned short next; /* pointeroffset to next RBD */
drivers/net/ethernet/i825xx/sun3_82586.h
311
unsigned short next; /* pointeroffset to next TBD */
drivers/net/ethernet/ibm/emac/core.c
1758
goto next;
drivers/net/ethernet/ibm/emac/core.c
1764
goto next;
drivers/net/ethernet/ibm/emac/core.c
1788
next:
drivers/net/ethernet/ibm/emac/core.c
1828
goto next;
drivers/net/ethernet/ibm/ibmvnic.c
3539
union sub_crq *next;
drivers/net/ethernet/ibm/ibmvnic.c
3553
next = ibmvnic_next_scrq(adapter, rx_scrq);
drivers/net/ethernet/ibm/ibmvnic.c
3555
be64_to_cpu(next->rx_comp.correlator);
drivers/net/ethernet/ibm/ibmvnic.c
3557
if (next->rx_comp.rc) {
drivers/net/ethernet/ibm/ibmvnic.c
3559
be16_to_cpu(next->rx_comp.rc));
drivers/net/ethernet/ibm/ibmvnic.c
3561
next->rx_comp.first = 0;
drivers/net/ethernet/ibm/ibmvnic.c
3567
next->rx_comp.first = 0;
drivers/net/ethernet/ibm/ibmvnic.c
3572
length = be32_to_cpu(next->rx_comp.len);
drivers/net/ethernet/ibm/ibmvnic.c
3573
offset = be16_to_cpu(next->rx_comp.off_frame_data);
drivers/net/ethernet/ibm/ibmvnic.c
3574
flags = next->rx_comp.flags;
drivers/net/ethernet/ibm/ibmvnic.c
3587
ntohs(next->rx_comp.vlan_tci));
drivers/net/ethernet/ibm/ibmvnic.c
3590
next->rx_comp.first = 0;
drivers/net/ethernet/ibm/ibmvnic.c
4249
union sub_crq *next;
drivers/net/ethernet/ibm/ibmvnic.c
4256
next = ibmvnic_next_scrq(adapter, scrq);
drivers/net/ethernet/ibm/ibmvnic.c
4257
for (i = 0; i < next->tx_comp.num_comps; i++) {
drivers/net/ethernet/ibm/ibmvnic.c
4258
index = be32_to_cpu(next->tx_comp.correlators[i]);
drivers/net/ethernet/ibm/ibmvnic.c
4271
if (next->tx_comp.rcs[i]) {
drivers/net/ethernet/ibm/ibmvnic.c
4273
next->tx_comp.rcs[i]);
drivers/net/ethernet/ibm/ibmvnic.c
4289
next->tx_comp.first = 0;
drivers/net/ethernet/intel/e100.c
1815
cb = nic->cb_to_clean = cb->next) {
drivers/net/ethernet/intel/e100.c
1859
nic->cb_to_clean = nic->cb_to_clean->next;
drivers/net/ethernet/intel/e100.c
1886
cb->next = (i + 1 < count) ? cb + 1 : nic->cbs;
drivers/net/ethernet/intel/e100.c
2056
for (rx = nic->rx_to_clean; rx->skb; rx = nic->rx_to_clean = rx->next) {
drivers/net/ethernet/intel/e100.c
2077
for (rx = nic->rx_to_use; !rx->skb; rx = nic->rx_to_use = rx->next) {
drivers/net/ethernet/intel/e100.c
2163
rx->next = (i + 1 < count) ? rx + 1 : nic->rxs;
drivers/net/ethernet/intel/e100.c
416
struct rx *next, *prev;
drivers/net/ethernet/intel/e100.c
495
struct cb *next, *prev;
drivers/net/ethernet/intel/e100.c
858
nic->cb_to_use = cb->next;
drivers/net/ethernet/intel/e100.c
891
nic->cb_to_send = nic->cb_to_send->next;
drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c
101
.next = fm10k_dbg_desc_seq_next,
drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c
108
.next = fm10k_dbg_desc_seq_next,
drivers/net/ethernet/intel/i40e/i40e_main.c
2300
struct i40e_new_mac_filter *i40e_next_filter(struct i40e_new_mac_filter *next)
drivers/net/ethernet/intel/i40e/i40e_main.c
2302
hlist_for_each_entry_continue(next, hlist) {
drivers/net/ethernet/intel/i40e/i40e_main.c
2303
if (!is_broadcast_ether_addr(next->f->macaddr))
drivers/net/ethernet/intel/i40e/i40e_main.c
2304
return next;
drivers/net/ethernet/intel/i40e/i40e_main.c
4537
tx_ring->next = q_vector->tx.ring;
drivers/net/ethernet/intel/i40e/i40e_main.c
4546
xdp_ring->next = q_vector->tx.ring;
drivers/net/ethernet/intel/i40e/i40e_main.c
4552
rx_ring->next = q_vector->rx.ring;
drivers/net/ethernet/intel/i40e/i40e_main.c
5023
u32 next;
drivers/net/ethernet/intel/i40e/i40e_main.c
5039
next = FIELD_GET(I40E_QINT_TQCTL_NEXTQ_INDX_MASK,
drivers/net/ethernet/intel/i40e/i40e_main.c
5051
qp = next;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2045
u32 next = rx_ring->next_to_clean, i = 0;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2051
rx_buffer = i40e_rx_bi(rx_ring, next);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2052
if (++next == rx_ring->count)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2053
next = 0;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2064
if (next == rx_ring->next_to_process)
drivers/net/ethernet/intel/i40e/i40e_txrx.h
328
struct i40e_ring *next; /* pointer to next ring in q_vector */
drivers/net/ethernet/intel/i40e/i40e_txrx.h
448
for (pos = (head).ring; pos != NULL; pos = pos->next)
drivers/net/ethernet/intel/iavf/iavf_main.c
436
rx_ring->next = q_vector->rx.ring;
drivers/net/ethernet/intel/iavf/iavf_main.c
462
tx_ring->next = q_vector->tx.ring;
drivers/net/ethernet/intel/iavf/iavf_txrx.h
219
struct iavf_ring *next; /* pointer to next ring in q_vector */
drivers/net/ethernet/intel/iavf/iavf_txrx.h
310
for (pos = (head).ring; pos != NULL; pos = pos->next)
drivers/net/ethernet/intel/ice/ice_base.c
955
tx_ring->next = q_vector->tx.tx_ring;
drivers/net/ethernet/intel/ice/ice_base.c
972
rx_ring->next = q_vector->rx.rx_ring;
drivers/net/ethernet/intel/ice/ice_lib.c
263
int next;
drivers/net/ethernet/intel/ice/ice_lib.c
266
next = curr + 1;
drivers/net/ethernet/intel/ice/ice_lib.c
273
next = ICE_NO_VSI;
drivers/net/ethernet/intel/ice/ice_lib.c
275
next = i;
drivers/net/ethernet/intel/ice/ice_lib.c
277
return next;
drivers/net/ethernet/intel/ice/ice_main.c
2712
xdp_ring->next = q_vector->tx.tx_ring;
drivers/net/ethernet/intel/ice/ice_txrx.c
863
goto next;
drivers/net/ethernet/intel/ice/ice_txrx.c
875
next:
drivers/net/ethernet/intel/ice/ice_txrx.h
301
struct ice_rx_ring *next; /* pointer to next ring in q_vector */
drivers/net/ethernet/intel/ice/ice_txrx.h
365
struct ice_tx_ring *next; /* pointer to next ring in q_vector */
drivers/net/ethernet/intel/ice/ice_txrx.h
441
for (pos = (head).rx_ring; pos; pos = pos->next)
drivers/net/ethernet/intel/ice/ice_txrx.h
444
for (pos = (head).tx_ring; pos; pos = pos->next)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2359
goto next;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2366
next:
drivers/net/ethernet/intel/idpf/xdp.c
272
goto next;
drivers/net/ethernet/intel/idpf/xdp.c
282
next:
drivers/net/ethernet/intel/igb/igb_main.c
181
.next = NULL,
drivers/net/ethernet/intel/ixgbe/ixgbe.h
357
struct ixgbe_ring *next; /* pointer to next ring in q_vector */
drivers/net/ethernet/intel/ixgbe/ixgbe.h
493
for (pos = (head).ring; pos != NULL; pos = pos->next)
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
812
ring->next = head->ring;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1209
unsigned int next)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1225
tx_ring->next_to_use, next,
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1226
tx_ring->tx_buffer_info[next].time_stamp, jiffies);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
153
.next = NULL,
drivers/net/ethernet/intel/ixgbevf/ixgbevf.h
226
for (pos = (head).ring; pos != NULL; pos = pos->next)
drivers/net/ethernet/intel/ixgbevf/ixgbevf.h
91
struct ixgbevf_ring *next;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
2736
ring->next = head->ring;
drivers/net/ethernet/korina.c
618
goto next;
drivers/net/ethernet/korina.c
635
goto next;
drivers/net/ethernet/korina.c
670
next:
drivers/net/ethernet/lantiq_xrx200.c
261
ch->skb_tail->next = skb;
drivers/net/ethernet/marvell/mvneta.c
2464
goto next;
drivers/net/ethernet/marvell/mvneta.c
2478
goto next;
drivers/net/ethernet/marvell/mvneta.c
2491
goto next;
drivers/net/ethernet/marvell/mvneta.c
2496
goto next;
drivers/net/ethernet/marvell/mvneta.c
2509
goto next;
drivers/net/ethernet/marvell/mvneta.c
2517
next:
drivers/net/ethernet/marvell/mvpp2/mvpp2.h
1190
u8 next;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4152
i = queue->next;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4153
queue->next = (i + 1) & 31;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3177
int mce, u8 op, u16 pcifunc, int next,
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3193
aq_req.mce.next = next;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2368
u16 index, next, end;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2378
next = find_next_bit(map, end, index);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2379
if (*max_area < (next - index)) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2380
*max_area = next - index;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2384
if (next < end) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2385
start = next + 1;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2397
u16 index, next;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2408
next = find_next_bit(map, end, index);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2409
if (next <= end) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2410
fcnt += next - index;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
2411
start = next + 1;
drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h
704
uint64_t next : 16;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
1066
sw_rx_sc = rcu_dereference_bh(sw_rx_sc->next)) {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
1099
sw_rx_sc = rcu_dereference_bh(sw_rx_sc->next)) {
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
443
for (tmp = head->next; tmp != head; tmp = tmp->next) {
drivers/net/ethernet/marvell/prestera/prestera_counter.c
353
goto next;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
398
goto next;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
401
next:
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
152
dma_addr_t next)
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
154
desc->next = cpu_to_le32(prestera_sdma_map(sdma, next));
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
22
__le32 next;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
379
struct prestera_sdma_buf *head, *tail, *next, *prev;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
390
next = head;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
391
prev = next;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
394
err = prestera_sdma_buf_init(sdma, next);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
398
err = prestera_sdma_rx_skb_alloc(sdma, next);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
402
prestera_sdma_rx_desc_init(sdma, next->desc,
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
403
next->buf_dma);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
406
next->desc_dma);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
408
prev = next;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
409
next++;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
437
dma_addr_t next)
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
439
desc->next = cpu_to_le32(prestera_sdma_map(sdma, next));
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
524
struct prestera_sdma_buf *head, *tail, *next, *prev;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
538
next = head;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
539
prev = next;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
546
err = prestera_sdma_buf_init(sdma, next);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
550
next->is_used = false;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
552
prestera_sdma_tx_desc_init(sdma, next->desc);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
555
next->desc_dma);
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
557
prev = next;
drivers/net/ethernet/marvell/prestera/prestera_rxtx.c
558
next++;
drivers/net/ethernet/marvell/skge.c
1005
} while ((e = e->next) != ring->start);
drivers/net/ethernet/marvell/skge.c
1032
} while ((e = e->next) != ring->start);
drivers/net/ethernet/marvell/skge.c
2802
e = e->next;
drivers/net/ethernet/marvell/skge.c
2829
skge->tx_ring.to_use = e->next;
drivers/net/ethernet/marvell/skge.c
2844
e = e->next;
drivers/net/ethernet/marvell/skge.c
2876
for (e = skge->tx_ring.to_clean; e != skge->tx_ring.to_use; e = e->next) {
drivers/net/ethernet/marvell/skge.c
3158
for (e = ring->to_clean; e != ring->to_use; e = e->next) {
drivers/net/ethernet/marvell/skge.c
3208
for (e = ring->to_clean; prefetch(e->next), work_done < budget; e = e->next) {
drivers/net/ethernet/marvell/skge.c
3705
for (e = skge->tx_ring.to_clean; e != skge->tx_ring.to_use; e = e->next) {
drivers/net/ethernet/marvell/skge.c
3713
for (e = skge->rx_ring.to_clean; ; e = e->next) {
drivers/net/ethernet/marvell/skge.c
928
e->next = ring->start;
drivers/net/ethernet/marvell/skge.c
931
e->next = e + 1;
drivers/net/ethernet/marvell/skge.h
2389
struct skge_element *next;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2659
int next = (i + 1) % ring_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2660
u32 next_ptr = ring->phys + next * sz;
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
275
goto next;
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
299
next:
drivers/net/ethernet/mellanox/mlx4/alloc.c
292
struct mlx4_zone_entry *next = list_first_entry(&entry->list,
drivers/net/ethernet/mellanox/mlx4/alloc.c
293
typeof(*next),
drivers/net/ethernet/mellanox/mlx4/alloc.c
296
if (next->priority == entry->priority)
drivers/net/ethernet/mellanox/mlx4/alloc.c
297
list_add_tail(&next->prio_list, &entry->prio_list);
drivers/net/ethernet/mellanox/mlx4/cmd.c
130
int next;
drivers/net/ethernet/mellanox/mlx4/cmd.c
2632
priv->cmd.context[i].next = i + 1;
drivers/net/ethernet/mellanox/mlx4/cmd.c
2640
priv->cmd.context[priv->cmd.max_cmds - 1].next = -1;
drivers/net/ethernet/mellanox/mlx4/cmd.c
352
cmd->free_head = context->next;
drivers/net/ethernet/mellanox/mlx4/cmd.c
399
context->next = cmd->free_head;
drivers/net/ethernet/mellanox/mlx4/cmd.c
692
cmd->free_head = context->next;
drivers/net/ethernet/mellanox/mlx4/cmd.c
766
context->next = cmd->free_head;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
164
struct list_head next;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
316
list_add_tail(&filter->next, &priv->filters);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
329
list_del(&filter->next);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
432
list_for_each_entry_safe(filter, tmp, &priv->filters, next) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
433
list_move(&filter->next, &del_list);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
438
list_for_each_entry_safe(filter, tmp, &del_list, next) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
451
list_for_each_entry_safe(filter, tmp, &priv->filters, next) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
460
list_move(&filter->next, &del_list);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
468
if (last_filter && (&last_filter->next != priv->filters.next))
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
469
list_move(&priv->filters, &last_filter->next);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
473
list_for_each_entry_safe(filter, tmp, &del_list, next)
drivers/net/ethernet/mellanox/mlx4/en_port.c
135
static unsigned long en_stats_adder(__be64 *start, __be64 *next, int num)
drivers/net/ethernet/mellanox/mlx4/en_port.c
140
int offset = next - start;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
739
goto next;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
743
goto next;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
771
goto next;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
778
goto next;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
826
goto next;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
836
goto next;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
849
goto next;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
858
goto next;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
931
next:
drivers/net/ethernet/mellanox/mlx4/icm.h
101
NULL : list_entry(icm->chunk_list.next,
drivers/net/ethernet/mellanox/mlx4/icm.h
114
if (iter->chunk->list.next == &iter->icm->chunk_list) {
drivers/net/ethernet/mellanox/mlx4/icm.h
119
iter->chunk = list_entry(iter->chunk->list.next,
drivers/net/ethernet/mellanox/mlx4/pd.c
184
uar = list_entry(priv->bf_list.next, struct mlx4_uar, bf_list);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1043
if (ent->in->next)
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1044
lay->in_ptr = cpu_to_be64(ent->in->next->dma);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1046
if (ent->out->next)
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1047
lay->out_ptr = cpu_to_be64(ent->out->next->dma);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1371
struct mlx5_cmd_mailbox *next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1382
next = to->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1384
if (!next) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1390
block = next->buf;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1395
next = next->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1404
struct mlx5_cmd_mailbox *next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1415
next = from->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1417
if (!next) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1423
block = next->buf;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1428
next = next->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1450
mailbox->next = NULL;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1489
tmp->next = head;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1490
block->next = cpu_to_be64(tmp->next ? tmp->next->dma : 0);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1495
msg->next = head;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1500
tmp = head->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1512
struct mlx5_cmd_mailbox *head = msg->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1513
struct mlx5_cmd_mailbox *next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1516
next = head->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1518
head = next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1914
msg = list_entry(ch->head.next, typeof(*msg), list);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
263
struct mlx5_cmd_mailbox *next = msg->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
267
for (i = 0; i < n && next; i++) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
268
calc_block_sig(next->buf);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
269
next = next->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
309
struct mlx5_cmd_mailbox *next = ent->out->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
319
for (i = 0; i < n && next; i++) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
320
err = verify_block_sig(next->buf);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
324
next = next->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
877
struct mlx5_cmd_mailbox *next = msg->next;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
911
for (i = 0; i < n && next; i++) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
914
dump_buf(next->buf, dump_len, 1, offset, ent->idx);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
918
dump_buf(next->buf, sizeof(struct mlx5_cmd_prot_block), 0, offset,
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
922
next = next->next;
drivers/net/ethernet/mellanox/mlx5/core/en.h
229
struct mlx5_wqe_srq_next_seg next;
drivers/net/ethernet/mellanox/mlx5/core/en/mapping.c
165
struct mapping_item *mi, *next;
drivers/net/ethernet/mellanox/mlx5/core/en/mapping.c
172
list_for_each_entry_safe(mi, next, &ctx->pending_list, list) {
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
67
struct mlx5e_neigh_hash_entry *next = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
71
for (next = nhe ?
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
79
next;
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
80
next = list_next_or_null_rcu(&rpriv->neigh_update.neigh_list,
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
81
&next->neigh_list,
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
84
if (mlx5e_rep_neigh_entry_hold(next))
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
92
return next;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
325
struct mlx5e_encap_entry *next = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
331
for (next = e ?
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
339
next;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
340
next = list_next_or_null_rcu(&nhe->encap_list,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
341
&next->encap_list,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
344
if (mlx5e_encap_take(next))
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
352
if (!next)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
353
return next;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
356
wait_for_completion(&next->res_ready);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
358
if (!match(next)) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
359
e = next;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
363
return next;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1303
&wqe->next.next_wqe_index);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1833
mlx5_wq_ll_pop(wq, cqe->wqe_id, &wqe->next.next_wqe_index);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2282
mlx5_wq_ll_pop(wq, cqe->wqe_id, &wqe->next.next_wqe_index);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2342
mlx5_wq_ll_pop(wq, cqe->wqe_id, &wqe->next.next_wqe_index);
drivers/net/ethernet/mellanox/mlx5/core/esw/bridge_debugfs.c
15
.next = mlx5_esw_bridge_debugfs_next,
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
699
fte->dup->children.next,
drivers/net/ethernet/mellanox/mlx5/core/health.c
745
unsigned long next;
drivers/net/ethernet/mellanox/mlx5/core/health.c
747
get_random_bytes(&next, sizeof(next));
drivers/net/ethernet/mellanox/mlx5/core/health.c
748
next %= HZ;
drivers/net/ethernet/mellanox/mlx5/core/health.c
749
next += jiffies + msecs_to_jiffies(mlx5_tout_ms(dev, HEALTH_POLL_INTERVAL));
drivers/net/ethernet/mellanox/mlx5/core/health.c
751
return next;
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1437
struct mlx5_core_dev *peer_dev, *next = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1456
next = peer_dev;
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1461
if (next) {
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1462
mdev->clock = next->clock;
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1464
mdev->clock_info = next->clock_info;
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1478
struct mlx5_core_dev *peer_dev, *next = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1485
next = peer_dev;
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1490
if (next) {
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1495
cpriv->mdev = next;
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1561
struct mlx5_core_dev *peer_dev, *next = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1576
next = peer_dev;
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c
1583
mlx5_clock_arm_pps_in_event(clock, next, mdev);
drivers/net/ethernet/mellanox/mlx5/core/lib/crypto.c
535
struct mlx5_crypto_dek *dek, *next;
drivers/net/ethernet/mellanox/mlx5/core/lib/crypto.c
537
list_for_each_entry_safe(dek, next, &pool->wait_for_free, entry) {
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1997
struct mlx5_roce_macsec_rx_rule *rx_rule, *next;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
2002
list_for_each_entry_safe(rx_rule, next, rx_rules_list, entry) {
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
2020
struct mlx5_roce_macsec_tx_rule *tx_rule, *next;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
2025
list_for_each_entry_safe(tx_rule, next, tx_rules_list, entry) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
227
struct mlx5hws_matcher *next = NULL;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
242
next = tmp_matcher;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
248
if (next)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
250
list_add_tail(&matcher->list_node, &next->list_node);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
256
if (next) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
261
next->match_ste.rtc_0_id,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
262
next->match_ste.rtc_1_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
312
struct mlx5hws_matcher *first, *last, *prev, *next;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
325
next = list_next_entry(matcher, list_node);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
362
if (!next)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
367
next->match_ste.rtc_0_id,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
368
next->match_ste.rtc_1_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
384
if (!prev || !next ||
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
392
next->match_ste.rtc_0_id,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
393
next->match_ste.rtc_1_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
398
struct mlx5hws_matcher *next = NULL, *prev = NULL;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
412
next = list_next_entry(matcher, list_node);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
416
if (next) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
421
next->match_ste.rtc_0_id,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
422
next->match_ste.rtc_1_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/table.c
375
list_for_each_entry(src_tbl, &dst_tbl->default_miss.head, default_miss.next) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/table.c
486
list_del_init(&tbl->default_miss.next);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/table.c
489
list_add(&tbl->default_miss.next, &miss_tbl->default_miss.head);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/table.h
10
struct list_head next;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
1153
.next = dr_dump_next,
drivers/net/ethernet/mellanox/mlx5/core/uar.c
232
up = list_entry(head->next, struct mlx5_uars_page, list);
drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c
28
void *next;
drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c
39
next = (void *) tlv + tlv_len;
drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c
40
return mlxfw_mfa2_tlv_get(mfa2_file, next);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1044
block->cur_set->next = set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
428
block->first_set->next = block->cur_set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
449
next_set = set->next;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
481
prev_set->next = set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
509
if (WARN_ON(!block->first_set->next))
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
511
return block->first_set->next->kvdl_index;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
99
struct mlxsw_afa_set *next; /* Pointer to the next set. */
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
219
if (region->list.next != &group->region_list &&
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
502
if (pos->next != &vgroup->vregion_list) { /* not last */
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
503
vregion2 = list_entry(pos->next, typeof(*vregion2),
drivers/net/ethernet/meta/fbnic/fbnic_fw_log.c
104
list_for_each_entry_safe_reverse(tail, next, &log->entries, list) {
drivers/net/ethernet/meta/fbnic/fbnic_fw_log.c
74
struct fbnic_fw_log_entry *entry, *head, *tail, *next;
drivers/net/ethernet/micrel/ksz884x.c
1554
*desc = &info->ring[info->next];
drivers/net/ethernet/micrel/ksz884x.c
1555
info->next++;
drivers/net/ethernet/micrel/ksz884x.c
1556
info->next &= info->mask;
drivers/net/ethernet/micrel/ksz884x.c
3468
previous->phw->next = cpu_to_le32(phys);
drivers/net/ethernet/micrel/ksz884x.c
3470
previous->phw->next = cpu_to_le32(desc_info->ring_phys);
drivers/net/ethernet/micrel/ksz884x.c
3475
desc_info->last = desc_info->next = 0;
drivers/net/ethernet/micrel/ksz884x.c
3499
info->last = info->next = 0;
drivers/net/ethernet/micrel/ksz884x.c
4657
int next;
drivers/net/ethernet/micrel/ksz884x.c
4666
next = info->next;
drivers/net/ethernet/micrel/ksz884x.c
4669
desc = &info->ring[next];
drivers/net/ethernet/micrel/ksz884x.c
4683
next++;
drivers/net/ethernet/micrel/ksz884x.c
4684
next &= info->mask;
drivers/net/ethernet/micrel/ksz884x.c
4686
info->next = next;
drivers/net/ethernet/micrel/ksz884x.c
4693
int next;
drivers/net/ethernet/micrel/ksz884x.c
4702
next = info->next;
drivers/net/ethernet/micrel/ksz884x.c
4705
desc = &info->ring[next];
drivers/net/ethernet/micrel/ksz884x.c
4728
next++;
drivers/net/ethernet/micrel/ksz884x.c
4729
next &= info->mask;
drivers/net/ethernet/micrel/ksz884x.c
4731
info->next = next;
drivers/net/ethernet/micrel/ksz884x.c
4738
int next;
drivers/net/ethernet/micrel/ksz884x.c
4747
next = info->next;
drivers/net/ethernet/micrel/ksz884x.c
4750
desc = &info->ring[next];
drivers/net/ethernet/micrel/ksz884x.c
4787
next++;
drivers/net/ethernet/micrel/ksz884x.c
4788
next &= info->mask;
drivers/net/ethernet/micrel/ksz884x.c
4790
info->next = next;
drivers/net/ethernet/micrel/ksz884x.c
876
u32 next;
drivers/net/ethernet/micrel/ksz884x.c
941
int next;
drivers/net/ethernet/microchip/lan743x_main.c
2520
rx->skb_tail->next = skb;
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
273
u32 base, next, max_list;
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
290
next = base;
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
293
clear_bit(next, free_list);
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
296
lan_rmw(QSYS_TAS_CFG_CTRL_GCL_ENTRY_NUM_SET(next),
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
300
next = lan_rd(lan966x, QSYS_TAS_GCL_CT_CFG2);
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
301
next = QSYS_TAS_GCL_CT_CFG2_NEXT_GCL_GET(next);
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
302
} while (base != next);
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
333
u32 i, base, next;
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
350
next = base;
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
352
lan_rmw(QSYS_TAS_CFG_CTRL_GCL_ENTRY_NUM_SET(next),
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
358
next = base;
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
360
next = find_next_bit(free_list, LAN966X_TAPRIO_NUM_GCL,
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
361
next + 1);
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
363
lan966x_taprio_gcl_setup_entry(port, &qopt->entries[i], next);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
119
u32 itr, next;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
124
next = sparx5_lg_get_next(sparx5, layer, group, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
125
if (itr == next)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
128
itr = next;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
172
u32 itr, next;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
182
next = sparx5_lg_get_next(sparx5, layer, i, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
188
if (itr == next)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
191
itr = next;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
216
u32 idx, u32 *prev, u32 *next, u32 *first)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
222
*next = *first;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
226
*next = sparx5_lg_get_next(sparx5, layer, group, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
231
if (itr == *next)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
235
itr = *next;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
273
u32 first, next, prev;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
277
WARN_ON(sparx5_lg_get_adjacent(sparx5, layer, group, idx, &prev, &next,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
285
next = prev;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
288
first = next;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
289
next = idx;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
295
return sparx5_lg_conf_set(sparx5, layer, group, first, idx, next,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
302
u32 first, next, old_group;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
322
next = idx;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
324
next = sparx5_lg_get_first(sparx5, layer, new_group);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
326
return sparx5_lg_conf_set(sparx5, layer, new_group, first, idx, next,
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
116
u32 itr, next;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
121
next = sparx5_sdlb_group_get_next(sparx5, group, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
122
if (itr == next)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
125
itr = next;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
139
u32 idx, u32 *prev, u32 *next,
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
146
*next = *first;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
150
*next = sparx5_sdlb_group_get_next(sparx5, group, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
155
if (itr == *next)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
159
itr = *next;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
165
u32 itr, next;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
171
next = sparx5_sdlb_group_get_next(sparx5, group, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
172
if (itr == next)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
175
itr = next;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
210
u32 itr, next;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
220
next = sparx5_sdlb_group_get_next(sparx5, i, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
226
if (itr == next)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
229
itr = next;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
237
u32 first, u32 next, bool empty)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
246
spx5_wr(ANA_AC_SDLB_XLB_NEXT_LBSET_NEXT_SET(next) |
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
262
u32 first, next;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
268
next = idx;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
270
next = sparx5_sdlb_group_get_first(sparx5, group);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
272
return sparx5_sdlb_group_link(sparx5, group, idx, first, next, false);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
277
u32 first, next, prev;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
280
if (sparx5_sdlb_group_get_adjacent(sparx5, group, idx, &prev, &next,
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
292
next = prev;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
295
first = next;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
296
next = idx;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
302
return sparx5_sdlb_group_link(sparx5, group, idx, first, next, empty);
drivers/net/ethernet/microchip/vcap/vcap_api_kunit.c
2033
struct vcap_client_keyfield *ckf, *next;
drivers/net/ethernet/microchip/vcap/vcap_api_kunit.c
2056
list_for_each_entry_safe(ckf, next, &ri.data.keyfields, ctrl.list) {
drivers/net/ethernet/microchip/vcap/vcap_api_kunit.c
2157
struct vcap_client_keyfield *ckf, *next;
drivers/net/ethernet/microchip/vcap/vcap_api_kunit.c
2181
list_for_each_entry_safe(ckf, next, &ri.data.keyfields, ctrl.list) {
drivers/net/ethernet/microsoft/mana/gdma_main.c
1587
const struct cpumask *next, *prev = cpu_none_mask;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1595
for_each_numa_hop_mask(next, node) {
drivers/net/ethernet/microsoft/mana/gdma_main.c
1596
weight = cpumask_weight_andnot(next, prev);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1598
cpumask_andnot(cpus, next, prev);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1614
prev = next;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2915
struct sk_buff *segs, *curr, *next;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2924
skb_list_walk_safe(segs, curr, next) {
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2929
skb_list_walk_safe(next, curr, next) {
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2930
curr->next = NULL;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
4038
.next = NULL,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
22
#define nfp_for_each_insn_walk2(nfp_prog, pos, next) \
drivers/net/ethernet/netronome/nfp/bpf/jit.c
24
next = list_next_entry(pos, l); \
drivers/net/ethernet/netronome/nfp/bpf/jit.c
26
&(nfp_prog)->insns != &next->l; \
drivers/net/ethernet/netronome/nfp/bpf/jit.c
28
next = nfp_meta_next(pos))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
30
#define nfp_for_each_insn_walk3(nfp_prog, pos, next, next2) \
drivers/net/ethernet/netronome/nfp/bpf/jit.c
32
next = list_next_entry(pos, l), \
drivers/net/ethernet/netronome/nfp/bpf/jit.c
33
next2 = list_next_entry(next, l); \
drivers/net/ethernet/netronome/nfp/bpf/jit.c
35
&(nfp_prog)->insns != &next->l && \
drivers/net/ethernet/netronome/nfp/bpf/jit.c
38
next = nfp_meta_next(pos), \
drivers/net/ethernet/netronome/nfp/bpf/jit.c
39
next2 = nfp_meta_next(next))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3968
struct bpf_insn insn, next;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3971
next = meta2->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3979
if (next.code != (BPF_ALU64 | BPF_AND | BPF_K))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3984
if (exp_mask[BPF_SIZE(insn.code)] != next.imm)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3987
if (next.src_reg || next.dst_reg)
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
508
goto next;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
539
next:
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
862
goto next;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
866
goto next;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
883
next:
drivers/net/ethernet/netronome/nfp/nfdk/rings.c
29
goto next;
drivers/net/ethernet/netronome/nfp/nfdk/rings.c
56
next:
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
595
struct nfp_net *nn, *next;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
633
list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
808
struct nfp_net *nn, *next;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
812
list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
91
struct nfp_net *nn, *next;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
93
list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list)
drivers/net/ethernet/ni/nixge.c
147
nixge_hw_dma_bd_set_addr((bd), next, (addr))
drivers/net/ethernet/pasemi/pasemi_mac.c
749
goto next;
drivers/net/ethernet/pasemi/pasemi_mac.c
772
next:
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
687
cur = list_entry(head->next, nx_mac_list_t, list);
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
721
cur = list_entry(head->next, nx_mac_list_t, list);
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1212
next:
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1221
goto next;
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1227
goto next;
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1819
buffer = list_entry(head->next, struct netxen_rx_buffer, list);
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1879
buffer = list_entry(head->next, struct netxen_rx_buffer, list);
drivers/net/ethernet/qlogic/qed/qed_chain.c
50
struct qed_chain_next *next;
drivers/net/ethernet/qlogic/qed/qed_chain.c
54
next = virt_curr + size;
drivers/net/ethernet/qlogic/qed/qed_chain.c
56
DMA_REGPAIR_LE(next->next_phys, phys_next);
drivers/net/ethernet/qlogic/qed/qed_chain.c
57
next->next_virt = virt_next;
drivers/net/ethernet/qlogic/qed/qed_chain.c
71
struct qed_chain_next *next;
drivers/net/ethernet/qlogic/qed/qed_chain.c
84
next = virt + size;
drivers/net/ethernet/qlogic/qed/qed_chain.c
85
virt_next = next->next_virt;
drivers/net/ethernet/qlogic/qed/qed_chain.c
86
phys_next = HILO_DMA_REGPAIR(next->next_phys);
drivers/net/ethernet/qlogic/qed/qed_cxt.c
77
__be64 next;
drivers/net/ethernet/qlogic/qed/qed_cxt.c
880
entries[j].next = cpu_to_be64(val);
drivers/net/ethernet/qlogic/qed/qed_cxt.c
887
entries[j].next = cpu_to_be64(val);
drivers/net/ethernet/qlogic/qede/qede_rdma.c
56
event_node = list_entry(head->next, struct qede_rdma_event_work,
drivers/net/ethernet/qlogic/qla3xxx.c
1835
lrg_buf_cb = lrg_buf_cb->next;
drivers/net/ethernet/qlogic/qla3xxx.c
298
lrg_buf_cb->next = NULL;
drivers/net/ethernet/qlogic/qla3xxx.c
303
qdev->lrg_buf_free_tail->next = lrg_buf_cb;
drivers/net/ethernet/qlogic/qla3xxx.c
354
qdev->lrg_buf_free_head = lrg_buf_cb->next;
drivers/net/ethernet/qlogic/qla3xxx.h
998
struct ql_rcv_buf_cb *next;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
3903
cmd = list_entry(head->next, struct qlcnic_cmd_args, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
4127
cmd = list_entry(head->next, struct qlcnic_cmd_args, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
1000
(next != QLC_83XX_IDC_DEV_READY)) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
1003
__func__, cur, prev, next);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
1008
if (next == QLC_83XX_IDC_DEV_INIT) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
1014
__func__, cur, prev, next);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
983
u32 cur, prev, next;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
987
next = state;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
989
if ((next < QLC_83XX_IDC_DEV_COLD) ||
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
990
(next > QLC_83XX_IDC_DEV_QUISCENT)) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
999
if ((next != QLC_83XX_IDC_DEV_COLD) &&
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
621
cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
1275
next:
drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
1285
goto next;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
1291
goto next;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
1451
buffer = list_entry(head->next, struct qlcnic_rx_buffer, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
859
buffer = list_entry(head->next, struct qlcnic_rx_buffer, list);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
1538
entry = list_entry(head->next, struct qlcnic_async_cmd,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
2109
cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list);
drivers/net/ethernet/qualcomm/rmnet/rmnet_map_data.c
637
port->skbagg_tail->next = skb;
drivers/net/ethernet/renesas/rswitch_main.c
1054
goto next;
drivers/net/ethernet/renesas/rswitch_main.c
1059
goto next;
drivers/net/ethernet/renesas/rswitch_main.c
1065
goto next;
drivers/net/ethernet/renesas/rswitch_main.c
1074
next:
drivers/net/ethernet/sfc/ef100_rep.c
337
struct efx_rep *efv, *next;
drivers/net/ethernet/sfc/ef100_rep.c
342
list_for_each_entry_safe(efv, next, &efx->vf_reps, list)
drivers/net/ethernet/sfc/ef100_sriov.c
20
struct efx_rep *efv, *next;
drivers/net/ethernet/sfc/ef100_sriov.c
39
list_for_each_entry_safe(efv, next, &efx->vf_reps, list)
drivers/net/ethernet/sfc/efx.c
199
struct efx_nic *other, *next;
drivers/net/ethernet/sfc/efx.c
207
list_for_each_entry_safe(other, next, &efx_unassociated_list,
drivers/net/ethernet/sfc/efx.c
244
struct efx_nic *other, *next;
drivers/net/ethernet/sfc/efx.c
249
list_for_each_entry_safe(other, next, &efx->secondary_list, node) {
drivers/net/ethernet/sfc/falcon/efx.c
1152
struct ef4_nic *other, *next;
drivers/net/ethernet/sfc/falcon/efx.c
1160
list_for_each_entry_safe(other, next, &ef4_unassociated_list,
drivers/net/ethernet/sfc/falcon/efx.c
1197
struct ef4_nic *other, *next;
drivers/net/ethernet/sfc/falcon/efx.c
1202
list_for_each_entry_safe(other, next, &efx->secondary_list, node) {
drivers/net/ethernet/sfc/falcon/mtd.c
107
list_for_each_entry_safe(part, next, &efx->mtd_list, node)
drivers/net/ethernet/sfc/falcon/mtd.c
97
struct ef4_mtd_partition *parts, *part, *next;
drivers/net/ethernet/sfc/mcdi.c
1009
if (mcdi->async_list.next == &async->list &&
drivers/net/ethernet/sfc/mcdi.c
1114
struct efx_mcdi_async_param *async, *next;
drivers/net/ethernet/sfc/mcdi.c
1141
list_for_each_entry_safe(async, next, &mcdi->async_list, list) {
drivers/net/ethernet/sfc/mtd.c
100
struct efx_mtd_partition *parts, *part, *next;
drivers/net/ethernet/sfc/mtd.c
110
list_for_each_entry_safe(part, next, &efx->mtd_list, node)
drivers/net/ethernet/sfc/siena/efx.c
208
struct efx_nic *other, *next;
drivers/net/ethernet/sfc/siena/efx.c
216
list_for_each_entry_safe(other, next, &efx_unassociated_list,
drivers/net/ethernet/sfc/siena/efx.c
253
struct efx_nic *other, *next;
drivers/net/ethernet/sfc/siena/efx.c
258
list_for_each_entry_safe(other, next, &efx->secondary_list, node) {
drivers/net/ethernet/sfc/siena/mcdi.c
1013
if (mcdi->async_list.next == &async->list &&
drivers/net/ethernet/sfc/siena/mcdi.c
1137
struct efx_mcdi_async_param *async, *next;
drivers/net/ethernet/sfc/siena/mcdi.c
1164
list_for_each_entry_safe(async, next, &mcdi->async_list, list) {
drivers/net/ethernet/sfc/siena/mtd.c
100
struct efx_mtd_partition *parts, *part, *next;
drivers/net/ethernet/sfc/siena/mtd.c
110
list_for_each_entry_safe(part, next, &efx->mtd_list, node)
drivers/net/ethernet/sfc/siena/ptp.c
1156
struct list_head *next;
drivers/net/ethernet/sfc/siena/ptp.c
1163
list_for_each_safe(cursor, next, &ptp->evt_list) {
drivers/net/ethernet/sfc/siena/ptp.c
1183
struct list_head *next;
drivers/net/ethernet/sfc/siena/ptp.c
1199
list_for_each_safe(cursor, next, &ptp->evt_list) {
drivers/net/ethernet/sfc/siena/ptp.c
1351
struct list_head *next;
drivers/net/ethernet/sfc/siena/ptp.c
1367
list_for_each_safe(cursor, next, &efx->ptp_data->evt_list) {
drivers/net/ethernet/sfc/siena/tx_common.c
433
struct sk_buff *segments, *next;
drivers/net/ethernet/sfc/siena/tx_common.c
441
skb_list_walk_safe(segments, skb, next) {
drivers/net/ethernet/sfc/tc.c
210
struct efx_tc_action_set *act, *next;
drivers/net/ethernet/sfc/tc.c
219
list_for_each_entry_safe(act, next, &acts->list, list)
drivers/net/ethernet/sfc/tc_conntrack.c
606
struct efx_tc_ct_entry *conn, *next;
drivers/net/ethernet/sfc/tc_conntrack.c
618
list_for_each_entry_safe(conn, next, &ct_zone->cts, list)
drivers/net/ethernet/sfc/tc_encap_actions.c
707
struct efx_tc_encap_action *encap, *next;
drivers/net/ethernet/sfc/tc_encap_actions.c
709
list_for_each_entry_safe(encap, next, &neigh->users, list) {
drivers/net/ethernet/sfc/tx_common.c
473
struct sk_buff *segments, *next;
drivers/net/ethernet/sfc/tx_common.c
481
skb_list_walk_safe(segments, skb, next) {
drivers/net/ethernet/sgi/ioc3-eth.c
397
goto next;
drivers/net/ethernet/sgi/ioc3-eth.c
428
next:
drivers/net/ethernet/silan/sc92031.c
793
goto next;
drivers/net/ethernet/silan/sc92031.c
814
next:
drivers/net/ethernet/sis/sis190.c
1255
struct sis190_phy *cur, *next;
drivers/net/ethernet/sis/sis190.c
1257
list_for_each_entry_safe(cur, next, first_phy, list) {
drivers/net/ethernet/sis/sis900.c
144
struct mii_phy * next;
drivers/net/ethernet/sis/sis900.c
2486
sis_priv->first_mii = phy->next;
drivers/net/ethernet/sis/sis900.c
627
mii_phy = mii_phy->next;
drivers/net/ethernet/sis/sis900.c
637
mii_phy->next = sis_priv->mii;
drivers/net/ethernet/sis/sis900.c
728
for (phy=sis_priv->first_mii; phy; phy=phy->next) {
drivers/net/ethernet/smsc/epic100.c
229
u32 next;
drivers/net/ethernet/smsc/epic100.c
236
u32 next;
drivers/net/ethernet/smsc/epic100.c
900
ep->rx_ring[i].next = ep->rx_ring_dma +
drivers/net/ethernet/smsc/epic100.c
905
ep->rx_ring[i-1].next = ep->rx_ring_dma;
drivers/net/ethernet/smsc/epic100.c
927
ep->tx_ring[i].next = ep->tx_ring_dma +
drivers/net/ethernet/smsc/epic100.c
930
ep->tx_ring[i-1].next = ep->tx_ring_dma;
drivers/net/ethernet/socionext/netsec.c
1030
goto next;
drivers/net/ethernet/socionext/netsec.c
1060
next:
drivers/net/ethernet/socionext/netsec.c
669
goto next;
drivers/net/ethernet/socionext/netsec.c
681
next:
drivers/net/ethernet/sun/cassini.c
2288
goto next;
drivers/net/ethernet/sun/cassini.c
2312
next:
drivers/net/ethernet/sun/cassini.c
3228
goto next;
drivers/net/ethernet/sun/cassini.c
3237
goto next;
drivers/net/ethernet/sun/cassini.c
3248
goto next;
drivers/net/ethernet/sun/cassini.c
3255
goto next;
drivers/net/ethernet/sun/cassini.c
3260
goto next;
drivers/net/ethernet/sun/cassini.c
3279
goto next;
drivers/net/ethernet/sun/cassini.c
3284
next:
drivers/net/ethernet/sun/cassini.c
628
entry = cp->rx_spare_list.next;
drivers/net/ethernet/sun/niu.c
10239
offsetof(union niu_page, next));
drivers/net/ethernet/sun/niu.c
3540
struct page *next = niu_next_page(page);
drivers/net/ethernet/sun/niu.c
3550
page = next;
drivers/net/ethernet/sun/niu.c
52
struct page *next; /* alias of "mapping" */
drivers/net/ethernet/sun/niu.c
55
#define niu_next_page(p) container_of(p, union niu_page, page)->next
drivers/net/ethernet/sun/niu.c
6461
struct page *next = niu_next_page(page);
drivers/net/ethernet/sun/niu.c
6465
page = next;
drivers/net/ethernet/sun/sunbmac.c
821
goto next;
drivers/net/ethernet/sun/sunbmac.c
879
next:
drivers/net/ethernet/sun/sungem.c
801
goto next;
drivers/net/ethernet/sun/sungem.c
860
next:
drivers/net/ethernet/sun/sunhme.c
1791
goto next;
drivers/net/ethernet/sun/sunhme.c
1854
next:
drivers/net/ethernet/sun/sunhme.c
2256
qp->next = qfe_sbus_list;
drivers/net/ethernet/sun/sunhme.c
2274
for (qp = qfe_pci_list; qp != NULL; qp = qp->next) {
drivers/net/ethernet/sun/sunhme.c
2289
qp->next = qfe_pci_list;
drivers/net/ethernet/sun/sunhme.c
2794
struct quattro *next = qfe->next;
drivers/net/ethernet/sun/sunhme.c
2798
qfe_pci_list = next;
drivers/net/ethernet/sun/sunhme.c
2860
struct quattro *next = qfe->next;
drivers/net/ethernet/sun/sunhme.c
2864
qfe_sbus_list = next;
drivers/net/ethernet/sun/sunhme.h
487
struct quattro *next;
drivers/net/ethernet/sun/sunqe.c
484
goto next;
drivers/net/ethernet/sun/sunqe.c
501
next:
drivers/net/ethernet/sun/sunqe.c
981
struct sunqec *next = root_qec_dev->next_module;
drivers/net/ethernet/sun/sunqe.c
989
root_qec_dev = next;
drivers/net/ethernet/sun/sunvnet_common.c
1016
BUG_ON(port->tx_bufs[txi].skb->next);
drivers/net/ethernet/sun/sunvnet_common.c
1018
port->tx_bufs[txi].skb->next = skb;
drivers/net/ethernet/sun/sunvnet_common.c
1035
struct sk_buff *next;
drivers/net/ethernet/sun/sunvnet_common.c
1038
next = skb->next;
drivers/net/ethernet/sun/sunvnet_common.c
1039
skb->next = NULL;
drivers/net/ethernet/sun/sunvnet_common.c
1041
skb = next;
drivers/net/ethernet/sun/sunvnet_common.c
1227
struct sk_buff *segs, *curr, *next;
drivers/net/ethernet/sun/sunvnet_common.c
1286
skb_list_walk_safe(segs, curr, next) {
drivers/net/ethernet/sun/sunvnet_common.c
1553
for (m = vp->mcast_list; m; m = m->next) {
drivers/net/ethernet/sun/sunvnet_common.c
1580
m->next = vp->mcast_list;
drivers/net/ethernet/sun/sunvnet_common.c
1601
for (m = vp->mcast_list; m; m = m->next) {
drivers/net/ethernet/sun/sunvnet_common.c
1627
pp = &m->next;
drivers/net/ethernet/sun/sunvnet_common.c
1640
*pp = m->next;
drivers/net/ethernet/sun/sunvnet_common.c
1775
port = list_entry(vp->port_list.next, struct vnet_port, list);
drivers/net/ethernet/sun/sunvnet_common.h
112
struct vnet_mcast_entry *next;
drivers/net/ethernet/ti/am65-cpts.c
235
struct list_head *this, *next;
drivers/net/ethernet/ti/am65-cpts.c
239
list_for_each_safe(this, next, events) {
drivers/net/ethernet/ti/am65-cpts.c
856
struct list_head *this, *next;
drivers/net/ethernet/ti/am65-cpts.c
865
list_for_each_safe(this, next, &events) {
drivers/net/ethernet/ti/am65-cpts.c
922
struct list_head *this, *next;
drivers/net/ethernet/ti/am65-cpts.c
930
list_for_each_safe(this, next, &cpts->events_rx) {
drivers/net/ethernet/ti/cpts.c
367
struct list_head *this, *next;
drivers/net/ethernet/ti/cpts.c
377
list_for_each_safe(this, next, &events) {
drivers/net/ethernet/ti/cpts.c
460
struct list_head *this, *next;
drivers/net/ethernet/ti/cpts.c
468
list_for_each_safe(this, next, &cpts->events) {
drivers/net/ethernet/ti/cpts.c
65
struct list_head *this, *next;
drivers/net/ethernet/ti/cpts.c
69
list_for_each_safe(this, next, &cpts->events) {
drivers/net/ethernet/ti/netcp_core.c
474
struct netcp_hook_list *next;
drivers/net/ethernet/ti/netcp_core.c
486
list_for_each_entry(next, &netcp_priv->txhook_list_head, list) {
drivers/net/ethernet/ti/netcp_core.c
487
if (next->order > order)
drivers/net/ethernet/ti/netcp_core.c
490
__list_add(&entry->list, next->list.prev, &next->list);
drivers/net/ethernet/ti/netcp_core.c
500
struct netcp_hook_list *next, *n;
drivers/net/ethernet/ti/netcp_core.c
504
list_for_each_entry_safe(next, n, &netcp_priv->txhook_list_head, list) {
drivers/net/ethernet/ti/netcp_core.c
505
if ((next->order == order) &&
drivers/net/ethernet/ti/netcp_core.c
506
(next->hook_rtn == hook_rtn) &&
drivers/net/ethernet/ti/netcp_core.c
507
(next->hook_data == hook_data)) {
drivers/net/ethernet/ti/netcp_core.c
508
list_del(&next->list);
drivers/net/ethernet/ti/netcp_core.c
510
devm_kfree(netcp_priv->dev, next);
drivers/net/ethernet/ti/netcp_core.c
523
struct netcp_hook_list *next;
drivers/net/ethernet/ti/netcp_core.c
535
list_for_each_entry(next, &netcp_priv->rxhook_list_head, list) {
drivers/net/ethernet/ti/netcp_core.c
536
if (next->order > order)
drivers/net/ethernet/ti/netcp_core.c
539
__list_add(&entry->list, next->list.prev, &next->list);
drivers/net/ethernet/ti/netcp_core.c
549
struct netcp_hook_list *next, *n;
drivers/net/ethernet/ti/netcp_core.c
553
list_for_each_entry_safe(next, n, &netcp_priv->rxhook_list_head, list) {
drivers/net/ethernet/ti/netcp_core.c
554
if ((next->order == order) &&
drivers/net/ethernet/ti/netcp_core.c
555
(next->hook_rtn == hook_rtn) &&
drivers/net/ethernet/ti/netcp_core.c
556
(next->hook_data == hook_data)) {
drivers/net/ethernet/ti/netcp_core.c
557
list_del(&next->list);
drivers/net/ethernet/ti/netcp_core.c
559
devm_kfree(netcp_priv->dev, next);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1082
chain->head = descr->next;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
226
for (descr = start_descr; start_descr != descr->next; descr++) {
drivers/net/ethernet/toshiba/ps3_gelic_net.c
229
= cpu_to_be32(descr->next->link.cpu_addr);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
293
descr = descr->next) {
drivers/net/ethernet/toshiba/ps3_gelic_net.c
340
descr->next = descr + 1;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
344
(descr - 1)->next = start_descr;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
351
cpu_to_be32(descr->next->link.cpu_addr);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
453
descr = descr->next;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
476
descr = descr->next;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
563
tx_chain->tail = tx_chain->tail->next) {
drivers/net/ethernet/toshiba/ps3_gelic_net.c
697
if (card->tx_chain.tail != card->tx_chain.head->next &&
drivers/net/ethernet/toshiba/ps3_gelic_net.c
815
card->tx_chain.head = descr->next;
drivers/net/ethernet/toshiba/ps3_gelic_net.h
253
struct gelic_descr *next;
drivers/net/ethernet/toshiba/ps3_gelic_wireless.c
1593
target = list_entry(wl->network_free_list.next,
drivers/net/ethernet/via/via-rhine.c
1246
dma_addr_t next;
drivers/net/ethernet/via/via-rhine.c
1250
next = rp->rx_ring_dma;
drivers/net/ethernet/via/via-rhine.c
1256
next += sizeof(struct rx_desc);
drivers/net/ethernet/via/via-rhine.c
1257
rp->rx_ring[i].next_desc = cpu_to_le32(next);
drivers/net/ethernet/via/via-rhine.c
1304
dma_addr_t next;
drivers/net/ethernet/via/via-rhine.c
1308
next = rp->tx_ring_dma;
drivers/net/ethernet/via/via-rhine.c
1313
next += sizeof(struct tx_desc);
drivers/net/ethernet/via/via-rhine.c
1314
rp->tx_ring[i].next_desc = cpu_to_le32(next);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2069
ring->next = head->ring;
drivers/net/ethernet/wangxun/libwx/wx_type.h
1052
for (posm = (headm).ring; posm; posm = posm->next)
drivers/net/ethernet/wangxun/libwx/wx_type.h
1063
struct wx_ring *next; /* pointer to next ring in q_vector */
drivers/net/ethernet/xilinx/ll_temac.h
331
u32 next; /* Physical address of next buffer descriptor */
drivers/net/ethernet/xilinx/ll_temac_main.c
352
lp->tx_bd_v[i].next = cpu_to_be32(lp->tx_bd_p
drivers/net/ethernet/xilinx/ll_temac_main.c
357
lp->rx_bd_v[i].next = cpu_to_be32(lp->rx_bd_p
drivers/net/ethernet/xilinx/xilinx_axienet.h
436
u32 next; /* Physical address of next buffer descriptor */
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
366
lp->tx_bd_v[i].next = lower_32_bits(addr);
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
376
lp->rx_bd_v[i].next = lower_32_bits(addr);
drivers/net/ethernet/xscale/ixp4xx_eth.c
223
u32 next; /* pointer to next buffer, unused */
drivers/net/ethernet/xscale/ixp4xx_eth.c
640
phys, desc->next, desc->buf_len, desc->pkt_len,
drivers/net/ethernet/xscale/ixp4xx_eth.c
665
BUG_ON(tab[n_desc].next);
drivers/net/fddi/skfp/smttimer.c
116
struct smt_timer *next ;
drivers/net/fddi/skfp/smttimer.c
137
next = smc->t.st_queue ;
drivers/net/fddi/skfp/smttimer.c
140
for ( tm = next ; tm ; tm = next) {
drivers/net/fddi/skfp/smttimer.c
141
next = tm->tm_next ;
drivers/net/geneve.c
101
struct list_head next; /* geneve's per namespace list */
drivers/net/geneve.c
1795
list_for_each_entry(geneve, &gn->geneve_list, next) {
drivers/net/geneve.c
1888
list_add(&geneve->next, &gn->geneve_list);
drivers/net/geneve.c
2251
list_del(&geneve->next);
drivers/net/geneve.c
2446
struct geneve_dev *geneve, *next;
drivers/net/geneve.c
2448
list_for_each_entry_safe(geneve, next, &gn->geneve_list, next)
drivers/net/gtp.c
1553
struct hlist_node *next;
drivers/net/gtp.c
1558
hlist_for_each_entry_safe(pctx, next, &gtp->tid_hash[i], hlist_tid)
drivers/net/hamradio/bpqether.c
420
.next = bpq_seq_next,
drivers/net/hamradio/bpqether.c
580
bpq = list_entry(bpq_devices.next, struct bpqdev, bpq_list);
drivers/net/hamradio/scc.c
2087
.next = scc_net_seq_next,
drivers/net/hamradio/yam.c
1172
yam_data = yam_data->next;
drivers/net/hamradio/yam.c
138
struct yam_mcs *next;
drivers/net/hamradio/yam.c
385
p = p->next;
drivers/net/hamradio/yam.c
395
p->next = yam_data;
drivers/net/hamradio/yam.c
410
p = p->next;
drivers/net/hamradio/yam.c
825
.next = yam_seq_next,
drivers/net/hyperv/hyperv_net.h
941
u32 next; /* next entry for writing */
drivers/net/hyperv/netvsc.c
1365
while (mrc->first != mrc->next) {
drivers/net/hyperv/netvsc.c
1397
if (mrc->next >= mrc->first)
drivers/net/hyperv/netvsc.c
1398
*filled = mrc->next - mrc->first;
drivers/net/hyperv/netvsc.c
1400
*filled = (count - mrc->first) + mrc->next;
drivers/net/hyperv/netvsc.c
1428
rcd = mrc->slots + mrc->next;
drivers/net/hyperv/netvsc.c
1432
if (++mrc->next == nvdev->recv_completion_cnt)
drivers/net/hyperv/netvsc.c
1433
mrc->next = 0;
drivers/net/hyperv/rndis_filter.c
1142
if (nvchan->mrc.first != nvchan->mrc.next)
drivers/net/ipvlan/ipvlan_main.c
659
struct ipvl_addr *addr, *next;
drivers/net/ipvlan/ipvlan_main.c
662
list_for_each_entry_safe(addr, next, &ipvlan->addrs, anode) {
drivers/net/ipvlan/ipvlan_main.c
731
struct ipvl_dev *ipvlan, *next;
drivers/net/ipvlan/ipvlan_main.c
767
list_for_each_entry_safe(ipvlan, next, &port->ipvlans, pnode)
drivers/net/macsec.c
1442
rx_scp = &rx_sc->next, rx_sc = rtnl_dereference(*rx_scp)) {
drivers/net/macsec.c
1446
rcu_assign_pointer(*rx_scp, rx_sc->next);
drivers/net/macsec.c
1483
rcu_assign_pointer(rx_sc->next, secy->rx_sc);
drivers/net/macsec.c
255
if (rx_sc && !rcu_dereference_bh(rx_sc->next))
drivers/net/macsec.c
3293
goto next;
drivers/net/macsec.c
3296
goto next;
drivers/net/macsec.c
3301
next:
drivers/net/macsec.c
3954
rcu_assign_pointer(macsec->secy.rx_sc, rx_sc->next);
drivers/net/macsec.c
57
sc = rcu_dereference_bh(sc->next))
drivers/net/macsec.c
61
sc = rtnl_dereference(sc->next))
drivers/net/macvlan.c
1797
struct macvlan_dev *vlan, *next;
drivers/net/macvlan.c
1844
list_for_each_entry_safe(vlan, next, &port->vlans, list)
drivers/net/macvlan.c
390
struct hlist_node *next;
drivers/net/macvlan.c
393
hash_for_each_safe(port->vlan_source_hash, i, next, entry, hlist)
drivers/net/mctp/mctp-i2c.c
704
if (!midev->unlock_marker.next)
drivers/net/mctp/mctp-serial.c
541
int next;
drivers/net/mctp/mctp-serial.c
550
next = next_chunk_len(dev);
drivers/net/mctp/mctp-serial.c
551
dev->txpos += next;
drivers/net/mctp/mctp-serial.c
552
KUNIT_EXPECT_EQ(test, next, params->chunks[i]);
drivers/net/mctp/mctp-serial.c
554
if (next == 0) {
drivers/net/mdio/mdio-mux.c
135
cb = cb->next;
drivers/net/mdio/mdio-mux.c
237
cb->next = pb->children;
drivers/net/mdio/mdio-mux.c
31
struct mdio_mux_child_bus *next;
drivers/net/mhi_net.c
158
tail->next = skb;
drivers/net/ovpn/io.c
336
struct sk_buff *curr, *next;
drivers/net/ovpn/io.c
341
skb_list_walk_safe(skb, curr, next) {
drivers/net/ovpn/io.c
356
struct sk_buff *segments, *curr, *next;
drivers/net/ovpn/io.c
407
skb_list_walk_safe(skb, curr, next) {
drivers/net/ovpn/io.c
430
skb_list.prev->next = NULL;
drivers/net/ovpn/io.c
433
ovpn_send(ovpn, skb_list.next, peer);
drivers/net/phy/bcm-phy-ptp.c
542
u64 ns, next;
drivers/net/phy/bcm-phy-ptp.c
556
next = NSEC_PER_SEC - ts.tv_nsec;
drivers/net/phy/bcm-phy-ptp.c
557
ts.tv_sec += next < NSEC_PER_MSEC ? 2 : 1;
drivers/net/phy/bcm-phy-ptp.c
580
next = next + NSEC_PER_MSEC;
drivers/net/phy/bcm-phy-ptp.c
581
schedule_delayed_work(&priv->pin_work, nsecs_to_jiffies(next));
drivers/net/phy/dp83640.c
1305
struct list_head *this, *next;
drivers/net/phy/dp83640.c
1324
list_for_each_safe(this, next, &dp83640->rxts) {
drivers/net/phy/dp83640.c
1468
struct list_head *this, *next;
drivers/net/phy/dp83640.c
1489
list_for_each_safe(this, next, &clock->phylist) {
drivers/net/phy/dp83640.c
591
struct list_head *this, *next;
drivers/net/phy/dp83640.c
594
list_for_each_safe(this, next, &dp83640->rxts) {
drivers/net/phy/mii_timestamper.c
55
struct list_head *this, *next;
drivers/net/phy/mii_timestamper.c
58
list_for_each_safe(this, next, &mii_timestamping_devices) {
drivers/net/ppp/ppp_generic.c
1909
list = list->next;
drivers/net/ppp/ppp_generic.c
2027
list = list->next;
drivers/net/ppp/ppp_generic.c
2037
list = list->next;
drivers/net/ppp/ppp_generic.c
2892
p->next = NULL;
drivers/net/ppp/ppp_generic.c
2893
fragpp = &p->next;
drivers/net/ppp/pppoe.c
1062
po = rcu_dereference(po->next);
drivers/net/ppp/pppoe.c
1083
struct pppox_sock *po, *next;
drivers/net/ppp/pppoe.c
1091
next = rcu_dereference(po->next);
drivers/net/ppp/pppoe.c
1092
if (next)
drivers/net/ppp/pppoe.c
1093
po = next;
drivers/net/ppp/pppoe.c
1117
.next = pppoe_seq_next,
drivers/net/ppp/pppoe.c
171
ret = rcu_dereference(ret->next);
drivers/net/ppp/pppoe.c
189
ret = rcu_dereference_protected(ret->next, lockdep_is_held(&pn->hash_lock));
drivers/net/ppp/pppoe.c
192
RCU_INIT_POINTER(po->next, first);
drivers/net/ppp/pppoe.c
210
struct pppox_sock *next;
drivers/net/ppp/pppoe.c
212
next = rcu_dereference_protected(ret->next,
drivers/net/ppp/pppoe.c
214
rcu_assign_pointer(*src, next);
drivers/net/ppp/pppoe.c
218
src = &ret->next;
drivers/net/ppp/pppoe.c
219
ret = rcu_dereference_protected(ret->next, lockdep_is_held(&pn->hash_lock));
drivers/net/ppp/pppoe.c
288
po = rcu_dereference_protected(po->next,
drivers/net/ppp/pppoe.c
661
po->next = NULL;
drivers/net/slip/slhc.c
136
ts[i].next = &(ts[i - 1]);
drivers/net/slip/slhc.c
138
ts[0].next = &(ts[comp->tslot_limit]);
drivers/net/slip/slhc.c
232
struct cstate *cs = lcs->next;
drivers/net/slip/slhc.c
307
cs = cs->next;
drivers/net/slip/slhc.c
334
lcs->next = cs->next;
drivers/net/slip/slhc.c
335
cs->next = ocs->next;
drivers/net/slip/slhc.c
336
ocs->next = cs;
drivers/net/tap.c
116
list_add_tail(&q->next, &tap->queue_list);
drivers/net/tap.c
1204
list_for_each_entry(q, &tap->queue_list, next)
drivers/net/tap.c
1231
list_add_tail_rcu(&tap_major->next, &major_list);
drivers/net/tap.c
1271
list_for_each_entry_safe(tap_major, tmp, &major_list, next) {
drivers/net/tap.c
1274
list_del_rcu(&tap_major->next);
drivers/net/tap.c
173
list_del_init(&q->next);
drivers/net/tap.c
240
list_for_each_entry_safe(q, tmp, &tap->queue_list, next) {
drivers/net/tap.c
241
list_del_init(&q->next);
drivers/net/tap.c
282
struct sk_buff *next;
drivers/net/tap.c
298
skb_list_walk_safe(segs, skb, next) {
drivers/net/tap.c
303
kfree_skb_list_reason(next, drop_reason);
drivers/net/tap.c
342
list_for_each_entry_rcu(tap_major, &major_list, next) {
drivers/net/tap.c
49
struct list_head next;
drivers/net/tun.c
1146
list_for_each_entry(tfile, &tun->disabled, next) {
drivers/net/tun.c
145
struct list_head next;
drivers/net/tun.c
3423
INIT_LIST_HEAD(&tfile->next);
drivers/net/tun.c
3616
list_for_each_entry(tfile, &tun->disabled, next)
drivers/net/tun.c
532
list_add_tail(&tfile->next, &tun->disabled);
drivers/net/tun.c
541
list_del_init(&tfile->next);
drivers/net/tun.c
661
list_for_each_entry(tfile, &tun->disabled, next) {
drivers/net/tun.c
677
list_for_each_entry_safe(tfile, tmp, &tun->disabled, next) {
drivers/net/usb/cdc_eem.c
219
goto next;
drivers/net/usb/cdc_eem.c
277
goto next;
drivers/net/usb/cdc_eem.c
323
next:
drivers/net/usb/r8152.c
2198
cursor = tp->tx_free.next;
drivers/net/usb/r8152.c
2216
struct sk_buff *segs, *seg, *next;
drivers/net/usb/r8152.c
2226
skb_list_walk_safe(segs, seg, next) {
drivers/net/usb/r8152.c
2525
struct list_head *cursor, *next, rx_queue;
drivers/net/usb/r8152.c
2556
list_for_each_safe(cursor, next, &rx_queue) {
drivers/net/usb/r8152.c
2698
list_add_tail(&agg->list, next);
drivers/net/virtio_net.c
2433
curr_skb->next = nskb;
drivers/net/vxlan/vxlan_core.c
3382
INIT_LIST_HEAD(&vxlan->next);
drivers/net/vxlan/vxlan_core.c
3709
list_for_each_entry(tmp, &vn->vxlan_list, next) {
drivers/net/vxlan/vxlan_core.c
4016
list_add(&vxlan->next, &vn->vxlan_list);
drivers/net/vxlan/vxlan_core.c
4515
list_del(&vxlan->next);
drivers/net/vxlan/vxlan_core.c
4728
struct vxlan_dev *vxlan, *next;
drivers/net/vxlan/vxlan_core.c
4731
list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) {
drivers/net/vxlan/vxlan_core.c
4939
struct vxlan_dev *vxlan, *next;
drivers/net/vxlan/vxlan_core.c
4941
list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next)
drivers/net/vxlan/vxlan_multicast.c
151
list_for_each_entry(vxlan, &vn->vxlan_list, next) {
drivers/net/vxlan/vxlan_private.h
69
return list_entry_rcu(fdb->remotes.next, struct vxlan_rdst, list);
drivers/net/wan/hd64572.h
170
unsigned long next; /* pointer to next block descriptor */
drivers/net/wan/hdlc.c
216
proto = proto->next;
drivers/net/wan/hdlc.c
326
proto->next = first_proto;
drivers/net/wan/hdlc.c
340
p = &((*p)->next);
drivers/net/wan/hdlc.c
342
*p = proto->next;
drivers/net/wan/hdlc_fr.c
1042
pvc = pvc->next;
drivers/net/wan/hdlc_fr.c
116
struct pvc_device *next; /* Sorted in ascending DLCI order */
drivers/net/wan/hdlc_fr.c
1161
struct pvc_device *next = pvc->next;
drivers/net/wan/hdlc_fr.c
1170
pvc = next;
drivers/net/wan/hdlc_fr.c
176
pvc = pvc->next;
drivers/net/wan/hdlc_fr.c
192
pvc_p = &(*pvc_p)->next;
drivers/net/wan/hdlc_fr.c
204
pvc->next = *pvc_p; /* Put it in the chain */
drivers/net/wan/hdlc_fr.c
243
*pvc_p = pvc->next;
drivers/net/wan/hdlc_fr.c
247
pvc_p = &(*pvc_p)->next;
drivers/net/wan/hdlc_fr.c
537
pvc = pvc->next;
drivers/net/wan/hdlc_fr.c
566
pvc = pvc->next;
drivers/net/wan/hdlc_fr.c
577
pvc = pvc->next;
drivers/net/wan/hdlc_fr.c
740
pvc = pvc->next;
drivers/net/wan/hdlc_fr.c
769
pvc = pvc->next;
drivers/net/wan/hdlc_fr.c
840
pvc = pvc->next;
drivers/net/wan/ixp4xx_hss.c
305
u32 next; /* pointer to next buffer, unused */
drivers/net/wan/ixp4xx_hss.c
583
phys, desc->next, desc->buf_len, desc->pkt_len,
drivers/net/wan/ixp4xx_hss.c
604
BUG_ON(tab[n_desc].next);
drivers/net/wireguard/device.c
142
struct sk_buff *next;
drivers/net/wireguard/device.c
189
skb_list_walk_safe(skb, skb, next) {
drivers/net/wireguard/device.h
36
struct { struct sk_buff *next, *prev; } empty; // Match first 2 members of struct sk_buff.
drivers/net/wireguard/queueing.c
100
if (next) {
drivers/net/wireguard/queueing.c
101
queue->tail = next;
drivers/net/wireguard/queueing.c
60
offsetof(struct sk_buff, next) != offsetof(struct prev_queue, empty.next) -
drivers/net/wireguard/queueing.c
82
struct sk_buff *tail = queue->tail, *next = smp_load_acquire(&NEXT(tail));
drivers/net/wireguard/queueing.c
85
if (!next)
drivers/net/wireguard/queueing.c
87
queue->tail = next;
drivers/net/wireguard/queueing.c
88
tail = next;
drivers/net/wireguard/queueing.c
89
next = smp_load_acquire(&NEXT(next));
drivers/net/wireguard/queueing.c
91
if (next) {
drivers/net/wireguard/queueing.c
92
queue->tail = next;
drivers/net/wireguard/queueing.c
99
next = smp_load_acquire(&NEXT(tail));
drivers/net/wireguard/receive.c
459
goto next;
drivers/net/wireguard/receive.c
467
goto next;
drivers/net/wireguard/receive.c
471
goto next;
drivers/net/wireguard/receive.c
477
next:
drivers/net/wireguard/send.c
244
struct sk_buff *skb, *next;
drivers/net/wireguard/send.c
249
skb_list_walk_safe(first, skb, next) {
drivers/net/wireguard/send.c
291
struct sk_buff *first, *skb, *next;
drivers/net/wireguard/send.c
296
skb_list_walk_safe(first, skb, next) {
drivers/net/wireguard/send.c
385
packets.prev->next = NULL;
drivers/net/wireguard/send.c
387
PACKET_CB(packets.next)->keypair = keypair;
drivers/net/wireguard/send.c
388
wg_packet_create_data(peer, packets.next);
drivers/net/wireless/ath/ar5523/ar5523.c
638
data = (struct ar5523_rx_data *) ar->rx_data_free.next;
drivers/net/wireless/ath/ar5523/ar5523.c
685
data = (struct ar5523_rx_data *) ar->rx_data_used.next;
drivers/net/wireless/ath/ar5523/ar5523.c
705
data = (struct ar5523_rx_data *) ar->rx_data_free.next;
drivers/net/wireless/ath/ar5523/ar5523.c
818
ar->tx_queue_pending.next;
drivers/net/wireless/ath/ath10k/htt_rx.c
555
prev_frag_buf->next = NULL;
drivers/net/wireless/ath/ath10k/htt_rx.c
563
prev_frag_buf->next = frag_buf;
drivers/net/wireless/ath/ath10k/htt_rx.c
574
prev_frag_buf->next = NULL;
drivers/net/wireless/ath/ath10k/htt_rx.c
632
prev_frag_buf->next = NULL;
drivers/net/wireless/ath/ath10k/htt_rx.c
640
prev_frag_buf->next = frag_buf;
drivers/net/wireless/ath/ath10k/htt_rx.c
651
prev_frag_buf->next = NULL;
drivers/net/wireless/ath/ath10k/mac.c
1622
const u8 *next;
drivers/net/wireless/ath/ath10k/mac.c
1637
next = ie + len;
drivers/net/wireless/ath/ath10k/mac.c
1639
if (WARN_ON(next > end))
drivers/net/wireless/ath/ath10k/mac.c
1642
memmove(ie, next, end - next);
drivers/net/wireless/ath/ath10k/mac.c
6958
goto next;
drivers/net/wireless/ath/ath10k/mac.c
6993
next:
drivers/net/wireless/ath/ath11k/core.c
1656
goto next;
drivers/net/wireless/ath/ath11k/core.c
1660
goto next;
drivers/net/wireless/ath/ath11k/core.c
1670
goto next;
drivers/net/wireless/ath/ath11k/core.c
1687
next:
drivers/net/wireless/ath/ath11k/core.c
1781
goto next;
drivers/net/wireless/ath/ath11k/core.c
1788
next:
drivers/net/wireless/ath/ath11k/dp_rx.c
4834
last->next = msdu;
drivers/net/wireless/ath/ath11k/dp_rx.c
4867
last->next = NULL;
drivers/net/wireless/ath/ath11k/dp_rx.c
4923
msdu = head_msdu->next;
drivers/net/wireless/ath/ath11k/dp_rx.c
4929
msdu = msdu->next;
drivers/net/wireless/ath/ath11k/dp_rx.c
4932
prev_buf->next = NULL;
drivers/net/wireless/ath/ath11k/dp_rx.c
4958
msdu = msdu->next;
drivers/net/wireless/ath/ath11k/dp_rx.c
5094
skb_next = mon_skb->next;
drivers/net/wireless/ath/ath11k/dp_rx.c
5119
skb_next = mon_skb->next;
drivers/net/wireless/ath/ath11k/dp_rx.c
5451
last->next = msdu;
drivers/net/wireless/ath/ath11k/dp_rx.c
5475
last->next = NULL;
drivers/net/wireless/ath/ath11k/dp_rx.c
5512
skb_next = tmp_msdu->next;
drivers/net/wireless/ath/ath11k/mac.c
1456
const u8 *next, *end;
drivers/net/wireless/ath/ath11k/mac.c
1470
next = ie + len;
drivers/net/wireless/ath/ath11k/mac.c
1472
if (WARN_ON(next > end))
drivers/net/wireless/ath/ath11k/mac.c
1475
memmove(ie, next, end - next);
drivers/net/wireless/ath/ath12k/core.c
349
goto next;
drivers/net/wireless/ath/ath12k/core.c
353
goto next;
drivers/net/wireless/ath/ath12k/core.c
363
goto next;
drivers/net/wireless/ath/ath12k/core.c
380
next:
drivers/net/wireless/ath/ath12k/core.c
474
goto next;
drivers/net/wireless/ath/ath12k/core.c
481
next:
drivers/net/wireless/ath/ath12k/dp_mon.c
260
msdu = head_msdu->next;
drivers/net/wireless/ath/ath12k/dp_mon.c
271
msdu = msdu->next;
drivers/net/wireless/ath/ath12k/dp_mon.c
274
prev_buf->next = NULL;
drivers/net/wireless/ath/ath12k/dp_mon.c
281
head_msdu->next = NULL;
drivers/net/wireless/ath/ath12k/dp_mon.c
307
msdu = msdu->next;
drivers/net/wireless/ath/ath12k/dp_mon.c
645
pmon->mon_mpdu->tail->next = msdu;
drivers/net/wireless/ath/ath12k/mac.c
1598
const u8 *next, *end;
drivers/net/wireless/ath/ath12k/mac.c
1613
next = ie + len;
drivers/net/wireless/ath/ath12k/mac.c
1615
if (WARN_ON(next > end))
drivers/net/wireless/ath/ath12k/mac.c
1618
memmove(ie, next, end - next);
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
1812
pmon->mon_mpdu->tail->next = NULL;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
2453
skb_next = mon_skb->next;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
2482
skb_next = mon_skb->next;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
2735
last->next = msdu;
drivers/net/wireless/ath/ath12k/wifi7/dp_mon.c
2763
last->next = NULL;
drivers/net/wireless/ath/ath5k/base.c
1627
goto next;
drivers/net/wireless/ath/ath5k/base.c
1640
next:
drivers/net/wireless/ath/ath5k/debug.c
166
.next = reg_next,
drivers/net/wireless/ath/ath6kl/htc_pipe.c
823
target->pipe.htc_packet_pool = (struct htc_packet *) packet->list.next;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
827
packet->list.next = NULL;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
840
packet->list.next = NULL;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
843
packet->list.next = lh;
drivers/net/wireless/ath/ath6kl/main.c
330
le32_to_cpu(debug_buf.next));
drivers/net/wireless/ath/ath6kl/target.h
343
__le32 next;
drivers/net/wireless/ath/ath9k/htc_drv_beacon.c
160
goto next;
drivers/net/wireless/ath/ath9k/htc_drv_beacon.c
170
goto next;
drivers/net/wireless/ath/ath9k/htc_drv_beacon.c
184
next:
drivers/net/wireless/ath/ath9k/recv.c
738
tbf = list_entry(bf->list.next, struct ath_rxbuf, list);
drivers/net/wireless/ath/ath9k/xmit.c
2662
bf = list_entry(bf_held->list.next, struct ath_buf,
drivers/net/wireless/ath/carl9170/tx.c
572
goto next;
drivers/net/wireless/ath/carl9170/tx.c
581
next:
drivers/net/wireless/ath/wcn36xx/dxe.c
220
cur_ctl = cur_ctl->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
256
cur = cur->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
326
cur_ctl = cur_ctl->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
340
cur = cur->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
447
ctl = ctl->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
59
struct wcn36xx_dxe_ctl *ctl = ch->head_blk_ctl, *next;
drivers/net/wireless/ath/wcn36xx/dxe.c
63
next = ctl->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
648
ctl = ctl->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
65
ctl = next;
drivers/net/wireless/ath/wcn36xx/dxe.c
761
ctl_skb = ctl_bd->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
822
ch->head_blk_ctl = ctl_skb->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
86
prev_ctl->next = cur_ctl;
drivers/net/wireless/ath/wcn36xx/dxe.c
865
ctl_skb_start = ctl_bd_start->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
87
cur_ctl->next = ch->head_blk_ctl;
drivers/net/wireless/ath/wcn36xx/dxe.c
872
ctl_bd = ctl_skb->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
873
ctl_skb = ctl_bd->next;
drivers/net/wireless/ath/wcn36xx/dxe.c
89
prev_ctl->next = cur_ctl;
drivers/net/wireless/ath/wcn36xx/dxe.h
420
struct wcn36xx_dxe_ctl *next;
drivers/net/wireless/ath/wil6210/cfg80211.c
2340
ret = vif->probe_client_pending.next;
drivers/net/wireless/ath/wil6210/wmi.c
3394
ret = wil->pending_wmi_ev.next;
drivers/net/wireless/broadcom/b43/pio.c
453
pack = list_entry(q->packets_list.next,
drivers/net/wireless/broadcom/b43/sdio.c
121
tuple = tuple->next;
drivers/net/wireless/broadcom/b43legacy/pio.c
461
packet = list_entry(queue->txfree.next, struct b43legacy_pio_txpacket,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
4800
goto next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
4806
goto next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
4813
goto next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
4831
next:
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
399
search = search->next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
451
search = search->next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
466
prev->next = search->next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
468
flow->tdls_entry = search->next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
487
tdls_entry->next = NULL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
494
while (search->next) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
495
search = search->next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
499
search->next = tdls_entry;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
52
search = search->next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.h
35
struct brcmf_flowring_tdls_entry *next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1294
struct sk_buff *cur, *next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1296
skb_queue_walk_safe(&bus->glom, cur, next) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1712
pfirst->len, pfirst->next,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c
404
req = list_entry(q->next, struct brcmf_usbreq, list);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c
405
list_del_init(q->next);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c
450
req = list_entry(q->next, struct brcmf_usbreq, list);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c
453
list_del(q->next);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c
462
struct brcmf_usbreq *req, *next;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/usb.c
464
list_for_each_entry_safe(req, next, q, list) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c
928
struct sk_buff *p, *next;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c
977
skb_queue_walk_safe(&dma_frames, p, next) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1510
t->next = wl->timers;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1576
wl->timers = wl->timers->next;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1587
if (tmp->next == t) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1588
tmp->next = t->next;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1595
tmp = tmp->next;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
288
struct brcms_timer *t, *next;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
318
for (t = wl->timers; t; t = next) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
319
next = t->next;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.h
44
struct brcms_timer *next; /* for freeing on unload */
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
7647
struct sk_buff *next = NULL;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
7670
skb_queue_walk_safe(&recv_frames, p, next) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c
542
pi->next = pi->sh->phy_head;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c
568
pi->sh->phy_head = pi->next;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c
569
else if (pi->sh->phy_head->next == pi)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c
570
pi->sh->phy_head->next = NULL;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_int.h
542
struct brcms_phy *next;
drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c
110
struct sk_buff *p, *next;
drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c
113
skb_queue_walk_safe(q, p, next) {
drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c
144
struct sk_buff *p, *next;
drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c
147
skb_queue_walk_safe(q, p, next) {
drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c
304
for (p = p0; p; p = p->next)
drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c
37
WARN_ON(skb->next);
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2664
i = (rxq->next + 1) % rxq->entries;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2743
rxq->next = (i ? i : rxq->entries) - 1;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2746
IPW_MEM_HOST_SHARED_RX_WRITE_INDEX, rxq->next);
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2802
element = priv->fw_pend_list.next;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2837
if (w != txq->next)
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2984
int next = txq->next;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2998
element = priv->msg_pend_list.next;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3005
&txq->drv[txq->next],
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3006
(u32) (txq->nic + txq->next *
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3009
packet->index = txq->next;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3011
tbd = &txq->drv[txq->next];
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3024
txq->next++;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3025
txq->next %= txq->entries;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3033
if (txq->next != next) {
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3039
txq->next);
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3053
int next = txq->next;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3065
element = priv->tx_pend_list.next;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3085
tbd = &txq->drv[txq->next];
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3087
packet->index = txq->next;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3124
txq->next++;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3125
txq->next %= txq->entries;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3136
tbd = &txq->drv[txq->next];
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3161
txq->next, tbd->host_addr,
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3169
txq->next++;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3170
txq->next %= txq->entries;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3180
if (txq->next != next) {
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3185
txq->next);
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3381
element = priv->tx_free_list.next;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
4378
write_register(priv->net_dev, w, q->next);
drivers/net/wireless/intel/ipw2x00/ipw2100.c
4488
priv->tx_queue.next = 0;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
4608
priv->rx_queue.next = priv->rx_queue.entries - 1;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
5791
element = priv->tx_pend_list.next;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
743
element = priv->msg_free_list.next;
drivers/net/wireless/intel/ipw2x00/ipw2100.h
186
u32 next;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
5102
element = rxq->rx_free.next;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
5140
element = rxq->rx_used.next;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
7554
element = priv->ieee->network_free_list.next;
drivers/net/wireless/intel/ipw2x00/libipw_crypto.c
82
struct libipw_crypt_data *entry, *next;
drivers/net/wireless/intel/ipw2x00/libipw_crypto.c
86
list_for_each_entry_safe(entry, next, &info->crypt_deinit_list, list) {
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
1486
target = list_entry(ieee->network_free_list.next,
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
1488
list_del(ieee->network_free_list.next);
drivers/net/wireless/intel/iwlegacy/3945-mac.c
1036
element = rxq->rx_used.next;
drivers/net/wireless/intel/iwlegacy/3945-mac.c
253
element = il->free_frames.next;
drivers/net/wireless/intel/iwlegacy/3945-mac.c
282
element = il->free_frames.next;
drivers/net/wireless/intel/iwlegacy/3945-mac.c
937
element = rxq->rx_free.next;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
265
element = rxq->rx_free.next;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
365
element = rxq->rx_used.next;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
3693
element = il->free_frames.next;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
3722
element = il->free_frames.next;
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
2850
list_entry(list->next, typeof(*entry), list);
drivers/net/wireless/intel/iwlwifi/fw/debugfs.c
416
.next = iwl_dbgfs_fw_info_seq_next,
drivers/net/wireless/intel/iwlwifi/iwl-utils.c
17
struct sk_buff *tmp, *next;
drivers/net/wireless/intel/iwlwifi/iwl-utils.c
30
next = skb_gso_segment(skb, netdev_flags);
drivers/net/wireless/intel/iwlwifi/iwl-utils.c
34
if (IS_ERR(next) && PTR_ERR(next) == -ENOMEM)
drivers/net/wireless/intel/iwlwifi/iwl-utils.c
37
if (WARN_ONCE(IS_ERR(next),
drivers/net/wireless/intel/iwlwifi/iwl-utils.c
38
"skb_gso_segment error: %d\n", (int)PTR_ERR(next)))
drivers/net/wireless/intel/iwlwifi/iwl-utils.c
39
return PTR_ERR(next);
drivers/net/wireless/intel/iwlwifi/iwl-utils.c
41
if (next)
drivers/net/wireless/intel/iwlwifi/iwl-utils.c
44
skb_list_walk_safe(next, tmp, next) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
608
struct page *next;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2807
.next = iwl_dbgfs_tx_queue_seq_next,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
47
info->next = *page_ptr;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1787
info->next = NULL;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
231
struct page *next;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
234
next = *page_ptr;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
237
while (next) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
239
struct page *tmp = next;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
241
info = IWL_TSO_PAGE_INFO(page_address(next));
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
242
next = info->next;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
245
if (!next && cmd_meta->sg_offset) {
drivers/net/wireless/intersil/p54/main.c
106
u8 *next;
drivers/net/wireless/intersil/p54/main.c
114
next = tim + 2 + dtim_len;
drivers/net/wireless/intersil/p54/main.c
119
memmove(tim, next, skb_tail_pointer(skb) - next);
drivers/net/wireless/intersil/p54/p54spi.c
434
entry = list_entry(priv->tx_pending.next,
drivers/net/wireless/marvell/libertas/firmware.c
107
goto next;
drivers/net/wireless/marvell/libertas/firmware.c
187
goto next;
drivers/net/wireless/marvell/libertas/firmware.c
192
goto next;
drivers/net/wireless/marvell/libertas/firmware.c
216
next:
drivers/net/wireless/marvell/libertas/firmware.c
98
next:
drivers/net/wireless/marvell/libertas/if_spi.c
901
cmd_packet_list.next);
drivers/net/wireless/marvell/libertas/if_spi.c
915
data_packet_list.next);
drivers/net/wireless/marvell/mwl8k.c
5297
return list_entry(priv->vif_list.next, struct mwl8k_vif, list);
drivers/net/wireless/mediatek/mt76/dma.c
314
u32 ctrl, next;
drivers/net/wireless/mediatek/mt76/dma.c
325
next = (q->head + 1) % q->ndesc;
drivers/net/wireless/mediatek/mt76/dma.c
364
q->head = next;
drivers/net/wireless/mediatek/mt76/dma.c
668
skb->prev = skb->next = NULL;
drivers/net/wireless/mediatek/mt76/mac80211.c
1527
nskb = nskb->next;
drivers/net/wireless/mediatek/mt76/mac80211.c
1528
skb->next = NULL;
drivers/net/wireless/mediatek/mt76/mac80211.c
933
phy->rx_amsdu[q].tail = &skb->next;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
2974
goto next;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
2992
next:
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
473
goto next;
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
498
goto next;
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
507
next:
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1368
goto next;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1373
goto next;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1390
next:
drivers/net/wireless/mediatek/mt76/sdio.c
532
skb->prev = skb->next = NULL;
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
259
goto next;
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
272
goto next;
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
297
next:
drivers/net/wireless/mediatek/mt76/testmode.c
167
frag_tail = &(*frag_tail)->next;
drivers/net/wireless/mediatek/mt76/tx.c
789
if (!iter->next) {
drivers/net/wireless/mediatek/mt76/usb.c
867
skb->prev = skb->next = NULL;
drivers/net/wireless/realtek/rtlwifi/base.c
1927
struct rtl_bssid_entry *entry, *next;
drivers/net/wireless/realtek/rtlwifi/base.c
1929
list_for_each_entry_safe(entry, next, &rtlpriv->scan_list.list, list) {
drivers/net/wireless/realtek/rtlwifi/base.c
1958
struct rtl_bssid_entry *entry, *next;
drivers/net/wireless/realtek/rtlwifi/base.c
1963
list_for_each_entry_safe(entry, next, &rtlpriv->scan_list.list, list) {
drivers/net/wireless/realtek/rtw88/main.c
431
desc->next = desc->data;
drivers/net/wireless/realtek/rtw88/main.c
440
u8 *next;
drivers/net/wireless/realtek/rtw88/main.c
447
next = desc->next + sizeof(struct rtw_fwcd_hdr);
drivers/net/wireless/realtek/rtw88/main.c
448
if (next - desc->data + size > desc->size) {
drivers/net/wireless/realtek/rtw88/main.c
453
hdr = (struct rtw_fwcd_hdr *)(desc->next);
drivers/net/wireless/realtek/rtw88/main.c
458
desc->next = next + size;
drivers/net/wireless/realtek/rtw88/main.c
460
return next;
drivers/net/wireless/realtek/rtw88/main.c
487
desc->next = NULL;
drivers/net/wireless/realtek/rtw88/main.h
1905
u8 *next;
drivers/net/wireless/realtek/rtw89/debug.c
4721
goto next;
drivers/net/wireless/realtek/rtw89/debug.c
4754
next:
drivers/net/wireless/realtek/rtw89/fw.c
1533
goto next;
drivers/net/wireless/realtek/rtw89/fw.c
1537
goto next;
drivers/net/wireless/realtek/rtw89/fw.c
1541
goto next;
drivers/net/wireless/realtek/rtw89/fw.c
1550
next:
drivers/net/wireless/realtek/rtw89/fw.c
8647
goto next;
drivers/net/wireless/realtek/rtw89/fw.c
8657
goto next;
drivers/net/wireless/realtek/rtw89/fw.c
8666
next:
drivers/net/wireless/realtek/rtw89/mac.c
5531
bool next = false;
drivers/net/wireless/realtek/rtw89/mac.c
5535
next = true;
drivers/net/wireless/realtek/rtw89/mac.c
5556
if (!retcode && next)
drivers/net/wireless/realtek/rtw89/mac.c
5812
bool next = false;
drivers/net/wireless/realtek/rtw89/mac.c
5831
next = true;
drivers/net/wireless/realtek/rtw89/mac.c
5886
if (next)
drivers/net/wireless/realtek/rtw89/phy.c
549
static bool __check_rate_pattern(struct rtw89_phy_rate_pattern *next,
drivers/net/wireless/realtek/rtw89/phy.c
565
if (next->enable)
drivers/net/wireless/realtek/rtw89/phy.c
569
next->rate = rate_base + c;
drivers/net/wireless/realtek/rtw89/phy.c
570
next->ra_mode = ra_mode;
drivers/net/wireless/realtek/rtw89/phy.c
571
next->ra_mask = ra_mask;
drivers/net/wireless/realtek/rtw89/phy.c
572
next->enable = true;
drivers/net/wireless/realtek/rtw89/rtw8851b.c
1054
goto next;
drivers/net/wireless/realtek/rtw89/rtw8851b.c
1062
next:
drivers/net/wireless/realtek/rtw89/rtw8852b_common.c
649
goto next;
drivers/net/wireless/realtek/rtw89/rtw8852b_common.c
657
next:
drivers/net/wireless/rsi/rsi_91x_core.c
122
skb = skb->next;
drivers/net/wireless/silabs/wfx/data_tx.c
422
WARN(skb->next || skb->prev, "skb is already member of a list");
drivers/net/wireless/silabs/wfx/main.c
204
goto next;
drivers/net/wireless/silabs/wfx/main.c
224
next:
drivers/net/wireless/ti/wlcore/main.c
3950
const u8 *next, *end = skb->data + skb->len;
drivers/net/wireless/ti/wlcore/main.c
3956
next = ie + len;
drivers/net/wireless/ti/wlcore/main.c
3957
memmove(ie, next, end - next);
drivers/net/wireless/ti/wlcore/main.c
3966
const u8 *next, *end = skb->data + skb->len;
drivers/net/wireless/ti/wlcore/main.c
3973
next = ie + len;
drivers/net/wireless/ti/wlcore/main.c
3974
memmove(ie, next, end - next);
drivers/net/wireless/ti/wlcore/tx.c
627
goto next;
drivers/net/wireless/ti/wlcore/tx.c
636
next:
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
269
td->next = 0;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.c
357
td->next = 0;
drivers/net/wwan/iosm/iosm_ipc_protocol_ops.h
324
__le32 next;
drivers/net/wwan/mhi_wwan_mbim.c
381
tail->next = skb;
drivers/net/wwan/t7xx/t7xx_port_wwan.c
121
cur = cur->next;
drivers/net/wwan/wwan_core.c
543
tail->next = skb;
drivers/net/wwan/wwan_core.c
924
tail->next = skb;
drivers/net/xen-netback/rx.c
349
pkt->frag_iter = frag_iter->next;
drivers/net/xen-netfront.c
1063
goto next;
drivers/net/xen-netfront.c
1073
goto next;
drivers/net/xen-netfront.c
1105
next:
drivers/nfc/pn533/pn533.c
1346
u8 *next, nfcid3[NFC_NFCID3_MAXSIZE];
drivers/nfc/pn533/pn533.c
1374
next = skb_put(skb, 1); /* Next */
drivers/nfc/pn533/pn533.c
1375
*next = 0;
drivers/nfc/pn533/pn533.c
1379
*next |= 1;
drivers/nfc/pn533/pn533.c
1383
*next |= 2;
drivers/nfc/pn533/pn533.c
1386
*next |= 4; /* We have some Gi */
drivers/nfc/pn533/pn533.c
1967
u8 *next, *arg, nfcid3[NFC_NFCID3_MAXSIZE];
drivers/nfc/pn533/pn533.c
2000
next = skb_put(skb, 1); /* Next */
drivers/nfc/pn533/pn533.c
2001
*next = 0;
drivers/nfc/pn533/pn533.c
2005
*next |= 1;
drivers/nfc/pn533/pn533.c
2013
*next |= 2;
drivers/nfc/pn533/pn533.c
2017
*next |= 4; /* We have some Gi */
drivers/nfc/pn533/pn533.c
2019
*next = 0;
drivers/nvdimm/badrange.c
106
struct badrange_entry *bre, *next;
drivers/nvdimm/badrange.c
118
list_for_each_entry_safe(bre, next, badrange_list, list) {
drivers/nvdimm/btt.c
804
struct arena_info *arena, *next;
drivers/nvdimm/btt.c
806
list_for_each_entry_safe(arena, next, &btt->arena_list, list) {
drivers/nvdimm/bus.c
407
struct badrange_entry *bre, *next;
drivers/nvdimm/bus.c
409
list_for_each_entry_safe(bre, next, badrange_list, list) {
drivers/nvdimm/namespace_devs.c
396
struct resource *next, struct resource *exist,
drivers/nvdimm/namespace_devs.c
454
struct resource *next = res->sibling, *new_res = NULL;
drivers/nvdimm/namespace_devs.c
470
space_valid(nd_region, ndd, label_id, NULL, next, exist,
drivers/nvdimm/namespace_devs.c
478
if (!loc && next) {
drivers/nvdimm/namespace_devs.c
480
valid.end = min(mapping_end, next->start - 1);
drivers/nvdimm/namespace_devs.c
481
space_valid(nd_region, ndd, label_id, res, next, exist,
drivers/nvdimm/namespace_devs.c
489
if (!loc && !next) {
drivers/nvdimm/namespace_devs.c
492
space_valid(nd_region, ndd, label_id, res, next, exist,
drivers/nvdimm/namespace_devs.c
513
if (strcmp(next->name, label_id->id) == 0) {
drivers/nvdimm/namespace_devs.c
515
rc = adjust_resource(next, next->start
drivers/nvdimm/namespace_devs.c
516
- allocate, resource_size(next)
drivers/nvdimm/namespace_devs.c
518
new_res = next;
drivers/nvdimm/namespace_devs.c
587
struct resource *next = res->sibling;
drivers/nvdimm/namespace_devs.c
590
if (!next || strcmp(res->name, label_id->id) != 0
drivers/nvdimm/namespace_devs.c
591
|| strcmp(next->name, label_id->id) != 0
drivers/nvdimm/namespace_devs.c
592
|| end != next->start)
drivers/nvdimm/namespace_devs.c
594
end += resource_size(next);
drivers/nvdimm/namespace_devs.c
595
nvdimm_free_dpa(ndd, next);
drivers/nvdimm/nd.h
364
#define for_each_dpa_resource_safe(ndd, res, next) \
drivers/nvdimm/nd.h
365
for (res = (ndd)->dpa.child, next = res ? res->sibling : NULL; \
drivers/nvdimm/nd.h
366
res; res = next, next = next ? next->sibling : NULL)
drivers/nvdimm/nd.h
442
static const unsigned next[] = { 0, 2, 3, 1 };
drivers/nvdimm/nd.h
444
return next[seq & 3];
drivers/nvme/host/core.c
4379
struct nvme_ns *ns, *next;
drivers/nvme/host/core.c
4383
list_for_each_entry_safe(ns, next, &ctrl->namespaces, list) {
drivers/nvme/host/core.c
4392
list_for_each_entry_safe(ns, next, &rm_list, list)
drivers/nvme/host/core.c
4549
struct nvme_ns *ns, *next;
drivers/nvme/host/core.c
4585
list_for_each_entry_safe(ns, next, &ns_list, list)
drivers/nvme/host/multipath.c
669
struct bio *bio, *next;
drivers/nvme/host/multipath.c
672
next = bio_list_get(&head->requeue_list);
drivers/nvme/host/multipath.c
675
while ((bio = next) != NULL) {
drivers/nvme/host/multipath.c
676
next = bio->bi_next;
drivers/nvme/host/rdma.c
1203
.next = NULL,
drivers/nvme/host/rdma.c
1602
wr.next = NULL;
drivers/nvme/host/rdma.c
1610
first->next = &wr;
drivers/nvme/host/rdma.c
1635
wr.next = NULL;
drivers/nvme/host/tcp.c
435
for (node = llist_del_all(&queue->req_list); node; node = node->next) {
drivers/nvme/target/fc.c
1522
struct nvmet_fc_tgt_assoc *assoc, *next;
drivers/nvme/target/fc.c
1527
list_for_each_entry_safe(assoc, next,
drivers/nvme/target/fc.c
1552
struct nvmet_fc_tgtport *tgtport, *next;
drivers/nvme/target/fc.c
1560
list_for_each_entry_safe(tgtport, next, &nvmet_fc_target_list,
drivers/nvme/target/loop.c
706
struct nvme_loop_ctrl *ctrl, *next;
drivers/nvme/target/loop.c
712
list_for_each_entry_safe(ctrl, next, &nvme_loop_ctrl_list, list)
drivers/nvme/target/rdma.c
520
rsp = list_entry(queue->rsp_wr_wait_list.next,
drivers/nvme/target/tcp.c
554
for (node = llist_del_all(&queue->resp_list); node; node = node->next) {
drivers/of/base.c
1698
struct property **next;
drivers/of/base.c
1700
for (next = list; *next; next = &(*next)->next) {
drivers/of/base.c
1701
if (*next == prop) {
drivers/of/base.c
1702
*next = prop->next;
drivers/of/base.c
1703
prop->next = NULL;
drivers/of/base.c
1719
struct property **next;
drivers/of/base.c
1725
prop->next = NULL;
drivers/of/base.c
1726
next = &np->properties;
drivers/of/base.c
1727
while (*next) {
drivers/of/base.c
1728
if (of_prop_cmp(prop->name, (*next)->name) == 0) {
drivers/of/base.c
1733
next = &(*next)->next;
drivers/of/base.c
1735
*next = prop;
drivers/of/base.c
1775
prop->next = np->deadprops;
drivers/of/base.c
1819
struct property **next, *oldprop;
drivers/of/base.c
1826
for (next = &np->properties; *next; next = &(*next)->next) {
drivers/of/base.c
1827
if (of_prop_cmp((*next)->name, newprop->name) == 0)
drivers/of/base.c
1830
*oldpropp = oldprop = *next;
drivers/of/base.c
1834
newprop->next = oldprop->next;
drivers/of/base.c
1835
*next = newprop;
drivers/of/base.c
1836
oldprop->next = np->deadprops;
drivers/of/base.c
1840
newprop->next = NULL;
drivers/of/base.c
1841
*next = newprop;
drivers/of/base.c
218
for (pp = np->properties; pp; pp = pp->next) {
drivers/of/base.c
659
struct device_node *next;
drivers/of/base.c
664
next = prev ? prev->sibling : node->child;
drivers/of/base.c
665
of_node_get(next);
drivers/of/base.c
667
return next;
drivers/of/base.c
685
struct device_node *next;
drivers/of/base.c
689
next = __of_get_next_child(node, prev);
drivers/of/base.c
691
return next;
drivers/of/base.c
711
struct device_node *next;
drivers/of/base.c
718
next = prev ? prev->sibling : node->child;
drivers/of/base.c
719
for (; next; next = next->sibling) {
drivers/of/base.c
720
if (!of_node_name_prefix(next, prefix))
drivers/of/base.c
722
if (of_node_get(next))
drivers/of/base.c
727
return next;
drivers/of/base.c
735
struct device_node *next;
drivers/of/base.c
742
next = prev ? prev->sibling : node->child;
drivers/of/base.c
743
for (; next; next = next->sibling) {
drivers/of/base.c
744
if (!checker(next))
drivers/of/base.c
746
if (of_node_get(next))
drivers/of/base.c
751
return next;
drivers/of/base.c
797
struct device_node *next = NULL;
drivers/of/base.c
806
next = prev->sibling;
drivers/of/base.c
808
next = node->child;
drivers/of/base.c
811
for (; next; next = next->sibling) {
drivers/of/base.c
812
if (__of_device_is_fail(next))
drivers/of/base.c
814
if (!(of_node_name_eq(next, "cpu") ||
drivers/of/base.c
815
__of_node_is_type(next, "cpu")))
drivers/of/base.c
817
if (of_node_get(next))
drivers/of/base.c
822
return next;
drivers/of/dynamic.c
319
struct property *prop, *next;
drivers/of/dynamic.c
321
for (prop = prop_list; prop != NULL; prop = next) {
drivers/of/dynamic.c
322
next = prop->next;
drivers/of/dynamic.c
943
new_pp->next = np->deadprops;
drivers/of/fdt.c
155
pprev = &pp->next;
drivers/of/fdt.c
239
struct device_node *child, *next;
drivers/of/fdt.c
253
next = child->sibling;
drivers/of/fdt.c
257
child = next;
drivers/of/of_reserved_mem.c
547
struct reserved_mem *this, *next;
drivers/of/of_reserved_mem.c
550
next = &reserved_mem[i + 1];
drivers/of/of_reserved_mem.c
552
if (this->base + this->size > next->base) {
drivers/of/of_reserved_mem.c
556
next_end = next->base + next->size;
drivers/of/of_reserved_mem.c
559
next->name, &next->base, &next_end);
drivers/of/overlay.c
353
new_prop->next = target->np->deadprops;
drivers/of/overlay.c
854
if (ovcs->cset.entries.next)
drivers/of/pdt.c
121
tail->next = of_pdt_build_one_prop(node, NULL, NULL, NULL, 0);
drivers/of/pdt.c
122
tail = tail->next;
drivers/of/pdt.c
124
tail->next = of_pdt_build_one_prop(node, tail->name,
drivers/of/pdt.c
126
tail = tail->next;
drivers/of/unittest.c
2035
save_next = prop->next;
drivers/of/unittest.c
2058
struct device_node *next, *dup, *child;
drivers/of/unittest.c
2093
next = child->sibling;
drivers/of/unittest.c
2095
child = next;
drivers/of/unittest.c
2164
struct device_node *next = np->sibling;
drivers/of/unittest.c
2169
np = next;
drivers/parisc/ccio-dma.c
1092
ioc = ioc->next;
drivers/parisc/ccio-dma.c
1106
ioc = ioc->next;
drivers/parisc/ccio-dma.c
1131
ioc = ioc->next;
drivers/parisc/ccio-dma.c
1535
ioc_p = &(*ioc_p)->next;
drivers/parisc/ccio-dma.c
242
struct ioc *next; /* Linked list of discovered iocs */
drivers/parisc/hppb.c
28
struct hppb_card *next;
drivers/parisc/hppb.c
33
.next = NULL,
drivers/parisc/hppb.c
52
while(card->next) {
drivers/parisc/hppb.c
53
card = card->next;
drivers/parisc/hppb.c
57
card->next = kzalloc_obj(struct hppb_card);
drivers/parisc/hppb.c
58
if(!card->next) {
drivers/parisc/hppb.c
62
card = card->next;
drivers/parisc/sba_iommu.c
1693
sba_dev->next = sba_list;
drivers/parport/daisy.c
211
p = &dev->next;
drivers/parport/daisy.c
214
*p = dev->next;
drivers/parport/daisy.c
251
p = p->next;
drivers/parport/daisy.c
34
struct daisydev *next;
drivers/parport/daisy.c
60
for (p = &topology; *p && (*p)->devnum<devnum; p = &(*p)->next)
drivers/parport/daisy.c
62
newdev->next = *p;
drivers/parport/parport_ip32.c
469
dma_addr_t next;
drivers/parport/parport_ip32.c
511
(unsigned int)parport_ip32_dma.next,
drivers/parport/parport_ip32.c
515
ctxval |= parport_ip32_dma.next &
drivers/parport/parport_ip32.c
520
parport_ip32_dma.next += count;
drivers/parport/parport_ip32.c
595
parport_ip32_dma.next = parport_ip32_dma.buf;
drivers/parport/parport_ip32.c
604
(parport_ip32_dma.next & (MACEPAR_CONTEXT_DATA_BOUND - 1));
drivers/parport/parport_pc.c
3394
priv = list_entry(ports_list.next,
drivers/parport/procfs.c
52
for (dev = port->devices; dev ; dev = dev->next) {
drivers/parport/share.c
1196
for (pd = port->devices; !port->cad && pd; pd = pd->next) {
drivers/parport/share.c
769
par_dev->next = port->physport->devices;
drivers/parport/share.c
845
if (dev->next)
drivers/parport/share.c
846
dev->next->prev = dev->prev;
drivers/parport/share.c
848
dev->prev->next = dev->next;
drivers/parport/share.c
850
port->devices = dev->next;
drivers/pci/controller/pci-tegra.c
2541
.next = tegra_pcie_ports_seq_next,
drivers/pci/hotplug/acpiphp_glue.c
153
struct acpiphp_slot *slot, *next;
drivers/pci/hotplug/acpiphp_glue.c
160
list_for_each_entry_safe(slot, next, &bridge->slots, node) {
drivers/pci/hotplug/cpqphp.h
226
struct pci_func *next;
drivers/pci/hotplug/cpqphp.h
248
struct slot *next;
drivers/pci/hotplug/cpqphp.h
267
struct pci_resource *next;
drivers/pci/hotplug/cpqphp.h
278
struct controller *next;
drivers/pci/hotplug/cpqphp.h
466
node->next = *head;
drivers/pci/hotplug/cpqphp_core.c
1166
ctrl->next = NULL;
drivers/pci/hotplug/cpqphp_core.c
1168
ctrl->next = cpqhp_ctrl_list;
drivers/pci/hotplug/cpqphp_core.c
1248
struct pci_func *next;
drivers/pci/hotplug/cpqphp_core.c
1279
res = res->next;
drivers/pci/hotplug/cpqphp_core.c
1286
res = res->next;
drivers/pci/hotplug/cpqphp_core.c
1293
res = res->next;
drivers/pci/hotplug/cpqphp_core.c
1300
res = res->next;
drivers/pci/hotplug/cpqphp_core.c
1307
ctrl = ctrl->next;
drivers/pci/hotplug/cpqphp_core.c
1312
next = cpqhp_slot_list[loop];
drivers/pci/hotplug/cpqphp_core.c
1313
while (next != NULL) {
drivers/pci/hotplug/cpqphp_core.c
1314
res = next->io_head;
drivers/pci/hotplug/cpqphp_core.c
1317
res = res->next;
drivers/pci/hotplug/cpqphp_core.c
1321
res = next->mem_head;
drivers/pci/hotplug/cpqphp_core.c
1324
res = res->next;
drivers/pci/hotplug/cpqphp_core.c
1328
res = next->p_mem_head;
drivers/pci/hotplug/cpqphp_core.c
1331
res = res->next;
drivers/pci/hotplug/cpqphp_core.c
1335
res = next->bus_head;
drivers/pci/hotplug/cpqphp_core.c
1338
res = res->next;
drivers/pci/hotplug/cpqphp_core.c
1342
TempSlot = next;
drivers/pci/hotplug/cpqphp_core.c
1343
next = next->next;
drivers/pci/hotplug/cpqphp_core.c
274
next_slot = old_slot->next;
drivers/pci/hotplug/cpqphp_core.c
666
slot->next = ctrl->slot;
drivers/pci/hotplug/cpqphp_ctrl.c
1001
while ((next->next != old_slot) && (next->next != NULL))
drivers/pci/hotplug/cpqphp_ctrl.c
1002
next = next->next;
drivers/pci/hotplug/cpqphp_ctrl.c
1004
if (next->next == old_slot) {
drivers/pci/hotplug/cpqphp_ctrl.c
1005
next->next = old_slot->next;
drivers/pci/hotplug/cpqphp_ctrl.c
1024
struct pci_func *next;
drivers/pci/hotplug/cpqphp_ctrl.c
1030
next = cpqhp_slot_list[tempBus];
drivers/pci/hotplug/cpqphp_ctrl.c
1032
while (!slot_remove(next))
drivers/pci/hotplug/cpqphp_ctrl.c
1033
next = cpqhp_slot_list[tempBus];
drivers/pci/hotplug/cpqphp_ctrl.c
1036
next = cpqhp_slot_list[bridge->bus];
drivers/pci/hotplug/cpqphp_ctrl.c
1038
if (next == NULL)
drivers/pci/hotplug/cpqphp_ctrl.c
1041
if (next == bridge) {
drivers/pci/hotplug/cpqphp_ctrl.c
1042
cpqhp_slot_list[bridge->bus] = bridge->next;
drivers/pci/hotplug/cpqphp_ctrl.c
1046
while ((next->next != bridge) && (next->next != NULL))
drivers/pci/hotplug/cpqphp_ctrl.c
1047
next = next->next;
drivers/pci/hotplug/cpqphp_ctrl.c
1049
if (next->next != bridge)
drivers/pci/hotplug/cpqphp_ctrl.c
1051
next->next = bridge->next;
drivers/pci/hotplug/cpqphp_ctrl.c
1079
while (func->next != NULL) {
drivers/pci/hotplug/cpqphp_ctrl.c
1080
func = func->next;
drivers/pci/hotplug/cpqphp_ctrl.c
1129
for (slot = ctrl->slot; slot; slot = slot->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
121
slot = slot->next;
drivers/pci/hotplug/cpqphp_ctrl.c
1740
for (ctrl = cpqhp_ctrl_list; ctrl; ctrl = ctrl->next)
drivers/pci/hotplug/cpqphp_ctrl.c
2402
io_node->length, io_node->next);
drivers/pci/hotplug/cpqphp_ctrl.c
2405
mem_node->length, mem_node->next);
drivers/pci/hotplug/cpqphp_ctrl.c
2408
p_mem_node->length, p_mem_node->next);
drivers/pci/hotplug/cpqphp_ctrl.c
2456
bus_node->next = NULL;
drivers/pci/hotplug/cpqphp_ctrl.c
2461
io_node->next = NULL;
drivers/pci/hotplug/cpqphp_ctrl.c
2474
mem_node->next = NULL;
drivers/pci/hotplug/cpqphp_ctrl.c
2484
p_mem_node->next = NULL;
drivers/pci/hotplug/cpqphp_ctrl.c
2554
hold_bus_node->next = func->bus_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2596
hold_IO_node->next = func->io_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2613
hold_IO_node->next = func->io_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2618
hold_IO_node->next = func->io_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2646
hold_mem_node->next = func->mem_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2665
hold_mem_node->next = func->mem_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2670
hold_mem_node->next = func->mem_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2699
hold_p_mem_node->next = func->p_mem_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2716
hold_p_mem_node->next = func->p_mem_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2721
hold_p_mem_node->next = func->p_mem_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2773
io_node->base, io_node->length, io_node->next);
drivers/pci/hotplug/cpqphp_ctrl.c
2778
io_node->next = func->io_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2792
p_mem_node->next = func->p_mem_head;
drivers/pci/hotplug/cpqphp_ctrl.c
2808
mem_node->next = func->mem_head;
drivers/pci/hotplug/cpqphp_ctrl.c
307
if (!((*head)->next))
drivers/pci/hotplug/cpqphp_ctrl.c
314
if (((*head)->next) &&
drivers/pci/hotplug/cpqphp_ctrl.c
315
((*head)->length > (*head)->next->length)) {
drivers/pci/hotplug/cpqphp_ctrl.c
318
*head = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
319
current_res->next = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
320
(*head)->next = current_res;
drivers/pci/hotplug/cpqphp_ctrl.c
325
while (current_res->next && current_res->next->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
326
if (current_res->next->length > current_res->next->next->length) {
drivers/pci/hotplug/cpqphp_ctrl.c
328
next_res = current_res->next;
drivers/pci/hotplug/cpqphp_ctrl.c
329
current_res->next = current_res->next->next;
drivers/pci/hotplug/cpqphp_ctrl.c
330
current_res = current_res->next;
drivers/pci/hotplug/cpqphp_ctrl.c
331
next_res->next = current_res->next;
drivers/pci/hotplug/cpqphp_ctrl.c
332
current_res->next = next_res;
drivers/pci/hotplug/cpqphp_ctrl.c
334
current_res = current_res->next;
drivers/pci/hotplug/cpqphp_ctrl.c
355
if (!((*head)->next))
drivers/pci/hotplug/cpqphp_ctrl.c
362
if (((*head)->next) &&
drivers/pci/hotplug/cpqphp_ctrl.c
363
((*head)->length < (*head)->next->length)) {
drivers/pci/hotplug/cpqphp_ctrl.c
366
*head = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
367
current_res->next = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
368
(*head)->next = current_res;
drivers/pci/hotplug/cpqphp_ctrl.c
373
while (current_res->next && current_res->next->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
374
if (current_res->next->length < current_res->next->next->length) {
drivers/pci/hotplug/cpqphp_ctrl.c
376
next_res = current_res->next;
drivers/pci/hotplug/cpqphp_ctrl.c
377
current_res->next = current_res->next->next;
drivers/pci/hotplug/cpqphp_ctrl.c
378
current_res = current_res->next;
drivers/pci/hotplug/cpqphp_ctrl.c
379
next_res->next = current_res->next;
drivers/pci/hotplug/cpqphp_ctrl.c
380
current_res->next = next_res;
drivers/pci/hotplug/cpqphp_ctrl.c
382
current_res = current_res->next;
drivers/pci/hotplug/cpqphp_ctrl.c
446
split_node->next = node;
drivers/pci/hotplug/cpqphp_ctrl.c
454
*head = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
457
while (prevnode->next != node)
drivers/pci/hotplug/cpqphp_ctrl.c
458
prevnode = prevnode->next;
drivers/pci/hotplug/cpqphp_ctrl.c
460
prevnode->next = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
462
node->next = NULL;
drivers/pci/hotplug/cpqphp_ctrl.c
487
while (node->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
489
node = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
542
for (node = *head; node; node = node->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
567
split_node->next = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
568
node->next = split_node;
drivers/pci/hotplug/cpqphp_ctrl.c
586
split_node->next = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
587
node->next = split_node;
drivers/pci/hotplug/cpqphp_ctrl.c
598
*head = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
601
while (prevnode->next != node)
drivers/pci/hotplug/cpqphp_ctrl.c
602
prevnode = prevnode->next;
drivers/pci/hotplug/cpqphp_ctrl.c
604
prevnode->next = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
606
node->next = NULL;
drivers/pci/hotplug/cpqphp_ctrl.c
636
for (max = *head; max; max = max->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
663
split_node->next = max->next;
drivers/pci/hotplug/cpqphp_ctrl.c
664
max->next = split_node;
drivers/pci/hotplug/cpqphp_ctrl.c
681
split_node->next = max->next;
drivers/pci/hotplug/cpqphp_ctrl.c
682
max->next = split_node;
drivers/pci/hotplug/cpqphp_ctrl.c
692
*head = max->next;
drivers/pci/hotplug/cpqphp_ctrl.c
694
while (temp && temp->next != max)
drivers/pci/hotplug/cpqphp_ctrl.c
695
temp = temp->next;
drivers/pci/hotplug/cpqphp_ctrl.c
698
temp->next = max->next;
drivers/pci/hotplug/cpqphp_ctrl.c
701
max->next = NULL;
drivers/pci/hotplug/cpqphp_ctrl.c
733
for (node = *head; node; node = node->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
760
split_node->next = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
761
node->next = split_node;
drivers/pci/hotplug/cpqphp_ctrl.c
780
split_node->next = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
781
node->next = split_node;
drivers/pci/hotplug/cpqphp_ctrl.c
788
*head = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
791
while (prevnode->next != node)
drivers/pci/hotplug/cpqphp_ctrl.c
792
prevnode = prevnode->next;
drivers/pci/hotplug/cpqphp_ctrl.c
794
prevnode->next = node->next;
drivers/pci/hotplug/cpqphp_ctrl.c
796
node->next = NULL;
drivers/pci/hotplug/cpqphp_ctrl.c
824
dbg("*head->next = %p\n", (*head)->next);
drivers/pci/hotplug/cpqphp_ctrl.c
826
if (!(*head)->next)
drivers/pci/hotplug/cpqphp_ctrl.c
830
dbg("*head->next->base = 0x%x\n", (*head)->next->base);
drivers/pci/hotplug/cpqphp_ctrl.c
835
if (((*head)->next) &&
drivers/pci/hotplug/cpqphp_ctrl.c
836
((*head)->base > (*head)->next->base)) {
drivers/pci/hotplug/cpqphp_ctrl.c
838
(*head) = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
839
node1->next = (*head)->next;
drivers/pci/hotplug/cpqphp_ctrl.c
840
(*head)->next = node1;
drivers/pci/hotplug/cpqphp_ctrl.c
846
while (node1->next && node1->next->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
847
if (node1->next->base > node1->next->next->base) {
drivers/pci/hotplug/cpqphp_ctrl.c
849
node2 = node1->next;
drivers/pci/hotplug/cpqphp_ctrl.c
850
node1->next = node1->next->next;
drivers/pci/hotplug/cpqphp_ctrl.c
851
node1 = node1->next;
drivers/pci/hotplug/cpqphp_ctrl.c
852
node2->next = node1->next;
drivers/pci/hotplug/cpqphp_ctrl.c
853
node1->next = node2;
drivers/pci/hotplug/cpqphp_ctrl.c
855
node1 = node1->next;
drivers/pci/hotplug/cpqphp_ctrl.c
861
while (node1 && node1->next) {
drivers/pci/hotplug/cpqphp_ctrl.c
862
if ((node1->base + node1->length) == node1->next->base) {
drivers/pci/hotplug/cpqphp_ctrl.c
865
node1->length += node1->next->length;
drivers/pci/hotplug/cpqphp_ctrl.c
866
node2 = node1->next;
drivers/pci/hotplug/cpqphp_ctrl.c
867
node1->next = node1->next->next;
drivers/pci/hotplug/cpqphp_ctrl.c
870
node1 = node1->next;
drivers/pci/hotplug/cpqphp_ctrl.c
956
struct pci_func *next;
drivers/pci/hotplug/cpqphp_ctrl.c
962
new_slot->next = NULL;
drivers/pci/hotplug/cpqphp_ctrl.c
968
next = cpqhp_slot_list[busnumber];
drivers/pci/hotplug/cpqphp_ctrl.c
969
while (next->next != NULL)
drivers/pci/hotplug/cpqphp_ctrl.c
970
next = next->next;
drivers/pci/hotplug/cpqphp_ctrl.c
971
next->next = new_slot;
drivers/pci/hotplug/cpqphp_ctrl.c
985
struct pci_func *next;
drivers/pci/hotplug/cpqphp_ctrl.c
990
next = cpqhp_slot_list[old_slot->bus];
drivers/pci/hotplug/cpqphp_ctrl.c
991
if (next == NULL)
drivers/pci/hotplug/cpqphp_ctrl.c
994
if (next == old_slot) {
drivers/pci/hotplug/cpqphp_ctrl.c
995
cpqhp_slot_list[old_slot->bus] = old_slot->next;
drivers/pci/hotplug/cpqphp_nvram.c
307
resNode = resNode->next;
drivers/pci/hotplug/cpqphp_nvram.c
332
resNode = resNode->next;
drivers/pci/hotplug/cpqphp_nvram.c
357
resNode = resNode->next;
drivers/pci/hotplug/cpqphp_nvram.c
382
resNode = resNode->next;
drivers/pci/hotplug/cpqphp_nvram.c
388
ctrl = ctrl->next;
drivers/pci/hotplug/cpqphp_nvram.c
452
p_byte = &(p_EV_header->next);
drivers/pci/hotplug/cpqphp_nvram.c
454
p_ev_ctrl = (struct ev_hrt_ctrl *) &(p_EV_header->next);
drivers/pci/hotplug/cpqphp_nvram.c
530
mem_node->next = ctrl->mem_head;
drivers/pci/hotplug/cpqphp_nvram.c
558
p_mem_node->next = ctrl->p_mem_head;
drivers/pci/hotplug/cpqphp_nvram.c
586
io_node->next = ctrl->io_head;
drivers/pci/hotplug/cpqphp_nvram.c
61
u8 next;
drivers/pci/hotplug/cpqphp_nvram.c
612
bus_node->next = ctrl->bus_head;
drivers/pci/hotplug/cpqphp_nvram.c
72
u8 next;
drivers/pci/hotplug/cpqphp_pci.c
1022
struct pci_func *next;
drivers/pci/hotplug/cpqphp_pci.c
1066
next = cpqhp_slot_list[secondary_bus];
drivers/pci/hotplug/cpqphp_pci.c
1068
while (next != NULL) {
drivers/pci/hotplug/cpqphp_pci.c
1069
rc = cpqhp_valid_replace(ctrl, next);
drivers/pci/hotplug/cpqphp_pci.c
1073
next = next->next;
drivers/pci/hotplug/cpqphp_pci.c
1301
io_node->next = ctrl->io_head;
drivers/pci/hotplug/cpqphp_pci.c
1304
io_node->next = func->io_head;
drivers/pci/hotplug/cpqphp_pci.c
1324
mem_node->next = ctrl->mem_head;
drivers/pci/hotplug/cpqphp_pci.c
1327
mem_node->next = func->mem_head;
drivers/pci/hotplug/cpqphp_pci.c
1349
p_mem_node->next = ctrl->p_mem_head;
drivers/pci/hotplug/cpqphp_pci.c
1352
p_mem_node->next = func->p_mem_head;
drivers/pci/hotplug/cpqphp_pci.c
1372
bus_node->next = ctrl->bus_head;
drivers/pci/hotplug/cpqphp_pci.c
1375
bus_node->next = func->bus_head;
drivers/pci/hotplug/cpqphp_pci.c
1418
t_node = node->next;
drivers/pci/hotplug/cpqphp_pci.c
1426
t_node = node->next;
drivers/pci/hotplug/cpqphp_pci.c
1434
t_node = node->next;
drivers/pci/hotplug/cpqphp_pci.c
1442
t_node = node->next;
drivers/pci/hotplug/cpqphp_pci.c
1470
res = res->next;
drivers/pci/hotplug/cpqphp_pci.c
1479
res = res->next;
drivers/pci/hotplug/cpqphp_pci.c
1488
res = res->next;
drivers/pci/hotplug/cpqphp_pci.c
1497
res = res->next;
drivers/pci/hotplug/cpqphp_pci.c
1517
res = res->next;
drivers/pci/hotplug/cpqphp_pci.c
1526
res = res->next;
drivers/pci/hotplug/cpqphp_pci.c
1535
res = res->next;
drivers/pci/hotplug/cpqphp_pci.c
1544
res = res->next;
drivers/pci/hotplug/cpqphp_pci.c
551
struct pci_func *next;
drivers/pci/hotplug/cpqphp_pci.c
570
next = cpqhp_slot_list[sub_bus];
drivers/pci/hotplug/cpqphp_pci.c
572
while (next != NULL) {
drivers/pci/hotplug/cpqphp_pci.c
573
rc = cpqhp_save_base_addr_length(ctrl, next);
drivers/pci/hotplug/cpqphp_pci.c
577
next = next->next;
drivers/pci/hotplug/cpqphp_pci.c
731
bus_node->next = func->bus_head;
drivers/pci/hotplug/cpqphp_pci.c
746
io_node->next = func->io_head;
drivers/pci/hotplug/cpqphp_pci.c
762
mem_node->next = func->mem_head;
drivers/pci/hotplug/cpqphp_pci.c
778
p_mem_node->next = func->p_mem_head;
drivers/pci/hotplug/cpqphp_pci.c
810
io_node->next = func->io_head;
drivers/pci/hotplug/cpqphp_pci.c
826
p_mem_node->next = func->p_mem_head;
drivers/pci/hotplug/cpqphp_pci.c
842
mem_node->next = func->mem_head;
drivers/pci/hotplug/cpqphp_pci.c
878
io_node->next = func->io_head;
drivers/pci/hotplug/cpqphp_pci.c
894
p_mem_node->next = func->p_mem_head;
drivers/pci/hotplug/cpqphp_pci.c
910
mem_node->next = func->mem_head;
drivers/pci/hotplug/cpqphp_pci.c
941
struct pci_func *next;
drivers/pci/hotplug/cpqphp_pci.c
968
next = cpqhp_slot_list[sub_bus];
drivers/pci/hotplug/cpqphp_pci.c
970
while (next != NULL) {
drivers/pci/hotplug/cpqphp_pci.c
971
rc = cpqhp_configure_board(ctrl, next);
drivers/pci/hotplug/cpqphp_pci.c
975
next = next->next;
drivers/pci/hotplug/cpqphp_sysfs.c
106
res = res->next;
drivers/pci/hotplug/cpqphp_sysfs.c
108
slot = slot->next;
drivers/pci/hotplug/cpqphp_sysfs.c
39
res = res->next;
drivers/pci/hotplug/cpqphp_sysfs.c
46
res = res->next;
drivers/pci/hotplug/cpqphp_sysfs.c
53
res = res->next;
drivers/pci/hotplug/cpqphp_sysfs.c
60
res = res->next;
drivers/pci/hotplug/cpqphp_sysfs.c
85
res = res->next;
drivers/pci/hotplug/cpqphp_sysfs.c
92
res = res->next;
drivers/pci/hotplug/cpqphp_sysfs.c
99
res = res->next;
drivers/pci/hotplug/ibmphp.h
215
struct ebda_rsrc_list *next;
drivers/pci/hotplug/ibmphp.h
324
struct range_node *next;
drivers/pci/hotplug/ibmphp.h
355
struct resource_node *next;
drivers/pci/hotplug/ibmphp.h
683
struct pci_func *next;
drivers/pci/hotplug/ibmphp_core.c
551
func_cur = func_cur->next;
drivers/pci/hotplug/ibmphp_core.c
565
struct slot *slot_cur, *next;
drivers/pci/hotplug/ibmphp_core.c
569
list_for_each_entry_safe(slot_cur, next, &ibmphp_slot_head,
drivers/pci/hotplug/ibmphp_ebda.c
1033
struct bus_info *bus_info, *next;
drivers/pci/hotplug/ibmphp_ebda.c
1035
list_for_each_entry_safe(bus_info, next, &bus_info_head,
drivers/pci/hotplug/ibmphp_ebda.c
1043
struct controller *controller = NULL, *next;
drivers/pci/hotplug/ibmphp_ebda.c
1046
list_for_each_entry_safe(controller, next, &ebda_hpc_head,
drivers/pci/hotplug/ibmphp_ebda.c
1060
struct ebda_pci_rsrc *resource, *next;
drivers/pci/hotplug/ibmphp_ebda.c
1062
list_for_each_entry_safe(resource, next, &ibmphp_ebda_pci_rsrc_head,
drivers/pci/hotplug/ibmphp_pci.c
142
cur_func->next = NULL;
drivers/pci/hotplug/ibmphp_pci.c
1583
temp_func = cur_func->next;
drivers/pci/hotplug/ibmphp_pci.c
161
cur_func->next = newfunc;
drivers/pci/hotplug/ibmphp_pci.c
202
for (prev_func = cur_func; prev_func->next; prev_func = prev_func->next) ;
drivers/pci/hotplug/ibmphp_pci.c
203
prev_func->next = newfunc;
drivers/pci/hotplug/ibmphp_pci.c
205
cur_func->next = newfunc;
drivers/pci/hotplug/ibmphp_pci.c
227
for (prev_func = cur_func; prev_func->next; prev_func = prev_func->next);
drivers/pci/hotplug/ibmphp_pci.c
228
prev_func->next = newfunc;
drivers/pci/hotplug/ibmphp_pci.c
274
for (prev_func = cur_func; prev_func->next; prev_func = prev_func->next);
drivers/pci/hotplug/ibmphp_pci.c
275
prev_func->next = newfunc;
drivers/pci/hotplug/ibmphp_pci.c
277
cur_func->next = newfunc;
drivers/pci/hotplug/ibmphp_res.c
1062
if (!res_cur->next) {
drivers/pci/hotplug/ibmphp_res.c
1181
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
1182
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1238
range = range->next;
drivers/pci/hotplug/ibmphp_res.c
1304
range = range->next;
drivers/pci/hotplug/ibmphp_res.c
1364
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
1365
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1377
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
1378
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1390
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
1391
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1404
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1438
range_cur = range_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1452
range_cur = range_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1466
range_cur = range_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1512
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
1513
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1526
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1552
struct bus_node *bus_cur = NULL, *next;
drivers/pci/hotplug/ibmphp_res.c
1561
list_for_each_entry_safe(bus_cur, next, &gbuses, bus_list) {
drivers/pci/hotplug/ibmphp_res.c
1568
range_cur = range_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1579
range_cur = range_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1590
range_cur = range_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1600
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
1601
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1613
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
1614
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1626
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
1627
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1640
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1672
for (pfmem_cur = bus_cur->firstPFMem, pfmem_prev = NULL; pfmem_cur; pfmem_prev = pfmem_cur, pfmem_cur = pfmem_cur->next) {
drivers/pci/hotplug/ibmphp_res.c
1675
pfmem_prev->next = pfmem_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1677
bus_cur->firstPFMem = pfmem_cur->next;
drivers/pci/hotplug/ibmphp_res.c
1680
pfmem_cur->next = NULL;
drivers/pci/hotplug/ibmphp_res.c
1686
pfmem_cur->next = bus_cur->firstPFMemFromMem;
drivers/pci/hotplug/ibmphp_res.c
1719
pfmem->next = bus_cur->firstPFMemFromMem;
drivers/pci/hotplug/ibmphp_res.c
1721
pfmem->next = NULL;
drivers/pci/hotplug/ibmphp_res.c
1778
range = range->next;
drivers/pci/hotplug/ibmphp_res.c
1788
range = range->next;
drivers/pci/hotplug/ibmphp_res.c
1799
range = range->next;
drivers/pci/hotplug/ibmphp_res.c
1812
if (res->next)
drivers/pci/hotplug/ibmphp_res.c
1813
res = res->next;
drivers/pci/hotplug/ibmphp_res.c
1827
if (res->next)
drivers/pci/hotplug/ibmphp_res.c
1828
res = res->next;
drivers/pci/hotplug/ibmphp_res.c
1842
if (res->next)
drivers/pci/hotplug/ibmphp_res.c
1843
res = res->next;
drivers/pci/hotplug/ibmphp_res.c
1858
res = res->next;
drivers/pci/hotplug/ibmphp_res.c
1886
range_cur = range_cur->next;
drivers/pci/hotplug/ibmphp_res.c
395
range_cur = range_cur->next;
drivers/pci/hotplug/ibmphp_res.c
411
range->next = range_cur;
drivers/pci/hotplug/ibmphp_res.c
416
range->next = NULL;
drivers/pci/hotplug/ibmphp_res.c
417
range_prev->next = range;
drivers/pci/hotplug/ibmphp_res.c
422
range_prev->next = range;
drivers/pci/hotplug/ibmphp_res.c
423
range->next = range_cur;
drivers/pci/hotplug/ibmphp_res.c
430
range_cur = range_cur->next;
drivers/pci/hotplug/ibmphp_res.c
467
if (res->next)
drivers/pci/hotplug/ibmphp_res.c
468
res = res->next;
drivers/pci/hotplug/ibmphp_res.c
481
res = res->next;
drivers/pci/hotplug/ibmphp_res.c
521
range = range->next;
drivers/pci/hotplug/ibmphp_res.c
524
if (res->next)
drivers/pci/hotplug/ibmphp_res.c
525
res = res->next;
drivers/pci/hotplug/ibmphp_res.c
621
range_cur = range_cur->next;
drivers/pci/hotplug/ibmphp_res.c
658
res->next = NULL;
drivers/pci/hotplug/ibmphp_res.c
670
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
671
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
680
res->next = NULL;
drivers/pci/hotplug/ibmphp_res.c
688
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
692
res_prev->next = res;
drivers/pci/hotplug/ibmphp_res.c
693
res->next = NULL;
drivers/pci/hotplug/ibmphp_res.c
711
res_prev->next = res;
drivers/pci/hotplug/ibmphp_res.c
715
res->next = res_cur;
drivers/pci/hotplug/ibmphp_res.c
722
res->next = NULL;
drivers/pci/hotplug/ibmphp_res.c
740
res->next = NULL;
drivers/pci/hotplug/ibmphp_res.c
800
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
801
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
823
if (mem_cur->next)
drivers/pci/hotplug/ibmphp_res.c
824
mem_cur = mem_cur->next;
drivers/pci/hotplug/ibmphp_res.c
835
bus_cur->firstPFMemFromMem = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
837
res_prev->next = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
842
if (res_cur->next)
drivers/pci/hotplug/ibmphp_res.c
843
res_cur = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
858
if (res_cur->next) {
drivers/pci/hotplug/ibmphp_res.c
861
bus_cur->firstIO = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
864
bus_cur->firstMem = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
867
bus_cur->firstPFMem = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
898
if (res_cur->next) {
drivers/pci/hotplug/ibmphp_res.c
900
res_prev->next = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
902
res_prev->nextRange = res_cur->next;
drivers/pci/hotplug/ibmphp_res.c
904
res_prev->next = NULL;
drivers/pci/hotplug/ibmphp_res.c
907
res_prev->next = NULL;
drivers/pci/hotplug/ibmphp_res.c
938
range = range->next;
drivers/pci/hotplug/octep_hp.c
369
struct pci_dev *tmp_pdev, *next;
drivers/pci/hotplug/octep_hp.c
389
list_for_each_entry_safe(tmp_pdev, next, &pdev->bus->devices, bus_list) {
drivers/pci/hotplug/pnv_php.c
197
struct device_node *child, *next;
drivers/pci/hotplug/pnv_php.c
207
next = child->sibling;
drivers/pci/hotplug/pnv_php.c
211
child = next;
drivers/pci/hotplug/rpadlpar_core.c
113
struct slot *slot, *next;
drivers/pci/hotplug/rpadlpar_core.c
115
list_for_each_entry_safe(slot, next, &rpaphp_slot_head,
drivers/pci/hotplug/rpaphp_core.c
452
struct slot *slot, *next;
drivers/pci/hotplug/rpaphp_core.c
459
list_for_each_entry_safe(slot, next, &rpaphp_slot_head,
drivers/pci/hotplug/shpchp_core.c
128
struct slot *slot, *next;
drivers/pci/hotplug/shpchp_core.c
130
list_for_each_entry_safe(slot, next, &ctrl->slot_list, slot_list) {
drivers/pci/pci-bridge-emul.c
389
bridge->pcie_conf.next = (bridge->ssid_start > bridge->pcie_start) ?
drivers/pci/pci-bridge-emul.h
45
u8 next;
drivers/pci/pci.c
1617
hlist_for_each_entry(tmp, &pci_dev->saved_cap_space, next) {
drivers/pci/pci.c
1880
hlist_for_each_entry(tmp, &dev->saved_cap_space, next)
drivers/pci/pci.c
1891
hlist_for_each_entry(tmp, &dev->saved_cap_space, next) {
drivers/pci/pci.c
3402
hlist_add_head(&new_cap->next, &pci_dev->saved_cap_space);
drivers/pci/pci.c
3479
hlist_for_each_entry_safe(tmp, n, &dev->saved_cap_space, next)
drivers/pci/pci.h
244
struct hlist_node next;
drivers/pci/pcie/aspm.c
419
child = list_entry(linkbus->devices.next, struct pci_dev, bus_list);
drivers/pci/probe.c
1090
next = list_next_entry(window, node);
drivers/pci/probe.c
1093
next_offset = next->offset;
drivers/pci/probe.c
1094
next_res = next->res;
drivers/pci/probe.c
990
struct resource_entry *window, *next, *n;
drivers/pci/proc.c
412
.next = pci_seq_next,
drivers/pci/quirks.c
5595
hlist_add_head(&state->next, &pdev->saved_cap_space);
drivers/pci/search.c
174
n = from ? from->node.next : pci_root_buses.next;
drivers/pci/switch/switchtec.c
168
stuser = list_entry(stdev->mrpc_queue.next, struct switchtec_user,
drivers/pci/switch/switchtec.c
208
struct switchtec_user *stuser = list_entry(stdev->mrpc_queue.next,
drivers/pci/switch/switchtec.c
229
stuser = list_entry(stdev->mrpc_queue.next, struct switchtec_user,
drivers/pci/switch/switchtec.c
286
stuser = list_entry(stdev->mrpc_queue.next,
drivers/pci/xen-pcifront.c
545
dev = container_of(bus->devices.next, struct pci_dev,
drivers/pcmcia/rsrc_nonstatic.c
1029
data->mem_db.next = &data->mem_db;
drivers/pcmcia/rsrc_nonstatic.c
1030
data->mem_db_valid.next = &data->mem_db_valid;
drivers/pcmcia/rsrc_nonstatic.c
1031
data->io_db.next = &data->io_db;
drivers/pcmcia/rsrc_nonstatic.c
1045
for (p = data->mem_db_valid.next; p != &data->mem_db_valid; p = q) {
drivers/pcmcia/rsrc_nonstatic.c
1046
q = p->next;
drivers/pcmcia/rsrc_nonstatic.c
1049
for (p = data->mem_db.next; p != &data->mem_db; p = q) {
drivers/pcmcia/rsrc_nonstatic.c
1050
q = p->next;
drivers/pcmcia/rsrc_nonstatic.c
1053
for (p = data->io_db.next; p != &data->io_db; p = q) {
drivers/pcmcia/rsrc_nonstatic.c
1054
q = p->next;
drivers/pcmcia/rsrc_nonstatic.c
1085
for (p = data->io_db.next; p != &data->io_db; p = p->next) {
drivers/pcmcia/rsrc_nonstatic.c
112
for (p = map; ; p = p->next) {
drivers/pcmcia/rsrc_nonstatic.c
1141
for (p = data->mem_db_valid.next; p != &data->mem_db_valid;
drivers/pcmcia/rsrc_nonstatic.c
1142
p = p->next) {
drivers/pcmcia/rsrc_nonstatic.c
1151
for (p = data->mem_db.next; p != &data->mem_db; p = p->next) {
drivers/pcmcia/rsrc_nonstatic.c
117
if ((p->next == map) || (p->next->base > base+num-1))
drivers/pcmcia/rsrc_nonstatic.c
126
q->next = p->next; p->next = q;
drivers/pcmcia/rsrc_nonstatic.c
137
q = p->next;
drivers/pcmcia/rsrc_nonstatic.c
144
p->next = q->next;
drivers/pcmcia/rsrc_nonstatic.c
166
p->next = q->next ; q->next = p;
drivers/pcmcia/rsrc_nonstatic.c
461
ok = inv_probe(m->next, s);
drivers/pcmcia/rsrc_nonstatic.c
490
if (inv_probe(s_data->mem_db.next, s) > 0)
drivers/pcmcia/rsrc_nonstatic.c
492
if (s_data->mem_db_valid.next != &s_data->mem_db_valid)
drivers/pcmcia/rsrc_nonstatic.c
499
for (m = s_data->mem_db.next; m != &s_data->mem_db; m = mm.next) {
drivers/pcmcia/rsrc_nonstatic.c
50
struct resource_map *next;
drivers/pcmcia/rsrc_nonstatic.c
543
for (m = s_data->mem_db.next; m != &s_data->mem_db; m = mm.next) {
drivers/pcmcia/rsrc_nonstatic.c
578
if (s_data->mem_db_valid.next != &s_data->mem_db_valid)
drivers/pcmcia/rsrc_nonstatic.c
613
for (m = data->map->next; m != data->map; m = m->next) {
drivers/pcmcia/rsrc_nonstatic.c
657
for (m = s_data->io_db.next; m != &s_data->io_db; m = m->next) {
drivers/pcmcia/sa1111_generic.c
181
s->next = dev_get_drvdata(&dev->dev);
drivers/pcmcia/sa1111_generic.c
234
struct sa1111_pcmcia_socket *next, *s = dev_get_drvdata(&dev->dev);
drivers/pcmcia/sa1111_generic.c
238
for (; s; s = next) {
drivers/pcmcia/sa1111_generic.c
239
next = s->next;
drivers/pcmcia/sa1111_generic.h
8
struct sa1111_pcmcia_socket *next;
drivers/perf/arm-cmn.c
2100
goto next;
drivers/perf/arm-cmn.c
2112
next:
drivers/perf/qcom_l2_pmu.c
136
struct list_head next;
drivers/perf/qcom_l2_pmu.c
752
list_for_each_entry(cluster, &l2cache_pmu->clusters, next) {
drivers/perf/qcom_l2_pmu.c
857
INIT_LIST_HEAD(&cluster->next);
drivers/perf/qcom_l2_pmu.c
883
list_add(&cluster->next, &l2cache_pmu->clusters);
drivers/perf/xgene_pmu.c
1242
list_for_each_entry(ctx, &xgene_pmu->mcpmus, next) {
drivers/perf/xgene_pmu.c
1247
list_for_each_entry(ctx, &xgene_pmu->mcbpmus, next) {
drivers/perf/xgene_pmu.c
1252
list_for_each_entry(ctx, &xgene_pmu->l3cpmus, next) {
drivers/perf/xgene_pmu.c
1257
list_for_each_entry(ctx, &xgene_pmu->iobpmus, next) {
drivers/perf/xgene_pmu.c
138
struct list_head next;
drivers/perf/xgene_pmu.c
1566
list_add(&ctx->next, &xgene_pmu->l3cpmus);
drivers/perf/xgene_pmu.c
1569
list_add(&ctx->next, &xgene_pmu->iobpmus);
drivers/perf/xgene_pmu.c
1572
list_add(&ctx->next, &xgene_pmu->iobpmus);
drivers/perf/xgene_pmu.c
1575
list_add(&ctx->next, &xgene_pmu->mcbpmus);
drivers/perf/xgene_pmu.c
1578
list_add(&ctx->next, &xgene_pmu->mcpmus);
drivers/perf/xgene_pmu.c
1688
list_add(&ctx->next, &xgene_pmu->l3cpmus);
drivers/perf/xgene_pmu.c
1691
list_add(&ctx->next, &xgene_pmu->iobpmus);
drivers/perf/xgene_pmu.c
1694
list_add(&ctx->next, &xgene_pmu->iobpmus);
drivers/perf/xgene_pmu.c
1697
list_add(&ctx->next, &xgene_pmu->mcbpmus);
drivers/perf/xgene_pmu.c
1700
list_add(&ctx->next, &xgene_pmu->mcpmus);
drivers/perf/xgene_pmu.c
1807
list_for_each_entry(ctx, &xgene_pmu->mcpmus, next) {
drivers/perf/xgene_pmu.c
1810
list_for_each_entry(ctx, &xgene_pmu->mcbpmus, next) {
drivers/perf/xgene_pmu.c
1813
list_for_each_entry(ctx, &xgene_pmu->l3cpmus, next) {
drivers/perf/xgene_pmu.c
1816
list_for_each_entry(ctx, &xgene_pmu->iobpmus, next) {
drivers/perf/xgene_pmu.c
1927
list_for_each_entry(ctx, pmus, next) {
drivers/phy/phy-core.c
1158
struct device_node *parent = of_node_get(children), *next;
drivers/phy/phy-core.c
1164
next = of_get_parent(parent);
drivers/phy/phy-core.c
1166
parent = next;
drivers/phy/socionext/phy-uniphier-usb2.c
110
struct uniphier_u2phy_priv *priv = NULL, *next = NULL;
drivers/phy/socionext/phy-uniphier-usb2.c
166
priv->next = next;
drivers/phy/socionext/phy-uniphier-usb2.c
167
next = priv;
drivers/phy/socionext/phy-uniphier-usb2.c
44
struct uniphier_u2phy_priv *next;
drivers/phy/socionext/phy-uniphier-usb2.c
89
priv = priv->next;
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
141
pfd->next = pfd->bitpos + c->x_bits > c->sz_reg ? c->x_addrs : 0;
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
174
mtk_rmw(hw, pf->index, pf->offset + pf->next, BIT(nbits_h) - 1,
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
187
h = (mtk_r32(hw, pf->index, pf->offset + pf->next))
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
206
if (!pf.next)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
226
if (!pf.next)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.h
128
u8 next;
drivers/pinctrl/pinctrl-at91.c
1708
if (!at91_gpio->next)
drivers/pinctrl/pinctrl-at91.c
1710
at91_gpio = at91_gpio->next;
drivers/pinctrl/pinctrl-at91.c
1785
if (prev->next) {
drivers/pinctrl/pinctrl-at91.c
1786
prev = prev->next;
drivers/pinctrl/pinctrl-at91.c
1788
prev->next = at91_gpio;
drivers/pinctrl/pinctrl-at91.c
54
struct at91_gpio_chip *next;
drivers/platform/mellanox/mlxbf-tmfifo.c
384
idx = virtio16_to_cpu(vdev, desc->next);
drivers/platform/mellanox/mlxbf-tmfifo.c
481
idx = virtio16_to_cpu(vdev, desc->next);
drivers/platform/mellanox/mlxbf-tmfifo.c
794
idx = virtio16_to_cpu(vdev, desc->next);
drivers/platform/raspberrypi/vchiq-interface/vchiq_arm.c
385
struct bulk_waiter_node *waiter, *next;
drivers/platform/raspberrypi/vchiq-interface/vchiq_arm.c
387
list_for_each_entry_safe(waiter, next,
drivers/platform/raspberrypi/vchiq-mmal/mmal-msg.h
265
u32 next; /* next header */
drivers/platform/surface/aggregator/ssh_packet_layer.c
1507
ktime_t next = KTIME_MAX;
drivers/platform/surface/aggregator/ssh_packet_layer.c
1531
next = ktime_before(expires, next) ? expires : next;
drivers/platform/surface/aggregator/ssh_packet_layer.c
1591
next = max(next, ktime_add(now, SSH_PTL_PACKET_TIMEOUT_RESOLUTION));
drivers/platform/surface/aggregator/ssh_packet_layer.c
1592
if (next != KTIME_MAX)
drivers/platform/surface/aggregator/ssh_packet_layer.c
1593
ssh_ptl_timeout_reaper_mod(ptl, now, next);
drivers/platform/surface/aggregator/ssh_packet_layer.c
729
head = head->next;
drivers/platform/surface/aggregator/ssh_request_layer.c
827
ktime_t next = KTIME_MAX;
drivers/platform/surface/aggregator/ssh_request_layer.c
848
next = ktime_before(expires, next) ? expires : next;
drivers/platform/surface/aggregator/ssh_request_layer.c
891
next = max(next, ktime_add(now, SSH_RTL_REQUEST_TIMEOUT_RESOLUTION));
drivers/platform/surface/aggregator/ssh_request_layer.c
892
if (next != KTIME_MAX)
drivers/platform/surface/aggregator/ssh_request_layer.c
893
ssh_rtl_timeout_reaper_mod(rtl, now, next);
drivers/platform/x86/dell/dcdbas.c
635
.next = NULL,
drivers/platform/x86/dell/dell-wmi-sysman/sysman.c
350
struct kobject *pos, *next;
drivers/platform/x86/dell/dell-wmi-sysman/sysman.c
352
list_for_each_entry_safe(pos, next, &kset->list, entry) {
drivers/platform/x86/hp/hp-bioscfg/bioscfg.c
537
struct kobject *pos, *next;
drivers/platform/x86/hp/hp-bioscfg/bioscfg.c
539
list_for_each_entry_safe(pos, next, &kset->list, entry)
drivers/platform/x86/tuxedo/nb04/wmi_ab.c
602
union tux_wmi_xx_496in_80out_in_t *next = &driver_data->next_kbl_set_multiple_keys_in;
drivers/platform/x86/tuxedo/nb04/wmi_ab.c
633
rgb_configs_j = &next->kbl_set_multiple_keys_in.rgb_configs[j];
drivers/platform/x86/tuxedo/nb04/wmi_ab.c
639
next->kbl_set_multiple_keys_in.rgb_configs_cnt =
drivers/platform/x86/tuxedo/nb04/wmi_ab.c
666
next, &out);
drivers/platform/x86/tuxedo/nb04/wmi_ab.c
667
memset(next, 0, sizeof(*next));
drivers/pmdomain/core.c
615
void dev_pm_genpd_set_next_wakeup(struct device *dev, ktime_t next)
drivers/pmdomain/core.c
626
td->next_wakeup = next;
drivers/pnp/card.c
124
dev_id->next = NULL;
drivers/pnp/card.c
126
while (ptr && ptr->next)
drivers/pnp/card.c
127
ptr = ptr->next;
drivers/pnp/card.c
129
ptr->next = dev_id;
drivers/pnp/card.c
139
struct pnp_id *next;
drivers/pnp/card.c
143
next = id->next;
drivers/pnp/card.c
145
id = next;
drivers/pnp/card.c
205
pos = pos->next;
drivers/pnp/card.c
310
pos = card->devices.next;
drivers/pnp/card.c
314
pos = from->card_list.next;
drivers/pnp/card.c
320
pos = pos->next;
drivers/pnp/core.c
205
for (id = dev->id; id; id = id->next)
drivers/pnp/core.c
84
struct pnp_id *next;
drivers/pnp/core.c
88
next = id->next;
drivers/pnp/core.c
90
id = next;
drivers/pnp/driver.c
165
pos = pos->next;
drivers/pnp/driver.c
331
dev_id->next = NULL;
drivers/pnp/driver.c
333
while (ptr && ptr->next)
drivers/pnp/driver.c
334
ptr = ptr->next;
drivers/pnp/driver.c
336
ptr->next = dev_id;
drivers/pnp/driver.c
39
pos = pos->next;
drivers/pnp/interface.c
450
pos = pos->next;
drivers/pnp/isapnp/compat.c
35
list = pnp_global.next;
drivers/pnp/isapnp/compat.c
37
list = from->global_list.next;
drivers/pnp/isapnp/compat.c
45
list = list->next;
drivers/pnp/isapnp/compat.c
50
list = card->devices.next;
drivers/pnp/isapnp/compat.c
52
list = from->card_list.next;
drivers/pnp/isapnp/compat.c
61
list = list->next;
drivers/pnp/quirks.c
198
new_option = list_entry(new_option->list.next,
drivers/power/sequencing/core.c
1011
struct device *next __free(put_device) =
drivers/power/sequencing/core.c
1013
return next;
drivers/power/sequencing/core.c
1066
.next = pwrseq_debugfs_seq_next,
drivers/power/sequencing/core.c
158
struct pwrseq_unit_dep *dep, *next;
drivers/power/sequencing/core.c
160
list_for_each_entry_safe(dep, next, list, list) {
drivers/ptp/ptp_ines.c
256
struct list_head *this, *next;
drivers/ptp/ptp_ines.c
266
list_for_each_safe(this, next, &port->events) {
drivers/pwm/core.c
2730
.next = pwm_seq_next,
drivers/rapidio/rio-scan.c
964
if (mport->nnode.next || mport->nnode.prev)
drivers/rapidio/rio.c
846
n = from ? from->global_list.next : rio_devices.next;
drivers/rapidio/rio.c
852
n = n->next;
drivers/ras/amd/fmpm.c
941
.next = fmpm_next,
drivers/remoteproc/qcom_q6v5_mss.c
460
struct qcom_scm_vmperm next[2];
drivers/remoteproc/qcom_q6v5_mss.c
471
next[perms].vmid = QCOM_SCM_VMID_HLOS;
drivers/remoteproc/qcom_q6v5_mss.c
472
next[perms].perm = QCOM_SCM_PERM_RWX;
drivers/remoteproc/qcom_q6v5_mss.c
477
next[perms].vmid = QCOM_SCM_VMID_MSS_MSA;
drivers/remoteproc/qcom_q6v5_mss.c
478
next[perms].perm = QCOM_SCM_PERM_RW;
drivers/remoteproc/qcom_q6v5_mss.c
483
current_perm, next, perms);
drivers/rpmsg/mtk_rpmsg.c
341
struct mtk_rpmsg_channel_info *info, *next;
drivers/rpmsg/mtk_rpmsg.c
369
list_for_each_entry_safe(info, next,
drivers/rtc/interface.c
832
struct timerqueue_node *next = timerqueue_getnext(&rtc->timerqueue);
drivers/rtc/interface.c
845
while (next) {
drivers/rtc/interface.c
846
if (next->expires >= now)
drivers/rtc/interface.c
848
next = timerqueue_iterate_next(next);
drivers/rtc/interface.c
853
if (!next || ktime_before(timer->node.expires, next->expires)) {
drivers/rtc/interface.c
895
struct timerqueue_node *next = timerqueue_getnext(&rtc->timerqueue);
drivers/rtc/interface.c
900
if (next == &timer->node) {
drivers/rtc/interface.c
904
next = timerqueue_getnext(&rtc->timerqueue);
drivers/rtc/interface.c
905
if (!next) {
drivers/rtc/interface.c
909
alarm.time = rtc_ktime_to_tm(next->expires);
drivers/rtc/interface.c
931
struct timerqueue_node *next;
drivers/rtc/interface.c
947
while ((next = timerqueue_getnext(&rtc->timerqueue))) {
drivers/rtc/interface.c
948
if (next->expires > now)
drivers/rtc/interface.c
952
timer = container_of(next, struct rtc_timer, node);
drivers/rtc/interface.c
971
if (next) {
drivers/rtc/interface.c
976
alarm.time = rtc_ktime_to_tm(next->expires);
drivers/rtc/interface.c
986
timer = container_of(next, struct rtc_timer, node);
drivers/rtc/rtc-88pm80x.c
76
static void rtc_next_alarm_time(struct rtc_time *next, struct rtc_time *now,
drivers/rtc/rtc-88pm80x.c
82
next->tm_year = now->tm_year;
drivers/rtc/rtc-88pm80x.c
83
next->tm_mon = now->tm_mon;
drivers/rtc/rtc-88pm80x.c
84
next->tm_mday = now->tm_mday;
drivers/rtc/rtc-88pm80x.c
85
next->tm_hour = alrm->tm_hour;
drivers/rtc/rtc-88pm80x.c
86
next->tm_min = alrm->tm_min;
drivers/rtc/rtc-88pm80x.c
87
next->tm_sec = alrm->tm_sec;
drivers/rtc/rtc-88pm80x.c
90
next_time = rtc_tm_to_time64(next);
drivers/rtc/rtc-88pm80x.c
95
rtc_time64_to_tm(next_time, next);
drivers/s390/block/dasd.c
1587
struct dasd_ccw_req *cqr, *next, *fcqr;
drivers/s390/block/dasd.c
1746
next = NULL;
drivers/s390/block/dasd.c
1754
if (cqr->devlist.next != &device->ccw_queue) {
drivers/s390/block/dasd.c
1755
next = list_entry(cqr->devlist.next,
drivers/s390/block/dasd.c
1780
next = cqr;
drivers/s390/block/dasd.c
1784
if (next && (next->status == DASD_CQR_QUEUED) &&
drivers/s390/block/dasd.c
1786
if (device->discipline->start_IO(next) == 0)
drivers/s390/block/dasd.c
1787
expires = next->expires;
drivers/s390/block/dasd.c
1939
cqr = list_entry(device->ccw_queue.next, struct dasd_ccw_req, devlist);
drivers/s390/block/dasd.c
2007
cqr = list_entry(device->ccw_queue.next, struct dasd_ccw_req, devlist);
drivers/s390/block/dasd.c
2513
cqr = list_entry(device->ccw_queue.next, struct dasd_ccw_req, devlist);
drivers/s390/block/dasd.c
2550
list_add(&cqr->devlist, device->ccw_queue.next);
drivers/s390/block/dasd_alias.c
365
if (group->next == device)
drivers/s390/block/dasd_alias.c
366
group->next = NULL;
drivers/s390/block/dasd_alias.c
696
alias_device = group->next;
drivers/s390/block/dasd_alias.c
708
group->next = list_first_entry(&group->aliaslist,
drivers/s390/block/dasd_alias.c
711
group->next = list_first_entry(&alias_device->alias_list,
drivers/s390/block/dasd_diag.c
236
struct dasd_ccw_req *cqr, *next;
drivers/s390/block/dasd_diag.c
288
next = list_entry(device->ccw_queue.next,
drivers/s390/block/dasd_diag.c
290
if (next->status == DASD_CQR_QUEUED) {
drivers/s390/block/dasd_diag.c
291
rc = dasd_start_diag(next);
drivers/s390/block/dasd_diag.c
293
expires = next->expires;
drivers/s390/block/dasd_eckd.h
659
struct dasd_device *next;
drivers/s390/block/dasd_int.h
777
if (left->next != chunk_list) {
drivers/s390/block/dasd_int.h
778
tmp = list_entry(left->next, struct dasd_mchunk, list);
drivers/s390/block/dasd_int.h
792
__list_add(&chunk->list, left, left->next);
drivers/s390/block/dasd_proc.c
124
.next = dasd_devices_next,
drivers/s390/char/con3215.c
117
raw3215_freelist = req->next;
drivers/s390/char/con3215.c
133
req->next = raw3215_freelist;
drivers/s390/char/con3215.c
75
struct raw3215_req *next; /* pointer to next request */
drivers/s390/char/con3215.c
922
req->next = raw3215_freelist;
drivers/s390/char/monwriter.c
204
struct mon_buf *entry, *next;
drivers/s390/char/monwriter.c
206
list_for_each_entry_safe(entry, next, &monpriv->list, list) {
drivers/s390/char/monwriter.c
87
struct mon_buf *entry, *next;
drivers/s390/char/monwriter.c
89
list_for_each_entry_safe(entry, next, &monpriv->list, list)
drivers/s390/char/raw3270.c
1254
v = list_entry(rp->view_list.next, struct raw3270_view, list);
drivers/s390/char/raw3270.c
392
rq = list_entry(rp->req_queue.next, struct raw3270_request, list);
drivers/s390/char/raw3270.c
688
rq = list_entry(rp->req_queue.next, struct raw3270_request, list);
drivers/s390/char/sclp.c
455
req = list_entry(sclp_req_queue.next, struct sclp_req, list);
drivers/s390/char/sclp_con.c
143
list = sclp_con_outqueue.next;
drivers/s390/char/sclp_con.c
146
list = list->next;
drivers/s390/char/sclp_con.c
186
page = sclp_con_pages.next;
drivers/s390/char/sclp_tty.c
122
buffer = list_entry(sclp_tty_outqueue.next,
drivers/s390/char/sclp_tty.c
194
page = sclp_tty_pages.next;
drivers/s390/char/sclp_vt220.c
131
request = list_entry(sclp_vt220_outqueue.next,
drivers/s390/char/sclp_vt220.c
373
list = sclp_vt220_outqueue.next;
drivers/s390/char/sclp_vt220.c
376
list = list->next;
drivers/s390/char/sclp_vt220.c
426
page = (void *) sclp_vt220_empty.next;
drivers/s390/char/tape_core.c
1137
req = list_entry(device->req_queue.next,
drivers/s390/char/tape_core.c
145
req = list_entry(tdev->req_queue.next, struct tape_request,
drivers/s390/char/tape_core.c
857
request = list_entry(device->req_queue.next, struct tape_request, list);
drivers/s390/char/tape_proc.c
102
.next = tape_proc_next,
drivers/s390/char/tape_proc.c
71
request = list_entry(device->req_queue.next,
drivers/s390/cio/blacklist.c
396
.next = cio_ignore_proc_seq_next,
drivers/s390/cio/vfio_ccw_chp.c
94
struct vfio_ccw_crw, next);
drivers/s390/cio/vfio_ccw_chp.c
97
list_del(&crw->next);
drivers/s390/cio/vfio_ccw_cp.c
32
struct list_head next;
drivers/s390/cio/vfio_ccw_cp.c
334
list_add_tail(&chain->next, &cp->ccwchain_list);
drivers/s390/cio/vfio_ccw_cp.c
346
list_del(&chain->next);
drivers/s390/cio/vfio_ccw_cp.c
409
list_for_each_entry(chain, &cp->ccwchain_list, next) {
drivers/s390/cio/vfio_ccw_cp.c
495
list_for_each_entry(iter, &cp->ccwchain_list, next) {
drivers/s390/cio/vfio_ccw_cp.c
766
list_for_each_entry_safe(chain, temp, &cp->ccwchain_list, next) {
drivers/s390/cio/vfio_ccw_cp.c
823
list_for_each_entry(chain, &cp->ccwchain_list, next) {
drivers/s390/cio/vfio_ccw_cp.c
839
list_for_each_entry_continue(chain, &cp->ccwchain_list, next) {
drivers/s390/cio/vfio_ccw_cp.c
881
chain = list_first_entry(&cp->ccwchain_list, struct ccwchain, next);
drivers/s390/cio/vfio_ccw_cp.c
917
list_for_each_entry(chain, &cp->ccwchain_list, next) {
drivers/s390/cio/vfio_ccw_cp.c
956
list_for_each_entry(chain, &cp->ccwchain_list, next) {
drivers/s390/cio/vfio_ccw_drv.c
295
list_add_tail(&crw->next, &private->crw);
drivers/s390/cio/vfio_ccw_ops.c
134
list_for_each_entry_safe(crw, temp, &private->crw, next) {
drivers/s390/cio/vfio_ccw_ops.c
135
list_del(&crw->next);
drivers/s390/cio/vfio_ccw_private.h
66
struct list_head next;
drivers/s390/crypto/ap_queue.c
1329
struct ap_message *ap_msg, *next;
drivers/s390/crypto/ap_queue.c
1331
list_for_each_entry_safe(ap_msg, next, &aq->pendingq, list) {
drivers/s390/crypto/ap_queue.c
1337
list_for_each_entry_safe(ap_msg, next, &aq->requestq, list) {
drivers/s390/crypto/ap_queue.c
272
ap_msg = list_entry(aq->requestq.next, struct ap_message, list);
drivers/s390/net/ctcm_main.c
1416
c = &(*c)->next;
drivers/s390/net/ctcm_main.c
1436
ch->next = *c;
drivers/s390/net/ctcm_main.c
213
*c = ch->next;
drivers/s390/net/ctcm_main.c
235
c = &((*c)->next);
drivers/s390/net/ctcm_main.c
257
ch = ch->next;
drivers/s390/net/ctcm_main.h
133
struct channel *next;
drivers/s390/net/qeth_l3_main.c
1107
im6 = rtnl_dereference(im6->next)) {
drivers/s390/scsi/zfcp_fc.c
64
unsigned long next = adapter->next_port_scan;
drivers/s390/scsi/zfcp_fc.c
68
if (time_before(now, next)) {
drivers/s390/scsi/zfcp_fc.c
69
delay = next - now;
drivers/sbus/char/bbc_i2c.h
54
struct bbc_i2c_bus *next;
drivers/sbus/char/openprom.c
163
!prop->next ||
drivers/sbus/char/openprom.c
164
(len = strlen(prop->next->name)) + 1 > bufsize)
drivers/sbus/char/openprom.c
167
prop = prop->next;
drivers/sbus/char/openprom.c
441
prop = prop->next;
drivers/scsi/BusLogic.c
177
ccb->next = adapter->free_ccbs;
drivers/scsi/BusLogic.c
2450
ccb->next = NULL;
drivers/scsi/BusLogic.c
2455
adapter->lastccb->next = ccb;
drivers/scsi/BusLogic.c
2607
adapter->firstccb = ccb->next;
drivers/scsi/BusLogic.c
286
adapter->free_ccbs = ccb->next;
drivers/scsi/BusLogic.c
287
ccb->next = NULL;
drivers/scsi/BusLogic.c
298
adapter->free_ccbs = ccb->next;
drivers/scsi/BusLogic.c
299
ccb->next = NULL;
drivers/scsi/BusLogic.c
322
ccb->next = adapter->free_ccbs;
drivers/scsi/BusLogic.c
3711
struct blogic_adapter *ha, *next;
drivers/scsi/BusLogic.c
3713
list_for_each_entry_safe(ha, next, &blogic_host_list, host_list)
drivers/scsi/BusLogic.h
847
struct blogic_ccb *next;
drivers/scsi/aacraid/aachba.c
3884
psg->sg[0].next = 0;
drivers/scsi/aacraid/aachba.c
3898
psg->sg[i].next = 0;
drivers/scsi/aacraid/aacraid.h
1227
struct list_head next; // used to link context's into a linked list
drivers/scsi/aacraid/aacraid.h
1289
void *next; /* this is used by the allocator */
drivers/scsi/aacraid/aacraid.h
478
__le32 next; /* reserved for F/W use */
drivers/scsi/aacraid/aacraid.h
486
u32 next; /* reserved for F/W use */
drivers/scsi/aacraid/commctrl.c
207
entry = dev->fib_list.next;
drivers/scsi/aacraid/commctrl.c
209
context = list_entry(entry, struct aac_fib_context, next);
drivers/scsi/aacraid/commctrl.c
213
entry = dev->fib_list.next;
drivers/scsi/aacraid/commctrl.c
215
entry = entry->next;
drivers/scsi/aacraid/commctrl.c
218
list_add_tail(&fibctx->next, &dev->fib_list);
drivers/scsi/aacraid/commctrl.c
273
entry = dev->fib_list.next;
drivers/scsi/aacraid/commctrl.c
277
fibctx = list_entry(entry, struct aac_fib_context, next);
drivers/scsi/aacraid/commctrl.c
284
entry = entry->next;
drivers/scsi/aacraid/commctrl.c
309
entry = fibctx->fib_list.next;
drivers/scsi/aacraid/commctrl.c
367
entry = fibctx->fib_list.next;
drivers/scsi/aacraid/commctrl.c
380
list_del(&fibctx->next);
drivers/scsi/aacraid/commctrl.c
414
entry = dev->fib_list.next;
drivers/scsi/aacraid/commctrl.c
418
fibctx = list_entry(entry, struct aac_fib_context, next);
drivers/scsi/aacraid/commctrl.c
424
entry = entry->next;
drivers/scsi/aacraid/commsup.c
168
fibptr->next = fibptr+1; /* Forward chain the fibs */
drivers/scsi/aacraid/commsup.c
1890
entry = dev->fib_list.next;
drivers/scsi/aacraid/commsup.c
1892
entry = entry->next;
drivers/scsi/aacraid/commsup.c
1948
entry = dev->fib_list.next;
drivers/scsi/aacraid/commsup.c
1963
next);
drivers/scsi/aacraid/commsup.c
198
dev->fibs[dev->scsi_host_ptr->can_queue + AAC_NUM_MGT_FIB - 1].next = NULL;
drivers/scsi/aacraid/commsup.c
1981
entry = entry->next;
drivers/scsi/aacraid/commsup.c
1992
entry = entry->next;
drivers/scsi/aacraid/commsup.c
2018
entry = entry->next;
drivers/scsi/aacraid/commsup.c
2048
entry = dev->queues->queue[HostNormCmdQueue].cmdq.next;
drivers/scsi/aacraid/commsup.c
251
dev->free_fib = fibptr->next;
drivers/scsi/aacraid/commsup.c
293
fibptr->next = fibptr->dev->free_fib;
drivers/scsi/aacraid/src.c
101
entry = dev->sync_fib_list.next;
drivers/scsi/aha152x.c
1136
struct scsi_cmnd *next;
drivers/scsi/aha152x.c
1139
next = SCNEXT(ptr);
drivers/scsi/aha152x.c
1143
next = NULL;
drivers/scsi/aha152x.c
1153
ptr = next;
drivers/scsi/aha152x.c
2297
struct scsi_cmnd *next = SCNEXT(ptr);
drivers/scsi/aha152x.c
2309
ptr = next;
drivers/scsi/aha152x.c
520
struct scsi_cmnd *next; /* next sc in queue */
drivers/scsi/aha152x.c
572
#define SCNEXT(SCpnt) SCDATA(SCpnt)->next
drivers/scsi/aic7xxx/aic79xx_core.c
200
u_int prev, u_int next, u_int tid);
drivers/scsi/aic7xxx/aic79xx_core.c
2896
u_int next;
drivers/scsi/aic7xxx/aic79xx_core.c
2948
next = SCB_LIST_NULL;
drivers/scsi/aic7xxx/aic79xx_core.c
2951
next = ahd_inw_scbram(ahd, SCB_NEXT2);
drivers/scsi/aic7xxx/aic79xx_core.c
2954
ahd_outw(ahd, SCB_NEXT2, next);
drivers/scsi/aic7xxx/aic79xx_core.c
8337
u_int next;
drivers/scsi/aic7xxx/aic79xx_core.c
8344
next = *list_head;
drivers/scsi/aic7xxx/aic79xx_core.c
8346
for (scbid = next; !SCBID_IS_NULL(scbid); scbid = next) {
drivers/scsi/aic7xxx/aic79xx_core.c
8362
next = ahd_inw_scbram(ahd, SCB_NEXT);
drivers/scsi/aic7xxx/aic79xx_core.c
8376
ahd_rem_wscb(ahd, scbid, prev, next, tid);
drivers/scsi/aic7xxx/aic79xx_core.c
8379
*list_head = next;
drivers/scsi/aic7xxx/aic79xx_core.c
8437
u_int prev, u_int next, u_int tid)
drivers/scsi/aic7xxx/aic79xx_core.c
8444
ahd_outw(ahd, SCB_NEXT, next);
drivers/scsi/aic7xxx/aic79xx_core.c
8455
if (SCBID_IS_NULL(next)
drivers/scsi/aic7xxx/aic79xx_core.c
8460
return (next);
drivers/scsi/aic7xxx/aic7xxx.h
472
/*31*/ uint8_t next; /*
drivers/scsi/aic7xxx/aic7xxx_core.c
5773
prev_scb->hscb->next = scb->hscb->tag;
drivers/scsi/aic7xxx/aic7xxx_core.c
5778
scb->hscb->next = ahc->next_queued_scb->hscb->tag;
drivers/scsi/aic7xxx/aic7xxx_core.c
5807
uint8_t next;
drivers/scsi/aic7xxx/aic7xxx_core.c
5917
next = scb->hscb->next;
drivers/scsi/aic7xxx/aic7xxx_core.c
5920
scb->hscb->next = next;
drivers/scsi/aic7xxx/aic7xxx_core.c
5929
scb->hscb->next = ahc->next_queued_scb->hscb->tag;
drivers/scsi/aic7xxx/aic7xxx_core.c
593
q_hscb->next = scb->hscb->tag;
drivers/scsi/aic7xxx/aic7xxx_core.c
5936
next = ahc_inb(ahc, WAITING_SCBH); /* Start at head of list. */
drivers/scsi/aic7xxx/aic7xxx_core.c
5939
while (next != SCB_LIST_NULL) {
drivers/scsi/aic7xxx/aic7xxx_core.c
5942
ahc_outb(ahc, SCBPTR, next);
drivers/scsi/aic7xxx/aic7xxx_core.c
5954
scb_index, next);
drivers/scsi/aic7xxx/aic7xxx_core.c
5982
next = ahc_rem_wscb(ahc, next, prev);
drivers/scsi/aic7xxx/aic7xxx_core.c
5985
prev = next;
drivers/scsi/aic7xxx/aic7xxx_core.c
5986
next = ahc_inb(ahc, SCB_NEXT);
drivers/scsi/aic7xxx/aic7xxx_core.c
5990
prev = next;
drivers/scsi/aic7xxx/aic7xxx_core.c
5991
next = ahc_inb(ahc, SCB_NEXT);
drivers/scsi/aic7xxx/aic7xxx_core.c
6108
u_int next;
drivers/scsi/aic7xxx/aic7xxx_core.c
6114
next = ahc_inb(ahc, DISCONNECTED_SCBH);
drivers/scsi/aic7xxx/aic7xxx_core.c
612
|| scb->hscb->next == SCB_LIST_NULL)
drivers/scsi/aic7xxx/aic7xxx_core.c
6124
while (next != SCB_LIST_NULL) {
drivers/scsi/aic7xxx/aic7xxx_core.c
6127
ahc_outb(ahc, SCBPTR, next);
drivers/scsi/aic7xxx/aic7xxx_core.c
6137
if (next == prev) {
drivers/scsi/aic7xxx/aic7xxx_core.c
614
scb->hscb->tag, scb->hscb->next);
drivers/scsi/aic7xxx/aic7xxx_core.c
6140
next, prev);
drivers/scsi/aic7xxx/aic7xxx_core.c
6147
next =
drivers/scsi/aic7xxx/aic7xxx_core.c
6148
ahc_rem_scb_from_disc_list(ahc, prev, next);
drivers/scsi/aic7xxx/aic7xxx_core.c
6150
prev = next;
drivers/scsi/aic7xxx/aic7xxx_core.c
6151
next = ahc_inb(ahc, SCB_NEXT);
drivers/scsi/aic7xxx/aic7xxx_core.c
6156
prev = next;
drivers/scsi/aic7xxx/aic7xxx_core.c
6157
next = ahc_inb(ahc, SCB_NEXT);
drivers/scsi/aic7xxx/aic7xxx_core.c
6172
u_int next;
drivers/scsi/aic7xxx/aic7xxx_core.c
6175
next = ahc_inb(ahc, SCB_NEXT);
drivers/scsi/aic7xxx/aic7xxx_core.c
6183
ahc_outb(ahc, SCB_NEXT, next);
drivers/scsi/aic7xxx/aic7xxx_core.c
6185
ahc_outb(ahc, DISCONNECTED_SCBH, next);
drivers/scsi/aic7xxx/aic7xxx_core.c
6187
return (next);
drivers/scsi/aic7xxx/aic7xxx_core.c
6217
u_int curscb, next;
drivers/scsi/aic7xxx/aic7xxx_core.c
6225
next = ahc_inb(ahc, SCB_NEXT);
drivers/scsi/aic7xxx/aic7xxx_core.c
6235
ahc_outb(ahc, WAITING_SCBH, next);
drivers/scsi/aic7xxx/aic7xxx_core.c
6248
ahc_outb(ahc, SCB_NEXT, next);
drivers/scsi/aic7xxx/aic7xxx_core.c
6255
return next;
drivers/scsi/aic94xx/aic94xx_hwi.c
1359
ascb = list_entry(ascb->list.next, struct asd_ascb, list);
drivers/scsi/aic94xx/aic94xx_hwi.c
320
for (i = 0; i < seq->num_escbs; i++, escb = list_entry(escb->list.next,
drivers/scsi/aic94xx/aic94xx_sds.c
427
u16 next;
drivers/scsi/aic94xx/aic94xx_sds.c
511
__le16 next;
drivers/scsi/aic94xx/aic94xx_sds.c
537
u16 next;
drivers/scsi/aic94xx/aic94xx_sds.c
546
__le16 next;
drivers/scsi/aic94xx/aic94xx_sds.c
730
el = start + le16_to_cpu(el->next);
drivers/scsi/arcmsr/arcmsr_hba.c
3256
ccb = list_entry(head->next, struct CommandControlBlock, list);
drivers/scsi/arm/msgqueue.c
102
for (mq = msgq->qe; mq && msgno; mq = mq->next, msgno--);
drivers/scsi/arm/msgqueue.c
131
mq->next = NULL;
drivers/scsi/arm/msgqueue.c
135
mqp = &(*mqp)->next;
drivers/scsi/arm/msgqueue.c
153
mqnext = mq->next;
drivers/scsi/arm/msgqueue.c
27
msgq->free = mq->next;
drivers/scsi/arm/msgqueue.c
41
mq->next = msgq->free;
drivers/scsi/arm/msgqueue.c
59
msgq->entries[i].next = &msgq->entries[i + 1];
drivers/scsi/arm/msgqueue.c
61
msgq->entries[NR_MESSAGES - 1].next = NULL;
drivers/scsi/arm/msgqueue.c
85
for (mq = msgq->qe; mq; mq = mq->next)
drivers/scsi/arm/msgqueue.h
20
struct msgqueue_entry *next;
drivers/scsi/arm/queue.c
118
l = queue->free.next;
drivers/scsi/arm/queue.c
195
SCpnt = __queue_remove(queue, queue->head.next);
drivers/scsi/bfa/bfa_cs.h
116
#define bfa_q_first(_q) ((void *)(((struct list_head *) (_q))->next))
drivers/scsi/bfa/bfa_cs.h
117
#define bfa_q_next(_qe) (((struct list_head *) (_qe))->next)
drivers/scsi/bnx2fc/57xx_hsi_bnx2fc.h
983
struct regpair next;
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
154
struct sk_buff *skb, *next;
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
159
skb_queue_walk_safe(list, skb, next) {
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
2746
struct bnx2fc_hba *hba, *next;
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
2763
list_for_each_entry_safe(hba, next, &to_be_deleted, list) {
drivers/scsi/bnx2fc/bnx2fc_hwi.c
2146
hba->t2_hash_tbl[i].next.lo = addr & 0xffffffff;
drivers/scsi/bnx2fc/bnx2fc_hwi.c
2147
hba->t2_hash_tbl[i].next.hi = addr >> 32;
drivers/scsi/bnx2fc/bnx2fc_io.c
436
cmd_mgr->free_list[index].next;
drivers/scsi/bnx2fc/bnx2fc_io.c
490
cmd_mgr->free_list[index].next;
drivers/scsi/bnx2fc/bnx2fc_tgt.c
616
u32 conn_id, next;
drivers/scsi/bnx2fc/bnx2fc_tgt.c
627
next = hba->next_conn_id;
drivers/scsi/bnx2fc/bnx2fc_tgt.c
637
if (conn_id == next) {
drivers/scsi/bnx2i/bnx2i_init.c
526
hba = list_entry(adapter_list.next, struct bnx2i_hba, link);
drivers/scsi/csiostor/csio_defs.h
69
return ((list->next == list) && (list->prev == list));
drivers/scsi/csiostor/csio_defs.h
72
#define csio_list_next(elem) (((struct list_head *)(elem))->next)
drivers/scsi/csiostor/csio_lnode.c
938
struct list_head *tmp, *next;
drivers/scsi/csiostor/csio_lnode.c
941
list_for_each_safe(tmp, next, &rnhead->sm.sm_list) {
drivers/scsi/csiostor/csio_scsi.c
1163
struct list_head *tmp, *next;
drivers/scsi/csiostor/csio_scsi.c
1167
list_for_each_safe(tmp, next, q) {
drivers/scsi/csiostor/csio_scsi.c
122
struct list_head *tmp, *next;
drivers/scsi/csiostor/csio_scsi.c
1236
struct list_head *tmp, *next;
drivers/scsi/csiostor/csio_scsi.c
1246
list_for_each_safe(tmp, next, q) {
drivers/scsi/csiostor/csio_scsi.c
133
list_for_each_safe(tmp, next, &scm->active_q) {
drivers/scsi/cxgbi/libcxgbi.c
472
start = idx = pmap->next;
drivers/scsi/cxgbi/libcxgbi.c
479
pmap->next = idx;
drivers/scsi/cxgbi/libcxgbi.c
487
pmap->sport_base + idx, pmap->next);
drivers/scsi/cxgbi/libcxgbi.c
496
pmap->next);
drivers/scsi/cxgbi/libcxgbi.h
465
unsigned int next;
drivers/scsi/dc395x.c
625
struct DeviceCtlBlk* next = NULL;
drivers/scsi/dc395x.c
634
next = i;
drivers/scsi/dc395x.c
640
if (!next)
drivers/scsi/dc395x.c
642
next = i;
drivers/scsi/dc395x.c
646
return next;
drivers/scsi/dc395x.c
714
start = list_entry(dcb_list_head->next, typeof(*start), list);
drivers/scsi/dc395x.c
736
srb = list_entry(waiting_list_head->next,
drivers/scsi/elx/efct/efct_hw.c
2014
if (io_to_abort->wqe.list_entry.next) {
drivers/scsi/elx/efct/efct_hw_queues.c
476
if (!hw->eq_list.next)
drivers/scsi/elx/efct/efct_lio.c
1035
struct efct_lio_vport_list_t *vport, *next;
drivers/scsi/elx/efct/efct_lio.c
1042
list_for_each_entry_safe(vport, next, &efct->tgt_efct.vport_list,
drivers/scsi/elx/libefc/efc_node.c
349
struct efc_hw_sequence *frame, *next;
drivers/scsi/elx/libefc/efc_node.c
354
list_for_each_entry_safe(frame, next, &node->pend_frames, list_entry) {
drivers/scsi/elx/libefc/efc_nport.c
625
struct efc_vport *next;
drivers/scsi/elx/libefc/efc_nport.c
631
list_for_each_entry_safe(vport, next, &efc->vport_list, list_entry) {
drivers/scsi/elx/libefc/efc_nport.c
686
struct efc_vport *next;
drivers/scsi/elx/libefc/efc_nport.c
691
list_for_each_entry_safe(vport, next, &efc->vport_list, list_entry) {
drivers/scsi/elx/libefc/efc_nport.c
723
struct efc_vport *next;
drivers/scsi/elx/libefc/efc_nport.c
727
list_for_each_entry_safe(vport, next, &efc->vport_list, list_entry) {
drivers/scsi/elx/libefc/efclib.c
62
struct efc_hw_sequence *frame, *next;
drivers/scsi/elx/libefc/efclib.c
67
list_for_each_entry_safe(frame, next, &efc->pend_frames, list_entry) {
drivers/scsi/esas2r/esas2r.h
1153
sgl = a->free_sg_list_head.next;
drivers/scsi/esas2r/esas2r.h
1405
struct list_head *element, *next;
drivers/scsi/esas2r/esas2r.h
1407
list_for_each_safe(element, next, comp_list) {
drivers/scsi/esas2r/esas2r_init.c
436
struct esas2r_mem_desc *memdesc, *next;
drivers/scsi/esas2r/esas2r_init.c
511
list_for_each_entry_safe(memdesc, next, &a->free_sg_list_head,
drivers/scsi/esas2r/esas2r_init.c
517
list_for_each_entry_safe(memdesc, next, &a->vrq_mds_head, next_desc) {
drivers/scsi/esas2r/esas2r_int.c
342
struct list_head *element, *next;
drivers/scsi/esas2r/esas2r_int.c
346
list_for_each_safe(element, next, &a->defer_list) {
drivers/scsi/esas2r/esas2r_io.c
771
struct list_head *next, *element;
drivers/scsi/esas2r/esas2r_io.c
782
list_for_each_safe(element, next, &a->defer_list) {
drivers/scsi/esas2r/esas2r_io.c
815
list_for_each_safe(element, next, &a->active_list) {
drivers/scsi/esas2r/esas2r_main.c
913
struct list_head *element, *next;
drivers/scsi/esas2r/esas2r_main.c
915
list_for_each_safe(element, next, queue) {
drivers/scsi/esp_scsi.c
886
ret = list_entry(head->next, struct esp_cmd_entry, list);
drivers/scsi/fcoe/fcoe_ctlr.c
273
struct fcoe_fcf *next;
drivers/scsi/fcoe/fcoe_ctlr.c
276
list_for_each_entry_safe(fcf, next, &fip->fcfs, list) {
drivers/scsi/fcoe/fcoe_ctlr.c
822
struct fcoe_fcf *next;
drivers/scsi/fcoe/fcoe_ctlr.c
830
list_for_each_entry_safe(fcf, next, &fip->fcfs, list) {
drivers/scsi/fcoe/fcoe_ctlr.c
867
list_for_each_entry_safe(fcf, next, &del_list, list) {
drivers/scsi/fcoe/fcoe_sysfs.c
850
struct fcoe_fcf_device *fcf, *next;
drivers/scsi/fcoe/fcoe_sysfs.c
853
list_for_each_entry_safe(fcf, next,
drivers/scsi/fnic/fdls_disc.c
1656
struct fnic_tport_s *tport, *next;
drivers/scsi/fnic/fdls_disc.c
1663
list_for_each_entry_safe(tport, next, &iport->tport_list, links) {
drivers/scsi/fnic/fdls_disc.c
1760
struct fnic_tport_s *tport, *next;
drivers/scsi/fnic/fdls_disc.c
1762
list_for_each_entry_safe(tport, next, &(iport->tport_list), links) {
drivers/scsi/fnic/fdls_disc.c
1810
struct fnic_tport_s *tport, *next;
drivers/scsi/fnic/fdls_disc.c
1812
list_for_each_entry_safe(tport, next, &(iport->tport_list), links) {
drivers/scsi/fnic/fdls_disc.c
194
struct reclaim_entry_s *reclaim_entry, *next;
drivers/scsi/fnic/fdls_disc.c
206
list_for_each_entry_safe(reclaim_entry, next,
drivers/scsi/fnic/fdls_disc.c
3167
struct fnic_tport_s *tport, *next;
drivers/scsi/fnic/fdls_disc.c
3242
list_for_each_entry_safe(tport, next, &iport->tport_list, links) {
drivers/scsi/fnic/fdls_disc.c
3269
struct fnic_tport_s *tport, *next;
drivers/scsi/fnic/fdls_disc.c
3369
list_for_each_entry_safe(tport, next, &iport->tport_list,
drivers/scsi/fnic/fdls_disc.c
4477
struct fnic_tport_s *tport, *next;
drivers/scsi/fnic/fdls_disc.c
4519
list_for_each_entry_safe(tport, next, &iport->tport_list, links) {
drivers/scsi/fnic/fdls_disc.c
4549
list_for_each_entry_safe(tport, next, &iport->tport_list,
drivers/scsi/fnic/fdls_disc.c
5065
struct fnic_tport_s *tport, *next;
drivers/scsi/fnic/fdls_disc.c
5077
list_for_each_entry_safe(tport, next, &iport->tport_list, links) {
drivers/scsi/fnic/fip.c
19
struct fcoe_vlan *vlan, *next;
drivers/scsi/fnic/fip.c
23
list_for_each_entry_safe(vlan, next, &fnic->vlan_list, list) {
drivers/scsi/fnic/fnic_fcs.c
1015
struct fnic_tport_event_s *cur_evt, *next;
drivers/scsi/fnic/fnic_fcs.c
1020
list_for_each_entry_safe(cur_evt, next, &fnic->tport_event_list, links) {
drivers/scsi/fnic/fnic_fcs.c
1065
struct fnic_tport_event_s *cur_evt, *next;
drivers/scsi/fnic/fnic_fcs.c
1069
list_for_each_entry_safe(cur_evt, next, &fnic->tport_event_list, links) {
drivers/scsi/fnic/fnic_fcs.c
286
struct fnic_frame_list *cur_frame, *next;
drivers/scsi/fnic/fnic_fcs.c
290
list_for_each_entry_safe(cur_frame, next, &fnic->frame_queue, links) {
drivers/scsi/fnic/fnic_fcs.c
328
struct fnic_frame_list *cur_frame, *next;
drivers/scsi/fnic/fnic_fcs.c
335
list_for_each_entry_safe(cur_frame, next, &fnic->fip_frame_queue,
drivers/scsi/fnic/fnic_fcs.c
779
struct fnic_frame_list *cur_frame, *next;
drivers/scsi/fnic/fnic_fcs.c
784
list_for_each_entry_safe(cur_frame, next, &fnic->tx_queue, links) {
drivers/scsi/fnic/fnic_fcs.c
841
struct fnic_frame_list *cur_frame, *next;
drivers/scsi/fnic/fnic_fcs.c
843
list_for_each_entry_safe(cur_frame, next, head, links) {
drivers/scsi/fnic/fnic_fcs.c
992
struct fnic_tport_s *tport, *next;
drivers/scsi/fnic/fnic_fcs.c
996
list_for_each_entry_safe(tport, next, &fnic->iport.tport_list, links) {
drivers/scsi/fnic/fnic_trace.c
468
struct fnic_tport_s *tport, *next;
drivers/scsi/fnic/fnic_trace.c
506
list_for_each_entry_safe(tport, next, &iport->tport_list, links) {
drivers/scsi/fnic/vnic_rq.c
167
buf = rq->to_clean = buf->next;
drivers/scsi/fnic/vnic_rq.c
36
buf->next = rq->bufs[0];
drivers/scsi/fnic/vnic_rq.c
39
buf->next = rq->bufs[i + 1];
drivers/scsi/fnic/vnic_rq.c
41
buf->next = buf + 1;
drivers/scsi/fnic/vnic_rq.h
131
buf = buf->next;
drivers/scsi/fnic/vnic_rq.h
187
rq->to_clean = buf->next;
drivers/scsi/fnic/vnic_rq.h
71
struct vnic_rq_buf *next;
drivers/scsi/fnic/vnic_wq.c
223
buf = wq->to_clean = buf->next;
drivers/scsi/fnic/vnic_wq.c
56
buf->next = wq->bufs[0];
drivers/scsi/fnic/vnic_wq.c
59
buf->next = wq->bufs[i + 1];
drivers/scsi/fnic/vnic_wq.c
61
buf->next = buf + 1;
drivers/scsi/fnic/vnic_wq.h
114
buf = buf->next;
drivers/scsi/fnic/vnic_wq.h
144
wq->to_clean = buf->next;
drivers/scsi/fnic/vnic_wq.h
58
struct vnic_wq_buf *next;
drivers/scsi/hpsa.c
9674
struct hpsa_sas_phy *next;
drivers/scsi/hpsa.c
9676
list_for_each_entry_safe(hpsa_sas_phy, next,
drivers/scsi/hpsa.c
9701
struct hpsa_sas_port *next;
drivers/scsi/hpsa.c
9706
list_for_each_entry_safe(hpsa_sas_port, next,
drivers/scsi/hptiop.c
1221
hba->u.mvfrey.internal_req.next = NULL;
drivers/scsi/hptiop.c
1449
hba->reqs[i].next = NULL;
drivers/scsi/hptiop.c
704
hba->req_list = ret->next;
drivers/scsi/hptiop.c
712
req->next = hba->req_list;
drivers/scsi/hptiop.h
247
struct hptiop_request *next;
drivers/scsi/ibmvscsi/ibmvfc.c
1545
evt = list_entry(queue->free.next, struct ibmvfc_event, queue_list);
drivers/scsi/ibmvscsi/ibmvfc.c
1549
evt = list_entry(queue->free.next, struct ibmvfc_event, queue_list);
drivers/scsi/ibmvscsi/ibmvscsi.c
449
pool->next = 0;
drivers/scsi/ibmvscsi/ibmvscsi.c
560
int offset = pool->next;
drivers/scsi/ibmvscsi/ibmvscsi.c
565
pool->next = offset;
drivers/scsi/ibmvscsi/ibmvscsi.h
72
int next;
drivers/scsi/initio.c
1247
int next;
drivers/scsi/initio.c
1249
next = host->active->next_state;
drivers/scsi/initio.c
1251
switch (next) {
drivers/scsi/initio.c
1253
next = initio_state_1(host);
drivers/scsi/initio.c
1256
next = initio_state_2(host);
drivers/scsi/initio.c
1259
next = initio_state_3(host);
drivers/scsi/initio.c
1262
next = initio_state_4(host);
drivers/scsi/initio.c
1265
next = initio_state_5(host);
drivers/scsi/initio.c
1268
next = initio_state_6(host);
drivers/scsi/initio.c
1271
next = initio_state_7(host);
drivers/scsi/initio.c
1278
if (next <= 0)
drivers/scsi/initio.c
1279
return next;
drivers/scsi/initio.c
2377
if ((host->first_busy = tmp->next) == NULL)
drivers/scsi/initio.c
2380
prev->next = tmp->next;
drivers/scsi/initio.c
2391
tmp = tmp->next;
drivers/scsi/initio.c
2873
prev->next = tmp;
drivers/scsi/initio.c
2876
prev->next = NULL;
drivers/scsi/initio.c
664
if ((host->first_avail = scb->next) == NULL)
drivers/scsi/initio.c
666
scb->next = NULL;
drivers/scsi/initio.c
691
cmnd->next = NULL;
drivers/scsi/initio.c
693
host->last_avail->next = cmnd;
drivers/scsi/initio.c
710
scbp->next = NULL;
drivers/scsi/initio.c
712
host->last_pending->next = scbp;
drivers/scsi/initio.c
728
if ((scbp->next = host->first_pending) != NULL) {
drivers/scsi/initio.c
753
first = first->next;
drivers/scsi/initio.c
758
first = first->next;
drivers/scsi/initio.c
775
if ((host->first_pending = tmp->next) == NULL)
drivers/scsi/initio.c
778
prev->next = tmp->next;
drivers/scsi/initio.c
782
tmp->next = NULL;
drivers/scsi/initio.c
786
tmp = tmp->next;
drivers/scsi/initio.c
801
scbp->next = NULL;
drivers/scsi/initio.c
803
host->last_busy->next = scbp;
drivers/scsi/initio.c
818
if ((host->first_busy = tmp->next) == NULL)
drivers/scsi/initio.c
820
tmp->next = NULL;
drivers/scsi/initio.c
845
if ((host->first_busy = tmp->next) == NULL)
drivers/scsi/initio.c
848
prev->next = tmp->next;
drivers/scsi/initio.c
852
tmp->next = NULL;
drivers/scsi/initio.c
860
tmp = tmp->next;
drivers/scsi/initio.c
877
tmp = tmp->next;
drivers/scsi/initio.c
892
scbp->next = NULL;
drivers/scsi/initio.c
894
host->last_done->next = scbp;
drivers/scsi/initio.c
907
if ((host->first_done = tmp->next) == NULL)
drivers/scsi/initio.c
909
tmp->next = NULL;
drivers/scsi/initio.c
944
if ((host->first_pending = tmp->next) == NULL)
drivers/scsi/initio.c
947
prev->next = tmp->next;
drivers/scsi/initio.c
959
tmp = tmp->next;
drivers/scsi/initio.c
974
if ((host->first_busy = tmp->next) == NULL)
drivers/scsi/initio.c
977
prev->next = tmp->next;
drivers/scsi/initio.c
981
tmp->next = NULL;
drivers/scsi/initio.c
993
tmp = tmp->next;
drivers/scsi/initio.h
350
struct scsi_ctrl_blk *next;
drivers/scsi/ipr.c
1372
res = list_entry(ioa_cfg->free_res_q.next,
drivers/scsi/ipr.c
6790
ipr_cmd->u.res = list_entry(ioa_cfg->used_res_q.next,
drivers/scsi/ipr.c
6864
ipr_cmd->u.res = list_entry(ioa_cfg->used_res_q.next,
drivers/scsi/ipr.c
690
ipr_cmd = list_entry(hrrq->hrrq_free_q.next,
drivers/scsi/ipr.c
7048
res = list_entry(ioa_cfg->free_res_q.next,
drivers/scsi/ipr.c
7765
hostrcb = list_entry(ioa_cfg->hostrcb_free_q.next,
drivers/scsi/ips.c
1099
scratch->next = NULL;
drivers/scsi/ips.c
2966
item->next = NULL;
drivers/scsi/ips.c
2969
queue->tail->next = item;
drivers/scsi/ips.c
3003
queue->head = item->next;
drivers/scsi/ips.c
3004
item->next = NULL;
drivers/scsi/ips.c
3041
while ((p) && (item != p->next))
drivers/scsi/ips.c
3042
p = p->next;
drivers/scsi/ips.c
3046
p->next = item->next;
drivers/scsi/ips.c
3048
if (!item->next)
drivers/scsi/ips.c
3051
item->next = NULL;
drivers/scsi/ips.c
802
item = item->next;
drivers/scsi/ips.h
979
struct ips_copp_wait_item *next;
drivers/scsi/isci/request.c
2144
enum sci_base_request_states next)
drivers/scsi/isci/request.c
2150
sci_change_state(&ireq->sm, next);
drivers/scsi/isci/task.c
103
task = list_entry(task->list.next, struct sas_task, list))
drivers/scsi/libfc/fc_disc.c
145
list_for_each_entry_safe(dp, next, &disc_ports, peers) {
drivers/scsi/libfc/fc_disc.c
80
struct fc_disc_port *dp, *next;
drivers/scsi/libfc/fc_exch.c
1939
struct fc_exch *next;
drivers/scsi/libfc/fc_exch.c
1943
list_for_each_entry_safe(ep, next, &pool->ex_list, ex_list) {
drivers/scsi/libfc/fc_exch.c
2559
struct fc_exch_mgr_anchor *ema, *next;
drivers/scsi/libfc/fc_exch.c
2562
list_for_each_entry_safe(ema, next, &lport->ema_list, ema_list)
drivers/scsi/libiscsi.c
1624
task = list_entry(conn->mgmtqueue.next, struct iscsi_task,
drivers/scsi/libiscsi.c
1647
task = list_entry(conn->requeue.next, struct iscsi_task,
drivers/scsi/libiscsi.c
1663
task = list_entry(conn->cmdqueue.next, struct iscsi_task,
drivers/scsi/libsas/sas_ata.c
903
list_del_init(sata_q.next);
drivers/scsi/libsas/sas_discover.c
66
phy = container_of(port->phy_list.next, struct asd_sas_phy, port_phy_el);
drivers/scsi/libsas/sas_event.c
114
sas_phy = container_of(port->phy_list.next, struct asd_sas_phy,
drivers/scsi/libsas/sas_scsi_host.c
669
ssp = list_entry(ha->eh_dev_q.next, typeof(*ssp), eh_list_node);
drivers/scsi/lpfc/lpfc_bsg.c
2898
struct list_head head, *curr, *next;
drivers/scsi/lpfc/lpfc_bsg.c
2935
list_for_each_safe(curr, next, &head) {
drivers/scsi/lpfc/lpfc_bsg.c
2979
dmp = list_entry(next, struct lpfc_dmabuf, list);
drivers/scsi/lpfc/lpfc_bsg.c
2993
dmp = list_entry(next, struct lpfc_dmabuf, list);
drivers/scsi/lpfc/lpfc_bsg.c
5461
struct lpfc_dmabuf *dmabuf, *next;
drivers/scsi/lpfc/lpfc_bsg.c
5503
list_for_each_entry_safe(dmabuf, next,
drivers/scsi/lpfc/lpfc_bsg.c
602
prsp = (struct lpfc_dmabuf *)pcmd->list.next;
drivers/scsi/lpfc/lpfc_bsg.c
800
ed = list_entry(evt->events_to_get.next, typeof(*ed), node);
drivers/scsi/lpfc/lpfc_bsg.c
807
ed = list_entry(evt->events_to_see.next, typeof(*ed), node);
drivers/scsi/lpfc/lpfc_bsg.c
873
struct list_head head, *curr, *next;
drivers/scsi/lpfc/lpfc_bsg.c
883
list_for_each_safe(curr, next, &head) {
drivers/scsi/lpfc/lpfc_debugfs.c
2200
struct lpfc_dmabuf *dmabuf, *next;
drivers/scsi/lpfc/lpfc_debugfs.c
2211
list_for_each_entry_safe(dmabuf, next,
drivers/scsi/lpfc/lpfc_hbadisc.c
2588
} else if (fcf_pri->list.next == &phba->fcf.fcf_pri_list
drivers/scsi/lpfc/lpfc_init.c
12884
struct lpfc_queue *eq, *next;
drivers/scsi/lpfc/lpfc_init.c
12903
list_for_each_entry_safe(eq, next, &eqlist, _poll_list) {
drivers/scsi/lpfc/lpfc_init.c
12914
struct lpfc_queue *eq, *next;
drivers/scsi/lpfc/lpfc_init.c
12928
list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list) {
drivers/scsi/lpfc/lpfc_init.c
14583
struct lpfc_dmabuf *dmabuf, *next;
drivers/scsi/lpfc/lpfc_init.c
14657
list_for_each_entry_safe(dmabuf, next, &dma_buffer_list, list) {
drivers/scsi/lpfc/lpfc_scsi.h
33
entry = list_entry((list)->next, type, member); \
drivers/scsi/lpfc/lpfc_scsi.h
40
list_entry((list)->next, type, member)
drivers/scsi/lpfc/lpfc_sli.c
12306
slp->next, slp->prev, pring->postbufq_cnt);
drivers/scsi/lpfc/lpfc_sli.c
12350
slp->next, slp->prev, pring->postbufq_cnt);
drivers/scsi/lpfc/lpfc_sli.c
15764
struct lpfc_queue *eq, *next;
drivers/scsi/lpfc/lpfc_sli.c
15766
list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list)
drivers/scsi/lpfc/lpfc_sli.c
20455
uint16_t next;
drivers/scsi/lpfc/lpfc_sli.c
20458
for_each_set_bit_wrap(next, phba->fcf.fcf_rr_bmask,
drivers/scsi/lpfc/lpfc_sli.c
20460
if (next == phba->fcf.current_rec.fcf_indx)
drivers/scsi/lpfc/lpfc_sli.c
20463
if (!(phba->fcf.fcf_pri[next].fcf_rec.flag & LPFC_FCF_FLOGI_FAILED)) {
drivers/scsi/lpfc/lpfc_sli.c
20465
"2845 Get next roundrobin failover FCF (x%x)\n", next);
drivers/scsi/lpfc/lpfc_sli.c
20466
return next;
drivers/scsi/lpfc/lpfc_sli.c
6883
struct lpfc_dmabuf *dmabuf, *next;
drivers/scsi/lpfc/lpfc_sli.c
6886
list_for_each_entry_safe(dmabuf, next,
drivers/scsi/megaraid.c
1071
struct list_head *pos, *next;
drivers/scsi/megaraid.c
1074
list_for_each_safe(pos, next, &adapter->pending_list) {
drivers/scsi/megaraid.c
1926
struct list_head *pos, *next;
drivers/scsi/megaraid.c
1940
list_for_each_safe(pos, next, &adapter->pending_list) {
drivers/scsi/megaraid.c
434
scb = list_entry(head->next, scb_t, list);
drivers/scsi/megaraid.c
436
list_del_init(head->next);
drivers/scsi/megaraid/megaraid_mbox.c
1290
scb = list_entry(head->next, scb_t, list);
drivers/scsi/megaraid/megaraid_mbox.c
1887
scb = list_entry(adapter->pend_list.next, scb_t, list);
drivers/scsi/megaraid/megaraid_mbox.c
3577
scb = list_entry(head->next, scb_t, list);
drivers/scsi/megaraid/megaraid_mm.c
608
kioc = list_entry(head->next, uioc_t, list);
drivers/scsi/megaraid/megaraid_sas_base.c
2788
reset_cmd = list_entry((&clist_local)->next,
drivers/scsi/megaraid/megaraid_sas_base.c
326
cmd = list_entry((&instance->cmd_pool)->next,
drivers/scsi/megaraid/megaraid_sas_base.c
370
list_add(&cmd->list, (&instance->cmd_pool)->next);
drivers/scsi/megaraid/megaraid_sas_base.c
3832
cmd = list_entry((&clist_local)->next,
drivers/scsi/mesh.c
150
struct mesh_state *next;
drivers/scsi/mesh.c
551
struct scsi_cmnd *cmd, *prev, *next;
drivers/scsi/mesh.c
568
next = (struct scsi_cmnd *) cmd->host_scribble;
drivers/scsi/mesh.c
570
ms->request_q = next;
drivers/scsi/mesh.c
572
prev->host_scribble = (void *) next;
drivers/scsi/mesh.c
573
if (next == NULL)
drivers/scsi/mpi3mr/mpi3mr_os.c
2348
_rmhs_node = list_entry(mrioc->delayed_rmhs_list.next,
drivers/scsi/mpi3mr/mpi3mr_os.c
2355
_evtack_node = list_entry(mrioc->delayed_evtack_cmds_list.next,
drivers/scsi/mpi3mr/mpi3mr_os.c
2408
delayed_dev_rmhs = list_entry(mrioc->delayed_rmhs_list.next,
drivers/scsi/mpi3mr/mpi3mr_os.c
2619
list_entry(mrioc->delayed_evtack_cmds_list.next,
drivers/scsi/mpi3mr/mpi3mr_transport.c
1330
struct mpi3mr_sas_phy *mr_sas_phy, *next;
drivers/scsi/mpi3mr/mpi3mr_transport.c
1501
list_for_each_entry_safe(mr_sas_phy, next, &mr_sas_port->phy_list,
drivers/scsi/mpi3mr/mpi3mr_transport.c
1525
struct mpi3mr_sas_port *mr_sas_port, *next;
drivers/scsi/mpi3mr/mpi3mr_transport.c
1541
list_for_each_entry_safe(mr_sas_port, next, &mr_sas_node->sas_port_list,
drivers/scsi/mpi3mr/mpi3mr_transport.c
2207
struct mpi3mr_sas_port *mr_sas_port, *next;
drivers/scsi/mpi3mr/mpi3mr_transport.c
2212
list_for_each_entry_safe(mr_sas_port, next,
drivers/scsi/mpi3mr/mpi3mr_transport.c
733
struct mpi3mr_sas_port *mr_sas_port, *next;
drivers/scsi/mpi3mr/mpi3mr_transport.c
739
list_for_each_entry_safe(mr_sas_port, next, &mr_sas_node->sas_port_list,
drivers/scsi/mpt3sas/mpt3sas_base.c
1962
struct adapter_reply_queue *reply_q, *next;
drivers/scsi/mpt3sas/mpt3sas_base.c
1967
list_for_each_entry_safe(reply_q, next, &ioc->reply_queue_list, list) {
drivers/scsi/mpt3sas/mpt3sas_base.c
3128
struct adapter_reply_queue *reply_q, *next;
drivers/scsi/mpt3sas/mpt3sas_base.c
3133
list_for_each_entry_safe(reply_q, next, &ioc->reply_queue_list, list) {
drivers/scsi/mpt3sas/mpt3sas_base.c
3945
request = list_entry(ioc->internal_free_list.next,
drivers/scsi/mpt3sas/mpt3sas_base.c
4013
request = list_entry(ioc->hpr_free_list.next,
drivers/scsi/mpt3sas/mpt3sas_scsih.c
12299
struct _sas_port *mpt3sas_port, *next;
drivers/scsi/mpt3sas/mpt3sas_scsih.c
12304
list_for_each_entry_safe(mpt3sas_port, next,
drivers/scsi/mpt3sas/mpt3sas_scsih.c
12517
struct _raid_device *raid_device, *next;
drivers/scsi/mpt3sas/mpt3sas_scsih.c
12554
list_for_each_entry_safe(raid_device, next, &ioc->raid_device_list,
drivers/scsi/mpt3sas/mpt3sas_scsih.c
4883
delayed_event_ack = list_entry(ioc->delayed_event_ack_list.next,
drivers/scsi/mpt3sas/mpt3sas_scsih.c
4893
delayed_sc = list_entry(ioc->delayed_sc_list.next,
drivers/scsi/mpt3sas/mpt3sas_scsih.c
4921
delayed_tr = list_entry(ioc->delayed_tr_volume_list.next,
drivers/scsi/mpt3sas/mpt3sas_scsih.c
4931
delayed_tr = list_entry(ioc->delayed_tr_list.next,
drivers/scsi/mpt3sas/mpt3sas_transport.c
616
struct _sas_port *mpt3sas_port, *next;
drivers/scsi/mpt3sas/mpt3sas_transport.c
622
list_for_each_entry_safe(mpt3sas_port, next, &sas_node->sas_port_list,
drivers/scsi/mpt3sas/mpt3sas_transport.c
685
struct _sas_phy *mpt3sas_phy, *next;
drivers/scsi/mpt3sas/mpt3sas_transport.c
869
list_for_each_entry_safe(mpt3sas_phy, next, &mpt3sas_port->phy_list,
drivers/scsi/mpt3sas/mpt3sas_transport.c
893
struct _sas_port *mpt3sas_port, *next;
drivers/scsi/mpt3sas/mpt3sas_transport.c
910
list_for_each_entry_safe(mpt3sas_port, next, &sas_node->sas_port_list,
drivers/scsi/mvsas/mv_sas.c
67
phy = container_of(sha->sas_port[i]->phy_list.next,
drivers/scsi/mvumi.c
321
cmd = list_entry((&mhba->cmd_pool)->next,
drivers/scsi/ncr53c8xx.c
159
struct list_head *elem = head->next;
drivers/scsi/ncr53c8xx.c
203
struct m_link *next;
drivers/scsi/ncr53c8xx.c
207
struct m_vtob *next;
drivers/scsi/ncr53c8xx.c
223
struct m_pool *next;
drivers/scsi/ncr53c8xx.c
244
while (!h[j].next) {
drivers/scsi/ncr53c8xx.c
246
h[j].next = (m_link_s *)mp->getp(mp);
drivers/scsi/ncr53c8xx.c
247
if (h[j].next)
drivers/scsi/ncr53c8xx.c
248
h[j].next->next = NULL;
drivers/scsi/ncr53c8xx.c
254
a = (m_addr_t) h[j].next;
drivers/scsi/ncr53c8xx.c
256
h[j].next = h[j].next->next;
drivers/scsi/ncr53c8xx.c
260
h[j].next = (m_link_s *) (a+s);
drivers/scsi/ncr53c8xx.c
261
h[j].next->next = NULL;
drivers/scsi/ncr53c8xx.c
301
while (q->next && q->next != (m_link_s *) b) {
drivers/scsi/ncr53c8xx.c
302
q = q->next;
drivers/scsi/ncr53c8xx.c
304
if (!q->next) {
drivers/scsi/ncr53c8xx.c
305
((m_link_s *) a)->next = h[i].next;
drivers/scsi/ncr53c8xx.c
306
h[i].next = (m_link_s *) a;
drivers/scsi/ncr53c8xx.c
309
q->next = q->next->next;
drivers/scsi/ncr53c8xx.c
391
vbp->next = mp->vtob[hc];
drivers/scsi/ncr53c8xx.c
409
vbpp = &(*vbpp)->next;
drivers/scsi/ncr53c8xx.c
412
*vbpp = (*vbpp)->next;
drivers/scsi/ncr53c8xx.c
423
for (mp = mp0.next; mp && mp->bush != bush; mp = mp->next);
drivers/scsi/ncr53c8xx.c
436
mp->next = mp0.next;
drivers/scsi/ncr53c8xx.c
437
mp0.next = mp;
drivers/scsi/ncr53c8xx.c
444
struct m_pool **pp = &mp0.next;
drivers/scsi/ncr53c8xx.c
447
pp = &(*pp)->next;
drivers/scsi/ncr53c8xx.c
449
*pp = (*pp)->next;
drivers/scsi/ncr53c8xx.c
500
vp = vp->next;
drivers/scsi/pmcraid.c
1608
res = list_entry(pinstance->free_res_q.next,
drivers/scsi/pmcraid.c
373
cmd = list_entry(pinstance->free_cmd_pool.next,
drivers/scsi/pmcraid.c
5062
res = list_entry(pinstance->free_res_q.next,
drivers/scsi/qedi/qedi.h
209
u16 next;
drivers/scsi/qedi/qedi_main.c
533
u16 start_id, u16 next)
drivers/scsi/qedi/qedi_main.c
537
id_tbl->next = next;
drivers/scsi/qedi/qedi_main.c
574
id = find_next_zero_bit(id_tbl->table, id_tbl->max, id_tbl->next);
drivers/scsi/qedi/qedi_main.c
577
if (id_tbl->next != 0) {
drivers/scsi/qedi/qedi_main.c
578
id = find_first_zero_bit(id_tbl->table, id_tbl->next);
drivers/scsi/qedi/qedi_main.c
579
if (id >= id_tbl->next)
drivers/scsi/qedi/qedi_main.c
586
id_tbl->next = (id + 1) & (id_tbl->max - 1);
drivers/scsi/qla1280.c
1246
sp = list_entry(done_q->next, struct srb, list);
drivers/scsi/qla1280.h
1013
struct scsi_qla_host *next;
drivers/scsi/qla2xxx/qla_def.h
5373
struct list_head next;
drivers/scsi/qla2xxx/qla_def.h
5392
struct list_head next;
drivers/scsi/qla2xxx/qla_edif.c
1409
INIT_LIST_HEAD(&sa_ctl->next);
drivers/scsi/qla2xxx/qla_edif.c
1421
list_add_tail(&sa_ctl->next, &fcport->edif.tx_sa_list);
drivers/scsi/qla2xxx/qla_edif.c
1423
list_add_tail(&sa_ctl->next, &fcport->edif.rx_sa_list);
drivers/scsi/qla2xxx/qla_edif.c
1438
next) {
drivers/scsi/qla2xxx/qla_edif.c
1439
list_del(&sa_ctl->next);
drivers/scsi/qla2xxx/qla_edif.c
1444
next) {
drivers/scsi/qla2xxx/qla_edif.c
1445
list_del(&sa_ctl->next);
drivers/scsi/qla2xxx/qla_edif.c
1463
list_for_each_entry_safe(sa_ctl, tsa_ctl, sa_list, next) {
drivers/scsi/qla2xxx/qla_edif.c
181
INIT_LIST_HEAD(&entry->next);
drivers/scsi/qla2xxx/qla_edif.c
189
list_add_tail(&entry->next, &fcport->edif.edif_indx_list);
drivers/scsi/qla2xxx/qla_edif.c
200
list_del(&entry->next);
drivers/scsi/qla2xxx/qla_edif.c
2715
list_for_each_entry_safe(entry, tentry, indx_list, next) {
drivers/scsi/qla2xxx/qla_edif.c
2768
list_del(&entry->next);
drivers/scsi/qla2xxx/qla_edif.c
3238
list_for_each_entry_safe(indx_lst, tindx_lst, indx_list, next) {
drivers/scsi/qla2xxx/qla_edif.c
3239
list_del(&indx_lst->next);
drivers/scsi/qla2xxx/qla_edif.c
325
list_del(&sa_ctl->next);
drivers/scsi/qla2xxx/qla_edif.c
3292
INIT_LIST_HEAD(&entry->next);
drivers/scsi/qla2xxx/qla_edif.c
3300
list_add_tail(&entry->next, sa_list);
drivers/scsi/qla2xxx/qla_edif.c
3365
list_for_each_entry_safe(entry, tmp, &ha->sadb_rx_index_list, next) {
drivers/scsi/qla2xxx/qla_edif.c
3366
list_del(&entry->next);
drivers/scsi/qla2xxx/qla_edif.c
3370
list_for_each_entry_safe(entry, tmp, &ha->sadb_tx_index_list, next) {
drivers/scsi/qla2xxx/qla_edif.c
3371
list_del(&entry->next);
drivers/scsi/qla2xxx/qla_edif.c
457
list_for_each_entry_safe(entry, tmp, &ha->sadb_rx_index_list, next) {
drivers/scsi/qla2xxx/qla_edif.c
459
list_del(&entry->next);
drivers/scsi/qla2xxx/qla_edif.c
468
list_for_each_entry_safe(entry, tmp, &ha->sadb_tx_index_list, next) {
drivers/scsi/qla2xxx/qla_edif.c
470
list_del(&entry->next);
drivers/scsi/qla2xxx/qla_edif.c
87
list_for_each_entry_safe(entry, tentry, indx_list, next) {
drivers/scsi/qla2xxx/qla_edif.h
14
struct list_head next;
drivers/scsi/qla2xxx/qla_os.c
3777
BUG_ON(base_vha->list.next == &ha->vp_list);
drivers/scsi/qla2xxx/qla_os.c
3979
struct purex_item *item, *next;
drivers/scsi/qla2xxx/qla_os.c
3983
list_for_each_entry_safe(item, next, &list->head, list) {
drivers/scsi/qla2xxx/qla_os.c
6474
struct purex_item *item, *next;
drivers/scsi/qla2xxx/qla_os.c
6481
list_for_each_entry_safe(item, next, &head, list) {
drivers/scsi/qla2xxx/qla_sup.c
1327
goto next;
drivers/scsi/qla2xxx/qla_sup.c
1337
next:
drivers/scsi/qla2xxx/qla_target.c
7413
tgt->sess_works_list.next, typeof(*prm),
drivers/scsi/qlogicfas.c
162
priv->next = cards;
drivers/scsi/qlogicfas.c
220
for (priv = cards; priv != NULL; priv = priv->next)
drivers/scsi/qlogicfas408.h
87
struct qlogicfas408_priv *next; /* next private struct */
drivers/scsi/qlogicpti.c
1215
struct scsi_cmnd *next;
drivers/scsi/qlogicpti.c
1217
next = (struct scsi_cmnd *) dq->host_scribble;
drivers/scsi/qlogicpti.c
1219
dq = next;
drivers/scsi/qlogicpti.c
690
while(qlink->next)
drivers/scsi/qlogicpti.c
691
qlink = qlink->next;
drivers/scsi/qlogicpti.c
692
qlink->next = qpti;
drivers/scsi/qlogicpti.c
696
qpti->next = NULL;
drivers/scsi/qlogicpti.c
704
qptichain = qpti->next;
drivers/scsi/qlogicpti.c
707
while(qlink->next != qpti)
drivers/scsi/qlogicpti.c
708
qlink = qlink->next;
drivers/scsi/qlogicpti.c
709
qlink->next = qpti->next;
drivers/scsi/qlogicpti.c
711
qpti->next = NULL;
drivers/scsi/qlogicpti.h
359
struct qlogicpti *next;
drivers/scsi/qlogicpti.h
505
for((qp) = qptichain; (qp); (qp) = (qp)->next)
drivers/scsi/raid_class.c
101
list_for_each_entry_safe(rc, next, &rd->component_list, node) {
drivers/scsi/raid_class.c
98
struct raid_component *rc, *next;
drivers/scsi/scsi.c
828
struct scsi_device *next = NULL;
drivers/scsi/scsi.c
832
while (list->next != &shost->__devices) {
drivers/scsi/scsi.c
833
next = list_entry(list->next, struct scsi_device, siblings);
drivers/scsi/scsi.c
838
if (!scsi_device_is_pseudo_dev(next) && !scsi_device_get(next))
drivers/scsi/scsi.c
840
next = NULL;
drivers/scsi/scsi.c
841
list = list->next;
drivers/scsi/scsi.c
847
return next;
drivers/scsi/scsi_debug.c
4799
unsigned long next;
drivers/scsi/scsi_debug.c
4805
next = find_next_zero_bit(sip->map_storep, map_size, index);
drivers/scsi/scsi_debug.c
4807
next = find_next_bit(sip->map_storep, map_size, index);
drivers/scsi/scsi_debug.c
4809
end = min_t(sector_t, sdebug_store_sectors, map_index_to_lba(next));
drivers/scsi/scsi_devinfo.c
497
char *vendor, *model, *strflags, *next;
drivers/scsi/scsi_devinfo.c
501
next = dev_list;
drivers/scsi/scsi_devinfo.c
502
if (next && next[0] == '"') {
drivers/scsi/scsi_devinfo.c
506
next++;
drivers/scsi/scsi_devinfo.c
516
for (vendor = strsep(&next, ":"); vendor && (vendor[0] != '\0')
drivers/scsi/scsi_devinfo.c
517
&& (res == 0); vendor = strsep(&next, ":")) {
drivers/scsi/scsi_devinfo.c
519
model = strsep(&next, ":");
drivers/scsi/scsi_devinfo.c
521
strflags = strsep(&next, next_check);
drivers/scsi/scsi_devinfo.c
607
if (devinfo_table->scsi_dev_info_list.next == dl->bottom &&
drivers/scsi/scsi_devinfo.c
644
dl->bottom = dl->bottom->next;
drivers/scsi/scsi_devinfo.c
646
dl->top = dl->top->next;
drivers/scsi/scsi_devinfo.c
654
dl->bottom = devinfo_table->scsi_dev_info_list.next;
drivers/scsi/scsi_devinfo.c
667
.next = devinfo_seq_next,
drivers/scsi/scsi_error.c
1363
struct scsi_cmnd *scmd, *next;
drivers/scsi/scsi_error.c
1371
list_for_each_entry_safe(scmd, next, work_q, eh_entry) {
drivers/scsi/scsi_error.c
1482
struct scsi_cmnd *scmd, *next;
drivers/scsi/scsi_error.c
1487
scmd = list_entry(cmd_list->next, struct scsi_cmnd, eh_entry);
drivers/scsi/scsi_error.c
1507
list_for_each_entry_safe(scmd, next, cmd_list, eh_entry)
drivers/scsi/scsi_error.c
1560
struct scsi_cmnd *scmd, *stu_scmd, *next;
drivers/scsi/scsi_error.c
1591
list_for_each_entry_safe(scmd, next,
drivers/scsi/scsi_error.c
1626
struct scsi_cmnd *scmd, *bdr_scmd, *next;
drivers/scsi/scsi_error.c
1657
list_for_each_entry_safe(scmd, next,
drivers/scsi/scsi_error.c
1694
struct scsi_cmnd *next, *scmd;
drivers/scsi/scsi_error.c
1709
scmd = list_entry(tmp_list.next, struct scsi_cmnd, eh_entry);
drivers/scsi/scsi_error.c
1723
list_for_each_entry_safe(scmd, next, &tmp_list, eh_entry) {
drivers/scsi/scsi_error.c
1750
struct scsi_cmnd *scmd, *chan_scmd, *next;
drivers/scsi/scsi_error.c
1792
list_for_each_entry_safe(scmd, next, work_q, eh_entry) {
drivers/scsi/scsi_error.c
1822
struct scsi_cmnd *scmd, *next;
drivers/scsi/scsi_error.c
1827
scmd = list_entry(work_q->next,
drivers/scsi/scsi_error.c
1839
list_for_each_entry_safe(scmd, next, work_q, eh_entry) {
drivers/scsi/scsi_error.c
1860
struct scsi_cmnd *scmd, *next;
drivers/scsi/scsi_error.c
1863
list_for_each_entry_safe(scmd, next, work_q, eh_entry) {
drivers/scsi/scsi_error.c
2255
struct scsi_cmnd *scmd, *next;
drivers/scsi/scsi_error.c
2257
list_for_each_entry_safe(scmd, next, done_q, eh_entry) {
drivers/scsi/scsi_lib.c
512
sdev = list_entry(starved_list.next,
drivers/scsi/scsi_proc.c
476
struct device *next = bus_find_next_device(&scsi_bus_type, start);
drivers/scsi/scsi_proc.c
479
return next;
drivers/scsi/scsi_proc.c
517
.next = scsi_seq_next,
drivers/scsi/scsi_scan.c
182
struct async_scan_data *next = list_entry(scanning_hosts.next,
drivers/scsi/scsi_scan.c
184
complete(&next->prev_finished);
drivers/scsi/scsi_scan.c
2021
struct async_scan_data *next = list_entry(scanning_hosts.next,
drivers/scsi/scsi_scan.c
2023
complete(&next->prev_finished);
drivers/scsi/scsi_transport_fc.c
2034
pos = list_entry((head)->next, typeof(*pos), member)
drivers/scsi/scsi_transport_sas.c
1077
struct list_head *ent = port->phy_list.next;
drivers/scsi/sg.c
2346
.next = dev_seq_next,
drivers/scsi/sg.c
2354
.next = dev_seq_next,
drivers/scsi/sg.c
2362
.next = dev_seq_next,
drivers/scsi/sim710.c
53
char *pos = str, *next;
drivers/scsi/sim710.c
56
while(pos != NULL && (next = strchr(pos, ':')) != NULL) {
drivers/scsi/sim710.c
57
int val = (int)simple_strtoul(++next, NULL, 0);
drivers/scsi/smartpqi/smartpqi_init.c
2219
struct pqi_scsi_dev *next;
drivers/scsi/smartpqi/smartpqi_init.c
2271
list_for_each_entry_safe(device, next, &ctrl_info->scsi_device_list,
drivers/scsi/smartpqi/smartpqi_init.c
2302
list_for_each_entry_safe(device, next, &delete_list, delete_list_entry)
drivers/scsi/smartpqi/smartpqi_init.c
2310
list_for_each_entry_safe(device, next, &delete_list, delete_list_entry) {
drivers/scsi/smartpqi/smartpqi_init.c
2349
list_for_each_entry_safe(device, next, &add_list, add_list_entry) {
drivers/scsi/smartpqi/smartpqi_init.c
2415
struct pqi_scsi_dev *next;
drivers/scsi/smartpqi/smartpqi_init.c
2587
list_for_each_entry_safe(device, next, &new_device_list_head,
drivers/scsi/smartpqi/smartpqi_init.c
4504
struct pqi_io_request *next;
drivers/scsi/smartpqi/smartpqi_init.c
4525
list_for_each_entry_safe(io_request, next,
drivers/scsi/smartpqi/smartpqi_init.c
6227
struct pqi_io_request *next;
drivers/scsi/smartpqi/smartpqi_init.c
6238
list_for_each_entry_safe(io_request, next,
drivers/scsi/smartpqi/smartpqi_sas_transport.c
168
struct pqi_sas_phy *next;
drivers/scsi/smartpqi/smartpqi_sas_transport.c
170
list_for_each_entry_safe(pqi_sas_phy, next,
drivers/scsi/smartpqi/smartpqi_sas_transport.c
195
struct pqi_sas_port *next;
drivers/scsi/smartpqi/smartpqi_sas_transport.c
200
list_for_each_entry_safe(pqi_sas_port, next,
drivers/scsi/snic/snic_debugfs.c
410
.next = snic_trc_seq_next,
drivers/scsi/snic/snic_io.c
101
SNIC_BUG_ON(rqi->list.next == NULL); /* if not added to spl_cmd_list */
drivers/scsi/snic/vnic_wq.c
212
buf = wq->to_clean = buf->next;
drivers/scsi/snic/vnic_wq.c
51
buf->next = wq->bufs[0];
drivers/scsi/snic/vnic_wq.c
54
buf->next = wq->bufs[i + 1];
drivers/scsi/snic/vnic_wq.c
56
buf->next = buf + 1;
drivers/scsi/snic/vnic_wq.h
129
wq->to_clean = buf->next;
drivers/scsi/snic/vnic_wq.h
37
struct vnic_wq_buf *next;
drivers/scsi/snic/vnic_wq.h
99
buf = buf->next;
drivers/scsi/sym53c8xx_2/sym_hipd.h
1121
struct sym_m_link *next;
drivers/scsi/sym53c8xx_2/sym_hipd.h
1129
struct sym_m_vtob *next;
drivers/scsi/sym53c8xx_2/sym_hipd.h
1160
struct sym_m_pool *next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
120
((m_link_p) a)->next = h[i].next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
121
h[i].next = (m_link_p) a;
drivers/scsi/sym53c8xx_2/sym_malloc.c
127
while (q->next && q->next != (m_link_p) b) {
drivers/scsi/sym53c8xx_2/sym_malloc.c
128
q = q->next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
130
if (!q->next) {
drivers/scsi/sym53c8xx_2/sym_malloc.c
131
((m_link_p) a)->next = h[i].next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
132
h[i].next = (m_link_p) a;
drivers/scsi/sym53c8xx_2/sym_malloc.c
135
q->next = q->next->next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
222
vbp->next = mp->vtob[hc];
drivers/scsi/sym53c8xx_2/sym_malloc.c
240
vbpp = &(*vbpp)->next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
243
*vbpp = (*vbpp)->next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
255
for (mp = mp0.next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
257
mp = mp->next);
drivers/scsi/sym53c8xx_2/sym_malloc.c
271
mp->next = mp0.next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
272
mp0.next = mp;
drivers/scsi/sym53c8xx_2/sym_malloc.c
282
m_pool_p *pp = &mp0.next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
285
pp = &(*pp)->next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
287
*pp = (*pp)->next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
358
vp = vp->next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
64
while (!h[j].next) {
drivers/scsi/sym53c8xx_2/sym_malloc.c
66
h[j].next = (m_link_p) M_GET_MEM_CLUSTER();
drivers/scsi/sym53c8xx_2/sym_malloc.c
67
if (h[j].next)
drivers/scsi/sym53c8xx_2/sym_malloc.c
68
h[j].next->next = NULL;
drivers/scsi/sym53c8xx_2/sym_malloc.c
74
a = h[j].next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
76
h[j].next = h[j].next->next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
80
h[j].next = (m_link_p) (a+s);
drivers/scsi/sym53c8xx_2/sym_malloc.c
81
h[j].next->next = NULL;
drivers/scsi/wd33c93.h
212
struct Scsi_Host *next;
drivers/sh/clk/core.c
421
if (clk->node.next || clk->node.prev)
drivers/sh/intc/virq.c
24
struct intc_virq_list *next;
drivers/sh/intc/virq.c
28
for (entry = head; entry; entry = entry->next)
drivers/sh/intc/virq.c
93
last = &entry->next;
drivers/soc/bcm/brcmstb/pm/pm-mips.c
67
struct brcmstb_mem_transfer *next;
drivers/soc/fsl/dpio/dpio-service.c
87
d = list_entry(dpio_list.next, struct dpaa2_io, node);
drivers/soc/fsl/qbman/bman.c
250
struct bman_pool *next;
drivers/soc/fsl/qbman/qman_test_stash.c
457
hp_cpu->iterator->node.next,
drivers/soc/fsl/qbman/qman_test_stash.c
500
hp_cpu->iterator->node.next,
drivers/soc/mediatek/mtk-mmsys.c
185
enum mtk_ddp_comp_id next)
drivers/soc/mediatek/mtk-mmsys.c
192
if (cur == routes[i].from_comp && next == routes[i].to_comp)
drivers/soc/mediatek/mtk-mmsys.c
204
enum mtk_ddp_comp_id next)
drivers/soc/mediatek/mtk-mmsys.c
211
if (cur == routes[i].from_comp && next == routes[i].to_comp)
drivers/spi/spi-bcm63xx.c
271
t = list_entry(t->transfer_list.next, struct spi_transfer,
drivers/spi/spi-bcm63xx.c
322
t = list_entry(t->transfer_list.next, struct spi_transfer,
drivers/spi/spi-fsi.c
439
struct spi_transfer *next = NULL;
drivers/spi/spi-fsi.c
466
next = list_next_entry(transfer, transfer_list);
drivers/spi/spi-fsi.c
469
if (next->rx_buf) {
drivers/spi/spi-fsi.c
472
if (next->len > SPI_FSI_MAX_RX_SIZE) {
drivers/spi/spi-fsi.c
478
next->len);
drivers/spi/spi-fsi.c
480
shift = SPI_FSI_SEQUENCE_SHIFT_IN(next->len);
drivers/spi/spi-fsi.c
482
} else if (next->tx_buf) {
drivers/spi/spi-fsi.c
483
if ((next->len + transfer->len) > (SPI_FSI_MAX_TX_SIZE + 8)) {
drivers/spi/spi-fsi.c
488
len = next->len;
drivers/spi/spi-fsi.c
496
next = NULL;
drivers/spi/spi-fsi.c
510
if (next) {
drivers/spi/spi-fsi.c
511
rc = fsi_spi_transfer_data(ctx, next);
drivers/spi/spi-fsi.c
515
transfer = next;
drivers/spi/spi-mpc52xx.c
198
ms->transfer = container_of(ms->message->transfers.next,
drivers/spi/spi-mpc52xx.c
295
if (ms->transfer->transfer_list.next == &ms->message->transfers) {
drivers/spi/spi-mpc52xx.c
310
ms->transfer = container_of(ms->transfer->transfer_list.next,
drivers/spi/spi-pic32-sqi.c
343
struct ring_desc *rdesc, *next;
drivers/spi/spi-pic32-sqi.c
430
list_for_each_entry_safe_reverse(rdesc, next,
drivers/spi/spi-topcliff-pch.c
1117
list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) {
drivers/spi/spi-topcliff-pch.c
1139
data->current_msg = list_entry(data->queue.next, struct spi_message,
drivers/spi/spi-topcliff-pch.c
1162
list_entry(data->current_msg->transfers.next,
drivers/spi/spi-topcliff-pch.c
1169
list_entry(data->cur_trans->transfer_list.next,
drivers/spi/spi-topcliff-pch.c
1222
if ((data->cur_trans->transfer_list.next) ==
drivers/spi/spi-topcliff-pch.c
563
list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) {
drivers/spi/spi-topcliff-pch.c
648
list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) {
drivers/spi/spi.c
2077
struct spi_message *next;
drivers/spi/spi.c
2082
next = list_first_entry_or_null(&ctlr->queue, struct spi_message,
drivers/spi/spi.c
2086
return next;
drivers/spi/spi.c
3747
if (rxfer->replaced_after->next == &msg->transfers) {
drivers/spi/spi.c
3765
list_move_tail(rxfer->replaced_after->next,
drivers/ssb/sdio.c
580
tuple = tuple->next;
drivers/staging/greybus/audio_manager.c
88
struct gb_audio_manager_module *module, *next;
drivers/staging/greybus/audio_manager.c
93
list_for_each_entry_safe(module, next, &modules_list, list) {
drivers/staging/greybus/bootrom.c
89
enum next_request_type next,
drivers/staging/greybus/bootrom.c
92
bootrom->next_request = next;
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
137
const struct gc2235_reg *next)
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
141
if (next->type != GC2235_8BIT)
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
145
ctrl->buffer.data[ctrl->index] = (u8)next->val;
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
149
ctrl->buffer.addr = next->reg;
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
165
const struct gc2235_reg *next)
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
170
return ctrl->buffer.addr + ctrl->index == next->reg;
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
176
const struct gc2235_reg *next = reglist;
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
181
for (; next->type != GC2235_TOK_TERM; next++) {
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
182
switch (next->type & GC2235_TOK_MASK) {
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
187
msleep(next->val);
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
195
next)) {
drivers/staging/media/atomisp/i2c/atomisp-gc2235.c
200
err = __gc2235_buf_reg_array(client, &ctrl, next);
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
169
const struct ov2722_reg *next)
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
174
switch (next->type) {
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
177
ctrl->buffer.data[ctrl->index] = (u8)next->val;
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
182
*data16 = cpu_to_be16((u16)next->val);
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
190
ctrl->buffer.addr = next->reg;
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
206
const struct ov2722_reg *next)
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
211
return ctrl->buffer.addr + ctrl->index == next->reg;
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
217
const struct ov2722_reg *next = reglist;
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
222
for (; next->type != OV2722_TOK_TERM; next++) {
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
223
switch (next->type & OV2722_TOK_MASK) {
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
228
msleep(next->val);
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
236
next)) {
drivers/staging/media/atomisp/i2c/atomisp-ov2722.c
241
err = __ov2722_buf_reg_array(client, &ctrl, next);
drivers/staging/media/atomisp/include/hmm/hmm_bo.h
128
struct hmm_buffer_object *next;
drivers/staging/media/atomisp/pci/atomisp_cmd.c
1700
s3a_buf = list_entry(asd->s3a_stats_ready.next,
drivers/staging/media/atomisp/pci/atomisp_cmd.c
623
param = list_entry(pipe->per_frame_params.next,
drivers/staging/media/atomisp/pci/atomisp_compat_css20.c
3043
dis_buf = list_entry(asd->dis_stats.next,
drivers/staging/media/atomisp/pci/atomisp_fops.c
120
metadata_buf = list_entry(metadata_list->next,
drivers/staging/media/atomisp/pci/atomisp_fops.c
158
s3a_buf = list_entry(s3a_list->next, struct atomisp_s3a_buf, list);
drivers/staging/media/atomisp/pci/atomisp_v4l2.c
806
struct v4l2_subdev *sd, *next;
drivers/staging/media/atomisp/pci/atomisp_v4l2.c
812
list_for_each_entry_safe(sd, next, &isp->v4l2_dev.subdevs, list)
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
106
temp_bo = this->next;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
107
this->next = temp_bo->next;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
108
if (temp_bo->next)
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
109
temp_bo->next->prev = this;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
110
temp_bo->next = NULL;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
183
bo->next = this->next;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
184
if (this->next)
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
185
this->next->prev = bo;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
186
this->next = bo;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
263
if (!bo->prev && !bo->next) {
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
270
} else if (!bo->prev && bo->next) {
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
271
bo->next->prev = NULL;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
273
__bo_insert_to_free_rbtree(&bdev->free_rbtree, bo->next);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
274
bo->next = NULL;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
280
} else if (bo->prev && !bo->next) {
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
281
bo->prev->next = NULL;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
288
} else if (bo->prev && bo->next) {
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
289
bo->next->prev = bo->prev;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
290
bo->prev->next = bo->next;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
291
bo->next = NULL;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
467
next_bo = list_entry(bo->list.next, struct hmm_buffer_object, list);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
476
if (bo->list.next != &bdev->entire_bo_list &&
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
511
bo = list_to_hmm_bo(bdev->entire_bo_list.next);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
59
bo->next = NULL;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
97
if (!this->next) {
drivers/staging/media/atomisp/pci/ia_css_acc_types.h
286
CSS_ALIGN(struct ia_css_binary_xinfo *next, 8);
drivers/staging/media/atomisp/pci/ia_css_acc_types.h
367
struct ia_css_fw_info *next;
drivers/staging/media/atomisp/pci/runtime/binary/src/binary.c
1003
xcandidate = xcandidate->next;
drivers/staging/media/atomisp/pci/runtime/binary/src/binary.c
1010
xcandidate = xcandidate->next) {
drivers/staging/media/atomisp/pci/runtime/binary/src/binary.c
398
metrics->next = NULL;
drivers/staging/media/atomisp/pci/runtime/binary/src/binary.c
492
binary->next = binary_infos[binary->sp.pipeline.mode];
drivers/staging/media/atomisp/pci/runtime/binary/src/binary.c
507
for (b = binary_infos[i]; b; b = b->next) {
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
872
for (stage = pipeline->stages; stage; stage = stage->next) {
drivers/staging/media/atomisp/pci/runtime/pipeline/interface/ia_css_pipeline.h
29
struct ia_css_pipeline_stage *next;
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
239
struct ia_css_pipeline_stage *next = s->next;
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
242
s = next;
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
285
while (last && last->next)
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
286
last = last->next;
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
312
last->next = new_stage;
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
332
for (stage = pipeline->stages; stage; stage = stage->next) {
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
352
for (s = pipeline->stages; s; s = s->next) {
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
371
for (s = pipeline->stages; s; s = s->next) {
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
391
for (s = pipeline->stages; s; s = s->next) {
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
414
for (s = pipeline->stages; s; s = s->next) {
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
608
stage->next = NULL;
drivers/staging/media/atomisp/pci/sh_css.c
2520
for (; fw; fw = fw->next) {
drivers/staging/media/atomisp/pci/sh_css.c
2549
for (; fw; fw = fw->next) {
drivers/staging/media/atomisp/pci/sh_css.c
3553
for (stage = pipeline->stages; stage; stage = stage->next) {
drivers/staging/media/atomisp/pci/sh_css.c
7285
for (stage = me->stages; stage; stage = stage->next)
drivers/staging/media/atomisp/pci/sh_css_metrics.c
87
for (; *l; l = &(*l)->next)
drivers/staging/media/atomisp/pci/sh_css_metrics.c
92
metrics->next = NULL;
drivers/staging/media/atomisp/pci/sh_css_metrics.h
24
struct sh_css_binary_metrics *next;
drivers/staging/media/atomisp/pci/sh_css_params.c
1174
for (stage = first_stage; stage; stage = stage->next) {
drivers/staging/media/atomisp/pci/sh_css_params.c
3133
for (stage = pipeline->stages; stage; stage = stage->next) {
drivers/staging/media/atomisp/pci/sh_css_params.c
877
for (stage = pipeline->stages; stage; stage = stage->next) {
drivers/staging/media/atomisp/pci/sh_css_sp.c
1207
for (stage = me->stages, num = 0; stage; stage = stage->next, num++) {
drivers/staging/media/atomisp/pci/sh_css_sp.c
1282
for (stage = me->stages, num = 0; stage; stage = stage->next, num++) {
drivers/staging/media/imx/imx-ic-prpencvf.c
201
struct imx_media_buffer *done, *next;
drivers/staging/media/imx/imx-ic-prpencvf.c
219
next = imx_media_capture_device_next_buf(vdev);
drivers/staging/media/imx/imx-ic-prpencvf.c
220
if (next) {
drivers/staging/media/imx/imx-ic-prpencvf.c
221
phys = vb2_dma_contig_plane_dma_addr(&next->vbuf.vb2_buf, 0);
drivers/staging/media/imx/imx-ic-prpencvf.c
222
priv->active_vb2_buf[priv->ipu_buf_num] = next;
drivers/staging/media/imx/imx-media-capture.c
847
buf = list_entry(priv->ready_q.next, struct imx_media_buffer,
drivers/staging/media/imx/imx-media-csi.c
266
struct imx_media_buffer *done, *next;
drivers/staging/media/imx/imx-media-csi.c
284
next = imx_media_capture_device_next_buf(vdev);
drivers/staging/media/imx/imx-media-csi.c
285
if (next) {
drivers/staging/media/imx/imx-media-csi.c
286
phys = vb2_dma_contig_plane_dma_addr(&next->vbuf.vb2_buf, 0);
drivers/staging/media/imx/imx-media-csi.c
287
priv->active_vb2_buf[priv->ipu_buf_num] = next;
drivers/staging/media/ipu3/ipu3-css-fw.h
102
struct imgu_fw_binary_xinfo *next __aligned(8);
drivers/staging/media/ipu3/ipu3-css-fw.h
159
u64 next;
drivers/staging/most/dim2/dim2.c
195
list_del(head->next);
drivers/staging/most/dim2/dim2.c
203
list_move_tail(head->next, &hdm_ch->started_list);
drivers/staging/most/dim2/dim2.c
302
list_del(head->next);
drivers/staging/most/dim2/dim2.c
429
list_del(head->next);
drivers/staging/octeon/ethernet-tx.c
107
to_free_list = to_free_list->next;
drivers/staging/octeon/ethernet-tx.c
442
skb->next = to_free_list;
drivers/staging/octeon/ethernet-tx.c
459
t->next = to_free_list;
drivers/staging/octeon/ethernet-tx.c
470
to_free_list = to_free_list->next;
drivers/staging/octeon/ethernet-tx.c
97
t->next = to_free_list;
drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c
366
next:
drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c
390
goto next;
drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c
397
goto next;
drivers/staging/rtl8723bs/include/osdep_service_linux.h
48
return list->next;
drivers/target/iscsi/cxgbit/cxgbit.h
120
struct np_info *next;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
105
for (p = cdev->np_hash_tab[bucket]; p; p = p->next) {
drivers/target/iscsi/cxgbit/cxgbit_cm.c
122
for (p = *prev; p; prev = &p->next, p = p->next) {
drivers/target/iscsi/cxgbit/cxgbit_cm.c
125
*prev = p->next;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
90
p->next = cdev->np_hash_tab[bucket];
drivers/target/iscsi/iscsi_target_tmr.c
639
goto next;
drivers/target/iscsi/iscsi_target_tmr.c
647
goto next;
drivers/target/iscsi/iscsi_target_tmr.c
689
goto next;
drivers/target/iscsi/iscsi_target_tmr.c
701
next:
drivers/target/target_core_alua.c
1711
struct se_lun *lun, *next;
drivers/target/target_core_alua.c
1742
list_for_each_entry_safe(lun, next,
drivers/target/target_core_pr.c
2478
list_entry(dev->t10_pr.registration_list.next,
drivers/target/target_core_pr.c
705
struct se_lun *lun_tmp, *next, *dest_lun;
drivers/target/target_core_pr.c
728
list_for_each_entry_safe(lun_tmp, next, &dev->dev_sep_list, lun_dev_link) {
drivers/target/target_core_sbc.c
1410
goto next;
drivers/target/target_core_sbc.c
1436
next:
drivers/target/target_core_spc.c
182
bool next = true;
drivers/target/target_core_spc.c
209
if (next) {
drivers/target/target_core_spc.c
210
next = false;
drivers/target/target_core_spc.c
213
next = true;
drivers/target/target_core_tmr.c
111
struct se_cmd *se_cmd, *next;
drivers/target/target_core_tmr.c
121
list_for_each_entry_safe(se_cmd, next, &dev->queues[i].state_list,
drivers/target/target_core_tmr.c
274
struct se_cmd *cmd, *next;
drivers/target/target_core_tmr.c
304
list_for_each_entry_safe(cmd, next, &dev->queues[i].state_list,
drivers/target/target_core_tmr.c
342
cmd = list_entry(drain_task_list.next, struct se_cmd, state_list);
drivers/target/target_core_transport.c
2329
cmd = list_entry(dev->delayed_cmd_list.next,
drivers/thermal/thermal_core.c
1403
struct thermal_trip_desc *td, *next;
drivers/thermal/thermal_core.c
1421
list_for_each_entry_safe(td, next, &tz->trips_invalid, list_node) {
drivers/thermal/thermal_core.c
1426
list_for_each_entry_safe(td, next, &tz->trips_reached, list_node) {
drivers/thermal/thermal_core.c
567
struct thermal_trip_desc *td, *next;
drivers/thermal/thermal_core.c
571
list_for_each_entry_safe_reverse(td, next, &tz->trips_reached, list_node) {
drivers/thermal/thermal_core.c
583
list_for_each_entry_safe(td, next, &tz->trips_high, list_node) {
drivers/thermal/thermal_core.c
591
list_for_each_entry_safe(td, next, &way_down_list, list_node)
drivers/thermal/thermal_core.c
937
struct thermal_instance *pos, *next;
drivers/thermal/thermal_core.c
939
list_for_each_entry_safe(pos, next, &td->thermal_instances, trip_node) {
drivers/thermal/thermal_debugfs.c
342
.next = cdev_seq_next,
drivers/thermal/thermal_debugfs.c
375
.next = cdev_seq_next,
drivers/thermal/thermal_debugfs.c
859
.next = tze_seq_next,
drivers/thunderbolt/cap.c
168
ret = header.basic.next;
drivers/thunderbolt/cap.c
173
ret = header.extended_long.next;
drivers/thunderbolt/cap.c
175
ret = header.extended_short.next;
drivers/thunderbolt/cap.c
88
return header.basic.next;
drivers/thunderbolt/switch.c
867
struct tb_port *next;
drivers/thunderbolt/switch.c
879
next = tb_port_at(tb_route(end->sw), prev->sw);
drivers/thunderbolt/switch.c
882
(next == prev || next->dual_link_port == prev))
drivers/thunderbolt/switch.c
883
next = prev->remote;
drivers/thunderbolt/switch.c
886
next = prev->remote;
drivers/thunderbolt/switch.c
888
next = tb_upstream_port(prev->sw);
drivers/thunderbolt/switch.c
893
if (next->dual_link_port &&
drivers/thunderbolt/switch.c
894
next->link_nr != prev->link_nr) {
drivers/thunderbolt/switch.c
895
next = next->dual_link_port;
drivers/thunderbolt/switch.c
900
return next != prev ? next : NULL;
drivers/thunderbolt/tb_regs.h
63
u8 next;
drivers/thunderbolt/tb_regs.h
77
u8 next;
drivers/thunderbolt/tb_regs.h
97
u16 next;
drivers/tty/goldfish.c
90
unsigned long next =
drivers/tty/goldfish.c
92
unsigned long avail = next - addr;
drivers/tty/hvc/hvc_console.c
257
list_del(&(hp->next));
drivers/tty/hvc/hvc_console.c
803
list_for_each_entry(hp, &hvc_structs, next) {
drivers/tty/hvc/hvc_console.c
967
list_add_tail(&(hp->next), &hvc_structs);
drivers/tty/hvc/hvc_console.c
97
list_for_each_entry(hp, &hvc_structs, next) {
drivers/tty/hvc/hvc_console.h
48
struct list_head next;
drivers/tty/hvc/hvc_iucv.c
190
struct iucv_tty_buffer *ent, *next;
drivers/tty/hvc/hvc_iucv.c
192
list_for_each_entry_safe(ent, next, list, list) {
drivers/tty/hvc/hvc_iucv.c
952
struct iucv_tty_buffer *ent, *next;
drivers/tty/hvc/hvc_iucv.c
956
list_for_each_entry_safe(ent, next, &priv->tty_outqueue, list)
drivers/tty/hvc/hvcs.c
1043
list_for_each_entry(hvcsd, &hvcs_structs, next) {
drivers/tty/hvc/hvcs.c
287
struct list_head next; /* list management */
drivers/tty/hvc/hvcs.c
621
list_for_each_entry(hvcsd, &hvcs_structs, next) {
drivers/tty/hvc/hvcs.c
675
list_del(&(hvcsd->next));
drivers/tty/hvc/hvcs.c
785
list_add_tail(&(hvcsd->next), &hvcs_structs);
drivers/tty/hvc/hvcs.c
905
list_for_each_entry(hvcsd, &hvcs_structs, next) {
drivers/tty/serial/8250/8250_core.c
114
i->head = i->head->next;
drivers/tty/serial/8250/8250_core.c
92
l = l->next;
drivers/tty/serial/8250/8250_hp300.c
273
port->next = hp300_ports;
drivers/tty/serial/8250/8250_hp300.c
29
struct hp300_port *next; /* next port */
drivers/tty/serial/8250/8250_hp300.c
309
port = port->next;
drivers/tty/serial/pch_uart.c
1009
int next = 1;
drivers/tty/serial/pch_uart.c
1014
while (next) {
drivers/tty/serial/pch_uart.c
1056
next = 0; /* MS ir prioirty is the lowest. So, MS ir
drivers/tty/serial/pch_uart.c
1066
next = 0;
drivers/tty/serial/sunzilog.c
1100
up->next = up + 1;
drivers/tty/serial/sunzilog.c
1102
up->next = NULL;
drivers/tty/serial/sunzilog.c
1590
up = up->next;
drivers/tty/serial/sunzilog.c
1627
up = up->next;
drivers/tty/serial/sunzilog.c
557
up = up->next;
drivers/tty/serial/sunzilog.c
579
up = up->next;
drivers/tty/serial/sunzilog.c
74
struct uart_sunzilog_port *next;
drivers/tty/synclink_gt.c
160
struct cond_wait *next;
drivers/tty/synclink_gt.c
175
__le32 next; /* physical address of next descriptor */
drivers/tty/synclink_gt.c
185
#define set_desc_next(a,b) (a).next = cpu_to_le32((unsigned int)(b))
drivers/tty/synclink_gt.c
2251
for (w = info->gpio_wait_q, prev = NULL ; w != NULL ; w = w->next) {
drivers/tty/synclink_gt.c
2256
prev->next = w->next;
drivers/tty/synclink_gt.c
2258
info->gpio_wait_q = w->next;
drivers/tty/synclink_gt.c
2929
w->next = *head;
drivers/tty/synclink_gt.c
2938
for (w = *head, prev = NULL ; w != NULL ; prev = w, w = w->next) {
drivers/tty/synclink_gt.c
2941
prev->next = w->next;
drivers/tty/synclink_gt.c
2943
*head = w->next;
drivers/tty/synclink_gt.c
2953
*head = (*head)->next;
drivers/tty/synclink_gt.c
3285
info->rbufs[i].next = cpu_to_le32(pbufs);
drivers/tty/synclink_gt.c
3287
info->rbufs[i].next = cpu_to_le32(pbufs + ((i+1) * sizeof(struct slgt_desc)));
drivers/tty/synclink_gt.c
3297
info->tbufs[i].next = cpu_to_le32(pbufs + info->rbuf_count * sizeof(struct slgt_desc));
drivers/tty/synclink_gt.c
3299
info->tbufs[i].next = cpu_to_le32(pbufs + ((info->rbuf_count + i + 1) * sizeof(struct slgt_desc)));
drivers/tty/tty_buffer.c
106
p->next = NULL;
drivers/tty/tty_buffer.c
123
struct tty_buffer *p, *next;
drivers/tty/tty_buffer.c
129
buf->head = p->next;
drivers/tty/tty_buffer.c
135
llist_for_each_entry_safe(p, next, llist, free)
drivers/tty/tty_buffer.c
225
struct tty_buffer *next;
drivers/tty/tty_buffer.c
233
while ((next = smp_load_acquire(&buf->head->next)) != NULL) {
drivers/tty/tty_buffer.c
235
buf->head = next;
drivers/tty/tty_buffer.c
289
smp_store_release(&b->next, n);
drivers/tty/tty_buffer.c
402
struct tty_buffer *next;
drivers/tty/tty_buffer.c
410
next = smp_load_acquire(&head->next);
drivers/tty/tty_buffer.c
417
head = next;
drivers/tty/tty_buffer.c
471
struct tty_buffer *next;
drivers/tty/tty_buffer.c
482
next = smp_load_acquire(&head->next);
drivers/tty/tty_buffer.c
488
if (next == NULL)
drivers/tty/tty_buffer.c
490
buf->head = next;
drivers/tty/tty_ldisc.c
217
.next = tty_ldiscs_seq_next,
drivers/tty/tty_ldsem.c
123
waiter = list_entry(sem->write_wait.next, struct ldsem_waiter, list);
drivers/tty/tty_ldsem.c
76
struct ldsem_waiter *waiter, *next;
drivers/tty/tty_ldsem.c
94
list_for_each_entry_safe(waiter, next, &sem->read_wait, list) {
drivers/usb/atm/usbatm.c
206
urb = list_entry(channel->list.next, struct urb, urb_list);
drivers/usb/c67x00/c67x00-sched.c
1034
tmp = list_entry(td->td_list.next, typeof(*td), td_list);
drivers/usb/c67x00/c67x00-sched.c
1074
if (td->td_list.next == &c67x00->td_list)
drivers/usb/c67x00/c67x00-sched.c
788
urb = list_entry(ep_data->queue.next,
drivers/usb/c67x00/c67x00-sched.c
895
while (td->td_list.next != &c67x00->td_list) {
drivers/usb/c67x00/c67x00-sched.c
896
td = list_entry(td->td_list.next, struct c67x00_td, td_list);
drivers/usb/cdns3/cdnsp-gadget.c
1411
seg = seg->next;
drivers/usb/cdns3/cdnsp-gadget.c
1452
event_deq_seg = event_deq_seg->next;
drivers/usb/cdns3/cdnsp-gadget.c
536
segment = pdev->event_ring->deq_seg->next;
drivers/usb/cdns3/cdnsp-gadget.c
618
event_deq_seg = event_deq_seg->next;
drivers/usb/cdns3/cdnsp-gadget.c
75
u32 next;
drivers/usb/cdns3/cdnsp-gadget.c
96
next = EXT_CAPS_NEXT(val);
drivers/usb/cdns3/cdnsp-gadget.c
97
offset += next << 2;
drivers/usb/cdns3/cdnsp-gadget.c
98
} while (next);
drivers/usb/cdns3/cdnsp-gadget.h
1182
struct cdnsp_segment *next;
drivers/usb/cdns3/cdnsp-mem.c
1047
seg = seg->next;
drivers/usb/cdns3/cdnsp-mem.c
107
struct cdnsp_segment *next,
drivers/usb/cdns3/cdnsp-mem.c
113
if (!prev || !next)
drivers/usb/cdns3/cdnsp-mem.c
116
prev->next = next;
drivers/usb/cdns3/cdnsp-mem.c
119
link->segment_ptr = cpu_to_le64(next->dma);
drivers/usb/cdns3/cdnsp-mem.c
142
struct cdnsp_segment *next;
drivers/usb/cdns3/cdnsp-mem.c
147
next = ring->enq_seg->next;
drivers/usb/cdns3/cdnsp-mem.c
149
cdnsp_link_segments(pdev, last, next, ring->type);
drivers/usb/cdns3/cdnsp-mem.c
245
seg = seg->next;
drivers/usb/cdns3/cdnsp-mem.c
257
seg = seg->next;
drivers/usb/cdns3/cdnsp-mem.c
270
seg = seg->next;
drivers/usb/cdns3/cdnsp-mem.c
343
struct cdnsp_segment *next;
drivers/usb/cdns3/cdnsp-mem.c
345
next = cdnsp_segment_alloc(pdev, cycle_state,
drivers/usb/cdns3/cdnsp-mem.c
347
if (!next) {
drivers/usb/cdns3/cdnsp-mem.c
352
cdnsp_link_segments(pdev, prev, next, type);
drivers/usb/cdns3/cdnsp-mem.c
354
prev = next;
drivers/usb/cdns3/cdnsp-mem.c
60
seg->next = NULL;
drivers/usb/cdns3/cdnsp-mem.c
86
seg = first->next;
drivers/usb/cdns3/cdnsp-mem.c
89
struct cdnsp_segment *next = seg->next;
drivers/usb/cdns3/cdnsp-mem.c
92
seg = next;
drivers/usb/cdns3/cdnsp-ring.c
101
return cdnsp_last_trb_on_seg(seg, trb) && (seg->next == ring->first_seg);
drivers/usb/cdns3/cdnsp-ring.c
1028
td = list_entry(ep_ring->td_list.next, struct cdnsp_td,
drivers/usb/cdns3/cdnsp-ring.c
1333
td = list_entry(ep_ring->td_list.next, struct cdnsp_td,
drivers/usb/cdns3/cdnsp-ring.c
135
*seg = (*seg)->next;
drivers/usb/cdns3/cdnsp-ring.c
158
ring->deq_seg = ring->deq_seg->next;
drivers/usb/cdns3/cdnsp-ring.c
1670
ep_ring->enq_seg = ep_ring->enq_seg->next;
drivers/usb/cdns3/cdnsp-ring.c
169
ring->deq_seg = ring->deq_seg->next;
drivers/usb/cdns3/cdnsp-ring.c
191
union cdnsp_trb *next;
drivers/usb/cdns3/cdnsp-ring.c
199
next = ++(ring->enqueue);
drivers/usb/cdns3/cdnsp-ring.c
202
while (cdnsp_trb_is_link(next)) {
drivers/usb/cdns3/cdnsp-ring.c
213
next->link.control &= cpu_to_le32(~TRB_CHAIN);
drivers/usb/cdns3/cdnsp-ring.c
214
next->link.control |= cpu_to_le32(chain);
drivers/usb/cdns3/cdnsp-ring.c
218
next->link.control ^= cpu_to_le32(TRB_CYCLE);
drivers/usb/cdns3/cdnsp-ring.c
221
if (cdnsp_link_trb_toggles_cycle(next))
drivers/usb/cdns3/cdnsp-ring.c
224
ring->enq_seg = ring->enq_seg->next;
drivers/usb/cdns3/cdnsp-ring.c
226
next = ring->enqueue;
drivers/usb/cdns3/cdnsp-ring.c
591
cur_seg = cur_seg->next;
drivers/usb/cdns3/cdnsp-ring.c
661
ep_ring->deq_seg = ep_ring->deq_seg->next;
drivers/usb/cdns3/cdnsp-ring.c
673
ep_ring->deq_seg = ep_ring->deq_seg->next;
drivers/usb/chipidea/otg_fsm.c
123
char *next;
drivers/usb/chipidea/otg_fsm.c
127
next = buf;
drivers/usb/chipidea/otg_fsm.c
129
t = scnprintf(next, size, "%d\n", ci->fsm.b_bus_req);
drivers/usb/chipidea/otg_fsm.c
131
next += t;
drivers/usb/chipidea/otg_fsm.c
34
char *next;
drivers/usb/chipidea/otg_fsm.c
38
next = buf;
drivers/usb/chipidea/otg_fsm.c
40
t = scnprintf(next, size, "%d\n", ci->fsm.a_bus_req);
drivers/usb/chipidea/otg_fsm.c
42
next += t;
drivers/usb/chipidea/otg_fsm.c
83
char *next;
drivers/usb/chipidea/otg_fsm.c
87
next = buf;
drivers/usb/chipidea/otg_fsm.c
89
t = scnprintf(next, size, "%d\n", ci->fsm.a_bus_drop);
drivers/usb/chipidea/otg_fsm.c
91
next += t;
drivers/usb/chipidea/trace.h
49
__field(u32, next)
drivers/usb/chipidea/trace.h
59
__entry->next = le32_to_cpu(td->ptr->next);
drivers/usb/chipidea/trace.h
66
__entry->td_remaining_size, __entry->next,
drivers/usb/chipidea/udc.c
1577
hwep->qh.ptr->td.next |= cpu_to_le32(TD_TERMINATE); /* needed? */
drivers/usb/chipidea/udc.c
409
lastnode->ptr->next = cpu_to_le32(node->dma);
drivers/usb/chipidea/udc.c
719
lastnode->ptr->next = cpu_to_le32(TD_TERMINATE);
drivers/usb/chipidea/udc.c
736
u32 next = firstnode->dma & TD_ADDR_MASK;
drivers/usb/chipidea/udc.c
743
prevlastnode->ptr->next = cpu_to_le32(next);
drivers/usb/chipidea/udc.c
770
hwep->qh.ptr->td.next = cpu_to_le32(firstnode->dma);
drivers/usb/chipidea/udc.c
805
hwep->qh.ptr->td.next = cpu_to_le32(node->dma);
drivers/usb/chipidea/udc.c
921
struct ci_hw_req *hwreq = list_entry(hwep->qh.queue.next,
drivers/usb/chipidea/udc.h
22
__le32 next;
drivers/usb/core/devio.c
2867
ps = list_entry(udev->filelist.next, struct usb_dev_state, list);
drivers/usb/core/devio.c
467
as = list_entry(ps->async_completed.next, struct async,
drivers/usb/core/hcd.c
1688
urb = list_entry(local_list.next, struct urb, urb_list);
drivers/usb/core/hub.c
845
struct list_head *next;
drivers/usb/core/hub.c
851
next = hub->tt.clear_list.next;
drivers/usb/core/hub.c
852
clear = list_entry(next, struct usb_tt_clear, clear_list);
drivers/usb/core/urb.c
967
victim = list_entry(anchor->urb_list.next, struct urb,
drivers/usb/dwc2/debugfs.c
260
ep->queue.next, ep->queue.prev);
drivers/usb/dwc2/gadget.c
4866
u32 next = DXEPCTL_NEXTEP((epnum + 1) % 15);
drivers/usb/dwc2/gadget.c
4869
dwc2_writel(hsotg, next, DIEPCTL(epnum));
drivers/usb/dwc2/gadget.c
4871
dwc2_writel(hsotg, next, DOEPCTL(epnum));
drivers/usb/dwc2/hcd.c
1592
if (!qh_list->next)
drivers/usb/dwc2/hcd.c
2694
qh_ptr = hsotg->periodic_sched_ready.next;
drivers/usb/dwc2/hcd.c
2714
qh_ptr = qh_ptr->next;
drivers/usb/dwc2/hcd.c
2726
qh_ptr = hsotg->non_periodic_sched_inactive.next;
drivers/usb/dwc2/hcd.c
2751
qh_ptr = qh_ptr->next;
drivers/usb/dwc2/hcd.c
2882
qh_ptr = hsotg->periodic_sched_assigned.next;
drivers/usb/dwc2/hcd.c
2894
qh_ptr = qh_ptr->next;
drivers/usb/dwc2/hcd.c
2900
qh_ptr = qh_ptr->next;
drivers/usb/dwc2/hcd.c
2930
qh_ptr = qh_ptr->next;
drivers/usb/dwc2/hcd.c
3014
hsotg->non_periodic_qh_ptr = hsotg->non_periodic_qh_ptr->next;
drivers/usb/dwc2/hcd.c
3033
goto next;
drivers/usb/dwc2/hcd.c
3037
goto next;
drivers/usb/dwc2/hcd.c
3049
next:
drivers/usb/dwc2/hcd.c
3051
hsotg->non_periodic_qh_ptr = hsotg->non_periodic_qh_ptr->next;
drivers/usb/dwc2/hcd.c
3055
hsotg->non_periodic_qh_ptr->next;
drivers/usb/dwc2/hcd_intr.c
128
qh_entry = hsotg->periodic_sched_inactive.next;
drivers/usb/dwc2/hcd_intr.c
131
qh_entry = qh_entry->next;
drivers/usb/dwc2/hcd_queue.c
1722
hsotg->non_periodic_qh_ptr->next;
drivers/usb/dwc3/gadget.c
4345
enum dwc3_link_state next = evtinfo & DWC3_LINK_STATE_MASK;
drivers/usb/dwc3/gadget.c
4371
(next == DWC3_LINK_STATE_RESUME)) {
drivers/usb/dwc3/gadget.c
4395
if (next == DWC3_LINK_STATE_U0) {
drivers/usb/dwc3/gadget.c
4422
switch (next) {
drivers/usb/dwc3/gadget.c
4446
dwc->link_state = next;
drivers/usb/dwc3/gadget.c
4449
while (dwc->wakeup_pending_funcs && (next == DWC3_LINK_STATE_U0)) {
drivers/usb/dwc3/gadget.c
4464
enum dwc3_link_state next = evtinfo & DWC3_LINK_STATE_MASK;
drivers/usb/dwc3/gadget.c
4466
if (!dwc->suspended && next == DWC3_LINK_STATE_U3) {
drivers/usb/dwc3/gadget.c
4471
dwc->link_state = next;
drivers/usb/fotg210/fotg210-hcd.c
2067
qtd = list_entry(qh->qtd_list.next,
drivers/usb/fotg210/fotg210-hcd.c
2955
qtd = list_entry(qtd_list->next, struct fotg210_qtd,
drivers/usb/fotg210/fotg210-hcd.c
3024
qtd = list_entry(qtd_list->next, struct fotg210_qtd, qtd_list);
drivers/usb/fotg210/fotg210-hcd.c
3153
struct fotg210_qh *qh, *next;
drivers/usb/fotg210/fotg210-hcd.c
3158
next = fotg210->async->qh_next.qh;
drivers/usb/fotg210/fotg210-hcd.c
3159
while (next) {
drivers/usb/fotg210/fotg210-hcd.c
3160
qh = next;
drivers/usb/fotg210/fotg210-hcd.c
3161
next = qh->qh_next.qh;
drivers/usb/fotg210/fotg210-hcd.c
3584
fotg210->qh_scan_next = list_entry(qh->intr_node.next,
drivers/usb/fotg210/fotg210-hcd.c
361
char *next = *nextp;
drivers/usb/fotg210/fotg210-hcd.c
380
temp = scnprintf(next, size,
drivers/usb/fotg210/fotg210-hcd.c
391
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
407
temp = scnprintf(next, size,
drivers/usb/fotg210/fotg210-hcd.c
4157
u32 now, next, start, period, span;
drivers/usb/fotg210/fotg210-hcd.c
4187
next = now + fotg210->i_thresh;
drivers/usb/fotg210/fotg210-hcd.c
4189
next = now;
drivers/usb/fotg210/fotg210-hcd.c
4195
excess = (stream->next_uframe - period - next) & (mod - 1);
drivers/usb/fotg210/fotg210-hcd.c
4197
start = next + excess - mod + period *
drivers/usb/fotg210/fotg210-hcd.c
4200
start = next + excess + period;
drivers/usb/fotg210/fotg210-hcd.c
4228
next = start;
drivers/usb/fotg210/fotg210-hcd.c
4236
} while (start > next && !done);
drivers/usb/fotg210/fotg210-hcd.c
428
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
431
temp = scnprintf(next, size, "\n");
drivers/usb/fotg210/fotg210-hcd.c
434
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
437
*nextp = next;
drivers/usb/fotg210/fotg210-hcd.c
4372
itd = list_entry(iso_sched->td_list.next,
drivers/usb/fotg210/fotg210-hcd.c
446
char *next;
drivers/usb/fotg210/fotg210-hcd.c
451
next = buf->output_buf;
drivers/usb/fotg210/fotg210-hcd.c
454
*next = 0;
drivers/usb/fotg210/fotg210-hcd.c
463
qh_lines(fotg210, qh, &next, &size);
drivers/usb/fotg210/fotg210-hcd.c
465
temp = scnprintf(next, size, "\nunlink =\n");
drivers/usb/fotg210/fotg210-hcd.c
467
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
471
qh_lines(fotg210, qh, &next, &size);
drivers/usb/fotg210/fotg210-hcd.c
514
char *next;
drivers/usb/fotg210/fotg210-hcd.c
526
next = buf->output_buf;
drivers/usb/fotg210/fotg210-hcd.c
529
temp = scnprintf(next, size, "size = %d\n", fotg210->periodic_size);
drivers/usb/fotg210/fotg210-hcd.c
531
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
544
temp = scnprintf(next, size, "%4d: ", i);
drivers/usb/fotg210/fotg210-hcd.c
546
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
554
temp = scnprintf(next, size, " qh%d-%04x/%p",
drivers/usb/fotg210/fotg210-hcd.c
562
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
568
temp = scnprintf(next, size,
drivers/usb/fotg210/fotg210-hcd.c
571
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
577
temp = output_buf_tds_dir(next,
drivers/usb/fotg210/fotg210-hcd.c
589
temp = scnprintf(next, size,
drivers/usb/fotg210/fotg210-hcd.c
596
temp = scnprintf(next, size,
drivers/usb/fotg210/fotg210-hcd.c
603
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
606
temp = scnprintf(next, size, "\n");
drivers/usb/fotg210/fotg210-hcd.c
608
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
638
char *next, scratch[80];
drivers/usb/fotg210/fotg210-hcd.c
644
next = buf->output_buf;
drivers/usb/fotg210/fotg210-hcd.c
650
size = scnprintf(next, size,
drivers/usb/fotg210/fotg210-hcd.c
663
temp = scnprintf(next, size,
drivers/usb/fotg210/fotg210-hcd.c
672
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
676
temp = scnprintf(next, size, "structural params 0x%08x\n", i);
drivers/usb/fotg210/fotg210-hcd.c
678
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
681
temp = scnprintf(next, size, "capability params 0x%08x\n", i);
drivers/usb/fotg210/fotg210-hcd.c
683
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
688
temp = scnprintf(next, size, fmt, temp, scratch);
drivers/usb/fotg210/fotg210-hcd.c
690
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
694
temp = scnprintf(next, size, fmt, temp, scratch);
drivers/usb/fotg210/fotg210-hcd.c
696
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
700
temp = scnprintf(next, size, fmt, temp, scratch);
drivers/usb/fotg210/fotg210-hcd.c
702
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
704
temp = scnprintf(next, size, "uframe %04x\n",
drivers/usb/fotg210/fotg210-hcd.c
707
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
710
temp = scnprintf(next, size, "async unlink qh %p\n",
drivers/usb/fotg210/fotg210-hcd.c
713
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
717
temp = scnprintf(next, size,
drivers/usb/fotg210/fotg210-hcd.c
722
next += temp;
drivers/usb/fotg210/fotg210-hcd.c
724
temp = scnprintf(next, size, "complete %ld unlink %ld\n",
drivers/usb/fotg210/fotg210-hcd.c
727
next += temp;
drivers/usb/fotg210/fotg210-udc.c
232
req = list_entry(ep->queue.next,
drivers/usb/fotg210/fotg210-udc.c
842
req = list_entry(ep->queue.next,
drivers/usb/fotg210/fotg210-udc.c
857
struct fotg210_request *req = list_entry(ep->queue.next,
drivers/usb/fotg210/fotg210-udc.c
867
struct fotg210_request *req = list_entry(ep->queue.next,
drivers/usb/gadget/composite.c
1391
int next = cdev->next_string_id;
drivers/usb/gadget/composite.c
1394
if (unlikely(next >= 254))
drivers/usb/gadget/composite.c
1396
str->id = ++next;
drivers/usb/gadget/composite.c
1399
cdev->next_string_id = next;
drivers/usb/gadget/composite.c
1539
unsigned next = c->next_string_id;
drivers/usb/gadget/composite.c
1540
if (unlikely(n > 254 || (unsigned)next + n > 254))
drivers/usb/gadget/composite.c
1543
return next + 1;
drivers/usb/gadget/composite.c
567
void *next = buf + USB_DT_CONFIG_SIZE;
drivers/usb/gadget/composite.c
586
status = usb_descriptor_fillbuf(next, len,
drivers/usb/gadget/composite.c
591
next += status;
drivers/usb/gadget/composite.c
601
status = usb_descriptor_fillbuf(next, len,
drivers/usb/gadget/composite.c
606
next += status;
drivers/usb/gadget/composite.c
609
len = next - buf;
drivers/usb/gadget/composite.c
643
while ((pos = pos->next) != &cdev->configs) {
drivers/usb/gadget/function/f_eem.c
440
goto next;
drivers/usb/gadget/function/f_eem.c
450
goto next;
drivers/usb/gadget/function/f_eem.c
457
goto next;
drivers/usb/gadget/function/f_eem.c
465
goto next;
drivers/usb/gadget/function/f_eem.c
525
goto next;
drivers/usb/gadget/function/f_eem.c
531
goto next;
drivers/usb/gadget/function/f_eem.c
541
goto next;
drivers/usb/gadget/function/f_eem.c
546
next:
drivers/usb/gadget/function/f_hid.c
981
struct f_hidg_req_list *list, *next;
drivers/usb/gadget/function/f_hid.c
990
list_for_each_entry_safe(list, next, &hidg->completed_out_req, list) {
drivers/usb/gadget/function/f_mass_storage.c
1538
common->next_buffhd_to_fill = bh2->next;
drivers/usb/gadget/function/f_mass_storage.c
1550
common->next_buffhd_to_drain = bh->next;
drivers/usb/gadget/function/f_mass_storage.c
1603
common->next_buffhd_to_fill = bh->next;
drivers/usb/gadget/function/f_mass_storage.c
1616
common->next_buffhd_to_fill = bh->next;
drivers/usb/gadget/function/f_mass_storage.c
1713
common->next_buffhd_to_fill = bh->next;
drivers/usb/gadget/function/f_mass_storage.c
2760
bh->next = bh + 1;
drivers/usb/gadget/function/f_mass_storage.c
2767
bh->next = buffhds;
drivers/usb/gadget/function/f_mass_storage.c
729
common->next_buffhd_to_fill = bh->next;
drivers/usb/gadget/function/f_mass_storage.c
837
common->next_buffhd_to_fill = bh->next;
drivers/usb/gadget/function/f_mass_storage.c
851
common->next_buffhd_to_drain = bh->next;
drivers/usb/gadget/function/f_printer.c
1173
req = container_of(dev->rx_reqs.next, struct usb_request, list);
drivers/usb/gadget/function/f_printer.c
1180
req = container_of(dev->tx_reqs.next, struct usb_request, list);
drivers/usb/gadget/function/f_printer.c
1447
req = container_of(dev->tx_reqs.next, struct usb_request,
drivers/usb/gadget/function/f_printer.c
1457
req = container_of(dev->rx_reqs.next,
drivers/usb/gadget/function/f_printer.c
1464
req = container_of(dev->rx_buffers.next,
drivers/usb/gadget/function/f_printer.c
402
req = container_of(dev->rx_reqs.next,
drivers/usb/gadget/function/f_printer.c
506
req = container_of(dev->rx_buffers.next,
drivers/usb/gadget/function/f_printer.c
635
req = container_of(dev->tx_reqs.next, struct usb_request,
drivers/usb/gadget/function/f_printer.c
924
req = container_of(dev->rx_buffers.next, struct usb_request,
drivers/usb/gadget/function/f_printer.c
931
req = container_of(dev->rx_buffers.next, struct usb_request,
drivers/usb/gadget/function/f_printer.c
938
req = container_of(dev->tx_reqs_active.next,
drivers/usb/gadget/function/storage_common.h
146
struct fsg_buffhd *next;
drivers/usb/gadget/function/u_ether.c
341
struct list_head *next;
drivers/usb/gadget/function/u_ether.c
343
next = req->list.next;
drivers/usb/gadget/function/u_ether.c
347
if (next == list)
drivers/usb/gadget/function/u_ether.c
350
req = container_of(next, struct usb_request, list);
drivers/usb/gadget/function/u_serial.c
256
req = list_entry(pool->next, struct usb_request, list);
drivers/usb/gadget/function/u_serial.c
328
req = list_entry(pool->next, struct usb_request, list);
drivers/usb/gadget/function/u_serial.c
503
req = list_entry(head->next, struct usb_request, list);
drivers/usb/gadget/udc/amd5536udc.h
472
u32 next;
drivers/usb/gadget/udc/aspeed_udc.c
312
req = list_entry(ep->queue.next, struct ast_udc_request,
drivers/usb/gadget/udc/aspeed_udc.c
796
req = list_entry(ep->queue.next, struct ast_udc_request, queue);
drivers/usb/gadget/udc/aspeed_udc.c
841
req = list_entry(ep->queue.next, struct ast_udc_request, queue);
drivers/usb/gadget/udc/at91_udc.c
1013
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/at91_udc.c
1300
req = list_entry(ep0->queue.next, struct at91_request, queue);
drivers/usb/gadget/udc/at91_udc.c
465
req = list_entry(ep->queue.next, struct at91_request, queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1443
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/atmel_usba_udc.c
1636
req = list_entry(ep->queue.next, struct usba_request, queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1700
req = list_entry(ep->queue.next, struct usba_request, queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
470
req = list_entry(ep->queue.next, struct usba_request, queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
500
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/atmel_usba_udc.c
888
if (ep->queue.next == &req->queue) {
drivers/usb/gadget/udc/atmel_usba_udc.h
257
dma_addr_t next;
drivers/usb/gadget/udc/bcm63xx_udc.c
1225
struct bcm63xx_req *next;
drivers/usb/gadget/udc/bcm63xx_udc.c
1227
next = list_first_entry(&bep->queue,
drivers/usb/gadget/udc/bcm63xx_udc.c
1229
iudma_write(udc, bep->iudma, next);
drivers/usb/gadget/udc/bcm63xx_udc.c
2091
struct bcm63xx_req *next;
drivers/usb/gadget/udc/bcm63xx_udc.c
2093
next = list_first_entry(&bep->queue,
drivers/usb/gadget/udc/bcm63xx_udc.c
2095
iudma_write(udc, iudma, next);
drivers/usb/gadget/udc/bdc/bdc_ep.c
1098
req = list_entry(ep->queue.next, struct bdc_req,
drivers/usb/gadget/udc/bdc/bdc_ep.c
575
req = list_entry(ep->queue.next, struct bdc_req,
drivers/usb/gadget/udc/bdc/bdc_ep.c
977
req = list_entry(ep->queue.next, struct bdc_req,
drivers/usb/gadget/udc/dummy_hcd.c
338
req = list_entry(ep->queue.next, struct dummy_request, queue);
drivers/usb/gadget/udc/fsl_qe_udc.c
1178
ep->tx_req = list_entry(ep->queue.next, struct qe_req,
drivers/usb/gadget/udc/fsl_qe_udc.c
133
req = list_entry(ep->queue.next, struct qe_req, queue);
drivers/usb/gadget/udc/fsl_qe_udc.c
910
req = list_entry(ep->queue.next, struct qe_req, queue);
drivers/usb/gadget/udc/fsl_udc_core.c
216
req = list_entry(ep->queue.next, struct fsl_req, queue);
drivers/usb/gadget/udc/fsl_udc_core.c
959
if (ep->queue.next == &req->queue) {
drivers/usb/gadget/udc/fsl_udc_core.c
964
if (req->queue.next != &ep->queue) {
drivers/usb/gadget/udc/fsl_udc_core.c
967
next_req = list_entry(req->queue.next, struct fsl_req,
drivers/usb/gadget/udc/goku_udc.c
492
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/goku_udc.c
527
req = list_entry(ep->queue.next, struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
607
req = list_entry(ep->queue.next, struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
634
req = list_entry(ep->queue.next, struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
652
req = list_entry(ep->queue.next, struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
804
req = list_entry(ep->queue.next, struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
847
if (ep->dma && ep->queue.next == &req->queue && !ep->stopped) {
drivers/usb/gadget/udc/goku_udc.c
875
req = list_entry(ep->queue.next, struct goku_request,
drivers/usb/gadget/udc/gr_udc.c
164
struct gr_dma_desc *next;
drivers/usb/gadget/udc/gr_udc.c
169
next = req->first_desc;
drivers/usb/gadget/udc/gr_udc.c
171
desc = next;
drivers/usb/gadget/udc/gr_udc.c
172
next = desc->next_desc;
drivers/usb/gadget/udc/gr_udc.c
257
struct gr_dma_desc *next;
drivers/usb/gadget/udc/gr_udc.c
259
next = req->first_desc;
drivers/usb/gadget/udc/gr_udc.c
260
if (!next)
drivers/usb/gadget/udc/gr_udc.c
264
desc = next;
drivers/usb/gadget/udc/gr_udc.c
265
next = desc->next_desc;
drivers/usb/gadget/udc/gr_udc.c
457
req->last_desc->next = desc->paddr;
drivers/usb/gadget/udc/gr_udc.h
115
u32 next;
drivers/usb/gadget/udc/lpc32xx_udc.c
1009
req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1429
req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1450
req = list_entry(ep0->queue.next, struct lpc32xx_request,
drivers/usb/gadget/udc/lpc32xx_udc.c
1487
req = list_entry(ep0->queue.next, struct lpc32xx_request,
drivers/usb/gadget/udc/lpc32xx_udc.c
1962
req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1988
req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
983
req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
1020
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/m66592-udc.c
1169
req = list_entry(ep->queue.next, struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
1318
req = list_entry(ep->queue.next, struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
733
req = list_entry(ep->queue.next, struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
907
req = list_entry(ep->queue.next, struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
915
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/m66592-udc.c
938
req = list_entry(ep->queue.next, struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
951
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
1135
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
1197
req = list_entry(ep->queue.next, struct net2280_request, queue);
drivers/usb/gadget/udc/net2280.c
1224
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
1276
if (ep->queue.next == &req->queue) {
drivers/usb/gadget/udc/net2280.c
1281
if (likely(ep->queue.next == &req->queue)) {
drivers/usb/gadget/udc/net2280.c
1308
start_dma(ep, list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
1654
char *next;
drivers/usb/gadget/udc/net2280.c
1662
next = buf;
drivers/usb/gadget/udc/net2280.c
1672
t = scnprintf(next, size, "%s version " DRIVER_VERSION
drivers/usb/gadget/udc/net2280.c
1686
next += t;
drivers/usb/gadget/udc/net2280.c
1701
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1707
next += t;
drivers/usb/gadget/udc/net2280.c
1723
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1745
next += t;
drivers/usb/gadget/udc/net2280.c
1747
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1756
next += t;
drivers/usb/gadget/udc/net2280.c
1761
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1770
next += t;
drivers/usb/gadget/udc/net2280.c
1777
t = scnprintf(next, size, "\nirqs: ");
drivers/usb/gadget/udc/net2280.c
1779
next += t;
drivers/usb/gadget/udc/net2280.c
1786
t = scnprintf(next, size, " %s/%lu", ep->ep.name, ep->irqs);
drivers/usb/gadget/udc/net2280.c
1788
next += t;
drivers/usb/gadget/udc/net2280.c
1791
t = scnprintf(next, size, "\n");
drivers/usb/gadget/udc/net2280.c
1793
next += t;
drivers/usb/gadget/udc/net2280.c
1805
char *next;
drivers/usb/gadget/udc/net2280.c
1811
next = buf;
drivers/usb/gadget/udc/net2280.c
1827
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1836
t = scnprintf(next, size, "ep0 max 64 pio %s\n",
drivers/usb/gadget/udc/net2280.c
1841
next += t;
drivers/usb/gadget/udc/net2280.c
1844
t = scnprintf(next, size, "\t(nothing queued)\n");
drivers/usb/gadget/udc/net2280.c
1848
next += t;
drivers/usb/gadget/udc/net2280.c
1853
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1860
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1867
next += t;
drivers/usb/gadget/udc/net2280.c
1873
t = scnprintf(next, size, "\t td %08x "
drivers/usb/gadget/udc/net2280.c
1882
next += t;
drivers/usb/gadget/udc/net2280.c
2524
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
2624
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
2736
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
3138
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
1028
if (use_dma && ep->dma_channel && ep->queue.next == &req->queue) {
drivers/usb/gadget/udc/omap_udc.c
1346
req = list_entry(ep->queue.next, struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
1422
req = container_of(ep0->queue.next, struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
1877
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
1918
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
1952
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
1985
req = list_entry(ep->queue.next, struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
647
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
654
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
666
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
673
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
790
req = container_of(ep->queue.next, struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
815
req = container_of(ep->queue.next, struct omap_req, queue);
drivers/usb/gadget/udc/pch_udc.c
1458
req = list_entry(ep->queue.next, struct pch_udc_request, queue);
drivers/usb/gadget/udc/pch_udc.c
1479
dma_addr_t addr = (dma_addr_t)td->next;
drivers/usb/gadget/udc/pch_udc.c
1480
td->next = 0x00;
drivers/usb/gadget/udc/pch_udc.c
1484
addr2 = (dma_addr_t)td->next;
drivers/usb/gadget/udc/pch_udc.c
1530
last->next = dma_addr;
drivers/usb/gadget/udc/pch_udc.c
1535
td->next = req->td_data_phys;
drivers/usb/gadget/udc/pch_udc.c
1621
td_data = phys_to_virt(td_data->next);
drivers/usb/gadget/udc/pch_udc.c
2024
req = list_entry(ep->queue.next, struct pch_udc_request, queue);
drivers/usb/gadget/udc/pch_udc.c
2038
td_data = phys_to_virt(td_data->next);
drivers/usb/gadget/udc/pch_udc.c
2058
req = list_entry(ep->queue.next, struct pch_udc_request, queue);
drivers/usb/gadget/udc/pch_udc.c
2102
req = list_entry(ep->queue.next, struct pch_udc_request, queue);
drivers/usb/gadget/udc/pch_udc.c
2128
addr = (dma_addr_t)td->next;
drivers/usb/gadget/udc/pch_udc.c
2142
req = list_entry(ep->queue.next, struct pch_udc_request, queue);
drivers/usb/gadget/udc/pch_udc.c
2209
req = list_entry(ep->queue.next, struct pch_udc_request,
drivers/usb/gadget/udc/pch_udc.c
222
u32 next;
drivers/usb/gadget/udc/pch_udc.c
2463
ep->td_data->next = ep->td_data_phys;
drivers/usb/gadget/udc/pxa25x_udc.c
1604
req = list_entry(ep->queue.next, struct pxa25x_request, queue);
drivers/usb/gadget/udc/pxa25x_udc.c
1824
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/pxa25x_udc.c
951
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/pxa27x_udc.c
1939
req = list_entry(ep->queue.next, struct pxa27x_request, queue);
drivers/usb/gadget/udc/pxa27x_udc.c
2035
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/pxa27x_udc.c
773
req = list_entry(ep->queue.next, struct pxa27x_request, queue);
drivers/usb/gadget/udc/r8a66597-udc.c
894
return list_entry(ep->queue.next, struct r8a66597_request, queue);
drivers/usb/gadget/udc/snps_udc_core.c
1000
td = phys_to_virt(td->next);
drivers/usb/gadget/udc/snps_udc_core.c
1017
td = phys_to_virt(td->next);
drivers/usb/gadget/udc/snps_udc_core.c
1241
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
1268
if (ep->queue.next == &req->queue) {
drivers/usb/gadget/udc/snps_udc_core.c
2098
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2116
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2193
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2314
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2354
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2680
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
566
dma_addr_t addr = (dma_addr_t)td->next;
drivers/usb/gadget/udc/snps_udc_core.c
573
addr_next = (dma_addr_t)td->next;
drivers/usb/gadget/udc/snps_udc_core.c
613
req->td_data->next = req->td_phys;
drivers/usb/gadget/udc/snps_udc_core.c
804
req->td_data->next);
drivers/usb/gadget/udc/snps_udc_core.c
807
td = (struct udc_data_dma *)phys_to_virt(last->next);
drivers/usb/gadget/udc/snps_udc_core.c
827
req->td_data->next = dma_addr;
drivers/usb/gadget/udc/snps_udc_core.c
846
last->next = dma_addr;
drivers/usb/host/ehci-dbg.c
379
char *next = *nextp;
drivers/usb/host/ehci-dbg.c
398
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
413
next += temp;
drivers/usb/host/ehci-dbg.c
446
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
456
next += temp;
drivers/usb/host/ehci-dbg.c
461
temp = scnprintf(next, size, "\n");
drivers/usb/host/ehci-dbg.c
463
next += temp;
drivers/usb/host/ehci-dbg.c
467
*nextp = next;
drivers/usb/host/ehci-dbg.c
476
char *next;
drivers/usb/host/ehci-dbg.c
481
next = buf->output_buf;
drivers/usb/host/ehci-dbg.c
484
*next = 0;
drivers/usb/host/ehci-dbg.c
493
qh_lines(ehci, qh, &next, &size);
drivers/usb/host/ehci-dbg.c
495
temp = scnprintf(next, size, "\nunlink =\n");
drivers/usb/host/ehci-dbg.c
497
next += temp;
drivers/usb/host/ehci-dbg.c
502
qh_lines(ehci, qh, &next, &size);
drivers/usb/host/ehci-dbg.c
516
char *next;
drivers/usb/host/ehci-dbg.c
523
next = buf->output_buf;
drivers/usb/host/ehci-dbg.c
526
*next = 0;
drivers/usb/host/ehci-dbg.c
531
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
534
next += temp;
drivers/usb/host/ehci-dbg.c
537
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
542
next += temp;
drivers/usb/host/ehci-dbg.c
547
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
552
next += temp;
drivers/usb/host/ehci-dbg.c
555
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
560
next += temp;
drivers/usb/host/ehci-dbg.c
562
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
565
next += temp;
drivers/usb/host/ehci-dbg.c
569
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
574
next += temp;
drivers/usb/host/ehci-dbg.c
577
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
585
next += temp;
drivers/usb/host/ehci-dbg.c
590
return next - buf->output_buf;
drivers/usb/host/ehci-dbg.c
628
char *next;
drivers/usb/host/ehci-dbg.c
639
next = buf->output_buf;
drivers/usb/host/ehci-dbg.c
642
temp = scnprintf(next, size, "size = %d\n", ehci->periodic_size);
drivers/usb/host/ehci-dbg.c
644
next += temp;
drivers/usb/host/ehci-dbg.c
657
temp = scnprintf(next, size, "%4d: ", i);
drivers/usb/host/ehci-dbg.c
659
next += temp;
drivers/usb/host/ehci-dbg.c
667
temp = scnprintf(next, size, " qh%d-%04x/%p",
drivers/usb/host/ehci-dbg.c
675
next += temp;
drivers/usb/host/ehci-dbg.c
681
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
684
next += temp;
drivers/usb/host/ehci-dbg.c
690
temp = output_buf_tds_dir(next, ehci,
drivers/usb/host/ehci-dbg.c
702
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
709
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
715
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
726
next += temp;
drivers/usb/host/ehci-dbg.c
729
temp = scnprintf(next, size, "\n");
drivers/usb/host/ehci-dbg.c
731
next += temp;
drivers/usb/host/ehci-dbg.c
761
char *next, scratch[80];
drivers/usb/host/ehci-dbg.c
767
next = buf->output_buf;
drivers/usb/host/ehci-dbg.c
773
size = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
785
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
794
next += temp;
drivers/usb/host/ehci-dbg.c
810
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
815
next += temp;
drivers/usb/host/ehci-dbg.c
819
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
822
next += temp;
drivers/usb/host/ehci-dbg.c
837
temp = scnprintf(next, size, "structural params 0x%08x\n", i);
drivers/usb/host/ehci-dbg.c
839
next += temp;
drivers/usb/host/ehci-dbg.c
842
temp = scnprintf(next, size, "capability params 0x%08x\n", i);
drivers/usb/host/ehci-dbg.c
844
next += temp;
drivers/usb/host/ehci-dbg.c
849
temp = scnprintf(next, size, fmt, temp, scratch);
drivers/usb/host/ehci-dbg.c
851
next += temp;
drivers/usb/host/ehci-dbg.c
855
temp = scnprintf(next, size, fmt, temp, scratch);
drivers/usb/host/ehci-dbg.c
857
next += temp;
drivers/usb/host/ehci-dbg.c
861
temp = scnprintf(next, size, fmt, temp, scratch);
drivers/usb/host/ehci-dbg.c
863
next += temp;
drivers/usb/host/ehci-dbg.c
865
temp = scnprintf(next, size, "uframe %04x\n",
drivers/usb/host/ehci-dbg.c
868
next += temp;
drivers/usb/host/ehci-dbg.c
874
temp = scnprintf(next, size, fmt, temp, scratch);
drivers/usb/host/ehci-dbg.c
876
next += temp;
drivers/usb/host/ehci-dbg.c
878
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
883
next += temp;
drivers/usb/host/ehci-dbg.c
888
temp = scnprintf(next, size, "async unlink qh %p\n",
drivers/usb/host/ehci-dbg.c
892
next += temp;
drivers/usb/host/ehci-dbg.c
896
temp = scnprintf(next, size,
drivers/usb/host/ehci-dbg.c
901
next += temp;
drivers/usb/host/ehci-dbg.c
903
temp = scnprintf(next, size, "complete %ld unlink %ld\n",
drivers/usb/host/ehci-dbg.c
906
next += temp;
drivers/usb/host/ehci-q.c
1040
qtd = list_entry (qtd_list->next, struct ehci_qtd,
drivers/usb/host/ehci-q.c
1115
qtd = list_entry(qtd_list->next, struct ehci_qtd, qtd_list);
drivers/usb/host/ehci-q.c
114
qtd = list_entry(qh->qtd_list.next, struct ehci_qtd, qtd_list);
drivers/usb/host/ehci-q.c
440
qh->qtd_list.next == &qtd->qtd_list &&
drivers/usb/host/ehci-sched.c
1476
u32 now, base, next, start, period, span, now2;
drivers/usb/host/ehci-sched.c
1511
next = start;
drivers/usb/host/ehci-sched.c
1526
} while (start > next && !done);
drivers/usb/host/ehci-sched.c
1555
next = now + ehci->i_thresh; /* uframe cache */
drivers/usb/host/ehci-sched.c
1557
next = (now + 2 + 7) & ~0x07; /* full frame cache */
drivers/usb/host/ehci-sched.c
1568
next = (next - base) & (mod - 1);
drivers/usb/host/ehci-sched.c
1594
if (unlikely(start < next &&
drivers/usb/host/ehci-sched.c
1635
start = next + ((start - next) & (period - 1));
drivers/usb/host/ehci-sched.c
1774
itd = list_entry(iso_sched->td_list.next,
drivers/usb/host/ehci-sched.c
2177
sitd = list_entry(sched->td_list.next,
drivers/usb/host/ehci-sched.c
629
ehci->qh_scan_next = list_entry(qh->intr_node.next,
drivers/usb/host/ehci-sched.c
676
else if (ehci->intr_unlink.next == &qh->unlink_node) {
drivers/usb/host/ehci-sched.c
697
else if (ehci->intr_unlink_wait.next == &qh->unlink_node) {
drivers/usb/host/fhci-mem.c
42
td = list_entry(fhci->empty_tds.next, struct td, node);
drivers/usb/host/fhci-mem.c
43
list_del(fhci->empty_tds.next);
drivers/usb/host/fhci-mem.c
66
ed = list_entry(fhci->empty_eds.next, struct ed, node);
drivers/usb/host/fhci-mem.c
67
list_del(fhci->empty_eds.next);
drivers/usb/host/fhci-q.c
100
td = list_entry(frame->tds_list.next, struct td, frame_lh);
drivers/usb/host/fhci-q.c
112
td = list_entry(ed->td_list.next, struct td, node);
drivers/usb/host/fhci-q.c
113
list_del_init(ed->td_list.next);
drivers/usb/host/fhci-q.c
117
ed->td_head = list_entry(ed->td_list.next, struct td,
drivers/usb/host/fhci-q.c
132
td = list_entry(p_list->done_list.next, struct td, node);
drivers/usb/host/fhci-q.c
133
list_del_init(p_list->done_list.next);
drivers/usb/host/fhci-q.c
149
ed->td_head = list_entry(ed->td_list.next, struct td, node);
drivers/usb/host/fhci-q.c
174
ed->td_head = list_entry(ed->td_list.next, struct td, node);
drivers/usb/host/fhci-q.c
75
td = list_entry(ed->td_list.next, struct td, node);
drivers/usb/host/fhci-q.c
87
td = list_entry(frame->tds_list.next, struct td, frame_lh);
drivers/usb/host/fhci-q.c
88
list_del_init(frame->tds_list.next);
drivers/usb/host/fhci-sched.c
257
struct list_head *node = list->next;
drivers/usb/host/fhci-sched.c
63
if (ed->td_list.next->next != &ed->td_list) {
drivers/usb/host/fhci-sched.c
65
list_entry(ed->td_list.next->next, struct td,
drivers/usb/host/isp116x-hcd.c
242
urb = container_of(ep->hep->urb_list.next,
drivers/usb/host/isp116x-hcd.c
329
prev = &temp->next;
drivers/usb/host/isp116x-hcd.c
331
*prev = ep->next;
drivers/usb/host/isp116x-hcd.c
361
container_of(ep->hep->urb_list.next, struct urb, urb_list);
drivers/usb/host/isp116x-hcd.c
510
while (last_ep->next)
drivers/usb/host/isp116x-hcd.c
511
last_ep = (last_ep->active = last_ep->next);
drivers/usb/host/isp116x-hcd.c
518
urb = container_of(ep->hep->urb_list.next,
drivers/usb/host/isp116x-hcd.c
565
if ((&isp116x->async)->next != (&isp116x->async)->prev)
drivers/usb/host/isp116x-hcd.c
566
list_move(&isp116x->async, (&isp116x->async)->next);
drivers/usb/host/isp116x-hcd.c
802
prev = &here->next;
drivers/usb/host/isp116x-hcd.c
806
ep->next = here;
drivers/usb/host/isp116x-hcd.c
855
if (ep->hep->urb_list.next == &urb->urb_list)
drivers/usb/host/isp116x.h
320
struct isp116x_ep *next;
drivers/usb/host/max3421-hcd.c
770
struct urb *urb, *next;
drivers/usb/host/max3421-hcd.c
777
list_for_each_entry_safe(urb, next, &ep->urb_list, urb_list) {
drivers/usb/host/octeon-hcd.c
2344
struct cvmx_usb_transaction *transaction, *next;
drivers/usb/host/octeon-hcd.c
2347
list_for_each_entry_safe(transaction, next, &pipe->transactions, node) {
drivers/usb/host/ohci-dbg.c
101
ohci_dump_status (struct ohci_hcd *controller, char **next, unsigned *size)
drivers/usb/host/ohci-dbg.c
107
ohci_dbg_sw (controller, next, size,
drivers/usb/host/ohci-dbg.c
114
ohci_dbg_sw (controller, next, size,
drivers/usb/host/ohci-dbg.c
129
ohci_dbg_sw (controller, next, size,
drivers/usb/host/ohci-dbg.c
140
next, size);
drivers/usb/host/ohci-dbg.c
143
next, size);
drivers/usb/host/ohci-dbg.c
148
next, size);
drivers/usb/host/ohci-dbg.c
152
next, size);
drivers/usb/host/ohci-dbg.c
155
next, size);
drivers/usb/host/ohci-dbg.c
159
next, size);
drivers/usb/host/ohci-dbg.c
162
next, size);
drivers/usb/host/ohci-dbg.c
165
ohci_readl (controller, &regs->donehead), next, size);
drivers/usb/host/ohci-dbg.c
168
#define dbg_port_sw(hc,num,value,next,size) \
drivers/usb/host/ohci-dbg.c
169
ohci_dbg_sw (hc, next, size, \
drivers/usb/host/ohci-dbg.c
194
char **next,
drivers/usb/host/ohci-dbg.c
204
ohci_dbg_sw (controller, next, size,
drivers/usb/host/ohci-dbg.c
215
ohci_dbg_sw (controller, next, size,
drivers/usb/host/ohci-dbg.c
222
ohci_dbg_sw (controller, next, size,
drivers/usb/host/ohci-dbg.c
23
#define ohci_dbg_sw(ohci, next, size, format, arg...) \
drivers/usb/host/ohci-dbg.c
236
dbg_port_sw (controller, i, temp, next, size);
drivers/usb/host/ohci-dbg.c
25
if (next != NULL) { \
drivers/usb/host/ohci-dbg.c
27
s_len = scnprintf (*next, *size, format, ## arg ); \
drivers/usb/host/ohci-dbg.c
28
*size -= s_len; *next += s_len; \
drivers/usb/host/ohci-dbg.c
34
#define ohci_dbg_nosw(ohci, next, size, format, arg...) \
drivers/usb/host/ohci-dbg.c
37
s_len = scnprintf(*next, *size, format, ## arg); \
drivers/usb/host/ohci-dbg.c
38
*size -= s_len; *next += s_len; \
drivers/usb/host/ohci-dbg.c
46
char **next,
drivers/usb/host/ohci-dbg.c
49
ohci_dbg_sw (ohci, next, size, "%s 0x%08x%s%s%s%s%s%s%s%s%s\n",
drivers/usb/host/ohci-dbg.c
492
char *next;
drivers/usb/host/ohci-dbg.c
501
next = buf->page;
drivers/usb/host/ohci-dbg.c
504
temp = scnprintf (next, size, "size = %d\n", NUM_INTS);
drivers/usb/host/ohci-dbg.c
506
next += temp;
drivers/usb/host/ohci-dbg.c
515
temp = scnprintf (next, size, "%2d [%3d]:", i, ohci->load [i]);
drivers/usb/host/ohci-dbg.c
517
next += temp;
drivers/usb/host/ohci-dbg.c
520
temp = scnprintf (next, size, " ed%d/%p",
drivers/usb/host/ohci-dbg.c
523
next += temp;
drivers/usb/host/ohci-dbg.c
539
temp = scnprintf (next, size,
drivers/usb/host/ohci-dbg.c
555
next += temp;
drivers/usb/host/ohci-dbg.c
570
temp = scnprintf (next, size, "\n");
drivers/usb/host/ohci-dbg.c
572
next += temp;
drivers/usb/host/ohci-dbg.c
588
char *next;
drivers/usb/host/ohci-dbg.c
594
next = buf->page;
drivers/usb/host/ohci-dbg.c
601
ohci_dbg_nosw(ohci, &next, &size,
drivers/usb/host/ohci-dbg.c
611
size -= scnprintf (next, size,
drivers/usb/host/ohci-dbg.c
616
ohci_dump_status(ohci, &next, &size);
drivers/usb/host/ohci-dbg.c
620
ohci_dbg_nosw(ohci, &next, &size,
drivers/usb/host/ohci-dbg.c
625
temp = scnprintf (next, size,
drivers/usb/host/ohci-dbg.c
630
next += temp;
drivers/usb/host/ohci-dbg.c
633
temp = scnprintf (next, size, "fmremaining 0x%08x %sFR=0x%04x\n",
drivers/usb/host/ohci-dbg.c
637
next += temp;
drivers/usb/host/ohci-dbg.c
640
temp = scnprintf (next, size, "periodicstart 0x%04x\n",
drivers/usb/host/ohci-dbg.c
643
next += temp;
drivers/usb/host/ohci-dbg.c
646
temp = scnprintf (next, size, "lsthresh 0x%04x\n",
drivers/usb/host/ohci-dbg.c
649
next += temp;
drivers/usb/host/ohci-dbg.c
651
temp = scnprintf (next, size, "hub poll timer %s\n",
drivers/usb/host/ohci-dbg.c
654
next += temp;
drivers/usb/host/ohci-dbg.c
657
ohci_dump_roothub (ohci, 1, &next, &size);
drivers/usb/host/ohci-dbg.c
68
char **next,
drivers/usb/host/ohci-dbg.c
72
ohci_dbg_sw (ohci, next, size, "%s %08x\n", label, value);
drivers/usb/host/ohci-hcd.c
254
u16 next = ohci_frame_no(ohci) + 1;
drivers/usb/host/ohci-hcd.c
259
if (unlikely(tick_before(frame, next))) {
drivers/usb/host/ohci-hcd.c
263
frame += (next - frame + ed->interval - 1) &
drivers/usb/host/ohci-hcd.c
279
(u16) (next - frame),
drivers/usb/host/ohci-hcd.c
285
next);
drivers/usb/host/ohci-q.c
1173
td = list_entry(ed->td_list.next, struct td, td_list);
drivers/usb/host/ohci-q.c
838
struct list_head *tmp = td->td_list.next;
drivers/usb/host/ohci-q.c
853
struct td *next;
drivers/usb/host/ohci-q.c
855
next = list_entry (tmp, struct td, td_list);
drivers/usb/host/ohci-q.c
856
tmp = next->td_list.next;
drivers/usb/host/ohci-q.c
858
if (next->urb != urb)
drivers/usb/host/ohci-q.c
869
list_del(&next->td_list);
drivers/usb/host/ohci-q.c
871
ed->hwHeadP = next->hwNextTD | toggle;
drivers/usb/host/oxu210hp-hcd.c
1261
qtd = list_entry(qh->qtd_list.next,
drivers/usb/host/oxu210hp-hcd.c
1956
qtd = list_entry(qtd_list->next, struct ehci_qtd,
drivers/usb/host/oxu210hp-hcd.c
2023
qtd = list_entry(qtd_list->next, struct ehci_qtd, qtd_list);
drivers/usb/host/oxu210hp-hcd.c
2061
struct ehci_qh *next;
drivers/usb/host/oxu210hp-hcd.c
2070
next = qh->reclaim;
drivers/usb/host/oxu210hp-hcd.c
2071
oxu->reclaim = next;
drivers/usb/host/oxu210hp-hcd.c
2091
if (next) {
drivers/usb/host/oxu210hp-hcd.c
2093
start_unlink_async(oxu, next);
drivers/usb/host/r8a66597-hcd.c
835
struct r8a66597_td *td, *next;
drivers/usb/host/r8a66597-hcd.c
842
list_for_each_entry_safe(td, next, list, queue) {
drivers/usb/host/r8a66597.h
147
return list_entry(r8a66597->pipe_queue[pipenum].next,
drivers/usb/host/sl811-hcd.c
1487
ep = ep->next;
drivers/usb/host/sl811-hcd.c
313
sl811->next_periodic = ep->next;
drivers/usb/host/sl811-hcd.c
318
ep = container_of(sl811->async.next,
drivers/usb/host/sl811-hcd.c
332
if (ep->schedule.next == &sl811->async)
drivers/usb/host/sl811-hcd.c
335
sl811->next_async = container_of(ep->schedule.next,
drivers/usb/host/sl811-hcd.c
345
urb = container_of(ep->hep->urb_list.next, struct urb, urb_list);
drivers/usb/host/sl811-hcd.c
458
prev = &temp->next;
drivers/usb/host/sl811-hcd.c
460
*prev = ep->next;
drivers/usb/host/sl811-hcd.c
468
sl811->next_periodic = ep->next;
drivers/usb/host/sl811-hcd.c
487
urb = container_of(ep->hep->urb_list.next, struct urb, urb_list);
drivers/usb/host/sl811-hcd.c
703
->hep->urb_list.next,
drivers/usb/host/sl811-hcd.c
713
->hep->urb_list.next,
drivers/usb/host/sl811-hcd.c
936
prev = &here->next;
drivers/usb/host/sl811-hcd.c
940
ep->next = here;
drivers/usb/host/sl811-hcd.c
980
if (ep->hep->urb_list.next != &urb->urb_list) {
drivers/usb/host/sl811.h
185
struct sl811h_ep *next;
drivers/usb/host/uhci-debug.c
220
struct urb_priv *urbp = list_entry(qh->queue.next,
drivers/usb/host/uhci-debug.c
222
struct uhci_td *td = list_entry(urbp->td_list.next,
drivers/usb/host/uhci-debug.c
435
tmp = tmp->next;
drivers/usb/host/uhci-debug.c
507
tmp = head->next;
drivers/usb/host/uhci-debug.c
511
tmp = tmp->next;
drivers/usb/host/uhci-q.c
1213
if (td->list.next != urbp->td_list.prev)
drivers/usb/host/uhci-q.c
1259
unsigned frame, next;
drivers/usb/host/uhci-q.c
1279
next = uhci->frame_number + 10;
drivers/usb/host/uhci-q.c
1283
frame += (next - frame + qh->period - 1) & -qh->period;
drivers/usb/host/uhci-q.c
1289
next = uhci->frame_number + 1;
drivers/usb/host/uhci-q.c
1305
if (!uhci_frame_before_eq(next, frame)) {
drivers/usb/host/uhci-q.c
1309
frame += (next - frame + qh->period - 1) &
drivers/usb/host/uhci-q.c
1316
else if (!uhci_frame_before_eq(next,
drivers/usb/host/uhci-q.c
1323
next);
drivers/usb/host/uhci-q.c
1462
if (qh->queue.next == &urbp->node && !qh->is_stopped) {
drivers/usb/host/uhci-q.c
1535
urbp->node.next != &qh->queue) {
drivers/usb/host/uhci-q.c
1536
struct urb *nurb = list_entry(urbp->node.next,
drivers/usb/host/uhci-q.c
1582
urbp = list_entry(qh->queue.next, struct urb_priv, node);
drivers/usb/host/uhci-q.c
1641
urbp = list_entry(qh->queue.next, struct urb_priv, node);
drivers/usb/host/uhci-q.c
1643
struct uhci_td *td = list_entry(urbp->td_list.next,
drivers/usb/host/uhci-q.c
1691
urbp = list_entry(qh->queue.next, struct urb_priv, node);
drivers/usb/host/uhci-q.c
1692
td = list_entry(urbp->td_list.next, struct uhci_td, list);
drivers/usb/host/uhci-q.c
1762
uhci->next_qh = list_entry(uhci->skelqh[i]->node.next,
drivers/usb/host/uhci-q.c
1765
uhci->next_qh = list_entry(qh->node.next,
drivers/usb/host/uhci-q.c
1772
list_entry(qh->queue.next, struct urb_priv, node));
drivers/usb/host/uhci-q.c
198
ntd = list_entry(td->fl_list.next,
drivers/usb/host/uhci-q.c
334
if (qh->queue.next != &urbp->node) {
drivers/usb/host/uhci-q.c
360
td = list_entry(urbp->td_list.next, struct uhci_td, list);
drivers/usb/host/uhci-q.c
383
urbp = list_entry(qh->queue.next, struct urb_priv, node);
drivers/usb/host/uhci-q.c
398
td = list_entry(urbp->td_list.next, struct uhci_td, list);
drivers/usb/host/uhci-q.c
415
pipe = list_entry(qh->queue.next, struct urb_priv, node)->urb->pipe;
drivers/usb/host/uhci-q.c
487
struct urb_priv *urbp = list_entry(qh->queue.next,
drivers/usb/host/uhci-q.c
489
struct uhci_td *td = list_entry(urbp->td_list.next,
drivers/usb/host/uhci-q.c
506
uhci->next_qh = list_entry(qh->node.next, struct uhci_qh,
drivers/usb/host/uhci-q.c
575
uhci->next_qh = list_entry(qh->node.next, struct uhci_qh,
drivers/usb/host/uhci-q.c
591
uhci->next_qh = list_entry(qh->node.next, struct uhci_qh,
drivers/usb/host/xen-hcd.c
830
urbp = list_entry(info->pending_submit_list.next,
drivers/usb/host/xhci-dbgcap.c
267
union xhci_trb *trb, *next;
drivers/usb/host/xhci-dbgcap.c
279
next = ++(ring->enqueue);
drivers/usb/host/xhci-dbgcap.c
280
if (TRB_TYPE_LINK_LE32(next->link.control)) {
drivers/usb/host/xhci-dbgcap.c
281
next->link.control ^= cpu_to_le32(TRB_CYCLE);
drivers/usb/host/xhci-dbgcap.c
519
seg->next = seg;
drivers/usb/host/xhci-dbgtty.c
113
req = list_entry(pool->next, struct dbc_request, list_pool);
drivers/usb/host/xhci-dbgtty.c
259
req = list_entry(head->next, struct dbc_request, list_pool);
drivers/usb/host/xhci-dbgtty.c
63
req = list_entry(pool->next, struct dbc_request, list_pool);
drivers/usb/host/xhci-ext-caps.h
133
u32 next;
drivers/usb/host/xhci-ext-caps.h
152
next = XHCI_EXT_CAPS_NEXT(val);
drivers/usb/host/xhci-ext-caps.h
153
offset += next << 2;
drivers/usb/host/xhci-ext-caps.h
154
} while (next);
drivers/usb/host/xhci-mem.c
101
if (!seg || !seg->next)
drivers/usb/host/xhci-mem.c
113
trb->link.segment_ptr = cpu_to_le64(seg->next->dma);
drivers/usb/host/xhci-mem.c
152
src->last_seg->next = dst->enq_seg->next;
drivers/usb/host/xhci-mem.c
153
dst->enq_seg->next = src->first_seg;
drivers/usb/host/xhci-mem.c
171
for (seg = dst->enq_seg; seg != dst->last_seg; seg = seg->next)
drivers/usb/host/xhci-mem.c
172
seg->next->num = seg->num + 1;
drivers/usb/host/xhci-mem.c
1815
seg = seg->next;
drivers/usb/host/xhci-mem.c
1925
list_del_init(ep->next);
drivers/usb/host/xhci-mem.c
342
struct xhci_segment *next;
drivers/usb/host/xhci-mem.c
344
next = xhci_segment_alloc(xhci, ring->bounce_buf_len, num, flags);
drivers/usb/host/xhci-mem.c
345
if (!next)
drivers/usb/host/xhci-mem.c
348
prev->next = next;
drivers/usb/host/xhci-mem.c
349
prev = next;
drivers/usb/host/xhci-mem.c
354
ring->last_seg->next = ring->first_seg;
drivers/usb/host/xhci-mem.c
59
seg->next = NULL;
drivers/usb/host/xhci-mem.c
76
struct xhci_segment *seg, *next;
drivers/usb/host/xhci-mem.c
78
ring->last_seg->next = NULL;
drivers/usb/host/xhci-mem.c
799
struct xhci_tt_bw_info *tt_info, *next;
drivers/usb/host/xhci-mem.c
811
list_for_each_entry_safe(tt_info, next, tt_list_head, tt_list) {
drivers/usb/host/xhci-mem.c
82
next = seg->next;
drivers/usb/host/xhci-mem.c
84
seg = next;
drivers/usb/host/xhci-mem.c
937
struct xhci_tt_bw_info *tt_info, *next;
drivers/usb/host/xhci-mem.c
950
list_for_each_entry_safe(tt_info, next, tt_list_head, tt_list) {
drivers/usb/host/xhci-ring.c
120
return last_trb_on_seg(seg, trb) && (seg->next == ring->first_seg);
drivers/usb/host/xhci-ring.c
176
*seg = (*seg)->next;
drivers/usb/host/xhci-ring.c
198
ring->deq_seg = ring->deq_seg->next;
drivers/usb/host/xhci-ring.c
215
ring->deq_seg = ring->deq_seg->next;
drivers/usb/host/xhci-ring.c
260
ring->enq_seg = ring->enq_seg->next;
drivers/usb/host/xhci-ring.c
353
enq_seg = enq_seg->next;
drivers/usb/host/xhci-ring.c
366
enq_seg = enq_seg->next;
drivers/usb/host/xhci-ring.c
402
if (trb_is_link(ring->enqueue) && ring->enq_seg->next->trbs == ring->dequeue)
drivers/usb/host/xhci-ring.c
409
seg = seg->next;
drivers/usb/host/xhci-sideband.c
55
seg = seg->next;
drivers/usb/host/xhci.c
2530
ep_entry = bw_table->interval_bw[i].endpoints.next;
drivers/usb/host/xhci.h
1279
for (seg = head; seg != NULL; seg = (seg->next != head ? seg->next : NULL))
drivers/usb/host/xhci.h
1284
struct xhci_segment *next;
drivers/usb/image/microtek.h
31
struct mts_desc *next;
drivers/usb/isp1760/isp1760-hcd.c
1137
if (usb_pipeint(list_entry(qh->qtd_list.next, struct isp1760_qtd,
drivers/usb/isp1760/isp1760-hcd.c
1428
qtd = list_entry(qtd->qtd_list.next,
drivers/usb/isp1760/isp1760-hcd.c
805
qtd = list_entry(qtd->qtd_list.next, typeof(*qtd), qtd_list);
drivers/usb/misc/usb-ljca.c
711
struct ljca_client *client, *next;
drivers/usb/misc/usb-ljca.c
743
list_for_each_entry_safe_reverse(client, next, &adap->client_list, link) {
drivers/usb/misc/usb-ljca.c
851
struct ljca_client *client, *next;
drivers/usb/misc/usb-ljca.c
857
list_for_each_entry_safe_reverse(client, next, &adap->client_list, link) {
drivers/usb/mon/mon_main.c
382
p = mon_buses.next;
drivers/usb/mon/mon_text.c
315
p = rp->e_list.next;
drivers/usb/mon/mon_text.c
669
p = rp->e_list.next;
drivers/usb/mtu3/mtu3_qmu.c
228
struct qmu_gpd *next;
drivers/usb/mtu3/mtu3_qmu.c
231
next = enq + 1;
drivers/usb/mtu3/mtu3_qmu.c
233
next = ring->start;
drivers/usb/mtu3/mtu3_qmu.c
236
return next == ring->dequeue;
drivers/usb/musb/cppi_dma.h
64
struct cppi_descriptor *next;
drivers/usb/musb/musb_gadget.c
1248
if (!musb_ep->busy && &request->list == musb_ep->req_list.next) {
drivers/usb/musb/musb_gadget.c
1294
if (musb_ep->req_list.next != &req->list || musb_ep->busy)
drivers/usb/musb/musb_gadget.h
107
return container_of(queue->next, struct musb_request, list);
drivers/usb/musb/musb_host.h
121
return list_entry(queue->next, struct urb, urb_list);
drivers/usb/musb/musb_host.h
50
return list_entry(q->next, struct musb_qh, ring);
drivers/usb/renesas_usbhs/mod_host.c
442
struct usbhsh_ep *uep, *next;
drivers/usb/renesas_usbhs/mod_host.c
444
list_for_each_entry_safe(uep, next, &udev->ep_list_head, ep_list)
drivers/usb/serial/garmin_gps.c
306
result = (struct garmin_packet *)garmin_data_p->pktlist.next;
drivers/usb/serial/garmin_gps.c
322
result = (struct garmin_packet *)garmin_data_p->pktlist.next;
drivers/vdpa/ifcvf/ifcvf_base.c
127
goto next;
drivers/vdpa/ifcvf/ifcvf_base.c
156
next:
drivers/vfio/mdev/mdev_core.c
116
list_del(&mdev->next);
drivers/vfio/mdev/mdev_core.c
138
list_for_each_entry(tmp, &mdev_list, next) {
drivers/vfio/mdev/mdev_core.c
173
list_add(&mdev->next, &mdev_list);
drivers/vfio/mdev/mdev_core.c
219
list_for_each_entry(tmp, &mdev_list, next) {
drivers/vfio/pci/pds/lm.c
363
enum vfio_device_mig_state next)
drivers/vfio/pci/pds/lm.c
368
if (cur == VFIO_DEVICE_STATE_STOP && next == VFIO_DEVICE_STATE_STOP_COPY) {
drivers/vfio/pci/pds/lm.c
382
if (cur == VFIO_DEVICE_STATE_STOP_COPY && next == VFIO_DEVICE_STATE_STOP) {
drivers/vfio/pci/pds/lm.c
388
if (cur == VFIO_DEVICE_STATE_STOP && next == VFIO_DEVICE_STATE_RESUMING) {
drivers/vfio/pci/pds/lm.c
396
if (cur == VFIO_DEVICE_STATE_RESUMING && next == VFIO_DEVICE_STATE_STOP) {
drivers/vfio/pci/pds/lm.c
405
if (cur == VFIO_DEVICE_STATE_RUNNING && next == VFIO_DEVICE_STATE_RUNNING_P2P) {
drivers/vfio/pci/pds/lm.c
416
if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == VFIO_DEVICE_STATE_RUNNING) {
drivers/vfio/pci/pds/lm.c
426
if (cur == VFIO_DEVICE_STATE_STOP && next == VFIO_DEVICE_STATE_RUNNING_P2P) {
drivers/vfio/pci/pds/lm.c
435
if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == VFIO_DEVICE_STATE_STOP) {
drivers/vfio/pci/pds/lm.h
37
enum vfio_device_mig_state next);
drivers/vfio/pci/vfio_pci_config.c
1562
u8 next;
drivers/vfio/pci/vfio_pci_config.c
1570
pos + PCI_CAP_LIST_NEXT, &next);
drivers/vfio/pci/vfio_pci_config.c
1591
*prev = next;
drivers/vfio/pci/vfio_pci_config.c
1592
pos = next;
drivers/vfio/pci/vfio_pci_config.c
1616
pos = next;
drivers/vfio/pci/vfio_pci_core.c
631
&vdev->ioeventfds_list, next) {
drivers/vfio/pci/vfio_pci_core.c
633
list_del(&ioeventfd->next);
drivers/vfio/pci/vfio_pci_priv.h
15
struct list_head next;
drivers/vfio/pci/vfio_pci_rdwr.c
461
list_for_each_entry(ioeventfd, &vdev->ioeventfds_list, next) {
drivers/vfio/pci/vfio_pci_rdwr.c
466
list_del(&ioeventfd->next);
drivers/vfio/pci/vfio_pci_rdwr.c
509
list_add(&ioeventfd->next, &vdev->ioeventfds_list);
drivers/vfio/vfio_iommu_spapr_tce.c
1098
list_for_each_entry(tcegrp, &container->group_list, next) {
drivers/vfio/vfio_iommu_spapr_tce.c
119
list_for_each_entry(tcemem, &container->prereg_list, next) {
drivers/vfio/vfio_iommu_spapr_tce.c
1273
list_for_each_entry(tcegrp, &container->group_list, next) {
drivers/vfio/vfio_iommu_spapr_tce.c
1305
list_add(&tcegrp->next, &container->group_list);
drivers/vfio/vfio_iommu_spapr_tce.c
1327
list_for_each_entry(tcegrp, &container->group_list, next) {
drivers/vfio/vfio_iommu_spapr_tce.c
1340
list_del(&tcegrp->next);
drivers/vfio/vfio_iommu_spapr_tce.c
150
list_for_each_entry(tcemem, &container->prereg_list, next) {
drivers/vfio/vfio_iommu_spapr_tce.c
169
list_add(&tcemem->next, &container->prereg_list);
drivers/vfio/vfio_iommu_spapr_tce.c
281
struct tce_iommu_group, next);
drivers/vfio/vfio_iommu_spapr_tce.c
353
struct tce_iommu_group, next);
drivers/vfio/vfio_iommu_spapr_tce.c
370
list_for_each_entry_safe(tcemem, tmtmp, &container->prereg_list, next)
drivers/vfio/vfio_iommu_spapr_tce.c
45
struct list_head next;
drivers/vfio/vfio_iommu_spapr_tce.c
54
struct list_head next;
drivers/vfio/vfio_iommu_spapr_tce.c
658
struct tce_iommu_group, next);
drivers/vfio/vfio_iommu_spapr_tce.c
683
list_for_each_entry(tcegrp, &container->group_list, next) {
drivers/vfio/vfio_iommu_spapr_tce.c
699
list_for_each_entry(tcegrp, &container->group_list, next) {
drivers/vfio/vfio_iommu_spapr_tce.c
725
list_for_each_entry(tcegrp, &container->group_list, next) {
drivers/vfio/vfio_iommu_spapr_tce.c
762
struct tce_iommu_group, next);
drivers/vfio/vfio_iommu_spapr_tce.c
866
struct tce_iommu_group, next);
drivers/vfio/vfio_iommu_spapr_tce.c
98
list_del(&tcemem->next);
drivers/vfio/vfio_iommu_type1.c
1064
struct vfio_regions *entry, *next;
drivers/vfio/vfio_iommu_type1.c
1068
list_for_each_entry_safe(entry, next, regions, list) {
drivers/vfio/vfio_iommu_type1.c
115
struct list_head next;
drivers/vfio/vfio_iommu_type1.c
1172
struct vfio_domain, next);
drivers/vfio/vfio_iommu_type1.c
1174
list_for_each_entry_continue(d, &iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
1182
phys_addr_t phys, next;
drivers/vfio/vfio_iommu_type1.c
1197
next = iommu_iova_to_phys(domain->domain, iova + len);
drivers/vfio/vfio_iommu_type1.c
1198
if (next != phys + len)
drivers/vfio/vfio_iommu_type1.c
1254
list_for_each_entry(domain, &iommu->domain_list, next)
drivers/vfio/vfio_iommu_type1.c
1560
list_for_each_entry(d, &iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
1573
list_for_each_entry_continue_reverse(d, &iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
1816
struct vfio_domain, next);
drivers/vfio/vfio_iommu_type1.c
1953
list_for_each_entry(g, &domain->group_list, next) {
drivers/vfio/vfio_iommu_type1.c
1968
list_for_each_entry(domain, &iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
1974
list_for_each_entry(group, &iommu->emulated_iommu_groups, next)
drivers/vfio/vfio_iommu_type1.c
2073
struct vfio_iova *node, *next;
drivers/vfio/vfio_iommu_type1.c
2079
list_for_each_entry_safe(node, next, iova, list) {
drivers/vfio/vfio_iommu_type1.c
2092
list_for_each_entry_safe(node, next, iova, list) {
drivers/vfio/vfio_iommu_type1.c
2135
struct vfio_iova *n, *next;
drivers/vfio/vfio_iommu_type1.c
2146
list_for_each_entry_safe(n, next, iova, list) {
drivers/vfio/vfio_iommu_type1.c
2181
struct iommu_resv_region *n, *next;
drivers/vfio/vfio_iommu_type1.c
2183
list_for_each_entry_safe(n, next, resv_regions, list) {
drivers/vfio/vfio_iommu_type1.c
2191
struct vfio_iova *n, *next;
drivers/vfio/vfio_iommu_type1.c
2193
list_for_each_entry_safe(n, next, iova, list) {
drivers/vfio/vfio_iommu_type1.c
2268
list_add(&group->next, &iommu->emulated_iommu_groups);
drivers/vfio/vfio_iommu_type1.c
2339
list_add(&group->next, &domain->group_list);
drivers/vfio/vfio_iommu_type1.c
2366
list_for_each_entry(d, &iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
2373
list_add(&group->next, &d->group_list);
drivers/vfio/vfio_iommu_type1.c
2397
list_add(&domain->next, &iommu->domain_list);
drivers/vfio/vfio_iommu_type1.c
2476
list_for_each_entry(domain, &iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
2511
list_for_each_entry(d, &iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
2512
list_for_each_entry(g, &d->group_list, next) {
drivers/vfio/vfio_iommu_type1.c
2549
list_for_each_entry(group, &iommu->emulated_iommu_groups, next) {
drivers/vfio/vfio_iommu_type1.c
2553
list_del(&group->next);
drivers/vfio/vfio_iommu_type1.c
2571
list_for_each_entry(domain, &iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
2578
list_del(&group->next);
drivers/vfio/vfio_iommu_type1.c
2598
list_del(&domain->next);
drivers/vfio/vfio_iommu_type1.c
2662
&domain->group_list, next) {
drivers/vfio/vfio_iommu_type1.c
2664
list_del(&group->next);
drivers/vfio/vfio_iommu_type1.c
2678
&iommu->emulated_iommu_groups, next) {
drivers/vfio/vfio_iommu_type1.c
2679
list_del(&group->next);
drivers/vfio/vfio_iommu_type1.c
2686
&iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
2688
list_del(&domain->next);
drivers/vfio/vfio_iommu_type1.c
2703
list_for_each_entry(domain, &iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
3242
list_for_each_entry(d, &iommu->domain_list, next) {
drivers/vfio/vfio_iommu_type1.c
647
struct rb_node *next;
drivers/vfio/vfio_iommu_type1.c
652
prev = next = &top->node;
drivers/vfio/vfio_iommu_type1.c
661
while ((next = rb_next(next))) {
drivers/vfio/vfio_iommu_type1.c
662
vpfn = rb_entry(next, struct vfio_pfn, node);
drivers/vfio/vfio_iommu_type1.c
83
struct list_head next;
drivers/vfio/vfio_main.c
1560
for (tmp = buf; tmp->next; tmp = buf + tmp->next)
drivers/vfio/vfio_main.c
1563
tmp->next = caps->size;
drivers/vfio/vfio_main.c
1578
for (tmp = buf; tmp->next; tmp = buf + tmp->next - offset)
drivers/vfio/vfio_main.c
1579
tmp->next += offset;
drivers/vhost/vdpa.c
1065
goto next;
drivers/vhost/vdpa.c
1082
next:
drivers/vhost/vhost.c
2688
unsigned int next;
drivers/vhost/vhost.c
2695
next = vhost16_to_cpu(vq, READ_ONCE(desc->next));
drivers/vhost/vhost.c
2696
return next;
drivers/vhost/vringh.c
207
*up_next = vringh16_to_cpu(vrh, desc->next);
drivers/vhost/vringh.c
421
i = vringh16_to_cpu(vrh, desc.next);
drivers/video/fbdev/aty/aty128fb.c
451
struct fb_info *next;
drivers/video/fbdev/core/fb_defio.c
264
struct fb_deferred_io_pageref *pageref, *next;
drivers/video/fbdev/core/fb_defio.c
283
list_for_each_entry_safe(pageref, next, &fbdefio->pagereflist, list)
drivers/video/fbdev/core/fb_procfs.c
42
.next = fb_seq_next,
drivers/video/fbdev/core/fbmem.c
334
if (info->modelist.prev && info->modelist.next &&
drivers/video/fbdev/core/fbmem.c
455
if (!fb_info->modelist.prev || !fb_info->modelist.next)
drivers/video/fbdev/core/modedb.c
1149
if (!head->prev || !head->next || list_empty(head))
drivers/video/fbdev/matrox/matroxfb_base.c
2014
for (drv = matroxfb_driver_l(matroxfb_driver_list.next);
drivers/video/fbdev/matrox/matroxfb_base.c
2016
drv = matroxfb_driver_l(drv->node.next)) {
drivers/video/fbdev/omap/hwa742.c
248
req = list_entry(hwa742.free_req_list.next,
drivers/video/fbdev/omap/hwa742.c
283
req = list_entry(hwa742.pending_req_list.next,
drivers/video/fbdev/omap2/omapfb/dss/display.c
230
dssdev = list_entry(l->next, struct omap_dss_device,
drivers/video/fbdev/pxa3xx-gcu.c
191
struct pxa3xx_gcu_batch *next;
drivers/video/fbdev/pxa3xx-gcu.c
194
next = running->next;
drivers/video/fbdev/pxa3xx-gcu.c
195
running->next = priv->free;
drivers/video/fbdev/pxa3xx-gcu.c
197
running = next;
drivers/video/fbdev/pxa3xx-gcu.c
219
ready = ready->next;
drivers/video/fbdev/pxa3xx-gcu.c
407
priv->free = buffer->next;
drivers/video/fbdev/pxa3xx-gcu.c
415
buffer->next = priv->free;
drivers/video/fbdev/pxa3xx-gcu.c
431
buffer->next = NULL;
drivers/video/fbdev/pxa3xx-gcu.c
436
priv->ready_last->next = buffer;
drivers/video/fbdev/pxa3xx-gcu.c
544
buffer->next = priv->free;
drivers/video/fbdev/pxa3xx-gcu.c
554
struct pxa3xx_gcu_batch *next, *buffer = priv->free;
drivers/video/fbdev/pxa3xx-gcu.c
557
next = buffer->next;
drivers/video/fbdev/pxa3xx-gcu.c
563
buffer = next;
drivers/video/fbdev/pxa3xx-gcu.c
78
struct pxa3xx_gcu_batch *next;
drivers/video/fbdev/riva/riva_hw.c
251
int last, next, cur;
drivers/video/fbdev/riva/riva_hw.c
291
next = VIDEO;
drivers/video/fbdev/riva/riva_hw.c
293
next = MPORT;
drivers/video/fbdev/riva/riva_hw.c
295
next = GRAPHICS;
drivers/video/fbdev/riva/riva_hw.c
301
next = VIDEO;
drivers/video/fbdev/riva/riva_hw.c
303
next = GRAPHICS;
drivers/video/fbdev/riva/riva_hw.c
305
next = MPORT;
drivers/video/fbdev/riva/riva_hw.c
310
next = GRAPHICS;
drivers/video/fbdev/riva/riva_hw.c
312
next = VIDEO;
drivers/video/fbdev/riva/riva_hw.c
314
next = MPORT;
drivers/video/fbdev/riva/riva_hw.c
319
next = MPORT;
drivers/video/fbdev/riva/riva_hw.c
321
next = GRAPHICS;
drivers/video/fbdev/riva/riva_hw.c
323
next = VIDEO;
drivers/video/fbdev/riva/riva_hw.c
328
cur = next;
drivers/video/fbdev/sis/sis.h
649
struct sis_video_info *next;
drivers/video/fbdev/sis/sis_main.c
5840
while((countvideo = countvideo->next) != NULL)
drivers/video/fbdev/sis/sis_main.c
6454
ivideo->next = card_list;
drivers/video/fbdev/smscufx.c
1825
node = dev->urbs.list.next; /* have reserved one with sem */
drivers/video/fbdev/smscufx.c
1916
entry = dev->urbs.list.next;
drivers/video/fbdev/udlfb.c
1819
node = dlfb->urbs.list.next; /* have reserved one with sem */
drivers/video/fbdev/udlfb.c
1915
entry = dlfb->urbs.list.next;
drivers/video/fbdev/udlfb.c
956
struct dlfb_deferred_free *d = list_entry(dlfb->deferred_free.next, struct dlfb_deferred_free, list);
drivers/video/fbdev/via/hw.c
901
bool next = true;
drivers/video/fbdev/via/hw.c
904
while (next) {
drivers/video/fbdev/via/hw.c
905
next = false;
drivers/video/fbdev/via/hw.c
913
next = true;
drivers/virt/acrn/ioeventfd.c
265
struct hsm_ioeventfd *p, *next;
drivers/virt/acrn/ioeventfd.c
270
list_for_each_entry_safe(p, next, &vm->ioeventfds, list)
drivers/virt/acrn/ioreq.c
467
struct acrn_ioreq_range *range, *next;
drivers/virt/acrn/ioreq.c
487
list_for_each_entry_safe(range, next, &client->range_list, list) {
drivers/virt/acrn/ioreq.c
638
struct acrn_ioreq_client *client, *next;
drivers/virt/acrn/ioreq.c
643
list_for_each_entry_safe(client, next, &vm->ioreq_clients, list)
drivers/virt/acrn/irqfd.c
219
struct hsm_irqfd *irqfd, *next;
drivers/virt/acrn/irqfd.c
224
list_for_each_entry_safe(irqfd, next, &vm->irqfds, list)
drivers/virtio/virtio_balloon.c
245
struct page *page, *next;
drivers/virtio/virtio_balloon.c
272
list_for_each_entry_safe(page, next, &pages, lru) {
drivers/virtio/virtio_balloon.c
293
struct page *page, *next;
drivers/virtio/virtio_balloon.c
295
list_for_each_entry_safe(page, next, pages, lru) {
drivers/virtio/virtio_balloon.c
479
struct page *page, *next;
drivers/virtio/virtio_balloon.c
486
list_for_each_entry_safe(page, next, &vb->free_page_list, lru) {
drivers/virtio/virtio_mem.c
1304
unsigned long addr, next, id, sb_id, count;
drivers/virtio/virtio_mem.c
1315
next = addr + PFN_PHYS(1 << order);
drivers/virtio/virtio_mem.c
1320
count = virtio_mem_phys_to_sb_id(vm, next - 1) - sb_id + 1;
drivers/virtio/virtio_mem.c
1354
addr = next;
drivers/virtio/virtio_mem.c
1364
list_for_each_entry_rcu(vm, &virtio_mem_devices, next) {
drivers/virtio/virtio_mem.c
270
struct list_head next;
drivers/virtio/virtio_mem.c
2951
INIT_LIST_HEAD(&vm->next);
drivers/virtio/virtio_mem.c
302
list_add_rcu(&vm->next, &virtio_mem_devices);
drivers/virtio/virtio_mem.c
316
list_del_rcu(&vm->next);
drivers/virtio/virtio_ring.c
103
u16 next; /* The next desc state in a list. */
drivers/virtio/virtio_ring.c
1481
extra[i].next = i + 1;
drivers/virtio/virtio_ring.c
1587
vq->free_head = vq->packed.desc_extra[id].next;
drivers/virtio/virtio_ring.c
1711
curr = vq->packed.desc_extra[curr].next;
drivers/virtio/virtio_ring.c
1763
curr = vq->packed.desc_extra[curr].next;
drivers/virtio/virtio_ring.c
1996
curr = vq->packed.desc_extra[curr].next;
drivers/virtio/virtio_ring.c
2030
vq->packed.desc_extra[state->last].next = vq->free_head;
drivers/virtio/virtio_ring.c
2359
desc_extra[i].next = i + 1;
drivers/virtio/virtio_ring.c
2361
desc_extra[num - 1].next = 0;
drivers/virtio/virtio_ring.c
3490
u16 next;
drivers/virtio/virtio_ring.c
3493
next = (vq->packed.next_avail_idx &
drivers/virtio/virtio_ring.c
3498
next = vq->split.avail_idx_shadow;
drivers/virtio/virtio_ring.c
3500
return next << 16 | _vq->index;
drivers/virtio/virtio_ring.c
541
return extra->next;
drivers/virtio/virtio_ring.c
568
extra[i].next = i + 1;
drivers/virtio/virtio_ring.c
582
u16 next;
drivers/virtio/virtio_ring.c
592
next = extra[i].next;
drivers/virtio/virtio_ring.c
594
desc[i].next = cpu_to_virtio16(vdev, next);
drivers/virtio/virtio_ring.c
596
return next;
drivers/virtio/virtio_ring.c
738
vq->free_head = vq->split.desc_extra[head].next;
drivers/virtio/virtio_ring.c
893
vq->split.desc_extra[i].next = vq->free_head;
drivers/w1/slaves/w1_ds2413.c
48
next:
drivers/w1/slaves/w1_ds2413.c
65
goto next;
drivers/xen/balloon.c
173
page = list_entry(ballooned_pages.next, struct page, lru);
drivers/xen/balloon.c
191
struct list_head *next = page->lru.next;
drivers/xen/balloon.c
192
if (next == &ballooned_pages)
drivers/xen/balloon.c
194
return list_entry(next, struct page, lru);
drivers/xen/gntalloc.c
127
struct gntalloc_gref *gref, *next;
drivers/xen/gntalloc.c
167
list_for_each_entry_safe(gref, next, &queue_file, next_file) {
drivers/xen/gntalloc.c
256
gref = list_entry(priv->list.next,
drivers/xen/gntalloc.c
354
n = list_entry(gref->next_file.next,
drivers/xen/gntalloc.c
463
struct gntalloc_gref *gref, *next;
drivers/xen/gntalloc.c
475
next = list_entry(gref->next_gref.next,
drivers/xen/gntalloc.c
479
gref = next;
drivers/xen/gntalloc.c
539
gref = list_entry(gref->next_file.next,
drivers/xen/gntdev-common.h
54
struct list_head next;
drivers/xen/gntdev-dmabuf.c
106
list_add(&obj->next, &priv->exp_wait_list);
drivers/xen/gntdev-dmabuf.c
117
list_del(&obj->next);
drivers/xen/gntdev-dmabuf.c
137
list_for_each_entry(obj, &priv->exp_wait_list, next)
drivers/xen/gntdev-dmabuf.c
151
list_for_each_entry(gntdev_dmabuf, &priv->exp_list, next)
drivers/xen/gntdev-dmabuf.c
31
struct list_head next;
drivers/xen/gntdev-dmabuf.c
312
list_del(&gntdev_dmabuf->next);
drivers/xen/gntdev-dmabuf.c
321
list_del(&map->next);
drivers/xen/gntdev-dmabuf.c
398
list_add(&gntdev_dmabuf->next, &args->dmabuf_priv->exp_list);
drivers/xen/gntdev-dmabuf.c
59
struct list_head next;
drivers/xen/gntdev-dmabuf.c
643
list_add(&gntdev_dmabuf->next, &priv->imp_list);
drivers/xen/gntdev-dmabuf.c
671
list_for_each_entry_safe(gntdev_dmabuf, q, &priv->imp_list, next) {
drivers/xen/gntdev-dmabuf.c
675
list_del(&gntdev_dmabuf->next);
drivers/xen/gntdev-dmabuf.c
715
list_for_each_entry_safe(gntdev_dmabuf, q, &priv->imp_list, next)
drivers/xen/gntdev.c
1003
batch->next = priv->batch;
drivers/xen/gntdev.c
226
list_for_each_entry(map, &priv->maps, next) {
drivers/xen/gntdev.c
228
list_add_tail(&add->next, &map->next);
drivers/xen/gntdev.c
233
list_add_tail(&add->next, &priv->maps);
drivers/xen/gntdev.c
244
list_for_each_entry(map, &priv->maps, next) {
drivers/xen/gntdev.c
623
map = list_entry(priv->maps.next,
drivers/xen/gntdev.c
624
struct gntdev_grant_map, next);
drivers/xen/gntdev.c
625
list_del(&map->next);
drivers/xen/gntdev.c
633
priv->batch = batch->next;
drivers/xen/gntdev.c
68
struct gntdev_copy_batch *next;
drivers/xen/gntdev.c
695
list_del(&map->next);
drivers/xen/gntdev.c
768
list_for_each_entry(map, &priv->maps, next) {
drivers/xen/gntdev.c
95
list_for_each_entry(map, &priv->maps, next)
drivers/xen/gntdev.c
972
priv->batch = batch->next;
drivers/xen/grant-table.c
307
struct gnttab_free_callback *callback, *next;
drivers/xen/grant-table.c
313
next = callback->next;
drivers/xen/grant-table.c
315
callback->next = NULL;
drivers/xen/grant-table.c
318
callback->next = gnttab_free_callback_list;
drivers/xen/grant-table.c
321
callback = next;
drivers/xen/grant-table.c
716
cb = cb->next;
drivers/xen/grant-table.c
722
callback->next = gnttab_free_callback_list;
drivers/xen/grant-table.c
736
for (pcb = &gnttab_free_callback_list; *pcb; pcb = &(*pcb)->next) {
drivers/xen/grant-table.c
738
*pcb = callback->next;
drivers/xen/mcelog.c
115
num = xen_mcelog.next;
drivers/xen/mcelog.c
131
xen_mcelog.next = 0;
drivers/xen/mcelog.c
146
if (xen_mcelog.next)
drivers/xen/mcelog.c
198
entry = xen_mcelog.next;
drivers/xen/mcelog.c
213
xen_mcelog.next++;
drivers/xen/privcmd.c
196
pos = pos->next;
drivers/xen/privcmd.c
230
pos = pos->next;
drivers/xen/xen-pciback/pci_stub.c
502
psdev = container_of(seized_devices.next,
drivers/xen/xen-pciback/vpci.c
28
return head->next;
drivers/xen/xenbus/xenbus_client.c
57
struct list_head next;
drivers/xen/xenbus/xenbus_client.c
711
list_add(&node->next, &xenbus_valloc_pages);
drivers/xen/xenbus/xenbus_client.c
787
list_add(&node->next, &xenbus_valloc_pages);
drivers/xen/xenbus/xenbus_client.c
814
list_for_each_entry(node, &xenbus_valloc_pages, next) {
drivers/xen/xenbus/xenbus_client.c
816
list_del(&node->next);
drivers/xen/xenbus/xenbus_client.c
901
list_for_each_entry(node, &xenbus_valloc_pages, next) {
drivers/xen/xenbus/xenbus_client.c
904
list_del(&node->next);
drivers/xen/xenbus/xenbus_dev_frontend.c
148
rb = list_entry(u->read_buffers.next, struct read_buffer, list);
drivers/xen/xenbus/xenbus_dev_frontend.c
170
rb = list_entry(u->read_buffers.next,
drivers/xen/xenbus/xenbus_dev_frontend.c
220
rb = list_entry(list->next, struct read_buffer, list);
drivers/xen/xenbus/xenbus_dev_frontend.c
221
list_del(list->next);
drivers/xen/xenfs/xensyms.c
106
.next = xensyms_next,
drivers/zorro/proc.c
93
.next = zorro_seq_next,
fs/affs/amigaffs.h
121
__be32 next;
fs/affs/file.c
608
u32 tmp_next = be32_to_cpu(AFFS_DATA_HEAD(prev_bh)->next);
fs/affs/file.c
614
AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr);
fs/affs/file.c
757
AFFS_DATA_HEAD(bh)->next = 0;
fs/affs/file.c
760
u32 tmp_next = be32_to_cpu(AFFS_DATA_HEAD(prev_bh)->next);
fs/affs/file.c
766
AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr);
fs/affs/file.c
791
AFFS_DATA_HEAD(bh)->next = 0;
fs/affs/file.c
794
u32 tmp_next = be32_to_cpu(AFFS_DATA_HEAD(prev_bh)->next);
fs/affs/file.c
800
AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr);
fs/affs/file.c
960
tmp = be32_to_cpu(AFFS_DATA_HEAD(bh)->next);
fs/affs/file.c
961
AFFS_DATA_HEAD(bh)->next = 0;
fs/afs/cell.c
733
for (p = &net->proc_cells.first; *p; p = &(*p)->next) {
fs/afs/cell.c
740
cell->proc_link.next = *p;
fs/afs/cell.c
741
rcu_assign_pointer(*p, &cell->proc_link.next);
fs/afs/cell.c
742
if (cell->proc_link.next)
fs/afs/cell.c
743
cell->proc_link.next->pprev = &cell->proc_link.next;
fs/afs/dir.c
373
unsigned int blknum, base, hdr, pos, next, nr_slots;
fs/afs/dir.c
385
for (unsigned int slot = hdr; slot < AFS_DIR_SLOTS_PER_BLOCK; slot = next) {
fs/afs/dir.c
390
next = slot + 1;
fs/afs/dir.c
391
if (next >= pos)
fs/afs/dir.c
392
ctx->pos = (base + next) * sizeof(union afs_xdr_dirent);
fs/afs/dir.c
411
next = slot + nr_slots;
fs/afs/dir.c
412
if (next > AFS_DIR_SLOTS_PER_BLOCK) {
fs/afs/dir.c
431
if (next > pos)
fs/afs/dir.c
432
ctx->pos = (base + next) * sizeof(union afs_xdr_dirent);
fs/afs/dir.c
446
ctx->pos = (base + next) * sizeof(union afs_xdr_dirent);
fs/afs/dir_edit.c
136
for (; fq; fq = fq->next) {
fs/afs/dir_edit.c
402
__be16 next;
fs/afs/dir_edit.c
459
next = de->u.hash_next;
fs/afs/dir_edit.c
476
meta->meta.hashtable[iter.bucket] = next;
fs/afs/dir_edit.c
495
pde->u.hash_next = next;
fs/afs/dir_search.c
93
for (; fq; fq = fq->next) {
fs/afs/flock.c
113
struct file_lock *p, *_p, *next = NULL;
fs/afs/flock.c
132
if (next && (lock_is_write(next) || lock_is_read(p)))
fs/afs/flock.c
134
next = p;
fs/afs/flock.c
140
if (next) {
fs/afs/flock.c
142
next->fl_u.afs.state = AFS_LOCK_YOUR_TRY;
fs/afs/flock.c
143
trace_afs_flock_op(vnode, next, afs_flock_op_wake);
fs/afs/flock.c
144
locks_wake_up(next);
fs/afs/flock.c
164
p = list_entry(vnode->pending_locks.next,
fs/afs/flock.c
590
ASSERTCMP(vnode->pending_locks.next, ==, &fl->fl_u.afs.link);
fs/afs/inode.c
771
struct afs_wb_key *wbk = list_entry(vnode->wb_keys.next,
fs/afs/proc.c
310
.next = afs_proc_cell_volumes_next,
fs/afs/proc.c
426
.next = afs_proc_cell_vlservers_next,
fs/afs/proc.c
504
.next = afs_proc_servers_next,
fs/afs/proc.c
559
.next = afs_proc_sysname_next,
fs/afs/proc.c
79
.next = afs_proc_cells_next,
fs/afs/write.c
305
wbk = list_entry(graveyard.next, struct afs_wb_key, vnode_link);
fs/aio.c
1192
struct aio_waiter *curr, *next;
fs/aio.c
1196
list_for_each_entry_safe(curr, next, &ctx->wait.head, w.entry)
fs/autofs/autofs_i.h
94
struct autofs_wait_queue *next;
fs/autofs/expire.c
464
goto next;
fs/autofs/expire.c
472
next:
fs/autofs/root.c
157
goto next;
fs/autofs/root.c
162
goto next;
fs/autofs/root.c
164
goto next;
fs/autofs/root.c
167
goto next;
fs/autofs/root.c
169
goto next;
fs/autofs/root.c
177
next:
fs/autofs/root.c
217
goto next;
fs/autofs/root.c
222
goto next;
fs/autofs/root.c
224
goto next;
fs/autofs/root.c
227
goto next;
fs/autofs/root.c
229
goto next;
fs/autofs/root.c
237
next:
fs/autofs/waitq.c
184
for (wq = sbi->queues; wq; wq = wq->next) {
fs/autofs/waitq.c
31
nwq = wq->next;
fs/autofs/waitq.c
389
wq->next = sbi->queues;
fs/autofs/waitq.c
493
for (wql = &sbi->queues; (wq = *wql) != NULL; wql = &wq->next) {
fs/autofs/waitq.c
503
*wql = wq->next; /* Unlink from chain */
fs/binfmt_elf.c
1683
struct elf_thread_core_info *next;
fs/binfmt_elf.c
1882
for (ct = dump_task->signal->core_state->dumper.next; ct; ct = ct->next) {
fs/binfmt_elf.c
1888
t->next = info->thread->next;
fs/binfmt_elf.c
1889
info->thread->next = t;
fs/binfmt_elf.c
1895
for (t = info->thread; t != NULL; t = t->next)
fs/binfmt_elf.c
1949
t = t->next;
fs/binfmt_elf.c
1961
threads = t->next;
fs/binfmt_elf_fdpic.c
1382
struct elf_thread_status *next;
fs/binfmt_elf_fdpic.c
1502
for (ct = current->signal->core_state->dumper.next;
fs/binfmt_elf_fdpic.c
1503
ct; ct = ct->next) {
fs/binfmt_elf_fdpic.c
1509
tmp->next = thread_list;
fs/binfmt_elf_fdpic.c
1518
tmp->next = thread_list;
fs/binfmt_elf_fdpic.c
1628
for (tmp = thread_list->next; tmp; tmp = tmp->next) {
fs/binfmt_elf_fdpic.c
1657
thread_list = thread_list->next;
fs/binfmt_misc.c
866
Node *e, *next;
fs/binfmt_misc.c
893
list_for_each_entry_safe(e, next, &misc->entries, list)
fs/btrfs/backref.c
1665
while (eie->next)
fs/btrfs/backref.c
1666
eie = eie->next;
fs/btrfs/backref.c
1667
eie->next = ref->inode_list;
fs/btrfs/backref.c
2398
for (eie = inode_list; eie; eie = eie->next) {
fs/btrfs/backref.c
305
while (eie && eie->next)
fs/btrfs/backref.c
306
eie = eie->next;
fs/btrfs/backref.c
311
eie->next = newref->inode_list;
fs/btrfs/backref.c
32
struct extent_inode_elem *next;
fs/btrfs/backref.c
541
goto next;
fs/btrfs/backref.c
551
goto next;
fs/btrfs/backref.c
559
goto next;
fs/btrfs/backref.c
565
while (old->next)
fs/btrfs/backref.c
566
old = old->next;
fs/btrfs/backref.c
567
old->next = eie;
fs/btrfs/backref.c
571
next:
fs/btrfs/backref.c
85
e->next = *eie;
fs/btrfs/backref.c
99
eie_next = eie->next;
fs/btrfs/block-group.c
1611
goto next;
fs/btrfs/block-group.c
1637
goto next;
fs/btrfs/block-group.c
1675
goto next;
fs/btrfs/block-group.c
1686
goto next;
fs/btrfs/block-group.c
1696
goto next;
fs/btrfs/block-group.c
1708
goto next;
fs/btrfs/block-group.c
1800
next:
fs/btrfs/block-group.c
1956
goto next;
fs/btrfs/block-group.c
1975
goto next;
fs/btrfs/block-group.c
1992
goto next;
fs/btrfs/block-group.c
2009
goto next;
fs/btrfs/block-group.c
2015
goto next;
fs/btrfs/block-group.c
2059
next:
fs/btrfs/block-group.c
2874
goto next;
fs/btrfs/block-group.c
2905
next:
fs/btrfs/block-group.c
4914
goto next;
fs/btrfs/block-group.c
4917
goto next;
fs/btrfs/block-group.c
4920
goto next;
fs/btrfs/block-group.c
4948
next:
fs/btrfs/block-group.c
772
next:
fs/btrfs/block-group.c
801
goto next;
fs/btrfs/block-group.c
819
goto next;
fs/btrfs/compression.c
157
struct list_head *tmp, *next;
fs/btrfs/compression.c
170
list_for_each_safe(tmp, next, &remove) {
fs/btrfs/compression.c
781
ws = gwsm->idle_ws.next;
fs/btrfs/compression.c
817
workspace = idle_ws->next;
fs/btrfs/ctree.c
4730
next:
fs/btrfs/ctree.c
4763
goto next;
fs/btrfs/ctree.c
4773
goto next;
fs/btrfs/ctree.c
4788
struct extent_buffer *next;
fs/btrfs/ctree.c
4810
next = NULL;
fs/btrfs/ctree.c
4890
next = c;
fs/btrfs/ctree.c
4891
ret = read_block_for_search(root, path, &next, slot, &key);
fs/btrfs/ctree.c
4901
ret = btrfs_try_tree_read_lock(next);
fs/btrfs/ctree.c
4914
free_extent_buffer(next);
fs/btrfs/ctree.c
4920
btrfs_tree_read_lock(next);
fs/btrfs/ctree.c
4927
path->nodes[level] = next;
fs/btrfs/ctree.c
4934
ret = read_block_for_search(root, path, &next, 0, &key);
fs/btrfs/ctree.c
4945
if (!btrfs_try_tree_read_lock(next)) {
fs/btrfs/ctree.c
4950
btrfs_tree_read_lock(next);
fs/btrfs/defrag.c
1026
goto next;
fs/btrfs/defrag.c
1037
goto next;
fs/btrfs/defrag.c
1044
goto next;
fs/btrfs/defrag.c
1062
goto next;
fs/btrfs/defrag.c
1067
goto next;
fs/btrfs/defrag.c
1088
goto next;
fs/btrfs/defrag.c
1104
next:
fs/btrfs/defrag.c
200
struct inode_defrag *defrag, *next;
fs/btrfs/defrag.c
204
rbtree_postorder_for_each_entry_safe(defrag, next,
fs/btrfs/defrag.c
668
goto next;
fs/btrfs/defrag.c
678
goto next;
fs/btrfs/defrag.c
711
goto next;
fs/btrfs/defrag.c
716
next:
fs/btrfs/defrag.c
794
struct extent_map *next;
fs/btrfs/defrag.c
808
next = defrag_lookup_extent(inode, em_end, newer_than, locked);
fs/btrfs/defrag.c
810
if (!next || next->disk_bytenr >= EXTENT_MAP_LAST_BYTE)
fs/btrfs/defrag.c
812
if (next->flags & EXTENT_FLAG_PREALLOC)
fs/btrfs/defrag.c
818
if (next->len >= get_extent_max_capacity(fs_info, em))
fs/btrfs/defrag.c
821
if (next->generation < newer_than)
fs/btrfs/defrag.c
824
if (next->len >= extent_thresh)
fs/btrfs/defrag.c
829
btrfs_free_extent_map(next);
fs/btrfs/defrag.c
984
goto next;
fs/btrfs/defrag.c
989
goto next;
fs/btrfs/defrag.c
993
goto next;
fs/btrfs/defrag.c
997
goto next;
fs/btrfs/delayed-inode.c
1748
struct btrfs_delayed_item *curr, *next;
fs/btrfs/delayed-inode.c
1750
list_for_each_entry_safe(curr, next, ins_list, readdir_list) {
fs/btrfs/delayed-inode.c
1756
list_for_each_entry_safe(curr, next, del_list, readdir_list) {
fs/btrfs/delayed-inode.c
1792
struct btrfs_delayed_item *curr, *next;
fs/btrfs/delayed-inode.c
1803
list_for_each_entry_safe(curr, next, ins_list, readdir_list) {
fs/btrfs/delayed-inode.c
2192
struct btrfs_delayed_item *next;
fs/btrfs/delayed-inode.c
2201
list_for_each_entry_safe(item, next, ins_list, log_list) {
fs/btrfs/delayed-inode.c
2208
list_for_each_entry_safe(item, next, del_list, log_list) {
fs/btrfs/delayed-inode.c
258
struct btrfs_delayed_node *next = NULL;
fs/btrfs/delayed-inode.c
266
p = delayed_root->node_list.next;
fs/btrfs/delayed-inode.c
270
p = node->n_list.next;
fs/btrfs/delayed-inode.c
272
next = list_entry(p, struct btrfs_delayed_node, n_list);
fs/btrfs/delayed-inode.c
273
refcount_inc(&next->refs);
fs/btrfs/delayed-inode.c
274
btrfs_delayed_node_ref_tracker_alloc(next, tracker, GFP_ATOMIC);
fs/btrfs/delayed-inode.c
278
return next;
fs/btrfs/delayed-inode.c
663
struct btrfs_delayed_item *next;
fs/btrfs/delayed-inode.c
704
next = __btrfs_next_delayed_item(curr);
fs/btrfs/delayed-inode.c
705
if (!next)
fs/btrfs/delayed-inode.c
712
if (continuous_keys_only && (next->index != curr->index + 1))
fs/btrfs/delayed-inode.c
715
ASSERT(next->bytes_reserved == 0);
fs/btrfs/delayed-inode.c
717
next_size = next->data_len + sizeof(struct btrfs_item);
fs/btrfs/delayed-inode.c
721
list_add_tail(&next->tree_list, &item_list);
fs/btrfs/delayed-inode.c
724
batch.total_data_size += next->data_len;
fs/btrfs/delayed-inode.c
725
curr = next;
fs/btrfs/delayed-inode.c
788
if (next && !continuous_keys_only) {
fs/btrfs/delayed-inode.c
797
} else if (!next) {
fs/btrfs/delayed-inode.c
809
list_for_each_entry_safe(curr, next, &item_list, tree_list) {
fs/btrfs/delayed-inode.c
847
struct btrfs_delayed_item *curr, *next;
fs/btrfs/delayed-inode.c
878
next = __btrfs_next_delayed_item(curr);
fs/btrfs/delayed-inode.c
879
if (!next)
fs/btrfs/delayed-inode.c
886
key.offset != next->index)
fs/btrfs/delayed-inode.c
889
curr = next;
fs/btrfs/delayed-inode.c
910
list_for_each_entry_safe(curr, next, &batch_list, tree_list) {
fs/btrfs/delayed-ref.c
390
struct btrfs_delayed_ref_node *next;
fs/btrfs/delayed-ref.c
397
next = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/delayed-ref.c
399
if (seq && next->seq >= seq)
fs/btrfs/delayed-ref.c
401
if (comp_refs(ref, next, false))
fs/btrfs/delayed-ref.c
404
if (ref->action == next->action) {
fs/btrfs/delayed-ref.c
405
mod = next->ref_mod;
fs/btrfs/delayed-ref.c
407
if (ref->ref_mod < next->ref_mod) {
fs/btrfs/delayed-ref.c
408
swap(ref, next);
fs/btrfs/delayed-ref.c
411
mod = -next->ref_mod;
fs/btrfs/delayed-ref.c
414
drop_delayed_ref(fs_info, delayed_refs, head, next);
fs/btrfs/dev-replace.c
827
struct rb_node *next = rb_next(node);
fs/btrfs/dev-replace.c
850
node = next;
fs/btrfs/discard.c
733
struct btrfs_block_group *block_group, *next;
fs/btrfs/discard.c
737
list_for_each_entry_safe(block_group, next, &fs_info->unused_bgs,
fs/btrfs/discard.c
763
struct btrfs_block_group *block_group, *next;
fs/btrfs/discard.c
768
list_for_each_entry_safe(block_group, next,
fs/btrfs/extent-io-tree.c
1548
struct extent_state *prev = NULL, *next = NULL;
fs/btrfs/extent-io-tree.c
1554
state = tree_search_prev_next(tree, start, &prev, &next);
fs/btrfs/extent-io-tree.c
1555
if (!state && !next && !prev) {
fs/btrfs/extent-io-tree.c
1563
} else if (!state && !next) {
fs/btrfs/extent-io-tree.c
1572
state = next;
fs/btrfs/extent-io-tree.c
1929
struct extent_state *next;
fs/btrfs/extent-io-tree.c
1933
next = next_state(state);
fs/btrfs/extent-io-tree.c
1934
if (next)
fs/btrfs/extent-io-tree.c
1935
refcount_inc(&next->refs);
fs/btrfs/extent-io-tree.c
1938
return next;
fs/btrfs/extent-io-tree.c
203
struct rb_node *next = rb_next(&state->rb_node);
fs/btrfs/extent-io-tree.c
205
return rb_entry_safe(next, struct extent_state, rb_node);
fs/btrfs/extent-io-tree.c
210
struct rb_node *next = rb_prev(&state->rb_node);
fs/btrfs/extent-io-tree.c
212
return rb_entry_safe(next, struct extent_state, rb_node);
fs/btrfs/extent-io-tree.c
356
struct extent_state *next;
fs/btrfs/extent-io-tree.c
358
next = next_state(state);
fs/btrfs/extent-io-tree.c
359
if (next && next->start == state->end + 1 && next->state == state->state) {
fs/btrfs/extent-io-tree.c
361
btrfs_merge_delalloc_extent(tree->inode, state, next);
fs/btrfs/extent-io-tree.c
362
state->end = next->end;
fs/btrfs/extent-io-tree.c
363
rb_erase(&next->rb_node, &tree->state);
fs/btrfs/extent-io-tree.c
364
RB_CLEAR_NODE(&next->rb_node);
fs/btrfs/extent-io-tree.c
365
btrfs_free_extent_state(next);
fs/btrfs/extent-io-tree.c
562
struct extent_state *next;
fs/btrfs/extent-io-tree.c
575
next = next_search_state(state, end);
fs/btrfs/extent-io-tree.c
585
next = next_search_state(state, end);
fs/btrfs/extent-io-tree.c
587
return next;
fs/btrfs/extent-io-tree.c
680
goto next;
fs/btrfs/extent-io-tree.c
711
goto next;
fs/btrfs/extent-io-tree.c
747
next:
fs/btrfs/extent-tree.c
5776
struct extent_buffer *next)
fs/btrfs/extent-tree.c
5783
btrfs_assert_tree_write_locked(next);
fs/btrfs/extent-tree.c
5787
if (btrfs_buffer_uptodate(next, generation, false))
fs/btrfs/extent-tree.c
5796
btrfs_tree_unlock(next);
fs/btrfs/extent-tree.c
5799
ret = btrfs_read_extent_buffer(next, &check);
fs/btrfs/extent-tree.c
5801
free_extent_buffer(next);
fs/btrfs/extent-tree.c
5804
btrfs_tree_lock(next);
fs/btrfs/extent-tree.c
5821
struct extent_buffer *next, u64 owner_root)
fs/btrfs/extent-tree.c
5825
.bytenr = next->start,
fs/btrfs/extent-tree.c
5854
ret = check_ref_exists(trans, root, next->start, ref.parent,
fs/btrfs/extent-tree.c
5872
ret = btrfs_qgroup_trace_subtree(trans, next, generation, level - 1);
fs/btrfs/extent-tree.c
5915
struct extent_buffer *next;
fs/btrfs/extent-tree.c
5934
next = btrfs_find_create_tree_block(fs_info, bytenr, btrfs_root_id(root),
fs/btrfs/extent-tree.c
5936
if (IS_ERR(next))
fs/btrfs/extent-tree.c
5937
return PTR_ERR(next);
fs/btrfs/extent-tree.c
5939
btrfs_tree_lock(next);
fs/btrfs/extent-tree.c
5971
ret = check_next_block_uptodate(trans, root, path, wc, next);
fs/btrfs/extent-tree.c
5976
ASSERT(level == btrfs_header_level(next));
fs/btrfs/extent-tree.c
5977
if (unlikely(level != btrfs_header_level(next))) {
fs/btrfs/extent-tree.c
5982
path->nodes[level] = next;
fs/btrfs/extent-tree.c
5990
ret = maybe_drop_reference(trans, root, path, wc, next, owner_root);
fs/btrfs/extent-tree.c
5999
btrfs_tree_unlock(next);
fs/btrfs/extent-tree.c
6000
free_extent_buffer(next);
fs/btrfs/extent_io.c
2859
goto next;
fs/btrfs/extent_io.c
2875
goto next;
fs/btrfs/extent_io.c
2888
next:
fs/btrfs/extent_io.c
70
if (!fs_info->allocated_ebs.next)
fs/btrfs/extent_map.c
1163
struct rb_node *next = rb_next(node);
fs/btrfs/extent_map.c
1170
goto next;
fs/btrfs/extent_map.c
1187
next:
fs/btrfs/extent_map.c
1199
node = next;
fs/btrfs/extent_map.c
1228
goto next;
fs/btrfs/extent_map.c
1240
goto next;
fs/btrfs/extent_map.c
1247
next:
fs/btrfs/extent_map.c
231
static bool mergeable_maps(const struct extent_map *prev, const struct extent_map *next)
fs/btrfs/extent_map.c
233
if (btrfs_extent_map_end(prev) != next->start)
fs/btrfs/extent_map.c
241
(next->flags & ~EXTENT_FLAG_MERGED))
fs/btrfs/extent_map.c
244
if (next->disk_bytenr < EXTENT_MAP_LAST_BYTE - 1)
fs/btrfs/extent_map.c
245
return btrfs_extent_map_block_start(next) == extent_map_block_end(prev);
fs/btrfs/extent_map.c
248
return next->disk_bytenr == prev->disk_bytenr;
fs/btrfs/extent_map.c
265
static void merge_ondisk_extents(const struct extent_map *prev, const struct extent_map *next,
fs/btrfs/extent_map.c
274
ASSERT(!btrfs_extent_map_is_compressed(next));
fs/btrfs/extent_map.c
295
new_disk_bytenr = min(prev->disk_bytenr, next->disk_bytenr);
fs/btrfs/extent_map.c
297
next->disk_bytenr + next->disk_num_bytes) -
fs/btrfs/extent_map.c
630
struct rb_node *next;
fs/btrfs/extent_map.c
632
next = rb_next(&em->rb_node);
fs/btrfs/extent_map.c
633
if (!next)
fs/btrfs/extent_map.c
635
return container_of(next, struct extent_map, rb_node);
fs/btrfs/extent_map.c
660
struct extent_map *next;
fs/btrfs/extent_map.c
669
next = existing;
fs/btrfs/extent_map.c
670
prev = prev_extent_map(next);
fs/btrfs/extent_map.c
673
next = next_extent_map(prev);
fs/btrfs/extent_map.c
678
end = next ? next->start : btrfs_extent_map_end(em);
fs/btrfs/extent_map.c
782
struct rb_node *next = rb_next(node);
fs/btrfs/extent_map.c
792
node = next;
fs/btrfs/extent_map.c
866
goto next;
fs/btrfs/extent_map.c
987
next:
fs/btrfs/free-space-cache.c
1244
struct btrfs_free_space *entry, *next;
fs/btrfs/free-space-cache.c
1248
list_for_each_entry_safe(entry, next, bitmap_list, list) {
fs/btrfs/free-space-cache.c
1273
struct btrfs_free_space *entry, *next;
fs/btrfs/free-space-cache.c
1275
list_for_each_entry_safe(entry, next, bitmap_list, list)
fs/btrfs/free-space-cache.c
2183
struct rb_node *next = rb_next(&bitmap_info->offset_index);
fs/btrfs/free-space-cache.c
2191
if (!next)
fs/btrfs/free-space-cache.c
2194
bitmap_info = rb_entry(next, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
3037
while ((head = block_group->cluster_list.next) !=
fs/btrfs/free-space-cache.c
3768
goto next;
fs/btrfs/free-space-cache.c
3793
goto next;
fs/btrfs/free-space-cache.c
3813
next:
fs/btrfs/free-space-cache.c
3970
goto next;
fs/btrfs/free-space-cache.c
3996
goto next;
fs/btrfs/free-space-cache.c
4013
goto next;
fs/btrfs/free-space-cache.c
4045
next:
fs/btrfs/free-space-tree.c
1358
goto next;
fs/btrfs/free-space-tree.c
1366
next:
fs/btrfs/inode.c
10182
struct rb_node *node, *next;
fs/btrfs/inode.c
10187
next = rb_next(node);
fs/btrfs/inode.c
10198
node = next;
fs/btrfs/inode.c
2797
goto next;
fs/btrfs/inode.c
2808
next:
fs/btrfs/inode.c
5355
goto next;
fs/btrfs/inode.c
5373
next:
fs/btrfs/inode.c
7295
goto next;
fs/btrfs/inode.c
7318
next:
fs/btrfs/inode.c
7337
goto next;
fs/btrfs/inode.c
7736
goto next;
fs/btrfs/inode.c
7747
goto next;
fs/btrfs/inode.c
7761
goto next;
fs/btrfs/inode.c
7800
next:
fs/btrfs/inode.c
8909
struct btrfs_delalloc_work *work, *next;
fs/btrfs/inode.c
8962
list_for_each_entry_safe(work, next, &works, list) {
fs/btrfs/inode.c
9277
goto next;
fs/btrfs/inode.c
9291
next:
fs/btrfs/ordered-data.c
1093
struct rb_node *next;
fs/btrfs/ordered-data.c
1128
next = rb_next(cur);
fs/btrfs/ordered-data.c
1131
next = cur;
fs/btrfs/ordered-data.c
1138
if (next) {
fs/btrfs/ordered-data.c
1139
entry = rb_entry(next, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
753
struct btrfs_ordered_extent *ordered, *next;
fs/btrfs/ordered-data.c
801
list_for_each_entry_safe(ordered, next, &works, work_list) {
fs/btrfs/print-tree.c
623
struct extent_buffer *next;
fs/btrfs/print-tree.c
626
next = read_tree_block(fs_info, btrfs_node_blockptr(c, i), &check);
fs/btrfs/print-tree.c
627
if (IS_ERR(next))
fs/btrfs/print-tree.c
629
if (!extent_buffer_uptodate(next)) {
fs/btrfs/print-tree.c
630
free_extent_buffer(next);
fs/btrfs/print-tree.c
634
if (btrfs_is_leaf(next) &&
fs/btrfs/print-tree.c
637
if (btrfs_header_level(next) !=
fs/btrfs/print-tree.c
640
btrfs_print_tree(next, follow);
fs/btrfs/print-tree.c
641
free_extent_buffer(next);
fs/btrfs/qgroup.c
4114
#define rbtree_iterate_from_safe(node, next, start) \
fs/btrfs/qgroup.c
4115
for (node = start; node && ({ next = rb_next(node); 1;}); node = next)
fs/btrfs/qgroup.c
4122
struct rb_node *next;
fs/btrfs/qgroup.c
4141
rbtree_iterate_from_safe(node, next, &entry->rb_node) {
fs/btrfs/qgroup.c
4693
struct btrfs_qgroup_swapped_block *next;
fs/btrfs/qgroup.c
4695
rbtree_postorder_for_each_entry_safe(entry, next, cur_root,
fs/btrfs/raid56.c
912
struct btrfs_raid_bio *next;
fs/btrfs/raid56.c
913
struct list_head *head = rbio->plug_list.next;
fs/btrfs/raid56.c
915
next = list_entry(head, struct btrfs_raid_bio,
fs/btrfs/raid56.c
920
list_add(&next->hash_list, &h->hash_list);
fs/btrfs/raid56.c
921
refcount_inc(&next->refs);
fs/btrfs/raid56.c
925
if (next->operation == BTRFS_RBIO_READ_REBUILD) {
fs/btrfs/raid56.c
926
start_async_work(next, recover_rbio_work_locked);
fs/btrfs/raid56.c
927
} else if (next->operation == BTRFS_RBIO_WRITE) {
fs/btrfs/raid56.c
928
steal_rbio(rbio, next);
fs/btrfs/raid56.c
929
start_async_work(next, rmw_rbio_work_locked);
fs/btrfs/raid56.c
930
} else if (next->operation == BTRFS_RBIO_PARITY_SCRUB) {
fs/btrfs/raid56.c
931
steal_rbio(rbio, next);
fs/btrfs/raid56.c
932
start_async_work(next, scrub_rbio_work_locked);
fs/btrfs/raid56.c
949
struct bio *next;
fs/btrfs/raid56.c
952
next = cur->bi_next;
fs/btrfs/raid56.c
956
cur = next;
fs/btrfs/relocation.c
1455
struct btrfs_root *next;
fs/btrfs/relocation.c
1459
list_for_each_entry_safe(root, next, &rc->dirty_subvol_roots,
fs/btrfs/relocation.c
1969
struct btrfs_backref_node *next;
fs/btrfs/relocation.c
1974
next = walk_up_backref(node, edges, &index);
fs/btrfs/relocation.c
1975
root = next->root;
fs/btrfs/relocation.c
2020
if (unlikely(next->new_bytenr)) {
fs/btrfs/relocation.c
2027
ASSERT(next->new_bytenr == 0);
fs/btrfs/relocation.c
2030
node->bytenr, next->bytenr);
fs/btrfs/relocation.c
2034
next->new_bytenr = root->node->start;
fs/btrfs/relocation.c
2035
btrfs_put_root(next->root);
fs/btrfs/relocation.c
2036
next->root = btrfs_grab_root(root);
fs/btrfs/relocation.c
2037
ASSERT(next->root);
fs/btrfs/relocation.c
2038
mark_block_processed(rc, next);
fs/btrfs/relocation.c
2040
next = node;
fs/btrfs/relocation.c
2043
rc->backref_cache.path[next->level] = next;
fs/btrfs/relocation.c
2046
next = edges[index]->node[UPPER];
fs/btrfs/relocation.c
2063
struct btrfs_backref_node *next;
fs/btrfs/relocation.c
2069
next = node;
fs/btrfs/relocation.c
2072
next = walk_up_backref(next, edges, &index);
fs/btrfs/relocation.c
2073
root = next->root;
fs/btrfs/relocation.c
2089
if (next != node)
fs/btrfs/relocation.c
2092
next = walk_down_backref(edges, &index);
fs/btrfs/relocation.c
2093
if (!next || next->level <= node->level)
fs/btrfs/relocation.c
2106
struct btrfs_backref_node *next = node;
fs/btrfs/relocation.c
2114
while (next) {
fs/btrfs/relocation.c
2117
if (next->processed)
fs/btrfs/relocation.c
2122
if (list_empty(&next->upper))
fs/btrfs/relocation.c
2125
edge = list_first_entry(&next->upper, struct btrfs_backref_edge,
fs/btrfs/relocation.c
2128
next = edge->node[UPPER];
fs/btrfs/relocation.c
2130
next = walk_down_backref(edges, &index);
fs/btrfs/relocation.c
2219
goto next;
fs/btrfs/relocation.c
2226
goto next;
fs/btrfs/relocation.c
2230
goto next;
fs/btrfs/relocation.c
2260
goto next;
fs/btrfs/relocation.c
2272
goto next;
fs/btrfs/relocation.c
2276
goto next;
fs/btrfs/relocation.c
2283
goto next;
fs/btrfs/relocation.c
2293
goto next;
fs/btrfs/relocation.c
2324
next:
fs/btrfs/relocation.c
2395
struct btrfs_backref_node *next = node;
fs/btrfs/relocation.c
2400
while (next) {
fs/btrfs/relocation.c
2403
if (next->processed)
fs/btrfs/relocation.c
2406
mark_block_processed(rc, next);
fs/btrfs/relocation.c
2408
if (list_empty(&next->upper))
fs/btrfs/relocation.c
2411
edge = list_first_entry(&next->upper, struct btrfs_backref_edge,
fs/btrfs/relocation.c
2414
next = edge->node[UPPER];
fs/btrfs/relocation.c
2416
next = walk_down_backref(edges, &index);
fs/btrfs/relocation.c
2599
struct tree_block *next;
fs/btrfs/relocation.c
2609
rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) {
fs/btrfs/relocation.c
2617
rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) {
fs/btrfs/relocation.c
2626
rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) {
fs/btrfs/relocation.c
3367
next:
fs/btrfs/relocation.c
3385
goto next;
fs/btrfs/relocation.c
3391
goto next;
fs/btrfs/relocation.c
3398
goto next;
fs/btrfs/relocation.c
4973
goto next;
fs/btrfs/relocation.c
5059
next:
fs/btrfs/scrub.c
1565
goto next;
fs/btrfs/scrub.c
1572
next:
fs/btrfs/scrub.c
2552
goto next;
fs/btrfs/scrub.c
2567
next:
fs/btrfs/send.c
5905
goto next;
fs/btrfs/send.c
6044
next:
fs/btrfs/send.c
6376
goto next;
fs/btrfs/send.c
6385
goto next;
fs/btrfs/send.c
6390
goto next;
fs/btrfs/send.c
6393
next:
fs/btrfs/sysfs.c
2570
struct btrfs_qgroup *next;
fs/btrfs/sysfs.c
2575
rbtree_postorder_for_each_entry_safe(qgroup, next,
fs/btrfs/sysfs.c
2590
struct btrfs_qgroup *next;
fs/btrfs/sysfs.c
2609
rbtree_postorder_for_each_entry_safe(qgroup, next,
fs/btrfs/transaction.c
1952
struct btrfs_pending_snapshot *pending, *next;
fs/btrfs/transaction.c
1956
list_for_each_entry_safe(pending, next, head, list) {
fs/btrfs/tree-log.c
1291
goto next;
fs/btrfs/tree-log.c
1320
next:
fs/btrfs/tree-log.c
1667
goto next;
fs/btrfs/tree-log.c
1736
next:
fs/btrfs/tree-log.c
2378
goto next;
fs/btrfs/tree-log.c
2391
next:
fs/btrfs/tree-log.c
3002
struct extent_buffer *next;
fs/btrfs/tree-log.c
3024
next = btrfs_find_create_tree_block(fs_info, bytenr,
fs/btrfs/tree-log.c
3027
if (IS_ERR(next)) {
fs/btrfs/tree-log.c
3028
ret = PTR_ERR(next);
fs/btrfs/tree-log.c
3037
ret = wc->process_func(next, wc, ptr_gen, *level - 1);
fs/btrfs/tree-log.c
3039
free_extent_buffer(next);
fs/btrfs/tree-log.c
3045
ret = btrfs_read_extent_buffer(next, &check);
fs/btrfs/tree-log.c
3047
free_extent_buffer(next);
fs/btrfs/tree-log.c
3055
ret = clean_log_buffer(trans, next);
fs/btrfs/tree-log.c
3057
free_extent_buffer(next);
fs/btrfs/tree-log.c
3061
free_extent_buffer(next);
fs/btrfs/tree-log.c
3064
ret = btrfs_read_extent_buffer(next, &check);
fs/btrfs/tree-log.c
3066
free_extent_buffer(next);
fs/btrfs/tree-log.c
3076
path->nodes[*level-1] = next;
fs/btrfs/tree-log.c
3077
*level = btrfs_header_level(next);
fs/btrfs/tree-log.c
6036
struct btrfs_dir_list *next;
fs/btrfs/tree-log.c
6038
list_for_each_entry_safe(dir_elem, next, &dir_list, list)
fs/btrfs/tree-log.c
6054
struct btrfs_ino_list *next;
fs/btrfs/tree-log.c
6056
list_for_each_entry_safe(curr, next, &ctx->conflict_inodes, list) {
fs/btrfs/tree-log.c
6653
const struct btrfs_delayed_item *next;
fs/btrfs/tree-log.c
6661
next = list_next_entry(curr, log_list);
fs/btrfs/tree-log.c
6662
while (!list_entry_is_head(next, delayed_del_list, log_list)) {
fs/btrfs/tree-log.c
6663
if (next->index != curr->index + 1)
fs/btrfs/tree-log.c
6665
curr = next;
fs/btrfs/tree-log.c
6666
next = list_next_entry(next, log_list);
fs/btrfs/tree-log.c
6691
const struct btrfs_delayed_item *next;
fs/btrfs/tree-log.c
6697
next = list_next_entry(first, log_list);
fs/btrfs/tree-log.c
6700
!list_entry_is_head(next, delayed_del_list, log_list)) {
fs/btrfs/tree-log.c
6706
key.offset != next->index)
fs/btrfs/tree-log.c
6710
*last_ret = next;
fs/btrfs/tree-log.c
6711
next = list_next_entry(next, log_list);
fs/btrfs/tree-log.c
7777
goto next;
fs/btrfs/tree-log.c
7784
goto next;
fs/btrfs/tree-log.c
7790
goto next;
fs/btrfs/tree-log.c
7801
goto next;
fs/btrfs/tree-log.c
7814
goto next;
fs/btrfs/tree-log.c
7817
next:
fs/btrfs/tree-mod-log.c
125
for (node = rb_first(tm_root); node; node = next) {
fs/btrfs/tree-mod-log.c
126
next = rb_next(node);
fs/btrfs/tree-mod-log.c
832
struct rb_node *next;
fs/btrfs/tree-mod-log.c
90
struct rb_node *next;
fs/btrfs/tree-mod-log.c
929
next = rb_next(&tm->node);
fs/btrfs/tree-mod-log.c
930
if (!next)
fs/btrfs/tree-mod-log.c
932
tm = rb_entry(next, struct tree_mod_elem, node);
fs/btrfs/ulist.c
289
if (uiter->cur_list && uiter->cur_list->next == &ulist->nodes)
fs/btrfs/ulist.c
292
uiter->cur_list = uiter->cur_list->next;
fs/btrfs/ulist.c
294
uiter->cur_list = ulist->nodes.next;
fs/btrfs/ulist.c
67
struct ulist_node *next;
fs/btrfs/ulist.c
69
list_for_each_entry_safe(node, next, &ulist->nodes, list) {
fs/btrfs/volumes.c
1060
struct btrfs_device *device, *next;
fs/btrfs/volumes.c
1063
list_for_each_entry_safe(device, next, &fs_devices->devices, dev_list) {
fs/btrfs/volumes.c
1606
goto next;
fs/btrfs/volumes.c
1630
goto next;
fs/btrfs/volumes.c
1649
next:
fs/btrfs/volumes.c
1871
goto next;
fs/btrfs/volumes.c
1877
goto next;
fs/btrfs/volumes.c
1912
next:
fs/btrfs/volumes.c
8195
struct btrfs_device *curr, *next;
fs/btrfs/volumes.c
8208
list_for_each_entry_safe(curr, next, &trans->dev_update_list,
fs/btrfs/xattr.c
321
goto next;
fs/btrfs/xattr.c
333
next:
fs/btrfs/zstd.c
115
struct list_head *pos, *next;
fs/btrfs/zstd.c
124
list_for_each_prev_safe(pos, next, &zwsm->lru_list) {
fs/btrfs/zstd.c
222
workspace = container_of(zwsm->idle_ws[i].next,
fs/btrfs/zstd.c
257
ws = zwsm->idle_ws[i].next;
fs/buffer.c
1630
struct buffer_head *head, *bh, *next;
fs/buffer.c
1648
next = bh->b_this_page;
fs/buffer.c
1662
bh = next;
fs/buffer.c
1768
goto next;
fs/buffer.c
1774
next:
fs/buffer.c
1928
struct buffer_head *next = bh->b_this_page;
fs/buffer.c
1934
bh = next;
fs/buffer.c
1982
struct buffer_head *next = bh->b_this_page;
fs/buffer.c
1989
bh = next;
fs/buffer.c
2901
struct buffer_head *next = bh->b_this_page;
fs/buffer.c
2905
bh = next;
fs/buffer.c
2982
struct buffer_head *next = bh->b_this_page;
fs/buffer.c
2984
bh = next;
fs/buffer.c
791
bh = BH_ENTRY(list->next);
fs/buffer.c
874
__remove_assoc_queue(BH_ENTRY(list->next));
fs/buffer.c
897
struct buffer_head *bh = BH_ENTRY(list->next);
fs/ceph/dir.c
1671
goto next;
fs/ceph/dir.c
1699
next:
fs/ceph/mds_client.c
1831
p = session->s_caps.next;
fs/ceph/mds_client.c
1838
p = p->next;
fs/ceph/mds_client.c
1858
p = p->next;
fs/ceph/mds_client.c
1940
cap = list_entry(session->s_caps.next,
fs/ceph/mds_client.c
3666
req = list_entry(tmp_list.next,
fs/ceph/mds_client.c
858
goto next;
fs/ceph/mds_client.c
860
goto next;
fs/ceph/mds_client.c
862
goto next;
fs/ceph/mds_client.c
869
goto next;
fs/ceph/mds_client.c
874
next:
fs/ceph/quota.c
220
struct ceph_snap_realm *realm, *next;
fs/ceph/quota.c
262
next = realm->parent;
fs/ceph/quota.c
263
if (has_quota || !next) {
fs/ceph/quota.c
269
ceph_get_snap_realm(mdsc, next);
fs/ceph/quota.c
271
realm = next;
fs/ceph/quota.c
332
struct ceph_snap_realm *realm, *next;
fs/ceph/quota.c
403
next = realm->parent;
fs/ceph/quota.c
404
if (exceeded || !next)
fs/ceph/quota.c
406
ceph_get_snap_realm(mdsc, next);
fs/ceph/quota.c
408
realm = next;
fs/char_dev.c
136
for (curr = chrdevs[i]; curr; prev = curr, curr = curr->next) {
fs/char_dev.c
158
cd->next = curr;
fs/char_dev.c
161
cd->next = prev->next;
fs/char_dev.c
162
prev->next = cd;
fs/char_dev.c
175
for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next)
fs/char_dev.c
182
*cp = cd->next;
fs/char_dev.c
201
dev_t n, next;
fs/char_dev.c
203
for (n = from; n < to; n = next) {
fs/char_dev.c
204
next = MKDEV(MAJOR(n)+1, 0);
fs/char_dev.c
205
if (next > to)
fs/char_dev.c
206
next = to;
fs/char_dev.c
208
next - n, name);
fs/char_dev.c
215
for (n = from; n < to; n = next) {
fs/char_dev.c
216
next = MKDEV(MAJOR(n)+1, 0);
fs/char_dev.c
217
kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n));
fs/char_dev.c
311
dev_t n, next;
fs/char_dev.c
313
for (n = from; n < to; n = next) {
fs/char_dev.c
314
next = MKDEV(MAJOR(n)+1, 0);
fs/char_dev.c
315
if (next > to)
fs/char_dev.c
316
next = to;
fs/char_dev.c
317
kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n));
fs/char_dev.c
36
struct char_device_struct *next;
fs/char_dev.c
437
inode = container_of(cdev->list.next, struct inode, i_devices);
fs/char_dev.c
57
for (cd = chrdevs[major_to_index(offset)]; cd; cd = cd->next) {
fs/char_dev.c
78
for (cd = chrdevs[major_to_index(i)]; cd; cd = cd->next)
fs/coda/psdev.c
240
req = list_entry(vcp->vc_pending.next, struct upc_req,uc_chain);
fs/configfs/dir.c
1647
for (p = q->next; p != &parent_sd->s_children; p = p->next) {
fs/configfs/dir.c
1648
struct configfs_dirent *next;
fs/configfs/dir.c
1653
next = list_entry(p, struct configfs_dirent, s_sibling);
fs/configfs/dir.c
1654
if (!next->s_element)
fs/configfs/dir.c
1670
dentry = next->s_dentry;
fs/configfs/dir.c
1679
name = configfs_get_name(next);
fs/configfs/dir.c
1683
fs_umode_to_dtype(next->s_mode)))
fs/configfs/dir.c
1720
p = sd->s_children.next;
fs/configfs/dir.c
1722
struct configfs_dirent *next;
fs/configfs/dir.c
1723
next = list_entry(p, struct configfs_dirent,
fs/configfs/dir.c
1725
if (next->s_element)
fs/configfs/dir.c
1727
p = p->next;
fs/coredump.c
528
core_state->dumper.next = NULL;
fs/coredump.c
541
ptr = core_state->dumper.next;
fs/coredump.c
544
ptr = ptr->next;
fs/coredump.c
553
struct core_thread *curr, *next;
fs/coredump.c
559
next = current->signal->core_state->dumper.next;
fs/coredump.c
563
while ((curr = next) != NULL) {
fs/coredump.c
564
next = curr->next;
fs/dcache.c
2454
goto next;
fs/dcache.c
2456
goto next;
fs/dcache.c
2459
goto next;
fs/dcache.c
2465
next:
fs/dcache.c
610
struct dentry *next;
fs/dcache.c
638
while (dentry->d_sib.next) {
fs/dcache.c
639
next = hlist_entry(dentry->d_sib.next, struct dentry, d_sib);
fs/dcache.c
640
if (likely(!(next->d_flags & DCACHE_DENTRY_CURSOR)))
fs/dcache.c
642
dentry->d_sib.next = next->d_sib.next;
fs/dlm/debug_fs.c
442
.next = table_seq_next,
fs/dlm/debug_fs.c
449
.next = table_seq_next,
fs/dlm/debug_fs.c
456
.next = table_seq_next,
fs/dlm/debug_fs.c
463
.next = table_seq_next,
fs/dlm/dir.c
323
list = r->res_masters_list.next;
fs/dlm/dir.c
341
list = ls->ls_masters_list.next;
fs/dlm/dir.c
345
for (offset = 0; list != &ls->ls_masters_list; list = list->next) {
fs/dlm/lock.c
2141
struct dlm_lkb *first = list_entry(head->next, struct dlm_lkb,
fs/dlm/lock.c
2739
lkb = list_entry(r->res_lookup.next, struct dlm_lkb,
fs/dlm/lock.c
6125
lkb = list_entry(proc->locks.next, struct dlm_lkb, lkb_ownqueue);
fs/dlm/lock.c
6197
lkb = list_entry(proc->locks.next, struct dlm_lkb,
fs/dlm/member.c
378
memb = list_entry(head->next, struct dlm_member, list);
fs/erofs/decompressor_deflate.c
126
z_erofs_deflate_head = strm->next;
fs/erofs/decompressor_deflate.c
174
strm->next = z_erofs_deflate_head;
fs/erofs/decompressor_deflate.c
34
struct z_erofs_deflate *n = strm->next;
fs/erofs/decompressor_deflate.c
6
struct z_erofs_deflate *next;
fs/erofs/decompressor_deflate.c
85
strm->next = z_erofs_deflate_head;
fs/erofs/decompressor_lzma.c
121
for (last = strm; last->next; last = last->next)
fs/erofs/decompressor_lzma.c
123
last->next = head;
fs/erofs/decompressor_lzma.c
129
for (strm = head; strm; strm = strm->next) {
fs/erofs/decompressor_lzma.c
177
z_erofs_lzma_head = strm->next;
fs/erofs/decompressor_lzma.c
222
strm->next = z_erofs_lzma_head;
fs/erofs/decompressor_lzma.c
37
struct z_erofs_lzma *n = strm->next;
fs/erofs/decompressor_lzma.c
6
struct z_erofs_lzma *next;
fs/erofs/decompressor_lzma.c
64
strm->next = z_erofs_lzma_head;
fs/erofs/decompressor_zstd.c
107
n = strm->next;
fs/erofs/decompressor_zstd.c
108
strm->next = head;
fs/erofs/decompressor_zstd.c
116
for (strm = head; strm; strm = strm->next) {
fs/erofs/decompressor_zstd.c
207
strm->next = z_erofs_zstd_head;
fs/erofs/decompressor_zstd.c
32
z_erofs_zstd_head = all ? NULL : strm->next;
fs/erofs/decompressor_zstd.c
43
n = strm->next;
fs/erofs/decompressor_zstd.c
6
struct z_erofs_zstd *next;
fs/erofs/decompressor_zstd.c
68
strm->next = z_erofs_zstd_head;
fs/erofs/zdata.c
1384
WRITE_ONCE(pcl->next, NULL);
fs/erofs/zdata.c
1404
struct z_erofs_pcluster *next;
fs/erofs/zdata.c
1407
for (; be.pcl != Z_EROFS_PCLUSTER_TAIL; be.pcl = next) {
fs/erofs/zdata.c
1409
next = READ_ONCE(be.pcl->next);
fs/erofs/zdata.c
1628
struct z_erofs_pcluster *next,
fs/erofs/zdata.c
1631
WRITE_ONCE(pcl->next, Z_EROFS_PCLUSTER_TAIL);
fs/erofs/zdata.c
1632
WRITE_ONCE(*qtail[JQ_SUBMIT], next);
fs/erofs/zdata.c
1634
qtail[JQ_BYPASS] = &pcl->next;
fs/erofs/zdata.c
1670
struct z_erofs_pcluster *pcl, *next;
fs/erofs/zdata.c
1686
q[JQ_SUBMIT]->head = next = f->head;
fs/erofs/zdata.c
1695
pcl = next;
fs/erofs/zdata.c
1696
next = READ_ONCE(pcl->next);
fs/erofs/zdata.c
1698
z_erofs_move_to_bypass_queue(pcl, next, qtail);
fs/erofs/zdata.c
1770
qtail[JQ_SUBMIT] = &pcl->next;
fs/erofs/zdata.c
1772
z_erofs_move_to_bypass_queue(pcl, next, qtail);
fs/erofs/zdata.c
1773
} while (next != Z_EROFS_PCLUSTER_TAIL);
fs/erofs/zdata.c
46
struct z_erofs_pcluster *next;
fs/erofs/zdata.c
759
pcl->next = fe->head;
fs/erofs/zdata.c
851
if (!cmpxchg(&fe->pcl->next, NULL, fe->head)) {
fs/erofs/zutil.c
264
p = erofs_sb_list.next;
fs/erofs/zutil.c
276
p = p->next;
fs/erofs/zutil.c
285
p = p->next;
fs/eventpoll.c
110
struct eppoll_entry *next;
fs/eventpoll.c
1285
if (epi->next == EP_UNACTIVE_PTR) {
fs/eventpoll.c
1286
epi->next = READ_ONCE(ep->ovflist);
fs/eventpoll.c
1381
pwq->next = epi->pwqlist;
fs/eventpoll.c
146
struct epitem *next;
fs/eventpoll.c
1477
for (p = tfile_check_list; p != EP_UNACTIVE_PTR; p = p->next) {
fs/eventpoll.c
1593
epi->next = EP_UNACTIVE_PTR;
fs/eventpoll.c
2158
tfile_check_list = head->next;
fs/eventpoll.c
277
struct epitems_head *next;
fs/eventpoll.c
294
if (!head->next) {
fs/eventpoll.c
295
head->next = tfile_check_list;
fs/eventpoll.c
309
head->next = NULL;
fs/eventpoll.c
690
*p = pwq->next;
fs/eventpoll.c
762
nepi = epi->next, epi->next = EP_UNACTIVE_PTR) {
fs/eventpoll.c
855
if (head->first == &epi->fllink && !epi->fllink.next) {
fs/eventpoll.c
861
if (!smp_load_acquire(&v->next))
fs/eventpoll.c
901
struct rb_node *rbp, *next;
fs/eventpoll.c
928
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = next) {
fs/eventpoll.c
929
next = rb_next(rbp);
fs/exfat/cache.c
224
cache = list_entry(ei->cache_lru.next,
fs/exfat/cache.c
79
if (ei->cache_lru.next != &cache->cache_list)
fs/exfat/fatent.c
301
unsigned int clu, next;
fs/exfat/fatent.c
304
next = p_chain->dir;
fs/exfat/fatent.c
306
*ret_clu = next + p_chain->size - 1;
fs/exfat/fatent.c
312
clu = next;
fs/exfat/fatent.c
313
if (exfat_ent_get(sb, clu, &next, &bh))
fs/exfat/fatent.c
315
} while (next != EXFAT_EOF_CLUSTER && count <= p_chain->size);
fs/ext2/balloc.c
1016
struct rb_node *next;
fs/ext2/balloc.c
1022
next = rb_next(&my_rsv->rsv_node);
fs/ext2/balloc.c
1024
if (!next)
fs/ext2/balloc.c
1027
next_rsv = rb_entry(next, struct ext2_reserve_window_node, rsv_node);
fs/ext2/balloc.c
586
ext2_grpblk_t next;
fs/ext2/balloc.c
588
next = ext2_find_next_zero_bit(bh->b_data, maxblocks, start);
fs/ext2/balloc.c
589
if (next >= maxblocks)
fs/ext2/balloc.c
591
return next;
fs/ext2/balloc.c
609
ext2_grpblk_t here, next;
fs/ext2/balloc.c
636
next = (r - ((char *)bh->b_data)) << 3;
fs/ext2/balloc.c
638
if (next < maxblocks && next >= here)
fs/ext2/balloc.c
639
return next;
fs/ext2/balloc.c
762
struct rb_node *next;
fs/ext2/balloc.c
791
next = rb_next(&rsv->rsv_node);
fs/ext2/balloc.c
792
rsv = rb_entry(next,struct ext2_reserve_window_node,rsv_node);
fs/ext2/balloc.c
798
if (!next)
fs/ext2/xattr.c
152
struct ext2_xattr_entry *next;
fs/ext2/xattr.c
155
next = EXT2_XATTR_NEXT(entry);
fs/ext2/xattr.c
156
if ((char *)next >= end)
fs/ext4/dir.c
421
struct fname *next;
fs/ext4/dir.c
434
struct fname *fname, *next;
fs/ext4/dir.c
436
rbtree_postorder_for_each_entry_safe(fname, next, root, rb_hash)
fs/ext4/dir.c
439
fname = fname->next;
fs/ext4/dir.c
503
new_fn->next = fname->next;
fs/ext4/dir.c
504
fname->next = new_fn;
fs/ext4/dir.c
552
fname = fname->next;
fs/ext4/extents.c
2000
ext4_lblk_t next;
fs/ext4/extents.c
2103
next = EXT_MAX_BLOCKS;
fs/ext4/extents.c
2105
next = ext4_ext_next_leaf_block(path);
fs/ext4/extents.c
2106
if (next != EXT_MAX_BLOCKS) {
fs/ext4/extents.c
2109
ext_debug(inode, "next leaf block - %u\n", next);
fs/ext4/extents.c
2110
npath = ext4_find_extent(inode, next, NULL, gb_flags);
fs/ext4/extents.c
2225
ext4_lblk_t next, end = block + num - 1;
fs/ext4/extents.c
2232
next = 0;
fs/ext4/extents.c
2234
if (!ext4_es_lookup_extent(inode, block, &next, &es, NULL))
fs/ext4/extents.c
2243
if (next == 0)
fs/ext4/extents.c
2255
if (next == 0)
fs/ext4/extents.c
2257
block = next;
fs/ext4/extents.c
2297
ext4_lblk_t next;
fs/ext4/extents.c
2300
next = ext4_ext_next_allocated_block(path);
fs/ext4/extents.c
2301
BUG_ON(next == *lblk);
fs/ext4/extents.c
2302
len = next - *lblk;
fs/ext4/extents.c
4172
ext4_lblk_t next = ext4_ext_next_allocated_block(path);
fs/ext4/extents.c
4173
map->m_len = min(map->m_len, next - map->m_lblk);
fs/ext4/extents_status.c
1922
goto next;
fs/ext4/extents_status.c
1925
goto next;
fs/ext4/extents_status.c
1931
next:
fs/ext4/fast_commit.c
1829
goto next;
fs/ext4/fast_commit.c
1853
goto next;
fs/ext4/fast_commit.c
1869
next:
fs/ext4/fsync.c
48
struct dentry *dentry, *next;
fs/ext4/fsync.c
59
next = dget_parent(dentry);
fs/ext4/fsync.c
61
dentry = next;
fs/ext4/ialloc.c
731
next:
fs/ext4/ialloc.c
742
goto next;
fs/ext4/inline.c
70
void *next = EXT4_XATTR_NEXT(entry);
fs/ext4/inline.c
72
if (next >= end) {
fs/ext4/inline.c
82
entry = next;
fs/ext4/inode.c
1119
struct buffer_head *next;
fs/ext4/inode.c
1123
block_start = block_end, bh = next) {
fs/ext4/inode.c
1124
next = bh->b_this_page;
fs/ext4/mballoc.c
2091
int max, order, next;
fs/ext4/mballoc.c
2122
next = (block + 1) * (1 << order);
fs/ext4/mballoc.c
2123
if (mb_test_bit(next, e4b->bd_bitmap))
fs/ext4/mballoc.c
2126
order = mb_find_order_for_block(e4b, next);
fs/ext4/mballoc.c
2128
block = next >> order;
fs/ext4/mballoc.c
3202
.next = ext4_mb_seq_groups_next,
fs/ext4/mballoc.c
3376
.next = ext4_mb_seq_structs_summary_next,
fs/ext4/mballoc.c
5429
unsigned int next;
fs/ext4/mballoc.c
5445
next = mb_find_next_bit(bitmap_bh->b_data, end, bit);
fs/ext4/mballoc.c
5448
(unsigned) next - bit, (unsigned) group);
fs/ext4/mballoc.c
5449
free += next - bit;
fs/ext4/mballoc.c
5451
trace_ext4_mballoc_discard(sb, NULL, group, bit, next - bit);
fs/ext4/mballoc.c
5454
next - bit);
fs/ext4/mballoc.c
5455
mb_free_blocks(pa->pa_inode, e4b, bit, next - bit);
fs/ext4/mballoc.c
5456
bit = next + 1;
fs/ext4/mballoc.c
6426
struct ext4_free_data *next;
fs/ext4/mballoc.c
6433
next = rb_entry(node, struct ext4_free_data, efd_node);
fs/ext4/mballoc.c
6434
if (ext4_freed_extents_can_be_merged(entry, next))
fs/ext4/mballoc.c
6435
ext4_merge_freed_extents(sbi, root, entry, next);
fs/ext4/mballoc.c
6930
ext4_grpblk_t next, count, free_count, last, origin_start;
fs/ext4/mballoc.c
6951
next = mb_find_next_bit(bitmap, last + 1, start);
fs/ext4/mballoc.c
6952
if (origin_start == 0 && next >= last)
fs/ext4/mballoc.c
6955
if ((next - start) >= minblocks) {
fs/ext4/mballoc.c
6956
int ret = ext4_trim_extent(sb, start, next - start, e4b);
fs/ext4/mballoc.c
6960
count += next - start;
fs/ext4/mballoc.c
6962
free_count += next - start;
fs/ext4/mballoc.c
6963
start = next + 1;
fs/ext4/mballoc.c
7145
ext4_grpblk_t start, next;
fs/ext4/mballoc.c
7173
next = mb_find_next_bit(bitmap, end + 1, start);
fs/ext4/mballoc.c
7176
error = formatter(sb, group, start, next - start, priv);
fs/ext4/mballoc.c
7181
start = next + 1;
fs/ext4/move_extent.c
623
goto next;
fs/ext4/move_extent.c
643
next:
fs/ext4/namei.c
1610
goto next;
fs/ext4/namei.c
1645
next:
fs/ext4/namei.c
1879
struct ext4_dir_entry_2 *next, *to, *prev, *de = (struct ext4_dir_entry_2 *) base;
fs/ext4/namei.c
1884
next = ext4_next_entry(de, blocksize);
fs/ext4/namei.c
1893
de = next;
fs/ext4/page-io.c
185
io_end, inode->i_ino, io_end->list.next, io_end->list.prev);
fs/ext4/page-io.c
229
after = cur->next;
fs/ext4/page-io.c
287
io_end = list_entry(unwritten.next, ext4_io_end_t, list);
fs/ext4/super.c
3866
n = pos->next;
fs/ext4/xattr.c
1633
struct ext4_xattr_entry *last, *next;
fs/ext4/xattr.c
1667
for (; !IS_LAST_ENTRY(last); last = next) {
fs/ext4/xattr.c
1668
next = EXT4_XATTR_NEXT(last);
fs/ext4/xattr.c
1669
if ((void *)next >= s->end) {
fs/ext4/xattr.c
228
struct ext4_xattr_entry *next = EXT4_XATTR_NEXT(e);
fs/ext4/xattr.c
229
if ((void *)next >= end) {
fs/ext4/xattr.c
237
e = next;
fs/ext4/xattr.c
327
struct ext4_xattr_entry *entry, *next;
fs/ext4/xattr.c
334
for (entry = *pentry; !IS_LAST_ENTRY(entry); entry = next) {
fs/ext4/xattr.c
335
next = EXT4_XATTR_NEXT(entry);
fs/ext4/xattr.c
336
if ((void *) next >= end) {
fs/f2fs/checkpoint.c
2021
struct ckpt_req *req, *next;
fs/f2fs/checkpoint.c
2034
llist_for_each_entry_safe(req, next, dispatch_list, llnode) {
fs/f2fs/data.c
1015
next:
fs/f2fs/data.c
1091
goto next;
fs/f2fs/data.c
2039
next:
fs/f2fs/data.c
2073
goto next;
fs/f2fs/data.c
2136
goto next;
fs/f2fs/data.c
3435
goto next;
fs/f2fs/data.c
3443
goto next;
fs/f2fs/data.c
3455
next:
fs/f2fs/dir.c
625
next:
fs/f2fs/dir.c
638
goto next;
fs/f2fs/extent_cache.c
384
struct rb_node *node, *next;
fs/f2fs/extent_cache.c
391
next = rb_next(node);
fs/f2fs/extent_cache.c
394
node = next;
fs/f2fs/extent_cache.c
972
struct extent_tree *et, *next;
fs/f2fs/extent_cache.c
984
list_for_each_entry_safe(et, next, &eti->zombie_list, list) {
fs/f2fs/file.c
1355
goto next;
fs/f2fs/file.c
1383
next:
fs/f2fs/file.c
3823
goto next;
fs/f2fs/file.c
3840
next:
fs/f2fs/file.c
3996
goto next;
fs/f2fs/file.c
4021
goto next;
fs/f2fs/file.c
4037
next:
fs/f2fs/file.c
679
goto next;
fs/f2fs/file.c
685
goto next;
fs/f2fs/file.c
688
goto next;
fs/f2fs/file.c
706
next:
fs/f2fs/gc.c
115
goto next;
fs/f2fs/gc.c
122
goto next;
fs/f2fs/gc.c
131
goto next;
fs/f2fs/gc.c
178
next:
fs/f2fs/gc.c
433
struct rb_node *cur = rb_first_cached(root), *next;
fs/f2fs/gc.c
437
next = rb_next(cur);
fs/f2fs/gc.c
438
if (!next)
fs/f2fs/gc.c
442
next_ve = rb_entry(next, struct victim_entry, rb_node);
fs/f2fs/gc.c
450
cur = next;
fs/f2fs/gc.c
584
next:
fs/f2fs/gc.c
617
goto next;
fs/f2fs/gc.c
886
goto next;
fs/f2fs/gc.c
892
goto next;
fs/f2fs/gc.c
902
goto next;
fs/f2fs/gc.c
910
goto next;
fs/f2fs/gc.c
915
goto next;
fs/f2fs/gc.c
918
goto next;
fs/f2fs/gc.c
922
goto next;
fs/f2fs/gc.c
931
next:
fs/f2fs/node.c
2785
struct free_nid *i, *next;
fs/f2fs/node.c
2789
list_for_each_entry_safe(i, next, &nm_i->free_nid_list, list) {
fs/f2fs/recovery.c
433
goto next;
fs/f2fs/recovery.c
461
goto next;
fs/f2fs/recovery.c
471
next:
fs/f2fs/recovery.c
820
goto next;
fs/f2fs/recovery.c
852
next:
fs/f2fs/segment.c
1002
struct rb_node *cur = rb_first_cached(&dcc->root), *next;
fs/f2fs/segment.c
1006
next = rb_next(cur);
fs/f2fs/segment.c
1007
if (!next)
fs/f2fs/segment.c
1011
next_dc = rb_entry(next, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1020
cur = next;
fs/f2fs/segment.c
1510
goto next;
fs/f2fs/segment.c
1541
next:
fs/f2fs/segment.c
1607
goto next;
fs/f2fs/segment.c
1619
next:
fs/f2fs/segment.c
1671
goto next;
fs/f2fs/segment.c
1694
next:
fs/f2fs/segment.c
1769
next:
fs/f2fs/segment.c
1794
goto next;
fs/f2fs/segment.c
2221
next:
fs/f2fs/segment.c
2231
goto next;
fs/f2fs/segment.c
334
goto next;
fs/f2fs/segment.c
3424
next:
fs/f2fs/segment.c
3462
goto next;
fs/f2fs/segment.c
368
next:
fs/f2fs/segment.c
4610
struct sit_entry_set *next = ses;
fs/f2fs/segment.c
4615
list_for_each_entry_continue(next, head, set_list)
fs/f2fs/segment.c
4616
if (ses->entry_cnt <= next->entry_cnt) {
fs/f2fs/segment.c
4617
list_move_tail(&ses->set_list, &next->set_list);
fs/f2fs/segment.c
588
struct flush_cmd *cmd, *next;
fs/f2fs/segment.c
599
llist_for_each_entry_safe(cmd, next,
fs/f2fs/segment.c
664
struct flush_cmd *tmp, *next;
fs/f2fs/segment.c
668
llist_for_each_entry_safe(tmp, next, list, llnode) {
fs/f2fs/segment.h
471
unsigned int next;
fs/f2fs/segment.h
477
next = find_next_bit(free_i->free_segmap,
fs/f2fs/segment.h
479
if (next >= start_segno + f2fs_usable_segs_in_sec(sbi)) {
fs/f2fs/segment.h
504
unsigned int next;
fs/f2fs/segment.h
518
next = find_next_bit(free_i->free_segmap,
fs/f2fs/segment.h
520
if (next < start_segno + f2fs_usable_segs_in_sec(sbi))
fs/f2fs/shrinker.c
101
p = p->next;
fs/f2fs/shrinker.c
123
p = p->next;
fs/f2fs/shrinker.c
140
p = f2fs_list.next;
fs/f2fs/shrinker.c
146
p = p->next;
fs/f2fs/shrinker.c
154
p = p->next;
fs/f2fs/shrinker.c
210
p = f2fs_list.next;
fs/f2fs/shrinker.c
216
p = p->next;
fs/f2fs/shrinker.c
224
p = p->next;
fs/f2fs/shrinker.c
48
p = f2fs_list.next;
fs/f2fs/shrinker.c
54
p = p->next;
fs/f2fs/shrinker.c
72
p = p->next;
fs/f2fs/shrinker.c
92
p = f2fs_list.next;
fs/f2fs/sysfs.c
1938
goto next;
fs/f2fs/sysfs.c
1947
next:
fs/f2fs/xattr.c
738
struct f2fs_xattr_entry *next = XATTR_NEXT_ENTRY(here);
fs/f2fs/xattr.c
741
memmove(here, next, (char *)last - (char *)next);
fs/fat/cache.c
187
cache = list_entry(i->cache_lru.next,
fs/fat/cache.c
71
if (MSDOS_I(inode)->cache_lru.next != &cache->cache_list)
fs/fat/dir.c
108
goto next;
fs/fat/dir.c
92
next:
fs/fat/fatent.c
123
int next;
fs/fat/fatent.c
127
next = (*ent12_p[0] >> 4) | (*ent12_p[1] << 4);
fs/fat/fatent.c
129
next = (*ent12_p[1] << 8) | *ent12_p[0];
fs/fat/fatent.c
132
next &= 0x0fff;
fs/fat/fatent.c
133
if (next >= BAD_FAT12)
fs/fat/fatent.c
134
next = FAT_ENT_EOF;
fs/fat/fatent.c
135
return next;
fs/fat/fatent.c
140
int next = le16_to_cpu(*fatent->u.ent16_p);
fs/fat/fatent.c
142
if (next >= BAD_FAT16)
fs/fat/fatent.c
143
next = FAT_ENT_EOF;
fs/fat/fatent.c
144
return next;
fs/fat/fatent.c
149
int next = le32_to_cpu(*fatent->u.ent32_p) & 0x0fffffff;
fs/fat/fatent.c
151
if (next >= BAD_FAT32)
fs/fat/fatent.c
152
next = FAT_ENT_EOF;
fs/fat/fatent.c
153
return next;
fs/fcntl.c
1139
goto next;
fs/fcntl.c
1146
next:
fs/filesystems.c
116
*tmp = fs->next;
fs/filesystems.c
117
fs->next = NULL;
fs/filesystems.c
122
tmp = &(*tmp)->next;
fs/filesystems.c
143
for (tmp=file_systems, index=0 ; tmp ; tmp=tmp->next, index++) {
fs/filesystems.c
159
for (tmp = file_systems; tmp; tmp = tmp->next, index--) {
fs/filesystems.c
183
for (tmp = file_systems, index = 0 ; tmp ; tmp = tmp->next, index++)
fs/filesystems.c
220
for (p = file_systems; p; p = p->next) {
fs/filesystems.c
248
tmp = tmp->next;
fs/filesystems.c
52
for (p = &file_systems; *p; p = &(*p)->next)
fs/filesystems.c
82
if (fs->next)
fs/fs-writeback.c
1444
tail = wb_inode(wb->b_dirty.next);
fs/fs-writeback.c
2291
work = list_entry(wb->work_list.next,
fs/fuse/dev.c
1231
return fiq->forget_list_head.next != NULL;
fs/fuse/dev.c
1282
struct fuse_forget_link *head = fiq->forget_list_head.next;
fs/fuse/dev.c
1287
newhead = &(*newhead)->next;
fs/fuse/dev.c
1289
fiq->forget_list_head.next = *newhead;
fs/fuse/dev.c
1291
if (fiq->forget_list_head.next == NULL)
fs/fuse/dev.c
1370
head = forget->next;
fs/fuse/dev.c
1387
if (fc->minor < 16 || fiq->forget_list_head.next->next == NULL)
fs/fuse/dev.c
1453
req = list_entry(fiq->interrupts.next, struct fuse_req,
fs/fuse/dev.c
1466
req = list_entry(fiq->pending.next, struct fuse_req, list);
fs/fuse/dev.c
1994
struct fuse_req *req, *next;
fs/fuse/dev.c
2015
list_for_each_entry_safe(req, next, &to_queue, list) {
fs/fuse/dev.c
2410
req = list_entry(head->next, struct fuse_req, list);
fs/fuse/dev.c
2458
struct fuse_req *req, *next;
fs/fuse/dev.c
2476
list_for_each_entry_safe(req, next, &fpq->io, list) {
fs/fuse/dev.c
343
fiq->forget_list_tail->next = forget;
fs/fuse/dev_uring.c
198
struct fuse_ring_ent *ent, *next;
fs/fuse/dev_uring.c
208
list_for_each_entry_safe(ent, next, &queue->ent_released,
fs/fuse/dev_uring.c
367
struct fuse_ring_ent *ent, *next;
fs/fuse/dev_uring.c
372
list_for_each_entry_safe(ent, next, head, list) {
fs/fuse/dev_uring.c
385
list_for_each_entry_safe(ent, next, &to_teardown, list) {
fs/fuse/file.c
1976
wpa = list_entry(fi->queued_writes.next,
fs/fuse/fuse_i.h
90
struct fuse_forget_link *next;
fs/fuse/virtio_fs.c
817
struct fuse_req *next;
fs/fuse/virtio_fs.c
835
list_for_each_entry_safe(req, next, &reqs, list) {
fs/gfs2/dir.c
1004
struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new;
fs/gfs2/dir.c
1082
next = dent;
fs/gfs2/dir.c
1083
if (dirent_next(dip, obh, &next))
fs/gfs2/dir.c
1084
next = NULL;
fs/gfs2/dir.c
1114
dent = next;
fs/gfs2/file.c
1264
loff_t next = (offset + len - 1) >> sdp->sd_sb.sb_bsize_shift;
fs/gfs2/file.c
1267
next = (next + 1) << sdp->sd_sb.sb_bsize_shift;
fs/gfs2/file.c
1271
len = next - offset;
fs/gfs2/glock.c
1119
gl->gl_node.next = NULL;
fs/gfs2/glock.c
1962
struct gfs2_glock *gl, *next;
fs/gfs2/glock.c
1967
list_for_each_entry_safe(gl, next, &lru_list, gl_lru) {
fs/gfs2/glock.c
2569
.next = gfs2_glock_seq_next,
fs/gfs2/glock.c
2576
.next = gfs2_glock_seq_next,
fs/gfs2/glock.c
2583
.next = gfs2_sbstats_seq_next,
fs/gfs2/glock.c
2786
.next = gfs2_glockfd_seq_next,
fs/gfs2/lops.c
174
struct buffer_head *bh, *next;
fs/gfs2/lops.c
183
next = bh->b_this_page;
fs/gfs2/lops.c
186
bh = next;
fs/gfs2/meta_io.c
213
struct buffer_head *next = bh->b_this_page;
fs/gfs2/meta_io.c
216
bh = next;
fs/gfs2/meta_io.c
76
struct buffer_head *next = bh->b_this_page;
fs/gfs2/meta_io.c
81
bh = next;
fs/gfs2/rgrp.c
1088
struct gfs2_rgrpd *next = gfs2_rgrpd_get_next(rgd);
fs/gfs2/rgrp.c
1095
if (next == NULL)
fs/gfs2/rgrp.c
1097
else if (next->rd_addr > rgd->rd_addr)
fs/gfs2/rgrp.c
1098
str->rg_skip = cpu_to_be32(next->rd_addr - rgd->rd_addr);
fs/gfs2/rgrp.c
526
struct rb_node *n, *next;
fs/gfs2/rgrp.c
533
next = NULL;
fs/gfs2/rgrp.c
535
next = n->rb_left;
fs/gfs2/rgrp.c
537
next = n->rb_right;
fs/gfs2/rgrp.c
538
if (next == NULL) {
fs/gfs2/rgrp.c
548
n = next;
fs/hfs/bfind.c
216
idx = bnode->next;
fs/hfs/bnode.c
249
be32_to_cpu(desc.next), be32_to_cpu(desc.prev),
fs/hfs/bnode.c
288
tmp->next = node->next;
fs/hfs/bnode.c
289
cnid = cpu_to_be32(tmp->next);
fs/hfs/bnode.c
290
hfs_bnode_write(tmp, &cnid, offsetof(struct hfs_bnode_desc, next), 4);
fs/hfs/bnode.c
293
tree->leaf_head = node->next;
fs/hfs/bnode.c
295
if (node->next) {
fs/hfs/bnode.c
296
tmp = hfs_bnode_find(tree, node->next);
fs/hfs/bnode.c
307
if (!node->prev && !node->next) {
fs/hfs/bnode.c
444
node->next = be32_to_cpu(desc->next);
fs/hfs/brec.c
265
node->this, new_node->this, node->next);
fs/hfs/brec.c
266
new_node->next = node->next;
fs/hfs/brec.c
272
if (node->next)
fs/hfs/brec.c
273
next_node = hfs_bnode_find(tree, node->next);
fs/hfs/brec.c
336
node_desc.next = cpu_to_be32(new_node->next);
fs/hfs/brec.c
345
node->next = new_node->this;
fs/hfs/brec.c
347
node_desc.next = cpu_to_be32(node->next);
fs/hfs/brec.c
506
new_node->next = 0;
fs/hfs/brec.c
510
node_desc.next = cpu_to_be32(new_node->next);
fs/hfs/btree.c
240
prev->next = idx;
fs/hfs/btree.c
242
hfs_bnode_write(prev, &cnid, offsetof(struct hfs_bnode_desc, next), 4);
fs/hfs/btree.c
247
desc.next = 0;
fs/hfs/btree.c
339
nidx = node->next;
fs/hfs/btree.c
378
i = node->next;
fs/hfs/btree.h
58
u32 next;
fs/hfsplus/bfind.c
259
idx = bnode->next;
fs/hfsplus/bnode.c
357
be32_to_cpu(desc.next), be32_to_cpu(desc.prev),
fs/hfsplus/bnode.c
396
tmp->next = node->next;
fs/hfsplus/bnode.c
397
cnid = cpu_to_be32(tmp->next);
fs/hfsplus/bnode.c
399
offsetof(struct hfs_bnode_desc, next), 4);
fs/hfsplus/bnode.c
402
tree->leaf_head = node->next;
fs/hfsplus/bnode.c
404
if (node->next) {
fs/hfsplus/bnode.c
405
tmp = hfs_bnode_find(tree, node->next);
fs/hfsplus/bnode.c
417
if (!node->prev && !node->next)
fs/hfsplus/bnode.c
555
node->next = be32_to_cpu(desc->next);
fs/hfsplus/brec.c
250
node->this, new_node->this, node->next);
fs/hfsplus/brec.c
251
new_node->next = node->next;
fs/hfsplus/brec.c
257
if (node->next)
fs/hfsplus/brec.c
258
next_node = hfs_bnode_find(tree, node->next);
fs/hfsplus/brec.c
321
node_desc.next = cpu_to_be32(new_node->next);
fs/hfsplus/brec.c
330
node->next = new_node->this;
fs/hfsplus/brec.c
332
node_desc.next = cpu_to_be32(node->next);
fs/hfsplus/brec.c
491
new_node->next = 0;
fs/hfsplus/brec.c
495
node_desc.next = cpu_to_be32(new_node->next);
fs/hfsplus/btree.c
325
prev->next = idx;
fs/hfsplus/btree.c
327
hfs_bnode_write(prev, &cnid, offsetof(struct hfs_bnode_desc, next), 4);
fs/hfsplus/btree.c
332
desc.next = 0;
fs/hfsplus/btree.c
435
nidx = node->next;
fs/hfsplus/btree.c
475
i = node->next;
fs/hfsplus/dir.c
220
goto next;
fs/hfsplus/dir.c
256
next:
fs/hfsplus/hfsplus_fs.h
75
u32 next;
fs/hfsplus/xattr.c
98
desc->next = cpu_to_be32(be32_to_cpu(head->leaf_tail) + 1);
fs/hostfs/hostfs_kern.c
274
unsigned long long next, ino;
fs/hostfs/hostfs_kern.c
285
next = ctx->pos;
fs/hostfs/hostfs_kern.c
286
seek_dir(dir, next);
fs/hostfs/hostfs_kern.c
287
while ((name = read_dir(dir, &next, &ino, &len, &type)) != NULL) {
fs/hostfs/hostfs_kern.c
290
ctx->pos = next;
fs/hugetlbfs/inode.c
584
pgoff_t next, index;
fs/hugetlbfs/inode.c
589
next = lstart >> PAGE_SHIFT;
fs/hugetlbfs/inode.c
590
while (filemap_get_folios(mapping, &next, end - 1, &fbatch)) {
fs/inode.c
1153
int next = atomic_add_return(LAST_INO_BATCH, &shared_last_ino);
fs/inode.c
1155
res = next - LAST_INO_BATCH;
fs/iomap/buffered-io.c
307
unsigned int next, blocks_skipped;
fs/iomap/buffered-io.c
309
next = ifs_next_nonuptodate_block(folio, first, last);
fs/iomap/buffered-io.c
310
blocks_skipped = next - first;
fs/iomap/buffered-io.c
321
first = next;
fs/iomap/buffered-io.c
324
if (++next <= last) {
fs/iomap/buffered-io.c
325
next = ifs_next_uptodate_block(folio, next, last);
fs/iomap/buffered-io.c
326
if (next <= last) {
fs/iomap/buffered-io.c
328
block_bits, last - next + 1);
fs/iomap/buffered-io.c
329
last = next - 1;
fs/iomap/ioend.c
353
struct iomap_ioend *next)
fs/iomap/ioend.c
360
bio_op(&next->io_bio) == REQ_OP_READ)
fs/iomap/ioend.c
363
if (ioend->io_bio.bi_status != next->io_bio.bi_status)
fs/iomap/ioend.c
365
if (next->io_flags & IOMAP_IOEND_BOUNDARY)
fs/iomap/ioend.c
368
(next->io_flags & IOMAP_IOEND_NOMERGE_FLAGS))
fs/iomap/ioend.c
370
if (ioend->io_offset + ioend->io_size != next->io_offset)
fs/iomap/ioend.c
383
next->io_sector)
fs/iomap/ioend.c
391
struct iomap_ioend *next;
fs/iomap/ioend.c
395
while ((next = list_first_entry_or_null(more_ioends, struct iomap_ioend,
fs/iomap/ioend.c
397
if (!iomap_ioend_can_merge(ioend, next))
fs/iomap/ioend.c
399
list_move_tail(&next->io_list, &ioend->io_list);
fs/iomap/ioend.c
400
ioend->io_size += next->io_size;
fs/jbd2/journal.c
1167
.next = jbd2_seq_info_next,
fs/jbd2/recovery.c
70
unsigned int max, nbufs, next;
fs/jbd2/recovery.c
86
for (next = start; next < max; next++) {
fs/jbd2/recovery.c
87
int err = jbd2_journal_bmap(journal, next, &blocknr);
fs/jbd2/recovery.c
91
next);
fs/jbd2/revoke.c
169
record = (struct jbd2_revoke_record_s *) hash_list->next;
fs/jbd2/revoke.c
175
record = (struct jbd2_revoke_record_s *) record->hash.next;
fs/jbd2/revoke.c
553
hash_list->next;
fs/jbd2/revoke.c
738
record = (struct jbd2_revoke_record_s*) hash_list->next;
fs/jbd2/transaction.c
2432
struct buffer_head *head, *bh, *next;
fs/jbd2/transaction.c
2454
next = bh->b_this_page;
fs/jbd2/transaction.c
2469
bh = next;
fs/jffs2/build.c
151
dead_fds = fd->next;
fs/jffs2/build.c
181
ic->scan_dents = fd->next;
fs/jffs2/build.c
232
ic->scan_dents = fd->next;
fs/jffs2/build.c
253
struct jffs2_raw_node_ref *next = raw->next_in_ino;
fs/jffs2/build.c
256
raw = next;
fs/jffs2/build.c
267
ic->scan_dents = fd->next;
fs/jffs2/build.c
295
fd->next = *dead_fds;
fs/jffs2/build.c
40
if (ic->next)
fs/jffs2/build.c
41
return ic->next;
fs/jffs2/build.c
61
for(fd = ic->scan_dents; fd; fd = fd->next) {
fs/jffs2/dir.c
102
for (fd_list = dir_f->dents; fd_list && fd_list->nhash <= nhash; fd_list = fd_list->next) {
fs/jffs2/dir.c
138
for (fd = f->dents; fd; fd = fd->next) {
fs/jffs2/dir.c
606
for (fd = f->dents ; fd; fd = fd->next) {
fs/jffs2/dir.c
799
for (fd = victim_f->dents; fd; fd = fd->next) {
fs/jffs2/erase.c
112
jeb = list_entry(c->erase_complete_list.next, struct jffs2_eraseblock, list);
fs/jffs2/erase.c
125
jeb = list_entry(c->erase_pending_list.next, struct jffs2_eraseblock, list);
fs/jffs2/fs.c
305
for (fd=f->dents; fd; fd = fd->next) {
fs/jffs2/gc.c
1002
fdp = &(*fdp)->next;
fs/jffs2/gc.c
159
for (ic = c->inocache_list[bucket]; ic; ic = ic->next) {
fs/jffs2/gc.c
566
for (fd = f->dents; fd; fd=fd->next) {
fs/jffs2/gc.c
97
ret = list_entry(nextlist->next, struct jffs2_eraseblock, list);
fs/jffs2/gc.c
999
*fdp = fd->next;
fs/jffs2/nodelist.c
395
struct jffs2_node_frag *next = frag_next(newfrag);
fs/jffs2/nodelist.c
397
if (next) {
fs/jffs2/nodelist.c
399
if (next->node)
fs/jffs2/nodelist.c
400
mark_ref_normal(next->node->raw);
fs/jffs2/nodelist.c
427
ret = ret->next;
fs/jffs2/nodelist.c
43
new->next = (*prev)->next;
fs/jffs2/nodelist.c
449
prev = &(*prev)->next;
fs/jffs2/nodelist.c
451
new->next = *prev;
fs/jffs2/nodelist.c
470
prev = &(*prev)->next;
fs/jffs2/nodelist.c
473
*prev = old->next;
fs/jffs2/nodelist.c
490
struct jffs2_inode_cache *this, *next;
fs/jffs2/nodelist.c
495
next = this->next;
fs/jffs2/nodelist.c
498
this = next;
fs/jffs2/nodelist.c
507
struct jffs2_raw_node_ref *this, *next;
fs/jffs2/nodelist.c
513
next = this[REFS_PER_BLOCK].next_in_ino;
fs/jffs2/nodelist.c
515
next = NULL;
fs/jffs2/nodelist.c
518
this = next;
fs/jffs2/nodelist.c
528
struct rb_node *next;
fs/jffs2/nodelist.c
53
prev = &((*prev)->next);
fs/jffs2/nodelist.c
534
next = fragtree->rb_node;
fs/jffs2/nodelist.c
536
while(next) {
fs/jffs2/nodelist.c
537
frag = rb_entry(next, struct jffs2_node_frag, rb);
fs/jffs2/nodelist.c
543
next = frag->rb.rb_right;
fs/jffs2/nodelist.c
545
next = frag->rb.rb_left;
fs/jffs2/nodelist.c
55
new->next = *prev;
fs/jffs2/nodelist.c
567
struct jffs2_node_frag *frag, *next;
fs/jffs2/nodelist.c
570
rbtree_postorder_for_each_entry_safe(frag, next, root, rb) {
fs/jffs2/nodelist.c
73
struct jffs2_node_frag *next = frag_next(frag);
fs/jffs2/nodelist.c
77
frag = next;
fs/jffs2/nodelist.h
175
struct jffs2_inode_cache *next;
fs/jffs2/nodelist.h
257
struct jffs2_full_dirent *next;
fs/jffs2/nodemgmt.c
286
struct list_head *next;
fs/jffs2/nodemgmt.c
296
ejeb = list_entry(c->erasable_list.next, struct jffs2_eraseblock, list);
fs/jffs2/nodemgmt.c
338
next = c->free_list.next;
fs/jffs2/nodemgmt.c
339
list_del(next);
fs/jffs2/nodemgmt.c
340
c->nextblock = list_entry(next, struct jffs2_eraseblock, list);
fs/jffs2/readinode.c
1436
fds = fd->next;
fs/jffs2/readinode.c
177
struct rb_node *next;
fs/jffs2/readinode.c
182
next = tn_root->rb_node;
fs/jffs2/readinode.c
184
while (next) {
fs/jffs2/readinode.c
185
tn = rb_entry(next, struct jffs2_tmp_dnode_info, rb);
fs/jffs2/readinode.c
188
next = tn->rb.rb_right;
fs/jffs2/readinode.c
190
next = tn->rb.rb_left;
fs/jffs2/readinode.c
303
struct jffs2_tmp_dnode_info *next = tn_next(this);
fs/jffs2/readinode.c
311
this = next;
fs/jffs2/readinode.c
546
struct jffs2_tmp_dnode_info *tn, *next;
fs/jffs2/readinode.c
548
rbtree_postorder_for_each_entry_safe(tn, next, list, rb) {
fs/jffs2/readinode.c
558
struct jffs2_full_dirent *next;
fs/jffs2/readinode.c
561
next = fd->next;
fs/jffs2/readinode.c
563
fd = next;
fs/jffs2/readinode.c
694
fd->next = NULL;
fs/jffs2/scan.c
1108
fd->next = NULL;
fs/jffs2/scan.c
1137
struct list_head *n = head->next;
fs/jffs2/scan.c
1141
n = n->next;
fs/jffs2/scan.c
432
ref->next = c->xref_temp;
fs/jffs2/summary.c
128
temp->next = NULL;
fs/jffs2/summary.c
150
temp->next = NULL;
fs/jffs2/summary.c
171
temp->next = NULL;
fs/jffs2/summary.c
186
temp->next = NULL;
fs/jffs2/summary.c
202
s->sum_list_head = s->sum_list_head->u.next;
fs/jffs2/summary.c
276
temp->next = NULL;
fs/jffs2/summary.c
296
temp->next = NULL;
fs/jffs2/summary.c
326
temp->next = NULL;
fs/jffs2/summary.c
337
temp->next = NULL;
fs/jffs2/summary.c
476
fd->next = NULL;
fs/jffs2/summary.c
537
ref->next = c->xref_temp;
fs/jffs2/summary.c
68
s->sum_list_tail->u.next = (union jffs2_sum_mem *) item;
fs/jffs2/summary.c
794
c->summary->sum_list_head = temp->u.next;
fs/jffs2/summary.h
101
union jffs2_sum_mem *next;
fs/jffs2/summary.h
111
union jffs2_sum_mem *next;
fs/jffs2/summary.h
125
union jffs2_sum_mem *next;
fs/jffs2/summary.h
135
union jffs2_sum_mem *next;
fs/jffs2/summary.h
95
union jffs2_sum_mem *next;
fs/jffs2/wbuf.c
103
new->next = c->wbuf_inodes;
fs/jffs2/wbuf.c
110
struct list_head *this, *next;
fs/jffs2/wbuf.c
116
list_for_each_safe(this, next, &c->erasable_pending_wbuf_list) {
fs/jffs2/wbuf.c
213
for (fd = f->dents; fd; fd = fd->next) {
fs/jffs2/wbuf.c
278
struct jffs2_raw_node_ref *raw, *next, *first_raw = NULL;
fs/jffs2/wbuf.c
298
for (next = raw = jeb->first_node; next; raw = next) {
fs/jffs2/wbuf.c
299
next = ref_next(raw);
fs/jffs2/wbuf.c
302
(next && ref_offset(next) <= c->wbuf_ofs)) {
fs/jffs2/wbuf.c
43
struct jffs2_inodirty *next;
fs/jffs2/wbuf.c
64
this = this->next;
fs/jffs2/wbuf.c
77
struct jffs2_inodirty *next = this->next;
fs/jffs2/wbuf.c
79
this = next;
fs/jffs2/write.c
609
for (fd = dir_f->dents; fd; fd = fd->next) {
fs/jffs2/write.c
641
dead_f->dents = fd->next;
fs/jffs2/xattr.c
1050
for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
fs/jffs2/xattr.c
1066
*pref = ref->next;
fs/jffs2/xattr.c
1125
for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
fs/jffs2/xattr.c
1132
*pref = ref->next;
fs/jffs2/xattr.c
1149
*pref = ref->next;
fs/jffs2/xattr.c
1151
ref->next = c->xref_dead_list;
fs/jffs2/xattr.c
1195
*pref = ref->next;
fs/jffs2/xattr.c
1199
ref->next = ic->xref;
fs/jffs2/xattr.c
1346
for (tmp=c->xref_dead_list, ptmp=&c->xref_dead_list; tmp; ptmp=&tmp->next, tmp=tmp->next) {
fs/jffs2/xattr.c
1348
*ptmp = tmp->next;
fs/jffs2/xattr.c
570
ref->next = ic->xref;
fs/jffs2/xattr.c
586
ref->next = c->xref_dead_list;
fs/jffs2/xattr.c
607
_ref = ref->next;
fs/jffs2/xattr.c
622
_ref = ref->next;
fs/jffs2/xattr.c
648
for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
fs/jffs2/xattr.c
652
*pref = ref->next;
fs/jffs2/xattr.c
658
for (cmp=ref->next, pcmp=&ref->next; cmp; pcmp=&cmp->next, cmp=cmp->next) {
fs/jffs2/xattr.c
664
*pcmp = cmp->next;
fs/jffs2/xattr.c
673
*pcmp = cmp->next;
fs/jffs2/xattr.c
676
*pref = ref->next;
fs/jffs2/xattr.c
747
_ref = ref->next;
fs/jffs2/xattr.c
752
_ref = ref->next;
fs/jffs2/xattr.c
795
_ref = ref->next;
fs/jffs2/xattr.c
807
for (tmp=xref_tmphash[i]; tmp; tmp=tmp->next) {
fs/jffs2/xattr.c
824
ref->next = xref_tmphash[i];
fs/jffs2/xattr.c
834
_ref = ref->next;
fs/jffs2/xattr.c
836
ref->next = c->xref_dead_list;
fs/jffs2/xattr.c
849
ref->next = c->xref_dead_list;
fs/jffs2/xattr.c
857
ref->next = ic->xref;
fs/jffs2/xattr.c
984
for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
fs/jffs2/xattr.c
997
*pref = ref->next;
fs/jffs2/xattr.h
62
struct jffs2_xattr_ref *next; /* chained from ic->xref_list */
fs/jfs/jfs_dtree.c
1406
nextbn = le64_to_cpu(sp->header.next);
fs/jfs/jfs_dtree.c
1407
rp->header.next = cpu_to_le64(nextbn);
fs/jfs/jfs_dtree.c
1409
sp->header.next = cpu_to_le64(rbn);
fs/jfs/jfs_dtree.c
1454
f->next = fsi;
fs/jfs/jfs_dtree.c
1455
f->next = -1;
fs/jfs/jfs_dtree.c
1571
f->next = fsi;
fs/jfs/jfs_dtree.c
1572
f->next = -1;
fs/jfs/jfs_dtree.c
1783
f->next = last;
fs/jfs/jfs_dtree.c
1796
f->next = fsi;
fs/jfs/jfs_dtree.c
1797
f->next = -1;
fs/jfs/jfs_dtree.c
1806
fsi = f->next;
fs/jfs/jfs_dtree.c
1809
f->next = n;
fs/jfs/jfs_dtree.c
1935
rp->header.next = 0;
fs/jfs/jfs_dtree.c
1968
f->next = fsi;
fs/jfs/jfs_dtree.c
1969
f->next = -1;
fs/jfs/jfs_dtree.c
1980
fsi = f->next;
fs/jfs/jfs_dtree.c
1983
f->next = n;
fs/jfs/jfs_dtree.c
2044
s->next = -1;
fs/jfs/jfs_dtree.c
2057
f->next = fsi;
fs/jfs/jfs_dtree.c
2058
f->next = -1;
fs/jfs/jfs_dtree.c
2122
|| (p->header.next == 0))
fs/jfs/jfs_dtree.c
2126
DT_GETPAGE(ip, le64_to_cpu(p->header.next),
fs/jfs/jfs_dtree.c
2454
nextbn = le64_to_cpu(p->header.next);
fs/jfs/jfs_dtree.c
2511
p->header.next = cpu_to_le64(nextbn);
fs/jfs/jfs_dtree.c
2597
f->next = fsi;
fs/jfs/jfs_dtree.c
2598
f->next = -1;
fs/jfs/jfs_dtree.c
2731
int i, next;
fs/jfs/jfs_dtree.c
2972
next = d->next;
fs/jfs/jfs_dtree.c
2973
while (next >= 0) {
fs/jfs/jfs_dtree.c
2974
t = (struct dtslot *) & p->slot[next];
fs/jfs/jfs_dtree.c
2991
next = t->next;
fs/jfs/jfs_dtree.c
3006
bn = le64_to_cpu(p->header.next);
fs/jfs/jfs_dtree.c
3199
bn = le64_to_cpu(p->header.next);
fs/jfs/jfs_dtree.c
3232
bn = le64_to_cpu(p->header.next);
fs/jfs/jfs_dtree.c
3273
bn = le64_to_cpu(p->header.next);
fs/jfs/jfs_dtree.c
3345
si = ih->next;
fs/jfs/jfs_dtree.c
3372
si = t->next;
fs/jfs/jfs_dtree.c
3428
si = lh->next;
fs/jfs/jfs_dtree.c
3441
si = ih->next;
fs/jfs/jfs_dtree.c
3482
si = t->next;
fs/jfs/jfs_dtree.c
3583
si = lh->next;
fs/jfs/jfs_dtree.c
3592
si = ih->next;
fs/jfs/jfs_dtree.c
3617
si = t->next;
fs/jfs/jfs_dtree.c
3653
p->header.freelist = h->next;
fs/jfs/jfs_dtree.c
3666
lh->next = h->next;
fs/jfs/jfs_dtree.c
3681
ih->next = h->next;
fs/jfs/jfs_dtree.c
3701
p->header.freelist = t->next;
fs/jfs/jfs_dtree.c
3741
lh->next = -1;
fs/jfs/jfs_dtree.c
3743
ih->next = -1;
fs/jfs/jfs_dtree.c
3746
t->next = -1;
fs/jfs/jfs_dtree.c
3792
int ssi, next; /* src slot index */
fs/jfs/jfs_dtree.c
3868
next = slh->next;
fs/jfs/jfs_dtree.c
3872
dlh->next = dsi;
fs/jfs/jfs_dtree.c
3880
next = sih->next;
fs/jfs/jfs_dtree.c
3883
dih->next = dsi;
fs/jfs/jfs_dtree.c
3887
s->next = sfsi;
fs/jfs/jfs_dtree.c
3899
while ((ssi = next) >= 0) {
fs/jfs/jfs_dtree.c
3934
d->next = dsi;
fs/jfs/jfs_dtree.c
3937
next = s->next;
fs/jfs/jfs_dtree.c
3938
s->next = sfsi;
fs/jfs/jfs_dtree.c
3949
dlh->next = -1;
fs/jfs/jfs_dtree.c
3951
dih->next = -1;
fs/jfs/jfs_dtree.c
3954
d->next = -1;
fs/jfs/jfs_dtree.c
4013
si = ((struct ldtentry *) t)->next;
fs/jfs/jfs_dtree.c
4015
si = ((struct idtentry *) t)->next;
fs/jfs/jfs_dtree.c
4016
t->next = si;
fs/jfs/jfs_dtree.c
4048
si = t->next;
fs/jfs/jfs_dtree.c
4058
t->next = p->header.freelist;
fs/jfs/jfs_dtree.c
4109
si = ((struct idtentry *) t)->next;
fs/jfs/jfs_dtree.c
4110
((struct idtentry *) t)->next = -1;
fs/jfs/jfs_dtree.c
4143
si = t->next;
fs/jfs/jfs_dtree.c
4155
t->next = p->header.freelist;
fs/jfs/jfs_dtree.c
4189
si = t->next;
fs/jfs/jfs_dtree.c
4215
si = t->next;
fs/jfs/jfs_dtree.c
484
static void free_index(tid_t tid, struct inode *ip, u32 index, u32 next)
fs/jfs/jfs_dtree.c
497
dirtab_slot->addr2 = cpu_to_le32(next);
fs/jfs/jfs_dtree.h
182
__le64 next; /* 8: next sibling */
fs/jfs/jfs_dtree.h
35
s8 next; /* 1: */
fs/jfs/jfs_dtree.h
55
s8 next; /* 1: */
fs/jfs/jfs_dtree.h
74
s8 next; /* 1: */
fs/jfs/jfs_logmgr.c
1862
struct lbuf *next = lbuf->l_freelist;
fs/jfs/jfs_logmgr.c
1865
lbuf = next;
fs/jfs/jfs_logmgr.c
2369
sbi = list_entry(log->sb_list.next, struct jfs_sb_info, log_list);
fs/jfs/jfs_logmgr.c
455
if ((i = linelock->next)) {
fs/jfs/jfs_logmgr.c
752
gcpn = list_entry(log->cqueue.next, struct tblock, cqueue)->pn;
fs/jfs/jfs_logmgr.c
951
lp = list_entry(log->synclist.next,
fs/jfs/jfs_txnmgr.c
1024
if (tblk->next)
fs/jfs/jfs_txnmgr.c
1025
lid_to_tlock(tblk->last)->next = lid;
fs/jfs/jfs_txnmgr.c
1027
tblk->next = lid;
fs/jfs/jfs_txnmgr.c
1028
tlck->next = 0;
fs/jfs/jfs_txnmgr.c
1035
tlck->next = jfs_ip->atlhead;
fs/jfs/jfs_txnmgr.c
1037
if (tlck->next == 0) {
fs/jfs/jfs_txnmgr.c
1049
maplock->next = 0;
fs/jfs/jfs_txnmgr.c
1077
linelock->next = 0;
fs/jfs/jfs_txnmgr.c
1085
linelock->next = tlock->next;
fs/jfs/jfs_txnmgr.c
1086
tlock->next = lid;
fs/jfs/jfs_txnmgr.c
1242
lid_to_tlock(jfs_ip->atltail)->next = tblk->next;
fs/jfs/jfs_txnmgr.c
1243
tblk->next = jfs_ip->atlhead;
fs/jfs/jfs_txnmgr.c
1383
for (lid = tblk->next; lid; lid = tlck->next) {
fs/jfs/jfs_txnmgr.c
1667
int next, lwm, hwm;
fs/jfs/jfs_txnmgr.c
1686
next = le16_to_cpu(p->header.nextindex);
fs/jfs/jfs_txnmgr.c
1720
if (lwm == next)
fs/jfs/jfs_txnmgr.c
1722
if (lwm > next) {
fs/jfs/jfs_txnmgr.c
1728
xadlock->count = next - lwm;
fs/jfs/jfs_txnmgr.c
191
TxAnchor.freelock = TxLock[lid].next;
fs/jfs/jfs_txnmgr.c
1924
if (twm == next - 1) {
fs/jfs/jfs_txnmgr.c
1944
if (hwm >= next) {
fs/jfs/jfs_txnmgr.c
1955
cpu_to_le16(hwm - next + 1);
fs/jfs/jfs_txnmgr.c
1957
xtlck->header.offset = next;
fs/jfs/jfs_txnmgr.c
1958
xtlck->header.length = hwm - next + 1;
fs/jfs/jfs_txnmgr.c
1972
if (lwm < next) {
fs/jfs/jfs_txnmgr.c
1980
xadlock->count = next - lwm;
fs/jfs/jfs_txnmgr.c
1984
tlck->ip, mp, xadlock->count, lwm, next);
fs/jfs/jfs_txnmgr.c
1992
if (twm == next - 1) {
fs/jfs/jfs_txnmgr.c
2013
if (hwm >= next) {
fs/jfs/jfs_txnmgr.c
2020
xadlock->count = hwm - next + 1;
fs/jfs/jfs_txnmgr.c
2021
xadlock->xdlist = &p->xad[next];
fs/jfs/jfs_txnmgr.c
2024
tlck->ip, mp, xadlock->count, next, hwm);
fs/jfs/jfs_txnmgr.c
205
TxLock[lid].next = TxAnchor.freelock;
fs/jfs/jfs_txnmgr.c
2181
lid_t lid, next;
fs/jfs/jfs_txnmgr.c
2189
tlck = lid_to_tlock(tblk->next);
fs/jfs/jfs_txnmgr.c
2190
lid = tlck->next;
fs/jfs/jfs_txnmgr.c
2191
tlck->next = 0;
fs/jfs/jfs_txnmgr.c
2194
next = tlck->next;
fs/jfs/jfs_txnmgr.c
2195
tlck->next = tblk->next;
fs/jfs/jfs_txnmgr.c
2196
tblk->next = lid;
fs/jfs/jfs_txnmgr.c
2197
lid = next;
fs/jfs/jfs_txnmgr.c
2204
for (lid = tblk->next; lid; lid = next) {
fs/jfs/jfs_txnmgr.c
2206
next = tlck->next;
fs/jfs/jfs_txnmgr.c
2275
for (lid = tblk->next; lid; lid = tlck->next) {
fs/jfs/jfs_txnmgr.c
2560
while ((lid = xtlck->next) != 0) {
fs/jfs/jfs_txnmgr.c
2563
xtlck->next = tlck->next;
fs/jfs/jfs_txnmgr.c
2596
lid_t lid, next;
fs/jfs/jfs_txnmgr.c
2604
for (lid = tblk->next; lid; lid = next) {
fs/jfs/jfs_txnmgr.c
2606
next = tlck->next;
fs/jfs/jfs_txnmgr.c
2632
tblk->next = tblk->last = 0;
fs/jfs/jfs_txnmgr.c
281
TxBlock[k].next = k + 1;
fs/jfs/jfs_txnmgr.c
2824
jfs_ip = list_entry(TxAnchor.anon_list.next,
fs/jfs/jfs_txnmgr.c
283
TxBlock[k].next = 0;
fs/jfs/jfs_txnmgr.c
2896
jfs_ip = list_entry(TxAnchor.anon_list.next,
fs/jfs/jfs_txnmgr.c
305
TxLock[k].next = k + 1;
fs/jfs/jfs_txnmgr.c
306
TxLock[k].next = 0;
fs/jfs/jfs_txnmgr.c
404
if ((tblk->next == 0) && !(flag & COMMIT_FORCE)) {
fs/jfs/jfs_txnmgr.c
412
TxAnchor.freetid = tblk->next;
fs/jfs/jfs_txnmgr.c
424
tblk->next = tblk->last = tblk->xflag = tblk->flag = tblk->lsn = 0;
fs/jfs/jfs_txnmgr.c
529
assert(tblk->next == 0);
fs/jfs/jfs_txnmgr.c
534
tblk->next = TxAnchor.freetid;
fs/jfs/jfs_txnmgr.c
656
jfs_ip->atlhead = tlck->next;
fs/jfs/jfs_txnmgr.c
660
lid_to_tlock(last)->next != lid;
fs/jfs/jfs_txnmgr.c
661
last = lid_to_tlock(last)->next) {
fs/jfs/jfs_txnmgr.c
664
lid_to_tlock(last)->next = tlck->next;
fs/jfs/jfs_txnmgr.c
671
if (tblk->next)
fs/jfs/jfs_txnmgr.c
672
lid_to_tlock(tblk->last)->next = lid;
fs/jfs/jfs_txnmgr.c
674
tblk->next = lid;
fs/jfs/jfs_txnmgr.c
675
tlck->next = 0;
fs/jfs/jfs_txnmgr.c
739
if (tblk->next)
fs/jfs/jfs_txnmgr.c
740
lid_to_tlock(tblk->last)->next = lid;
fs/jfs/jfs_txnmgr.c
742
tblk->next = lid;
fs/jfs/jfs_txnmgr.c
743
tlck->next = 0;
fs/jfs/jfs_txnmgr.c
750
tlck->next = jfs_ip->atlhead;
fs/jfs/jfs_txnmgr.c
752
if (tlck->next == 0) {
fs/jfs/jfs_txnmgr.c
764
linelock->next = 0;
fs/jfs/jfs_txnmgr.c
877
for (lid = tblk->next; lid; lid = tlck->next) {
fs/jfs/jfs_txnmgr.c
905
lid_t lid, next, llid, k;
fs/jfs/jfs_txnmgr.c
917
for (lid = tblk->next; lid; lid = next) {
fs/jfs/jfs_txnmgr.c
919
next = tlck->next;
fs/jfs/jfs_txnmgr.c
956
llid = ((struct linelock *) & tlck->lock)->next;
fs/jfs/jfs_txnmgr.c
959
k = linelock->next;
fs/jfs/jfs_txnmgr.c
967
tblk->next = tblk->last = 0;
fs/jfs/jfs_txnmgr.h
158
lid_t next; /* 2: next linelock */
fs/jfs/jfs_txnmgr.h
174
lid_t next; /* 2: */
fs/jfs/jfs_txnmgr.h
205
lid_t next; /* 2: */
fs/jfs/jfs_txnmgr.h
233
lid_t next; /* 2: */
fs/jfs/jfs_txnmgr.h
35
lid_t next; /* index of first tlock of tid */
fs/jfs/jfs_txnmgr.h
83
lid_t next; /* 2: index next lockword on tid locklist
fs/jfs/jfs_xtree.c
1027
nextbn = le64_to_cpu(sp->header.next);
fs/jfs/jfs_xtree.c
1028
rp->header.next = cpu_to_le64(nextbn);
fs/jfs/jfs_xtree.c
1030
sp->header.next = cpu_to_le64(rbn);
fs/jfs/jfs_xtree.c
1270
rp->header.next = 0;
fs/jfs/jfs_xtree.c
149
s64 next, size, xoff, xend;
fs/jfs/jfs_xtree.c
168
if ((rc = xtSearch(ip, lstart, &next, &cmp, &btstack, 0))) {
fs/jfs/jfs_xtree.c
1802
bn = le64_to_cpu(p->header.next);
fs/jfs/jfs_xtree.c
187
if (next)
fs/jfs/jfs_xtree.c
188
*plen = min(next - lstart, llen);
fs/jfs/jfs_xtree.c
1976
s64 next;
fs/jfs/jfs_xtree.c
1991
if ((rc = xtSearch(ip, xoff, &next, &cmp, &btstack, XT_INSERT)))
fs/jfs/jfs_xtree.c
2002
if (next)
fs/jfs/jfs_xtree.c
2003
xlen = min(xlen, (int)(next - xoff));
fs/jfs/jfs_xtree.c
2297
if (p->header.next) {
fs/jfs/jfs_xtree.c
2306
p->header.next = 0;
fs/jfs/jfs_xtree.c
244
s64 next = 0;
fs/jfs/jfs_xtree.c
2592
if (tblk->next == lid)
fs/jfs/jfs_xtree.c
2593
tblk->next = tlck->next;
fs/jfs/jfs_xtree.c
2595
for (prev = lid_to_tlock(tblk->next);
fs/jfs/jfs_xtree.c
2596
prev->next != lid;
fs/jfs/jfs_xtree.c
2597
prev = lid_to_tlock(prev->next)) {
fs/jfs/jfs_xtree.c
2598
assert(prev->next);
fs/jfs/jfs_xtree.c
2600
prev->next = tlck->next;
fs/jfs/jfs_xtree.c
2602
lid_to_tlock(tblk->last)->next = lid;
fs/jfs/jfs_xtree.c
2603
tlck->next = 0;
fs/jfs/jfs_xtree.c
313
next = t64;
fs/jfs/jfs_xtree.c
359
*nextp = next;
fs/jfs/jfs_xtree.c
417
next = offsetXAD(&p->xad[index + 1]);
fs/jfs/jfs_xtree.c
418
goto next;
fs/jfs/jfs_xtree.c
434
next = offsetXAD(&p->xad[base]);
fs/jfs/jfs_xtree.c
469
*nextp = next;
fs/jfs/jfs_xtree.c
485
next:
fs/jfs/jfs_xtree.c
542
s64 next;
fs/jfs/jfs_xtree.c
556
if ((rc = xtSearch(ip, xoff, &next, &cmp, &btstack, XT_INSERT)))
fs/jfs/jfs_xtree.c
564
if ((cmp == 0) || (next && (xlen > next - xoff))) {
fs/jfs/jfs_xtree.c
84
static int xtSearch(struct inode *ip, s64 xoff, s64 *next, int *cmpp,
fs/jfs/jfs_xtree.h
69
__le64 next; /* 8: */
fs/kernfs/file.c
178
void *next = ops->seq_start(sf, ppos);
fs/kernfs/file.c
180
if (next == ERR_PTR(-ENODEV))
fs/kernfs/file.c
181
kernfs_seq_stop_active(sf, next);
fs/kernfs/file.c
182
return next;
fs/kernfs/file.c
193
void *next = ops->seq_next(sf, v, ppos);
fs/kernfs/file.c
195
if (next == ERR_PTR(-ENODEV))
fs/kernfs/file.c
196
kernfs_seq_stop_active(sf, next);
fs/kernfs/file.c
197
return next;
fs/kernfs/file.c
228
.next = kernfs_seq_next,
fs/libfs.c
119
p = &d->d_sib.next;
fs/libfs.c
136
p = &cursor->d_sib.next;
fs/libfs.c
195
struct dentry *next = NULL;
fs/libfs.c
204
p = &cursor->d_sib.next;
fs/libfs.c
206
while ((next = scan_positives(cursor, p, 1, next)) != NULL) {
fs/libfs.c
207
if (!dir_emit(ctx, next->d_name.name, next->d_name.len,
fs/libfs.c
208
d_inode(next)->i_ino,
fs/libfs.c
209
fs_umode_to_dtype(d_inode(next)->i_mode)))
fs/libfs.c
212
p = &next->d_sib.next;
fs/libfs.c
216
if (next)
fs/libfs.c
217
hlist_add_before(&cursor->d_sib, &next->d_sib);
fs/libfs.c
219
dput(next);
fs/libfs.c
453
bool next)
fs/libfs.c
458
if (next)
fs/libfs.c
513
struct dentry *next;
fs/libfs.c
519
next = find_positive_dentry(dir, dentry, true);
fs/libfs.c
522
if (!next)
fs/libfs.c
524
dentry = next;
fs/lockd/clntlock.c
232
struct file_lock *fl, *next;
fs/lockd/clntlock.c
258
list_for_each_entry_safe(fl, next, &host->h_reclaim, fl_u.nfs_fl.list) {
fs/lockd/host.c
42
#define for_each_host_safe(host, next, chain, table) \
fs/lockd/host.c
45
hlist_for_each_entry_safe((host), (next), \
fs/lockd/host.c
660
struct hlist_node *next;
fs/lockd/host.c
674
for_each_host_safe(host, next, chain, nlm_server_hosts) {
fs/lockd/svclock.c
1053
block = list_entry(nlm_blocked.next, struct nlm_block, b_list);
fs/lockd/svclock.c
323
struct nlm_block *block, *next;
fs/lockd/svclock.c
328
list_for_each_entry_safe(block, next, &file->f_blocks, b_flist) {
fs/lockd/svcsubs.c
302
struct hlist_node *next;
fs/lockd/svcsubs.c
308
hlist_for_each_entry_safe(file, next, &nlm_files[i], f_list) {
fs/locks.c
3044
.next = locks_next,
fs/mbcache.c
163
node = entry->e_hash_list.next;
fs/mbcache.c
173
node = node->next;
fs/mbcache.c
407
struct mb_cache_entry *entry, *next;
fs/mbcache.c
415
list_for_each_entry_safe(entry, next, &cache->c_list, e_list) {
fs/namespace.c
1106
struct list_head *next = p->mnt_mounts.next;
fs/namespace.c
1107
if (next == &p->mnt_mounts) {
fs/namespace.c
1111
next = p->mnt_child.next;
fs/namespace.c
1112
if (next != &p->mnt_parent->mnt_mounts)
fs/namespace.c
1117
return list_entry(next, struct mount, mnt_child);
fs/namespace.c
1550
struct mount *next = node_to_mount(node);
fs/namespace.c
1551
*pos = next->mnt_id_unique;
fs/namespace.c
1552
return next;
fs/namespace.c
1577
.next = m_next,
fs/namespace.c
1988
while (mp.node.next) {
fs/namespace.c
1989
mnt = hlist_entry(mp.node.next, struct mount, mnt_mp_list);
fs/namespace.c
3928
struct mount *mnt, *next;
fs/namespace.c
3944
list_for_each_entry_safe(mnt, next, mounts, mnt_expire) {
fs/namespace.c
3970
struct list_head *next;
fs/namespace.c
3974
next = this_parent->mnt_mounts.next;
fs/namespace.c
3976
while (next != &this_parent->mnt_mounts) {
fs/namespace.c
3977
struct list_head *tmp = next;
fs/namespace.c
3980
next = tmp->next;
fs/namespace.c
4000
next = this_parent->mnt_child.next;
fs/namespace.c
6357
goto next;
fs/namespace.c
6363
next: ;
fs/namespace.c
665
struct mount *next = m->mnt_next_for_sb;
fs/namespace.c
667
if (next)
fs/namespace.c
668
next->mnt_pprev_for_sb = p;
fs/namespace.c
669
*p = next;
fs/netfs/fscache_cache.c
425
.next = fscache_caches_seq_next,
fs/netfs/fscache_cookie.c
1180
.next = fscache_cookies_seq_next,
fs/netfs/fscache_volume.c
528
.next = fscache_volumes_seq_next,
fs/netfs/iterator.c
252
folioq = folioq->next;
fs/netfs/iterator.c
272
folioq = folioq->next;
fs/netfs/main.c
101
.next = netfs_requests_seq_next,
fs/netfs/misc.c
30
while (tail->next)
fs/netfs/misc.c
31
tail = tail->next;
fs/netfs/misc.c
42
tail->next = p;
fs/netfs/misc.c
80
struct folio_queue *next;
fs/netfs/misc.c
85
for (; fq; fq = next) {
fs/netfs/misc.c
99
next = fq->next;
fs/netfs/read_retry.c
101
list_for_each_continue(next, &stream->subrequests) {
fs/netfs/read_retry.c
102
subreq = list_entry(next, struct netfs_io_subrequest, rreq_link);
fs/netfs/read_retry.c
238
} while (!list_is_head(next, &stream->subrequests));
fs/netfs/read_retry.c
286
for (p = rreq->buffer.tail; p; p = p->next) {
fs/netfs/read_retry.c
30
struct list_head *next;
fs/netfs/read_retry.c
73
next = stream->subrequests.next;
fs/netfs/read_retry.c
86
from = list_entry(next, struct netfs_io_subrequest, rreq_link);
fs/netfs/rolling_buffer.c
109
smp_store_release(&head->next, fq);
fs/netfs/rolling_buffer.c
184
struct folio_queue *spent = roll->tail, *next = READ_ONCE(spent->next);
fs/netfs/rolling_buffer.c
186
if (!next)
fs/netfs/rolling_buffer.c
188
next->prev = NULL;
fs/netfs/rolling_buffer.c
190
roll->tail = next;
fs/netfs/rolling_buffer.c
191
return next;
fs/netfs/rolling_buffer.c
205
roll->tail = p->next;
fs/netfs/write_issue.c
857
for (fq = (struct folio_queue *)iter->folioq; fq; fq = fq->next) {
fs/netfs/write_retry.c
191
} while (!list_is_head(next, &stream->subrequests));
fs/netfs/write_retry.c
20
struct list_head *next;
fs/netfs/write_retry.c
53
next = stream->subrequests.next;
fs/netfs/write_retry.c
66
from = list_entry(next, struct netfs_io_subrequest, rreq_link);
fs/netfs/write_retry.c
75
list_for_each_continue(next, &stream->subrequests) {
fs/netfs/write_retry.c
76
subreq = list_entry(next, struct netfs_io_subrequest, rreq_link);
fs/nfs/blocklayout/extent_tree.c
236
struct pnfs_block_extent *next = ext_tree_next(be);
fs/nfs/blocklayout/extent_tree.c
240
be = next;
fs/nfs/client.c
1302
.next = nfs_server_list_next,
fs/nfs/client.c
1314
.next = nfs_volume_list_next,
fs/nfs/dir.c
2897
cache = list_entry(head->next, struct nfs_access_entry, lru);
fs/nfs/dir.c
2907
struct nfs_inode *nfsi, *next;
fs/nfs/dir.c
2912
list_for_each_entry_safe(nfsi, next, &nfs_access_lru_list, access_cache_inode_lru) {
fs/nfs/dir.c
2921
cache = list_entry(nfsi->access_cache_entry_lru.next,
fs/nfs/direct.c
293
struct nfs_page *req = nfs_list_entry(hdr->pages.next);
fs/nfs/direct.c
314
req = nfs_list_entry(head->next);
fs/nfs/direct.c
584
req = nfs_list_entry(reqs.next);
fs/nfs/direct.c
606
req = nfs_list_entry(reqs.next);
fs/nfs/direct.c
646
req = nfs_list_entry(data->pages.next);
fs/nfs/direct.c
722
req = nfs_list_entry(reqs.next);
fs/nfs/direct.c
790
req = nfs_list_entry(hdr->pages.next);
fs/nfs/direct.c
816
req = nfs_list_entry(head->next);
fs/nfs/direct.c
837
req = nfs_list_entry(hdr->pages.next);
fs/nfs/nfs4client.c
158
dss = list_entry(clp->cl_ds_clients.next,
fs/nfs/nfs4client.c
172
cp_state = list_entry(clp->pending_cb_stateids.next,
fs/nfs/nfs4session.c
137
p = &slot->next;
fs/nfs/nfs4session.c
284
p = &(*p)->next;
fs/nfs/nfs4session.c
47
p = &(*p)->next;
fs/nfs/nfs4session.c
51
*p = slot->next;
fs/nfs/nfs4session.h
23
struct nfs4_slot *next;
fs/nfs/nfs4state.c
1036
struct nfs_seqid *next = list_next_entry(seqid, list);
fs/nfs/nfs4state.c
1037
rpc_wake_up_queued_task(&sequence->wait, next->task);
fs/nfs/pagelist.c
102
hdr->req = nfs_list_entry(mirror->pg_list.next);
fs/nfs/pagelist.c
1386
struct nfs_page *req = nfs_list_entry(pages.next);
fs/nfs/pagelist.c
357
struct nfs_page *tmp, *next;
fs/nfs/pagelist.c
364
next = tmp->wb_this_page;
fs/nfs/pagelist.c
369
tmp = next;
fs/nfs/pagelist.c
914
req = nfs_list_entry(head->next);
fs/nfs/pnfs.c
1545
struct pnfs_layout_segment *lseg, *next;
fs/nfs/pnfs.c
1617
list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list) {
fs/nfs/pnfs.c
2659
struct pnfs_layout_segment *lseg, *next;
fs/nfs/pnfs.c
2670
list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list)
fs/nfs/pnfs.c
464
struct pnfs_layout_segment *lseg, *next;
fs/nfs/pnfs.c
469
list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list)
fs/nfs/pnfs.c
725
struct pnfs_layout_segment *lseg, *next;
fs/nfs/pnfs.c
733
list_for_each_entry_safe(lseg, next, &lo->plh_segs, pls_list)
fs/nfs/pnfs.c
762
struct pnfs_layout_segment *lseg, *next;
fs/nfs/pnfs.c
764
list_for_each_entry_safe(lseg, next, &lo->plh_return_segs, pls_list) {
fs/nfs/pnfs.c
851
struct pnfs_layout_hdr *lo, *next;
fs/nfs/pnfs.c
854
list_for_each_entry_safe(lo, next, &server->layouts, plh_layouts) {
fs/nfs/pnfs.c
892
lo = list_entry(layout_list->next, struct pnfs_layout_hdr,
fs/nfs/read.c
147
struct nfs_page *req = nfs_list_entry(hdr->pages.next);
fs/nfs/read.c
205
req = nfs_list_entry(head->next);
fs/nfs/write.c
1377
req = nfs_list_entry(head->next);
fs/nfs/write.c
1701
first = nfs_list_entry(data->pages.next);
fs/nfs/write.c
1735
req = nfs_list_entry(page_list->next);
fs/nfs/write.c
1807
req = nfs_list_entry(data->pages.next);
fs/nfs/write.c
1825
goto next;
fs/nfs/write.c
1835
goto next;
fs/nfs/write.c
1841
next:
fs/nfs/write.c
918
struct nfs_page *req = nfs_list_entry(hdr->pages.next);
fs/nfs/write.c
935
goto next;
fs/nfs/write.c
939
next:
fs/nfsd/export.c
1489
.next = cache_seq_next_rcu,
fs/nfsd/filecache.c
472
list_move(l->freeme.next, &dispose);
fs/nfsd/nfs4proc.c
1464
struct list_head *pos, *next;
fs/nfsd/nfs4proc.c
1467
list_for_each_safe(pos, next, &clp->async_copies) {
fs/nfsd/nfs4proc.c
2966
struct nfsd4_op *next = &argp->ops[resp->opcnt];
fs/nfsd/nfs4proc.c
2983
if (next->opnum == OP_ILLEGAL)
fs/nfsd/nfs4proc.c
2985
nextd = OPDESC(next);
fs/nfsd/nfs4state.c
1656
stp = list_entry(open_stp->st_locks.next,
fs/nfsd/nfs4state.c
2431
ses = list_entry(clp->cl_sessions.next, struct nfsd4_session,
fs/nfsd/nfs4state.c
2509
dp = list_entry(clp->cl_delegations.next, struct nfs4_delegation, dl_perclnt);
fs/nfsd/nfs4state.c
2515
dp = list_entry(reaplist.next, struct nfs4_delegation, dl_recall_lru);
fs/nfsd/nfs4state.c
2520
dp = list_entry(clp->cl_revoked.next, struct nfs4_delegation, dl_recall_lru);
fs/nfsd/nfs4state.c
2525
oo = list_entry(clp->cl_openowners.next, struct nfs4_openowner, oo_perclient);
fs/nfsd/nfs4state.c
3090
.next = states_next,
fs/nfsd/nfs4state.c
6812
struct list_head *pos, *next;
fs/nfsd/nfs4state.c
6819
list_for_each_safe(pos, next, &nn->client_lru) {
fs/nfsd/nfs4state.c
6849
struct list_head *pos, *next;
fs/nfsd/nfs4state.c
6856
list_for_each_safe(pos, next, &nn->client_lru) {
fs/nfsd/nfs4state.c
6873
struct list_head *pos, *next;
fs/nfsd/nfs4state.c
6876
list_for_each_safe(pos, next, reaplist) {
fs/nfsd/nfs4state.c
6928
struct list_head *pos, *next, reaplist;
fs/nfsd/nfs4state.c
6958
list_for_each_safe(pos, next, &nn->del_recall_lru) {
fs/nfsd/nfs4state.c
7054
struct list_head *pos, *next;
fs/nfsd/nfs4state.c
7058
list_for_each_safe(pos, next, &nn->client_lru) {
fs/nfsd/nfs4state.c
8899
crp = list_entry(nn->reclaim_str_hashtbl[i].next,
fs/nfsd/nfs4state.c
9035
clp = list_entry(nn->conf_id_hashtbl[i].next, struct nfs4_client, cl_idhash);
fs/nfsd/nfs4state.c
9044
clp = list_entry(nn->unconf_id_hashtbl[i].next, struct nfs4_client, cl_idhash);
fs/nfsd/nfs4state.c
9111
struct list_head *pos, *next, reaplist;
fs/nfsd/nfs4state.c
9121
list_for_each_safe(pos, next, &nn->del_recall_lru) {
fs/nfsd/nfs4state.c
9127
list_for_each_safe(pos, next, &reaplist) {
fs/nfsd/nfs4xdr.c
130
tb->next = argp->to_free;
fs/nfsd/nfs4xdr.c
2770
char *str, *end, *next;
fs/nfsd/nfs4xdr.c
2787
next = end + 1;
fs/nfsd/nfs4xdr.c
2788
if (*end && (!*next || *next == sep)) {
fs/nfsd/nfs4xdr.c
2805
end = next;
fs/nfsd/nfs4xdr.c
5449
struct nfsd4_test_stateid_id *stateid, *next;
fs/nfsd/nfs4xdr.c
5455
list_for_each_entry_safe(stateid, next,
fs/nfsd/nfs4xdr.c
6339
args->to_free = tb->next;
fs/nfsd/nfssvc.c
887
struct svc_xprt *perm_sock = list_entry(rqstp->rq_server->sv_permsocks.next, typeof(struct svc_xprt), xpt_list);
fs/nfsd/xdr4.h
901
struct svcxdr_tmpbuf *next;
fs/nilfs2/cpfile.c
628
__u64 curr = *cnop, next;
fs/nilfs2/cpfile.c
669
next = le64_to_cpu(cp->cp_snapshot_list.ssl_next);
fs/nilfs2/cpfile.c
670
if (next == 0)
fs/nilfs2/cpfile.c
674
next_blkoff = nilfs_cpfile_get_blkoff(cpfile, next);
fs/nilfs2/cpfile.c
677
ret = nilfs_cpfile_get_checkpoint_block(cpfile, next,
fs/nilfs2/cpfile.c
684
offset = nilfs_cpfile_checkpoint_offset(cpfile, next, bh);
fs/nilfs2/cpfile.c
686
curr = next;
fs/nilfs2/cpfile.c
901
__u64 next, prev;
fs/nilfs2/cpfile.c
931
next = le64_to_cpu(list->ssl_next);
fs/nilfs2/cpfile.c
935
if (next != 0) {
fs/nilfs2/cpfile.c
936
ret = nilfs_cpfile_get_checkpoint_block(cpfile, next, 0,
fs/nilfs2/cpfile.c
942
cpfile, next, next_bh);
fs/nilfs2/cpfile.c
969
list->ssl_next = cpu_to_le64(next);
fs/nilfs2/mdt.c
289
__u64 next;
fs/nilfs2/mdt.c
303
ret = nilfs_bmap_seek_key(NILFS_I(inode)->i_bmap, start + 1, &next);
fs/nilfs2/mdt.c
305
if (next <= end) {
fs/nilfs2/mdt.c
306
ret = nilfs_mdt_read_block(inode, next, true, out_bh);
fs/nilfs2/mdt.c
308
*blkoff = next;
fs/nilfs2/recovery.c
569
goto next;
fs/nilfs2/recovery.c
582
next:
fs/nilfs2/segbuf.c
155
bh_sum = list_entry(segbuf->sb_segsum_buffers.next,
fs/nilfs2/segbuf.c
164
raw_sum->ss_next = cpu_to_le64(segbuf->sb_sum.next);
fs/nilfs2/segbuf.c
183
bh = list_entry(segbuf->sb_segsum_buffers.next, struct buffer_head,
fs/nilfs2/segbuf.c
210
bh = list_entry(segbuf->sb_segsum_buffers.next, struct buffer_head,
fs/nilfs2/segbuf.c
92
segbuf->sb_sum.next = nilfs_get_segment_start_blocknr(nilfs, nextnum);
fs/nilfs2/segbuf.h
100
(list_entry((bh)->b_assoc_buffers.next, struct buffer_head, \
fs/nilfs2/segbuf.h
102
#define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next == head)
fs/nilfs2/segbuf.h
41
sector_t next;
fs/nilfs2/segbuf.h
87
#define NILFS_NEXT_SEGBUF(segbuf) NILFS_LIST_SEGBUF((segbuf)->sb_list.next)
fs/nilfs2/segbuf.h
90
#define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next)
fs/nilfs2/segbuf.h
91
#define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head))
fs/nilfs2/segbuf.h
98
(list_entry((head)->next, struct buffer_head, b_assoc_buffers))
fs/nls/nls_base.c
250
if (nls->next)
fs/nls/nls_base.c
260
tmp = &(*tmp)->next;
fs/nls/nls_base.c
262
nls->next = tables;
fs/nls/nls_base.c
276
*tmp = nls->next;
fs/nls/nls_base.c
280
tmp = &(*tmp)->next;
fs/nls/nls_base.c
290
for (nls = tables; nls; nls = nls->next) {
fs/notify/fsnotify.c
396
node = srcu_dereference(mark->obj_list.next,
fs/notify/mark.c
1098
struct fsnotify_mark *mark, *next;
fs/notify/mark.c
1108
list_for_each_entry_safe(mark, next, &private_destroy_list, g_list) {
fs/ntfs3/attrib.c
265
char *next;
fs/ntfs3/attrib.c
278
next = Add2Ptr(attr, asize);
fs/ntfs3/attrib.c
365
memmove(next, attr, used - aoff);
fs/ntfs3/attrib.c
394
char *next = Add2Ptr(attr, asize);
fs/ntfs3/attrib.c
398
memmove(next + dsize, next, tail);
fs/ntfs3/attrib.c
404
memmove(next + dsize, next, tail);
fs/ntfs3/attrib.c
405
memset(next, 0, dsize);
fs/ntfs3/bitmap.c
125
struct rb_node *node, *next;
fs/ntfs3/bitmap.c
134
next = rb_next(node);
fs/ntfs3/bitmap.c
138
node = next;
fs/ntfs3/frecord.c
113
struct rb_node *next = rb_next(node);
fs/ntfs3/frecord.c
118
node = next;
fs/ntfs3/frecord.c
1575
struct rb_node *next = rb_next(node);
fs/ntfs3/frecord.c
1585
node = next;
fs/ntfs3/frecord.c
1615
next:
fs/ntfs3/frecord.c
1622
goto next;
fs/ntfs3/frecord.c
1625
goto next;
fs/ntfs3/frecord.c
1631
goto next;
fs/ntfs3/frecord.c
1635
goto next;
fs/ntfs3/frecord.c
3103
struct rb_node *node, *next;
fs/ntfs3/frecord.c
3197
for (node = rb_first(&ni->mi_tree); node; node = next) {
fs/ntfs3/frecord.c
3201
next = rb_next(node);
fs/ntfs3/fslog.c
113
__le32 next; // 0x00: RESTART_ENTRY_ALLOCATED if allocated
fs/ntfs3/fslog.c
127
__le32 next; // 0x00: RESTART_ENTRY_ALLOCATED if allocated
fs/ntfs3/fslog.c
150
__le32 next; // 0x00: RESTART_ENTRY_ALLOCATED if allocated
fs/ntfs3/fslog.c
163
__le32 next; // 0x00: RESTART_ENTRY_ALLOCATED if allocated
fs/ntfs3/fslog.c
187
__le32 next; // 0x00: RESTART_ENTRY_ALLOCATED if allocated
fs/ntfs3/fslog.c
2626
struct NTFS_DE *next = Add2Ptr(e, esize);
fs/ntfs3/fslog.c
2628
if (esize < min_de || PtrOffset(hdr, next) > used ||
fs/ntfs3/fslog.c
2636
e = next;
fs/ntfs3/fslog.c
2940
u8 *next = Add2Ptr(attr, asize);
fs/ntfs3/fslog.c
2943
memmove(Add2Ptr(attr, nsize), next, used - PtrOffset(rec, next));
fs/ntfs3/fslog.c
4233
struct DIR_PAGE_ENTRY *next = dp;
fs/ntfs3/fslog.c
4235
while ((next = enum_rstbl(dptbl, next))) {
fs/ntfs3/fslog.c
4236
if (next->target_attr == dp->target_attr &&
fs/ntfs3/fslog.c
4237
next->vcn == dp->vcn) {
fs/ntfs3/fslog.c
4238
if (le64_to_cpu(next->oldest_lsn) <
fs/ntfs3/fslog.c
4240
dp->oldest_lsn = next->oldest_lsn;
fs/ntfs3/fslog.c
4243
free_rsttbl_idx(dptbl, PtrOffset(dptbl, next));
fs/ntfs3/fslog.c
4452
tr->next != RESTART_ENTRY_ALLOCATED_LE) {
fs/ntfs3/fslog.c
4828
if (oe->next != RESTART_ENTRY_ALLOCATED_LE)
fs/ntfs3/fslog.c
4932
if (oe->next != RESTART_ENTRY_ALLOCATED_LE) {
fs/ntfs3/fsntfs.c
2100
u32 next, left;
fs/ntfs3/fsntfs.c
2175
next = sbi->security.next_off & (SecurityDescriptorsBlockSize - 1);
fs/ntfs3/fsntfs.c
2176
left = SecurityDescriptorsBlockSize - next;
fs/ntfs3/fsntfs.c
644
next:
fs/ntfs3/fsntfs.c
653
goto next;
fs/ntfs3/fsntfs.c
661
goto next;
fs/ntfs3/fsntfs.c
664
goto next;
fs/ntfs3/fsntfs.c
667
goto next;
fs/ntfs3/fsntfs.c
671
goto next;
fs/ntfs3/index.c
1341
goto next;
fs/ntfs3/index.c
1385
next:
fs/ntfs3/index.c
2285
struct NTFS_DE *e, *re, *next, *prev, *me;
fs/ntfs3/index.c
2366
next = de_get_next(e);
fs/ntfs3/index.c
2368
err = indx_get_entry_to_replace(indx, ni, next, &re, fnd2);
fs/ntfs3/index.c
2394
indx_free_children(indx, ni, next, true);
fs/ntfs3/index.c
2396
de_set_vbn_le(next, de_get_vbn_le(e));
fs/ntfs3/lib/decompress_common.h
146
if (unlikely(is->end == is->next))
fs/ntfs3/lib/decompress_common.h
148
return *is->next++;
fs/ntfs3/lib/decompress_common.h
157
if (unlikely(is->end - is->next < 2))
fs/ntfs3/lib/decompress_common.h
159
v = get_unaligned_le16(is->next);
fs/ntfs3/lib/decompress_common.h
160
is->next += 2;
fs/ntfs3/lib/decompress_common.h
170
if (unlikely(is->end - is->next < 4))
fs/ntfs3/lib/decompress_common.h
172
v = get_unaligned_le32(is->next);
fs/ntfs3/lib/decompress_common.h
173
is->next += 4;
fs/ntfs3/lib/decompress_common.h
184
if ((size_t)(is->end - is->next) < count)
fs/ntfs3/lib/decompress_common.h
186
memcpy(dst_buffer, is->next, count);
fs/ntfs3/lib/decompress_common.h
187
is->next += count;
fs/ntfs3/lib/decompress_common.h
66
const u8 *next;
fs/ntfs3/lib/decompress_common.h
78
is->next = buffer;
fs/ntfs3/lib/decompress_common.h
79
is->end = is->next + size;
fs/ntfs3/lib/decompress_common.h
91
if (is->end - is->next >= 2) {
fs/ntfs3/lib/decompress_common.h
92
is->bitbuf |= (u32)get_unaligned_le16(is->next)
fs/ntfs3/lib/decompress_common.h
94
is->next += 2;
fs/ntfs3/lznt.c
250
goto next;
fs/ntfs3/lznt.c
279
next:
fs/ntfs3/record.c
579
char *next;
fs/ntfs3/record.c
588
next = Add2Ptr(attr, asize);
fs/ntfs3/record.c
596
memmove(next + dsize, next, tail);
fs/ntfs3/record.c
597
memset(next, 0, dsize);
fs/ntfs3/record.c
605
memmove(next - dsize, next, tail);
fs/ntfs3/record.c
635
char *next = Add2Ptr(attr, asize);
fs/ntfs3/record.c
642
memmove(next + dsize, next, tail);
fs/ntfs3/record.c
648
memmove(next, next + dsize, tail);
fs/ntfs3/record.c
654
memmove(next + new_run_size - run_size, next + dsize, tail);
fs/ocfs2/alloc.c
7512
int ret = 0, count = 0, next;
fs/ocfs2/alloc.c
7525
next = ocfs2_find_next_bit(bitmap, max, start);
fs/ocfs2/alloc.c
7527
if ((next - start) >= minbits) {
fs/ocfs2/alloc.c
7529
start, next - start);
fs/ocfs2/alloc.c
7534
count += next - start;
fs/ocfs2/alloc.c
7536
start = next + 1;
fs/ocfs2/aops.c
406
struct buffer_head *next;
fs/ocfs2/aops.c
410
block_start = block_end, bh = next)
fs/ocfs2/aops.c
412
next = bh->b_this_page;
fs/ocfs2/cluster/heartbeat.c
765
event = list_entry(o2hb_node_events.next,
fs/ocfs2/cluster/netdebug.c
159
.next = nst_seq_next,
fs/ocfs2/cluster/netdebug.c
373
.next = sc_seq_next,
fs/ocfs2/dir.c
3521
goto next;
fs/ocfs2/dir.c
3531
next:
fs/ocfs2/dir.c
734
goto next;
fs/ocfs2/dir.c
742
goto next;
fs/ocfs2/dir.c
757
next:
fs/ocfs2/dlm/dlmdebug.c
554
.next = lockres_seq_next,
fs/ocfs2/dlm/dlmdomain.c
433
iter = res->hash_node.next;
fs/ocfs2/dlm/dlmmaster.c
2916
struct dlm_lock *lock, *next;
fs/ocfs2/dlm/dlmmaster.c
2923
list_for_each_entry_safe(lock, next, queue, list) {
fs/ocfs2/dlm/dlmrecovery.c
1036
struct dlm_lock_resource *res, *next;
fs/ocfs2/dlm/dlmrecovery.c
1040
list_for_each_entry_safe(res, next, &dlm->reco.resources, recovering) {
fs/ocfs2/dlm/dlmrecovery.c
133
struct dlm_work_item *item, *next;
fs/ocfs2/dlm/dlmrecovery.c
146
list_for_each_entry_safe(item, next, &tmp_list, list) {
fs/ocfs2/dlm/dlmrecovery.c
2047
struct dlm_lock *lock, *next;
fs/ocfs2/dlm/dlmrecovery.c
2066
list_for_each_entry_safe(lock, next, queue, list) {
fs/ocfs2/dlm/dlmrecovery.c
2131
struct dlm_lock_resource *res, *next;
fs/ocfs2/dlm/dlmrecovery.c
2135
list_for_each_entry_safe(res, next, &dlm->reco.resources, recovering) {
fs/ocfs2/dlm/dlmrecovery.c
2253
struct dlm_lock *lock, *next;
fs/ocfs2/dlm/dlmrecovery.c
2267
list_for_each_entry_safe(lock, next, &res->granted, list) {
fs/ocfs2/dlm/dlmrecovery.c
2276
list_for_each_entry_safe(lock, next, &res->converting, list) {
fs/ocfs2/dlm/dlmrecovery.c
2285
list_for_each_entry_safe(lock, next, &res->blocked, list) {
fs/ocfs2/dlm/dlmrecovery.c
764
struct dlm_reco_node_data *ndata, *next;
fs/ocfs2/dlm/dlmrecovery.c
771
list_for_each_entry_safe(ndata, next, &tmplist, list) {
fs/ocfs2/dlm/dlmthread.c
289
lockres = list_entry(dlm->purge_list.next,
fs/ocfs2/dlm/dlmthread.c
364
target = list_entry(res->converting.next, struct dlm_lock, list);
fs/ocfs2/dlm/dlmthread.c
435
target = list_entry(res->blocked.next, struct dlm_lock, list);
fs/ocfs2/dlm/dlmthread.c
578
lock = list_entry(dlm->pending_asts.next,
fs/ocfs2/dlm/dlmthread.c
622
lock = list_entry(dlm->pending_basts.next,
fs/ocfs2/dlm/dlmthread.c
707
res = list_entry(dlm->dirty_list.next,
fs/ocfs2/dlmglue.c
3259
.next = ocfs2_dlm_seq_next,
fs/ocfs2/dlmglue.c
4395
lockres = list_entry(osb->blocked_lock_list.next,
fs/ocfs2/file.c
1500
goto next;
fs/ocfs2/file.c
1510
next:
fs/ocfs2/move_extents.c
849
goto next;
fs/ocfs2/move_extents.c
860
goto next;
fs/ocfs2/move_extents.c
888
next:
fs/ocfs2/quota_local.c
239
struct ocfs2_quota_chunk *pos, *next;
fs/ocfs2/quota_local.c
241
list_for_each_entry_safe(pos, next, head, qc_chunk) {
fs/ocfs2/quota_local.c
318
struct ocfs2_recovery_chunk *next;
fs/ocfs2/quota_local.c
321
list_for_each_entry_safe(rchunk, next, head, rc_list) {
fs/ocfs2/quota_local.c
470
struct ocfs2_recovery_chunk *rchunk, *next;
fs/ocfs2/quota_local.c
476
list_for_each_entry_safe(rchunk, next, &(rec->r_list[type]), rc_list) {
fs/ocfs2/refcounttree.c
1362
struct ocfs2_refcount_rec *next)
fs/ocfs2/refcounttree.c
1365
ocfs2_get_ref_rec_low_cpos(next))
fs/ocfs2/reservations.c
462
struct rb_node *prev, *next;
fs/ocfs2/reservations.c
522
next = rb_first(root);
fs/ocfs2/reservations.c
523
next_resv = rb_entry(next, struct ocfs2_alloc_reservation,
fs/ocfs2/reservations.c
558
next = rb_next(prev);
fs/ocfs2/reservations.c
559
if (next) {
fs/ocfs2/reservations.c
560
next_resv = rb_entry(next,
fs/ocfs2/reservations.c
578
trace_ocfs2_resv_find_window_next(next ? next_resv->r_start: -1,
fs/ocfs2/reservations.c
579
next ? ocfs2_resv_end(next_resv) : -1);
fs/ocfs2/reservations.c
599
if (!next)
fs/ocfs2/reservations.c
602
prev = next;
fs/ocfs2/super.c
1227
.next = NULL,
fs/omfs/dir.c
164
__be64 *entry, next;
fs/omfs/dir.c
184
next = oi->i_sibling;
fs/omfs/dir.c
198
*entry = next;
fs/omfs/file.c
221
sector_t next, offset;
fs/omfs/file.c
239
next = inode->i_ino;
fs/omfs/file.c
243
if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next))
fs/omfs/file.c
247
next = be64_to_cpu(oe->e_next);
fs/omfs/file.c
262
if (next == ~0)
fs/omfs/file.c
266
bh = omfs_bread(inode->i_sb, next);
fs/omfs/file.c
37
u64 next, last;
fs/omfs/file.c
45
next = inode->i_ino;
fs/omfs/file.c
52
bh = omfs_bread(inode->i_sb, next);
fs/omfs/file.c
61
if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next))
fs/omfs/file.c
69
last = next;
fs/omfs/file.c
70
next = be64_to_cpu(oe->e_next);
fs/omfs/file.c
89
if (next == ~0)
fs/omfs/file.c
92
bh = omfs_bread(inode->i_sb, next);
fs/openpromfs/inode.c
140
.next = property_next,
fs/openpromfs/inode.c
228
prop = prop->next;
fs/openpromfs/inode.c
315
prop = prop->next;
fs/openpromfs/inode.c
324
prop = prop->next;
fs/orangefs/devorangefs-req.c
61
struct orangefs_kernel_op_s *op, *next;
fs/orangefs/devorangefs-req.c
68
next,
fs/orangefs/dir.c
12
struct orangefs_dir_part *next;
fs/orangefs/dir.c
142
if (part->next)
fs/orangefs/dir.c
143
part = part->next;
fs/orangefs/dir.c
149
new->next = NULL;
fs/orangefs/dir.c
155
part->next = new;
fs/orangefs/dir.c
221
goto next;
fs/orangefs/dir.c
224
goto next;
fs/orangefs/dir.c
235
next:
fs/orangefs/dir.c
249
while (part->next && count) {
fs/orangefs/dir.c
251
part = part->next;
fs/orangefs/dir.c
274
part = part->next;
fs/orangefs/dir.c
291
struct orangefs_dir_part *next = part->next;
fs/orangefs/dir.c
293
part = next;
fs/orangefs/dir.c
382
struct orangefs_dir_part *next = part->next;
fs/orangefs/dir.c
384
part = next;
fs/orangefs/orangefs-debugfs.c
154
.next = help_next,
fs/orangefs/orangefs-mod.c
212
struct orangefs_kernel_op_s *next;
fs/orangefs/orangefs-mod.c
216
next,
fs/overlayfs/copy_up.c
1224
struct dentry *next;
fs/overlayfs/copy_up.c
1230
next = dget(dentry);
fs/overlayfs/copy_up.c
1233
parent = dget_parent(next);
fs/overlayfs/copy_up.c
1238
dput(next);
fs/overlayfs/copy_up.c
1239
next = parent;
fs/overlayfs/copy_up.c
1243
err = ovl_copy_up_one(parent, next, flags);
fs/overlayfs/copy_up.c
1246
dput(next);
fs/overlayfs/export.c
102
struct dentry *next, *parent = NULL;
fs/overlayfs/export.c
116
next = dget(dentry);
fs/overlayfs/export.c
118
parent = dget_parent(next);
fs/overlayfs/export.c
119
if (WARN_ON(parent == next)) {
fs/overlayfs/export.c
129
err = ovl_encode_maybe_copy_up(next);
fs/overlayfs/export.c
138
dput(next);
fs/overlayfs/export.c
139
next = parent;
fs/overlayfs/export.c
143
dput(next);
fs/overlayfs/export.c
493
struct dentry *next, *parent = NULL;
fs/overlayfs/export.c
500
next = dget(real);
fs/overlayfs/export.c
502
parent = dget_parent(next);
fs/overlayfs/export.c
508
ancestor = ovl_lookup_real_inode(sb, next, layer);
fs/overlayfs/export.c
522
if (parent == next) {
fs/overlayfs/export.c
527
dput(next);
fs/overlayfs/export.c
528
next = parent;
fs/overlayfs/export.c
532
dput(next);
fs/overlayfs/export.c
554
struct dentry *next, *this;
fs/overlayfs/export.c
563
next = dget(real);
fs/overlayfs/export.c
565
parent = dget_parent(next);
fs/overlayfs/export.c
590
if (parent == next) {
fs/overlayfs/export.c
595
dput(next);
fs/overlayfs/export.c
596
next = parent;
fs/overlayfs/export.c
600
this = ovl_lookup_real_one(connected, next, layer);
fs/overlayfs/export.c
626
dput(next);
fs/overlayfs/namei.c
370
const char *next = strchrnul(s, '/');
fs/overlayfs/namei.c
371
size_t thislen = next - s;
fs/overlayfs/namei.c
372
bool end = !next[0];
fs/overlayfs/namei.c
379
d->name.len - rem, next, &base,
fs/overlayfs/readdir.c
1288
goto next;
fs/overlayfs/readdir.c
1292
goto next;
fs/overlayfs/readdir.c
1317
next:
fs/overlayfs/readdir.c
436
int idx, next;
fs/overlayfs/readdir.c
440
for (idx = 0; idx != -1; idx = next) {
fs/overlayfs/readdir.c
441
next = ovl_path_next(idx, dentry, &realpath, &layer);
fs/overlayfs/readdir.c
450
if (next != -1) {
fs/overlayfs/readdir.c
869
od->cursor = p->l_node.next;
fs/overlayfs/super.c
1244
struct dentry *next = dentry, *parent;
fs/overlayfs/super.c
1250
parent = dget_parent(next);
fs/overlayfs/super.c
1253
while (!err && parent != next) {
fs/overlayfs/super.c
1260
next = parent;
fs/overlayfs/super.c
1261
parent = dget_parent(next);
fs/overlayfs/super.c
1262
dput(next);
fs/overlayfs/util.c
1333
char *s, *next, *buf = NULL;
fs/overlayfs/util.c
1354
for (s = buf; *s++ == '/'; s = next) {
fs/overlayfs/util.c
1355
next = strchrnul(s, '/');
fs/overlayfs/util.c
1356
if (s == next)
fs/pnode.c
129
struct mount *next = next_peer(m);
fs/pnode.c
131
if (next != m) {
fs/pnode.c
134
m->mnt_master = next;
fs/pnode.c
138
next = m->mnt_master;
fs/pnode.c
144
if (!next || !will_be_unmounted(next))
fs/pnode.c
145
return next;
fs/pnode.c
146
if (IS_MNT_MARKED(next))
fs/pnode.c
147
return next->mnt_master;
fs/pnode.c
148
m = next;
fs/pnode.c
154
struct mount *next;
fs/pnode.c
156
while ((next = m->mnt_master) != master) {
fs/pnode.c
158
m = next;
fs/pnode.c
184
struct mount *next = next_peer(m);
fs/pnode.c
185
return (next == origin) ? NULL : next;
fs/pnode.c
186
} else if (m->mnt_slave.next)
fs/pnode.c
19
return list_entry(p->mnt_share.next, struct mount, mnt_share);
fs/pnode.c
232
struct mount *next;
fs/pnode.c
235
next = next_peer(m);
fs/pnode.c
237
if (next == origin)
fs/pnode.c
239
} else if (m->mnt_slave.next != &next->mnt_slave)
fs/pnode.c
241
m = next;
fs/pnode.c
246
if (m->mnt_slave.next)
fs/pnode.c
251
if (master->mnt_slave.next == &m->mnt_slave)
fs/pnode.c
29
return hlist_entry(p->mnt_slave.next, struct mount, mnt_slave);
fs/proc/array.c
800
.next = children_seq_next,
fs/proc/base.c
2570
.next = timers_next,
fs/proc/consoles.c
106
.next = c_next,
fs/proc/consoles.c
95
return hlist_entry_safe(con->node.next, struct console, node);
fs/proc/devices.c
51
.next = devinfo_next,
fs/proc/generic.c
165
const char *cp = name, *next;
fs/proc/generic.c
169
while ((next = strchr(cp, '/')) != NULL) {
fs/proc/generic.c
170
de = pde_subdir_find(de, cp, next - cp);
fs/proc/generic.c
175
cp = next + 1;
fs/proc/generic.c
310
struct proc_dir_entry *next;
fs/proc/generic.c
320
next = pde_subdir_next(de);
fs/proc/generic.c
322
de = next;
fs/proc/generic.c
751
struct proc_dir_entry *root = NULL, *de, *next;
fs/proc/generic.c
777
next = pde_subdir_first(de);
fs/proc/generic.c
778
if (next) {
fs/proc/generic.c
779
if (unlikely(pde_is_permanent(next))) {
fs/proc/generic.c
782
next->parent->name, next->name);
fs/proc/generic.c
785
pde_erase(next, de);
fs/proc/generic.c
786
de = next;
fs/proc/generic.c
789
next = de->parent;
fs/proc/generic.c
791
next->nlink--;
fs/proc/generic.c
800
de = next;
fs/proc/interrupts.c
32
.next = int_seq_next,
fs/proc/nommu.c
105
.next = nommu_region_list_next,
fs/proc/proc_tty.c
122
.next = t_next,
fs/proc/proc_tty.c
73
if (&p->tty_drivers == tty_drivers.next) {
fs/proc/proc_tty.c
94
dev_t next = MKDEV(MAJOR(from)+1, 0);
fs/proc/proc_tty.c
95
show_tty_range(m, p, from, next - from);
fs/proc/proc_tty.c
96
from = next;
fs/proc/stat.c
72
unsigned int i, next = 0;
fs/proc/stat.c
75
show_irq_gap(p, i - next);
fs/proc/stat.c
77
next = i + 1;
fs/proc/stat.c
79
show_irq_gap(p, irq_get_nr_irqs() - next);
fs/proc/task_mmu.c
1522
.next = m_next,
fs/proc/task_mmu.c
2777
unsigned long next = addr + PAGE_SIZE;
fs/proc/task_mmu.c
2784
p, addr, &next);
fs/proc/task_mmu.c
2785
if (next == addr)
fs/proc/task_mmu.c
2792
flush_end = next;
fs/proc/task_mmu.c
2801
unsigned long next = addr + PAGE_SIZE;
fs/proc/task_mmu.c
2806
ret = pagemap_scan_output(categories, p, addr, &next);
fs/proc/task_mmu.c
2807
if (next == addr)
fs/proc/task_mmu.c
2818
flush_end = next;
fs/proc/task_mmu.c
3377
.next = m_next,
fs/proc/task_mmu.c
507
.next = m_next,
fs/proc/task_nommu.c
247
.next = m_next,
fs/proc/vmcore.c
104
list_del_rcu(&cb->next);
fs/proc/vmcore.c
123
list_for_each_entry_srcu(cb, &vmcore_cb_list, next,
fs/proc/vmcore.c
86
INIT_LIST_HEAD(&cb->next);
fs/proc/vmcore.c
88
list_add_tail(&cb->next, &vmcore_cb_list);
fs/pstore/inode.c
125
.next = pstore_ftrace_seq_next,
fs/resctrl/ctrlmondata.c
254
next:
fs/resctrl/ctrlmondata.c
287
goto next;
fs/resctrl/ctrlmondata.c
962
next:
fs/resctrl/ctrlmondata.c
991
goto next;
fs/resctrl/monitor.c
1710
next:
fs/resctrl/monitor.c
1740
goto next;
fs/resctrl/pseudo_lock.c
123
struct pseudo_lock_pm_req *pm_req, *next;
fs/resctrl/pseudo_lock.c
125
list_for_each_entry_safe(pm_req, next, &plr->pm_reqs, list) {
fs/resctrl/rdtgroup.c
1739
next:
fs/resctrl/rdtgroup.c
1767
goto next;
fs/romfs/super.c
187
nextfh = be32_to_cpu(ri.next);
fs/romfs/super.c
240
if ((be32_to_cpu(ri.next) & ROMFH_TYPE) == ROMFH_HRD)
fs/romfs/super.c
247
offset = be32_to_cpu(ri.next) & ROMFH_MASK;
fs/romfs/super.c
288
nextfh = be32_to_cpu(ri.next);
fs/select.c
1010
for (walk = head; walk; walk = walk->next) {
fs/select.c
1021
walk = head->next;
fs/select.c
1024
walk = walk->next;
fs/select.c
152
p = p->next;
fs/select.c
174
new_table->next = table;
fs/select.c
835
struct poll_list *next;
fs/select.c
895
for (walk = list; walk != NULL; walk = walk->next) {
fs/select.c
96
struct poll_table_page * next;
fs/select.c
982
walk->next = NULL;
fs/select.c
996
walk = walk->next = kmalloc_flex(*walk, entries, len);
fs/seq_file.c
1012
return node->next;
fs/seq_file.c
1083
return rcu_dereference(node->next);
fs/seq_file.c
1127
if (node->next)
fs/seq_file.c
1128
return node->next;
fs/seq_file.c
121
p = m->op->next(m, p, &m->index);
fs/seq_file.c
237
p = m->op->next(m, p, &m->index);
fs/seq_file.c
263
p = m->op->next(m, p, &m->index);
fs/seq_file.c
266
m->op->next);
fs/seq_file.c
581
op->next = single_next;
fs/seq_file.c
922
lh = ((struct list_head *)v)->next;
fs/smb/client/file.c
1734
(lock->blist.next == &lock->blist));
fs/smb/client/file.c
1936
el = locks_to_send.next;
fs/smb/client/reparse.c
1127
u32 next = 0;
fs/smb/client/reparse.c
1156
ea = (void *)((u8 *)ea + next);
fs/smb/client/reparse.c
1157
next = le32_to_cpu(ea->next_entry_offset);
fs/smb/client/reparse.c
1178
} while (next);
fs/smb/client/reparse.c
596
u32 next;
fs/smb/client/reparse.c
616
u32 dlen = 0, next = 0;
fs/smb/client/reparse.c
629
xattrs[i].next = ALIGN(sizeof(*ea) + name_size +
fs/smb/client/reparse.c
631
dlen += xattrs[i].next;
fs/smb/client/reparse.c
640
ea = (void *)((u8 *)ea + next);
fs/smb/client/reparse.c
641
next = xattrs[i].next;
fs/smb/client/reparse.c
642
ea->next_entry_offset = cpu_to_le32(next);
fs/smb/client/smb1ops.c
647
u32 next = 0;
fs/smb/client/smb1ops.c
651
ea = (void *)((u8 *)ea + next);
fs/smb/client/smb1ops.c
652
next = le32_to_cpu(ea->next_entry_offset);
fs/smb/client/smb1ops.c
653
} while (next);
fs/smb/client/smb1ops.c
693
u32 next = 0;
fs/smb/client/smb1ops.c
697
ea = (void *)((u8 *)ea + next);
fs/smb/client/smb1ops.c
698
next = le32_to_cpu(ea->next_entry_offset);
fs/smb/client/smb1ops.c
699
} while (next);
fs/smb/client/smb2inode.c
100
{ .next = NEXT_OFF, .name_len = SMB2_WSL_XATTR_NAME_LEN, .name = SMB2_WSL_XATTR_UID, },
fs/smb/client/smb2inode.c
101
{ .next = NEXT_OFF, .name_len = SMB2_WSL_XATTR_NAME_LEN, .name = SMB2_WSL_XATTR_GID, },
fs/smb/client/smb2inode.c
102
{ .next = NEXT_OFF, .name_len = SMB2_WSL_XATTR_NAME_LEN, .name = SMB2_WSL_XATTR_MODE, },
fs/smb/client/smb2inode.c
103
{ .next = 0, .name_len = SMB2_WSL_XATTR_NAME_LEN, .name = SMB2_WSL_XATTR_DEV, },
fs/smb/client/smb2inode.c
111
u32 outlen, next;
fs/smb/client/smb2inode.c
156
next = le32_to_cpu(ea->next_entry_offset);
fs/smb/client/smb2inode.c
157
if (!next)
fs/smb/client/smb2inode.c
159
if (!IS_ALIGNED(next, 4) ||
fs/smb/client/smb2inode.c
160
check_add_overflow((unsigned long)ea, next, &addr))
fs/smb/client/smb2inode.c
92
__le32 next;
fs/smb/client/smb2ops.c
3964
u64 next;
fs/smb/client/smb2ops.c
4018
next = le64_to_cpu(out_data[num - 1].file_offset) +
fs/smb/client/smb2ops.c
4020
len = len - (next - start);
fs/smb/client/smb2ops.c
4021
start = next;
fs/smb/client/smb2ops.c
4523
for (; buffer; buffer = buffer->next) {
fs/smb/client/smb2ops.c
4680
for (; folioq; folioq = folioq->next) {
fs/smb/client/smb2ops.c
633
size_t next = 0;
fs/smb/client/smb2ops.c
771
next = le32_to_cpu(p->Next);
fs/smb/client/smb2ops.c
772
if (!next) {
fs/smb/client/smb2ops.c
777
if (next > bytes_left) {
fs/smb/client/smb2ops.c
779
__func__, next, bytes_left);
fs/smb/client/smb2ops.c
783
p = (struct network_interface_info_ioctl_rsp *)((u8 *)p+next);
fs/smb/client/smb2ops.c
784
bytes_left -= next;
fs/smb/client/smbdirect.c
1243
request->wr.next = NULL;
fs/smb/client/smbdirect.c
1257
last->wr.next = &request->wr;
fs/smb/client/smbdirect.c
1321
struct smbdirect_send_io *sibling, *next;
fs/smb/client/smbdirect.c
1323
list_for_each_entry_safe(sibling, next, &last->sibling_list, sibling_list) {
fs/smb/client/smbdirect.c
1641
recv_wr.next = NULL;
fs/smb/client/smbdirect.c
3259
folioq = folioq->next;
fs/smb/client/smbdirect.c
3284
if (!folioq->next) {
fs/smb/client/smbdirect.c
3288
folioq = folioq->next;
fs/smb/client/smbdirect.c
566
struct smbdirect_send_io *sibling, *next;
fs/smb/client/smbdirect.c
601
list_for_each_entry_safe(sibling, next, &request->sibling_list, sibling_list) {
fs/smb/server/crypto_ctx.c
106
ctx = list_entry(ctx_list.idle_ctx.next,
fs/smb/server/crypto_ctx.c
212
ctx = list_entry(ctx_list.idle_ctx.next,
fs/smb/server/mgmt/share_config.c
41
p = list_entry(share->veto_list.next,
fs/smb/server/oplock.c
1394
goto next;
fs/smb/server/oplock.c
1399
goto next;
fs/smb/server/oplock.c
1406
goto next;
fs/smb/server/oplock.c
1413
goto next;
fs/smb/server/oplock.c
1416
next:
fs/smb/server/oplock.c
1578
unsigned int next = 0;
fs/smb/server/oplock.c
1592
cc = (struct create_context *)((char *)cc + next);
fs/smb/server/oplock.c
1596
next = le32_to_cpu(cc->Next);
fs/smb/server/oplock.c
1601
cc_len = next ? next : remain_len;
fs/smb/server/oplock.c
1603
if ((next & 0x7) != 0 ||
fs/smb/server/oplock.c
1604
next > remain_len ||
fs/smb/server/oplock.c
1617
remain_len -= next;
fs/smb/server/oplock.c
1618
} while (next != 0);
fs/smb/server/smb2pdu.c
2388
unsigned int next = 0;
fs/smb/server/smb2pdu.c
2400
goto next;
fs/smb/server/smb2pdu.c
2456
next:
fs/smb/server/smb2pdu.c
2457
next = le32_to_cpu(eabuf->NextEntryOffset);
fs/smb/server/smb2pdu.c
2458
if (next == 0 || buf_len < next)
fs/smb/server/smb2pdu.c
2460
buf_len -= next;
fs/smb/server/smb2pdu.c
2461
eabuf = (struct smb2_ea_info *)((char *)eabuf + next);
fs/smb/server/smb2pdu.c
2472
} while (next != 0);
fs/smb/server/smb2pdu.c
5055
int nbytes = 0, streamlen, stream_name_len, next, idx = 0;
fs/smb/server/smb2pdu.c
5106
next = sizeof(struct smb2_file_stream_info) + streamlen * 2;
fs/smb/server/smb2pdu.c
5107
if (next > buf_free_len) {
fs/smb/server/smb2pdu.c
5122
nbytes += next;
fs/smb/server/smb2pdu.c
5123
buf_free_len -= next;
fs/smb/server/smb2pdu.c
5124
file_info->NextEntryOffset = cpu_to_le32(next);
fs/smb/server/smb_common.c
226
int i, seq_num, bcount, next;
fs/smb/server/smb_common.c
231
next = 0;
fs/smb/server/smb_common.c
235
dialect = next_dialect(dialect, &next, bcount);
fs/smb/server/smb_common.c
251
bcount -= (++next);
fs/smb/server/transport_rdma.c
1064
struct smbdirect_send_io *sendmsg, *sibling, *next;
fs/smb/server/transport_rdma.c
1103
list_for_each_entry_safe(sibling, next, &sendmsg->sibling_list, sibling_list) {
fs/smb/server/transport_rdma.c
1248
struct smbdirect_send_io *sibling, *next;
fs/smb/server/transport_rdma.c
1250
list_for_each_entry_safe(sibling, next, &last->sibling_list, sibling_list) {
fs/smb/server/transport_rdma.c
1478
msg->wr.next = NULL;
fs/smb/server/transport_rdma.c
1489
last->wr.next = &msg->wr;
fs/smb/server/transport_rdma.c
882
wr.next = NULL;
fs/smb/server/vfs.c
1163
char *next = strchrnul(filename, '/');
fs/smb/server/vfs.c
1164
size_t filename_len = next - filename;
fs/smb/server/vfs.c
1165
bool is_last = !next[0];
fs/smb/server/vfs.c
1180
next[0] = '\0';
fs/smb/server/vfs.c
1187
next[0] = '/';
fs/smb/server/xattr.h
78
int next;
fs/ubifs/commit.c
725
i = list_entry(list.next, struct idx_node, list);
fs/ubifs/debug.c
2336
for (cur = head->next; cur->next != head; cur = cur->next) {
fs/ubifs/debug.c
2342
sb = container_of(cur->next, struct ubifs_scan_node, list);
fs/ubifs/debug.c
2403
for (cur = head->next; cur->next != head; cur = cur->next) {
fs/ubifs/debug.c
2409
sb = container_of(cur->next, struct ubifs_scan_node, list);
fs/ubifs/gc.c
1011
idx_gc = list_entry(c->idx_gc.next, struct ubifs_gced_idx_leb, list);
fs/ubifs/gc.c
543
snod = list_entry(sleb->nodes.next, struct ubifs_scan_node, list);
fs/ubifs/gc.c
990
idx_gc = list_entry(c->idx_gc.next, struct ubifs_gced_idx_leb,
fs/ubifs/log.c
522
bud = list_entry(c->old_buds.next, struct ubifs_bud, list);
fs/ubifs/lprops.c
787
lprops = list_entry(c->empty_list.next, struct ubifs_lprops, list);
fs/ubifs/lprops.c
810
lprops = list_entry(c->freeable_list.next, struct ubifs_lprops, list);
fs/ubifs/lprops.c
834
lprops = list_entry(c->frdi_idx_list.next, struct ubifs_lprops, list);
fs/ubifs/recovery.c
1082
ucleb = list_entry(c->unclean_leb_list.next,
fs/ubifs/recovery.c
878
snod = list_entry(sleb->nodes.next,
fs/ubifs/replay.c
1012
snod = list_entry(sleb->nodes.next, struct ubifs_scan_node, list);
fs/ubifs/replay.c
519
struct ubifs_bud *next;
fs/ubifs/replay.c
553
next = list_entry(bud->list.next, struct ubifs_bud, list);
fs/ubifs/replay.c
554
if (!list_is_last(&next->list, &jh->buds_list))
fs/ubifs/replay.c
557
err = ubifs_leb_read(c, next->lnum, (char *)&data, next->start, 4, 1);
fs/ubifs/replay.c
874
b = list_entry(c->replay_buds.next, struct bud_entry, list);
fs/ubifs/scan.c
361
node = list_entry(head->next, struct ubifs_scan_node, list);
fs/ubifs/shrinker.c
156
p = ubifs_infos.next;
fs/ubifs/shrinker.c
168
p = p->next;
fs/ubifs/shrinker.c
178
p = p->next;
fs/ubifs/shrinker.c
191
p = p->next;
fs/ubifs/super.c
1126
ucleb = list_entry(c->unclean_leb_list.next,
fs/ubifs/super.c
1134
bud = list_entry(c->old_buds.next, struct ubifs_bud, list);
fs/ubifs/super.c
903
orph = list_entry(c->orph_list.next, struct ubifs_orphan, list);
fs/udf/inode.c
1053
int next = laarr[start].extLocation.logicalBlockNum +
fs/udf/inode.c
1059
next, (UDF_DEFAULT_PREALLOC_BLOCKS > length ?
fs/udf/inode.c
1071
laarr[c + 1].extLocation.logicalBlockNum = next;
fs/udf/unicode.c
54
unicode_t next;
fs/udf/unicode.c
69
next = str_i[str_i_idx++] << 8;
fs/udf/unicode.c
70
next |= str_i[str_i_idx++];
fs/udf/unicode.c
71
if ((next & SURROGATE_MASK) != SURROGATE_PAIR ||
fs/udf/unicode.c
72
!(next & SURROGATE_LOW)) {
fs/udf/unicode.c
79
(next & SURROGATE_CHAR_MASK);
fs/unicode/mkutf8data.c
1050
struct tree *next;
fs/unicode/mkutf8data.c
1094
next = tree->next;
fs/unicode/mkutf8data.c
1096
assert(next);
fs/unicode/mkutf8data.c
1097
n = next->root;
fs/unicode/mkutf8data.c
1116
next = next->next;
fs/unicode/mkutf8data.c
1650
trees[trees_count-2].next = &trees[trees_count-1];
fs/unicode/mkutf8data.c
1654
trees[i].next = &trees[trees_count-2];
fs/unicode/mkutf8data.c
1656
trees[i+1].next = &trees[trees_count-1];
fs/unicode/mkutf8data.c
350
struct tree *next;
fs/verity/enable.c
25
struct block_buffer *next = cur + 1;
fs/verity/enable.c
32
if (WARN_ON_ONCE(next->is_root_hash && next->filled != 0))
fs/verity/enable.c
38
fsverity_hash_block(params, cur->data, &next->data[next->filled]);
fs/verity/enable.c
39
next->filled += params->digest_size;
fs/xfs/libxfs/xfs_attr_leaf.c
950
void *next, *end;
fs/xfs/libxfs/xfs_attr_leaf.c
970
next = xfs_attr_sf_nextentry(sfe);
fs/xfs/libxfs/xfs_attr_leaf.c
972
if (next < end)
fs/xfs/libxfs/xfs_attr_leaf.c
973
memmove(sfe, next, end - next);
fs/xfs/libxfs/xfs_bmap.c
5782
struct xfs_bmbt_irec got, next;
fs/xfs/libxfs/xfs_bmap.c
5832
if (xfs_iext_peek_next_extent(ifp, &icur, &next)) {
fs/xfs/libxfs/xfs_bmap.c
5833
if (new_startoff + got.br_blockcount > next.br_startoff) {
fs/xfs/libxfs/xfs_bmap.c
5844
if (xfs_bmse_can_merge(ip, whichfork, &got, &next,
fs/xfs/libxfs/xfs_fs.h
972
void *next = ((char *)gpr + gpr->gpr_reclen);
fs/xfs/libxfs/xfs_fs.h
975
if (next >= end)
fs/xfs/libxfs/xfs_fs.h
978
return (struct xfs_getparents_rec *)next;
fs/xfs/libxfs/xfs_ialloc.c
2042
goto next;
fs/xfs/libxfs/xfs_ialloc.c
2067
next:
fs/xfs/libxfs/xfs_iext_tree.c
125
struct xfs_iext_leaf *next;
fs/xfs/libxfs/xfs_iext_tree.c
241
cur->leaf->next) {
fs/xfs/libxfs/xfs_iext_tree.c
242
cur->leaf = cur->leaf->next;
fs/xfs/libxfs/xfs_iext_tree.c
578
if (leaf->next)
fs/xfs/libxfs/xfs_iext_tree.c
579
leaf->next->prev = new;
fs/xfs/libxfs/xfs_iext_tree.c
580
new->next = leaf->next;
fs/xfs/libxfs/xfs_iext_tree.c
582
leaf->next = new;
fs/xfs/libxfs/xfs_iext_tree.c
717
struct xfs_iext_node *next = parent->ptrs[*pos + 1];
fs/xfs/libxfs/xfs_iext_tree.c
718
int nr_next = xfs_iext_node_nr_entries(next, 0), i;
fs/xfs/libxfs/xfs_iext_tree.c
727
node->keys[nr_entries + i] = next->keys[i];
fs/xfs/libxfs/xfs_iext_tree.c
728
node->ptrs[nr_entries + i] = next->ptrs[i];
fs/xfs/libxfs/xfs_iext_tree.c
732
return next;
fs/xfs/libxfs/xfs_iext_tree.c
835
if (leaf->next) {
fs/xfs/libxfs/xfs_iext_tree.c
836
int nr_next = xfs_iext_leaf_nr_entries(ifp, leaf->next, 0), i;
fs/xfs/libxfs/xfs_iext_tree.c
846
leaf->next->recs[i];
fs/xfs/libxfs/xfs_iext_tree.c
849
if (cur->leaf == leaf->next) {
fs/xfs/libxfs/xfs_iext_tree.c
854
offset = xfs_iext_leaf_key(leaf->next, 0);
fs/xfs/libxfs/xfs_iext_tree.c
855
leaf = leaf->next;
fs/xfs/libxfs/xfs_iext_tree.c
863
leaf->prev->next = leaf->next;
fs/xfs/libxfs/xfs_iext_tree.c
864
if (leaf->next)
fs/xfs/libxfs/xfs_iext_tree.c
865
leaf->next->prev = leaf->prev;
fs/xfs/libxfs/xfs_iext_tree.c
908
if (ifp->if_height > 1 && leaf->next)
fs/xfs/libxfs/xfs_iext_tree.c
909
cur->leaf = leaf->next;
fs/xfs/libxfs/xfs_iext_tree.c
961
if (ifp->if_height == 1 || !cur->leaf->next)
fs/xfs/libxfs/xfs_iext_tree.c
963
cur->leaf = cur->leaf->next;
fs/xfs/libxfs/xfs_rtbitmap.c
643
unsigned int next)
fs/xfs/libxfs/xfs_rtbitmap.c
649
last = ((void *)xfs_rbmblock_wordptr(args, next) - 1) - bp->b_addr;
fs/xfs/scrub/attr.c
579
struct xfs_attr_sf_entry *next;
fs/xfs/scrub/attr.c
599
next = xfs_attr_sf_nextentry(sfe);
fs/xfs/scrub/attr.c
600
if ((unsigned char *)next > end) {
fs/xfs/scrub/attr.c
635
sfe = next;
fs/xfs/scrub/attr_repair.c
459
struct xfs_attr_sf_entry *next;
fs/xfs/scrub/attr_repair.c
480
next = xfs_attr_sf_nextentry(sfe);
fs/xfs/scrub/attr_repair.c
481
if ((unsigned char *)next > end)
fs/xfs/scrub/attr_repair.c
496
sfe = next;
fs/xfs/scrub/cow_repair.c
184
goto next;
fs/xfs/scrub/cow_repair.c
192
next:
fs/xfs/scrub/dir_repair.c
519
struct xfs_dir2_sf_entry *next;
fs/xfs/scrub/dir_repair.c
537
next = xfs_dir2_sf_nextentry(rd->sc->mp, hdr, sfep);
fs/xfs/scrub/dir_repair.c
538
if ((unsigned char *)next > end)
fs/xfs/scrub/dir_repair.c
546
sfep = next;
fs/xfs/scrub/iscan.c
175
int next = xfs_lowbit64(allocmask);
fs/xfs/scrub/iscan.c
177
ASSERT(next >= 0);
fs/xfs/scrub/iscan.c
178
*cursor = rec.ir_startino + next;
fs/xfs/scrub/iscan.c
179
*allocmaskp = allocmask >> next;
fs/xfs/scrub/iscan.c
180
*nr_inodesp = XFS_INODES_PER_CHUNK - next;
fs/xfs/xfs_buf_item_recover.c
517
goto next;
fs/xfs/xfs_buf_item_recover.c
523
goto next;
fs/xfs/xfs_buf_item_recover.c
530
goto next;
fs/xfs/xfs_buf_item_recover.c
538
next:
fs/xfs/xfs_extent_busy.c
565
struct xfs_extent_busy *busyp, *next;
fs/xfs/xfs_extent_busy.c
578
next = list_next_entry(busyp, list);
fs/xfs/xfs_extent_busy.c
581
busyp = next;
fs/xfs/xfs_healthmon.c
278
event->next = hm->first_event;
fs/xfs/xfs_healthmon.c
303
hm->last_event->next = event;
fs/xfs/xfs_healthmon.c
305
event->next = NULL;
fs/xfs/xfs_healthmon.c
901
hm->first_event = event->next;
fs/xfs/xfs_healthmon.c
91
struct xfs_healthmon_event *next = hm->first_event;
fs/xfs/xfs_healthmon.c
93
while ((event = next) != NULL) {
fs/xfs/xfs_healthmon.c
95
next = event->next;
fs/xfs/xfs_healthmon.h
113
struct xfs_healthmon_event *next;
fs/xfs/xfs_log_cil.c
1782
struct xfs_log_item *lip, *ilip, *next;
fs/xfs/xfs_log_cil.c
1785
list_for_each_entry_safe(lip, next, &tp->t_items, li_trans) {
fs/xfs/xfs_log_cil.c
1825
struct xfs_log_item *lip, *next;
fs/xfs/xfs_log_cil.c
1862
list_for_each_entry_safe(lip, next, &tp->t_items, li_trans) {
fs/xfs/xfs_log_recover.c
2034
struct xlog_recover_item *next;
fs/xfs/xfs_log_recover.c
2046
list_for_each_entry_safe(item, next, &trans->r_itemq, ri_list) {
fs/xfs/xfs_log_recover.c
2516
struct xfs_defer_capture *dfc, *next;
fs/xfs/xfs_log_recover.c
2520
list_for_each_entry_safe(dfc, next, capture_list, dfc_list) {
fs/xfs/xfs_log_recover.c
2564
struct xfs_defer_capture *next;
fs/xfs/xfs_log_recover.c
2566
list_for_each_entry_safe(dfc, next, capture_list, dfc_list) {
fs/xfs/xfs_mru_cache.c
232
struct xfs_mru_cache_elem *elem, *next;
fs/xfs/xfs_mru_cache.c
235
list_for_each_entry_safe(elem, next, &mru->reap_list, list_node) {
fs/xfs/xfs_mru_cache.c
248
list_for_each_entry_safe(elem, next, &tmp, list_node) {
fs/xfs/xfs_mru_cache.c
269
unsigned long now, next;
fs/xfs/xfs_mru_cache.c
276
next = _xfs_mru_cache_migrate(mru, jiffies);
fs/xfs/xfs_mru_cache.c
279
mru->queued = next;
fs/xfs/xfs_mru_cache.c
282
if (next <= now)
fs/xfs/xfs_mru_cache.c
283
next = 0;
fs/xfs/xfs_mru_cache.c
285
next -= now;
fs/xfs/xfs_mru_cache.c
286
queue_delayed_work(xfs_mru_reap_wq, &mru->work, next);
fs/xfs/xfs_reflink.c
1758
goto next;
fs/xfs/xfs_reflink.c
1774
next:
fs/xfs/xfs_rtalloc.c
257
xfs_rtxnum_t next; /* next rtext to try */
fs/xfs/xfs_rtalloc.c
279
error = xfs_rtcheck_range(args, i, scanlen, 1, &next, &stat);
fs/xfs/xfs_rtalloc.c
300
thislen = next - i;
fs/xfs/xfs_rtalloc.c
309
if (next >= end)
fs/xfs/xfs_rtalloc.c
311
error = xfs_rtfind_forw(args, next, end, &i);
fs/xfs/xfs_rtalloc.c
336
*nextp = next;
fs/xfs/xfs_rtalloc.c
356
xfs_rtxnum_t next; /* next rtext to try (dummy) */
fs/xfs/xfs_rtalloc.c
371
error = xfs_rtcheck_range(args, start, scanlen, 1, &next, &isfree);
fs/xfs/xfs_rtalloc.c
385
alloclen = next - start;
fs/xfs/xfs_trans.c
709
struct xfs_log_item *lip, *next;
fs/xfs/xfs_trans.c
713
list_for_each_entry_safe(lip, next, &tp->t_items, li_trans) {
fs/xfs/xfs_trans_ail.c
56
next_lip = list_entry(lip->li_ail.next, struct xfs_log_item, li_ail);
fs/xfs/xfs_trans_ail.c
99
if (lip->li_ail.next == &ailp->ail_head)
fs/xfs/xfs_zone_gc.c
1035
struct xfs_gc_bio *chunk, *next;
fs/xfs/xfs_zone_gc.c
1049
list_for_each_entry_safe(chunk, next, &data->resetting, entry) {
fs/xfs/xfs_zone_gc.c
1056
list_for_each_entry_safe(chunk, next, &data->writing, entry) {
fs/xfs/xfs_zone_gc.c
1064
list_for_each_entry_safe(chunk, next, &data->reading, entry) {
fs/xfs/xfs_zone_gc.c
966
struct xfs_group *next = reset_list;
fs/xfs/xfs_zone_gc.c
974
struct xfs_rtgroup *rtg = to_rtg(next);
fs/xfs/xfs_zone_gc.c
980
next = rtg_group(rtg)->xg_next_reset;
fs/xfs/xfs_zone_gc.c
993
} while (next);
fs/zonefs/super.c
941
struct blk_zone *zone, *next, *end;
fs/zonefs/super.c
960
for (zone = &zd->zones[1]; zone < end; zone = next) {
fs/zonefs/super.c
962
next = zone + 1;
fs/zonefs/super.c
979
for (; next < end; next++) {
fs/zonefs/super.c
980
if (zonefs_zone_type(next) != ztype)
fs/zonefs/super.c
982
zone->len += next->len;
fs/zonefs/super.c
983
zone->capacity += next->capacity;
fs/zonefs/super.c
984
if (next->cond == BLK_ZONE_COND_READONLY &&
fs/zonefs/super.c
987
else if (next->cond == BLK_ZONE_COND_OFFLINE)
include/asm-generic/mcs_spinlock.h
5
struct mcs_spinlock *next;
include/asm-generic/nommu_context.h
12
struct mm_struct *next,
include/asm-generic/switch_to.h
21
#define switch_to(prev, next, last) \
include/asm-generic/switch_to.h
23
((last) = __switch_to((prev), (next))); \
include/asm-generic/tlb.h
273
struct mmu_gather_batch *next;
include/crypto/algapi.h
94
struct crypto_spawn *next;
include/drm/display/drm_dp_mst_helper.h
125
struct list_head next;
include/drm/display/drm_dp_mst_helper.h
504
struct list_head next;
include/drm/display/drm_dp_mst_helper.h
585
struct list_head next;
include/drm/drm_bridge.h
1471
struct drm_bridge *next = drm_bridge_get_next_bridge(bridge);
include/drm/drm_bridge.h
1475
return next;
include/drm/drm_colorop.h
266
struct drm_colorop *next;
include/drm/drm_colorop.h
462
void drm_colorop_set_next_property(struct drm_colorop *colorop, struct drm_colorop *next);
include/drm/drm_gpuvm.h
1042
#define drm_gpuva_for_each_op_safe(op, next, ops) \
include/drm/drm_gpuvm.h
1043
list_for_each_entry_safe(op, next, &(ops)->list, entry)
include/drm/drm_gpuvm.h
1271
struct drm_gpuva *next,
include/drm/drm_gpuvm.h
1294
const u64 va_end = op->next ?
include/drm/drm_gpuvm.h
1295
op->next->va.addr :
include/drm/drm_gpuvm.h
949
struct drm_gpuva_op_map *next;
include/drm/drm_mm.h
278
return READ_ONCE(mm->hole_stack.next);
include/drm/drm_mm.h
375
#define drm_mm_for_each_node_safe(entry, next, mm) \
include/drm/drm_mm.h
376
list_for_each_entry_safe(entry, next, drm_mm_nodes(mm), node_list)
include/drm/spsc_queue.h
103
next = READ_ONCE(node->next);
include/drm/spsc_queue.h
104
WRITE_ONCE(queue->head, next);
include/drm/spsc_queue.h
106
if (unlikely(!next)) {
include/drm/spsc_queue.h
110
(long)&node->next, (long) &queue->head) != (long)&node->next) {
include/drm/spsc_queue.h
114
} while (unlikely(!(queue->head = READ_ONCE(node->next))));
include/drm/spsc_queue.h
35
struct spsc_node *next;
include/drm/spsc_queue.h
69
node->next = NULL;
include/drm/spsc_queue.h
76
tail = (struct spsc_node **)atomic_long_xchg(&queue->tail, (long)&node->next);
include/drm/spsc_queue.h
93
struct spsc_node *next, *node;
include/keys/rxrpc-type.h
56
struct rxrpc_key_token *next; /* the next token in the list */
include/linux/adb.h
21
struct adb_request *next;
include/linux/agp_backend.h
70
struct agp_memory *next;
include/linux/agpgart.h
113
struct agp_file_private *next;
include/linux/agpgart.h
90
struct agp_client *next;
include/linux/agpgart.h
98
struct agp_controller *next;
include/linux/atalk.h
14
struct atalk_route *next;
include/linux/atalk.h
32
struct atalk_iface *next;
include/linux/bio.h
321
u8 bio_seg_gap(struct request_queue *q, struct bio *prev, struct bio *next,
include/linux/bootconfig.h
52
uint16_t next;
include/linux/bpf_verifier.h
721
struct bpf_scc_backedge *next;
include/linux/comedi/comedidev.h
437
struct comedi_driver *next; /* Next in list of COMEDI drivers. */
include/linux/compat.h
389
compat_uptr_t next;
include/linux/cookie.h
39
s64 next = atomic64_add_return(COOKIE_LOCAL_BATCH,
include/linux/cookie.h
41
val = next - COOKIE_LOCAL_BATCH;
include/linux/crash_dump.h
123
struct list_head next;
include/linux/damon.h
835
return container_of(r->list.next, struct damon_region, list);
include/linux/damon.h
865
#define damon_for_each_region_safe(r, next, t) \
include/linux/damon.h
866
list_for_each_entry_safe(r, next, &t->regions_list, list)
include/linux/damon.h
871
#define damon_for_each_target_safe(t, next, ctx) \
include/linux/damon.h
872
list_for_each_entry_safe(t, next, &(ctx)->adaptive_targets, list)
include/linux/damon.h
877
#define damon_for_each_scheme_safe(s, next, ctx) \
include/linux/damon.h
878
list_for_each_entry_safe(s, next, &(ctx)->schemes, list)
include/linux/damon.h
883
#define damos_for_each_quota_goal_safe(goal, next, quota) \
include/linux/damon.h
884
list_for_each_entry_safe(goal, next, &(quota)->goals, list)
include/linux/damon.h
889
#define damos_for_each_core_filter_safe(f, next, scheme) \
include/linux/damon.h
890
list_for_each_entry_safe(f, next, &(scheme)->core_filters, list)
include/linux/damon.h
895
#define damos_for_each_ops_filter_safe(f, next, scheme) \
include/linux/damon.h
896
list_for_each_entry_safe(f, next, &(scheme)->ops_filters, list)
include/linux/damon.h
906
struct damon_region *prev, struct damon_region *next,
include/linux/damon.h
909
__list_add(&r->list, &prev->list, &next->list);
include/linux/dcache.h
611
return hlist_entry_safe(dentry->d_sib.next, struct dentry, d_sib);
include/linux/debugobjects.h
87
unsigned int expect, unsigned int next);
include/linux/dm-io.h
26
struct page_list *next;
include/linux/dmaengine.h
628
struct dma_async_tx_descriptor *next;
include/linux/dmaengine.h
676
static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *next)
include/linux/dmaengine.h
704
static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *next)
include/linux/dmaengine.h
706
txd->next = next;
include/linux/dmaengine.h
707
next->parent = txd;
include/linux/dmaengine.h
715
txd->next = NULL;
include/linux/dmaengine.h
723
return txd->next;
include/linux/efi.h
1283
phys_addr_t next; // pa of next struct instance
include/linux/folio_queue.h
33
struct folio_queue *next; /* Next queue segment or NULL */
include/linux/folio_queue.h
56
folioq->next = NULL;
include/linux/fs.h
2288
struct file_system_type * next;
include/linux/ftrace.h
449
struct ftrace_ops __rcu *next;
include/linux/ftrace.h
489
while (likely(op = rcu_dereference_raw_check((op)->next)) && \
include/linux/hdlc.h
30
struct hdlc_proto *next; /* next protocol in the list */
include/linux/hfs_common.h
465
__be32 next; /* (V) Number of the next node at this level */
include/linux/huge_mm.h
29
pmd_t *pmd, unsigned long addr, unsigned long next);
include/linux/huge_mm.h
488
unsigned long end, struct vm_area_struct *next);
include/linux/huge_mm.h
691
struct vm_area_struct *next)
include/linux/if_pppox.h
46
struct pppox_sock __rcu *next; /* for hash table */
include/linux/if_tap.h
72
struct list_head next;
include/linux/if_tunnel.h
15
for (pos = rcu_dereference(start); pos; pos = rcu_dereference(pos->next))
include/linux/if_vlan.h
154
struct vlan_priority_tci_mapping *next;
include/linux/if_vlan.h
222
mp = mp->next;
include/linux/igmp.h
76
struct ip_mc_list *next;
include/linux/interrupt.h
130
struct irqaction *next;
include/linux/interrupt.h
693
struct tasklet_struct *next;
include/linux/io_uring_types.h
45
struct io_wq_work_node *next;
include/linux/iommu.h
248
struct list_head next;
include/linux/iov_iter.h
157
folioq = folioq->next;
include/linux/iov_iter.h
184
if (slot == folioq_nr_slots(folioq) && folioq->next) {
include/linux/iov_iter.h
185
folioq = folioq->next;
include/linux/jump_label.h
105
struct static_key_mod *next;
include/linux/kho/abi/kexec_handover.h
131
DECLARE_KHOSER_PTR(next, struct kho_vmalloc_chunk *);
include/linux/list.h
1007
n->next = LIST_POISON1;
include/linux/list.h
1036
WRITE_ONCE(n->next, first);
include/linux/list.h
1038
WRITE_ONCE(first->pprev, &n->next);
include/linux/list.h
1049
struct hlist_node *next)
include/linux/list.h
1051
WRITE_ONCE(n->pprev, next->pprev);
include/linux/list.h
1052
WRITE_ONCE(n->next, next);
include/linux/list.h
1053
WRITE_ONCE(next->pprev, &n->next);
include/linux/list.h
1065
WRITE_ONCE(n->next, prev->next);
include/linux/list.h
1066
WRITE_ONCE(prev->next, n);
include/linux/list.h
1067
WRITE_ONCE(n->pprev, &prev->next);
include/linux/list.h
1069
if (n->next)
include/linux/list.h
1070
WRITE_ONCE(n->next->pprev, &n->next);
include/linux/list.h
1083
n->pprev = &n->next;
include/linux/list.h
1092
return h->pprev == &h->next;
include/linux/list.h
1106
return !n->next && n->pprev == &h->first;
include/linux/list.h
1139
to->first->pprev = &last->next;
include/linux/list.h
1140
last->next = to->first;
include/linux/list.h
1149
for (pos = (head)->first; pos ; pos = pos->next)
include/linux/list.h
1152
for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \
include/linux/list.h
1169
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
include/linux/list.h
1177
for (pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member);\
include/linux/list.h
1179
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
include/linux/list.h
1188
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
include/linux/list.h
1199
pos && ({ n = pos->member.next; 1; }); \
include/linux/list.h
120
struct list_head *next = entry->next;
include/linux/list.h
127
if (likely(prev->next == entry && next->prev == entry))
include/linux/list.h
138
struct list_head *next)
include/linux/list.h
156
struct list_head *next)
include/linux/list.h
158
if (!__list_add_valid(new, prev, next))
include/linux/list.h
161
next->prev = new;
include/linux/list.h
162
new->next = next;
include/linux/list.h
164
WRITE_ONCE(prev->next, new);
include/linux/list.h
177
__list_add(new, head, head->next);
include/linux/list.h
201
static inline void __list_del(struct list_head * prev, struct list_head * next)
include/linux/list.h
203
next->prev = prev;
include/linux/list.h
204
WRITE_ONCE(prev->next, next);
include/linux/list.h
217
__list_del(entry->prev, entry->next);
include/linux/list.h
226
__list_del(entry->prev, entry->next);
include/linux/list.h
238
entry->next = LIST_POISON1;
include/linux/list.h
252
new->next = old->next;
include/linux/list.h
253
new->next->prev = new;
include/linux/list.h
255
new->prev->next = new;
include/linux/list.h
335
first->prev->next = last->next;
include/linux/list.h
336
last->next->prev = first->prev;
include/linux/list.h
338
head->prev->next = first;
include/linux/list.h
341
last->next = head;
include/linux/list.h
362
return list->next == head;
include/linux/list.h
381
return READ_ONCE(head->next) == head;
include/linux/list.h
399
smp_store_release(&entry->next, entry);
include/linux/list.h
417
struct list_head *next = smp_load_acquire(&head->next);
include/linux/list.h
418
return list_is_head(next, head) && (next == READ_ONCE(head->prev));
include/linux/list.h
430
first = head->next;
include/linux/list.h
45
WRITE_ONCE(list->next, list);
include/linux/list.h
459
return !list_empty(head) && (head->next == head->prev);
include/linux/list.h
465
struct list_head *new_first = entry->next;
include/linux/list.h
466
list->next = head->next;
include/linux/list.h
467
list->next->prev = list;
include/linux/list.h
469
entry->next = list;
include/linux/list.h
470
head->next = new_first;
include/linux/list.h
493
if (list_is_singular(head) && !list_is_head(entry, head) && (entry != head->next))
include/linux/list.h
519
if (head->next == entry) {
include/linux/list.h
523
list->next = head->next;
include/linux/list.h
524
list->next->prev = list;
include/linux/list.h
526
list->prev->next = list;
include/linux/list.h
527
head->next = entry;
include/linux/list.h
533
struct list_head *next)
include/linux/list.h
535
struct list_head *first = list->next;
include/linux/list.h
539
prev->next = first;
include/linux/list.h
541
last->next = next;
include/linux/list.h
542
next->prev = last;
include/linux/list.h
554
__list_splice(list, head, head->next);
include/linux/list.h
580
__list_splice(list, head, head->next);
include/linux/list.h
620
list_entry((ptr)->next, type, member)
include/linux/list.h
63
struct list_head *next);
include/linux/list.h
643
struct list_head *pos__ = READ_ONCE(head__->next); \
include/linux/list.h
667
list_entry((pos)->member.next, typeof(*(pos)), member)
include/linux/list.h
709
for (pos = (head)->next; !list_is_head(pos, (head)); pos = pos->next)
include/linux/list.h
719
for (pos = pos->next; !list_is_head(pos, (head)); pos = pos->next)
include/linux/list.h
736
for (pos = (head)->next, n = pos->next; \
include/linux/list.h
738
pos = n, n = pos->next)
include/linux/list.h
75
struct list_head *next)
include/linux/list.h
91
if (likely(next->prev == prev && prev->next == next && new != prev && new != next))
include/linux/list.h
948
h->next = NULL;
include/linux/list.h
96
ret &= __list_add_valid_or_report(new, prev, next);
include/linux/list.h
989
struct hlist_node *next = n->next;
include/linux/list.h
992
WRITE_ONCE(*pprev, next);
include/linux/list.h
993
if (next)
include/linux/list.h
994
WRITE_ONCE(next->pprev, pprev);
include/linux/list_bl.h
107
n->next = prev->next;
include/linux/list_bl.h
108
n->pprev = &prev->next;
include/linux/list_bl.h
109
prev->next = n;
include/linux/list_bl.h
111
if (n->next)
include/linux/list_bl.h
112
n->next->pprev = &n->next;
include/linux/list_bl.h
117
struct hlist_bl_node *next = n->next;
include/linux/list_bl.h
125
((unsigned long)next |
include/linux/list_bl.h
127
if (next)
include/linux/list_bl.h
128
next->pprev = pprev;
include/linux/list_bl.h
134
n->next = LIST_POISON1;
include/linux/list_bl.h
175
pos = pos->next)
include/linux/list_bl.h
187
pos && ({ n = pos->next; 1; }) && \
include/linux/list_bl.h
39
struct hlist_bl_node *next, **pprev;
include/linux/list_bl.h
46
h->next = NULL;
include/linux/list_bl.h
82
n->next = first;
include/linux/list_bl.h
84
first->pprev = &n->next;
include/linux/list_bl.h
90
struct hlist_bl_node *next)
include/linux/list_bl.h
92
struct hlist_bl_node **pprev = next->pprev;
include/linux/list_bl.h
95
n->next = next;
include/linux/list_bl.h
96
next->pprev = &n->next;
include/linux/list_nulls.h
101
WRITE_ONCE(first->pprev, &n->next);
include/linux/list_nulls.h
106
struct hlist_nulls_node *next = n->next;
include/linux/list_nulls.h
109
WRITE_ONCE(*pprev, next);
include/linux/list_nulls.h
110
if (!is_a_nulls(next))
include/linux/list_nulls.h
111
WRITE_ONCE(next->pprev, pprev);
include/linux/list_nulls.h
132
pos = pos->next)
include/linux/list_nulls.h
144
pos = pos->next)
include/linux/list_nulls.h
26
struct hlist_nulls_node *next, **pprev;
include/linux/list_nulls.h
97
n->next = first;
include/linux/list_private.h
42
list_private_entry((ptr)->next, type, member)
include/linux/list_private.h
59
list_private_entry(ACCESS_PRIVATE(pos, member).next, typeof(*(pos)), member)
include/linux/llist.h
100
return READ_ONCE(node->next) != node;
include/linux/llist.h
144
for ((pos) = (node); pos; (pos) = (pos)->next)
include/linux/llist.h
163
for ((pos) = (node); (pos) && ((n) = (pos)->next, true); (pos) = (n))
include/linux/llist.h
183
(pos) = llist_entry((pos)->member.next, typeof(*(pos)), member))
include/linux/llist.h
205
(n = llist_entry(pos->member.next, typeof(*n), member), true); \
include/linux/llist.h
223
return READ_ONCE(node->next);
include/linux/llist.h
241
new_last->next = first;
include/linux/llist.h
251
new_last->next = head->first;
include/linux/llist.h
253
return new_last->next == NULL;
include/linux/llist.h
61
struct llist_node *next;
include/linux/llist.h
86
WRITE_ONCE(node->next, node);
include/linux/lwq.h
85
_t2 = ((*_t1)->next), \
include/linux/lwq.h
88
(_n) ? (_t1 = &(_n)->_member.node.next, 0) \
include/linux/maple_tree.h
134
struct maple_enode *next; /* Overlaps the pivot */
include/linux/maple_tree.h
329
unsigned long *next, gfp_t gfp);
include/linux/maple_tree.h
525
unsigned long *next, gfp_t gfp);
include/linux/mdev.h
21
struct list_head next;
include/linux/mlx5/device.h
799
__be64 next;
include/linux/mlx5/driver.h
254
struct mlx5_cmd_mailbox *next;
include/linux/mlx5/driver.h
343
struct mlx5_cmd_mailbox *next;
include/linux/mm.h
2756
return (uintptr_t)page->lru.next & BIT(1);
include/linux/mm.h
2771
return (uintptr_t)folio->lru.next & BIT(1);
include/linux/mm.h
2780
page->lru.next = (void *)BIT(1);
include/linux/mm.h
2785
page->lru.next = NULL;
include/linux/mm.h
4502
unsigned long addr, unsigned long next);
include/linux/mm.h
4504
unsigned long addr, unsigned long next);
include/linux/mmc/sdio_func.h
25
struct sdio_func_tuple *next;
include/linux/msg.h
13
struct msg_msgseg *next;
include/linux/netdevice.h
3328
lh = dev->dev_list.next;
include/linux/netdevice.h
3345
net_device_entry(net->dev_base_head.next);
include/linux/netdevice.h
5160
for (iter = (dev)->adj_list.lower.next, \
include/linux/netdevice.h
5175
for (iter = (dev)->adj_list.lower.next, \
include/linux/nls.h
34
struct nls_table *next;
include/linux/notifier.h
56
struct notifier_block __rcu *next;
include/linux/of.h
32
struct property *next;
include/linux/of.h
959
for (pp = dn->properties; pp != NULL; pp = pp->next)
include/linux/page_ext.h
108
void *next = curr;
include/linux/page_ext.h
109
next += page_ext_size;
include/linux/page_ext.h
110
return next;
include/linux/pagewalk.h
72
unsigned long next, struct mm_walk *walk);
include/linux/pagewalk.h
74
unsigned long next, struct mm_walk *walk);
include/linux/pagewalk.h
76
unsigned long next, struct mm_walk *walk);
include/linux/pagewalk.h
78
unsigned long next, struct mm_walk *walk);
include/linux/pagewalk.h
80
unsigned long next, struct mm_walk *walk);
include/linux/pagewalk.h
81
int (*pte_hole)(unsigned long addr, unsigned long next,
include/linux/pagewalk.h
84
unsigned long addr, unsigned long next,
include/linux/pagewalk.h
86
int (*test_walk)(unsigned long addr, unsigned long next,
include/linux/pagewalk.h
91
int (*install_pte)(unsigned long addr, unsigned long next,
include/linux/parport.h
142
struct pardevice *next;
include/linux/perf_event.h
1228
struct task_struct *next);
include/linux/perf_event.h
1409
frag = frag->next;
include/linux/perf_event.h
1649
struct task_struct *next)
include/linux/perf_event.h
1657
perf_cgroup_from_task(next, NULL))
include/linux/perf_event.h
1662
__perf_event_task_sched_out(prev, next);
include/linux/perf_event.h
1960
struct task_struct *next) { }
include/linux/perf_event.h
77
struct perf_raw_frag *next;
include/linux/plist.h
275
return list_entry(head->node_list.next,
include/linux/pm_domain.h
326
void dev_pm_genpd_set_next_wakeup(struct device *dev, ktime_t next);
include/linux/pm_domain.h
400
static inline void dev_pm_genpd_set_next_wakeup(struct device *dev, ktime_t next)
include/linux/pnp.h
371
struct pnp_id *next;
include/linux/preempt.h
339
struct task_struct *next);
include/linux/preempt.h
363
notifier->link.next = NULL;
include/linux/raid/xor.h
11
struct xor_block_template *next;
include/linux/rculist.h
100
if (!__list_add_valid(new, prev, next))
include/linux/rculist.h
103
new->next = next;
include/linux/rculist.h
106
next->prev = new;
include/linux/rculist.h
127
__list_add_rcu(new, head, head->next);
include/linux/rculist.h
24
WRITE_ONCE(list->next, list);
include/linux/rculist.h
257
new->next = old->next;
include/linux/rculist.h
260
new->next->prev = new;
include/linux/rculist.h
285
struct list_head *next,
include/linux/rculist.h
288
struct list_head *first = list->next;
include/linux/rculist.h
318
last->next = next;
include/linux/rculist.h
32
#define list_next_rcu(list) (*((struct list_head __rcu **)(&(list)->next)))
include/linux/rculist.h
321
next->prev = last;
include/linux/rculist.h
336
__list_splice_init_rcu(list, head, head->next, sync);
include/linux/rculist.h
410
struct list_head *__next = READ_ONCE(__ptr->next); \
include/linux/rculist.h
430
struct list_head *__next = READ_ONCE(__ptr->next); \
include/linux/rculist.h
448
pos = list_entry_rcu((head)->next, typeof(*pos), member); \
include/linux/rculist.h
450
pos = list_entry_rcu(pos->member.next, typeof(*pos), member))
include/linux/rculist.h
467
pos = list_entry_rcu((head)->next, typeof(*pos), member); \
include/linux/rculist.h
469
pos = list_entry_rcu(pos->member.next, typeof(*pos), member))
include/linux/rculist.h
501
for (pos = list_entry_lockless((head)->next, typeof(*pos), member); \
include/linux/rculist.h
503
pos = list_entry_lockless(pos->member.next, typeof(*pos), member))
include/linux/rculist.h
51
for (pos = rcu_dereference((head)->next); \
include/linux/rculist.h
524
for (pos = list_entry_rcu(pos->member.next, typeof(*pos), member); \
include/linux/rculist.h
526
pos = list_entry_rcu(pos->member.next, typeof(*pos), member))
include/linux/rculist.h
53
pos = rcu_dereference(pos->next))
include/linux/rculist.h
547
pos = list_entry_rcu(pos->member.next, typeof(*(pos)), member))
include/linux/rculist.h
586
struct hlist_node *next = old->next;
include/linux/rculist.h
588
new->next = next;
include/linux/rculist.h
591
if (next)
include/linux/rculist.h
592
WRITE_ONCE(new->next->pprev, &new->next);
include/linux/rculist.h
621
#define hlist_next_rcu(node) (*((struct hlist_node __rcu **)(&(node)->next)))
include/linux/rculist.h
648
n->next = first;
include/linux/rculist.h
652
WRITE_ONCE(first->pprev, &n->next);
include/linux/rculist.h
680
for (i = h->first; i; i = i->next)
include/linux/rculist.h
684
n->next = last->next;
include/linux/rculist.h
685
WRITE_ONCE(n->pprev, &last->next);
include/linux/rculist.h
711
struct hlist_node *next)
include/linux/rculist.h
713
WRITE_ONCE(n->pprev, next->pprev);
include/linux/rculist.h
714
n->next = next;
include/linux/rculist.h
716
WRITE_ONCE(next->pprev, &n->next);
include/linux/rculist.h
740
n->next = prev->next;
include/linux/rculist.h
741
WRITE_ONCE(n->pprev, &prev->next);
include/linux/rculist.h
743
if (n->next)
include/linux/rculist.h
744
WRITE_ONCE(n->next->pprev, &n->next);
include/linux/rculist.h
98
struct list_head *prev, struct list_head *next)
include/linux/rculist_bl.h
79
n->next = first;
include/linux/rculist_bl.h
81
first->pprev = &n->next;
include/linux/rculist_bl.h
99
pos = rcu_dereference_raw(pos->next))
include/linux/rculist_nulls.h
111
WRITE_ONCE(n->next, first);
include/linux/rculist_nulls.h
115
WRITE_ONCE(first->pprev, &n->next);
include/linux/rculist_nulls.h
143
for (i = h->first; !is_a_nulls(i); i = i->next)
include/linux/rculist_nulls.h
147
WRITE_ONCE(n->next, last->next);
include/linux/rculist_nulls.h
148
WRITE_ONCE(n->pprev, &last->next);
include/linux/rculist_nulls.h
158
WRITE_ONCE(n->pprev, &n->next);
include/linux/rculist_nulls.h
159
WRITE_ONCE(n->next, (struct hlist_nulls_node *)NULLS_MARKER(NULL));
include/linux/rculist_nulls.h
180
struct hlist_nulls_node *next = old->next;
include/linux/rculist_nulls.h
182
WRITE_ONCE(new->next, next);
include/linux/rculist_nulls.h
185
if (!is_a_nulls(next))
include/linux/rculist_nulls.h
186
WRITE_ONCE(next->pprev, &new->next);
include/linux/rculist_nulls.h
53
(*((struct hlist_nulls_node __rcu __force **)&(node)->next))
include/linux/rhashtable-types.h
19
struct rhash_head __rcu *next;
include/linux/rhashtable-types.h
24
struct rhlist_head __rcu *next;
include/linux/rhashtable.h
1078
pprev = &he->next;
include/linux/rhashtable.h
1084
lpprev = &list->next;
include/linux/rhashtable.h
1085
list = rht_dereference_bucket(list->next,
include/linux/rhashtable.h
1092
list = rht_dereference_bucket(list->next, tbl, hash);
include/linux/rhashtable.h
1098
obj = rht_dereference_bucket(obj->next, tbl, hash);
include/linux/rhashtable.h
1102
list = rht_dereference_bucket(list->next, tbl, hash);
include/linux/rhashtable.h
1104
RCU_INIT_POINTER(list->rhead.next, obj);
include/linux/rhashtable.h
1232
pprev = &he->next;
include/linux/rhashtable.h
1236
rcu_assign_pointer(obj_new->next, obj_old->next);
include/linux/rhashtable.h
452
pos = rht_dereference_bucket((pos)->next, tbl, hash))
include/linux/rhashtable.h
476
pos = rht_dereference_bucket((pos)->next, tbl, hash))
include/linux/rhashtable.h
503
#define rht_for_each_entry_safe(tpos, pos, next, tbl, hash, member) \
include/linux/rhashtable.h
505
next = !rht_is_a_nulls(pos) ? \
include/linux/rhashtable.h
506
rht_dereference_bucket(pos->next, tbl, hash) : NULL; \
include/linux/rhashtable.h
508
pos = next, \
include/linux/rhashtable.h
509
next = !rht_is_a_nulls(pos) ? \
include/linux/rhashtable.h
510
rht_dereference_bucket(pos->next, tbl, hash) : NULL)
include/linux/rhashtable.h
527
pos = rcu_dereference_all(pos->next))
include/linux/rhashtable.h
543
pos = rcu_dereference_all(pos->next))
include/linux/rhashtable.h
562
pos = rht_dereference_bucket_rcu(pos->next, tbl, hash))
include/linux/rhashtable.h
590
for (pos = list; pos; pos = rcu_dereference_all(pos->next))
include/linux/rhashtable.h
604
pos = rcu_dereference_all(pos->next))
include/linux/rhashtable.h
806
pprev = &head->next;
include/linux/rhashtable.h
819
RCU_INIT_POINTER(list->next, plist);
include/linux/rhashtable.h
820
head = rht_dereference_bucket(head->next, tbl, hash);
include/linux/rhashtable.h
821
RCU_INIT_POINTER(list->rhead.next, head);
include/linux/rhashtable.h
844
RCU_INIT_POINTER(obj->next, head);
include/linux/rhashtable.h
849
RCU_INIT_POINTER(list->next, NULL);
include/linux/sbitmap.h
257
goto next;
include/linux/sbitmap.h
274
next:
include/linux/sched.h
810
struct wake_q_node *next;
include/linux/sched/mm.h
559
struct mm_struct *next,
include/linux/sched/signal.h
637
list_entry_rcu((p)->tasks.next, struct task_struct, tasks)
include/linux/sched/signal.h
78
struct core_thread *next;
include/linux/seq_file.h
34
void * (*next) (struct seq_file *m, void *v, loff_t *pos);
include/linux/skbuff.h
1897
skb->next = NULL;
include/linux/skbuff.h
1903
skb->next = SKB_LIST_POISON_NEXT;
include/linux/skbuff.h
1909
for ((skb) = (first), (next_skb) = (skb) ? (skb)->next : NULL; (skb); \
include/linux/skbuff.h
1910
(skb) = (next_skb), (next_skb) = (skb) ? (skb)->next : NULL)
include/linux/skbuff.h
1926
return list->next == (const struct sk_buff *) list;
include/linux/skbuff.h
1938
return READ_ONCE(list->next) == (const struct sk_buff *) list;
include/linux/skbuff.h
1952
return skb->next == (const struct sk_buff *) list;
include/linux/skbuff.h
1983
return skb->next;
include/linux/skbuff.h
2195
struct sk_buff *skb = list_->next;
include/linux/skbuff.h
2210
return list_->next;
include/linux/skbuff.h
2225
struct sk_buff *next = skb->next;
include/linux/skbuff.h
2227
if (next == (struct sk_buff *)list_)
include/linux/skbuff.h
2228
next = NULL;
include/linux/skbuff.h
2229
return next;
include/linux/skbuff.h
2290
list->prev = list->next = (struct sk_buff *)list;
include/linux/skbuff.h
2322
struct sk_buff *prev, struct sk_buff *next,
include/linux/skbuff.h
2328
WRITE_ONCE(newsk->next, next);
include/linux/skbuff.h
2330
WRITE_ONCE(((struct sk_buff_list *)next)->prev, newsk);
include/linux/skbuff.h
2331
WRITE_ONCE(((struct sk_buff_list *)prev)->next, newsk);
include/linux/skbuff.h
2337
struct sk_buff *next)
include/linux/skbuff.h
2339
struct sk_buff *first = list->next;
include/linux/skbuff.h
2343
WRITE_ONCE(prev->next, first);
include/linux/skbuff.h
2345
WRITE_ONCE(last->next, next);
include/linux/skbuff.h
2346
WRITE_ONCE(next->prev, last);
include/linux/skbuff.h
2358
__skb_queue_splice(list, (struct sk_buff *) head, head->next);
include/linux/skbuff.h
2374
__skb_queue_splice(list, (struct sk_buff *) head, head->next);
include/linux/skbuff.h
2427
__skb_insert(newsk, prev, ((struct sk_buff_list *)prev)->next, list);
include/linux/skbuff.h
2434
struct sk_buff *next,
include/linux/skbuff.h
2437
__skb_insert(newsk, ((struct sk_buff_list *)next)->prev, next, list);
include/linux/skbuff.h
2481
struct sk_buff *next, *prev;
include/linux/skbuff.h
2484
next = skb->next;
include/linux/skbuff.h
2486
skb->next = skb->prev = NULL;
include/linux/skbuff.h
2487
WRITE_ONCE(next->prev, prev);
include/linux/skbuff.h
2488
WRITE_ONCE(prev->next, next);
include/linux/skbuff.h
340
struct sk_buff *next;
include/linux/skbuff.h
4134
for (skb = (queue)->next; \
include/linux/skbuff.h
4136
skb = skb->next)
include/linux/skbuff.h
4139
for (skb = (queue)->next, tmp = skb->next; \
include/linux/skbuff.h
4141
skb = tmp, tmp = skb->next)
include/linux/skbuff.h
4145
skb = skb->next)
include/linux/skbuff.h
4160
for (tmp = skb->next; \
include/linux/skbuff.h
4162
skb = tmp, tmp = skb->next)
include/linux/skbuff.h
4190
for (iter = skb_shinfo(skb)->frag_list; iter; iter = iter->next)
include/linux/skbuff.h
889
struct sk_buff *next;
include/linux/soc/mediatek/mtk-mmsys.h
89
enum mtk_ddp_comp_id next);
include/linux/soc/mediatek/mtk-mmsys.h
93
enum mtk_ddp_comp_id next);
include/linux/static_call.h
170
struct static_call_mod *next;
include/linux/suspend.h
313
struct pbe *next;
include/linux/swap.h
255
unsigned int next[SWAP_NR_ORDERS]; /* Likely next allocation offset */
include/linux/timer.h
57
.entry = { .next = TIMER_ENTRY_STATIC }, \
include/linux/trace.h
24
struct trace_export __rcu *next;
include/linux/tty_buffer.h
12
struct tty_buffer *next;
include/linux/types.h
205
struct list_head *next, *prev;
include/linux/types.h
213
struct hlist_node *next, **pprev;
include/linux/types.h
247
struct callback_head *next;
include/linux/uprobes.h
167
struct return_instance *next; /* keep as stack */
include/linux/user-return-notifier.h
20
struct task_struct *next)
include/linux/user-return-notifier.h
24
set_tsk_thread_flag(next, TIF_USER_RETURN_NOTIFY);
include/linux/user-return-notifier.h
40
struct task_struct *next)
include/linux/user_events.h
24
struct user_event_mm *next;
include/linux/vmalloc.h
54
struct vm_struct *next; /* Early registration of vm_areas. */
include/linux/xarray.h
1016
struct xa_limit limit, u32 *next, gfp_t gfp)
include/linux/xarray.h
1022
err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
include/linux/xarray.h
1055
struct xa_limit limit, u32 *next, gfp_t gfp)
include/linux/xarray.h
1061
err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
include/linux/xarray.h
571
struct xa_limit, u32 *next, gfp_t);
include/linux/xarray.h
977
struct xa_limit limit, u32 *next, gfp_t gfp)
include/linux/xarray.h
983
err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
include/media/dmxdev.h
86
struct list_head next;
include/media/drv-intf/saa7146_vv.h
154
void saa7146_set_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next);
include/media/drv-intf/saa7146_vv.h
66
struct saa7146_buf *next);
include/media/dvb_demux.h
84
struct dvb_demux_filter *next;
include/media/v4l2-ctrls.h
367
struct v4l2_ctrl_ref *next;
include/net/atmclip.h
34
struct clip_vcc *next; /* next VCC */
include/net/atmclip.h
50
struct net_device *next; /* next CLIP interface */
include/net/ax25.h
187
struct ax25_route *next;
include/net/ax25.h
369
struct ax25_protocol *next;
include/net/bluetooth/bluetooth.h
612
frag = &(*frag)->next;
include/net/bond_alb.h
70
u32 next; /* The next Hash table entry index, assigned
include/net/bonding.h
61
netdev_adjacent_get_private(bond_slave_list(bond)->next) : \
include/net/fq_impl.h
203
struct sk_buff *next;
include/net/fq_impl.h
218
skb_list_walk_safe(skb, skb, next) {
include/net/gtp.h
46
__u8 next;
include/net/if_inet6.h
121
struct ifmcaddr6 __rcu *next;
include/net/if_inet6.h
97
struct ipv6_mc_socklist __rcu *next;
include/net/ip.h
132
struct ip_ra_chain __rcu *next;
include/net/ip.h
192
iter->frag = skb->next;
include/net/ip6_tunnel.h
47
struct ip6_tnl __rcu *next; /* next tunnel in list */
include/net/ip_tunnels.h
134
struct ip_tunnel_prl_entry __rcu *next;
include/net/ip_tunnels.h
154
struct ip_tunnel __rcu *next;
include/net/ip_vs.h
477
struct ip_vs_protocol *next;
include/net/ip_vs.h
537
struct ip_vs_proto_data *next;
include/net/ipv6.h
173
iter->frag = skb->next;
include/net/ipv6.h
277
struct ip6_ra_chain *next;
include/net/ipv6.h
317
struct ip6_flowlabel __rcu *next;
include/net/ipv6.h
342
struct ipv6_fl_socklist __rcu *next;
include/net/macsec.h
160
struct macsec_rx_sc __rcu *next;
include/net/neighbour.h
181
struct pneigh_entry __rcu *next;
include/net/net_namespace.h
358
return READ_ONCE(net->list.next);
include/net/netlabel.h
156
struct netlbl_lsm_catmap *next;
include/net/netlabel.h
337
catmap = catmap->next;
include/net/rose.h
111
struct rose_node *next;
include/net/rose.h
120
struct rose_route *next;
include/net/rose.h
94
struct rose_neigh *next;
include/net/sch_generic.h
1056
skb->next = NULL;
include/net/sch_generic.h
1057
last->next = skb;
include/net/sch_generic.h
1076
skb->next = qh->head;
include/net/sch_generic.h
1089
qh->head = skb->next;
include/net/sch_generic.h
1093
skb->next = NULL;
include/net/sch_generic.h
1162
struct sk_buff *next = skb->next;
include/net/sch_generic.h
1164
prefetch(next);
include/net/sch_generic.h
1166
skb = next;
include/net/sch_generic.h
1177
skb->next = q->to_free;
include/net/sch_generic.h
1186
skb->next = *to_free;
include/net/sch_generic.h
1194
skb->prev->next = *to_free;
include/net/sch_generic.h
1196
skb->next = *to_free;
include/net/sch_generic.h
305
struct Qdisc_ops *next;
include/net/sch_generic.h
33
struct qdisc_rate_table *next;
include/net/sch_generic.h
425
struct tcf_proto __rcu *next;
include/net/sctp/sctp.h
383
result = list->next;
include/net/sctp/structs.h
1381
struct sctp_stream_out_ext *next;
include/net/slhc_vj.h
131
struct cstate *next; /* next in ring (xmit) */
include/net/sock.h
1137
sk->sk_backlog.tail->next = skb;
include/net/sock.h
1140
skb->next = NULL;
include/net/sock.h
1270
if (offsetof(struct sock, sk_node.next) != 0)
include/net/sock.h
1271
memset(sk, 0, offsetof(struct sock, sk_node.next));
include/net/sock.h
784
return hlist_entry_safe(sk->sk_node.next, struct sock, sk_node);
include/net/sock.h
789
return (!is_a_nulls(sk->sk_nulls_node.next)) ?
include/net/sock.h
790
hlist_nulls_entry(sk->sk_nulls_node.next,
include/net/tcp.h
2239
if (sk->sk_write_queue.next == skb)
include/net/udp.h
254
for (skb = skb_shinfo(skb)->frag_list; skb; skb = skb->next) {
include/net/vxlan.h
289
struct list_head next; /* vxlan's per namespace list */
include/net/xfrm.h
1122
struct xfrm_if __rcu *next; /* next interface in list */
include/net/xfrm.h
1607
struct xfrm4_protocol __rcu *next;
include/net/xfrm.h
1619
struct xfrm6_protocol __rcu *next;
include/net/xfrm.h
1629
struct xfrm_tunnel __rcu *next;
include/net/xfrm.h
1638
struct xfrm6_tunnel __rcu *next;
include/rdma/ib_mad.h
467
struct ib_mad_send_buf *next;
include/rdma/ib_verbs.h
1416
struct ib_send_wr *next;
include/rdma/ib_verbs.h
1487
struct ib_recv_wr *next;
include/rdma/ib_verbs.h
2190
struct ib_flow_spec_list *next;
include/rdma/rdmavt_mr.h
59
u32 next; /* next unused index (speeds search) */
include/rdma/rdmavt_qp.h
339
struct rvt_qp __rcu *next; /* link list for QPN hash table */
include/rdma/rdmavt_qp.h
713
qp = rcu_dereference(qp->next))
include/rv/ltl_monitor.h
111
char states[32], next[32];
include/rv/ltl_monitor.h
118
snprintf(next, sizeof(next), "%*pbl", RV_MAX_BA_STATES, next_state);
include/rv/ltl_monitor.h
127
CONCATENATE(trace_event_, MONITOR_NAME)(task, states, atoms.buffer, next);
include/sound/control.h
118
struct snd_ctl_layer_ops *next;
include/sound/emu10k1.h
1576
struct snd_emu10k1_fx8010_irq *next;
include/sound/gus.h
183
struct snd_gf1_mem_block *next;
include/sound/gus.h
203
struct snd_gf1_dma_block *next;
include/sound/opl3.h
258
struct fm_patch *next;
include/sound/pcm.h
1484
for (s = info->pcm->streams[info->stream].substream; s; s = s->next)
include/sound/pcm.h
484
struct snd_pcm_substream *next;
include/sound/pcm_oss.h
21
struct snd_pcm_oss_setup *next;
include/sound/soundfont.h
20
struct snd_sf_zone *next; /* Link to next */
include/sound/soundfont.h
38
struct snd_sf_sample *next;
include/sound/soundfont.h
45
struct snd_soundfont *next; /* Link to next */
include/trace/events/ext4.h
2679
ext4_group_t next, unsigned int prefetch_ios),
include/trace/events/ext4.h
2681
TP_ARGS(sb, group, next, prefetch_ios),
include/trace/events/ext4.h
2686
__field( __u32, next )
include/trace/events/ext4.h
2693
__entry->next = next;
include/trace/events/ext4.h
2699
__entry->group, __entry->next, __entry->ios)
include/trace/events/sched.h
224
struct task_struct *next,
include/trace/events/sched.h
227
TP_ARGS(preempt, prev, next, prev_state),
include/trace/events/sched.h
244
memcpy(__entry->next_comm, next->comm, TASK_COMM_LEN);
include/trace/events/sched.h
245
__entry->next_pid = next->pid;
include/trace/events/sched.h
246
__entry->next_prio = next->prio;
include/uapi/drm/drm.h
115
unsigned char next;
include/uapi/drm/v3d_drm.h
78
__u64 next;
include/uapi/linux/dm-ioctl.h
168
__u32 next;
include/uapi/linux/dm-ioctl.h
193
__u32 next; /* offset to the next record from
include/uapi/linux/dm-ioctl.h
216
__u32 next;
include/uapi/linux/fd.h
356
struct floppy_raw_cmd *next; /* used for chaining of raw cmd's
include/uapi/linux/futex.h
118
struct robust_list __user *next;
include/uapi/linux/romfs_fs.h
36
__be32 next; /* low 4 bits see ROMFH_ */
include/uapi/linux/vfio.h
89
__u32 next; /* Offset of next capability */
include/uapi/linux/videodev2.h
1332
struct v4l2_clip __user *next;
include/uapi/linux/virtio_ring.h
108
__virtio16 next;
include/xen/grant_table.h
64
struct gnttab_free_callback *next;
include/xen/interface/xen-mca.h
377
unsigned next;
init/initramfs.c
113
q->next = NULL;
init/initramfs.c
125
*p = q->next;
init/initramfs.c
243
static void __init read_into(char *buf, unsigned size, enum state next)
init/initramfs.c
248
state = next;
init/initramfs.c
252
next_state = next;
init/initramfs.c
78
struct hash *next;
init/initramfs.c
94
for (p = head + hash(major, minor, ino); *p; p = &(*p)->next) {
init/main.c
1240
struct list_head next;
init/main.c
1261
list_add(&entry->next, &blacklisted_initcalls);
init/main.c
1286
list_for_each_entry(entry, &blacklisted_initcalls, next) {
io_uring/bpf_filter.c
116
struct io_bpf_filter *next = f->next;
io_uring/bpf_filter.c
131
f = next;
io_uring/bpf_filter.c
20
struct io_bpf_filter *next;
io_uring/bpf_filter.c
430
filter->next = old_filter;
io_uring/bpf_filter.c
91
filter = filter->next;
io_uring/io-wq.c
436
struct io_wq_work *next;
io_uring/io-wq.c
449
next = container_of(acct->work_list.first, struct io_wq_work, list);
io_uring/io-wq.c
450
work_flags = atomic_read(&next->flags);
io_uring/io_uring.c
1088
node = req->comp_list.next;
io_uring/io_uring.c
1094
node = req->comp_list.next;
io_uring/io_uring.c
1115
node = req->comp_list.next;
io_uring/io_uring.c
166
req->comp_list.next = IO_URING_PTR_POISON;
io_uring/io_uring.c
283
ctx->submit_state.free_list.next = NULL;
io_uring/io_uring.c
366
req->work.list.next = NULL;
io_uring/io_uring.h
515
return !ctx->submit_state.free_list.next;
io_uring/io_uring.h
524
req = container_of(ctx->submit_state.free_list.next, struct io_kiocb, comp_list);
io_uring/notif.c
103
nd->next = prev_nd->next;
io_uring/notif.c
104
prev_nd->next = nd;
io_uring/notif.c
134
nd->next = NULL;
io_uring/notif.c
38
nd = nd->next;
io_uring/notif.c
65
tw_flags = nd->next ? 0 : IOU_F_TWQ_LAZY_WAKE;
io_uring/notif.c
88
if (unlikely(nd->head != nd || nd->next))
io_uring/notif.h
17
struct io_notif_data *next;
io_uring/slist.h
11
for (pos = (head)->first, prv = NULL; pos; prv = pos, pos = (pos)->next)
io_uring/slist.h
23
struct io_wq_work_node *next = pos->next;
io_uring/slist.h
25
pos->next = node;
io_uring/slist.h
26
node->next = next;
io_uring/slist.h
27
if (!next)
io_uring/slist.h
34
node->next = NULL;
io_uring/slist.h
39
list->last->next = node;
io_uring/slist.h
50
WRITE_ONCE(list->first, last->next);
io_uring/slist.h
52
prev->next = last->next;
io_uring/slist.h
56
last->next = NULL;
io_uring/slist.h
62
node->next = stack->next;
io_uring/slist.h
63
stack->next = node;
io_uring/slist.h
76
struct io_wq_work_node *node = stack->next;
io_uring/slist.h
78
stack->next = node->next;
io_uring/slist.h
8
for (pos = (head)->first; pos; pos = (pos)->next)
io_uring/slist.h
84
if (!work->list.next)
io_uring/slist.h
87
return container_of(work->list.next, struct io_wq_work, list);
io_uring/tw.c
211
req->io_task_work.node.next = head;
io_uring/tw.c
302
struct llist_node *next = (*node)->next;
io_uring/tw.c
308
*node = next;
io_uring/tw.c
61
struct llist_node *next = node->next;
io_uring/tw.c
75
node = next;
io_uring/tw.c
95
node = node->next;
ipc/msg.c
208
if (mss->list.next)
ipc/msg.c
221
mss->list.next = NULL;
ipc/msgutil.c
107
for (seg = msg->next; seg != NULL; seg = seg->next) {
ipc/msgutil.c
138
for (dst_pseg = dst->next, src_pseg = src->next;
ipc/msgutil.c
140
dst_pseg = dst_pseg->next, src_pseg = src_pseg->next) {
ipc/msgutil.c
167
for (seg = msg->next; seg != NULL; seg = seg->next) {
ipc/msgutil.c
183
seg = msg->next;
ipc/msgutil.c
186
struct msg_msgseg *tmp = seg->next;
ipc/msgutil.c
35
struct msg_msgseg *next;
ipc/msgutil.c
65
msg->next = NULL;
ipc/msgutil.c
69
pseg = &msg->next;
ipc/msgutil.c
80
seg->next = NULL;
ipc/msgutil.c
81
pseg = &seg->next;
ipc/sem.c
2356
un = list_entry_rcu(ulp->list_proc.next,
ipc/util.c
893
.next = sysvipc_proc_next,
kernel/audit_tree.c
1018
owner = list_entry(chunk->trees.next,
kernel/audit_tree.c
546
struct audit_krule *rule, *next;
kernel/audit_tree.c
549
list_for_each_entry_safe(rule, next, &tree->rules, rlist) {
kernel/audit_tree.c
620
for (p = tree->chunks.next; p != &tree->chunks; p = q) {
kernel/audit_tree.c
622
q = p->next;
kernel/audit_tree.c
679
while (cursor.next != &tree_list) {
kernel/audit_tree.c
687
tree = container_of(cursor.next, struct audit_tree, list);
kernel/audit_tree.c
773
victim = list_entry(prune_list.next,
kernel/audit_tree.c
903
while (cursor.next != &tree_list) {
kernel/audit_tree.c
907
tree = container_of(cursor.next, struct audit_tree, list);
kernel/audit_tree.c
989
victim = list_entry(list->next, struct audit_tree, list);
kernel/auditsc.c
117
struct audit_tree_refs *next;
kernel/auditsc.c
1695
for (aux = context->aux; aux; aux = aux->next) {
kernel/auditsc.c
1749
for (aux = context->aux_pids; aux; aux = aux->next) {
kernel/auditsc.c
244
p = p->next;
kernel/auditsc.c
264
p->next = ctx->trees;
kernel/auditsc.c
2712
axp->d.next = ctx->aux_pids;
kernel/auditsc.c
2751
ax->d.next = context->aux;
kernel/auditsc.c
286
for (q = p; q != ctx->trees; q = q->next, n = 31) {
kernel/auditsc.c
305
q = p->next;
kernel/auditsc.c
318
for (p = ctx->first_trees; p != ctx->trees; p = p->next) {
kernel/auditsc.c
90
struct audit_aux_data *next;
kernel/auditsc.c
932
struct audit_names *n, *next;
kernel/auditsc.c
934
list_for_each_entry_safe(n, next, &context->names_list, list) {
kernel/auditsc.c
952
context->aux = aux->next;
kernel/auditsc.c
957
context->aux_pids = aux->next;
kernel/bpf/arraymap.c
348
u32 *next = (u32 *)next_key;
kernel/bpf/arraymap.c
351
*next = 0;
kernel/bpf/arraymap.c
358
*next = index + 1;
kernel/bpf/arraymap.c
735
.next = bpf_array_map_seq_next,
kernel/bpf/bpf_iter.c
158
p = seq->op->next(seq, p, &seq->index);
kernel/bpf/bpf_iter.c
162
seq->op->next);
kernel/bpf/bpf_lru_list.c
167
struct list_head *cur, *last, *next = inactive;
kernel/bpf/bpf_lru_list.c
174
last = l->next_inactive_rotation->next;
kernel/bpf/bpf_lru_list.c
176
last = last->next;
kernel/bpf/bpf_lru_list.c
186
next = cur->prev;
kernel/bpf/bpf_lru_list.c
191
cur = next;
kernel/bpf/bpf_lru_list.c
195
l->next_inactive_rotation = next;
kernel/bpf/cgroup.c
724
if (pltmp->node.next)
kernel/bpf/cgroup_iter.c
160
.next = cgroup_iter_seq_next,
kernel/bpf/cpumap.c
677
u32 *next = next_key;
kernel/bpf/cpumap.c
680
*next = 0;
kernel/bpf/cpumap.c
686
*next = index + 1;
kernel/bpf/devmap.c
1109
struct hlist_node *next;
kernel/bpf/devmap.c
1113
hlist_for_each_entry_safe(dev, next, head, index_hlist) {
kernel/bpf/devmap.c
222
struct hlist_node *next;
kernel/bpf/devmap.c
226
hlist_for_each_entry_safe(dev, next, head, index_hlist) {
kernel/bpf/devmap.c
260
u32 *next = next_key;
kernel/bpf/devmap.c
263
*next = 0;
kernel/bpf/devmap.c
269
*next = index + 1;
kernel/bpf/devmap.c
295
u32 idx, *next = next_key;
kernel/bpf/devmap.c
313
*next = next_dev->idx;
kernel/bpf/devmap.c
328
*next = next_dev->idx;
kernel/bpf/devmap.c
750
struct hlist_node *next;
kernel/bpf/devmap.c
790
hlist_for_each_entry_safe(dst, next, head, index_hlist) {
kernel/bpf/dmabuf_iter.c
89
.next = dmabuf_iter_seq_next,
kernel/bpf/hashtab.c
2247
.next = bpf_hash_map_seq_next,
kernel/bpf/hashtab.c
423
BUILD_BUG_ON(offsetof(struct htab_elem, fnode.next) !=
kernel/bpf/helpers.c
2247
if (!head->next || list_empty(head))
kernel/bpf/helpers.c
2249
head = head->next;
kernel/bpf/helpers.c
2258
head = head->next;
kernel/bpf/helpers.c
2391
if (unlikely(!h->next))
kernel/bpf/helpers.c
2437
if (unlikely(!h->next))
kernel/bpf/helpers.c
2442
n = tail ? h->prev : h->next;
kernel/bpf/helpers.c
2466
if (list_empty(h) || unlikely(!h->next))
kernel/bpf/helpers.c
2469
return (struct bpf_list_node *)h->next;
kernel/bpf/helpers.c
2476
if (list_empty(h) || unlikely(!h->next))
kernel/bpf/inode.c
266
.next = map_seq_next,
kernel/bpf/kmem_cache_iter.c
199
.next = kmem_cache_iter_seq_next,
kernel/bpf/kmem_cache_iter.c
39
struct kmem_cache *next;
kernel/bpf/kmem_cache_iter.c
53
next = list_first_entry(&slab_caches, struct kmem_cache, list);
kernel/bpf/kmem_cache_iter.c
55
next = NULL;
kernel/bpf/kmem_cache_iter.c
57
next = list_next_entry(prev, list);
kernel/bpf/kmem_cache_iter.c
60
if (next && next->refcount > 0)
kernel/bpf/kmem_cache_iter.c
61
next->refcount++;
kernel/bpf/kmem_cache_iter.c
76
kit->pos = next;
kernel/bpf/kmem_cache_iter.c
77
return next;
kernel/bpf/link_iter.c
76
.next = bpf_link_seq_next,
kernel/bpf/local_storage.c
283
struct bpf_cgroup_storage_key *next = _next_key;
kernel/bpf/local_storage.c
284
*next = storage->key;
kernel/bpf/local_storage.c
286
__u64 *next = _next_key;
kernel/bpf/local_storage.c
287
*next = storage->key.cgroup_inode_id;
kernel/bpf/map_iter.c
76
.next = bpf_map_seq_next,
kernel/bpf/memalloc.c
132
struct llist_node *entry, *next;
kernel/bpf/memalloc.c
137
next = entry->next;
kernel/bpf/memalloc.c
138
head->first = next;
kernel/bpf/percpu_freelist.c
114
WRITE_ONCE(head->first, node->next);
kernel/bpf/percpu_freelist.c
31
node->next = head->first;
kernel/bpf/percpu_freelist.h
20
struct pcpu_freelist_node *next;
kernel/bpf/prog_iter.c
76
.next = bpf_prog_seq_next,
kernel/bpf/reuseport_array.c
321
u32 *next = (u32 *)next_key;
kernel/bpf/reuseport_array.c
324
*next = 0;
kernel/bpf/reuseport_array.c
331
*next = index + 1;
kernel/bpf/rqspinlock.c
341
struct mcs_spinlock *prev, *next, *node;
kernel/bpf/rqspinlock.c
497
node->next = NULL;
kernel/bpf/rqspinlock.c
522
next = NULL;
kernel/bpf/rqspinlock.c
534
WRITE_ONCE(prev->next, node);
kernel/bpf/rqspinlock.c
548
next = READ_ONCE(node->next);
kernel/bpf/rqspinlock.c
549
if (next)
kernel/bpf/rqspinlock.c
550
prefetchw(next);
kernel/bpf/rqspinlock.c
575
if (!next)
kernel/bpf/rqspinlock.c
576
next = smp_cond_load_relaxed(&node->next, (VAL));
kernel/bpf/rqspinlock.c
577
arch_mcs_spin_unlock_contended(&next->locked);
kernel/bpf/rqspinlock.c
608
next = smp_cond_load_relaxed(&node->next, VAL);
kernel/bpf/rqspinlock.c
609
WRITE_ONCE(next->locked, RES_TIMEOUT_VAL);
kernel/bpf/rqspinlock.c
646
if (!next)
kernel/bpf/rqspinlock.c
647
next = smp_cond_load_relaxed(&node->next, (VAL));
kernel/bpf/rqspinlock.c
649
arch_mcs_spin_unlock_contended(&next->locked);
kernel/bpf/stream.c
118
stream->backlog_head = node->next;
kernel/bpf/stream.c
137
stream->backlog_tail->next = head;
kernel/bpf/task_iter.c
242
.next = task_seq_next,
kernel/bpf/task_iter.c
401
.next = task_file_seq_next,
kernel/bpf/task_iter.c
655
.next = task_vma_seq_next,
kernel/bpf/verifier.c
190
struct bpf_verifier_stack_elem *next;
kernel/bpf/verifier.c
2022
backedge->next = visit->backedges;
kernel/bpf/verifier.c
20354
for (backedge = visit->backedges; backedge; backedge = backedge->next)
kernel/bpf/verifier.c
20359
for (backedge = visit->backedges; backedge; backedge = backedge->next) {
kernel/bpf/verifier.c
2050
struct bpf_scc_backedge *backedge, *next;
kernel/bpf/verifier.c
2052
for (backedge = visit->backedges; backedge; backedge = next) {
kernel/bpf/verifier.c
2054
next = backedge->next;
kernel/bpf/verifier.c
2109
elem = head->next;
kernel/bpf/verifier.c
2143
elem->next = env->head;
kernel/bpf/verifier.c
3019
elem->next = env->head;
kernel/bpf/verifier.c
3709
goto next;
kernel/bpf/verifier.c
3711
goto next;
kernel/bpf/verifier.c
3714
goto next;
kernel/bpf/verifier.c
3721
next:
kernel/cgroup/cgroup.c
1060
l1 = l1->next;
kernel/cgroup/cgroup.c
1061
l2 = l2->next;
kernel/cgroup/cgroup.c
4695
unsigned long next = last + CGROUP_FILE_NOTIFY_MIN_INTV;
kernel/cgroup/cgroup.c
4697
if (time_in_range(jiffies, last, next)) {
kernel/cgroup/cgroup.c
4698
timer_reduce(&cfile->notify_timer, next);
kernel/cgroup/cgroup.c
4748
struct cgroup_subsys_state *next;
kernel/cgroup/cgroup.c
4773
next = list_entry_rcu(parent->children.next, struct cgroup_subsys_state, sibling);
kernel/cgroup/cgroup.c
4775
next = list_entry_rcu(pos->sibling.next, struct cgroup_subsys_state, sibling);
kernel/cgroup/cgroup.c
4777
list_for_each_entry_rcu(next, &parent->children, sibling,
kernel/cgroup/cgroup.c
4779
if (next->serial_nr > pos->serial_nr)
kernel/cgroup/cgroup.c
4787
if (&next->sibling != &parent->children)
kernel/cgroup/cgroup.c
4788
return next;
kernel/cgroup/cgroup.c
4818
struct cgroup_subsys_state *next;
kernel/cgroup/cgroup.c
4827
next = css_next_child(NULL, pos);
kernel/cgroup/cgroup.c
4828
if (next)
kernel/cgroup/cgroup.c
4829
return next;
kernel/cgroup/cgroup.c
4833
next = css_next_child(pos, pos->parent);
kernel/cgroup/cgroup.c
4834
if (next)
kernel/cgroup/cgroup.c
4835
return next;
kernel/cgroup/cgroup.c
4914
struct cgroup_subsys_state *next;
kernel/cgroup/cgroup.c
4927
next = css_next_child(pos, pos->parent);
kernel/cgroup/cgroup.c
4928
if (next)
kernel/cgroup/cgroup.c
4929
return css_leftmost_descendant(next);
kernel/cgroup/cgroup.c
4969
l = it->tcset_pos->next;
kernel/cgroup/cgroup.c
4982
l = l->next;
kernel/cgroup/cgroup.c
5040
it->task_pos = it->cur_tasks_head->next;
kernel/cgroup/cgroup.c
5072
it->task_pos = it->task_pos->next;
kernel/cgroup/cgroup.c
5092
it->task_pos = it->task_pos->next;
kernel/cgroup/cgroup.c
5096
it->task_pos = it->cur_tasks_head->next;
kernel/cgroup/cgroup.c
5100
it->task_pos = it->cur_tasks_head->next;
kernel/cgroup/cgroup.c
7120
struct task_struct *task, *next;
kernel/cgroup/cgroup.c
7123
llist_for_each_entry_safe(task, next, lnode, cg_dead_lnode) {
kernel/cgroup/dmem.c
206
struct dmem_cgroup_pool_state *pool, *next;
kernel/cgroup/dmem.c
209
list_for_each_entry_safe(pool, next, &dmemcs->pools, css_node) {
kernel/cgroup/dmem.c
447
struct dmem_cgroup_pool_state *pool, *next;
kernel/cgroup/dmem.c
449
list_for_each_entry_safe(pool, next, &region->pools, region_node)
kernel/cgroup/dmem.c
470
struct dmem_cgroup_pool_state *pool, *next;
kernel/cgroup/dmem.c
480
list_for_each_entry_safe(pool, next, &region->pools, region_node) {
kernel/cgroup/rstat.c
121
if (!try_cmpxchg(&rstatc->lnode.next, &self, NULL))
kernel/crash_reserve.c
242
goto next;
kernel/crash_reserve.c
250
next:
kernel/debug/debug_core.c
1057
.next = NULL,
kernel/dma/debug.c
648
entry = list_entry(free_entries.next, struct dma_debug_entry, list);
kernel/events/callchain.c
212
ri = ri->next;
kernel/events/core.c
14518
struct perf_event *child_event, *next;
kernel/events/core.c
14568
list_for_each_entry_safe(child_event, next, &ctx->event_list, event_entry)
kernel/events/core.c
1941
struct rb_node *next;
kernel/events/core.c
1943
next = rb_next_match(&key, &event->group_node, __group_cmp);
kernel/events/core.c
1944
if (next)
kernel/events/core.c
1945
return __node_2_pe(next);
kernel/events/core.c
3765
perf_event_context_sched_out(struct task_struct *task, struct task_struct *next)
kernel/events/core.c
3776
next_ctx = rcu_dereference(next->perf_event_ctxp);
kernel/events/core.c
3820
WRITE_ONCE(ctx->task, next);
kernel/events/core.c
3835
RCU_INIT_POINTER(next->perf_event_ctxp, ctx);
kernel/events/core.c
3916
struct task_struct *next,
kernel/events/core.c
3923
if (prev == next || cpuctx->task_ctx)
kernel/events/core.c
3927
__perf_pmu_sched_task(cpc, sched_in ? next : prev, sched_in);
kernel/events/core.c
3945
struct task_struct *next)
kernel/events/core.c
3948
perf_pmu_sched_task(task, next, false);
kernel/events/core.c
3951
perf_event_switch(task, next, false);
kernel/events/core.c
3953
perf_event_context_sched_out(task, next);
kernel/events/core.c
3960
perf_cgroup_switch(next);
kernel/events/core.c
4740
struct perf_event *event, *next;
kernel/events/core.c
4749
list_for_each_entry_safe(event, next, &ctx->event_list, event_entry) {
kernel/events/core.c
8246
frag = frag->next;
kernel/events/uprobes.c
1189
struct map_info *next;
kernel/events/uprobes.c
1196
struct map_info *next = info->next;
kernel/events/uprobes.c
1198
return next;
kernel/events/uprobes.c
1225
prev->next = NULL;
kernel/events/uprobes.c
1236
prev = prev->next;
kernel/events/uprobes.c
1237
info->next = curr;
kernel/events/uprobes.c
1251
curr = curr->next;
kernel/events/uprobes.c
1260
info->next = prev;
kernel/events/uprobes.c
1949
ri->next = utask->ri_pool;
kernel/events/uprobes.c
1958
utask->ri_pool = ri->next;
kernel/events/uprobes.c
2021
ri_next = ri->next;
kernel/events/uprobes.c
2029
ri_next = ri->next;
kernel/events/uprobes.c
2040
for (pos = rcu_dereference_raw(head); pos; pos = rcu_dereference_raw(pos->next))
kernel/events/uprobes.c
2148
for (o = o_utask->return_instances; o; o = o->next) {
kernel/events/uprobes.c
2164
n->next = NULL;
kernel/events/uprobes.c
2166
p = &n->next;
kernel/events/uprobes.c
2243
ri_next = ri->next;
kernel/events/uprobes.c
2307
ri->next = utask->return_instances;
kernel/events/uprobes.c
2629
ri = ri->next; /* can't be NULL if chained */
kernel/events/uprobes.c
2671
ri_next = ri->next;
kernel/exit.c
431
self.next = xchg(&core_state->dumper.next, &self);
kernel/fail_function.c
209
.next = fei_seq_next,
kernel/fork.c
972
tsk->wake_q.next = NULL;
kernel/futex/core.c
1167
if (fetch_robust_entry(&entry, &head->list.next, &pi))
kernel/futex/core.c
1187
rc = fetch_robust_entry(&next_entry, &entry->next, &next_pi);
kernel/futex/core.c
1262
if (compat_fetch_robust_entry(&uentry, &entry, &head->list.next, &pi))
kernel/futex/core.c
1284
(compat_uptr_t __user *)&entry->next, &next_pi);
kernel/futex/core.c
1326
struct list_head *next, *head = &curr->pi_state_list;
kernel/futex/core.c
1348
next = head->next;
kernel/futex/core.c
1349
pi_state = list_entry(next, struct futex_pi_state, list);
kernel/futex/core.c
1379
if (head->next != next) {
kernel/futex/requeue.c
386
struct futex_q *this, *next;
kernel/futex/requeue.c
586
plist_for_each_entry_safe(this, next, &hb1->chain, list) {
kernel/futex/waitwake.c
157
struct futex_q *this, *next;
kernel/futex/waitwake.c
180
plist_for_each_entry_safe(this, next, &hb->chain, list) {
kernel/futex/waitwake.c
255
struct futex_q *this, *next;
kernel/futex/waitwake.c
299
plist_for_each_entry_safe(this, next, &hb1->chain, list) {
kernel/futex/waitwake.c
313
plist_for_each_entry_safe(this, next, &hb2->chain, list) {
kernel/gcov/fs.c
245
.next = gcov_seq_next,
kernel/gcov/fs.c
677
char *next;
kernel/gcov/fs.c
686
for (curr = filename; (next = strchr(curr, '/')); curr = next + 1) {
kernel/gcov/fs.c
687
if (curr == next)
kernel/gcov/fs.c
689
*next = 0;
kernel/gcov/gcc_4_7.c
138
return info->next;
kernel/gcov/gcc_4_7.c
147
info->next = gcov_info_head;
kernel/gcov/gcc_4_7.c
159
prev->next = info->next;
kernel/gcov/gcc_4_7.c
161
gcov_info_head = info->next;
kernel/gcov/gcc_4_7.c
293
dup->next = NULL;
kernel/gcov/gcc_4_7.c
96
struct gcov_info *next;
kernel/irq/chip.c
914
for (action = desc->action; action; action = action->next)
kernel/irq/internals.h
147
for (act = desc->action; act; act = act->next)
kernel/irq/ipi.c
56
unsigned int next;
kernel/irq/ipi.c
69
next = cpumask_next_zero(offset, dest);
kernel/irq/ipi.c
70
if (next < nr_cpu_ids)
kernel/irq/ipi.c
71
next = cpumask_next(next, dest);
kernel/irq/ipi.c
72
if (next < nr_cpu_ids) {
kernel/irq/manage.c
1450
old = old->next;
kernel/irq/manage.c
1650
old_ptr = &old->next;
kernel/irq/manage.c
1888
action_ptr = &action->next;
kernel/irq/manage.c
1892
*action_ptr = action->next;
kernel/irq/manage.c
2418
action_ptr = &action->next;
kernel/irq/manage.c
2428
*action_ptr = action->next;
kernel/irq/proc.c
519
while ((action = action->next) != NULL)
kernel/jump_label.c
623
struct static_key_mod *next;
kernel/jump_label.c
647
key->next = mod;
kernel/jump_label.c
678
for (mod = static_key_mod(key); mod; mod = mod->next) {
kernel/jump_label.c
752
jlm2->next = NULL;
kernel/jump_label.c
758
jlm->next = static_key_mod(key);
kernel/jump_label.c
796
prev = &key->next;
kernel/jump_label.c
800
prev = &jlm->next;
kernel/jump_label.c
801
jlm = jlm->next;
kernel/jump_label.c
808
if (prev == &key->next)
kernel/jump_label.c
809
static_key_set_mod(key, jlm->next);
kernel/jump_label.c
811
*prev = jlm->next;
kernel/jump_label.c
817
if (jlm->next == NULL) {
kernel/kallsyms.c
790
.next = s_next,
kernel/kallsyms.c
833
.next = s_next,
kernel/kallsyms_selftest.c
260
int next = 0, nr_failed = 0;
kernel/kallsyms_selftest.c
337
if (i >= next) {
kernel/kallsyms_selftest.c
361
next = i + (rand & 0xff) + 1;
kernel/kexec_core.c
322
struct page *page, *next;
kernel/kexec_core.c
324
list_for_each_entry_safe(page, next, list, lru) {
kernel/kprobes.c
1874
ap = list_entry(p->list.next, struct kprobe, list);
kernel/kprobes.c
1993
node = node->next;
kernel/kprobes.c
2022
node = node->next;
kernel/kprobes.c
2030
node = node->next;
kernel/kprobes.c
2122
first = first->next;
kernel/kprobes.c
2129
current->kretprobe_instances.first = node->next;
kernel/kprobes.c
2130
node->next = NULL;
kernel/kprobes.c
2135
first = first->next;
kernel/kprobes.c
231
struct kprobe_insn_page *kip, *next;
kernel/kprobes.c
236
list_for_each_entry_safe(kip, next, &c->pages, list) {
kernel/kprobes.c
2881
.next = kprobe_seq_next,
kernel/kprobes.c
2925
.next = kprobe_blacklist_seq_next,
kernel/kthread.c
1352
kthread_insert_work(worker, &fwork.work, work->node.next);
kernel/kthread.c
1355
worker->work_list.next);
kernel/kthread.c
812
create = list_entry(kthread_create_list.next,
kernel/livepatch/patch.c
106
func = list_entry_rcu(func->stack_node.next,
kernel/liveupdate/kexec_handover.c
1061
chunk = KHOSER_LOAD_PTR(chunk->hdr.next);
kernel/liveupdate/kexec_handover.c
1123
chunk = KHOSER_LOAD_PTR(chunk->hdr.next);
kernel/liveupdate/kexec_handover.c
344
DECLARE_KHOSER_PTR(next, struct khoser_mem_chunk *);
kernel/liveupdate/kexec_handover.c
374
KHOSER_STORE_PTR(cur_chunk->hdr.next, chunk);
kernel/liveupdate/kexec_handover.c
385
chunk = KHOSER_LOAD_PTR(chunk->hdr.next);
kernel/liveupdate/kexec_handover.c
504
chunk = KHOSER_LOAD_PTR(chunk->hdr.next);
kernel/liveupdate/kexec_handover.c
955
KHOSER_STORE_PTR(cur->hdr.next, chunk);
kernel/locking/lockdep.c
1631
__calc_dep_bit(struct held_lock *prev, struct held_lock *next)
kernel/locking/lockdep.c
1633
return (prev->read == 0) + ((next->read != 2) << 1);
kernel/locking/lockdep.c
1636
static inline u8 calc_dep(struct held_lock *prev, struct held_lock *next)
kernel/locking/lockdep.c
1638
return 1U << __calc_dep_bit(prev, next);
kernel/locking/lockdep.c
1646
__calc_dep_bitb(struct held_lock *prev, struct held_lock *next)
kernel/locking/lockdep.c
1648
return (next->read != 2) + ((prev->read == 0) << 1);
kernel/locking/lockdep.c
1651
static inline u8 calc_depb(struct held_lock *prev, struct held_lock *next)
kernel/locking/lockdep.c
1653
return 1U << __calc_dep_bitb(prev, next);
kernel/locking/lockdep.c
2553
struct held_lock *next,
kernel/locking/lockdep.c
2575
print_lock(next);
kernel/locking/lockdep.c
2582
print_lock_name(next, hlock_class(next));
kernel/locking/lockdep.c
2601
hlock_class(prev), hlock_class(next));
kernel/locking/lockdep.c
2781
struct held_lock *next)
kernel/locking/lockdep.c
2812
bfs_init_root(&that, next);
kernel/locking/lockdep.c
2859
prev, next,
kernel/locking/lockdep.c
2869
struct held_lock *prev, struct held_lock *next)
kernel/locking/lockdep.c
2993
struct lock_class *next = hlock_class(nxt);
kernel/locking/lockdep.c
3003
__print_lock_name(nxt, next);
kernel/locking/lockdep.c
3011
struct held_lock *next)
kernel/locking/lockdep.c
3027
print_lock(next);
kernel/locking/lockdep.c
3033
class->cmp_fn(prev->instance, next->instance));
kernel/locking/lockdep.c
3037
print_deadlock_scenario(next, prev);
kernel/locking/lockdep.c
3057
check_deadlock(struct task_struct *curr, struct held_lock *next)
kernel/locking/lockdep.c
3067
if (prev->instance == next->nest_lock)
kernel/locking/lockdep.c
3070
if (hlock_class(prev) != hlock_class(next))
kernel/locking/lockdep.c
3077
if ((next->read == 2) && prev->read)
kernel/locking/lockdep.c
3083
class->cmp_fn(prev->instance, next->instance) < 0)
kernel/locking/lockdep.c
3093
print_deadlock_bug(curr, prev, next);
kernel/locking/lockdep.c
3123
struct held_lock *next, u16 distance,
kernel/locking/lockdep.c
3129
if (!hlock_class(prev)->key || !hlock_class(next)->key) {
kernel/locking/lockdep.c
3140
WARN_ONCE(!debug_locks_silent && !hlock_class(next)->key,
kernel/locking/lockdep.c
3142
hlock_class(next),
kernel/locking/lockdep.c
3143
hlock_class(next)->name);
kernel/locking/lockdep.c
3147
if (prev->class_idx == next->class_idx) {
kernel/locking/lockdep.c
3151
class->cmp_fn(prev->instance, next->instance) < 0)
kernel/locking/lockdep.c
3165
ret = check_noncircular(next, prev, trace);
kernel/locking/lockdep.c
3169
if (!check_irq_usage(curr, prev, next))
kernel/locking/lockdep.c
3181
if (entry->class == hlock_class(next)) {
kernel/locking/lockdep.c
3184
entry->dep |= calc_dep(prev, next);
kernel/locking/lockdep.c
3202
list_for_each_entry(entry, &hlock_class(next)->locks_before, entry) {
kernel/locking/lockdep.c
3206
entry->dep |= calc_depb(prev, next);
kernel/locking/lockdep.c
3219
ret = check_redundant(prev, next);
kernel/locking/lockdep.c
3235
ret = add_lock_to_list(hlock_class(next), hlock_class(prev),
kernel/locking/lockdep.c
3237
calc_dep(prev, next), *trace);
kernel/locking/lockdep.c
3242
ret = add_lock_to_list(hlock_class(prev), hlock_class(next),
kernel/locking/lockdep.c
3243
&hlock_class(next)->locks_before, distance,
kernel/locking/lockdep.c
3244
calc_depb(prev, next), *trace);
kernel/locking/lockdep.c
3258
check_prevs_add(struct task_struct *curr, struct held_lock *next)
kernel/locking/lockdep.c
3284
int ret = check_prev_add(curr, hlock, next, distance, &trace);
kernel/locking/lockdep.c
3378
int next = chain_hlocks[offset];
kernel/locking/lockdep.c
3380
WARN_ON_ONCE(!(next & CHAIN_BLK_FLAG));
kernel/locking/lockdep.c
3382
if (next == CHAIN_BLK_LIST_END)
kernel/locking/lockdep.c
3385
next &= ~CHAIN_BLK_FLAG;
kernel/locking/lockdep.c
3386
next <<= 16;
kernel/locking/lockdep.c
3387
next |= chain_hlocks[offset + 1];
kernel/locking/lockdep.c
3389
return next;
kernel/locking/lockdep.c
3400
static inline void init_chain_block(int offset, int next, int bucket, int size)
kernel/locking/lockdep.c
3402
chain_hlocks[offset] = (next >> 16) | CHAIN_BLK_FLAG;
kernel/locking/lockdep.c
3403
chain_hlocks[offset + 1] = (u16)next;
kernel/locking/lockdep.c
3414
int next = chain_block_buckets[bucket];
kernel/locking/lockdep.c
3451
init_chain_block(offset, next, bucket, size);
kernel/locking/lockdep.c
3466
static inline void del_chain_block(int bucket, int size, int next)
kernel/locking/lockdep.c
3469
chain_block_buckets[bucket] = next;
kernel/locking/lockdep.c
4852
static int check_wait_context(struct task_struct *curr, struct held_lock *next)
kernel/locking/lockdep.c
4854
u8 next_inner = hlock_class(next)->wait_type_inner;
kernel/locking/lockdep.c
4855
u8 next_outer = hlock_class(next)->wait_type_outer;
kernel/locking/lockdep.c
4859
if (!next_inner || next->trylock)
kernel/locking/lockdep.c
4870
if (prev->irq_context != next->irq_context)
kernel/locking/lockdep.c
4902
return print_lock_invalid_wait_context(curr, next);
kernel/locking/lockdep.c
4927
struct held_lock *next)
kernel/locking/lockdep.c
6287
WARN_ON_ONCE(!class->lock_entry.next);
kernel/locking/lockdep.c
6291
WARN_ON_ONCE(!class->lock_entry.next);
kernel/locking/lockdep_proc.c
122
.next = l_next,
kernel/locking/lockdep_proc.c
187
.next = lc_next,
kernel/locking/lockdep_proc.c
636
.next = ls_next,
kernel/locking/mcs_spinlock.h
105
while (!(next = READ_ONCE(node->next)))
kernel/locking/mcs_spinlock.h
110
arch_mcs_spin_unlock_contended(&next->locked);
kernel/locking/mcs_spinlock.h
63
node->next = NULL;
kernel/locking/mcs_spinlock.h
83
WRITE_ONCE(prev->next, node);
kernel/locking/mcs_spinlock.h
96
struct mcs_spinlock *next = READ_ONCE(node->next);
kernel/locking/mcs_spinlock.h
98
if (likely(!next)) {
kernel/locking/mutex-debug.c
77
DEBUG_LOCKS_WARN_ON(!lock->wait_list.prev && !lock->wait_list.next);
kernel/locking/mutex.c
933
struct task_struct *next = NULL;
kernel/locking/mutex.c
971
next = waiter->task;
kernel/locking/mutex.c
974
__clear_task_blocked_on(next, lock);
kernel/locking/mutex.c
975
wake_q_add(&wake_q, next);
kernel/locking/mutex.c
979
__mutex_handoff(lock, next);
kernel/locking/osq_lock.c
101
node->next = NULL;
kernel/locking/osq_lock.c
129
WRITE_ONCE(prev->next, node);
kernel/locking/osq_lock.c
16
struct optimistic_spin_node *next, *prev;
kernel/locking/osq_lock.c
164
if (data_race(prev->next) == node &&
kernel/locking/osq_lock.c
165
cmpxchg(&prev->next, node, NULL) == node)
kernel/locking/osq_lock.c
192
next = osq_wait_next(lock, node, prev->cpu);
kernel/locking/osq_lock.c
193
if (!next)
kernel/locking/osq_lock.c
204
WRITE_ONCE(next->prev, prev);
kernel/locking/osq_lock.c
205
WRITE_ONCE(prev->next, next);
kernel/locking/osq_lock.c
212
struct optimistic_spin_node *node, *next;
kernel/locking/osq_lock.c
225
next = xchg(&node->next, NULL);
kernel/locking/osq_lock.c
226
if (next) {
kernel/locking/osq_lock.c
227
WRITE_ONCE(next->locked, 1);
kernel/locking/osq_lock.c
231
next = osq_wait_next(lock, node, OSQ_UNLOCKED_VAL);
kernel/locking/osq_lock.c
232
if (next)
kernel/locking/osq_lock.c
233
WRITE_ONCE(next->locked, 1);
kernel/locking/osq_lock.c
81
if (node->next) {
kernel/locking/osq_lock.c
82
struct optimistic_spin_node *next;
kernel/locking/osq_lock.c
84
next = xchg(&node->next, NULL);
kernel/locking/osq_lock.c
85
if (next)
kernel/locking/osq_lock.c
86
return next;
kernel/locking/osq_lock.c
96
struct optimistic_spin_node *prev, *next;
kernel/locking/qspinlock.c
132
struct mcs_spinlock *prev, *next, *node;
kernel/locking/qspinlock.c
252
node->next = NULL;
kernel/locking/qspinlock.c
278
next = NULL;
kernel/locking/qspinlock.c
288
WRITE_ONCE(prev->next, node);
kernel/locking/qspinlock.c
299
next = READ_ONCE(node->next);
kernel/locking/qspinlock.c
300
if (next)
kernel/locking/qspinlock.c
301
prefetchw(next);
kernel/locking/qspinlock.c
367
if (!next)
kernel/locking/qspinlock.c
368
next = smp_cond_load_relaxed(&node->next, (VAL));
kernel/locking/qspinlock.c
370
arch_mcs_spin_unlock_contended(&next->locked);
kernel/locking/qspinlock.c
371
pv_kick_node(lock, next);
kernel/module/main.c
3640
struct hlist_node *next;
kernel/module/main.c
3644
hlist_for_each_entry_safe(pos, next, head, entry) {
kernel/module/procfs.c
114
.next = m_next,
kernel/module/tracking.c
103
.next = unloaded_tainted_modules_seq_next,
kernel/notifier.c
31
nl = &((*nl)->next);
kernel/notifier.c
33
n->next = *nl;
kernel/notifier.c
44
rcu_assign_pointer(*nl, n->next);
kernel/notifier.c
48
nl = &((*nl)->next);
kernel/notifier.c
75
next_nb = rcu_dereference_raw(nb->next);
kernel/nstree.c
587
first_ns = list_entry_rcu(head->next, typeof(*first_ns), ns_owner_node.ns_list_entry);
kernel/nstree.c
674
return list_entry_rcu(head->next, struct ns_common, ns_tree_node.ns_list_entry);
kernel/nstree.c
675
return list_entry_rcu(head->next, struct ns_common, ns_unified_node.ns_list_entry);
kernel/nstree.c
682
return list_entry_rcu(ns->ns_tree_node.ns_list_entry.next, struct ns_common, ns_tree_node.ns_list_entry);
kernel/nstree.c
683
return list_entry_rcu(ns->ns_unified_node.ns_list_entry.next, struct ns_common, ns_unified_node.ns_list_entry);
kernel/padata.c
137
struct padata_work *cur, *next;
kernel/padata.c
143
list_for_each_entry_safe(cur, next, works, pw_list) {
kernel/padata.c
260
padata = list_entry(reorder->list.next, struct padata_priv, list);
kernel/padata.c
338
padata = list_entry(local_list.next,
kernel/pid_namespace.c
291
int ret, next;
kernel/pid_namespace.c
296
next = idr_get_cursor(&pid_ns->idr) - 1;
kernel/pid_namespace.c
298
tmp.data = &next;
kernel/pid_namespace.c
302
idr_set_cursor(&pid_ns->idr, next + 1);
kernel/power/snapshot.c
157
struct linked_page *next;
kernel/power/snapshot.c
215
safe_pages_list = safe_pages_list->next;
kernel/power/snapshot.c
2398
struct highmem_pbe *next;
kernel/power/snapshot.c
243
lp->next = safe_pages_list;
kernel/power/snapshot.c
2544
pbe->next = highmem_pblist;
kernel/power/snapshot.c
2680
lp->next = safe_pages_list;
kernel/power/snapshot.c
2694
lp->next = safe_pages_list;
kernel/power/snapshot.c
274
struct linked_page *lp = list->next;
kernel/power/snapshot.c
2748
pbe->next = restore_pblist;
kernel/power/snapshot.c
2774
next:
kernel/power/snapshot.c
2840
goto next;
kernel/power/snapshot.c
2925
pbe = pbe->next;
kernel/power/snapshot.c
322
lp->next = ca->chain;
kernel/power/snapshot.c
590
bm->cur.zone = list_entry(bm->zones.next, struct mem_zone_bm_rtree,
kernel/power/snapshot.c
592
bm->cur.node = list_entry(bm->cur.zone->leaves.next,
kernel/power/snapshot.c
895
bm->cur.node = list_entry(bm->cur.node->list.next,
kernel/power/snapshot.c
905
bm->cur.zone = list_entry(bm->cur.zone->list.next,
kernel/power/snapshot.c
907
bm->cur.node = list_entry(bm->cur.zone->leaves.next,
kernel/power/snapshot.c
986
p_list = lp->next;
kernel/power/swap.c
1027
last->next = tmp;
kernel/power/swap.c
1071
handle->maps = handle->maps->next;
kernel/power/swap.c
87
struct swap_map_page_list *next;
kernel/power/swap.c
997
handle->maps = handle->maps->next;
kernel/printk/index.c
120
.next = pi_next,
kernel/printk/printk.c
1011
VMCOREINFO_OFFSET(prb_data_blk_lpos, next);
kernel/printk/printk.c
1425
char *next;
kernel/printk/printk.c
1443
next = memchr(text, '\n', text_len);
kernel/printk/printk.c
1444
if (next) {
kernel/printk/printk.c
1445
line_len = next - text;
kernel/printk/printk_ringbuffer.c
1068
blk_lpos->next = EMPTY_LINE_LPOS;
kernel/printk/printk_ringbuffer.c
1091
blk_lpos->next = FAILED_LPOS;
kernel/printk/printk_ringbuffer.c
1130
blk_lpos->next = next_lpos;
kernel/printk/printk_ringbuffer.c
1158
if (head_lpos != blk_lpos->next)
kernel/printk/printk_ringbuffer.c
1162
wrapped = is_blk_wrapped(data_ring, blk_lpos->begin, blk_lpos->next);
kernel/printk/printk_ringbuffer.c
1224
(blk_lpos->next - blk_lpos->begin) - sizeof(blk->id));
kernel/printk/printk_ringbuffer.c
1228
blk_lpos->next = next_lpos;
kernel/printk/printk_ringbuffer.c
1241
if (!is_blk_wrapped(data_ring, blk_lpos->begin, blk_lpos->next)) {
kernel/printk/printk_ringbuffer.c
1243
return (DATA_INDEX(data_ring, blk_lpos->next) -
kernel/printk/printk_ringbuffer.c
1251
return (DATA_INDEX(data_ring, blk_lpos->next) +
kernel/printk/printk_ringbuffer.c
1278
blk_lpos->next == EMPTY_LINE_LPOS) {
kernel/printk/printk_ringbuffer.c
1288
if (!is_blk_wrapped(data_ring, blk_lpos->begin, blk_lpos->next)) {
kernel/printk/printk_ringbuffer.c
1290
*data_size = blk_lpos->next - blk_lpos->begin;
kernel/printk/printk_ringbuffer.c
1295
blk_lpos->next)) {
kernel/printk/printk_ringbuffer.c
1297
*data_size = DATA_INDEX(data_ring, blk_lpos->next);
kernel/printk/printk_ringbuffer.c
1311
WARN_ON_ONCE(blk_lpos->next != ALIGN(blk_lpos->next, sizeof(db->id)))) {
kernel/printk/printk_ringbuffer.c
1851
const char *next = text;
kernel/printk/printk_ringbuffer.c
1854
next = memchr(next, '\n', next_size);
kernel/printk/printk_ringbuffer.c
1855
if (!next)
kernel/printk/printk_ringbuffer.c
1858
next++;
kernel/printk/printk_ringbuffer.c
1859
next_size = text_size - (next - text);
kernel/printk/printk_ringbuffer.c
1951
(blk_lpos->begin == FAILED_LPOS && blk_lpos->next == FAILED_LPOS)) {
kernel/printk/printk_ringbuffer.c
2389
descs[_DESCS_COUNT(descbits) - 1].text_blk_lpos.next = FAILED_LPOS;
kernel/printk/printk_ringbuffer.c
331
LPOS_DATALESS((blk)->next))
kernel/printk/printk_ringbuffer.c
638
lpos_begin = blk_lpos->next;
kernel/printk/printk_ringbuffer.c
830
if (!data_push_tail(rb, desc.text_blk_lpos.next))
kernel/printk/printk_ringbuffer.h
158
.next = FAILED_LPOS, \
kernel/printk/printk_ringbuffer.h
58
unsigned long next;
kernel/rcu/rcu.h
269
return rhp->next == rhp;
kernel/rcu/rcu_segcblist.c
31
rclp->tail = &rhp->next;
kernel/rcu/rcu_segcblist.c
334
rhp->next = NULL;
kernel/rcu/rcu_segcblist.c
336
WRITE_ONCE(rsclp->tails[RCU_NEXT_TAIL], &rhp->next);
kernel/rcu/rcu_segcblist.c
358
rhp->next = NULL;
kernel/rcu/rcu_segcblist.c
365
WRITE_ONCE(rsclp->tails[i], &rhp->next);
kernel/rcu/rcu_segcblist.c
56
rhp->next = NULL;
kernel/rcu/rcu_segcblist.c
58
srclp->tail = &rhp->next;
kernel/rcu/rcu_segcblist.c
75
rclp->head = rhp->next;
kernel/rcu/rcutorture.c
353
p = rcu_torture_freelist.next;
kernel/rcu/srcutiny.c
161
lh = lh->next;
kernel/rcu/srcutiny.c
225
rhp->next = NULL;
kernel/rcu/srcutiny.c
229
ssp->srcu_cb_tail = &rhp->next;
kernel/rcu/srcutree.c
105
sdp->srcu_barrier_head.next = &sdp->srcu_barrier_head;
kernel/rcu/srcutree.c
1663
rhp->next = rhp; // Mark the callback as having been invoked.
kernel/rcu/tasks.h
1075
t1 = list_entry(tmp.next, struct task_struct, rcu_tasks_exit_list);
kernel/rcu/tasks.h
1297
WARN_ON_ONCE(!rtpcp->rtp_exit_list.next);
kernel/rcu/tasks.h
280
if (!rtpcp->rtp_blkd_tasks.next)
kernel/rcu/tasks.h
282
if (!rtpcp->rtp_exit_list.next)
kernel/rcu/tasks.h
284
rtpcp->barrier_q_head.next = &rtpcp->barrier_q_head;
kernel/rcu/tasks.h
354
rhp->next = NULL;
kernel/rcu/tasks.h
410
rhp->next = rhp; // Mark the callback as having been invoked.
kernel/rcu/tiny.c
101
struct rcu_head *next, *list;
kernel/rcu/tiny.c
121
next = list->next;
kernel/rcu/tiny.c
122
prefetch(next);
kernel/rcu/tiny.c
125
list = next;
kernel/rcu/tiny.c
172
head->next = NULL;
kernel/rcu/tiny.c
176
rcu_ctrlblk.curtail = &head->next;
kernel/rcu/tree.c
1657
struct llist_node *done, *rcu, *next, *head;
kernel/rcu/tree.c
1675
head = done->next;
kernel/rcu/tree.c
1676
done->next = NULL;
kernel/rcu/tree.c
1687
llist_for_each_safe(rcu, next, head) {
kernel/rcu/tree.c
1705
struct llist_node *wait_tail, *next = NULL, *rcu = NULL;
kernel/rcu/tree.c
1719
llist_for_each_safe(rcu, next, wait_tail->next) {
kernel/rcu/tree.c
1725
wait_tail->next = next;
kernel/rcu/tree.c
1738
if (wait_tail->next && wait_tail->next->next == NULL &&
kernel/rcu/tree.c
1739
rcu_sr_is_wait_head(wait_tail->next) &&
kernel/rcu/tree.c
1741
rcu_sr_put_wait_head(wait_tail->next);
kernel/rcu/tree.c
1742
wait_tail->next = NULL;
kernel/rcu/tree.c
1754
if (wait_tail->next) {
kernel/rcu/tree.c
3130
head->next = NULL;
kernel/rcu/tree.c
3742
rhp->next = rhp; // Mark the callback as having been invoked.
kernel/rcu/tree.c
4729
per_cpu_ptr(&rcu_data, i)->barrier_head.next =
kernel/rcu/tree_exp.h
150
WRITE_ONCE(rnp->exp_tasks, rnp->blkd_tasks.next);
kernel/rcu/tree_plugin.h
457
np = t->rcu_node_entry.next;
kernel/rcu/tree_plugin.h
794
WRITE_ONCE(rnp->gp_tasks, rnp->blkd_tasks.next);
kernel/resource.c
1474
struct resource *child, *next, **p;
kernel/resource.c
1478
next = child->sibling;
kernel/resource.c
1481
*p = next;
kernel/resource.c
155
.next = r_next,
kernel/resource.c
749
goto next;
kernel/resource.c
775
next: if (!this || this->end == root->end)
kernel/resource.c
926
struct resource *first, *next;
kernel/resource.c
944
for (next = first; ; next = next->sibling) {
kernel/resource.c
946
if (next->start < new->start || next->end > new->end)
kernel/resource.c
947
return next;
kernel/resource.c
948
if (!next->sibling)
kernel/resource.c
950
if (next->sibling->start > new->end)
kernel/resource.c
955
new->sibling = next->sibling;
kernel/resource.c
958
next->sibling = NULL;
kernel/resource.c
959
for (next = first; next; next = next->sibling)
kernel/resource.c
960
next->parent = new;
kernel/resource.c
965
next = parent->child;
kernel/resource.c
966
while (next->sibling != first)
kernel/resource.c
967
next = next->sibling;
kernel/resource.c
968
next->sibling = new;
kernel/scftorture.c
175
node = node->next;
kernel/sched/autogroup.c
235
static unsigned long next = INITIAL_JIFFIES;
kernel/sched/autogroup.c
251
if (!capable(CAP_SYS_ADMIN) && time_before(jiffies, next))
kernel/sched/autogroup.c
254
next = HZ / 10 + jiffies;
kernel/sched/core.c
1031
if (unlikely(cmpxchg_relaxed(&node->next, NULL, WAKE_Q_TAIL)))
kernel/sched/core.c
1038
head->lastp = &node->next;
kernel/sched/core.c
1091
node = node->next;
kernel/sched/core.c
1093
WRITE_ONCE(task->wake_q.next, NULL);
kernel/sched/core.c
4858
struct task_struct *next)
kernel/sched/core.c
4863
notifier->ops->sched_out(notifier, next);
kernel/sched/core.c
4868
struct task_struct *next)
kernel/sched/core.c
4871
__fire_sched_out_preempt_notifiers(curr, next);
kernel/sched/core.c
4882
struct task_struct *next)
kernel/sched/core.c
4888
static inline void prepare_task(struct task_struct *next)
kernel/sched/core.c
4897
WRITE_ONCE(next->on_cpu, 1);
kernel/sched/core.c
4919
struct balance_callback *next;
kernel/sched/core.c
4925
next = head->next;
kernel/sched/core.c
4926
head->next = NULL;
kernel/sched/core.c
4927
head = next;
kernel/sched/core.c
4947
.next = NULL,
kernel/sched/core.c
5002
prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf)
kernel/sched/core.c
5016
rq_lockp(rq)->owner = next;
kernel/sched/core.c
5043
# define prepare_arch_switch(next) do { } while (0)
kernel/sched/core.c
5081
struct task_struct *next)
kernel/sched/core.c
5085
sched_info_switch(rq, prev, next);
kernel/sched/core.c
5086
perf_event_task_sched_out(prev, next);
kernel/sched/core.c
5087
fire_sched_out_preempt_notifiers(prev, next);
kernel/sched/core.c
5089
prepare_task(next);
kernel/sched/core.c
5090
prepare_arch_switch(next);
kernel/sched/core.c
5240
struct task_struct *next, struct rq_flags *rf)
kernel/sched/core.c
5243
prepare_task_switch(rq, prev, next);
kernel/sched/core.c
5259
if (!next->mm) { // to kernel
kernel/sched/core.c
5260
enter_lazy_tlb(prev->active_mm, next);
kernel/sched/core.c
5262
next->active_mm = prev->active_mm;
kernel/sched/core.c
5268
membarrier_switch_mm(rq, prev->active_mm, next->mm);
kernel/sched/core.c
5277
switch_mm_irqs_off(prev->active_mm, next->mm, next);
kernel/sched/core.c
5278
lru_gen_use_mm(next->mm);
kernel/sched/core.c
5287
mm_cid_switch_to(prev, next);
kernel/sched/core.c
5293
rseq_sched_switch_event(next);
kernel/sched/core.c
5295
prepare_lock_switch(rq, next, rf);
kernel/sched/core.c
5298
switch_to(prev, next, prev);
kernel/sched/core.c
6013
struct task_struct *next, *p, *max;
kernel/sched/core.c
6053
next = rq->core_pick;
kernel/sched/core.c
6099
next = pick_task(rq, rf);
kernel/sched/core.c
6100
if (unlikely(next == RETRY_TASK))
kernel/sched/core.c
6102
if (!next->core_cookie) {
kernel/sched/core.c
6110
task_vruntime_update(rq, next, false);
kernel/sched/core.c
6183
next = rq->core_pick;
kernel/sched/core.c
6187
WARN_ON_ONCE(!next);
kernel/sched/core.c
6230
WARN_ON_ONCE(!cookie_match(next, rq_i->core_pick));
kernel/sched/core.c
6242
put_prev_set_next_task(rq, prev, next);
kernel/sched/core.c
6243
if (rq->core->core_forceidle_count && next == rq->idle)
kernel/sched/core.c
6246
return next;
kernel/sched/core.c
6272
goto next;
kernel/sched/core.c
6275
goto next;
kernel/sched/core.c
6278
goto next;
kernel/sched/core.c
6286
goto next;
kernel/sched/core.c
6294
next:
kernel/sched/core.c
6766
struct task_struct *prev, *next;
kernel/sched/core.c
6835
next = prev;
kernel/sched/core.c
6852
next = pick_next_task(rq, rq->donor, &rf);
kernel/sched/core.c
6853
rq_set_donor(rq, next);
kernel/sched/core.c
6854
rq->next_class = next->sched_class;
kernel/sched/core.c
6855
if (unlikely(task_is_blocked(next))) {
kernel/sched/core.c
6856
next = find_proxy_task(rq, next, &rf);
kernel/sched/core.c
6857
if (!next)
kernel/sched/core.c
6859
if (next == rq->idle)
kernel/sched/core.c
6868
is_switch = prev != next;
kernel/sched/core.c
6875
RCU_INIT_POINTER(rq->curr, next);
kernel/sched/core.c
6877
if (!task_current_donor(rq, next))
kernel/sched/core.c
6878
proxy_tag_curr(rq, next);
kernel/sched/core.c
6904
psi_account_irqtime(rq, prev, next);
kernel/sched/core.c
6905
psi_sched_switch(prev, next, !task_on_rq_queued(prev) ||
kernel/sched/core.c
6908
trace_sched_switch(preempt, prev, next, prev_state);
kernel/sched/core.c
6911
rq = context_switch(rq, prev, next, &rf);
kernel/sched/core.c
6914
if (!task_current_donor(rq, next))
kernel/sched/core.c
6915
proxy_tag_curr(rq, next);
kernel/sched/deadline.c
2051
dl_rq->earliest_dl.next = 0;
kernel/sched/deadline.c
2634
static void put_prev_task_dl(struct rq *rq, struct task_struct *p, struct task_struct *next)
kernel/sched/deadline.c
3015
src_rq->dl.earliest_dl.next))
kernel/sched/deadline.c
3465
goto next;
kernel/sched/deadline.c
3475
next:
kernel/sched/deadline.c
521
dl_rq->earliest_dl.curr = dl_rq->earliest_dl.next = 0;
kernel/sched/deadline.c
588
rq->dl.earliest_dl.next = p->dl.deadline;
kernel/sched/deadline.c
607
dl_rq->earliest_dl.next = __node_2_pdl(leftmost)->dl.deadline;
kernel/sched/debug.c
1213
.next = sched_debug_next,
kernel/sched/ext.c
2381
static void switch_class(struct rq *rq, struct task_struct *next)
kernel/sched/ext.c
2384
const struct sched_class *next_class = next->sched_class;
kernel/sched/ext.c
2411
.task = next,
kernel/sched/ext.c
2422
struct task_struct *next)
kernel/sched/ext.c
2456
if (next && sched_class_above(&ext_sched_class, next->sched_class)) {
kernel/sched/ext.c
2465
if (next && next->sched_class != &ext_sched_class)
kernel/sched/ext.c
2466
switch_class(rq, next);
kernel/sched/ext.c
459
list_node = list_node->next;
kernel/sched/ext.c
4759
goto next;
kernel/sched/ext.c
4807
goto next;
kernel/sched/ext.c
4824
next:
kernel/sched/fair.c
10177
group = group->next;
kernel/sched/fair.c
1028
cfs_rq->next && entity_eligible(cfs_rq, cfs_rq->next)) {
kernel/sched/fair.c
1030
WARN_ON_ONCE(cfs_rq->next->sched_delayed);
kernel/sched/fair.c
1031
return cfs_rq->next;
kernel/sched/fair.c
10938
} while (group = group->next, group != sd->groups);
kernel/sched/fair.c
11173
sg = sg->next;
kernel/sched/fair.c
12179
unsigned long interval, next;
kernel/sched/fair.c
12183
next = sd->last_balance + interval;
kernel/sched/fair.c
12185
if (time_after(*next_balance, next))
kernel/sched/fair.c
12186
*next_balance = next;
kernel/sched/fair.c
3379
work->next = work;
kernel/sched/fair.c
3637
p->numa_work.next = &p->numa_work;
kernel/sched/fair.c
3678
if (!curr->mm || (curr->flags & (PF_EXITING | PF_KTHREAD)) || work->next != work)
kernel/sched/fair.c
4241
struct cfs_rq *prev, struct cfs_rq *next)
kernel/sched/fair.c
4260
n_last_update_time = cfs_rq_last_update_time(next);
kernel/sched/fair.c
5348
if (cfs_rq->next != se)
kernel/sched/fair.c
5351
cfs_rq->next = NULL;
kernel/sched/fair.c
5357
if (cfs_rq->next == se)
kernel/sched/fair.c
5781
p->sched_throttle_work.next = &p->sched_throttle_work;
kernel/sched/fair.c
5822
p->sched_throttle_work.next = &p->sched_throttle_work;
kernel/sched/fair.c
5949
return p->sched_throttle_work.next != &p->sched_throttle_work;
kernel/sched/fair.c
6187
goto next;
kernel/sched/fair.c
6191
goto next;
kernel/sched/fair.c
6223
next:
kernel/sched/fair.c
6602
if (!cfs_b->throttled_cfs_rq.next)
kernel/sched/fair.c
8422
for (; pd; pd = pd->next) {
kernel/sched/fair.c
8736
cfs_rq_of(se)->next = se;
kernel/sched/fair.c
8757
if (cfs_rq->next && entity_before(cfs_rq->next, pse))
kernel/sched/fair.c
9077
static void put_prev_task_fair(struct rq *rq, struct task_struct *prev, struct task_struct *next)
kernel/sched/fair.c
9375
(&p->se == cfs_rq_of(&p->se)->next))
kernel/sched/fair.c
9692
goto next;
kernel/sched/fair.c
9707
goto next;
kernel/sched/fair.c
9716
goto next;
kernel/sched/fair.c
9725
goto next;
kernel/sched/fair.c
9737
goto next;
kernel/sched/fair.c
9766
next:
kernel/sched/idle.c
480
static void put_prev_task_idle(struct rq *rq, struct task_struct *prev, struct task_struct *next)
kernel/sched/idle.c
487
static void set_next_task_idle(struct rq *rq, struct task_struct *next, bool first)
kernel/sched/idle.c
492
next->se.exec_start = rq_clock_task(rq);
kernel/sched/psi.c
926
void psi_task_switch(struct task_struct *prev, struct task_struct *next,
kernel/sched/psi.c
936
if (next->pid) {
kernel/sched/psi.c
937
psi_flags_change(next, 0, TSK_ONCPU);
kernel/sched/psi.c
943
for_each_group(group, task_psi_group(next)) {
kernel/sched/psi.c
996
if ((prev->psi_flags ^ next->psi_flags) & ~TSK_ONCPU) {
kernel/sched/rt.c
1674
struct sched_rt_entity *next = NULL;
kernel/sched/rt.c
1684
next = list_entry(queue->next, struct sched_rt_entity, run_list);
kernel/sched/rt.c
1686
return next;
kernel/sched/rt.c
1716
static void put_prev_task_rt(struct rq *rq, struct task_struct *p, struct task_struct *next)
kernel/sched/rt.c
2110
int next;
kernel/sched/rt.c
2148
next = atomic_read_acquire(&rd->rto_loop_next);
kernel/sched/rt.c
2150
if (rd->rto_loop == next)
kernel/sched/rt.c
2153
rd->rto_loop = next;
kernel/sched/rt.c
2282
if (src_rq->rt.highest_prio.next >=
kernel/sched/rt.c
2494
unsigned long next;
kernel/sched/rt.c
2501
next = DIV_ROUND_UP(min(soft, hard), USEC_PER_SEC/HZ);
kernel/sched/rt.c
2502
if (p->rt.timeout > next) {
kernel/sched/rt.c
2546
if (rt_se->run_list.prev != rt_se->run_list.next) {
kernel/sched/rt.c
404
if (p->prio < rq->rt.highest_prio.next)
kernel/sched/rt.c
405
rq->rt.highest_prio.next = p->prio;
kernel/sched/rt.c
421
rq->rt.highest_prio.next = p->prio;
kernel/sched/rt.c
423
rq->rt.highest_prio.next = MAX_RT_PRIO-1;
kernel/sched/rt.c
501
tg = list_entry_rcu(tg->list.next,
kernel/sched/rt.c
633
goto next;
kernel/sched/rt.c
651
next:
kernel/sched/rt.c
82
rt_rq->highest_prio.next = MAX_RT_PRIO-1;
kernel/sched/sched.h
2078
if (unlikely(head->next || rq->balance_callback == &balance_push_callback))
kernel/sched/sched.h
2082
head->next = rq->balance_callback;
kernel/sched/sched.h
2185
struct sched_group *next; /* Must be a circular list */
kernel/sched/sched.h
2567
void (*put_prev_task)(struct rq *rq, struct task_struct *p, struct task_struct *next);
kernel/sched/sched.h
2669
static inline void set_next_task(struct rq *rq, struct task_struct *next)
kernel/sched/sched.h
2671
next->sched_class->set_next_task(rq, next, false);
kernel/sched/sched.h
2677
struct task_struct *next)
kernel/sched/sched.h
2680
next->dl_server = rq->dl_server;
kernel/sched/sched.h
2686
struct task_struct *next)
kernel/sched/sched.h
2690
__put_prev_set_next_dl_server(rq, prev, next);
kernel/sched/sched.h
2692
if (next == prev)
kernel/sched/sched.h
2695
prev->sched_class->put_prev_task(rq, prev, next);
kernel/sched/sched.h
2696
next->sched_class->set_next_task(rq, next, true);
kernel/sched/sched.h
3958
static __always_inline void mm_cid_schedin(struct task_struct *next)
kernel/sched/sched.h
3960
struct mm_struct *mm = next->mm;
kernel/sched/sched.h
3963
if (!next->mm_cid.active)
kernel/sched/sched.h
3969
mm_cid_from_task(next, cpu_cid, mode);
kernel/sched/sched.h
3971
mm_cid_from_cpu(next, cpu_cid, mode);
kernel/sched/sched.h
4003
static inline void mm_cid_switch_to(struct task_struct *prev, struct task_struct *next)
kernel/sched/sched.h
4006
mm_cid_schedin(next);
kernel/sched/sched.h
4010
static inline void mm_cid_switch_to(struct task_struct *prev, struct task_struct *next) { }
kernel/sched/sched.h
615
struct cfs_rq *prev, struct cfs_rq *next);
kernel/sched/sched.h
673
struct balance_callback *next;
kernel/sched/sched.h
701
struct sched_entity *next;
kernel/sched/sched.h
837
int next; /* next highest */
kernel/sched/sched.h
880
u64 next;
kernel/sched/sched.h
978
struct perf_domain *next;
kernel/sched/stats.c
210
.next = schedstat_next,
kernel/sched/stats.h
111
void psi_task_switch(struct task_struct *prev, struct task_struct *next,
kernel/sched/stats.h
219
struct task_struct *next,
kernel/sched/stats.h
225
psi_task_switch(prev, next, sleep);
kernel/sched/stats.h
233
struct task_struct *next,
kernel/sched/stats.h
328
sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next)
kernel/sched/stats.h
338
if (next != rq->idle)
kernel/sched/stats.h
339
sched_info_arrive(rq, next);
kernel/sched/stats.h
345
# define sched_info_switch(rq, t, next) do { } while (0)
kernel/sched/stop_task.c
61
static void put_prev_task_stop(struct rq *rq, struct task_struct *prev, struct task_struct *next)
kernel/sched/topology.c
1109
last->next = sg;
kernel/sched/topology.c
1111
last->next = first;
kernel/sched/topology.c
122
group = group->next;
kernel/sched/topology.c
1267
last->next = sg;
kernel/sched/topology.c
1270
last->next = first;
kernel/sched/topology.c
1308
goto next;
kernel/sched/topology.c
1318
next:
kernel/sched/topology.c
1319
sg = sg->next;
kernel/sched/topology.c
1472
struct asym_cap_data *entry, *next;
kernel/sched/topology.c
1481
list_for_each_entry_safe(entry, next, &asym_cap_list, link) {
kernel/sched/topology.c
177
(sd->groups != sd->groups->next))
kernel/sched/topology.c
199
if (parent->groups == parent->groups->next)
kernel/sched/topology.c
2412
sg->next = sg;
kernel/sched/topology.c
325
tmp = pd->next;
kernel/sched/topology.c
336
pd = pd->next;
kernel/sched/topology.c
374
pd = pd->next;
kernel/sched/topology.c
431
tmp->next = pd;
kernel/sched/topology.c
613
tmp = sg->next;
kernel/sched/wait.c
104
list_for_each_entry_safe_from(curr, next, &wq_head->head, entry) {
kernel/sched/wait.c
95
wait_queue_entry_t *curr, *next;
kernel/signal.c
484
q = list_entry(queue->list.next, struct sigqueue , list);
kernel/signal.c
688
goto next;
kernel/signal.c
692
next:
kernel/smp.c
540
prev->next = &csd_next->node.llist;
kernel/smp.c
566
prev->next = &csd_next->node.llist;
kernel/softirq.c
818
t->next = NULL;
kernel/softirq.c
820
head->tail = &(t->next);
kernel/softirq.c
918
list = list->next;
kernel/softirq.c
941
t->next = NULL;
kernel/softirq.c
943
tl_head->tail = &t->next;
kernel/softirq.c
965
t->next = NULL;
kernel/softirq.c
977
t->next = NULL;
kernel/static_call_inline.c
157
.next = static_call_key_next(key),
kernel/static_call_inline.c
162
for (site_mod = &first; site_mod; site_mod = site_mod->next) {
kernel/static_call_inline.c
269
site_mod->next = NULL;
kernel/static_call_inline.c
281
site_mod->next = static_call_key_next(key);
kernel/static_call_inline.c
431
prev = &site_mod->next, site_mod = site_mod->next)
kernel/static_call_inline.c
437
*prev = site_mod->next;
kernel/task_work.c
136
pprev = &work->next;
kernel/task_work.c
138
} else if (try_cmpxchg(pprev, &work, work->next))
kernel/task_work.c
203
struct callback_head *work, *head, *next;
kernel/task_work.c
232
next = work->next;
kernel/task_work.c
234
work = next;
kernel/task_work.c
77
work->next = head;
kernel/time/alarmtimer.c
240
struct timerqueue_node *next;
kernel/time/alarmtimer.c
244
next = timerqueue_getnext(&base->timerqueue);
kernel/time/alarmtimer.c
245
if (!next)
kernel/time/alarmtimer.c
247
delta = ktime_sub(next->expires, base->get_ktime());
kernel/time/alarmtimer.c
249
expires = next->expires;
kernel/time/clockevents.c
356
dev = list_entry(clockevents_released.next,
kernel/time/hrtimer.c
521
struct timerqueue_node *next;
kernel/time/hrtimer.c
524
next = timerqueue_getnext(&base->active);
kernel/time/hrtimer.c
525
timer = container_of(next, struct hrtimer, node);
kernel/time/hrtimer.c
528
next = timerqueue_iterate_next(next);
kernel/time/hrtimer.c
529
if (!next)
kernel/time/hrtimer.c
532
timer = container_of(next, struct hrtimer, node);
kernel/time/hrtimer.c
916
struct timerqueue_node *next;
kernel/time/hrtimer.c
918
next = timerqueue_getnext(&base->active);
kernel/time/hrtimer.c
919
expires = ktime_sub(next->expires, base->offset);
kernel/time/posix-cpu-timers.c
1297
struct k_itimer *timer, *next;
kernel/time/posix-cpu-timers.c
1370
list_for_each_entry_safe(timer, next, &firing, it.cpu.elist) {
kernel/time/posix-cpu-timers.c
803
struct timerqueue_node *next;
kernel/time/posix-cpu-timers.c
806
while ((next = timerqueue_getnext(head))) {
kernel/time/posix-cpu-timers.c
810
ctmr = container_of(next, struct cpu_timer, node);
kernel/time/posix-timers.c
1073
struct hlist_node *next;
kernel/time/posix-timers.c
1087
hlist_for_each_entry_safe(timer, next, &timers, list) {
kernel/time/tick-broadcast.c
1007
ktime_t next;
kernel/time/tick-broadcast.c
1016
next = tick_next_period;
kernel/time/tick-broadcast.c
1018
return next;
kernel/time/tick-broadcast.c
416
ktime_t next = ktime_add_ns(dev->next_event, TICK_NSEC);
kernel/time/tick-broadcast.c
418
clockevents_program_event(dev, next, true);
kernel/time/tick-common.c
111
ktime_t next = dev->next_event;
kernel/time/tick-common.c
130
next = ktime_add_ns(next, TICK_NSEC);
kernel/time/tick-common.c
132
if (!clockevents_program_event(dev, next, false))
kernel/time/tick-common.c
164
ktime_t next;
kernel/time/tick-common.c
168
next = tick_next_period;
kernel/time/tick-common.c
174
if (!clockevents_program_event(dev, next, false))
kernel/time/tick-common.c
176
next = ktime_add_ns(next, TICK_NSEC);
kernel/time/timer.c
1859
unsigned long clk, next, adj;
kernel/time/timer.c
1862
next = base->clk + TIMER_NEXT_MAX_DELTA;
kernel/time/timer.c
1872
if (time_before(tmp, next))
kernel/time/timer.c
1873
next = tmp;
kernel/time/timer.c
1923
WRITE_ONCE(base->next_expiry, next);
kernel/time/timer.c
1925
base->timers_pending = !(next == base->clk + TIMER_NEXT_MAX_DELTA);
kernel/time/timer.c
692
timer->entry.next == TIMER_ENTRY_STATIC);
kernel/time/timer.c
894
entry->next = LIST_POISON2;
kernel/time/timer_list.c
346
.next = timer_list_next,
kernel/time/timer_list.c
64
unsigned long next = 0, i;
kernel/time/timer_list.c
80
while (curr && i < next) {
kernel/time/timer_list.c
92
next++;
kernel/trace/bpf_trace.c
724
.next = ctx_size ? &frag : NULL,
kernel/trace/fgraph.c
1089
struct task_struct *next,
kernel/trace/fgraph.c
1106
if (!next->ftrace_timestamp)
kernel/trace/fgraph.c
1109
next->ftrace_sleeptime += timestamp - next->ftrace_timestamp;
kernel/trace/ftrace.c
1143
struct ftrace_page *next;
kernel/trace/ftrace.c
1599
for (pg = ftrace_pages_start; pg; pg = pg->next) { \
kernel/trace/ftrace.c
1630
for (pg = ftrace_pages_start; pg; pg = pg->next) {
kernel/trace/ftrace.c
1724
ops != &ftrace_list_end; ops = ops->next) {
kernel/trace/ftrace.c
230
} else if (rcu_dereference_protected(ftrace_ops_list->next,
kernel/trace/ftrace.c
2821
iter->pg = iter->pg->next;
kernel/trace/ftrace.c
2840
iter->pg = iter->pg->next;
kernel/trace/ftrace.c
2845
iter->pg = iter->pg->next;
kernel/trace/ftrace.c
289
rcu_assign_pointer(ops->next, *list);
kernel/trace/ftrace.c
311
rcu_dereference_protected(ops->next,
kernel/trace/ftrace.c
317
for (p = list; *p != &ftrace_list_end; p = &(*p)->next)
kernel/trace/ftrace.c
324
*p = (*p)->next;
kernel/trace/ftrace.c
3815
for (pg = new_pgs; pg; pg = pg->next) {
kernel/trace/ftrace.c
3895
pages = pg->next;
kernel/trace/ftrace.c
3932
pg->next = kzalloc_obj(*pg);
kernel/trace/ftrace.c
3933
if (!pg->next)
kernel/trace/ftrace.c
3936
pg = pg->next;
kernel/trace/ftrace.c
3974
struct list_head *next;
kernel/trace/ftrace.c
3990
next = func_probes->next;
kernel/trace/ftrace.c
3991
iter->probe = list_entry(next, struct ftrace_func_probe, list);
kernel/trace/ftrace.c
4010
if (iter->probe->list.next == func_probes)
kernel/trace/ftrace.c
4012
next = iter->probe->list.next;
kernel/trace/ftrace.c
4013
iter->probe = list_entry(next, struct ftrace_func_probe, list);
kernel/trace/ftrace.c
4030
hnd = hnd->next;
kernel/trace/ftrace.c
4107
iter->mod_list = iter->mod_list->next;
kernel/trace/ftrace.c
4184
if (iter->pg->next) {
kernel/trace/ftrace.c
4185
iter->pg = iter->pg->next;
kernel/trace/ftrace.c
426
struct ftrace_profile_page *next;
kernel/trace/ftrace.c
4551
.next = t_next,
kernel/trace/ftrace.c
468
pg = pg->next;
kernel/trace/ftrace.c
5775
char *func, *command, *next = buff;
kernel/trace/ftrace.c
5779
func = strsep(&next, ":");
kernel/trace/ftrace.c
5781
if (!next) {
kernel/trace/ftrace.c
5792
command = strsep(&next, ":");
kernel/trace/ftrace.c
5798
return p->func(tr, hash, func, command, next, enable);
kernel/trace/ftrace.c
623
pg = pg->next;
kernel/trace/ftrace.c
663
pg->next = (void *)get_zeroed_page(GFP_KERNEL);
kernel/trace/ftrace.c
664
if (!pg->next)
kernel/trace/ftrace.c
666
pg = pg->next;
kernel/trace/ftrace.c
6722
char *mod = NULL, *func, *command, *next = buf;
kernel/trace/ftrace.c
6727
func = strsep(&next, ":");
kernel/trace/ftrace.c
6730
if (next) {
kernel/trace/ftrace.c
6734
command = strsep(&next, ":");
kernel/trace/ftrace.c
6738
mod = next;
kernel/trace/ftrace.c
676
pg = pg->next;
kernel/trace/ftrace.c
7157
.next = g_next,
kernel/trace/ftrace.c
7591
if (WARN_ON(ftrace_pages->next)) {
kernel/trace/ftrace.c
7593
while (ftrace_pages->next)
kernel/trace/ftrace.c
7594
ftrace_pages = ftrace_pages->next;
kernel/trace/ftrace.c
7597
ftrace_pages->next = start_pg;
kernel/trace/ftrace.c
7634
if (WARN_ON(!pg->next))
kernel/trace/ftrace.c
7636
pg = pg->next;
kernel/trace/ftrace.c
7643
if (pg->next) {
kernel/trace/ftrace.c
7644
pg_unuse = pg->next;
kernel/trace/ftrace.c
7645
pg->next = NULL;
kernel/trace/ftrace.c
7680
for (pg = pg_unuse; pg && skip > 0; pg = pg->next) {
kernel/trace/ftrace.c
7780
#define next_to_ftrace_page(p) container_of(p, struct ftrace_page, next)
kernel/trace/ftrace.c
7789
for (ops = ftrace_ops_list; ops != &ftrace_list_end; ops = ops->next) {
kernel/trace/ftrace.c
786
if (!stat->pages->next)
kernel/trace/ftrace.c
788
stat->pages = stat->pages->next;
kernel/trace/ftrace.c
7910
*last_pg = pg->next;
kernel/trace/ftrace.c
7912
pg->next = tmp_page;
kernel/trace/ftrace.c
7915
last_pg = &pg->next;
kernel/trace/ftrace.c
7932
tmp_page = pg->next;
kernel/trace/ftrace.c
8282
for (pg = ftrace_pages_start; pg; last_pg = &pg->next, pg = *last_pg) {
kernel/trace/ftrace.c
8302
*last_pg = pg->next;
kernel/trace/ftrace.c
8303
pg->next = tmp_page;
kernel/trace/ftrace.c
8305
pg = container_of(last_pg, struct ftrace_page, next);
kernel/trace/ftrace.c
8590
struct task_struct *next,
kernel/trace/ftrace.c
8600
if (trace_ignore_this_task(pid_list, no_pid_list, next))
kernel/trace/ftrace.c
8605
next->pid);
kernel/trace/ftrace.c
8762
.next = fpid_next,
kernel/trace/ftrace.c
8798
.next = fnpid_next,
kernel/trace/pid_list.c
23
pid_list->lower_list = chunk->next;
kernel/trace/pid_list.c
26
chunk->next = NULL;
kernel/trace/pid_list.c
277
unsigned int *next)
kernel/trace/pid_list.c
316
*next = pid_join(upper1, upper2, lower);
kernel/trace/pid_list.c
366
upper_next = &chunk->next;
kernel/trace/pid_list.c
377
lower_next = &chunk->next;
kernel/trace/pid_list.c
441
chunk->next = pid_list->upper_list;
kernel/trace/pid_list.c
452
chunk->next = pid_list->lower_list;
kernel/trace/pid_list.c
47
pid_list->upper_list = chunk->next;
kernel/trace/pid_list.c
481
pid_list->lower_list = pid_list->lower_list->next;
kernel/trace/pid_list.c
489
pid_list->upper_list = pid_list->upper_list->next;
kernel/trace/pid_list.c
50
chunk->next = NULL;
kernel/trace/pid_list.c
66
chunk->next = pid_list->lower_list;
kernel/trace/pid_list.c
76
chunk->next = pid_list->upper_list;
kernel/trace/pid_list.h
69
union lower_chunk *next;
kernel/trace/pid_list.h
74
union upper_chunk *next;
kernel/trace/rethook.c
214
node = node->next;
kernel/trace/rethook.c
222
node = node->next;
kernel/trace/rethook.c
28
node = node->next;
kernel/trace/rethook.c
317
first = first->next;
kernel/trace/rethook.c
325
current->rethooks.first = node->next;
kernel/trace/rethook.c
326
node->next = NULL;
kernel/trace/rethook.c
330
first = first->next;
kernel/trace/ring_buffer.c
1262
val = (unsigned long)list->next;
kernel/trace/ring_buffer.c
1281
return rb_list_head(list->next) != &page->list;
kernel/trace/ring_buffer.c
1291
ptr = (unsigned long *)&list->next;
kernel/trace/ring_buffer.c
1320
unsigned long *ptr = (unsigned long *)&list->next;
kernel/trace/ring_buffer.c
1353
ret = cmpxchg((unsigned long *)&list->next,
kernel/trace/ring_buffer.c
1392
struct list_head *p = rb_list_head((*bpage)->list.next);
kernel/trace/ring_buffer.c
1417
if (RB_WARN_ON(cpu_buffer, rb_list_head(list->prev->next) != list))
kernel/trace/ring_buffer.c
1445
unsigned long *ptr = (unsigned long *)&old->list.prev->next;
kernel/trace/ring_buffer.c
1530
rb_list_head(rb_list_head(list->next)->prev) != list))
kernel/trace/ring_buffer.c
1534
rb_list_head(rb_list_head(list->prev)->next) != list))
kernel/trace/ring_buffer.c
1594
tmp = rb_list_head(tmp->next);
kernel/trace/ring_buffer.c
1995
cpu_buffer->reader_page->list.next = &orig_head->list;
kernel/trace/ring_buffer.c
1998
bpage->list.next = &cpu_buffer->reader_page->list;
kernel/trace/ring_buffer.c
2006
bpage->list.next = &head_page->list;
kernel/trace/ring_buffer.c
2206
.next = rbm_next,
kernel/trace/ring_buffer.c
2342
cpu_buffer->pages = pages.next;
kernel/trace/ring_buffer.c
2750
tail_page = rb_list_head(tail_page->next);
kernel/trace/ring_buffer.c
2754
first_page = list_entry(rb_list_head(to_remove->next),
kernel/trace/ring_buffer.c
2758
to_remove = rb_list_head(to_remove)->next;
kernel/trace/ring_buffer.c
2764
next_page = rb_list_head(to_remove)->next;
kernel/trace/ring_buffer.c
2771
tail_page->next = (struct list_head *)((unsigned long)next_page |
kernel/trace/ring_buffer.c
2867
first_page = pages->next;
kernel/trace/ring_buffer.c
2873
last_page->next = head_page_with_bit;
kernel/trace/ring_buffer.c
2877
if (try_cmpxchg(&prev_page->next,
kernel/trace/ring_buffer.c
5489
cpu_buffer->reader_page->list.next = rb_list_head(reader->list.next);
kernel/trace/ring_buffer.c
5541
rb_list_head(reader->list.next)->prev = &cpu_buffer->reader_page->list;
kernel/trace/ring_buffer.c
667
struct list_head *next;
kernel/trace/ring_buffer.c
683
next = rb_list_head(page->list.next);
kernel/trace/ring_buffer.c
684
page = list_entry(next, struct buffer_page, list);
kernel/trace/ring_buffer.c
6953
cpu_buffer->reader_page = list_entry(cpu_buffer->new_pages.next,
kernel/trace/ring_buffer.c
6958
cpu_buffer->pages = cpu_buffer->new_pages.next;
kernel/trace/rv/monitors/pagefault/pagefault.h
50
ltl_possible_next_states(struct ltl_monitor *mon, unsigned int state, unsigned long *next)
kernel/trace/rv/monitors/pagefault/pagefault.h
61
__set_bit(S0, next);
kernel/trace/rv/monitors/pagefault/pagefault_trace.h
10
TP_ARGS(task, states, atoms, next));
kernel/trace/rv/monitors/pagefault/pagefault_trace.h
9
TP_PROTO(struct task_struct *task, char *states, char *atoms, char *next),
kernel/trace/rv/monitors/sleep/sleep.h
125
ltl_possible_next_states(struct ltl_monitor *mon, unsigned int state, unsigned long *next)
kernel/trace/rv/monitors/sleep/sleep.h
166
__set_bit(S0, next);
kernel/trace/rv/monitors/sleep/sleep.h
168
__set_bit(S1, next);
kernel/trace/rv/monitors/sleep/sleep.h
170
__set_bit(S4, next);
kernel/trace/rv/monitors/sleep/sleep.h
172
__set_bit(S5, next);
kernel/trace/rv/monitors/sleep/sleep.h
176
__set_bit(S1, next);
kernel/trace/rv/monitors/sleep/sleep.h
178
__set_bit(S2, next);
kernel/trace/rv/monitors/sleep/sleep.h
180
__set_bit(S3, next);
kernel/trace/rv/monitors/sleep/sleep.h
182
__set_bit(S4, next);
kernel/trace/rv/monitors/sleep/sleep.h
184
__set_bit(S6, next);
kernel/trace/rv/monitors/sleep/sleep.h
186
__set_bit(S7, next);
kernel/trace/rv/monitors/sleep/sleep.h
190
__set_bit(S1, next);
kernel/trace/rv/monitors/sleep/sleep.h
192
__set_bit(S2, next);
kernel/trace/rv/monitors/sleep/sleep.h
194
__set_bit(S3, next);
kernel/trace/rv/monitors/sleep/sleep.h
196
__set_bit(S4, next);
kernel/trace/rv/monitors/sleep/sleep.h
198
__set_bit(S6, next);
kernel/trace/rv/monitors/sleep/sleep.h
200
__set_bit(S7, next);
kernel/trace/rv/monitors/sleep/sleep.h
204
__set_bit(S0, next);
kernel/trace/rv/monitors/sleep/sleep.h
206
__set_bit(S1, next);
kernel/trace/rv/monitors/sleep/sleep.h
208
__set_bit(S4, next);
kernel/trace/rv/monitors/sleep/sleep.h
210
__set_bit(S5, next);
kernel/trace/rv/monitors/sleep/sleep.h
214
__set_bit(S0, next);
kernel/trace/rv/monitors/sleep/sleep.h
216
__set_bit(S1, next);
kernel/trace/rv/monitors/sleep/sleep.h
218
__set_bit(S4, next);
kernel/trace/rv/monitors/sleep/sleep.h
220
__set_bit(S5, next);
kernel/trace/rv/monitors/sleep/sleep.h
224
__set_bit(S0, next);
kernel/trace/rv/monitors/sleep/sleep.h
226
__set_bit(S1, next);
kernel/trace/rv/monitors/sleep/sleep.h
228
__set_bit(S4, next);
kernel/trace/rv/monitors/sleep/sleep.h
230
__set_bit(S5, next);
kernel/trace/rv/monitors/sleep/sleep.h
234
__set_bit(S1, next);
kernel/trace/rv/monitors/sleep/sleep.h
236
__set_bit(S2, next);
kernel/trace/rv/monitors/sleep/sleep.h
238
__set_bit(S3, next);
kernel/trace/rv/monitors/sleep/sleep.h
240
__set_bit(S4, next);
kernel/trace/rv/monitors/sleep/sleep.h
242
__set_bit(S6, next);
kernel/trace/rv/monitors/sleep/sleep.h
244
__set_bit(S7, next);
kernel/trace/rv/monitors/sleep/sleep.h
248
__set_bit(S0, next);
kernel/trace/rv/monitors/sleep/sleep.h
250
__set_bit(S1, next);
kernel/trace/rv/monitors/sleep/sleep.h
252
__set_bit(S4, next);
kernel/trace/rv/monitors/sleep/sleep.h
254
__set_bit(S5, next);
kernel/trace/rv/monitors/sleep/sleep_trace.h
10
TP_ARGS(task, states, atoms, next));
kernel/trace/rv/monitors/sleep/sleep_trace.h
9
TP_PROTO(struct task_struct *task, char *states, char *atoms, char *next),
kernel/trace/rv/monitors/snroc/snroc.c
27
struct task_struct *next,
kernel/trace/rv/monitors/snroc/snroc.c
31
da_handle_event(next, sched_switch_in_snroc);
kernel/trace/rv/monitors/sssw/sssw.c
31
struct task_struct *next,
kernel/trace/rv/monitors/sssw/sssw.c
43
da_handle_event(next, sched_switch_in_sssw);
kernel/trace/rv/monitors/sts/sts.c
77
struct task_struct *next,
kernel/trace/rv/rv.c
228
struct rv_monitor *next;
kernel/trace/rv/rv.c
233
next = list_next_entry(mon, list);
kernel/trace/rv/rv.c
235
return next->parent == mon || !mon->enable;
kernel/trace/rv/rv.c
526
.next = available_monitors_next,
kernel/trace/rv/rv.c
533
.next = enabled_monitors_next,
kernel/trace/rv/rv_reactors.c
117
.next = reactors_next,
kernel/trace/rv/rv_reactors.c
157
.next = reactors_next,
kernel/trace/rv/rv_trace.h
135
TP_PROTO(struct task_struct *task, char *states, char *atoms, char *next),
kernel/trace/rv/rv_trace.h
137
TP_ARGS(task, states, atoms, next),
kernel/trace/rv/rv_trace.h
144
__string(next, next)
kernel/trace/rv/rv_trace.h
152
__assign_str(next);
kernel/trace/rv/rv_trace.h
156
__get_str(states), __get_str(atoms), __get_str(next))
kernel/trace/trace.c
10117
last = &map->tail.next;
kernel/trace/trace.c
10118
map = map->tail.next;
kernel/trace/trace.c
10123
*last = trace_eval_jmp_to_tail(map)->tail.next;
kernel/trace/trace.c
1876
for (t = trace_types; t; t = t->next) {
kernel/trace/trace.c
1878
*last = t->next;
kernel/trace/trace.c
1881
last = &t->next;
kernel/trace/trace.c
192
union trace_eval_map_item *next;
kernel/trace/trace.c
1946
for (t = trace_types; t; t = t->next) {
kernel/trace/trace.c
1973
type->next = trace_types;
kernel/trace/trace.c
2848
struct trace_entry *ent, *next = NULL;
kernel/trace/trace.c
2880
if (ent && (!next || ts < next_ts)) {
kernel/trace/trace.c
2881
next = ent;
kernel/trace/trace.c
2900
return next;
kernel/trace/trace.c
3868
.next = s_next,
kernel/trace/trace.c
4192
t = t->next;
kernel/trace/trace.c
4206
t = get_tracer_for_array(tr, t->next);
kernel/trace/trace.c
4239
if (t->next)
kernel/trace/trace.c
4249
.next = t_next,
kernel/trace/trace.c
436
export = rcu_dereference_raw_check(export->next);
kernel/trace/trace.c
443
rcu_assign_pointer(export->next, *list);
kernel/trace/trace.c
458
for (p = list; *p != NULL; p = &(*p)->next)
kernel/trace/trace.c
465
rcu_assign_pointer(*p, (*p)->next);
kernel/trace/trace.c
5023
if (ptr->tail.next) {
kernel/trace/trace.c
5024
ptr = ptr->tail.next;
kernel/trace/trace.c
5088
.next = eval_map_next,
kernel/trace/trace.c
5148
if (!ptr->tail.next)
kernel/trace/trace.c
5150
ptr = ptr->tail.next;
kernel/trace/trace.c
5153
ptr->tail.next = map_array;
kernel/trace/trace.c
6415
.next = l_next,
kernel/trace/trace.c
7691
struct tracing_log_err *err, *next;
kernel/trace/trace.c
7695
list_for_each_entry_safe(err, next, &tr->err_log, list) {
kernel/trace/trace.c
7756
.next = tracing_err_log_seq_next,
kernel/trace/trace.c
9479
for (t = trace_types; t && !ret; t = t->next)
kernel/trace/trace.h
239
unsigned int *next);
kernel/trace/trace.h
650
struct tracer *next;
kernel/trace/trace_dynevent.c
192
.next = dyn_event_seq_next,
kernel/trace/trace_events.c
1084
struct task_struct *next,
kernel/trace/trace_events.c
1100
trace_ignore_this_task(NULL, no_pid_list, next);
kernel/trace/trace_events.c
1104
trace_ignore_this_task(pid_list, NULL, next)));
kernel/trace/trace_events.c
1110
struct task_struct *next,
kernel/trace/trace_events.c
1121
trace_ignore_this_task(pid_list, no_pid_list, next));
kernel/trace/trace_events.c
2161
.next = f_next,
kernel/trace/trace_events.c
222
struct ftrace_event_field *field, *next;
kernel/trace/trace_events.c
226
list_for_each_entry_safe(field, next, head, link) {
kernel/trace/trace_events.c
2597
.next = t_next,
kernel/trace/trace_events.c
2604
.next = s_next,
kernel/trace/trace_events.c
2611
.next = t_next,
kernel/trace/trace_events.c
2618
.next = t_next,
kernel/trace/trace_events.c
2625
.next = p_next,
kernel/trace/trace_events.c
2632
.next = np_next,
kernel/trace/trace_events.c
3422
list_add(&modstr->next, &module_strings);
kernel/trace/trace_events.c
3431
char *attr, *tmp, *next, *ret = (char *)type;
kernel/trace/trace_events.c
3434
next = (char *)type;
kernel/trace/trace_events.c
3435
while ((attr = strstr(next, ATTRIBUTE_STR))) {
kernel/trace/trace_events.c
3437
if (attr != next && !isspace(attr[-1])) {
kernel/trace/trace_events.c
3438
next = attr + ATTRIBUTE_STR_LEN;
kernel/trace/trace_events.c
3451
next = attr + ATTRIBUTE_STR_LEN;
kernel/trace/trace_events.c
3453
tmp = strpbrk(next, "()");
kernel/trace/trace_events.c
3464
next = tmp + 1;
kernel/trace/trace_events.c
3466
next = skip_spaces(next);
kernel/trace/trace_events.c
3467
strcpy(attr, next);
kernel/trace/trace_events.c
3468
next = attr;
kernel/trace/trace_events.c
3943
list_for_each_entry_safe(modstr, m, &module_strings, next) {
kernel/trace/trace_events.c
3946
list_del(&modstr->next);
kernel/trace/trace_events.c
4475
struct trace_event_file *file, *next;
kernel/trace/trace_events.c
4477
list_for_each_entry_safe(file, next, &tr->events, list)
kernel/trace/trace_events.c
46
struct list_head next;
kernel/trace/trace_events_filter.c
508
const char *next = ptr++;
kernel/trace/trace_events_filter.c
510
if (isspace(*next))
kernel/trace/trace_events_filter.c
513
switch (*next) {
kernel/trace/trace_events_filter.c
522
if (!is_not(next))
kernel/trace/trace_events_filter.c
529
parse_error(pe, FILT_ERR_TOO_MANY_PREDS, next - str);
kernel/trace/trace_events_filter.c
536
len = parse_pred(next, data, ptr - str, pe, &prog[N].pred);
kernel/trace/trace_events_filter.c
541
ptr = next + len;
kernel/trace/trace_events_filter.c
547
next = ptr++;
kernel/trace/trace_events_filter.c
548
if (isspace(*next))
kernel/trace/trace_events_filter.c
551
switch (*next) {
kernel/trace/trace_events_filter.c
558
if (next[1] == next[0]) {
kernel/trace/trace_events_filter.c
565
next - str);
kernel/trace/trace_events_filter.c
575
if (*next == '&') { /* #8 */
kernel/trace/trace_events_filter.c
583
if (*next == '|') { /* #10 */
kernel/trace/trace_events_filter.c
587
if (!*next) /* #11 */
kernel/trace/trace_events_synth.c
2286
.next = dyn_event_seq_next,
kernel/trace/trace_events_trigger.c
320
.next = trigger_next,
kernel/trace/trace_events_trigger.c
364
char *command, *next;
kernel/trace/trace_events_trigger.c
367
next = buff = strim(buff);
kernel/trace/trace_events_trigger.c
369
command = strsep(&next, ": \t");
kernel/trace/trace_events_trigger.c
370
if (next) {
kernel/trace/trace_events_trigger.c
371
next = skip_spaces(next);
kernel/trace/trace_events_trigger.c
372
if (!*next)
kernel/trace/trace_events_trigger.c
373
next = NULL;
kernel/trace/trace_events_trigger.c
381
return p->parse(p, file, buff, command, next);
kernel/trace/trace_events_user.c
1088
struct user_event_validator *validator, *next;
kernel/trace/trace_events_user.c
1091
list_for_each_entry_safe(validator, next, head, user_event_link) {
kernel/trace/trace_events_user.c
1099
struct ftrace_event_field *field, *next;
kernel/trace/trace_events_user.c
1102
list_for_each_entry_safe(field, next, head, link) {
kernel/trace/trace_events_user.c
2614
struct user_event_enabler *enabler, *next;
kernel/trace/trace_events_user.c
2639
list_for_each_entry_safe(enabler, next, &mm->enablers, mm_enablers_link) {
kernel/trace/trace_events_user.c
2812
.next = user_seq_next,
kernel/trace/trace_events_user.c
595
struct user_event_mm *next;
kernel/trace/trace_events_user.c
615
next = mm->next;
kernel/trace/trace_events_user.c
627
mm = next;
kernel/trace/trace_events_user.c
693
mm->next = found;
kernel/trace/trace_events_user.c
764
struct user_event_enabler *enabler, *next;
kernel/trace/trace_events_user.c
766
list_for_each_entry_safe(enabler, next, &mm->enablers, mm_enablers_link)
kernel/trace/trace_functions_graph.c
643
struct ftrace_graph_ret_entry *next;
kernel/trace/trace_functions_graph.c
651
next = &data->ret;
kernel/trace/trace_functions_graph.c
673
next = ring_buffer_event_data(event);
kernel/trace/trace_functions_graph.c
688
if (next->ent.type == TRACE_GRAPH_RET)
kernel/trace/trace_functions_graph.c
689
data->ret = *next;
kernel/trace/trace_functions_graph.c
691
data->ret.ent.type = next->ent.type;
kernel/trace/trace_functions_graph.c
695
if (next->ent.type != TRACE_GRAPH_RET)
kernel/trace/trace_functions_graph.c
698
if (curr->ent.pid != next->ent.pid ||
kernel/trace/trace_functions_graph.c
699
curr->graph_ent.func != next->ret.func)
kernel/trace/trace_functions_graph.c
706
return next;
kernel/trace/trace_hwlat.c
648
.next = s_mode_next,
kernel/trace/trace_kprobe.c
1329
.next = dyn_event_seq_next,
kernel/trace/trace_kprobe.c
1395
.next = dyn_event_seq_next,
kernel/trace/trace_osnoise.c
2184
.next = s_options_next,
kernel/trace/trace_output.c
833
int next;
kernel/trace/trace_output.c
836
next = ida_alloc_range(&trace_event_ida, __TRACE_LAST_TYPE,
kernel/trace/trace_output.c
838
if (next < 0)
kernel/trace/trace_output.c
840
return next;
kernel/trace/trace_pid.c
102
pid = next;
kernel/trace/trace_pid.c
94
unsigned int next;
kernel/trace/trace_pid.c
99
if (trace_pid_list_next(pid_list, pid, &next) < 0)
kernel/trace/trace_printk.c
154
if (mod_fmt->list.next == &trace_bprintk_fmt_list)
kernel/trace/trace_printk.c
157
mod_fmt = container_of(mod_fmt->list.next, typeof(*mod_fmt), list);
kernel/trace/trace_printk.c
355
.next = t_next,
kernel/trace/trace_probe.c
557
char *next;
kernel/trace/trace_probe.c
577
next = NULL;
kernel/trace/trace_probe.c
578
is_ptr = split_next_field(fieldname, &next, ctx);
kernel/trace/trace_probe.c
611
ctx->offset += next - fieldname;
kernel/trace/trace_probe.c
612
fieldname = next;
kernel/trace/trace_recursion_record.c
176
.next = recursed_function_seq_next,
kernel/trace/trace_sched_switch.c
26
struct task_struct *prev, struct task_struct *next,
kernel/trace/trace_sched_switch.c
36
tracing_record_taskinfo_sched_switch(prev, next, flags);
kernel/trace/trace_sched_switch.c
405
struct task_struct *next, int flags)
kernel/trace/trace_sched_switch.c
417
done &= !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(next);
kernel/trace/trace_sched_switch.c
419
done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(next);
kernel/trace/trace_sched_switch.c
495
.next = saved_tgids_next,
kernel/trace/trace_sched_switch.c
574
.next = saved_cmdlines_next,
kernel/trace/trace_sched_wakeup.c
390
struct task_struct *next,
kernel/trace/trace_sched_wakeup.c
405
entry->next_pid = next->pid;
kernel/trace/trace_sched_wakeup.c
406
entry->next_prio = next->prio;
kernel/trace/trace_sched_wakeup.c
407
entry->next_state = task_state_index(next);
kernel/trace/trace_sched_wakeup.c
408
entry->next_cpu = task_cpu(next);
kernel/trace/trace_sched_wakeup.c
441
struct task_struct *prev, struct task_struct *next,
kernel/trace/trace_sched_wakeup.c
465
if (next != wakeup_task)
kernel/trace/trace_sched_wakeup.c
480
if (unlikely(!tracer_enabled || next != wakeup_task))
kernel/trace/trace_sched_wakeup.c
487
tracing_sched_switch_trace(wakeup_trace, prev, next, trace_ctx);
kernel/trace/trace_stack.c
471
.next = t_next,
kernel/trace/trace_stat.c
224
.next = stat_seq_next,
kernel/trace/trace_uprobe.c
783
.next = dyn_event_seq_next,
kernel/trace/trace_uprobe.c
846
.next = dyn_event_seq_next,
kernel/ucount.c
280
struct ucounts *iter, *next;
kernel/ucount.c
281
for (iter = ucounts; iter != last; iter = next) {
kernel/ucount.c
284
next = iter->ns->ucounts;
kernel/user_namespace.c
727
.next = m_next,
kernel/user_namespace.c
734
.next = m_next,
kernel/user_namespace.c
741
.next = m_next,
kernel/vmcore_info.c
214
VMCOREINFO_OFFSET(list_head, next);
kernel/watchdog.c
113
next:
kernel/watchdog.c
127
goto next;
kernel/workqueue.c
1041
(worker->hentry.next || worker->hentry.pprev)))
kernel/workqueue.c
3879
head = worker->scheduled.next;
kernel/workqueue.c
3884
head = target->entry.next;
kernel/workqueue.c
4107
struct wq_flusher *next, *tmp;
kernel/workqueue.c
4110
list_for_each_entry_safe(next, tmp, &wq->flusher_queue, list) {
kernel/workqueue.c
4111
if (next->flush_color != wq->flush_color)
kernel/workqueue.c
4113
list_del_init(&next->list);
kernel/workqueue.c
4114
complete(&next->done);
kernel/workqueue.c
4151
WARN_ON_ONCE(wq->flush_color != next->flush_color);
kernel/workqueue.c
4153
list_del_init(&next->list);
kernel/workqueue.c
4154
wq->first_flusher = next;
kernel/workqueue.c
5612
WARN(!ret && (wq->pwqs.next != &dfl_pwq->pwqs_node ||
lib/842/842_compress.c
482
u64 last, next, pad, total;
lib/842/842_compress.c
517
next = get_unaligned((u64 *)p->in);
lib/842/842_compress.c
528
if (next == last) {
lib/842/842_compress.c
538
if (next == last) /* reached max repeat bits */
lib/842/842_compress.c
542
if (next == 0)
lib/842/842_compress.c
551
last = next;
lib/alloc_tag.c
121
.next = allocinfo_next,
lib/bootconfig.c
171
return node->next ? &xbc_nodes[node->next] : NULL;
lib/bootconfig.c
343
struct xbc_node *next;
lib/bootconfig.c
354
next = xbc_node_get_subkey(node);
lib/bootconfig.c
355
if (next) {
lib/bootconfig.c
356
node = next;
lib/bootconfig.c
363
while (!node->next) {
lib/bootconfig.c
419
node->next = 0;
lib/bootconfig.c
440
while (node->next)
lib/bootconfig.c
463
sib->next = xbc_node_index(node);
lib/bootconfig.c
467
node->next = last_parent->child;
lib/bootconfig.c
472
sib->next = xbc_node_index(node);
lib/bootconfig.c
609
char *next;
lib/bootconfig.c
617
next = skip_spaces(*__v);
lib/bootconfig.c
618
while (*next == '#') {
lib/bootconfig.c
619
next = skip_comment(next);
lib/bootconfig.c
620
next = skip_spaces(next);
lib/bootconfig.c
622
*__v = next;
lib/bootconfig.c
623
c = __xbc_parse_value(__v, &next);
lib/bootconfig.c
630
*__v = next;
lib/bootconfig.c
700
char *next;
lib/bootconfig.c
710
next = skip_comment(v);
lib/bootconfig.c
714
c = __xbc_parse_value(&v, &next);
lib/bootconfig.c
724
unsigned short nidx = child->next;
lib/bootconfig.c
728
child->next = nidx; /* keep subkeys */
lib/bootconfig.c
740
c = xbc_parse_array(&next);
lib/bootconfig.c
748
ret = __xbc_close_brace(next - 1);
lib/bootconfig.c
753
*k = next;
lib/bootconfig.c
818
if (xbc_nodes[i].next > xbc_node_num) {
lib/cpumask.c
137
unsigned int next, prev;
lib/cpumask.c
142
next = cpumask_next_and_wrap(prev, src1p, src2p);
lib/cpumask.c
143
if (next < nr_cpu_ids)
lib/cpumask.c
144
__this_cpu_write(distribute_cpu_mask_prev, next);
lib/cpumask.c
146
return next;
lib/cpumask.c
158
unsigned int next, prev;
lib/cpumask.c
162
next = cpumask_next_wrap(prev, srcp);
lib/cpumask.c
163
if (next < nr_cpu_ids)
lib/cpumask.c
164
__this_cpu_write(distribute_cpu_mask_prev, next);
lib/cpumask.c
166
return next;
lib/crypto/mpi/mpi-internal.h
159
struct karatsuba_ctx *next;
lib/crypto/mpi/mpih-mul.c
374
if (!ctx->next) {
lib/crypto/mpi/mpih-mul.c
375
ctx->next = kzalloc_obj(*ctx);
lib/crypto/mpi/mpih-mul.c
376
if (!ctx->next)
lib/crypto/mpi/mpih-mul.c
382
ctx->next) < 0)
lib/crypto/mpi/mpih-mul.c
401
for (ctx = ctx->next; ctx; ctx = ctx2) {
lib/crypto/mpi/mpih-mul.c
402
ctx2 = ctx->next;
lib/debugobjects.c
1051
unsigned int expect, unsigned int next)
lib/debugobjects.c
1071
obj->astate = next;
lib/debugobjects.c
163
next_batch = last->next;
lib/debugobjects.c
171
last->next = dst->objects.first;
lib/debugobjects.c
172
if (last->next)
lib/debugobjects.c
173
last->next->pprev = &last->next;
lib/debugobjects.c
200
struct hlist_node *last, *next;
lib/debugobjects.c
211
next = last->next;
lib/debugobjects.c
213
last->next = NULL;
lib/debugobjects.c
216
src->objects.first = next;
lib/debugobjects.c
217
if (next)
lib/debugobjects.c
218
next->pprev = &src->objects.first;
lib/dynamic_debug.c
1042
iter->table = list_entry(ddebug_tables.next,
lib/dynamic_debug.c
1064
iter->table = list_entry(iter->table->link.next,
lib/dynamic_debug.c
1176
.next = ddebug_proc_next,
lib/dynamic_debug.c
1375
struct ddebug_table *dt = list_entry(ddebug_tables.next,
lib/error-inject.c
208
.next = ei_seq_next,
lib/idr.c
398
next:
lib/idr.c
436
goto next;
lib/iov_iter.c
1511
folioq = folioq->next;
lib/iov_iter.c
1545
if (slot == folioq_nr_slots(folioq) && folioq->next) {
lib/iov_iter.c
1546
folioq = folioq->next;
lib/iov_iter.c
1867
size_t next = min_t(size_t, PAGE_SIZE, left);
lib/iov_iter.c
1872
contig_sz += next;
lib/iov_iter.c
1873
left -= next;
lib/iov_iter.c
551
folioq = folioq->next;
lib/iov_iter.c
563
if (slot >= folioq_nr_slots(folioq) && folioq->next) {
lib/iov_iter.c
564
folioq = folioq->next;
lib/iov_iter.c
923
folioq = folioq->next;
lib/iov_iter.c
958
if (slot == folioq_nr_slots(folioq) && folioq->next) {
lib/iov_iter.c
959
folioq = folioq->next;
lib/irq_poll.c
105
iop = list_entry(list->next, struct irq_poll, list);
lib/klist.c
379
struct klist_node *next;
lib/klist.c
385
next = to_klist_node(last->n_node.next);
lib/klist.c
389
next = to_klist_node(i->i_klist->k_list.next);
lib/klist.c
392
while (next != to_klist_node(&i->i_klist->k_list)) {
lib/klist.c
393
if (likely(!knode_dead(next))) {
lib/klist.c
394
kref_get(&next->n_ref);
lib/klist.c
395
i->i_cur = next;
lib/klist.c
398
next = to_klist_node(next->n_node.next);
lib/kobject_uevent.c
101
const char *next = buf;
lib/kobject_uevent.c
103
while (next <= buf_end && *next != delim)
lib/kobject_uevent.c
104
if (!isalnum(*next++))
lib/kobject_uevent.c
107
if (next == buf)
lib/kobject_uevent.c
110
return next;
lib/kobject_uevent.c
117
const char *next, *buf_end, *key;
lib/kobject_uevent.c
140
next = buf + UUID_STRING_LEN;
lib/kobject_uevent.c
143
while (next <= buf_end) {
lib/kobject_uevent.c
144
if (*next != ' ')
lib/kobject_uevent.c
148
key = ++next;
lib/kobject_uevent.c
152
buf = next;
lib/kobject_uevent.c
153
next = action_arg_word_end(buf, buf_end, '=');
lib/kobject_uevent.c
154
if (!next || next > buf_end || *next != '=')
lib/kobject_uevent.c
156
key_len = next - buf;
lib/kobject_uevent.c
159
if (++next > buf_end)
lib/kobject_uevent.c
162
buf = next;
lib/kobject_uevent.c
163
next = action_arg_word_end(buf, buf_end, ' ');
lib/kobject_uevent.c
164
if (!next)
lib/kobject_uevent.c
168
key_len, key, (int) (next - buf), buf))
lib/list_debug.c
23
struct list_head *next)
lib/list_debug.c
27
CHECK_DATA_CORRUPTION(next == NULL, NULL,
lib/list_debug.c
29
CHECK_DATA_CORRUPTION(next->prev != prev, next,
lib/list_debug.c
31
prev, next->prev, next) ||
lib/list_debug.c
32
CHECK_DATA_CORRUPTION(prev->next != next, prev,
lib/list_debug.c
34
next, prev->next, prev) ||
lib/list_debug.c
35
CHECK_DATA_CORRUPTION(new == prev || new == next, NULL,
lib/list_debug.c
37
new, prev, next))
lib/list_debug.c
47
struct list_head *prev, *next;
lib/list_debug.c
50
next = entry->next;
lib/list_debug.c
52
if (CHECK_DATA_CORRUPTION(next == NULL, NULL,
lib/list_debug.c
56
CHECK_DATA_CORRUPTION(next == LIST_POISON1, next,
lib/list_debug.c
62
CHECK_DATA_CORRUPTION(prev->next != entry, prev,
lib/list_debug.c
64
entry, prev->next, prev) ||
lib/list_debug.c
65
CHECK_DATA_CORRUPTION(next->prev != entry, next,
lib/list_debug.c
67
entry, next->prev, next))
lib/list_sort.c
191
struct list_head *list = head->next, *pending = NULL;
lib/list_sort.c
198
head->prev->next = NULL;
lib/list_sort.c
22
tail = &a->next;
lib/list_sort.c
23
a = a->next;
lib/list_sort.c
238
list = list->next;
lib/list_sort.c
239
pending->next = NULL;
lib/list_sort.c
247
struct list_head *next = pending->prev;
lib/list_sort.c
249
if (!next)
lib/list_sort.c
252
pending = next;
lib/list_sort.c
30
tail = &b->next;
lib/list_sort.c
31
b = b->next;
lib/list_sort.c
58
tail->next = a;
lib/list_sort.c
61
a = a->next;
lib/list_sort.c
65
tail->next = b;
lib/list_sort.c
68
b = b->next;
lib/list_sort.c
77
tail->next = b;
lib/list_sort.c
90
b = b->next;
lib/list_sort.c
94
tail->next = head;
lib/llist.c
33
struct llist_node *entry, *next;
lib/llist.c
39
next = READ_ONCE(entry->next);
lib/llist.c
40
} while (!try_cmpxchg(&head->first, &entry, next));
lib/llist.c
60
struct llist_node *entry, *next;
lib/llist.c
67
next = READ_ONCE(entry->next);
lib/llist.c
68
} while (!try_cmpxchg(&head->first, &entry, next));
lib/llist.c
87
head = head->next;
lib/llist.c
88
tmp->next = new_head;
lib/lru_cache.c
317
n = lc->free.next;
lib/lwq.c
63
ep = &(*ep)->next;
lib/lzo/lzo1x_compress.c
325
goto next;
lib/lzo/lzo1x_compress.c
58
next:
lib/lzo/lzo1x_decompress_safe.c
121
next = t & 3;
lib/lzo/lzo1x_decompress_safe.c
132
next = t & 3;
lib/lzo/lzo1x_decompress_safe.c
139
next = t & 3;
lib/lzo/lzo1x_decompress_safe.c
163
next = get_unaligned_le16(ip);
lib/lzo/lzo1x_decompress_safe.c
165
m_pos -= next >> 2;
lib/lzo/lzo1x_decompress_safe.c
166
next &= 3;
lib/lzo/lzo1x_decompress_safe.c
169
next = get_unaligned_le16(ip);
lib/lzo/lzo1x_decompress_safe.c
170
if (((next & 0xfffc) == 0xfffc) &&
lib/lzo/lzo1x_decompress_safe.c
180
next &= 3;
lib/lzo/lzo1x_decompress_safe.c
202
next = get_unaligned_le16(ip);
lib/lzo/lzo1x_decompress_safe.c
205
m_pos -= next >> 2;
lib/lzo/lzo1x_decompress_safe.c
206
next &= 3;
lib/lzo/lzo1x_decompress_safe.c
227
state = next;
lib/lzo/lzo1x_decompress_safe.c
229
op += next;
lib/lzo/lzo1x_decompress_safe.c
230
ip += next;
lib/lzo/lzo1x_decompress_safe.c
253
state = next;
lib/lzo/lzo1x_decompress_safe.c
254
t = next;
lib/lzo/lzo1x_decompress_safe.c
44
size_t t, next;
lib/lzo/lzo1x_decompress_safe.c
68
next = t;
lib/maple_tree.c
2543
struct maple_enode *next, *last;
lib/maple_tree.c
2550
next = mas->node;
lib/maple_tree.c
2554
last = next;
lib/maple_tree.c
2555
node = mte_to_node(next);
lib/maple_tree.c
2556
type = mte_node_type(next);
lib/maple_tree.c
2564
goto next;
lib/maple_tree.c
2577
next:
lib/maple_tree.c
2579
next = mt_slot(mas->tree, slots, offset);
lib/maple_tree.c
2591
return (void *)next;
lib/maple_tree.c
3315
struct maple_enode *next;
lib/maple_tree.c
3320
next = mas->node;
lib/maple_tree.c
3322
node = mte_to_node(next);
lib/maple_tree.c
3323
type = mte_node_type(next);
lib/maple_tree.c
3333
next = mt_slot(mas->tree, slots, offset);
lib/maple_tree.c
3338
return (void *)next;
lib/maple_tree.c
4027
unsigned long *next, gfp_t gfp)
lib/maple_tree.c
4032
range_lo = max(min, *next);
lib/maple_tree.c
4054
*next = *startp + 1;
lib/maple_tree.c
4055
if (*next == 0)
lib/maple_tree.c
4885
struct maple_node *node, *next;
lib/maple_tree.c
4888
next = mte_to_node(*enode);
lib/maple_tree.c
4890
*enode = ma_enode_ptr(next);
lib/maple_tree.c
4893
next = rcu_dereference_protected(slots[offset],
lib/maple_tree.c
4896
} while (!ma_is_leaf(next->type));
lib/maple_tree.c
4951
struct maple_enode *next = *enode;
lib/maple_tree.c
4957
*enode = next;
lib/maple_tree.c
4961
next = mt_slot_locked(mt, slots, next_offset);
lib/maple_tree.c
4962
if ((mte_dead_node(next)))
lib/maple_tree.c
4963
next = mt_slot_locked(mt, slots, ++next_offset);
lib/maple_tree.c
4972
} while (!mte_is_leaf(next));
lib/maple_tree.c
5009
goto next;
lib/maple_tree.c
5017
next:
lib/maple_tree.c
6081
unsigned long *next, gfp_t gfp)
lib/maple_tree.c
6093
next, gfp);
lib/maple_tree.c
838
void *next;
lib/maple_tree.c
845
next = mt_slot_locked(mt, slots,
lib/maple_tree.c
847
if (unlikely((mte_to_node(next) &&
lib/maple_tree.c
848
mte_node_type(next))))
lib/maple_tree.c
911
mte_to_mat(dead_enode)->next = NULL;
lib/maple_tree.c
917
mte_to_mat(mat->tail)->next = dead_enode;
lib/maple_tree.c
933
struct maple_enode *next;
lib/maple_tree.c
938
next = mte_to_mat(mat->head)->next;
lib/maple_tree.c
943
mat->head = next;
lib/math/tests/prime_numbers_kunit.c
22
unsigned long x, last, next;
lib/math/tests/prime_numbers_kunit.c
33
next = next_prime_number(last);
lib/math/tests/prime_numbers_kunit.c
34
KUNIT_ASSERT_EQ_MSG(test, next, x, "next-prime(%lu)", last);
lib/math/tests/prime_numbers_kunit.c
35
last = next;
lib/oid_registry.c
80
goto next;
lib/oid_registry.c
84
goto next;
lib/oid_registry.c
88
next:
lib/plist.c
102
iter = list_entry(iter->prio_list.next,
lib/plist.c
127
if (node->node_list.next != &head->node_list) {
lib/plist.c
128
struct plist_node *next;
lib/plist.c
130
next = list_entry(node->node_list.next,
lib/plist.c
134
if (list_empty(&next->prio_list))
lib/plist.c
135
list_add(&next->prio_list, &node->prio_list);
lib/plist.c
179
iter = list_entry(iter->prio_list.next, struct plist_node,
lib/plist.c
226
BUG_ON(prio_pos->prio_list.next != &node_pos->prio_list);
lib/plist.c
231
BUG_ON(prio_pos->prio_list.next != &first->prio_list);
lib/plist.c
35
WARN(n->prev != p || p->next != n,
lib/plist.c
39
t, t->next, t->prev,
lib/plist.c
40
p, p->next, p->prev,
lib/plist.c
41
n, n->next, n->prev);
lib/plist.c
46
struct list_head *prev = top, *next = top->next;
lib/plist.c
48
plist_check_prev_next(top, prev, next);
lib/plist.c
49
while (next != top) {
lib/plist.c
50
prev = next;
lib/plist.c
51
next = prev->next;
lib/plist.c
52
plist_check_prev_next(top, prev, next);
lib/plist.c
94
iter = list_entry(reverse_iter->prio_list.next,
lib/rhashtable.c
1125
list = rht_dereference(list->next, ht);
lib/rhashtable.c
1159
struct rhash_head *pos, *next;
lib/rhashtable.c
1163
next = !rht_is_a_nulls(pos) ?
lib/rhashtable.c
1164
rht_dereference(pos->next, ht) : NULL;
lib/rhashtable.c
1166
pos = next,
lib/rhashtable.c
1167
next = !rht_is_a_nulls(pos) ?
lib/rhashtable.c
1168
rht_dereference(pos->next, ht) : NULL)
lib/rhashtable.c
235
struct rhash_head *head, *next, *entry;
lib/rhashtable.c
248
next = rht_dereference_bucket(entry->next, old_tbl, old_hash);
lib/rhashtable.c
250
if (rht_is_a_nulls(next))
lib/rhashtable.c
253
pprev = &entry->next;
lib/rhashtable.c
266
RCU_INIT_POINTER(entry->next, head);
lib/rhashtable.c
271
rcu_assign_pointer(*pprev, next);
lib/rhashtable.c
274
rht_assign_locked(bkt, next);
lib/rhashtable.c
519
pprev = &head->next;
lib/rhashtable.c
529
RCU_INIT_POINTER(list->next, plist);
lib/rhashtable.c
530
head = rht_dereference_bucket(head->next, tbl, hash);
lib/rhashtable.c
531
RCU_INIT_POINTER(list->rhead.next, head);
lib/rhashtable.c
576
RCU_INIT_POINTER(obj->next, head);
lib/rhashtable.c
581
RCU_INIT_POINTER(list->next, NULL);
lib/rhashtable.c
775
list = rcu_dereference(list->next)) {
lib/rhashtable.c
821
goto next;
lib/rhashtable.c
823
list = rcu_dereference(list->next);
lib/rhashtable.c
833
next:
lib/rhashtable.c
881
if (!rhlist || !(list = rcu_dereference(list->next))) {
lib/rhashtable.c
882
p = rcu_dereference(p->next);
lib/sbitmap.c
540
goto next;
lib/sbitmap.c
558
next:
lib/scatterlist.c
1289
folioq = folioq->next;
lib/scatterlist.c
1314
if (!folioq->next) {
lib/scatterlist.c
1318
folioq = folioq->next;
lib/scatterlist.c
202
struct scatterlist *sgl, *next;
lib/scatterlist.c
220
next = sg_chain_ptr(&sgl[curr_max_ents - 1]);
lib/scatterlist.c
225
next = NULL;
lib/scatterlist.c
233
sgl = next;
lib/test_hmm.c
1018
next:
lib/test_hmm.c
1047
unsigned long next;
lib/test_hmm.c
1066
for (addr = start; addr < end; addr = next) {
lib/test_hmm.c
1072
next = min(end, addr + (PTRS_PER_PTE << PAGE_SHIFT));
lib/test_hmm.c
1073
if (next > vma->vm_end)
lib/test_hmm.c
1074
next = vma->vm_end;
lib/test_hmm.c
1080
args.end = next;
lib/test_hmm.c
1113
unsigned long next;
lib/test_hmm.c
1140
for (addr = start; addr < end; addr = next) {
lib/test_hmm.c
1146
next = min(end, addr + (PTRS_PER_PTE << PAGE_SHIFT));
lib/test_hmm.c
1147
if (next > vma->vm_end)
lib/test_hmm.c
1148
next = vma->vm_end;
lib/test_hmm.c
1154
args.end = next;
lib/test_hmm.c
1331
unsigned long next;
lib/test_hmm.c
1355
for (addr = start; addr < end; addr = next) {
lib/test_hmm.c
1358
next = min(addr + (ARRAY_SIZE(pfns) << PAGE_SHIFT), end);
lib/test_hmm.c
1360
range.end = next;
lib/test_hmm.c
693
goto next;
lib/test_hmm.c
703
goto next;
lib/test_hmm.c
718
goto next;
lib/test_hmm.c
790
next:
lib/test_hmm.c
959
goto next;
lib/test_hmm.c
965
goto next;
lib/test_lockup.c
321
struct page *page, *next;
lib/test_lockup.c
323
list_for_each_entry_safe(page, next, pages, lru)
lib/test_maple_tree.c
3614
unsigned long next;
lib/test_maple_tree.c
3618
next = 0;
lib/test_maple_tree.c
3621
mas_alloc_cyclic(&mas, &location, mt, 2, ULONG_MAX, &next, GFP_KERNEL);
lib/test_maple_tree.c
3625
MAS_BUG_ON(&mas, i != next - 3);
lib/test_maple_tree.c
3630
next = 0;
lib/test_maple_tree.c
3633
mtree_alloc_cyclic(mt, &location, mt, 2, ULONG_MAX, &next, GFP_KERNEL);
lib/test_maple_tree.c
3635
MT_BUG_ON(mt, i != next - 3);
lib/test_maple_tree.c
3647
next = 0;
lib/test_maple_tree.c
3650
mtree_alloc_cyclic(mt, &location, mt, 2, 1024, &next, GFP_KERNEL);
lib/test_maple_tree.c
3652
MT_BUG_ON(mt, i != next - 3);
lib/test_maple_tree.c
3657
mtree_alloc_cyclic(mt, &location, mt, 2, 1024, &next, GFP_KERNEL);
lib/test_maple_tree.c
3659
MT_BUG_ON(mt, 124 != next);
lib/test_maple_tree.c
3662
mtree_alloc_cyclic(mt, &location, mt, 2, 1024, &next, GFP_KERNEL);
lib/test_maple_tree.c
3664
MT_BUG_ON(mt, 101 != next);
lib/test_maple_tree.c
3669
next = ULONG_MAX - 1;
lib/test_maple_tree.c
3670
ret = mtree_alloc_cyclic(mt, &location, mt, 2, ULONG_MAX, &next, GFP_KERNEL);
lib/test_maple_tree.c
3672
ret = mtree_alloc_cyclic(mt, &location, mt, 2, ULONG_MAX, &next, GFP_KERNEL);
lib/test_maple_tree.c
3674
ret = mtree_alloc_cyclic(mt, &location, mt, 2, ULONG_MAX, &next, GFP_KERNEL);
lib/test_rhashtable.c
483
struct rhash_head *pos, *next;
lib/test_rhashtable.c
487
next = !rht_is_a_nulls(pos) ? rht_dereference(pos->next, ht) : NULL;
lib/test_rhashtable.c
498
list = rht_dereference(list->next, ht);
lib/test_rhashtable.c
506
pos = next,
lib/test_rhashtable.c
507
next = !rht_is_a_nulls(pos) ?
lib/test_rhashtable.c
508
rht_dereference(pos->next, ht) : NULL;
lib/test_xarray.c
1039
u32 next = 0;
lib/test_xarray.c
1046
&next, GFP_KERNEL) != 0);
lib/test_xarray.c
1049
next = 0x3ffd;
lib/test_xarray.c
1051
&next, GFP_KERNEL) != 0);
lib/test_xarray.c
1063
&next, GFP_KERNEL) != 0);
lib/test_xarray.c
1071
next = UINT_MAX;
lib/test_xarray.c
1073
xa_limit_32b, &next, GFP_KERNEL) != 0);
lib/test_xarray.c
1076
xa_limit_32b, &next, GFP_KERNEL) != 0);
lib/test_xarray.c
1079
xa_limit_32b, &next, GFP_KERNEL) != 0);
lib/test_xarray.c
1087
next = UINT_MAX;
lib/test_xarray.c
1089
xa_limit_32b, &next, GFP_KERNEL) != 0);
lib/test_xarray.c
1092
&next, GFP_KERNEL);
lib/test_xarray.c
1184
unsigned long next = 4 << order;
lib/test_xarray.c
1188
XA_BUG_ON(xa, xa_store_index(xa, next, GFP_KERNEL) != NULL);
lib/test_xarray.c
1189
XA_BUG_ON(xa, xa_store_index(xa, next + 1, GFP_KERNEL) != NULL);
lib/test_xarray.c
1198
XA_BUG_ON(xa, (index < multi) || (index >= next));
lib/test_xarray.c
1200
xa_mk_value(next));
lib/test_xarray.c
1201
XA_BUG_ON(xa, index != next);
lib/test_xarray.c
1202
XA_BUG_ON(xa, xa_find_after(xa, &index, next, XA_PRESENT) != NULL);
lib/test_xarray.c
1203
XA_BUG_ON(xa, index != next);
lib/test_xarray.c
1206
xa_erase_index(xa, next);
lib/test_xarray.c
1207
xa_erase_index(xa, next + 1);
lib/test_xarray.c
209
unsigned long next = base + (1UL << order);
lib/test_xarray.c
216
XA_BUG_ON(xa, xa_store_index(xa, next, GFP_KERNEL));
lib/test_xarray.c
219
for (i = base; i < next; i++) {
lib/test_xarray.c
244
XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_0));
lib/test_xarray.c
245
XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_1));
lib/test_xarray.c
246
XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_2));
lib/test_xarray.c
248
xa_erase_index(xa, next);
lib/tests/kunit_iov_iter.c
368
struct folio_queue *folioq, *next;
lib/tests/kunit_iov_iter.c
370
for (folioq = data; folioq; folioq = next) {
lib/tests/kunit_iov_iter.c
371
next = folioq->next;
lib/tests/kunit_iov_iter.c
390
p->next = kzalloc_obj(struct folio_queue);
lib/tests/kunit_iov_iter.c
391
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, p->next);
lib/tests/kunit_iov_iter.c
392
folioq_init(p->next, 0);
lib/tests/kunit_iov_iter.c
393
p->next->prev = p;
lib/tests/kunit_iov_iter.c
394
p = p->next;
lib/tests/list-test.c
103
KUNIT_EXPECT_PTR_EQ(test, list.next, &a_new);
lib/tests/list-test.c
105
KUNIT_EXPECT_PTR_EQ(test, a_new.next, &b);
lib/tests/list-test.c
121
KUNIT_EXPECT_PTR_EQ(test, list.next, &a_new);
lib/tests/list-test.c
123
KUNIT_EXPECT_PTR_EQ(test, a_new.next, &b);
lib/tests/list-test.c
142
KUNIT_EXPECT_PTR_EQ(test, &b, list.next);
lib/tests/list-test.c
145
KUNIT_EXPECT_PTR_EQ(test, &a, b.next);
lib/tests/list-test.c
148
KUNIT_EXPECT_PTR_EQ(test, &list, a.next);
lib/tests/list-test.c
164
KUNIT_EXPECT_PTR_EQ(test, list.next, &b);
lib/tests/list-test.c
184
KUNIT_EXPECT_PTR_EQ(test, list.next, &b);
lib/tests/list-test.c
204
KUNIT_EXPECT_PTR_EQ(test, &a, list2.next);
lib/tests/list-test.c
205
KUNIT_EXPECT_PTR_EQ(test, &b, a.next);
lib/tests/list-test.c
223
KUNIT_EXPECT_PTR_EQ(test, &b, list2.next);
lib/tests/list-test.c
224
KUNIT_EXPECT_PTR_EQ(test, &a, b.next);
lib/tests/list-test.c
341
KUNIT_EXPECT_PTR_EQ(test, list.next, &b);
lib/tests/list-test.c
343
KUNIT_EXPECT_PTR_EQ(test, b.next, &a);
lib/tests/list-test.c
56
KUNIT_EXPECT_PTR_EQ(test, list.next, &b);
lib/tests/list-test.c
58
KUNIT_EXPECT_PTR_EQ(test, b.next, &a);
lib/tests/list-test.c
70
KUNIT_EXPECT_PTR_EQ(test, list.next, &a);
lib/tests/list-test.c
72
KUNIT_EXPECT_PTR_EQ(test, a.next, &b);
lib/tests/list-test.c
87
KUNIT_EXPECT_PTR_EQ(test, list.next, &b);
lib/tests/list-test.c
924
KUNIT_EXPECT_PTR_EQ(test, a.next, NULL);
lib/tests/list-test.c
942
KUNIT_EXPECT_PTR_EQ(test, c.pprev, &(b.next));
lib/tests/list-test.c
943
KUNIT_EXPECT_PTR_EQ(test, b.next, &c);
lib/tests/list-test.c
945
KUNIT_EXPECT_PTR_EQ(test, a.pprev, &(c.next));
lib/tests/list-test.c
946
KUNIT_EXPECT_PTR_EQ(test, c.next, &a);
lib/tests/list-test.c
948
KUNIT_EXPECT_PTR_EQ(test, d.pprev, &(a.next));
lib/tests/list-test.c
949
KUNIT_EXPECT_PTR_EQ(test, a.next, &d);
lib/tests/test_list_sort.c
84
for (cur = head.next; cur->next != &head; cur = cur->next) {
lib/tests/test_list_sort.c
88
KUNIT_ASSERT_PTR_EQ_MSG(test, cur->next->prev, cur,
lib/tests/test_list_sort.c
91
cmp_result = cmp(test, cur, cur->next);
lib/tests/test_list_sort.c
95
el1 = container_of(cur->next, struct debug_el, list);
lib/timerqueue.c
75
struct rb_node *next;
lib/timerqueue.c
79
next = rb_next(&node->node);
lib/timerqueue.c
80
if (!next)
lib/timerqueue.c
82
return container_of(next, struct timerqueue_node, node);
lib/ts_fsm.c
137
struct ts_fsm_token *cur = NULL, *next;
lib/ts_fsm.c
169
next = &fsm->tokens[tok_idx + 1];
lib/ts_fsm.c
171
next = NULL;
lib/ts_fsm.c
199
if (next == NULL)
lib/ts_fsm.c
205
while (!match_token(next, data[block_idx])) {
lib/ts_fsm.c
222
while (!match_token(next, data[block_idx])) {
lib/vsprintf.c
3454
char *next;
lib/vsprintf.c
3670
val.s = simple_strntoll(str, &next, base,
lib/vsprintf.c
3673
val.u = simple_strntoull(str, &next, base,
lib/vsprintf.c
3713
if (!next)
lib/vsprintf.c
3715
str = next;
lib/xarray.c
2038
struct xa_limit limit, u32 *next, gfp_t gfp)
lib/xarray.c
2043
limit.min = max(min, *next);
lib/xarray.c
2058
*next = *id + 1;
lib/xarray.c
2059
if (*next == 0)
lib/xarray.c
2419
node->array, node->private_list.prev, node->private_list.next);
lib/xarray.c
272
struct xa_node *next, *node = xas->xa_alloc;
lib/xarray.c
276
next = rcu_dereference_raw(node->parent);
lib/xarray.c
278
xas->xa_alloc = node = next;
lib/xarray.c
790
void *first, *next;
lib/xarray.c
808
next = first;
lib/xarray.c
828
if (xa_is_node(next) && (!node || node->shift))
lib/xarray.c
829
xas_free_nodes(xas, xa_to_node(next));
lib/xarray.c
832
count += !next - !entry;
lib/xarray.c
843
next = xa_entry_locked(xas->xa, node, ++offset);
lib/xarray.c
844
if (!xa_is_sibling(next)) {
lib/xarray.c
847
first = next;
lib/zlib_deflate/deflate.c
197
char *next;
lib/zlib_deflate/deflate.c
226
next = (char *) mem;
lib/zlib_deflate/deflate.c
227
next += sizeof(*mem);
lib/zlib_deflate/deflate.c
233
mem->window_memory = (Byte *) PTR_ALIGN(next, PAGE_SIZE);
lib/zlib_deflate/deflate.c
235
mem->window_memory = (Byte *) next;
lib/zlib_deflate/deflate.c
237
next += zlib_deflate_window_memsize(windowBits);
lib/zlib_deflate/deflate.c
238
mem->prev_memory = (Pos *) next;
lib/zlib_deflate/deflate.c
239
next += zlib_deflate_prev_memsize(windowBits);
lib/zlib_deflate/deflate.c
240
mem->head_memory = (Pos *) next;
lib/zlib_deflate/deflate.c
241
next += zlib_deflate_head_memsize(memLevel);
lib/zlib_deflate/deflate.c
242
mem->overlay_memory = next;
lib/zlib_inflate/inflate.c
189
next = strm->next_in; \
lib/zlib_inflate/inflate.c
200
strm->next_in = next; \
lib/zlib_inflate/inflate.c
219
hold += (unsigned long)(*next++) << bits; \
lib/zlib_inflate/inflate.c
334
const unsigned char *next; /* next input */
lib/zlib_inflate/inflate.c
456
memcpy(put, next, copy);
lib/zlib_inflate/inflate.c
458
next += copy;
lib/zlib_inflate/inflate.c
48
state->lencode = state->distcode = state->next = state->codes;
lib/zlib_inflate/inflate.c
492
state->next = state->codes;
lib/zlib_inflate/inflate.c
493
state->lencode = (code const *)(state->next);
lib/zlib_inflate/inflate.c
495
ret = zlib_inflate_table(CODES, state->lens, 19, &(state->next),
lib/zlib_inflate/inflate.c
558
state->next = state->codes;
lib/zlib_inflate/inflate.c
559
state->lencode = (code const *)(state->next);
lib/zlib_inflate/inflate.c
561
ret = zlib_inflate_table(LENS, state->lens, state->nlen, &(state->next),
lib/zlib_inflate/inflate.c
568
state->distcode = (code const *)(state->next);
lib/zlib_inflate/inflate.c
571
&(state->next), &(state->distbits), state->work);
lib/zlib_inflate/inflate.h
108
code *next; /* next available space in codes[] */
lib/zlib_inflate/inftrees.c
190
next = *table; /* current table to fill in */
lib/zlib_inflate/inftrees.c
224
next[(huff >> drop) + fill] = this;
lib/zlib_inflate/inftrees.c
252
next += min; /* here min is 1 << curr */
lib/zlib_inflate/inftrees.c
273
(*table)[low].val = (unsigned short)(next - *table);
lib/zlib_inflate/inftrees.c
292
next = *table;
lib/zlib_inflate/inftrees.c
297
next[huff >> drop] = this;
lib/zlib_inflate/inftrees.c
40
code *next; /* next available space in table */
lib/zstd/compress/zstd_lazy.c
800
U32 next = (*tagRow-1) & rowMask;
lib/zstd/compress/zstd_lazy.c
801
next += (next == 0) ? rowMask : 0; /* skip first position */
lib/zstd/compress/zstd_lazy.c
802
*tagRow = (BYTE)next;
lib/zstd/compress/zstd_lazy.c
803
return next;
mm/backing-dev.c
728
blkcg_cgwb_list->next && memcg_cgwb_list->next) {
mm/backing-dev.c
934
struct bdi_writeback *wb, *next;
mm/backing-dev.c
937
list_for_each_entry_safe(wb, next, memcg_cgwb_list, memcg_node)
mm/backing-dev.c
939
memcg_cgwb_list->next = NULL; /* prevent new wb's */
mm/backing-dev.c
953
struct bdi_writeback *wb, *next;
mm/backing-dev.c
957
list_for_each_entry_safe(wb, next, list, blkcg_node)
mm/backing-dev.c
959
list->next = NULL; /* prevent new wb's */
mm/cma.c
538
struct list_head *mp, *next;
mm/cma.c
647
list_for_each_safe(mp, next, &ranges) {
mm/compaction.c
97
struct page *page, *next;
mm/compaction.c
99
list_for_each_entry_safe(page, next, &freepages[order], lru) {
mm/damon/core.c
1091
struct damos *dst_scheme, *next, *src_scheme, *new_scheme;
mm/damon/core.c
1094
damon_for_each_scheme_safe(dst_scheme, next, dst) {
mm/damon/core.c
1105
damon_for_each_scheme_safe(src_scheme, next, src) {
mm/damon/core.c
1188
struct damon_target *dst_target, *next, *src_target, *new_target;
mm/damon/core.c
1191
damon_for_each_target_safe(dst_target, next, dst) {
mm/damon/core.c
1217
damon_for_each_target_safe(src_target, next, src) {
mm/damon/core.c
186
struct damon_region *next, *newr;
mm/damon/core.c
190
next = damon_next_region(r);
mm/damon/core.c
191
if (r->ar.end != next->ar.start) {
mm/damon/core.c
192
newr = damon_new_region(r->ar.end, next->ar.start);
mm/damon/core.c
195
damon_insert_region(newr, r, next, t);
mm/damon/core.c
216
struct damon_region *r, *next;
mm/damon/core.c
221
damon_for_each_region_safe(r, next, t) {
mm/damon/core.c
2427
struct damon_region *r, *prev = NULL, *next;
mm/damon/core.c
2429
damon_for_each_region_safe(r, next, t) {
mm/damon/core.c
2512
struct damon_region *r, *next;
mm/damon/core.c
2516
damon_for_each_region_safe(r, next, t) {
mm/damon/core.c
2663
struct damon_call_control *control, *next;
mm/damon/core.c
2670
list_for_each_entry_safe(control, next, &controls, list) {
mm/damon/core.c
450
struct damos_filter *f, *next;
mm/damon/core.c
455
damos_for_each_core_filter_safe(f, next, s)
mm/damon/core.c
458
damos_for_each_ops_filter_safe(f, next, s)
mm/damon/core.c
506
struct damon_region *r, *next;
mm/damon/core.c
508
damon_for_each_region_safe(r, next, t)
mm/damon/core.c
765
struct damos *s, *next;
mm/damon/core.c
768
damon_for_each_scheme_safe(s, next, ctx)
mm/damon/core.c
830
struct damos_quota_goal *dst_goal, *next, *src_goal, *new_goal;
mm/damon/core.c
833
damos_for_each_quota_goal_safe(dst_goal, next, dst) {
mm/damon/core.c
840
damos_for_each_quota_goal_safe(src_goal, next, src) {
mm/damon/core.c
925
struct damos_filter *dst_filter, *next, *src_filter, *new_filter;
mm/damon/core.c
928
damos_for_each_core_filter_safe(dst_filter, next, dst) {
mm/damon/core.c
936
damos_for_each_core_filter_safe(src_filter, next, src) {
mm/damon/core.c
953
struct damos_filter *dst_filter, *next, *src_filter, *new_filter;
mm/damon/core.c
956
damos_for_each_ops_filter_safe(dst_filter, next, dst) {
mm/damon/core.c
964
damos_for_each_ops_filter_safe(src_filter, next, src) {
mm/damon/sysfs-schemes.c
164
struct damon_sysfs_scheme_region *r, *next;
mm/damon/sysfs-schemes.c
166
list_for_each_entry_safe(r, next, &regions->regions_list, list) {
mm/damon/sysfs-schemes.c
2772
struct damos *scheme, *next;
mm/damon/sysfs-schemes.c
2776
damon_for_each_scheme_safe(scheme, next, ctx)
mm/damon/tests/core-kunit.h
522
KUNIT_EXPECT_PTR_EQ(test, filter->list.next, &filter->list);
mm/damon/tests/core-kunit.h
615
struct damos_quota_goal *goal, *next;
mm/damon/tests/core-kunit.h
649
damos_for_each_quota_goal_safe(goal, next, (&dst))
mm/damon/vaddr.c
139
goto next;
mm/damon/vaddr.c
151
next:
mm/damon/vaddr.c
307
unsigned long next, struct mm_walk *walk)
mm/damon/vaddr.c
432
unsigned long next, struct mm_walk *walk)
mm/damon/vaddr.c
69
struct damon_region *n = NULL, *next;
mm/damon/vaddr.c
701
unsigned long next, struct mm_walk *walk)
mm/damon/vaddr.c
736
for (; addr < next; pte += nr, addr += nr * PAGE_SIZE) {
mm/damon/vaddr.c
864
unsigned long next, struct mm_walk *walk)
mm/damon/vaddr.c
87
next = damon_next_region(r);
mm/damon/vaddr.c
901
for (; addr < next; pte += nr, addr += nr * PAGE_SIZE) {
mm/damon/vaddr.c
92
damon_insert_region(n, r, next, t);
mm/debug.c
258
vmg->prev, vmg->middle, vmg->next, vmg->target,
mm/debug.c
293
if (vmg->next) {
mm/debug.c
295
dump_vma(vmg->next);
mm/gup.c
1273
unsigned long now, next;
mm/gup.c
1286
now = jiffies; next = next_warn;
mm/gup.c
1287
if (next && time_before(now, next))
mm/gup.c
235
struct page *next = start + i;
mm/gup.c
236
struct folio *folio = page_folio(next);
mm/gup.c
241
folio_nr_pages(folio) - folio_page_idx(folio, next));
mm/gup.c
3015
unsigned long next;
mm/gup.c
3022
next = pmd_addr_end(addr, end);
mm/gup.c
3031
if (!gup_fast_pmd_leaf(pmd, pmdp, addr, next, flags,
mm/gup.c
3035
} else if (!gup_fast_pte_range(pmd, pmdp, addr, next, flags,
mm/gup.c
3038
} while (pmdp++, addr = next, addr != end);
mm/gup.c
3047
unsigned long next;
mm/gup.c
3054
next = pud_addr_end(addr, end);
mm/gup.c
3058
if (!gup_fast_pud_leaf(pud, pudp, addr, next, flags,
mm/gup.c
3061
} else if (!gup_fast_pmd_range(pudp, pud, addr, next, flags,
mm/gup.c
3064
} while (pudp++, addr = next, addr != end);
mm/gup.c
3073
unsigned long next;
mm/gup.c
3080
next = p4d_addr_end(addr, end);
mm/gup.c
3084
if (!gup_fast_pud_range(p4dp, p4d, addr, next, flags,
mm/gup.c
3087
} while (p4dp++, addr = next, addr != end);
mm/gup.c
3095
unsigned long next;
mm/gup.c
3102
next = pgd_addr_end(addr, end);
mm/gup.c
3106
if (!gup_fast_p4d_range(pgdp, pgd, addr, next, flags,
mm/gup.c
3109
} while (pgdp++, addr = next, addr != end);
mm/gup_test.c
105
unsigned long i, nr_pages, addr, next;
mm/gup_test.c
128
for (addr = gup->addr; addr < gup->addr + gup->size; addr = next) {
mm/gup_test.c
132
next = addr + nr * PAGE_SIZE;
mm/gup_test.c
133
if (next > gup->addr + gup->size) {
mm/gup_test.c
134
next = gup->addr + gup->size;
mm/gup_test.c
135
nr = (next - addr) / PAGE_SIZE;
mm/huge_memory.c
2266
pmd_t *pmd, unsigned long addr, unsigned long next)
mm/huge_memory.c
2305
if (next - addr != HPAGE_PMD_SIZE) {
mm/huge_memory.c
3319
struct vm_area_struct *next)
mm/huge_memory.c
3328
if (next)
mm/huge_memory.c
3329
split_huge_pmd_if_needed(next, end);
mm/huge_memory.c
3786
struct folio *new_folio, *next;
mm/huge_memory.c
3857
new_folio = next) {
mm/huge_memory.c
3860
next = folio_next(new_folio);
mm/huge_memory.c
3954
struct folio *new_folio, *next;
mm/huge_memory.c
4084
for (new_folio = folio; new_folio != end_folio; new_folio = next) {
mm/huge_memory.c
4085
next = folio_next(new_folio);
mm/huge_memory.c
4412
struct folio *folio, *next;
mm/huge_memory.c
4421
list_for_each_entry_safe(folio, next, &ds_queue->split_queue,
mm/huge_memory.c
4453
goto next;
mm/huge_memory.c
4456
goto next;
mm/huge_memory.c
4459
goto next;
mm/huge_memory.c
4467
next:
mm/huge_memory.c
4551
goto next;
mm/huge_memory.c
4554
goto next;
mm/huge_memory.c
4559
goto next;
mm/huge_memory.c
4568
next:
mm/huge_memory.c
4640
goto next;
mm/huge_memory.c
4649
goto next;
mm/huge_memory.c
4659
goto next;
mm/huge_memory.c
4662
goto next;
mm/huge_memory.c
4687
next:
mm/huge_memory.c
4732
goto next;
mm/huge_memory.c
4738
goto next;
mm/huge_memory.c
4741
goto next;
mm/huge_memory.c
4758
next:
mm/hugetlb.c
1592
node = node->next;
mm/hugetlb.c
1990
folio = list_entry(h->hugepage_freelists[node].next,
mm/hugetlb.c
3722
struct folio *folio, *next;
mm/hugetlb.c
3724
list_for_each_entry_safe(folio, next, freel, lru) {
mm/hugetlb.c
3973
struct folio *folio, *next;
mm/hugetlb.c
3990
list_for_each_entry_safe(folio, next, src_list, lru) {
mm/hugetlb.c
4048
struct folio *folio, *next;
mm/hugetlb.c
4050
list_for_each_entry_safe(folio, next, &src->hugepage_freelists[node], lru) {
mm/hugetlb.c
4067
list_for_each_entry_safe(folio, next, &list, lru) {
mm/hugetlb.c
4947
goto next;
mm/hugetlb.c
5031
goto next;
mm/hugetlb.c
5053
next:
mm/hugetlb.c
6487
goto next;
mm/hugetlb.c
6540
next:
mm/hugetlb_vmemmap.c
140
unsigned long next, struct mm_walk *walk)
mm/hugetlb_vmemmap.c
202
struct page *page, *next;
mm/hugetlb_vmemmap.c
204
list_for_each_entry_safe(page, next, list, lru)
mm/hugetlb_vmemmap.c
393
struct page *page, *next;
mm/hugetlb_vmemmap.c
406
list_for_each_entry_safe(page, next, list, lru)
mm/hugetlb_vmemmap.c
99
unsigned long next, struct mm_walk *walk)
mm/kasan/init.c
113
unsigned long next;
mm/kasan/init.c
116
next = pmd_addr_end(addr, end);
mm/kasan/init.c
138
zero_pte_populate(pmd, addr, next);
mm/kasan/init.c
139
} while (pmd++, addr = next, addr != end);
mm/kasan/init.c
148
unsigned long next;
mm/kasan/init.c
151
next = pud_addr_end(addr, end);
mm/kasan/init.c
176
zero_pmd_populate(pud, addr, next);
mm/kasan/init.c
177
} while (pud++, addr = next, addr != end);
mm/kasan/init.c
186
unsigned long next;
mm/kasan/init.c
189
next = p4d_addr_end(addr, end);
mm/kasan/init.c
218
zero_pud_populate(p4d, addr, next);
mm/kasan/init.c
219
} while (p4d++, addr = next, addr != end);
mm/kasan/init.c
236
unsigned long next;
mm/kasan/init.c
239
next = pgd_addr_end(addr, end);
mm/kasan/init.c
278
zero_p4d_populate(pgd, addr, next);
mm/kasan/init.c
279
} while (pgd++, addr = next, addr != end);
mm/kasan/init.c
347
unsigned long next;
mm/kasan/init.c
350
for (; addr < end; addr = next, pte++) {
mm/kasan/init.c
351
next = (addr + PAGE_SIZE) & PAGE_MASK;
mm/kasan/init.c
352
if (next > end)
mm/kasan/init.c
353
next = end;
mm/kasan/init.c
369
unsigned long next;
mm/kasan/init.c
371
for (; addr < end; addr = next, pmd++) {
mm/kasan/init.c
374
next = pmd_addr_end(addr, end);
mm/kasan/init.c
381
IS_ALIGNED(next, PMD_SIZE)) {
mm/kasan/init.c
387
kasan_remove_pte_table(pte, addr, next);
mm/kasan/init.c
395
unsigned long next;
mm/kasan/init.c
397
for (; addr < end; addr = next, pud++) {
mm/kasan/init.c
400
next = pud_addr_end(addr, end);
mm/kasan/init.c
407
IS_ALIGNED(next, PUD_SIZE)) {
mm/kasan/init.c
414
kasan_remove_pmd_table(pmd, addr, next);
mm/kasan/init.c
422
unsigned long next;
mm/kasan/init.c
424
for (; addr < end; addr = next, p4d++) {
mm/kasan/init.c
427
next = p4d_addr_end(addr, end);
mm/kasan/init.c
434
IS_ALIGNED(next, P4D_SIZE)) {
mm/kasan/init.c
440
kasan_remove_pud_table(pud, addr, next);
mm/kasan/init.c
447
unsigned long addr, end, next;
mm/kasan/init.c
457
for (; addr < end; addr = next) {
mm/kasan/init.c
460
next = pgd_addr_end(addr, end);
mm/kasan/init.c
468
IS_ALIGNED(next, PGDIR_SIZE)) {
mm/kasan/init.c
475
kasan_remove_p4d_table(p4d, addr, next);
mm/kasan/kasan.h
271
struct qlist_node *next;
mm/kasan/quarantine.c
177
struct qlist_node *next = qlink->next;
mm/kasan/quarantine.c
180
qlink = next;
mm/kasan/quarantine.c
302
struct qlist_node *next = curr->next;
mm/kasan/quarantine.c
310
curr = next;
mm/kasan/quarantine.c
61
q->tail->next = qlink;
mm/kasan/quarantine.c
63
qlink->next = NULL;
mm/kasan/quarantine.c
78
to->tail->next = from->head;
mm/kasan/tags.c
115
next:
mm/kasan/tags.c
122
goto next; /* Busy slot. */
mm/kasan/tags.c
124
goto next; /* Busy slot. */
mm/kfence/core.c
292
metadata_update_state(struct kfence_metadata *meta, enum kfence_object_state next,
mm/kfence/core.c
297
next == KFENCE_OBJECT_ALLOCATED ? &meta->alloc_track : &meta->free_track;
mm/kfence/core.c
326
WRITE_ONCE(meta->state, next);
mm/kfence/core.c
433
meta = list_entry(kfence_freelist.next, struct kfence_metadata, list);
mm/kfence/core.c
796
.next = next_object,
mm/khugepaged.c
594
goto next;
mm/khugepaged.c
643
next:
mm/kmemleak.c
1551
void *next;
mm/kmemleak.c
1554
next = min(start + MAX_SCAN_SIZE, end);
mm/kmemleak.c
1555
scan_block(start, next, NULL);
mm/kmemleak.c
1556
start = next;
mm/kmemleak.c
1603
void *next;
mm/kmemleak.c
1606
next = min(start + MAX_SCAN_SIZE, end);
mm/kmemleak.c
1607
scan_block(start, next, object);
mm/kmemleak.c
1609
start = next;
mm/kmemleak.c
1640
object = list_entry(gray_list.next, typeof(*object), gray_list);
mm/kmemleak.c
1648
tmp = list_entry(object->gray_list.next, typeof(*object),
mm/kmemleak.c
2024
.next = kmemleak_seq_next,
mm/ksm.c
1168
struct ksm_stable_node *stable_node, *next;
mm/ksm.c
1184
list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) {
mm/ksm.c
1201
slot = list_entry(ksm_mm_head.slot.mm_node.next,
mm/ksm.c
1233
slot = list_entry(mm_slot->slot.mm_node.next,
mm/ksm.c
1719
BUG_ON(stable_node->hlist.first->next);
mm/ksm.c
2231
if (rmap_item->hlist.next)
mm/ksm.c
2611
struct ksm_stable_node *stable_node, *next;
mm/ksm.c
2614
list_for_each_entry_safe(stable_node, next,
mm/ksm.c
2628
slot = list_entry(mm_slot->slot.mm_node.next,
mm/ksm.c
2724
slot = list_entry(mm_slot->slot.mm_node.next,
mm/ksm.c
3347
struct ksm_stable_node *stable_node, *next;
mm/ksm.c
3365
list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) {
mm/ksm.c
887
if (rmap_item->hlist.next) {
mm/madvise.c
1075
unsigned long next, struct mm_walk *walk)
mm/madvise.c
1084
unsigned long next, struct mm_walk *walk)
mm/madvise.c
1093
unsigned long next, struct mm_walk *walk)
mm/madvise.c
1109
static int guard_install_set_pte(unsigned long addr, unsigned long next,
mm/madvise.c
1213
unsigned long next, struct mm_walk *walk)
mm/madvise.c
1225
unsigned long next, struct mm_walk *walk)
mm/madvise.c
1237
unsigned long next, struct mm_walk *walk)
mm/madvise.c
378
unsigned long next = pmd_addr_end(addr, end);
mm/madvise.c
404
if (next - addr != HPAGE_PMD_SIZE) {
mm/madvise.c
663
unsigned long next;
mm/madvise.c
666
next = pmd_addr_end(addr, end);
mm/madvise.c
668
if (madvise_free_huge_pmd(tlb, vma, pmd, addr, next))
mm/memblock.c
544
struct memblock_region *next = &type->regions[i + 1];
mm/memblock.c
546
if (this->base + this->size != next->base ||
mm/memblock.c
548
memblock_get_region_node(next) ||
mm/memblock.c
549
this->flags != next->flags) {
mm/memblock.c
550
BUG_ON(this->base + this->size > next->base);
mm/memblock.c
555
this->size += next->size;
mm/memblock.c
557
memmove(next, next + 1, (type->cnt - (i + 2)) * sizeof(*next));
mm/memcontrol-v1.c
552
unsigned long val, next;
mm/memcontrol-v1.c
555
next = __this_cpu_read(memcg->events_percpu->targets[target]);
mm/memcontrol-v1.c
557
if ((long)(next - val) < 0) {
mm/memcontrol-v1.c
560
next = val + THRESHOLDS_EVENTS_TARGET;
mm/memcontrol-v1.c
563
next = val + SOFTLIMIT_EVENTS_TARGET;
mm/memcontrol-v1.c
568
__this_cpu_write(memcg->events_percpu->targets[target], next);
mm/memcontrol.c
1028
struct mem_cgroup *next;
mm/memcontrol.c
1038
next = NULL;
mm/memcontrol.c
1072
next = mem_cgroup_from_css(css);
mm/memcontrol.c
1080
if (cmpxchg(&iter->position, pos, next) != pos) {
mm/memcontrol.c
1086
if (!next) {
mm/memcontrol.c
1105
return next;
mm/memory-failure.c
1875
struct raw_hwp_page *p, *next;
mm/memory-failure.c
1879
llist_for_each_entry_safe(p, next, head, node) {
mm/memory-failure.c
438
struct to_kill *tk, *next;
mm/memory-failure.c
440
list_for_each_entry_safe(tk, next, to_kill, nd) {
mm/memory-failure.c
465
struct to_kill *tk, *next;
mm/memory-failure.c
467
list_for_each_entry_safe(tk, next, to_kill, nd) {
mm/memory.c
1383
unsigned long next;
mm/memory.c
1390
next = pmd_addr_end(addr, end);
mm/memory.c
1394
VM_BUG_ON_VMA(next-addr != HPAGE_PMD_SIZE, src_vma);
mm/memory.c
1406
addr, next))
mm/memory.c
1408
} while (dst_pmd++, src_pmd++, addr = next, addr != end);
mm/memory.c
1420
unsigned long next;
mm/memory.c
1427
next = pud_addr_end(addr, end);
mm/memory.c
1431
VM_BUG_ON_VMA(next-addr != HPAGE_PUD_SIZE, src_vma);
mm/memory.c
1443
addr, next))
mm/memory.c
1445
} while (dst_pud++, src_pud++, addr = next, addr != end);
mm/memory.c
1456
unsigned long next;
mm/memory.c
1463
next = p4d_addr_end(addr, end);
mm/memory.c
1467
addr, next))
mm/memory.c
1469
} while (dst_p4d++, src_p4d++, addr = next, addr != end);
mm/memory.c
1512
unsigned long next;
mm/memory.c
1549
next = pgd_addr_end(addr, end);
mm/memory.c
1553
addr, next))) {
mm/memory.c
1557
} while (dst_pgd++, src_pgd++, addr = next, addr != end);
mm/memory.c
1993
unsigned long next;
mm/memory.c
1997
next = pmd_addr_end(addr, end);
mm/memory.c
1999
if (next - addr != HPAGE_PMD_SIZE)
mm/memory.c
2002
addr = next;
mm/memory.c
2008
next - addr == HPAGE_PMD_SIZE && pmd_none(*pmd)) {
mm/memory.c
2018
addr = next;
mm/memory.c
2021
addr = zap_pte_range(tlb, vma, pmd, addr, next, details);
mm/memory.c
2022
if (addr != next)
mm/memory.c
203
unsigned long next;
mm/memory.c
2035
unsigned long next;
mm/memory.c
2039
next = pud_addr_end(addr, end);
mm/memory.c
2041
if (next - addr != HPAGE_PUD_SIZE)
mm/memory.c
2044
goto next;
mm/memory.c
2049
next = zap_pmd_range(tlb, vma, pud, addr, next, details);
mm/memory.c
2050
next:
mm/memory.c
2052
} while (pud++, addr = next, addr != end);
mm/memory.c
2063
unsigned long next;
mm/memory.c
2067
next = p4d_addr_end(addr, end);
mm/memory.c
2070
next = zap_pud_range(tlb, vma, p4d, addr, next, details);
mm/memory.c
2071
} while (p4d++, addr = next, addr != end);
mm/memory.c
2082
unsigned long next;
mm/memory.c
2088
next = pgd_addr_end(addr, end);
mm/memory.c
209
next = pmd_addr_end(addr, end);
mm/memory.c
2091
next = zap_p4d_range(tlb, vma, pgd, addr, next, details);
mm/memory.c
2092
} while (pgd++, addr = next, addr != end);
mm/memory.c
213
} while (pmd++, addr = next, addr != end);
mm/memory.c
237
unsigned long next;
mm/memory.c
243
next = pud_addr_end(addr, end);
mm/memory.c
246
free_pmd_range(tlb, pud, addr, next, floor, ceiling);
mm/memory.c
247
} while (pud++, addr = next, addr != end);
mm/memory.c
271
unsigned long next;
mm/memory.c
277
next = p4d_addr_end(addr, end);
mm/memory.c
280
free_pud_range(tlb, p4d, addr, next, floor, ceiling);
mm/memory.c
281
} while (p4d++, addr = next, addr != end);
mm/memory.c
2898
unsigned long next;
mm/memory.c
2907
next = pmd_addr_end(addr, end);
mm/memory.c
2908
err = remap_pte_range(mm, pmd, addr, next,
mm/memory.c
2912
} while (pmd++, addr = next, addr != end);
mm/memory.c
2921
unsigned long next;
mm/memory.c
2929
next = pud_addr_end(addr, end);
mm/memory.c
2930
err = remap_pmd_range(mm, pud, addr, next,
mm/memory.c
2934
} while (pud++, addr = next, addr != end);
mm/memory.c
2943
unsigned long next;
mm/memory.c
2951
next = p4d_addr_end(addr, end);
mm/memory.c
2952
err = remap_pud_range(mm, p4d, addr, next,
mm/memory.c
2956
} while (p4d++, addr = next, addr != end);
mm/memory.c
2983
unsigned long next;
mm/memory.c
2998
next = pgd_addr_end(addr, end);
mm/memory.c
2999
err = remap_p4d_range(mm, pgd, addr, next,
mm/memory.c
3003
} while (pgd++, addr = next, addr != end);
mm/memory.c
316
unsigned long next;
mm/memory.c
3264
unsigned long next;
mm/memory.c
3277
next = pmd_addr_end(addr, end);
mm/memory.c
3287
err = apply_to_pte_range(mm, pmd, addr, next,
mm/memory.c
3291
} while (pmd++, addr = next, addr != end);
mm/memory.c
3302
unsigned long next;
mm/memory.c
3313
next = pud_addr_end(addr, end);
mm/memory.c
3323
err = apply_to_pmd_range(mm, pud, addr, next,
mm/memory.c
3327
} while (pud++, addr = next, addr != end);
mm/memory.c
3338
unsigned long next;
mm/memory.c
3349
next = p4d_addr_end(addr, end);
mm/memory.c
3359
err = apply_to_pud_range(mm, p4d, addr, next,
mm/memory.c
3363
} while (p4d++, addr = next, addr != end);
mm/memory.c
3373
unsigned long start = addr, next;
mm/memory.c
3383
next = pgd_addr_end(addr, end);
mm/memory.c
3395
err = apply_to_p4d_range(mm, pgd, addr, next,
mm/memory.c
3399
} while (pgd++, addr = next, addr != end);
mm/memory.c
366
next = pgd_addr_end(addr, end);
mm/memory.c
369
free_p4d_range(tlb, pgd, addr, next, floor, ceiling);
mm/memory.c
370
} while (pgd++, addr = next, addr != end);
mm/memory.c
404
struct vm_area_struct *next;
mm/memory.c
406
next = mas_find(mas, unmap->tree_end - 1);
mm/memory.c
422
while (next && next->vm_start <= vma->vm_end + PMD_SIZE) {
mm/memory.c
423
vma = next;
mm/memory.c
424
next = mas_find(mas, unmap->tree_end - 1);
mm/memory.c
433
next ? next->vm_start : unmap->pg_end);
mm/memory.c
434
vma = next;
mm/memory.c
5188
goto next;
mm/memory.c
5202
next:
mm/mempolicy.c
3147
struct rb_node *next = rb_next(&n->nd);
mm/mempolicy.c
3170
if (!next)
mm/mempolicy.c
3172
n = rb_entry(next, struct sp_node, nd);
mm/mempolicy.c
3273
struct rb_node *next;
mm/mempolicy.c
3278
next = rb_first(&sp->root);
mm/mempolicy.c
3279
while (next) {
mm/mempolicy.c
3280
n = rb_entry(next, struct sp_node, nd);
mm/mempolicy.c
3281
next = rb_next(&n->nd);
mm/mempolicy.c
913
struct vm_area_struct *next, *vma = walk->vma;
mm/mempolicy.c
927
next = find_vma(vma->vm_mm, vma->vm_end);
mm/mempolicy.c
930
(!next || vma->vm_end < next->vm_start)))
mm/migrate.c
2033
list_move(from->next, &folios);
mm/migrate_device.c
1120
goto next;
mm/migrate_device.c
1127
goto next;
mm/migrate_device.c
1160
goto next;
mm/migrate_device.c
1179
goto next;
mm/migrate_device.c
1186
goto next;
mm/migrate_device.c
1208
goto next;
mm/migrate_device.c
1216
goto next;
mm/migrate_device.c
1233
next:
mm/migrate_device.c
292
goto next;
mm/migrate_device.c
303
goto next;
mm/migrate_device.c
310
goto next;
mm/migrate_device.c
341
goto next;
mm/migrate_device.c
346
goto next;
mm/migrate_device.c
353
goto next;
mm/migrate_device.c
379
goto next;
mm/migrate_device.c
422
goto next;
mm/migrate_device.c
479
next:
mm/migrate_device.c
586
goto next;
mm/migrate_device.c
607
goto next;
mm/migrate_device.c
626
goto next;
mm/migrate_device.c
630
next:
mm/mmap.c
1169
struct vm_area_struct *next, *prev = vma;
mm/mmap.c
1171
for_each_vma_range(vmi, next, start + size) {
mm/mmap.c
1173
if (next->vm_start != prev->vm_end)
mm/mmap.c
1176
if (next->vm_file != vma->vm_file)
mm/mmap.c
1179
if (next->vm_flags != vma->vm_flags)
mm/mmap.c
1182
if (start + size <= next->vm_end)
mm/mmap.c
1185
prev = next;
mm/mmap.c
1188
if (!next)
mm/mmap.c
120
struct vm_area_struct *brkvma, *next = NULL;
mm/mmap.c
189
next = vma_find(&vmi, newbrk + PAGE_SIZE + stack_guard_gap);
mm/mmap.c
190
if (next && newbrk + PAGE_SIZE > vm_start_gap(next))
mm/mmu_gather.c
150
for (batch = &tlb->local; batch && batch->nr; batch = batch->next)
mm/mmu_gather.c
157
struct mmu_gather_batch *batch, *next;
mm/mmu_gather.c
159
for (batch = tlb->local.next; batch; batch = next) {
mm/mmu_gather.c
160
next = batch->next;
mm/mmu_gather.c
163
tlb->local.next = NULL;
mm/mmu_gather.c
29
if (batch->next) {
mm/mmu_gather.c
30
tlb->active = batch->next;
mm/mmu_gather.c
416
tlb->local.next = NULL;
mm/mmu_gather.c
42
batch->next = NULL;
mm/mmu_gather.c
46
tlb->active->next = batch;
mm/mmu_notifier.c
134
struct hlist_node *next;
mm/mmu_notifier.c
152
hlist_for_each_entry_safe(interval_sub, next,
mm/mprotect.c
456
unsigned long next;
mm/mprotect.c
465
next = pmd_addr_end(addr, end);
mm/mprotect.c
474
goto next;
mm/mprotect.c
478
if ((next - addr != HPAGE_PMD_SIZE) ||
mm/mprotect.c
501
goto next;
mm/mprotect.c
507
ret = change_pte_range(tlb, vma, pmd, addr, next, newprot,
mm/mprotect.c
512
next:
mm/mprotect.c
514
} while (pmd++, addr = next, addr != end);
mm/mprotect.c
527
unsigned long next;
mm/mprotect.c
535
next = pud_addr_end(addr, end);
mm/mprotect.c
554
if ((next - addr != PUD_SIZE) ||
mm/mprotect.c
570
pages += change_pmd_range(tlb, vma, pudp, addr, next, newprot,
mm/mprotect.c
572
} while (pudp++, addr = next, addr != end);
mm/mprotect.c
585
unsigned long next;
mm/mprotect.c
590
next = p4d_addr_end(addr, end);
mm/mprotect.c
596
pages += change_pud_range(tlb, vma, p4d, addr, next, newprot,
mm/mprotect.c
598
} while (p4d++, addr = next, addr != end);
mm/mprotect.c
609
unsigned long next;
mm/mprotect.c
616
next = pgd_addr_end(addr, end);
mm/mprotect.c
624
pages += change_p4d_range(tlb, vma, pgd, addr, next, newprot,
mm/mprotect.c
626
} while (pgd++, addr = next, addr != end);
mm/mprotect.c
665
unsigned long next, struct mm_walk *walk)
mm/mprotect.c
673
unsigned long addr, unsigned long next,
mm/mprotect.c
681
static int prot_none_test(unsigned long addr, unsigned long next,
mm/mremap.c
1170
struct vm_area_struct *next = vma_next(&vmi);
mm/mremap.c
1172
vm_flags_set(next, VM_ACCOUNT); /* Acquires VMA lock. */
mm/mremap.c
544
unsigned long next, extent, mask, size;
mm/mremap.c
565
next = (old_addr + size) & mask;
mm/mremap.c
567
extent = next - old_addr;
mm/mremap.c
570
next = (new_addr + size) & mask;
mm/mremap.c
571
if (extent > next - new_addr)
mm/mremap.c
572
extent = next - new_addr;
mm/oom_kill.c
315
goto next;
mm/oom_kill.c
319
goto next;
mm/oom_kill.c
329
goto next;
mm/oom_kill.c
344
goto next;
mm/oom_kill.c
352
next:
mm/page_alloc.c
6907
struct page *page, *next;
mm/page_alloc.c
6910
list_for_each_entry_safe(page, next, &list[order], lru) {
mm/page_owner.c
138
dummy_stack.next = &failure_stack;
mm/page_owner.c
192
stack->next = NULL;
mm/page_owner.c
195
stack->next = stack_list;
mm/page_owner.c
41
struct stack *next;
mm/page_owner.c
883
stack = stack->next;
mm/page_owner.c
931
.next = stack_next,
mm/page_reporting.c
153
struct page *page, *next;
mm/page_reporting.c
183
list_for_each_entry_safe(page, next, list, lru) {
mm/page_reporting.c
195
next = page;
mm/page_reporting.c
202
next = page;
mm/page_reporting.c
243
next = list_first_entry(list, struct page, lru);
mm/page_reporting.c
251
if (!list_entry_is_head(next, list, lru) && !list_is_first(&next->lru, list))
mm/page_reporting.c
252
list_rotate_to_front(&next->lru, list);
mm/pagewalk.c
102
unsigned long next;
mm/pagewalk.c
130
next = pmd_addr_end(addr, end);
mm/pagewalk.c
135
err = ops->pte_hole(addr, next, depth, walk);
mm/pagewalk.c
149
err = ops->pmd_entry(pmd, addr, next, walk);
mm/pagewalk.c
174
err = walk_pte_range(pmd, addr, next, walk);
mm/pagewalk.c
181
} while (pmd++, addr = next, addr != end);
mm/pagewalk.c
190
unsigned long next;
mm/pagewalk.c
200
next = pud_addr_end(addr, end);
mm/pagewalk.c
205
err = ops->pte_hole(addr, next, depth, walk);
mm/pagewalk.c
215
err = ops->pud_entry(pud, addr, next, walk);
mm/pagewalk.c
240
err = walk_pmd_range(pud, addr, next, walk);
mm/pagewalk.c
246
} while (pud++, addr = next, addr != end);
mm/pagewalk.c
255
unsigned long next;
mm/pagewalk.c
264
next = p4d_addr_end(addr, end);
mm/pagewalk.c
269
err = ops->pte_hole(addr, next, depth, walk);
mm/pagewalk.c
276
err = ops->p4d_entry(p4d, addr, next, walk);
mm/pagewalk.c
281
err = walk_pud_range(p4d, addr, next, walk);
mm/pagewalk.c
284
} while (p4d++, addr = next, addr != end);
mm/pagewalk.c
293
unsigned long next;
mm/pagewalk.c
305
next = pgd_addr_end(addr, end);
mm/pagewalk.c
310
err = ops->pte_hole(addr, next, 0, walk);
mm/pagewalk.c
317
err = ops->pgd_entry(pgd, addr, next, walk);
mm/pagewalk.c
322
err = walk_p4d_range(pgd, addr, next, walk);
mm/pagewalk.c
325
} while (pgd++, addr = next, addr != end);
mm/pagewalk.c
344
unsigned long next;
mm/pagewalk.c
353
next = hugetlb_entry_end(h, addr, end);
mm/pagewalk.c
356
err = ops->hugetlb_entry(pte, hmask, addr, next, walk);
mm/pagewalk.c
358
err = ops->pte_hole(addr, next, -1, walk);
mm/pagewalk.c
361
} while (addr = next, addr != end);
mm/pagewalk.c
480
unsigned long next;
mm/pagewalk.c
500
next = end;
mm/pagewalk.c
502
err = ops->pte_hole(start, next, -1, &walk);
mm/pagewalk.c
505
next = min(end, vma->vm_start);
mm/pagewalk.c
507
err = ops->pte_hole(start, next, -1, &walk);
mm/pagewalk.c
511
next = min(end, vma->vm_end);
mm/pagewalk.c
514
err = walk_page_test(start, next, &walk);
mm/pagewalk.c
526
err = __walk_page_range(start, next, &walk);
mm/pagewalk.c
530
} while (start = next, start < end);
mm/percpu.c
1744
struct pcpu_chunk *chunk, *next;
mm/percpu.c
1817
list_for_each_entry_safe(chunk, next, &pcpu_chunk_lists[slot],
mm/percpu.c
1955
struct pcpu_chunk *chunk, *next;
mm/percpu.c
1963
list_for_each_entry_safe(chunk, next, free_head, list) {
mm/percpu.c
1978
list_for_each_entry_safe(chunk, next, &to_free, list) {
mm/pgtable-generic.c
427
struct ptdesc *pt, *next;
mm/pgtable-generic.c
435
list_for_each_entry_safe(pt, next, &page_list, pt_list)
mm/ptdump.c
100
unsigned long next, struct mm_walk *walk)
mm/ptdump.c
121
unsigned long next, struct mm_walk *walk)
mm/ptdump.c
134
static int ptdump_hole(unsigned long addr, unsigned long next,
mm/ptdump.c
31
unsigned long next, struct mm_walk *walk)
mm/ptdump.c
54
unsigned long next, struct mm_walk *walk)
mm/ptdump.c
77
unsigned long next, struct mm_walk *walk)
mm/rmap.c
454
struct anon_vma_chain *avc, *next;
mm/rmap.c
456
list_for_each_entry_safe(avc, next, &vma->anon_vma_chain, same_vma) {
mm/rmap.c
481
struct anon_vma_chain *avc, *next;
mm/rmap.c
499
list_for_each_entry_safe(avc, next, &vma->anon_vma_chain, same_vma) {
mm/rmap.c
531
list_for_each_entry_safe(avc, next, &vma->anon_vma_chain, same_vma) {
mm/shmem.c
108
pgoff_t next; /* the next page offset to be fallocated */
mm/shmem.c
1541
struct shmem_inode_info *info, *next;
mm/shmem.c
1549
list_for_each_entry_safe(info, next, &shmem_swaplist, swaplist) {
mm/shmem.c
1573
next = list_next_entry(info, swaplist);
mm/shmem.c
1666
index < shmem_falloc->next)
mm/shmem.c
2719
vmf->pgoff < shmem_falloc->next) {
mm/shmem.c
3719
shmem_falloc.next = (unmap_end + 1) >> PAGE_SHIFT;
mm/shmem.c
3759
shmem_falloc.next = start;
mm/shmem.c
3818
shmem_falloc.nr_falloced += index - shmem_falloc.next;
mm/shmem.c
3819
shmem_falloc.next = index;
mm/shmem.c
730
LIST_HEAD(list), *pos, *next;
mm/shmem.c
741
list_for_each_safe(pos, next, &sbinfo->shrinklist) {
mm/shmem.c
750
goto next;
mm/shmem.c
754
next:
mm/shmem.c
761
list_for_each_safe(pos, next, &list) {
mm/shmem.c
762
pgoff_t next, end;
mm/shmem.c
784
next = folio_next_index(folio);
mm/shmem.c
786
if (end <= folio->index || end >= next) {
mm/shmem.c
811
freed += next - end;
mm/slab_common.c
1134
if (p == slab_caches.next)
mm/slab_common.c
1189
.next = slab_next,
mm/slab_common.c
1528
struct rcu_head *next;
mm/slab_common.c
1530
for (; head; head = next) {
mm/slab_common.c
1534
next = head->next;
mm/slab_common.c
1980
head->next = krcp->head;
mm/slub.c
2694
void *next = *head;
mm/slub.c
2698
if (is_kfence_address(next)) {
mm/slub.c
2699
slab_free_hook(s, next, false, false);
mm/slub.c
2710
object = next;
mm/slub.c
2711
next = get_freepointer(s, object);
mm/slub.c
3376
void *next;
mm/slub.c
3406
next = next_freelist_entry(s, &pos, start, page_limit,
mm/slub.c
3408
next = setup_object(s, next);
mm/slub.c
3409
set_freepointer(s, cur, next);
mm/slub.c
3410
cur = next;
mm/slub.c
3461
void *start, *p, *next;
mm/slub.c
3521
next = p + s->size;
mm/slub.c
3522
next = setup_object(s, next);
mm/slub.c
3523
set_freepointer(s, p, next);
mm/slub.c
3524
p = next;
mm/slub.c
9532
struct saved_alias *next;
mm/slub.c
9560
al->next = alias_list;
mm/slub.c
9592
alias_list = alias_list->next;
mm/slub.c
9706
.next = slab_debugfs_next,
mm/sparse-vmemmap.c
406
unsigned long addr, unsigned long next)
mm/sparse-vmemmap.c
411
unsigned long addr, unsigned long next)
mm/sparse-vmemmap.c
420
unsigned long next;
mm/sparse-vmemmap.c
426
for (addr = start; addr < end; addr = next) {
mm/sparse-vmemmap.c
427
next = pmd_addr_end(addr, end);
mm/sparse-vmemmap.c
447
vmemmap_set_pmd(pmd, p, node, addr, next);
mm/sparse-vmemmap.c
460
} else if (vmemmap_check_pmd(pmd, node, addr, next))
mm/sparse-vmemmap.c
462
if (vmemmap_populate_basepages(addr, next, node, altmap))
mm/sparse-vmemmap.c
531
unsigned long next, last = addr + size;
mm/sparse-vmemmap.c
539
next = addr + PAGE_SIZE;
mm/sparse-vmemmap.c
540
pte = vmemmap_populate_address(next, node, NULL, -1, 0);
mm/sparse-vmemmap.c
548
next += PAGE_SIZE;
mm/sparse-vmemmap.c
549
rc = vmemmap_populate_range(next, last, node, NULL,
mm/swapfile.c
1058
offset = si->global_cluster->next[order];
mm/swapfile.c
1350
struct swap_info_struct *si, *next;
mm/swapfile.c
1354
plist_for_each_entry_safe(si, next, &swap_avail_head, avail_list) {
mm/swapfile.c
1378
if (plist_node_empty(&next->avail_list))
mm/swapfile.c
1391
struct swap_info_struct *si, *next;
mm/swapfile.c
1395
plist_for_each_entry_safe(si, next, &swap_active_head, list) {
mm/swapfile.c
1406
if (plist_node_empty(&next->list))
mm/swapfile.c
2265
unsigned long next;
mm/swapfile.c
2271
next = pmd_addr_end(addr, end);
mm/swapfile.c
2272
ret = unuse_pte_range(vma, pmd, addr, next, type);
mm/swapfile.c
2275
} while (pmd++, addr = next, addr != end);
mm/swapfile.c
2284
unsigned long next;
mm/swapfile.c
2289
next = pud_addr_end(addr, end);
mm/swapfile.c
2292
ret = unuse_pmd_range(vma, pud, addr, next, type);
mm/swapfile.c
2295
} while (pud++, addr = next, addr != end);
mm/swapfile.c
2304
unsigned long next;
mm/swapfile.c
2309
next = p4d_addr_end(addr, end);
mm/swapfile.c
2312
ret = unuse_pud_range(vma, p4d, addr, next, type);
mm/swapfile.c
2315
} while (p4d++, addr = next, addr != end);
mm/swapfile.c
2322
unsigned long addr, end, next;
mm/swapfile.c
2330
next = pgd_addr_end(addr, end);
mm/swapfile.c
2333
ret = unuse_p4d_range(vma, pgd, addr, next, type);
mm/swapfile.c
2336
} while (pgd++, addr = next, addr != end);
mm/swapfile.c
2425
(p = p->next) != &init_mm.mmlist) {
mm/swapfile.c
2509
struct list_head *p, *next;
mm/swapfile.c
2516
list_for_each_safe(p, next, &init_mm.mmlist)
mm/swapfile.c
3000
.next = swap_next,
mm/swapfile.c
3272
si->global_cluster->next[i] = SWAP_ENTRY_INVALID;
mm/swapfile.c
3898
struct page *page, *next;
mm/swapfile.c
3900
list_for_each_entry_safe(page, next, &head->lru, lru) {
mm/swapfile.c
918
unsigned int next = SWAP_ENTRY_INVALID, found = SWAP_ENTRY_INVALID;
mm/swapfile.c
950
next = offset;
mm/swapfile.c
957
this_cpu_write(percpu_swap_cluster.offset[order], next);
mm/swapfile.c
960
si->global_cluster->next[order] = next;
mm/vma.c
1049
struct vm_area_struct *next = vmg->next;
mm/vma.c
105
struct vm_area_struct *tgt = merge_next ? vmg->next : vmg->prev;
mm/vma.c
1062
if ((vmg->vm_flags & VM_SPECIAL) || (!prev && !next))
mm/vma.c
1070
vmg->end = next->vm_end;
mm/vma.c
1071
vmg->target = next;
mm/vma.c
1085
if (can_merge_right && !can_merge_remove_vma(next))
mm/vma.c
1155
struct vm_area_struct *next = vmg->next;
mm/vma.c
1163
if (next && target != next && vmg->end == next->vm_end)
mm/vma.c
1169
VM_WARN_ON_VMG(remove_next && !can_merge_remove_vma(next), vmg);
mm/vma.c
1171
VM_WARN_ON_VMG(next && !remove_next &&
mm/vma.c
1172
next != target && vmg->end > next->vm_start, vmg);
mm/vma.c
1180
sticky_flags |= next->vm_flags & VM_STICKY;
mm/vma.c
1190
ret = dup_anon_vma(target, next, &anon_dup);
mm/vma.c
1197
vma_start_write(next);
mm/vma.c
1382
struct vm_area_struct *next = NULL;
mm/vma.c
1420
for_each_vma_range(*(vms->vmi), next, vms->end) {
mm/vma.c
1423
if (vma_is_sealed(next)) {
mm/vma.c
1428
if (next->vm_end > vms->end) {
mm/vma.c
1429
error = __split_vma(vms->vmi, next, vms->end, 0);
mm/vma.c
1433
vma_start_write(next);
mm/vma.c
1435
error = mas_store_gfp(mas_detach, next, GFP_KERNEL);
mm/vma.c
1439
vma_mark_detached(next);
mm/vma.c
1440
nrpages = vma_pages(next);
mm/vma.c
1443
if (next->vm_flags & VM_LOCKED)
mm/vma.c
1446
if (next->vm_flags & VM_ACCOUNT)
mm/vma.c
1449
if (is_exec_mapping(next->vm_flags))
mm/vma.c
1451
else if (is_stack_mapping(next->vm_flags))
mm/vma.c
1453
else if (is_data_mapping(next->vm_flags))
mm/vma.c
1466
error = userfaultfd_unmap_prep(next, vms->start,
mm/vma.c
1472
BUG_ON(next->vm_start < vms->start);
mm/vma.c
1473
BUG_ON(next->vm_start > vms->end);
mm/vma.c
1477
vms->next = vma_next(vms->vmi);
mm/vma.c
1478
if (vms->next)
mm/vma.c
1479
vms->unmap_end = vms->next->vm_start;
mm/vma.c
158
*remove = vmg->next;
mm/vma.c
163
adjust = vmg->next;
mm/vma.c
1763
vmg.next = vma_iter_next_rewind(vmi, NULL);
mm/vma.c
1878
vmg.next = vma_iter_next_rewind(&vmi, NULL);
mm/vma.c
2006
struct vm_area_struct *prev, *next;
mm/vma.c
201
if (vmg->next->vm_pgoff == vmg->pgoff + pglen)
mm/vma.c
2010
next = vma_iter_load(&vmi);
mm/vma.c
2011
if (next) {
mm/vma.c
2012
anon_vma = reusable_anon_vma(next, vma, next);
mm/vma.c
2418
map->next = vms->next;
mm/vma.c
2421
map->next = vma_iter_next_rewind(vmi, &map->prev);
mm/vma.c
2471
map->prev, map->next);
mm/vma.c
2751
if (map.prev || map.next) {
mm/vma.c
3093
struct vm_area_struct *next;
mm/vma.c
3116
next = find_vma_intersection(mm, vma->vm_end, gap_addr);
mm/vma.c
3117
if (next && vma_is_accessible(next)) {
mm/vma.c
3118
if (!(next->vm_flags & VM_GROWSUP))
mm/vma.c
3123
if (next)
mm/vma.c
32
struct vm_area_struct *next;
mm/vma.c
439
struct vm_area_struct *next = vmg->next;
mm/vma.c
442
if (!next || vmg->end != next->vm_start || !can_vma_merge_before(vmg))
mm/vma.c
456
return !prev->anon_vma || !next->anon_vma ||
mm/vma.c
457
prev->anon_vma == next->anon_vma;
mm/vma.c
71
.next = (vma_) ? NULL : (map_)->next, \
mm/vma.c
711
adjust = vmg->next;
mm/vma.c
736
vma_iter_config(vmg->vmi, vmg->end, vmg->next->vm_end);
mm/vma.c
811
struct vm_area_struct *next;
mm/vma.c
822
VM_WARN_ON_VMG(vmg->next, vmg); /* We set this. */
mm/vma.c
856
next = vmg->next = vma_iter_next_range(vmg->vmi);
mm/vma.c
862
next = NULL;
mm/vma.c
88
struct vm_area_struct *vma = merge_next ? vmg->next : vmg->prev;
mm/vma.c
891
if (vmg->__remove_next && !can_merge_remove_vma(next)) {
mm/vma.c
901
vma_start_write(next);
mm/vma.c
902
vmg->target = next;
mm/vma.c
903
sticky_flags |= (next->vm_flags & VM_STICKY);
mm/vma.c
921
vmg->end = next->vm_end;
mm/vma.c
929
err = dup_anon_vma(prev, next->anon_vma ? next : middle,
mm/vma.c
963
vmg->end = next->vm_end;
mm/vma.c
964
vmg->pgoff = next->vm_pgoff - pglen;
mm/vma.c
973
err = dup_anon_vma(next, middle, &anon_dup);
mm/vma.h
253
.next = NULL, \
mm/vma.h
38
struct vm_area_struct *next; /* vma after the munmap area */
mm/vma.h
641
struct vm_area_struct *next = vma_next(vmi);
mm/vma.h
658
return next;
mm/vma.h
86
struct vm_area_struct *next;
mm/vma_exec.c
40
struct vm_area_struct *next;
mm/vma_exec.c
70
next = vma_next(&vmi);
mm/vma_exec.c
76
next ? next->vm_start : USER_PGTABLES_CEILING);
mm/vma_exec.c
85
next ? next->vm_start : USER_PGTABLES_CEILING);
mm/vmalloc.c
1243
return (&parent->rb_right == link ? list->next : list);
mm/vmalloc.c
1446
struct list_head *next;
mm/vmalloc.c
1462
next = get_va_next_sibling(parent, link);
mm/vmalloc.c
1463
if (unlikely(next == NULL))
mm/vmalloc.c
1473
if (next != head) {
mm/vmalloc.c
1474
sibling = list_entry(next, struct vmap_area, list);
mm/vmalloc.c
1494
if (next->prev != head) {
mm/vmalloc.c
1495
sibling = list_entry(next->prev, struct vmap_area, list);
mm/vmalloc.c
172
unsigned long next;
mm/vmalloc.c
179
next = pmd_addr_end(addr, end);
mm/vmalloc.c
181
if (vmap_try_huge_pmd(pmd, addr, next, phys_addr, prot,
mm/vmalloc.c
187
err = vmap_pte_range(pmd, addr, next, phys_addr, prot, max_page_shift, mask);
mm/vmalloc.c
190
} while (pmd++, phys_addr += (next - addr), addr = next, addr != end);
mm/vmalloc.c
224
unsigned long next;
mm/vmalloc.c
231
next = pud_addr_end(addr, end);
mm/vmalloc.c
233
if (vmap_try_huge_pud(pud, addr, next, phys_addr, prot,
mm/vmalloc.c
239
err = vmap_pmd_range(pud, addr, next, phys_addr, prot, max_page_shift, mask);
mm/vmalloc.c
242
} while (pud++, phys_addr += (next - addr), addr = next, addr != end);
mm/vmalloc.c
276
unsigned long next;
mm/vmalloc.c
283
next = p4d_addr_end(addr, end);
mm/vmalloc.c
285
if (vmap_try_huge_p4d(p4d, addr, next, phys_addr, prot,
mm/vmalloc.c
291
err = vmap_pud_range(p4d, addr, next, phys_addr, prot, max_page_shift, mask);
mm/vmalloc.c
294
} while (p4d++, phys_addr += (next - addr), addr = next, addr != end);
mm/vmalloc.c
304
unsigned long next;
mm/vmalloc.c
3149
for (p = &vmlist; (tmp = *p) != NULL; p = &tmp->next) {
mm/vmalloc.c
3156
vm->next = *p;
mm/vmalloc.c
3179
for (p = &vmlist; (cur = *p) != NULL; p = &cur->next) {
mm/vmalloc.c
3187
vm->next = *p;
mm/vmalloc.c
319
next = pgd_addr_end(addr, end);
mm/vmalloc.c
320
err = vmap_p4d_range(pgd, addr, next, phys_addr, prot,
mm/vmalloc.c
324
} while (pgd++, phys_addr += (next - addr), addr = next, addr != end);
mm/vmalloc.c
401
unsigned long next;
mm/vmalloc.c
406
next = pmd_addr_end(addr, end);
mm/vmalloc.c
413
WARN_ON(next - addr < PMD_SIZE);
mm/vmalloc.c
418
vunmap_pte_range(pmd, addr, next, mask);
mm/vmalloc.c
421
} while (pmd++, addr = next, addr != end);
mm/vmalloc.c
428
unsigned long next;
mm/vmalloc.c
433
next = pud_addr_end(addr, end);
mm/vmalloc.c
440
WARN_ON(next - addr < PUD_SIZE);
mm/vmalloc.c
445
vunmap_pmd_range(pud, addr, next, mask);
mm/vmalloc.c
446
} while (pud++, addr = next, addr != end);
mm/vmalloc.c
453
unsigned long next;
mm/vmalloc.c
457
next = p4d_addr_end(addr, end);
mm/vmalloc.c
4602
unsigned long next;
mm/vmalloc.c
465
vunmap_pud_range(p4d, addr, next, mask);
mm/vmalloc.c
466
} while (p4d++, addr = next, addr != end);
mm/vmalloc.c
4678
next = va->va_end;
mm/vmalloc.c
4680
} while ((vn = find_vmap_area_exceed_addr_lock(next, &va)));
mm/vmalloc.c
483
unsigned long next;
mm/vmalloc.c
491
next = pgd_addr_end(addr, end);
mm/vmalloc.c
496
vunmap_p4d_range(pgd, addr, next, &mask);
mm/vmalloc.c
497
} while (pgd++, addr = next, addr != end);
mm/vmalloc.c
5317
for (busy = vmlist; busy; busy = busy->next) {
mm/vmalloc.c
5457
for (tmp = vmlist; tmp; tmp = tmp->next) {
mm/vmalloc.c
574
unsigned long next;
mm/vmalloc.c
580
next = pmd_addr_end(addr, end);
mm/vmalloc.c
581
if (vmap_pages_pte_range(pmd, addr, next, prot, pages, nr, mask))
mm/vmalloc.c
583
} while (pmd++, addr = next, addr != end);
mm/vmalloc.c
592
unsigned long next;
mm/vmalloc.c
598
next = pud_addr_end(addr, end);
mm/vmalloc.c
599
if (vmap_pages_pmd_range(pud, addr, next, prot, pages, nr, mask))
mm/vmalloc.c
601
} while (pud++, addr = next, addr != end);
mm/vmalloc.c
610
unsigned long next;
mm/vmalloc.c
616
next = p4d_addr_end(addr, end);
mm/vmalloc.c
617
if (vmap_pages_pud_range(p4d, addr, next, prot, pages, nr, mask))
mm/vmalloc.c
619
} while (p4d++, addr = next, addr != end);
mm/vmalloc.c
628
unsigned long next;
mm/vmalloc.c
636
next = pgd_addr_end(addr, end);
mm/vmalloc.c
639
err = vmap_pages_p4d_range(pgd, addr, next, prot, pages, &nr, &mask);
mm/vmalloc.c
642
} while (pgd++, addr = next, addr != end);
mm/vmscan.c
1629
struct folio *folio, *next;
mm/vmscan.c
1633
list_for_each_entry_safe(folio, next, folio_list, lru) {
mm/vmscan.c
2956
mm_state->tail = mm_state->tail->next;
mm/vmscan.c
3077
mm_state->head = mm_state->head->next;
mm/vmscan.c
3086
mm_state->tail = mm_state->head->next;
mm/vmscan.c
3617
goto next;
mm/vmscan.c
3623
goto next;
mm/vmscan.c
3628
goto next;
mm/vmscan.c
3632
goto next;
mm/vmscan.c
3635
goto next;
mm/vmscan.c
3648
next:
mm/vmscan.c
3665
unsigned long next;
mm/vmscan.c
3684
for (i = pmd_index(start), addr = start; addr != end; i++, addr = next) {
mm/vmscan.c
3687
next = pmd_addr_end(addr, end);
mm/vmscan.c
3718
if (!walk_pte_range(&val, addr, next, args))
mm/vmscan.c
3739
unsigned long next;
mm/vmscan.c
3746
for (i = pud_index(start), addr = start; addr != end; i++, addr = next) {
mm/vmscan.c
3749
next = pud_addr_end(addr, end);
mm/vmscan.c
3754
walk_pmd_range(&val, addr, next, args);
mm/vmscan.c
3916
goto next;
mm/vmscan.c
3922
next:
mm/vmscan.c
3959
int prev, next;
mm/vmscan.c
3993
next = lru_gen_from_seq(lrugen->max_seq + 1);
mm/vmscan.c
3999
lrugen->nr_pages[next][type][zone];
mm/vmscan.c
4012
WRITE_ONCE(lrugen->timestamps[next], jiffies);
mm/vmscan.c
4695
struct folio *next;
mm/vmscan.c
4725
list_for_each_entry_safe_reverse(folio, next, &list, lru) {
mm/vmscan.c
5468
.next = lru_gen_seq_next,
mm/vmscan.c
5558
char *cur, *next;
mm/vmscan.c
5588
next = buf;
mm/vmscan.c
5589
next[len] = '\0';
mm/vmscan.c
5591
while ((cur = strsep(&next, ",;\n"))) {
mm/vmscan.c
5722
lruvec->lrugen.list.next = LIST_POISON1;
mm/vmstat.c
1725
.next = frag_next,
mm/vmstat.c
1732
.next = frag_next,
mm/vmstat.c
1866
.next = frag_next,
mm/vmstat.c
1958
.next = vmstat_next,
mm/vmstat.c
2366
.next = frag_next,
mm/vmstat.c
2408
.next = frag_next,
mm/zpdesc.h
37
struct zpdesc *next;
mm/zpdesc.h
57
ZPDESC_MATCH(__folio_index, next);
mm/zsmalloc.c
1269
set_freeobj(zspage, link->next >> OBJ_TAG_BITS);
mm/zsmalloc.c
1375
link->next = get_freeobj(zspage) << OBJ_TAG_BITS;
mm/zsmalloc.c
188
unsigned long next;
mm/zsmalloc.c
712
return zpdesc->next;
mm/zsmalloc.c
779
zpdesc->next = NULL;
mm/zsmalloc.c
807
struct zpdesc *zpdesc, *next;
mm/zsmalloc.c
814
next = zpdesc = get_first_zpdesc(zspage);
mm/zsmalloc.c
817
next = get_next_zpdesc(zpdesc);
mm/zsmalloc.c
822
zpdesc = next;
mm/zsmalloc.c
869
link->next = freeobj++ << OBJ_TAG_BITS;
mm/zsmalloc.c
880
link->next = freeobj++ << OBJ_TAG_BITS;
mm/zsmalloc.c
886
link->next = -1UL << OBJ_TAG_BITS;
mm/zsmalloc.c
915
zpdesc->next = NULL;
mm/zsmalloc.c
923
prev_zpdesc->next = zpdesc;
net/802/garp.c
209
struct rb_node *node, *next;
net/802/garp.c
213
next = node ? rb_next(node) : NULL, node != NULL;
net/802/garp.c
214
node = next) {
net/802/garp.c
396
struct rb_node *node, *next;
net/802/garp.c
400
next = node ? rb_next(node) : NULL, node != NULL;
net/802/garp.c
401
node = next) {
net/802/mrp.c
298
struct rb_node *node, *next;
net/802/mrp.c
302
next = node ? rb_next(node) : NULL, node != NULL;
net/802/mrp.c
303
node = next) {
net/802/mrp.c
581
struct rb_node *node, *next;
net/802/mrp.c
585
next = node ? rb_next(node) : NULL, node != NULL;
net/802/mrp.c
586
node = next) {
net/8021q/vlan_dev.c
190
mp = mp->next;
net/8021q/vlan_dev.c
199
np->next = mp;
net/8021q/vlan_dev.c
608
vlan->egress_priority_map[i] = pm->next;
net/8021q/vlan_netlink.c
264
pm = pm->next) {
net/8021q/vlanproc.c
271
mp = mp->next;
net/8021q/vlanproc.c
68
.next = vlan_seq_next,
net/9p/trans_fd.c
432
req = list_entry(m->unsent_req_list.next, struct p9_req_t,
net/9p/trans_fd.c
802
m, m->mux_list.prev, m->mux_list.next);
net/9p/trans_rdma.c
290
wr.next = NULL;
net/9p/trans_rdma.c
387
wr.next = NULL;
net/appletalk/aarp.c
1023
.next = aarp_seq_next,
net/appletalk/aarp.c
274
*n = (*n)->next;
net/appletalk/aarp.c
277
n = &((*n)->next);
net/appletalk/aarp.c
293
*n = (*n)->next;
net/appletalk/aarp.c
297
n = &((*n)->next);
net/appletalk/aarp.c
314
*n = (*n)->next;
net/appletalk/aarp.c
317
n = &((*n)->next);
net/appletalk/aarp.c
371
*n = (*n)->next;
net/appletalk/aarp.c
420
list = list->next;
net/appletalk/aarp.c
499
entry->next = proxies[hash];
net/appletalk/aarp.c
628
a->next = unresolved[hash];
net/appletalk/aarp.c
680
*list = a->next;
net/appletalk/aarp.c
683
a->next = resolved[hash];
net/appletalk/aarp.c
693
list = &((*list)->next);
net/appletalk/aarp.c
70
struct aarp_entry *next;
net/appletalk/aarp.c
915
for (entry = table[ct]; entry; entry = entry->next) {
net/appletalk/aarp.c
962
else if (entry->next)
net/appletalk/aarp.c
963
entry = entry->next;
net/appletalk/atalk_proc.c
109
r = r->next;
net/appletalk/atalk_proc.c
191
.next = atalk_seq_interface_next,
net/appletalk/atalk_proc.c
198
.next = atalk_seq_route_next,
net/appletalk/atalk_proc.c
205
.next = atalk_seq_socket_next,
net/appletalk/atalk_proc.c
21
for (i = atalk_interfaces; pos && i; i = i->next)
net/appletalk/atalk_proc.c
48
i = i->next;
net/appletalk/atalk_proc.c
82
for (r = atalk_routes; pos && r; r = r->next)
net/appletalk/ddp.c
223
*iface = tmp->next;
net/appletalk/ddp.c
228
iface = &tmp->next;
net/appletalk/ddp.c
248
iface->next = atalk_interfaces;
net/appletalk/ddp.c
365
for (iface = atalk_interfaces; iface; iface = iface->next) {
net/appletalk/ddp.c
413
for (iface = atalk_interfaces; iface; iface = iface->next) {
net/appletalk/ddp.c
448
for (r = atalk_routes; r; r = r->next) {
net/appletalk/ddp.c
529
for (rt = atalk_routes; rt; rt = rt->next) {
net/appletalk/ddp.c
545
for (iface = atalk_interfaces; iface; iface = iface->next) {
net/appletalk/ddp.c
573
rt->next = atalk_routes;
net/appletalk/ddp.c
604
*r = tmp->next;
net/appletalk/ddp.c
609
r = &tmp->next;
net/appletalk/ddp.c
629
*r = tmp->next;
net/appletalk/ddp.c
633
r = &tmp->next;
net/atm/br2684.c
288
return list_empty(&brdev->brvccs) ? NULL : list_entry_brvcc(brdev->brvccs.next); /* 1 vcc/dev right now */
net/atm/br2684.c
376
brdev->brvccs.next != brdev->brvccs.prev) /* >1 VCC */
net/atm/br2684.c
379
brvcc = list_entry_brvcc(brdev->brvccs.next);
net/atm/br2684.c
818
.next = br2684_seq_next,
net/atm/br2684.c
854
net_dev = list_entry_brdev(br2684_devs.next);
net/atm/br2684.c
857
brvcc = list_entry_brvcc(brdev->brvccs.next);
net/atm/clip.c
109
for (walk = &entry->vccs; *walk; walk = &(*walk)->next)
net/atm/clip.c
113
*walk = clip_vcc->next; /* atomic */
net/atm/clip.c
140
for (cv = entry->vccs; cv; cv = cv->next) {
net/atm/clip.c
530
for (dev = clip_devs; dev; dev = PRIV(dev)->next)
net/atm/clip.c
535
for (dev = clip_devs; dev; dev = PRIV(dev)->next)
net/atm/clip.c
552
clip_priv->next = clip_devs;
net/atm/clip.c
832
curr = curr->next;
net/atm/clip.c
893
.next = neigh_seq_next,
net/atm/clip.c
928
struct net_device *dev, *next;
net/atm/clip.c
93
clip_vcc->next = entry->vccs;
net/atm/clip.c
942
next = PRIV(dev)->next;
net/atm/clip.c
945
dev = next;
net/atm/common.c
502
dev = list_entry(atm_devs.next,
net/atm/lec.c
1002
next);
net/atm/lec.c
1012
.next = lec_seq_next,
net/atm/lec.c
1321
hlist_add_head(&entry->next, tmp);
net/atm/lec.c
1338
hlist_del(&to_remove->next);
net/atm/lec.c
1351
&priv->lec_arp_tables[i], next) {
net/atm/lec.c
1395
&priv->lec_arp_tables[i], next) {
net/atm/lec.c
1422
hlist_for_each_entry(rulla, &priv->lec_no_forward, next) {
net/atm/lec.c
1444
hlist_for_each_entry(rulla, &priv->lec_arp_empty_ones, next) {
net/atm/lec.c
1466
hlist_for_each_entry(rulla, &priv->mcast_fwds, next) {
net/atm/lec.c
1497
struct hlist_node *next;
net/atm/lec.c
1509
hlist_for_each_entry_safe(entry, next,
net/atm/lec.c
1510
&priv->lec_arp_tables[i], next) {
net/atm/lec.c
1517
hlist_for_each_entry_safe(entry, next,
net/atm/lec.c
1518
&priv->lec_arp_empty_ones, next) {
net/atm/lec.c
1521
hlist_del(&entry->next);
net/atm/lec.c
1526
hlist_for_each_entry_safe(entry, next,
net/atm/lec.c
1527
&priv->lec_no_forward, next) {
net/atm/lec.c
1530
hlist_del(&entry->next);
net/atm/lec.c
1535
hlist_for_each_entry_safe(entry, next, &priv->mcast_fwds, next) {
net/atm/lec.c
1538
hlist_del(&entry->next);
net/atm/lec.c
1558
hlist_for_each_entry(entry, head, next) {
net/atm/lec.c
1574
INIT_HLIST_NODE(&to_return->next);
net/atm/lec.c
1621
hlist_del(&to_remove->next);
net/atm/lec.c
1693
struct hlist_node *next;
net/atm/lec.c
1703
hlist_for_each_entry_safe(entry, next,
net/atm/lec.c
1704
&priv->lec_arp_tables[i], next) {
net/atm/lec.c
1830
struct hlist_node *next;
net/atm/lec.c
1837
hlist_for_each_entry_safe(entry, next,
net/atm/lec.c
1838
&priv->lec_arp_tables[i], next) {
net/atm/lec.c
1862
struct hlist_node *next;
net/atm/lec.c
1877
hlist_for_each_entry_safe(entry, next,
net/atm/lec.c
1878
&priv->lec_arp_empty_ones, next) {
net/atm/lec.c
1880
hlist_del(&entry->next);
net/atm/lec.c
1924
&priv->lec_arp_tables[i], next) {
net/atm/lec.c
1988
hlist_add_head(&entry->next, &priv->mcast_fwds);
net/atm/lec.c
2007
hlist_add_head(&entry->next, &priv->lec_no_forward);
net/atm/lec.c
2016
&priv->lec_arp_tables[i], next) {
net/atm/lec.c
2079
hlist_add_head(&entry->next, &priv->lec_arp_empty_ones);
net/atm/lec.c
2100
&priv->lec_arp_tables[i], next) {
net/atm/lec.c
2134
&priv->lec_arp_tables[i], next) {
net/atm/lec.c
2185
struct hlist_node *next;
net/atm/lec.c
2195
hlist_for_each_entry_safe(entry, next,
net/atm/lec.c
2196
&priv->lec_arp_tables[i], next) {
net/atm/lec.c
2206
hlist_for_each_entry_safe(entry, next,
net/atm/lec.c
2207
&priv->lec_arp_empty_ones, next) {
net/atm/lec.c
2211
hlist_del(&entry->next);
net/atm/lec.c
2216
hlist_for_each_entry_safe(entry, next,
net/atm/lec.c
2217
&priv->lec_no_forward, next) {
net/atm/lec.c
2221
hlist_del(&entry->next);
net/atm/lec.c
2226
hlist_for_each_entry_safe(entry, next, &priv->mcast_fwds, next) {
net/atm/lec.c
2230
hlist_del(&entry->next);
net/atm/lec.c
2244
struct hlist_node *next;
net/atm/lec.c
2250
hlist_for_each_entry_safe(entry, next,
net/atm/lec.c
2251
&priv->lec_arp_empty_ones, next) {
net/atm/lec.c
2263
hlist_del(&entry->next);
net/atm/lec.c
865
for (; e; e = e->next) {
net/atm/lec_arpc.h
15
struct hlist_node next; /* Linked entry list */
net/atm/mpc.c
1300
entry = entry->next;
net/atm/mpc.c
136
mpc = mpc->next;
net/atm/mpc.c
1405
entry = entry->next;
net/atm/mpc.c
1441
mpc = mpc->next;
net/atm/mpc.c
150
mpc = mpc->next;
net/atm/mpc.c
1505
tmp = mpc->next;
net/atm/mpc.c
1527
nextqos = qos->next;
net/atm/mpc.c
164
mpc = mpc->next;
net/atm/mpc.c
196
entry->next = qos_head;
net/atm/mpc.c
210
qos = qos->next;
net/atm/mpc.c
226
qos_head = qos_head->next;
net/atm/mpc.c
233
if (curr->next == entry) {
net/atm/mpc.c
234
curr->next = entry->next;
net/atm/mpc.c
238
curr = curr->next;
net/atm/mpc.c
266
qos = qos->next;
net/atm/mpc.c
290
mpc->next = mpcs;
net/atm/mpc.h
16
struct mpoa_client *next;
net/atm/mpc.h
42
struct atm_mpoa_qos *next;
net/atm/mpoa_caches.c
112
entry->next = client->in_cache;
net/atm/mpoa_caches.c
201
entry->prev->next = entry->next;
net/atm/mpoa_caches.c
203
client->in_cache = entry->next;
net/atm/mpoa_caches.c
204
if (entry->next != NULL)
net/atm/mpoa_caches.c
205
entry->next->prev = entry->prev;
net/atm/mpoa_caches.c
237
next_entry = entry->next;
net/atm/mpoa_caches.c
266
entry = entry->next; /* Entry in hold down */
net/atm/mpoa_caches.c
278
entry = entry->next;
net/atm/mpoa_caches.c
293
entry = entry->next;
net/atm/mpoa_caches.c
319
entry = entry->next;
net/atm/mpoa_caches.c
345
entry = entry->next;
net/atm/mpoa_caches.c
366
entry = entry->next;
net/atm/mpoa_caches.c
388
entry = entry->next;
net/atm/mpoa_caches.c
408
entry = entry->next;
net/atm/mpoa_caches.c
434
entry->prev->next = entry->next;
net/atm/mpoa_caches.c
436
client->eg_cache = entry->next;
net/atm/mpoa_caches.c
437
if (entry->next != NULL)
net/atm/mpoa_caches.c
438
entry->next->prev = entry->prev;
net/atm/mpoa_caches.c
472
entry->next = client->eg_cache;
net/atm/mpoa_caches.c
48
entry = entry->next;
net/atm/mpoa_caches.c
511
next_entry = entry->next;
net/atm/mpoa_caches.c
69
entry = entry->next;
net/atm/mpoa_caches.c
90
entry = entry->next;
net/atm/mpoa_caches.h
18
struct in_cache_entry *next;
net/atm/mpoa_caches.h
55
struct eg_cache_entry *next;
net/atm/mpoa_proc.c
111
for (mpc = mpcs; mpc; mpc = mpc->next)
net/atm/mpoa_proc.c
121
return v == SEQ_START_TOKEN ? mpcs : p->next;
net/atm/mpoa_proc.c
149
for (in_entry = mpc->in_cache; in_entry; in_entry = in_entry->next) {
net/atm/mpoa_proc.c
168
for (eg_entry = mpc->eg_cache; eg_entry; eg_entry = eg_entry->next) {
net/atm/mpoa_proc.c
196
.next = mpc_next,
net/atm/proc.c
242
.next = atm_dev_seq_next,
net/atm/proc.c
265
.next = vcc_seq_next,
net/atm/proc.c
287
.next = vcc_seq_next,
net/atm/proc.c
310
.next = vcc_seq_next,
net/ax25/af_ax25.c
2001
.next = ax25_info_next,
net/ax25/ax25_iface.c
115
listen->next = listen_list;
net/ax25/ax25_iface.c
136
listen_list = listen->next;
net/ax25/ax25_iface.c
142
while (listen != NULL && listen->next != NULL) {
net/ax25/ax25_iface.c
143
if (ax25cmp(&listen->next->callsign, callsign) == 0 && listen->next->dev == dev) {
net/ax25/ax25_iface.c
144
s = listen->next;
net/ax25/ax25_iface.c
145
listen->next = listen->next->next;
net/ax25/ax25_iface.c
151
listen = listen->next;
net/ax25/ax25_iface.c
164
for (protocol = protocol_list; protocol != NULL; protocol = protocol->next)
net/ax25/ax25_iface.c
179
for (listen = listen_list; listen != NULL; listen = listen->next)
net/ax25/ax25_iface.c
206
for (protocol = protocol_list; protocol != NULL; protocol = protocol->next)
net/ax25/ax25_iface.c
35
struct listen_struct *next;
net/ax25/ax25_iface.c
48
ap->next = protocol_list;
net/ax25/ax25_iface.c
65
protocol_list = protocol->next;
net/ax25/ax25_iface.c
69
while (protocol != NULL && protocol->next != NULL) {
net/ax25/ax25_iface.c
70
if (protocol->next->pid == pid) {
net/ax25/ax25_iface.c
71
protocol->next = protocol->next->next;
net/ax25/ax25_iface.c
75
protocol = protocol->next;
net/ax25/ax25_out.c
255
unsigned short start, end, next;
net/ax25/ax25_out.c
297
next = (ax25->vs + 1) % ax25->modulus;
net/ax25/ax25_out.c
298
last = (next == end);
net/ax25/ax25_out.c
318
ax25->vs = next;
net/ax25/ax25_route.c
110
ax25_rt = ax25_rt->next;
net/ax25/ax25_route.c
137
ax25_rt->next = ax25_route_list;
net/ax25/ax25_route.c
164
ax25_rt = ax25_rt->next;
net/ax25/ax25_route.c
168
ax25_route_list = s->next;
net/ax25/ax25_route.c
171
for (t = ax25_route_list; t != NULL; t = t->next) {
net/ax25/ax25_route.c
172
if (t->next == s) {
net/ax25/ax25_route.c
173
t->next = s->next;
net/ax25/ax25_route.c
220
ax25_rt = ax25_rt->next;
net/ax25/ax25_route.c
267
for (ax25_rt = ax25_route_list; ax25_rt != NULL; ax25_rt = ax25_rt->next) {
net/ax25/ax25_route.c
280
((struct ax25_route *) v)->next;
net/ax25/ax25_route.c
333
.next = ax25_rt_seq_next,
net/ax25/ax25_route.c
355
for (ax25_rt = ax25_route_list; ax25_rt != NULL; ax25_rt = ax25_rt->next) {
net/ax25/ax25_route.c
410
ax25_rt = ax25_rt->next;
net/ax25/ax25_route.c
50
ax25_rt = ax25_rt->next;
net/ax25/ax25_route.c
54
ax25_route_list = s->next;
net/ax25/ax25_route.c
58
for (t = ax25_route_list; t != NULL; t = t->next) {
net/ax25/ax25_route.c
59
if (t->next == s) {
net/ax25/ax25_route.c
60
t->next = s->next;
net/ax25/ax25_uid.c
183
.next = ax25_uid_seq_next,
net/batman-adv/bat_iv_ogm.c
2266
goto next;
net/batman-adv/bat_iv_ogm.c
2269
goto next;
net/batman-adv/bat_iv_ogm.c
2312
next:
net/batman-adv/bat_v.c
570
goto next;
net/batman-adv/bat_v.c
573
goto next;
net/batman-adv/bat_v.c
581
next:
net/batman-adv/mesh-interface.c
1092
hard_iface = netdev_adjacent_get_private(mesh_iface->adj_list.lower.next);
net/batman-adv/multicast.c
456
pmc6 = rcu_dereference(pmc6->next)) {
net/batman-adv/routing.c
569
goto next;
net/batman-adv/routing.c
573
goto next;
net/batman-adv/routing.c
582
goto next;
net/batman-adv/routing.c
586
goto next;
net/batman-adv/routing.c
608
next:
net/bluetooth/af_bluetooth.c
841
.next = bt_seq_next,
net/bluetooth/hci_core.c
3259
skb = list; list = list->next;
net/bluetooth/hci_core.c
3355
skb = list; list = list->next;
net/bluetooth/hci_sync.c
2101
struct adv_info *next = NULL;
net/bluetooth/hci_sync.c
2113
next = hci_get_next_instance(hdev, instance);
net/bluetooth/hci_sync.c
2124
if (next && next->instance == instance)
net/bluetooth/hci_sync.c
2125
next = NULL;
net/bluetooth/hci_sync.c
2136
if (next && !ext_adv_capable(hdev))
net/bluetooth/hci_sync.c
2137
hci_schedule_adv_instance_sync(hdev, next->instance, false);
net/bluetooth/iso.c
1533
frag = &(*frag)->next;
net/bluetooth/l2cap_core.c
2248
frag = &(*frag)->next;
net/bluetooth/l2cap_core.c
3043
frag = &(*frag)->next;
net/bluetooth/l2cap_core.c
5735
new_frag->next = NULL;
net/bluetooth/l2cap_core.c
5737
(*last_frag)->next = new_frag;
net/bluetooth/l2cap_core.c
7334
c = list_entry(chan_list.next, typeof(*c), global_l);
net/bluetooth/l2cap_core.c
7390
struct l2cap_chan *chan, *next;
net/bluetooth/l2cap_core.c
7394
goto next;
net/bluetooth/l2cap_core.c
7408
next:
net/bluetooth/l2cap_core.c
7409
next = l2cap_global_fixed_chan(pchan, hcon);
net/bluetooth/l2cap_core.c
7411
pchan = next;
net/bluetooth/l2cap_sock.c
1812
struct l2cap_rx_busy *rx_busy, *next;
net/bluetooth/l2cap_sock.c
1821
list_for_each_entry_safe(rx_busy, next, &l2cap_pi(sk)->rx_busy, list) {
net/bluetooth/rfcomm/core.c
574
struct sk_buff *frag, *next;
net/bluetooth/rfcomm/core.c
590
for (; frag; frag = next) {
net/bluetooth/rfcomm/core.c
593
next = frag->next;
net/bpf/test_run.c
608
struct prog_test_ref_kfunc *next;
net/bridge/br_forward.c
345
p = rcu_dereference(p->next);
net/bridge/br_mdb.c
1385
pp = &p->next) {
net/bridge/br_mdb.c
1509
pp = &p->next;
net/bridge/br_mdb.c
1514
pp = &p->next;
net/bridge/br_mdb.c
1521
pp = &p->next;
net/bridge/br_mdb.c
1616
pg = mlock_dereference(pg->next, mp->br))
net/bridge/br_mdb.c
1658
pg = mlock_dereference(pg->next, mp->br)) {
net/bridge/br_mdb.c
354
pp = &p->next) {
net/bridge/br_mdb.c
772
pp = &p->next) {
net/bridge/br_mdb.c
997
pp = &p->next) {
net/bridge/br_multicast.c
1403
struct net_bridge_port_group __rcu *next,
net/bridge/br_multicast.c
1440
rcu_assign_pointer(p->next, next);
net/bridge/br_multicast.c
1529
pp = &p->next) {
net/bridge/br_multicast.c
2836
pg = mlock_dereference(pg->next, br))
net/bridge/br_multicast.c
3550
pp = &p->next) {
net/bridge/br_multicast.c
3639
pp = &p->next) {
net/bridge/br_multicast.c
365
pg_lst = mlock_dereference(pg_lst->next, br)) {
net/bridge/br_multicast.c
3680
pp = &p->next) {
net/bridge/br_multicast.c
3707
p = mlock_dereference(p->next, brmctx->br)) {
net/bridge/br_multicast.c
3739
p = mlock_dereference(p->next, brmctx->br)) {
net/bridge/br_multicast.c
418
pg = mlock_dereference(pg->next, br)) {
net/bridge/br_multicast.c
447
pp = &p->next)
net/bridge/br_multicast.c
463
pp = &p->next;
net/bridge/br_multicast.c
487
pg = mlock_dereference(pg->next, br)) {
net/bridge/br_multicast.c
572
pp = &p->next) {
net/bridge/br_multicast.c
810
rcu_assign_pointer(*pp, pg->next);
net/bridge/br_multicast.c
844
pp = &p->next) {
net/bridge/br_private.h
1005
struct net_bridge_port_group __rcu *next,
net/bridge/br_private.h
351
struct net_bridge_port_group __rcu *next;
net/bridge/br_switchdev.c
520
pp = &p->next) {
net/bridge/br_switchdev.c
736
pp = &p->next) {
net/can/bcm.c
1593
struct bcm_op *op, *next;
net/can/bcm.c
1620
list_for_each_entry_safe(op, next, &bo->tx_ops, list)
net/can/bcm.c
1623
list_for_each_entry_safe(op, next, &bo->rx_ops, list) {
net/can/bcm.c
1652
list_for_each_entry_safe(op, next, &bo->rx_ops, list)
net/core/bpf_sk_storage.c
889
.next = bpf_sk_storage_map_seq_next,
net/core/datagram.c
154
skb->prev->next = nskb;
net/core/datagram.c
155
skb->next->prev = nskb;
net/core/datagram.c
157
nskb->next = skb->next;
net/core/datagram.c
338
if (skb->next) {
net/core/datagram.c
800
__wsum next, *csum = priv2;
net/core/datagram.c
802
next = csum_and_copy_to_user(from + progress, iter_to, len);
net/core/datagram.c
803
*csum = csum_block_add(*csum, next, progress);
net/core/datagram.c
804
return next ? 0 : len;
net/core/datagram.c
812
__wsum next = csum_partial_copy_nocheck(from + progress, iter_to, len);
net/core/datagram.c
814
*csum = csum_block_add(*csum, next, progress);
net/core/dev.c
12715
list_skb = &(*list_skb)->next;
net/core/dev.c
3457
skb->next = __this_cpu_read(softnet_data.completion_queue);
net/core/dev.c
3901
struct sk_buff *next = skb->next;
net/core/dev.c
3904
rc = xmit_one(skb, dev, txq, next != NULL);
net/core/dev.c
3906
skb->next = next;
net/core/dev.c
3910
skb = next;
net/core/dev.c
4080
struct sk_buff *next, *head = NULL, *tail;
net/core/dev.c
4082
for (; skb != NULL; skb = next) {
net/core/dev.c
4083
next = skb->next;
net/core/dev.c
4096
tail->next = skb;
net/core/dev.c
4178
struct sk_buff *next, *to_free = NULL, *to_free2 = NULL;
net/core/dev.c
4232
skb->ll_node.next = first_n;
net/core/dev.c
4253
llist_for_each_entry_safe(skb, next, ll_list, ll_node)
net/core/dev.c
4277
llist_for_each_entry_safe(skb, next, ll_list, ll_node) {
net/core/dev.c
4278
if (next) {
net/core/dev.c
4279
prefetch(next);
net/core/dev.c
4280
prefetch(&next->priority);
net/core/dev.c
4851
is_list = !!skb->next;
net/core/dev.c
5767
clist = clist->next;
net/core/dev.c
6217
struct sk_buff *skb, *next;
net/core/dev.c
6227
list_for_each_entry_safe(skb, next, head, list) {
net/core/dev.c
6248
struct sk_buff *skb, *next;
net/core/dev.c
6251
list_for_each_entry_safe(skb, next, head, list) {
net/core/dev.c
6302
struct sk_buff *skb, *next;
net/core/dev.c
6305
list_for_each_entry_safe(skb, next, head, list) {
net/core/dev.c
6387
struct sk_buff *skb, *next;
net/core/dev.c
6390
list_for_each_entry_safe(skb, next, head, list) {
net/core/dev.c
6402
list_for_each_entry_safe(skb, next, head, list) {
net/core/dev.c
6584
struct softnet_data *next = remsd->rps_ipi_next;
net/core/dev.c
6588
remsd = next;
net/core/dev.c
6823
struct sk_buff *skb, *next;
net/core/dev.c
6835
llist_for_each_entry_safe(skb, next, free_list, ll_node) {
net/core/dev.c
6836
prefetch(next);
net/core/dev.c
8137
upper = list_entry_rcu((*iter)->next, struct netdev_adjacent, list);
net/core/dev.c
8154
upper = list_entry((*iter)->next, struct netdev_adjacent, list);
net/core/dev.c
8172
upper = list_entry_rcu((*iter)->next, struct netdev_adjacent, list);
net/core/dev.c
8187
struct net_device *udev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
net/core/dev.c
8202
next = NULL;
net/core/dev.c
8210
next = udev;
net/core/dev.c
8217
if (!next) {
net/core/dev.c
8220
next = dev_stack[--cur];
net/core/dev.c
8224
now = next;
net/core/dev.c
8236
struct net_device *udev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
net/core/dev.c
8250
next = NULL;
net/core/dev.c
8256
next = udev;
net/core/dev.c
8263
if (!next) {
net/core/dev.c
8266
next = dev_stack[--cur];
net/core/dev.c
8270
now = next;
net/core/dev.c
8313
*iter = lower->list.next;
net/core/dev.c
8336
lower = list_entry_rcu((*iter)->next, struct netdev_adjacent, list);
net/core/dev.c
8367
*iter = lower->list.next;
net/core/dev.c
8378
lower = list_entry((*iter)->next, struct netdev_adjacent, list);
net/core/dev.c
8394
lower = list_entry((*iter)->next, struct netdev_adjacent, list);
net/core/dev.c
8410
struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
net/core/dev.c
8424
next = NULL;
net/core/dev.c
8430
next = ldev;
net/core/dev.c
8437
if (!next) {
net/core/dev.c
8440
next = dev_stack[--cur];
net/core/dev.c
8444
now = next;
net/core/dev.c
8457
struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
net/core/dev.c
8472
next = NULL;
net/core/dev.c
8480
next = ldev;
net/core/dev.c
8487
if (!next) {
net/core/dev.c
8490
next = dev_stack[--cur];
net/core/dev.c
8494
now = next;
net/core/dev.c
8506
lower = list_entry_rcu((*iter)->next, struct netdev_adjacent, list);
net/core/dev.c
8595
struct net_device *ldev, *next, *now, *dev_stack[MAX_NEST_DEV + 1];
net/core/dev.c
8609
next = NULL;
net/core/dev.c
8615
next = ldev;
net/core/dev.c
8622
if (!next) {
net/core/dev.c
8625
next = dev_stack[--cur];
net/core/dev.c
8629
now = next;
net/core/dev.h
114
#define netdev_for_each_altname_safe(dev, namenode, next) \
net/core/dev.h
115
list_for_each_entry_safe((namenode), (next), &(dev)->name_node->list, \
net/core/dev.h
359
if (!napi->poll_list.next)
net/core/devmem.c
137
if (binding->list.next)
net/core/fib_rules.c
89
pos = ops->rules_list.next;
net/core/fib_rules.c
90
if (pos->next != &ops->rules_list) {
net/core/fib_rules.c
91
rule = list_entry(pos->next, struct fib_rule, list);
net/core/flow_offload.c
464
struct flow_block_cb *this, *next;
net/core/flow_offload.c
466
list_for_each_entry_safe(this, next, &flow_block_indr_list, indr.list) {
net/core/flow_offload.c
475
struct flow_block_cb *this, *next;
net/core/flow_offload.c
477
list_for_each_entry_safe(this, next, cleanup_list, indr.list) {
net/core/flow_offload.c
486
struct flow_indr_dev *this, *next, *indr_dev = NULL;
net/core/flow_offload.c
490
list_for_each_entry_safe(this, next, &flow_block_indr_dev_list, list) {
net/core/gro.c
206
NAPI_GRO_CB(p)->last->next = skb;
net/core/gro.c
233
NAPI_GRO_CB(p)->last->next = skb;
net/core/neighbour.c
1106
unsigned long now, next;
net/core/neighbour.c
1116
next = now + HZ;
net/core/neighbour.c
1125
next = neigh->confirmed + neigh->parms->reachable_time;
net/core/neighbour.c
1133
next = now + NEIGH_VAR(neigh->parms, DELAY_PROBE_TIME);
net/core/neighbour.c
1150
next = neigh->confirmed + neigh->parms->reachable_time;
net/core/neighbour.c
1157
next = now + max(NEIGH_VAR(neigh->parms, RETRANS_TIME),
net/core/neighbour.c
1162
next = now + max(NEIGH_VAR(neigh->parms, RETRANS_TIME), HZ/100);
net/core/neighbour.c
1186
if (time_before(next, jiffies + HZ/100))
net/core/neighbour.c
1187
next = jiffies + HZ/100;
net/core/neighbour.c
1188
if (!mod_timer(&neigh->timer, next))
net/core/neighbour.c
1223
unsigned long next, now = jiffies;
net/core/neighbour.c
1231
next = now + 1;
net/core/neighbour.c
1234
next = now + max(NEIGH_VAR(neigh->parms,
net/core/neighbour.c
1238
neigh_add_timer(neigh, next);
net/core/neighbour.c
2628
goto next;
net/core/neighbour.c
2636
next:
net/core/neighbour.c
2822
goto next;
net/core/neighbour.c
2825
goto next;
net/core/neighbour.c
2831
next:
net/core/neighbour.c
2859
n = rcu_dereference(n->next)) {
net/core/neighbour.c
2861
goto next;
net/core/neighbour.c
2864
goto next;
net/core/neighbour.c
2870
next:
net/core/neighbour.c
3358
pn = rcu_dereference(pn->next);
net/core/neighbour.c
3376
pn = rcu_dereference(pn->next);
net/core/neighbour.c
3386
pn = rcu_dereference(pn->next);
net/core/neighbour.c
3559
.next = neigh_stat_seq_next,
net/core/neighbour.c
776
n = rcu_dereference_check(n->next, lockdep_is_held(&tbl->phash_lock));
net/core/neighbour.c
818
n->next = tbl->phash_buckets[hash_val];
net/core/neighbour.c
852
np = &n->next) {
net/core/neighbour.c
855
rcu_assign_pointer(*np, n->next);
net/core/neighbour.c
886
rcu_assign_pointer(*np, n->next);
net/core/neighbour.c
891
np = &n->next;
net/core/net-procfs.c
161
.next = dev_seq_next,
net/core/net-procfs.c
168
.next = softnet_seq_next,
net/core/net-procfs.c
243
nxt = READ_ONCE(pt->list.next);
net/core/net-procfs.c
250
nxt = READ_ONCE(dev->ptype_all.next);
net/core/net-procfs.c
257
nxt = READ_ONCE(net->ptype_all.next);
net/core/net-procfs.c
268
nxt = READ_ONCE(net->ptype_specific.next);
net/core/net-procfs.c
274
nxt = READ_ONCE(ptype_base[0].next);
net/core/net-procfs.c
281
nxt = READ_ONCE(ptype_base[hash].next);
net/core/net-procfs.c
320
.next = ptype_seq_next,
net/core/net-procfs.c
388
.next = dev_seq_next,
net/core/net_namespace.c
1505
first_device = first_device->next;
net/core/net_namespace.c
520
struct net *net, *next;
net/core/net_namespace.c
525
llist_for_each_entry_safe(net, next, kill_list, defer_free_list)
net/core/net_test.c
150
struct sk_buff *skb, *segs, *cur, *next, *last;
net/core/net_test.c
202
prev->next = frag_skb;
net/core/net_test.c
239
for (cur = segs, i = 0; cur; cur = next, i++) {
net/core/net_test.c
240
next = cur->next;
net/core/net_test.c
252
if (!next)
net/core/netpoll.c
258
clist = clist->next;
net/core/rtnetlink.c
146
struct sk_buff *next = head->next;
net/core/rtnetlink.c
150
head = next;
net/core/rtnetlink.c
98
tail->next = defer_kfree_skb_list;
net/core/skbuff.c
1291
struct sk_buff *next = segs->next;
net/core/skbuff.c
1298
segs = next;
net/core/skbuff.c
1612
n->next = n->prev = NULL;
net/core/skbuff.c
1770
u32 bytelen, next;
net/core/skbuff.c
1793
next = (u32)atomic_read(&sk->sk_zckey);
net/core/skbuff.c
1794
if ((u32)(uarg_zc->id + uarg_zc->len) == next) {
net/core/skbuff.c
1800
atomic_set(&sk->sk_zckey, ++next);
net/core/skbuff.c
2027
struct page *next = (struct page *)page_private(head);
net/core/skbuff.c
2029
head = next;
net/core/skbuff.c
2774
fragp = &frag->next) {
net/core/skbuff.c
2784
nfrag->next = frag->next;
net/core/skbuff.c
2799
if (frag->next)
net/core/skbuff.c
2800
skb_drop_list(&frag->next);
net/core/skbuff.c
2918
list = list->next;
net/core/skbuff.c
2931
insp = list->next;
net/core/skbuff.c
2948
skb_shinfo(skb)->frag_list = list->next;
net/core/skbuff.c
2953
clone->next = list;
net/core/skbuff.c
3393
} else if (skb->next) {
net/core/skbuff.c
3394
skb = skb->next;
net/core/skbuff.c
4050
struct sk_buff *skb, *next;
net/core/skbuff.c
4057
skb_queue_walk_safe(list, skb, next) {
net/core/skbuff.c
4506
} else if (st->cur_skb->next) {
net/core/skbuff.c
4507
st->cur_skb = st->cur_skb->next;
net/core/skbuff.c
4688
list_skb = list_skb->next;
net/core/skbuff.c
4705
skb->next = nskb;
net/core/skbuff.c
4707
tail->next = nskb;
net/core/skbuff.c
4710
nskb->next = list_skb;
net/core/skbuff.c
4751
kfree_skb_list(skb->next);
net/core/skbuff.c
4752
skb->next = NULL;
net/core/skbuff.c
4791
for (check_skb = list_skb; check_skb; check_skb = check_skb->next) {
net/core/skbuff.c
4835
if (frag_len != iter->len && iter->next)
net/core/skbuff.c
4912
list_skb = list_skb->next;
net/core/skbuff.c
4946
tail->next = nskb;
net/core/skbuff.c
5013
list_skb = list_skb->next;
net/core/skbuff.c
5087
for (iter = segs; iter; iter = iter->next) {
net/core/skbuff.c
5404
if (skb1->next == NULL && tailbits) {
net/core/skbuff.c
5435
skb2->next = skb1->next;
net/core/skbuff.c
5442
skb_p = &skb1->next;
net/core/skbuff.c
6876
list = list->next;
net/core/skbuff.c
6884
insp = list->next;
net/core/skbuff.c
6900
shinfo->frag_list = list->next;
net/core/skbuff.c
6905
clone->next = list;
net/core/skbuff.c
7411
__wsum next = csum_partial_copy_nocheck(iter_from, to + progress, len);
net/core/skbuff.c
7413
*csum = csum_block_add(*csum, next, progress);
net/core/skbuff.c
7421
__wsum next, *csum = priv2;
net/core/skbuff.c
7423
next = csum_and_copy_from_user(iter_from, to + progress, len);
net/core/skbuff.c
7424
*csum = csum_block_add(*csum, next, progress);
net/core/skbuff.c
7425
return next ? 0 : len;
net/core/sock.c
3200
struct sk_buff *skb, *next;
net/core/sock.c
3209
next = skb->next;
net/core/sock.c
3210
prefetch(next);
net/core/sock.c
3215
skb = next;
net/core/sock.c
4370
.next = proto_seq_next,
net/core/sock_map.c
1404
.next = sock_hash_seq_next,
net/core/sock_map.c
455
static int sock_map_get_next_key(struct bpf_map *map, void *key, void *next)
net/core/sock_map.c
459
u32 *key_next = next;
net/core/sock_map.c
791
.next = sock_map_seq_next,
net/devlink/core.c
111
static u32 next;
net/devlink/core.c
119
xa_limit_32b, &next, GFP_KERNEL);
net/devlink/core.c
335
goto next;
net/devlink/core.c
338
goto next;
net/devlink/core.c
344
next:
net/dsa/dsa.c
1488
struct dsa_port *dp, *next;
net/dsa/dsa.c
1491
dsa_switch_for_each_port_safe(dp, next, ds) {
net/dsa/dsa.c
862
struct dsa_link *dl, *next;
net/dsa/dsa.c
864
list_for_each_entry_safe(dl, next, &dst->rtable, list) {
net/handshake/handshake-test.c
348
struct handshake_req *req, *next;
net/handshake/handshake-test.c
376
next = handshake_req_next(hn, HANDSHAKE_HANDLER_CLASS_TLSHD);
net/handshake/handshake-test.c
377
KUNIT_ASSERT_PTR_EQ(test, req, next);
net/handshake/handshake-test.c
390
struct handshake_req *req, *next;
net/handshake/handshake-test.c
418
next = handshake_req_next(hn, HANDSHAKE_HANDLER_CLASS_TLSHD);
net/handshake/handshake-test.c
419
KUNIT_ASSERT_PTR_EQ(test, req, next);
net/handshake/handshake-test.c
422
handshake_complete(next, -ETIMEDOUT, NULL);
net/ipv4/af_inet.c
1432
if (skb->next)
net/ipv4/af_inet.c
1459
} while ((skb = skb->next));
net/ipv4/arp.c
1491
.next = neigh_seq_next,
net/ipv4/cipso_ipv4.c
266
__list_del(entry->list.prev, entry->list.next);
net/ipv4/devinet.c
711
unsigned long now, next, next_sec, next_sched;
net/ipv4/devinet.c
719
next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY);
net/ipv4/devinet.c
750
if (time_before(tstamp + valid_lft * HZ, next))
net/ipv4/devinet.c
751
next = tstamp + valid_lft * HZ;
net/ipv4/devinet.c
756
next)) {
net/ipv4/devinet.c
757
next = tstamp + preferred_lft * HZ;
net/ipv4/devinet.c
802
next_sec = round_jiffies_up(next);
net/ipv4/devinet.c
803
next_sched = next;
net/ipv4/devinet.c
806
if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ))
net/ipv4/fib_frontend.c
1071
goto next;
net/ipv4/fib_frontend.c
1079
next:
net/ipv4/fib_semantics.c
168
struct fib_nh_exception *next;
net/ipv4/fib_semantics.c
170
next = rcu_dereference_protected(fnhe->fnhe_next, 1);
net/ipv4/fib_semantics.c
177
fnhe = next;
net/ipv4/fib_trie.c
1631
struct fib_alias *fa = hlist_entry(pprev, typeof(*fa), fa_list.next);
net/ipv4/fib_trie.c
1674
fa_next = hlist_entry_safe(fa_to_delete->fa_list.next,
net/ipv4/fib_trie.c
2264
goto next;
net/ipv4/fib_trie.c
2269
goto next;
net/ipv4/fib_trie.c
2273
goto next;
net/ipv4/fib_trie.c
2277
goto next;
net/ipv4/fib_trie.c
2281
goto next;
net/ipv4/fib_trie.c
2315
next:
net/ipv4/fib_trie.c
2816
.next = fib_trie_seq_next,
net/ipv4/fib_trie.c
3005
.next = fib_route_seq_next,
net/ipv4/fib_trie.c
474
tn_info(tn)->rcu.next = NULL;
net/ipv4/fib_trie.c
480
tn_info(n)->rcu.next = tn_info(tn)->rcu.next;
net/ipv4/fib_trie.c
481
tn_info(tn)->rcu.next = &tn_info(n)->rcu;
net/ipv4/fib_trie.c
489
head = head->next;
net/ipv4/fou_core.c
1229
struct fou *fou, *next;
net/ipv4/fou_core.c
1233
list_for_each_entry_safe(fou, next, &fn->fou_list, list)
net/ipv4/gre_offload.c
117
} while ((skb = skb->next));
net/ipv4/igmp.c
1212
pmc->next = in_dev->mc_tomb;
net/ipv4/igmp.c
1229
for (pmc = in_dev->mc_tomb; pmc; pmc = pmc->next) {
net/ipv4/igmp.c
1236
pmc_prev->next = pmc->next;
net/ipv4/igmp.c
1238
in_dev->mc_tomb = pmc->next;
net/ipv4/igmp.c
1274
nextpmc = pmc->next;
net/ipv4/igmp.c
190
struct ip_sf_list *next;
net/ipv4/igmp.c
193
next = psf->sf_next;
net/ipv4/igmp.c
195
psf = next;
net/ipv4/igmp.c
2976
.next = igmp_mc_seq_next,
net/ipv4/igmp.c
3026
state->im = state->im->next;
net/ipv4/igmp.c
3107
.next = igmp_mcf_seq_next,
net/ipv4/igmp.c
668
pmc_next = pmc->next;
net/ipv4/igmp.c
688
pmc_prev->next = pmc_next;
net/ipv4/inet_connection_sock.c
1460
struct request_sock *next, *req;
net/ipv4/inet_connection_sock.c
1519
next = req->dl_next;
net/ipv4/inet_connection_sock.c
1521
req = next;
net/ipv4/inet_fragment.c
313
struct sk_buff *next = FRAG_CB(skb)->next_frag;
net/ipv4/inet_fragment.c
317
skb = next;
net/ipv4/inet_fragment.c
579
nextp = &clone->next;
net/ipv4/inet_fragment.c
650
nextp = &fp->next;
net/ipv4/inetpeer.c
105
next = rcu_dereference_raw(*pp);
net/ipv4/inetpeer.c
106
if (!next)
net/ipv4/inetpeer.c
108
parent = next;
net/ipv4/inetpeer.c
124
pp = &next->rb_left;
net/ipv4/inetpeer.c
126
pp = &next->rb_right;
net/ipv4/inetpeer.c
96
struct rb_node **pp, *parent, *next;
net/ipv4/ip_input.c
158
for (ra = rcu_dereference(net->ipv4.ra_chain); ra; ra = rcu_dereference(ra->next)) {
net/ipv4/ip_input.c
580
struct sk_buff *skb, *next;
net/ipv4/ip_input.c
582
list_for_each_entry_safe(skb, next, head, list) {
net/ipv4/ip_input.c
604
struct sk_buff *skb, *next, *hint = NULL;
net/ipv4/ip_input.c
608
list_for_each_entry_safe(skb, next, head, list) {
net/ipv4/ip_input.c
653
struct sk_buff *skb, *next;
net/ipv4/ip_input.c
656
list_for_each_entry_safe(skb, next, head, list) {
net/ipv4/ip_output.c
1422
tail_skb = &(tmp_skb->next);
net/ipv4/ip_output.c
636
if (frag->next)
net/ipv4/ip_output.c
819
((frag->len & 7) && frag->next) ||
net/ipv4/ip_sockglue.c
361
rap = &ra->next) {
net/ipv4/ip_sockglue.c
370
RCU_INIT_POINTER(*rap, ra->next);
net/ipv4/ip_sockglue.c
392
RCU_INIT_POINTER(new_ra->next, ra);
net/ipv4/ipconfig.c
1056
d = d->next;
net/ipv4/ipconfig.c
1259
if (!d->next) {
net/ipv4/ipconfig.c
1282
if ((d = d->next))
net/ipv4/ipconfig.c
1528
ic_first_dev->next) {
net/ipv4/ipconfig.c
190
struct ic_device *next;
net/ipv4/ipconfig.c
253
last = &d->next;
net/ipv4/ipconfig.c
321
struct ic_device *d, *next;
net/ipv4/ipconfig.c
325
next = ic_first_dev;
net/ipv4/ipconfig.c
326
while ((d = next)) {
net/ipv4/ipconfig.c
331
next = d->next;
net/ipv4/ipconfig.c
551
d = d->next;
net/ipv4/ipmr.c
130
ret = list_entry_rcu(net->ipv4.mr_tables.next,
net/ipv4/ipmr.c
133
ret = list_entry_rcu(mrt->list.next,
net/ipv4/ipmr.c
286
struct mr_table *mrt, *next;
net/ipv4/ipmr.c
289
list_for_each_entry_safe(mrt, next, &net->ipv4.mr_tables, list) {
net/ipv4/ipmr.c
3114
.next = mr_vif_seq_next,
net/ipv4/ipmr.c
3174
.next = mr_mfc_seq_next,
net/ipv4/ipmr.c
770
struct mr_mfc *c, *next;
net/ipv4/ipmr.c
785
list_for_each_entry_safe(c, next, &mrt->mfc_unres_queue, list) {
net/ipv4/ipmr_base.c
185
if (c->list.next != it->cache)
net/ipv4/ipmr_base.c
186
return list_entry(c->list.next, struct mr_mfc, list);
net/ipv4/netfilter/arp_tables.c
350
goto next;
net/ipv4/netfilter/arp_tables.c
383
next: ;
net/ipv4/netfilter/ip_tables.c
412
goto next;
net/ipv4/netfilter/ip_tables.c
445
next: ;
net/ipv4/nexthop.c
2603
struct rb_node **pp, *parent = NULL, *next;
net/ipv4/nexthop.c
2615
next = *pp;
net/ipv4/nexthop.c
2616
if (!next)
net/ipv4/nexthop.c
2619
parent = next;
net/ipv4/nexthop.c
2623
pp = &next->rb_left;
net/ipv4/nexthop.c
2625
pp = &next->rb_right;
net/ipv4/nexthop.c
591
struct rb_node **pp, *parent = NULL, *next;
net/ipv4/nexthop.c
597
next = rcu_dereference_raw(*pp);
net/ipv4/nexthop.c
598
if (!next)
net/ipv4/nexthop.c
600
parent = next;
net/ipv4/nexthop.c
604
pp = &next->rb_left;
net/ipv4/nexthop.c
606
pp = &next->rb_right;
net/ipv4/ping.c
1144
.next = ping_seq_next,
net/ipv4/raw.c
1067
.next = raw_seq_next,
net/ipv4/raw_diag.c
164
goto next;
net/ipv4/raw_diag.c
166
goto next;
net/ipv4/raw_diag.c
169
goto next;
net/ipv4/raw_diag.c
172
goto next;
net/ipv4/raw_diag.c
175
next:
net/ipv4/route.c
229
.next = rt_cache_seq_next,
net/ipv4/route.c
307
.next = rt_cpu_seq_next,
net/ipv4/route.c
3101
goto next;
net/ipv4/route.c
3104
goto next;
net/ipv4/route.c
3108
goto next;
net/ipv4/route.c
3114
goto next;
net/ipv4/route.c
3122
next:
net/ipv4/tcp.c
2292
skb = skb->next;
net/ipv4/tcp.c
2661
skb = skb_shinfo(skb)->frag_list ?: skb->next;
net/ipv4/tcp.c
3094
int next = (int)new_state[sk->sk_state];
net/ipv4/tcp.c
3095
int ns = next & TCP_STATE_MASK;
net/ipv4/tcp.c
3099
return next & TCP_ACTION_FIN;
net/ipv4/tcp_ao.c
1098
struct hlist_node *next;
net/ipv4/tcp_ao.c
1121
hlist_for_each_entry_safe(key, next, &ao_info->head, node) {
net/ipv4/tcp_input.c
3611
struct sk_buff *skb, *next;
net/ipv4/tcp_input.c
3622
for (skb = skb_rb_first(&sk->tcp_rtx_queue); skb; skb = next) {
net/ipv4/tcp_input.c
3694
next = skb_rb_next(skb);
net/ipv4/tcp_input.c
3697
tcp_highest_sack_replace(sk, skb, next);
net/ipv4/tcp_input.c
5761
return !skb_queue_is_last(list, skb) ? skb->next : NULL;
net/ipv4/tcp_input.c
5770
struct sk_buff *next = tcp_skb_next(skb, list);
net/ipv4/tcp_input.c
5780
return next;
net/ipv4/tcp_ipv4.c
3357
.next = bpf_iter_tcp_seq_next,
net/ipv4/tcp_ipv4.c
3379
.next = tcp_seq_next,
net/ipv4/tcp_offload.c
208
while (skb->next) {
net/ipv4/tcp_offload.c
223
skb = skb->next;
net/ipv4/tcp_offload.c
29
skb = skb->next;
net/ipv4/tcp_offload.c
66
th2 = tcp_hdr(seg->next);
net/ipv4/tcp_offload.c
67
iph2 = ip_hdr(seg->next);
net/ipv4/tcp_offload.c
73
while ((seg = seg->next)) {
net/ipv4/tcp_output.c
2571
struct sk_buff *skb, *next;
net/ipv4/tcp_output.c
2574
tcp_for_write_queue_from_safe(skb, next, sk) {
net/ipv4/tcp_output.c
2578
if (tcp_has_tx_tstamp(skb) || !tcp_skb_can_collapse(skb, next))
net/ipv4/tcp_output.c
2666
struct sk_buff *skb, *nskb, *next;
net/ipv4/tcp_output.c
2754
tcp_for_write_queue_from_safe(skb, next, sk) {
net/ipv4/tcp_output.c
2928
struct sk_buff *next_skb = skb->next;
net/ipv4/tunnel4.c
44
pprev = &t->next) {
net/ipv4/tunnel4.c
51
handler->next = *pprev;
net/ipv4/tunnel4.c
74
pprev = &t->next) {
net/ipv4/tunnel4.c
76
*pprev = handler->next;
net/ipv4/tunnel4.c
93
handler = rcu_dereference(handler->next)) \
net/ipv4/udp.c
1704
struct sk_buff *next, *to_drop = NULL;
net/ipv4/udp.c
1759
llist_for_each_entry_safe(skb, next, ll_list, ll_node) {
net/ipv4/udp.c
1765
skb->next = to_drop;
net/ipv4/udp.c
1806
to_drop = skb->next;
net/ipv4/udp.c
2514
struct sk_buff *next, *segs;
net/ipv4/udp.c
2523
skb_list_walk_safe(segs, skb, next) {
net/ipv4/udp.c
3598
sk = hlist_entry_safe(sk->__sk_common.skc_portaddr_node.next,
net/ipv4/udp.c
3730
.next = bpf_iter_udp_seq_next,
net/ipv4/udp.c
3753
.next = udp_seq_next,
net/ipv4/udp_diag.c
116
goto next;
net/ipv4/udp_diag.c
118
goto next;
net/ipv4/udp_diag.c
121
goto next;
net/ipv4/udp_diag.c
124
goto next;
net/ipv4/udp_diag.c
127
goto next;
net/ipv4/udp_diag.c
133
next:
net/ipv4/udp_offload.c
304
} while ((skb = skb->next));
net/ipv4/udp_offload.c
385
if ((udp_hdr(seg)->dest == udp_hdr(seg->next)->dest) &&
net/ipv4/udp_offload.c
386
(udp_hdr(seg)->source == udp_hdr(seg->next)->source) &&
net/ipv4/udp_offload.c
387
(ip_hdr(seg)->daddr == ip_hdr(seg->next)->daddr) &&
net/ipv4/udp_offload.c
388
(ip_hdr(seg)->saddr == ip_hdr(seg->next)->saddr))
net/ipv4/udp_offload.c
391
while ((seg = seg->next)) {
net/ipv4/udp_offload.c
441
uh2 = udp_hdr(seg->next);
net/ipv4/udp_offload.c
442
iph2 = ipv6_hdr(seg->next);
net/ipv4/udp_offload.c
449
while ((seg = seg->next)) {
net/ipv4/udp_offload.c
587
if (!seg->next)
net/ipv4/udp_offload.c
599
seg = seg->next;
net/ipv4/udp_offload.c
66
!udp_tunnel_gro->list.first->next) {
net/ipv4/xfrm4_protocol.c
238
pprev = &t->next) {
net/ipv4/xfrm4_protocol.c
245
handler->next = *pprev;
net/ipv4/xfrm4_protocol.c
279
pprev = &t->next) {
net/ipv4/xfrm4_protocol.c
281
*pprev = handler->next;
net/ipv4/xfrm4_protocol.c
43
handler = rcu_dereference(handler->next)) \
net/ipv6/addrconf.c
4502
.next = if6_seq_next,
net/ipv6/addrconf.c
4614
unsigned long now, next, next_sec, next_sched;
net/ipv6/addrconf.c
4622
next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY);
net/ipv6/addrconf.c
4653
if (time_before(ifp->tstamp + ifp->prefered_lft * HZ, next))
net/ipv6/addrconf.c
4654
next = ifp->tstamp + ifp->prefered_lft * HZ;
net/ipv6/addrconf.c
4670
} else if (time_before(ifp->tstamp + ifp->prefered_lft * HZ - regen_advance * HZ, next))
net/ipv6/addrconf.c
4671
next = ifp->tstamp + ifp->prefered_lft * HZ - regen_advance * HZ;
net/ipv6/addrconf.c
4695
(time_before(ifp->tstamp + ifp->valid_lft * HZ, next)))
net/ipv6/addrconf.c
4696
next = ifp->tstamp + ifp->valid_lft * HZ;
net/ipv6/addrconf.c
4709
if (time_before(ifp->tstamp + ifp->prefered_lft * HZ, next))
net/ipv6/addrconf.c
4710
next = ifp->tstamp + ifp->prefered_lft * HZ;
net/ipv6/addrconf.c
4716
next_sec = round_jiffies_up(next);
net/ipv6/addrconf.c
4717
next_sched = next;
net/ipv6/addrconf.c
4720
if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ))
net/ipv6/addrconf.c
4728
now, next, next_sec, next_sched);
net/ipv6/addrconf.c
5299
goto next;
net/ipv6/addrconf.c
5304
next:
net/ipv6/addrconf.c
5315
ifmca = rcu_dereference(ifmca->next), ip_idx++) {
net/ipv6/anycast.c
216
struct ipv6_ac_socklist *next = pac->acl_next;
net/ipv6/anycast.c
226
pac = next;
net/ipv6/anycast.c
617
.next = ac6_seq_next,
net/ipv6/calipso.c
232
__list_del(entry->list.prev, entry->list.next);
net/ipv6/ila/ila_xlat.c
132
ila = rcu_access_pointer(ila->next);
net/ipv6/ila/ila_xlat.c
150
ila = rcu_access_pointer(ila->next);
net/ipv6/ila/ila_xlat.c
163
struct ila_map *next;
net/ipv6/ila/ila_xlat.c
167
next = rcu_access_pointer(ila->next);
net/ipv6/ila/ila_xlat.c
169
ila = next;
net/ipv6/ila/ila_xlat.c
20
struct ila_map __rcu *next;
net/ipv6/ila/ila_xlat.c
255
tila = rcu_dereference_protected(tila->next,
net/ipv6/ila/ila_xlat.c
261
RCU_INIT_POINTER(ila->next, tila);
net/ipv6/ila/ila_xlat.c
262
rcu_assign_pointer(prev->next, ila);
net/ipv6/ila/ila_xlat.c
265
RCU_INIT_POINTER(ila->next, head);
net/ipv6/ila/ila_xlat.c
301
ila = rcu_dereference_protected(ila->next,
net/ipv6/ila/ila_xlat.c
310
rcu_assign_pointer(prev->next, ila->next);
net/ipv6/ila/ila_xlat.c
315
head = rcu_dereference_protected(ila->next,
net/ipv6/ila/ila_xlat.c
555
ila = rcu_access_pointer(ila->next);
net/ipv6/ila/ila_xlat.c
589
ila = rcu_access_pointer(ila->next);
net/ipv6/ip6_fib.c
1543
struct fib6_info *iter, *next;
net/ipv6/ip6_fib.c
1545
list_for_each_entry_safe(iter, next, &purge_list, purge_link) {
net/ipv6/ip6_fib.c
1634
struct fib6_node *next;
net/ipv6/ip6_fib.c
1638
next = dir ? rcu_dereference(fn->right) :
net/ipv6/ip6_fib.c
1641
if (next) {
net/ipv6/ip6_fib.c
1642
fn = next;
net/ipv6/ip6_fib.c
1743
goto next;
net/ipv6/ip6_fib.c
1761
next:
net/ipv6/ip6_fib.c
2840
.next = ipv6_route_seq_next,
net/ipv6/ip6_fib.c
709
goto next;
net/ipv6/ip6_fib.c
713
next:
net/ipv6/ip6_flowlabel.c
164
*flp = fl->next;
net/ipv6/ip6_flowlabel.c
172
flp = &fl->next;
net/ipv6/ip6_flowlabel.c
197
*flp = fl->next;
net/ipv6/ip6_flowlabel.c
202
flp = &fl->next;
net/ipv6/ip6_flowlabel.c
245
fl->next = fl_ht[FL_HASH(fl->label)];
net/ipv6/ip6_flowlabel.c
289
inet->ipv6_fl_list = sfl->next;
net/ipv6/ip6_flowlabel.c
495
sfl->next = inet->ipv6_fl_list;
net/ipv6/ip6_flowlabel.c
559
sflp = &sfl->next) {
net/ipv6/ip6_flowlabel.c
568
*sflp = sfl->next;
net/ipv6/ip6_flowlabel.c
63
fl = rcu_dereference(fl->next))
net/ipv6/ip6_flowlabel.c
65
for (fl = rcu_dereference(fl->next); \
net/ipv6/ip6_flowlabel.c
67
fl = rcu_dereference(fl->next))
net/ipv6/ip6_flowlabel.c
72
sfl = rcu_dereference(sfl->next))
net/ipv6/ip6_flowlabel.c
857
.next = ip6fl_seq_next,
net/ipv6/ip6_gre.c
1578
t = rtnl_net_dereference(net, t->next);
net/ipv6/ip6_gre.c
271
rcu_assign_pointer(t->next, rtnl_dereference(*tp));
net/ipv6/ip6_gre.c
282
tp = &iter->next) {
net/ipv6/ip6_gre.c
284
rcu_assign_pointer(*tp, t->next);
net/ipv6/ip6_gre.c
304
tp = &t->next)
net/ipv6/ip6_input.c
112
struct sk_buff *skb, *next, *hint = NULL;
net/ipv6/ip6_input.c
116
list_for_each_entry_safe(skb, next, head, list) {
net/ipv6/ip6_input.c
330
struct sk_buff *skb, *next;
net/ipv6/ip6_input.c
333
list_for_each_entry_safe(skb, next, head, list) {
net/ipv6/ip6_input.c
84
struct sk_buff *skb, *next;
net/ipv6/ip6_input.c
86
list_for_each_entry_safe(skb, next, head, list) {
net/ipv6/ip6_offload.c
145
for (skb = segs; skb; skb = skb->next) {
net/ipv6/ip6_offload.c
165
if (skb->next)
net/ipv6/ip6_output.c
1917
tail_skb = &(tmp_skb->next);
net/ipv6/ip6_output.c
385
for (ra = ip6_ra_chain; ra; ra = ra->next) {
net/ipv6/ip6_output.c
756
if (frag->next)
net/ipv6/ip6_output.c
934
((frag->len & 7) && frag->next) ||
net/ipv6/ip6_tunnel.c
100
for (t = rcu_dereference(start); t; t = rcu_dereference(t->next))
net/ipv6/ip6_tunnel.c
217
rcu_assign_pointer(t->next , rtnl_dereference(*tp));
net/ipv6/ip6_tunnel.c
2295
t = rtnl_net_dereference(net, t->next);
net/ipv6/ip6_tunnel.c
2307
t = rtnl_net_dereference(net, t->next);
net/ipv6/ip6_tunnel.c
238
tp = &iter->next) {
net/ipv6/ip6_tunnel.c
240
rcu_assign_pointer(*tp, t->next);
net/ipv6/ip6_tunnel.c
349
tp = &t->next) {
net/ipv6/ip6_vti.c
1125
t = rtnl_net_dereference(net, t->next);
net/ipv6/ip6_vti.c
158
rcu_assign_pointer(t->next, rtnl_dereference(*tp));
net/ipv6/ip6_vti.c
170
tp = &iter->next) {
net/ipv6/ip6_vti.c
172
rcu_assign_pointer(*tp, t->next);
net/ipv6/ip6_vti.c
261
tp = &t->next) {
net/ipv6/ip6_vti.c
76
for (t = rcu_dereference(start); t; t = rcu_dereference(t->next))
net/ipv6/ip6mr.c
117
ret = list_entry_rcu(net->ipv6.mr6_tables.next,
net/ipv6/ip6mr.c
120
ret = list_entry_rcu(mrt->list.next,
net/ipv6/ip6mr.c
273
struct mr_table *mrt, *next;
net/ipv6/ip6mr.c
276
list_for_each_entry_safe(mrt, next, &net->ipv6.mr6_tables, list) {
net/ipv6/ip6mr.c
476
.next = mr_vif_seq_next,
net/ipv6/ip6mr.c
537
.next = mr_mfc_seq_next,
net/ipv6/ip6mr.c
820
struct mr_mfc *c, *next;
net/ipv6/ip6mr.c
822
list_for_each_entry_safe(c, next, &mrt->mfc_unres_queue, list) {
net/ipv6/ipv6_sockglue.c
74
for (rap = &ip6_ra_chain; (ra = *rap) != NULL; rap = &ra->next) {
net/ipv6/ipv6_sockglue.c
82
*rap = ra->next;
net/ipv6/ipv6_sockglue.c
96
new_ra->next = ra;
net/ipv6/mcast.c
123
pmc = sock_dereference(pmc->next, sk))
net/ipv6/mcast.c
128
pmc = rcu_dereference(pmc->next))
net/ipv6/mcast.c
148
mc = mc_dereference(mc->next, idev))
net/ipv6/mcast.c
153
mc = rcu_dereference(mc->next))
net/ipv6/mcast.c
158
mc = mc_dereference(mc->next, idev))
net/ipv6/mcast.c
2099
pmc_next = mc_dereference(pmc->next, idev);
net/ipv6/mcast.c
2121
rcu_assign_pointer(pmc_prev->next, pmc_next);
net/ipv6/mcast.c
220
mc_lst->next = NULL;
net/ipv6/mcast.c
243
mc_lst->next = np->ipv6_mc_list;
net/ipv6/mcast.c
2851
rcu_assign_pointer(idev->mc_list, mc_dereference(i->next, idev));
net/ipv6/mcast.c
2931
im = rcu_dereference(im->next);
net/ipv6/mcast.c
2998
.next = igmp6_mc_seq_next,
net/ipv6/mcast.c
300
lnk = &mc_lst->next) {
net/ipv6/mcast.c
303
*lnk = mc_lst->next;
net/ipv6/mcast.c
3046
state->im = rcu_dereference(state->im->next);
net/ipv6/mcast.c
3126
.next = igmp6_mcf_seq_next,
net/ipv6/mcast.c
336
np->ipv6_mc_list = mc_lst->next;
net/ipv6/mcast.c
766
rcu_assign_pointer(pmc->next, idev->mc_tomb);
net/ipv6/mcast.c
787
rcu_assign_pointer(pmc_prev->next, pmc->next);
net/ipv6/mcast.c
789
rcu_assign_pointer(idev->mc_tomb, pmc->next);
net/ipv6/mcast.c
822
nextpmc = mc_dereference(pmc->next, idev);
net/ipv6/mcast.c
968
rcu_assign_pointer(mc->next, idev->mc_list);
net/ipv6/mcast.c
996
map = &ma->next) {
net/ipv6/mcast.c
999
*map = ma->next;
net/ipv6/netfilter/ip6_tables.c
430
goto next;
net/ipv6/netfilter/ip6_tables.c
463
next: ;
net/ipv6/netfilter/nft_fib_ipv6.c
144
static bool nft_fib_v6_skip_icmpv6(const struct sk_buff *skb, u8 next, const struct ipv6hdr *iph)
net/ipv6/netfilter/nft_fib_ipv6.c
146
if (likely(next != IPPROTO_ICMPV6))
net/ipv6/ping.c
250
.next = ping_seq_next,
net/ipv6/raw.c
1239
.next = raw_seq_next,
net/ipv6/route.c
935
struct fib6_info *next = rcu_dereference(rt0->fib6_next);
net/ipv6/route.c
938
if (!next || next->fib6_metric != rt0->fib6_metric)
net/ipv6/route.c
939
next = leaf;
net/ipv6/route.c
941
if (next != rt0) {
net/ipv6/route.c
944
if (next->fib6_node)
net/ipv6/route.c
945
rcu_assign_pointer(fn->rr_ptr, next);
net/ipv6/sit.c
169
tp = &iter->next) {
net/ipv6/sit.c
171
rcu_assign_pointer(*tp, t->next);
net/ipv6/sit.c
181
rcu_assign_pointer(t->next, rtnl_dereference(*tp));
net/ipv6/sit.c
1822
t = rtnl_net_dereference(net, t->next);
net/ipv6/sit.c
243
tp = &t->next) {
net/ipv6/sit.c
291
prl = rcu_dereference(prl->next))
net/ipv6/sit.c
381
for (p = rtnl_dereference(t->prl); p; p = rtnl_dereference(p->next)) {
net/ipv6/sit.c
403
p->next = t->prl;
net/ipv6/sit.c
418
n = rcu_dereference_protected(p->next, 1);
net/ipv6/sit.c
436
p = &x->next) {
net/ipv6/sit.c
438
*p = x->next;
net/ipv6/tcp_ipv6.c
2256
.next = tcp_seq_next,
net/ipv6/tcpv6_offload.c
126
th2 = tcp_hdr(seg->next);
net/ipv6/tcpv6_offload.c
127
iph2 = ipv6_hdr(seg->next);
net/ipv6/tcpv6_offload.c
134
while ((seg = seg->next)) {
net/ipv6/tunnel6.c
100
pprev = &t->next) {
net/ipv6/tunnel6.c
102
*pprev = handler->next;
net/ipv6/tunnel6.c
120
handler = rcu_dereference(handler->next)) \
net/ipv6/tunnel6.c
57
pprev = &t->next) {
net/ipv6/tunnel6.c
64
handler->next = *pprev;
net/ipv6/udp.c
1357
} while ((frags = frags->next));
net/ipv6/udp.c
1877
.next = udp_seq_next,
net/ipv6/udp.c
920
struct sk_buff *next, *segs;
net/ipv6/udp.c
928
skb_list_walk_safe(segs, skb, next) {
net/ipv6/xfrm6_protocol.c
254
pprev = &t->next) {
net/ipv6/xfrm6_protocol.c
261
handler->next = *pprev;
net/ipv6/xfrm6_protocol.c
295
pprev = &t->next) {
net/ipv6/xfrm6_protocol.c
297
*pprev = handler->next;
net/ipv6/xfrm6_protocol.c
44
handler = rcu_dereference(handler->next)) \
net/kcm/kcmproc.c
238
.next = kcm_seq_next,
net/kcm/kcmsock.c
685
} else if (skb->next) {
net/kcm/kcmsock.c
686
txm->frag_skb = skb->next;
net/kcm/kcmsock.c
825
skb->next = tskb;
net/kcm/kcmsock.c
947
frag_prev->next = NULL;
net/key/af_key.c
3844
.next = pfkey_seq_next,
net/l2tp/l2tp_debugfs.c
261
.next = l2tp_dfs_seq_next,
net/l2tp/l2tp_ppp.c
1589
.next = pppol2tp_seq_next,
net/lapb/lapb_iface.c
66
if (lapb->node.next) {
net/llc/llc_proc.c
205
.next = llc_seq_next,
net/llc/llc_proc.c
212
.next = llc_seq_next,
net/llc/llc_proc.c
82
struct sock* sk, *next;
net/llc/llc_proc.c
92
next = sk_nulls_next(sk);
net/llc/llc_proc.c
93
if (next) {
net/llc/llc_proc.c
94
sk = next;
net/mac80211/mlme.c
3250
goto next;
net/mac80211/mlme.c
3262
next:
net/mac80211/rx.c
2273
entry = &cache->entries[cache->next++];
net/mac80211/rx.c
2274
if (cache->next >= IEEE80211_FRAGMENT_MAX)
net/mac80211/rx.c
2275
cache->next = 0;
net/mac80211/rx.c
2299
idx = cache->next;
net/mac80211/sta_info.h
466
unsigned int next;
net/mac80211/tx.c
1096
struct sk_buff *next = skb_queue_next(&tx->skbs, skb);
net/mac80211/tx.c
1097
next_len = next->len;
net/mac80211/tx.c
1346
struct sk_buff *next;
net/mac80211/tx.c
1349
skb_list_walk_safe(skb, skb, next)
net/mac80211/tx.c
3482
frag_tail = &(*frag_tail)->next;
net/mac80211/tx.c
3770
struct sk_buff *next;
net/mac80211/tx.c
3804
skb_list_walk_safe(skb, skb, next) {
net/mac80211/tx.c
4288
struct sk_buff *next;
net/mac80211/tx.c
4334
skb_list_walk_safe(skb, skb, next) {
net/mac80211/tx.c
4343
kfree_skb_list(next);
net/mac80211/tx.c
4597
struct sk_buff *next;
net/mac80211/tx.c
4603
skb_list_walk_safe(skb, skb, next) {
net/mac80211/tx.c
4619
struct sk_buff *seg, *next;
net/mac80211/tx.c
4671
skb_list_walk_safe(skb, seg, next) {
net/mac802154/llsec.c
926
struct mac802154_llsec_device_key *next;
net/mac802154/llsec.c
928
next = kzalloc_obj(*devkey, GFP_ATOMIC);
net/mac802154/llsec.c
929
if (!next)
net/mac802154/llsec.c
932
next->devkey.key_id = *in_key;
net/mac802154/llsec.c
938
list_add_rcu(&next->devkey.list, &dev->dev.keys);
net/mac802154/llsec.c
940
kfree_sensitive(next);
net/mctp/route.c
416
skb->next = NULL;
net/mctp/route.c
419
key->reasm_tailp = &skb->next;
net/mpls/af_mpls.c
1576
goto next;
net/mpls/af_mpls.c
1593
next:
net/mpls/mpls_gso.c
70
} while ((skb = skb->next));
net/mptcp/mptcp_diag.c
166
goto next;
net/mptcp/mptcp_diag.c
169
goto next;
net/mptcp/mptcp_diag.c
172
goto next;
net/mptcp/mptcp_diag.c
175
goto next;
net/mptcp/mptcp_diag.c
178
next:
net/mptcp/pm_kernel.c
1084
goto next;
net/mptcp/pm_kernel.c
1106
next:
net/mptcp/pm_kernel.c
1128
goto next;
net/mptcp/pm_kernel.c
1132
goto next;
net/mptcp/pm_kernel.c
1142
next:
net/mptcp/pm_kernel.c
1269
cur = list_entry(list->next,
net/mptcp/pm_kernel.c
1297
if (free_list.next == &pernet->endp_list)
net/mptcp/pm_kernel.c
1303
free_list.prev->next = &free_list;
net/mptcp/pm_kernel.c
1304
free_list.next->prev = &free_list;
net/mptcp/pm_kernel.c
1466
goto next;
net/mptcp/pm_kernel.c
1476
next:
net/mptcp/pm_kernel.c
959
goto next;
net/mptcp/pm_kernel.c
972
next:
net/mptcp/protocol.c
2179
prefetch(skb->next);
net/mptcp/protocol.c
3166
int next = (int)new_state[sk->sk_state];
net/mptcp/protocol.c
3167
int ns = next & TCP_STATE_MASK;
net/mptcp/protocol.c
3171
return next & TCP_ACTION_FIN;
net/mptcp/subflow.c
1088
skb = skb->next;
net/netfilter/core.c
654
struct sk_buff *skb, *next;
net/netfilter/core.c
658
list_for_each_entry_safe(skb, next, head, list) {
net/netfilter/ipset/ip_set_hash_gen.h
305
struct mtype_elem next; /* temporary storage for uadd */
net/netfilter/ipset/ip_set_hash_gen.h
945
mtype_data_next(&h->next, d);
net/netfilter/ipset/ip_set_hash_ip.c
153
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_ip.c
157
hash_ip4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_ip.c
202
hash_ip6_data_next(struct hash_ip6_elem *next, const struct hash_ip6_elem *e)
net/netfilter/ipset/ip_set_hash_ip.c
70
hash_ip4_data_next(struct hash_ip4_elem *next, const struct hash_ip4_elem *e)
net/netfilter/ipset/ip_set_hash_ip.c
72
next->ip = e->ip;
net/netfilter/ipset/ip_set_hash_ipmac.c
179
hash_ipmac6_data_next(struct hash_ipmac6_elem *next,
net/netfilter/ipset/ip_set_hash_ipmac.c
71
hash_ipmac4_data_next(struct hash_ipmac4_elem *next,
net/netfilter/ipset/ip_set_hash_ipmac.c
74
next->ip = e->ip;
net/netfilter/ipset/ip_set_hash_ipmark.c
152
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
156
hash_ipmark4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_ipmark.c
201
hash_ipmark6_data_next(struct hash_ipmark6_elem *next,
net/netfilter/ipset/ip_set_hash_ipmark.c
69
hash_ipmark4_data_next(struct hash_ipmark4_elem *next,
net/netfilter/ipset/ip_set_hash_ipmark.c
72
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_ipport.c
188
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
190
p = retried && ip == ntohl(h->next.ip) ? ntohs(h->next.port)
net/netfilter/ipset/ip_set_hash_ipport.c
196
hash_ipport4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_ipport.c
246
hash_ipport6_data_next(struct hash_ipport6_elem *next,
net/netfilter/ipset/ip_set_hash_ipport.c
249
next->port = d->port;
net/netfilter/ipset/ip_set_hash_ipport.c
348
port = ntohs(h->next.port);
net/netfilter/ipset/ip_set_hash_ipport.c
79
hash_ipport4_data_next(struct hash_ipport4_elem *next,
net/netfilter/ipset/ip_set_hash_ipport.c
82
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_ipport.c
83
next->port = d->port;
net/netfilter/ipset/ip_set_hash_ipportip.c
184
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
186
p = retried && ip == ntohl(h->next.ip) ? ntohs(h->next.port)
net/netfilter/ipset/ip_set_hash_ipportip.c
192
hash_ipportip4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_ipportip.c
245
hash_ipportip6_data_next(struct hash_ipportip6_elem *next,
net/netfilter/ipset/ip_set_hash_ipportip.c
248
next->port = d->port;
net/netfilter/ipset/ip_set_hash_ipportip.c
343
port = ntohs(h->next.port);
net/netfilter/ipset/ip_set_hash_ipportip.c
77
hash_ipportip4_data_next(struct hash_ipportip4_elem *next,
net/netfilter/ipset/ip_set_hash_ipportip.c
80
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_ipportip.c
81
next->port = d->port;
net/netfilter/ipset/ip_set_hash_ipportnet.c
121
hash_ipportnet4_data_next(struct hash_ipportnet4_elem *next,
net/netfilter/ipset/ip_set_hash_ipportnet.c
124
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_ipportnet.c
125
next->port = d->port;
net/netfilter/ipset/ip_set_hash_ipportnet.c
126
next->ip2 = d->ip2;
net/netfilter/ipset/ip_set_hash_ipportnet.c
270
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
271
p = ntohs(h->next.port);
net/netfilter/ipset/ip_set_hash_ipportnet.c
272
ip2 = ntohl(h->next.ip2);
net/netfilter/ipset/ip_set_hash_ipportnet.c
287
hash_ipportnet4_data_next(&h->next,
net/netfilter/ipset/ip_set_hash_ipportnet.c
376
hash_ipportnet6_data_next(struct hash_ipportnet6_elem *next,
net/netfilter/ipset/ip_set_hash_ipportnet.c
379
next->port = d->port;
net/netfilter/ipset/ip_set_hash_ipportnet.c
501
port = ntohs(h->next.port);
net/netfilter/ipset/ip_set_hash_mac.c
60
hash_mac4_data_next(struct hash_mac4_elem *next,
net/netfilter/ipset/ip_set_hash_net.c
102
hash_net4_data_next(struct hash_net4_elem *next,
net/netfilter/ipset/ip_set_hash_net.c
105
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_net.c
193
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_net.c
198
hash_net4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_net.c
273
hash_net6_data_next(struct hash_net6_elem *next,
net/netfilter/ipset/ip_set_hash_netiface.c
129
hash_netiface4_data_next(struct hash_netiface4_elem *next,
net/netfilter/ipset/ip_set_hash_netiface.c
132
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_netiface.c
261
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_netiface.c
266
hash_netiface4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_netiface.c
363
hash_netiface6_data_next(struct hash_netiface6_elem *next,
net/netfilter/ipset/ip_set_hash_netnet.c
125
hash_netnet4_data_next(struct hash_netnet4_elem *next,
net/netfilter/ipset/ip_set_hash_netnet.c
128
next->ipcmp = d->ipcmp;
net/netfilter/ipset/ip_set_hash_netnet.c
252
ip = ntohl(h->next.ip[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
253
ip2 = ntohl(h->next.ip[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
265
hash_netnet4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_netnet.c
361
hash_netnet6_data_next(struct hash_netnet6_elem *next,
net/netfilter/ipset/ip_set_hash_netport.c
117
hash_netport4_data_next(struct hash_netport4_elem *next,
net/netfilter/ipset/ip_set_hash_netport.c
120
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_netport.c
121
next->port = d->port;
net/netfilter/ipset/ip_set_hash_netport.c
240
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_netport.c
241
p = ntohs(h->next.port);
net/netfilter/ipset/ip_set_hash_netport.c
252
hash_netport4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_netport.c
334
hash_netport6_data_next(struct hash_netport6_elem *next,
net/netfilter/ipset/ip_set_hash_netport.c
337
next->port = d->port;
net/netfilter/ipset/ip_set_hash_netport.c
447
port = ntohs(h->next.port);
net/netfilter/ipset/ip_set_hash_netportnet.c
132
hash_netportnet4_data_next(struct hash_netportnet4_elem *next,
net/netfilter/ipset/ip_set_hash_netportnet.c
135
next->ipcmp = d->ipcmp;
net/netfilter/ipset/ip_set_hash_netportnet.c
136
next->port = d->port;
net/netfilter/ipset/ip_set_hash_netportnet.c
300
ip = ntohl(h->next.ip[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
301
p = ntohs(h->next.port);
net/netfilter/ipset/ip_set_hash_netportnet.c
302
ip2 = ntohl(h->next.ip[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
317
hash_netportnet4_data_next(&h->next,
net/netfilter/ipset/ip_set_hash_netportnet.c
424
hash_netportnet6_data_next(struct hash_netportnet6_elem *next,
net/netfilter/ipset/ip_set_hash_netportnet.c
427
next->port = d->port;
net/netfilter/ipset/ip_set_hash_netportnet.c
557
port = ntohs(h->next.port);
net/netfilter/ipset/ip_set_list_set.c
190
struct set_elem *e, *next, *prev = NULL;
net/netfilter/ipset/ip_set_list_set.c
207
next = list_next_entry(e, list);
net/netfilter/ipset/ip_set_list_set.c
209
next->id == d->refid;
net/netfilter/ipset/ip_set_list_set.c
241
struct set_elem *e, *n, *prev, *next;
net/netfilter/ipset/ip_set_list_set.c
245
n = prev = next = NULL;
net/netfilter/ipset/ip_set_list_set.c
255
next = e;
net/netfilter/ipset/ip_set_list_set.c
261
if ((d->before > 0 && !next) ||
net/netfilter/ipset/ip_set_list_set.c
284
if (!list_is_last(&next->list, &map->members))
net/netfilter/ipset/ip_set_list_set.c
285
n = list_next_entry(next, list);
net/netfilter/ipset/ip_set_list_set.c
306
else if (next)
net/netfilter/ipset/ip_set_list_set.c
307
list_add_tail_rcu(&e->list, &next->list);
net/netfilter/ipset/ip_set_list_set.c
323
struct set_elem *e, *n, *next, *prev = NULL;
net/netfilter/ipset/ip_set_list_set.c
335
next = list_next_entry(e, list);
net/netfilter/ipset/ip_set_list_set.c
337
next->id != d->refid)
net/netfilter/ipvs/ip_vs_app.c
556
if ((e = inc->a_list.next) != &app->incs_list)
net/netfilter/ipvs/ip_vs_app.c
560
for (e = app->a_list.next; e != &ipvs->app_list; e = e->next) {
net/netfilter/ipvs/ip_vs_app.c
592
.next = ip_vs_app_seq_next,
net/netfilter/ipvs/ip_vs_conn.c
1188
.next = ip_vs_conn_seq_next,
net/netfilter/ipvs/ip_vs_conn.c
1254
.next = ip_vs_conn_seq_next,
net/netfilter/ipvs/ip_vs_ctl.c
1335
struct ip_vs_dest *dest, *next;
net/netfilter/ipvs/ip_vs_ctl.c
1339
list_for_each_entry_safe(dest, next, &ipvs->dest_trash, t_list) {
net/netfilter/ipvs/ip_vs_ctl.c
2476
.next = ip_vs_info_seq_next,
net/netfilter/ipvs/ip_vs_dh.c
116
p = p->next;
net/netfilter/ipvs/ip_vs_dh.c
122
p = p->next;
net/netfilter/ipvs/ip_vs_lblc.c
231
struct hlist_node *next;
net/netfilter/ipvs/ip_vs_lblc.c
237
hlist_for_each_entry_safe(en, next, &tbl->bucket[i], list) {
net/netfilter/ipvs/ip_vs_lblc.c
258
struct hlist_node *next;
net/netfilter/ipvs/ip_vs_lblc.c
266
hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) {
net/netfilter/ipvs/ip_vs_lblc.c
301
struct hlist_node *next;
net/netfilter/ipvs/ip_vs_lblc.c
323
hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) {
net/netfilter/ipvs/ip_vs_lblcr.c
398
struct hlist_node *next;
net/netfilter/ipvs/ip_vs_lblcr.c
403
hlist_for_each_entry_safe(en, next, &tbl->bucket[i], list) {
net/netfilter/ipvs/ip_vs_lblcr.c
425
struct hlist_node *next;
net/netfilter/ipvs/ip_vs_lblcr.c
431
hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) {
net/netfilter/ipvs/ip_vs_lblcr.c
465
struct hlist_node *next;
net/netfilter/ipvs/ip_vs_lblcr.c
487
hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) {
net/netfilter/ipvs/ip_vs_mh.c
139
while ((p = p->next) != &svc->destinations) {
net/netfilter/ipvs/ip_vs_mh.c
185
p = p->next;
net/netfilter/ipvs/ip_vs_mh.c
191
p = p->next;
net/netfilter/ipvs/ip_vs_mh.c
221
p = p->next;
net/netfilter/ipvs/ip_vs_proto.c
101
for (; *pp_p; pp_p = &(*pp_p)->next) {
net/netfilter/ipvs/ip_vs_proto.c
103
*pp_p = pp->next;
net/netfilter/ipvs/ip_vs_proto.c
123
for (; *pd_p; pd_p = &(*pd_p)->next) {
net/netfilter/ipvs/ip_vs_proto.c
125
*pd_p = pd->next;
net/netfilter/ipvs/ip_vs_proto.c
144
for (pp = ip_vs_proto_table[hash]; pp; pp = pp->next) {
net/netfilter/ipvs/ip_vs_proto.c
162
for (pd = ipvs->proto_data_table[hash]; pd; pd = pd->next) {
net/netfilter/ipvs/ip_vs_proto.c
180
for (pd = ipvs->proto_data_table[i]; pd; pd = pd->next) {
net/netfilter/ipvs/ip_vs_proto.c
52
pp->next = ip_vs_proto_table[hash];
net/netfilter/ipvs/ip_vs_proto.c
75
pd->next = ipvs->proto_data_table[hash];
net/netfilter/ipvs/ip_vs_proto.c
83
ipvs->proto_data_table[hash] = pd->next;
net/netfilter/ipvs/ip_vs_rr.c
42
svc->sched_data = p->next->prev;
net/netfilter/ipvs/ip_vs_sh.c
183
p = p->next;
net/netfilter/ipvs/ip_vs_sh.c
195
p = p->next;
net/netfilter/ipvs/ip_vs_sync.c
312
sb = list_entry(ms->sync_queue.next, struct ip_vs_sync_buff,
net/netfilter/nf_conncount.c
575
goto next; /* do not bother */
net/netfilter/nf_conncount.c
594
next:
net/netfilter/nf_conntrack_ecache.c
62
next:
net/netfilter/nf_conntrack_ecache.c
88
goto next;
net/netfilter/nf_conntrack_expect.c
237
struct hlist_node *next;
net/netfilter/nf_conntrack_expect.c
244
hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) {
net/netfilter/nf_conntrack_expect.c
463
struct hlist_node *next;
net/netfilter/nf_conntrack_expect.c
474
hlist_for_each_entry_safe(i, next, &nf_ct_expect_hash[h], hnode) {
net/netfilter/nf_conntrack_expect.c
539
const struct hlist_node *next;
net/netfilter/nf_conntrack_expect.c
545
hlist_for_each_entry_safe(exp, next,
net/netfilter/nf_conntrack_expect.c
565
const struct hlist_node *next;
net/netfilter/nf_conntrack_expect.c
571
hlist_for_each_entry_safe(exp, next,
net/netfilter/nf_conntrack_expect.c
701
.next = exp_seq_next,
net/netfilter/nf_conntrack_sip.c
808
struct hlist_node *next;
net/netfilter/nf_conntrack_sip.c
812
hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) {
net/netfilter/nf_conntrack_sip.c
832
struct hlist_node *next;
net/netfilter/nf_conntrack_sip.c
835
hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) {
net/netfilter/nf_conntrack_standalone.c
397
.next = ct_seq_next,
net/netfilter/nf_conntrack_standalone.c
478
.next = ct_cpu_seq_next,
net/netfilter/nf_flow_table_offload.c
1167
struct flow_block_cb *block_cb, *next;
net/netfilter/nf_flow_table_offload.c
1176
list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) {
net/netfilter/nf_flow_table_procfs.c
62
.next = nf_flow_table_cpu_seq_next,
net/netfilter/nf_log.c
416
.next = seq_next,
net/netfilter/nf_nat_sip.c
191
goto next;
net/netfilter/nf_nat_sip.c
196
goto next;
net/netfilter/nf_nat_sip.c
259
next:
net/netfilter/nf_synproxy_core.c
302
.next = synproxy_cpu_seq_next,
net/netfilter/nf_tables_api.c
10153
struct nft_trans *trans, *next;
net/netfilter/nf_tables_api.c
10165
list_for_each_entry_safe(trans, next, &head, list) {
net/netfilter/nf_tables_api.c
10264
struct nft_trans *trans, *next;
net/netfilter/nf_tables_api.c
10266
list_for_each_entry_safe(trans, next, &nft_net->commit_list, list) {
net/netfilter/nf_tables_api.c
10428
struct nft_trans_gc *trans, *next;
net/netfilter/nf_tables_api.c
10435
list_for_each_entry_safe(trans, next, &trans_gc_list, list) {
net/netfilter/nf_tables_api.c
10563
struct nft_set_elem_catchall *catchall, *next;
net/netfilter/nf_tables_api.c
10571
list_for_each_entry_safe(catchall, next, &set->catchall_list, list) {
net/netfilter/nf_tables_api.c
10593
struct nft_module_request *req, *next;
net/netfilter/nf_tables_api.c
10596
list_for_each_entry_safe(req, next, &nft_net->module_list, list) {
net/netfilter/nf_tables_api.c
10751
struct nft_set *set, *next;
net/netfilter/nf_tables_api.c
10753
list_for_each_entry_safe(set, next, set_update_list, pending_update) {
net/netfilter/nf_tables_api.c
10784
struct nft_trans *trans, *next;
net/netfilter/nf_tables_api.c
10837
list_for_each_entry_safe(trans, next, &nft_net->commit_list, list) {
net/netfilter/nf_tables_api.c
10882
list_for_each_entry_safe(trans, next, &nft_net->commit_list, list) {
net/netfilter/nf_tables_api.c
11096
struct nft_module_request *req, *next;
net/netfilter/nf_tables_api.c
11101
list_for_each_entry_safe(req, next, &module_list, list) {
net/netfilter/nf_tables_api.c
11149
struct nft_set *set, *next;
net/netfilter/nf_tables_api.c
11151
list_for_each_entry_safe(set, next, set_update_list, pending_update) {
net/netfilter/nf_tables_api.c
11164
struct nft_trans *trans, *next;
net/netfilter/nf_tables_api.c
11176
list_for_each_entry_safe_reverse(trans, next, &nft_net->commit_list,
net/netfilter/nf_tables_api.c
11366
list_for_each_entry_safe_reverse(trans, next,
net/netfilter/nf_tables_api.c
2313
struct nft_hook *hook, *next;
net/netfilter/nf_tables_api.c
2325
list_for_each_entry_safe(hook, next,
net/netfilter/nf_tables_api.c
2410
struct nft_hook *hook, *next;
net/netfilter/nf_tables_api.c
2452
list_for_each_entry_safe(hook, next, hook_list, list) {
net/netfilter/nf_tables_api.c
2593
struct nft_hook *h, *next;
net/netfilter/nf_tables_api.c
2595
list_for_each_entry_safe(h, next, &hook->list, list) {
net/netfilter/nf_tables_api.c
2852
struct nft_hook *h, *next;
net/netfilter/nf_tables_api.c
2882
list_for_each_entry_safe(h, next, &hook.list, list) {
net/netfilter/nf_tables_api.c
3024
list_for_each_entry_safe(h, next, &hook.list, list) {
net/netfilter/nf_tables_api.c
351
struct nf_hook_ops *ops, *next;
net/netfilter/nf_tables_api.c
353
list_for_each_entry_safe(ops, next, &hook->ops_list, list) {
net/netfilter/nf_tables_api.c
3693
const struct nft_expr *expr, *next;
net/netfilter/nf_tables_api.c
3722
nft_rule_for_each_expr(expr, next, rule) {
net/netfilter/nf_tables_api.c
381
struct nft_hook *hook, *next;
net/netfilter/nf_tables_api.c
384
list_for_each_entry_safe(hook, next, hook_list, list) {
net/netfilter/nf_tables_api.c
4035
struct nft_expr *expr, *next;
net/netfilter/nf_tables_api.c
4043
next = nft_expr_next(expr);
net/netfilter/nf_tables_api.c
4045
expr = next;
net/netfilter/nf_tables_api.c
5647
struct nft_set_elem_catchall *next, *catchall;
net/netfilter/nf_tables_api.c
5649
list_for_each_entry_safe(catchall, next, &set->catchall_list, list) {
net/netfilter/nf_tables_api.c
7078
struct nft_set_elem_catchall *catchall, *next;
net/netfilter/nf_tables_api.c
7080
list_for_each_entry_safe(catchall, next, &set->catchall_list, list) {
net/netfilter/nf_tables_api.c
8900
struct nft_hook *hook, *next;
net/netfilter/nf_tables_api.c
8903
list_for_each_entry_safe(hook, next, hook_list, list) {
net/netfilter/nf_tables_api.c
8945
struct nft_hook *hook, *next;
net/netfilter/nf_tables_api.c
8973
list_for_each_entry_safe(hook, next, hook_list, list) {
net/netfilter/nf_tables_api.c
8989
struct nft_hook *hook, *next;
net/netfilter/nf_tables_api.c
8991
list_for_each_entry_safe(hook, next, hook_list, list) {
net/netfilter/nf_tables_api.c
9004
struct nft_hook *hook, *next;
net/netfilter/nf_tables_api.c
9016
list_for_each_entry_safe(hook, next, &flowtable_hook.list, list) {
net/netfilter/nf_tables_api.c
9076
list_for_each_entry_safe(hook, next, &flowtable_hook.list, list) {
net/netfilter/nf_tables_api.c
9226
struct nft_hook *this, *next;
net/netfilter/nf_tables_api.c
9228
list_for_each_entry_safe(this, next, &flowtable_hook->list, list) {
net/netfilter/nf_tables_api.c
9589
struct nft_hook *hook, *next;
net/netfilter/nf_tables_api.c
9592
list_for_each_entry_safe(hook, next, &flowtable->hook_list, list) {
net/netfilter/nf_tables_offload.c
302
struct nft_expr *expr, *next;
net/netfilter/nf_tables_offload.c
310
nft_rule_for_each_expr(expr, next, rule) {
net/netfilter/nf_tables_offload.c
328
struct flow_block_cb *block_cb, *next;
net/netfilter/nf_tables_offload.c
342
list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) {
net/netfilter/nfnetlink.c
348
struct nfnl_err *nfnl_err, *next;
net/netfilter/nfnetlink.c
350
list_for_each_entry_safe(nfnl_err, next, err_list, head)
net/netfilter/nfnetlink.c
356
struct nfnl_err *nfnl_err, *next;
net/netfilter/nfnetlink.c
358
list_for_each_entry_safe(nfnl_err, next, err_list, head) {
net/netfilter/nfnetlink_log.c
1106
.next = seq_next,
net/netfilter/nfnetlink_queue.c
1015
if (skb->next == NULL) { /* last packet, no need to copy entry */
net/netfilter/nfnetlink_queue.c
1741
h = h->next;
net/netfilter/nfnetlink_queue.c
1799
.next = seq_next,
net/netfilter/nfnetlink_queue.c
480
struct nf_queue_entry *entry, *next;
net/netfilter/nfnetlink_queue.c
483
list_for_each_entry_safe(entry, next, &queue->queue_list, list) {
net/netfilter/nft_bitwise.c
634
const struct nft_expr *next;
net/netfilter/nft_bitwise.c
639
next = nft_expr_next(expr);
net/netfilter/nft_bitwise.c
640
if (next->ops == &nft_bitwise_ops)
net/netfilter/nft_bitwise.c
641
return nft_bitwise_reduce(track, next);
net/netfilter/nft_bitwise.c
642
else if (next->ops == &nft_bitwise_fast_ops)
net/netfilter/nft_bitwise.c
643
return nft_bitwise_fast_reduce(track, next);
net/netfilter/nft_set_hash.c
797
struct hlist_node *next;
net/netfilter/nft_set_hash.c
801
hlist_for_each_entry_safe(he, next, &priv->table[i], node) {
net/netfilter/nft_set_pipapo.c
1760
struct nft_trans_gc *gc, *next;
net/netfilter/nft_set_pipapo.c
1771
list_for_each_entry_safe(gc, next, &priv->gc_head, list) {
net/netfilter/nft_set_rbtree.c
343
struct rb_node *node, *next, *parent, **p, *first = NULL;
net/netfilter/nft_set_rbtree.c
383
for (node = first; node != NULL; node = next) {
net/netfilter/nft_set_rbtree.c
384
next = rb_next(node);
net/netfilter/nft_set_rbtree.c
900
struct rb_node *node, *next;
net/netfilter/nft_set_rbtree.c
902
for (node = rb_first(&priv->root); node ; node = next) {
net/netfilter/nft_set_rbtree.c
903
next = rb_next(node);
net/netfilter/nft_set_rbtree.c
994
struct nft_rbtree_elem *rbe, *next;
net/netfilter/nft_set_rbtree.c
998
list_for_each_entry_safe(rbe, next, &priv->expired, list) {
net/netfilter/x_tables.c
1585
.next = xt_table_seq_next,
net/netfilter/x_tables.c
1627
trav->curr = trav->curr->next;
net/netfilter/x_tables.c
1637
trav->curr = trav->curr->next;
net/netfilter/x_tables.c
1704
.next = xt_match_seq_next,
net/netfilter/x_tables.c
1738
.next = xt_target_seq_next,
net/netfilter/xt_hashlimit.c
1216
.next = dl_seq_next,
net/netfilter/xt_hashlimit.c
1223
.next = dl_seq_next,
net/netfilter/xt_hashlimit.c
1230
.next = dl_seq_next,
net/netfilter/xt_recent.c
163
e = list_entry(t->lru_list.next, struct recent_entry, lru_list);
net/netfilter/xt_recent.c
186
e = list_entry(t->lru_list.next, struct recent_entry, lru_list);
net/netfilter/xt_recent.c
231
struct recent_entry *e, *next;
net/netfilter/xt_recent.c
235
list_for_each_entry_safe(e, next, &t->iphash[i], list)
net/netfilter/xt_recent.c
501
const struct list_head *head = e->list.next;
net/netfilter/xt_recent.c
507
head = t->iphash[st->bucket].next;
net/netfilter/xt_recent.c
541
.next = recent_seq_next,
net/netlabel/netlabel_addrlist.h
131
i = i->next;
net/netlabel/netlabel_addrlist.h
150
for (iter = __af6list_valid((head)->next, head); \
net/netlabel/netlabel_addrlist.h
152
iter = __af6list_valid(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
155
for (iter = __af6list_valid_rcu((head)->next, head); \
net/netlabel/netlabel_addrlist.h
157
iter = __af6list_valid_rcu(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
160
for (iter = __af6list_valid((head)->next, head), \
net/netlabel/netlabel_addrlist.h
161
tmp = __af6list_valid(iter->list.next, head); \
net/netlabel/netlabel_addrlist.h
163
iter = tmp, tmp = __af6list_valid(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
64
i = i->next;
net/netlabel/netlabel_addrlist.h
83
for (iter = __af4list_valid((head)->next, head); \
net/netlabel/netlabel_addrlist.h
85
iter = __af4list_valid(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
88
for (iter = __af4list_valid_rcu((head)->next, head); \
net/netlabel/netlabel_addrlist.h
90
iter = __af4list_valid_rcu(iter->list.next, head))
net/netlabel/netlabel_addrlist.h
93
for (iter = __af4list_valid((head)->next, head), \
net/netlabel/netlabel_addrlist.h
94
tmp = __af4list_valid(iter->list.next, head); \
net/netlabel/netlabel_addrlist.h
96
iter = tmp, tmp = __af4list_valid(iter->list.next, head))
net/netlabel/netlabel_kapi.c
568
iter = iter->next;
net/netlabel/netlabel_kapi.c
588
iter->next = *catmap;
net/netlabel/netlabel_kapi.c
591
iter->next = prev->next;
net/netlabel/netlabel_kapi.c
592
prev->next = iter;
net/netlabel/netlabel_kapi.c
638
if (iter->next != NULL) {
net/netlabel/netlabel_kapi.c
639
iter = iter->next;
net/netlabel/netlabel_kapi.c
698
if (iter->next == NULL)
net/netlabel/netlabel_kapi.c
701
iter = iter->next;
net/netlink/af_netlink.c
2784
.next = netlink_seq_next,
net/netlink/policy.c
455
goto next;
net/netlink/policy.c
463
next:
net/netrom/af_netrom.c
1341
.next = nr_info_next,
net/netrom/nr_route.c
902
.next = nr_node_next,
net/netrom/nr_route.c
958
.next = nr_neigh_next,
net/nfc/llcp_core.c
1436
goto next;
net/nfc/llcp_core.c
1449
next:
net/nsh/nsh.c
117
for (skb = segs; skb; skb = skb->next) {
net/openvswitch/conntrack.c
1625
struct hlist_node *next;
net/openvswitch/conntrack.c
1627
hlist_for_each_entry_safe(ct_limit, next, head, hlist_node)
net/openvswitch/flow_netlink.c
1745
nlattr_set(nla, val, tbl[nla_type(nla)].next ? : tbl);
net/openvswitch/flow_netlink.c
404
.next = ovs_vxlan_ext_key_lens },
net/openvswitch/flow_netlink.c
440
.next = ovs_tunnel_key_lens, },
net/openvswitch/flow_netlink.c
451
.next = ovs_nsh_key_attr_lens, },
net/openvswitch/flow_netlink.c
46
const struct ovs_len_tbl *next;
net/packet/af_packet.c
248
struct sk_buff *next, *head = NULL, *tail;
net/packet/af_packet.c
252
for (; skb != NULL; skb = next) {
net/packet/af_packet.c
253
next = skb->next;
net/packet/af_packet.c
262
tail->next = skb;
net/packet/af_packet.c
3671
*mlp = ml->next;
net/packet/af_packet.c
3673
mlp = &ml->next;
net/packet/af_packet.c
3701
for (ml = po->mclist; ml; ml = ml->next) {
net/packet/af_packet.c
3720
i->next = po->mclist;
net/packet/af_packet.c
3724
po->mclist = i->next;
net/packet/af_packet.c
3739
for (mlp = &pkt_sk(sk)->mclist; (ml = *mlp) != NULL; mlp = &ml->next) {
net/packet/af_packet.c
3746
*mlp = ml->next;
net/packet/af_packet.c
3771
po->mclist = ml->next;
net/packet/af_packet.c
4742
.next = packet_seq_next,
net/packet/diag.c
205
goto next;
net/packet/diag.c
214
next:
net/packet/diag.c
48
for (ml = po->mclist; ml; ml = ml->next) {
net/packet/internal.h
8
struct packet_mclist *next;
net/phonet/pep-gprs.c
109
skb->next = skb_shinfo(skb)->frag_list;
net/phonet/pep.c
1245
skb->next = skb_shinfo(skb)->frag_list;
net/phonet/socket.c
598
.next = pn_sock_seq_next,
net/phonet/socket.c
768
.next = pn_res_seq_next,
net/psp/psp_main.c
119
struct psp_assoc *pas, *next;
net/psp/psp_main.c
135
list_for_each_entry_safe(pas, next, &psd->stale_assocs, assocs_list)
net/psp/psp_sock.c
251
struct psp_assoc *pas, *next;
net/psp/psp_sock.c
256
list_for_each_entry_safe(pas, next, &psd->prev_assocs, assocs_list) {
net/rds/af_rds.c
662
WARN_ON((&rs->rs_item != rs->rs_item.next ||
net/rds/bind.c
127
rs->rs_bound_node.next = NULL;
net/rds/ib_frmr.c
364
struct rds_ib_mr *ibmr, *next;
net/rds/ib_frmr.c
382
list_for_each_entry_safe(ibmr, next, list, unmap_list) {
net/rds/ib_rdma.c
296
struct llist_node *next;
net/rds/ib_rdma.c
301
next = node->next;
net/rds/ib_rdma.c
304
node = next;
net/rds/ib_rdma.c
321
struct llist_node **next = nodes_head;
net/rds/ib_rdma.c
325
*next = cur;
net/rds/ib_rdma.c
326
next = &cur->next;
net/rds/ib_rdma.c
328
*next = NULL;
net/rds/ib_rdma.c
413
clean_nodes = clean_nodes->next;
net/rds/ib_recv.c
525
cache->ready = head->next;
net/rds/ib_recv.c
545
frag = list_entry(ibinc->ii_frags.next, struct rds_page_frag, f_item);
net/rds/ib_recv.c
550
frag = list_entry(frag->f_item.next,
net/rds/ib_recv.c
59
recv->r_wr.next = NULL;
net/rds/ib_recv.c
805
frag = list_entry(ibinc->ii_frags.next, struct rds_page_frag, f_item);
net/rds/ib_recv.c
839
frag = list_entry(frag->f_item.next,
net/rds/ib_send.c
657
send->s_wr.next = NULL;
net/rds/ib_send.c
709
&send->s_wr, send->s_wr.num_sge, send->s_wr.next);
net/rds/ib_send.c
726
prev->s_wr.next = &send->s_wr;
net/rds/ib_send.c
829
send->s_atomic_wr.wr.next = NULL;
net/rds/ib_send.c
963
send->s_rdma_wr.wr.next = NULL;
net/rds/ib_send.c
966
prev->s_rdma_wr.wr.next = &send->s_rdma_wr.wr;
net/rds/ib_send.c
991
send->s_rdma_wr.wr.next);
net/rds/recv.c
444
*inc = list_entry(rs->rs_recv_queue.next,
net/rds/recv.c
514
notifier = list_entry(rs->rs_notify_queue.next,
net/rds/recv.c
525
notifier = list_entry(copy.next, struct rds_notifier, n_list);
net/rds/recv.c
652
info = list_entry(q->zcookie_head.next,
net/rds/send.c
141
rm0 = list_entry(cp0->cp_retrans.next, struct rds_message,
net/rds/send.c
158
rm0 = list_entry(cp0->cp_send_queue.next, struct rds_message,
net/rds/send.c
313
rm = list_entry(cp->cp_send_queue.next,
net/rds/send.c
673
rm = list_entry(messages->next, struct rds_message,
net/rds/send.c
851
rm = list_entry(list.next, struct rds_message, m_sock_item);
net/rose/af_rose.c
1514
.next = rose_info_next,
net/rose/rose_route.c
1011
rose_route = rose_route->next;
net/rose/rose_route.c
1045
rose_route = rose_route->next;
net/rose/rose_route.c
1076
rose_route->next = rose_route_list;
net/rose/rose_route.c
1108
rose_node = rose_node->next, ++i);
net/rose/rose_route.c
1118
: ((struct rose_node *)v)->next;
net/rose/rose_route.c
1151
.next = rose_node_next,
net/rose/rose_route.c
1167
rose_neigh = rose_neigh->next, ++i);
net/rose/rose_route.c
1177
: ((struct rose_neigh *)v)->next;
net/rose/rose_route.c
1222
.next = rose_neigh_next,
net/rose/rose_route.c
1238
rose_route = rose_route->next, ++i);
net/rose/rose_route.c
1248
: ((struct rose_route *)v)->next;
net/rose/rose_route.c
126
rose_neigh->next = rose_neigh_list;
net/rose/rose_route.c
1294
.next = rose_route_next,
net/rose/rose_route.c
1312
rose_neigh = rose_neigh->next;
net/rose/rose_route.c
1320
rose_node = rose_node->next;
net/rose/rose_route.c
1329
rose_route = rose_route->next;
net/rose/rose_route.c
144
rose_tmpn = rose_tmpn->next;
net/rose/rose_route.c
166
rose_node->next = NULL;
net/rose/rose_route.c
168
rose_tmpp->next = rose_node;
net/rose/rose_route.c
169
rose_node->next = NULL;
net/rose/rose_route.c
173
rose_node->next = rose_node_list;
net/rose/rose_route.c
176
rose_tmpp->next = rose_node;
net/rose/rose_route.c
177
rose_node->next = rose_tmpn;
net/rose/rose_route.c
209
rose_node_list = rose_node->next;
net/rose/rose_route.c
214
while (s != NULL && s->next != NULL) {
net/rose/rose_route.c
215
if (s->next == rose_node) {
net/rose/rose_route.c
216
s->next = rose_node->next;
net/rose/rose_route.c
221
s = s->next;
net/rose/rose_route.c
238
rose_neigh_list = rose_neigh->next;
net/rose/rose_route.c
242
while (s != NULL && s->next != NULL) {
net/rose/rose_route.c
243
if (s->next == rose_neigh) {
net/rose/rose_route.c
244
s->next = rose_neigh->next;
net/rose/rose_route.c
248
s = s->next;
net/rose/rose_route.c
266
rose_route_list = rose_route->next;
net/rose/rose_route.c
271
while (s != NULL && s->next != NULL) {
net/rose/rose_route.c
272
if (s->next == rose_route) {
net/rose/rose_route.c
273
s->next = rose_route->next;
net/rose/rose_route.c
278
s = s->next;
net/rose/rose_route.c
302
rose_node = rose_node->next;
net/rose/rose_route.c
316
rose_neigh = rose_neigh->next;
net/rose/rose_route.c
393
sn->next = rose_neigh_list;
net/rose/rose_route.c
414
rose_node = rose_node->next;
net/rose/rose_route.c
432
rose_node->next = rose_node_list;
net/rose/rose_route.c
459
rose_node = rose_node->next;
net/rose/rose_route.c
488
rose_neigh = rose_neigh->next;
net/rose/rose_route.c
497
rose_node = rose_node->next;
net/rose/rose_route.c
534
rose_route = rose_route->next;
net/rose/rose_route.c
562
rose_node = rose_node->next;
net/rose/rose_route.c
573
rose_neigh = rose_neigh->next;
net/rose/rose_route.c
668
for (rose_route = rose_route_list; rose_route != NULL; rose_route = rose_route->next)
net/rose/rose_route.c
67
rose_node = rose_node->next;
net/rose/rose_route.c
688
for (node = rose_node_list; node != NULL; node = node->next) {
net/rose/rose_route.c
700
for (node = rose_node_list; node != NULL; node = node->next) {
net/rose/rose_route.c
788
s = rose_route->next;
net/rose/rose_route.c
806
rose_route = rose_route->next;
net/rose/rose_route.c
81
rose_neigh = rose_neigh->next;
net/rose/rose_route.c
825
rose_neigh = rose_neigh->next;
net/rose/rose_route.c
846
for (rose_neigh = rose_neigh_list; rose_neigh != NULL; rose_neigh = rose_neigh->next) {
net/rose/rose_route.c
896
rose_neigh = rose_neigh->next;
net/rxrpc/ar-internal.h
951
struct rxrpc_txqueue *next;
net/rxrpc/call_event.c
149
for (req.tq = call->tx_queue; req.tq; req.tq = req.tq->next) {
net/rxrpc/call_event.c
236
tq = tq->next;
net/rxrpc/call_event.c
442
ktime_t next = READ_ONCE(call->expect_term_by), delay;
net/rxrpc/call_event.c
444
#define set(T) { ktime_t _t = (T); if (ktime_before(_t, next)) next = _t; }
net/rxrpc/call_event.c
454
delay = ktime_sub(next, now);
net/rxrpc/call_event.c
99
for (tq = call->tx_queue; tq; tq = tq->next) {
net/rxrpc/call_object.c
536
struct rxrpc_txqueue *tq, *next;
net/rxrpc/call_object.c
538
for (tq = call->tx_queue; tq; tq = next) {
net/rxrpc/call_object.c
539
next = tq->next;
net/rxrpc/call_object.c
613
call = list_entry(rx->to_be_accepted.next,
net/rxrpc/call_object.c
622
call = list_entry(rx->sock_calls.next,
net/rxrpc/conn_client.c
417
struct rxrpc_call *call = list_entry(bundle->waiting_calls.next,
net/rxrpc/conn_client.c
759
next:
net/rxrpc/conn_client.c
793
goto next;
net/rxrpc/conn_object.c
451
conn = list_entry(graveyard.next, struct rxrpc_connection,
net/rxrpc/input.c
244
call->tx_queue = tq->next;
net/rxrpc/input.c
307
prefetch(tq->next);
net/rxrpc/input.c
310
call->tx_queue = tq->next;
net/rxrpc/input.c
974
tq = tq->next;
net/rxrpc/input.c
994
tq = tq->next;
net/rxrpc/input_rack.c
207
for (tq = call->tx_queue; tq; tq = tq->next) {
net/rxrpc/input_rack.c
267
for (tq = call->tx_queue; tq; tq = tq->next) {
net/rxrpc/key.c
122
pptoken = &(*pptoken)->next)
net/rxrpc/key.c
267
pptoken = &(*pptoken)->next)
net/rxrpc/key.c
529
pp = &(*pp)->next;
net/rxrpc/key.c
548
struct rxrpc_key_token *next;
net/rxrpc/key.c
550
for (; token; token = next) {
net/rxrpc/key.c
551
next = token->next;
net/rxrpc/key.c
596
for (token = key->payload.data[0]; token; token = token->next) {
net/rxrpc/key.c
759
for (token = key->payload.data[0]; token; token = token->next) {
net/rxrpc/key.c
836
for (token = key->payload.data[0]; token; token = token->next) {
net/rxrpc/output.c
568
prefetch(tq->next);
net/rxrpc/output.c
599
tq = tq->next;
net/rxrpc/peer_event.c
272
peer = list_entry(collector->next,
net/rxrpc/proc.c
111
.next = rxrpc_call_seq_next,
net/rxrpc/proc.c
198
.next = rxrpc_connection_seq_next,
net/rxrpc/proc.c
274
.next = rxrpc_bundle_seq_next,
net/rxrpc/proc.c
397
.next = rxrpc_peer_seq_next,
net/rxrpc/proc.c
467
.next = rxrpc_local_seq_next,
net/rxrpc/recvmsg.c
440
l = rx->recvmsg_q.next;
net/rxrpc/security.c
109
for (token = key->payload.data[0]; token; token = token->next) {
net/rxrpc/security.c
85
for (token = key->payload.data[0]; token; token = token->next) {
net/rxrpc/sendmsg.c
295
call->send_queue->next = tq;
net/sched/act_gate.c
113
gact->next_entry = next;
net/sched/act_gate.c
75
struct tcfg_gate_entry *next;
net/sched/act_gate.c
83
next = gact->next_entry;
net/sched/act_gate.c
86
gact->current_gate_status = next->gate_state ? GATE_ACT_GATE_OPEN : 0;
net/sched/act_gate.c
88
gact->current_max_octets = next->maxoctets;
net/sched/act_gate.c
91
next->interval);
net/sched/act_gate.c
95
if (list_is_last(&next->list, &p->entries))
net/sched/act_gate.c
96
next = list_first_entry(&p->entries,
net/sched/act_gate.c
99
next = list_next_entry(next, list);
net/sched/cls_api.c
1126
for (; tp; tp = tcf_chain_dereference(tp->next, chain))
net/sched/cls_api.c
1130
tp = tcf_chain_dereference(tp->next, chain);
net/sched/cls_api.c
1629
struct flow_block_cb *block_cb, *next;
net/sched/cls_api.c
1651
list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) {
net/sched/cls_api.c
1671
struct flow_block_cb *block_cb, *next;
net/sched/cls_api.c
1675
list_for_each_entry_safe(block_cb, next, &bo->cb_list, list) {
net/sched/cls_api.c
1728
for (; tp; tp = rcu_dereference_bh(tp->next)) {
net/sched/cls_api.c
1889
struct tcf_proto __rcu *next;
net/sched/cls_api.c
1905
RCU_INIT_POINTER(tp->next, tcf_chain_tp_prev(chain, chain_info));
net/sched/cls_api.c
1918
struct tcf_proto *next = tcf_chain_dereference(chain_info->next, chain);
net/sched/cls_api.c
1922
tcf_chain0_head_change(chain, next);
net/sched/cls_api.c
1923
RCU_INIT_POINTER(*chain_info->pprev, next);
net/sched/cls_api.c
1977
struct tcf_proto *next;
net/sched/cls_api.c
1984
pprev = &tp_iter->next) {
net/sched/cls_api.c
1987
chain_info.next = tp_iter->next;
net/sched/cls_api.c
2002
next = tcf_chain_dereference(chain_info.next, chain);
net/sched/cls_api.c
2004
tcf_chain0_head_change(chain, next);
net/sched/cls_api.c
2005
RCU_INIT_POINTER(*chain_info.pprev, next);
net/sched/cls_api.c
2023
pprev = &tp->next) {
net/sched/cls_api.c
2042
chain_info->next = tp->next;
net/sched/cls_api.c
2045
chain_info->next = NULL;
net/sched/cls_api.c
756
tp_next = rcu_dereference_protected(tp->next, 1);
net/sched/cls_api.c
767
tp_next = rcu_dereference_protected(tp->next, 1);
net/sched/cls_api.c
81
static u32 next;
net/sched/cls_api.c
99
n, xa_limit_32b, &next, GFP_KERNEL);
net/sched/cls_flow.c
606
struct flow_filter *f, *next;
net/sched/cls_flow.c
608
list_for_each_entry_safe(f, next, &head->filters, list) {
net/sched/cls_flower.c
604
struct cls_fl_filter *f, *next;
net/sched/cls_flower.c
608
list_for_each_entry_safe(f, next, &mask->filters, list) {
net/sched/cls_fw.c
100
for (; f; f = rtnl_dereference(f->next)) {
net/sched/cls_fw.c
145
rtnl_dereference(f->next));
net/sched/cls_fw.c
172
fp = &pfp->next, pfp = rtnl_dereference(*fp)) {
net/sched/cls_fw.c
174
RCU_INIT_POINTER(*fp, rtnl_dereference(f->next));
net/sched/cls_fw.c
299
fp = &pfp->next, pfp = rtnl_dereference(*fp))
net/sched/cls_fw.c
303
RCU_INIT_POINTER(fnew->next, rtnl_dereference(pfp->next));
net/sched/cls_fw.c
343
RCU_INIT_POINTER(f->next, head->ht[fw_hash(handle)]);
net/sched/cls_fw.c
35
struct fw_filter __rcu *next;
net/sched/cls_fw.c
371
f = rtnl_dereference(f->next)) {
net/sched/cls_fw.c
64
f = rcu_dereference_bh(f->next)) {
net/sched/cls_route.c
168
f = rcu_dereference_bh(f->next))
net/sched/cls_route.c
174
f = rcu_dereference_bh(f->next))
net/sched/cls_route.c
180
f = rcu_dereference_bh(f->next))
net/sched/cls_route.c
236
f = rtnl_dereference(f->next))
net/sched/cls_route.c
295
struct route4_filter *next;
net/sched/cls_route.c
297
next = rtnl_dereference(f->next);
net/sched/cls_route.c
298
RCU_INIT_POINTER(b->ht[h2], next);
net/sched/cls_route.c
332
fp = &nf->next, nf = rtnl_dereference(*fp)) {
net/sched/cls_route.c
335
RCU_INIT_POINTER(*fp, rtnl_dereference(f->next));
net/sched/cls_route.c
451
fp = rtnl_dereference(fp->next))
net/sched/cls_route.c
48
struct route4_filter __rcu *next;
net/sched/cls_route.c
537
fp = &f1->next)
net/sched/cls_route.c
542
rcu_assign_pointer(f->next, f1);
net/sched/cls_route.c
552
fp = &pfp->next, pfp = rtnl_dereference(*fp)) {
net/sched/cls_route.c
554
rcu_assign_pointer(*fp, fold->next);
net/sched/cls_route.c
595
f = rtnl_dereference(f->next)) {
net/sched/cls_u32.c
1008
RCU_INIT_POINTER(ht->next, tp_c->hlist);
net/sched/cls_u32.c
1172
ins = &pins->next, pins = rtnl_dereference(*ins))
net/sched/cls_u32.c
1176
RCU_INIT_POINTER(n->next, pins);
net/sched/cls_u32.c
1215
ht = rtnl_dereference(ht->next)) {
net/sched/cls_u32.c
1225
n = rtnl_dereference(n->next)) {
net/sched/cls_u32.c
1298
ht = rtnl_dereference(ht->next)) {
net/sched/cls_u32.c
1316
n = rtnl_dereference(n->next)) {
net/sched/cls_u32.c
147
n = rcu_dereference_bh(n->next);
net/sched/cls_u32.c
153
n = rcu_dereference_bh(n->next);
net/sched/cls_u32.c
169
n = rcu_dereference_bh(n->next);
net/sched/cls_u32.c
185
n = rcu_dereference_bh(n->next);
net/sched/cls_u32.c
193
n = rcu_dereference_bh(n->next);
net/sched/cls_u32.c
199
n = rcu_dereference_bh(n->next);
net/sched/cls_u32.c
271
ht = rtnl_dereference(ht->next))
net/sched/cls_u32.c
289
n = rtnl_dereference(n->next))
net/sched/cls_u32.c
393
RCU_INIT_POINTER(root_ht->next, tp_c->hlist);
net/sched/cls_u32.c
45
struct tc_u_knode __rcu *next;
net/sched/cls_u32.c
471
kp = &pkp->next, pkp = rtnl_dereference(*kp)) {
net/sched/cls_u32.c
473
RCU_INIT_POINTER(*kp, key->next);
net/sched/cls_u32.c
595
rtnl_dereference(n->next));
net/sched/cls_u32.c
620
hn = &phn->next, phn = rtnl_dereference(*hn)) {
net/sched/cls_u32.c
625
RCU_INIT_POINTER(*hn, ht->next);
net/sched/cls_u32.c
652
RCU_INIT_POINTER(tp_c->hlist, ht->next);
net/sched/cls_u32.c
71
struct tc_u_hnode __rcu *next;
net/sched/cls_u32.c
812
ins = &pins->next, pins = rtnl_dereference(*ins))
net/sched/cls_u32.c
817
RCU_INIT_POINTER(n->next, pins->next);
net/sched/cls_u32.c
832
RCU_INIT_POINTER(new->next, n->next);
net/sched/sch_api.c
139
for (qp = &qdisc_base; (q = *qp) != NULL; qp = &q->next)
net/sched/sch_api.c
164
qops->next = NULL;
net/sched/sch_api.c
183
for (qp = &qdisc_base; (q = *qp) != NULL; qp = &q->next)
net/sched/sch_api.c
187
*qp = q->next;
net/sched/sch_api.c
188
q->next = NULL;
net/sched/sch_api.c
209
for (q = qdisc_base; q; q = q->next) {
net/sched/sch_api.c
366
for (q = qdisc_base; q; q = q->next) {
net/sched/sch_api.c
432
for (rtab = qdisc_rtab_list; rtab; rtab = rtab->next) {
net/sched/sch_api.c
447
rtab->next = qdisc_rtab_list;
net/sched/sch_api.c
465
rtabp = &rtab->next) {
net/sched/sch_api.c
467
*rtabp = rtab->next;
net/sched/sch_api.c
683
struct hlist_node *next;
net/sched/sch_api.c
702
hlist_for_each_entry_safe(cl, next, &ohash[i], hnode) {
net/sched/sch_cake.c
1254
skb_prev = skb_check, skb_check = skb_check->next) {
net/sched/sch_cake.c
1349
if (elig_ack && aggressive && elig_ack->next == skb &&
net/sched/sch_cake.c
1358
elig_ack_prev->next = elig_ack->next;
net/sched/sch_cake.c
1360
flow->head = elig_ack->next;
net/sched/sch_cake.c
1787
u64 next = \
net/sched/sch_cake.c
1792
qdisc_watchdog_schedule_ns(&q->watchdog, next);
net/sched/sch_cake.c
2055
u64 next = min(ktime_to_ns(q->time_next_packet),
net/sched/sch_cake.c
2059
qdisc_watchdog_schedule_ns(&q->watchdog, next);
net/sched/sch_cake.c
2262
u64 next = min(ktime_to_ns(q->time_next_packet),
net/sched/sch_cake.c
2265
qdisc_watchdog_schedule_ns(&q->watchdog, next);
net/sched/sch_cake.c
2271
ktime_t next = \
net/sched/sch_cake.c
2276
ktime_to_ns(next));
net/sched/sch_cake.c
3136
skb = skb->next;
net/sched/sch_cake.c
917
flow->head = skb->next;
net/sched/sch_cake.c
931
flow->tail->next = skb;
net/sched/sch_cake.c
933
skb->next = NULL;
net/sched/sch_drr.c
448
struct hlist_node *next;
net/sched/sch_drr.c
454
hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
net/sched/sch_dualpi2.c
457
struct sk_buff *nskb, *next;
net/sched/sch_dualpi2.c
469
skb_list_walk_safe(nskb, nskb, next) {
net/sched/sch_etf.c
124
ktime_t next;
net/sched/sch_etf.c
131
next = ktime_sub_ns(skb->tstamp, q->delta);
net/sched/sch_etf.c
132
qdisc_watchdog_schedule_ns(&q->watchdog, ktime_to_ns(next));
net/sched/sch_etf.c
216
skb->next = NULL;
net/sched/sch_etf.c
240
skb->next = NULL;
net/sched/sch_etf.c
257
ktime_t now, next;
net/sched/sch_etf.c
281
next = ktime_sub_ns(skb->tstamp, q->delta);
net/sched/sch_etf.c
284
if (ktime_after(now, next))
net/sched/sch_fq.c
189
return f->next == &throttled;
net/sched/sch_fq.c
206
head->last->next = flow;
net/sched/sch_fq.c
210
flow->next = NULL;
net/sched/sch_fq.c
239
f->next = &throttled;
net/sched/sch_fq.c
274
f->next = tofree;
net/sched/sch_fq.c
290
tofree = f->next;
net/sched/sch_fq.c
481
struct sk_buff *next = skb->next;
net/sched/sch_fq.c
483
prefetch(next);
net/sched/sch_fq.c
484
flow->head = next;
net/sched/sch_fq.c
515
flow->tail->next = skb;
net/sched/sch_fq.c
517
skb->next = NULL;
net/sched/sch_fq.c
699
head->first = f->next;
net/sched/sch_fq.c
710
head->first = f->next;
net/sched/sch_fq.c
725
head->first = f->next;
net/sched/sch_fq.c
91
struct fq_flow *next; /* next pointer in RR lists */
net/sched/sch_fq_codel.c
120
flow->head = skb->next;
net/sched/sch_fq_codel.c
132
flow->tail->next = skb;
net/sched/sch_fq_codel.c
134
skb->next = NULL;
net/sched/sch_fq_codel.c
677
skb = skb->next;
net/sched/sch_fq_pie.c
125
flow->tail->next = skb;
net/sched/sch_fq_pie.c
127
skb->next = NULL;
net/sched/sch_fq_pie.c
233
flow->head = skb->next;
net/sched/sch_fq_pie.c
234
skb->next = NULL;
net/sched/sch_fq_pie.c
390
unsigned long next, tupdate;
net/sched/sch_fq_pie.c
409
next = 0;
net/sched/sch_fq_pie.c
412
next = tupdate;
net/sched/sch_fq_pie.c
415
mod_timer(&q->adapt_timer, jiffies + next);
net/sched/sch_generic.c
154
struct sk_buff *next = skb->next;
net/sched/sch_generic.c
169
skb = next;
net/sched/sch_generic.c
195
skb->next = nskb;
net/sched/sch_generic.c
223
skb->next = nskb;
net/sched/sch_generic.c
673
.next = (struct sk_buff *)&noop_qdisc.gso_skb,
net/sched/sch_generic.c
679
.next = (struct sk_buff *)&noop_qdisc.skb_bad_txq,
net/sched/sch_hfsc.c
1501
struct hlist_node *next;
net/sched/sch_hfsc.c
1512
hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
net/sched/sch_hhf.c
184
struct hh_flow_state *flow, *next;
net/sched/sch_hhf.c
190
list_for_each_entry_safe(flow, next, head, flowchain) {
net/sched/sch_hhf.c
333
bucket->head = skb->next;
net/sched/sch_hhf.c
344
bucket->tail->next = skb;
net/sched/sch_hhf.c
346
skb->next = NULL;
net/sched/sch_hhf.c
485
struct hh_flow_state *flow, *next;
net/sched/sch_hhf.c
490
list_for_each_entry_safe(flow, next, head, flowchain) {
net/sched/sch_htb.c
1619
struct hlist_node *next;
net/sched/sch_htb.c
1644
hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
net/sched/sch_htb.c
886
next:
net/sched/sch_htb.c
896
struct htb_class *next;
net/sched/sch_htb.c
903
next = htb_lookup_leaf(hprio, prio);
net/sched/sch_htb.c
906
start = next;
net/sched/sch_htb.c
907
cl = next;
net/sched/sch_htb.c
908
goto next;
net/sched/sch_multiq.c
384
.next = NULL,
net/sched/sch_netem.c
398
q->t_tail->next = nskb;
net/sched/sch_netem.c
505
segs = skb->next;
net/sched/sch_netem.c
529
skb->next = segs;
net/sched/sch_netem.c
630
skb2 = segs->next;
net/sched/sch_netem.c
695
q->t_head = skb->next;
net/sched/sch_netem.c
729
skb->next = NULL;
net/sched/sch_prio.c
407
.next = NULL,
net/sched/sch_qfq.c
1067
struct qfq_group *next = qfq_ffs(q, mask);
net/sched/sch_qfq.c
1068
if (qfq_gt(roundedF, next->F)) {
net/sched/sch_qfq.c
1069
if (qfq_gt(limit, next->F))
net/sched/sch_qfq.c
1070
agg->S = next->F;
net/sched/sch_qfq.c
1369
hlist_del(&agg->next);
net/sched/sch_qfq.c
143
struct hlist_node next; /* Link for the slot list. */
net/sched/sch_qfq.c
1498
struct hlist_node *next;
net/sched/sch_qfq.c
1504
hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
net/sched/sch_qfq.c
769
struct qfq_group *next;
net/sched/sch_qfq.c
772
next = qfq_ffs(q, mask);
net/sched/sch_qfq.c
773
if (qfq_gt(grp->F, next->F))
net/sched/sch_qfq.c
797
struct qfq_group *next;
net/sched/sch_qfq.c
800
next = qfq_ffs(q, mask);
net/sched/sch_qfq.c
801
if (!qfq_gt(next->F, old_F))
net/sched/sch_qfq.c
910
hlist_add_head(&agg->next, &grp->slots[i]);
net/sched/sch_qfq.c
918
struct qfq_aggregate, next);
net/sched/sch_qfq.c
929
hlist_del(&agg->next);
net/sched/sch_sfq.c
203
n = q->dep[qlen].next;
net/sched/sch_sfq.c
205
slot->dep.next = n;
net/sched/sch_sfq.c
208
q->dep[qlen].next = x; /* sfq_dep_head(q, p)->next = x */
net/sched/sch_sfq.c
214
n = q->slots[x].dep.next; \
net/sched/sch_sfq.c
216
sfq_dep_head(q, p)->next = n; \
net/sched/sch_sfq.c
255
skb->prev->next = (struct sk_buff *)slot;
net/sched/sch_sfq.c
256
skb->next = skb->prev = NULL;
net/sched/sch_sfq.c
265
slot->skblist_next = skb->next;
net/sched/sch_sfq.c
266
skb->next->prev = (struct sk_buff *)slot;
net/sched/sch_sfq.c
267
skb->next = skb->prev = NULL;
net/sched/sch_sfq.c
281
skb->next = (struct sk_buff *)slot;
net/sched/sch_sfq.c
282
slot->skblist_prev->next = skb;
net/sched/sch_sfq.c
296
x = q->dep[d].next;
net/sched/sch_sfq.c
311
x = q->tail->next;
net/sched/sch_sfq.c
313
if (slot->next == x)
net/sched/sch_sfq.c
316
q->tail->next = slot->next;
net/sched/sch_sfq.c
364
x = q->dep[0].next; /* get a free slot */
net/sched/sch_sfq.c
444
slot->next = x;
net/sched/sch_sfq.c
446
slot->next = q->tail->next;
net/sched/sch_sfq.c
447
q->tail->next = x;
net/sched/sch_sfq.c
488
a = q->tail->next;
net/sched/sch_sfq.c
504
next_a = slot->next;
net/sched/sch_sfq.c
509
q->tail->next = next_a;
net/sched/sch_sfq.c
564
x = q->dep[0].next; /* get a free slot */
net/sched/sch_sfq.c
588
slot->next = x;
net/sched/sch_sfq.c
590
slot->next = q->tail->next;
net/sched/sch_sfq.c
591
q->tail->next = x;
net/sched/sch_sfq.c
789
q->dep[i].next = i + SFQ_MAX_FLOWS;
net/sched/sch_sfq.c
90
sfq_index next;
net/sched/sch_sfq.c
98
sfq_index next; /* next slot in sfq RR chain */
net/sched/sch_taprio.c
925
struct sched_entry *entry, *next;
net/sched/sch_taprio.c
948
next = list_first_entry(&oper->entries, struct sched_entry,
net/sched/sch_taprio.c
950
end_time = next->end_time;
net/sched/sch_taprio.c
955
next = list_first_entry(&oper->entries, struct sched_entry,
net/sched/sch_taprio.c
960
next = list_next_entry(entry, list);
net/sched/sch_taprio.c
963
end_time = ktime_add_ns(entry->end_time, next->interval);
net/sched/sch_taprio.c
967
if (next->gate_duration[tc] == oper->cycle_time)
net/sched/sch_taprio.c
968
next->gate_close_time[tc] = KTIME_MAX;
net/sched/sch_taprio.c
970
next->gate_close_time[tc] = ktime_add_ns(entry->end_time,
net/sched/sch_taprio.c
971
next->gate_duration[tc]);
net/sched/sch_taprio.c
982
next->end_time = end_time;
net/sched/sch_taprio.c
983
taprio_set_budgets(q, oper, next);
net/sched/sch_taprio.c
986
rcu_assign_pointer(q->current_entry, next);
net/sched/sch_tbf.c
604
.next = NULL,
net/sched/sch_teql.c
207
q->next = NEXT_SLAVE(m->slaves);
net/sched/sch_teql.c
210
q->next = sch;
net/sched/sch_teql.c
64
struct Qdisc *next;
net/sched/sch_teql.c
69
#define NEXT_SLAVE(q) (((struct teql_sched_data *)qdisc_priv(q))->next)
net/sctp/associola.c
504
pos = asoc->peer.transport_addr_list.next;
net/sctp/bind_addr.c
286
goto next;
net/sctp/bind_addr.c
293
next:
net/sctp/bind_addr.c
349
goto next;
net/sctp/bind_addr.c
354
next:
net/sctp/bind_addr.c
553
addr = list_entry(bp->address_list.next,
net/sctp/diag.c
312
goto next;
net/sctp/diag.c
316
goto next;
net/sctp/diag.c
319
goto next;
net/sctp/diag.c
34
laddr = list_entry(asoc->base.bind_addr.address_list.next,
net/sctp/diag.c
341
next:
net/sctp/diag.c
383
goto next;
net/sctp/diag.c
386
goto next;
net/sctp/diag.c
390
goto next;
net/sctp/diag.c
394
goto next;
net/sctp/diag.c
398
goto next;
net/sctp/diag.c
408
next:
net/sctp/inqueue.c
136
if (chunk->skb->next) {
net/sctp/inqueue.c
137
chunk->skb = chunk->skb->next;
net/sctp/objcnt.c
91
.next = sctp_objcnt_seq_next,
net/sctp/offload.c
78
for (skb = segs; skb; skb = skb->next) {
net/sctp/output.c
425
SCTP_OUTPUT_CB(head)->last->next = skb;
net/sctp/output.c
582
chunk = list_entry(packet->chunk_list.next, struct sctp_chunk, list);
net/sctp/proc.c
193
.next = sctp_eps_seq_next,
net/sctp/proc.c
295
.next = sctp_transport_seq_next,
net/sctp/proc.c
370
.next = sctp_transport_seq_next,
net/sctp/sm_statefuns.c
3998
entry = asoc->addip_chunk_list.next;
net/sctp/socket.c
248
asoc = list_entry(sctp_sk(sk)->ep->asocs.next,
net/sctp/socket.c
4952
asoc = list_entry(sctp_sk(sk)->ep->asocs.next,
net/sctp/socket.c
5210
asoc = list_entry(ep->asocs.next,
net/sctp/socket.c
619
p = bp->address_list.next;
net/sctp/socket.c
6332
addr = list_entry(bp->address_list.next,
net/sctp/socket.c
8418
goto next;
net/sctp/socket.c
8420
next:
net/sctp/stream_sched.c
68
ch = list_entry(stream->out_curr->ext->outq.next,
net/sctp/stream_sched.c
71
entry = q->out_chunk_list.next;
net/sctp/stream_sched_fc.c
125
soute = list_entry(stream->fc_list.next, struct sctp_stream_out_ext, fc_list);
net/sctp/stream_sched_fc.c
126
ch = list_entry(soute->outq.next, struct sctp_chunk, stream_list);
net/sctp/stream_sched_prio.c
100
p->next = list_entry(pos, struct sctp_stream_out_ext, prio_list);
net/sctp/stream_sched_prio.c
113
if (prio_head->next == soute)
net/sctp/stream_sched_prio.c
123
prio_head->next = NULL;
net/sctp/stream_sched_prio.c
145
if (prio_head->next) {
net/sctp/stream_sched_prio.c
146
list_add(&soute->prio_list, prio_head->next->prio_list.prev);
net/sctp/stream_sched_prio.c
151
prio_head->next = soute;
net/sctp/stream_sched_prio.c
245
prio = list_entry(stream->prio_list.next,
net/sctp/stream_sched_prio.c
247
soute = prio->next;
net/sctp/stream_sched_prio.c
249
ch = list_entry(soute->outq.next, struct sctp_chunk, stream_list);
net/sctp/stream_sched_prio.c
51
p->next = NULL;
net/sctp/stream_sched_prio.c
97
pos = p->next->prio_list.next;
net/sctp/stream_sched_prio.c
99
pos = pos->next;
net/sctp/stream_sched_rr.c
125
ch = list_entry(soute->outq.next, struct sctp_chunk, stream_list);
net/sctp/stream_sched_rr.c
31
pos = stream->rr_next->rr_list.next;
net/sctp/stream_sched_rr.c
33
pos = pos->next;
net/sctp/ulpqueue.c
327
pos = f_frag->next;
net/sctp/ulpqueue.c
330
for (last = list; list; last = list, list = list->next)
net/sctp/ulpqueue.c
337
last->next = pos;
net/sctp/ulpqueue.c
367
pnext = pos->next;
net/sctp/ulpqueue.c
379
pos->next = pnext;
net/sctp/ulpqueue.c
983
for (last = flist; flist; flist = flist->next) {
net/smc/smc_core.c
525
goto next;
net/smc/smc_core.c
528
next:
net/smc/smc_core.c
620
goto next;
net/smc/smc_core.c
624
next:
net/smc/smc_core.c
647
goto next;
net/smc/smc_core.c
651
next:
net/smc/smc_diag.c
216
goto next;
net/smc/smc_diag.c
220
next:
net/smc/smc_ib.c
617
goto next;
net/smc/smc_ib.c
620
next:
net/smc/smc_ism.c
377
goto next;
net/smc/smc_ism.c
379
goto next;
net/smc/smc_ism.c
382
next:
net/smc/smc_pnet.c
932
lower = lower->next;
net/smc/smc_wr.c
574
lnk->wr_tx_ibs[i].next = NULL;
net/smc/smc_wr.c
595
lnk->wr_tx_v2_ib->next = NULL;
net/smc/smc_wr.c
625
lnk->wr_rx_ibs[i].next = NULL;
net/smc/smc_wr.c
629
lnk->wr_reg.wr.next = NULL;
net/strparser/strparser.c
147
if (WARN_ON(head->next)) {
net/strparser/strparser.c
159
strp->skb_nextp = &head->next;
net/strparser/strparser.c
204
strp->skb_nextp = &skb->next;
net/sunrpc/auth.c
329
cred = list_entry(head->next, struct rpc_cred, cr_lru);
net/sunrpc/auth.c
429
struct rpc_cred *cred, *next;
net/sunrpc/auth.c
433
list_for_each_entry_safe(cred, next, &cred_unused, cr_lru) {
net/sunrpc/cache.c
1020
cq = list_entry(cq->list.next, struct cache_queue, list))
net/sunrpc/cache.c
1071
cq = list_entry(cq->list.next,
net/sunrpc/cache.c
1131
cr = list_entry(dequeued.next, struct cache_request, q.list);
net/sunrpc/cache.c
1403
else if (ch->cache_list.next == NULL) {
net/sunrpc/cache.c
1470
.next = cache_seq_next_rcu,
net/sunrpc/cache.c
443
struct list_head *next;
net/sunrpc/cache.c
451
next = current_detail->others.next;
net/sunrpc/cache.c
453
next = cache_list.next;
net/sunrpc/cache.c
454
if (next == &cache_list) {
net/sunrpc/cache.c
459
current_detail = list_entry(next, struct cache_detail, others);
net/sunrpc/cache.c
694
discard = list_entry(cache_defer_list.next,
net/sunrpc/cache.c
762
dreq = list_entry(pending.next, struct cache_deferred_req, recent);
net/sunrpc/cache.c
784
dreq = list_entry(pending.next, struct cache_deferred_req, recent);
net/sunrpc/cache.c
852
while (rp->q.list.next != &cd->queue &&
net/sunrpc/cache.c
853
list_entry(rp->q.list.next, struct cache_queue, list)
net/sunrpc/cache.c
855
struct list_head *next = rp->q.list.next;
net/sunrpc/cache.c
856
list_move(&rp->q.list, next);
net/sunrpc/cache.c
858
if (rp->q.list.next == &cd->queue) {
net/sunrpc/cache.c
864
rq = container_of(rp->q.list.next, struct cache_request, q.list);
net/sunrpc/cache.c
994
cq = list_entry(cq->list.next, struct cache_queue, list))
net/sunrpc/debugfs.c
61
struct list_head *next = task->tk_task.next;
net/sunrpc/debugfs.c
66
if (next == &clnt->cl_tasks)
net/sunrpc/debugfs.c
68
return list_entry(next, struct rpc_task, tk_task);
net/sunrpc/debugfs.c
84
.next = tasks_next,
net/sunrpc/rpc_pipe.c
299
msg = list_entry(pipe->pipe.next,
net/sunrpc/rpc_pipe.c
72
msg = list_entry(head->next, struct rpc_pipe_msg, list);
net/sunrpc/sched.c
156
task->u.tk_wait.list.next = q;
net/sunrpc/sched.c
183
q = t->u.tk_wait.list.next;
net/sunrpc/stats.c
271
struct rpc_clnt *next = clnt;
net/sunrpc/stats.c
273
_add_rpc_iostats(&stats, &next->cl_metrics[op]);
net/sunrpc/stats.c
274
if (next == next->cl_parent)
net/sunrpc/stats.c
276
next = next->cl_parent;
net/sunrpc/stats.c
277
} while (next);
net/sunrpc/svc_xprt.c
1002
struct list_head *le, *next;
net/sunrpc/svc_xprt.c
1006
list_for_each_safe(le, next, &serv->sv_tempsocks) {
net/sunrpc/svc_xprt.c
1017
le = to_be_closed.next;
net/sunrpc/svc_xprt.c
1297
dr = list_entry(xprt->xpt_deferred.next,
net/sunrpc/svc_xprt.c
1513
.next = svc_pool_stats_next,
net/sunrpc/svc_xprt.c
960
struct list_head *le, *next;
net/sunrpc/svc_xprt.c
971
list_for_each_safe(le, next, &serv->sv_tempsocks) {
net/sunrpc/xprt.c
1409
struct rpc_rqst *next = list_first_entry(&req->rq_xmit2,
net/sunrpc/xprt.c
1412
list_add_tail(&next->rq_xmit, &next->rq_xprt->xmit_queue);
net/sunrpc/xprt.c
1622
struct rpc_rqst *next, *req = task->tk_rqstp;
net/sunrpc/xprt.c
1628
next = list_first_entry_or_null(&xprt->xmit_queue,
net/sunrpc/xprt.c
1630
if (!next)
net/sunrpc/xprt.c
1632
xprt_pin_rqst(next);
net/sunrpc/xprt.c
1634
status = xprt_request_transmit(next, task);
net/sunrpc/xprt.c
1635
if (status == -EBADMSG && next != req)
net/sunrpc/xprt.c
1638
xprt_unpin_rqst(next);
net/sunrpc/xprt.c
1738
req = list_entry(xprt->free.next, struct rpc_rqst, rq_list);
net/sunrpc/xprtrdma/frwr_ops.c
398
mr->mr_regwr.wr.next = post_wr;
net/sunrpc/xprtrdma/frwr_ops.c
517
last->next = NULL;
net/sunrpc/xprtrdma/frwr_ops.c
528
prev = &last->next;
net/sunrpc/xprtrdma/frwr_ops.c
620
last->next = NULL;
net/sunrpc/xprtrdma/frwr_ops.c
631
prev = &last->next;
net/sunrpc/xprtrdma/frwr_ops.c
690
mr->mr_regwr.wr.next = NULL;
net/sunrpc/xprtrdma/svc_rdma_pcl.c
270
struct svc_rdma_chunk *chunk, *next;
net/sunrpc/xprtrdma/svc_rdma_pcl.c
286
while ((next = pcl_next_chunk(pcl, chunk))) {
net/sunrpc/xprtrdma/svc_rdma_pcl.c
287
if (!next->ch_payload_length)
net/sunrpc/xprtrdma/svc_rdma_pcl.c
291
ret = pcl_process_region(xdr, start, next->ch_position - start,
net/sunrpc/xprtrdma/svc_rdma_pcl.c
296
chunk = next;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
147
ctxt->rc_recv_wr.next = NULL;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
273
ctxt->rc_recv_wr.next = recv_chain;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
290
bad_wr = bad_wr->next;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1028
struct svc_rdma_chunk *chunk, *next;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1048
next = pcl_next_chunk(pcl, chunk);
net/sunrpc/xprtrdma/svc_rdma_rw.c
1049
if (!next)
net/sunrpc/xprtrdma/svc_rdma_rw.c
1053
length = next->ch_position - head->rc_readbytes;
net/sunrpc/xprtrdma/svc_rdma_rw.c
212
ctxt->rw_node.next = first;
net/sunrpc/xprtrdma/svc_rdma_rw.c
906
struct svc_rdma_chunk *chunk, *next;
net/sunrpc/xprtrdma/svc_rdma_rw.c
922
next = pcl_next_chunk(pcl, chunk);
net/sunrpc/xprtrdma/svc_rdma_rw.c
923
if (!next)
net/sunrpc/xprtrdma/svc_rdma_rw.c
927
length = next->ch_position - head->rc_readbytes;
net/sunrpc/xprtrdma/svc_rdma_sendto.c
147
ctxt->sc_send_wr.next = NULL;
net/sunrpc/xprtrdma/svc_rdma_transport.c
424
newxprt = list_entry(listen_rdma->sc_accept_q.next,
net/sunrpc/xprtrdma/verbs.c
1382
rep->rr_recv_wr.next = wr;
net/sunrpc/xprtrdma/verbs.c
1398
wr = wr->next;
net/sunrpc/xprtrdma/verbs.c
972
rep->rr_recv_wr.next = NULL;
net/tipc/crypto.c
1360
WARN_ON(skb->next);
net/tipc/crypto.c
1361
skb->next = skb_clone(skb, GFP_ATOMIC);
net/tipc/crypto.c
1362
if (unlikely(!skb->next))
net/tipc/crypto.c
1912
(*skb)->next, skb_cb->flags);
net/tipc/crypto.c
1918
if (err == -EBADMSG && (*skb)->next)
net/tipc/crypto.c
1919
tipc_rcv(net, (*skb)->next, b);
net/tipc/crypto.c
1923
if (likely((*skb)->next)) {
net/tipc/crypto.c
1924
kfree_skb((*skb)->next);
net/tipc/crypto.c
1925
(*skb)->next = NULL;
net/tipc/crypto.c
2034
for (p = tn->node_list.next; p != &tn->node_list; p = p->next) {
net/tipc/crypto.c
2061
for (p = tn->node_list.next; p != &tn->node_list; p = p->next) {
net/tipc/msg.c
139
frag->next = NULL;
net/tipc/msg.c
171
tail->next = frag;
net/tipc/msg.c
747
skb->next = NULL;
net/tipc/msg.c
76
skb->next = NULL;
net/tipc/msg.c
787
frag->next = NULL;
net/tls/tls_strp.c
172
goto next;
net/tls/tls_strp.c
184
next:
net/tls/tls_strp.c
185
iter = iter->next;
net/tls/tls_strp.c
312
last->next = nskb;
net/tls/tls_strp.c
450
skb = skb->next;
net/unix/af_unix.c
3579
.next = unix_seq_next,
net/unix/af_unix.c
3775
.next = bpf_iter_unix_seq_next,
net/unix/diag.c
197
goto next;
net/unix/diag.c
200
goto next;
net/unix/diag.c
204
goto next;
net/unix/diag.c
213
next:
net/vmw_vsock/af_vsock.c
698
vconnected = list_entry(vlistener->accept_queue.next,
net/wireless/scan.c
2599
const struct element *next;
net/wireless/scan.c
2607
next = (void *)(elem->data + elem->datalen);
net/wireless/scan.c
2634
for (elem = next;
net/wireless/scan.c
2637
elem = next) {
net/wireless/scan.c
2639
next = (void *)(elem->data + elem->datalen);
net/wireless/scan.c
3645
const u8 *pos, *end, *next;
net/wireless/scan.c
3659
next = pos + 2 + pos[1];
net/wireless/scan.c
3660
while (next + 2 + next[1] - pos < IW_GENERIC_IE_MAX)
net/wireless/scan.c
3661
next = next + 2 + next[1];
net/wireless/scan.c
3665
iwe.u.data.length = next - pos;
net/wireless/scan.c
3671
pos = next;
net/wireless/sme.c
1000
next = ((u8 *)ev) + sizeof(*ev);
net/wireless/sme.c
1002
ev->cr.ap_mld_addr = next;
net/wireless/sme.c
1005
next += ETH_ALEN;
net/wireless/sme.c
1008
ev->cr.req_ie = next;
net/wireless/sme.c
1012
next += params->req_ie_len;
net/wireless/sme.c
1015
ev->cr.resp_ie = next;
net/wireless/sme.c
1019
next += params->resp_ie_len;
net/wireless/sme.c
1022
ev->cr.fils.kek = next;
net/wireless/sme.c
1026
next += params->fils.kek_len;
net/wireless/sme.c
1029
ev->cr.fils.pmk = next;
net/wireless/sme.c
1033
next += params->fils.pmk_len;
net/wireless/sme.c
1036
ev->cr.fils.pmkid = next;
net/wireless/sme.c
1039
next += WLAN_PMKID_LEN;
net/wireless/sme.c
1053
ev->cr.links[link].addr = next;
net/wireless/sme.c
1057
next += ETH_ALEN;
net/wireless/sme.c
1060
ev->cr.links[link].bssid = next;
net/wireless/sme.c
1064
next += ETH_ALEN;
net/wireless/sme.c
1173
u8 *next;
net/wireless/sme.c
1211
next = ((u8 *)ev) + sizeof(*ev);
net/wireless/sme.c
1213
ev->rm.req_ie = next;
net/wireless/sme.c
1216
next += info->req_ie_len;
net/wireless/sme.c
1219
ev->rm.resp_ie = next;
net/wireless/sme.c
1223
next += info->resp_ie_len;
net/wireless/sme.c
1226
ev->rm.fils.kek = next;
net/wireless/sme.c
1230
next += info->fils.kek_len;
net/wireless/sme.c
1233
ev->rm.fils.pmk = next;
net/wireless/sme.c
1237
next += info->fils.pmk_len;
net/wireless/sme.c
1240
ev->rm.fils.pmkid = next;
net/wireless/sme.c
1243
next += WLAN_PMKID_LEN;
net/wireless/sme.c
1249
ev->rm.ap_mld_addr = next;
net/wireless/sme.c
1252
next += ETH_ALEN;
net/wireless/sme.c
1259
ev->rm.links[link].addr = next;
net/wireless/sme.c
1263
next += ETH_ALEN;
net/wireless/sme.c
1267
ev->rm.links[link].bssid = next;
net/wireless/sme.c
1271
next += ETH_ALEN;
net/wireless/sme.c
976
u8 *next;
net/wireless/wext-proc.c
124
.next = wireless_dev_seq_next,
net/x25/x25_link.c
301
if (nb->node.next) {
net/x25/x25_proc.c
149
.next = x25_seq_route_next,
net/x25/x25_proc.c
156
.next = x25_seq_socket_next,
net/x25/x25_proc.c
163
.next = x25_seq_forward_next,
net/x25/x25_route.c
68
if (rt->node.next) {
net/xdp/xskmap.c
108
u32 *next = next_key;
net/xdp/xskmap.c
111
*next = 0;
net/xdp/xskmap.c
117
*next = index + 1;
net/xfrm/xfrm_device.c
176
if (!skb->next) {
net/xfrm/xfrm_device.c
208
skb2->next = nskb;
net/xfrm/xfrm_device.c
211
skb2->next = nskb;
net/xfrm/xfrm_device.c
218
pskb->next = nskb;
net/xfrm/xfrm_interface_core.c
147
for (xi = rcu_dereference(start); xi; xi = rcu_dereference(xi->next))
net/xfrm/xfrm_interface_core.c
219
rcu_assign_pointer(xi->next , rtnl_dereference(*xip));
net/xfrm/xfrm_interface_core.c
230
xip = &iter->next) {
net/xfrm/xfrm_interface_core.c
232
rcu_assign_pointer(*xip, xi->next);
net/xfrm/xfrm_interface_core.c
275
xip = &xi->next)
net/xfrm/xfrm_interface_core.c
961
xip = &xi->next)
net/xfrm/xfrm_iptfs.c
1186
list_for_each_entry_safe(skb, next, sublist, list) {
net/xfrm/xfrm_iptfs.c
1667
struct sk_buff *skb, *next;
net/xfrm/xfrm_iptfs.c
1695
list_for_each_entry_safe(skb, next, &list, list) {
net/xfrm/xfrm_iptfs.c
1718
struct sk_buff *next;
net/xfrm/xfrm_iptfs.c
1736
list_for_each_entry_safe(skb, next, &list, list) {
net/xfrm/xfrm_iptfs.c
1741
list_for_each_entry_safe(skb, next, &freelist, list) {
net/xfrm/xfrm_iptfs.c
2143
nextp = &(*nextp)->next;
net/xfrm/xfrm_iptfs.c
2244
nextp = iptfs_rehome_fraglist(&(*nextp)->next, *nextp);
net/xfrm/xfrm_iptfs.c
2246
nextp = &(*nextp)->next;
net/xfrm/xfrm_iptfs.c
2304
nextp = &skb2->next;
net/xfrm/xfrm_iptfs.c
949
struct sk_buff *defer, *first_skb, *next, *skb;
net/xfrm/xfrm_policy.c
358
time64_t next = TIME64_MAX;
net/xfrm/xfrm_policy.c
374
if (tmo < next)
net/xfrm/xfrm_policy.c
375
next = tmo;
net/xfrm/xfrm_policy.c
382
if (tmo < next)
net/xfrm/xfrm_policy.c
383
next = tmo;
net/xfrm/xfrm_policy.c
392
if (tmo < next)
net/xfrm/xfrm_policy.c
393
next = tmo;
net/xfrm/xfrm_policy.c
402
if (tmo < next)
net/xfrm/xfrm_policy.c
403
next = tmo;
net/xfrm/xfrm_policy.c
408
if (next != TIME64_MAX &&
net/xfrm/xfrm_policy.c
409
!mod_timer(&xp->timer, jiffies + make_jiffies(next)))
net/xfrm/xfrm_state.c
1590
xfrm_hash_grow_check(net, x->bydst.next != NULL);
net/xfrm/xfrm_state.c
1759
xfrm_hash_grow_check(net, x->bydst.next != NULL);
net/xfrm/xfrm_state.c
1879
xfrm_hash_grow_check(net, x->bydst.next != NULL);
net/xfrm/xfrm_state.c
2598
goto next;
net/xfrm/xfrm_state.c
2615
next:
net/xfrm/xfrm_state.c
644
time64_t next = TIME64_MAX;
net/xfrm/xfrm_state.c
669
if (tmo < next)
net/xfrm/xfrm_state.c
670
next = tmo;
net/xfrm/xfrm_state.c
677
if (tmo < next)
net/xfrm/xfrm_state.c
678
next = tmo;
net/xfrm/xfrm_state.c
688
} else if (tmo < next) {
net/xfrm/xfrm_state.c
689
next = tmo;
net/xfrm/xfrm_state.c
699
else if (tmo < next)
net/xfrm/xfrm_state.c
700
next = tmo;
net/xfrm/xfrm_state.c
707
if (next != TIME64_MAX) {
net/xfrm/xfrm_state.c
708
hrtimer_forward_now(&x->mtimer, ktime_set(next, 0));
samples/bpf/test_lru_dist.c
36
struct list_head *next, *prev;
samples/bpf/test_lru_dist.c
41
list->next = list;
samples/bpf/test_lru_dist.c
47
struct list_head *next)
samples/bpf/test_lru_dist.c
49
next->prev = new;
samples/bpf/test_lru_dist.c
50
new->next = next;
samples/bpf/test_lru_dist.c
52
prev->next = new;
samples/bpf/test_lru_dist.c
57
__list_add(new, head, head->next);
samples/bpf/test_lru_dist.c
60
static inline void __list_del(struct list_head *prev, struct list_head *next)
samples/bpf/test_lru_dist.c
62
next->prev = prev;
samples/bpf/test_lru_dist.c
63
prev->next = next;
samples/bpf/test_lru_dist.c
68
__list_del(entry->prev, entry->next);
samples/trace_events/trace_custom_sched.h
29
struct task_struct *next,
samples/trace_events/trace_custom_sched.h
32
TP_ARGS(preempt, prev, next, prev_state),
samples/trace_events/trace_custom_sched.h
50
__entry->next_pid = next->pid;
samples/trace_events/trace_custom_sched.h
51
__entry->next_prio = next->prio;
samples/vfio-mdev/mbochs.c
1292
list_for_each_entry_safe(dmabuf, tmp, &mdev_state->dmabufs, next) {
samples/vfio-mdev/mbochs.c
1293
list_del(&dmabuf->next);
samples/vfio-mdev/mbochs.c
164
struct list_head next;
samples/vfio-mdev/mbochs.c
957
list_add(&dmabuf->next, &mdev_state->dmabufs);
samples/vfio-mdev/mbochs.c
979
list_for_each_entry(dmabuf, &mdev_state->dmabufs, next)
samples/vfio-mdev/mbochs.c
993
list_for_each_entry(dmabuf, &mdev_state->dmabufs, next)
samples/vfio-mdev/mtty.c
163
struct list_head next;
samples/vfs/test-list-all-mounts.c
116
goto next;
samples/vfs/test-list-all-mounts.c
94
next:
scripts/asn1_compiler.c
1098
ppaction = &(*ppaction)->next
scripts/asn1_compiler.c
1107
action->next = *ppaction;
scripts/asn1_compiler.c
1113
action->next = NULL;
scripts/asn1_compiler.c
1182
child_p = &element->next;
scripts/asn1_compiler.c
1246
for (c = e->children; c; c = c->next)
scripts/asn1_compiler.c
1328
for (action = action_list; action; action = action->next) {
scripts/asn1_compiler.c
1338
for (action = action_list; action; action = action->next)
scripts/asn1_compiler.c
1347
for (action = action_list; action; action = action->next)
scripts/asn1_compiler.c
1406
for (ce = e->children; ce; ce = ce->next)
scripts/asn1_compiler.c
1549
for (ec = e->children; ec; ec = ec->next)
scripts/asn1_compiler.c
1597
for (ec = e->children; ec; ec = ec->next)
scripts/asn1_compiler.c
292
struct action *next;
scripts/asn1_compiler.c
687
struct element *next;
scripts/basic/fixdep.c
111
struct item *next;
scripts/basic/fixdep.c
142
aux->next = hashtab[hash % HASHSZ];
scripts/basic/fixdep.c
155
for (aux = hashtab[hash % HASHSZ]; aux; aux = aux->next) {
scripts/dtc/checks.c
107
while (pos->next) {
scripts/dtc/checks.c
108
pos = pos->next;
scripts/dtc/checks.c
296
for (prop2 = prop->next; prop2; prop2 = prop2->next) {
scripts/dtc/checks.c
582
for (pp = &node->proplist; *pp; pp = &((*pp)->next))
scripts/dtc/checks.c
598
*pp = prop->next;
scripts/dtc/data.c
124
m = m->next;
scripts/dtc/data.c
136
mp = &((*mp)->next);
scripts/dtc/data.c
14
nm = m->next;
scripts/dtc/data.c
257
struct marker *next = m->next;
scripts/dtc/data.c
272
m->next = alloc_marker(marker->offset + offset, marker->type,
scripts/dtc/data.c
274
m = m->next;
scripts/dtc/data.c
276
m->next = next;
scripts/dtc/data.c
290
m->next = NULL;
scripts/dtc/dtc.h
131
struct marker *next;
scripts/dtc/dtc.h
144
for (; (m); (m) = (m)->next)
scripts/dtc/dtc.h
159
struct marker *next = next_type_marker(m->next);
scripts/dtc/dtc.h
161
if (next)
scripts/dtc/dtc.h
162
return next->offset - m->offset;
scripts/dtc/dtc.h
202
struct label *next;
scripts/dtc/dtc.h
214
struct property *next;
scripts/dtc/dtc.h
243
for ((l) = (l0); (l); (l) = (l)->next)
scripts/dtc/dtc.h
250
for ((p) = (n)->proplist; (p); (p) = (p)->next)
scripts/dtc/dtc.h
314
struct reserve_info *next;
scripts/dtc/flattree.c
301
for (re = reservelist; re; re = re->next) {
scripts/dtc/flattree.c
517
for (re = dti->reservelist; re; re = re->next) {
scripts/dtc/livetree.c
159
new_node->proplist = new_prop->next;
scripts/dtc/livetree.c
160
new_prop->next = NULL;
scripts/dtc/livetree.c
27
new->next = *labels;
scripts/dtc/livetree.c
271
prop->next = NULL;
scripts/dtc/livetree.c
275
p = &((*p)->next);
scripts/dtc/livetree.c
289
prop = prop->next;
scripts/dtc/livetree.c
427
assert(first->next == NULL);
scripts/dtc/livetree.c
429
first->next = list;
scripts/dtc/livetree.c
438
new->next = NULL;
scripts/dtc/livetree.c
443
for (last = list; last->next; last = last->next)
scripts/dtc/livetree.c
446
last->next = new;
scripts/dtc/livetree.c
67
assert(first->next == NULL);
scripts/dtc/livetree.c
69
first->next = list;
scripts/dtc/livetree.c
760
ri = ri->next)
scripts/dtc/livetree.c
77
struct property *next;
scripts/dtc/livetree.c
770
ri = ri->next)
scripts/dtc/livetree.c
777
tbl[i]->next = tbl[i+1];
scripts/dtc/livetree.c
778
tbl[n-1]->next = NULL;
scripts/dtc/livetree.c
80
next = p->next;
scripts/dtc/livetree.c
81
p->next = head;
scripts/dtc/livetree.c
813
tbl[i]->next = tbl[i+1];
scripts/dtc/livetree.c
814
tbl[n-1]->next = NULL;
scripts/dtc/livetree.c
83
p = next;
scripts/dtc/srcpos.c
162
for (node = search_path_head; !*fp && node; node = node->next)
scripts/dtc/srcpos.c
17
struct search_path *next; /* next node in list, NULL for end */
scripts/dtc/srcpos.c
245
node->next = NULL;
scripts/dtc/srcpos.c
253
search_path_tail = &node->next;
scripts/dtc/srcpos.c
287
assert(pos->next == NULL);
scripts/dtc/srcpos.c
305
for (p = pos; p->next != NULL; p = p->next);
scripts/dtc/srcpos.c
306
p->next = newtail;
scripts/dtc/srcpos.c
315
p_next = pos->next;
scripts/dtc/srcpos.c
386
if (pos->next != NULL) {
scripts/dtc/srcpos.c
387
rest = srcpos_string_comment(pos->next, first_line, level);
scripts/dtc/srcpos.h
63
struct srcpos *next;
scripts/dtc/srcpos.h
83
(Current).next = NULL; \
scripts/dtc/treesource.c
155
mi = &(*mi)->next;
scripts/dtc/treesource.c
160
mi = &(*mi)->next;
scripts/dtc/treesource.c
170
nm->next = *mi;
scripts/dtc/treesource.c
173
return &nm->next;
scripts/dtc/treesource.c
254
for (m = &prop->val.markers; *m; m = &((*m)->next)) {
scripts/dtc/treesource.c
303
size_t chunk_len = (m->next ? m->next->offset : len) - m->offset;
scripts/dtc/treesource.c
426
for (re = dti->reservelist; re; re = re->next) {
scripts/gcc-plugins/stackleak_plugin.c
270
rtx_insn *insn, *next;
scripts/gcc-plugins/stackleak_plugin.c
283
for (insn = get_insns(); insn; insn = next) {
scripts/gcc-plugins/stackleak_plugin.c
286
next = NEXT_INSN(insn);
scripts/gcc-plugins/stackleak_plugin.c
324
if (GET_CODE(next) == NOTE &&
scripts/gcc-plugins/stackleak_plugin.c
325
NOTE_KIND(next) == NOTE_INSN_CALL_ARG_LOCATION) {
scripts/gcc-plugins/stackleak_plugin.c
326
insn = next;
scripts/gcc-plugins/stackleak_plugin.c
327
next = NEXT_INSN(insn);
scripts/gcc-plugins/stackleak_plugin.c
337
rtx_insn *insn, *next;
scripts/gcc-plugins/stackleak_plugin.c
353
for (insn = get_insns(); insn; insn = next) {
scripts/gcc-plugins/stackleak_plugin.c
356
next = NEXT_INSN(insn);
scripts/genksyms/genksyms.c
178
(defn = defn->next) && defn->tag == SYM_NORMAL &&
scripts/genksyms/genksyms.c
180
(defn = defn->next) && defn->tag == SYM_NORMAL &&
scripts/genksyms/genksyms.c
336
struct string_list *next = s->next;
scripts/genksyms/genksyms.c
338
s = next;
scripts/genksyms/genksyms.c
349
newnode->next = NULL;
scripts/genksyms/genksyms.c
362
for (n2 = n; n2->next; n2 = n2->next)
scripts/genksyms/genksyms.c
364
n2->next = start;
scripts/genksyms/genksyms.c
390
for (start = start->next; start != end; start = start->next) {
scripts/genksyms/genksyms.c
391
n->next = copy_node(start);
scripts/genksyms/genksyms.c
392
n = n->next;
scripts/genksyms/genksyms.c
394
n->next = NULL;
scripts/genksyms/genksyms.c
403
a = a->next;
scripts/genksyms/genksyms.c
404
b = b->next;
scripts/genksyms/genksyms.c
485
def->next = defn;
scripts/genksyms/genksyms.c
517
while ((tmp = tmp->next) != NULL)
scripts/genksyms/genksyms.c
525
while ((list = list->next) != NULL)
scripts/genksyms/genksyms.c
545
while ((tmp = tmp->next) != NULL)
scripts/genksyms/genksyms.c
553
while ((list = list->next) != NULL)
scripts/genksyms/genksyms.h
30
struct string_list *next;
scripts/genksyms/parse.y
146
struct string_list *decl = (*$3)->next;
scripts/genksyms/parse.y
147
(*$3)->next = NULL;
scripts/genksyms/parse.y
31
*p = node->next;
scripts/genksyms/parse.y
440
{ remove_list($2, &(*$1)->next); $$ = $2; }
scripts/genksyms/parse.y
58
r->next = (*keyw)->next; *body = r; (*keyw)->next = NULL;
scripts/include/list.h
106
static inline void __list_del(struct list_head *prev, struct list_head *next)
scripts/include/list.h
108
next->prev = prev;
scripts/include/list.h
109
prev->next = next;
scripts/include/list.h
114
__list_del(entry->prev, entry->next);
scripts/include/list.h
126
entry->next = LIST_POISON1;
scripts/include/list.h
140
new->next = old->next;
scripts/include/list.h
141
new->next->prev = new;
scripts/include/list.h
143
new->prev->next = new;
scripts/include/list.h
200
return list->next == head;
scripts/include/list.h
219
return head->next == head;
scripts/include/list.h
240
list_entry((ptr)->next, type, member)
scripts/include/list.h
259
list_entry((pos)->member.next, typeof(*(pos)), member)
scripts/include/list.h
324
h->next = NULL;
scripts/include/list.h
343
struct hlist_node *next = n->next;
scripts/include/list.h
346
*pprev = next;
scripts/include/list.h
347
if (next)
scripts/include/list.h
348
next->pprev = pprev;
scripts/include/list.h
361
n->next = LIST_POISON1;
scripts/include/list.h
391
n->next = first;
scripts/include/list.h
393
first->pprev = &n->next;
scripts/include/list.h
414
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
scripts/include/list.h
425
pos && ({ n = pos->member.next; 1; }); \
scripts/include/list.h
53
list->next = list;
scripts/include/list.h
65
struct list_head *next)
scripts/include/list.h
67
next->prev = new;
scripts/include/list.h
68
new->next = next;
scripts/include/list.h
70
prev->next = new;
scripts/include/list.h
83
__list_add(new, head, head->next);
scripts/include/list_types.h
14
struct hlist_node *next, **pprev;
scripts/include/list_types.h
6
struct list_head *next, *prev;
scripts/kconfig/conf.c
436
for (child = menu->list; child; child = child->next) {
scripts/kconfig/conf.c
492
for (child = menu->list; child; child = child->next) {
scripts/kconfig/conf.c
568
for (child = menu->list; child; child = child->next)
scripts/kconfig/conf.c
603
for (child = menu->list; child; child = child->next)
scripts/kconfig/confdata.c
852
goto next;
scripts/kconfig/confdata.c
863
goto next;
scripts/kconfig/confdata.c
872
next:
scripts/kconfig/confdata.c
885
if (menu->next) {
scripts/kconfig/confdata.c
886
menu = menu->next;
scripts/kconfig/expr.h
189
struct property *next; /* next property - null if last */
scripts/kconfig/expr.h
202
for (st = sym->prop; st; st = st->next) \
scripts/kconfig/expr.h
206
for (st = sym->prop; st; st = st->next) \
scripts/kconfig/expr.h
230
struct menu *next;
scripts/kconfig/gconf.c
181
for (struct menu *child = menu->list; child; child = child->next) {
scripts/kconfig/gconf.c
897
for (child = menu->list; child; child = child->next) {
scripts/kconfig/gconf.c
942
for (child = menu->list; child; child = child->next) {
scripts/kconfig/lxdialog/dialog.h
143
struct dialog_list *next;
scripts/kconfig/lxdialog/dialog.h
158
item_cur && (item_cur != &item_nil); item_cur = item_cur->next)
scripts/kconfig/lxdialog/dialog.h
75
struct subtitle_list *next;
scripts/kconfig/lxdialog/util.c
231
for (pos = dlg.subtitles; pos != NULL; pos = pos->next) {
scripts/kconfig/lxdialog/util.c
243
for (pos = dlg.subtitles; pos != NULL; pos = pos->next) {
scripts/kconfig/lxdialog/util.c
555
struct dialog_list *p, *next;
scripts/kconfig/lxdialog/util.c
557
for (p = item_head; p; p = next) {
scripts/kconfig/lxdialog/util.c
558
next = p->next;
scripts/kconfig/lxdialog/util.c
571
item_cur->next = p;
scripts/kconfig/lxdialog/util.c
633
for (p = item_head; p; p = p->next)
scripts/kconfig/lxdialog/util.c
651
for (p = item_head; p; p = p->next) {
scripts/kconfig/mconf.c
321
tmp = pos->next;
scripts/kconfig/mconf.c
329
pos->next = xcalloc(1, sizeof(*pos));
scripts/kconfig/mconf.c
330
pos = pos->next;
scripts/kconfig/mconf.c
346
tmp = pos->next;
scripts/kconfig/mconf.c
523
for (child = menu->list; child; child = child->next) {
scripts/kconfig/mconf.c
594
for (child = menu->list; child; child = child->next)
scripts/kconfig/mconf.c
612
for (child = menu->list; child; child = child->next) {
scripts/kconfig/menu.c
181
propp = &(*propp)->next)
scripts/kconfig/menu.c
249
for (prop = sym->prop; prop; prop = prop->next) {
scripts/kconfig/menu.c
320
for (menu = parent->list; menu; menu = menu->next) {
scripts/kconfig/menu.c
33
while (menu != root && !menu->next)
scripts/kconfig/menu.c
345
for (; prop; prop = prop->next) {
scripts/kconfig/menu.c
39
return menu->next;
scripts/kconfig/menu.c
395
for (menu = parent->list; menu; menu = menu->next)
scripts/kconfig/menu.c
426
for (menu = parent->next; menu; menu = menu->next) {
scripts/kconfig/menu.c
433
goto next;
scripts/kconfig/menu.c
453
next:
scripts/kconfig/menu.c
459
parent->list = parent->next;
scripts/kconfig/menu.c
460
parent->next = last_menu->next;
scripts/kconfig/menu.c
461
last_menu->next = NULL;
scripts/kconfig/menu.c
466
for (menu = parent->list; menu; menu = menu->next) {
scripts/kconfig/menu.c
496
for (last_menu = menu->list; ; last_menu = last_menu->next) {
scripts/kconfig/menu.c
498
if (!last_menu->next)
scripts/kconfig/menu.c
501
last_menu->next = menu->next;
scripts/kconfig/menu.c
502
menu->next = menu->list;
scripts/kconfig/menu.c
538
for (child = menu->list; child; child = child->next) {
scripts/kconfig/menu.c
572
for (child = menu->list; child; child = child->next)
scripts/kconfig/menu.c
81
last_entry_ptr = &menu->next;
scripts/kconfig/menu.c
852
if (menu->next)
scripts/kconfig/menu.c
860
while (menu && !menu->next) {
scripts/kconfig/menu.c
867
menu = menu->next;
scripts/kconfig/menu.c
868
if (menu->next)
scripts/kconfig/menu.c
96
last_entry_ptr = &current_menu->next;
scripts/kconfig/nconf.c
1251
for (i = 0, child = menu->list; child; child = child->next) {
scripts/kconfig/nconf.c
826
for (child = menu->list; child; child = child->next) {
scripts/kconfig/nconf.c
901
for (child = menu->list; child; child = child->next)
scripts/kconfig/parser.y
517
for (prop = menu->sym->prop; prop; prop = prop->next) {
scripts/kconfig/parser.y
541
for (prop = menu->sym->prop; prop; prop = prop->next) {
scripts/kconfig/parser.y
704
for (prop = sym->prop; prop; prop = prop->next) {
scripts/kconfig/parser.y
793
else if (menu->next)
scripts/kconfig/parser.y
794
menu = menu->next;
scripts/kconfig/parser.y
798
if (menu->next) {
scripts/kconfig/parser.y
799
menu = menu->next;
scripts/kconfig/qconf.cc
1076
for (struct property *prop = sym->prop; prop; prop = prop->next) {
scripts/kconfig/qconf.cc
1804
for (child = menu->list; child; child = child->next) {
scripts/kconfig/qconf.cc
480
ConfigList *list = it.next();
scripts/kconfig/qconf.cc
491
ConfigList *list = it.next();
scripts/kconfig/qconf.cc
625
for (child = menu->list; child; child = child->next) {
scripts/kconfig/qconf.cc
688
for (child = menu->list; child; child = child->next) {
scripts/kconfig/symbol.c
1083
struct dep_stack *prev, *next;
scripts/kconfig/symbol.c
1093
check_top->next = stack;
scripts/kconfig/symbol.c
1103
check_top->next = NULL;
scripts/kconfig/symbol.c
1133
for (; stack; stack = stack->next) {
scripts/kconfig/symbol.c
1135
next_sym = stack->next ? stack->next->sym : last_sym;
scripts/kconfig/symbol.c
1239
for (prop = sym->prop; prop; prop = prop->next) {
scripts/mod/modpost.c
157
char *orig = *stringp, *next;
scripts/mod/modpost.c
164
next = strchr(orig, '\n');
scripts/mod/modpost.c
165
if (next)
scripts/mod/modpost.c
166
*next++ = '\0';
scripts/mod/modpost.c
168
*stringp = next;
scripts/mod/modpost.c
2079
struct module_alias *alias, *next;
scripts/mod/modpost.c
2091
list_for_each_entry_safe(alias, next, &mod->aliases, node) {
scripts/mod/modpost.c
2107
struct module_alias *alias, *next;
scripts/mod/modpost.c
2118
list_for_each_entry_safe(alias, next, &mod->aliases, node) {
security/apparmor/apparmorfs.c
2157
struct aa_ns *parent, *next;
security/apparmor/apparmorfs.c
2165
next = list_first_entry(&ns->sub_ns, typeof(*ns), base.list);
security/apparmor/apparmorfs.c
2166
mutex_lock_nested(&next->lock, next->level);
security/apparmor/apparmorfs.c
2167
return next;
security/apparmor/apparmorfs.c
2174
next = list_next_entry(ns, base.list);
security/apparmor/apparmorfs.c
2175
if (!list_entry_is_head(next, &parent->sub_ns, base.list)) {
security/apparmor/apparmorfs.c
2176
mutex_lock_nested(&next->lock, next->level);
security/apparmor/apparmorfs.c
2177
return next;
security/apparmor/apparmorfs.c
2259
struct aa_profile *next = __next_profile(profile);
security/apparmor/apparmorfs.c
2260
if (next)
security/apparmor/apparmorfs.c
2261
return next;
security/apparmor/apparmorfs.c
2354
.next = p_next,
security/apparmor/domain.c
145
goto next;
security/apparmor/domain.c
152
next:
security/apparmor/domain.c
209
goto next;
security/apparmor/domain.c
215
next:
security/apparmor/domain.c
524
const char *next;
security/apparmor/domain.c
532
for (next = rules->file->trans.table[index].strs; next;
security/apparmor/domain.c
533
next = next_name(xtype, next)) {
security/apparmor/domain.c
534
const char *lookup = (*next == '&') ? next + 1 : next;
security/apparmor/domain.c
535
*name = next;
security/apparmor/label.c
1050
struct aa_profile *next;
security/apparmor/label.c
1063
label_for_each_in_merge(i, a, b, next) {
security/apparmor/label.c
1064
AA_BUG(!next);
security/apparmor/label.c
1065
if (profile_is_stale(next)) {
security/apparmor/label.c
1066
new->vec[k] = aa_get_newest_profile(next);
security/apparmor/label.c
1069
if (next->label.proxy != new->vec[k]->label.proxy)
security/apparmor/label.c
1074
new->vec[k++] = aa_get_profile(next);
security/apparmor/label.c
1307
goto next;
security/apparmor/label.c
1314
next:
security/apparmor/label.c
1365
goto next;
security/apparmor/label.c
1371
next:
security/apparmor/lib.c
68
const char *next;
security/apparmor/lib.c
74
next = str + n;
security/apparmor/lib.c
75
ent = val_table_find_ent(debug_values_table, str, next - str);
security/apparmor/lib.c
80
(int)(next - str), str);
security/apparmor/lib.c
81
str = next + 1;
security/apparmor/lib.c
82
} while (*next != 0);
security/apparmor/match.c
436
#define match_char(state, def, base, next, check, C) \
security/apparmor/match.c
446
(state) = (next)[pos]; \
security/apparmor/match.c
471
u32 *next = NEXT_TABLE(dfa);
security/apparmor/match.c
485
match_char(state, def, base, next, check, c);
security/apparmor/match.c
491
match_char(state, def, base, next, check, (u8) *str);
security/apparmor/match.c
515
u32 *next = NEXT_TABLE(dfa);
security/apparmor/match.c
530
match_char(state, def, base, next, check, c);
security/apparmor/match.c
536
match_char(state, def, base, next, check, (u8) *str);
security/apparmor/match.c
558
u32 *next = NEXT_TABLE(dfa);
security/apparmor/match.c
565
match_char(state, def, base, next, check, equiv[(u8) c]);
security/apparmor/match.c
567
match_char(state, def, base, next, check, (u8) c);
security/apparmor/match.c
576
u32 *next = NEXT_TABLE(dfa);
security/apparmor/match.c
584
match_char(state, def, base, next, check, -1);
security/apparmor/match.c
607
u32 *next = NEXT_TABLE(dfa);
security/apparmor/match.c
623
state = next[pos];
security/apparmor/match.c
634
state = next[pos];
security/apparmor/match.c
668
u32 *next = NEXT_TABLE(dfa);
security/apparmor/match.c
685
state = next[pos];
security/apparmor/match.c
696
state = next[pos];
security/apparmor/match.c
745
u32 *next = NEXT_TABLE(dfa);
security/apparmor/match.c
769
state = next[pos];
security/apparmor/match.c
788
state = next[pos];
security/apparmor/procattr.c
126
char *next = hat + strlen(hat) + 1;
security/apparmor/procattr.c
131
hat = next;
security/integrity/ima/ima_fs.c
102
qe = list_entry_rcu(qe->later.next, struct ima_queue_entry, later);
security/integrity/ima/ima_fs.c
200
.next = ima_measurements_next,
security/integrity/ima/ima_fs.c
275
.next = ima_measurements_next,
security/integrity/ima/ima_fs.c
389
.next = ima_policy_next,
security/integrity/ima/ima_policy.c
2060
entry = list_entry_rcu(entry->list.next, struct ima_rule_entry, list);
security/integrity/ima/ima_policy.c
322
char *cur, *next;
security/integrity/ima/ima_policy.c
329
next = src_copy;
security/integrity/ima/ima_policy.c
330
while ((cur = strsep(&next, "|"))) {
security/ipe/audit.c
100
list_for_each_entry(ptr, &r->props, next) {
security/ipe/eval.c
347
list_for_each_entry(rule, &rules->rules, next) {
security/ipe/eval.c
350
list_for_each_entry(prop, &rule->props, next) {
security/ipe/policy.h
48
struct list_head next;
security/ipe/policy.h
57
struct list_head next;
security/ipe/policy_parser.c
223
list_for_each_entry_safe(p, t, &r->props, next) {
security/ipe/policy_parser.c
224
list_del(&p->next);
security/ipe/policy_parser.c
342
list_add_tail(&p->next, &r->props);
security/ipe/policy_parser.c
380
INIT_LIST_HEAD(&r->next);
security/ipe/policy_parser.c
435
list_add_tail(&r->next, &p->rules[op].rules);
security/ipe/policy_parser.c
458
list_for_each_entry_safe(pp, t, &p->rules[i].rules, next) {
security/ipe/policy_parser.c
459
list_del(&pp->next);
security/keys/gc.c
139
list_entry(keys->next, struct key, graveyard_link);
security/keys/keyring.c
417
if (keyring->name_link.next != NULL &&
security/keys/proc.c
23
.next = proc_keys_next,
security/keys/proc.c
35
.next = proc_key_users_next,
security/landlock/ruleset.c
483
struct landlock_rule *freeme, *next;
security/landlock/ruleset.c
486
rbtree_postorder_for_each_entry_safe(freeme, next, &ruleset->root_inode,
security/landlock/ruleset.c
491
rbtree_postorder_for_each_entry_safe(freeme, next,
security/lsm_init.c
201
char *sep, *name, *next;
security/lsm_init.c
229
next = sep;
security/lsm_init.c
231
while ((name = strsep(&next, ",")) != NULL) {
security/safesetid/lsm.c
40
hash_for_each_possible(policy->rules, rule, next, __kuid_val(src.uid)) {
security/safesetid/lsm.c
48
hash_for_each_possible(policy->rules, rule, next, __kgid_val(src.gid)) {
security/safesetid/lsm.h
45
struct hlist_node next;
security/safesetid/securityfs.c
104
hash_for_each(pol->rules, bucket, rule, next) {
security/safesetid/securityfs.c
78
hash_for_each_safe(pol->rules, bucket, tmp, rule, next)
security/safesetid/securityfs.c
91
hash_add(pol->rules, &rule->next, __kuid_val(rule->src_id.uid));
security/safesetid/securityfs.c
93
hash_add(pol->rules, &rule->next, __kgid_val(rule->src_id.gid));
security/selinux/avc.c
805
c->next = avc_callbacks;
security/selinux/avc.c
83
struct avc_callback_node *next;
security/selinux/avc.c
961
for (c = avc_callbacks; c; c = c->next) {
security/selinux/selinuxfs.c
1586
.next = sel_avc_stats_seq_next,
security/selinux/ss/avtab.c
108
newnode = avtab_insert_node(h, prev ? &prev->next : &h->htable[hvalue],
security/selinux/ss/avtab.c
132
prev = cur, cur = cur->next) {
security/selinux/ss/avtab.c
137
return avtab_insert_node(h, prev ? &prev->next : &h->htable[hvalue],
security/selinux/ss/avtab.c
155
for (cur = h->htable[hvalue]; cur; cur = cur->next) {
security/selinux/ss/avtab.c
176
for (cur = node->next; cur; cur = cur->next) {
security/selinux/ss/avtab.c
198
cur = cur->next;
security/selinux/ss/avtab.c
275
cur = cur->next;
security/selinux/ss/avtab.c
569
for (cur = a->htable[i]; cur; cur = cur->next) {
security/selinux/ss/avtab.c
58
newnode->next = *dst;
security/selinux/ss/avtab.c
99
prev = cur, cur = cur->next) {
security/selinux/ss/avtab.h
81
struct avtab_node *next;
security/selinux/ss/constraint.h
55
struct constraint_expr *next; /* next expression */
security/selinux/ss/constraint.h
61
struct constraint_node *next; /* next constraint */
security/selinux/ss/ebitmap.c
139
e_iter = e_iter->next;
security/selinux/ss/ebitmap.c
192
e_prev->next = e_iter;
security/selinux/ss/ebitmap.c
232
n1 = n1->next;
security/selinux/ss/ebitmap.c
250
n1 = n1->next;
security/selinux/ss/ebitmap.c
251
n2 = n2->next;
security/selinux/ss/ebitmap.c
271
n = n->next;
security/selinux/ss/ebitmap.c
297
if (!n->next) {
security/selinux/ss/ebitmap.c
309
prev->next = n->next;
security/selinux/ss/ebitmap.c
311
e->node = n->next;
security/selinux/ss/ebitmap.c
317
n = n->next;
security/selinux/ss/ebitmap.c
335
new->next = prev->next;
security/selinux/ss/ebitmap.c
336
prev->next = new;
security/selinux/ss/ebitmap.c
338
new->next = e->node;
security/selinux/ss/ebitmap.c
355
n = n->next;
security/selinux/ss/ebitmap.c
39
n1 = n1->next;
security/selinux/ss/ebitmap.c
40
n2 = n2->next;
security/selinux/ss/ebitmap.c
435
n->next = tmp;
security/selinux/ss/ebitmap.c
566
for (node = e->node; node; node = node->next) {
security/selinux/ss/ebitmap.c
65
new->next = NULL;
security/selinux/ss/ebitmap.c
67
prev->next = new;
security/selinux/ss/ebitmap.c
71
n = n->next;
security/selinux/ss/ebitmap.h
37
struct ebitmap_node *next;
security/selinux/ss/ebitmap.h
54
for (*n = e->node; *n; *n = (*n)->next) {
security/selinux/ss/ebitmap.h
76
for (*n = (*n)->next; *n; *n = (*n)->next) {
security/selinux/ss/hashtab.c
101
cur = cur->next;
security/selinux/ss/hashtab.c
124
cur = cur->next;
security/selinux/ss/hashtab.c
161
orig_cur = orig_cur->next) {
security/selinux/ss/hashtab.c
171
tmp->next = NULL;
security/selinux/ss/hashtab.c
175
tail->next = tmp;
security/selinux/ss/hashtab.c
186
tmp = cur->next;
security/selinux/ss/hashtab.c
62
newnode->next = *dst;
security/selinux/ss/hashtab.c
78
cur = cur->next;
security/selinux/ss/hashtab.h
115
cur = cur->next;
security/selinux/ss/hashtab.h
29
struct hashtab_node *next;
security/selinux/ss/hashtab.h
84
cur = cur->next;
security/selinux/ss/hashtab.h
87
return __hashtab_insert(h, prev ? &prev->next : &h->htable[hvalue], key,
security/selinux/ss/policydb.c
1244
lc->next = c;
security/selinux/ss/policydb.c
1261
le->next = e;
security/selinux/ss/policydb.c
1961
datum = datum->next;
security/selinux/ss/policydb.c
1971
datum->next = NULL;
security/selinux/ss/policydb.c
1974
last->next = datum;
security/selinux/ss/policydb.c
2044
datum->next = NULL;
security/selinux/ss/policydb.c
2058
dst = &datum->next;
security/selinux/ss/policydb.c
2084
first = first->next;
security/selinux/ss/policydb.c
2164
genfs_p = genfs, genfs = genfs->next) {
security/selinux/ss/policydb.c
2174
newgenfs->next = genfs;
security/selinux/ss/policydb.c
2176
genfs_p->next = newgenfs;
security/selinux/ss/policydb.c
2212
for (l = NULL, c = genfs->head; c; l = c, c = c->next) {
security/selinux/ss/policydb.c
2227
newc->next = c;
security/selinux/ss/policydb.c
2229
l->next = newc;
security/selinux/ss/policydb.c
2270
l->next = c;
security/selinux/ss/policydb.c
239
e = e->next;
security/selinux/ss/policydb.c
243
constraint = constraint->next;
security/selinux/ss/policydb.c
252
e = e->next;
security/selinux/ss/policydb.c
256
constraint = constraint->next;
security/selinux/ss/policydb.c
2675
lra->next = ra;
security/selinux/ss/policydb.c
2923
for (ra = r; ra; ra = ra->next)
security/selinux/ss/policydb.c
2929
for (ra = r; ra; ra = ra->next) {
security/selinux/ss/policydb.c
3048
for (c = node; c; c = c->next) {
security/selinux/ss/policydb.c
3050
for (e = c->expr; e; e = e->next)
security/selinux/ss/policydb.c
3057
for (e = c->expr; e; e = e->next) {
security/selinux/ss/policydb.c
3106
for (c = cladatum->constraints; c; c = c->next)
security/selinux/ss/policydb.c
3139
for (c = cladatum->validatetrans; c; c = c->next)
security/selinux/ss/policydb.c
3318
for (c = p->ocontexts[i]; c; c = c->next)
security/selinux/ss/policydb.c
3324
for (c = p->ocontexts[i]; c; c = c->next) {
security/selinux/ss/policydb.c
338
struct filename_trans_datum *next, *d = datum;
security/selinux/ss/policydb.c
344
next = d->next;
security/selinux/ss/policydb.c
3452
for (genfs = p->genfs; genfs; genfs = genfs->next)
security/selinux/ss/policydb.c
3458
for (genfs = p->genfs; genfs; genfs = genfs->next) {
security/selinux/ss/policydb.c
346
d = next;
security/selinux/ss/policydb.c
3468
for (c = genfs->head; c; c = c->next)
security/selinux/ss/policydb.c
3474
for (c = genfs->head; c; c = c->next) {
security/selinux/ss/policydb.c
3577
datum = datum->next;
security/selinux/ss/policydb.c
3605
datum = datum->next;
security/selinux/ss/policydb.c
3626
datum = datum->next;
security/selinux/ss/policydb.c
811
c = c->next;
security/selinux/ss/policydb.c
824
c = c->next;
security/selinux/ss/policydb.c
828
g = g->next;
security/selinux/ss/policydb.c
838
for (ra = p->role_allow; ra; ra = ra->next) {
security/selinux/ss/policydb.c
883
for (c = head; c; c = c->next) {
security/selinux/ss/policydb.h
101
struct filename_trans_datum *next; /* record for next otype*/
security/selinux/ss/policydb.h
107
struct role_allow *next;
security/selinux/ss/policydb.h
203
struct ocontext *next;
security/selinux/ss/policydb.h
209
struct genfs *next;
security/selinux/ss/services.c
1749
datum = datum->next;
security/selinux/ss/services.c
2138
oc = oc->next;
security/selinux/ss/services.c
2472
c = c->next;
security/selinux/ss/services.c
2525
c = c->next;
security/selinux/ss/services.c
2578
c = c->next;
security/selinux/ss/services.c
2634
c = c->next;
security/selinux/ss/services.c
2707
c = c->next;
security/selinux/ss/services.c
2721
c = c->next;
security/selinux/ss/services.c
279
for (e = cexpr; e; e = e->next) {
security/selinux/ss/services.c
2909
for (genfs = policydb->genfs; genfs; genfs = genfs->next) {
security/selinux/ss/services.c
2920
for (c = genfs->head; c; c = c->next) {
security/selinux/ss/services.c
3013
c = c->next;
security/selinux/ss/services.c
696
constraint = constraint->next;
security/selinux/ss/services.c
707
for (ra = policydb->role_allow; ra; ra = ra->next) {
security/selinux/ss/services.c
830
constraint = constraint->next;
security/smack/smack_lsm.c
730
char *next = strchr(from, ',');
security/smack/smack_lsm.c
734
if (next)
security/smack/smack_lsm.c
735
len = next - from;
security/smack/smackfs.c
1003
.next = cipso_seq_next,
security/smack/smackfs.c
1079
.next = net4addr_seq_next,
security/smack/smackfs.c
1115
m = list_entry_rcu(smk_net4addr_list.next,
security/smack/smackfs.c
1129
m_next = list_entry_rcu(m->list.next,
security/smack/smackfs.c
1341
.next = net6addr_seq_next,
security/smack/smackfs.c
1377
m = list_entry_rcu(smk_net6addr_list.next,
security/smack/smackfs.c
1390
m_next = list_entry_rcu(m->list.next,
security/smack/smackfs.c
1903
.next = onlycap_seq_next,
security/smack/smackfs.c
1931
first = public->next;
security/smack/smackfs.c
1935
private->prev->next = public;
security/smack/smackfs.c
1936
private->next->prev = public;
security/smack/smackfs.c
1937
rcu_assign_pointer(public->next, private->next);
security/smack/smackfs.c
1944
private->next = first;
security/smack/smackfs.c
1947
last->next = private;
security/smack/smackfs.c
2244
.next = load_self_seq_next,
security/smack/smackfs.c
2376
.next = load2_seq_next,
security/smack/smackfs.c
2453
.next = load_self2_seq_next,
security/smack/smackfs.c
2725
.next = relabel_self_seq_next,
security/smack/smackfs.c
618
.next = load2_seq_next,
security/smack/smackfs.c
803
.next = cipso_seq_next,
security/tomoyo/audit.c
450
ptr = list_entry(tomoyo_log.next, typeof(*ptr), list);
security/tomoyo/common.c
2692
if (!ns || (head->r.eof && ns->next != &tomoyo_namespace_list)) {
security/tomoyo/common.c
2695
head->r.ns = ns ? ns->next : tomoyo_namespace_list.next;
security/tomoyo/common.c
2710
head->r.ns->next != &tomoyo_namespace_list;
security/tomoyo/common.c
751
next:
security/tomoyo/common.c
831
goto next;
security/tomoyo/common.h
1289
pos = srcu_dereference((head)->next, &tomoyo_ss); \
security/tomoyo/common.h
1290
for ( ; pos != (head); pos = srcu_dereference(pos->next, &tomoyo_ss))
sound/core/control.c
103
cread = snd_kctl_event(ctl->events.next);
sound/core/control.c
2013
kev = snd_kctl_event(ctl->events.next);
sound/core/control.c
210
for (lops = snd_ctl_layer; lops; lops = lops->next)
sound/core/control.c
2201
for (lops = snd_ctl_layer; lops; lops = lops->next)
sound/core/control.c
2222
lops->next = snd_ctl_layer;
sound/core/control.c
2249
for (lops2 = snd_ctl_layer, prev_lops2 = NULL; lops2; lops2 = lops2->next) {
sound/core/control.c
2252
snd_ctl_layer = lops->next;
sound/core/control.c
2254
prev_lops2->next = lops->next;
sound/core/control.c
2284
for (lops = snd_ctl_layer; lops; lops = lops->next) \
sound/core/control.c
2333
control = snd_kcontrol(card->controls.next);
sound/core/device.c
224
struct snd_device *dev, *next;
sound/core/device.c
228
list_for_each_entry_safe_reverse(dev, next, &card->devices, list) {
sound/core/device.c
237
list_for_each_entry_safe_reverse(dev, next, &card->devices, list)
sound/core/info.c
300
size_t next;
sound/core/info.c
307
next = pos + count;
sound/core/info.c
309
if (next > 16 * 1024)
sound/core/info.c
318
if (next > buf->len) {
sound/core/info.c
319
char *nbuf = kvzalloc(PAGE_ALIGN(next), GFP_KERNEL);
sound/core/info.c
324
buf->len = PAGE_ALIGN(next);
sound/core/info.c
328
buf->size = next;
sound/core/info.c
329
*offset = next;
sound/core/oss/pcm_oss.c
2342
setup = setup->next) {
sound/core/oss/pcm_oss.c
2967
setup = setup->next;
sound/core/oss/pcm_oss.c
2977
setupn = setup->next;
sound/core/oss/pcm_oss.c
3001
for (setup = pstr->oss.setup_list; setup; setup = setup->next) {
sound/core/oss/pcm_oss.c
3044
setup1->next; setup1 = setup1->next);
sound/core/oss/pcm_oss.c
3045
setup1->next = setup;
sound/core/oss/pcm_oss.c
536
struct snd_pcm_plugin *plugin, *next;
sound/core/oss/pcm_oss.c
540
next = plugin->next;
sound/core/oss/pcm_oss.c
542
plugin = next;
sound/core/oss/pcm_oss.c
551
plugin->next = runtime->oss.plugin_first;
sound/core/oss/pcm_oss.c
566
plugin->next = NULL;
sound/core/oss/pcm_oss.c
569
runtime->oss.plugin_last->next = plugin;
sound/core/oss/pcm_plugin.c
100
while (plugin->next) {
sound/core/oss/pcm_plugin.c
105
plugin = plugin->next;
sound/core/oss/pcm_plugin.c
196
plugin_next = plugin->next;
sound/core/oss/pcm_plugin.c
584
struct snd_pcm_plugin *plugin, *next;
sound/core/oss/pcm_plugin.c
593
next = plugin->next;
sound/core/oss/pcm_plugin.c
594
if (next) {
sound/core/oss/pcm_plugin.c
601
err = next->client_channels(next, frames1, &dst_channels);
sound/core/oss/pcm_plugin.c
619
plugin = next;
sound/core/oss/pcm_plugin.c
626
struct snd_pcm_plugin *plugin, *next;
sound/core/oss/pcm_plugin.c
638
next = plugin->next;
sound/core/oss/pcm_plugin.c
639
if (next) {
sound/core/oss/pcm_plugin.c
651
plugin = next;
sound/core/oss/pcm_plugin.h
60
struct snd_pcm_plugin *next;
sound/core/pcm.c
1079
for (substream = pcm->streams[cidx].substream; substream; substream = substream->next)
sound/core/pcm.c
129
substream = substream->next)
sound/core/pcm.c
672
prev->next = substream;
sound/core/pcm.c
682
prev->next = NULL;
sound/core/pcm.c
826
substream_next = substream->next;
sound/core/pcm.c
833
setupn = setup->next;
sound/core/pcm.c
903
substream = substream->next) {
sound/core/pcm.c
916
substream = substream->next)
sound/core/pcm.c
929
for (substream = pstr->substream; substream; substream = substream->next) {
sound/core/pcm_lib.c
516
for (substream = stream->substream; substream != NULL; substream = substream->next)
sound/core/pcm_local.h
74
(subs) = (subs)->next)
sound/core/rawmidi.c
1914
substream = list_entry(stream->substreams.next, struct snd_rawmidi_substream, list);
sound/core/seq/seq_fifo.c
120
f->tail->next = cell;
sound/core/seq/seq_fifo.c
124
cell->next = NULL;
sound/core/seq/seq_fifo.c
143
f->head = cell->next;
sound/core/seq/seq_fifo.c
149
cell->next = NULL;
sound/core/seq/seq_fifo.c
199
cell->next = f->head;
sound/core/seq/seq_fifo.c
221
struct snd_seq_event_cell *cell, *next, *oldhead;
sound/core/seq/seq_fifo.c
252
for (cell = oldhead; cell; cell = next) {
sound/core/seq/seq_fifo.c
253
next = cell->next;
sound/core/seq/seq_memory.c
105
for (; len > 0 && cell; cell = cell->next) {
sound/core/seq/seq_memory.c
228
cell->next = pool->free;
sound/core/seq/seq_memory.c
250
nextptr = curp->next;
sound/core/seq/seq_memory.c
251
curp->next = pool->free;
sound/core/seq/seq_memory.c
315
pool->free = cell->next;
sound/core/seq/seq_memory.c
322
cell->next = NULL;
sound/core/seq/seq_memory.c
397
tail->next = tmp;
sound/core/seq/seq_memory.c
402
src = src->next;
sound/core/seq/seq_memory.c
461
cellptr->next = pool->free;
sound/core/seq/seq_memory.h
35
struct snd_seq_event_cell *next; /* next cell */
sound/core/seq/seq_prioq.c
152
f->tail->next = cell;
sound/core/seq/seq_prioq.c
154
cell->next = NULL;
sound/core/seq/seq_prioq.c
178
cur = cur->next;
sound/core/seq/seq_prioq.c
187
prev->next = cell;
sound/core/seq/seq_prioq.c
188
cell->next = cur;
sound/core/seq/seq_prioq.c
223
f->head = cell->next;
sound/core/seq/seq_prioq.c
229
cell->next = NULL;
sound/core/seq/seq_prioq.c
252
register struct snd_seq_event_cell *cell, *next;
sound/core/seq/seq_prioq.c
258
for (cell = f->head; cell; cell = next) {
sound/core/seq/seq_prioq.c
259
next = cell->next;
sound/core/seq/seq_prioq.c
267
f->head = cell->next;
sound/core/seq/seq_prioq.c
269
prev->next = cell->next;
sound/core/seq/seq_prioq.c
271
f->tail = cell->next;
sound/core/seq/seq_prioq.c
275
cell->next = NULL;
sound/core/seq/seq_prioq.c
279
freeprev->next = cell;
sound/core/seq/seq_prioq.c
286
freenext = freefirst->next;
sound/core/timer.c
1527
timer = list_entry(snd_timer_list.next,
sound/core/timer.c
346
list_entry(timer->open_list_head.next,
sound/core/timer.c
798
list_del_init(head->next);
sound/drivers/opl3/opl3_synth.c
309
for (patch = opl3->patch_table[key]; patch; patch = patch->next) {
sound/drivers/opl3/opl3_synth.c
321
patch->next = opl3->patch_table[key];
sound/drivers/opl3/opl3_synth.c
334
struct fm_patch *patch, *next;
sound/drivers/opl3/opl3_synth.c
335
for (patch = opl3->patch_table[i]; patch; patch = next) {
sound/drivers/opl3/opl3_synth.c
336
next = patch->next;
sound/drivers/opl4/opl4_synth.c
459
return list_entry(opl4->off_voices.next, struct opl4_voice, list);
sound/drivers/opl4/opl4_synth.c
462
return list_entry(opl4->on_voices.next, struct opl4_voice, list);
sound/firewire/amdtp-stream.c
2133
struct amdtp_stream *s, *next;
sound/firewire/amdtp-stream.c
2138
list_for_each_entry_safe(s, next, &d->streams, list) {
sound/hda/common/controller.c
1245
struct hda_codec *codec, *next;
sound/hda/common/controller.c
1255
list_for_each_codec_safe(codec, next, &chip->bus) {
sound/hda/common/hda_jack.h
27
struct hda_jack_callback *next;
sound/hda/common/jack.c
178
struct hda_jack_callback *cb, *next;
sound/hda/common/jack.c
184
for (cb = jack->callback; cb; cb = next) {
sound/hda/common/jack.c
185
next = cb->next;
sound/hda/common/jack.c
298
for (cb = jack->callback; cb; cb = cb->next) {
sound/hda/common/jack.c
338
callback->next = jack->callback;
sound/hda/common/jack.c
689
for (cb = jack->callback; cb; cb = cb->next) {
sound/hda/common/jack.c
699
for (cb = gated->callback; cb; cb = cb->next) {
sound/i2c/i2c.c
42
device = snd_i2c_device(bus->devices.next);
sound/i2c/i2c.c
49
slave = snd_i2c_slave_bus(bus->buses.next);
sound/isa/gus/gus_dma.c
163
gus->gf1.dma_data_pcm = block->next;
sound/isa/gus/gus_dma.c
167
gus->gf1.dma_data_synth = block->next;
sound/isa/gus/gus_dma.c
189
block->next = NULL;
sound/isa/gus/gus_dma.c
206
gus->gf1.dma_data_synth_last->next = block;
sound/isa/gus/gus_dma.c
214
gus->gf1.dma_data_pcm_last->next = block;
sound/isa/gus/gus_dma.c
90
gus->gf1.dma_data_pcm = block->next;
sound/isa/gus/gus_dma.c
98
gus->gf1.dma_data_synth = block->next;
sound/isa/gus/gus_mem.c
110
for (block = alloc->first; block; block = block->next)
sound/isa/gus/gus_mem.c
135
block->prev = block->next = NULL;
sound/isa/gus/gus_mem.c
136
for (pblock = alloc->first, idx = 0; pblock; pblock = pblock->next) {
sound/isa/gus/gus_mem.c
142
if (pblock->next) {
sound/isa/gus/gus_mem.c
143
if (pblock->ptr + pblock->size == pblock->next->ptr)
sound/isa/gus/gus_mem.c
145
if (pblock->next->ptr < boundary)
sound/isa/gus/gus_mem.c
146
ptr2 = pblock->next->ptr;
sound/isa/gus/gus_mem.c
246
nblock = block->next;
sound/isa/gus/gus_mem.c
277
for (block = alloc->first, i = 0; block; block = block->next, i++) {
sound/isa/gus/gus_mem.c
38
nblock->next = pblock;
sound/isa/gus/gus_mem.c
43
nblock->prev->next = nblock;
sound/isa/gus/gus_mem.c
46
pblock = pblock->next;
sound/isa/gus/gus_mem.c
48
nblock->next = NULL;
sound/isa/gus/gus_mem.c
54
alloc->last->next = nblock;
sound/isa/gus/gus_mem.c
67
alloc->first = block->next;
sound/isa/gus/gus_mem.c
68
if (block->next)
sound/isa/gus/gus_mem.c
69
block->next->prev = NULL;
sound/isa/gus/gus_mem.c
71
block->prev->next = block->next;
sound/isa/gus/gus_mem.c
72
if (block->next)
sound/isa/gus/gus_mem.c
73
block->next->prev = block->prev;
sound/isa/gus/gus_mem.c
78
block->prev->next = NULL;
sound/isa/gus/gus_mem.c
80
block->next->prev = block->prev;
sound/isa/gus/gus_mem.c
82
block->prev->next = block->next;
sound/isa/gus/gus_mem.c
94
for (block = alloc->first; block; block = block->next) {
sound/isa/gus/gus_pcm.c
826
for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next)
sound/pci/atiixp.c
198
__le32 next; /* address of the next packet descriptor */
sound/pci/atiixp.c
382
desc->next = cpu_to_le32((u32)dma->desc_buf.addr);
sound/pci/atiixp.c
384
desc->next = cpu_to_le32(desc_addr);
sound/pci/atiixp_modem.c
174
__le32 next; /* address of the next packet descriptor */
sound/pci/atiixp_modem.c
349
desc->next = cpu_to_le32((u32)dma->desc_buf.addr);
sound/pci/atiixp_modem.c
351
desc->next = cpu_to_le32(desc_addr);
sound/pci/ca0106/ca0106_main.c
1307
substream = substream->next) {
sound/pci/ca0106/ca0106_main.c
1315
substream = substream->next) {
sound/pci/ctxfi/ctdaio.c
635
hw->daio_mgr_set_imapnxt(mgr->ctrl_blk, entry->next);
sound/pci/ctxfi/ctdaio.c
699
entry->slot = entry->addr = entry->next = entry->user = 0;
sound/pci/ctxfi/cthardware.h
129
int (*srcimp_mgr_set_imapnxt)(void *blk, unsigned int next);
sound/pci/ctxfi/cthardware.h
175
int (*daio_mgr_set_imapnxt)(void *blk, unsigned int next);
sound/pci/ctxfi/cthw20k1.c
1080
static int daio_mgr_set_imapnxt(void *blk, unsigned int next)
sound/pci/ctxfi/cthw20k1.c
1084
set_field(&ctl->daoimap.aim, AIM_NXT, next);
sound/pci/ctxfi/cthw20k1.c
541
static int srcimp_mgr_set_imapnxt(void *blk, unsigned int next)
sound/pci/ctxfi/cthw20k1.c
545
set_field(&ctl->srcimap.srcaim, SRCAIM_NXT, next);
sound/pci/ctxfi/cthw20k2.c
1036
static int daio_mgr_set_imapnxt(void *blk, unsigned int next)
sound/pci/ctxfi/cthw20k2.c
1040
set_field(&ctl->daoimap.aim, AIM_NXT, next);
sound/pci/ctxfi/cthw20k2.c
541
static int srcimp_mgr_set_imapnxt(void *blk, unsigned int next)
sound/pci/ctxfi/cthw20k2.c
545
set_field(&ctl->srcimap.srcaim, SRCAIM_NXT, next);
sound/pci/ctxfi/ctimap.c
102
pos = head->next;
sound/pci/ctxfi/ctimap.c
27
entry->next = entry->addr;
sound/pci/ctxfi/ctimap.c
49
pos = head->next;
sound/pci/ctxfi/ctimap.c
56
entry->next = pos_ent->addr;
sound/pci/ctxfi/ctimap.c
58
pre_ent->next = entry->addr;
sound/pci/ctxfi/ctimap.c
67
struct list_head *next, *pre, *head;
sound/pci/ctxfi/ctimap.c
76
next = (entry->list.next == head) ? head->next : entry->list.next;
sound/pci/ctxfi/ctimap.c
80
entry->next = entry->addr = entry->user = entry->slot = 0;
sound/pci/ctxfi/ctimap.c
87
next_ent = list_entry(next, struct imapper, list);
sound/pci/ctxfi/ctimap.c
89
pre_ent->next = next_ent->addr;
sound/pci/ctxfi/ctimap.h
24
unsigned short next; /* the next input mapper ram id */
sound/pci/ctxfi/ctsrc.c
773
hw->srcimp_mgr_set_imapnxt(mgr->ctrl_blk, entry->next);
sound/pci/ctxfi/ctsrc.c
829
entry->slot = entry->addr = entry->next = entry->user = 0;
sound/pci/ctxfi/ctvmem.c
219
pos = vm->used.next;
sound/pci/ctxfi/ctvmem.c
225
pos = vm->unused.next;
sound/pci/echoaudio/echoaudio.c
878
for (ss = pcm->streams[stream].substream; ss; ss = ss->next)
sound/pci/emu10k1/emufx.c
427
nirq = irq->next; /* irq ptr can be removed from list */
sound/pci/emu10k1/emufx.c
446
irq->next = NULL;
sound/pci/emu10k1/emufx.c
453
irq->next = emu->fx8010.irq_handlers;
sound/pci/emu10k1/emufx.c
467
emu->fx8010.irq_handlers = tmp->next;
sound/pci/emu10k1/emufx.c
473
while (tmp && tmp->next != irq)
sound/pci/emu10k1/emufx.c
474
tmp = tmp->next;
sound/pci/emu10k1/emufx.c
476
tmp->next = tmp->next->next;
sound/pci/emu10k1/emupcm.c
1422
for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next)
sound/pci/emu10k1/emupcm.c
1427
for (substream = pcm->streams[SNDRV_PCM_STREAM_CAPTURE].substream; substream; substream = substream->next)
sound/pci/emu10k1/emupcm.c
1453
for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next)
sound/pci/emu10k1/memory.c
137
struct list_head *next;
sound/pci/emu10k1/memory.c
139
page = search_empty_map_area(emu, blk->pages, &next);
sound/pci/emu10k1/memory.c
147
list_add_tail(&blk->mapped_link, next);
sound/pci/emu10k1/memory.c
179
p = blk->mapped_link.next;
sound/pci/emu10k1/memory.c
276
p = emu->mapped_order_link_head.next;
sound/pci/emu10k1/memory.c
278
nextp = p->next;
sound/pci/emu10k1/memory.c
453
p = blk->mem.list.next;
sound/pci/emu10k1/p16v.c
578
substream = substream->next) {
sound/pci/emu10k1/p16v.c
591
substream = substream->next) {
sound/pci/emu10k1/voice.c
46
goto next;
sound/pci/emu10k1/voice.c
62
next: ;
sound/pci/es1968.c
1341
if (buf->list.next != &chip->buf_list) {
sound/pci/es1968.c
1342
chunk = list_entry(buf->list.next, struct esm_memory, list);
sound/pci/es1968.c
1358
while ((p = chip->buf_list.next) != &chip->buf_list) {
sound/pci/mixart/mixart.c
918
for (subs = pcm->streams[stream].substream; subs; subs = subs->next, idx++)
sound/pci/trident/trident_main.c
2126
for (substream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream; substream; substream = substream->next)
sound/pci/trident/trident_main.c
2176
substream = substream->next;
sound/pci/trident/trident_main.c
2178
substream = substream->next;
sound/pci/trident/trident_main.c
2181
substream = substream->next;
sound/pci/ymfpci/ymfpci_main.c
1832
substream = substream->next;
sound/soc/atmel/mchp-pdmc.c
228
for (s = info->pcm->streams[SNDRV_PCM_STREAM_CAPTURE].substream; s; s = s->next)
sound/soc/au1x/ac97c.c
100
goto next;
sound/soc/au1x/ac97c.c
116
next:
sound/soc/au1x/ac97c.c
139
goto next;
sound/soc/au1x/ac97c.c
148
next:
sound/soc/au1x/dma.c
100
pointer->next = stream->buffer;
sound/soc/au1x/dma.c
120
set_dma_addr1(stream->dma, stream->buffer->next->start);
sound/soc/au1x/dma.c
126
set_dma_addr0(stream->dma, stream->buffer->next->start);
sound/soc/au1x/dma.c
140
stream->buffer = stream->buffer->next;
sound/soc/au1x/dma.c
142
set_dma_addr0(stream->dma, stream->buffer->next->start);
sound/soc/au1x/dma.c
147
stream->buffer = stream->buffer->next;
sound/soc/au1x/dma.c
149
set_dma_addr1(stream->dma, stream->buffer->next->start);
sound/soc/au1x/dma.c
30
struct pcm_period *next;
sound/soc/au1x/dma.c
56
pointer_next = pointer->next;
sound/soc/au1x/dma.c
92
pointer->next = kmalloc_obj(struct pcm_period);
sound/soc/au1x/dma.c
93
if (!pointer->next) {
sound/soc/au1x/dma.c
97
pointer = pointer->next;
sound/soc/fsl/fsl_dma.c
433
dma_private->link[i].next = cpu_to_be64(temp_link);
sound/soc/fsl/fsl_dma.c
438
dma_private->link[i - 1].next = cpu_to_be64(dma_private->ld_buf_phys);
sound/soc/fsl/fsl_dma.h
103
__be64 next; /* Address of next list descriptor */
sound/soc/fsl/fsl_dma.h
121
__be64 next; /* Address of next link descriptor */
sound/soc/intel/atom/sst/sst_ipc.c
155
msg = list_entry(sst_drv_ctx->ipc_dispatch_list.next,
sound/soc/mediatek/common/mtk-afe-platform-driver.c
140
if (afe->sub_dais.next && afe->sub_dais.prev) {
sound/soc/renesas/rcar/core.c
1318
substream = substream->next) {
sound/soc/soc-dapm.c
1620
list_del(widgets.next);
sound/soc/soc-pcm.c
2203
goto next;
sound/soc/soc-pcm.c
2214
goto next;
sound/soc/soc-pcm.c
2218
goto next;
sound/soc/soc-pcm.c
2228
goto next;
sound/soc/soc-pcm.c
2235
goto next;
sound/soc/soc-pcm.c
2239
goto next;
sound/soc/soc-pcm.c
2244
goto next;
sound/soc/soc-pcm.c
2253
goto next;
sound/soc/soc-pcm.c
2260
goto next;
sound/soc/soc-pcm.c
2265
goto next;
sound/soc/soc-pcm.c
2273
goto next;
sound/soc/soc-pcm.c
2279
goto next;
sound/soc/soc-pcm.c
2300
goto next;
sound/soc/soc-pcm.c
2311
goto next;
sound/soc/soc-pcm.c
2315
goto next;
sound/soc/soc-pcm.c
2320
goto next;
sound/soc/soc-pcm.c
2327
goto next;
sound/soc/soc-pcm.c
2334
goto next;
sound/soc/soc-pcm.c
2339
goto next;
sound/soc/soc-pcm.c
2345
next:
sound/sound_core.c
114
struct sound_unit *next;
sound/sound_core.c
166
list=&((*list)->next);
sound/sound_core.c
173
list=&((*list)->next);
sound/sound_core.c
186
list=&((*list)->next);
sound/sound_core.c
201
s->next=*list;
sound/sound_core.c
219
*list=p->next;
sound/sound_core.c
222
list=&(p->next);
sound/sound_core.c
528
s=s->next;
sound/sparc/amd7930.c
1011
amd->next = amd7930_list;
sound/sparc/amd7930.c
1049
struct snd_amd7930 *next = p->next;
sound/sparc/amd7930.c
1053
p = next;
sound/sparc/amd7930.c
341
struct snd_amd7930 *next;
sound/sparc/cs4231.c
117
struct snd_cs4231 *next;
sound/synth/emux/soundfont.c
1151
for (sf = sflist->fonts; sf; sf = sf->next) {
sound/synth/emux/soundfont.c
1152
for (cur = sf->zones; cur; cur = cur->next) {
sound/synth/emux/soundfont.c
1368
nextsf = sf->next;
sound/synth/emux/soundfont.c
1370
nextzp = zp->next;
sound/synth/emux/soundfont.c
1374
nextsp = sp->next;
sound/synth/emux/soundfont.c
1462
for (sf = sflist->fonts; sf; sf = sf->next) {
sound/synth/emux/soundfont.c
1466
nextzp = zp->next;
sound/synth/emux/soundfont.c
1474
nextsp = sp->next;
sound/synth/emux/soundfont.c
225
for (sf = sflist->fonts; sf; sf = sf->next) {
sound/synth/emux/soundfont.c
240
sf->next = sflist->fonts;
sound/synth/emux/soundfont.c
315
zp->next = sf->zones;
sound/synth/emux/soundfont.c
349
sp->next = sf->samples;
sound/synth/emux/soundfont.c
366
sf->samples = sp->next;
sound/synth/emux/soundfont.c
394
for (zp = sf->zones; zp; prevp = zp, zp = zp->next) {
sound/synth/emux/soundfont.c
405
prevp->next = zp->next;
sound/synth/emux/soundfont.c
406
zp->next = sf->zones;
sound/synth/emux/soundfont.c
443
struct snd_sf_zone *prev, *next, *p;
sound/synth/emux/soundfont.c
447
for (p = sf->zones; p; p = next) {
sound/synth/emux/soundfont.c
448
next = p->next;
sound/synth/emux/soundfont.c
453
prev->next = next;
sound/synth/emux/soundfont.c
455
sf->zones = next;
sound/synth/emux/soundfont.c
515
for (zone = sf->zones; zone; zone = zone->next) {
sound/synth/emux/soundfont.c
645
for (p = sf->samples; p; p = p->next) {
sound/synth/util_mem.c
49
while ((p = hdr->block.next) != &hdr->block) {
sound/usb/helper.c
33
u8 *p, *end, *next;
sound/usb/helper.c
40
next = p + p[0];
sound/usb/helper.c
41
if (next > end)
sound/usb/helper.c
46
p = next;
sound/usb/midi.c
1814
goto next;
sound/usb/midi.c
1822
goto next;
sound/usb/midi.c
1832
next:
sound/usb/midi2.c
1229
struct snd_usb_midi2_interface *umidi, *next;
sound/usb/midi2.c
1231
list_for_each_entry_safe(umidi, next, &chip->midi_v2_list, list)
sound/usb/mixer.c
2948
struct usb_mixer_elem_list *list, *next;
sound/usb/mixer.c
2957
for (list = mixer->id_elems[id]; list; list = next) {
sound/usb/mixer.c
2958
next = list->next_id_elem;
sound/usb/qcom/qc_audio_offload.c
426
&info->list; curr_head = curr_head->next) {
sound/virtio/virtio_pcm.c
482
for (kss = ks->substream; kss; kss = kss->next)
tools/bootconfig/main.c
102
if (!xbc_node_get_child(node)->next)
tools/bootconfig/main.c
53
while (cnode && xbc_node_is_key(cnode) && !cnode->next) {
tools/bootconfig/main.c
66
if (vnode && xbc_node_is_value(vnode) && vnode->next)
tools/bootconfig/main.c
85
if (cnode->next) {
tools/bootconfig/main.c
94
if (node->next) {
tools/bootconfig/main.c
98
while (!node->next) {
tools/bpf/bpf_dbg.c
888
uint32_t pkt_len, int next)
tools/bpf/bpf_dbg.c
896
if (i++ == next)
tools/bpf/resolve_btfids/main.c
358
struct rb_node *next;
tools/bpf/resolve_btfids/main.c
361
next = rb_first(root);
tools/bpf/resolve_btfids/main.c
362
while (next) {
tools/bpf/resolve_btfids/main.c
363
id = rb_entry(next, struct btf_id, rb_node);
tools/bpf/resolve_btfids/main.c
364
next = rb_next(&id->rb_node);
tools/bpf/resolve_btfids/main.c
721
struct rb_node *next;
tools/bpf/resolve_btfids/main.c
724
next = rb_first(root);
tools/bpf/resolve_btfids/main.c
725
while (next) {
tools/bpf/resolve_btfids/main.c
726
id = rb_entry(next, struct btf_id, rb_node);
tools/bpf/resolve_btfids/main.c
731
next = rb_next(next);
tools/bpf/resolve_btfids/main.c
746
struct rb_node *next;
tools/bpf/resolve_btfids/main.c
749
next = rb_first(&obj->sets);
tools/bpf/resolve_btfids/main.c
750
while (next) {
tools/bpf/resolve_btfids/main.c
756
id = rb_entry(next, struct btf_id, rb_node);
tools/bpf/resolve_btfids/main.c
790
next = rb_next(next);
tools/bpf/resolve_btfids/main.c
959
struct rb_node *next;
tools/bpf/resolve_btfids/main.c
963
for (next = rb_first(&obj->sets); next; next = rb_next(next)) {
tools/bpf/resolve_btfids/main.c
964
set_id = rb_entry(next, struct btf_id, rb_node);
tools/firewire/list.h
16
return list->next == list;
tools/firewire/list.h
23
new_link->next = link;
tools/firewire/list.h
24
new_link->prev->next = new_link;
tools/firewire/list.h
25
new_link->next->prev = new_link;
tools/firewire/list.h
3
struct list *next, *prev;
tools/firewire/list.h
37
list_insert(list->next, new_link);
tools/firewire/list.h
43
link->prev->next = link->next;
tools/firewire/list.h
44
link->next->prev = link->prev;
tools/firewire/list.h
51
list_entry((list)->next, type, member)
tools/firewire/list.h
57
list_entry((elm)->member.next, typeof(*elm), member)
tools/firewire/list.h
9
list->next = list;
tools/firmware/ihex2fw.c
259
p = &((*p)->next);
tools/firmware/ihex2fw.c
261
record->next = *p;
tools/firmware/ihex2fw.c
282
p = p->next;
tools/firmware/ihex2fw.c
29
struct ihex_binrec *next; /* not part of the real data structure */
tools/include/linux/list.h
102
__list_del(entry->prev, entry->next);
tools/include/linux/list.h
107
__list_del(entry->prev, entry->next);
tools/include/linux/list.h
108
entry->next = LIST_POISON1;
tools/include/linux/list.h
126
new->next = old->next;
tools/include/linux/list.h
127
new->next->prev = new;
tools/include/linux/list.h
129
new->prev->next = new;
tools/include/linux/list.h
190
return list->next == head;
tools/include/linux/list.h
199
return head->next == head;
tools/include/linux/list.h
217
struct list_head *next = head->next;
tools/include/linux/list.h
218
return (next == head) && (next == head->prev);
tools/include/linux/list.h
230
first = head->next;
tools/include/linux/list.h
241
return !list_empty(head) && (head->next == head->prev);
tools/include/linux/list.h
247
struct list_head *new_first = entry->next;
tools/include/linux/list.h
248
list->next = head->next;
tools/include/linux/list.h
249
list->next->prev = list;
tools/include/linux/list.h
251
entry->next = list;
tools/include/linux/list.h
252
head->next = new_first;
tools/include/linux/list.h
27
list->next = list;
tools/include/linux/list.h
276
(head->next != entry && head != entry))
tools/include/linux/list.h
286
struct list_head *next)
tools/include/linux/list.h
288
struct list_head *first = list->next;
tools/include/linux/list.h
292
prev->next = first;
tools/include/linux/list.h
294
last->next = next;
tools/include/linux/list.h
295
next->prev = last;
tools/include/linux/list.h
307
__list_splice(list, head, head->next);
tools/include/linux/list.h
333
__list_splice(list, head, head->next);
tools/include/linux/list.h
373
list_entry((ptr)->next, type, member)
tools/include/linux/list.h
40
struct list_head *next)
tools/include/linux/list.h
414
list_entry((pos)->member.next, typeof(*(pos)), member)
tools/include/linux/list.h
42
next->prev = new;
tools/include/linux/list.h
43
new->next = next;
tools/include/linux/list.h
430
for (pos = (head)->next; pos != (head); pos = pos->next)
tools/include/linux/list.h
447
for (pos = (head)->next, n = pos->next; pos != (head); \
tools/include/linux/list.h
448
pos = n, n = pos->next)
tools/include/linux/list.h
45
prev->next = new;
tools/include/linux/list.h
50
struct list_head *next);
tools/include/linux/list.h
621
h->next = NULL;
tools/include/linux/list.h
63
__list_add(new, head, head->next);
tools/include/linux/list.h
637
struct hlist_node *next = n->next;
tools/include/linux/list.h
640
WRITE_ONCE(*pprev, next);
tools/include/linux/list.h
641
if (next)
tools/include/linux/list.h
642
next->pprev = pprev;
tools/include/linux/list.h
648
n->next = LIST_POISON1;
tools/include/linux/list.h
663
n->next = first;
tools/include/linux/list.h
665
first->pprev = &n->next;
tools/include/linux/list.h
672
struct hlist_node *next)
tools/include/linux/list.h
674
n->pprev = next->pprev;
tools/include/linux/list.h
675
n->next = next;
tools/include/linux/list.h
676
next->pprev = &n->next;
tools/include/linux/list.h
683
n->next = prev->next;
tools/include/linux/list.h
684
prev->next = n;
tools/include/linux/list.h
685
n->pprev = &prev->next;
tools/include/linux/list.h
687
if (n->next)
tools/include/linux/list.h
688
n->next->pprev = &n->next;
tools/include/linux/list.h
694
n->pprev = &n->next;
tools/include/linux/list.h
699
return h->pprev == &h->next;
tools/include/linux/list.h
718
for (pos = (head)->first; pos ; pos = pos->next)
tools/include/linux/list.h
721
for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \
tools/include/linux/list.h
738
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
tools/include/linux/list.h
746
for (pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member);\
tools/include/linux/list.h
748
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
tools/include/linux/list.h
757
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
tools/include/linux/list.h
768
pos && ({ n = pos->member.next; 1; }); \
tools/include/linux/list.h
781
begin->prev->next = end->next;
tools/include/linux/list.h
782
end->next->prev = begin->prev;
tools/include/linux/list.h
791
for (; pos != (head); pos = pos->next)
tools/include/linux/list.h
87
static inline void __list_del(struct list_head * prev, struct list_head * next)
tools/include/linux/list.h
89
next->prev = prev;
tools/include/linux/list.h
90
WRITE_ONCE(prev->next, next);
tools/include/linux/types.h
100
struct list_head *next, *prev;
tools/include/linux/types.h
108
struct hlist_node *next, **pprev;
tools/include/uapi/drm/drm.h
115
unsigned char next;
tools/lib/api/io_dir.h
55
struct io_dirent64 *next;
tools/lib/api/io_dir.h
81
iod->next = iod->buff;
tools/lib/api/io_dir.h
83
entry = iod->next;
tools/lib/api/io_dir.h
84
iod->next = (struct io_dirent64 *)((char *)entry + entry->d_reclen);
tools/lib/bpf/hashmap.c
142
prev_ptr = &cur->next, cur = cur->next) {
tools/lib/bpf/hashmap.c
27
entry->next = *pprev;
tools/lib/bpf/hashmap.c
34
*pprev = entry->next;
tools/lib/bpf/hashmap.c
35
entry->next = NULL;
tools/lib/bpf/hashmap.h
170
for (cur = (map)->buckets[bkt]; cur; cur = cur->next)
tools/lib/bpf/hashmap.h
183
cur && ({tmp = cur->next; true; }); \
tools/lib/bpf/hashmap.h
197
cur = cur->next) \
tools/lib/bpf/hashmap.h
204
cur && ({ tmp = cur->next; true; }); \
tools/lib/bpf/hashmap.h
69
struct hashmap_entry *next;
tools/lib/bpf/libbpf.c
10889
bpf_object__prev_map(const struct bpf_object *obj, const struct bpf_map *next)
tools/lib/bpf/libbpf.c
10891
if (next == NULL && obj != NULL) {
tools/lib/bpf/libbpf.c
10897
return __bpf_map__iter(next, obj, -1);
tools/lib/bpf/libbpf.c
9587
bpf_object__prev_program(const struct bpf_object *obj, struct bpf_program *next)
tools/lib/bpf/libbpf.c
9589
struct bpf_program *prog = next;
tools/lib/list_sort.c
174
struct list_head *list = head->next, *pending = NULL;
tools/lib/list_sort.c
181
head->prev->next = NULL;
tools/lib/list_sort.c
22
tail = &a->next;
tools/lib/list_sort.c
221
list = list->next;
tools/lib/list_sort.c
222
pending->next = NULL;
tools/lib/list_sort.c
23
a = a->next;
tools/lib/list_sort.c
230
struct list_head *next = pending->prev;
tools/lib/list_sort.c
232
if (!next)
tools/lib/list_sort.c
235
pending = next;
tools/lib/list_sort.c
30
tail = &b->next;
tools/lib/list_sort.c
31
b = b->next;
tools/lib/list_sort.c
57
tail->next = a;
tools/lib/list_sort.c
60
a = a->next;
tools/lib/list_sort.c
64
tail->next = b;
tools/lib/list_sort.c
67
b = b->next;
tools/lib/list_sort.c
76
tail->next = b;
tools/lib/list_sort.c
80
b = b->next;
tools/lib/list_sort.c
84
tail->next = head;
tools/lib/perf/evlist.c
134
struct perf_evsel *next = perf_evlist__next(evlist, evsel);
tools/lib/perf/evlist.c
138
if (next)
tools/lib/perf/evlist.c
139
list_for_each_entry_from(next, &evlist->entries, node)
tools/lib/perf/evlist.c
140
next->idx--;
tools/lib/perf/evlist.c
207
struct perf_evsel *next;
tools/lib/perf/evlist.c
210
next = list_first_entry(&evlist->entries,
tools/lib/perf/evlist.c
214
next = list_next_entry(prev, node);
tools/lib/perf/evlist.c
218
if (&next->node == &evlist->entries)
tools/lib/perf/evlist.c
221
return next;
tools/lib/perf/evlist.c
789
return map->next;
tools/lib/perf/evlist.c
809
struct perf_evsel *first = list_entry(evlist->entries.next,
tools/lib/perf/include/internal/evlist.h
117
return list_entry(evlist->entries.next, struct perf_evsel, node);
tools/lib/perf/include/internal/mmap.h
38
struct perf_mmap *next;
tools/lib/perf/mmap.c
28
prev->next = map;
tools/net/ynl/lib/ynl-priv.h
67
struct ynl_dump_list_type *next;
tools/net/ynl/lib/ynl-priv.h
87
uptr = (unsigned long)list->next;
tools/net/ynl/lib/ynl-priv.h
96
struct ynl_ntf_base_type *next;
tools/net/ynl/lib/ynl.c
1025
ds->last->next = obj;
tools/net/ynl/lib/ynl.c
1039
ds->last->next = YNL_LIST_END;
tools/net/ynl/lib/ynl.c
861
ys->ntf_first = ntf->next;
tools/net/ynl/lib/ynl.c
862
if (ys->ntf_last_next == &ntf->next)
tools/net/ynl/lib/ynl.c
904
ys->ntf_last_next = &rsp->next;
tools/objtool/arch/loongarch/decode.c
85
else for (*ops_list = op, ops_list = &op->next; op; op = NULL)
tools/objtool/arch/x86/decode.c
172
else for (*ops_list = op, ops_list = &op->next; op; op = NULL)
tools/objtool/check.c
1374
struct stack_op *op, *next;
tools/objtool/check.c
1376
for (op = insn->stack_ops; op; op = next) {
tools/objtool/check.c
1377
next = op->next;
tools/objtool/check.c
2022
alt->next = NULL;
tools/objtool/check.c
2031
for (a = orig_insn->alts; a->next; a = a->next)
tools/objtool/check.c
2033
a->next = alt;
tools/objtool/check.c
2093
goto next;
tools/objtool/check.c
2110
alt->next = insn->alts;
tools/objtool/check.c
2112
next:
tools/objtool/check.c
3348
for (op = insn->stack_ops; op; op = op->next) {
tools/objtool/check.c
3844
for (alt = insn->alts; alt; alt = alt->next) {
tools/objtool/check.c
4138
struct instruction *next, *dest;
tools/objtool/check.c
4142
next = next_insn_to_validate(file, insn);
tools/objtool/check.c
4151
for (alt = insn->alts; alt; alt = alt->next) {
tools/objtool/check.c
4231
if (!next) {
tools/objtool/check.c
4235
insn = next;
tools/objtool/check.c
4943
struct insn_chunk *next;
tools/objtool/check.c
4960
chunk->next = chunks;
tools/objtool/check.c
4965
for (chunk = chunks; chunk; chunk = chunk->next)
tools/objtool/check.c
70
struct instruction *next = next_insn_same_sec(file, insn);
tools/objtool/check.c
76
if (next && insn_func(next) == func)
tools/objtool/check.c
77
return next;
tools/objtool/disas.c
1117
for (alt = orig_insn->alts; alt; alt = alt->next) {
tools/objtool/disas.c
630
for (alt = insn->alts; alt; alt = alt->next) {
tools/objtool/elf.c
42
__node->next = __elf_table_entry(name, key); \
tools/objtool/elf.c
52
*head = node->next;
tools/objtool/elf.c
56
for (prev = NULL, cur = *head; cur; prev = cur, cur = cur->next) {
tools/objtool/elf.c
58
prev->next = cur->next;
tools/objtool/elf.c
76
obj = elf_list_entry(obj->member.next, typeof(*(obj)), member))
tools/objtool/include/objtool/arch.h
67
struct stack_op *next;
tools/objtool/include/objtool/check.h
49
struct alternative *next;
tools/objtool/include/objtool/elf.h
39
struct elf_hash_node *next;
tools/objtool/include/objtool/elf.h
445
static inline void set_sym_next_reloc(struct reloc *reloc, struct reloc *next)
tools/objtool/include/objtool/elf.h
449
reloc->_sym_next_reloc = (unsigned long)next | bit;
tools/objtool/trace.c
103
#define TRACE_CFI_REG_VAL(reg, prev, next) \
tools/objtool/trace.c
105
next.base, next.offset)
tools/objtool/trace.c
107
#define TRACE_CFI_REG_REF(reg, prev, next) \
tools/objtool/trace.c
109
next.base, next.offset)
tools/objtool/trace.c
15
#define TRACE_CFI_ATTR(attr, prev, next, fmt, ...) \
tools/objtool/trace.c
17
if ((prev)->attr != (next)->attr) \
tools/objtool/trace.c
21
#define TRACE_CFI_ATTR_BOOL(attr, prev, next) \
tools/objtool/trace.c
22
TRACE_CFI_ATTR(attr, prev, next, \
tools/objtool/trace.c
23
"%s", (next)->attr ? "true" : "false")
tools/objtool/trace.c
25
#define TRACE_CFI_ATTR_NUM(attr, prev, next, fmt) \
tools/objtool/trace.c
26
TRACE_CFI_ATTR(attr, prev, next, fmt, (next)->attr)
tools/perf/builtin-annotate.c
413
struct rb_node *nd = rb_first_cached(&hists->entries), *next;
tools/perf/builtin-annotate.c
487
next = rb_next(nd);
tools/perf/builtin-annotate.c
491
next = rb_prev(nd);
tools/perf/builtin-annotate.c
497
if (use_browser == 0 || next != NULL)
tools/perf/builtin-annotate.c
498
nd = next;
tools/perf/builtin-annotate.c
532
next = rb_next(nd);
tools/perf/builtin-annotate.c
536
next = rb_prev(nd);
tools/perf/builtin-annotate.c
542
if (next != NULL)
tools/perf/builtin-annotate.c
543
nd = next;
tools/perf/builtin-c2c.c
2385
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/builtin-c2c.c
2388
while (next) {
tools/perf/builtin-c2c.c
2391
he = rb_entry(next, struct hist_entry, rb_node);
tools/perf/builtin-c2c.c
2395
next = rb_next(&he->rb_node);
tools/perf/builtin-diff.c
529
struct rb_node *next;
tools/perf/builtin-diff.c
536
next = rb_first_cached(root);
tools/perf/builtin-diff.c
537
while (next != NULL) {
tools/perf/builtin-diff.c
538
struct hist_entry *he = rb_entry(next, struct hist_entry, rb_node_in);
tools/perf/builtin-diff.c
540
next = rb_next(&he->rb_node_in);
tools/perf/builtin-diff.c
587
struct rb_node *next = rb_first_cached(root);
tools/perf/builtin-diff.c
590
while (next != NULL) {
tools/perf/builtin-diff.c
591
struct hist_entry *he_pair = rb_entry(next, struct hist_entry,
tools/perf/builtin-diff.c
594
next = rb_next(&he_pair->rb_node_in);
tools/perf/builtin-diff.c
654
struct rb_node *next = rb_first_cached(root);
tools/perf/builtin-diff.c
656
while (next != NULL) {
tools/perf/builtin-diff.c
657
struct hist_entry *he = rb_entry(next, struct hist_entry,
tools/perf/builtin-diff.c
661
next = rb_next(&he->rb_node_in);
tools/perf/builtin-diff.c
673
struct rb_node *next;
tools/perf/builtin-diff.c
680
next = rb_first_cached(root);
tools/perf/builtin-diff.c
681
while (next != NULL) {
tools/perf/builtin-diff.c
687
he = rb_entry(next, struct hist_entry, rb_node_in);
tools/perf/builtin-diff.c
688
next = rb_next(&he->rb_node_in);
tools/perf/builtin-ftrace.c
886
goto next;
tools/perf/builtin-ftrace.c
898
goto next;
tools/perf/builtin-ftrace.c
902
goto next;
tools/perf/builtin-ftrace.c
933
next:
tools/perf/builtin-help.c
196
p = &((*p)->next);
tools/perf/builtin-help.c
216
new->next = man_viewer_info_list;
tools/perf/builtin-help.c
30
struct man_viewer_list *next;
tools/perf/builtin-help.c
35
struct man_viewer_info_list *next;
tools/perf/builtin-help.c
411
for (viewer = man_viewer_list; viewer; viewer = viewer->next)
tools/perf/builtin-help.c
64
for (viewer = man_viewer_info_list; viewer; viewer = viewer->next) {
tools/perf/builtin-inject.c
449
node = node->next;
tools/perf/builtin-kmem.c
1003
struct rb_node *next;
tools/perf/builtin-kmem.c
1011
next = rb_first(root);
tools/perf/builtin-kmem.c
1013
while (next && n_lines--) {
tools/perf/builtin-kmem.c
1014
struct alloc_stat *data = rb_entry(next, struct alloc_stat,
tools/perf/builtin-kmem.c
1044
next = rb_next(next);
tools/perf/builtin-kmem.c
1064
struct rb_node *next = rb_first(&page_alloc_sorted);
tools/perf/builtin-kmem.c
1080
while (next && n_lines--) {
tools/perf/builtin-kmem.c
1087
data = rb_entry(next, struct page_stat, node);
tools/perf/builtin-kmem.c
1100
next = rb_next(next);
tools/perf/builtin-kmem.c
1113
struct rb_node *next = rb_first(&page_caller_sorted);
tools/perf/builtin-kmem.c
1122
while (next && n_lines--) {
tools/perf/builtin-kmem.c
1129
data = rb_entry(next, struct page_stat, node);
tools/perf/builtin-kmem.c
1142
next = rb_next(next);
tools/perf/builtin-kwork.c
1110
if ((args == NULL) || (args->next == NULL))
tools/perf/builtin-kwork.c
1114
for (sym = args->next->symbol.symbols; sym != NULL; sym = sym->next) {
tools/perf/builtin-kwork.c
1906
struct rb_node *next;
tools/perf/builtin-kwork.c
1922
next = rb_first_cached(&kwork->sorted_work_root);
tools/perf/builtin-kwork.c
1923
while (next) {
tools/perf/builtin-kwork.c
1924
work = rb_entry(next, struct kwork_work, node);
tools/perf/builtin-kwork.c
1934
next = rb_next(next);
tools/perf/builtin-kwork.c
2001
struct rb_node *next;
tools/perf/builtin-kwork.c
2008
next = rb_first_cached(&class->work_root);
tools/perf/builtin-kwork.c
2009
while (next) {
tools/perf/builtin-kwork.c
2010
work = rb_entry(next, struct kwork_work, node);
tools/perf/builtin-kwork.c
2014
next = rb_next(next);
tools/perf/builtin-kwork.c
2073
struct rb_node *next;
tools/perf/builtin-kwork.c
2080
next = rb_first_cached(&class->work_root);
tools/perf/builtin-kwork.c
2081
while (next) {
tools/perf/builtin-kwork.c
2082
work = rb_entry(next, struct kwork_work, node);
tools/perf/builtin-kwork.c
2085
goto next;
tools/perf/builtin-kwork.c
2095
next:
tools/perf/builtin-kwork.c
2096
next = rb_next(next);
tools/perf/builtin-kwork.c
2153
struct rb_node *next;
tools/perf/builtin-kwork.c
2159
next = rb_first_cached(&kwork->sorted_work_root);
tools/perf/builtin-kwork.c
2160
while (next) {
tools/perf/builtin-kwork.c
2161
work = rb_entry(next, struct kwork_work, node);
tools/perf/builtin-kwork.c
2165
goto next;
tools/perf/builtin-kwork.c
2169
next:
tools/perf/builtin-kwork.c
2170
next = rb_next(next);
tools/perf/builtin-kwork.c
550
struct rb_node *next;
tools/perf/builtin-kwork.c
553
next = rb_first_cached(root);
tools/perf/builtin-kwork.c
554
while (next) {
tools/perf/builtin-kwork.c
555
work = rb_entry(next, struct kwork_work, node);
tools/perf/builtin-kwork.c
560
next = rb_next(next);
tools/perf/builtin-lock.c
884
goto next;
tools/perf/builtin-lock.c
893
next:
tools/perf/builtin-lock.c
932
goto next;
tools/perf/builtin-lock.c
935
goto next;
tools/perf/builtin-lock.c
939
next:
tools/perf/builtin-sched.c
3507
struct rb_node *next;
tools/perf/builtin-sched.c
3524
next = rb_first_cached(&sched->sorted_atom_root);
tools/perf/builtin-sched.c
3526
while (next) {
tools/perf/builtin-sched.c
3529
work_list = rb_entry(next, struct work_atoms, node);
tools/perf/builtin-sched.c
3531
next = rb_next(next);
tools/perf/builtin-sched.c
3545
while ((next = rb_first_cached(&sched->sorted_atom_root))) {
tools/perf/builtin-sched.c
3548
data = rb_entry(next, struct work_atoms, node);
tools/perf/builtin-sched.c
3549
rb_erase_cached(next, &sched->sorted_atom_root);
tools/perf/builtin-sched.c
858
struct task_desc *prev, __maybe_unused *next;
tools/perf/builtin-sched.c
884
next = register_pid(sched, next_pid, next_comm);
tools/perf/builtin-timechart.c
1002
p->next = new_list;
tools/perf/builtin-timechart.c
1010
cursor = cursor->next;
tools/perf/builtin-timechart.c
1012
prev->next = p;
tools/perf/builtin-timechart.c
1030
pwr = pwr->next;
tools/perf/builtin-timechart.c
1040
pwr = pwr->next;
tools/perf/builtin-timechart.c
1071
c = c->next;
tools/perf/builtin-timechart.c
1083
c = c->next;
tools/perf/builtin-timechart.c
1086
p = p->next;
tools/perf/builtin-timechart.c
110
struct per_pidcomm *next;
tools/perf/builtin-timechart.c
1105
we = we->next;
tools/perf/builtin-timechart.c
1132
sample = sample->next;
tools/perf/builtin-timechart.c
1134
c = c->next;
tools/perf/builtin-timechart.c
1136
p = p->next;
tools/perf/builtin-timechart.c
1156
c = c->next;
tools/perf/builtin-timechart.c
1161
for (sample = c->io_samples; sample; sample = sample->next) {
tools/perf/builtin-timechart.c
1248
c = c->next;
tools/perf/builtin-timechart.c
1250
p = p->next;
tools/perf/builtin-timechart.c
1269
c = c->next;
tools/perf/builtin-timechart.c
1291
sample = sample->next;
tools/perf/builtin-timechart.c
1305
c = c->next;
tools/perf/builtin-timechart.c
1307
p = p->next;
tools/perf/builtin-timechart.c
131
struct sample_wrapper *next;
tools/perf/builtin-timechart.c
1321
filt->next = process_filter;
tools/perf/builtin-timechart.c
1338
filt = filt->next;
tools/perf/builtin-timechart.c
1376
c = c->next;
tools/perf/builtin-timechart.c
1378
p = p->next;
tools/perf/builtin-timechart.c
1417
c = c->next;
tools/perf/builtin-timechart.c
1419
p = p->next;
tools/perf/builtin-timechart.c
143
struct cpu_sample *next;
tools/perf/builtin-timechart.c
1449
c = c->next;
tools/perf/builtin-timechart.c
1451
p = p->next;
tools/perf/builtin-timechart.c
162
struct io_sample *next;
tools/perf/builtin-timechart.c
177
struct power_event *next;
tools/perf/builtin-timechart.c
186
struct wake_event *next;
tools/perf/builtin-timechart.c
196
struct process_filter *next;
tools/perf/builtin-timechart.c
209
cursor = cursor->next;
tools/perf/builtin-timechart.c
214
cursor->next = tchart->all_data;
tools/perf/builtin-timechart.c
227
c->next = p->all;
tools/perf/builtin-timechart.c
248
c = c->next;
tools/perf/builtin-timechart.c
300
sample->next = c->samples;
tools/perf/builtin-timechart.c
375
pwr->next = tchart->power_events;
tools/perf/builtin-timechart.c
393
pwr->next = tchart->power_events;
tools/perf/builtin-timechart.c
442
we->next = tchart->wake_events;
tools/perf/builtin-timechart.c
715
pwr->next = tchart->power_events;
tools/perf/builtin-timechart.c
752
c->io_samples = prev->next;
tools/perf/builtin-timechart.c
763
sample->next = c->io_samples;
tools/perf/builtin-timechart.c
801
prev = sample->next;
tools/perf/builtin-timechart.c
838
sample->next = prev->next;
tools/perf/builtin-timechart.c
93
struct per_pid *next;
tools/perf/builtin-timechart.c
982
tchart->all_data = p->next;
tools/perf/builtin-timechart.c
983
p->next = NULL;
tools/perf/builtin-timechart.c
987
p->next = NULL;
tools/perf/builtin-timechart.c
997
p->next = prev->next;
tools/perf/builtin-timechart.c
998
prev->next = p;
tools/perf/builtin-top.c
399
struct rb_node *next;
tools/perf/builtin-top.c
416
next = rb_first_cached(&hists->entries);
tools/perf/builtin-top.c
417
while (next) {
tools/perf/builtin-top.c
418
n = rb_entry(next, struct hist_entry, rb_node);
tools/perf/builtin-top.c
423
next = rb_next(&n->rb_node);
tools/perf/builtin-trace.c
2088
for (; field; field = field->next, ++arg) {
tools/perf/builtin-trace.c
2214
sc->args = sc->args->next;
tools/perf/builtin-trace.c
2222
field = field->next;
tools/perf/builtin-trace.c
2438
field = field->next, ++arg.idx, bit <<= 1) {
tools/perf/builtin-trace.c
3207
for (; field && arg; field = field->next, ++syscall_arg.idx, bit <<= 1, ++arg) {
tools/perf/builtin-trace.c
3822
for (i = 0, field = sc->args; field; ++i, field = field->next) {
tools/perf/builtin-trace.c
3871
for (j = 0, field_tmp = sc->args; field_tmp; ++j, field_tmp = field_tmp->next) {
tools/perf/builtin-trace.c
3898
for (field = sc->args; field; field = field->next) {
tools/perf/builtin-trace.c
3917
field && candidate_field; field = field->next, candidate_field = candidate_field->next) {
tools/perf/builtin-trace.c
3962
for (candidate_field = candidate_field->next; candidate_field; candidate_field = candidate_field->next)
tools/perf/builtin-trace.c
4246
field = field->next, ++fmt) {
tools/perf/check-header_ignore_hunks/lib/list_sort.c
11
tail->next = b;
tools/perf/check-header_ignore_hunks/lib/list_sort.c
24
b = b->next;
tools/perf/jvmti/libjvmti.c
105
for (hdr = compile_info; hdr != NULL; hdr = hdr->next) {
tools/perf/jvmti/libjvmti.c
88
for (hdr = compile_info; hdr != NULL; hdr = hdr->next) {
tools/perf/tests/parse-events.c
819
term = list_entry(terms->terms.next, struct parse_events_term, list);
tools/perf/tests/parse-events.c
828
term = list_entry(term->list.next, struct parse_events_term, list);
tools/perf/tests/parse-events.c
837
term = list_entry(term->list.next, struct parse_events_term, list);
tools/perf/tests/parse-events.c
846
term = list_entry(term->list.next, struct parse_events_term, list);
tools/perf/tests/parse-events.c
855
term = list_entry(term->list.next, struct parse_events_term, list);
tools/perf/tests/parse-events.c
864
term = list_entry(term->list.next, struct parse_events_term, list);
tools/perf/tests/parse-events.c
879
term = list_entry(term->list.next, struct parse_events_term, list);
tools/perf/tests/parse-events.c
893
term = list_entry(term->list.next, struct parse_events_term, list);
tools/perf/tests/switch-tracking.c
255
node = list_entry(events->next, struct event_node, list);
tools/perf/ui/browser.c
113
pos = ui_browser__list_head_filter_entries(browser, head->next);
tools/perf/ui/browser.c
129
pos = ui_browser__list_head_filter_entries(browser, pos->next);
tools/perf/ui/browser.c
517
browser->top = ui_browser__list_head_filter_entries(browser, head->next);
tools/perf/ui/browser.c
84
pos = pos->next;
tools/perf/ui/browsers/annotate-data.c
307
struct browser_entry *entry, *next;
tools/perf/ui/browsers/annotate-data.c
324
next = browser__next_entry(uib, entry);
tools/perf/ui/browsers/annotate-data.c
325
if (next == entry)
tools/perf/ui/browsers/annotate-data.c
328
entry = next;
tools/perf/ui/browsers/hists.c
1030
struct rb_node *next = rb_next(node);
tools/perf/ui/browsers/hists.c
1038
goto next;
tools/perf/ui/browsers/hists.c
1089
next:
tools/perf/ui/browsers/hists.c
1092
node = next;
tools/perf/ui/browsers/hists.c
1119
struct rb_node *next = rb_next(node);
tools/perf/ui/browsers/hists.c
1158
node = next;
tools/perf/ui/browsers/hists.c
2976
goto next;
tools/perf/ui/browsers/hists.c
2990
next:
tools/perf/ui/browsers/hists.c
341
chain->has_children = chain->list.next != &child->val ||
tools/perf/ui/browsers/hists.c
344
chain->has_children = chain->list.next == &child->val &&
tools/perf/ui/browsers/hists.c
3562
if (pos->core.node.next == &evlist->core.entries)
tools/perf/ui/browsers/hists.c
357
chain = list_entry(node->val.next, struct callchain_list, list);
tools/perf/ui/browsers/hists.c
926
struct rb_node *next = rb_next(node);
tools/perf/ui/browsers/hists.c
952
goto next;
tools/perf/ui/browsers/hists.c
978
next:
tools/perf/ui/browsers/hists.c
981
node = next;
tools/perf/ui/hist.c
1014
goto next;
tools/perf/ui/hist.c
1018
next:
tools/perf/ui/hist.c
831
if (fmt->sort_list.next == NULL)
tools/perf/ui/hist.c
995
goto next;
tools/perf/ui/hist.c
999
next:
tools/perf/ui/progress.c
22
if (p->curr >= p->next) {
tools/perf/ui/progress.c
25
p->next += nr * p->step;
tools/perf/ui/progress.c
34
p->next = p->step = total / 16 ?: 1;
tools/perf/ui/progress.h
11
u64 curr, next, step, total;
tools/perf/ui/stdio/hist.c
117
struct rb_node *node, *next;
tools/perf/ui/stdio/hist.c
146
next = rb_next(node);
tools/perf/ui/stdio/hist.c
147
if (!next && (callchain_param.mode != CHAIN_GRAPH_REL || !remaining))
tools/perf/ui/stdio/hist.c
173
node = next;
tools/perf/ui/tui/progress.c
12
p->next = p->step = p->total / (SLtt_Screen_Cols - 2) ?: 1;
tools/perf/util/annotate-data.c
1057
var_types = var_types->next;
tools/perf/util/annotate-data.c
1064
struct die_var_type *next = var_types->next;
tools/perf/util/annotate-data.c
1067
var_types = next;
tools/perf/util/annotate-data.c
769
for (pos = var_types; pos; pos = pos->next) {
tools/perf/util/annotate-data.c
875
for (var = var_types; var != NULL; var = var->next) {
tools/perf/util/annotate.c
1138
struct rb_node *next;
tools/perf/util/annotate.c
1141
next = rb_next(node);
tools/perf/util/annotate.c
1145
node = next;
tools/perf/util/annotate.c
1305
queue = list_entry(queue->node.next, typeof(*queue), node);
tools/perf/util/annotate.c
2740
struct disasm_line *next;
tools/perf/util/annotate.c
2745
next = list_next_entry(curr, al.node);
tools/perf/util/annotate.c
2746
while (next->al.offset == -1 &&
tools/perf/util/annotate.c
2747
next != list_last_entry(sources, struct disasm_line, al.node))
tools/perf/util/annotate.c
2748
next = list_next_entry(next, al.node);
tools/perf/util/annotate.c
2750
if (next->al.offset == -1)
tools/perf/util/annotate.c
2753
return next;
tools/perf/util/annotate.c
2760
struct disasm_line *next;
tools/perf/util/annotate.c
2771
next = annotation__next_asm_line(notes, dl);
tools/perf/util/annotate.c
2772
if (next == NULL)
tools/perf/util/annotate.c
2775
addr = ip + (next->al.offset - dl->al.offset) + offset;
tools/perf/util/annotate.c
950
struct annotation_line *al, *next;
tools/perf/util/annotate.c
960
next = annotation_line__next(al, &notes->src->source);
tools/perf/util/annotate.c
961
end = next ? next->offset : len;
tools/perf/util/auxtrace.c
1103
return list_entry(buffer->list.next, struct auxtrace_buffer,
tools/perf/util/auxtrace.c
1108
return list_entry(queue->head.next, struct auxtrace_buffer,
tools/perf/util/auxtrace.c
478
buffer = list_entry(queues->queue_array[i].head.next,
tools/perf/util/block-info.c
443
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/block-info.c
453
while (next) {
tools/perf/util/block-info.c
454
he = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/block-info.c
457
next = rb_next(&he->rb_node);
tools/perf/util/block-range.c
113
next = rb_entry(n, struct block_range, node);
tools/perf/util/block-range.c
115
if (next->start <= end) { /* add head: [start...][n->start...] */
tools/perf/util/block-range.c
122
.end = next->start - 1,
tools/perf/util/block-range.c
127
rb_link_left_of_node(&head->node, &next->node);
tools/perf/util/block-range.c
238
next = block_range__next(entry);
tools/perf/util/block-range.c
239
if (!next)
tools/perf/util/block-range.c
245
if (end < next->start) { /* add tail: [...e->end][...end] */
tools/perf/util/block-range.c
270
if (entry->end + 1 != next->start) {
tools/perf/util/block-range.c
277
.end = next->start - 1,
tools/perf/util/block-range.c
282
rb_link_left_of_node(&hole->node, &next->node);
tools/perf/util/block-range.c
287
entry = next;
tools/perf/util/block-range.c
81
struct block_range *next, *entry = NULL;
tools/perf/util/bpf_lock_contention.c
532
goto next;
tools/perf/util/bpf_lock_contention.c
542
next:
tools/perf/util/bpf_lock_contention.c
772
goto next;
tools/perf/util/bpf_lock_contention.c
788
goto next;
tools/perf/util/bpf_lock_contention.c
802
goto next;
tools/perf/util/bpf_lock_contention.c
824
next:
tools/perf/util/bpf_skel/kwork_top.bpf.c
173
struct task_struct *prev, *next;
tools/perf/util/bpf_skel/kwork_top.bpf.c
176
next = (struct task_struct *)ctx[2];
tools/perf/util/bpf_skel/kwork_top.bpf.c
189
on_sched_in(next, ts);
tools/perf/util/bpf_skel/off_cpu.bpf.c
257
struct task_struct *next, int state)
tools/perf/util/bpf_skel/off_cpu.bpf.c
266
goto next;
tools/perf/util/bpf_skel/off_cpu.bpf.c
274
goto next;
tools/perf/util/bpf_skel/off_cpu.bpf.c
290
next:
tools/perf/util/bpf_skel/off_cpu.bpf.c
291
pelem = bpf_task_storage_get(&tstamp, next, NULL, 0);
tools/perf/util/bpf_skel/off_cpu.bpf.c
295
.pid = next->pid,
tools/perf/util/bpf_skel/off_cpu.bpf.c
296
.tgid = next->tgid,
tools/perf/util/bpf_skel/off_cpu.bpf.c
299
.cgroup_id = needs_cgroup ? get_cgroup_id(next) : 0,
tools/perf/util/bpf_skel/off_cpu.bpf.c
355
struct task_struct *prev, *next;
tools/perf/util/bpf_skel/off_cpu.bpf.c
362
next = (struct task_struct *)ctx[2];
tools/perf/util/bpf_skel/off_cpu.bpf.c
369
return off_cpu_stat(ctx, prev, next, prev_state & 0xff);
tools/perf/util/call-path.c
58
if (cpr->next < cpr->sz) {
tools/perf/util/call-path.c
69
n = cpr->next++ & CALL_PATH_BLOCK_MASK;
tools/perf/util/call-path.h
57
size_t next;
tools/perf/util/callchain.c
1119
cursor->last = &node->next;
tools/perf/util/callchain.c
1606
struct callchain_cursor_node *node, *next;
tools/perf/util/callchain.c
1609
for (node = cursor->first; node != NULL; node = next) {
tools/perf/util/callchain.c
1610
next = node->next;
tools/perf/util/callchain.c
1680
for (node = cursor->first; node != NULL; node = node->next)
tools/perf/util/callchain.c
1782
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/callchain.c
1785
while (next) {
tools/perf/util/callchain.c
1786
struct hist_entry *he = rb_entry(next, struct hist_entry,
tools/perf/util/callchain.c
1790
next = rb_next(&he->rb_node);
tools/perf/util/callchain.c
199
goto next;
tools/perf/util/callchain.c
209
goto next;
tools/perf/util/callchain.c
232
next:
tools/perf/util/callchain.c
841
new->val.next = &to_split->list;
tools/perf/util/callchain.c
844
old_tail->next = &new->val;
tools/perf/util/callchain.h
156
struct callchain_cursor_node *next;
tools/perf/util/callchain.h
231
cursor->curr = cursor->curr->next;
tools/perf/util/data-convert-bt.c
1040
for (field = fields; field; field = field->next) {
tools/perf/util/data-convert-bt.c
421
for (field = fields; field; field = field->next) {
tools/perf/util/demangle-rust-v0.c
107
size_t next;
tools/perf/util/demangle-rust-v0.c
1675
if (printer->status == DemangleOk && /* safety */ printer->parser.next > 0) {
tools/perf/util/demangle-rust-v0.c
1676
printer->parser.next--;
tools/perf/util/demangle-rust-v0.c
176
char next = parser.sym[parser.next];
tools/perf/util/demangle-rust-v0.c
179
if (parser.next < parser.sym_len && next >= 'A' && next <= 'Z') {
tools/perf/util/demangle-rust-v0.c
187
*rest = parser.sym + parser.next;
tools/perf/util/demangle-rust-v0.c
677
if (parser->next == parser->sym_len) {
tools/perf/util/demangle-rust-v0.c
680
return parser->sym[parser->next];
tools/perf/util/demangle-rust-v0.c
687
parser->next++;
tools/perf/util/demangle-rust-v0.c
697
if (parser->next == parser->sym_len) {
tools/perf/util/demangle-rust-v0.c
700
return parser->sym[parser->next++];
tools/perf/util/demangle-rust-v0.c
704
static NODISCARD demangle_status parser_ch(struct parser *parser, uint8_t *next) {
tools/perf/util/demangle-rust-v0.c
706
if (parser->next == parser->sym_len) {
tools/perf/util/demangle-rust-v0.c
709
*next = parser->sym[parser->next++];
tools/perf/util/demangle-rust-v0.c
720
size_t start = parser->next;
tools/perf/util/demangle-rust-v0.c
731
buf->len = parser->next - start - 1; // skip final _
tools/perf/util/demangle-rust-v0.c
739
parser->next++;
tools/perf/util/demangle-rust-v0.c
750
parser->next++;
tools/perf/util/demangle-rust-v0.c
754
parser->next++;
tools/perf/util/demangle-rust-v0.c
758
parser->next++;
tools/perf/util/demangle-rust-v0.c
818
uint8_t next = parser_next(parser);
tools/perf/util/demangle-rust-v0.c
819
if ('A' <= next && next <= 'Z') {
tools/perf/util/demangle-rust-v0.c
820
*out = next;
tools/perf/util/demangle-rust-v0.c
822
} else if ('a' <= next && next <= 'z') {
tools/perf/util/demangle-rust-v0.c
831
size_t start = parser->next;
tools/perf/util/demangle-rust-v0.c
847
.next = (size_t)i,
tools/perf/util/demangle-rust-v0.c
887
size_t start = parser->next;
tools/perf/util/demangle-rust-v0.c
888
if (parser->sym_len - parser->next < len) {
tools/perf/util/demangle-rust-v0.c
891
parser->next += len;
tools/perf/util/dso.c
898
struct rb_node *next = rb_first(root);
tools/perf/util/dso.c
901
while (next) {
tools/perf/util/dso.c
904
cache = rb_entry(next, struct dso_cache, rb_node);
tools/perf/util/dso.c
905
next = rb_next(&cache->rb_node);
tools/perf/util/dwarf-aux.c
1650
vt->next = *var_types;
tools/perf/util/dwarf-aux.c
1712
vt->next = *var_types;
tools/perf/util/dwarf-aux.h
146
struct die_var_type *next;
tools/perf/util/env.c
177
struct rb_node *next;
tools/perf/util/env.c
182
next = rb_first(root);
tools/perf/util/env.c
184
while (next) {
tools/perf/util/env.c
187
node = rb_entry(next, struct bpf_prog_info_node, rb_node);
tools/perf/util/env.c
188
next = rb_next(&node->rb_node);
tools/perf/util/env.c
198
next = rb_first(root);
tools/perf/util/env.c
200
while (next) {
tools/perf/util/env.c
203
node = rb_entry(next, struct btf_node, rb_node);
tools/perf/util/env.c
204
next = rb_next(&node->rb_node);
tools/perf/util/evsel.h
432
return list_entry(evsel->core.node.next, struct evsel, core.node);
tools/perf/util/evsel.h
493
for ((_evsel) = list_entry((_leader)->core.node.next, struct evsel, core.node); \
tools/perf/util/evsel.h
496
(_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node))
tools/perf/util/evsel.h
506
(_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node))
tools/perf/util/evsel_fprintf.c
100
field = field->next;
tools/perf/util/evsel_fprintf.c
103
field = field->next;
tools/perf/util/evsel_fprintf.c
150
goto next;
tools/perf/util/evsel_fprintf.c
208
next:
tools/perf/util/hashmap.c
142
prev_ptr = &cur->next, cur = cur->next) {
tools/perf/util/hashmap.c
27
entry->next = *pprev;
tools/perf/util/hashmap.c
34
*pprev = entry->next;
tools/perf/util/hashmap.c
35
entry->next = NULL;
tools/perf/util/hashmap.h
170
for (cur = (map)->buckets[bkt]; cur; cur = cur->next)
tools/perf/util/hashmap.h
183
cur && ({tmp = cur->next; true; }); \
tools/perf/util/hashmap.h
197
cur = cur->next) \
tools/perf/util/hashmap.h
204
cur && ({ tmp = cur->next; true; }); \
tools/perf/util/hashmap.h
69
struct hashmap_entry *next;
tools/perf/util/header.c
1035
struct rb_node *next;
tools/perf/util/header.c
1046
next = rb_first(root);
tools/perf/util/header.c
1047
while (next) {
tools/perf/util/header.c
1051
node = rb_entry(next, struct bpf_prog_info_node, rb_node);
tools/perf/util/header.c
1052
next = rb_next(&node->rb_node);
tools/perf/util/header.c
1077
struct rb_node *next;
tools/perf/util/header.c
1089
next = rb_first(root);
tools/perf/util/header.c
1090
while (next) {
tools/perf/util/header.c
1093
node = rb_entry(next, struct btf_node, rb_node);
tools/perf/util/header.c
1094
next = rb_next(&node->rb_node);
tools/perf/util/header.c
1988
struct rb_node *next;
tools/perf/util/header.c
1993
next = rb_first(root);
tools/perf/util/header.c
1995
if (!next)
tools/perf/util/header.c
1998
while (next) {
tools/perf/util/header.c
2001
node = rb_entry(next, struct bpf_prog_info_node, rb_node);
tools/perf/util/header.c
2002
next = rb_next(&node->rb_node);
tools/perf/util/header.c
2015
struct rb_node *next;
tools/perf/util/header.c
2020
next = rb_first(root);
tools/perf/util/header.c
2022
if (!next)
tools/perf/util/header.c
2025
while (next) {
tools/perf/util/header.c
2028
node = rb_entry(next, struct btf_node, rb_node);
tools/perf/util/header.c
2029
next = rb_next(&node->rb_node);
tools/perf/util/hist.c
1829
struct rb_node *next;
tools/perf/util/hist.c
1840
next = rb_first_cached(root);
tools/perf/util/hist.c
1842
while (next) {
tools/perf/util/hist.c
1845
n = rb_entry(next, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
1846
next = rb_next(&n->rb_node_in);
tools/perf/util/hist.c
2066
struct rb_node *next;
tools/perf/util/hist.c
2095
next = rb_first_cached(root);
tools/perf/util/hist.c
2098
while (next) {
tools/perf/util/hist.c
2099
n = rb_entry(next, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
2100
next = rb_next(&n->rb_node_in);
tools/perf/util/hist.c
2242
goto next;
tools/perf/util/hist.c
2249
next:
tools/perf/util/hist.c
265
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/hist.c
271
while (next && row++ < max_rows) {
tools/perf/util/hist.c
272
n = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/hist.c
275
next = rb_next(&n->rb_node);
tools/perf/util/hist.c
473
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/hist.c
476
while (next) {
tools/perf/util/hist.c
477
n = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/hist.c
478
next = rb_next(&n->rb_node);
tools/perf/util/hist.c
489
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/hist.c
492
while (next) {
tools/perf/util/hist.c
493
n = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/hist.c
494
next = rb_next(&n->rb_node);
tools/perf/util/hist.c
502
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/hist.c
506
while (next) {
tools/perf/util/hist.c
507
n = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/hist.c
511
next = rb_next(&n->rb_node);
tools/perf/util/hist.h
334
return list_entry(he->pairs.node.next, struct hist_entry, pairs.node);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3209
next:
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3224
goto next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3324
goto next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3380
goto next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3391
goto next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3414
goto next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3452
goto next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3492
goto next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3515
goto next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3873
unsigned char *next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3884
next = memmem(decoder->buf, decoder->len, INTEL_PT_PSB_STR,
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3886
if (!next) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3901
decoder->pkt_step = next - decoder->buf;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4091
unsigned char *next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4093
next = memmem(*buf, *len, INTEL_PT_PSB_STR, INTEL_PT_PSB_LEN);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4094
if (next) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4095
*len -= next - *buf;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4096
*buf = next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4116
unsigned char *next;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4121
next = memmem(*buf + 1, *len - 1, INTEL_PT_PSB_STR, INTEL_PT_PSB_LEN);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4122
if (next) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4123
*len -= next - *buf;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4124
*buf = next;
tools/perf/util/intel-tpebs.c
584
if (t && &t->nd == tpebs_results.next) {
tools/perf/util/machine.c
1261
struct rb_node *next = rb_first_cached(&machines->guests);
tools/perf/util/machine.c
1265
while (next) {
tools/perf/util/machine.c
1266
struct machine *pos = rb_entry(next, struct machine, rb_node);
tools/perf/util/machine.c
1268
next = rb_next(&pos->rb_node);
tools/perf/util/machine.c
1588
struct map *next = maps__find_next_entry(machine__kernel_maps(machine),
tools/perf/util/machine.c
1591
if (next) {
tools/perf/util/machine.c
1592
machine__set_kernel_mmap(machine, start, map__start(next));
tools/perf/util/machine.c
1593
map__put(next);
tools/perf/util/machine.c
2381
cursor->curr = cursor->curr->next;
tools/perf/util/machine.c
2421
cursor->curr = cursor->curr->next;
tools/perf/util/machine.c
2453
cursor->curr = cursor->curr->next;
tools/perf/util/machine.c
3100
struct thread_list *pos, *next;
tools/perf/util/machine.c
3102
list_for_each_entry_safe(pos, next, list, list) {
tools/perf/util/maps.c
966
struct map *next = NULL;
tools/perf/util/maps.c
970
next = maps_by_address[i + 1];
tools/perf/util/maps.c
972
if (!next || map__start(next) >= map__end(new)) {
tools/perf/util/ordered-events.c
141
new = list_entry(cache->next, struct ordered_event, list);
tools/perf/util/ordered-events.c
297
first = list_entry(head->next, struct ordered_event, list);
tools/perf/util/ordered-events.c
42
p = last->list.next;
tools/perf/util/print-events.c
290
struct rb_node *node, *next;
tools/perf/util/print-events.c
300
for (node = rb_first_cached(&groups.entries); node; node = next) {
tools/perf/util/print-events.c
312
next = rb_next(node);
tools/perf/util/probe-event.c
1801
fieldp = &(*fieldp)->next;
tools/perf/util/probe-event.c
2053
field = field->next;
tools/perf/util/probe-event.c
2147
if (ref->next) {
tools/perf/util/probe-event.c
2148
depth = __synthesize_probe_trace_arg_ref(ref->next, buf,
tools/perf/util/probe-event.c
2176
ref = ref->next;
tools/perf/util/probe-event.c
2400
struct perf_probe_arg_field *field, *next;
tools/perf/util/probe-event.c
2414
next = field->next;
tools/perf/util/probe-event.c
2417
field = next;
tools/perf/util/probe-event.c
2462
field = field->next;
tools/perf/util/probe-event.c
2463
ppfield = &((*ppfield)->next);
tools/perf/util/probe-event.c
2500
struct probe_trace_arg_ref *ref, *next;
tools/perf/util/probe-event.c
2514
next = ref->next;
tools/perf/util/probe-event.c
2516
ref = next;
tools/perf/util/probe-event.c
2728
goto next;
tools/perf/util/probe-event.c
2732
goto next;
tools/perf/util/probe-event.c
2737
next:
tools/perf/util/probe-event.h
42
struct probe_trace_arg_ref *next; /* Next reference */
tools/perf/util/probe-event.h
79
struct perf_probe_arg_field *next; /* Next field */
tools/perf/util/probe-finder.c
262
ref_ptr = &(*ref_ptr)->next;
tools/perf/util/probe-finder.c
356
(*ref_ptr)->next = ref;
tools/perf/util/probe-finder.c
362
goto next;
tools/perf/util/probe-finder.c
387
(*ref_ptr)->next = ref;
tools/perf/util/probe-finder.c
440
next:
tools/perf/util/probe-finder.c
442
if (field->next)
tools/perf/util/probe-finder.c
444
field->next, &ref, die_mem, user_access);
tools/perf/util/rblist.c
112
struct rb_node *pos, *next = rb_first_cached(&rblist->entries);
tools/perf/util/rblist.c
114
while (next) {
tools/perf/util/rblist.c
115
pos = next;
tools/perf/util/rblist.c
116
next = rb_next(pos);
tools/perf/util/scripting-engines/trace-event-perl.c
108
if (field->next)
tools/perf/util/scripting-engines/trace-event-perl.c
109
define_symbolic_values(field->next, ev_name, field_name);
tools/perf/util/scripting-engines/trace-event-perl.c
166
if (field->next)
tools/perf/util/scripting-engines/trace-event-perl.c
167
define_flag_values(field->next, ev_name, field_name);
tools/perf/util/scripting-engines/trace-event-perl.c
255
if (args->next)
tools/perf/util/scripting-engines/trace-event-perl.c
256
define_event_symbols(event, ev_name, args->next);
tools/perf/util/scripting-engines/trace-event-perl.c
396
for (field = event->format.fields; field; field = field->next) {
tools/perf/util/scripting-engines/trace-event-perl.c
641
for (f = event->format.fields; f; f = f->next) {
tools/perf/util/scripting-engines/trace-event-perl.c
660
for (f = event->format.fields; f; f = f->next) {
tools/perf/util/scripting-engines/trace-event-perl.c
684
for (f = event->format.fields; f; f = f->next) {
tools/perf/util/scripting-engines/trace-event-python.c
1017
for (field = event->format.fields; field; field = field->next) {
tools/perf/util/scripting-engines/trace-event-python.c
2062
for (f = event->format.fields; f; f = f->next) {
tools/perf/util/scripting-engines/trace-event-python.c
2087
for (f = event->format.fields; f; f = f->next) {
tools/perf/util/scripting-engines/trace-event-python.c
2112
for (f = event->format.fields; f; f = f->next) {
tools/perf/util/scripting-engines/trace-event-python.c
220
if (field->next)
tools/perf/util/scripting-engines/trace-event-python.c
221
define_values(field_type, field->next, ev_name, field_name);
tools/perf/util/scripting-engines/trace-event-python.c
317
if (args->next)
tools/perf/util/scripting-engines/trace-event-python.c
318
define_event_symbols(event, ev_name, args->next);
tools/perf/util/session.c
232
static void perf_decomp__release_events(struct decomp *next)
tools/perf/util/session.c
238
decomp = next;
tools/perf/util/session.c
241
next = decomp->next;
tools/perf/util/session.h
97
struct decomp *next;
tools/perf/util/sort.c
3405
field = field->next;
tools/perf/util/srccode.c
107
srcfile_list.next != &srcfile_list) {
tools/perf/util/srcline.c
415
struct rb_node *next = rb_first_cached(tree);
tools/perf/util/srcline.c
417
while (next) {
tools/perf/util/srcline.c
418
pos = rb_entry(next, struct srcline_node, rb_node);
tools/perf/util/srcline.c
419
next = rb_next(&pos->rb_node);
tools/perf/util/srcline.c
499
struct rb_node *next = rb_first_cached(tree);
tools/perf/util/srcline.c
501
while (next) {
tools/perf/util/srcline.c
502
pos = rb_entry(next, struct inline_node, rb_node);
tools/perf/util/srcline.c
503
next = rb_next(&pos->rb_node);
tools/perf/util/stream.c
114
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/stream.c
116
while (next) {
tools/perf/util/stream.c
119
he = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/stream.c
121
next = rb_next(&he->rb_node);
tools/perf/util/symbol-elf.c
2813
size_t shstrndx, next;
tools/perf/util/symbol-elf.c
2842
for (offset = 0; (next = gelf_getnote(data, offset, &nhdr, &name_off,
tools/perf/util/symbol-elf.c
2843
&desc_off)) > 0; offset = next) {
tools/perf/util/symbol.c
1060
struct rb_node *next = rb_first(modules);
tools/perf/util/symbol.c
1062
while (next) {
tools/perf/util/symbol.c
1063
mi = rb_entry(next, struct module_info, rb_node);
tools/perf/util/symbol.c
1064
next = rb_next(&mi->rb_node);
tools/perf/util/symbol.c
1387
replacement_map = list_entry(md.maps.next, struct map_list_node, node)->map;
tools/perf/util/symbol.c
1409
struct map_list_node *new_node = list_entry(md.maps.next, struct map_list_node, node);
tools/perf/util/symbol.c
1466
list_node = list_entry(md.maps.next, struct map_list_node, node);
tools/perf/util/symbol.c
210
struct symbol *curr, *next;
tools/perf/util/symbol.c
224
next = rb_entry(nd, struct symbol, rb_node);
tools/perf/util/symbol.c
225
if (curr->start != next->start)
tools/perf/util/symbol.c
228
if (choose_best_symbol(curr, next) == SYMBOL_A) {
tools/perf/util/symbol.c
229
if (next->type == STT_GNU_IFUNC)
tools/perf/util/symbol.c
231
rb_erase_cached(&next->rb_node, symbols);
tools/perf/util/symbol.c
232
symbol__delete(next);
tools/perf/util/symbol.c
236
next->ifunc_alias = true;
tools/perf/util/symbol.c
350
struct rb_node *next = rb_first_cached(symbols);
tools/perf/util/symbol.c
352
while (next) {
tools/perf/util/symbol.c
353
pos = rb_entry(next, struct symbol, rb_node);
tools/perf/util/symbol.c
354
next = rb_next(&pos->rb_node);
tools/perf/util/symbol.c
809
struct rb_node *next = rb_first_cached(root);
tools/perf/util/symbol.c
816
while (next) {
tools/perf/util/symbol.c
821
pos = rb_entry(next, struct symbol, rb_node);
tools/perf/util/symbol.c
822
next = rb_next(&pos->rb_node);
tools/perf/util/symbol.c
866
struct rb_node *next = rb_first_cached(root);
tools/perf/util/symbol.c
877
while (next) {
tools/perf/util/symbol.c
880
pos = rb_entry(next, struct symbol, rb_node);
tools/perf/util/symbol.c
881
next = rb_next(&pos->rb_node);
tools/perf/util/tool.c
73
session->active_decomp->decomp_last->next = decomp;
tools/perf/util/trace-event-info.c
149
tps = tps->next;
tools/perf/util/trace-event-info.c
247
tps = tps->next;
tools/perf/util/trace-event-info.c
39
struct tracepoint_path *next;
tools/perf/util/trace-event-info.c
397
tps = tps->next;
tools/perf/util/trace-event-info.c
425
goto next;
tools/perf/util/trace-event-info.c
462
next:
tools/perf/util/trace-event-info.c
517
ppath->next = tracepoint_name_to_path(pos->name);
tools/perf/util/trace-event-info.c
518
if (ppath->next)
tools/perf/util/trace-event-info.c
519
goto next;
tools/perf/util/trace-event-info.c
528
ppath->next = tracepoint_id_to_path(pos->core.attr.config);
tools/perf/util/trace-event-info.c
529
if (!ppath->next) {
tools/perf/util/trace-event-info.c
532
put_tracepoints_path(path.next);
tools/perf/util/trace-event-info.c
535
next:
tools/perf/util/trace-event-info.c
536
ppath = ppath->next;
tools/perf/util/trace-event-info.c
539
return nr_tracepoints > 0 ? path.next : NULL;
tools/perf/util/trace-event-parse.c
134
for (; sym; sym = sym->next) {
tools/perf/util/trace-event-parse.c
219
for (arg = event->print_fmt.args; arg; arg = arg->next) {
tools/perf/util/trace-event-parse.c
238
char *next = NULL;
tools/perf/util/trace-event-parse.c
242
line = strtok_r(file, "\n", &next);
tools/perf/util/trace-event-parse.c
252
line = strtok_r(NULL, "\n", &next);
tools/perf/util/trace-event-parse.c
263
char *next = NULL;
tools/perf/util/trace-event-parse.c
266
line = strtok_r(file, "\n", &next);
tools/perf/util/trace-event-parse.c
270
line = strtok_r(NULL, "\n", &next);
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
20
struct osl_table_info *next;
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
280
struct osl_table_info *next;
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
294
next = gbl_table_list_head;
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
296
if (ACPI_COMPARE_NAMESEG(next->signature, signature)) {
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
297
if (next->instance == instance) {
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
300
if (next->instance >= next_instance) {
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
301
next_instance = next->instance + 1;
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
305
if (!next->next) {
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
308
next = next->next;
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
310
next->next = new_info;
tools/power/acpi/os_specific/service_layers/oslinuxtbl.c
373
info = info->next;
tools/power/cpupower/lib/cpufreq.c
308
current->next = malloc(sizeof(*current));
tools/power/cpupower/lib/cpufreq.c
309
if (!current->next)
tools/power/cpupower/lib/cpufreq.c
311
current = current->next;
tools/power/cpupower/lib/cpufreq.c
319
current->next = NULL;
tools/power/cpupower/lib/cpufreq.c
335
current = first->next;
tools/power/cpupower/lib/cpufreq.c
346
struct cpufreq_available_governors *tmp, *next;
tools/power/cpupower/lib/cpufreq.c
353
next = tmp->next;
tools/power/cpupower/lib/cpufreq.c
357
tmp = next;
tools/power/cpupower/lib/cpufreq.c
385
current->next = malloc(sizeof(*current));
tools/power/cpupower/lib/cpufreq.c
386
if (!current->next)
tools/power/cpupower/lib/cpufreq.c
388
current = current->next;
tools/power/cpupower/lib/cpufreq.c
396
current->next = NULL;
tools/power/cpupower/lib/cpufreq.c
411
current = first->next;
tools/power/cpupower/lib/cpufreq.c
441
current->next = malloc(sizeof(*current));
tools/power/cpupower/lib/cpufreq.c
442
if (!current->next)
tools/power/cpupower/lib/cpufreq.c
444
current = current->next;
tools/power/cpupower/lib/cpufreq.c
452
current->next = NULL;
tools/power/cpupower/lib/cpufreq.c
467
current = first->next;
tools/power/cpupower/lib/cpufreq.c
476
struct cpufreq_available_frequencies *tmp, *next;
tools/power/cpupower/lib/cpufreq.c
483
next = tmp->next;
tools/power/cpupower/lib/cpufreq.c
485
tmp = next;
tools/power/cpupower/lib/cpufreq.c
516
current->next = malloc(sizeof(*current));
tools/power/cpupower/lib/cpufreq.c
517
if (!current->next)
tools/power/cpupower/lib/cpufreq.c
519
current = current->next;
tools/power/cpupower/lib/cpufreq.c
527
current->next = NULL;
tools/power/cpupower/lib/cpufreq.c
543
current = first->next;
tools/power/cpupower/lib/cpufreq.c
557
struct cpufreq_affected_cpus *tmp, *next;
tools/power/cpupower/lib/cpufreq.c
564
next = tmp->next;
tools/power/cpupower/lib/cpufreq.c
566
tmp = next;
tools/power/cpupower/lib/cpufreq.c
749
current->next = malloc(sizeof(*current));
tools/power/cpupower/lib/cpufreq.c
750
if (!current->next)
tools/power/cpupower/lib/cpufreq.c
752
current = current->next;
tools/power/cpupower/lib/cpufreq.c
760
current->next = NULL;
tools/power/cpupower/lib/cpufreq.c
778
current = first->next;
tools/power/cpupower/lib/cpufreq.c
787
struct cpufreq_stats *tmp, *next;
tools/power/cpupower/lib/cpufreq.c
794
next = tmp->next;
tools/power/cpupower/lib/cpufreq.c
796
tmp = next;
tools/power/cpupower/lib/cpufreq.h
19
struct cpufreq_available_governors *next;
tools/power/cpupower/lib/cpufreq.h
25
struct cpufreq_available_frequencies *next;
tools/power/cpupower/lib/cpufreq.h
32
struct cpufreq_affected_cpus *next;
tools/power/cpupower/lib/cpufreq.h
39
struct cpufreq_stats *next;
tools/power/cpupower/utils/cpufreq-info.c
235
while (freqs->next) {
tools/power/cpupower/utils/cpufreq-info.c
238
freqs = freqs->next;
tools/power/cpupower/utils/cpufreq-info.c
362
while (governors->next) {
tools/power/cpupower/utils/cpufreq-info.c
364
governors = governors->next;
tools/power/cpupower/utils/cpufreq-info.c
384
while (cpus->next) {
tools/power/cpupower/utils/cpufreq-info.c
386
cpus = cpus->next;
tools/power/cpupower/utils/cpufreq-info.c
405
while (cpus->next) {
tools/power/cpupower/utils/cpufreq-info.c
407
cpus = cpus->next;
tools/power/cpupower/utils/cpufreq-info.c
429
stats = stats->next;
tools/power/cpupower/utils/cpufreq-info.c
504
while (freqs->next) {
tools/power/cpupower/utils/cpufreq-info.c
507
freqs = freqs->next;
tools/power/cpupower/utils/cpufreq-set.c
308
while (cpus->next) {
tools/power/cpupower/utils/cpufreq-set.c
310
cpus = cpus->next;
tools/power/cpupower/utils/helpers/pci.c
47
for (device = (*pacc)->devices; device; device = device->next) {
tools/power/x86/intel-speed-select/isst-config.c
2792
char *next;
tools/power/x86/intel-speed-select/isst-config.c
2794
next = optarg;
tools/power/x86/intel-speed-select/isst-config.c
2797
while (next && *next) {
tools/power/x86/intel-speed-select/isst-config.c
2798
if (*next == '-') /* no negative cpu numbers */
tools/power/x86/intel-speed-select/isst-config.c
2801
start = strtoul(next, &next, 10);
tools/power/x86/intel-speed-select/isst-config.c
2808
if (*next == '\0')
tools/power/x86/intel-speed-select/isst-config.c
2811
if (*next == ',') {
tools/power/x86/intel-speed-select/isst-config.c
2812
next += 1;
tools/power/x86/intel-speed-select/isst-config.c
2816
if (*next == '-') {
tools/power/x86/intel-speed-select/isst-config.c
2817
next += 1; /* start range */
tools/power/x86/intel-speed-select/isst-config.c
2818
} else if (*next == '.') {
tools/power/x86/intel-speed-select/isst-config.c
2819
next += 1;
tools/power/x86/intel-speed-select/isst-config.c
2820
if (*next == '.')
tools/power/x86/intel-speed-select/isst-config.c
2821
next += 1; /* start range */
tools/power/x86/intel-speed-select/isst-config.c
2826
end = strtoul(next, &next, 10);
tools/power/x86/intel-speed-select/isst-config.c
2837
if (*next == ',')
tools/power/x86/intel-speed-select/isst-config.c
2838
next += 1;
tools/power/x86/intel-speed-select/isst-config.c
2839
else if (*next != '\0')
tools/power/x86/turbostat/turbostat.c
10025
pinfo = pinfo->next;
tools/power/x86/turbostat/turbostat.c
10155
new_pmt->next = pmt_mmios;
tools/power/x86/turbostat/turbostat.c
10158
last->next = new_pmt;
tools/power/x86/turbostat/turbostat.c
10190
pmmio = pmmio->next;
tools/power/x86/turbostat/turbostat.c
10225
ret = ret->next;
tools/power/x86/turbostat/turbostat.c
10243
pcounter = pcounter->next;
tools/power/x86/turbostat/turbostat.c
10343
pcounter->next = *pmt_root;
tools/power/x86/turbostat/turbostat.c
10565
for (mp = head; mp; mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
10643
msrp->next = sys.tp;
tools/power/x86/turbostat/turbostat.c
10647
msrp->next = sys.cp;
tools/power/x86/turbostat/turbostat.c
10651
msrp->next = sys.pp;
tools/power/x86/turbostat/turbostat.c
10667
sp->next = msrp->sp;
tools/power/x86/turbostat/turbostat.c
107
struct perf_counter_info *next;
tools/power/x86/turbostat/turbostat.c
10739
pinfo->next = sys.perf_tp;
tools/power/x86/turbostat/turbostat.c
10745
pinfo->next = sys.perf_cp;
tools/power/x86/turbostat/turbostat.c
10751
pinfo->next = sys.perf_pp;
tools/power/x86/turbostat/turbostat.c
10780
goto next;
tools/power/x86/turbostat/turbostat.c
10783
goto next;
tools/power/x86/turbostat/turbostat.c
10788
goto next;
tools/power/x86/turbostat/turbostat.c
10792
goto next;
tools/power/x86/turbostat/turbostat.c
10797
goto next;
tools/power/x86/turbostat/turbostat.c
10802
goto next;
tools/power/x86/turbostat/turbostat.c
10806
goto next;
tools/power/x86/turbostat/turbostat.c
10810
goto next;
tools/power/x86/turbostat/turbostat.c
10814
goto next;
tools/power/x86/turbostat/turbostat.c
10818
goto next;
tools/power/x86/turbostat/turbostat.c
10822
goto next;
tools/power/x86/turbostat/turbostat.c
10826
goto next;
tools/power/x86/turbostat/turbostat.c
10830
goto next;
tools/power/x86/turbostat/turbostat.c
10834
goto next;
tools/power/x86/turbostat/turbostat.c
10838
goto next;
tools/power/x86/turbostat/turbostat.c
10848
goto next;
tools/power/x86/turbostat/turbostat.c
10851
next:
tools/power/x86/turbostat/turbostat.c
11013
goto next;
tools/power/x86/turbostat/turbostat.c
11018
goto next;
tools/power/x86/turbostat/turbostat.c
11040
goto next;
tools/power/x86/turbostat/turbostat.c
11045
goto next;
tools/power/x86/turbostat/turbostat.c
11050
goto next;
tools/power/x86/turbostat/turbostat.c
11055
goto next;
tools/power/x86/turbostat/turbostat.c
11060
goto next;
tools/power/x86/turbostat/turbostat.c
11065
goto next;
tools/power/x86/turbostat/turbostat.c
11069
goto next;
tools/power/x86/turbostat/turbostat.c
11073
goto next;
tools/power/x86/turbostat/turbostat.c
11075
next:
tools/power/x86/turbostat/turbostat.c
129
struct sysfs_path *next;
tools/power/x86/turbostat/turbostat.c
139
struct msr_counter *next;
tools/power/x86/turbostat/turbostat.c
1932
struct pmt_mmio *next;
tools/power/x86/turbostat/turbostat.c
1964
struct pmt_counter *next;
tools/power/x86/turbostat/turbostat.c
2342
*pp = p->next;
tools/power/x86/turbostat/turbostat.c
2350
pp = &p->next;
tools/power/x86/turbostat/turbostat.c
2933
for (mp = sys.tp; mp; mp = mp->next)
tools/power/x86/turbostat/turbostat.c
2936
for (pp = sys.perf_tp; pp; pp = pp->next)
tools/power/x86/turbostat/turbostat.c
2952
ppmt = ppmt->next;
tools/power/x86/turbostat/turbostat.c
2981
for (mp = sys.cp; mp; mp = mp->next)
tools/power/x86/turbostat/turbostat.c
2984
for (pp = sys.perf_cp; pp; pp = pp->next)
tools/power/x86/turbostat/turbostat.c
3001
ppmt = ppmt->next;
tools/power/x86/turbostat/turbostat.c
3084
for (mp = sys.pp; mp; mp = mp->next)
tools/power/x86/turbostat/turbostat.c
3087
for (pp = sys.perf_pp; pp; pp = pp->next)
tools/power/x86/turbostat/turbostat.c
3103
ppmt = ppmt->next;
tools/power/x86/turbostat/turbostat.c
3174
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3193
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3226
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3455
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3469
for (i = 0, pp = sys.perf_tp; pp; ++i, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
3483
for (i = 0, ppmt = sys.pmt_tp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
3529
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3539
for (i = 0, pp = sys.perf_cp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
3549
for (i = 0, ppmt = sys.pmt_cp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
3681
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3693
for (i = 0, pp = sys.perf_pp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
3705
for (i = 0, ppmt = sys.pmt_pp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
3835
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3844
for (i = 0, pp = sys.perf_pp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
3853
for (i = 0, ppmt = sys.pmt_pp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
3879
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3886
for (i = 0, pp = sys.perf_cp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
3893
for (i = 0, ppmt = sys.pmt_cp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
4003
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4010
for (i = 0, pp = sys.perf_tp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
4017
for (i = 0, ppmt = sys.pmt_tp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
4128
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next)
tools/power/x86/turbostat/turbostat.c
4131
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next)
tools/power/x86/turbostat/turbostat.c
4134
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next)
tools/power/x86/turbostat/turbostat.c
4197
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4203
for (i = 0, pp = sys.perf_tp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
4209
for (i = 0, ppmt = sys.pmt_tp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
4227
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4233
for (i = 0, pp = sys.perf_cp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
4239
for (i = 0, ppmt = sys.pmt_cp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
4289
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4296
for (i = 0, pp = sys.perf_pp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
4303
for (i = 0, ppmt = sys.pmt_pp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
4366
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4376
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4385
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4395
for (i = 0, pp = sys.perf_tp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
4405
for (i = 0, pp = sys.perf_cp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
4414
for (i = 0, pp = sys.perf_pp; pp; i++, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
4424
for (i = 0, ppmt = sys.pmt_tp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
4427
for (i = 0, ppmt = sys.pmt_cp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
4430
for (i = 0, ppmt = sys.pmt_pp; ppmt; i++, ppmt = ppmt->next) {
tools/power/x86/turbostat/turbostat.c
4707
goto next;
tools/power/x86/turbostat/turbostat.c
4712
goto next;
tools/power/x86/turbostat/turbostat.c
4715
next:
tools/power/x86/turbostat/turbostat.c
4883
sp = sp->next;
tools/power/x86/turbostat/turbostat.c
5107
for (size_t i = 0; pp; ++i, pp = pp->next) {
tools/power/x86/turbostat/turbostat.c
5215
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
5223
for (i = 0, pp = sys.pmt_tp; pp; i++, pp = pp->next)
tools/power/x86/turbostat/turbostat.c
5259
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
5267
for (i = 0, pp = sys.pmt_cp; pp; i++, pp = pp->next)
tools/power/x86/turbostat/turbostat.c
5329
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
5346
for (i = 0, pp = sys.pmt_pp; pp; i++, pp = pp->next)
tools/power/x86/turbostat/turbostat.c
5911
pp = pp->next;
tools/power/x86/turbostat/turbostat.c
6111
char *next = cpu_str;
tools/power/x86/turbostat/turbostat.c
6113
while (next && *next) {
tools/power/x86/turbostat/turbostat.c
6115
if (*next == '-') /* no negative cpu numbers */
tools/power/x86/turbostat/turbostat.c
6118
if (*next == '\0' || *next == '\n')
tools/power/x86/turbostat/turbostat.c
6121
start = strtoul(next, &next, 10);
tools/power/x86/turbostat/turbostat.c
6127
if (*next == '\0' || *next == '\n')
tools/power/x86/turbostat/turbostat.c
6130
if (*next == ',') {
tools/power/x86/turbostat/turbostat.c
6131
next += 1;
tools/power/x86/turbostat/turbostat.c
6135
if (*next == '-') {
tools/power/x86/turbostat/turbostat.c
6136
next += 1; /* start range */
tools/power/x86/turbostat/turbostat.c
6137
} else if (*next == '.') {
tools/power/x86/turbostat/turbostat.c
6138
next += 1;
tools/power/x86/turbostat/turbostat.c
6139
if (*next == '.')
tools/power/x86/turbostat/turbostat.c
6140
next += 1; /* start range */
tools/power/x86/turbostat/turbostat.c
6145
end = strtoul(next, &next, 10);
tools/power/x86/turbostat/turbostat.c
6155
if (*next == ',')
tools/power/x86/turbostat/turbostat.c
6156
next += 1;
tools/power/x86/turbostat/turbostat.c
6157
else if (*next != '\0' && *next != '\n')
tools/power/x86/turbostat/turbostat.c
7293
goto next;
tools/power/x86/turbostat/turbostat.c
7297
goto next;
tools/power/x86/turbostat/turbostat.c
7306
goto next;
tools/power/x86/turbostat/turbostat.c
7323
next:
tools/sched_ext/scx_qmap.bpf.c
537
struct task_struct *next, unsigned long prev_state)
tools/sched_ext/scx_qmap.bpf.c
549
switch (next->policy) {
tools/testing/radix-tree/maple.c
33940
reader->next = (reader->mod + 4) % 10;
tools/testing/radix-tree/maple.c
33988
alt_start = test->index[index + reader->next];
tools/testing/radix-tree/maple.c
33993
i = reader->next;
tools/testing/radix-tree/maple.c
489
struct maple_enode *next;
tools/testing/radix-tree/maple.c
493
next = mas->node;
tools/testing/radix-tree/maple.c
495
node = mte_to_node(next);
tools/testing/radix-tree/maple.c
496
type = mte_node_type(next);
tools/testing/radix-tree/maple.c
498
next = mas_slot(mas, ma_slots(node, type), mas->offset);
tools/testing/radix-tree/maple.c
506
mas->node = next;
tools/testing/radix-tree/maple.c
56
int next;
tools/testing/selftests/alsa/alsa-local.h
32
struct card_cfg_data *next;
tools/testing/selftests/alsa/conf.c
195
snd_config_iterator_t i, next;
tools/testing/selftests/alsa/conf.c
201
snd_config_for_each(i, next, config) {
tools/testing/selftests/alsa/conf.c
230
for (data = conf_cards; data; data = data->next) {
tools/testing/selftests/alsa/conf.c
272
snd_config_iterator_t i, next;
tools/testing/selftests/alsa/conf.c
283
snd_config_for_each(i, next, card_config) {
tools/testing/selftests/alsa/conf.c
297
data->next = conf_cards;
tools/testing/selftests/alsa/conf.c
334
conf_cards = conf->next;
tools/testing/selftests/alsa/conf.c
383
snd_config_iterator_t i, next;
tools/testing/selftests/alsa/conf.c
396
snd_config_for_each(i, next, cfg)
tools/testing/selftests/alsa/conf.c
97
for (conf = conf_cards; conf; conf = conf->next) {
tools/testing/selftests/alsa/mixer-test.c
1128
for (ctl = ctl_list; ctl != NULL; ctl = ctl->next) {
tools/testing/selftests/alsa/mixer-test.c
127
card_data->next = card_list;
tools/testing/selftests/alsa/mixer-test.c
169
ctl_data->next = ctl_list;
tools/testing/selftests/alsa/mixer-test.c
41
struct card_data *next;
tools/testing/selftests/alsa/mixer-test.c
53
struct ctl_data *next;
tools/testing/selftests/alsa/pcm-test.c
100
for (pcm_data = pcm_list; pcm_data != NULL; pcm_data = pcm_data->next) {
tools/testing/selftests/alsa/pcm-test.c
118
pcm_data->next = pcm_missing;
tools/testing/selftests/alsa/pcm-test.c
216
card_data->next = card_list;
tools/testing/selftests/alsa/pcm-test.c
264
pcm_data->next = pcm_list;
tools/testing/selftests/alsa/pcm-test.c
30
struct card_data *next;
tools/testing/selftests/alsa/pcm-test.c
43
struct pcm_data *next;
tools/testing/selftests/alsa/pcm-test.c
559
snd_config_iterator_t i, next;
tools/testing/selftests/alsa/pcm-test.c
568
snd_config_for_each(i, next, cfg) {
tools/testing/selftests/alsa/pcm-test.c
585
for (pcm = pcm_list; pcm != NULL; pcm = pcm->next) {
tools/testing/selftests/alsa/pcm-test.c
617
for (conf = conf_cards; conf; conf = conf->next)
tools/testing/selftests/alsa/pcm-test.c
623
for (pcm = pcm_list; pcm != NULL; pcm = pcm->next) {
tools/testing/selftests/alsa/pcm-test.c
636
for (conf = conf_cards; conf; conf = conf->next)
tools/testing/selftests/alsa/pcm-test.c
641
for (pcm = pcm_missing; pcm != NULL; pcm = pcm->next) {
tools/testing/selftests/alsa/pcm-test.c
647
for (card = card_list; card != NULL; card = card->next) {
tools/testing/selftests/alsa/pcm-test.c
656
for (card = card_list; card != NULL; card = card->next) {
tools/testing/selftests/bpf/bpf_arena_list.h
11
arena_list_node_t *next;
tools/testing/selftests/bpf/bpf_arena_list.h
40
pos && ({ ___tmp = (void *)pos->member.next; 1; }) && can_loop; \
tools/testing/selftests/bpf/bpf_arena_list.h
49
WRITE_ONCE(n->next, first);
tools/testing/selftests/bpf/bpf_arena_list.h
52
tmp = &n->next;
tools/testing/selftests/bpf/bpf_arena_list.h
67
arena_list_node_t *next = n->next;
tools/testing/selftests/bpf/bpf_arena_list.h
70
cast_user(next);
tools/testing/selftests/bpf/bpf_arena_list.h
72
WRITE_ONCE(*pprev, next);
tools/testing/selftests/bpf/bpf_arena_list.h
73
if (next) {
tools/testing/selftests/bpf/bpf_arena_list.h
75
cast_kern(next);
tools/testing/selftests/bpf/bpf_arena_list.h
76
WRITE_ONCE(next->pprev, pprev);
tools/testing/selftests/bpf/bpf_arena_list.h
88
n->next = LIST_POISON1;
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
111
for ( ; list; list = list->next) {
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
138
node = best->next;
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
143
for (node = list; node; node = node->next) {
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
144
if (node->next == best) {
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
145
node->next = best->next;
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
204
for (t1 = l1; t1; t1 = t1->next)
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
302
for (i = 0, t = list; t; i++, t = t->next)
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
33
struct tlpm_node *next;
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
79
node->next = list;
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
93
list = list->next;
tools/testing/selftests/bpf/netlink_helpers.c
181
next:
tools/testing/selftests/bpf/netlink_helpers.c
232
goto next;
tools/testing/selftests/bpf/prog_tests/arena_spin_lock.c
11
unsigned long next;
tools/testing/selftests/bpf/prog_tests/btf_dump.c
699
{ .next = (struct list_head *)1 });
tools/testing/selftests/bpf/prog_tests/btf_dump.c
704
{ .next = (struct list_head *)1 });
tools/testing/selftests/bpf/prog_tests/btf_dump.c
708
{ .next = (struct list_head *)0 });
tools/testing/selftests/bpf/prog_tests/btf_dump.c
712
{ .next = (struct list_head *)0 });
tools/testing/selftests/bpf/prog_tests/core_read_macros.c
17
struct callback_head___shuffled *next;
tools/testing/selftests/bpf/prog_tests/core_read_macros.c
41
u_probe_in.next = &u_probe_in;
tools/testing/selftests/bpf/prog_tests/core_read_macros.c
45
u_core_in.next = &u_core_in;
tools/testing/selftests/bpf/prog_tests/core_read_macros.c
7
struct callback_head *next;
tools/testing/selftests/bpf/prog_tests/ctx_rewrite.c
243
char *next;
tools/testing/selftests/bpf/prog_tests/ctx_rewrite.c
244
int snum = strtol(str, &next, 10);
tools/testing/selftests/bpf/prog_tests/ctx_rewrite.c
246
if (next - str == 0 || num != snum)
tools/testing/selftests/bpf/prog_tests/ctx_rewrite.c
249
return next;
tools/testing/selftests/bpf/prog_tests/ctx_rewrite.c
366
char *next = str;
tools/testing/selftests/bpf/prog_tests/ctx_rewrite.c
368
while (next) {
tools/testing/selftests/bpf/prog_tests/ctx_rewrite.c
369
next = strchr(str, '\n');
tools/testing/selftests/bpf/prog_tests/ctx_rewrite.c
370
if (next) {
tools/testing/selftests/bpf/prog_tests/ctx_rewrite.c
371
max_line = max_t(u32, max_line, (next - str));
tools/testing/selftests/bpf/prog_tests/ctx_rewrite.c
372
str = next + 1;
tools/testing/selftests/bpf/prog_tests/unpriv_bpf_disabled.c
160
__u32 next;
tools/testing/selftests/bpf/prog_tests/unpriv_bpf_disabled.c
184
ASSERT_EQ(bpf_prog_get_next_id(prog_id, &next), -EPERM, "prog_get_next_id_fails");
tools/testing/selftests/bpf/prog_tests/unpriv_bpf_disabled.c
185
ASSERT_EQ(bpf_prog_get_next_id(0, &next), -EPERM, "prog_get_next_id_fails");
tools/testing/selftests/bpf/prog_tests/unpriv_bpf_disabled.c
190
ASSERT_EQ(bpf_map_get_next_id(map_info.id, &next), -EPERM,
tools/testing/selftests/bpf/prog_tests/unpriv_bpf_disabled.c
193
ASSERT_EQ(bpf_map_get_next_id(0, &next), -EPERM, "map_get_next_id_fails");
tools/testing/selftests/bpf/prog_tests/unpriv_bpf_disabled.c
199
ASSERT_EQ(bpf_link_get_next_id(link_info.id, &next), -EPERM,
tools/testing/selftests/bpf/prog_tests/unpriv_bpf_disabled.c
202
ASSERT_EQ(bpf_link_get_next_id(0, &next), -EPERM, "link_get_next_id_fails");
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
246
struct arena_mcs_spinlock __arena *prev, *next, *node0, *node;
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
346
node->next = NULL;
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
371
next = NULL;
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
381
WRITE_ONCE(prev->next, node);
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
390
next = READ_ONCE(node->next);
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
443
if (!next)
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
444
next = smp_cond_load_relaxed_label(&node->next, (VAL), release_node_err);
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
446
arch_mcs_spin_unlock_contended(&next->locked);
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
63
struct arena_mcs_spinlock __arena *next;
tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c
32
struct list_head *next;
tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c
37
struct hlist_node *next;
tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c
46
struct callback_head *next;
tools/testing/selftests/bpf/progs/kfree_skb.c
20
struct callback_head *next;
tools/testing/selftests/bpf/progs/kfree_skb.c
78
ptr = dev->ifalias->rcuhead.next;
tools/testing/selftests/bpf/progs/linked_list_fail.c
275
struct test2 *next;
tools/testing/selftests/bpf/progs/map_kptr.c
317
p_st = p->next;
tools/testing/selftests/bpf/progs/map_kptr_fail.c
199
unref_ptr = unref_ptr->next;
tools/testing/selftests/bpf/progs/mptcp_bpf.h
18
list_entry((ptr)->next, type, member)
tools/testing/selftests/bpf/progs/mptcp_bpf.h
21
list_entry((pos)->member.next, typeof(*(pos)), member)
tools/testing/selftests/bpf/progs/nested_trust_failure.c
39
bpf_sk_storage_get(&sk_storage_map, skb->next, 0, 0);
tools/testing/selftests/bpf/progs/netif_receive_skb.c
205
{ .next = (struct list_head *)1 });
tools/testing/selftests/bpf/progs/netif_receive_skb.c
209
{ .next = (struct list_head *)0 });
tools/testing/selftests/bpf/progs/test_cls_redirect.c
291
uint8_t next;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
294
.next = ipv6->nexthdr,
tools/testing/selftests/bpf/progs/test_cls_redirect.c
300
switch (exthdr.next) {
tools/testing/selftests/bpf/progs/test_cls_redirect.c
334
*upper_proto = exthdr.next;
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
197
uint8_t next;
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
200
.next = ipv6->nexthdr,
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
205
switch (exthdr.next) {
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
235
*upper_proto = exthdr.next;
tools/testing/selftests/bpf/progs/test_core_read_macros.c
13
struct callback_head___shuffled *next;
tools/testing/selftests/bpf/progs/test_core_read_macros.c
41
k_probe_in.next = &k_probe_in;
tools/testing/selftests/bpf/progs/test_core_read_macros.c
42
__builtin_preserve_access_index(({k_core_in.next = &k_core_in;}));
tools/testing/selftests/bpf/progs/test_core_read_macros.c
44
k_probe_out = (long)BPF_PROBE_READ(&k_probe_in, next, next, func);
tools/testing/selftests/bpf/progs/test_core_read_macros.c
45
k_core_out = (long)BPF_CORE_READ(&k_core_in, next, next, func);
tools/testing/selftests/bpf/progs/test_core_read_macros.c
46
u_probe_out = (long)BPF_PROBE_READ_USER(u_probe_in, next, next, func);
tools/testing/selftests/bpf/progs/test_core_read_macros.c
47
u_core_out = (long)BPF_CORE_READ_USER(u_core_in, next, next, func);
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
228
opcode = next(ctx, 1);
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
236
goto next;
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
238
opsize = next(ctx, 1);
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
247
mss = next(ctx, 2);
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
252
wscale = next(ctx, 1);
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
259
tsval = next(ctx, 4);
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
260
tsecr = next(ctx, 4);
tools/testing/selftests/bpf/progs/test_tcp_custom_syncookie.c
278
next:
tools/testing/selftests/bpf/progs/xdp_synproxy_kern.c
222
opcode = next(ctx, 1);
tools/testing/selftests/bpf/progs/xdp_synproxy_kern.c
231
opsize = next(ctx, 1);
tools/testing/selftests/bpf/progs/xdp_synproxy_kern.c
237
wscale = next(ctx, 1);
tools/testing/selftests/bpf/progs/xdp_synproxy_kern.c
244
tsecr = next(ctx, 4);
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
772
.next = &prog_test_struct,
tools/testing/selftests/bpf/test_kmods/bpf_testmod_kfunc.h
24
struct prog_test_ref_kfunc *next;
tools/testing/selftests/bpf/test_loader.c
858
const char *start = log, *end, *next, *match;
tools/testing/selftests/bpf/test_loader.c
882
next = start;
tools/testing/selftests/bpf/test_loader.c
883
match = match_msg(msg, &next);
tools/testing/selftests/bpf/test_loader.c
884
if (match && next <= end) {
tools/testing/selftests/bpf/test_loader.c
886
matches[j].end = next;
tools/testing/selftests/bpf/test_lru_map.c
124
int next = *next_to_try;
tools/testing/selftests/bpf/test_lru_map.c
127
while (next < nr_cpus) {
tools/testing/selftests/bpf/test_lru_map.c
129
CPU_SET(next, &cpuset);
tools/testing/selftests/bpf/test_lru_map.c
130
next++;
tools/testing/selftests/bpf/test_lru_map.c
137
*next_to_try = next;
tools/testing/selftests/bpf/testing_helpers.c
19
char *next;
tools/testing/selftests/bpf/testing_helpers.c
23
num = strtol(s, &next, 10);
tools/testing/selftests/bpf/testing_helpers.c
32
if (!parsing_end && *next == '-') {
tools/testing/selftests/bpf/testing_helpers.c
33
s = next + 1;
tools/testing/selftests/bpf/testing_helpers.c
36
} else if (*next == ',') {
tools/testing/selftests/bpf/testing_helpers.c
38
s = next + 1;
tools/testing/selftests/bpf/testing_helpers.c
40
} else if (*next == '\0') {
tools/testing/selftests/bpf/testing_helpers.c
42
s = next;
tools/testing/selftests/bpf/veristat.c
2693
char *input = line, *state = NULL, *next;
tools/testing/selftests/bpf/veristat.c
2713
while ((next = strtok_r(cnt++ ? NULL : input, ",\n", &state))) {
tools/testing/selftests/bpf/veristat.c
2716
err = parse_stat(next, specs);
tools/testing/selftests/bpf/veristat.c
2729
err = parse_stat_value(next, specs->ids[col], st);
tools/testing/selftests/bpf/veristat.c
959
char *input, *state = NULL, *next;
tools/testing/selftests/bpf/veristat.c
966
while ((next = strtok_r(cnt++ ? NULL : input, ",", &state))) {
tools/testing/selftests/bpf/veristat.c
967
err = parse_stat(next, specs);
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
142
self->next = (void *) self->buf;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
145
meta = self->next;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
150
self->next += thislen;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
40
void *next;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
169
self->next = (void *) self->buf;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
172
meta = self->next;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
177
self->next += thislen;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
46
void *next;
tools/testing/selftests/filesystems/utils.c
108
struct list *next, *prev;
tools/testing/selftests/filesystems/utils.c
110
next = list->next;
tools/testing/selftests/filesystems/utils.c
112
next->prev = prev;
tools/testing/selftests/filesystems/utils.c
113
prev->next = next;
tools/testing/selftests/filesystems/utils.c
55
for (__iterator = (__list)->next; __iterator != __list; __iterator = __iterator->next)
tools/testing/selftests/filesystems/utils.c
71
struct list *next;
tools/testing/selftests/filesystems/utils.c
85
list->next = list->prev = list;
tools/testing/selftests/filesystems/utils.c
90
return list == list->next;
tools/testing/selftests/filesystems/utils.c
93
static inline void __list_add(struct list *new, struct list *prev, struct list *next)
tools/testing/selftests/filesystems/utils.c
95
next->prev = new;
tools/testing/selftests/filesystems/utils.c
96
new->next = next;
tools/testing/selftests/filesystems/utils.c
98
prev->next = new;
tools/testing/selftests/futex/functional/robust_list.c
102
head->list.next = &head->list;
tools/testing/selftests/futex/functional/robust_list.c
134
while (list->next != &head->list)
tools/testing/selftests/futex/functional/robust_list.c
135
list = list->next;
tools/testing/selftests/futex/functional/robust_list.c
137
list->next = &lock->list;
tools/testing/selftests/futex/functional/robust_list.c
138
lock->list.next = &head->list;
tools/testing/selftests/futex/functional/robust_list.c
522
head.list.next = &a.list;
tools/testing/selftests/futex/functional/robust_list.c
527
a.list.next = &b.list;
tools/testing/selftests/futex/functional/robust_list.c
528
b.list.next = &c.list;
tools/testing/selftests/futex/functional/robust_list.c
529
c.list.next = &a.list;
tools/testing/selftests/iommu/iommufd.c
2584
if (cap->next)
tools/testing/selftests/iommu/iommufd.c
2585
cap_size = (buffer + cap->next) - (void *)cap;
tools/testing/selftests/iommu/iommufd.c
2613
if (!cap->next)
tools/testing/selftests/iommu/iommufd.c
2616
ASSERT_GE(info_cmd->argsz, cap->next + sizeof(*cap));
tools/testing/selftests/iommu/iommufd.c
2617
ASSERT_GE(buffer + cap->next, (void *)cap);
tools/testing/selftests/iommu/iommufd.c
2618
cap = buffer + cap->next;
tools/testing/selftests/kselftest_harness.h
1077
for (f = __fixture_list; f; f = f->next) {
tools/testing/selftests/kselftest_harness.h
1093
v = v ? v->next : NULL;
tools/testing/selftests/kselftest_harness.h
1094
t = t ? t->next : NULL;
tools/testing/selftests/kselftest_harness.h
1238
for (xfail = variant->xfails; xfail; xfail = xfail->next)
tools/testing/selftests/kselftest_harness.h
1271
for (f = __fixture_list; f; f = f->next) {
tools/testing/selftests/kselftest_harness.h
1272
for (v = f->variant ?: &no_variant; v; v = v->next) {
tools/testing/selftests/kselftest_harness.h
1275
for (t = f->tests; t; t = t->next)
tools/testing/selftests/kselftest_harness.h
1291
for (f = __fixture_list; f; f = f->next) {
tools/testing/selftests/kselftest_harness.h
1292
for (v = f->variant ?: &no_variant; v; v = v->next) {
tools/testing/selftests/kselftest_harness.h
1293
for (t = f->tests; t; t = t->next) {
tools/testing/selftests/kselftest_harness.h
827
item->next = NULL; \
tools/testing/selftests/kselftest_harness.h
832
item->next = NULL; \
tools/testing/selftests/kselftest_harness.h
834
item->prev->next = item; \
tools/testing/selftests/kselftest_harness.h
837
item->next = head; \
tools/testing/selftests/kselftest_harness.h
838
item->next->prev = item; \
tools/testing/selftests/kselftest_harness.h
856
struct __fixture_metadata *prev, *next;
tools/testing/selftests/kselftest_harness.h
866
struct __test_xfail *prev, *next;
tools/testing/selftests/kselftest_harness.h
906
struct __fixture_variant_metadata *prev, *next;
tools/testing/selftests/kselftest_harness.h
934
struct __test_metadata *prev, *next;
tools/testing/selftests/kvm/arm64/arch_timer_edge_cases.c
853
uint32_t next = cur;
tools/testing/selftests/kvm/arm64/arch_timer_edge_cases.c
859
next = (next + 1) % CPU_SETSIZE;
tools/testing/selftests/kvm/arm64/arch_timer_edge_cases.c
860
} while (!CPU_ISSET(next, &cpuset));
tools/testing/selftests/kvm/arm64/arch_timer_edge_cases.c
862
return next;
tools/testing/selftests/kvm/lib/arm64/gic_v3_its.c
181
u64 next;
tools/testing/selftests/kvm/lib/arm64/gic_v3_its.c
190
next = (cwriter + sizeof(*cmd)) % cmdq_size;
tools/testing/selftests/kvm/lib/arm64/gic_v3_its.c
191
its_write_u64(GITS_CWRITER, next);
tools/testing/selftests/kvm/lib/arm64/gic_v3_its.c
198
for (i = 0; its_read_u64(GITS_CREADR) != next; i++) {
tools/testing/selftests/kvm/lib/sparsebit.c
1373
struct node *nodep, *next;
tools/testing/selftests/kvm/lib/sparsebit.c
1420
for (next = node_next(s, nodep);
tools/testing/selftests/kvm/lib/sparsebit.c
1421
next && (next->idx < middle_end);
tools/testing/selftests/kvm/lib/sparsebit.c
1422
next = node_next(s, nodep)) {
tools/testing/selftests/kvm/lib/sparsebit.c
1423
assert(next->idx + MASK_BITS + next->num_after - 1 <= middle_end);
tools/testing/selftests/kvm/lib/sparsebit.c
1424
node_rm(s, next);
tools/testing/selftests/kvm/lib/sparsebit.c
1425
next = NULL;
tools/testing/selftests/kvm/lib/sparsebit.c
1455
struct node *nodep, *next;
tools/testing/selftests/kvm/lib/sparsebit.c
1483
for (next = node_next(s, nodep);
tools/testing/selftests/kvm/lib/sparsebit.c
1484
next && (next->idx < middle_end);
tools/testing/selftests/kvm/lib/sparsebit.c
1485
next = node_next(s, nodep)) {
tools/testing/selftests/kvm/lib/sparsebit.c
1486
assert(next->idx + MASK_BITS + next->num_after - 1 <= middle_end);
tools/testing/selftests/kvm/lib/sparsebit.c
1487
node_rm(s, next);
tools/testing/selftests/kvm/lib/sparsebit.c
1488
next = NULL;
tools/testing/selftests/kvm/lib/sparsebit.c
1916
sparsebit_idx_t next;
tools/testing/selftests/kvm/lib/sparsebit.c
1980
next = sparsebit_next_set(s, first);
tools/testing/selftests/kvm/lib/sparsebit.c
1981
assert(next == 0 || next > first);
tools/testing/selftests/kvm/lib/sparsebit.c
1982
assert(next == 0 || get_value(next));
tools/testing/selftests/kvm/lib/sparsebit.c
1985
next = sparsebit_next_clear(s, first);
tools/testing/selftests/kvm/lib/sparsebit.c
1986
assert(next == 0 || next > first);
tools/testing/selftests/kvm/lib/sparsebit.c
1987
assert(next == 0 || !get_value(next));
tools/testing/selftests/kvm/lib/sparsebit.c
1990
next = sparsebit_next_clear(s, first);
tools/testing/selftests/kvm/lib/sparsebit.c
1992
assert(next == 0 || next > last);
tools/testing/selftests/kvm/lib/sparsebit.c
1994
next = sparsebit_next_set(s, first - 1);
tools/testing/selftests/kvm/lib/sparsebit.c
1996
next = sparsebit_first_set(s);
tools/testing/selftests/kvm/lib/sparsebit.c
1999
assert(next == first);
tools/testing/selftests/kvm/lib/sparsebit.c
2001
assert(sparsebit_is_clear(s, first) || next <= last);
tools/testing/selftests/kvm/lib/sparsebit.c
2005
next = sparsebit_next_set(s, first);
tools/testing/selftests/kvm/lib/sparsebit.c
2007
assert(next == 0 || next > last);
tools/testing/selftests/kvm/lib/sparsebit.c
2009
next = sparsebit_next_clear(s, first - 1);
tools/testing/selftests/kvm/lib/sparsebit.c
2011
next = sparsebit_first_clear(s);
tools/testing/selftests/kvm/lib/sparsebit.c
2014
assert(next == first);
tools/testing/selftests/kvm/lib/sparsebit.c
2016
assert(sparsebit_is_set(s, first) || next <= last);
tools/testing/selftests/kvm/lib/sparsebit.c
605
struct node *prev, *next, *tmp;
tools/testing/selftests/kvm/lib/sparsebit.c
745
next = node_next(s, nodep);
tools/testing/selftests/kvm/lib/sparsebit.c
746
if (next) {
tools/testing/selftests/kvm/lib/sparsebit.c
748
if (next->mask == 0 && next->num_after == 0) {
tools/testing/selftests/kvm/lib/sparsebit.c
749
node_rm(s, next);
tools/testing/selftests/kvm/lib/sparsebit.c
758
if (next->idx == nodep->idx + MASK_BITS + nodep->num_after &&
tools/testing/selftests/kvm/lib/sparsebit.c
759
next->mask == ~(mask_t) 0) {
tools/testing/selftests/kvm/lib/sparsebit.c
761
next->mask = 0;
tools/testing/selftests/kvm/lib/sparsebit.c
762
nodep->num_after += next->num_after;
tools/testing/selftests/kvm/lib/sparsebit.c
763
next->num_after = 0;
tools/testing/selftests/kvm/lib/sparsebit.c
765
node_rm(s, next);
tools/testing/selftests/kvm/lib/sparsebit.c
766
next = NULL;
tools/testing/selftests/mm/compaction_test.c
252
entry->next = list;
tools/testing/selftests/mm/compaction_test.c
26
struct map_list *next;
tools/testing/selftests/mm/compaction_test.c
264
for (entry = list; entry != NULL; entry = entry->next) {
tools/testing/selftests/mm/compaction_test.c
266
if (!entry->next)
tools/testing/selftests/mm/compaction_test.c
268
entry = entry->next;
tools/testing/selftests/mm/mlock2.h
37
goto next;
tools/testing/selftests/mm/mlock2.h
42
next:
tools/testing/selftests/net/mptcp/mptcp_connect.c
1267
char *next = strchr(type, ',');
tools/testing/selftests/net/mptcp/mptcp_connect.c
1272
if (next) {
tools/testing/selftests/net/mptcp/mptcp_connect.c
1273
parse_cmsg_types(next + 1);
tools/testing/selftests/net/mptcp/mptcp_connect.c
1274
len = next - type;
tools/testing/selftests/net/mptcp/mptcp_connect.c
1295
char *next = strchr(name, ',');
tools/testing/selftests/net/mptcp/mptcp_connect.c
1298
if (next) {
tools/testing/selftests/net/mptcp/mptcp_connect.c
1299
parse_setsock_options(next + 1);
tools/testing/selftests/net/mptcp/mptcp_connect.c
1300
len = next - name;
tools/testing/selftests/net/mptcp/mptcp_connect.c
565
int next = read(fd, msg_buf, sizeof(msg_buf));
tools/testing/selftests/net/mptcp/mptcp_connect.c
568
ret, (unsigned int)len, last_hint, tcp_inq.last, next);
tools/testing/selftests/net/mptcp/pm_nl_ctl.c
1242
goto next;
tools/testing/selftests/net/mptcp/pm_nl_ctl.c
1248
next:
tools/testing/selftests/net/psock_tpacket.c
419
void *next = get_next_frame(ring, frame_num);
tools/testing/selftests/net/psock_tpacket.c
421
while (__tx_kernel_ready(next, ring->version) &&
tools/testing/selftests/net/psock_tpacket.c
423
ppd.raw = next;
tools/testing/selftests/net/psock_tpacket.c
446
struct tpacket3_hdr *tx = next;
tools/testing/selftests/net/psock_tpacket.c
463
__tx_user_ready(next, ring->version);
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
114
opts_list = opts_list->next;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
123
opts_list = opts_list->next;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
34
struct test_ftracer *next;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
402
tracer->next = ftracers;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
441
ftracers = tracer->next;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
445
while (f->next != tracer) {
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
446
if (!f->next)
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
448
f = f->next;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
450
f->next = tracer->next;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
452
tracer->next = NULL;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
467
struct test_ftracer *n = f->next;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
469
f->next = NULL;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
65
struct opts_list_t *next;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
98
tmp->next = opts_list;
tools/testing/selftests/net/tcp_ao/lib/proc.c
16
struct netstat *next;
tools/testing/selftests/net/tcp_ao/lib/proc.c
198
ns = ns->next;
tools/testing/selftests/net/tcp_ao/lib/proc.c
225
nsb = nsb->next;
tools/testing/selftests/net/tcp_ao/lib/proc.c
249
nsb = nsb->next;
tools/testing/selftests/net/tcp_ao/lib/proc.c
250
nsa = nsa->next;
tools/testing/selftests/net/tcp_ao/lib/proc.c
267
ns = ns->next;
tools/testing/selftests/net/tcp_ao/lib/proc.c
29
ns = ns->next;
tools/testing/selftests/net/tcp_ao/lib/proc.c
50
ret->next = ns;
tools/testing/selftests/net/tcp_ao/lib/setup.c
77
struct dlist_t *next;
tools/testing/selftests/net/tcp_ao/lib/setup.c
89
p->next = destructors_list;
tools/testing/selftests/net/tcp_ao/lib/setup.c
98
struct dlist_t *p = destructors_list->next;
tools/testing/selftests/net/timestamping.c
335
struct timeval next;
tools/testing/selftests/net/timestamping.c
497
gettimeofday(&next, 0);
tools/testing/selftests/net/timestamping.c
498
next.tv_sec = (next.tv_sec + 1) / 5 * 5;
tools/testing/selftests/net/timestamping.c
499
next.tv_usec = 0;
tools/testing/selftests/net/timestamping.c
508
delta_us = (long)(next.tv_sec - now.tv_sec) * 1000000 +
tools/testing/selftests/net/timestamping.c
509
(long)(next.tv_usec - now.tv_usec);
tools/testing/selftests/net/timestamping.c
545
next.tv_sec += 5;
tools/testing/selftests/openat2/openat2_test.c
270
goto next;
tools/testing/selftests/openat2/openat2_test.c
317
next:
tools/testing/selftests/powerpc/dscr/dscr_explicit_test.c
31
sem_t *next = (sem_t *)args + 1;
tools/testing/selftests/powerpc/dscr/dscr_explicit_test.c
46
FAIL_IF_EXIT(sem_post(next));
tools/testing/selftests/powerpc/dscr/dscr_explicit_test.c
57
sem_t *next = &semaphores[0];
tools/testing/selftests/powerpc/dscr/dscr_explicit_test.c
66
FAIL_IF(sem_init(next, 0, 1)); /* other thread starts first */
tools/testing/selftests/powerpc/dscr/dscr_explicit_test.c
79
FAIL_IF(sem_post(next));
tools/testing/selftests/powerpc/dscr/dscr_explicit_test.c
84
FAIL_IF(sem_destroy(next));
tools/testing/selftests/powerpc/syscalls/rtas_filter.c
51
struct region *next;
tools/testing/selftests/proc/fd-001-lookup.c
122
next:
tools/testing/selftests/proc/fd-001-lookup.c
137
goto next;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
185
node->next = (struct percpu_list_node *)expect;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
213
offset = offsetof(struct percpu_list_node, next);
tools/testing/selftests/rseq/basic_percpu_ops_test.c
240
list->c[cpu].head = node->next;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
297
node->next = list.c[i].head;
tools/testing/selftests/rseq/basic_percpu_ops_test.c
71
struct percpu_list_node *next;
tools/testing/selftests/rseq/param_test.c
1301
node->next = NULL;
tools/testing/selftests/rseq/param_test.c
394
struct percpu_list_node *next;
tools/testing/selftests/rseq/param_test.c
646
node->next = (struct percpu_list_node *)expect;
tools/testing/selftests/rseq/param_test.c
677
offset = offsetof(struct percpu_list_node, next);
tools/testing/selftests/rseq/param_test.c
706
list->c[cpu].head = node->next;
tools/testing/selftests/rseq/param_test.c
762
node->next = list.c[i].head;
tools/testing/selftests/timers/nanosleep.c
195
goto next;
tools/testing/selftests/timers/nanosleep.c
209
next:
tools/testing/selftests/vfio/lib/drivers/ioat/hw.h
124
uint64_t next;
tools/testing/selftests/vfio/lib/drivers/ioat/hw.h
135
uint64_t next;
tools/testing/selftests/vfio/lib/drivers/ioat/hw.h
178
uint64_t next;
tools/testing/selftests/vfio/lib/drivers/ioat/hw.h
192
uint64_t next;
tools/testing/selftests/vfio/lib/drivers/ioat/hw.h
222
uint64_t next;
tools/testing/selftests/vfio/lib/drivers/ioat/hw.h
91
uint64_t next;
tools/testing/selftests/vfio/lib/drivers/ioat/ioat.c
194
.next = desc_iova,
tools/testing/selftests/vfio/lib/iommu.c
222
struct dma_region *curr, *next;
tools/testing/selftests/vfio/lib/iommu.c
234
list_for_each_entry_safe(curr, next, &iommu->dma_regions, link)
tools/testing/selftests/vfio/lib/iommu.c
252
*cap_offset = hdr->next;
tools/testing/selftests/x86/test_shadow_stack.c
543
struct node *next;
tools/testing/selftests/x86/test_shadow_stack.c
581
cur->next = head;
tools/testing/selftests/x86/test_shadow_stack.c
587
head = cur->next;
tools/testing/selftests/x86/test_shadow_stack.c
630
cur->next = head;
tools/testing/selftests/x86/test_shadow_stack.c
642
head = cur->next;
tools/testing/selftests/x86/xstate.c
148
pthread_mutex_unlock(&finfo->next->mutex);
tools/testing/selftests/x86/xstate.c
173
finfo[i].next = &finfo[next_nr];
tools/testing/selftests/x86/xstate.c
54
struct futex_info *next;
tools/testing/vma/include/stubs.h
424
struct vm_area_struct *next)
tools/testing/vma/tests/merge.c
11
vmg->next = vma_next(vmg->vmi);
tools/testing/vma/tests/merge.c
1184
vmg.next = vma_next;
tools/testing/vma/tests/merge.c
1367
vmg.next = vma;
tools/testing/vma/tests/merge.c
42
vmg->next = NULL;
tools/testing/vsock/control.c
61
goto next;
tools/testing/vsock/control.c
72
goto next;
tools/testing/vsock/control.c
74
goto next;
tools/testing/vsock/control.c
90
next:
tools/testing/vsock/vsock_diag_test.c
324
struct vsock_stat *next;
tools/testing/vsock/vsock_diag_test.c
326
list_for_each_entry_safe(st, next, sockets, list)
tools/tracing/latency/latency-collector.c
1247
static char *find_next_optstr(const char *allopt, const char **next)
tools/tracing/latency/latency-collector.c
1271
*next = begin + s;
tools/tracing/latency/latency-collector.c
1280
const char *next = allopt;
tools/tracing/latency/latency-collector.c
1286
str = find_next_optstr(next, &next);
tools/tracing/rtla/src/common.c
99
tevent->next = common->events;
tools/tracing/rtla/src/trace.c
259
tevent = tevent->next;
tools/tracing/rtla/src/trace.c
453
tevent = tevent->next;
tools/tracing/rtla/src/trace.c
555
tevent = tevent->next;
tools/tracing/rtla/src/trace.h
6
struct trace_events *next;
tools/usb/testusb.c
122
struct testdev *next;
tools/usb/testusb.c
274
entry->next = testdevs;
tools/usb/testusb.c
491
if (testdevs && !testdevs->next && !device)
tools/usb/testusb.c
493
for (entry = testdevs; entry; entry = entry->next) {
tools/usb/testusb.c
525
for (entry = testdevs; entry; entry = entry->next) {
tools/usb/usbip/libsrc/list.h
112
for (pos = (head)->next; pos != (head); pos = pos->next)
tools/usb/usbip/libsrc/list.h
121
for (pos = (head)->next, n = pos->next; pos != (head); \
tools/usb/usbip/libsrc/list.h
122
pos = n, n = pos->next)
tools/usb/usbip/libsrc/list.h
20
struct list_head *next, *prev;
tools/usb/usbip/libsrc/list.h
30
list->next = list;
tools/usb/usbip/libsrc/list.h
42
struct list_head *next)
tools/usb/usbip/libsrc/list.h
44
next->prev = new;
tools/usb/usbip/libsrc/list.h
45
new->next = next;
tools/usb/usbip/libsrc/list.h
47
prev->next = new;
tools/usb/usbip/libsrc/list.h
60
__list_add(new, head, head->next);
tools/usb/usbip/libsrc/list.h
70
static inline void __list_del(struct list_head * prev, struct list_head * next)
tools/usb/usbip/libsrc/list.h
72
next->prev = prev;
tools/usb/usbip/libsrc/list.h
73
prev->next = next;
tools/usb/usbip/libsrc/list.h
82
__list_del(entry->prev, entry->next);
tools/usb/usbip/libsrc/list.h
93
__list_del(entry->prev, entry->next);
tools/usb/usbip/libsrc/list.h
94
entry->next = LIST_POISON1;
tools/usb/usbip/libsrc/names.c
110
for (; c; c = c->next)
tools/usb/usbip/libsrc/names.c
121
for (; s; s = s->next)
tools/usb/usbip/libsrc/names.c
134
for (; p; p = p->next)
tools/usb/usbip/libsrc/names.c
143
struct pool *next;
tools/usb/usbip/libsrc/names.c
163
p->next = pool_head;
tools/usb/usbip/libsrc/names.c
183
pool = pool->next;
tools/usb/usbip/libsrc/names.c
194
for (; v; v = v->next)
tools/usb/usbip/libsrc/names.c
202
v->next = vendors[h];
tools/usb/usbip/libsrc/names.c
214
for (; p; p = p->next)
tools/usb/usbip/libsrc/names.c
223
p->next = products[h];
tools/usb/usbip/libsrc/names.c
234
for (; c; c = c->next)
tools/usb/usbip/libsrc/names.c
242
c->next = classes[h];
tools/usb/usbip/libsrc/names.c
253
for (; s; s = s->next)
tools/usb/usbip/libsrc/names.c
26
struct vendor *next;
tools/usb/usbip/libsrc/names.c
262
s->next = subclasses[h];
tools/usb/usbip/libsrc/names.c
275
for (; p; p = p->next)
tools/usb/usbip/libsrc/names.c
286
p->next = protocols[h];
tools/usb/usbip/libsrc/names.c
32
struct product *next;
tools/usb/usbip/libsrc/names.c
38
struct class *next;
tools/usb/usbip/libsrc/names.c
44
struct subclass *next;
tools/usb/usbip/libsrc/names.c
50
struct protocol *next;
tools/usb/usbip/libsrc/names.c
56
struct genericstrtable *next;
tools/usb/usbip/libsrc/names.c
88
for (; v; v = v->next)
tools/usb/usbip/libsrc/names.c
99
for (; p; p = p->next)
tools/verification/rv/src/in_kernel.c
341
char *curr, *next;
tools/verification/rv/src/in_kernel.c
353
next = strstr(curr, "\n");
tools/verification/rv/src/in_kernel.c
354
*next = '\0';
tools/verification/rv/src/in_kernel.c
368
curr = ++next;
tools/verification/rvgen/rvgen/templates/ltl2k/trace.h
10
TP_ARGS(task, states, atoms, next));
tools/verification/rvgen/rvgen/templates/ltl2k/trace.h
9
TP_PROTO(struct task_struct *task, char *states, char *atoms, char *next),
tools/virtio/ringtest/ring.c
20
unsigned short next,
tools/virtio/ringtest/ring.c
23
return (unsigned short)(next - event - 1) < (unsigned short)(next - prev);
tools/virtio/ringtest/virtio_ring_0_9.c
127
guest.free_head = desc[head].next;
tools/virtio/ringtest/virtio_ring_0_9.c
189
ring.desc[index].next = guest.free_head;
tools/virtio/ringtest/virtio_ring_0_9.c
86
ring.desc[i].next = i + 1;
tools/virtio/vringh_test.c
705
d[0].next = 1;
tools/virtio/vringh_test.c
719
d[3].next = 1;
tools/virtio/vringh_test.c
723
d[4].next = 2;
virt/kvm/kvm_main.c
6396
struct task_struct *next)