Symbol: READ_ONCE
arch/alpha/include/asm/atomic.h
29
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/alpha/include/asm/atomic.h
30
#define arch_atomic64_read(v) READ_ONCE((v)->counter)
arch/alpha/include/asm/pgtable.h
278
pte_t pte = READ_ONCE(*ptep);
arch/alpha/include/asm/tlbflush.h
62
if (READ_ONCE(*mmc))
arch/alpha/mm/tlbflush.c
105
if (READ_ONCE(mm->context[cpu]))
arch/arc/include/asm/atomic.h
17
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/arm/include/asm/atomic.h
25
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/arm/include/asm/pgtable.h
153
#define pgdp_get(pgpd) READ_ONCE(*pgdp)
arch/arm/include/asm/spinlock.h
121
return !arch_spin_value_unlocked(READ_ONCE(*lock));
arch/arm/include/asm/spinlock.h
126
struct __raw_tickets tickets = READ_ONCE(lock->tickets);
arch/arm/include/asm/spinlock.h
75
lockval.tickets.owner = READ_ONCE(lock->tickets.owner);
arch/arm/mm/kasan_init.c
56
if (!pte_none(READ_ONCE(*ptep)))
arch/arm/mm/kasan_init.c
68
} else if (pte_none(READ_ONCE(*ptep))) {
arch/arm64/include/asm/barrier.h
201
VAL = READ_ONCE(*__PTR); \
arch/arm64/include/asm/efi.h
49
on_task_stack(current, READ_ONCE(efi_rt_stack_top[-1]), 1))
arch/arm64/include/asm/kvm_host.h
1234
#define vcpu_has_run_once(vcpu) (!!READ_ONCE((vcpu)->pid))
arch/arm64/include/asm/kvm_host.h
958
READ_ONCE(v->arch.flagset) & (m); \
arch/arm64/include/asm/percpu.h
58
return READ_ONCE(*(u##sz *)ptr); \
arch/arm64/include/asm/pgtable.h
1008
return p4d_page_paddr(READ_ONCE(*p4dp)) + pud_index(addr) * sizeof(pud_t);
arch/arm64/include/asm/pgtable.h
1022
return pud_offset_lockless(p4dp, READ_ONCE(*p4dp), addr);
arch/arm64/include/asm/pgtable.h
1131
return pgd_page_paddr(READ_ONCE(*pgdp)) + p4d_index(addr) * sizeof(p4d_t);
arch/arm64/include/asm/pgtable.h
1145
return p4d_offset_lockless(pgdp, READ_ONCE(*pgdp), addr);
arch/arm64/include/asm/pgtable.h
1336
VM_WARN_ON(pmd_table(READ_ONCE(*pmdp)) && !system_supports_haft());
arch/arm64/include/asm/pgtable.h
387
return READ_ONCE(*ptep);
arch/arm64/include/asm/pgtable.h
855
#define pte_offset_phys(dir,addr) (pmd_page_paddr(READ_ONCE(*(dir))) + pte_index(addr) * sizeof(pte_t))
arch/arm64/include/asm/pgtable.h
915
#define pmd_offset_phys(dir, addr) (pud_page_paddr(READ_ONCE(*(dir))) + pmd_index(addr) * sizeof(pmd_t))
arch/arm64/include/asm/preempt.h
12
return READ_ONCE(current_thread_info()->preempt.count);
arch/arm64/include/asm/preempt.h
46
u32 pc = READ_ONCE(current_thread_info()->preempt.count);
arch/arm64/include/asm/preempt.h
53
u32 pc = READ_ONCE(current_thread_info()->preempt.count);
arch/arm64/include/asm/preempt.h
61
u64 pc = READ_ONCE(ti->preempt_count);
arch/arm64/include/asm/preempt.h
73
return !pc || !READ_ONCE(ti->preempt_count);
arch/arm64/include/asm/preempt.h
78
u64 pc = READ_ONCE(current_thread_info()->preempt_count);
arch/arm64/include/asm/rqspinlock.h
39
VAL = READ_ONCE(*__PTR); \
arch/arm64/include/asm/stacktrace/common.h
165
state->fp = READ_ONCE(record->fp);
arch/arm64/include/asm/stacktrace/common.h
166
state->pc = READ_ONCE(record->lr);
arch/arm64/include/asm/uaccess.h
84
ttbr0 = READ_ONCE(current_thread_info()->ttbr0);
arch/arm64/kernel/armv8_deprecated.c
464
bool enable = READ_ONCE(insn->current_mode) == INSN_HW;
arch/arm64/kernel/armv8_deprecated.c
590
if (READ_ONCE(ie->current_mode) == INSN_UNDEF)
arch/arm64/kernel/fpsimd.c
163
return READ_ONCE(vl_config[type].__default_vl);
arch/arm64/kernel/paravirt.c
64
ret = le64_to_cpu(READ_ONCE(kaddr->stolen_time));
arch/arm64/kernel/probes/simulate-insn.c
214
set_x_reg(regs, xn, READ_ONCE(*(u64 *)load_addr));
arch/arm64/kernel/probes/simulate-insn.c
216
set_w_reg(regs, xn, READ_ONCE(*(u32 *)load_addr));
arch/arm64/kernel/probes/simulate-insn.c
229
set_x_reg(regs, xn, READ_ONCE(*(s32 *)load_addr));
arch/arm64/kernel/proton-pack.c
45
state = READ_ONCE(*oldp);
arch/arm64/kernel/smp.c
142
status = READ_ONCE(secondary_data.status);
arch/arm64/kernel/smp.c
144
status = READ_ONCE(__early_cpu_boot_status);
arch/arm64/kernel/stacktrace.c
202
switch (READ_ONCE(meta->type)) {
arch/arm64/kernel/stacktrace.c
232
new_fp = READ_ONCE(record->fp);
arch/arm64/kernel/stacktrace.c
233
new_pc = READ_ONCE(record->lr);
arch/arm64/kvm/arm.c
748
return READ_ONCE(vcpu->arch.mp_state.mp_state) == KVM_MP_STATE_STOPPED;
arch/arm64/kvm/arm.c
760
return READ_ONCE(vcpu->arch.mp_state.mp_state) == KVM_MP_STATE_SUSPENDED;
arch/arm64/kvm/arm.c
766
*mp_state = READ_ONCE(vcpu->arch.mp_state);
arch/arm64/kvm/hyp/include/nvhe/spinlock.h
103
hyp_spinlock_t lockval = READ_ONCE(*lock);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
131
hyp_vcpu->vcpu.arch.hcr_el2 |= READ_ONCE(host_vcpu->arch.hcr_el2) &
arch/arm64/kvm/hyp/nvhe/pkvm.c
333
unsigned long host_arch_flags = READ_ONCE(host_kvm->arch.flags);
arch/arm64/kvm/hyp/nvhe/pkvm.c
426
hyp_vm->kvm.arch.pkvm.is_protected = READ_ONCE(host_kvm->arch.pkvm.is_protected);
arch/arm64/kvm/hyp/nvhe/pkvm.c
453
sve_max_vl = min(READ_ONCE(host_vcpu->arch.sve_max_vl), kvm_host_sve_max_vl);
arch/arm64/kvm/hyp/nvhe/pkvm.c
455
sve_state = kern_hyp_va(READ_ONCE(host_vcpu->arch.sve_state));
arch/arm64/kvm/hyp/nvhe/pkvm.c
516
hyp_vcpu->vcpu.vcpu_id = READ_ONCE(host_vcpu->vcpu_id);
arch/arm64/kvm/hyp/nvhe/pkvm.c
517
hyp_vcpu->vcpu.vcpu_idx = READ_ONCE(host_vcpu->vcpu_idx);
arch/arm64/kvm/hyp/nvhe/pkvm.c
520
hyp_vcpu->vcpu.arch.cflags = READ_ONCE(host_vcpu->arch.cflags);
arch/arm64/kvm/hyp/nvhe/pkvm.c
750
nr_vcpus = READ_ONCE(host_kvm->created_vcpus);
arch/arm64/kvm/hyp/nvhe/pkvm.c
756
handle = READ_ONCE(host_kvm->arch.pkvm.handle);
arch/arm64/kvm/hyp/pgtable.c
163
.old = READ_ONCE(*ptep),
arch/arm64/kvm/hyp/pgtable.c
193
ctx.old = READ_ONCE(*ptep);
arch/arm64/kvm/hypercalls.c
499
val = READ_ONCE(smccc_feat->std_bmap);
arch/arm64/kvm/hypercalls.c
502
val = READ_ONCE(smccc_feat->std_hyp_bmap);
arch/arm64/kvm/hypercalls.c
505
val = READ_ONCE(smccc_feat->vendor_hyp_bmap);
arch/arm64/kvm/hypercalls.c
508
val = READ_ONCE(smccc_feat->vendor_hyp_bmap_2);
arch/arm64/kvm/pkvm.c
188
return READ_ONCE(kvm->arch.pkvm.is_created);
arch/arm64/kvm/pvtime.c
28
vcpu->arch.steal.last_steal = READ_ONCE(current->sched_info.run_delay);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
232
ret = extract_bytes(READ_ONCE(irq->mpidr), addr & 7, len);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
499
old_propbaser = READ_ONCE(dist->propbaser);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
530
old_pendbaser = READ_ONCE(vgic_cpu->pendbaser);
arch/arm64/mm/fault.c
155
pgd = READ_ONCE(*pgdp);
arch/arm64/mm/fault.c
168
p4d = READ_ONCE(*p4dp);
arch/arm64/mm/fault.c
174
pud = READ_ONCE(*pudp);
arch/arm64/mm/fault.c
180
pmd = READ_ONCE(*pmdp);
arch/arm64/mm/fixmap.c
45
pmd_t pmd = READ_ONCE(*pmdp);
arch/arm64/mm/fixmap.c
59
pud_t pud = READ_ONCE(*pudp);
arch/arm64/mm/fixmap.c
77
p4d_t p4d = READ_ONCE(*p4dp);
arch/arm64/mm/hugetlbpage.c
265
if (want_pmd_share(vma, addr) && pud_none(READ_ONCE(*pudp)))
arch/arm64/mm/hugetlbpage.c
287
if (!pgd_present(READ_ONCE(*pgdp)))
arch/arm64/mm/hugetlbpage.c
291
if (!p4d_present(READ_ONCE(*p4dp)))
arch/arm64/mm/hugetlbpage.c
295
pud = READ_ONCE(*pudp);
arch/arm64/mm/hugetlbpage.c
307
pmd = READ_ONCE(*pmdp);
arch/arm64/mm/kasan_init.c
105
if (pgd_none(READ_ONCE(*pgdp))) {
arch/arm64/mm/kasan_init.c
141
} while (pmdp++, addr = next, addr != end && pmd_none(READ_ONCE(*pmdp)));
arch/arm64/mm/kasan_init.c
153
} while (pudp++, addr = next, addr != end && pud_none(READ_ONCE(*pudp)));
arch/arm64/mm/kasan_init.c
165
} while (p4dp++, addr = next, addr != end && p4d_none(READ_ONCE(*p4dp)));
arch/arm64/mm/kasan_init.c
259
pgd_t pgd = READ_ONCE(swapper_pg_dir[idx]);
arch/arm64/mm/kasan_init.c
283
pgd_t pgd = READ_ONCE(swapper_pg_dir[pgd_idx]);
arch/arm64/mm/kasan_init.c
65
if (pmd_none(READ_ONCE(*pmdp))) {
arch/arm64/mm/kasan_init.c
79
if (pud_none(READ_ONCE(*pudp))) {
arch/arm64/mm/kasan_init.c
92
if (p4d_none(READ_ONCE(*p4dp))) {
arch/arm64/mm/mmu.c
1478
pmd = READ_ONCE(*pmdp);
arch/arm64/mm/mmu.c
1511
pud = READ_ONCE(*pudp);
arch/arm64/mm/mmu.c
1544
p4d = READ_ONCE(*p4dp);
arch/arm64/mm/mmu.c
1570
pgd = READ_ONCE(*pgdp);
arch/arm64/mm/mmu.c
1626
pmd = READ_ONCE(*pmdp);
arch/arm64/mm/mmu.c
1647
if (!pmd_none(READ_ONCE(pmdp[i])))
arch/arm64/mm/mmu.c
1666
pud = READ_ONCE(*pudp);
arch/arm64/mm/mmu.c
1687
if (!pud_none(READ_ONCE(pudp[i])))
arch/arm64/mm/mmu.c
1706
p4d = READ_ONCE(*p4dp);
arch/arm64/mm/mmu.c
1727
if (!p4d_none(READ_ONCE(p4dp[i])))
arch/arm64/mm/mmu.c
1745
pgd = READ_ONCE(*pgdp);
arch/arm64/mm/mmu.c
1766
return pmd_sect(READ_ONCE(*pmdp));
arch/arm64/mm/mmu.c
1799
if (!pgattr_change_is_safe(READ_ONCE(pud_val(*pudp)),
arch/arm64/mm/mmu.c
1813
if (!pgattr_change_is_safe(READ_ONCE(pmd_val(*pmdp)),
arch/arm64/mm/mmu.c
1830
if (!pud_sect(READ_ONCE(*pudp)))
arch/arm64/mm/mmu.c
1838
if (!pmd_sect(READ_ONCE(*pmdp)))
arch/arm64/mm/mmu.c
1850
pmd = READ_ONCE(*pmdp);
arch/arm64/mm/mmu.c
1883
pud = READ_ONCE(*pudp);
arch/arm64/mm/mmu.c
204
pmd_t pmd = READ_ONCE(*pmdp);
arch/arm64/mm/mmu.c
260
pmd_t old_pmd = READ_ONCE(*pmdp);
arch/arm64/mm/mmu.c
274
READ_ONCE(pmd_val(*pmdp))));
arch/arm64/mm/mmu.c
284
pmd_val(old_pmd) != READ_ONCE(pmd_val(*pmdp)));
arch/arm64/mm/mmu.c
300
pud_t pud = READ_ONCE(*pudp);
arch/arm64/mm/mmu.c
357
p4d_t p4d = READ_ONCE(*p4dp);
arch/arm64/mm/mmu.c
380
pud_t old_pud = READ_ONCE(*pudp);
arch/arm64/mm/mmu.c
397
READ_ONCE(pud_val(*pudp))));
arch/arm64/mm/mmu.c
405
pud_val(old_pud) != READ_ONCE(pud_val(*pudp)));
arch/arm64/mm/mmu.c
423
pgd_t pgd = READ_ONCE(*pgdp);
arch/arm64/mm/mmu.c
446
p4d_t old_p4d = READ_ONCE(*p4dp);
arch/arm64/mm/mmu.c
456
p4d_val(old_p4d) != READ_ONCE(p4d_val(*p4dp)));
arch/arm64/mm/pageattr.c
396
if (pgd_none(READ_ONCE(*pgdp)))
arch/arm64/mm/pageattr.c
400
if (p4d_none(READ_ONCE(*p4dp)))
arch/arm64/mm/pageattr.c
404
pud = READ_ONCE(*pudp);
arch/arm64/mm/pageattr.c
411
pmd = READ_ONCE(*pmdp);
arch/arm64/mm/trans_pgd.c
103
pmd_t pmd = READ_ONCE(*src_pmdp);
arch/arm64/mm/trans_pgd.c
129
if (p4d_none(READ_ONCE(*dst_p4dp))) {
arch/arm64/mm/trans_pgd.c
139
pud_t pud = READ_ONCE(*src_pudp);
arch/arm64/mm/trans_pgd.c
165
if (pgd_none(READ_ONCE(*dst_pgdp))) {
arch/arm64/mm/trans_pgd.c
176
if (p4d_none(READ_ONCE(*src_p4dp)))
arch/arm64/mm/trans_pgd.c
195
if (pgd_none(READ_ONCE(*src_pgdp)))
arch/arm64/mm/trans_pgd.c
93
if (pud_none(READ_ONCE(*dst_pudp))) {
arch/csky/include/asm/atomic.h
18
return READ_ONCE(v->counter);
arch/hexagon/include/asm/atomic.h
31
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/loongarch/include/asm/atomic.h
110
#define arch_atomic64_read(v) READ_ONCE((v)->counter)
arch/loongarch/include/asm/atomic.h
44
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/loongarch/include/asm/barrier.h
93
typeof(*p) ___p1 = READ_ONCE(*p); \
arch/loongarch/include/asm/pgtable.h
124
#define ptep_get(ptep) READ_ONCE(*(ptep))
arch/loongarch/include/asm/pgtable.h
125
#define pmdp_get(pmdp) READ_ONCE(*(pmdp))
arch/loongarch/mm/kasan_init.c
208
} while (pudp++, addr = next, addr != end && __pud_none(early, READ_ONCE(*pudp)));
arch/loongarch/mm/kasan_init.c
220
} while (p4dp++, addr = next, addr != end && __p4d_none(early, READ_ONCE(*p4dp)));
arch/m68k/amiga/pcmcia.c
32
READ_ONCE(gayle_reset);
arch/m68k/apollo/config.c
153
READ_ONCE(*(at + 3));
arch/m68k/apollo/config.c
154
READ_ONCE(*(at + 5));
arch/m68k/include/asm/atomic.h
19
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/m68k/virt/ints.c
94
if (READ_ONCE(in_nmi))
arch/mips/cavium-octeon/executive/cvmx-l2c.c
308
READ_ONCE(*ptr);
arch/mips/include/asm/atomic.h
29
return READ_ONCE(v->counter); \
arch/mips/mm/context.c
210
old_active_mmid = READ_ONCE(cpu_data[cpu].asid_cache);
arch/openrisc/include/asm/atomic.h
128
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/openrisc/include/asm/cmpxchg.h
106
oldv = READ_ONCE(*p);
arch/openrisc/include/asm/cmpxchg.h
73
load32 = READ_ONCE(*p);
arch/parisc/include/asm/atomic.h
232
return READ_ONCE((v)->counter);
arch/parisc/include/asm/atomic.h
73
return READ_ONCE((v)->counter);
arch/parisc/include/asm/pgtable.h
448
return READ_ONCE(*ptep);
arch/parisc/include/asm/spinlock.h
25
lock_val = READ_ONCE(*a);
arch/powerpc/include/asm/barrier.h
81
typeof(*p) ___p1 = READ_ONCE(*p); \
arch/powerpc/include/asm/kvm_book3s_64.h
430
old_pte = READ_ONCE(*ptep);
arch/powerpc/include/asm/local.h
22
return READ_ONCE(l->v);
arch/powerpc/include/asm/nohash/32/pte-8xx.h
231
pte_basic_t val = READ_ONCE(ptep->pte);
arch/powerpc/include/asm/nohash/tlbflush.h
37
unsigned int pid = READ_ONCE(mm->context.id);
arch/powerpc/include/asm/paravirt.h
37
__be32 yield_count = READ_ONCE(lppaca_of(cpu).yield_count);
arch/powerpc/include/asm/qspinlock.h
73
return READ_ONCE(lock->val);
arch/powerpc/include/asm/qspinlock.h
83
return !!(READ_ONCE(lock->val) & _Q_TAIL_CPU_MASK);
arch/powerpc/include/asm/simple_spinlock.h
42
return !arch_spin_value_unlocked(READ_ONCE(*lock));
arch/powerpc/kernel/interrupt.c
133
if (!READ_ONCE(*validp))
arch/powerpc/kernel/interrupt.c
143
if (!READ_ONCE(*validp))
arch/powerpc/kernel/interrupt.c
169
if (!READ_ONCE(*validp))
arch/powerpc/kernel/irq_64.c
258
irq_happened = READ_ONCE(local_paca->irq_happened);
arch/powerpc/kernel/irq_64.c
282
irq_happened = READ_ONCE(local_paca->irq_happened);
arch/powerpc/kernel/mce_power.c
47
pte = READ_ONCE(*ptep);
arch/powerpc/kernel/smp.c
346
} while (READ_ONCE(info->messages));
arch/powerpc/kernel/smp.c
458
fn = READ_ONCE(nmi_ipi_function);
arch/powerpc/kernel/watchdog.c
541
threshold += (READ_ONCE(wd_timeout_pct) * threshold) / 100;
arch/powerpc/kvm/book3s_64_mmu_hv.c
619
pte = READ_ONCE(*ptep);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1098
pte = READ_ONCE(*ptep);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1365
p4d = READ_ONCE(*p4dp);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1372
pud = READ_ONCE(*pudp);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1384
pmd = READ_ONCE(*pmdp);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1396
pte = pte_val(READ_ONCE(*ptep));
arch/powerpc/kvm/book3s_64_mmu_radix.c
856
pte = READ_ONCE(*ptep);
arch/powerpc/kvm/book3s_hv.c
1520
pcpu = READ_ONCE(v->cpu);
arch/powerpc/kvm/book3s_hv.c
170
vcpu = READ_ONCE(vc->runnable_threads[i]);
arch/powerpc/kvm/book3s_hv.c
243
cpu = READ_ONCE(vcpu->arch.thread_cpu);
arch/powerpc/kvm/book3s_hv_builtin.c
414
host_ipi = READ_ONCE(local_paca->kvm_hstate.host_ipi);
arch/powerpc/kvm/book3s_hv_builtin.c
474
host_ipi = READ_ONCE(local_paca->kvm_hstate.host_ipi);
arch/powerpc/kvm/book3s_hv_rm_xics.c
246
old_state = new_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_hv_rm_xics.c
445
old_state = new_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_hv_rm_xics.c
509
old_state = new_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_hv_rm_xics.c
575
old_state = new_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_hv_rm_xics.c
650
old_state = new_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_hv_rm_xics.c
89
old = new = READ_ONCE(rm_core[core].rm_state);
arch/powerpc/kvm/book3s_xics.c
1159
old_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_xics.c
346
old_state = new_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_xics.c
549
old_state = new_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_xics.c
603
old_state = new_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_xics.c
670
old_state = new_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_xics.c
715
state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_xics.c
756
old_state = new_state = READ_ONCE(icp->state);
arch/powerpc/kvm/book3s_xics.c
971
state.raw = READ_ONCE(icp->state.raw);
arch/powerpc/kvm/book3s_xive.c
171
qpage = READ_ONCE(q->qpage);
arch/powerpc/kvm/book3s_xive.c
407
qpage = READ_ONCE(q->qpage);
arch/powerpc/kvm/e500_mmu_host.c
380
pte_t pte = READ_ONCE(*ptep);
arch/powerpc/lib/qspinlock.c
319
if (READ_ONCE(lock->val) == val) {
arch/powerpc/lib/qspinlock.c
364
next = READ_ONCE(node->next);
arch/powerpc/lib/qspinlock.c
393
u32 val = READ_ONCE(lock->val);
arch/powerpc/lib/qspinlock.c
430
if (!READ_ONCE(node->locked)) {
arch/powerpc/lib/qspinlock.c
474
val = READ_ONCE(lock->val);
arch/powerpc/lib/qspinlock.c
584
while (!READ_ONCE(node->locked)) {
arch/powerpc/lib/qspinlock.c
601
next = READ_ONCE(node->next);
arch/powerpc/lib/qspinlock.c
613
val = READ_ONCE(lock->val);
arch/powerpc/lib/qspinlock.c
671
next = READ_ONCE(node->next);
arch/powerpc/lib/qspinlock.c
674
while (!(next = READ_ONCE(node->next)))
arch/powerpc/mm/book3s64/hash_4k.c
36
pte_t pte = READ_ONCE(*ptep);
arch/powerpc/mm/book3s64/hash_64k.c
243
pte_t pte = READ_ONCE(*ptep);
arch/powerpc/mm/book3s64/hash_64k.c
54
pte_t pte = READ_ONCE(*ptep);
arch/powerpc/mm/book3s64/hash_hugepage.c
36
pmd_t pmd = READ_ONCE(*pmdp);
arch/powerpc/mm/book3s64/radix_pgtable.c
1157
if (pmd_none(READ_ONCE(*pmd))) {
arch/powerpc/mm/book3s64/radix_pgtable.c
1336
if (pmd_leaf(READ_ONCE(*pmd))) {
arch/powerpc/mm/pgtable.c
471
p4d = READ_ONCE(*p4dp);
arch/powerpc/mm/pgtable.c
489
pud = READ_ONCE(*pudp);
arch/powerpc/mm/pgtable.c
504
pmd = READ_ONCE(*pmdp);
arch/powerpc/perf/imc-pmu.c
1077
data = be64_to_cpu(READ_ONCE(*addr));
arch/powerpc/perf/imc-pmu.c
1284
if (be64_to_cpu(READ_ONCE(mem->tb1)) > *prev_tb)
arch/powerpc/perf/imc-pmu.c
1285
*prev_tb = be64_to_cpu(READ_ONCE(mem->tb1));
arch/powerpc/perf/imc-pmu.c
1289
if ((be64_to_cpu(READ_ONCE(mem->tb1)) & IMC_TRACE_RECORD_TB1_MASK) !=
arch/powerpc/perf/imc-pmu.c
1290
be64_to_cpu(READ_ONCE(mem->tb2)))
arch/powerpc/perf/imc-pmu.c
1294
data->ip = be64_to_cpu(READ_ONCE(mem->ip));
arch/powerpc/perf/imc-pmu.c
1302
switch (IMC_TRACE_RECORD_VAL_HVPR(be64_to_cpu(READ_ONCE(mem->val)))) {
arch/powerpc/platforms/powernv/idle.c
262
u64 s = READ_ONCE(*state);
arch/powerpc/platforms/powernv/idle.c
265
BUG_ON(!(READ_ONCE(*lock) & PNV_CORE_IDLE_LOCK_BIT));
arch/powerpc/platforms/powernv/opal-irqchip.c
46
e = READ_ONCE(last_outstanding_events) & opal_event_irqchip.mask;
arch/powerpc/platforms/powernv/opal-irqchip.c
72
if (READ_ONCE(last_outstanding_events) & opal_event_irqchip.mask)
arch/powerpc/platforms/powernv/opal-msglog.c
44
out_pos = be32_to_cpu(READ_ONCE(mc->out_pos));
arch/powerpc/platforms/powernv/pci-ioda-tce.c
92
unsigned long oldtce, tce = be64_to_cpu(READ_ONCE(tmp[n]));
arch/powerpc/platforms/pseries/lpar.c
670
return tb_to_ns(be64_to_cpu(READ_ONCE(lppaca->enqueue_dispatch_tb)) +
arch/powerpc/platforms/pseries/lpar.c
671
be64_to_cpu(READ_ONCE(lppaca->ready_enqueue_tb)));
arch/powerpc/platforms/pseries/mobility.c
603
if (READ_ONCE(info->done) == false) {
arch/powerpc/platforms/pseries/papr_scm.c
1066
READ_ONCE(p->health_bitmap_inject_mask));
arch/powerpc/platforms/pseries/papr_scm.c
1128
health = READ_ONCE(p->health_bitmap);
arch/powerpc/platforms/pseries/papr_scm.c
868
mask = READ_ONCE(p->health_bitmap_inject_mask);
arch/powerpc/xmon/xmon.c
3245
unsigned int p_state = READ_ONCE(tsk->__state);
arch/riscv/include/asm/atomic.h
29
return READ_ONCE(v->counter);
arch/riscv/include/asm/atomic.h
40
return READ_ONCE(v->counter);
arch/riscv/include/asm/barrier.h
63
typeof(*p) ___p1 = READ_ONCE(*p); \
arch/riscv/include/asm/barrier.h
74
VAL = READ_ONCE(*__PTR); \
arch/riscv/include/asm/pgtable.h
686
pte_t read_pte = READ_ONCE(*ptep);
arch/riscv/include/asm/vector.h
58
return READ_ONCE(current->thread.riscv_v_flags);
arch/riscv/kernel/hibernate.c
174
if (pmd_none(READ_ONCE(*dst_pmdp))) {
arch/riscv/kernel/hibernate.c
186
pte_t pte = READ_ONCE(*src_ptep);
arch/riscv/kernel/hibernate.c
203
if (pud_none(READ_ONCE(*dst_pudp))) {
arch/riscv/kernel/hibernate.c
215
pmd_t pmd = READ_ONCE(*src_pmdp);
arch/riscv/kernel/hibernate.c
242
if (p4d_none(READ_ONCE(*dst_p4dp))) {
arch/riscv/kernel/hibernate.c
254
pud_t pud = READ_ONCE(*src_pudp);
arch/riscv/kernel/hibernate.c
281
if (pgd_none(READ_ONCE(*dst_pgdp))) {
arch/riscv/kernel/hibernate.c
293
p4d_t p4d = READ_ONCE(*src_p4dp);
arch/riscv/kernel/hibernate.c
320
pgd_t pgd = READ_ONCE(*src_pgdp);
arch/riscv/kernel/paravirt.c
93
sequence = READ_ONCE(st->sequence);
arch/riscv/kernel/paravirt.c
95
steal = READ_ONCE(st->steal);
arch/riscv/kernel/paravirt.c
98
sequence != READ_ONCE(st->sequence));
arch/riscv/kernel/stacktrace.c
29
val = READ_ONCE(addr); \
arch/riscv/kernel/vector.c
240
if (READ_ONCE(riscv_v_implicit_uacc))
arch/riscv/kvm/aia.c
57
if (READ_ONCE(vcpu->arch.irqs_pending_mask[1])) {
arch/riscv/kvm/aia.c
59
val = READ_ONCE(vcpu->arch.irqs_pending[1]) & mask;
arch/riscv/kvm/aia.c
83
if (READ_ONCE(vcpu->arch.irqs_pending[1]) &
arch/riscv/kvm/mmu.c
114
gstage.vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/kvm/mmu.c
146
gstage.vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/kvm/mmu.c
255
gstage.vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/kvm/mmu.c
280
gstage.vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/kvm/mmu.c
29
gstage.vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/kvm/mmu.c
303
gstage.vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/kvm/mmu.c
468
gstage.vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/kvm/mmu.c
54
gstage.vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/kvm/mmu.c
591
gstage.vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/kvm/mmu.c
594
pgd = READ_ONCE(kvm->arch.pgd);
arch/riscv/kvm/mmu.c
609
hgatp |= (READ_ONCE(k->vmid.vmid) << HGATP_VMID_SHIFT) & HGATP_VMID;
arch/riscv/kvm/mmu.c
94
gstage.vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/kvm/tlb.c
180
vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
arch/riscv/kvm/tlb.c
200
unsigned long vmid = READ_ONCE(v->vmid);
arch/riscv/kvm/tlb.c
211
unsigned long vmid = READ_ONCE(v->vmid);
arch/riscv/kvm/tlb.c
464
PAGE_SHIFT, READ_ONCE(kvm->arch.vmid.vmid));
arch/riscv/kvm/vcpu.c
352
if (READ_ONCE(vcpu->arch.irqs_pending_mask[0])) {
arch/riscv/kvm/vcpu.c
354
val = READ_ONCE(vcpu->arch.irqs_pending[0]) & mask;
arch/riscv/kvm/vcpu.c
453
if (READ_ONCE(vcpu->arch.irqs_pending[0]) & ie)
arch/riscv/kvm/vcpu.c
489
return READ_ONCE(vcpu->arch.mp_state.mp_state) == KVM_MP_STATE_STOPPED;
arch/riscv/kvm/vcpu.c
495
*mp_state = READ_ONCE(vcpu->arch.mp_state);
arch/riscv/kvm/vcpu_sbi_replace.c
107
vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
arch/riscv/kvm/vcpu_sbi_replace.c
116
vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
arch/riscv/kvm/vcpu_sbi_sta.c
70
vcpu->arch.sta.last_steal = READ_ONCE(current->sched_info.run_delay);
arch/riscv/kvm/vcpu_sbi_v01.c
81
vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
arch/riscv/kvm/vcpu_sbi_v01.c
88
vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
arch/riscv/kvm/vmid.c
117
WRITE_ONCE(vmid->vmid_version, READ_ONCE(vmid_version));
arch/riscv/kvm/vmid.c
62
return unlikely(READ_ONCE(vmid->vmid_version) !=
arch/riscv/kvm/vmid.c
63
READ_ONCE(vmid_version));
arch/riscv/kvm/vmid.c
93
WRITE_ONCE(vmid_version, READ_ONCE(vmid_version) + 1);
arch/s390/include/asm/barrier.h
49
typeof(*p) ___p1 = READ_ONCE(*p); \
arch/s390/include/asm/cmpxchg.h
183
old = READ_ONCE(*(u32 *)ptr);
arch/s390/include/asm/cmpxchg.h
198
old = READ_ONCE(*(u32 *)ptr);
arch/s390/include/asm/cmpxchg.h
214
u32 old = READ_ONCE(*(u32 *)ptr);
arch/s390/include/asm/cmpxchg.h
221
u64 old = READ_ONCE(*(u64 *)ptr);
arch/s390/include/asm/cmpxchg.h
49
prev = READ_ONCE(*(u32 *)ptr);
arch/s390/include/asm/cmpxchg.h
72
prev = READ_ONCE(*(u32 *)ptr);
arch/s390/include/asm/fpu.h
180
flags = ~READ_ONCE(thread->ufpu_flags) & (KERNEL_FPC | KERNEL_VXR);
arch/s390/include/asm/fpu.h
194
uflags = READ_ONCE(thread->ufpu_flags);
arch/s390/include/asm/percpu.h
26
prev__ = READ_ONCE(*ptr__); \
arch/s390/include/asm/pgtable.h
1414
#define pgd_offset(mm, address) pgd_offset_raw(READ_ONCE((mm)->pgd), address)
arch/s390/include/asm/preempt.h
139
return unlikely(READ_ONCE(get_lowcore()->preempt_count) == preempt_offset);
arch/s390/include/asm/preempt.h
49
old = READ_ONCE(get_lowcore()->preempt_count);
arch/s390/include/asm/preempt.h
76
return !(READ_ONCE(get_lowcore()->preempt_count) & PREEMPT_NEED_RESCHED);
arch/s390/include/asm/skey.h
27
if (READ_ONCE(skey_regions_initialized))
arch/s390/include/asm/spinlock.h
156
old = READ_ONCE(rw->cnts);
arch/s390/include/asm/spinlock.h
164
old = READ_ONCE(rw->cnts);
arch/s390/include/asm/spinlock.h
71
return READ_ONCE(lp->lock) != 0;
arch/s390/include/asm/tlbflush.h
56
gmap_asce = READ_ONCE(mm->context.gmap_asce);
arch/s390/kernel/diag/diag310.c
156
if (READ_ONCE(features_available))
arch/s390/kernel/diag/diag310.c
179
stride = READ_ONCE(memtop_stride);
arch/s390/kernel/diag/diag310.c
198
pages = READ_ONCE(memtop_pages[level - 1]);
arch/s390/kernel/facility.c
15
r = READ_ONCE(size);
arch/s390/kernel/idle.c
60
return sysfs_emit(buf, "%lu\n", READ_ONCE(idle->idle_count));
arch/s390/kernel/idle.c
69
return sysfs_emit(buf, "%lu\n", READ_ONCE(idle->idle_time) >> 12);
arch/s390/kernel/time.c
547
while (READ_ONCE(sync->in_sync) == 0)
arch/s390/kvm/dat.c
1006
oldcrste = READ_ONCE(*crstep);
arch/s390/kvm/dat.c
1228
if (!READ_ONCE(*pgste_of(ptep)).cmma_d)
arch/s390/kvm/dat.c
191
union pte old = READ_ONCE(*ptep);
arch/s390/kvm/dat.c
243
old = READ_ONCE(*pmdp);
arch/s390/kvm/dat.c
287
old = READ_ONCE(*pmdp);
arch/s390/kvm/dat.c
308
old = READ_ONCE(*crstep);
arch/s390/kvm/dat.c
337
old = READ_ONCE(*crstep);
arch/s390/kvm/dat.c
406
entry = READ_ONCE(**last);
arch/s390/kvm/dat.c
419
entry = READ_ONCE(**last);
arch/s390/kvm/dat.c
426
entry = READ_ONCE(**last);
arch/s390/kvm/dat.c
439
entry = READ_ONCE(**last);
arch/s390/kvm/dat.c
446
entry = READ_ONCE(**last);
arch/s390/kvm/dat.c
464
entry = READ_ONCE(**last);
arch/s390/kvm/dat.c
472
entry = READ_ONCE(**last);
arch/s390/kvm/dat.c
478
entry = READ_ONCE(**last);
arch/s390/kvm/dat.c
496
entry = READ_ONCE(**last);
arch/s390/kvm/dat.c
504
entry = READ_ONCE(**last);
arch/s390/kvm/dat.c
519
if (pte_hole(READ_ONCE(table->ptes[idx]))) {
arch/s390/kvm/dat.c
549
crste = READ_ONCE(*walk->last);
arch/s390/kvm/dat.c
561
crste = READ_ONCE(*walk->last);
arch/s390/kvm/dat.c
628
crste = READ_ONCE(*crstep);
arch/s390/kvm/dat.c
767
union crste crste = READ_ONCE(*crstep);
arch/s390/kvm/dat.c
845
union pte new_pte, pte = READ_ONCE(*ptep);
arch/s390/kvm/dat.c
859
union crste new_crste, crste = READ_ONCE(*crstep);
arch/s390/kvm/dat.h
849
if (READ_ONCE(pgstep->val) & PGSTE_PCL_BIT)
arch/s390/kvm/dat.h
944
oldcrste = READ_ONCE(*crstep);
arch/s390/kvm/diag.c
205
tcpu_cpu = READ_ONCE(tcpu->cpu);
arch/s390/kvm/gaccess.c
1484
oldcrste = READ_ONCE(*host);
arch/s390/kvm/gaccess.c
1500
while (!dat_crstep_xchg_atomic(table, READ_ONCE(*table), newcrste, gfn, sg->asce))
arch/s390/kvm/gaccess.c
152
old = READ_ONCE(*ic);
arch/s390/kvm/gaccess.c
1598
parent = READ_ONCE(sg->parent);
arch/s390/kvm/gaccess.c
1602
if (READ_ONCE(sg->parent) != parent)
arch/s390/kvm/gaccess.c
174
old = READ_ONCE(*ic);
arch/s390/kvm/gaccess.c
190
old = READ_ONCE(*ic);
arch/s390/kvm/gaccess.c
207
old = READ_ONCE(*ic);
arch/s390/kvm/gmap.c
1075
if (READ_ONCE(*ptep).val != _PTE_EMPTY.val)
arch/s390/kvm/gmap.c
1171
parent = READ_ONCE(sg->parent);
arch/s390/kvm/gmap.c
1175
if (READ_ONCE(sg->parent) != parent)
arch/s390/kvm/gmap.c
288
pte = READ_ONCE(*ptep);
arch/s390/kvm/gmap.c
317
crste = READ_ONCE(*crstep);
arch/s390/kvm/gmap.c
444
union pte pte = READ_ONCE(*ptep);
arch/s390/kvm/gmap.c
485
crste = READ_ONCE(*table);
arch/s390/kvm/gmap.c
524
union crste newcrste, oldcrste = READ_ONCE(*f->crstep);
arch/s390/kvm/gmap.c
557
union pte newpte, oldpte = READ_ONCE(*f->ptep);
arch/s390/kvm/gmap.c
665
oldval = READ_ONCE(*f->crstep);
arch/s390/kvm/gmap.c
730
oldcrste = READ_ONCE(*crstep);
arch/s390/kvm/gmap.c
839
while (!dat_crstep_xchg_atomic(crstep, READ_ONCE(*crstep), _PMD_EMPTY, c_gfn, gmap->asce))
arch/s390/kvm/gmap.c
856
crste = READ_ONCE(*crstep);
arch/s390/kvm/gmap.c
866
crste = READ_ONCE(*crstep);
arch/s390/kvm/interrupt.c
1936
u64 type = READ_ONCE(inti->type);
arch/s390/kvm/interrupt.c
196
word = READ_ONCE(gisa->u64.word[0]);
arch/s390/kvm/interrupt.c
219
word = READ_ONCE(gisa->u64.word[0]);
arch/s390/kvm/interrupt.c
240
word = READ_ONCE(gi->origin->u64.word[0]);
arch/s390/kvm/interrupt.c
242
alert_mask = READ_ONCE(gi->alert.mask);
arch/s390/kvm/interrupt.c
259
return READ_ONCE(gisa->ipm);
arch/s390/kvm/interrupt.c
70
old_val = READ_ONCE(*sigp_ctrl);
arch/s390/kvm/kvm-s390.c
1924
old = READ_ONCE(sca->utility);
arch/s390/kvm/kvm-s390.c
3844
READ_ONCE(halt_poll_max_steal)) {
arch/s390/kvm/vsie.c
1576
struct kvm_s390_sie_block *scb = READ_ONCE(vcpu->arch.vsie_block);
arch/s390/kvm/vsie.c
317
const uint32_t crycbd_o = READ_ONCE(scb_o->crycbd);
arch/s390/kvm/vsie.c
395
const uint32_t new_ibc = READ_ONCE(__new_ibc) & 0x0fffU;
arch/s390/kvm/vsie.c
469
const uint32_t new_prefix = READ_ONCE(__new_prefix);
arch/s390/kvm/vsie.c
470
const bool wants_tx = READ_ONCE(scb_o->ecb) & ECB_TE;
arch/s390/kvm/vsie.c
518
new_mso = READ_ONCE(scb_o->mso) & 0xfffffffffff00000UL;
arch/s390/kvm/vsie.c
750
gpa = READ_ONCE(scb_o->scaol) & ~0xfUL;
arch/s390/kvm/vsie.c
752
gpa |= (u64) READ_ONCE(scb_o->scaoh) << 32;
arch/s390/kvm/vsie.c
773
gpa = READ_ONCE(scb_o->itdba) & ~0xffUL;
arch/s390/kvm/vsie.c
789
gpa = READ_ONCE(scb_o->gvrd) & ~0x1ffUL;
arch/s390/kvm/vsie.c
808
gpa = READ_ONCE(scb_o->riccbd) & ~0x3fUL;
arch/s390/kvm/vsie.c
828
gpa = READ_ONCE(scb_o->sdnxo) & ~0xfUL;
arch/s390/kvm/vsie.c
829
sdnxc = READ_ONCE(scb_o->sdnxo) & 0xfUL;
arch/s390/kvm/vsie.c
992
__u32 fac = READ_ONCE(vsie_page->scb_o->fac);
arch/s390/lib/spinlock.c
169
old = READ_ONCE(lp->lock);
arch/s390/lib/spinlock.c
207
while (READ_ONCE(node->prev) != NULL) {
arch/s390/lib/spinlock.c
221
old = READ_ONCE(lp->lock);
arch/s390/lib/spinlock.c
241
while ((next = READ_ONCE(node->next)) == NULL)
arch/s390/lib/spinlock.c
257
owner = arch_spin_yield_target(READ_ONCE(lp->lock), NULL);
arch/s390/lib/spinlock.c
297
owner = READ_ONCE(lp->lock);
arch/s390/lib/spinlock.c
311
while (READ_ONCE(rw->cnts) & 0x10000)
arch/s390/lib/spinlock.c
323
while (READ_ONCE(rw->cnts) & 0x10000)
arch/s390/lib/spinlock.c
340
old = READ_ONCE(rw->cnts);
arch/s390/lib/spinlock.c
356
cpu = READ_ONCE(lp->lock) & _Q_LOCK_CPU_MASK;
arch/s390/lib/test_unwind.c
281
return READ_ONCE(u)->ret;
arch/s390/lib/test_unwind.c
377
struct unwindme *u = READ_ONCE(unwindme);
arch/s390/mm/pgtable.c
46
asce = READ_ONCE(mm->context.gmap_asce);
arch/s390/mm/pgtable.c
66
asce = READ_ONCE(mm->context.gmap_asce);
arch/s390/pci/pci_insn.c
220
u64 req = ZPCI_CREATE_REQ(READ_ONCE(entry->fh), entry->bar, len);
arch/s390/pci/pci_insn.c
313
u64 req = ZPCI_CREATE_REQ(READ_ONCE(entry->fh), entry->bar, len);
arch/sh/include/asm/atomic.h
22
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/sh/include/asm/cmpxchg-xchg.h
32
oldv = READ_ONCE(*p);
arch/sparc/include/asm/atomic_32.h
49
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/sparc/include/asm/atomic_64.h
17
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/sparc/include/asm/atomic_64.h
18
#define arch_atomic64_read(v) READ_ONCE((v)->counter)
arch/sparc/include/asm/barrier_64.h
50
typeof(*p) ___p1 = READ_ONCE(*p); \
arch/sparc/include/asm/vvar.h
46
ret = READ_ONCE(s->seq);
arch/x86/events/core.c
1095
READ_ONCE(cpuc->excl_cntrs->exclusive_present))
arch/x86/events/core.c
2538
if (READ_ONCE(x86_pmu.attr_rdpmc) &&
arch/x86/events/intel/bts.c
311
int state = READ_ONCE(bts->state);
arch/x86/events/intel/bts.c
349
state = READ_ONCE(bts->state);
arch/x86/events/intel/bts.c
378
if (READ_ONCE(bts->state) != BTS_STATE_ACTIVE)
arch/x86/events/intel/bts.c
486
if (READ_ONCE(bts->state) == BTS_STATE_STOPPED)
arch/x86/events/intel/pt.c
1516
if (!READ_ONCE(pt->handle_nmi))
arch/x86/events/intel/pt.c
1607
if (READ_ONCE(pt->resume_allowed)) {
arch/x86/events/intel/pt.c
1620
READ_ONCE(pt->resume_allowed))
arch/x86/events/intel/pt.c
1654
if (READ_ONCE(pt->pause_allowed))
arch/x86/events/intel/pt.c
1747
if (READ_ONCE(pt->handle_nmi)) {
arch/x86/events/intel/pt.c
424
if (READ_ONCE(event->hw.aux_paused))
arch/x86/events/intel/pt.c
428
if (READ_ONCE(pt->vmx_on))
arch/x86/events/intel/pt.c
566
u64 ctl = READ_ONCE(event->hw.aux_config);
arch/x86/events/intel/pt.c
573
if (!READ_ONCE(pt->vmx_on))
arch/x86/hyperv/hv_spinlock.c
50
if (READ_ONCE(*byte) == val) {
arch/x86/include/asm/barrier.h
68
typeof(*p) ___p1 = READ_ONCE(*p); \
arch/x86/include/asm/clocksource.h
13
return READ_ONCE(vclocks_used) & (1U << vclock);
arch/x86/include/asm/clocksource.h
18
WRITE_ONCE(vclocks_used, READ_ONCE(vclocks_used) | (1 << which));
arch/x86/include/asm/mmu_context.h
97
return READ_ONCE(mm->context.lam_cr3_mask);
arch/x86/include/asm/pgtable-3level.h
138
old.pmd_high = READ_ONCE(pmdp->pmd_high);
arch/x86/include/asm/pgtable.h
1289
old_pte = READ_ONCE(*ptep);
arch/x86/include/asm/pgtable.h
1349
old_pmd = READ_ONCE(*pmdp);
arch/x86/include/asm/pkru.h
13
#define pkru_get_init_value() READ_ONCE(init_pkru_value)
arch/x86/include/asm/posted_intr.h
76
pir_vals[i] = READ_ONCE(pir[i]);
arch/x86/include/asm/resctrl.h
103
u32 closid = READ_ONCE(state->default_closid);
arch/x86/include/asm/resctrl.h
104
u32 rmid = READ_ONCE(state->default_rmid);
arch/x86/include/asm/resctrl.h
112
tmp = READ_ONCE(tsk->closid);
arch/x86/include/asm/resctrl.h
118
tmp = READ_ONCE(tsk->rmid);
arch/x86/include/asm/resctrl.h
155
return READ_ONCE(tsk->closid) == closid;
arch/x86/include/asm/resctrl.h
161
return READ_ONCE(tsk->rmid) == rmid;
arch/x86/include/asm/tlbflush.h
293
return mm && READ_ONCE(mm->context.asid_transition);
arch/x86/include/asm/unwind.h
139
val = READ_ONCE(x); \
arch/x86/kernel/acpi/boot.c
1802
old = READ_ONCE(*lock);
arch/x86/kernel/acpi/boot.c
1818
old = READ_ONCE(*lock);
arch/x86/kernel/acpi/madt_wakeup.c
180
while (READ_ONCE(acpi_mp_wake_mailbox->command))
arch/x86/kernel/acpi/madt_wakeup.c
55
while (READ_ONCE(acpi_mp_wake_mailbox->command) && --timeout)
arch/x86/kernel/apic/apic.c
841
jif_start = READ_ONCE(jiffies);
arch/x86/kernel/apic/apic.c
864
unsigned long jif_now = READ_ONCE(jiffies);
arch/x86/kernel/cpu/aperfmperf.c
431
return READ_ONCE(per_cpu_ptr(arch_cpu_scale, cpu)->capacity);
arch/x86/kernel/cpu/aperfmperf.c
448
freq_ratio = READ_ONCE(this_cpu_ptr(arch_cpu_scale)->freq_ratio);
arch/x86/kernel/cpu/bus_lock.c
246
unsigned int saved_sld_mitigate = READ_ONCE(sysctl_sld_mitigate);
arch/x86/kernel/cpu/mce/dev-mcelog.c
247
if (READ_ONCE(mcelog->next))
arch/x86/kernel/cpu/resctrl/pseudo_lock.c
255
mem_r = READ_ONCE(plr->kmem);
arch/x86/kernel/cpu/resctrl/pseudo_lock.c
359
line_size = READ_ONCE(plr->line_size);
arch/x86/kernel/cpu/resctrl/pseudo_lock.c
360
mem_r = READ_ONCE(plr->kmem);
arch/x86/kernel/cpu/resctrl/pseudo_lock.c
361
size = READ_ONCE(plr->size);
arch/x86/kernel/cpu/umwait.c
133
u32 ctrl = READ_ONCE(umwait_control_cached);
arch/x86/kernel/cpu/umwait.c
152
ctrl = READ_ONCE(umwait_control_cached);
arch/x86/kernel/cpu/umwait.c
165
u32 ctrl = READ_ONCE(umwait_control_cached);
arch/x86/kernel/cpu/umwait.c
187
ctrl = READ_ONCE(umwait_control_cached);
arch/x86/kernel/cpu/umwait.c
36
wrmsrq(MSR_IA32_UMWAIT_CONTROL, READ_ONCE(umwait_control_cached));
arch/x86/kernel/cpu/vmware.c
234
clock = READ_ONCE(steal->clock);
arch/x86/kernel/cpu/vmware.c
239
initial_high = READ_ONCE(steal->clock_high);
arch/x86/kernel/cpu/vmware.c
242
low = READ_ONCE(steal->clock_low);
arch/x86/kernel/cpu/vmware.c
245
high = READ_ONCE(steal->clock_high);
arch/x86/kernel/espfix_64.c
148
stack_page = READ_ONCE(espfix_pages[page]);
arch/x86/kernel/espfix_64.c
155
stack_page = READ_ONCE(espfix_pages[page]);
arch/x86/kernel/fpu/xstate.c
1896
timestamp = READ_ONCE(x86_task_fpu(task)->avx512_timestamp);
arch/x86/kernel/fpu/xstate.h
31
return READ_ONCE(perm->__state_perm);
arch/x86/kernel/hpet.c
804
old.lockval = READ_ONCE(hpet.lockval);
arch/x86/kernel/hpet.c
837
new.lockval = READ_ONCE(hpet.lockval);
arch/x86/kernel/kvm.c
1094
if (READ_ONCE(*ptr) == val)
arch/x86/kernel/kvm.c
1100
if (READ_ONCE(*ptr) == val)
arch/x86/kernel/kvm.c
685
state = READ_ONCE(src->preempted);
arch/x86/kernel/ldt.c
47
ldt = READ_ONCE(mm->context.ldt);
arch/x86/kernel/nmi.c
647
nsp->idt_seq_snap = READ_ONCE(nsp->idt_seq);
arch/x86/kernel/nmi.c
648
nsp->idt_nmi_seq_snap = READ_ONCE(nsp->idt_nmi_seq);
arch/x86/kernel/nmi.c
649
nsp->idt_ignored_snap = READ_ONCE(nsp->idt_ignored);
arch/x86/kernel/nmi.c
669
nmi_seq = READ_ONCE(nsp->idt_nmi_seq);
arch/x86/kernel/nmi.c
678
if (nsp->idt_ignored_snap != READ_ONCE(nsp->idt_ignored) && (idx & 0x1))
arch/x86/kernel/nmi.c
691
__func__, j - READ_ONCE(nsp->recv_jiffies));
arch/x86/kernel/smpboot.c
1429
if (READ_ONCE(md->control) == CPUDEAD_MWAIT_KEXEC_HLT) {
arch/x86/kernel/smpboot.c
1463
if (READ_ONCE(md->status) != CPUDEAD_MWAIT_WAIT)
arch/x86/kernel/smpboot.c
1467
for (i = 0; READ_ONCE(md->status) != newstate && i < 1000; i++) {
arch/x86/kernel/smpboot.c
1473
if (READ_ONCE(md->status) != newstate)
arch/x86/kvm/cpuid.c
1350
if (!READ_ONCE(max_cpuid_80000000))
arch/x86/kvm/cpuid.c
1352
if (function > READ_ONCE(max_cpuid_80000000))
arch/x86/kvm/lapic.c
681
irr_val = READ_ONCE(*p_irr);
arch/x86/kvm/mmu/mmu.c
1442
if (READ_ONCE(eager_page_split))
arch/x86/kvm/mmu/mmu.c
1761
return !tdp_mmu_enabled || READ_ONCE(kvm->arch.indirect_shadow_pages);
arch/x86/kvm/mmu/mmu.c
2885
if (!READ_ONCE(kvm->arch.indirect_shadow_pages))
arch/x86/kvm/mmu/mmu.c
2978
if (READ_ONCE(sp->unsync))
arch/x86/kvm/mmu/mmu.c
3266
pgd = READ_ONCE(*pgd_offset(kvm->mm, hva));
arch/x86/kvm/mmu/mmu.c
3270
p4d = READ_ONCE(*p4d_offset(&pgd, hva));
arch/x86/kvm/mmu/mmu.c
3274
pud = READ_ONCE(*pud_offset(&p4d, hva));
arch/x86/kvm/mmu/mmu.c
3283
pmd = READ_ONCE(*pmd_offset(&pud, hva));
arch/x86/kvm/mmu/mmu.c
357
return READ_ONCE(*sptep);
arch/x86/kvm/mmu/mmu.c
7060
u64 huge_spte = READ_ONCE(*huge_sptep);
arch/x86/kvm/mmu/mmu.c
7485
struct vhost_task *nx_thread = READ_ONCE(kvm->arch.nx_huge_page_recovery_thread);
arch/x86/kvm/mmu/mmu.c
7647
bool enabled = READ_ONCE(nx_huge_pages);
arch/x86/kvm/mmu/mmu.c
7648
uint ratio = READ_ONCE(nx_huge_pages_recovery_ratio);
arch/x86/kvm/mmu/mmu.c
7653
*period = READ_ONCE(nx_huge_pages_recovery_period_ms);
arch/x86/kvm/mmu/mmu.c
7697
unsigned long pages = READ_ONCE(kvm->arch.possible_nx_huge_pages[mmu_type].nr_pages);
arch/x86/kvm/mmu/mmu.c
7698
unsigned int ratio = READ_ONCE(nx_huge_pages_recovery_ratio);
arch/x86/kvm/mmu/mmu_internal.h
223
return READ_ONCE(nx_huge_pages) && !kvm->arch.disable_nx_huge_pages;
arch/x86/kvm/mmu/page_track.c
148
return !!READ_ONCE(slot->arch.gfn_write_track[index]);
arch/x86/kvm/mmu/spte.h
298
return READ_ONCE(root->has_mapped_host_mmio);
arch/x86/kvm/mmu/spte.h
300
return READ_ONCE(vcpu->kvm->arch.has_mapped_host_mmio);
arch/x86/kvm/mmu/tdp_iter.h
19
return READ_ONCE(*rcu_dereference(sptep));
arch/x86/kvm/svm/avic.c
449
int cpu = READ_ONCE(vcpu->cpu);
arch/x86/kvm/svm/avic.c
704
new_entry = READ_ONCE(*entry);
arch/x86/kvm/svm/sev.c
4166
switch (READ_ONCE(vcpu->run->snp_req_certs.ret)) {
arch/x86/kvm/svm/svm.c
3752
if (!READ_ONCE(vcpu->arch.apic->apicv_active)) {
arch/x86/kvm/svm/svm.h
949
return READ_ONCE(svm->sev_es.ghcb->save.field); \
arch/x86/kvm/trace.h
412
__entry->requests = READ_ONCE(vcpu->requests); \
arch/x86/kvm/vmx/posted_intr.c
113
old.control = READ_ONCE(pi_desc->control);
arch/x86/kvm/vmx/posted_intr.c
155
READ_ONCE(kvm->arch.nr_possible_bypass_irqs);
arch/x86/kvm/vmx/posted_intr.c
190
old.control = READ_ONCE(pi_desc->control);
arch/x86/kvm/vmx/tdx.c
1085
if (unlikely(READ_ONCE(to_kvm_tdx(vcpu->kvm)->wait_for_sept_zap)))
arch/x86/kvm/x86.c
10295
return (READ_ONCE(kvm->arch.apicv_inhibit_reasons) == 0);
arch/x86/kvm/x86.c
10301
ulong vm_reasons = READ_ONCE(vcpu->kvm->arch.apicv_inhibit_reasons);
arch/x86/kvm/x86.c
10355
if (!target || !READ_ONCE(target->ready))
arch/x86/kvm/x86.c
11787
if (READ_ONCE(vcpu->arch.pv.pv_unhalted))
arch/x86/kvm/x86.c
13651
if (READ_ONCE(eager_page_split))
arch/x86/kvm/x86.c
14008
if (!READ_ONCE(vcpu->arch.apf.pageready_pending))
arch/x86/kvm/x86.c
2273
return READ_ONCE(vcpu->mode) == EXITING_GUEST_MODE ||
arch/x86/kvm/x86.c
7550
r = READ_ONCE(kvm->arch.default_tsc_khz);
arch/x86/kvm/x86.c
8029
int fep_flags = READ_ONCE(force_emulation_prefix);
arch/x86/kvm/xen.c
1802
vcpu_idx = READ_ONCE(xe->vcpu_idx);
arch/x86/kvm/xen.c
641
unsigned long evtchn_pending_sel = READ_ONCE(v->arch.xen.evtchn_pending_sel);
arch/x86/mm/pat/memtype.c
165
old_flags = READ_ONCE(pg->flags.f);
arch/x86/mm/tlb.c
1337
if (time_after(jiffies, READ_ONCE(mm->context.next_trim_cpumask))) {
arch/x86/mm/tlb.c
478
while (READ_ONCE(per_cpu(cpu_tlbstate.loaded_mm, cpu)) == LOADED_MM_SWITCHING)
arch/x86/mm/tlb.c
481
if (READ_ONCE(per_cpu(cpu_tlbstate.loaded_mm, cpu)) != mm)
arch/x86/mm/tlb.c
492
if (READ_ONCE(per_cpu(cpu_tlbstate.loaded_mm_asid, cpu)) != bc_asid) {
arch/x86/net/bpf_jit_comp.c
883
WARN_ON_ONCE(READ_ONCE(poke->tailcall_target_stable));
arch/x86/platform/efi/efi_64.c
835
efi.runtime = READ_ONCE(systab->runtime);
arch/x86/platform/efi/quirks.c
781
if (READ_ONCE(efi_rts_work.efi_rts_id) == EFI_NONE ||
arch/x86/xen/p2m.c
545
mid_mfn = READ_ONCE(p2m_top_mfn_p[topidx]);
arch/x86/xen/p2m.c
574
p2m_pfn = pte_pfn(READ_ONCE(*ptep));
arch/x86/xen/spinlock.c
50
} else if (READ_ONCE(*byte) == val) {
arch/xtensa/include/asm/atomic.h
46
#define arch_atomic_read(v) READ_ONCE((v)->counter)
arch/xtensa/include/asm/cmpxchg.h
191
oldv = READ_ONCE(*p);
arch/xtensa/kernel/smp.c
221
ccount = READ_ONCE(cpu_start_ccount);
arch/xtensa/kernel/smp.c
318
if (READ_ONCE(cpu_start_id) == -cpu) {
arch/xtensa/platforms/iss/simdisk.c
88
READ_ONCE(*buffer);
block/bfq-iosched.c
5153
READ_ONCE(bfqd->queued);
block/bio.c
495
if (READ_ONCE(cache->nr_irq) >= ALLOC_CACHE_THRESHOLD)
block/bio.c
779
if (!READ_ONCE(cache->free_list)) {
block/bio.c
821
if (READ_ONCE(cache->nr_irq) + cache->nr > ALLOC_CACHE_MAX)
block/blk-cgroup.c
2220
if (!READ_ONCE(bis->lqueued)) {
block/blk-core.c
1015
bio = READ_ONCE(kiocb->private);
block/blk-core.c
1028
stamp = READ_ONCE(part->bd_stamp);
block/blk-core.c
1068
unsigned long now = READ_ONCE(jiffies);
block/blk-core.c
944
blk_qc_t cookie = READ_ONCE(bio->bi_cookie);
block/blk-core.c
949
bdev = READ_ONCE(bio->bi_bdev);
block/blk-iocost.c
1201
u64 active_sum = READ_ONCE(parent->child_active_sum);
block/blk-iocost.c
1202
u64 inuse_sum = READ_ONCE(parent->child_inuse_sum);
block/blk-iocost.c
1203
u32 active = READ_ONCE(child->active);
block/blk-iocost.c
1204
u32 inuse = READ_ONCE(child->inuse);
block/blk-iocost.c
1586
bool pay_debt = READ_ONCE(iocg->abs_vdebt);
block/blk-iocost.c
2652
ioc_locked = use_debt || READ_ONCE(iocg->abs_vdebt);
block/blk-iolatency.c
414
scale_lat = READ_ONCE(lat_info->scale_lat);
block/blk-mq-debugfs.c
395
READ_ONCE(tags->active_queues));
block/blk-mq-sched.c
217
struct blk_mq_ctx *ctx = READ_ONCE(hctx->dispatch_from);
block/blk-mq.c
1687
deadline = READ_ONCE(rq->deadline);
block/blk-mq.h
427
users = READ_ONCE(hctx->tags->active_queues);
block/blk-sysfs.c
507
jiffies_to_msecs(READ_ONCE(disk->queue->rq_timeout)));
block/blk-wbt.c
282
u64 issue = READ_ONCE(rwb->sync_issue);
block/bsg.c
89
return put_user(READ_ONCE(bd->max_queue), uarg);
block/fops.c
272
if (!READ_ONCE(dio->waiter))
block/ioctl.c
967
start = READ_ONCE(sqe->addr);
block/ioctl.c
968
len = READ_ONCE(sqe->addr3);
block/kyber-iosched.c
684
struct blk_mq_hw_ctx *hctx = READ_ONCE(wqe->private);
drivers/accel/amdxdna/amdxdna_mailbox.c
354
if (READ_ONCE(mb_chann->bad_state)) {
drivers/accel/amdxdna/amdxdna_mailbox.c
414
if (READ_ONCE(mb_chann->bad_state)) {
drivers/accel/ivpu/ivpu_fw_log.c
148
log->read_index = READ_ONCE(log->write_index);
drivers/accel/ivpu/ivpu_fw_log.c
149
log->read_wrap_count = READ_ONCE(log->wrap_count);
drivers/accel/ivpu/ivpu_fw_log.c
154
log->read_index = READ_ONCE(log->write_index);
drivers/accel/ivpu/ivpu_fw_log.c
155
log->read_wrap_count = READ_ONCE(log->wrap_count);
drivers/accel/ivpu/ivpu_fw_log.c
98
u32 log_start = only_new_msgs ? READ_ONCE(log->read_index) : 0;
drivers/accel/ivpu/ivpu_fw_log.c
99
u32 log_end = READ_ONCE(log->write_index);
drivers/accel/ivpu/ivpu_job.c
428
u32 tail = READ_ONCE(header->tail);
drivers/accel/ivpu/ivpu_mmu.c
932
val = READ_ONCE(entry[0]);
drivers/accel/ivpu/ivpu_mmu.c
966
if (!READ_ONCE(file_priv->has_mmu_faults)) {
drivers/android/binder.c
6942
int debug_id = READ_ONCE(e->debug_id_done);
drivers/android/binder.c
6961
seq_printf(m, debug_id && debug_id == READ_ONCE(e->debug_id_done) ?
drivers/base/core.c
2406
drv = READ_ONCE(dev->driver);
drivers/base/core.c
2645
struct device_driver *drv = READ_ONCE(dev->driver);
drivers/base/platform.c
1453
struct device_driver *drv = READ_ONCE(dev->driver);
drivers/base/power/main.c
1016
if (READ_ONCE(async_error))
drivers/base/power/main.c
1195
if (READ_ONCE(async_error))
drivers/base/power/main.c
1362
if (READ_ONCE(link->status) != DL_STATE_DORMANT)
drivers/base/power/main.c
1443
if (READ_ONCE(async_error))
drivers/base/power/main.c
1502
if (error || READ_ONCE(async_error))
drivers/base/power/main.c
1557
if (READ_ONCE(async_error)) {
drivers/base/power/main.c
1572
error = READ_ONCE(async_error);
drivers/base/power/main.c
1638
if (READ_ONCE(async_error))
drivers/base/power/main.c
1701
if (error || READ_ONCE(async_error))
drivers/base/power/main.c
1762
if (READ_ONCE(async_error)) {
drivers/base/power/main.c
1777
error = READ_ONCE(async_error);
drivers/base/power/main.c
1878
if (READ_ONCE(async_error)) {
drivers/base/power/main.c
1990
if (error || READ_ONCE(async_error))
drivers/base/power/main.c
2053
if (READ_ONCE(async_error)) {
drivers/base/power/main.c
2068
error = READ_ONCE(async_error);
drivers/base/power/main.c
284
if (READ_ONCE(link->status) != DL_STATE_DORMANT &&
drivers/base/power/main.c
342
if (READ_ONCE(link->status) != DL_STATE_DORMANT &&
drivers/base/power/main.c
692
if (READ_ONCE(link->status) != DL_STATE_DORMANT)
drivers/base/power/main.c
866
if (READ_ONCE(async_error))
drivers/base/power/runtime.c
185
autosuspend_delay = READ_ONCE(dev->power.autosuspend_delay);
drivers/base/power/runtime.c
189
expires = READ_ONCE(dev->power.last_busy);
drivers/base/power/runtime.c
628
u64 slack = (u64)READ_ONCE(dev->power.autosuspend_delay) *
drivers/block/loop.c
1860
if (data_race(READ_ONCE(lo->lo_state)) != Lo_bound)
drivers/block/loop.c
2204
data_race(READ_ONCE(lo->lo_state)) == Lo_unbound)
drivers/block/nbd.c
814
unsigned long deadline = READ_ONCE(req->deadline);
drivers/block/nbd.c
830
if (READ_ONCE(jiffies) + msecs_to_jiffies(wait_ms) >= deadline) {
drivers/block/ublk_drv.c
1901
if (READ_ONCE(ubq->active_fcmd)) {
drivers/block/ublk_drv.c
2059
if (unlikely(READ_ONCE(ubq->fail_io)))
drivers/block/ublk_drv.c
2072
unlikely(READ_ONCE(ubq->force_abort)))
drivers/block/ublk_drv.c
2714
done = (READ_ONCE(ubq->active_fcmd) != fcmd) && !list_empty(&fcmd->node);
drivers/block/ublk_drv.c
2733
fcmd = READ_ONCE(ubq->active_fcmd);
drivers/block/ublk_drv.c
2980
buf = ublk_sqe_addr_to_auto_buf_reg(READ_ONCE(cmd->sqe->addr));
drivers/block/ublk_drv.c
3265
u16 q_id = READ_ONCE(ub_src->q_id);
drivers/block/ublk_drv.c
3266
u16 tag = READ_ONCE(ub_src->tag);
drivers/block/ublk_drv.c
3267
s32 result = READ_ONCE(ub_src->result);
drivers/block/ublk_drv.c
3268
u64 addr = READ_ONCE(ub_src->addr); /* unioned with zone_append_lba */
drivers/block/ublk_drv.c
3312
if (READ_ONCE(io->task) != current) {
drivers/block/ublk_drv.c
3617
.uaddr = u64_to_user_ptr(READ_ONCE(cmd->sqe->addr)),
drivers/block/ublk_drv.c
3698
.uaddr = u64_to_user_ptr(READ_ONCE(cmd->sqe->addr)),
drivers/block/ublk_drv.c
3840
unsigned tag = READ_ONCE(ub_cmd->tag);
drivers/block/ublk_drv.c
3841
unsigned q_id = READ_ONCE(ub_cmd->q_id);
drivers/block/ublk_drv.c
3842
unsigned index = READ_ONCE(ub_cmd->addr);
drivers/block/ublk_drv.c
3874
.q_id = READ_ONCE(uc->q_id),
drivers/block/ublk_drv.c
3875
.flags = READ_ONCE(uc->flags),
drivers/block/ublk_drv.c
3876
.nr_elem = READ_ONCE(uc->nr_elem),
drivers/block/ublk_drv.c
3877
.elem_bytes = READ_ONCE(uc->elem_bytes),
drivers/block/ublk_drv.c
3972
on_daemon = current == READ_ONCE(io->task);
drivers/block/ublk_drv.c
5282
header.dev_id = READ_ONCE(ub_src->dev_id);
drivers/block/ublk_drv.c
5283
header.queue_id = READ_ONCE(ub_src->queue_id);
drivers/block/ublk_drv.c
5284
header.len = READ_ONCE(ub_src->len);
drivers/block/ublk_drv.c
5285
header.addr = READ_ONCE(ub_src->addr);
drivers/block/ublk_drv.c
5286
header.data[0] = READ_ONCE(ub_src->data[0]);
drivers/block/ublk_drv.c
5287
header.dev_path_len = READ_ONCE(ub_src->dev_path_len);
drivers/block/ublk_drv.c
717
fcmd->buf_group = READ_ONCE(cmd->sqe->buf_index);
drivers/block/xen-blkback/blkback.c
1079
dst->operation = READ_ONCE(src->operation);
drivers/block/xen-blkback/blkback.c
1086
dst->u.rw.nr_segments = READ_ONCE(src->u.rw.nr_segments);
drivers/block/xen-blkback/blkback.c
1106
READ_ONCE(src->u.indirect.nr_segments);
drivers/block/xen-blkback/blkback.c
1132
dst->operation = READ_ONCE(src->operation);
drivers/block/xen-blkback/blkback.c
1139
dst->u.rw.nr_segments = READ_ONCE(src->u.rw.nr_segments);
drivers/block/xen-blkback/blkback.c
1159
READ_ONCE(src->u.indirect.nr_segments);
drivers/block/xen-blkback/blkback.c
946
first_sect = READ_ONCE(segments[i].first_sect);
drivers/block/xen-blkback/blkback.c
947
last_sect = READ_ONCE(segments[i].last_sect);
drivers/block/xen-blkback/xenbus.c
210
rsp_prod = READ_ONCE(sring_common->rsp_prod);
drivers/block/xen-blkback/xenbus.c
211
req_prod = READ_ONCE(sring_common->req_prod);
drivers/block/xen-blkfront.c
1530
rp = READ_ONCE(rinfo->ring.sring->rsp_prod);
drivers/block/zloop.c
711
if (data_race(READ_ONCE(zlo->state)) == Zlo_deleting)
drivers/bluetooth/hci_bcm4377.c
823
while (tail != le16_to_cpu(READ_ONCE(heads[ring->ring_id]))) {
drivers/bus/fsl-mc/fsl-mc-bus.c
171
const struct device_driver *drv = READ_ONCE(dev->driver);
drivers/bus/mhi/ep/main.c
70
if (READ_ONCE(ring->irq_pending))
drivers/bus/mhi/ep/main.c
74
} else if (ring->intmodt && !READ_ONCE(ring->irq_pending)) {
drivers/char/ipmi/ipmi_msghandler.c
1892
int run_to_completion = READ_ONCE(intf->run_to_completion);
drivers/char/ipmi/ipmi_msghandler.c
2020
chans = READ_ONCE(intf->channel_list)->c;
drivers/char/ipmi/ipmi_msghandler.c
2208
chans = READ_ONCE(intf->channel_list)->c;
drivers/char/ipmi/ipmi_msghandler.c
2324
int run_to_completion = READ_ONCE(intf->run_to_completion);
drivers/char/ipmi/ipmi_msghandler.c
4650
chans = READ_ONCE(intf->channel_list)->c;
drivers/char/ipmi/ipmi_msghandler.c
4743
chans = READ_ONCE(intf->channel_list)->c;
drivers/char/ipmi/ipmi_msghandler.c
4819
int run_to_completion = READ_ONCE(intf->run_to_completion);
drivers/char/ipmi/ipmi_msghandler.c
4859
int run_to_completion = READ_ONCE(intf->run_to_completion);
drivers/char/ipmi/ipmi_msghandler.c
4955
int run_to_completion = READ_ONCE(intf->run_to_completion);
drivers/char/ipmi/ipmi_msghandler.c
795
int intf_num = READ_ONCE(intf->intf_num);
drivers/char/random.c
1169
delta = now - READ_ONCE(state->last_time);
drivers/char/random.c
1172
delta2 = delta - READ_ONCE(state->last_delta);
drivers/char/random.c
1175
delta3 = delta2 - READ_ONCE(state->last_delta2);
drivers/char/random.c
232
if (unlikely(READ_ONCE(early_boot))) {
drivers/char/random.c
370
if (unlikely(crng->generation != READ_ONCE(base_crng.generation))) {
drivers/char/random.c
526
next_gen = READ_ONCE(base_crng.generation); \
drivers/char/random.c
727
orig = READ_ONCE(input_pool.init_bits);
drivers/char/xillybus/xillyusb.c
1533
if (!READ_ONCE(chan->read_data_ok)) { /* FPGA has sent EOF */
drivers/char/xillybus/xillyusb.c
1536
if (READ_ONCE(fifo->fill))
drivers/char/xillybus/xillyusb.c
1595
!READ_ONCE(fifo->fill))
drivers/char/xillybus/xillyusb.c
1880
(READ_ONCE(chan->in_fifo->fill) || !chan->read_data_ok))
drivers/char/xillybus/xillyusb.c
1884
(READ_ONCE(chan->out_ep->fifo.fill) != chan->out_ep->fifo.size))
drivers/char/xillybus/xillyusb.c
2069
idt_len = READ_ONCE(idt_fifo.fill);
drivers/char/xillybus/xillyusb.c
236
nmax = fifo->size - READ_ONCE(fifo->fill);
drivers/char/xillybus/xillyusb.c
780
fill = READ_ONCE(fifo->fill) & ep->fill_mask;
drivers/clocksource/bcm2835_timer.c
60
event_handler = READ_ONCE(timer->evt.event_handler);
drivers/comedi/comedi_fops.c
316
return READ_ONCE(cfp->read_subdev);
drivers/comedi/comedi_fops.c
324
return READ_ONCE(cfp->write_subdev);
drivers/cpufreq/amd-pstate-ut.c
154
cur_perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
1019
perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
1032
policy->boost_supported = READ_ONCE(cpudata->boost_supported);
drivers/cpufreq/amd-pstate.c
1072
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
1097
perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
1110
perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
1136
perf = READ_ONCE(cpudata->prefcore_ranking);
drivers/cpufreq/amd-pstate.c
1147
hw_prefcore = READ_ONCE(cpudata->hw_prefcore);
drivers/cpufreq/amd-pstate.c
1501
perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
1519
policy->boost_supported = READ_ONCE(cpudata->boost_supported);
drivers/cpufreq/amd-pstate.c
1553
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
1581
perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
1618
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
1635
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
1660
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
238
value = prev = READ_ONCE(cpudata->cppc_req_cached);
drivers/cpufreq/amd-pstate.c
248
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
294
value = prev = READ_ONCE(cpudata->cppc_req_cached);
drivers/cpufreq/amd-pstate.c
366
value = READ_ONCE(cpudata->cppc_req_cached);
drivers/cpufreq/amd-pstate.c
393
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
437
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
498
value = prev = READ_ONCE(cpudata->cppc_req_cached);
drivers/cpufreq/amd-pstate.c
508
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
572
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
616
perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
633
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
675
perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
734
perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
760
union perf_cached perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/amd-pstate.c
764
nominal_freq = READ_ONCE(cpudata->nominal_freq);
drivers/cpufreq/amd-pstate.c
854
sched_set_itmt_core_prio((int)READ_ONCE(cpudata->prefcore_ranking), cpudata->cpu);
drivers/cpufreq/amd-pstate.c
872
prev_high = READ_ONCE(cpudata->prefcore_ranking);
drivers/cpufreq/amd-pstate.c
936
perf = READ_ONCE(cpudata->perf);
drivers/cpufreq/cpufreq.c
533
unsigned int min = READ_ONCE(policy->min);
drivers/cpufreq/cpufreq.c
534
unsigned int max = READ_ONCE(policy->max);
drivers/cpufreq/cpufreq_governor.c
294
lst = READ_ONCE(policy_dbs->last_sample_time);
drivers/cpufreq/cpufreq_governor.c
312
if (unlikely(lst != READ_ONCE(policy_dbs->last_sample_time))) {
drivers/cpufreq/cpufreq_stats.c
127
bool pending = READ_ONCE(stats->reset_pending);
drivers/cpufreq/cpufreq_stats.c
274
if (unlikely(READ_ONCE(stats->reset_pending)))
drivers/cpufreq/cpufreq_stats.c
55
cpufreq_stats_update(stats, READ_ONCE(stats->reset_time));
drivers/cpufreq/cpufreq_stats.c
62
if (READ_ONCE(stats->reset_pending))
drivers/cpufreq/cpufreq_stats.c
72
bool pending = READ_ONCE(stats->reset_pending);
drivers/cpufreq/cpufreq_stats.c
85
time = local_clock() - READ_ONCE(stats->reset_time);
drivers/cpufreq/intel_pstate.c
1061
if (READ_ONCE(global.no_turbo)) {
drivers/cpufreq/intel_pstate.c
1066
cpu->capacity_perf = HWP_HIGHEST_PERF(READ_ONCE(cpu->hwp_cap_cached));
drivers/cpufreq/intel_pstate.c
1325
u64 value = READ_ONCE(cpu->hwp_req_cached);
drivers/cpufreq/intel_pstate.c
1355
min_perf = HWP_LOWEST_PERF(READ_ONCE(cpu->hwp_cap_cached));
drivers/cpufreq/intel_pstate.c
1414
wrmsrq_on_cpu(cpu->cpu, MSR_HWP_REQUEST, READ_ONCE(cpu->hwp_req_cached));
drivers/cpufreq/intel_pstate.c
1473
policy->cpuinfo.max_freq = READ_ONCE(global.no_turbo) ?
drivers/cpufreq/intel_pstate.c
2058
if (READ_ONCE(global.no_turbo) && !READ_ONCE(global.turbo_disabled) &&
drivers/cpufreq/intel_pstate.c
2371
u64 hwp_req = READ_ONCE(cpu->hwp_req_cached);
drivers/cpufreq/intel_pstate.c
2372
u64 hwp_cap = READ_ONCE(cpu->hwp_cap_cached);
drivers/cpufreq/intel_pstate.c
2546
target = READ_ONCE(global.no_turbo) ?
drivers/cpufreq/intel_pstate.c
2818
return READ_ONCE(global.no_turbo) ?
drivers/cpufreq/intel_pstate.c
2955
max_freq = READ_ONCE(global.no_turbo) ?
drivers/cpufreq/intel_pstate.c
3049
policy->cpuinfo.max_freq = READ_ONCE(global.no_turbo) ?
drivers/cpufreq/intel_pstate.c
3149
u64 prev = READ_ONCE(cpu->hwp_req_cached), value = prev;
drivers/cpufreq/intel_pstate.c
3248
u64 hwp_cap = READ_ONCE(cpu->hwp_cap_cached);
drivers/cpufreq/intel_pstate.c
3252
cap_pstate = READ_ONCE(global.no_turbo) ?
drivers/cpufreq/intel_pstate.c
3380
u64 value = READ_ONCE(cpu->hwp_req_cached);
drivers/cpufreq/intel_pstate.c
377
cppc_perf.highest_perf = HWP_HIGHEST_PERF(READ_ONCE(all_cpu_data[cpu]->hwp_cap_cached));
drivers/cpufreq/intel_pstate.c
729
u64 value = READ_ONCE(cpu->hwp_req_cached);
drivers/crypto/amcc/crypto4xx_core.c
1041
((READ_ONCE(pd->pd_ctl.w) &
drivers/crypto/caam/jr.c
273
head = READ_ONCE(jrp->head);
drivers/crypto/caam/jr.c
442
tail = READ_ONCE(jrp->tail);
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
515
u64 orh = READ_ONCE(*sr->resp.orh);
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
521
while (READ_ONCE(*sr->resp.completion) == PENDING_SIG) {
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
563
READ_ONCE(*sr->resp.orh));
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
572
err = READ_ONCE(*sr->resp.orh) & 0xff;
drivers/crypto/inside-secure/eip93/eip93-main.c
220
pe_ctrl_stat = READ_ONCE(rdesc->pe_ctrl_stat_word);
drivers/crypto/inside-secure/eip93/eip93-main.c
221
pe_length = READ_ONCE(rdesc->pe_length_word);
drivers/crypto/nx/nx-common-powernv.c
191
while (!(READ_ONCE(csb->flags) & CSB_V)) {
drivers/cxl/core/port.c
203
int part = READ_ONCE(cxled->part);
drivers/cxl/core/region.c
3902
int rc, part = READ_ONCE(cxled->part);
drivers/devfreq/tegra30-devfreq.c
657
cur_freq = READ_ONCE(tegra->cur_freq);
drivers/dma-buf/st-dma-fence.c
463
if (!READ_ONCE(cb.seen)) {
drivers/dma/mediatek/mtk-hsdma.c
567
desc2 = READ_ONCE(rxd->desc2);
drivers/dma/plx_dma.c
145
flags = le32_to_cpu(READ_ONCE(desc->hw->flags_and_size));
drivers/dma/xilinx/xilinx_dpdma.c
374
testcase = READ_ONCE(dpdma_debugfs.testcase);
drivers/edac/altera_edac.c
174
reg = READ_ONCE(ptemp[0]);
drivers/edac/altera_edac.c
175
read_reg = READ_ONCE(ptemp[1]);
drivers/edac/altera_edac.c
616
if (READ_ONCE(ptemp[i]))
drivers/edac/altera_edac.c
633
if (READ_ONCE(ptemp[i]) != i)
drivers/firewire/core-card.c
330
irm_device_quirks = READ_ONCE(irm_device->quirks);
drivers/firewire/ohci.c
1364
if (last->transfer_status == 0 && !READ_ONCE(ctx->flushing))
drivers/firewire/ohci.c
1396
if (READ_ONCE(ctx->flushing))
drivers/firewire/ohci.c
1418
if (READ_ONCE(ctx->flushing)) {
drivers/firewire/ohci.c
1437
int quirks = READ_ONCE(device->quirks);
drivers/firewire/ohci.c
2826
res_count = le16_to_cpu(READ_ONCE(last->res_count));
drivers/firewire/ohci.c
596
res_count = READ_ONCE(ctx->descriptors[i].res_count);
drivers/firewire/ohci.c
604
next_res_count = READ_ONCE(ctx->descriptors[next_i].res_count);
drivers/firewire/ohci.c
620
next_res_count = READ_ONCE(ctx->descriptors[next_i].res_count);
drivers/firmware/arm_scmi/notify.c
153
__pd = READ_ONCE(ni_->registered_protocols[(__pid)]); \
drivers/firmware/arm_scmi/notify.c
164
__revt = READ_ONCE(pd_->registered_events[eid_]); \
drivers/firmware/qemu_fw_cfg.c
119
if (be32_to_cpu(READ_ONCE(d->control)) & FW_CFG_DMA_CTL_ERROR) {
drivers/firmware/qemu_fw_cfg.c
80
u32 ctrl = be32_to_cpu(READ_ONCE(d->control));
drivers/gpio/gpio-virtuser.c
575
*val = READ_ONCE(ld->debounce);
drivers/gpio/gpiolib-cdev.c
1028
if (!eflags || READ_ONCE(line->sw_debounced))
drivers/gpio/gpiolib-cdev.c
1067
u64 active_edflags = READ_ONCE(line->edflags);
drivers/gpio/gpiolib-cdev.c
1072
(READ_ONCE(line->desc->debounce_period_us) == debounce_period_us))
drivers/gpio/gpiolib-cdev.c
1076
if (debounce_period_us && READ_ONCE(line->sw_debounced)) {
drivers/gpio/gpiolib-cdev.c
1089
if ((line->irq && !READ_ONCE(line->sw_debounced)) ||
drivers/gpio/gpiolib-cdev.c
1091
(!debounce_period_us && READ_ONCE(line->sw_debounced)))
drivers/gpio/gpiolib-cdev.c
1214
unsigned long flags = READ_ONCE(*flagsp);
drivers/gpio/gpiolib-cdev.c
145
unsigned long flags = READ_ONCE(*flagsp);
drivers/gpio/gpiolib-cdev.c
2228
dflags = READ_ONCE(desc->flags);
drivers/gpio/gpiolib-cdev.c
2290
debounce_period_us = READ_ONCE(desc->debounce_period_us);
drivers/gpio/gpiolib-cdev.c
631
edflags = READ_ONCE(line->edflags);
drivers/gpio/gpiolib-cdev.c
676
if (READ_ONCE(line->sw_debounced)) {
drivers/gpio/gpiolib-cdev.c
680
usecs_to_jiffies(READ_ONCE(line->desc->debounce_period_us)));
drivers/gpio/gpiolib-cdev.c
756
switch (READ_ONCE(line->edflags) & GPIO_V2_LINE_EDGE_FLAGS) {
drivers/gpio/gpiolib-cdev.c
808
value = READ_ONCE(line->level);
drivers/gpio/gpiolib-cdev.c
821
usecs_to_jiffies(READ_ONCE(line->desc->debounce_period_us)));
drivers/gpio/gpiolib-cdev.c
831
u64 eflags, edflags = READ_ONCE(line->edflags);
drivers/gpio/gpiolib-cdev.c
846
if (READ_ONCE(line->level) == level)
drivers/gpio/gpiolib-cdev.c
983
if (READ_ONCE(line->edflags) & GPIO_V2_LINE_FLAG_EVENT_CLOCK_HTE)
drivers/gpio/gpiolib-of.c
886
if (READ_ONCE(desc->hog) == hog)
drivers/gpio/gpiolib.c
126
flags = READ_ONCE(desc->flags);
drivers/gpio/gpiolib.c
2522
flags = READ_ONCE(desc->flags);
drivers/gpio/gpiolib.c
2746
flags = READ_ONCE(desc->flags);
drivers/gpio/gpiolib.c
3031
flags = READ_ONCE(desc->flags);
drivers/gpio/gpiolib.c
460
flags = READ_ONCE(desc->flags);
drivers/gpio/gpiolib.c
5185
dflags = READ_ONCE(desc->flags);
drivers/gpio/gpiolib.c
5339
flags = READ_ONCE(desc->flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1651
if (!READ_ONCE(pinfo->block_mmu_notifications)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2508
if (READ_ONCE(process_info->block_mmu_notifications))
drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
299
uint64_t seq = READ_ONCE(ring->fence_drv.sync_seq);
drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
359
emitted += READ_ONCE(ring->fence_drv.sync_seq);
drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
377
sync_seq = READ_ONCE(ring->fence_drv.sync_seq);
drivers/gpu/drm/amd/amdgpu/amdgpu_ih.c
163
if (wptr != READ_ONCE(ih->rptr)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
1671
pin_count = READ_ONCE(bo->tbo.pin_count);
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
1675
dma_buf = READ_ONCE(bo->tbo.base.dma_buf);
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
1676
attachment = READ_ONCE(bo->tbo.base.import_attach);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
537
chunk->sync_seq = READ_ONCE(ring->fence_drv.sync_seq);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
163
fence = READ_ONCE(queue->hang_detect_fence);
drivers/gpu/drm/amd/amdgpu/amdgpu_vkms.c
103
*vblank_time = READ_ONCE(amdgpu_crtc->vblank_timer.node.expires);
drivers/gpu/drm/amd/amdgpu/amdgpu_vkms.c
98
if (!READ_ONCE(vblank->enabled)) {
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
2653
!READ_ONCE(adev->barrier_has_auto_waitcnt));
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
678
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
747
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
225
rptr = READ_ONCE(*((u64 *)&ring->adev->wb.wb[ring->rptr_offs]));
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
245
wptr = READ_ONCE(*((u64 *)&adev->wb.wb[ring->wptr_offs]));
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
314
wptr = READ_ONCE(*((u64 *)&adev->wb.wb[ring->wptr_offs]));
drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
352
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
192
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
193
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
192
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
186
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
drivers/gpu/drm/amd/amdkfd/kfd_events.c
769
if (READ_ONCE(slots[id]) != UNSIGNALED_EVENT_SLOT)
drivers/gpu/drm/amd/amdkfd/kfd_events.c
778
if (READ_ONCE(slots[id]) != UNSIGNALED_EVENT_SLOT) {
drivers/gpu/drm/amd/amdkfd/kfd_events.c
850
if (!READ_ONCE(event_waiters[i].event))
drivers/gpu/drm/amd/amdkfd/kfd_events.c
853
if (READ_ONCE(event_waiters[i].activated)) {
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
971
if (READ_ONCE(p->svms.faulting_task) == current) {
drivers/gpu/drm/amd/amdkfd/kfd_process.c
2297
!READ_ONCE(pdd->process->irq_drain_is_open));
drivers/gpu/drm/amd/amdkfd/kfd_process.c
455
return sysfs_emit(buf, "%llu\n", READ_ONCE(pdd->faults));
drivers/gpu/drm/amd/amdkfd/kfd_process.c
460
return sysfs_emit(buf, "%llu\n", READ_ONCE(pdd->page_in));
drivers/gpu/drm/amd/amdkfd/kfd_process.c
465
return sysfs_emit(buf, "%llu\n", READ_ONCE(pdd->page_out));
drivers/gpu/drm/amd/amdkfd/kfd_smi_events.c
166
uint64_t events = READ_ONCE(client->events);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2129
max_pages = READ_ONCE(max_svm_range_pages);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2267
READ_ONCE(max_svm_range_pages),
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2280
READ_ONCE(max_svm_range_pages),
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
496
if (READ_ONCE(prange->svm_bo->evicting)) {
drivers/gpu/drm/amd/display/dmub/inc/dmub_cmd.h
7477
(void)READ_ONCE(*data++);
drivers/gpu/drm/display/drm_dp_helper.c
729
return !aux->is_remote && !READ_ONCE(aux->dpcd_probe_disabled);
drivers/gpu/drm/display/drm_dp_mst_topology.c
1260
state = READ_ONCE(txmsg->state);
drivers/gpu/drm/drm_fb_helper.c
125
if (READ_ONCE(fb_helper->deferred_setup))
drivers/gpu/drm/drm_gem.c
340
if (drm_WARN_ON(dev, READ_ONCE(obj->handle_count) == 0))
drivers/gpu/drm/drm_lease.c
387
bool universal_planes = READ_ONCE(lessor_priv->universal_planes);
drivers/gpu/drm/drm_sysfs.c
236
status = READ_ONCE(connector->status);
drivers/gpu/drm/drm_sysfs.c
249
dpms = READ_ONCE(connector->dpms);
drivers/gpu/drm/drm_sysfs.c
261
enabled = READ_ONCE(connector->encoder);
drivers/gpu/drm/drm_vblank.c
1463
drm_WARN_ON(dev, !READ_ONCE(vblank->inmodeset));
drivers/gpu/drm/drm_vblank.c
1644
if (!READ_ONCE(vblank->enabled)) {
drivers/gpu/drm/drm_vblank.c
1794
READ_ONCE(vblank->enabled)) {
drivers/gpu/drm/drm_vblank.c
1843
!READ_ONCE(vblank->enabled),
drivers/gpu/drm/drm_vblank.c
2030
vblank_enabled = READ_ONCE(vblank->config.disable_immediate) &&
drivers/gpu/drm/drm_vblank.c
2031
READ_ONCE(vblank->enabled);
drivers/gpu/drm/drm_vblank.c
2131
if (!READ_ONCE(vblank->enabled)) {
drivers/gpu/drm/drm_vblank.c
2292
if (!READ_ONCE(vblank->enabled)) {
drivers/gpu/drm/drm_vblank.c
2305
*vblank_time = READ_ONCE(vtimer->timer.node.expires);
drivers/gpu/drm/drm_vblank.c
517
drm_WARN_ON(dev, READ_ONCE(vblank->enabled) &&
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
288
unsigned int new_flush_seq = READ_ONCE(mmu_context->flush_seq);
drivers/gpu/drm/exynos/exynos_drm_vidi.c
325
return READ_ONCE(ctx->connected) ? connector_status_connected :
drivers/gpu/drm/i915/display/intel_display_irq.c
1775
int vblank_enable_count = READ_ONCE(display->irq.vblank_enable_count);
drivers/gpu/drm/i915/display/intel_dp.c
3732
if (READ_ONCE(intel_dp->oui_valid))
drivers/gpu/drm/i915/display/intel_fbc.c
2204
if (READ_ONCE(fbc->underrun_detected))
drivers/gpu/drm/i915/display/intel_hdmi.c
1854
READ_ONCE(to_intel_digital_connector_state(conn_state)->force_audio) != HDMI_AUDIO_OFF_DVI;
drivers/gpu/drm/i915/display/intel_hotplug.c
793
enabled = READ_ONCE(display->hotplug.poll_enabled);
drivers/gpu/drm/i915/display/intel_hotplug.c
804
READ_ONCE(display->hotplug.poll_enabled));
drivers/gpu/drm/i915/display/intel_psr.c
3464
if (READ_ONCE(intel_dp->psr.irq_aux_error)) {
drivers/gpu/drm/i915/display/intel_psr.c
4398
*val = READ_ONCE(intel_dp->psr.debug);
drivers/gpu/drm/i915/display/intel_psr.c
935
READ_ONCE(vblank->enabled);
drivers/gpu/drm/i915/display/intel_sdvo.c
1322
READ_ONCE(to_intel_digital_connector_state(conn_state)->force_audio) != HDMI_AUDIO_OFF_DVI;
drivers/gpu/drm/i915/gem/i915_gem_domain.c
633
if (READ_ONCE(obj->write_domain) == read_domains)
drivers/gpu/drm/i915/gem/i915_gem_lmem.c
44
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
drivers/gpu/drm/i915/gem/i915_gem_object.c
622
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
drivers/gpu/drm/i915/gem/i915_gem_object.h
311
return READ_ONCE(obj->frontbuffer) || obj->is_dpt;
drivers/gpu/drm/i915/gem/i915_gem_object.h
661
return !IS_ERR_OR_NULL(READ_ONCE(obj->mm.pages));
drivers/gpu/drm/i915/gem/i915_gem_pages.c
670
if (n < READ_ONCE(iter->sg_idx))
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
296
count = READ_ONCE(i915->mm.shrink_memory) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
297
num_objects = READ_ONCE(i915->mm.shrink_count);
drivers/gpu/drm/i915/gem/i915_gem_tiling.c
441
READ_ONCE(obj->tiling_and_stride) & TILING_MASK;
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
208
if (!signal && READ_ONCE(b->irq_armed) && list_empty(&b->signalers))
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
269
if (!READ_ONCE(b->irq_armed) && !list_empty(&b->signalers))
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
273
if (READ_ONCE(b->irq_armed) && !atomic_read(&b->active))
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
321
if (!READ_ONCE(b->irq_armed))
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
352
struct intel_breadcrumbs *b = READ_ONCE(rq->engine)->breadcrumbs;
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
408
if (!READ_ONCE(b->irq_armed) || __i915_request_is_complete(rq))
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
439
struct intel_breadcrumbs *b = READ_ONCE(rq->engine)->breadcrumbs;
drivers/gpu/drm/i915/gt/intel_context.c
608
active = READ_ONCE(ce->stats.active);
drivers/gpu/drm/i915/gt/intel_context_types.h
94
__intel_context_inflight(READ_ONCE((ce)->inflight))
drivers/gpu/drm/i915/gt/intel_context_types.h
96
__intel_context_inflight_count(READ_ONCE((ce)->inflight))
drivers/gpu/drm/i915/gt/intel_engine.h
138
cur = READ_ONCE(execlists->active);
drivers/gpu/drm/i915/gt/intel_engine.h
143
active = READ_ONCE(*cur);
drivers/gpu/drm/i915/gt/intel_engine.h
144
cur = READ_ONCE(execlists->active);
drivers/gpu/drm/i915/gt/intel_engine.h
156
return READ_ONCE(engine->status_page.addr[reg]);
drivers/gpu/drm/i915/gt/intel_engine.h
346
return READ_ONCE(engine->props.heartbeat_interval_ms);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1620
return READ_ONCE(engine->props.stop_timeout_ms);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1926
if (!READ_ONCE(gt->awake))
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2066
if (!READ_ONCE(t->expires))
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2155
write = READ_ONCE(*execlists->csb_write);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2402
engine->fw_domain, READ_ONCE(engine->fw_active));
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2405
rq = READ_ONCE(engine->heartbeat.systole);
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
169
long delay = READ_ONCE(engine->props.heartbeat_interval_ms);
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
212
serial = READ_ONCE(engine->serial);
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
31
delay = READ_ONCE(engine->props.heartbeat_interval_ms);
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
53
longer = READ_ONCE(engine->props.preempt_timeout_ms) * 2;
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
84
engine->wakeref_serial = READ_ONCE(engine->serial) + 1;
drivers/gpu/drm/i915/gt/intel_engine_pm.c
85
READ_ONCE(*ce->timeline->hwsp_seqno),
drivers/gpu/drm/i915/gt/intel_engine_pm.c
88
READ_ONCE(*ce->timeline->hwsp_seqno));
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1030
struct i915_request *rq = READ_ONCE(ve->request);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1054
GEM_BUG_ON(READ_ONCE(ve->context.inflight));
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1148
return rq->context->lrc.ccid == READ_ONCE(el->yield);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1162
if (READ_ONCE(engine->execlists.pending[0]))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1202
return READ_ONCE(engine->props.timeslice_duration_ms);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1249
return READ_ONCE(engine->props.preempt_timeout_ms);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1763
if (wait_for_atomic_us((entry = READ_ONCE(*csb)) != -1, 10)) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1784
u64 entry = READ_ONCE(*csb);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1841
tail = READ_ONCE(*execlists->csb_write);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2460
if (unlikely(READ_ONCE(engine->execlists.error_interrupt))) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
264
return READ_ONCE(rq->sched.attr.priority);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2843
GEM_BUG_ON(READ_ONCE(*execlists->csb_write) != reset_value);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3293
locked = READ_ONCE(rq->engine);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3295
while (unlikely(locked != (engine = READ_ONCE(rq->engine)))) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3405
if (READ_ONCE(stats->active))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3787
rq = READ_ONCE(ve->request);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3812
const int prio = READ_ONCE(sched_engine->queue_priority_hint);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3823
struct intel_engine_cs *sibling = READ_ONCE(ve->siblings[n]);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3828
if (!READ_ONCE(ve->request))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
4109
READ_ONCE(sched_engine->queue_priority_hint));
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
4137
struct i915_request *rq = READ_ONCE(ve->request);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
495
unsigned int tag = ffs(READ_ONCE(engine->context_tag));
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
582
if (READ_ONCE(ve->request))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
838
!READ_ONCE(prev->fence.error)) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
961
return READ_ONCE(rq->fence.flags);
drivers/gpu/drm/i915/gt/intel_gt_buffer_pool.c
193
age = READ_ONCE(node->age);
drivers/gpu/drm/i915/gt/intel_gt_buffer_pool.c
60
age = READ_ONCE(node->age);
drivers/gpu/drm/i915/gt/intel_gt_pm_debugfs.c
495
seq_printf(m, "Interactive? %d\n", READ_ONCE(rps->power.interactive));
drivers/gpu/drm/i915/gt/intel_gt_pm_debugfs.c
509
seq_printf(m, "Wait boosts: %d\n", READ_ONCE(rps->boosts));
drivers/gpu/drm/i915/gt/intel_gt_pm_debugfs.c
84
READ_ONCE(fw_domain->wake_count));
drivers/gpu/drm/i915/gt/intel_gt_requests.c
104
first = READ_ONCE(engine->retire);
drivers/gpu/drm/i915/gt/intel_gtt.h
452
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
drivers/gpu/drm/i915/gt/intel_lrc.c
1950
return READ_ONCE(ce->lrc_reg_state[CTX_TIMESTAMP]);
drivers/gpu/drm/i915/gt/intel_reset.c
851
vma = READ_ONCE(gt->ggtt->fence_regs[i].vma);
drivers/gpu/drm/i915/gt/intel_ring.h
92
unsigned int head = READ_ONCE(ring->head);
drivers/gpu/drm/i915/gt/intel_rps.c
1033
struct intel_rps *rps = &READ_ONCE(rq->engine)->gt->rps;
drivers/gpu/drm/i915/gt/intel_rps.c
1069
if (READ_ONCE(rps->cur_freq) < rps->boost_freq)
drivers/gpu/drm/i915/gt/mock_engine.c
273
locked = READ_ONCE(rq->engine);
drivers/gpu/drm/i915/gt/mock_engine.c
275
while (unlikely(locked != (engine = READ_ONCE(rq->engine)))) {
drivers/gpu/drm/i915/gt/selftest_engine_heartbeat.c
134
GEM_BUG_ON(READ_ONCE(engine->serial) != engine->wakeref_serial);
drivers/gpu/drm/i915/gt/selftest_engine_heartbeat.c
224
if (READ_ONCE(engine->heartbeat.systole)) {
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
111
while (READ_ONCE(sema[2]) == 0)
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
117
while (READ_ONCE(sema[1]) == 0) /* wait for the gpu to catch up */
drivers/gpu/drm/i915/gt/selftest_execlists.c
1220
while (!READ_ONCE(slot[i]) &&
drivers/gpu/drm/i915/gt/selftest_execlists.c
1381
} while (READ_ONCE(engine->execlists.pending[0]));
drivers/gpu/drm/i915/gt/selftest_execlists.c
1642
if (wait_for(READ_ONCE(*map), 10)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
1694
GEM_BUG_ON(READ_ONCE(*map));
drivers/gpu/drm/i915/gt/selftest_execlists.c
3321
if (READ_ONCE(result[0]) != NUM_GPR) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
3428
while (READ_ONCE(engine->execlists.pending[0]))
drivers/gpu/drm/i915/gt/selftest_execlists.c
61
if (!READ_ONCE(engine->execlists.pending[0]) && is_active(rq))
drivers/gpu/drm/i915/gt/selftest_execlists.c
81
if (READ_ONCE(engine->execlists.pending[0]))
drivers/gpu/drm/i915/gt/selftest_execlists.c
87
if (READ_ONCE(rq->fence.error))
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
1196
ret = READ_ONCE(threads[tmp].result);
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
255
return READ_ONCE(h->seqno[rq->fence.context % (PAGE_SIZE/sizeof(u32))]);
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
914
while (!READ_ONCE(arg->stop)) {
drivers/gpu/drm/i915/gt/selftest_lrc.c
194
u32 lri = READ_ONCE(hw[dw]);
drivers/gpu/drm/i915/gt/selftest_lrc.c
240
u32 offset = READ_ONCE(hw[dw]);
drivers/gpu/drm/i915/gt/selftest_lrc.c
69
if (!READ_ONCE(engine->execlists.pending[0]) && is_active(rq))
drivers/gpu/drm/i915/gt/selftest_lrc.c
836
timestamp = READ_ONCE(arg->ce[0]->lrc_reg_state[CTX_TIMESTAMP]);
drivers/gpu/drm/i915/gt/selftest_rc6.c
269
engine->name, READ_ONCE(*res));
drivers/gpu/drm/i915/gt/selftest_ring_submission.c
113
if (err || READ_ONCE(*result)) {
drivers/gpu/drm/i915/gt/selftest_ring_submission.c
124
if (READ_ONCE(*result) != STACK_MAGIC) {
drivers/gpu/drm/i915/gt/selftest_ring_submission.c
135
if (READ_ONCE(*result) != STACK_MAGIC) {
drivers/gpu/drm/i915/gt/selftest_ring_submission.c
167
if (READ_ONCE(*result)) {
drivers/gpu/drm/i915/gt/selftest_ring_submission.c
198
if (READ_ONCE(*result)) {
drivers/gpu/drm/i915/gt/selftest_rps.c
546
dc = READ_ONCE(*cntr);
drivers/gpu/drm/i915/gt/selftest_rps.c
549
dc = READ_ONCE(*cntr) - dc;
drivers/gpu/drm/i915/gt/selftest_rps.c
809
if (wait_for(READ_ONCE(*cntr), 10)) {
drivers/gpu/drm/i915/gt/selftest_slpc.c
526
status = READ_ONCE(threads[i].result);
drivers/gpu/drm/i915/gt/selftest_timeline.c
1389
if (READ_ONCE(*tl->hwsp_seqno) != count) {
drivers/gpu/drm/i915/gt/selftest_timeline.c
496
if (READ_ONCE(*tl->hwsp_seqno) != tl->seqno) {
drivers/gpu/drm/i915/gt/selftest_timeline.c
582
if (!err && READ_ONCE(*tl->hwsp_seqno) != n) {
drivers/gpu/drm/i915/gt/selftest_timeline.c
652
if (!err && READ_ONCE(*tl->hwsp_seqno) != n) {
drivers/gpu/drm/i915/gt/selftest_timeline.c
752
if (READ_ONCE(*hwsp_seqno[0]) != seqno[0] ||
drivers/gpu/drm/i915/gt/selftest_timeline.c
753
READ_ONCE(*hwsp_seqno[1]) != seqno[1]) {
drivers/gpu/drm/i915/gt/sysfs_engines.c
333
if (READ_ONCE(engine->execlists.pending[0]))
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
449
if (unlikely(tail != READ_ONCE(desc->tail))) {
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
455
if (unlikely(READ_ONCE(desc->head) >= size)) {
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
553
FIELD_GET(GUC_HXG_MSG_0_ORIGIN, READ_ONCE(req->status)) == \
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
631
head = READ_ONCE(desc->head);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
889
u32 tail = READ_ONCE(desc->tail);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
920
if (unlikely(head != READ_ONCE(desc->head))) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4996
!READ_ONCE(wq->busy),
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
506
ce->parallel.guc.wqi_head = READ_ONCE(*ce->parallel.guc.wq_head);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5588
READ_ONCE(*ce->parallel.guc.wq_head));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5590
READ_ONCE(*ce->parallel.guc.wq_tail));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5592
READ_ONCE(*ce->parallel.guc.wq_status));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
6004
if (READ_ONCE(engine->props.heartbeat_interval_ms))
drivers/gpu/drm/i915/i915_active.c
245
it = READ_ONCE(ref->cache);
drivers/gpu/drm/i915/i915_active.c
247
u64 cached = READ_ONCE(it->timeline);
drivers/gpu/drm/i915/i915_active.c
33
#define fetch_node(x) rb_entry(READ_ONCE(x), typeof(struct active_node), node)
drivers/gpu/drm/i915/i915_active.c
57
return (struct intel_engine_cs *)READ_ONCE(node->base.cb.node.prev);
drivers/gpu/drm/i915/i915_active.c
787
p = READ_ONCE(p->rb_right);
drivers/gpu/drm/i915/i915_active.c
789
p = READ_ONCE(p->rb_left);
drivers/gpu/drm/i915/i915_debugfs.c
387
seq_printf(m, "Interactive? %d\n", READ_ONCE(rps->power.interactive));
drivers/gpu/drm/i915/i915_debugfs.c
401
seq_printf(m, "Wait boosts: %d\n", READ_ONCE(rps->boosts));
drivers/gpu/drm/i915/i915_debugfs.c
93
return READ_ONCE(obj->userfault_count) ? 'g' : ' ';
drivers/gpu/drm/i915/i915_gpu_error.c
1019
sg = READ_ONCE(error->fit);
drivers/gpu/drm/i915/i915_gpu_error.c
2176
error = READ_ONCE(i915->gpu_error.first_error);
drivers/gpu/drm/i915/i915_gpu_error.c
2338
error = READ_ONCE(i915->gpu_error.first_error);
drivers/gpu/drm/i915/i915_gpu_error.c
974
if (READ_ONCE(error->sgl))
drivers/gpu/drm/i915/i915_list_util.h
20
return READ_ONCE(list->next) == head;
drivers/gpu/drm/i915/i915_mitigations.c
100
unsigned long local = READ_ONCE(mitigations);
drivers/gpu/drm/i915/i915_mitigations.c
27
return READ_ONCE(mitigations) & BIT(CLEAR_RESIDUALS);
drivers/gpu/drm/i915/i915_perf.c
3405
stream = READ_ONCE(engine->oa_group->exclusive_stream);
drivers/gpu/drm/i915/i915_pmu.c
514
if (!READ_ONCE(pmu->timer_enabled))
drivers/gpu/drm/i915/i915_pmu.c
702
val = READ_ONCE(pmu->irq_count);
drivers/gpu/drm/i915/i915_request.c
1083
struct list_head *pos = READ_ONCE(signal->link.prev);
drivers/gpu/drm/i915/i915_request.c
1105
if (unlikely(READ_ONCE(prev->link.next) != &signal->link)) {
drivers/gpu/drm/i915/i915_request.c
1141
return rq->sched.semaphores | READ_ONCE(rq->engine->saturated);
drivers/gpu/drm/i915/i915_request.c
1206
const intel_engine_mask_t mask = READ_ONCE(from->engine)->mask;
drivers/gpu/drm/i915/i915_request.c
1460
if (to->engine == READ_ONCE(from->engine))
drivers/gpu/drm/i915/i915_request.c
1490
is_power_of_2(to->execution_mask | READ_ONCE(from->execution_mask)))
drivers/gpu/drm/i915/i915_request.c
1683
bool pow2 = is_power_of_2(READ_ONCE(prev->engine)->mask |
drivers/gpu/drm/i915/i915_request.c
1932
timeout_ns = READ_ONCE(rq->engine->props.max_busywait_duration_ns);
drivers/gpu/drm/i915/i915_request.c
2092
if (READ_ONCE(wait.tsk))
drivers/gpu/drm/i915/i915_request.c
259
locked = READ_ONCE(rq->engine);
drivers/gpu/drm/i915/i915_request.c
261
while (unlikely(locked != (engine = READ_ONCE(rq->engine)))) {
drivers/gpu/drm/i915/i915_request.c
437
return READ_ONCE(engine->execlists.active);
drivers/gpu/drm/i915/i915_request.c
489
(rq = READ_ONCE(*port)); /* may race with promotion of pending[] */
drivers/gpu/drm/i915/i915_request.c
581
old = READ_ONCE(rq->fence.error);
drivers/gpu/drm/i915/i915_request.h
487
const u32 *hwsp = READ_ONCE(rq->hwsp_seqno);
drivers/gpu/drm/i915/i915_request.h
489
return READ_ONCE(*hwsp);
drivers/gpu/drm/i915/i915_scheduler.c
144
while (locked != (sched_engine = READ_ONCE(rq->engine)->sched_engine)) {
drivers/gpu/drm/i915/i915_scheduler.c
212
if (prio > READ_ONCE(p->signaler->attr.priority))
drivers/gpu/drm/i915/i915_timer_util.h
15
return READ_ONCE(t->expires);
drivers/gpu/drm/i915/i915_vma.c
595
ptr = READ_ONCE(vma->iomap);
drivers/gpu/drm/i915/i915_vma.h
340
return READ_ONCE(vma->iomap);
drivers/gpu/drm/i915/intel_uncore.c
900
unsigned int actual = READ_ONCE(domain->wake_count);
drivers/gpu/drm/i915/intel_wakeref.h
241
return READ_ONCE(wf->wakeref);
drivers/gpu/drm/i915/selftests/i915_active.c
126
if (READ_ONCE(active->retired) && count) {
drivers/gpu/drm/i915/selftests/i915_active.c
160
if (!READ_ONCE(active->retired)) {
drivers/gpu/drm/i915/selftests/i915_active.c
193
if (!READ_ONCE(active->retired)) {
drivers/gpu/drm/i915/selftests/i915_active.c
238
if (!READ_ONCE(active->retired)) {
drivers/gpu/drm/i915/selftests/i915_request.c
1673
status = READ_ONCE(threads[idx].result);
drivers/gpu/drm/i915/selftests/i915_request.c
1841
err = READ_ONCE(threads[i].result);
drivers/gpu/drm/i915/selftests/i915_request.c
2058
if (wait_for(READ_ONCE(*sema) == 0, 50)) {
drivers/gpu/drm/i915/selftests/i915_request.c
2069
if (wait_for(READ_ONCE(*sema) == 0, 50)) {
drivers/gpu/drm/i915/selftests/i915_request.c
2203
if (i > 1 && wait_for(READ_ONCE(sema[i - 1]), 500)) {
drivers/gpu/drm/i915/selftests/i915_request.c
2217
wait_for(READ_ONCE(sema[i - 1]), 500);
drivers/gpu/drm/i915/selftests/i915_request.c
2500
if (wait_for(READ_ONCE(sema[2 * i]) == -1, 500)) {
drivers/gpu/drm/i915/selftests/i915_request.c
2528
if (wait_for(READ_ONCE(sema[2 * i - 2]) != -1, 500)) {
drivers/gpu/drm/i915/selftests/i915_request.c
2615
if (wait_for(READ_ONCE(sema[i]) == -1, 50)) {
drivers/gpu/drm/i915/selftests/i915_request.c
2622
while (!READ_ONCE(cb.seen))
drivers/gpu/drm/i915/selftests/i915_request.c
3246
status = READ_ONCE(engines[idx].result);
drivers/gpu/drm/i915/selftests/i915_request.c
344
while (!READ_ONCE(thread->stop)) {
drivers/gpu/drm/i915/selftests/i915_request.c
520
err = READ_ONCE(threads[n].result);
drivers/gpu/drm/i915/selftests/i915_sw_fence.c
555
if (READ_ONCE(ipc.value)) {
drivers/gpu/drm/i915/selftests/i915_sw_fence.c
563
if (!READ_ONCE(ipc.value)) {
drivers/gpu/drm/i915/selftests/igt_spinner.c
225
return READ_ONCE(*seqno);
drivers/gpu/drm/imagination/pvr_ccb.c
123
u32 next_write_offset = (READ_ONCE(ctrl->write_offset) + 1) & READ_ONCE(ctrl->wrap_mask);
drivers/gpu/drm/imagination/pvr_ccb.c
127
if (READ_ONCE(ctrl->read_offset) != next_write_offset) {
drivers/gpu/drm/imagination/pvr_ccb.c
172
while ((read_offset = READ_ONCE(ctrl->read_offset)) != READ_ONCE(ctrl->write_offset)) {
drivers/gpu/drm/imagination/pvr_ccb.c
175
WRITE_ONCE(ctrl->read_offset, (read_offset + 1) & READ_ONCE(ctrl->wrap_mask));
drivers/gpu/drm/imagination/pvr_ccb.c
219
u32 wr_offset = READ_ONCE(ctrl->write_offset);
drivers/gpu/drm/imagination/pvr_ccb.c
220
u32 rd_offset = READ_ONCE(ctrl->read_offset);
drivers/gpu/drm/imagination/pvr_ccb.c
258
old_write_offset = READ_ONCE(ctrl->write_offset);
drivers/gpu/drm/imagination/pvr_ccb.c
406
int ret = wait_event_timeout(pvr_dev->kccb.rtn_q, READ_ONCE(pvr_dev->kccb.rtn[slot_nr]) &
drivers/gpu/drm/imagination/pvr_ccb.c
410
*rtn_out = READ_ONCE(pvr_dev->kccb.rtn[slot_nr]);
drivers/gpu/drm/imagination/pvr_ccb.c
431
idle = (READ_ONCE(ctrl->write_offset) == READ_ONCE(ctrl->read_offset));
drivers/gpu/drm/imagination/pvr_cccb.c
114
read_offset = READ_ONCE(ctrl->read_offset);
drivers/gpu/drm/imagination/pvr_cccb.c
176
read_offset = READ_ONCE(ctrl->read_offset);
drivers/gpu/drm/imagination/pvr_fw.c
875
if (READ_ONCE(fw_dev->fwif_sysinit->firmware_started))
drivers/gpu/drm/imagination/pvr_fw_trace.c
475
trace_seq_data->start_offset = READ_ONCE(tracebuf_space->trace_pointer);
drivers/gpu/drm/imagination/pvr_power.c
159
enum rogue_fwif_pow_state pow_state = READ_ONCE(pvr_dev->fw_dev.fwif_sysdata->pow_state);
drivers/gpu/drm/msm/adreno/adreno_gpu.c
268
!READ_ONCE(gpu->crashstate)) {
drivers/gpu/drm/msm/adreno/adreno_gpu.c
292
!READ_ONCE(gpu->crashstate);
drivers/gpu/drm/panthor/panthor_drv.c
1562
user_mmio_offset = READ_ONCE(pfile->user_mmio.offset);
drivers/gpu/drm/panthor/panthor_fw.c
1299
u32 ack, req = READ_ONCE(*req_ptr) & req_mask;
drivers/gpu/drm/panthor/panthor_fw.c
1304
ret = read_poll_timeout_atomic(READ_ONCE, ack,
drivers/gpu/drm/panthor/panthor_fw.c
1311
if (wait_event_timeout(*wq, (READ_ONCE(*ack_ptr) & req_mask) == req,
drivers/gpu/drm/panthor/panthor_fw.c
1316
ack = READ_ONCE(*ack_ptr);
drivers/gpu/drm/panthor/panthor_fw.h
436
__cur_val = READ_ONCE((__iface)->input->__in_reg); \
drivers/gpu/drm/panthor/panthor_fw.h
437
__out_val = READ_ONCE((__iface)->output->__out_reg); \
drivers/gpu/drm/panthor/panthor_fw.h
462
__cur_val = READ_ONCE((__iface)->input->__in_reg); \
drivers/gpu/drm/panthor/panthor_fw.h
472
__cur_val = READ_ONCE((__iface)->input->__in_reg); \
drivers/gpu/drm/panthor/panthor_sched.c
1793
req = READ_ONCE(csg_iface->input->req);
drivers/gpu/drm/panthor/panthor_sched.c
1794
ack = READ_ONCE(csg_iface->output->ack);
drivers/gpu/drm/panthor/panthor_sched.c
1795
cs_irq_req = READ_ONCE(csg_iface->output->cs_irq_req);
drivers/gpu/drm/panthor/panthor_sched.c
1796
cs_irq_ack = READ_ONCE(csg_iface->input->cs_irq_ack);
drivers/gpu/drm/panthor/panthor_sched.c
1863
req = READ_ONCE(glb_iface->input->req);
drivers/gpu/drm/panthor/panthor_sched.c
1864
ack = READ_ONCE(glb_iface->output->ack);
drivers/gpu/drm/radeon/radeon_gem.c
515
cur_placement = READ_ONCE(robj->tbo.resource->mem_type);
drivers/gpu/drm/radeon/radeon_gem.c
546
cur_placement = READ_ONCE(robj->tbo.resource->mem_type);
drivers/gpu/drm/scheduler/sched_internal.h
85
if (READ_ONCE(entity->dependency))
drivers/gpu/drm/scheduler/sched_main.c
1457
if (!READ_ONCE(s_entity->stopped))
drivers/gpu/drm/scheduler/sched_main.c
1582
return READ_ONCE(sched->pause_submit);
drivers/gpu/drm/vmwgfx/vmwgfx_drv.c
1445
if (READ_ONCE(dev_priv->suspend_locked)) {
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
1449
return READ_ONCE(*(vmw->fifo_mem + fifo_reg));
drivers/gpu/drm/vmwgfx/vmwgfx_irq.c
95
masked_status = status & READ_ONCE(dev_priv->irq_mask);
drivers/gpu/drm/vmwgfx/vmwgfx_vkms.c
256
if (!READ_ONCE(vblank->enabled)) {
drivers/gpu/drm/vmwgfx/vmwgfx_vkms.c
261
*vblank_time = READ_ONCE(du->vkms.timer.node.expires);
drivers/gpu/drm/xe/xe_bo.c
3011
if (!READ_ONCE(bo->ttm.pin_count))
drivers/gpu/drm/xe/xe_exec_queue.c
1660
lrc = READ_ONCE(q->lrc[i]);
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
1492
return READ_ONCE(gt->sriov.vf.migration.recovery_inprogress);
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
1503
if (irq_pending || READ_ONCE(gt->sriov.vf.migration.ggtt_need_fixes))
drivers/gpu/drm/xe/xe_guc_ct.c
1308
ret = wait_event_timeout(ct->g2h_fence_wq, READ_ONCE(g2h_fence.done), HZ);
drivers/gpu/drm/xe/xe_guc_ct.c
1311
if (READ_ONCE(g2h_fence.done)) {
drivers/gpu/drm/xe/xe_guc_ct.c
2009
snapshot->g2h_outstanding = READ_ONCE(ct->g2h_outstanding);
drivers/gpu/drm/xe/xe_guc_ct.h
34
return READ_ONCE(ct->state) != XE_GUC_CT_STATE_NOT_INITIALIZED;
drivers/gpu/drm/xe/xe_guc_ct.h
40
return READ_ONCE(ct->state) == XE_GUC_CT_STATE_ENABLED;
drivers/gpu/drm/xe/xe_guc_submit.c
1475
READ_ONCE(q->multi_queue.group->banned))
drivers/gpu/drm/xe/xe_guc_submit.c
2184
(!READ_ONCE(q->guc->suspend_pending) || exec_queue_killed(q) || \
drivers/gpu/drm/xe/xe_guc_submit.c
2459
struct xe_lrc *lrc = READ_ONCE(q->lrc[0]);
drivers/gpu/drm/xe/xe_guc_submit.c
772
!READ_ONCE(group->sync_pending) ||
drivers/gpu/drm/xe/xe_map.h
55
return READ_ONCE(*(u32 *)map->vaddr);
drivers/gpu/drm/xe/xe_pm.c
528
return READ_ONCE(xe->pm_callback_task);
drivers/gpu/drm/xe/xe_svm.c
68
.__flags = READ_ONCE(range->base.pages.flags.__flags),
drivers/gpu/drm/xe/xe_sync.c
445
return READ_ONCE(ufence->signalled);
drivers/gpu/drm/xe/xe_tlb_inval.c
219
int seqno_recv = READ_ONCE(tlb_inval->seqno_recv);
drivers/gpu/drm/xe/xe_vm.h
363
if (READ_ONCE(vm->validation.validating) == current) {
drivers/gpu/drm/xe/xe_vm.h
415
((READ_ONCE(tile_present) & ~READ_ONCE(tile_invalidated)) & BIT((tile)->id))
drivers/greybus/gb-beagleplay.c
256
tail = READ_ONCE(bg->tx_circ_buf.tail);
drivers/greybus/gb-beagleplay.c
561
if (READ_ONCE(bg->flashing_mode))
drivers/greybus/gb-beagleplay.c
571
if (!READ_ONCE(bg->flashing_mode))
drivers/greybus/gb-beagleplay.c
651
switch (READ_ONCE(bg->fwl_ack)) {
drivers/greybus/gb-beagleplay.c
690
switch (READ_ONCE(bg->fwl_cmd_response)) {
drivers/greybus/gb-beagleplay.c
775
*crc32 = READ_ONCE(bg->fwl_cmd_response);
drivers/hid/i2c-hid/i2c-hid-core.c
1178
if (!READ_ONCE(ihid->panel_follower_work_finished))
drivers/hid/uhid.c
197
!uhid->report_running || !READ_ONCE(uhid->running),
drivers/hid/uhid.c
199
if (!ret || !READ_ONCE(uhid->running) || uhid->report_running)
drivers/hid/uhid.c
240
if (!READ_ONCE(uhid->running))
drivers/hid/uhid.c
282
if (!READ_ONCE(uhid->running) || count > UHID_DATA_MAX)
drivers/hid/uhid.c
595
if (!READ_ONCE(uhid->running))
drivers/hid/uhid.c
606
if (!READ_ONCE(uhid->running))
drivers/hid/uhid.c
618
if (!READ_ONCE(uhid->running))
drivers/hid/uhid.c
628
if (!READ_ONCE(uhid->running))
drivers/hv/channel_mgmt.c
1235
while (READ_ONCE(channel->probe_done) == false) {
drivers/hv/channel_mgmt.c
1297
while (!READ_ONCE(channel->probe_done) || !READ_ONCE(channel->rescind))
drivers/hv/channel_mgmt.c
861
message_type = READ_ONCE(msg->header.message_type);
drivers/hv/connection.c
390
return READ_ONCE(vmbus_connection.channels[relid]);
drivers/hv/connection.c
420
callback_fn = READ_ONCE(channel->onchannel_callback);
drivers/hv/mshv_vtl_main.c
306
message_type = READ_ONCE(msg->header.message_type);
drivers/hv/mshv_vtl_main.c
317
eventfd = READ_ONCE(flag_eventfds[i]);
drivers/hv/mshv_vtl_main.c
480
poll_file = per_cpu_ptr(&mshv_vtl_poll_file, READ_ONCE(input.cpu));
drivers/hv/mshv_vtl_main.c
614
u32 offset = READ_ONCE(mshv_vtl_this_run()->vtl_ret_action_size);
drivers/hv/mshv_vtl_main.c
643
message_type = READ_ONCE(msg->header.message_type);
drivers/hv/mshv_vtl_main.c
670
if (READ_ONCE(mshv_vtl_this_run()->cancel)) {
drivers/hv/mshv_vtl_main.c
872
u32 message_type = READ_ONCE(msg->header.message_type);
drivers/hv/mshv_vtl_main.c
897
if (READ_ONCE(vtl_synic_mask_vmbus_sint_masked))
drivers/hv/mshv_vtl_main.c
904
READ_ONCE(has_message) ||
drivers/hv/mshv_vtl_main.c
905
READ_ONCE(vtl_synic_mask_vmbus_sint_masked));
drivers/hv/mshv_vtl_main.c
921
if (READ_ONCE(has_message) || READ_ONCE(vtl_synic_mask_vmbus_sint_masked))
drivers/hv/mshv_vtl_main.c
984
old_eventfd = READ_ONCE(flag_eventfds[set_eventfd.flag]);
drivers/hv/ring_buffer.c
140
read_loc = READ_ONCE(rbi->ring_buffer->read_index);
drivers/hv/ring_buffer.c
141
write_loc = READ_ONCE(rbi->ring_buffer->write_index);
drivers/hv/ring_buffer.c
482
pkt_len = READ_ONCE(desc->len8) << 3;
drivers/hv/ring_buffer.c
483
pkt_offset = READ_ONCE(desc->offset8) << 3;
drivers/hv/ring_buffer.c
58
if (READ_ONCE(rbi->ring_buffer->interrupt_mask))
drivers/hv/ring_buffer.c
609
pending_sz = READ_ONCE(rbi->ring_buffer->pending_send_sz);
drivers/hv/ring_buffer.c
67
if (old_write == READ_ONCE(rbi->ring_buffer->read_index)) {
drivers/hwmon/bt1-pvt.c
375
*val = !!READ_ONCE(pvt->cache[type].thres_sts_lo);
drivers/hwmon/bt1-pvt.c
377
*val = !!READ_ONCE(pvt->cache[type].thres_sts_hi);
drivers/hwmon/bt1-pvt.c
493
data = READ_ONCE(cache->data);
drivers/hwmon/max16065.c
188
int val = READ_ONCE(data->fault[attr2->nr]);
drivers/hwmon/max16065.c
206
int adc = READ_ONCE(data->adc[attr->index]);
drivers/hwmon/max16065.c
219
int curr_sense = READ_ONCE(data->curr_sense);
drivers/hwmon/xgene-hwmon.c
116
val = le16_to_cpu(READ_ONCE(*addr));
drivers/hwmon/xgene-hwmon.c
145
val = le16_to_cpu(READ_ONCE(generic_comm_base->status));
drivers/hwtracing/coresight/coresight-etm-perf.c
571
if (READ_ONCE(handle->event)) {
drivers/hwtracing/coresight/coresight-etm-perf.c
618
if (READ_ONCE(handle->event)) {
drivers/hwtracing/coresight/coresight-etm-perf.c
718
if (READ_ONCE(handle->event))
drivers/hwtracing/coresight/coresight-etm4x-core.c
893
drvdata->paused = !!READ_ONCE(event->hw.aux_paused);
drivers/hwtracing/coresight/coresight-tmc-etr.c
1258
sysfs_buf = READ_ONCE(drvdata->sysfs_buf);
drivers/hwtracing/coresight/coresight-tmc-etr.c
1280
sysfs_buf = READ_ONCE(drvdata->sysfs_buf);
drivers/i2c/busses/i2c-designware-amdpsp.c
44
status = READ_ONCE(req->hdr.status);
drivers/i2c/busses/i2c-mv64xxx.c
592
while (READ_ONCE(drv_data->block) &&
drivers/i2c/busses/i2c-rk3x.c
1050
while (READ_ONCE(i2c->busy) &&
drivers/i2c/busses/i2c-xgene-slimpro.c
121
val = le16_to_cpu(READ_ONCE(*addr));
drivers/i2c/busses/i2c-xgene-slimpro.c
183
status = le16_to_cpu(READ_ONCE(generic_comm_base->status));
drivers/i2c/muxes/i2c-mux-pca954x.c
342
idle_state = READ_ONCE(data->idle_state);
drivers/i2c/muxes/i2c-mux-pca954x.c
367
return sprintf(buf, "%d\n", READ_ONCE(data->idle_state));
drivers/infiniband/core/addr.c
466
struct net_device *ndev = READ_ONCE(dst->dev);
drivers/infiniband/core/cma.c
2195
state = READ_ONCE(id_priv->state);
drivers/infiniband/core/cma.c
2428
if (READ_ONCE(listen_id->state) != RDMA_CM_LISTEN) {
drivers/infiniband/core/cma.c
2470
if (READ_ONCE(conn_id->state) == RDMA_CM_CONNECT &&
drivers/infiniband/core/cma.c
2531
if (READ_ONCE(id_priv->state) != RDMA_CM_CONNECT)
drivers/infiniband/core/cma.c
2604
if (READ_ONCE(listen_id->state) != RDMA_CM_LISTEN)
drivers/infiniband/core/cma.c
3013
if (READ_ONCE(id_priv->state) == RDMA_CM_DESTROYING ||
drivers/infiniband/core/cma.c
3014
READ_ONCE(id_priv->state) == RDMA_CM_DEVICE_REMOVAL)
drivers/infiniband/core/cma.c
3032
if (READ_ONCE(id_priv->state) == RDMA_CM_DESTROYING ||
drivers/infiniband/core/cma.c
3033
READ_ONCE(id_priv->state) == RDMA_CM_DEVICE_REMOVAL)
drivers/infiniband/core/cma.c
4201
if (READ_ONCE(id_priv->state) != RDMA_CM_IDLE)
drivers/infiniband/core/cma.c
4260
if (READ_ONCE(id_priv->state) != RDMA_CM_CONNECT)
drivers/infiniband/core/cma.c
4702
if (READ_ONCE(id_priv->state) != RDMA_CM_CONNECT)
drivers/infiniband/core/cma.c
4906
if (READ_ONCE(id_priv->state) == RDMA_CM_DEVICE_REMOVAL ||
drivers/infiniband/core/cma.c
4907
READ_ONCE(id_priv->state) == RDMA_CM_DESTROYING)
drivers/infiniband/core/cma.c
5097
if (!id->device || (READ_ONCE(id_priv->state) != RDMA_CM_ADDR_BOUND &&
drivers/infiniband/core/cma.c
5098
READ_ONCE(id_priv->state) != RDMA_CM_ADDR_RESOLVED))
drivers/infiniband/core/cma.c
5219
if (READ_ONCE(id_priv->state) == RDMA_CM_DESTROYING ||
drivers/infiniband/core/cma.c
5220
READ_ONCE(id_priv->state) == RDMA_CM_DEVICE_REMOVAL)
drivers/infiniband/core/cq.c
455
(READ_ONCE(default_comp_vector) + 1) % num_comp_vectors;
drivers/infiniband/core/netlink.c
104
cb_table = READ_ONCE(rdma_nl_types[type].cb_table);
drivers/infiniband/core/netlink.c
115
WARN_ON(READ_ONCE(rdma_nl_types[index].cb_table)))
drivers/infiniband/core/netlink.c
93
cb_table = READ_ONCE(rdma_nl_types[type].cb_table);
drivers/infiniband/core/security.c
741
if (!READ_ONCE(map->agent.smp_allowed))
drivers/infiniband/core/ucma.c
1646
else if (READ_ONCE(mc->ctx->file) != file)
drivers/infiniband/core/uverbs.h
292
async_ev_file = READ_ONCE(attrs->ufile->default_async_file);
drivers/infiniband/core/uverbs_cmd.c
1077
obj->uevent.event_file = READ_ONCE(attrs->ufile->default_async_file);
drivers/infiniband/core/uverbs_cmd.c
1471
obj->uevent.event_file = READ_ONCE(attrs->ufile->default_async_file);
drivers/infiniband/core/uverbs_cmd.c
2972
obj->uevent.event_file = READ_ONCE(attrs->ufile->default_async_file);
drivers/infiniband/core/uverbs_cmd.c
3460
obj->uevent.event_file = READ_ONCE(attrs->ufile->default_async_file);
drivers/infiniband/hw/efa/efa_com.c
1184
while ((READ_ONCE(eqe->common) & EFA_ADMIN_EQE_PHASE_MASK) == phase) {
drivers/infiniband/hw/efa/efa_com.c
456
while ((READ_ONCE(cqe->acq_common_descriptor.flags) &
drivers/infiniband/hw/efa/efa_com.c
859
while ((READ_ONCE(aenq_common->flags) &
drivers/infiniband/hw/efa/efa_com.c
99
if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num)
drivers/infiniband/hw/erdma/erdma_cmdq.c
252
be32_to_cpu(READ_ONCE(*cqe)));
drivers/infiniband/hw/erdma/erdma_cq.c
14
be32_to_cpu(READ_ONCE(*cqe)));
drivers/infiniband/hw/erdma/erdma_eq.c
158
cqn = FIELD_GET(ERDMA_CEQE_HDR_CQN_MASK, READ_ONCE(*ceqe));
drivers/infiniband/hw/erdma/erdma_eq.c
25
u32 owner = FIELD_GET(ERDMA_CEQE_HDR_O_MASK, READ_ONCE(*eqe));
drivers/infiniband/hw/hfi1/file_ops.c
1654
!(READ_ONCE(dd->flags) & HFI1_FROZEN),
drivers/infiniband/hw/hfi1/pio.c
1424
sc->alloc_free = READ_ONCE(sc->free);
drivers/infiniband/hw/hfi1/pio.c
1431
sc->alloc_free = READ_ONCE(sc->free);
drivers/infiniband/hw/hfi1/pio.c
1671
head = READ_ONCE(sc->sr_head); /* snapshot the head */
drivers/infiniband/hw/hfi1/rc.c
2327
if (cmp_psn(psn, READ_ONCE(qp->s_next_psn)) >= 0)
drivers/infiniband/hw/hfi1/rc.c
441
if (qp->s_last == READ_ONCE(qp->s_head))
drivers/infiniband/hw/hfi1/rc.c
485
if (qp->s_tail == READ_ONCE(qp->s_head)) {
drivers/infiniband/hw/hfi1/sdma.c
1669
swtail = READ_ONCE(sde->descq_tail) & sde->sdma_mask;
drivers/infiniband/hw/hfi1/sdma.c
1814
swtail = READ_ONCE(sde->descq_tail) & sde->sdma_mask;
drivers/infiniband/hw/hfi1/sdma.c
2162
tail = READ_ONCE(sde->descq_tail) & sde->sdma_mask;
drivers/infiniband/hw/hfi1/sdma.c
3241
nr = ffz(READ_ONCE(sde->ahg_bits));
drivers/infiniband/hw/hfi1/sdma.h
399
READ_ONCE(sde->descq_head)) - 1;
drivers/infiniband/hw/hfi1/uc.c
41
if (qp->s_last == READ_ONCE(qp->s_head))
drivers/infiniband/hw/hfi1/uc.c
80
if (qp->s_cur == READ_ONCE(qp->s_head)) {
drivers/infiniband/hw/hfi1/ud.c
453
if (qp->s_last == READ_ONCE(qp->s_head))
drivers/infiniband/hw/hfi1/ud.c
466
if (qp->s_cur == READ_ONCE(qp->s_head))
drivers/infiniband/hw/hfi1/user_sdma.c
688
if (READ_ONCE(req->has_error))
drivers/infiniband/hw/hfi1/user_sdma.c
711
if (READ_ONCE(req->has_error))
drivers/infiniband/hw/hfi1/user_sdma.c
737
if (READ_ONCE(iovec->offset) == iovec->iov.iov_len) {
drivers/infiniband/hw/hfi1/vnic_sdma.c
135
if (unlikely(READ_ONCE(vnic_sdma->state) != HFI1_VNIC_SDMA_Q_ACTIVE))
drivers/infiniband/hw/hfi1/vnic_sdma.c
233
return (READ_ONCE(vnic_sdma->state) == HFI1_VNIC_SDMA_Q_ACTIVE);
drivers/infiniband/hw/irdma/hw.c
113
struct irdma_cq *icq = READ_ONCE(rf->cq_table[cq_idx]);
drivers/infiniband/hw/irdma/utils.c
348
(READ_ONCE(ip_dev->flags) & IFF_UP)) {
drivers/infiniband/hw/irdma/utils.c
388
dev == iwdev->netdev) && (READ_ONCE(dev->flags) & IFF_UP)) {
drivers/infiniband/hw/irdma/utils.c
613
READ_ONCE(cqp_request->request_done),
drivers/infiniband/hw/mlx5/mr.c
517
if (ent->disabled || READ_ONCE(ent->dev->fill_delay) || ent->is_tmp)
drivers/infiniband/hw/mlx5/mr.c
569
!READ_ONCE(dev->fill_delay)) {
drivers/infiniband/hw/mlx5/mr.c
608
READ_ONCE(cache->last_add) + 300 * HZ);
drivers/infiniband/hw/qedr/verbs.c
1107
while (oparams.num_cq_notif != READ_ONCE(cq->cnq_notif) && iter) {
drivers/infiniband/hw/qedr/verbs.c
1113
while (oparams.num_cq_notif != READ_ONCE(cq->cnq_notif) && iter) {
drivers/infiniband/sw/rdmavt/mr.c
762
if (!READ_ONCE(mr->lkey_published))
drivers/infiniband/sw/rdmavt/mr.c
871
if (!READ_ONCE(mr->lkey_published))
drivers/infiniband/sw/rdmavt/qp.c
1308
if (READ_ONCE(qp->s_last) != qp->s_head)
drivers/infiniband/sw/rdmavt/qp.c
1808
if (next == READ_ONCE(wq->tail)) {
drivers/infiniband/sw/rdmavt/qp.c
2167
call_send = qp->s_head == READ_ONCE(qp->s_last) && !wr->next;
drivers/infiniband/sw/rdmavt/qp.c
2224
if (next == READ_ONCE(wq->tail)) {
drivers/infiniband/sw/rdmavt/qp.c
2938
if (sqp->s_last == READ_ONCE(sqp->s_head))
drivers/infiniband/sw/rdmavt/rc.c
68
credits = READ_ONCE(qp->r_rq.kwq->count);
drivers/infiniband/sw/rdmavt/rc.c
75
head = READ_ONCE(qp->r_rq.kwq->head);
drivers/infiniband/sw/rdmavt/rc.c
76
tail = READ_ONCE(qp->r_rq.kwq->tail);
drivers/infiniband/sw/siw/siw.h
628
return READ_ONCE(sqe->flags) == 0;
drivers/infiniband/sw/siw/siw.h
635
if (READ_ONCE(sqe->flags) & SIW_WQE_VALID)
drivers/infiniband/sw/siw/siw.h
650
if (READ_ONCE(orq_e->flags) == 0)
drivers/infiniband/sw/siw/siw.h
665
if (READ_ONCE(irq_e->flags) == 0) {
drivers/infiniband/sw/siw/siw_cq.c
56
if (READ_ONCE(cqe->flags) & SIW_WQE_VALID) {
drivers/infiniband/sw/siw/siw_qp.c
1013
cq_notify = READ_ONCE(cq->notify->flags);
drivers/infiniband/sw/siw/siw_qp.c
1048
if (!READ_ONCE(cqe->flags)) {
drivers/infiniband/sw/siw/siw_qp.c
1105
if (!READ_ONCE(cqe->flags)) {
drivers/infiniband/sw/siw/siw_qp.c
1170
if (!READ_ONCE(sqe->flags))
drivers/infiniband/sw/siw/siw_qp.c
1206
if (!READ_ONCE(sqe->flags))
drivers/infiniband/sw/siw/siw_qp.c
1271
if (!READ_ONCE(rqe->flags))
drivers/infiniband/sw/siw/siw_qp_rx.c
751
if (READ_ONCE(orqe->flags) & SIW_WQE_VALID) {
drivers/infiniband/sw/siw/siw_verbs.c
191
attr->active_mtu = ib_mtu_int_to_enum(READ_ONCE(ndev->mtu));
drivers/infiniband/sw/siw/siw_verbs.c
533
qp_attr->path_mtu = ib_mtu_int_to_enum(READ_ONCE(ndev->mtu));
drivers/infiniband/ulp/ipoib/ipoib_main.c
1318
return READ_ONCE(dev->ifindex);
drivers/infiniband/ulp/ipoib/ipoib_main.c
1321
return READ_ONCE(priv->parent->ifindex);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1294
if (READ_ONCE(alive_path->state) != RTRS_CLT_CONNECTED)
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2124
if (READ_ONCE(clt_path->state) != RTRS_CLT_CONNECTING)
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2182
WARN_ON(READ_ONCE(clt_path->state) == RTRS_CLT_CONNECTED);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2605
if (READ_ONCE(clt_path->state) != RTRS_CLT_CONNECTED) {
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2606
if (READ_ONCE(clt_path->state) == RTRS_CLT_CONNECTING_ERR)
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2677
if (READ_ONCE(clt_path->state) != RTRS_CLT_RECONNECTING)
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2944
err = (READ_ONCE(clt_path->state) ==
drivers/infiniband/ulp/rtrs/rtrs-clt.c
3045
if (READ_ONCE(clt_path->state) != RTRS_CLT_CONNECTED)
drivers/infiniband/ulp/rtrs/rtrs-clt.c
3088
if (READ_ONCE(clt_path->state) != RTRS_CLT_CONNECTED)
drivers/infiniband/ulp/rtrs/rtrs-clt.c
59
if (READ_ONCE(clt_path->state) == RTRS_CLT_CONNECTED) {
drivers/infiniband/ulp/rtrs/rtrs-clt.c
766
READ_ONCE((&clt_path->s.entry)->next),
drivers/infiniband/ulp/rtrs/rtrs-clt.c
829
if (READ_ONCE(clt_path->state) != RTRS_CLT_CONNECTED)
drivers/infiniband/ulp/rtrs/rtrs-clt.c
880
if (READ_ONCE(clt_path->state) != RTRS_CLT_CONNECTED)
drivers/input/keyboard/imx-sm-bbm-key.c
71
if (READ_ONCE(bbnsm->suspended)) {
drivers/input/misc/pwm-beeper.c
72
unsigned long period = READ_ONCE(beeper->period);
drivers/input/misc/regulator-haptic.c
228
magnitude = READ_ONCE(haptic->magnitude);
drivers/input/touchscreen/zforce_ts.c
460
suspending = READ_ONCE(ts->suspending);
drivers/iommu/amd/iommu.c
1204
while ((__s64)(READ_ONCE(*iommu->cmd_sem) - data) < 0 &&
drivers/iommu/amd/iommu.c
2166
old_domid = READ_ONCE(dte->data[1]) & DTE_DOMID_MASK;
drivers/iommu/amd/iommu.c
3230
new = READ_ONCE(dte->data[2]);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
663
llq->val = READ_ONCE(cmdq->q.llq.val);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
669
llq->val = READ_ONCE(cmdq->q.llq.val);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
716
llq->val = READ_ONCE(cmdq->q.llq.val);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
816
llq.val = READ_ONCE(cmdq->q.llq.val);
drivers/iommu/arm/arm-smmu-v3/tegra241-cmdqv.c
392
if (READ_ONCE(bypass_vcmdq))
drivers/iommu/arm/arm-smmu-v3/tegra241-cmdqv.c
396
if (!READ_ONCE(vintf->enabled))
drivers/iommu/arm/arm-smmu-v3/tegra241-cmdqv.c
408
if (!vcmdq || !READ_ONCE(vcmdq->enabled))
drivers/iommu/dma-iommu.c
2053
iotlb_gather.queued = free_iova && READ_ONCE(cookie->fq_domain);
drivers/iommu/dma-iommu.c
825
iotlb_gather.queued = READ_ONCE(cookie->fq_domain);
drivers/iommu/generic_pt/fmt/amdv1.h
182
pts->entry = entry = READ_ONCE(*tablep);
drivers/iommu/generic_pt/fmt/amdv1.h
295
if (READ_ONCE(*tablep) & AMDV1PT_FMT_D)
drivers/iommu/generic_pt/fmt/amdv1.h
309
WRITE_ONCE(*tablep, READ_ONCE(*tablep) & ~(u64)AMDV1PT_FMT_D);
drivers/iommu/generic_pt/fmt/vtdss.h
148
return READ_ONCE(*tablep) & VTDSS_FMT_D;
drivers/iommu/generic_pt/fmt/vtdss.h
156
WRITE_ONCE(*tablep, READ_ONCE(*tablep) & ~(u64)VTDSS_FMT_D);
drivers/iommu/generic_pt/fmt/vtdss.h
93
pts->entry = entry = READ_ONCE(tablep[pts->index]);
drivers/iommu/generic_pt/fmt/x86_64.h
115
pts->entry = entry = READ_ONCE(tablep[pts->index]);
drivers/iommu/generic_pt/iommu_pt.h
672
uintptr_t top_of_table = READ_ONCE(common->top_of_table);
drivers/iommu/generic_pt/pt_defs.h
325
return READ_ONCE(common->top_of_table) % (1 << PT_TOP_LEVEL_BITS);
drivers/iommu/generic_pt/pt_iter.h
257
return _pt_top_range(common, READ_ONCE(common->top_of_table));
drivers/iommu/generic_pt/pt_iter.h
314
_pt_top_range(common, READ_ONCE(common->top_of_table));
drivers/iommu/intel/dmar.c
1447
while (READ_ONCE(qi->desc_status[wait_index]) != QI_DONE) {
drivers/iommu/intel/iommu.h
906
val = READ_ONCE(context->lo) | 1;
drivers/iommu/intel/iommu.h
920
val = READ_ONCE(context->lo) & GENMASK_ULL(63, 1);
drivers/iommu/intel/pasid.c
256
WARN_ON(READ_ONCE(pte->val[0]) != 0);
drivers/iommu/intel/pasid.h
116
old = READ_ONCE(*ptr);
drivers/iommu/intel/pasid.h
122
return READ_ONCE(*ptr);
drivers/iommu/intel/pasid.h
141
return (u16)(READ_ONCE(pe->val[1]) & GENMASK_ULL(15, 0));
drivers/iommu/intel/pasid.h
57
return READ_ONCE(pde->val) & PASID_PTE_PRESENT;
drivers/iommu/intel/pasid.h
67
return phys_to_virt(READ_ONCE(pde->val) & PDE_PFN_MASK);
drivers/iommu/intel/pasid.h
73
return READ_ONCE(pte->val[0]) & PASID_PTE_PRESENT;
drivers/iommu/intel/pasid.h
79
return READ_ONCE(pte->val[0]) & PASID_PTE_FPD;
drivers/iommu/intel/pasid.h
85
return (u16)((READ_ONCE(pte->val[0]) >> 6) & 0x7);
drivers/iommu/io-pgtable-arm-v7s.c
481
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-arm-v7s.c
573
pte[i] = READ_ONCE(ptep[i]);
drivers/iommu/io-pgtable-arm-v7s.c
654
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-arm.c
453
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-arm.c
637
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-arm.c
650
pte = READ_ONCE(ptep[i]);
drivers/iommu/io-pgtable-arm.c
782
arm_lpae_iopte pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-dart.c
189
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-dart.c
255
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-dart.c
268
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-dart.c
318
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-dart.c
350
pte = READ_ONCE(*ptep);
drivers/iommu/iommu.c
3007
if (!dev->iommu && !READ_ONCE(iommu->ready))
drivers/iommu/iommufd/vfio_compat.c
230
if (READ_ONCE(ioas->iopt.disable_large_pages)) {
drivers/iommu/iommufd/vfio_compat.c
312
bool no_iommu_mode = READ_ONCE(ictx->no_iommu_mode);
drivers/iommu/riscv/iommu.c
1024
tc = READ_ONCE(dc->tc);
drivers/iommu/riscv/iommu.c
1046
tc = READ_ONCE(dc->tc);
drivers/iommu/riscv/iommu.c
1104
pte = READ_ONCE(ptr[i]);
drivers/iommu/riscv/iommu.c
1136
pte = READ_ONCE(*ptr);
drivers/iommu/riscv/iommu.c
1177
pte = READ_ONCE(*ptr);
drivers/iommu/riscv/iommu.c
1216
old = READ_ONCE(*ptr);
drivers/iommu/riscv/iommu.c
1266
old = READ_ONCE(*ptr);
drivers/iommu/riscv/iommu.c
1523
if (READ_ONCE(dc->tc) & RISCV_IOMMU_DC_TC_V)
drivers/iommu/riscv/iommu.c
618
ddt = READ_ONCE(*(unsigned long *)ddtp);
drivers/iommu/s390-iommu.c
1010
rte = READ_ONCE(rto[rtx]);
drivers/iommu/s390-iommu.c
1013
ste = READ_ONCE(sto[sx]);
drivers/iommu/s390-iommu.c
1016
pte = READ_ONCE(pto[px]);
drivers/iommu/s390-iommu.c
327
rse = READ_ONCE(*rsep);
drivers/iommu/s390-iommu.c
357
rfe = READ_ONCE(*rfep);
drivers/iommu/s390-iommu.c
388
rte = READ_ONCE(*rtep);
drivers/iommu/s390-iommu.c
415
ste = READ_ONCE(*step);
drivers/iommu/s390-iommu.c
479
pte = READ_ONCE(*ptep);
drivers/iommu/s390-iommu.c
955
rfe = READ_ONCE(rfo[rfx]);
drivers/iommu/s390-iommu.c
978
rse = READ_ONCE(rso[rsx]);
drivers/irqchip/irq-riscv-imsic-state.c
141
if (READ_ONCE(vec->enable))
drivers/irqchip/irq-riscv-imsic-state.c
150
mvec = READ_ONCE(vec->move_prev);
drivers/irqchip/irq-riscv-imsic-state.c
156
if (READ_ONCE(mvec->move_next)) {
drivers/irqchip/irq-riscv-imsic-state.c
170
mvec = READ_ONCE(vec->move_next);
drivers/irqchip/irq-riscv-imsic-state.c
185
if (READ_ONCE(tvec->enable)) {
drivers/irqchip/irq-riscv-imsic-state.c
345
mvec = READ_ONCE(vec->move_prev);
drivers/irqchip/irq-riscv-imsic-state.c
363
enabled = READ_ONCE(vec->enable);
drivers/irqchip/irq-riscv-imsic-state.h
87
return READ_ONCE(vec->enable);
drivers/irqchip/irq-riscv-imsic-state.h
92
return READ_ONCE(vec->move_prev);
drivers/isdn/mISDN/timerdev.c
112
wait_event_interruptible(dev->wait, (READ_ONCE(dev->work) ||
drivers/isdn/mISDN/timerdev.c
148
work = READ_ONCE(dev->work);
drivers/md/dm-bufio.c
1140
dm_bufio_cache_size_latch = READ_ONCE(dm_bufio_cache_size);
drivers/md/dm-bufio.c
2407
unsigned long retain_bytes = READ_ONCE(dm_bufio_retain_bytes);
drivers/md/dm-bufio.c
2764
last_accessed = READ_ONCE(b->last_accessed);
drivers/md/dm-clone-target.c
1100
if (current_volume > READ_ONCE(clone->hydration_threshold))
drivers/md/dm-clone-target.c
1409
READ_ONCE(clone->hydration_threshold),
drivers/md/dm-clone-target.c
1410
READ_ONCE(clone->hydration_batch_size));
drivers/md/dm-clone-target.c
154
return READ_ONCE(clone->mode);
drivers/md/dm-clone-target.c
982
unsigned int max_batch_size = READ_ONCE(clone->hydration_batch_size);
drivers/md/dm-crypt.c
267
val = min_not_zero(READ_ONCE(max_write_size),
drivers/md/dm-crypt.c
270
val = min_not_zero(READ_ONCE(max_read_size),
drivers/md/dm-integrity.c
2260
if (READ_ONCE(ic->free_sectors) <= ic->free_sectors_threshold)
drivers/md/dm-integrity.c
2884
if (READ_ONCE(ic->free_sectors) <= ic->free_sectors_threshold)
drivers/md/dm-integrity.c
390
return READ_ONCE(ic->failed);
drivers/md/dm-kcopyd.c
134
throttle = READ_ONCE(t->throttle);
drivers/md/dm-kcopyd.c
185
if (likely(READ_ONCE(t->throttle) >= 100))
drivers/md/dm-mpath.c
2109
if (pg != READ_ONCE(m->current_pg) ||
drivers/md/dm-mpath.c
2110
READ_ONCE(m->is_suspending))
drivers/md/dm-mpath.c
2155
pgpath = READ_ONCE(m->current_pgpath);
drivers/md/dm-mpath.c
2177
if (!READ_ONCE(m->current_pg)) {
drivers/md/dm-mpath.c
2256
pg = READ_ONCE(m->current_pg);
drivers/md/dm-mpath.c
2257
next_pg = READ_ONCE(m->next_pg);
drivers/md/dm-mpath.c
2258
if (unlikely(!READ_ONCE(m->current_pgpath) && next_pg))
drivers/md/dm-mpath.c
391
if (unlikely(READ_ONCE(m->current_pg) != pg)) {
drivers/md/dm-mpath.c
417
pg = READ_ONCE(m->current_pg);
drivers/md/dm-mpath.c
425
if (READ_ONCE(m->next_pg)) {
drivers/md/dm-mpath.c
519
pgpath = READ_ONCE(m->current_pgpath);
drivers/md/dm-mpath.c
616
pgpath = READ_ONCE(m->current_pgpath);
drivers/md/dm-mpath.c
801
READ_ONCE(queue_if_no_path_timeout_secs) * HZ;
drivers/md/dm-stats.c
695
(bi_sector == (READ_ONCE(last->last_sector) &&
drivers/md/dm-stats.c
697
(READ_ONCE(last->last_rw) == WRITE))
drivers/md/dm-stats.c
749
shared->tmp.sectors[READ] += READ_ONCE(p->sectors[READ]);
drivers/md/dm-stats.c
750
shared->tmp.sectors[WRITE] += READ_ONCE(p->sectors[WRITE]);
drivers/md/dm-stats.c
751
shared->tmp.ios[READ] += READ_ONCE(p->ios[READ]);
drivers/md/dm-stats.c
752
shared->tmp.ios[WRITE] += READ_ONCE(p->ios[WRITE]);
drivers/md/dm-stats.c
753
shared->tmp.merges[READ] += READ_ONCE(p->merges[READ]);
drivers/md/dm-stats.c
754
shared->tmp.merges[WRITE] += READ_ONCE(p->merges[WRITE]);
drivers/md/dm-stats.c
755
shared->tmp.ticks[READ] += READ_ONCE(p->ticks[READ]);
drivers/md/dm-stats.c
756
shared->tmp.ticks[WRITE] += READ_ONCE(p->ticks[WRITE]);
drivers/md/dm-stats.c
757
shared->tmp.io_ticks[READ] += READ_ONCE(p->io_ticks[READ]);
drivers/md/dm-stats.c
758
shared->tmp.io_ticks[WRITE] += READ_ONCE(p->io_ticks[WRITE]);
drivers/md/dm-stats.c
759
shared->tmp.io_ticks_total += READ_ONCE(p->io_ticks_total);
drivers/md/dm-stats.c
760
shared->tmp.time_in_queue += READ_ONCE(p->time_in_queue);
drivers/md/dm-stats.c
765
shared->tmp.histogram[i] += READ_ONCE(p->histogram[i]);
drivers/md/dm-switch.c
148
return (READ_ONCE(sctx->region_table[region_index]) >> bit) &
drivers/md/dm-thin.c
2545
unsigned long no_space_timeout = READ_ONCE(no_space_timeout_secs) * HZ;
drivers/md/dm-vdo/admin-state.h
72
return READ_ONCE(state->current_state);
drivers/md/dm-vdo/block-map.c
3327
totals.dirty_pages += READ_ONCE(stats->dirty_pages);
drivers/md/dm-vdo/block-map.c
3328
totals.clean_pages += READ_ONCE(stats->clean_pages);
drivers/md/dm-vdo/block-map.c
3329
totals.free_pages += READ_ONCE(stats->free_pages);
drivers/md/dm-vdo/block-map.c
3330
totals.failed_pages += READ_ONCE(stats->failed_pages);
drivers/md/dm-vdo/block-map.c
3331
totals.incoming_pages += READ_ONCE(stats->incoming_pages);
drivers/md/dm-vdo/block-map.c
3332
totals.outgoing_pages += READ_ONCE(stats->outgoing_pages);
drivers/md/dm-vdo/block-map.c
3333
totals.cache_pressure += READ_ONCE(stats->cache_pressure);
drivers/md/dm-vdo/block-map.c
3334
totals.read_count += READ_ONCE(stats->read_count);
drivers/md/dm-vdo/block-map.c
3335
totals.write_count += READ_ONCE(stats->write_count);
drivers/md/dm-vdo/block-map.c
3336
totals.failed_reads += READ_ONCE(stats->failed_reads);
drivers/md/dm-vdo/block-map.c
3337
totals.failed_writes += READ_ONCE(stats->failed_writes);
drivers/md/dm-vdo/block-map.c
3338
totals.reclaimed += READ_ONCE(stats->reclaimed);
drivers/md/dm-vdo/block-map.c
3339
totals.read_outgoing += READ_ONCE(stats->read_outgoing);
drivers/md/dm-vdo/block-map.c
3340
totals.found_in_cache += READ_ONCE(stats->found_in_cache);
drivers/md/dm-vdo/block-map.c
3341
totals.discard_required += READ_ONCE(stats->discard_required);
drivers/md/dm-vdo/block-map.c
3342
totals.wait_for_page += READ_ONCE(stats->wait_for_page);
drivers/md/dm-vdo/block-map.c
3343
totals.fetch_required += READ_ONCE(stats->fetch_required);
drivers/md/dm-vdo/block-map.c
3344
totals.pages_loaded += READ_ONCE(stats->pages_loaded);
drivers/md/dm-vdo/block-map.c
3345
totals.pages_saved += READ_ONCE(stats->pages_saved);
drivers/md/dm-vdo/block-map.c
3346
totals.flush_count += READ_ONCE(stats->flush_count);
drivers/md/dm-vdo/data-vio.c
1082
return READ_ONCE(pool->limiter.busy);
drivers/md/dm-vdo/data-vio.c
1087
return READ_ONCE(pool->limiter.limit);
drivers/md/dm-vdo/data-vio.c
1092
return READ_ONCE(pool->limiter.max_busy);
drivers/md/dm-vdo/data-vio.c
1303
(READ_ONCE(vdo->read_only_notifier.read_only_error) == VDO_SUCCESS) &&
drivers/md/dm-vdo/data-vio.c
464
if (!data_vio->write && READ_ONCE(lock_holder->allocation_succeeded)) {
drivers/md/dm-vdo/dedupe.c
2290
VDO_ASSERT_LOG_ONLY(READ_ONCE(zone->active) == 0, "all contexts inactive");
drivers/md/dm-vdo/dedupe.c
2638
tally->dedupe_advice_valid += READ_ONCE(stats->dedupe_advice_valid);
drivers/md/dm-vdo/dedupe.c
2639
tally->dedupe_advice_stale += READ_ONCE(stats->dedupe_advice_stale);
drivers/md/dm-vdo/dedupe.c
2640
tally->concurrent_data_matches += READ_ONCE(stats->concurrent_data_matches);
drivers/md/dm-vdo/dedupe.c
2641
tally->concurrent_hash_collisions += READ_ONCE(stats->concurrent_hash_collisions);
drivers/md/dm-vdo/dedupe.c
2642
tally->curr_dedupe_queries += READ_ONCE(zone->active);
drivers/md/dm-vdo/dedupe.c
2766
return READ_ONCE(zones->dedupe_flag) ? ONLINE : OFFLINE;
drivers/md/dm-vdo/dedupe.c
2909
if (!READ_ONCE(vdo->hash_zones->dedupe_flag)) {
drivers/md/dm-vdo/flush.c
324
READ_ONCE(zone->oldest_active_generation));
drivers/md/dm-vdo/funnel-queue.c
116
queue->oldest = READ_ONCE(oldest->next);
drivers/md/dm-vdo/funnel-queue.c
166
if (READ_ONCE(queue->newest) != &queue->stub)
drivers/md/dm-vdo/funnel-queue.c
46
struct funnel_queue_entry *next = READ_ONCE(oldest->next);
drivers/md/dm-vdo/funnel-queue.c
61
next = READ_ONCE(oldest->next);
drivers/md/dm-vdo/funnel-queue.c
69
struct funnel_queue_entry *newest = READ_ONCE(queue->newest);
drivers/md/dm-vdo/funnel-queue.c
86
next = READ_ONCE(oldest->next);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
104
if (!READ_ONCE(queue->running)) {
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
145
} else if (!READ_ONCE(queue->running)) {
drivers/md/dm-vdo/indexer/index-session.c
151
WRITE_ONCE(*count_ptr, READ_ONCE(*count_ptr) + 1);
drivers/md/dm-vdo/indexer/index-session.c
691
stats->posts_found = READ_ONCE(session_stats->posts_found);
drivers/md/dm-vdo/indexer/index-session.c
692
stats->in_memory_posts_found = READ_ONCE(session_stats->posts_found_open_chapter);
drivers/md/dm-vdo/indexer/index-session.c
693
stats->dense_posts_found = READ_ONCE(session_stats->posts_found_dense);
drivers/md/dm-vdo/indexer/index-session.c
694
stats->sparse_posts_found = READ_ONCE(session_stats->posts_found_sparse);
drivers/md/dm-vdo/indexer/index-session.c
695
stats->posts_not_found = READ_ONCE(session_stats->posts_not_found);
drivers/md/dm-vdo/indexer/index-session.c
696
stats->updates_found = READ_ONCE(session_stats->updates_found);
drivers/md/dm-vdo/indexer/index-session.c
697
stats->updates_not_found = READ_ONCE(session_stats->updates_not_found);
drivers/md/dm-vdo/indexer/index-session.c
698
stats->deletions_found = READ_ONCE(session_stats->deletions_found);
drivers/md/dm-vdo/indexer/index-session.c
699
stats->deletions_not_found = READ_ONCE(session_stats->deletions_not_found);
drivers/md/dm-vdo/indexer/index-session.c
700
stats->queries_found = READ_ONCE(session_stats->queries_found);
drivers/md/dm-vdo/indexer/index-session.c
701
stats->queries_not_found = READ_ONCE(session_stats->queries_not_found);
drivers/md/dm-vdo/indexer/index-session.c
702
stats->requests = READ_ONCE(session_stats->requests);
drivers/md/dm-vdo/indexer/sparse-cache.c
313
if (READ_ONCE(chapter->skip_search) != skip_search)
drivers/md/dm-vdo/indexer/sparse-cache.c
555
return READ_ONCE(chapter->skip_search);
drivers/md/dm-vdo/indexer/volume.c
111
.value = READ_ONCE(cache->search_pending_counters[zone_number].atomic_value),
drivers/md/dm-vdo/indexer/volume.c
218
if (atomic64_read(&cache->clock) != READ_ONCE(page->last_used))
drivers/md/dm-vdo/indexer/volume.c
237
last_used = READ_ONCE(cache->cache[i].last_used);
drivers/md/dm-vdo/indexer/volume.c
662
index_value = READ_ONCE(cache->index[physical_page]);
drivers/md/dm-vdo/logger.c
23
int log_level_latch = READ_ONCE(vdo_log_level);
drivers/md/dm-vdo/logical-zone.c
356
(unsigned long long) READ_ONCE(zone->flush_generation),
drivers/md/dm-vdo/logical-zone.c
357
(unsigned long long) READ_ONCE(zone->oldest_active_generation),
drivers/md/dm-vdo/logical-zone.c
358
(unsigned long long) READ_ONCE(zone->notification_generation),
drivers/md/dm-vdo/logical-zone.c
359
vdo_bool_to_string(READ_ONCE(zone->notifying)),
drivers/md/dm-vdo/logical-zone.c
360
(unsigned long long) READ_ONCE(zone->ios_in_flush_generation));
drivers/md/dm-vdo/packer.c
230
.compressed_fragments_written = READ_ONCE(stats->compressed_fragments_written),
drivers/md/dm-vdo/packer.c
231
.compressed_blocks_written = READ_ONCE(stats->compressed_blocks_written),
drivers/md/dm-vdo/packer.c
232
.compressed_fragments_in_packer = READ_ONCE(stats->compressed_fragments_in_packer),
drivers/md/dm-vdo/slab-depot.c
3717
READ_ONCE(scrubber->slab_count),
drivers/md/dm-vdo/slab-depot.c
4513
total += READ_ONCE(depot->allocators[zone].allocated_blocks);
drivers/md/dm-vdo/slab-depot.c
4530
return (READ_ONCE(depot->slab_count) * depot->slab_config.data_blocks);
drivers/md/dm-vdo/slab-depot.c
5099
totals.slabs_opened += READ_ONCE(stats->slabs_opened);
drivers/md/dm-vdo/slab-depot.c
5100
totals.slabs_reopened += READ_ONCE(stats->slabs_reopened);
drivers/md/dm-vdo/slab-depot.c
5122
READ_ONCE(depot->allocators[zone].ref_counts_statistics.blocks_written);
drivers/md/dm-vdo/slab-depot.c
5146
totals.disk_full_count += READ_ONCE(stats->disk_full_count);
drivers/md/dm-vdo/slab-depot.c
5147
totals.flush_count += READ_ONCE(stats->flush_count);
drivers/md/dm-vdo/slab-depot.c
5148
totals.blocked_count += READ_ONCE(stats->blocked_count);
drivers/md/dm-vdo/slab-depot.c
5149
totals.blocks_written += READ_ONCE(stats->blocks_written);
drivers/md/dm-vdo/slab-depot.c
5150
totals.tail_busy_count += READ_ONCE(stats->tail_busy_count);
drivers/md/dm-vdo/slab-depot.c
5165
slab_count_t slab_count = READ_ONCE(depot->slab_count);
drivers/md/dm-vdo/slab-depot.c
5171
unrecovered += READ_ONCE(depot->allocators[zone].scrubber.slab_count);
drivers/md/dm-vdo/slab-depot.c
5192
(unsigned int) depot->old_zone_count, READ_ONCE(depot->slab_count),
drivers/md/dm-vdo/vdo.c
1151
vdo_log_error_strerror(READ_ONCE(notifier->read_only_error),
drivers/md/dm-vdo/vdo.c
1407
return READ_ONCE(vdo->compressing);
drivers/md/dm-verity-target.c
673
bytes <= READ_ONCE(dm_verity_use_bh_bytes[ioprio]) &&
drivers/md/dm-verity-target.c
725
unsigned int cluster = READ_ONCE(dm_verity_prefetch_cluster);
drivers/md/dm-writecache.c
1819
if (READ_ONCE(wc->writeback_size) - wbl->size >= wc->max_writeback_jobs) {
drivers/md/dm-writecache.c
454
#define writecache_has_error(wc) (unlikely(READ_ONCE((wc)->error)))
drivers/md/dm-zone.c
53
struct dm_table *zone_revalidate_map = READ_ONCE(md->zone_revalidate_map);
drivers/md/dm.c
132
int latch = READ_ONCE(swap_bios);
drivers/md/dm.c
153
int param = READ_ONCE(*module_param);
drivers/md/dm.c
174
unsigned int param = READ_ONCE(*module_param);
drivers/md/md.c
3028
unsigned long flags = READ_ONCE(rdev->flags);
drivers/md/md.c
3729
struct mddev *mddev = READ_ONCE(rdev->mddev);
drivers/md/md.c
388
if (READ_ONCE(mddev->suspend_lo) >= READ_ONCE(mddev->suspend_hi))
drivers/md/md.c
390
if (bio->bi_iter.bi_sector >= READ_ONCE(mddev->suspend_hi))
drivers/md/md.c
392
if (bio_end_sector(bio) < READ_ONCE(mddev->suspend_lo))
drivers/md/md.c
5616
(unsigned long long)READ_ONCE(mddev->suspend_lo));
drivers/md/md.c
5647
(unsigned long long)READ_ONCE(mddev->suspend_hi));
drivers/md/md.c
9868
mddev = READ_ONCE(rdev->mddev);
drivers/md/raid1.c
1062
if (!READ_ONCE(conf->array_frozen) &&
drivers/md/raid1.c
1113
if (!READ_ONCE(conf->array_frozen))
drivers/md/raid1.c
1726
struct md_rdev *rdev = READ_ONCE(conf->mirrors[i].rdev);
drivers/md/raid1.c
838
(READ_ONCE(conf->nonrot_disks) || ctl.min_pending == 0))
drivers/md/raid10.c
1923
struct md_rdev *rdev = READ_ONCE(conf->mirrors[i].rdev);
drivers/md/raid10.c
992
if (READ_ONCE(conf->barrier))
drivers/md/raid5-cache.c
1152
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
1400
if (!READ_ONCE(conf->log))
drivers/md/raid5-cache.c
1421
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
1550
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
1566
target = READ_ONCE(log->reclaim_target);
drivers/md/raid5-cache.c
1592
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
2627
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
2794
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
2929
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
3036
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
326
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
353
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
398
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
451
struct r5l_log *log = READ_ONCE(conf->log);
drivers/md/raid5-cache.c
493
struct r5l_log *log = READ_ONCE(sh->raid_conf->log);
drivers/md/raid5-cache.c
693
!READ_ONCE(conf->log) ||
drivers/md/raid5-cache.c
696
log = READ_ONCE(conf->log);
drivers/md/raid5.c
702
struct md_rdev *rdev = READ_ONCE(conf->disks[i].rdev);
drivers/md/raid5.c
705
rdev = READ_ONCE(conf->disks[i].replacement);
drivers/md/raid5.c
727
struct md_rdev *rdev = READ_ONCE(conf->disks[i].rdev);
drivers/md/raid5.c
730
rdev = READ_ONCE(conf->disks[i].replacement);
drivers/md/raid5.c
7450
int max_stripes = READ_ONCE(conf->max_nr_stripes);
drivers/md/raid5.c
7451
int min_stripes = READ_ONCE(conf->min_nr_stripes);
drivers/md/raid5.c
8127
struct md_rdev *rdev = READ_ONCE(conf->disks[i].rdev);
drivers/media/dvb-core/dvb_ringbuffer.c
75
free = READ_ONCE(rbuf->pread) - rbuf->pwrite;
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
2096
if (READ_ONCE(stream->stopping)) {
drivers/misc/kgdbts.c
288
READ_ONCE(hw_break_val);
drivers/misc/lkdtm/bugs.c
123
while (READ_ONCE(wait_for_panic))
drivers/misc/lkdtm/bugs.c
153
while (READ_ONCE(wait_for_bug))
drivers/misc/mei/dma-ring.c
186
rd_idx = READ_ONCE(ctrl->dbuf_rd_idx) & (dbuf_depth - 1);
drivers/misc/mei/dma-ring.c
230
rd_idx = READ_ONCE(ctrl->hbuf_rd_idx);
drivers/misc/mei/dma-ring.c
231
wr_idx = READ_ONCE(ctrl->hbuf_wr_idx);
drivers/misc/mei/dma-ring.c
259
wr_idx = READ_ONCE(ctrl->hbuf_wr_idx) & (hbuf_depth - 1);
drivers/misc/vmw_balloon.c
1487
atomic64_read(&b->size), READ_ONCE(b->target));
drivers/misc/vmw_balloon.c
1664
seq_printf(f, "%-22s: %16lu\n", "target", READ_ONCE(b->target));
drivers/misc/vmw_balloon.c
957
target = READ_ONCE(b->target);
drivers/misc/vmw_balloon.c
973
if (target > size && time_before(jiffies, READ_ONCE(b->shrink_timeout)))
drivers/mmc/core/core.c
373
struct mmc_request *ongoing_mrq = READ_ONCE(host->ongoing_mrq);
drivers/net/amt.c
1254
if ((!v6 && !READ_ONCE(amt->ready4)) ||
drivers/net/amt.c
1255
(v6 && !READ_ONCE(amt->ready6)))
drivers/net/amt.c
2290
if (READ_ONCE(amt->status) != AMT_STATUS_SENT_UPDATE)
drivers/net/amt.c
2384
if (READ_ONCE(amt->ready4))
drivers/net/amt.c
2409
if (READ_ONCE(amt->ready6))
drivers/net/amt.c
2932
if (READ_ONCE(amt->status) >= AMT_STATUS_RECEIVED_ADVERTISEMENT)
drivers/net/amt.c
3229
amt->qrv = READ_ONCE(amt->net->ipv4.sysctl_igmp_qrv);
drivers/net/amt.c
568
ihv3->qrv = READ_ONCE(amt->net->ipv4.sysctl_igmp_qrv);
drivers/net/bonding/bond_alb.c
1398
count = slaves ? READ_ONCE(slaves->count) : 0;
drivers/net/bonding/bond_alb.c
1514
count = slaves ? READ_ONCE(slaves->count) : 0;
drivers/net/bonding/bond_main.c
1191
if (!usable || !READ_ONCE(usable->count))
drivers/net/bonding/bond_main.c
1524
slave_ops = READ_ONCE(slave->dev->header_ops);
drivers/net/bonding/bond_main.c
1545
slave_ops = READ_ONCE(slave->dev->header_ops);
drivers/net/bonding/bond_main.c
1619
recv_probe = READ_ONCE(bond->recv_probe);
drivers/net/bonding/bond_main.c
2902
if (commit || READ_ONCE(bond->send_peer_notif)) {
drivers/net/bonding/bond_main.c
3496
bond_time_in_interval(bond, READ_ONCE(slave->last_rx), 1)) {
drivers/net/bonding/bond_main.c
3522
!bond_time_in_interval(bond, READ_ONCE(slave->last_rx),
drivers/net/bonding/bond_main.c
3859
if (READ_ONCE(bond->send_peer_notif) || should_notify_rtnl) {
drivers/net/bonding/bond_main.c
4982
slave_cnt = READ_ONCE(bond->slave_cnt);
drivers/net/bonding/bond_main.c
5023
slave_cnt = READ_ONCE(bond->slave_cnt);
drivers/net/bonding/bond_main.c
5229
count = slaves ? READ_ONCE(slaves->count) : 0;
drivers/net/bonding/bond_main.c
5246
count = slaves ? READ_ONCE(slaves->count) : 0;
drivers/net/bonding/bond_main.c
5321
slaves_count = slaves ? READ_ONCE(slaves->count) : 0;
drivers/net/bonding/bond_main.c
5369
if (READ_ONCE(slave->queue_id) == skb_get_queue_mapping(skb)) {
drivers/net/bonding/bond_main.c
5502
count = slaves ? READ_ONCE(slaves->count) : 0;
drivers/net/bonding/bond_netlink.c
56
READ_ONCE(slave->queue_id)))
drivers/net/bonding/bond_options.c
1225
WRITE_ONCE(targets_rx[i], READ_ONCE(targets_rx[i+1]));
drivers/net/bonding/bond_procfs.c
212
seq_printf(seq, "Slave queue ID: %d\n", READ_ONCE(slave->queue_id));
drivers/net/bonding/bond_sysfs.c
638
READ_ONCE(slave->queue_id));
drivers/net/bonding/bond_sysfs_slave.c
56
return sysfs_emit(buf, "%d\n", READ_ONCE(slave->queue_id));
drivers/net/can/esd/esdacc.c
727
if (READ_ONCE(*ov->bmfifo.irq_cnt) != ov->bmfifo.local_irq_cnt) {
drivers/net/can/esd/esdacc.c
729
ov->bmfifo.local_irq_cnt = READ_ONCE(*ov->bmfifo.irq_cnt);
drivers/net/can/esd/esdacc.c
735
if (READ_ONCE(*core->bmfifo.irq_cnt) != core->bmfifo.local_irq_cnt) {
drivers/net/can/esd/esdacc.c
737
core->bmfifo.local_irq_cnt = READ_ONCE(*core->bmfifo.irq_cnt);
drivers/net/can/kvaser_pciefd/kvaser_pciefd_core.c
717
return can->tx_max_count - (READ_ONCE(can->tx_idx) - READ_ONCE(can->ack_idx));
drivers/net/can/rockchip/rockchip_canfd.h
515
return READ_ONCE(priv->tx_head) & (RKCANFD_TXFIFO_DEPTH - 1);
drivers/net/can/rockchip/rockchip_canfd.h
521
return READ_ONCE(priv->tx_tail) & (RKCANFD_TXFIFO_DEPTH - 1);
drivers/net/can/rockchip/rockchip_canfd.h
527
return READ_ONCE(priv->tx_head) - READ_ONCE(priv->tx_tail);
drivers/net/can/vxcan.c
135
iflink = peer ? READ_ONCE(peer->ifindex) : 0;
drivers/net/dsa/mv88e6xxx/pcs-639x.c
89
handler = READ_ONCE(mpcs->handle_irq);
drivers/net/ethernet/actions/owl-emac.c
518
status = READ_ONCE(desc->status);
drivers/net/ethernet/actions/owl-emac.c
519
control = READ_ONCE(desc->control);
drivers/net/ethernet/actions/owl-emac.c
589
status = READ_ONCE(desc->status);
drivers/net/ethernet/actions/owl-emac.c
590
control = READ_ONCE(desc->control);
drivers/net/ethernet/actions/owl-emac.c
643
status = READ_ONCE(desc->status);
drivers/net/ethernet/actions/owl-emac.c
725
status = READ_ONCE(ring->descs[tx_next].status);
drivers/net/ethernet/actions/owl-emac.c
733
status = READ_ONCE(ring->descs[ring->tail].status);
drivers/net/ethernet/actions/owl-emac.c
764
status = READ_ONCE(desc->status);
drivers/net/ethernet/airoha/airoha_ppe.c
948
ib1 = READ_ONCE(hwe->ib1);
drivers/net/ethernet/amazon/ena/ena_com.c
1814
if (READ_ONCE(resp->req_id) != phc->req_id) {
drivers/net/ethernet/amazon/ena/ena_com.c
1873
if (READ_ONCE(resp->req_id) != phc->req_id) {
drivers/net/ethernet/amazon/ena/ena_com.c
2309
while ((READ_ONCE(aenq_common->flags) & ENA_ADMIN_AENQ_COMMON_DESC_PHASE_MASK) == phase) {
drivers/net/ethernet/amazon/ena/ena_com.c
469
while ((READ_ONCE(cqe->acq_common_descriptor.flags) &
drivers/net/ethernet/amazon/ena/ena_com.c
824
if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num)
drivers/net/ethernet/amazon/ena/ena_eth_com.c
21
desc_phase = (READ_ONCE(cdesc->status) & ENA_ETH_IO_RX_CDESC_BASE_PHASE_MASK) >>
drivers/net/ethernet/amazon/ena/ena_eth_com.c
246
status = READ_ONCE(cdesc->status);
drivers/net/ethernet/amazon/ena/ena_eth_com.h
218
cdesc_phase = READ_ONCE(cdesc->flags) & ENA_ETH_IO_TX_CDESC_PHASE_MASK;
drivers/net/ethernet/amazon/ena/ena_eth_com.h
224
*req_id = READ_ONCE(cdesc->req_id);
drivers/net/ethernet/amazon/ena/ena_netdev.c
1480
READ_ONCE(ena_napi->interrupts_masked)) {
drivers/net/ethernet/amazon/ena/ena_netdev.c
3396
if (likely(READ_ONCE(ena_napi->first_interrupt)))
drivers/net/ethernet/amazon/ena/ena_netdev.c
3443
if (unlikely(!READ_ONCE(ena_napi->first_interrupt) && is_tx_comp_time_expired)) {
drivers/net/ethernet/amazon/ena/ena_xdp.c
135
if (!READ_ONCE(adapter->rx_ring->xdp_bpf_prog))
drivers/net/ethernet/amazon/ena/ena_xdp.h
90
xdp_prog = READ_ONCE(rx_ring->xdp_bpf_prog);
drivers/net/ethernet/aquantia/atlantic/aq_main.c
136
prog = READ_ONCE(aq_nic->xdp_prog);
drivers/net/ethernet/aquantia/atlantic/aq_main.c
228
prog = READ_ONCE(aq_nic->xdp_prog);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
190
if (READ_ONCE(self->xdp_prog)) {
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
422
prog = READ_ONCE(rx_ring->xdp_prog);
drivers/net/ethernet/atheros/ag71xx.c
749
timestamp = READ_ONCE(netdev_get_tx_queue(ag->ndev, 0)->trans_start);
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
355
while (READ_ONCE(token->state) < BNGE_HWRM_COMPLETE &&
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
367
if (READ_ONCE(token->state) != BNGE_HWRM_COMPLETE) {
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
372
len = le16_to_cpu(READ_ONCE(ctx->resp->resp_len));
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
381
READ_ONCE(token->state) == BNGE_HWRM_DEFERRED) {
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
386
len = le16_to_cpu(READ_ONCE(ctx->resp->resp_len));
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
388
__le16 resp_seq = READ_ONCE(ctx->resp->seq_id);
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1094
(READ_ONCE(txr->dev_state) ==
drivers/net/ethernet/broadcom/bnge/bnge_txrx.h
13
u32 used = READ_ONCE(txr->tx_prod) - READ_ONCE(txr->tx_cons);
drivers/net/ethernet/broadcom/bnx2.c
253
diff = READ_ONCE(txr->tx_prod) - READ_ONCE(txr->tx_cons);
drivers/net/ethernet/broadcom/bnx2.c
2834
cons = READ_ONCE(*bnapi->hw_tx_cons_ptr);
drivers/net/ethernet/broadcom/bnx2.c
3133
cons = READ_ONCE(*bnapi->hw_rx_cons_ptr);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
10414
u64 hw = le64_to_cpu(READ_ONCE(hw_stats[i]));
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15439
if (READ_ONCE(bp->xdp_prog))
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16552
if (sh && READ_ONCE(bp->xdp_prog)) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
884
READ_ONCE(txr->dev_state) == BNXT_DEV_STATE_CLOSING);
drivers/net/ethernet/broadcom/bnxt/bnxt.h
2832
u32 used = READ_ONCE(txr->tx_prod) - READ_ONCE(txr->tx_cons);
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
562
while (READ_ONCE(token->state) < BNXT_HWRM_COMPLETE &&
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
584
if (READ_ONCE(token->state) != BNXT_HWRM_COMPLETE) {
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
589
len = le16_to_cpu(READ_ONCE(ctx->resp->resp_len));
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
604
READ_ONCE(token->state) == BNXT_HWRM_DEFERRED) {
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
609
len = le16_to_cpu(READ_ONCE(ctx->resp->resp_len));
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
611
__le16 resp_seq = READ_ONCE(ctx->resp->seq_id);
drivers/net/ethernet/broadcom/bnxt/bnxt_ptp.c
760
num_requests = BNXT_MAX_TX_TS - READ_ONCE(ptp->tx_avail);
drivers/net/ethernet/broadcom/bnxt/bnxt_ptp.h
195
time = (u64)READ_ONCE(ptp->old_time) << BNXT_HI_TIMER_SHIFT;
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
177
struct bpf_prog *xdp_prog = READ_ONCE(rxr->xdp_prog);
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
228
struct bpf_prog *xdp_prog = READ_ONCE(rxr->xdp_prog);
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
334
struct bpf_prog *xdp_prog = READ_ONCE(bp->xdp_prog);
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
350
if (READ_ONCE(txr->dev_state) == BNXT_DEV_STATE_CLOSING)
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
201
while ((READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) &&
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
202
!(READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_QUIET) &&
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
213
WRITE_ONCE(reg_val, READ_ONCE(reg_val) &
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
216
READ_ONCE(reg_val));
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
220
if (READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) {
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
713
WRITE_ONCE(d64, READ_ONCE(d64) &
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
715
WRITE_ONCE(d64, READ_ONCE(d64) | CN23XX_PKT_INPUT_CTL_RST);
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
717
READ_ONCE(d64));
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
724
while (!(READ_ONCE(d64) & BIT_ULL(q_no)) && loop--) {
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
748
while (!(READ_ONCE(d64) & BIT_ULL(q_no)) && loop--) {
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
767
READ_ONCE(d32));
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c
881
lio_pci_writeq(oct, (READ_ONCE(bar1) & 0xFFFFFFFEULL),
drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
70
while ((READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) &&
drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
71
!(READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_QUIET) &&
drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
83
WRITE_ONCE(reg_val, READ_ONCE(reg_val) &
drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
86
READ_ONCE(reg_val));
drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
90
if (READ_ONCE(reg_val) & CN23XX_PKT_INPUT_CTL_RST) {
drivers/net/ethernet/cavium/liquidio/lio_main.c
3998
cores_crashed = READ_ONCE(oct->cores_crashed);
drivers/net/ethernet/cavium/liquidio/lio_main.c
800
vfs_mask1 = READ_ONCE(oct->sriov_info.vf_drv_loaded_mask);
drivers/net/ethernet/cavium/liquidio/lio_main.c
801
vfs_mask2 = READ_ONCE(other_oct->sriov_info.vf_drv_loaded_mask);
drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c
93
err = READ_ONCE(rep_resp->status) ? -EBUSY : 0;
drivers/net/ethernet/cavium/liquidio/request_manager.c
788
if (READ_ONCE(sc->caller_is_done)) {
drivers/net/ethernet/cavium/liquidio/response_manager.c
153
if (unlikely(READ_ONCE(sc->caller_is_done))) {
drivers/net/ethernet/chelsio/cxgb4/sge.c
1417
hw_cidx = ntohs(READ_ONCE(q->stat->cidx));
drivers/net/ethernet/chelsio/cxgb4/sge.c
2097
int hw_cidx = ntohs(READ_ONCE(q->stat->cidx));
drivers/net/ethernet/chelsio/cxgb4/sge.c
343
int hw_cidx = ntohs(READ_ONCE(q->stat->cidx));
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
1240
READ_ONCE(sock_net(newsk)->
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
1388
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_window_scaling)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
1396
ecn_ok = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.h
93
#define USER_MSS(tp) (READ_ONCE((tp)->rx_opt.user_mss))
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1380
if (copied >= target && !READ_ONCE(sk->sk_backlog.tail))
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1413
if (READ_ONCE(sk->sk_backlog.tail)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1587
if (READ_ONCE(sk->sk_backlog.tail)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1719
if (copied >= target && !READ_ONCE(sk->sk_backlog.tail))
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1750
if (READ_ONCE(sk->sk_backlog.tail)) {
drivers/net/ethernet/engleder/tsnep_main.c
1451
prog = READ_ONCE(rx->adapter->xdp_prog);
drivers/net/ethernet/engleder/tsnep_main.c
1554
prog = READ_ONCE(rx->adapter->xdp_prog);
drivers/net/ethernet/engleder/tsnep_main.c
375
while (READ_ONCE(tx->read) != tx->write) {
drivers/net/ethernet/engleder/tsnep_main.c
863
READ_ONCE(entry->skb->sk->sk_tsflags) & SOF_TIMESTAMPING_BIND_PHC)
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2601
xdp_prog = READ_ONCE(priv->xdp_prog);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2792
if (READ_ONCE(priv->xdp_prog)) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
436
xdp_prog = READ_ONCE(ch->xdp.prog);
drivers/net/ethernet/freescale/dpaa2/dpaa2-xsk.c
44
xdp_prog = READ_ONCE(ch->xdp.prog);
drivers/net/ethernet/freescale/fec_main.c
1582
while (bdp != READ_ONCE(txq->bd.cur)) {
drivers/net/ethernet/freescale/fec_main.c
1585
status = fec16_to_cpu(READ_ONCE(bdp->cbd_sc));
drivers/net/ethernet/freescale/fec_main.c
2455
struct bpf_prog *prog = READ_ONCE(fep->xdp_prog);
drivers/net/ethernet/freescale/fec_main.c
473
struct bpf_prog *xdp_prog = READ_ONCE(fep->xdp_prog);
drivers/net/ethernet/fungible/funeth/funeth_rx.c
155
xdp_prog = READ_ONCE(q->xdp_prog);
drivers/net/ethernet/fungible/funeth/funeth_tx.c
478
db_val = READ_ONCE(q->irq_db_val) | (q->cons_cnt & q->mask);
drivers/net/ethernet/google/gve/gve_adminq.c
483
status = be32_to_cpu(READ_ONCE(cmd->status));
drivers/net/ethernet/google/gve/gve_adminq.c
528
opcode = be32_to_cpu(READ_ONCE(cmd->opcode));
drivers/net/ethernet/google/gve/gve_ethtool.c
715
ori_flags = READ_ONCE(priv->ethtool_flags);
drivers/net/ethernet/google/gve/gve_main.c
1545
old_prog = READ_ONCE(priv->xdp_prog);
drivers/net/ethernet/google/gve/gve_rx.c
858
xprog = READ_ONCE(priv->xdp_prog);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
465
u64 last_read = READ_ONCE(gve->last_sync_nic_counter);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
812
xprog = READ_ONCE(priv->xdp_prog);
drivers/net/ethernet/google/gve/gve_tx.c
942
__be32 counter = READ_ONCE(priv->counter_array[counter_index]);
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.h
32
u32 len = READ_ONCE(ring->len);
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.h
37
return (READ_ONCE(ring->ntu) + len - READ_ONCE(ring->ntc)) % len;
drivers/net/ethernet/hisilicon/hip04_eth.c
456
count = tx_count(READ_ONCE(priv->tx_head), tx_tail);
drivers/net/ethernet/hisilicon/hip04_eth.c
515
count = tx_count(tx_head, READ_ONCE(priv->tx_tail));
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_cmd.h
468
u8 __iomem *reg_addr = READ_ONCE(base);
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
778
READ_ONCE(ring->page_pool->pages_state_hold_cnt),
drivers/net/ethernet/hisilicon/hns3/hns3_enet.h
632
int end = READ_ONCE(ring->next_to_use);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.h
650
u8 __iomem *reg_addr = READ_ONCE(base);
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.h
54
return READ_ONCE(*sq->cons_idx_addr) & wq->idx_mask;
drivers/net/ethernet/huawei/hinic3/hinic3_wq.h
49
return READ_ONCE(wq->prod_idx) - READ_ONCE(wq->cons_idx);
drivers/net/ethernet/intel/e1000/e1000.h
191
unsigned int use = READ_ONCE((R)->next_to_use); \
drivers/net/ethernet/intel/fm10k/fm10k_common.h
20
u32 __iomem *hw_addr = READ_ONCE((hw)->hw_addr); \
drivers/net/ethernet/intel/fm10k/fm10k_common.h
28
u32 __iomem *sw_addr = READ_ONCE((hw)->sw_addr); \
drivers/net/ethernet/intel/fm10k/fm10k_iov.c
160
if (!READ_ONCE(interface->iov_data))
drivers/net/ethernet/intel/fm10k/fm10k_iov.c
204
if (!READ_ONCE(interface->iov_data))
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1387
itr_round = READ_ONCE(ring_container->itr_scale) + 8;
drivers/net/ethernet/intel/fm10k/fm10k_netdev.c
1226
ring = READ_ONCE(interface->rx_ring[i]);
drivers/net/ethernet/intel/fm10k/fm10k_netdev.c
1242
ring = READ_ONCE(interface->tx_ring[i]);
drivers/net/ethernet/intel/fm10k/fm10k_netdev.c
1447
struct fm10k_l2_accel *l2_accel = READ_ONCE(interface->l2_accel);
drivers/net/ethernet/intel/fm10k/fm10k_netdev.c
513
int num_tx_queues = READ_ONCE(interface->num_tx_queues);
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
358
hw_addr = READ_ONCE(interface->uc_addr);
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
50
u32 __iomem *hw_addr = READ_ONCE(hw->hw_addr);
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
557
struct fm10k_ring *tx_ring = READ_ONCE(interface->tx_ring[i]);
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
579
struct fm10k_ring *rx_ring = READ_ONCE(interface->rx_ring[i]);
drivers/net/ethernet/intel/i40e/i40e.h
1331
return !!READ_ONCE(vsi->xdp_prog);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
190
struct i40e_ring *rx_ring = READ_ONCE(vsi->rx_rings[i]);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
232
struct i40e_ring *tx_ring = READ_ONCE(vsi->tx_rings[i]);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
272
struct i40e_ring *xdp_ring = READ_ONCE(vsi->xdp_rings[i]);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2410
i40e_add_queue_stats(&data, READ_ONCE(vsi->tx_rings[i]));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2411
i40e_add_queue_stats(&data, READ_ONCE(vsi->rx_rings[i]));
drivers/net/ethernet/intel/i40e/i40e_main.c
488
ring = READ_ONCE(vsi->tx_rings[i]);
drivers/net/ethernet/intel/i40e/i40e_main.c
494
ring = READ_ONCE(vsi->xdp_rings[i]);
drivers/net/ethernet/intel/i40e/i40e_main.c
500
ring = READ_ONCE(vsi->rx_rings[i]);
drivers/net/ethernet/intel/i40e/i40e_main.c
913
p = READ_ONCE(vsi->tx_rings[q]);
drivers/net/ethernet/intel/i40e/i40e_main.c
931
p = READ_ONCE(vsi->rx_rings[q]);
drivers/net/ethernet/intel/i40e/i40e_main.c
951
p = READ_ONCE(vsi->xdp_rings[q]);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
354
base_adj = I40E_PTP_40GB_INCVAL * READ_ONCE(pf->ptp_adj_mult);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2450
xdp_prog = READ_ONCE(rx_ring->xdp_prog);
drivers/net/ethernet/intel/i40e/i40e_xsk.c
424
xdp_prog = READ_ONCE(rx_ring->xdp_prog);
drivers/net/ethernet/intel/ice/ice.h
751
return !!READ_ONCE(vsi->xdp_prog);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1959
tx_ring = READ_ONCE(vsi->tx_rings[j]);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1975
rx_ring = READ_ONCE(vsi->rx_rings[j]);
drivers/net/ethernet/intel/ice/ice_main.c
2925
if (READ_ONCE(rx_ring->xsk_pool))
drivers/net/ethernet/intel/ice/ice_main.c
6916
ring = READ_ONCE(rings[i]);
drivers/net/ethernet/intel/ice/ice_main.c
6947
ring = READ_ONCE(rings[i]);
drivers/net/ethernet/intel/ice/ice_ptp.c
2311
cached_time = READ_ONCE(pkt_ctx->cached_phctime);
drivers/net/ethernet/intel/ice/ice_txrx.c
1280
struct xsk_buff_pool *xsk_pool = READ_ONCE(tx_ring->xsk_pool);
drivers/net/ethernet/intel/ice/ice_txrx.c
1310
struct xsk_buff_pool *xsk_pool = READ_ONCE(rx_ring->xsk_pool);
drivers/net/ethernet/intel/ice/ice_txrx.c
959
xdp_prog = READ_ONCE(rx_ring->xdp_prog);
drivers/net/ethernet/intel/ice/ice_xsk.c
585
xdp_prog = READ_ONCE(rx_ring->xdp_prog);
drivers/net/ethernet/intel/ice/ice_xsk.c
859
if (!READ_ONCE(ring->xsk_pool))
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3377
cached_time = READ_ONCE(rxq->cached_phc_time);
drivers/net/ethernet/intel/idpf/xdp.c
424
cached_time = READ_ONCE(rxq->cached_phc_time);
drivers/net/ethernet/intel/igb/e1000_regs.h
357
u8 __iomem *hw_addr = READ_ONCE((hw)->hw_addr); \
drivers/net/ethernet/intel/igb/igb.h
846
return !!READ_ONCE(adapter->xdp_prog);
drivers/net/ethernet/intel/igb/igb_main.c
711
u8 __iomem *hw_addr = READ_ONCE(hw->hw_addr);
drivers/net/ethernet/intel/igb/igb_main.c
8298
xsk_pool = READ_ONCE(q_vector->rx.ring->xsk_pool);
drivers/net/ethernet/intel/igb/igb_main.c
8440
xsk_pool = READ_ONCE(tx_ring->xsk_pool);
drivers/net/ethernet/intel/igb/igb_main.c
8715
xdp_prog = READ_ONCE(rx_ring->xdp_prog);
drivers/net/ethernet/intel/igb/igb_xsk.c
355
xdp_prog = READ_ONCE(rx_ring->xdp_prog);
drivers/net/ethernet/intel/igb/igb_xsk.c
555
if (!READ_ONCE(ring->xsk_pool))
drivers/net/ethernet/intel/igc/igc_main.c
1677
READ_ONCE(skb->sk->sk_tsflags) & SOF_TIMESTAMPING_BIND_PHC)
drivers/net/ethernet/intel/igc/igc_main.c
2549
prog = READ_ONCE(adapter->xdp_prog);
drivers/net/ethernet/intel/igc/igc_main.c
2814
prog = READ_ONCE(adapter->xdp_prog);
drivers/net/ethernet/intel/igc/igc_main.c
7006
u8 __iomem *hw_addr = READ_ONCE(hw->hw_addr);
drivers/net/ethernet/intel/igc/igc_regs.h
351
u8 __iomem *hw_addr = READ_ONCE((hw)->hw_addr); \
drivers/net/ethernet/intel/ixgbe/ixgbe_common.h
145
u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr);
drivers/net/ethernet/intel/ixgbe/ixgbe_common.h
164
u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2410
xdp_prog = READ_ONCE(rx_ring->xdp_prog);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
314
reg_addr = READ_ONCE(hw->hw_addr);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
351
u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
7682
struct ixgbe_ring *rx_ring = READ_ONCE(adapter->rx_ring[i]);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
7706
struct ixgbe_ring *tx_ring = READ_ONCE(adapter->tx_ring[i]);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
7716
struct ixgbe_ring *xdp_ring = READ_ONCE(adapter->xdp_ring[i]);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
9698
struct ixgbe_ring *ring = READ_ONCE(adapter->rx_ring[i]);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
9714
struct ixgbe_ring *ring = READ_ONCE(adapter->tx_ring[i]);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
9719
struct ixgbe_ring *ring = READ_ONCE(adapter->xdp_ring[i]);
drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c
459
incval = READ_ONCE(adapter->base_incval);
drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c
107
xdp_prog = READ_ONCE(rx_ring->xdp_prog);
drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c
14
bool xdp_on = READ_ONCE(adapter->xdp_prog);
drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c
519
if (!READ_ONCE(adapter->xdp_prog))
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1057
xdp_prog = READ_ONCE(rx_ring->xdp_prog);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
152
u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr);
drivers/net/ethernet/intel/ixgbevf/vf.h
162
u8 __iomem *reg_addr = READ_ONCE(hw->hw_addr);
drivers/net/ethernet/intel/libeth/rx.c
126
mtu = READ_ONCE(pp->netdev->mtu);
drivers/net/ethernet/intel/libeth/rx.c
25
len = READ_ONCE(pp->netdev->mtu) + LIBETH_RX_LL_LEN;
drivers/net/ethernet/marvell/mvneta.c
2444
xdp_prog = READ_ONCE(pp->xdp_prog);
drivers/net/ethernet/marvell/mvneta.c
3359
struct bpf_prog *xdp_prog = READ_ONCE(pp->xdp_prog);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3907
xdp_prog = READ_ONCE(port->xdp_prog);
drivers/net/ethernet/marvell/mvpp2/mvpp2_tai.c
355
ts.tv_sec = READ_ONCE(tai->stamp.tv_sec);
drivers/net/ethernet/marvell/octeon_ep/octep_main.c
564
u32 pkts_pend = READ_ONCE(oq->pkts_pending);
drivers/net/ethernet/marvell/octeon_ep/octep_main.c
565
u32 last_pkt_count = READ_ONCE(oq->last_pkt_count);
drivers/net/ethernet/marvell/octeon_ep/octep_main.c
566
u32 pkts_processed = READ_ONCE(iq->pkts_processed);
drivers/net/ethernet/marvell/octeon_ep/octep_main.c
567
u32 pkt_in_done = READ_ONCE(iq->pkt_in_done);
drivers/net/ethernet/marvell/octeon_ep/octep_rx.c
330
last_pkt_count = READ_ONCE(oq->last_pkt_count);
drivers/net/ethernet/marvell/octeon_ep/octep_rx.c
348
pkts_pending = READ_ONCE(oq->pkts_pending);
drivers/net/ethernet/marvell/octeon_ep/octep_rx.c
424
read_idx = READ_ONCE(oq->host_read_idx);
drivers/net/ethernet/marvell/octeon_ep/octep_rx.c
539
pkts_pending = READ_ONCE(oq->pkts_pending);
drivers/net/ethernet/marvell/octeon_ep/octep_rx.c
542
pkts_pending = READ_ONCE(oq->pkts_pending);
drivers/net/ethernet/marvell/octeon_ep/octep_rx.c
550
pkts_pending = READ_ONCE(oq->pkts_pending);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_main.c
297
u32 pkts_pend = READ_ONCE(oq->pkts_pending);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_main.c
298
u32 last_pkt_count = READ_ONCE(oq->last_pkt_count);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_main.c
299
u32 pkts_processed = READ_ONCE(iq->pkts_processed);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_main.c
300
u32 pkt_in_done = READ_ONCE(iq->pkt_in_done);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_rx.c
331
last_pkt_count = READ_ONCE(oq->last_pkt_count);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_rx.c
350
pkts_pending = READ_ONCE(oq->pkts_pending);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_rx.c
380
read_idx = READ_ONCE(oq->host_read_idx);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_rx.c
504
pkts_pending = READ_ONCE(oq->pkts_pending);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_rx.c
507
pkts_pending = READ_ONCE(oq->pkts_pending);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_rx.c
515
pkts_pending = READ_ONCE(oq->pkts_pending);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
414
} while ((READ_ONCE(*completion_ptr) & CN10K_CPT_COMP_E_MASK) ==
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2266
classid = READ_ONCE(pf->qos.defcls);
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
603
qid = READ_ONCE(node->qid);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1245
rxd->rxd2 = READ_ONCE(dma_rxd->rxd2);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1249
rxd->rxd1 = READ_ONCE(dma_rxd->rxd1);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1250
rxd->rxd3 = READ_ONCE(dma_rxd->rxd3);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1251
rxd->rxd4 = READ_ONCE(dma_rxd->rxd4);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1253
rxd->rxd5 = READ_ONCE(dma_rxd->rxd5);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1254
rxd->rxd6 = READ_ONCE(dma_rxd->rxd6);
drivers/net/ethernet/mediatek/mtk_ppe.c
569
ib1 = READ_ONCE(hwe->ib1);
drivers/net/ethernet/mediatek/mtk_star_emac.c
316
status = READ_ONCE(desc->status);
drivers/net/ethernet/mediatek/mtk_star_emac.c
347
status = READ_ONCE(desc->status);
drivers/net/ethernet/mediatek/mtk_wed_wo.c
124
le32_to_cpu(READ_ONCE(desc->ctrl)));
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1469
pkt_rate_low = READ_ONCE(priv->pkt_rate_low);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1470
pkt_rate_high = READ_ONCE(priv->pkt_rate_high);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1473
rx_packets = READ_ONCE(priv->rx_ring[ring]->packets);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1474
rx_bytes = READ_ONCE(priv->rx_ring[ring]->bytes);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3102
stats->packets = READ_ONCE(ring->packets);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3103
stats->bytes = READ_ONCE(ring->bytes);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3104
stats->alloc_fail = READ_ONCE(ring->alloc_fail);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3122
stats->packets = READ_ONCE(ring->packets);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3123
stats->bytes = READ_ONCE(ring->bytes);
drivers/net/ethernet/mellanox/mlx4/en_port.c
165
packets += READ_ONCE(ring->packets);
drivers/net/ethernet/mellanox/mlx4/en_port.c
166
bytes += READ_ONCE(ring->bytes);
drivers/net/ethernet/mellanox/mlx4/en_port.c
176
packets += READ_ONCE(ring->packets);
drivers/net/ethernet/mellanox/mlx4/en_port.c
177
bytes += READ_ONCE(ring->bytes);
drivers/net/ethernet/mellanox/mlx4/en_port.c
254
sw_rx_dropped += READ_ONCE(ring->dropped);
drivers/net/ethernet/mellanox/mlx4/en_port.c
255
priv->port_stats.rx_chksum_good += READ_ONCE(ring->csum_ok);
drivers/net/ethernet/mellanox/mlx4/en_port.c
256
priv->port_stats.rx_chksum_none += READ_ONCE(ring->csum_none);
drivers/net/ethernet/mellanox/mlx4/en_port.c
257
priv->port_stats.rx_chksum_complete += READ_ONCE(ring->csum_complete);
drivers/net/ethernet/mellanox/mlx4/en_port.c
258
priv->port_stats.rx_alloc_pages += READ_ONCE(ring->rx_alloc_pages);
drivers/net/ethernet/mellanox/mlx4/en_port.c
259
priv->xdp_stats.rx_xdp_drop += READ_ONCE(ring->xdp_drop);
drivers/net/ethernet/mellanox/mlx4/en_port.c
260
priv->xdp_stats.rx_xdp_redirect += READ_ONCE(ring->xdp_redirect);
drivers/net/ethernet/mellanox/mlx4/en_port.c
261
priv->xdp_stats.rx_xdp_redirect_fail += READ_ONCE(ring->xdp_redirect_fail);
drivers/net/ethernet/mellanox/mlx4/en_port.c
262
priv->xdp_stats.rx_xdp_tx += READ_ONCE(ring->xdp_tx);
drivers/net/ethernet/mellanox/mlx4/en_port.c
263
priv->xdp_stats.rx_xdp_tx_full += READ_ONCE(ring->xdp_tx_full);
drivers/net/ethernet/mellanox/mlx4/en_port.c
274
sw_tx_dropped += READ_ONCE(ring->tx_dropped);
drivers/net/ethernet/mellanox/mlx4/en_port.c
275
priv->port_stats.tx_chksum_offload += READ_ONCE(ring->tx_csum);
drivers/net/ethernet/mellanox/mlx4/en_port.c
276
priv->port_stats.queue_stopped += READ_ONCE(ring->queue_stopped);
drivers/net/ethernet/mellanox/mlx4/en_port.c
277
priv->port_stats.wake_queue += READ_ONCE(ring->wake_queue);
drivers/net/ethernet/mellanox/mlx4/en_port.c
278
priv->port_stats.tso_packets += READ_ONCE(ring->tso_packets);
drivers/net/ethernet/mellanox/mlx4/en_port.c
279
priv->port_stats.xmit_more += READ_ONCE(ring->xmit_more);
drivers/net/ethernet/mellanox/mlx4/en_tx.c
232
u32 used = READ_ONCE(ring->prod) - READ_ONCE(ring->cons);
drivers/net/ethernet/mellanox/mlx4/en_tx.c
455
last_nr_txbb = READ_ONCE(ring->last_nr_txbb);
drivers/net/ethernet/mellanox/mlx4/en_tx.c
456
ring_cons = READ_ONCE(ring->cons);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
292
own = READ_ONCE(ent->lay->status_own);
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
160
qid = READ_ONCE(node->qid);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
148
same_dev = READ_ONCE(nhe->neigh_dev) == n->dev;
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
250
if (p->dev == READ_ONCE(nhe->neigh_dev)) {
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c
561
jiffies_to_usecs(jiffies - READ_ONCE(sq->txq->trans_start)));
drivers/net/ethernet/mellanox/mlx5/core/en/selq.c
157
if (READ_ONCE(priv->dcbx_dp.trust_state) == MLX5_QPTS_TRUST_DSCP)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
455
n = neigh_lookup(tbl, &m_neigh->dst_ip, READ_ONCE(nhe->neigh_dev));
drivers/net/ethernet/mellanox/mlx5/core/en_stats.c
2490
stats = READ_ONCE(priv->htb_qos_sq_stats);
drivers/net/ethernet/mellanox/mlx5/core/en_stats.c
2493
struct mlx5e_sq_stats *s = READ_ONCE(stats[qid]);
drivers/net/ethernet/mellanox/mlx5/core/en_stats.c
483
stats = READ_ONCE(priv->htb_qos_sq_stats);
drivers/net/ethernet/mellanox/mlx5/core/en_stats.c
486
mlx5e_stats_grp_sw_update_stats_sq(s, READ_ONCE(stats[i]));
drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c
155
qos_sqs_size = READ_ONCE(c->qos_sqs_size);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
343
return READ_ONCE(devcom->comp->ready);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
355
if (!READ_ONCE(comp->ready)) {
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
406
if (!READ_ONCE(comp->ready))
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
11268
switch (READ_ONCE(net->ipv4.sysctl_fib_multipath_hash_policy)) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
11286
hash_fields = READ_ONCE(net->ipv4.sysctl_fib_multipath_hash_fields);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
11407
seed = READ_ONCE(net->ipv4.sysctl_fib_multipath_hash_seed).user_seed;
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
11490
usp = READ_ONCE(net->ipv4.sysctl_ip_fwd_update_priority);
drivers/net/ethernet/meta/fbnic/fbnic.h
116
return !!READ_ONCE(fbd->uc_addr0);
drivers/net/ethernet/meta/fbnic/fbnic.h
121
u32 __iomem *csr = READ_ONCE(fbd->uc_addr0);
drivers/net/ethernet/meta/fbnic/fbnic_pci.c
39
u32 __iomem *csr = READ_ONCE(fbd->uc_addr0);
drivers/net/ethernet/meta/fbnic/fbnic_pci.c
72
return !!READ_ONCE(fbd->uc_addr4);
drivers/net/ethernet/meta/fbnic/fbnic_pci.c
77
u32 __iomem *csr = READ_ONCE(fbd->uc_addr4);
drivers/net/ethernet/meta/fbnic/fbnic_pci.c
85
u32 __iomem *csr = READ_ONCE(fbd->uc_addr4);
drivers/net/ethernet/meta/fbnic/fbnic_phylink.c
285
if (!time_before(READ_ONCE(fbd->end_of_pmd_training), jiffies))
drivers/net/ethernet/meta/fbnic/fbnic_phylink.c
299
if (!time_before(READ_ONCE(fbd->end_of_pmd_training), jiffies))
drivers/net/ethernet/meta/fbnic/fbnic_time.c
138
WRITE_ONCE(fbn->time_offset, READ_ONCE(fbn->time_offset) + delta);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1154
xdp_prog = READ_ONCE(fbn->xdp_prog);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
85
offset = READ_ONCE(fbn->time_offset);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
88
high = READ_ONCE(fbn->time_high);
drivers/net/ethernet/microsoft/mana/mana_en.c
116
link_event = READ_ONCE(ac->link_event);
drivers/net/ethernet/microsoft/mana/mana_en.c
3488
if (READ_ONCE(gd->rdma_teardown))
drivers/net/ethernet/netronome/nfp/flower/action.c
478
set_tun->ttl = READ_ONCE(net->ipv6.devconf_all->hop_limit);
drivers/net/ethernet/netronome/nfp/flower/action.c
499
set_tun->ttl = READ_ONCE(net->ipv4.sysctl_ip_default_ttl);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
918
xdp_prog = READ_ONCE(dp->xdp_prog);
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
217
skb = READ_ONCE(tx_ring->txbufs[i].skb);
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
222
xdp = READ_ONCE(tx_ring->txbufs[i].xdp);
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
125
xdp_prog = READ_ONCE(dp->xdp_prog);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1040
xdp_prog = READ_ONCE(dp->xdp_prog);
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
47
frag = READ_ONCE(rx_ring->rxbufs[i].frag);
drivers/net/ethernet/pensando/ionic/ionic_dev.c
63
now = READ_ONCE(jiffies);
drivers/net/ethernet/pensando/ionic/ionic_lif.c
1808
xdp_prog = READ_ONCE(lif->xdp_prog);
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2200
READ_ONCE(lif->rxqcqs[i]->q.xdp_prog));
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2738
xdp_prog = READ_ONCE(lif->xdp_prog);
drivers/net/ethernet/pensando/ionic/ionic_lif.c
473
adminqcq = READ_ONCE(lif->adminqcq);
drivers/net/ethernet/pensando/ionic/ionic_main.c
354
now = READ_ONCE(jiffies);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
60
now = READ_ONCE(jiffies);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
85
now = READ_ONCE(jiffies);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
987
xdp_prog = READ_ONCE(q->xdp_prog);
drivers/net/ethernet/qlogic/qed/qed_iwarp.c
522
while (READ_ONCE(ep->state) != QED_IWARP_EP_CLOSED &&
drivers/net/ethernet/qlogic/qede/qede_fp.c
1248
struct bpf_prog *xdp_prog = READ_ONCE(rxq->xdp_prog);
drivers/net/ethernet/qualcomm/rmnet/rmnet_handlers.c
167
if (READ_ONCE(port->egress_agg_params.count) > 1) {
drivers/net/ethernet/qualcomm/rmnet/rmnet_vnd.c
101
return READ_ONCE(priv->real_dev->ifindex);
drivers/net/ethernet/realtek/r8169_main.c
2215
unsigned int timer_val = READ_ONCE(tp->dev->mtu) + ETH_HLEN + 0x20;
drivers/net/ethernet/realtek/r8169_main.c
4529
return READ_ONCE(tp->dirty_tx) + NUM_TX_DESC - READ_ONCE(tp->cur_tx);
drivers/net/ethernet/realtek/r8169_main.c
4710
while (READ_ONCE(tp->cur_tx) != dirty_tx) {
drivers/net/ethernet/realtek/r8169_main.c
4714
status = le32_to_cpu(READ_ONCE(tp->TxDescArray[entry].opts1));
drivers/net/ethernet/realtek/r8169_main.c
4744
if (READ_ONCE(tp->cur_tx) != dirty_tx && skb)
drivers/net/ethernet/realtek/r8169_main.c
4777
status = le32_to_cpu(READ_ONCE(desc->opts1));
drivers/net/ethernet/realtek/rtase/rtase_main.c
258
return READ_ONCE(ring->dirty_idx) + RTASE_NUM_DESC -
drivers/net/ethernet/realtek/rtase/rtase_main.c
259
READ_ONCE(ring->cur_idx);
drivers/net/ethernet/realtek/rtase/rtase_main.c
272
tx_left = READ_ONCE(ring->cur_idx) - dirty_tx;
drivers/net/ethernet/sfc/ef10.c
2146
if (likely(READ_ONCE(efx->irq_soft_enabled))) {
drivers/net/ethernet/sfc/ef10.c
2161
bool soft_enabled = READ_ONCE(efx->irq_soft_enabled);
drivers/net/ethernet/sfc/ef10.c
2821
if (unlikely(READ_ONCE(efx->reset_pending)))
drivers/net/ethernet/sfc/ef10.c
2971
if (unlikely(READ_ONCE(efx->reset_pending)))
drivers/net/ethernet/sfc/ef100_nic.c
336
if (likely(READ_ONCE(efx->irq_soft_enabled))) {
drivers/net/ethernet/sfc/ef100_rep.c
441
if (efv->write_index - READ_ONCE(efv->read_index) > efv->rx_pring_size) {
drivers/net/ethernet/sfc/ef100_tx.c
441
txq2->old_read_count = READ_ONCE(txq2->read_count);
drivers/net/ethernet/sfc/ef100_tx.c
478
txq2->old_read_count = READ_ONCE(txq2->read_count);
drivers/net/ethernet/sfc/efx_common.c
861
pending = READ_ONCE(efx->reset_pending);
drivers/net/ethernet/sfc/efx_common.c
926
if (!efx_net_active(READ_ONCE(efx->state)))
drivers/net/ethernet/sfc/falcon/efx.c
2496
pending = READ_ONCE(efx->reset_pending);
drivers/net/ethernet/sfc/falcon/efx.c
2556
if (READ_ONCE(efx->state) != STATE_READY)
drivers/net/ethernet/sfc/falcon/falcon.c
1372
isolate = !!READ_ONCE(efx->reset_pending);
drivers/net/ethernet/sfc/falcon/falcon.c
452
if (!likely(READ_ONCE(efx->irq_soft_enabled)))
drivers/net/ethernet/sfc/falcon/farch.c
1503
bool soft_enabled = READ_ONCE(efx->irq_soft_enabled);
drivers/net/ethernet/sfc/falcon/farch.c
1595
if (!likely(READ_ONCE(efx->irq_soft_enabled)))
drivers/net/ethernet/sfc/falcon/farch.c
834
if (unlikely(READ_ONCE(efx->reset_pending)))
drivers/net/ethernet/sfc/falcon/farch.c
989
if (unlikely(READ_ONCE(efx->reset_pending)))
drivers/net/ethernet/sfc/falcon/nic.h
466
return READ_ONCE(channel->event_test_cpu);
drivers/net/ethernet/sfc/falcon/nic.h
470
return READ_ONCE(efx->last_irq_cpu);
drivers/net/ethernet/sfc/falcon/nic.h
83
unsigned int empty_read_count = READ_ONCE(tx_queue->empty_read_count);
drivers/net/ethernet/sfc/falcon/tx.c
127
txq1->old_read_count = READ_ONCE(txq1->read_count);
drivers/net/ethernet/sfc/falcon/tx.c
128
txq2->old_read_count = READ_ONCE(txq2->read_count);
drivers/net/ethernet/sfc/falcon/tx.c
517
tx_queue->old_write_count = READ_ONCE(tx_queue->write_count);
drivers/net/ethernet/sfc/mae.c
212
credits = READ_ONCE(rx_queue->notified_count) - rx_queue->granted_count;
drivers/net/ethernet/sfc/mcdi_filters.c
662
if (READ_ONCE(table->entry[i].spec) &
drivers/net/ethernet/sfc/nic_common.h
226
return READ_ONCE(channel->event_test_cpu);
drivers/net/ethernet/sfc/nic_common.h
230
return READ_ONCE(efx->last_irq_cpu);
drivers/net/ethernet/sfc/nic_common.h
70
unsigned int empty_read_count = READ_ONCE(tx_queue->empty_read_count);
drivers/net/ethernet/sfc/ptp.c
1038
while (!READ_ONCE(*start) && (time_before(jiffies, timeout))) {
drivers/net/ethernet/sfc/ptp.c
1048
if (READ_ONCE(*start))
drivers/net/ethernet/sfc/ptp.c
857
READ_ONCE(*mc_running)) {
drivers/net/ethernet/sfc/ptp.c
867
READ_ONCE(*mc_running));
drivers/net/ethernet/sfc/siena/efx_common.c
885
pending = READ_ONCE(efx->reset_pending);
drivers/net/ethernet/sfc/siena/efx_common.c
950
if (READ_ONCE(efx->state) != STATE_READY)
drivers/net/ethernet/sfc/siena/farch.c
1517
bool soft_enabled = READ_ONCE(efx->irq_soft_enabled);
drivers/net/ethernet/sfc/siena/farch.c
1609
if (!likely(READ_ONCE(efx->irq_soft_enabled)))
drivers/net/ethernet/sfc/siena/farch.c
833
if (unlikely(READ_ONCE(efx->reset_pending)))
drivers/net/ethernet/sfc/siena/farch.c
982
if (unlikely(READ_ONCE(efx->reset_pending)))
drivers/net/ethernet/sfc/siena/nic_common.h
225
return READ_ONCE(channel->event_test_cpu);
drivers/net/ethernet/sfc/siena/nic_common.h
229
return READ_ONCE(efx->last_irq_cpu);
drivers/net/ethernet/sfc/siena/nic_common.h
70
unsigned int empty_read_count = READ_ONCE(tx_queue->empty_read_count);
drivers/net/ethernet/sfc/siena/ptp.c
1051
while (!READ_ONCE(*start) && (time_before(jiffies, timeout))) {
drivers/net/ethernet/sfc/siena/ptp.c
1061
if (READ_ONCE(*start))
drivers/net/ethernet/sfc/siena/ptp.c
870
READ_ONCE(*mc_running)) {
drivers/net/ethernet/sfc/siena/ptp.c
880
READ_ONCE(*mc_running));
drivers/net/ethernet/sfc/siena/tx.c
72
txq2->old_read_count = READ_ONCE(txq2->read_count);
drivers/net/ethernet/sfc/siena/tx_common.c
229
tx_queue->old_write_count = READ_ONCE(tx_queue->write_count);
drivers/net/ethernet/sfc/tc_counters.c
96
touched = READ_ONCE(cnt->touched);
drivers/net/ethernet/sfc/tx.c
80
txq2->old_read_count = READ_ONCE(txq2->read_count);
drivers/net/ethernet/sfc/tx_common.c
257
tx_queue->old_write_count = READ_ONCE(tx_queue->write_count);
drivers/net/ethernet/socionext/netsec.c
1297
struct bpf_prog *xdp_prog = READ_ONCE(priv->xdp_prog);
drivers/net/ethernet/socionext/netsec.c
965
xdp_prog = READ_ONCE(priv->xdp_prog);
drivers/net/ethernet/spacemit/k1_emac.c
492
if (READ_ONCE(tx_desc->desc0) & TX_DESC_0_OWN)
drivers/net/ethernet/spacemit/k1_emac.c
615
if (READ_ONCE(rx_desc->desc0) & RX_DESC_0_OWN)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5242
prog = READ_ONCE(priv->xdp_prog);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5534
prog = READ_ONCE(priv->xdp_prog);
drivers/net/ethernet/sun/niu.c
6244
rx_rings = READ_ONCE(np->rx_rings);
drivers/net/ethernet/sun/niu.c
6275
tx_rings = READ_ONCE(np->tx_rings);
drivers/net/ethernet/ti/am65-cpsw-nuss.c
1185
prog = READ_ONCE(port->xdp_prog);
drivers/net/ethernet/ti/am65-cpsw-nuss.c
419
trans_start = READ_ONCE(netif_txq->trans_start);
drivers/net/ethernet/ti/cpsw_new.c
1126
if (xdpf->len < READ_ONCE(priv->tx_packet_min))
drivers/net/ethernet/ti/cpsw_new.c
950
if (skb_put_padto(skb, READ_ONCE(priv->tx_packet_min))) {
drivers/net/ethernet/ti/cpsw_priv.c
1314
prog = READ_ONCE(priv->xdp_prog);
drivers/net/ethernet/ti/cpts.c
188
return READ_ONCE(cpts->cur_timestamp);
drivers/net/ethernet/ti/icssg/icssg_common.c
790
xdp_prog = READ_ONCE(emac->xdp_prog);
drivers/net/ethernet/ti/icssg/icssg_prueth.h
522
return !!READ_ONCE(emac->xdp_prog);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
739
skb->offload_fwd_mark = READ_ONCE(emac->offload_fwd_mark);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
3058
struct wx_ring *ring = READ_ONCE(wx->rx_ring[i]);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
3074
struct wx_ring *ring = READ_ONCE(wx->tx_ring[i]);
drivers/net/ethernet/wangxun/libwx/wx_ptp.c
71
incval = READ_ONCE(wx->base_incval);
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
1305
if (READ_ONCE(lp->rx_dim_enabled)) {
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
1311
.event_ctr = READ_ONCE(lp->rx_irqs),
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
1409
WRITE_ONCE(lp->rx_irqs, READ_ONCE(lp->rx_irqs) + 1);
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
2689
if (READ_ONCE(lp->stopping))
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
854
cur_p = &lp->tx_bd_v[(READ_ONCE(lp->tx_bd_tail) + num_frag) %
drivers/net/gtp.c
913
switch (READ_ONCE(udp_sk(sk)->encap_type)) {
drivers/net/ipvlan/ipvlan_main.c
324
rx_errs += READ_ONCE(pcptr->rx_errs);
drivers/net/ipvlan/ipvlan_main.c
325
tx_drps += READ_ONCE(pcptr->tx_drps);
drivers/net/ipvlan/ipvlan_main.c
356
return READ_ONCE(ipvlan->phy_dev->ifindex);
drivers/net/macsec.c
3738
return READ_ONCE(macsec_priv(dev)->real_dev->ifindex);
drivers/net/macvlan.c
1003
rx_errors += READ_ONCE(p->rx_errors);
drivers/net/macvlan.c
1004
tx_dropped += READ_ONCE(p->tx_dropped);
drivers/net/macvlan.c
1169
return READ_ONCE(vlan->lowerdev->ifindex);
drivers/net/mctp/mctp-usb.c
251
if (READ_ONCE(mctp_usb->stopped))
drivers/net/mhi_net.c
278
size = mhi_netdev->mru ? mhi_netdev->mru : READ_ONCE(ndev->mtu);
drivers/net/netdevsim/dev.c
236
READ_ONCE(nsim_dev->nsim_bus_dev->max_vfs));
drivers/net/netdevsim/netdev.c
406
iflink = peer ? READ_ONCE(peer->netdev->ifindex) :
drivers/net/netdevsim/netdev.c
407
READ_ONCE(dev->ifindex);
drivers/net/netdevsim/netdev.c
423
xdp_prog = READ_ONCE(ns->xdp.prog);
drivers/net/netkit.c
167
iflink = READ_ONCE(peer->ifindex);
drivers/net/netkit.c
83
enum netkit_action ret = READ_ONCE(nk->policy);
drivers/net/ovpn/crypto.h
83
idx = READ_ONCE(cs->primary_idx);
drivers/net/ovpn/peer.c
1228
last_recv = READ_ONCE(peer->last_recv);
drivers/net/ovpn/peer.c
1251
last_sent = READ_ONCE(peer->last_sent);
drivers/net/ovpn/socket.c
154
u8 type = READ_ONCE(udp_sk(sk)->encap_type);
drivers/net/ovpn/tcp.c
355
READ_ONCE(net_hotdata.max_backlog)) {
drivers/net/ovpn/udp.c
347
skb->mark = READ_ONCE(sk->sk_mark);
drivers/net/ovpn/udp.c
38
if (unlikely(READ_ONCE(udp_sk(sk)->encap_type) != UDP_ENCAP_OVPNINUDP))
drivers/net/ovpn/udp.c
416
if ((READ_ONCE(udp_sk(ovpn_sock->sk)->encap_type) == UDP_ENCAP_OVPNINUDP) &&
drivers/net/phy/phy.c
642
link_stats->link_down_events = READ_ONCE(phydev->link_down_events);
drivers/net/ppp/ppp_generic.c
1383
return READ_ONCE(ppp->ppp_net);
drivers/net/ppp/ppp_generic.c
1606
chan = READ_ONCE(pch->chan);
drivers/net/ppp/pppoe.c
884
skb->priority = READ_ONCE(sk->sk_priority);
drivers/net/tap.c
1060
vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz);
drivers/net/tap.c
198
int numvtaps = READ_ONCE(tap->numvtaps);
drivers/net/tap.c
591
vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz);
drivers/net/tap.c
716
vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz);
drivers/net/team/team_core.c
1838
rx_dropped += READ_ONCE(p->rx_dropped);
drivers/net/team/team_core.c
1839
tx_dropped += READ_ONCE(p->tx_dropped);
drivers/net/team/team_core.c
1840
rx_nohandler += READ_ONCE(p->rx_nohandler);
drivers/net/team/team_core.c
2089
port_ops = READ_ONCE(port->dev->header_ops);
drivers/net/team/team_core.c
2110
port_ops = READ_ONCE(port->dev->header_ops);
drivers/net/tun.c
1217
numqueues = READ_ONCE(tun->numqueues);
drivers/net/tun.c
1732
int vnet_hdr_sz = READ_ONCE(tun->vnet_hdr_sz);
drivers/net/tun.c
2017
vnet_hdr_sz = READ_ONCE(tun->vnet_hdr_sz);
drivers/net/tun.c
2049
vnet_hdr_sz = READ_ONCE(tun->vnet_hdr_sz);
drivers/net/tun.c
2453
features = tun_vnet_hdr_guest_features(READ_ONCE(tun->vnet_hdr_sz));
drivers/net/tun.c
426
if (READ_ONCE(e->queue_index) != queue_index)
drivers/net/tun.c
466
numqueues = READ_ONCE(tun->numqueues);
drivers/net/tun.c
486
numqueues = READ_ONCE(tun->numqueues);
drivers/net/usb/cdc_mbim.c
342
is_router = !!READ_ONCE(in6_dev->cnf.forwarding);
drivers/net/veth.c
1491
iflink = peer ? READ_ONCE(peer->ifindex) : 0;
drivers/net/veth.c
305
if (!READ_ONCE(rq->rx_notify_masked) &&
drivers/net/virtio_net.c
6159
jiffies_to_usecs(jiffies - READ_ONCE(txq->trans_start)));
drivers/net/vxlan/vxlan_core.c
1446
if (READ_ONCE(f->updated) != now)
drivers/net/vxlan/vxlan_core.c
1885
if (!(READ_ONCE(n->nud_state) & NUD_CONNECTED)) {
drivers/net/vxlan/vxlan_core.c
2057
if (!(READ_ONCE(n->nud_state) & NUD_CONNECTED)) {
drivers/net/vxlan/vxlan_core.c
236
ci.ndm_used = jiffies_to_clock_t(now - READ_ONCE(fdb->used));
drivers/net/vxlan/vxlan_core.c
238
ci.ndm_updated = jiffies_to_clock_t(now - READ_ONCE(fdb->updated));
drivers/net/vxlan/vxlan_core.c
2864
timeout = READ_ONCE(f->updated) + vxlan->cfg.age_interval * HZ;
drivers/net/vxlan/vxlan_core.c
405
if (READ_ONCE(f->used) != now)
drivers/net/vxlan/vxlan_core.c
4673
return READ_ONCE(vxlan->net);
drivers/net/vxlan/vxlan_mdb.c
1696
READ_ONCE(remote->filter_mode) == MCAST_INCLUDE) ||
drivers/net/vxlan/vxlan_mdb.c
1697
(READ_ONCE(remote->flags) & VXLAN_MDB_REMOTE_F_BLOCKED))
drivers/net/wireguard/device.c
165
family = READ_ONCE(peer->endpoint.addr.sa_family);
drivers/net/wireguard/noise.c
843
if (likely(!READ_ONCE(container_of(handshake, struct wg_peer,
drivers/net/wireguard/queueing.c
96
if (tail != READ_ONCE(queue->head))
drivers/net/wireguard/queueing.h
122
int cpu = cpumask_next(READ_ONCE(*last_cpu), cpu_online_mask);
drivers/net/wireguard/receive.c
230
send = keypair && READ_ONCE(keypair->sending.is_valid) &&
drivers/net/wireguard/receive.c
252
if (unlikely(!READ_ONCE(keypair->receiving.is_valid) ||
drivers/net/wireguard/receive.c
254
READ_ONCE(keypair->receiving_counter.counter) >= REJECT_AFTER_MESSAGES)) {
drivers/net/wireguard/receive.c
466
READ_ONCE(keypair->receiving_counter.counter));
drivers/net/wireguard/receive.c
523
if (unlikely(READ_ONCE(peer->is_dead)))
drivers/net/wireguard/send.c
131
send = keypair && READ_ONCE(keypair->sending.is_valid) &&
drivers/net/wireguard/send.c
317
if (unlikely(READ_ONCE(peer->is_dead)))
drivers/net/wireguard/send.c
363
if (unlikely(!READ_ONCE(keypair->sending.is_valid)))
drivers/net/wireguard/send.c
68
unlikely(READ_ONCE(peer->is_dead)))
drivers/net/wireguard/timers.c
113
if (!READ_ONCE(peer->is_dead)) {
drivers/net/wireguard/timers.c
36
!READ_ONCE(peer->is_dead)))
drivers/net/wireless/ath/ath11k/hal.c
835
hp = READ_ONCE(*srng->u.dst_ring.hp_addr);
drivers/net/wireless/ath/ath12k/hal.c
506
hp = READ_ONCE(*srng->u.dst_ring.hp_addr);
drivers/net/wireless/ath/ath12k/wifi7/hal.c
487
len = le32_get_bits(READ_ONCE(desc->flags), HAL_CE_DST_STATUS_DESC_FLAGS_LEN);
drivers/net/wireless/ath/ath5k/desc.c
503
txstat1 = READ_ONCE(tx_status->tx_status_1);
drivers/net/wireless/ath/ath5k/desc.c
509
txstat0 = READ_ONCE(tx_status->tx_status_0);
drivers/net/wireless/ath/ath5k/desc.c
703
rxstat1 = READ_ONCE(rx_status->rx_status_1);
drivers/net/wireless/ath/ath5k/desc.c
710
rxstat0 = READ_ONCE(rx_status->rx_status_0);
drivers/net/wireless/ath/ath9k/ar9002_mac.c
322
status = READ_ONCE(ads->ds_txstatus9);
drivers/net/wireless/ath/ath9k/ar9002_mac.c
336
status = READ_ONCE(ads->ds_txstatus0);
drivers/net/wireless/ath/ath9k/ar9002_mac.c
346
status = READ_ONCE(ads->ds_txstatus1);
drivers/net/wireless/ath/ath9k/ar9002_mac.c
375
status = READ_ONCE(ads->ds_txstatus5);
drivers/net/wireless/ath/ath9k/ar9002_mac.c
394
return MS(READ_ONCE(ads->ds_ctl4), AR_PacketDur0);
drivers/net/wireless/ath/ath9k/ar9002_mac.c
396
return MS(READ_ONCE(ads->ds_ctl4), AR_PacketDur1);
drivers/net/wireless/ath/ath9k/ar9002_mac.c
398
return MS(READ_ONCE(ads->ds_ctl5), AR_PacketDur2);
drivers/net/wireless/ath/ath9k/ar9002_mac.c
400
return MS(READ_ONCE(ads->ds_ctl5), AR_PacketDur3);
drivers/net/wireless/ath/ath9k/ar9003_mac.c
363
status = READ_ONCE(ads->status8);
drivers/net/wireless/ath/ath9k/ar9003_mac.c
389
status = READ_ONCE(ads->status2);
drivers/net/wireless/ath/ath9k/ar9003_mac.c
399
status = READ_ONCE(ads->status3);
drivers/net/wireless/ath/ath9k/ar9003_mac.c
424
status = READ_ONCE(ads->status7);
drivers/net/wireless/ath/ath9k/ar9003_mac.c
441
return MS(READ_ONCE(adc->ctl15), AR_PacketDur0);
drivers/net/wireless/ath/ath9k/ar9003_mac.c
443
return MS(READ_ONCE(adc->ctl15), AR_PacketDur1);
drivers/net/wireless/ath/ath9k/ar9003_mac.c
445
return MS(READ_ONCE(adc->ctl16), AR_PacketDur2);
drivers/net/wireless/ath/ath9k/ar9003_mac.c
447
return MS(READ_ONCE(adc->ctl16), AR_PacketDur3);
drivers/net/wireless/ath/wcn36xx/dxe.c
414
if (READ_ONCE(ctl->desc->ctrl) & WCN36xx_DXE_CTRL_VLD)
drivers/net/wireless/ath/wcn36xx/dxe.c
418
READ_ONCE(ctl->desc->ctrl) & WCN36xx_DXE_CTRL_EOP) {
drivers/net/wireless/ath/wcn36xx/dxe.c
624
while (!(READ_ONCE(dxe->ctrl) & WCN36xx_DXE_CTRL_VLD)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
3762
while (READ_ONCE(bus->dpc_triggered)) {
drivers/net/wireless/intel/iwlwifi/mei/main.c
1175
u32 rd = le32_to_cpu(READ_ONCE(notif_q->rd_ptr));
drivers/net/wireless/intel/iwlwifi/mei/main.c
1176
u32 wr = le32_to_cpu(READ_ONCE(notif_q->wr_ptr));
drivers/net/wireless/intel/iwlwifi/mei/main.c
337
u32 rd = le32_to_cpu(READ_ONCE(notif_q->rd_ptr));
drivers/net/wireless/intel/iwlwifi/mei/main.c
338
u32 wr = le32_to_cpu(READ_ONCE(notif_q->wr_ptr));
drivers/net/wireless/intel/iwlwifi/mei/main.c
376
if (READ_ONCE(notif_q->wr_ptr) != READ_ONCE(notif_q->rd_ptr))
drivers/net/wireless/intel/iwlwifi/mei/main.c
380
return READ_ONCE(notif_q->wr_ptr) != READ_ONCE(notif_q->rd_ptr);
drivers/net/wireless/intel/iwlwifi/mei/main.c
512
rd = le32_to_cpu(READ_ONCE(notif_q->rd_ptr));
drivers/net/wireless/intel/iwlwifi/mei/main.c
513
wr = le32_to_cpu(READ_ONCE(notif_q->wr_ptr));
drivers/net/wireless/intel/iwlwifi/mld/debugfs.c
670
ll_causes = READ_ONCE(mld_vif->low_latency_causes);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2205
READ_ONCE(mld->rxq_sync.state) == 0,
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2258
if (READ_ONCE(mld->rxq_sync.state) == 0)
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
6230
READ_ONCE(mvm->queue_sync_state) == 0,
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1919
bool rfkill_safe_init_done = READ_ONCE(mvm->rfkill_safe_init_done);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
770
if (READ_ONCE(mvm->queue_sync_state) == 0)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
201
return le16_to_cpu(READ_ONCE(*rb_stts));
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
205
return le16_to_cpu(READ_ONCE(rb_stts->closed_rb_num)) & 0xFFF;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2557
wr_ptr = READ_ONCE(txq->write_ptr);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2559
while ((txq->read_ptr != READ_ONCE(txq->write_ptr) ||
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2563
u8 write_ptr = READ_ONCE(txq->write_ptr);
drivers/net/wireless/marvell/mwifiex/pcie.c
669
READ_ONCE(adapter->int_status) != 0,
drivers/net/wireless/marvell/mwifiex/pcie.c
674
READ_ONCE(adapter->int_status) != 0,
drivers/net/wireless/mediatek/mt76/dma.c
505
ctrl = le32_to_cpu(READ_ONCE(desc->ctrl));
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
1142
rate_set_tsf = READ_ONCE(sta->rate_set_tsf);
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1358
rate_set_tsf = READ_ONCE(sta->rate_set_tsf);
drivers/net/wireless/mediatek/mt76/mt7615/pci_mac.c
159
(READ_ONCE(dev->reset_state) & state),
drivers/net/wireless/mediatek/mt76/mt7615/pci_mac.c
212
if (!(READ_ONCE(dev->reset_state) & MT_MCU_CMD_STOP_PDMA))
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
1261
(READ_ONCE(dev->recovery.state) & state),
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
1503
if (READ_ONCE(dev->recovery.state) & MT_MCU_CMD_WA_WDT)
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
1523
if (!(READ_ONCE(dev->recovery.state) & MT_MCU_CMD_STOP_DMA))
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
1687
if (READ_ONCE(dev->recovery.state) & MT_MCU_CMD_WDT_MASK) {
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
1698
if ((READ_ONCE(dev->recovery.state) & MT_MCU_CMD_STOP_DMA)) {
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2179
(READ_ONCE(dev->recovery.state) & state),
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2516
if (READ_ONCE(dev->recovery.state) & MT_MCU_CMD_WA_WDT)
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2535
if (!(READ_ONCE(dev->recovery.state) & MT_MCU_CMD_STOP_DMA))
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2729
if (READ_ONCE(dev->recovery.state) & MT_MCU_CMD_WDT_MASK) {
drivers/net/wireless/quantenna/qtnfmac/shm_ipc.c
69
if (!READ_ONCE(ipc->waiting_for_ack))
drivers/net/wireless/realtek/rtw89/chan.c
2612
if (READ_ONCE(rtwvif_link->sync_bcn_tsf) > rtwvif_link->last_sync_bcn_tsf)
drivers/net/wireless/realtek/rtw89/chan.c
2656
rtwvif_link->last_sync_bcn_tsf = READ_ONCE(rtwvif_link->sync_bcn_tsf);
drivers/net/wireless/realtek/rtw89/chan.c
2675
READ_ONCE(rtwvif_link->sync_bcn_tsf) == rtwvif_link->last_sync_bcn_tsf)
drivers/net/wireless/realtek/rtw89/chan.c
2680
rtwvif_link->last_sync_bcn_tsf = READ_ONCE(rtwvif_link->sync_bcn_tsf);
drivers/net/wireless/realtek/rtw89/chan.c
705
u64 sync_tsf = READ_ONCE(rtwvif_link->sync_bcn_tsf);
drivers/net/wireless/realtek/rtw89/chan.h
124
return READ_ONCE(hal->entity_active[phy_idx]);
drivers/net/wireless/realtek/rtw89/chan.h
141
return READ_ONCE(hal->entity_mode);
drivers/net/wireless/virtual/mac80211_hwsim.c
1456
u32 _portid = READ_ONCE(data->wmediumd);
drivers/net/wireless/virtual/mac80211_hwsim.c
2125
_portid = READ_ONCE(data->wmediumd);
drivers/net/wireless/virtual/mac80211_hwsim.c
2248
u32 _portid = READ_ONCE(data->wmediumd);
drivers/net/wireless/virtual/mac80211_hwsim.c
3584
_portid = READ_ONCE(data->wmediumd);
drivers/net/wireless/virtual/mac80211_hwsim.c
3651
_portid = READ_ONCE(data->wmediumd);
drivers/net/wireless/virtual/virt_wifi.c
472
return READ_ONCE(priv->lowerdev->ifindex);
drivers/net/wwan/t7xx/t7xx_modem_ops.c
801
enum t7xx_mode mode = READ_ONCE(t7xx_dev->mode);
drivers/net/wwan/t7xx/t7xx_netdev.c
423
ccmni = READ_ONCE(ccmni_ctlb->ccmni_inst[netif_id]);
drivers/net/wwan/t7xx/t7xx_netdev.c
445
struct t7xx_ccmni *ccmni = READ_ONCE(ctlb->ccmni_inst[0]);
drivers/net/wwan/t7xx/t7xx_netdev.c
457
struct t7xx_ccmni *ccmni = READ_ONCE(ctlb->ccmni_inst[0]);
drivers/net/wwan/t7xx/t7xx_netdev.c
475
if (!READ_ONCE(ctlb->ccmni_inst[0])) {
drivers/net/wwan/t7xx/t7xx_pci.c
116
mode = READ_ONCE(t7xx_dev->mode);
drivers/net/wwan/t7xx/t7xx_pci.c
162
show = READ_ONCE(t7xx_dev->debug_ports_show);
drivers/net/wwan/t7xx/t7xx_pci.c
416
READ_ONCE(t7xx_dev->mode) != T7XX_READY) {
drivers/net/wwan/t7xx/t7xx_pci.c
544
enum t7xx_mode mode = READ_ONCE(t7xx_dev->mode);
drivers/net/wwan/t7xx/t7xx_pci.c
82
mode = READ_ONCE(t7xx_dev->mode);
drivers/net/wwan/t7xx/t7xx_state_monitor.c
268
mode = READ_ONCE(ctl->md->t7xx_dev->mode);
drivers/net/xen-netback/interface.c
187
num_queues = READ_ONCE(vif->num_queues);
drivers/net/xen-netback/interface.c
219
num_queues = READ_ONCE(vif->num_queues);
drivers/net/xen-netback/interface.c
283
num_queues = READ_ONCE(vif->num_queues);
drivers/net/xen-netback/interface.c
435
num_queues = READ_ONCE(vif->num_queues);
drivers/net/xen-netback/interface.c
632
rsp_prod = READ_ONCE(shared->rsp_prod);
drivers/net/xen-netback/interface.c
633
req_prod = READ_ONCE(shared->req_prod);
drivers/net/xen-netback/netback.c
1557
rsp_prod = READ_ONCE(txs->rsp_prod);
drivers/net/xen-netback/netback.c
1558
req_prod = READ_ONCE(txs->req_prod);
drivers/net/xen-netback/netback.c
1572
rsp_prod = READ_ONCE(rxs->rsp_prod);
drivers/net/xen-netback/netback.c
1573
req_prod = READ_ONCE(rxs->req_prod);
drivers/net/xen-netback/rx.c
522
unsigned int needed = READ_ONCE(queue->rx_slots_needed);
drivers/net/xen-netback/rx.c
532
unsigned int needed = READ_ONCE(queue->rx_slots_needed);
drivers/net/xen-netback/rx.c
63
needed = READ_ONCE(queue->rx_slots_needed);
drivers/nvme/host/apple.c
604
return (le16_to_cpu(READ_ONCE(hcqe->status)) & 1) == q->cq_phase;
drivers/nvme/host/apple.c
621
__u16 command_id = READ_ONCE(cqe->command_id);
drivers/nvme/host/apple.c
680
if (!READ_ONCE(q->enabled) && !force)
drivers/nvme/host/apple.c
783
if (unlikely(!READ_ONCE(q->enabled)))
drivers/nvme/host/apple.c
866
if (READ_ONCE(anv->ioq.enabled)) {
drivers/nvme/host/ioctl.c
462
c.common.opcode = READ_ONCE(cmd->opcode);
drivers/nvme/host/ioctl.c
463
c.common.flags = READ_ONCE(cmd->flags);
drivers/nvme/host/ioctl.c
472
c.common.cdw2[0] = cpu_to_le32(READ_ONCE(cmd->cdw2));
drivers/nvme/host/ioctl.c
473
c.common.cdw2[1] = cpu_to_le32(READ_ONCE(cmd->cdw3));
drivers/nvme/host/ioctl.c
476
c.common.cdw10 = cpu_to_le32(READ_ONCE(cmd->cdw10));
drivers/nvme/host/ioctl.c
477
c.common.cdw11 = cpu_to_le32(READ_ONCE(cmd->cdw11));
drivers/nvme/host/ioctl.c
478
c.common.cdw12 = cpu_to_le32(READ_ONCE(cmd->cdw12));
drivers/nvme/host/ioctl.c
479
c.common.cdw13 = cpu_to_le32(READ_ONCE(cmd->cdw13));
drivers/nvme/host/ioctl.c
480
c.common.cdw14 = cpu_to_le32(READ_ONCE(cmd->cdw14));
drivers/nvme/host/ioctl.c
481
c.common.cdw15 = cpu_to_le32(READ_ONCE(cmd->cdw15));
drivers/nvme/host/ioctl.c
486
d.metadata = READ_ONCE(cmd->metadata);
drivers/nvme/host/ioctl.c
487
d.addr = READ_ONCE(cmd->addr);
drivers/nvme/host/ioctl.c
488
d.data_len = READ_ONCE(cmd->data_len);
drivers/nvme/host/ioctl.c
489
d.metadata_len = READ_ONCE(cmd->metadata_len);
drivers/nvme/host/ioctl.c
490
d.timeout_ms = READ_ONCE(cmd->timeout_ms);
drivers/nvme/host/multipath.c
1028
nvme_iopolicy_names[READ_ONCE(subsys->iopolicy)]);
drivers/nvme/host/multipath.c
1035
int old_iopolicy = READ_ONCE(subsys->iopolicy);
drivers/nvme/host/multipath.c
185
if ((READ_ONCE(ns->head->subsys->iopolicy) == NVME_IOPOLICY_QD) &&
drivers/nvme/host/multipath.c
322
READ_ONCE(head->subsys->iopolicy) == NVME_IOPOLICY_NUMA)
drivers/nvme/host/multipath.c
468
switch (READ_ONCE(head->subsys->iopolicy)) {
drivers/nvme/host/nvme.h
473
return READ_ONCE(ctrl->state);
drivers/nvme/host/pci.c
1512
return (le16_to_cpu(READ_ONCE(hcqe->status)) & 1) == nvmeq->cq_phase;
drivers/nvme/host/pci.c
1535
__u16 command_id = READ_ONCE(cqe->command_id);
drivers/nvme/target/admin-cmd.c
1494
nvmet_set_result(req, READ_ONCE(req->sq->ctrl->aen_enabled));
drivers/nvme/target/core.c
740
old_sqhd = READ_ONCE(req->sq->sqhd);
drivers/nvme/target/nvmet.h
537
if (!(READ_ONCE(ctrl->aen_enabled) & (1 << bn)))
drivers/nvme/target/pci-epf.c
1791
cqe->sq_head = cpu_to_le16(READ_ONCE(iod->sq->head));
drivers/nvme/target/pci-epf.c
231
return le32_to_cpu(READ_ONCE(*bar_reg));
drivers/nvme/target/pr.c
84
nvmet_set_result(req, READ_ONCE(req->ns->pr.notify_mask));
drivers/nvme/target/tcp.c
599
queue_cmd = READ_ONCE(queue->cmd);
drivers/pci/endpoint/functions/pci-epf-test.c
986
command = le32_to_cpu(READ_ONCE(reg->command));
drivers/pci/endpoint/functions/pci-epf-test.c
993
if ((le32_to_cpu(READ_ONCE(reg->flags)) & FLAG_USE_DMA) &&
drivers/pci/ide.c
746
int nr = READ_ONCE(hb->nr_ide_streams);
drivers/pci/pci-driver.c
1667
const struct device_driver *drv = READ_ONCE(dev->driver);
drivers/pinctrl/core.c
1282
struct pinctrl_state *old_state = READ_ONCE(p->state);
drivers/platform/surface/aggregator/controller.c
1703
if (WARN_ON(READ_ONCE(ctrl->state) != SSAM_CONTROLLER_STARTED)) {
drivers/platform/surface/aggregator/controller.c
57
u8 old = READ_ONCE(c->value);
drivers/platform/surface/aggregator/controller.c
87
u16 old = READ_ONCE(c->value);
drivers/platform/surface/aggregator/core.c
161
if (READ_ONCE(link->status) == DL_STATE_SUPPLIER_UNBIND) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1334
ptl_old = READ_ONCE(p->ptl);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1478
if (READ_ONCE(p->ptl)) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
853
struct ssh_ptl *ptl = READ_ONCE(p->ptl);
drivers/platform/surface/aggregator/ssh_request_layer.c
607
fixed = READ_ONCE(r->state) & SSH_REQUEST_FLAGS_TY_MASK;
drivers/platform/surface/aggregator/ssh_request_layer.c
619
if (flags == fixed && !READ_ONCE(r->packet.ptl)) {
drivers/platform/surface/aggregator/ssh_request_layer.c
674
if (!READ_ONCE(r->packet.ptl)) {
drivers/platform/surface/aggregator/ssh_request_layer.c
812
ktime_t timestamp = READ_ONCE(r->timestamp);
drivers/platform/surface/aggregator/ssh_request_layer.h
125
ptl = READ_ONCE(rqst->packet.ptl);
drivers/platform/surface/aggregator/trace.h
407
__entry->state = READ_ONCE(packet->state);
drivers/platform/surface/aggregator/trace.h
409
__entry->priority = READ_ONCE(packet->priority);
drivers/platform/surface/aggregator/trace.h
445
__entry->state = READ_ONCE(packet->state);
drivers/platform/surface/aggregator/trace.h
448
__entry->priority = READ_ONCE(packet->priority);
drivers/platform/surface/aggregator/trace.h
490
__entry->state = READ_ONCE(request->state);
drivers/platform/surface/aggregator/trace.h
540
__entry->state = READ_ONCE(request->state);
drivers/platform/surface/surface_acpi_notify.c
104
if (READ_ONCE(link->status) == DL_STATE_SUPPLIER_UNBIND) {
drivers/pmdomain/governor.c
379
next_hrtimer = READ_ONCE(dev->next_hrtimer);
drivers/powercap/idle_inject.c
133
duration_us = READ_ONCE(ii_dev->run_duration_us);
drivers/powercap/idle_inject.c
134
duration_us += READ_ONCE(ii_dev->idle_duration_us);
drivers/powercap/idle_inject.c
161
play_idle_precise(READ_ONCE(ii_dev->idle_duration_us) * NSEC_PER_USEC,
drivers/powercap/idle_inject.c
162
READ_ONCE(ii_dev->latency_us) * NSEC_PER_USEC);
drivers/powercap/idle_inject.c
194
*run_duration_us = READ_ONCE(ii_dev->run_duration_us);
drivers/powercap/idle_inject.c
195
*idle_duration_us = READ_ONCE(ii_dev->idle_duration_us);
drivers/powercap/idle_inject.c
223
unsigned int idle_duration_us = READ_ONCE(ii_dev->idle_duration_us);
drivers/powercap/idle_inject.c
224
unsigned int run_duration_us = READ_ONCE(ii_dev->run_duration_us);
drivers/ptp/ptp_private.h
100
int cnt = READ_ONCE(q->tail) - READ_ONCE(q->head);
drivers/ptp/ptp_vmclock.c
246
if (READ_ONCE(st->sys_cs_id) == CSID_X86_KVM_CLK)
drivers/rpmsg/qcom_glink_native.c
1457
READ_ONCE(channel->intent_req_result) == 0 ||
drivers/rpmsg/qcom_glink_native.c
1458
(READ_ONCE(channel->intent_req_result) > 0 &&
drivers/rpmsg/qcom_glink_native.c
1459
READ_ONCE(channel->intent_received)) ||
drivers/rpmsg/qcom_glink_native.c
1468
ret = READ_ONCE(channel->intent_req_result) ? 0 : -EAGAIN;
drivers/s390/char/sclp_mem.c
186
return sysfs_emit(buf, "%u\n", READ_ONCE(sclp_mem->config));
drivers/s390/char/sclp_mem.c
275
return sysfs_emit(buf, "%u\n", READ_ONCE(sclp_mem->memmap_on_memory));
drivers/s390/net/ism_drv.c
463
while ((ism->ieq_idx + 1) != READ_ONCE(ism->ieq->header.idx)) {
drivers/s390/net/qeth_core_main.c
1431
if (drain || READ_ONCE(priv->state) == QETH_QAOB_DONE) {
drivers/s390/net/qeth_core_main.c
3676
frames = READ_ONCE(queue->max_coalesced_frames);
drivers/s390/net/qeth_core_main.c
3677
usecs = READ_ONCE(queue->coalesce_usecs);
drivers/s390/net/qeth_core_main.c
5647
(skb_len > READ_ONCE(priv->rx_copybreak) &&
drivers/s390/net/qeth_core_main.c
6892
READ_ONCE(card->options.isolation) != ISOLATION_MODE_FWD) {
drivers/s390/net/qeth_l2_main.c
1386
if (READ_ONCE(card->info.pnso_mode) == QETH_PNSO_NONE)
drivers/s390/net/qeth_l2_main.c
1460
if (READ_ONCE(card->info.pnso_mode) == QETH_PNSO_NONE)
drivers/s390/scsi/zfcp_erp.c
623
act = READ_ONCE(fsf_req->erp_action);
drivers/scsi/fcoe/fcoe.c
1663
if (READ_ONCE(stats->ErrorFrames) < 5)
drivers/scsi/fcoe/fcoe_transport.c
186
lfc += READ_ONCE(stats->LinkFailureCount);
drivers/scsi/fcoe/fcoe_transport.c
187
vlfc += READ_ONCE(stats->VLinkFailureCount);
drivers/scsi/fcoe/fcoe_transport.c
188
mdac += READ_ONCE(stats->MissDiscAdvCount);
drivers/scsi/libfc/fc_lport.c
317
fc_stats->tx_frames += READ_ONCE(stats->TxFrames);
drivers/scsi/libfc/fc_lport.c
318
fc_stats->tx_words += READ_ONCE(stats->TxWords);
drivers/scsi/libfc/fc_lport.c
319
fc_stats->rx_frames += READ_ONCE(stats->RxFrames);
drivers/scsi/libfc/fc_lport.c
320
fc_stats->rx_words += READ_ONCE(stats->RxWords);
drivers/scsi/libfc/fc_lport.c
321
fc_stats->error_frames += READ_ONCE(stats->ErrorFrames);
drivers/scsi/libfc/fc_lport.c
322
fc_stats->invalid_crc_count += READ_ONCE(stats->InvalidCRCCount);
drivers/scsi/libfc/fc_lport.c
323
fc_stats->fcp_input_requests += READ_ONCE(stats->InputRequests);
drivers/scsi/libfc/fc_lport.c
324
fc_stats->fcp_output_requests += READ_ONCE(stats->OutputRequests);
drivers/scsi/libfc/fc_lport.c
325
fc_stats->fcp_control_requests += READ_ONCE(stats->ControlRequests);
drivers/scsi/libfc/fc_lport.c
326
fcp_in_bytes += READ_ONCE(stats->InputBytes);
drivers/scsi/libfc/fc_lport.c
327
fcp_out_bytes += READ_ONCE(stats->OutputBytes);
drivers/scsi/libfc/fc_lport.c
328
fc_stats->fcp_packet_alloc_failures += READ_ONCE(stats->FcpPktAllocFails);
drivers/scsi/libfc/fc_lport.c
329
fc_stats->fcp_packet_aborts += READ_ONCE(stats->FcpPktAborts);
drivers/scsi/libfc/fc_lport.c
330
fc_stats->fcp_frame_alloc_failures += READ_ONCE(stats->FcpFrameAllocFails);
drivers/scsi/libfc/fc_lport.c
331
fc_stats->link_failure_count += READ_ONCE(stats->LinkFailureCount);
drivers/scsi/libiscsi.c
1022
if (READ_ONCE(conn->ping_task))
drivers/scsi/libiscsi.c
1075
if (READ_ONCE(conn->ping_task) != task) {
drivers/scsi/libiscsi.c
2068
if (READ_ONCE(conn->ping_task) &&
drivers/scsi/libiscsi.c
2219
if (READ_ONCE(conn->ping_task)) {
drivers/scsi/libiscsi.c
540
if (READ_ONCE(conn->ping_task) == task)
drivers/scsi/lpfc/lpfc_sli.c
11343
if (READ_ONCE(eq->mode) == LPFC_EQ_POLL)
drivers/scsi/mpi3mr/mpi3mr_fw.c
56
ci = READ_ONCE(op_req_q->ci);
drivers/scsi/qla2xxx/qla_edif.c
3694
while (!READ_ONCE(sess->edif.sess_down_acked) &&
drivers/scsi/qla2xxx/qla_target.c
1036
while (!READ_ONCE(sess->logout_completed)) {
drivers/scsi/scsi_transport_iscsi.c
2355
if (READ_ONCE(conn->state) != ISCSI_CONN_DOWN)
drivers/scsi/scsi_transport_iscsi.c
2633
state = READ_ONCE(conn->state);
drivers/scsi/scsi_transport_iscsi.c
3020
state = READ_ONCE(conn->state);
drivers/scsi/scsi_transport_iscsi.c
4170
int conn_state = READ_ONCE(conn->state);
drivers/scsi/sg.c
2580
duration = READ_ONCE(hp->duration);
drivers/scsi/storvsc_drv.c
1476
outgoing_channel = READ_ONCE(stor_device->stor_chns[q_num]);
drivers/scsi/storvsc_drv.c
1490
channel = READ_ONCE(stor_device->stor_chns[tgt_cpu]);
drivers/scsi/storvsc_drv.c
1507
channel = READ_ONCE(stor_device->stor_chns[tgt_cpu]);
drivers/scsi/xen-scsifront.c
390
rp = READ_ONCE(info->ring.sring->rsp_prod);
drivers/soc/tegra/regulators-tegra20.c
359
tegra->sys_reboot_mode = READ_ONCE(tegra->sys_reboot_mode_req);
drivers/soc/tegra/regulators-tegra20.c
360
tegra->sys_suspend_mode = READ_ONCE(tegra->sys_suspend_mode_req);
drivers/soc/tegra/regulators-tegra30.c
354
tegra->sys_reboot_mode = READ_ONCE(tegra->sys_reboot_mode_req);
drivers/soc/tegra/regulators-tegra30.c
355
tegra->sys_suspend_mode = READ_ONCE(tegra->sys_suspend_mode_req);
drivers/spi/spi-fsl-dspi.c
1237
if (READ_ONCE(message->status))
drivers/spi/spi-imx.c
1795
READ_ONCE(spi_imx->target_aborted)) {
drivers/spi/spi-imx.c
1803
READ_ONCE(spi_imx->target_aborted)) {
drivers/spi/spi.c
1816
if (READ_ONCE(ctlr->cur_msg_incomplete))
drivers/spi/spi.c
2170
if (READ_ONCE(ctlr->cur_msg_need_completion))
drivers/spi/spi.c
4669
if (READ_ONCE(ctlr->queue_empty) && !ctlr->must_async) {
drivers/target/target_core_user.c
1457
while (udev->cmdr_last_cleaned != READ_ONCE(mb->cmd_tail)) {
drivers/thermal/testing/zone.c
370
*temp = READ_ONCE(tt_zone->tz_temp);
drivers/thermal/thermal_sysfs.c
145
return sysfs_emit(buf, "%d\n", READ_ONCE(trip->temperature));
drivers/thermal/thermal_sysfs.c
191
return sysfs_emit(buf, "%d\n", READ_ONCE(trip->hysteresis));
drivers/tty/n_tty.c
198
if (unlikely(READ_ONCE(ldata->no_room))) {
drivers/tty/serial/amba-pl011.c
2569
unsigned int len = READ_ONCE(wctxt->len);
drivers/tty/serial/imx.c
2176
int len = READ_ONCE(wctxt->len);
drivers/tty/serial/sifive.c
851
int len = READ_ONCE(wctxt->len);
drivers/tty/serial/tegra-utc.c
445
unsigned int len = READ_ONCE(wctxt->len);
drivers/tty/tty_audit.c
212
audit_tty = READ_ONCE(current->signal->audit_tty);
drivers/tty/tty_port.c
29
tty = READ_ONCE(port->itty);
drivers/tty/tty_port.c
50
tty = READ_ONCE(port->itty);
drivers/ufs/core/ufs-mcq.c
118
struct blk_mq_hw_ctx *hctx = READ_ONCE(req->mq_hctx);
drivers/ufs/core/ufshcd.c
4549
const u32 cur_ahit = READ_ONCE(hba->ahit);
drivers/usb/class/cdc-wdm.c
536
cntr = READ_ONCE(desc->length);
drivers/usb/core/devio.c
152
lim = READ_ONCE(usbfs_memory_mb);
drivers/usb/core/devio.c
2586
READ_ONCE(ps->not_yet_resumed) == 0);
drivers/usb/core/port.c
204
enum usb_device_state state = READ_ONCE(port_dev->state);
drivers/usb/core/sysfs.c
1150
string = READ_ONCE(intf->cur_altsetting->string);
drivers/usb/core/sysfs.c
1166
alt = READ_ONCE(intf->cur_altsetting);
drivers/usb/gadget/function/f_ncm.c
1150
struct net_device *netdev = READ_ONCE(ncm->netdev);
drivers/usb/gadget/udc/gr_udc.c
1244
if (READ_ONCE(req->last_desc->ctrl) & GR_DESC_IN_CTRL_EN)
drivers/usb/gadget/udc/gr_udc.c
1273
ctrl = READ_ONCE(req->curr_desc->ctrl);
drivers/usb/host/ohci-hcd.c
811
head = hc32_to_cpu(ohci, READ_ONCE(ed->hwHeadP)) & TD_MASK;
drivers/usb/host/uhci-hcd.h
191
#define qh_element(qh) READ_ONCE((qh)->element)
drivers/usb/host/uhci-hcd.h
279
READ_ONCE((td)->status))
drivers/vfio/pci/vfio_pci_intrs.c
94
struct eventfd_ctx *trigger = READ_ONCE(ctx->trigger);
drivers/vfio/vfio_main.c
490
return !WARN_ON_ONCE(!READ_ONCE(device->open_count));
drivers/vhost/scsi.c
1229
tpg = READ_ONCE(vs_tpg[*vc->target]);
drivers/vhost/vhost.c
2695
next = vhost16_to_cpu(vq, READ_ONCE(desc->next));
drivers/vhost/vringh.c
1222
tmp = READ_ONCE(*from);
drivers/vhost/vringh.c
861
*val = vringh16_to_cpu(vrh, READ_ONCE(*p));
drivers/virt/coco/tdx-guest/tdx-guest.c
312
out_len = READ_ONCE(quote_buf->out_len);
drivers/virtio/virtio_ring.c
2063
return virtqueue_poll_packed(vq, READ_ONCE(vq->last_used_idx));
drivers/virtio/virtio_ring.c
2095
return virtqueue_poll_packed(vq, READ_ONCE(vq->last_used_idx));
drivers/virtio/virtio_ring.c
2186
last_used_idx = READ_ONCE(vq->last_used_idx);
drivers/virtio/virtio_ring.c
2276
last_used_idx = READ_ONCE(vq->last_used_idx);
drivers/virtio/virtio_ring.c
2309
last_used_idx = READ_ONCE(vq->last_used_idx);
drivers/virtio/virtio_ring.c
3580
return READ_ONCE(vq->broken);
drivers/xen/events/events_base.c
275
irq = READ_ONCE(evtchn_to_irq[EVTCHN_ROW(evtchn)][EVTCHN_COL(evtchn)]);
drivers/xen/evtchn.c
180
prod = READ_ONCE(u->ring_prod);
drivers/xen/evtchn.c
181
cons = READ_ONCE(u->ring_cons);
drivers/xen/evtchn.c
223
c = READ_ONCE(u->ring_cons);
drivers/xen/evtchn.c
224
p = READ_ONCE(u->ring_prod);
drivers/xen/evtchn.c
234
READ_ONCE(u->ring_cons) != READ_ONCE(u->ring_prod));
drivers/xen/evtchn.c
628
if (READ_ONCE(u->ring_cons) != READ_ONCE(u->ring_prod))
drivers/xen/grant-table.c
506
unsigned int nr = READ_ONCE(free_per_iteration);
drivers/xen/pvcalls-back.c
796
data = READ_ONCE(queue->rskq_accept_head) != NULL;
drivers/xen/pvcalls-front.c
1057
READ_ONCE(bedata->rsp[req_id].req_id) == req_id);
drivers/xen/pvcalls-front.c
1087
if (READ_ONCE(map->passive.inflight_req_id) != PVCALLS_INVALID_ID &&
drivers/xen/pvcalls-front.c
1088
READ_ONCE(map->passive.inflight_req_id) != 0) {
drivers/xen/pvcalls-front.c
334
READ_ONCE(bedata->rsp[req_id].req_id) == req_id);
drivers/xen/pvcalls-front.c
481
READ_ONCE(bedata->rsp[req_id].req_id) == req_id);
drivers/xen/pvcalls-front.c
715
READ_ONCE(bedata->rsp[req_id].req_id) == req_id);
drivers/xen/pvcalls-front.c
765
READ_ONCE(bedata->rsp[req_id].req_id) == req_id);
drivers/xen/pvcalls-front.c
805
req_id = READ_ONCE(map->passive.inflight_req_id);
drivers/xen/pvcalls-front.c
807
READ_ONCE(bedata->rsp[req_id].req_id) == req_id) {
drivers/xen/pvcalls-front.c
883
READ_ONCE(bedata->rsp[req_id].req_id) == req_id)) {
drivers/xen/pvcalls-front.c
926
uint32_t req_id = READ_ONCE(map->passive.inflight_req_id);
drivers/xen/pvcalls-front.c
929
READ_ONCE(bedata->rsp[req_id].req_id) == req_id)
drivers/xen/time.c
43
h = READ_ONCE(p32[1]);
drivers/xen/time.c
44
l = READ_ONCE(p32[0]);
drivers/xen/time.c
45
h2 = READ_ONCE(p32[1]);
drivers/xen/time.c
50
ret = READ_ONCE(*p);
fs/afs/cell.c
471
unsigned int min_ttl = READ_ONCE(afs_cell_min_ttl);
fs/afs/cell.c
472
unsigned int max_ttl = READ_ONCE(afs_cell_max_ttl);
fs/afs/dir.c
1042
dir_version = (long)READ_ONCE(dvnode->status.data_version);
fs/afs/dir.c
1043
de_version = (long)READ_ONCE(dentry->d_fsdata);
fs/afs/dir.c
1045
dir_version = (long)READ_ONCE(dvnode->invalid_before);
fs/afs/dir.c
781
afs_dataversion_t data_version = READ_ONCE(dvnode->status.data_version);
fs/afs/dynroot.c
297
if (READ_ONCE(cell->state) == AFS_CELL_REMOVING ||
fs/afs/dynroot.c
298
READ_ONCE(cell->state) == AFS_CELL_DEAD) {
fs/afs/file.c
274
state = READ_ONCE(call->state);
fs/afs/file.c
279
while (state < AFS_CALL_COMPLETE && READ_ONCE(call->need_attention)) {
fs/afs/file.c
282
state = READ_ONCE(call->state);
fs/afs/flock.c
743
lock_count = READ_ONCE(vnode->status.lock_count);
fs/afs/internal.h
1472
return READ_ONCE(call->state) == state;
fs/afs/rotate.c
608
set = READ_ONCE(op->estate->responsive_set);
fs/afs/rotate.c
609
failed = READ_ONCE(op->estate->failed_set);
fs/afs/rotate.c
698
error = READ_ONCE(estate->error);
fs/afs/rxrpc.c
525
while (state = READ_ONCE(call->state),
fs/afs/rxrpc.c
551
state = READ_ONCE(call->state);
fs/afs/rxrpc.c
944
state = READ_ONCE(call->state);
fs/afs/server.c
256
ret = READ_ONCE(server->create_error);
fs/afs/server.c
361
READ_ONCE(server->cell->state) >= AFS_CELL_REMOVING)
fs/afs/server.c
379
READ_ONCE(server->cell->state) < AFS_CELL_REMOVING) {
fs/afs/vl_probe.c
289
rtt_s = READ_ONCE(server->rtt);
fs/afs/vl_rotate.c
257
set = READ_ONCE(alist->responded);
fs/afs/vl_rotate.c
258
failed = READ_ONCE(alist->probe_failed);
fs/afs/vl_rotate.c
259
vc->addr_index = READ_ONCE(alist->preferred);
fs/afs/vl_rotate.c
303
afs_prioritise_error(&vc->cumul_error, READ_ONCE(s->probe.error),
fs/aio.c
1731
if (!READ_ONCE(req->cancelled))
fs/aio.c
1743
if (!mask && !READ_ONCE(req->cancelled)) {
fs/autofs/expire.c
211
unsigned int ino_count = READ_ONCE(ino->count);
fs/autofs/expire.c
379
ino_count = READ_ONCE(ino->count) + 1;
fs/autofs/expire.c
396
ino_count = READ_ONCE(ino->count) + 1;
fs/binfmt_elf.c
1021
const int snapshot_randomize_va_space = READ_ONCE(randomize_va_space);
fs/binfmt_elf.c
1549
state = READ_ONCE(p->__state);
fs/binfmt_elf_fdpic.c
1362
state = READ_ONCE(p->__state);
fs/btrfs/bio.c
141
bbio->bio.bi_status = READ_ONCE(bbio->status);
fs/btrfs/block-group.c
2060
if (ret && !READ_ONCE(space_info->periodic_reclaim))
fs/btrfs/block-group.c
3861
if (READ_ONCE(space_info->periodic_reclaim))
fs/btrfs/block-group.c
3873
if (READ_ONCE(space_info->periodic_reclaim))
fs/btrfs/btrfs_inode.h
347
return READ_ONCE(inode->first_dir_index_to_log);
fs/btrfs/compression.c
149
ret = READ_ONCE(compr_pool.count) - READ_ONCE(compr_pool.thresh);
fs/btrfs/ctree.h
344
return READ_ONCE(root->log_transid);
fs/btrfs/ctree.h
354
return READ_ONCE(root->last_log_commit);
fs/btrfs/ctree.h
364
return READ_ONCE(root->last_trans);
fs/btrfs/delayed-inode.c
1303
delayed_node = READ_ONCE(inode->delayed_node);
fs/btrfs/delayed-inode.c
76
node = READ_ONCE(btrfs_inode->delayed_node);
fs/btrfs/discard.c
424
u32 kbps_limit = READ_ONCE(discard_ctl->kbps_limit);
fs/btrfs/discard.c
659
iops_limit = READ_ONCE(discard_ctl->iops_limit);
fs/btrfs/free-space-cache.c
2703
bg_reclaim_threshold = READ_ONCE(sinfo->bg_reclaim_threshold);
fs/btrfs/free-space-cache.c
3731
const u64 max_discard_size = READ_ONCE(discard_ctl->max_discard_size);
fs/btrfs/free-space-cache.c
3843
const u64 max_discard_size = READ_ONCE(discard_ctl->max_discard_size);
fs/btrfs/free-space-cache.c
3939
const u64 max_discard_size = READ_ONCE(discard_ctl->max_discard_size);
fs/btrfs/fs.h
1008
return READ_ONCE(fs_info->last_root_drop_gen);
fs/btrfs/fs.h
1179
#define BTRFS_FS_ERROR(fs_info) (READ_ONCE((fs_info)->fs_error))
fs/btrfs/fs.h
982
return READ_ONCE(fs_info->generation);
fs/btrfs/fs.h
992
return READ_ONCE(fs_info->last_trans_committed);
fs/btrfs/inode.c
9562
int err = blk_status_to_errno(READ_ONCE(priv->status));
fs/btrfs/inode.c
9629
ret = blk_status_to_errno(READ_ONCE(priv->status));
fs/btrfs/inode.c
9640
return blk_status_to_errno(READ_ONCE(priv->status));
fs/btrfs/ioctl.c
4705
sqe_addr = u64_to_user_ptr(READ_ONCE(cmd->sqe->addr));
fs/btrfs/ioctl.c
4846
sqe_addr = u64_to_user_ptr(READ_ONCE(cmd->sqe->addr));
fs/btrfs/messages.c
34
unsigned long fs_state = READ_ONCE(info->fs_state);
fs/btrfs/relocation.c
4033
return blk_status_to_errno(READ_ONCE(priv->status));
fs/btrfs/scrub.c
1387
bwlimit = READ_ONCE(device->scrub_speed_max);
fs/btrfs/space-info.c
2081
if (READ_ONCE(space_info->dynamic_reclaim))
fs/btrfs/space-info.c
2083
return READ_ONCE(space_info->bg_reclaim_threshold);
fs/btrfs/space-info.c
2163
if (!READ_ONCE(space_info->periodic_reclaim))
fs/btrfs/space-info.c
2178
if (!READ_ONCE(space_info->periodic_reclaim))
fs/btrfs/sysfs.c
1271
switch (READ_ONCE(fs_info->exclusive_operation)) {
fs/btrfs/sysfs.c
1395
const enum btrfs_read_policy policy = READ_ONCE(fs_devices->read_policy);
fs/btrfs/sysfs.c
1411
READ_ONCE(fs_devices->rr_min_contig_read));
fs/btrfs/sysfs.c
1415
READ_ONCE(fs_devices->read_devid));
fs/btrfs/sysfs.c
1459
if (index != READ_ONCE(fs_devices->read_policy) ||
fs/btrfs/sysfs.c
1460
value != READ_ONCE(fs_devices->rr_min_contig_read)) {
fs/btrfs/sysfs.c
1486
if (index != READ_ONCE(fs_devices->read_policy) ||
fs/btrfs/sysfs.c
1487
value != READ_ONCE(fs_devices->read_devid)) {
fs/btrfs/sysfs.c
1498
if (index != READ_ONCE(fs_devices->read_policy)) {
fs/btrfs/sysfs.c
1514
return sysfs_emit(buf, "%d\n", READ_ONCE(fs_info->bg_reclaim_threshold));
fs/btrfs/sysfs.c
2028
return sysfs_emit(buf, "%llu\n", READ_ONCE(device->scrub_speed_max));
fs/btrfs/sysfs.c
533
READ_ONCE(fs_info->discard_ctl.iops_limit));
fs/btrfs/sysfs.c
564
READ_ONCE(fs_info->discard_ctl.kbps_limit));
fs/btrfs/sysfs.c
594
READ_ONCE(fs_info->discard_ctl.max_discard_size));
fs/btrfs/sysfs.c
756
return sysfs_emit(buf, "%llu\n", READ_ONCE(sinfo->chunk_size));
fs/btrfs/sysfs.c
933
if (READ_ONCE(space_info->dynamic_reclaim))
fs/btrfs/sysfs.c
958
return sysfs_emit(buf, "%d\n", READ_ONCE(space_info->dynamic_reclaim));
fs/btrfs/sysfs.c
991
return sysfs_emit(buf, "%d\n", READ_ONCE(space_info->periodic_reclaim));
fs/btrfs/transaction.c
2332
ret = READ_ONCE(prev_trans->aborted);
fs/btrfs/transaction.h
173
#define TRANS_ABORTED(trans) (unlikely(READ_ONCE((trans)->aborted)))
fs/btrfs/tree-log.h
68
return READ_ONCE(trans->fs_info->last_trans_log_full_commit) ==
fs/btrfs/volumes.c
5448
ctl->max_chunk_size = READ_ONCE(space_info->chunk_size);
fs/btrfs/volumes.c
6212
if (device->devid == READ_ONCE(device->fs_devices->read_devid))
fs/btrfs/volumes.c
6260
min_reads_per_dev = READ_ONCE(fs_info->fs_devices->rr_min_contig_read) >>
fs/btrfs/volumes.c
6280
const enum btrfs_read_policy policy = READ_ONCE(fs_info->fs_devices->read_policy);
fs/buffer.c
1606
b_state = READ_ONCE(bh->b_state);
fs/buffer.c
1800
1 << READ_ONCE(inode->i_blkbits), b_state);
fs/ceph/caps.c
1315
ceph_encode_32(&p, READ_ONCE(osdc->epoch_barrier));
fs/ceph/caps.c
3052
fi->filp_gen != READ_ONCE(fsc->filp_gen))
fs/ceph/caps.c
3111
fi->filp_gen != READ_ONCE(fsc->filp_gen)) {
fs/ceph/file.c
240
fi->filp_gen = READ_ONCE(ceph_inode_to_fs_client(inode)->filp_gen);
fs/ceph/file.c
535
READ_ONCE(di->lease_shared_gen)) {
fs/ceph/io.c
141
is_odirect = READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT;
fs/ceph/io.c
185
is_odirect = READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT;
fs/ceph/io.c
31
is_odirect = READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT;
fs/ceph/io.c
73
is_odirect = READ_ONCE(ci->i_ceph_flags) & CEPH_I_ODIRECT;
fs/ceph/mds_client.c
1333
parent = READ_ONCE(req->r_dentry->d_parent);
fs/ceph/mds_client.c
1670
if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_FENCE_IO)
fs/ceph/mds_client.c
3457
if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_FENCE_IO) {
fs/ceph/mds_client.c
3468
if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_SHUTDOWN) {
fs/ceph/mds_client.c
3473
if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_MOUNTING) {
fs/ceph/mds_client.c
5402
if (READ_ONCE(fsc->mount_state) != CEPH_MOUNT_MOUNTED)
fs/ceph/mds_client.c
5405
if (!READ_ONCE(fsc->blocklisted))
fs/ceph/mds_client.c
5941
if (READ_ONCE(mdsc->fsc->mount_state) >= CEPH_MOUNT_SHUTDOWN)
fs/ceph/mds_client.c
5972
if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_SHUTDOWN)
fs/ceph/mds_client.c
6288
if (READ_ONCE(mdsc->fsc->mount_state) != CEPH_MOUNT_FENCE_IO &&
fs/ceph/mds_client.c
970
if (READ_ONCE(mdsc->fsc->mount_state) == CEPH_MOUNT_FENCE_IO)
fs/ceph/super.h
1141
unsigned long flags = READ_ONCE(ceph_inode(inode)->i_ceph_flags);
fs/ceph/super.h
1143
int state = READ_ONCE(fsc->mount_state);
fs/ceph/super.h
694
if (!(READ_ONCE(ci->i_ceph_flags) & CEPH_I_ERROR_WRITE)) {
fs/ceph/super.h
703
if (READ_ONCE(ci->i_ceph_flags) & CEPH_I_ERROR_WRITE) {
fs/coredump.c
1651
if ((READ_ONCE(file_inode(vma->vm_file)->i_mode) & 0111) != 0)
fs/crypto/hooks.c
369
pstr.name = READ_ONCE(inode->i_link);
fs/crypto/hooks.c
463
link = READ_ONCE(inode->i_link);
fs/crypto/hooks.c
54
dentry_parent = READ_ONCE(dentry->d_parent);
fs/crypto/keysetup.c
849
return !READ_ONCE(ci->ci_master_key->mk_present);
fs/d_path.c
108
const struct dentry *parent = READ_ONCE(dentry->d_parent);
fs/d_path.c
111
struct mount *m = READ_ONCE(mnt->mnt_parent);
fs/d_path.c
115
dentry = READ_ONCE(mnt->mnt_mountpoint);
fs/d_path.c
120
mnt_ns = READ_ONCE(mnt->mnt_ns);
fs/d_path.c
99
u32 dlen = READ_ONCE(name->len);
fs/dax.c
508
struct address_space *mapping = READ_ONCE(folio->mapping);
fs/dcache.c
2567
unsigned n = READ_ONCE(dir->i_dir_seq);
fs/dcache.c
2649
if (unlikely(READ_ONCE(parent->d_inode->i_dir_seq) != seq)) {
fs/dcache.c
320
const unsigned char *cs = READ_ONCE(dentry->d_name.name);
fs/dcache.c
3221
parent = READ_ONCE(dentry->d_parent);
fs/dcache.c
369
s = READ_ONCE(dentry->d_name.name);
fs/dcache.c
408
flags = READ_ONCE(dentry->d_flags);
fs/dcache.c
416
unsigned flags = READ_ONCE(dentry->d_flags);
fs/dcache.c
750
d_flags = READ_ONCE(dentry->d_flags);
fs/dcache.c
977
ret = READ_ONCE(dentry->d_parent);
fs/debugfs/file.c
177
struct debugfs_fsdata *fsd = READ_ONCE(dentry->d_fsdata);
fs/debugfs/file.c
220
fsd = READ_ONCE(dentry->d_fsdata);
fs/debugfs/file.c
251
fsd = READ_ONCE(dentry->d_fsdata);
fs/debugfs/file.c
75
d_fsd = READ_ONCE(dentry->d_fsdata);
fs/debugfs/inode.c
704
fsd = READ_ONCE(dentry->d_fsdata);
fs/direct-io.c
1110
unsigned i_blkbits = READ_ONCE(inode->i_blkbits);
fs/dlm/lock.c
325
return jiffies + (READ_ONCE(dlm_config.ci_toss_secs) * HZ);
fs/dlm/lowcomms.c
1495
buflen = READ_ONCE(dlm_config.ci_buffer_size);
fs/dlm/lowcomms.c
605
READ_ONCE(sk->sk_err_soft));
fs/dlm/lowcomms.c
614
READ_ONCE(sk->sk_err_soft));
fs/dlm/lowcomms.c
622
READ_ONCE(sk->sk_err_soft));
fs/ecryptfs/inode.c
351
lower_inode = READ_ONCE(lower_dentry->d_inode);
fs/erofs/decompressor_deflate.c
123
wait_event(z_erofs_deflate_wq, READ_ONCE(z_erofs_deflate_head));
fs/erofs/decompressor_lzma.c
115
READ_ONCE(z_erofs_lzma_head));
fs/erofs/decompressor_lzma.c
174
wait_event(z_erofs_lzma_wq, READ_ONCE(z_erofs_lzma_head));
fs/erofs/decompressor_zstd.c
29
wait_event(z_erofs_zstd_wq, READ_ONCE(z_erofs_zstd_head));
fs/erofs/zdata.c
1409
next = READ_ONCE(be.pcl->next);
fs/erofs/zdata.c
1520
mapping = READ_ONCE(folio->mapping);
fs/erofs/zdata.c
1696
next = READ_ONCE(pcl->next);
fs/erofs/zdata.c
553
if (READ_ONCE(pcl->compressed_bvecs[i].page))
fs/eventfd.c
164
count = READ_ONCE(ctx->count);
fs/eventpoll.c
1284
if (READ_ONCE(ep->ovflist) != EP_UNACTIVE_PTR) {
fs/eventpoll.c
1286
epi->next = READ_ONCE(ep->ovflist);
fs/eventpoll.c
1538
} else if (!READ_ONCE(file->f_ep)) {
fs/eventpoll.c
2309
if (READ_ONCE(fd_file(f)->f_ep) || ep->gen == loop_check_gen ||
fs/eventpoll.c
393
READ_ONCE(ep->ovflist) != EP_UNACTIVE_PTR;
fs/eventpoll.c
410
unsigned long bp_usec = READ_ONCE(ep->busy_poll_usecs);
fs/eventpoll.c
424
return !!READ_ONCE(ep->busy_poll_usecs) ||
fs/eventpoll.c
425
READ_ONCE(ep->prefer_busy_poll) ||
fs/eventpoll.c
444
unsigned int napi_id = READ_ONCE(ep->napi_id);
fs/eventpoll.c
445
u16 budget = READ_ONCE(ep->busy_poll_budget);
fs/eventpoll.c
446
bool prefer_busy_poll = READ_ONCE(ep->prefer_busy_poll);
fs/eventpoll.c
490
napi_id = READ_ONCE(sk->sk_napi_id);
fs/eventpoll.c
535
epoll_params.busy_poll_usecs = READ_ONCE(ep->busy_poll_usecs);
fs/eventpoll.c
536
epoll_params.busy_poll_budget = READ_ONCE(ep->busy_poll_budget);
fs/eventpoll.c
537
epoll_params.prefer_busy_poll = READ_ONCE(ep->prefer_busy_poll);
fs/eventpoll.c
548
unsigned int napi_id = READ_ONCE(ep->napi_id);
fs/eventpoll.c
550
if (napi_id_valid(napi_id) && READ_ONCE(ep->prefer_busy_poll))
fs/eventpoll.c
556
unsigned int napi_id = READ_ONCE(ep->napi_id);
fs/eventpoll.c
558
if (napi_id_valid(napi_id) && READ_ONCE(ep->prefer_busy_poll))
fs/eventpoll.c
761
for (nepi = READ_ONCE(ep->ovflist); (epi = nepi) != NULL;
fs/exec.c
1544
mode = READ_ONCE(inode->i_mode);
fs/ext4/ext4_jbd2.c
221
if (errseq_check(&mapping->wb_err, READ_ONCE(sbi->s_bdev_wb_err))) {
fs/ext4/extents_status.c
1113
es1 = READ_ONCE(tree->cache_es);
fs/ext4/extents_status.c
328
es1 = READ_ONCE(tree->cache_es);
fs/ext4/fast_commit.c
978
ret = jbd2_submit_inode_data(journal, READ_ONCE(ei->jinode));
fs/ext4/fast_commit.c
984
ret = jbd2_wait_inode_data(journal, READ_ONCE(ei->jinode));
fs/ext4/file.c
399
if (pos + size <= READ_ONCE(EXT4_I(inode)->i_disksize) &&
fs/ext4/inode.c
131
struct jbd2_inode *jinode = READ_ONCE(EXT4_I(inode)->jinode);
fs/ext4/inode.c
1991
map->m_seq = READ_ONCE(EXT4_I(inode)->i_es_seq);
fs/ext4/inode.c
2534
if (disksize > READ_ONCE(EXT4_I(inode)->i_disksize)) {
fs/ext4/inode.c
4624
u64 i_blocks = READ_ONCE(inode->i_blocks);
fs/ext4/inode.c
5591
if (READ_ONCE(ei->i_disksize) != ext4_isize(inode->i_sb, raw_inode))
fs/ext4/inode.c
597
map->m_seq = READ_ONCE(EXT4_I(inode)->i_es_seq);
fs/ext4/inode.c
668
map->m_seq = READ_ONCE(EXT4_I(inode)->i_es_seq);
fs/ext4/inode.c
881
old_state = READ_ONCE(bh->b_state);
fs/ext4/mballoc.c
3043
ac->ac_g_ex.fe_group = READ_ONCE(sbi->s_mb_last_groups[hash]);
fs/ext4/move_extent.c
221
if (mext->orig_map.m_seq != READ_ONCE(EXT4_I(orig_inode)->i_es_seq)) {
fs/ext4/super.c
363
return le16_to_cpu(READ_ONCE(bg->bg_free_inodes_count_lo)) |
fs/ext4/super.c
365
(__u32)le16_to_cpu(READ_ONCE(bg->bg_free_inodes_count_hi)) << 16 : 0);
fs/fcntl.c
362
u64 hint = READ_ONCE(inode->i_write_hint);
fs/fcntl.c
863
int signum = READ_ONCE(fown->signum);
fs/fs_pin.c
76
p = READ_ONCE(m->mnt_pins.first);
fs/fs_pin.c
90
q = READ_ONCE(p->first);
fs/fuse/control.c
114
val = READ_ONCE(fc->max_background);
fs/fuse/control.c
156
val = READ_ONCE(fc->congestion_threshold);
fs/fuse/dev.c
1560
READ_ONCE(file->private_data) != FUSE_DEV_SYNC_INIT);
fs/fuse/dev.c
186
!READ_ONCE(fc->connected)) {
fs/fuse/dev_uring.c
1016
struct iovec __user *uiov = u64_to_user_ptr(READ_ONCE(sqe->addr));
fs/fuse/dev_uring.c
1093
unsigned int qid = READ_ONCE(cmd_req->qid);
fs/fuse/dev_uring.c
129
queue = READ_ONCE(ring->queues[qid]);
fs/fuse/dev_uring.c
170
queue = READ_ONCE(ring->queues[qid]);
fs/fuse/dev_uring.c
440
struct fuse_ring_queue *queue = READ_ONCE(ring->queues[qid]);
fs/fuse/dev_uring.c
475
struct fuse_ring_queue *queue = READ_ONCE(ring->queues[qid]);
fs/fuse/dev_uring.c
888
uint64_t commit_id = READ_ONCE(cmd_req->commit_id);
fs/fuse/dev_uring.c
889
unsigned int qid = READ_ONCE(cmd_req->qid);
fs/fuse/dev_uring.c
905
if (!READ_ONCE(fc->connected) || READ_ONCE(queue->stopped))
fs/fuse/dir.c
1528
u32 inval_mask = READ_ONCE(fi->inval_mask);
fs/fuse/dir.c
1773
if (perm_mask & READ_ONCE(fi->inval_mask) ||
fs/fuse/fuse_dev_i.h
51
struct fuse_dev *fud = READ_ONCE(file->private_data);
fs/fuse/fuse_i.h
1581
return READ_ONCE(fi->fb);
fs/fuse/iomode.c
20
return READ_ONCE(fi->iocachectr) < 0 && !fuse_inode_backing(fi);
fs/gfs2/glock.c
2751
if (!READ_ONCE(fl_gh->gh_gl))
fs/gfs2/quota.c
1185
u64 sync_gen = READ_ONCE(sdp->sd_quota_sync_gen);
fs/iomap/direct-io.c
865
if (!READ_ONCE(dio->submit.waiter))
fs/jbd2/commit.c
1005
mapping = READ_ONCE(bh->b_folio->mapping);
fs/jbd2/journal.c
776
return tid_geq(READ_ONCE(journal->j_commit_sequence), tid);
fs/jbd2/transaction.c
1173
jh = READ_ONCE(bh->b_private);
fs/jbd2/transaction.c
1179
if (READ_ONCE(jh->b_transaction) != handle->h_transaction &&
fs/jbd2/transaction.c
1180
READ_ONCE(jh->b_next_transaction) != handle->h_transaction)
fs/jbd2/transaction.c
127
oldts = READ_ONCE(transaction->t_max_wait);
fs/kernfs/inode.c
32
attr = READ_ONCE(kn->iattr);
fs/kernfs/inode.c
54
return READ_ONCE(kn->iattr);
fs/libfs.c
1825
parent = READ_ONCE(dentry->d_parent);
fs/libfs.c
1826
dir = READ_ONCE(parent->d_inode);
fs/libfs.c
1861
const struct inode *dir = READ_ONCE(dentry->d_inode);
fs/mnt_idmapping.c
223
u32 nr_extents = READ_ONCE(map_from->nr_extents);
fs/mount.h
185
struct mnt_namespace *ns = READ_ONCE(m->mnt_ns);
fs/namei.c
2055
res = READ_ONCE(inode->i_link);
fs/namei.c
3776
if (READ_ONCE(c1->d_parent) == p2) {
fs/namei.c
411
return likely(!READ_ONCE(inode->i_acl));
fs/namei.c
6241
link = READ_ONCE(inode->i_link);
fs/namespace.c
1250
mnt->mnt.mnt_flags = READ_ONCE(old->mnt.mnt_flags) &
fs/namespace.c
1392
if (likely(READ_ONCE(mnt->mnt_ns))) {
fs/namespace.c
2468
unsigned int max = READ_ONCE(sysctl_mount_max);
fs/namespace.c
401
if (READ_ONCE(mnt->mnt_sb->s_readonly_remount))
fs/namespace.c
445
while (__test_write_hold(READ_ONCE(mnt->mnt_pprev_for_sb))) {
fs/namespace.c
5212
unsigned int mnt_flags = READ_ONCE(mnt->mnt_flags);
fs/netfs/fscache_cookie.c
379
return READ_ONCE(cookie->state) == FSCACHE_COOKIE_STATE_DROPPED;
fs/netfs/read_collect.c
225
transferred = READ_ONCE(front->transferred);
fs/netfs/rolling_buffer.c
184
struct folio_queue *spent = roll->tail, *next = READ_ONCE(spent->next);
fs/nfs/dir.c
1962
unsigned long gencount = READ_ONCE(NFS_I(inode)->attr_gencount);
fs/nfs/dir.c
2558
unsigned long gencount = READ_ONCE(NFS_I(inode)->attr_gencount);
fs/nfs/flexfilelayout/flexfilelayoutdev.c
386
if (READ_ONCE(ds->ds_clp))
fs/nfs/inode.c
1041
cache_validity = READ_ONCE(NFS_I(inode)->cache_validity);
fs/nfs/inode.c
167
unsigned long cache_validity = READ_ONCE(NFS_I(inode)->cache_validity);
fs/nfs/inode.c
955
unsigned long cache_validity = READ_ONCE(NFS_I(inode)->cache_validity);
fs/nfs/namespace.c
153
int timeout = READ_ONCE(nfs_mountpoint_expiry_timeout);
fs/nfs/namespace.c
259
int timeout = READ_ONCE(nfs_mountpoint_expiry_timeout);
fs/nfs/nfs4proc.c
312
cache_validity = READ_ONCE(NFS_I(inode)->cache_validity) | flags;
fs/nfs/nfs4proc.c
5782
cache_validity |= READ_ONCE(NFS_I(inode)->cache_validity);
fs/nfsd/filecache.c
253
filemap_check_wb_err(file->f_mapping, READ_ONCE(file->f_wb_err)))
fs/nfsd/nfs4callback.c
486
if (likely(target == READ_ONCE(ses->se_cb_highest_slot)))
fs/nfsd/nfs4proc.c
1962
since = READ_ONCE(dst->f_wb_err);
fs/nfsd/nfs4state.c
1234
if ((READ_ONCE(f->f_mode) & FMODE_NOCMTIME) == 0)
fs/nfsd/nfs4state.c
6646
if (READ_ONCE(nn->grace_ended))
fs/nfsd/nfs4state.c
6666
if (READ_ONCE(nn->grace_end_forced))
fs/nfsd/vfs.c
1483
since = READ_ONCE(file->f_wb_err);
fs/nfsd/vfs.c
1683
errseq_t since = READ_ONCE(nf->nf_file->f_wb_err);
fs/nfsd/vfs.c
693
since = READ_ONCE(dst->f_wb_err);
fs/notify/fanotify/fanotify_user.c
1386
conn = READ_ONCE(old->connector);
fs/notify/fsnotify.c
136
READ_ONCE(inode->i_fsnotify_mask));
fs/notify/fsnotify.c
138
READ_ONCE(inode->i_sb->s_fsnotify_mask));
fs/notify/fsnotify.c
149
__u32 marks_mask = READ_ONCE(inode->i_fsnotify_mask) | mnt_mask |
fs/notify/fsnotify.c
150
READ_ONCE(inode->i_sb->s_fsnotify_mask);
fs/notify/fsnotify.c
186
READ_ONCE(real_mount(path->mnt)->mnt_fsnotify_mask) : 0;
fs/notify/fsnotify.c
543
marks_mask |= READ_ONCE(sb->s_fsnotify_mask);
fs/notify/fsnotify.c
545
marks_mask |= READ_ONCE(mnt->mnt_fsnotify_mask);
fs/notify/fsnotify.c
547
marks_mask |= READ_ONCE(inode->i_fsnotify_mask);
fs/notify/fsnotify.c
549
marks_mask |= READ_ONCE(inode2->i_fsnotify_mask);
fs/notify/fsnotify.c
551
marks_mask |= READ_ONCE(mnt_data->ns->n_fsnotify_mask);
fs/notify/fsnotify.c
640
mnt_mask = READ_ONCE(real_mount(file->f_path.mnt)->mnt_fsnotify_mask);
fs/notify/inotify/inotify_fsnotify.c
87
wd = READ_ONCE(i_mark->wd);
fs/notify/inotify/inotify_user.c
572
int do_inode = (new_mask & ~READ_ONCE(inode->i_fsnotify_mask));
fs/notify/mark.c
136
return READ_ONCE(*fsnotify_conn_mask_p(conn));
fs/notify/mark.c
393
struct fsnotify_mark_connector *conn = READ_ONCE(mark->connector);
fs/nsfs.c
162
kinfo->nr_mounts = READ_ONCE(mnt_ns->nr_mounts);
fs/nsfs.c
602
} else if (!READ_ONCE(to_pid_ns(ns)->child_reaper)) {
fs/overlayfs/file.c
138
struct file *upperfile = READ_ONCE(of->upperfile);
fs/overlayfs/ovl_entry.h
150
return lowerdata ? READ_ONCE(lowerdata->dentry) : NULL;
fs/overlayfs/ovl_entry.h
192
return READ_ONCE(oi->__upperdentry);
fs/overlayfs/readdir.c
984
realfile = READ_ONCE(od->upperfile);
fs/overlayfs/super.c
99
parent = READ_ONCE(d->d_parent);
fs/overlayfs/util.c
280
path->mnt = READ_ONCE(lowerdata->layer)->mnt;
fs/pidfs.c
1010
xattrs = READ_ONCE(attr->xattrs);
fs/pidfs.c
1032
xattrs = READ_ONCE(attr->xattrs);
fs/pidfs.c
365
attr = READ_ONCE(pid->attr);
fs/pidfs.c
723
attr = READ_ONCE(pid->attr);
fs/pidfs.c
767
xattrs = READ_ONCE(attr->xattrs);
fs/pidfs.c
833
attr = READ_ONCE(pid->attr);
fs/pidfs.c
957
attr = READ_ONCE(pid->attr);
fs/pipe.c
232
union pipe_index idx = { .head_tail = READ_ONCE(pipe->head_tail) };
fs/pipe.c
233
unsigned int writers = READ_ONCE(pipe->writers);
fs/pipe.c
423
union pipe_index idx = { .head_tail = READ_ONCE(pipe->head_tail) };
fs/pipe.c
424
unsigned int max_usage = READ_ONCE(pipe->max_usage);
fs/pipe.c
427
!READ_ONCE(pipe->readers);
fs/pipe.c
685
idx.head_tail = READ_ONCE(pipe->head_tail);
fs/pipe.c
775
unsigned long soft_limit = READ_ONCE(pipe_user_pages_soft);
fs/pipe.c
782
unsigned long hard_limit = READ_ONCE(pipe_user_pages_hard);
fs/pipe.c
798
unsigned int max_size = READ_ONCE(pipe_max_size);
fs/proc/array.c
524
rsslim = READ_ONCE(sig->rlim[RLIMIT_RSS].rlim_cur);
fs/proc/base.c
888
ptrace_active = READ_ONCE(task->ptrace) &&
fs/proc/base.c
889
READ_ONCE(task->mm) == mm &&
fs/proc/base.c
890
READ_ONCE(task->parent) == current;
fs/proc/generic.c
143
nlink_t nlink = READ_ONCE(de->nlink);
fs/proc/proc_sysctl.c
934
head = READ_ONCE(PROC_I(inode)->sysctl);
fs/proc_namespace.c
31
event = READ_ONCE(ns->event);
fs/pstore/platform.c
281
unsigned int remaining = READ_ONCE(kmsg_bytes);
fs/select.c
240
if (!READ_ONCE(pwq->triggered))
fs/smb/client/cifsglob.h
2342
unsigned long delay = READ_ONCE(server->reconn_delay);
fs/smb/client/cifsglob.h
2354
unsigned int oplock = READ_ONCE(cinode->oplock);
fs/smb/client/dfs_cache.c
182
READ_ONCE(ce->tgthint) == t ? " (target hint)" : "");
fs/smb/client/dfs_cache.c
234
READ_ONCE(ce->tgthint) == t ? " (target hint)" : "");
fs/smb/client/dfs_cache.c
343
struct cache_dfs_tgt *t = READ_ONCE(ce->tgthint);
fs/smb/client/dfs_cache.c
627
target = READ_ONCE(ce->tgthint);
fs/smb/client/dfs_cache.c
813
if (READ_ONCE(ce->tgthint) == t)
fs/smb/client/dfs_cache.c
960
t = READ_ONCE(ce->tgthint);
fs/smb/client/file.c
1421
unsigned int oplock = READ_ONCE(cinode->oplock);
fs/smb/client/file.c
3184
oplock = READ_ONCE(cinode->oplock);
fs/smb/client/smb2ops.c
1831
u32 max_chunks = READ_ONCE(tcon->max_chunks);
fs/smb/client/smb2ops.c
1832
u32 max_bytes_copy = READ_ONCE(tcon->max_bytes_copy);
fs/smb/client/smb2ops.c
1833
u32 max_bytes_chunk = READ_ONCE(tcon->max_bytes_chunk);
fs/smb/client/smb2ops.c
4155
unsigned int old_oplock = READ_ONCE(cinode->oplock);
fs/smb/client/smb2ops.c
4159
new_oplock = READ_ONCE(cinode->oplock);
fs/smb/client/smb2transport.c
689
switch (READ_ONCE(server->tcpStatus)) {
fs/smb/client/smb2transport.c
703
switch (READ_ONCE(ses->ses_status)) {
fs/smb/server/connection.c
414
READ_ONCE(conn->status));
fs/smb/server/connection.h
196
return READ_ONCE(conn->status) == KSMBD_SESS_GOOD;
fs/smb/server/connection.h
201
return READ_ONCE(conn->status) == KSMBD_SESS_NEED_NEGOTIATE;
fs/smb/server/connection.h
206
return READ_ONCE(conn->status) == KSMBD_SESS_NEED_SETUP;
fs/smb/server/connection.h
211
return READ_ONCE(conn->status) == KSMBD_SESS_NEED_RECONNECT;
fs/smb/server/connection.h
216
return READ_ONCE(conn->status) == KSMBD_SESS_EXITING;
fs/smb/server/connection.h
221
return READ_ONCE(conn->status) == KSMBD_SESS_RELEASING;
fs/smb/server/server.h
65
return READ_ONCE(server_conf.state) == SERVER_STATE_RUNNING;
fs/smb/server/server.h
70
return READ_ONCE(server_conf.state) < SERVER_STATE_RESETTING;
fs/splice.c
272
unsigned int max_usage = READ_ONCE(pipe->max_usage);
fs/splice.c
50
fmode_t fmode = READ_ONCE(file->f_mode);
fs/tracefs/event_inode.c
604
ei = READ_ONCE(ti->private);
fs/tracefs/inode.c
347
tmp_mode = READ_ONCE(inode->i_mode) & ~S_IALLUGO;
fs/userfaultfd.c
451
if (unlikely(READ_ONCE(ctx->released))) {
fs/userfaultfd.c
520
if (likely(must_wait && !READ_ONCE(ctx->released))) {
fs/userfaultfd.c
582
if (READ_ONCE(ctx->released) ||
fs/xfs/libxfs/xfs_bmap.c
4410
*seq = READ_ONCE(ifp->if_seq);
fs/xfs/libxfs/xfs_bmap.c
4458
*seq = READ_ONCE(ifp->if_seq);
fs/xfs/libxfs/xfs_iext_tree.c
629
WRITE_ONCE(ifp->if_seq, READ_ONCE(ifp->if_seq) + 1);
fs/xfs/xfs_aops.c
309
if (XFS_WPC(wpc)->data_seq != READ_ONCE(ip->i_df.if_seq)) {
fs/xfs/xfs_aops.c
315
XFS_WPC(wpc)->cow_seq != READ_ONCE(ip->i_cowfp->if_seq)) {
fs/xfs/xfs_aops.c
385
XFS_WPC(wpc)->cow_seq = READ_ONCE(ip->i_cowfp->if_seq);
fs/xfs/xfs_aops.c
408
XFS_WPC(wpc)->data_seq = READ_ONCE(ip->i_df.if_seq);
fs/xfs/xfs_error.c
144
randfactor = READ_ONCE(mp->m_errortag[error_tag]);
fs/xfs/xfs_error.c
161
unsigned int delay = READ_ONCE(mp->m_errortag[error_tag]);
fs/xfs/xfs_error.c
222
val = READ_ONCE(src_mp->m_errortag[i]);
fs/xfs/xfs_error.c
83
READ_ONCE(mp->m_errortag[to_attr(attr)->tag]));
fs/xfs/xfs_extent_busy.c
628
if (busy_gen != READ_ONCE(eb->eb_gen))
fs/xfs/xfs_extent_busy.c
638
if (busy_gen != READ_ONCE(eb->eb_gen))
fs/xfs/xfs_extent_busy.c
710
*busy_gen = READ_ONCE(eb->eb_gen);
fs/xfs/xfs_icache.c
1763
first_index = READ_ONCE(pag->pag_ici_reclaim_cursor);
fs/xfs/xfs_icache.c
1985
trace_xfs_inodegc_worker(mp, READ_ONCE(gc->shrinker_hits));
fs/xfs/xfs_icache.c
2193
items = READ_ONCE(gc->items);
fs/xfs/xfs_icache.c
2195
shrinker_hits = READ_ONCE(gc->shrinker_hits);
fs/xfs/xfs_icache.c
2316
unsigned int h = READ_ONCE(gc->shrinker_hits);
fs/xfs/xfs_iomap.c
65
return READ_ONCE(ip->i_af.if_seq);
fs/xfs/xfs_iomap.c
67
cookie = (u64)READ_ONCE(ip->i_cowfp->if_seq) << 32;
fs/xfs/xfs_iomap.c
68
return cookie | READ_ONCE(ip->i_df.if_seq);
fs/xfs/xfs_log.c
136
free_bytes = log->l_logsize - READ_ONCE(log->l_tail_space) -
fs/xfs/xfs_log_cil.c
82
return lip->li_seq == READ_ONCE(cil->xc_current_sequence);
fs/xfs/xfs_log_priv.h
659
cur_cycle = READ_ONCE(log->l_curr_cycle);
fs/xfs/xfs_log_priv.h
661
cur_block = READ_ONCE(log->l_curr_block);
fs/xfs/xfs_trace.h
1554
__entry->tail_space = READ_ONCE(log->l_tail_space);
fs/xfs/xfs_trace.h
4101
__entry->forkseq = READ_ONCE(xfs_ifork_ptr(ip, whichfork)->if_seq);
fs/xfs/xfs_trans_priv.h
118
return READ_ONCE(ailp->ail_target);
fs/xfs/xfs_zone_alloc.c
404
if (!READ_ONCE(rtg->rtg_open_zone))
fs/xfs/xfs_zone_gc.c
1050
if (READ_ONCE(chunk->state) != XFS_GC_BIO_DONE)
fs/xfs/xfs_zone_gc.c
1057
if (READ_ONCE(chunk->state) != XFS_GC_BIO_DONE)
fs/xfs/xfs_zone_gc.c
1065
if (READ_ONCE(chunk->state) != XFS_GC_BIO_DONE)
include/asm-generic/atomic.h
130
#define arch_atomic_read(v) READ_ONCE((v)->counter)
include/asm-generic/barrier.h
150
__unqual_scalar_typeof(*p) ___p1 = READ_ONCE(*p); \
include/asm-generic/barrier.h
204
__unqual_scalar_typeof(*p) ___p1 = READ_ONCE(*p); \
include/asm-generic/barrier.h
250
VAL = READ_ONCE(*__PTR); \
include/asm-generic/bitops/lock.h
25
if (READ_ONCE(*p) & mask)
include/asm-generic/bitops/lock.h
64
old = READ_ONCE(*p);
include/asm-generic/percpu.h
142
___ret = READ_ONCE(*raw_cpu_ptr(&(pcp))); \
include/asm-generic/preempt.h
11
return READ_ONCE(current_thread_info()->preempt_count);
include/asm-generic/ticket_spinlock.h
80
arch_spinlock_t val = READ_ONCE(*lock);
include/clocksource/hyperv_timer.h
67
sequence = READ_ONCE(tsc_pg->tsc_sequence);
include/clocksource/hyperv_timer.h
76
scale = READ_ONCE(tsc_pg->tsc_scale);
include/clocksource/hyperv_timer.h
77
offset = READ_ONCE(tsc_pg->tsc_offset);
include/clocksource/hyperv_timer.h
86
} while (READ_ONCE(tsc_pg->tsc_sequence) != sequence);
include/drm/drm_connector.h
2459
return READ_ONCE(connector->registration_state) ==
include/drm/drm_mm.h
278
return READ_ONCE(mm->hole_stack.next);
include/drm/spsc_queue.h
103
next = READ_ONCE(node->next);
include/drm/spsc_queue.h
114
} while (unlikely(!(queue->head = READ_ONCE(node->next))));
include/drm/spsc_queue.h
98
node = READ_ONCE(queue->head);
include/linux/average.h
56
unsigned long internal = READ_ONCE(e->internal); \
include/linux/bitops.h
363
old__ = READ_ONCE(*(ptr)); \
include/linux/bitops.h
378
old__ = READ_ONCE(*(ptr)); \
include/linux/blk-mq.h
828
return READ_ONCE(rq->state);
include/linux/bpf.h
2473
while ((prog = READ_ONCE(item->prog))) {
include/linux/bpf.h
2515
while ((prog = READ_ONCE(item->prog))) {
include/linux/bpf.h
3590
while ((prog = READ_ONCE(item->prog))) {
include/linux/bpf_mprog.h
118
t.prog = READ_ONCE(fp->prog); \
include/linux/bpf_mprog.h
126
(p = READ_ONCE(fp->prog)); \
include/linux/cgroup-defs.h
937
return READ_ONCE(skcd->prioidx);
include/linux/cgroup-defs.h
946
return READ_ONCE(skcd->classid);
include/linux/console.h
530
return data_race(READ_ONCE(con->flags));
include/linux/cpufreq.h
260
return READ_ONCE(per_cpu(cpufreq_pressure, cpu));
include/linux/dcache.h
546
return READ_ONCE(dentry->d_inode);
include/linux/dma-resv.h
452
return READ_ONCE(obj->lock.ctx);
include/linux/dynamic_queue_limits.h
142
if (READ_ONCE(dql->stall_thrs))
include/linux/dynamic_queue_limits.h
149
return READ_ONCE(dql->adj_limit) - READ_ONCE(dql->num_queued);
include/linux/entry-common.h
153
unsigned long work = READ_ONCE(current_thread_info()->syscall_work);
include/linux/entry-common.h
268
unsigned long work = READ_ONCE(current_thread_info()->syscall_work);
include/linux/eventpoll.h
49
if (likely(!READ_ONCE(file->f_ep)))
include/linux/filelock.h
252
return READ_ONCE(inode->i_flctx);
include/linux/filter.h
1645
while ((_prog = READ_ONCE(_item->prog))) { \
include/linux/fs.h
1345
return READ_ONCE(file->f_owner);
include/linux/fs.h
884
return READ_ONCE(inode->i_state.__state);
include/linux/fsnotify_backend.h
582
return READ_ONCE(sb->s_fsnotify_info);
include/linux/fsnotify_backend.h
676
__u32 parent_mask = READ_ONCE(inode->i_fsnotify_mask);
include/linux/generic-radix-tree.h
187
struct genradix_root *r = READ_ONCE(radix->root);
include/linux/hrtimer.h
297
return !!(READ_ONCE(timer->state) & HRTIMER_STATE_ENQUEUED);
include/linux/huge_mm.h
299
unsigned long mask = READ_ONCE(huge_anon_orders_always);
include/linux/huge_mm.h
302
mask |= READ_ONCE(huge_anon_orders_madvise);
include/linux/huge_mm.h
305
mask |= READ_ONCE(huge_anon_orders_inherit);
include/linux/huge_mm.h
532
return READ_ONCE(huge_zero_folio) == folio;
include/linux/huge_mm.h
537
return READ_ONCE(huge_zero_pfn) == (pfn & ~(HPAGE_PMD_NR - 1));
include/linux/hung_task.h
46
WARN_ON_ONCE(READ_ONCE(current->blocker));
include/linux/hyperv.h
217
write_loc = READ_ONCE(rbi->ring_buffer->write_index);
include/linux/hyperv.h
230
read_loc = READ_ONCE(rbi->ring_buffer->read_index);
include/linux/idr.h
69
return READ_ONCE(idr->idr_next);
include/linux/if_team.h
262
int en_port_count = READ_ONCE(team->en_port_count);
include/linux/inetdevice.h
56
#define IPV4_DEVCONF_RO(cnf, attr) READ_ONCE(IPV4_DEVCONF(cnf, attr))
include/linux/inetdevice.h
59
#define IPV4_DEVCONF_ALL_RO(net, attr) READ_ONCE(IPV4_DEVCONF_ALL(net, attr))
include/linux/inetdevice.h
64
return READ_ONCE(in_dev->cnf.data[index]);
include/linux/iommu.h
1619
return READ_ONCE(mm->iommu_mm);
include/linux/iommu.h
1624
struct iommu_mm_data *iommu_mm = READ_ONCE(mm->iommu_mm);
include/linux/iosys-map.h
364
val__ = READ_ONCE(*(type__ *)(vaddr__))
include/linux/jbd2.h
1712
READ_ONCE(journal->j_fs_dev_wb_err));
include/linux/kdb.h
217
return unlikely(READ_ONCE(kdb_printf_cpu) == raw_smp_processor_id());
include/linux/kvm_host.h
2175
if (unlikely(READ_ONCE(kvm->mmu_invalidate_in_progress)) &&
include/linux/kvm_host.h
2180
return READ_ONCE(kvm->mmu_invalidate_seq) != mmu_seq;
include/linux/kvm_host.h
2272
return READ_ONCE(vcpu->requests);
include/linux/list.h
381
return READ_ONCE(head->next) == head;
include/linux/list.h
418
return list_is_head(next, head) && (next == READ_ONCE(head->prev));
include/linux/list.h
643
struct list_head *pos__ = READ_ONCE(head__->next); \
include/linux/list.h
657
struct list_head *pos__ = READ_ONCE(head__->prev); \
include/linux/list.h
975
return !READ_ONCE(h->pprev);
include/linux/list.h
984
return !READ_ONCE(h->first);
include/linux/list_bl.h
74
return !((unsigned long)READ_ONCE(h->first) & ~LIST_BL_LOCKMASK);
include/linux/list_nulls.h
84
return !READ_ONCE(h->pprev);
include/linux/list_nulls.h
89
return is_a_nulls(READ_ONCE(h->first));
include/linux/livepatch_sched.h
17
READ_ONCE(curr->__state) & TASK_FREEZABLE)
include/linux/llist.h
100
return READ_ONCE(node->next) != node;
include/linux/llist.h
218
return READ_ONCE(head->first) == NULL;
include/linux/llist.h
223
return READ_ONCE(node->next);
include/linux/llist.h
238
struct llist_node *first = READ_ONCE(head->first);
include/linux/local_lock_internal.h
151
if (READ_ONCE(__tl->acquired)) { \
include/linux/local_lock_internal.h
168
if (READ_ONCE(__tl->acquired)) { \
include/linux/local_lock_internal.h
180
#define __local_lock_is_locked(lock) READ_ONCE(this_cpu_ptr(lock)->acquired)
include/linux/memcontrol.h
1628
return READ_ONCE(memcg->socket_pressure);
include/linux/memcontrol.h
382
return READ_ONCE(objcg->memcg);
include/linux/memcontrol.h
490
unsigned long memcg_data = READ_ONCE(folio->memcg_data);
include/linux/memcontrol.h
604
*min = READ_ONCE(memcg->memory.emin);
include/linux/memcontrol.h
605
*low = READ_ONCE(memcg->memory.elow);
include/linux/memcontrol.h
629
return READ_ONCE(memcg->memory.elow) >=
include/linux/memcontrol.h
639
return READ_ONCE(memcg->memory.emin) >=
include/linux/memcontrol.h
908
return READ_ONCE(mz->lru_zone_size[zone_idx][lru]);
include/linux/mempool.h
37
return READ_ONCE(pool->curr_nr) >= pool->min_nr;
include/linux/mm.h
2184
old_flags = READ_ONCE(page->flags.f);
include/linux/mm_inline.h
146
unsigned long flags = READ_ONCE(folio->flags.f);
include/linux/mm_inline.h
159
unsigned long flags = READ_ONCE(folio->flags.f);
include/linux/mm_inline.h
251
return max(READ_ONCE(lrugen->max_seq) - gen + 1, READ_ONCE(lrugen->min_seq[type]));
include/linux/mm_inline.h
308
unsigned long refs = READ_ONCE(old->flags.f) & LRU_REFS_MASK;
include/linux/mmu_notifier.h
374
return READ_ONCE(interval_sub->invalidate_seq) != seq;
include/linux/mmzone.h
2128
struct mem_section_usage *usage = READ_ONCE(ms->usage);
include/linux/mmzone.h
2135
struct mem_section_usage *usage = READ_ONCE(ms->usage);
include/linux/mount.h
68
return READ_ONCE(mnt->mnt_idmap);
include/linux/netdevice.h
304
((((dev)->hard_header_len + READ_ONCE((dev)->needed_headroom)) \
include/linux/netdevice.h
307
((((dev)->hard_header_len + READ_ONCE((dev)->needed_headroom) + (extra)) \
include/linux/netdevice.h
4648
unsigned int operstate = READ_ONCE(dev->operstate);
include/linux/netdevice.h
4808
if (READ_ONCE(txq->trans_start) != now)
include/linux/netdevice.h
4888
return READ_ONCE(txq->xmit_lock_owner) == cpu;
include/linux/netdevice.h
5455
READ_ONCE(dev->gro_max_size) :
include/linux/netdevice.h
5456
READ_ONCE(dev->gro_ipv4_max_size);
include/linux/netdevice.h
5464
READ_ONCE(dev->gso_max_size) :
include/linux/netdevice.h
5465
READ_ONCE(dev->gso_ipv4_max_size);
include/linux/netdevice.h
5615
u8 reg_state = READ_ONCE(dev->reg_state);
include/linux/netdevice.h
636
val = READ_ONCE(n->state);
include/linux/netdevice.h
742
int fb_tunnels_only_for_init_net = READ_ONCE(sysctl_fb_tunnels_only_for_init_net);
include/linux/netdevice.h
754
return READ_ONCE(sysctl_devconf_inherit_init_net);
include/linux/objpool.h
132
while (head != READ_ONCE(slot->last)) {
include/linux/objpool.h
147
if (READ_ONCE(slot->last) - head - 1 >= pool->nr_objs) {
include/linux/objpool.h
148
head = READ_ONCE(slot->head);
include/linux/objpool.h
153
obj = READ_ONCE(slot->entries[head & slot->mask]);
include/linux/objpool.h
197
tail = READ_ONCE(slot->tail);
include/linux/objpool.h
200
head = READ_ONCE(slot->head);
include/linux/page-flags.h
226
unsigned long head = READ_ONCE(page[1].compound_head);
include/linux/page-flags.h
284
unsigned long head = READ_ONCE(page->compound_head);
include/linux/page-flags.h
323
return READ_ONCE(page->compound_head) & 1 || page_is_fake_head(page);
include/linux/page-flags.h
329
READ_ONCE(page->compound_head) & 1;
include/linux/page-flags.h
335
return READ_ONCE(page->flags.f) == PAGE_POISON_PATTERN;
include/linux/pci.h
2671
return READ_ONCE(dev->error_state) == pci_channel_io_perm_failure;
include/linux/percpu-refcount.h
174
percpu_ptr = READ_ONCE(ref->percpu_count_ptr);
include/linux/percpu_counter.h
121
s64 ret = READ_ONCE(fbc->count);
include/linux/pgalloc_tag.h
153
old_flags = READ_ONCE(page->flags.f);
include/linux/pgtable.h
1992
pud_t pudval = READ_ONCE(*pud);
include/linux/pgtable.h
461
return READ_ONCE(*ptep);
include/linux/pgtable.h
468
return READ_ONCE(*pmdp);
include/linux/pgtable.h
475
return READ_ONCE(*pudp);
include/linux/pgtable.h
482
return READ_ONCE(*p4dp);
include/linux/pgtable.h
489
return READ_ONCE(*pgdp);
include/linux/pid_namespace.h
74
scope = max(scope, READ_ONCE(ns->memfd_noexec_scope));
include/linux/ptr_ring.h
172
return READ_ONCE(r->queue[r->consumer_head]);
include/linux/ptr_ring.h
197
return !r->queue[READ_ONCE(r->consumer_head)];
include/linux/rbtree.h
30
#define RB_EMPTY_ROOT(root) (READ_ONCE((root)->rb_node) == NULL)
include/linux/rcu_sync.h
36
return !READ_ONCE(rsp->gp_state); /* GP_IDLE */
include/linux/rculist.h
364
container_of(READ_ONCE(ptr), type, member)
include/linux/rculist.h
410
struct list_head *__next = READ_ONCE(__ptr->next); \
include/linux/rculist.h
430
struct list_head *__next = READ_ONCE(__ptr->next); \
include/linux/rculist.h
485
container_of((typeof(ptr))READ_ONCE(ptr), type, member)
include/linux/rcupdate.h
1181
rcu_callback_t func = READ_ONCE(rhp->func);
include/linux/rcupdate.h
176
if (!(preempt) && READ_ONCE((t)->rcu_tasks_holdout)) \
include/linux/rcupdate.h
507
typeof(*p) *local = (typeof(*p) *__force)READ_ONCE(p); \
include/linux/rcupdate.h
514
typeof(*p) *local = (typeof(*p) *__force)READ_ONCE(p); \
include/linux/rcupdate.h
528
typeof(p) local = READ_ONCE(p); \
include/linux/rcupdate.h
91
#define rcu_preempt_depth() READ_ONCE(current->rcu_read_lock_nesting)
include/linux/rtmutex.h
44
return READ_ONCE(lock->owner) != NULL;
include/linux/rtmutex.h
52
unsigned long owner = (unsigned long) READ_ONCE(lock->owner);
include/linux/sched.h
152
#define task_is_running(task) (READ_ONCE((task)->__state) == TASK_RUNNING)
include/linux/sched.h
154
#define task_is_traced(task) ((READ_ONCE(task->jobctl) & JOBCTL_TRACED) != 0)
include/linux/sched.h
155
#define task_is_stopped(task) ((READ_ONCE(task->jobctl) & JOBCTL_STOPPED) != 0)
include/linux/sched.h
156
#define task_is_stopped_or_traced(task) ((READ_ONCE(task->jobctl) & (JOBCTL_STOPPED | JOBCTL_TRACED)) != 0)
include/linux/sched.h
1689
return __task_state_index(READ_ONCE(tsk->__state), tsk->exit_state);
include/linux/sched.h
2193
struct mutex *blocked_on = READ_ONCE(p->blocked_on);
include/linux/sched.h
2218
struct mutex *blocked_on = READ_ONCE(p->blocked_on);
include/linux/sched.h
2259
return READ_ONCE(task_thread_info(p)->cpu);
include/linux/sched.h
2314
return READ_ONCE(owner->on_cpu) && !vcpu_is_preempted(task_cpu(owner));
include/linux/sched.h
318
#define get_current_state() READ_ONCE(current->__state)
include/linux/sched/cputime.h
80
if (!READ_ONCE(tsk->signal->posix_cputimers.timers_active))
include/linux/sched/mm.h
251
unsigned int pflags = READ_ONCE(current->flags);
include/linux/sched/signal.h
760
return READ_ONCE(task->signal->rlim[limit].rlim_cur);
include/linux/sched/signal.h
766
return READ_ONCE(task->signal->rlim[limit].rlim_max);
include/linux/seqlock.h
392
return unlikely(READ_ONCE(s->sequence) != start);
include/linux/seqlock.h
645
return READ_ONCE(s->seqcount.sequence);
include/linux/seqlock.h
676
return unlikely(READ_ONCE(s->seqcount.sequence) != start);
include/linux/serial_core.h
1043
u64 fifo_timeout = (u64)READ_ONCE(port->frame_time) * port->fifosize;
include/linux/skbuff.h
1420
READ_ONCE(fclones->skb2.sk) == sk;
include/linux/skbuff.h
1938
return READ_ONCE(list->next) == (const struct sk_buff *) list;
include/linux/skbuff.h
2247
struct sk_buff *skb = READ_ONCE(list_->prev);
include/linux/skbuff.h
2275
return READ_ONCE(list_->qlen);
include/linux/skmsg.h
329
return READ_ONCE(psock->msg_tot_len);
include/linux/srcutiny.h
148
idx = ((data_race(READ_ONCE(ssp->srcu_idx)) + 1) & 0x2) >> 1;
include/linux/srcutiny.h
151
data_race(READ_ONCE(ssp->srcu_lock_nesting[!idx])),
include/linux/srcutiny.h
152
data_race(READ_ONCE(ssp->srcu_lock_nesting[idx])),
include/linux/srcutiny.h
153
data_race(READ_ONCE(ssp->srcu_idx)),
include/linux/srcutiny.h
154
data_race(READ_ONCE(ssp->srcu_idx_max)));
include/linux/srcutiny.h
85
idx = ((READ_ONCE(ssp->srcu_idx) + 1) & 0x2) >> 1;
include/linux/srcutiny.h
86
WRITE_ONCE(ssp->srcu_lock_nesting[idx], READ_ONCE(ssp->srcu_lock_nesting[idx]) + 1);
include/linux/srcutree.h
292
struct srcu_ctr __percpu *scp = READ_ONCE(ssp->srcu_ctrp);
include/linux/srcutree.h
336
struct srcu_ctr __percpu *scp = READ_ONCE(ssp->srcu_ctrp);
include/linux/static_call.h
159
#define static_call_query(name) (READ_ONCE(STATIC_CALL_KEY(name).func))
include/linux/static_call.h
328
void *func = READ_ONCE(STATIC_CALL_KEY(name).func); \
include/linux/sunrpc/sched.h
174
#define RPC_SIGNALLED(t) (READ_ONCE(task->tk_rpc_status) == -ERESTARTSYS)
include/linux/swap.h
566
return READ_ONCE(vm_swappiness);
include/linux/swap.h
570
return READ_ONCE(vm_swappiness);
include/linux/swap.h
572
return READ_ONCE(memcg->swappiness);
include/linux/swap.h
577
return READ_ONCE(vm_swappiness);
include/linux/swiotlb.h
162
if (READ_ONCE(dev->dma_uses_io_tlb))
include/linux/task_work.h
26
return READ_ONCE(task->task_works);
include/linux/tcp.h
593
int somaxconn = READ_ONCE(sock_net(sk)->core.sysctl_somaxconn);
include/linux/tcp.h
626
u16 user_mss = READ_ONCE(tp->rx_opt.user_mss);
include/linux/thread_info.h
142
return READ_ONCE(ti->flags);
include/linux/trace_events.h
453
return !!READ_ONCE(call->prog_array);
include/linux/trace_recursion.h
129
unsigned int val = READ_ONCE(current->trace_recursion);
include/linux/tracepoint.h
373
it_func = READ_ONCE((it_func_ptr)->func); \
include/linux/udp.h
211
READ_ONCE(udp_sk(sk)->encap_rcv) &&
include/linux/user_namespace.h
160
return READ_ONCE(ns->rlimit_max[type]);
include/linux/vmw_vmci_defs.h
806
return READ_ONCE(*(unsigned long *)var);
include/net/act_api.h
82
if (READ_ONCE(tm->lastuse) != now)
include/net/act_api.h
84
if (unlikely(!READ_ONCE(tm->firstuse)))
include/net/act_api.h
92
dtm->install = jiffies_to_clock_t(now - READ_ONCE(stm->install));
include/net/act_api.h
93
dtm->lastuse = jiffies_to_clock_t(now - READ_ONCE(stm->lastuse));
include/net/act_api.h
95
firstuse = READ_ONCE(stm->firstuse);
include/net/act_api.h
99
dtm->expires = jiffies_to_clock_t(READ_ONCE(stm->expires));
include/net/addrconf.h
454
return !!READ_ONCE(idev->cnf.ignore_routes_with_linkdown);
include/net/af_vsock.h
271
return READ_ONCE(net->vsock.mode);
include/net/af_vsock.h
296
return READ_ONCE(net->vsock.child_ns_mode);
include/net/bluetooth/bluetooth.h
576
skb->priority = READ_ONCE(sk->sk_priority);
include/net/bluetooth/hci_core.h
1640
.tsflags = READ_ONCE(sk->sk_tsflags),
include/net/bonding.h
525
unsigned long tmp, ret = READ_ONCE(slave->target_last_arp_rx[0]);
include/net/bonding.h
529
tmp = READ_ONCE(slave->target_last_arp_rx[i]);
include/net/bonding.h
542
return READ_ONCE(slave->last_rx);
include/net/bonding.h
552
return READ_ONCE(slave->last_tx);
include/net/busy_poll.h
105
unsigned long bp_usec = READ_ONCE(sk->sk_ll_usec);
include/net/busy_poll.h
120
unsigned int napi_id = READ_ONCE(sk->sk_napi_id);
include/net/busy_poll.h
124
READ_ONCE(sk->sk_prefer_busy_poll),
include/net/busy_poll.h
125
READ_ONCE(sk->sk_busy_poll_budget) ?: BUSY_POLL_BUDGET);
include/net/busy_poll.h
152
if (unlikely(READ_ONCE(sk->sk_napi_id) != skb->napi_id))
include/net/busy_poll.h
174
if (!READ_ONCE(sk->sk_napi_id))
include/net/busy_poll.h
42
return READ_ONCE(sysctl_net_busy_poll);
include/net/busy_poll.h
47
return READ_ONCE(sk->sk_ll_usec) && !signal_pending(current);
include/net/busy_poll.h
89
unsigned long bp_usec = READ_ONCE(sysctl_net_busy_poll);
include/net/dst.h
252
if (unlikely(time != READ_ONCE(dst->lastuse))) {
include/net/dst.h
448
old = READ_ONCE(dst->expires);
include/net/dst.h
470
return INDIRECT_CALL_INET(READ_ONCE(skb_dst(skb)->output),
include/net/dst.h
480
return INDIRECT_CALL_INET(READ_ONCE(skb_dst(skb)->input),
include/net/dst.h
490
if (READ_ONCE(dst->obsolete))
include/net/dst.h
577
return READ_ONCE(dst->dev);
include/net/gro.h
542
if (gro->rx_count >= READ_ONCE(net_hotdata.gro_normal_batch))
include/net/inet6_hashtables.h
178
READ_ONCE(sk->sk_portpair) != ports ||
include/net/inet6_hashtables.h
184
return inet_sk_bound_dev_eq(net, READ_ONCE(sk->sk_bound_dev_if), dif,
include/net/inet_connection_sock.h
196
return READ_ONCE(sk->tcp_retransmit_timer.expires);
include/net/inet_connection_sock.h
202
return READ_ONCE(icsk->icsk_delack_timer.expires);
include/net/inet_connection_sock.h
293
return inet_csk_reqsk_queue_len(sk) > READ_ONCE(sk->sk_max_ack_backlog);
include/net/inet_connection_sock.h
325
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pingpong_thresh);
include/net/inet_connection_sock.h
336
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pingpong_thresh);
include/net/inet_hashtables.h
362
READ_ONCE(sk->sk_portpair) != ports ||
include/net/inet_hashtables.h
367
return inet_sk_bound_dev_eq(net, READ_ONCE(sk->sk_bound_dev_if), dif,
include/net/inet_sock.h
107
u32 mark = READ_ONCE(sk->sk_mark);
include/net/inet_sock.h
109
if (!mark && READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fwmark_accept))
include/net/inet_sock.h
118
int bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
include/net/inet_sock.h
122
if (!bound_dev_if && READ_ONCE(net->ipv4.sysctl_tcp_l3mdev_accept))
include/net/inet_sock.h
134
if (!READ_ONCE(net->ipv4.sysctl_tcp_l3mdev_accept))
include/net/inet_sock.h
155
return inet_bound_dev_eq(!!READ_ONCE(net->ipv4.sysctl_tcp_l3mdev_accept),
include/net/inet_sock.h
310
return READ_ONCE(inet->inet_flags) & IP_CMSG_ALL;
include/net/inet_sock.h
315
return inet_dsfield_to_dscp(READ_ONCE(inet->tos));
include/net/inet_sock.h
337
if (sk && READ_ONCE(sk->sk_state) == TCP_NEW_SYN_RECV)
include/net/inet_sock.h
339
if (sk && READ_ONCE(sk->sk_state) == TCP_TIME_WAIT)
include/net/inet_sock.h
349
if (sk && READ_ONCE(sk->sk_state) == TCP_NEW_SYN_RECV)
include/net/inet_sock.h
351
if (sk && READ_ONCE(sk->sk_state) == TCP_TIME_WAIT)
include/net/inet_sock.h
436
return READ_ONCE(net->ipv4.sysctl_ip_nonlocal_bind) ||
include/net/ip.h
102
ipcm->oif = READ_ONCE(inet->sk.sk_bound_dev_if);
include/net/ip.h
104
ipcm->protocol = READ_ONCE(inet->inet_num);
include/net/ip.h
355
u32 range = READ_ONCE(net->ipv4.ip_local_ports.range);
include/net/ip.h
377
return port < READ_ONCE(net->ipv4.sysctl_ip_prot_sock);
include/net/ip.h
404
(READ_ONCE((net)->ipv4.sysctl_fwmark_reflect) ? (mark) : 0)
include/net/ip.h
440
u8 pmtudisc = READ_ONCE(inet_sk(sk)->pmtudisc);
include/net/ip.h
449
u8 pmtudisc = READ_ONCE(inet_sk(sk)->pmtudisc);
include/net/ip.h
457
return READ_ONCE(inet_sk(sk)->pmtudisc) < IP_PMTUDISC_PROBE;
include/net/ip.h
462
u8 pmtudisc = READ_ONCE(inet_sk(sk)->pmtudisc);
include/net/ip.h
479
if (READ_ONCE(net->ipv4.sysctl_ip_fwd_use_pmtu) ||
include/net/ip.h
483
if (mtu && time_before(jiffies, READ_ONCE(rt->dst.expires)))
include/net/ip.h
492
mtu = READ_ONCE(dev->mtu);
include/net/ip.h
521
mtu = min(READ_ONCE(dst_dev(dst)->mtu), IP_MAX_MTU);
include/net/ip.h
97
.tos = READ_ONCE(inet->tos),
include/net/ip6_fib.h
290
*cookie = READ_ONCE(fn->fn_sernum);
include/net/ip6_fib.h
558
gc_args.timeout = READ_ONCE(net->ipv6.sysctl.ip6_rt_gc_interval);
include/net/ip6_route.h
282
if (np && READ_ONCE(np->pmtudisc) >= IPV6_PMTUDISC_PROBE) {
include/net/ip6_route.h
283
mtu = READ_ONCE(dst_dev(dst)->mtu);
include/net/ip6_route.h
293
u8 pmtudisc = READ_ONCE(inet6_sk(sk)->pmtudisc);
include/net/ip6_route.h
301
u8 pmtudisc = READ_ONCE(inet6_sk(sk)->pmtudisc);
include/net/ip6_route.h
348
mtu = READ_ONCE(idev->cnf.mtu6);
include/net/ip_fib.h
562
mp_seed = READ_ONCE(net->ipv4.sysctl_fib_multipath_hash_seed.mp_seed);
include/net/ip_tunnels.h
643
if (headroom > READ_ONCE(dev->needed_headroom))
include/net/ip_vs.h
1069
return READ_ONCE(ipvs->sysctl_sync_threshold[1]);
include/net/ip_vs.h
1074
return READ_ONCE(ipvs->sysctl_sync_refresh_period);
include/net/ip_vs.h
1099
return READ_ONCE(ipvs->sysctl_sync_ports);
include/net/ipv6.h
1289
(READ_ONCE(inet6_sk(sk)->srcprefs) & prefmask) | pref);
include/net/ipv6.h
398
READ_ONCE(sock_net(sk)->ipv6.flowlabel_has_excl))
include/net/ipv6.h
456
s32 accept_ra = READ_ONCE(idev->cnf.accept_ra);
include/net/ipv6.h
461
return READ_ONCE(idev->cnf.forwarding) ? accept_ra == 2 :
include/net/ipv6.h
831
hlimit = READ_ONCE(np->mcast_hops);
include/net/ipv6.h
833
hlimit = READ_ONCE(np->hop_limit);
include/net/ipv6.h
858
return READ_ONCE(net->ipv6.sysctl.ip_nonlocal_bind) ||
include/net/ipv6.h
889
auto_flowlabels = READ_ONCE(net->ipv6.sysctl.auto_flowlabels);
include/net/ipv6.h
904
if (READ_ONCE(net->ipv6.sysctl.flowlabel_state_ranges))
include/net/ipv6.h
912
switch (READ_ONCE(net->ipv6.sysctl.auto_flowlabels)) {
include/net/ipv6.h
938
return READ_ONCE(net->ipv6.sysctl.multipath_hash_policy);
include/net/ipv6.h
942
return READ_ONCE(net->ipv6.sysctl.multipath_hash_fields);
include/net/ipv6_frag.h
81
if (READ_ONCE(fq->q.fqdir->dead)) {
include/net/lwtunnel.h
141
dst->lwtstate->orig_output = READ_ONCE(dst->output);
include/net/lwtunnel.h
145
dst->lwtstate->orig_input = READ_ONCE(dst->input);
include/net/neighbour.h
337
if (READ_ONCE(n->confirmed) != now)
include/net/neighbour.h
477
if (READ_ONCE(neigh->used) != now)
include/net/neighbour.h
479
if (!(READ_ONCE(neigh->nud_state) & (NUD_CONNECTED | NUD_DELAY | NUD_PROBE)))
include/net/neighbour.h
511
hh_len = READ_ONCE(hh->hh_len);
include/net/neighbour.h
552
(READ_ONCE(n->nud_state) & NUD_CONNECTED) &&
include/net/neighbour.h
553
READ_ONCE(hh->hh_len))
include/net/neighbour.h
556
return READ_ONCE(n->output)(n, skb);
include/net/neighbour.h
99
#define NEIGH_VAR(p, attr) READ_ONCE(__NEIGH_VAR(p, attr))
include/net/net_namespace.h
358
return READ_ONCE(net->list.next);
include/net/netdev_queues.h
349
unsigned long trans_start = READ_ONCE(txq->trans_start);
include/net/netfilter/nf_conntrack.h
295
s32 timeout = READ_ONCE(ct->timeout) - nfct_time_stamp;
include/net/netfilter/nf_conntrack.h
302
return (__s32)(READ_ONCE(ct->timeout) - nfct_time_stamp) <= 0;
include/net/netfilter/nf_conntrack_ecache.h
119
if (local64_read(&e->timestamp) && READ_ONCE(e->cache) == 0)
include/net/netfilter/nf_tables.h
1577
return 1 << READ_ONCE(net->nft.gencursor);
include/net/netfilter/nf_tables.h
657
u32 gc_int = READ_ONCE(set->gc_int);
include/net/netfilter/nf_tables.h
844
READ_ONCE(nft_set_ext_timeout(ext)->timeout) == 0)
include/net/netfilter/nf_tables.h
847
return time_after_eq64(tstamp, READ_ONCE(nft_set_ext_timeout(ext)->expiration));
include/net/pkt_sched.h
133
return READ_ONCE(dev->mtu) + dev->hard_header_len;
include/net/proto_memory.h
21
return !!READ_ONCE(*prot->memory_pressure);
include/net/proto_memory.h
41
return !!READ_ONCE(*sk->sk_prot->memory_pressure);
include/net/proto_memory.h
71
if (unlikely(val >= READ_ONCE(net_hotdata.sysctl_mem_pcpu_rsv)))
include/net/proto_memory.h
82
if (unlikely(val <= -READ_ONCE(net_hotdata.sysctl_mem_pcpu_rsv)))
include/net/psp/functions.h
132
state = READ_ONCE(sk->sk_state);
include/net/raw.h
93
return inet_bound_dev_eq(READ_ONCE(net->ipv4.sysctl_raw_l3mdev_accept),
include/net/request_sock.h
197
return READ_ONCE(queue->rskq_accept_head) == NULL;
include/net/route.h
152
flowi4_init_output(fl4, sk->sk_bound_dev_if, READ_ONCE(sk->sk_mark),
include/net/route.h
207
flowi4_init_output(fl4, oif, sk ? READ_ONCE(sk->sk_mark) : 0, tos,
include/net/route.h
332
flowi4_init_output(fl4, oif, READ_ONCE(sk->sk_mark), ip_sock_rt_tos(sk),
include/net/route.h
394
hoplimit = READ_ONCE(net->ipv4.sysctl_ip_default_ttl);
include/net/route.h
50
return READ_ONCE(inet_sk(sk)->tos) & INET_DSCP_MASK;
include/net/rps.h
117
_sock_rps_record_flow_hash(READ_ONCE(sk->sk_rxhash));
include/net/rps.h
126
hash = READ_ONCE(sk->sk_rxhash);
include/net/rps.h
134
if (READ_ONCE(table->ents[index]) != RPS_NO_CPU)
include/net/rps.h
84
if (READ_ONCE(table->ents[index]) != val)
include/net/sch_generic.h
176
return READ_ONCE(qdisc->running);
include/net/sch_generic.h
181
return !(READ_ONCE(qdisc->state) & QDISC_STATE_NON_EMPTY);
include/net/sch_generic.h
193
return !READ_ONCE(qdisc->q.qlen);
include/net/sch_generic.h
219
if (READ_ONCE(qdisc->running))
include/net/sock.h
1099
return READ_ONCE(sk->sk_ack_backlog) > READ_ONCE(sk->sk_max_ack_backlog);
include/net/sock.h
1107
return READ_ONCE(sk->sk_wmem_queued) >> 1;
include/net/sock.h
1112
return READ_ONCE(sk->sk_sndbuf) - READ_ONCE(sk->sk_wmem_queued);
include/net/sock.h
1195
if (unlikely(READ_ONCE(sk->sk_incoming_cpu) != cpu))
include/net/sock.h
1207
if (unlikely(READ_ONCE(sk->sk_rxhash) != skb->hash))
include/net/sock.h
1247
if (unlikely(READ_ONCE(sk->sk_backlog.tail))) {
include/net/sock.h
1411
if (READ_ONCE(sk->sk_wmem_queued) >= READ_ONCE(sk->sk_sndbuf))
include/net/sock.h
1559
return READ_ONCE(sk->sk_prot->sysctl_mem[index]);
include/net/sock.h
1924
.mark = READ_ONCE(sk->sk_mark),
include/net/sock.h
1925
.tsflags = READ_ONCE(sk->sk_tsflags),
include/net/sock.h
1926
.priority = READ_ONCE(sk->sk_priority),
include/net/sock.h
2030
if (READ_ONCE(sk->sk_tx_queue_mapping) != tx_queue) {
include/net/sock.h
2037
if (time_is_before_jiffies(READ_ONCE(sk->sk_tx_queue_mapping_jiffies) + HZ))
include/net/sock.h
2062
unlikely(READ_ONCE(sk->sk_rx_queue_mapping) != rx_queue))
include/net/sock.h
2089
int res = READ_ONCE(sk->sk_rx_queue_mapping);
include/net/sock.h
2146
return READ_ONCE(sk->sk_ino);
include/net/sock.h
2152
return READ_ONCE(sk->sk_uid);
include/net/sock.h
2258
if (!READ_ONCE(sk->sk_dst_pending_confirm))
include/net/sock.h
2267
if (sk && READ_ONCE(sk->sk_dst_pending_confirm))
include/net/sock.h
2429
u32 txhash = READ_ONCE(sk->sk_txhash);
include/net/sock.h
2633
return wmem_alloc < (READ_ONCE(sk->sk_sndbuf) >> 1);
include/net/sock.h
2703
return noblock ? 0 : READ_ONCE(sk->sk_rcvtimeo);
include/net/sock.h
2708
return noblock ? 0 : READ_ONCE(sk->sk_sndtimeo);
include/net/sock.h
2713
int v = waitall ? len : min_t(int, READ_ONCE(sk->sk_rcvlowat), len);
include/net/sock.h
2805
return READ_ONCE(sk->sk_stamp);
include/net/sock.h
2833
u32 tsflags = READ_ONCE(sk->sk_tsflags);
include/net/sock.h
2869
if (READ_ONCE(sk->sk_flags) & FLAGS_RECV_CMSGS)
include/net/sock.h
2922
int family = READ_ONCE(sk->sk_family);
include/net/sock.h
3022
return (1 << READ_ONCE(sk->sk_state)) &
include/net/sock.h
3060
return READ_ONCE(*(int *)((void *)sock_net(sk) + proto->sysctl_wmem_offset));
include/net/sock.h
3062
return READ_ONCE(*proto->sysctl_wmem);
include/net/sock.h
3069
return READ_ONCE(*(int *)((void *)sock_net(sk) + proto->sysctl_rmem_offset));
include/net/sock.h
3071
return READ_ONCE(*proto->sysctl_rmem);
include/net/sock.h
3080
if (!sk || !sk_fullsock(sk) || READ_ONCE(sk->sk_pacing_shift) == val)
include/net/sock.h
3092
int bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
include/net/sock.h
3139
const struct proto *prot = READ_ONCE(sk->sk_prot);
include/net/sock.h
733
return READ_ONCE(sk->sk_peek_off);
include/net/sock.h
741
s32 off = READ_ONCE(sk->sk_peek_off);
include/net/tc_act/tc_gact.h
59
return READ_ONCE(a->tcfa_action) & TC_ACT_EXT_VAL_MASK;
include/net/tcp.h
1665
if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_slow_start_after_idle) ||
include/net/tcp.h
1718
return tcp_win_from_space(sk, READ_ONCE(sk->sk_rcvbuf) -
include/net/tcp.h
1719
READ_ONCE(sk->sk_backlog.len) -
include/net/tcp.h
1725
return tcp_win_from_space(sk, READ_ONCE(sk->sk_rcvbuf));
include/net/tcp.h
1760
rcvbuf = READ_ONCE(sk->sk_rcvbuf);
include/net/tcp.h
1769
int avail = READ_ONCE(tp->rcv_nxt) - READ_ONCE(tp->copied_seq);
include/net/tcp.h
1793
val = READ_ONCE(tp->keepalive_intvl);
include/net/tcp.h
1795
return val ? : READ_ONCE(net->ipv4.sysctl_tcp_keepalive_intvl);
include/net/tcp.h
1804
val = READ_ONCE(tp->keepalive_time);
include/net/tcp.h
1806
return val ? : READ_ONCE(net->ipv4.sysctl_tcp_keepalive_time);
include/net/tcp.h
1817
val = READ_ONCE(tp->keepalive_probes);
include/net/tcp.h
1819
return val ? : READ_ONCE(net->ipv4.sysctl_tcp_keepalive_probes);
include/net/tcp.h
1833
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fin_timeout);
include/net/tcp.h
2395
val = READ_ONCE(tp->notsent_lowat);
include/net/tcp.h
2397
return val ?: READ_ONCE(net->ipv4.sysctl_tcp_notsent_lowat);
include/net/tcp.h
2532
out->ao_key = READ_ONCE(ao->current_key);
include/net/tcp.h
310
return READ_ONCE(tcp_memory_pressure);
include/net/tcp.h
625
last_overflow = READ_ONCE(reuse->synq_overflow_ts);
include/net/tcp.h
633
last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp);
include/net/tcp.h
649
last_overflow = READ_ONCE(reuse->synq_overflow_ts);
include/net/tcp.h
656
last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp);
include/net/tcp.h
695
return READ_ONCE(net->ipv4.sysctl_tcp_ecn) ||
include/net/tcp.h
855
return READ_ONCE(inet_csk(sk)->icsk_rto_max);
include/net/tcp.h
882
u32 rto_min = READ_ONCE(inet_csk(sk)->icsk_rto_min);
include/net/tcp_ecn.h
139
if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn_fallback))
include/net/tcp_ecn.h
602
u8 tcp_ecn = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn);
include/net/tcp_ecn.h
636
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn_fallback)) {
include/net/tcp_ecn.h
665
u32 ecn_beacon = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn_option_beacon);
include/net/udp.h
384
return sk_rmem_alloc_get(sk) - READ_ONCE(udp_sk(sk)->forward_deficit);
include/net/udp.h
391
return inet_bound_dev_eq(!!READ_ONCE(net->ipv4.sysctl_udp_l3mdev_accept),
include/net/udp_tunnel.h
232
if (READ_ONCE(sk->sk_family) == PF_INET6)
include/net/udplite.h
74
u16 pcslen = READ_ONCE(udp_sk(sk)->pcslen);
include/net/xfrm.h
1706
struct net_device *dev = READ_ONCE(xdo->dev);
include/net/xfrm.h
2083
struct net_device *dev = READ_ONCE(xso->dev);
include/rv/da_monitor.h
115
curr_state = READ_ONCE(da_mon->curr_state);
include/rv/da_monitor.h
156
curr_state = READ_ONCE(da_mon->curr_state);
include/sound/core.h
155
return READ_ONCE(card->power_state);
include/trace/events/mptcp.h
53
__entry->pace = READ_ONCE(ssk->sk_pacing_rate);
include/trace/events/netfs.h
496
struct address_space *__m = READ_ONCE(folio->mapping);
include/trace/events/sock.h
113
__entry->sysctl_mem[0] = READ_ONCE(prot->sysctl_mem[0]);
include/trace/events/sock.h
114
__entry->sysctl_mem[1] = READ_ONCE(prot->sysctl_mem[1]);
include/trace/events/sock.h
115
__entry->sysctl_mem[2] = READ_ONCE(prot->sysctl_mem[2]);
include/trace/events/sock.h
121
__entry->wmem_queued = READ_ONCE(sk->sk_wmem_queued);
include/trace/events/sock.h
86
__entry->sk_rcvbuf = READ_ONCE(sk->sk_rcvbuf);
include/vdso/helpers.h
14
while (unlikely((seq = READ_ONCE(vc->seq)) & 1))
include/vdso/helpers.h
27
seq = READ_ONCE(vc->seq);
io_uring/advise.c
39
ma->addr = READ_ONCE(sqe->addr);
io_uring/advise.c
40
ma->len = READ_ONCE(sqe->off);
io_uring/advise.c
42
ma->len = READ_ONCE(sqe->len);
io_uring/advise.c
43
ma->advice = READ_ONCE(sqe->fadvise_advice);
io_uring/advise.c
86
fa->offset = READ_ONCE(sqe->off);
io_uring/advise.c
87
fa->len = READ_ONCE(sqe->addr);
io_uring/advise.c
89
fa->len = READ_ONCE(sqe->len);
io_uring/advise.c
90
fa->advice = READ_ONCE(sqe->fadvise_advice);
io_uring/cancel.c
149
cancel->addr = READ_ONCE(sqe->addr);
io_uring/cancel.c
150
cancel->flags = READ_ONCE(sqe->cancel_flags);
io_uring/cancel.c
156
cancel->fd = READ_ONCE(sqe->fd);
io_uring/cancel.c
161
cancel->opcode = READ_ONCE(sqe->len);
io_uring/cmd_net.c
148
uaddr = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/cmd_net.c
149
ulen = u64_to_user_ptr(READ_ONCE(sqe->addr3));
io_uring/cmd_net.c
150
peer = READ_ONCE(sqe->optlen);
io_uring/cmd_net.c
160
struct proto *prot = READ_ONCE(sk->sk_prot);
io_uring/cmd_net.c
19
level = READ_ONCE(sqe->level);
io_uring/cmd_net.c
23
optval = u64_to_user_ptr(READ_ONCE(sqe->optval));
io_uring/cmd_net.c
24
optname = READ_ONCE(sqe->optname);
io_uring/cmd_net.c
25
optlen = READ_ONCE(sqe->optlen);
io_uring/cmd_net.c
47
optval = u64_to_user_ptr(READ_ONCE(sqe->optval));
io_uring/cmd_net.c
48
optname = READ_ONCE(sqe->optname);
io_uring/cmd_net.c
49
optlen = READ_ONCE(sqe->optlen);
io_uring/cmd_net.c
50
level = READ_ONCE(sqe->level);
io_uring/epoll.c
36
epoll->epfd = READ_ONCE(sqe->fd);
io_uring/epoll.c
37
epoll->op = READ_ONCE(sqe->len);
io_uring/epoll.c
38
epoll->fd = READ_ONCE(sqe->off);
io_uring/epoll.c
43
ev = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/epoll.c
74
iew->maxevents = READ_ONCE(sqe->len);
io_uring/epoll.c
75
iew->events = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/eventfd.c
86
if (READ_ONCE(rings->cq_flags) & IORING_CQ_EVENTFD_DISABLED)
io_uring/fdinfo.c
104
sq_idx = READ_ONCE(ctx->sq_array[entry & sq_mask]);
io_uring/fdinfo.c
109
opcode = READ_ONCE(sqe->opcode);
io_uring/fdinfo.c
37
unsigned int mode = READ_ONCE(ctx->napi_track_mode);
io_uring/fdinfo.c
65
unsigned int sq_head = READ_ONCE(r->sq.head);
io_uring/fdinfo.c
66
unsigned int sq_tail = READ_ONCE(r->sq.tail);
io_uring/fdinfo.c
67
unsigned int cq_head = READ_ONCE(r->cq.head);
io_uring/fdinfo.c
68
unsigned int cq_tail = READ_ONCE(r->cq.tail);
io_uring/fs.c
118
un->dfd = READ_ONCE(sqe->fd);
io_uring/fs.c
120
un->flags = READ_ONCE(sqe->unlink_flags);
io_uring/fs.c
124
fname = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/fs.c
170
mkd->dfd = READ_ONCE(sqe->fd);
io_uring/fs.c
171
mkd->mode = READ_ONCE(sqe->len);
io_uring/fs.c
173
fname = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/fs.c
216
sl->new_dfd = READ_ONCE(sqe->fd);
io_uring/fs.c
217
oldpath = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/fs.c
218
newpath = u64_to_user_ptr(READ_ONCE(sqe->addr2));
io_uring/fs.c
262
lnk->old_dfd = READ_ONCE(sqe->fd);
io_uring/fs.c
263
lnk->new_dfd = READ_ONCE(sqe->len);
io_uring/fs.c
264
oldf = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/fs.c
265
newf = u64_to_user_ptr(READ_ONCE(sqe->addr2));
io_uring/fs.c
266
lnk->flags = READ_ONCE(sqe->hardlink_flags);
io_uring/fs.c
61
ren->old_dfd = READ_ONCE(sqe->fd);
io_uring/fs.c
62
oldf = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/fs.c
63
newf = u64_to_user_ptr(READ_ONCE(sqe->addr2));
io_uring/fs.c
64
ren->new_dfd = READ_ONCE(sqe->len);
io_uring/fs.c
65
ren->flags = READ_ONCE(sqe->rename_flags);
io_uring/futex.c
136
iof->uaddr = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/futex.c
137
iof->futex_val = READ_ONCE(sqe->addr2);
io_uring/futex.c
138
iof->futex_mask = READ_ONCE(sqe->addr3);
io_uring/futex.c
139
flags = READ_ONCE(sqe->fd);
io_uring/futex.c
183
iof->uaddr = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/futex.c
184
iof->futex_nr = READ_ONCE(sqe->len);
io_uring/io_uring.c
1200
check_cq = READ_ONCE(ctx->check_cq);
io_uring/io_uring.c
1725
req->opcode = opcode = READ_ONCE(sqe->opcode);
io_uring/io_uring.c
1727
sqe_flags = READ_ONCE(sqe->flags);
io_uring/io_uring.c
1729
req->cqe.user_data = READ_ONCE(sqe->user_data);
io_uring/io_uring.c
1767
req->buf_index = READ_ONCE(sqe->buf_group);
io_uring/io_uring.c
1799
req->cqe.fd = READ_ONCE(sqe->fd);
io_uring/io_uring.c
1812
personality = READ_ONCE(sqe->personality);
io_uring/io_uring.c
1983
head = READ_ONCE(ctx->sq_array[head]);
io_uring/io_uring.c
1986
READ_ONCE(ctx->rings->sq_dropped) + 1);
io_uring/io_uring.c
2497
ext_arg->min_time = READ_ONCE(w->min_wait_usec) * NSEC_PER_USEC;
io_uring/io_uring.c
2498
ext_arg->sig = u64_to_user_ptr(READ_ONCE(w->sigmask));
io_uring/io_uring.c
2499
ext_arg->argsz = READ_ONCE(w->sigmask_sz);
io_uring/io_uring.c
2501
ext_arg->ts.tv_sec = READ_ONCE(w->ts.tv_sec);
io_uring/io_uring.c
2502
ext_arg->ts.tv_nsec = READ_ONCE(w->ts.tv_nsec);
io_uring/io_uring.c
3084
int disabled = READ_ONCE(sysctl_io_uring_disabled);
io_uring/io_uring.c
641
WRITE_ONCE(r->cq_overflow, READ_ONCE(r->cq_overflow) + 1);
io_uring/io_uring.h
166
dist = READ_ONCE(rings->cq.tail) - (int) iowq->cq_tail;
io_uring/io_uring.h
457
return READ_ONCE(r->sq.tail) - READ_ONCE(r->sq.head) == ctx->sq_entries;
io_uring/kbuf.c
210
buf_len = READ_ONCE(buf->len);
io_uring/kbuf.c
214
req->buf_index = READ_ONCE(buf->bid);
io_uring/kbuf.c
216
sel.addr = u64_to_user_ptr(READ_ONCE(buf->addr));
io_uring/kbuf.c
266
u32 len = READ_ONCE(buf->len);
io_uring/kbuf.c
297
req->buf_index = READ_ONCE(buf->bid);
io_uring/kbuf.c
299
u32 len = READ_ONCE(buf->len);
io_uring/kbuf.c
312
iov->iov_base = u64_to_user_ptr(READ_ONCE(buf->addr));
io_uring/kbuf.c
46
buf_len = READ_ONCE(buf->len);
io_uring/kbuf.c
489
tmp = READ_ONCE(sqe->fd);
io_uring/kbuf.c
495
p->bgid = READ_ONCE(sqe->buf_group);
io_uring/kbuf.c
508
tmp = READ_ONCE(sqe->fd);
io_uring/kbuf.c
51
WRITE_ONCE(buf->addr, READ_ONCE(buf->addr) + this_len);
io_uring/kbuf.c
512
p->addr = READ_ONCE(sqe->addr);
io_uring/kbuf.c
513
p->len = READ_ONCE(sqe->len);
io_uring/kbuf.c
525
p->bgid = READ_ONCE(sqe->buf_group);
io_uring/kbuf.c
526
tmp = READ_ONCE(sqe->off);
io_uring/mock_file.c
225
uarg = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/mock_file.c
226
uarg_size = READ_ONCE(sqe->len);
io_uring/mock_file.c
284
uarg = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/mock_file.c
285
uarg_size = READ_ONCE(sqe->len);
io_uring/mock_file.c
74
ubuf = u64_to_user_ptr(READ_ONCE(sqe->addr3));
io_uring/mock_file.c
75
iovec = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/mock_file.c
76
iovec_len = READ_ONCE(sqe->len);
io_uring/mock_file.c
77
flags = READ_ONCE(sqe->file_index);
io_uring/msg_ring.c
267
msg->user_data = READ_ONCE(sqe->off);
io_uring/msg_ring.c
268
msg->len = READ_ONCE(sqe->len);
io_uring/msg_ring.c
269
msg->cmd = READ_ONCE(sqe->addr);
io_uring/msg_ring.c
270
msg->src_fd = READ_ONCE(sqe->addr3);
io_uring/msg_ring.c
271
msg->dst_fd = READ_ONCE(sqe->file_index);
io_uring/msg_ring.c
272
msg->flags = READ_ONCE(sqe->msg_ring_flags);
io_uring/napi.c
117
if (time_after(jiffies, READ_ONCE(e->timeout))) {
io_uring/napi.c
187
if (time_after(jiffies, READ_ONCE(e->timeout)))
io_uring/napi.c
199
if (READ_ONCE(ctx->napi_track_mode) == IO_URING_NAPI_TRACKING_STATIC)
io_uring/napi.c
239
u64 sys_dt = READ_ONCE(sysctl_net_busy_poll) * NSEC_PER_USEC;
io_uring/napi.c
362
iowq->napi_busy_poll_dt = READ_ONCE(ctx->napi_busy_poll_dt);
io_uring/napi.c
369
iowq->napi_prefer_busy_poll = READ_ONCE(ctx->napi_prefer_busy_poll);
io_uring/napi.c
383
if (!READ_ONCE(ctx->napi_busy_poll_dt))
io_uring/napi.h
47
if (READ_ONCE(ctx->napi_track_mode) != IO_URING_NAPI_TRACKING_DYNAMIC)
io_uring/napi.h
52
__io_napi_add_id(ctx, READ_ONCE(sock->sk->sk_napi_id));
io_uring/net.c
1258
ifq_idx = READ_ONCE(sqe->zcrx_ifq_idx);
io_uring/net.c
1263
zc->len = READ_ONCE(sqe->len);
io_uring/net.c
1264
zc->flags = READ_ONCE(sqe->ioprio);
io_uring/net.c
1265
if (READ_ONCE(sqe->msg_flags))
io_uring/net.c
131
shutdown->how = READ_ONCE(sqe->len);
io_uring/net.c
1340
if (unlikely(READ_ONCE(sqe->__pad2[0]) || READ_ONCE(sqe->addr3)))
io_uring/net.c
1354
zc->flags = READ_ONCE(sqe->ioprio);
io_uring/net.c
1367
zc->len = READ_ONCE(sqe->len);
io_uring/net.c
1368
zc->msg_flags = READ_ONCE(sqe->msg_flags) | MSG_NOSIGNAL | MSG_ZEROCOPY;
io_uring/net.c
1369
req->buf_index = READ_ONCE(sqe->buf_index);
io_uring/net.c
1622
accept->addr = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/net.c
1623
accept->addr_len = u64_to_user_ptr(READ_ONCE(sqe->addr2));
io_uring/net.c
1624
accept->flags = READ_ONCE(sqe->accept_flags);
io_uring/net.c
1626
accept->iou_flags = READ_ONCE(sqe->ioprio);
io_uring/net.c
1630
accept->file_slot = READ_ONCE(sqe->file_index);
io_uring/net.c
1726
sock->domain = READ_ONCE(sqe->fd);
io_uring/net.c
1727
sock->type = READ_ONCE(sqe->off);
io_uring/net.c
1728
sock->protocol = READ_ONCE(sqe->len);
io_uring/net.c
1729
sock->file_slot = READ_ONCE(sqe->file_index);
io_uring/net.c
1781
conn->addr = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/net.c
1782
conn->addr_len = READ_ONCE(sqe->addr2);
io_uring/net.c
1853
uaddr = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/net.c
1854
bind->addr_len = READ_ONCE(sqe->addr2);
io_uring/net.c
1887
listen->backlog = READ_ONCE(sqe->len);
io_uring/net.c
357
sr->buf = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/net.c
359
if (READ_ONCE(sqe->__pad3[0]))
io_uring/net.c
368
addr = u64_to_user_ptr(READ_ONCE(sqe->addr2));
io_uring/net.c
369
addr_len = READ_ONCE(sqe->addr_len);
io_uring/net.c
399
sr->umsg = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/net.c
423
sr->len = READ_ONCE(sqe->len);
io_uring/net.c
426
sr->flags = READ_ONCE(sqe->ioprio);
io_uring/net.c
429
sr->msg_flags = READ_ONCE(sqe->msg_flags) | MSG_NOSIGNAL;
io_uring/net.c
794
sr->umsg = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/net.c
795
sr->len = READ_ONCE(sqe->len);
io_uring/net.c
798
sr->flags = READ_ONCE(sqe->ioprio);
io_uring/net.c
801
sr->msg_flags = READ_ONCE(sqe->msg_flags);
io_uring/net.c
816
sr->mshot_total_len = READ_ONCE(sqe->optlen);
io_uring/nop.c
32
nop->flags = READ_ONCE(sqe->nop_flags);
io_uring/nop.c
37
nop->result = READ_ONCE(sqe->len);
io_uring/nop.c
41
nop->fd = READ_ONCE(sqe->fd);
io_uring/nop.c
45
req->buf_index = READ_ONCE(sqe->buf_index);
io_uring/nop.c
51
nop->extra1 = READ_ONCE(sqe->off);
io_uring/nop.c
52
nop->extra2 = READ_ONCE(sqe->addr);
io_uring/openclose.c
111
how = u64_to_user_ptr(READ_ONCE(sqe->addr2));
io_uring/openclose.c
112
len = READ_ONCE(sqe->len);
io_uring/openclose.c
228
close->fd = READ_ONCE(sqe->fd);
io_uring/openclose.c
229
close->file_slot = READ_ONCE(sqe->file_index);
io_uring/openclose.c
288
flags = READ_ONCE(sqe->install_fd_flags);
io_uring/openclose.c
333
p->fds = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/openclose.c
334
p->flags = READ_ONCE(sqe->pipe_flags);
io_uring/openclose.c
338
p->file_slot = READ_ONCE(sqe->file_index);
io_uring/openclose.c
68
open->dfd = READ_ONCE(sqe->fd);
io_uring/openclose.c
69
fname = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/openclose.c
75
open->file_slot = READ_ONCE(sqe->file_index);
io_uring/openclose.c
97
u64 mode = READ_ONCE(sqe->len);
io_uring/openclose.c
98
u64 flags = READ_ONCE(sqe->open_flags);
io_uring/poll.c
837
events = READ_ONCE(sqe->poll32_events);
io_uring/poll.c
856
flags = READ_ONCE(sqe->len);
io_uring/poll.c
864
upd->old_user_data = READ_ONCE(sqe->addr);
io_uring/poll.c
868
upd->new_user_data = READ_ONCE(sqe->off);
io_uring/poll.c
886
flags = READ_ONCE(sqe->len);
io_uring/register.c
597
tail = READ_ONCE(o.rings->sq.tail);
io_uring/register.c
598
old_head = READ_ONCE(o.rings->sq.head);
io_uring/register.c
610
tail = READ_ONCE(o.rings->cq.tail);
io_uring/register.c
611
old_head = READ_ONCE(o.rings->cq.head);
io_uring/register.c
632
WRITE_ONCE(n.rings->sq_dropped, READ_ONCE(o.rings->sq_dropped));
io_uring/register.c
634
WRITE_ONCE(n.rings->cq_flags, READ_ONCE(o.rings->cq_flags));
io_uring/register.c
635
WRITE_ONCE(n.rings->cq_overflow, READ_ONCE(o.rings->cq_overflow));
io_uring/rsrc.c
420
up->offset = READ_ONCE(sqe->off);
io_uring/rsrc.c
421
up->nr_args = READ_ONCE(sqe->len);
io_uring/rsrc.c
424
up->arg = READ_ONCE(sqe->addr);
io_uring/rw.c
1310
iopoll_start = READ_ONCE(req->iopoll_start);
io_uring/rw.c
1353
if (READ_ONCE(req->iopoll_completed))
io_uring/rw.c
1368
READ_ONCE(req->iopoll_completed))
io_uring/rw.c
272
rw->kiocb.ki_pos = READ_ONCE(sqe->off);
io_uring/rw.c
274
req->buf_index = READ_ONCE(sqe->buf_index);
io_uring/rw.c
277
ioprio = READ_ONCE(sqe->ioprio);
io_uring/rw.c
288
rw->kiocb.ki_write_stream = READ_ONCE(sqe->write_stream);
io_uring/rw.c
295
rw->addr = READ_ONCE(sqe->addr);
io_uring/rw.c
296
rw->len = READ_ONCE(sqe->len);
io_uring/rw.c
297
rw->flags = (__force rwf_t) READ_ONCE(sqe->rw_flags);
io_uring/rw.c
299
attr_type_mask = READ_ONCE(sqe->attr_type_mask);
io_uring/rw.c
307
attr_ptr = READ_ONCE(sqe->attr_ptr);
io_uring/slist.h
13
#define wq_list_empty(list) (READ_ONCE((list)->first) == NULL)
io_uring/splice.c
114
sp->off_in = READ_ONCE(sqe->splice_off_in);
io_uring/splice.c
115
sp->off_out = READ_ONCE(sqe->off);
io_uring/splice.c
34
sp->len = READ_ONCE(sqe->len);
io_uring/splice.c
35
sp->flags = READ_ONCE(sqe->splice_flags);
io_uring/splice.c
38
sp->splice_fd_in = READ_ONCE(sqe->splice_fd_in);
io_uring/splice.c
46
if (READ_ONCE(sqe->splice_off_in) || READ_ONCE(sqe->off))
io_uring/sqpoll.c
171
return READ_ONCE(sqd->state);
io_uring/statx.c
34
sx->dfd = READ_ONCE(sqe->fd);
io_uring/statx.c
35
sx->mask = READ_ONCE(sqe->len);
io_uring/statx.c
36
path = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/statx.c
37
sx->buffer = u64_to_user_ptr(READ_ONCE(sqe->addr2));
io_uring/statx.c
38
sx->flags = READ_ONCE(sqe->statx_flags);
io_uring/sync.c
32
sync->off = READ_ONCE(sqe->off);
io_uring/sync.c
33
sync->len = READ_ONCE(sqe->len);
io_uring/sync.c
34
sync->flags = READ_ONCE(sqe->sync_range_flags);
io_uring/sync.c
60
sync->flags = READ_ONCE(sqe->fsync_flags);
io_uring/sync.c
64
sync->off = READ_ONCE(sqe->off);
io_uring/sync.c
67
sync->len = READ_ONCE(sqe->len);
io_uring/sync.c
94
sync->off = READ_ONCE(sqe->off);
io_uring/sync.c
95
sync->len = READ_ONCE(sqe->addr);
io_uring/sync.c
96
sync->mode = READ_ONCE(sqe->len);
io_uring/timeout.c
133
seq = READ_ONCE(ctx->cached_cq_tail) - atomic_read(&ctx->cq_timeouts);
io_uring/timeout.c
456
tr->addr = READ_ONCE(sqe->addr);
io_uring/timeout.c
457
tr->flags = READ_ONCE(sqe->timeout_flags);
io_uring/timeout.c
465
if (get_timespec64(&tr->ts, u64_to_user_ptr(READ_ONCE(sqe->addr2))))
io_uring/timeout.c
522
u32 off = READ_ONCE(sqe->off);
io_uring/timeout.c
528
flags = READ_ONCE(sqe->timeout_flags);
io_uring/timeout.c
560
if (get_timespec64(&data->ts, u64_to_user_ptr(READ_ONCE(sqe->addr))))
io_uring/truncate.c
31
ft->len = READ_ONCE(sqe->off);
io_uring/tw.c
188
head = READ_ONCE(ctx->work_llist.first);
io_uring/tw.c
199
nr_tw_prev = READ_ONCE(first_req->nr_tw);
io_uring/uring_cmd.c
192
ioucmd->flags = READ_ONCE(sqe->uring_cmd_flags);
io_uring/uring_cmd.c
199
req->buf_index = READ_ONCE(sqe->buf_index);
io_uring/uring_cmd.c
206
ioucmd->cmd_op = READ_ONCE(sqe->cmd_op);
io_uring/wait.c
132
if (!READ_ONCE(iowq->hit_timeout))
io_uring/wait.c
139
return READ_ONCE(iowq->hit_timeout) ? -ETIME : 0;
io_uring/wait.c
170
if (unlikely(READ_ONCE(ctx->check_cq)))
io_uring/wait.c
219
iowq.cq_tail = READ_ONCE(rings->cq.head) + min_events;
io_uring/wait.c
220
iowq.cq_min_tail = READ_ONCE(rings->cq.tail);
io_uring/wait.c
221
nr_wait = (int) iowq.cq_tail - READ_ONCE(rings->cq.tail);
io_uring/wait.c
288
check_cq = READ_ONCE(ctx->check_cq);
io_uring/wait.c
309
READ_ONCE(io_get_rings(ctx)->cq.tail);
io_uring/wait.c
319
return READ_ONCE(io_get_rings(ctx)->cq.head) == READ_ONCE(io_get_rings(ctx)->cq.tail) ? ret : 0;
io_uring/wait.c
85
if (iowq->cq_min_tail != READ_ONCE(rings->cq.tail))
io_uring/wait.h
32
return ctx->cached_cq_tail - READ_ONCE(rings->cq.head);
io_uring/wait.h
39
return READ_ONCE(rings->cq.tail) - READ_ONCE(rings->cq.head);
io_uring/waitid.c
273
iw->which = READ_ONCE(sqe->len);
io_uring/waitid.c
274
iw->upid = READ_ONCE(sqe->fd);
io_uring/waitid.c
275
iw->options = READ_ONCE(sqe->file_index);
io_uring/waitid.c
277
iw->infop = u64_to_user_ptr(READ_ONCE(sqe->addr2));
io_uring/xattr.c
130
name = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/xattr.c
131
ix->ctx.cvalue = u64_to_user_ptr(READ_ONCE(sqe->addr2));
io_uring/xattr.c
133
ix->ctx.size = READ_ONCE(sqe->len);
io_uring/xattr.c
134
ix->ctx.flags = READ_ONCE(sqe->xattr_flags);
io_uring/xattr.c
164
path = u64_to_user_ptr(READ_ONCE(sqe->addr3));
io_uring/xattr.c
51
name = u64_to_user_ptr(READ_ONCE(sqe->addr));
io_uring/xattr.c
52
ix->ctx.value = u64_to_user_ptr(READ_ONCE(sqe->addr2));
io_uring/xattr.c
53
ix->ctx.size = READ_ONCE(sqe->len);
io_uring/xattr.c
54
ix->ctx.flags = READ_ONCE(sqe->xattr_flags);
io_uring/xattr.c
92
path = u64_to_user_ptr(READ_ONCE(sqe->addr3));
io_uring/zcrx.c
1515
const struct proto *prot = READ_ONCE(sk->sk_prot);
io_uring/zcrx.c
930
__u64 off = READ_ONCE(rqe->off);
ipc/mqueue.c
725
if (READ_ONCE(ewp->state) == STATE_READY) {
ipc/mqueue.c
734
if (READ_ONCE(ewp->state) == STATE_READY) {
ipc/msg.c
1221
msg = READ_ONCE(msr_d.r_msg);
ipc/msg.c
1235
msg = READ_ONCE(msr_d.r_msg);
ipc/sem.c
2181
error = READ_ONCE(queue.status);
ipc/sem.c
2197
error = READ_ONCE(queue.status);
ipc/sem.c
418
if (!READ_ONCE(sma->use_global_lock)) {
kernel/acct.c
165
res = to_acct(READ_ONCE(ns->bacct));
kernel/acct.c
177
if (res != to_acct(READ_ONCE(ns->bacct))) {
kernel/audit.c
1541
t = READ_ONCE(current->signal->audit_tty);
kernel/audit.c
1562
t = READ_ONCE(current->signal->audit_tty);
kernel/audit_tree.c
255
if (READ_ONCE(p->key) == key) {
kernel/bpf/arraymap.c
1032
ptr = READ_ONCE(*elem);
kernel/bpf/arraymap.c
1143
if (!READ_ONCE(poke->tailcall_target_stable))
kernel/bpf/arraymap.c
1308
ee = READ_ONCE(array->ptrs[i]);
kernel/bpf/arraymap.c
1410
return READ_ONCE(*inner_map);
kernel/bpf/arraymap.c
890
if (elem && (ptr = READ_ONCE(*elem)))
kernel/bpf/bpf_lru_list.c
36
return READ_ONCE(node->ref);
kernel/bpf/bpf_lru_list.c
502
u8 node_type = READ_ONCE(node->type);
kernel/bpf/bpf_lru_list.h
68
if (!READ_ONCE(node->ref))
kernel/bpf/btf.c
8127
seq_printf(m, "btf_id:\t%u\n", READ_ONCE(btf->id));
kernel/bpf/btf.c
8209
info.id = READ_ONCE(btf->id);
kernel/bpf/btf.c
8272
return READ_ONCE(btf->id);
kernel/bpf/cgroup.c
1768
ptr = &READ_ONCE(storage->buf)->data[0];
kernel/bpf/cgroup.c
79
while ((prog = READ_ONCE(item->prog))) {
kernel/bpf/core.c
1049
if (atomic_long_add_return(size, &bpf_jit_current) > READ_ONCE(bpf_jit_limit)) {
kernel/bpf/core.c
2092
prog = READ_ONCE(array->ptrs[index]);
kernel/bpf/hashtab.c
2637
*value = map->ops->map_fd_sys_lookup_elem(READ_ONCE(*ptr));
kernel/bpf/hashtab.c
2695
return READ_ONCE(*inner_map);
kernel/bpf/helpers.c
1212
callback_fn = READ_ONCE(cb->callback_fn);
kernel/bpf/helpers.c
1322
old_cb = READ_ONCE(async->cb);
kernel/bpf/helpers.c
1433
} while (READ_ONCE(cb->prog) != prog ||
kernel/bpf/helpers.c
1434
(void __force *)READ_ONCE(cb->callback_fn) != callback_fn);
kernel/bpf/helpers.c
1479
cb = READ_ONCE(async->cb);
kernel/bpf/helpers.c
1513
t = READ_ONCE(async->timer);
kernel/bpf/helpers.c
1514
if (!t || !READ_ONCE(t->cb.prog))
kernel/bpf/helpers.c
1559
t = READ_ONCE(async->timer);
kernel/bpf/helpers.c
1672
if (!READ_ONCE(async->cb))
kernel/bpf/helpers.c
2444
if (WARN_ON_ONCE(READ_ONCE(node->owner) != head))
kernel/bpf/helpers.c
2492
if (READ_ONCE(node_internal->owner) != root)
kernel/bpf/helpers.c
2566
if (READ_ONCE(node_internal->owner) != root)
kernel/bpf/helpers.c
2576
if (READ_ONCE(node_internal->owner) != root)
kernel/bpf/helpers.c
2706
cgrp = READ_ONCE(array->ptrs[idx]);
kernel/bpf/helpers.c
3185
w = READ_ONCE(async->work);
kernel/bpf/helpers.c
3186
if (!w || !READ_ONCE(w->cb.prog))
kernel/bpf/helpers.c
4269
ctx = READ_ONCE(twk->ctx);
kernel/bpf/helpers.c
4478
cb = READ_ONCE(async->cb);
kernel/bpf/local_storage.c
141
return &READ_ONCE(storage->buf)->data[0];
kernel/bpf/local_storage.c
437
&READ_ONCE(storage->buf)->data[0], m);
kernel/bpf/lpm_trie.c
771
return elem_size * READ_ONCE(trie->n_entries);
kernel/bpf/map_iter.c
209
ret += READ_ONCE(*pcount);
kernel/bpf/memalloc.c
313
if (unlikely(READ_ONCE(c->draining))) {
kernel/bpf/memalloc.c
324
if (unlikely(READ_ONCE(c->draining))) {
kernel/bpf/memalloc.c
423
if (unlikely(READ_ONCE(c->draining))) {
kernel/bpf/mprog.c
117
oprog = READ_ONCE(fp->prog);
kernel/bpf/mprog.c
187
if (tuple->prog == READ_ONCE(fp->prog))
kernel/bpf/mprog.c
202
if (tuple->prog == READ_ONCE(fp->prog) &&
kernel/bpf/mprog.c
218
if (tuple->prog == READ_ONCE(fp->prog) &&
kernel/bpf/mprog.c
311
prog = READ_ONCE(fp->prog);
kernel/bpf/mprog.c
432
prog = READ_ONCE(fp->prog);
kernel/bpf/percpu_freelist.c
108
if (!READ_ONCE(head->first))
kernel/bpf/reuseport_array.c
221
if (READ_ONCE(nsk->sk_user_data))
kernel/bpf/ringbuf.c
487
hdr_len = READ_ONCE(hdr->len);
kernel/bpf/ringbuf.c
508
hdr_len = READ_ONCE(hdr->len);
kernel/bpf/rqspinlock.c
136
int real_cnt = READ_ONCE(rqh_cpu->cnt);
kernel/bpf/rqspinlock.c
163
remote_lock = READ_ONCE(rqh_cpu->locks[cnt - 1]);
kernel/bpf/rqspinlock.c
178
if (READ_ONCE(rqh_cpu->locks[i]) != lock)
kernel/bpf/rqspinlock.c
548
next = READ_ONCE(node->next);
kernel/bpf/stackmap.c
189
u64 ip = READ_ONCE(id_offs[i].ip);
kernel/bpf/stackmap.c
270
bucket = READ_ONCE(smap->buckets[id]);
kernel/bpf/stackmap.c
57
u32 curr_sysctl_max_stack = READ_ONCE(sysctl_perf_event_max_stack);
kernel/bpf/syscall.c
1038
READ_ONCE(map->frozen));
kernel/bpf/syscall.c
2257
if (READ_ONCE(map->frozen)) {
kernel/bpf/syscall.c
5353
if (!READ_ONCE(map->frozen))
kernel/bpf/syscall.c
962
if (READ_ONCE(map->free_after_mult_rcu_gp))
kernel/bpf/syscall.c
964
else if (READ_ONCE(map->free_after_rcu_gp))
kernel/bpf/syscall.c
996
if (READ_ONCE(map->frozen))
kernel/bpf/verifier.c
7215
READ_ONCE(map->frozen) &&
kernel/cgroup/cgroup-v1.c
746
switch (READ_ONCE(tsk->__state)) {
kernel/cgroup/cgroup.c
3734
int descendants = READ_ONCE(cgrp->max_descendants);
kernel/cgroup/cgroup.c
3777
int depth = READ_ONCE(cgrp->max_depth);
kernel/cgroup/cgroup.c
6646
if (root == &cgrp_dfl_root && !READ_ONCE(cgrp_dfl_visible))
kernel/cgroup/cpuset.c
3322
err = perr_strings[READ_ONCE(cs->prs_err)];
kernel/cgroup/dmem.c
167
return pool ? READ_ONCE(pool->cnt.low) : 0;
kernel/cgroup/dmem.c
172
return pool ? READ_ONCE(pool->cnt.min) : 0;
kernel/cgroup/dmem.c
177
return pool ? READ_ONCE(pool->cnt.max) : PAGE_COUNTER_MAX;
kernel/cgroup/dmem.c
337
min = READ_ONCE(ctest->emin);
kernel/cgroup/dmem.c
343
low = READ_ONCE(ctest->elow);
kernel/cgroup/dmem.c
579
if (pool && !READ_ONCE(pool->inited))
kernel/cgroup/misc.c
159
if (!(valid_type(type) && cg && READ_ONCE(misc_res_capacity[type])))
kernel/cgroup/misc.c
169
if (new_usage > READ_ONCE(res->max) ||
kernel/cgroup/misc.c
170
new_usage > READ_ONCE(misc_res_capacity[type])) {
kernel/cgroup/misc.c
223
if (READ_ONCE(misc_res_capacity[i])) {
kernel/cgroup/misc.c
224
max = READ_ONCE(cg->res[i].max);
kernel/cgroup/misc.c
289
if (READ_ONCE(misc_res_capacity[type]))
kernel/cgroup/misc.c
313
if (READ_ONCE(misc_res_capacity[i]) || usage)
kernel/cgroup/misc.c
336
if (READ_ONCE(misc_res_capacity[i]) || watermark)
kernel/cgroup/misc.c
359
cap = READ_ONCE(misc_res_capacity[i]);
kernel/cgroup/misc.c
378
if (READ_ONCE(misc_res_capacity[i]) || events)
kernel/cgroup/pids.c
102
if (nr_pids > READ_ONCE(p->watermark))
kernel/cgroup/pids.c
358
return READ_ONCE(pids->watermark);
kernel/events/core.c
1052
if (READ_ONCE(cpuctx->cgrp) == NULL)
kernel/events/core.c
1056
if (READ_ONCE(cpuctx->cgrp) == cgrp)
kernel/events/core.c
10599
prog = READ_ONCE(event->prog);
kernel/events/core.c
1063
if (READ_ONCE(cpuctx->cgrp) == NULL)
kernel/events/core.c
10760
READ_ONCE(event->overflow_handler)(event, data, regs);
kernel/events/core.c
11834
struct task_struct *task = READ_ONCE(event->ctx->task);
kernel/events/core.c
14474
attach_state = READ_ONCE(event->attach_state);
kernel/events/core.c
1472
ctx = READ_ONCE(event->ctx);
kernel/events/core.c
241
return READ_ONCE(event->owner) == TASK_TOMBSTONE;
kernel/events/core.c
3124
struct task_struct *task = READ_ONCE(ctx->task);
kernel/events/core.c
329
struct task_struct *task = READ_ONCE(ctx->task); /* verified in event_function */
kernel/events/core.c
3335
if (READ_ONCE(event->state) != PERF_EVENT_STATE_ACTIVE)
kernel/events/core.c
3345
if (READ_ONCE(event->oncpu) != smp_processor_id())
kernel/events/core.c
3374
if (READ_ONCE(event->state) != PERF_EVENT_STATE_ACTIVE)
kernel/events/core.c
3385
ret = cpu_function_call(READ_ONCE(event->oncpu),
kernel/events/core.c
387
struct task_struct *task = READ_ONCE(ctx->task);
kernel/events/core.c
4958
enum perf_event_state state = READ_ONCE(event->state);
kernel/events/core.c
4977
event_cpu = READ_ONCE(event->oncpu);
kernel/events/core.c
5865
owner = READ_ONCE(event->owner);
kernel/events/core.c
5969
ctx = READ_ONCE(child->ctx);
kernel/events/core.c
6297
if (unlikely(READ_ONCE(event->state) == PERF_EVENT_STATE_ERROR &&
kernel/events/core.c
643
u64 max_len = READ_ONCE(perf_sample_allowed_ns);
kernel/events/core.c
6988
pmu = READ_ONCE(event->pmu); \
kernel/events/core.c
7362
aux_offset = READ_ONCE(rb->user_page->aux_offset);
kernel/events/core.c
7363
aux_size = READ_ONCE(rb->user_page->aux_size);
kernel/events/core.c
7563
int cpu = READ_ONCE(event->oncpu);
kernel/events/core.c
7882
if (WARN_ON_ONCE(READ_ONCE(sampler->state) != PERF_EVENT_STATE_ACTIVE))
kernel/events/core.c
7885
if (WARN_ON_ONCE(READ_ONCE(sampler->oncpu) != smp_processor_id()))
kernel/events/core.c
7896
if (READ_ONCE(rb->aux_in_sampling)) {
kernel/events/core.c
804
__unqual_scalar_typeof(*(ptr)) ___p = READ_ONCE(*(ptr)); \
kernel/events/core.c
8423
pgd = READ_ONCE(*pgdp);
kernel/events/core.c
8431
p4d = READ_ONCE(*p4dp);
kernel/events/core.c
8439
pud = READ_ONCE(*pudp);
kernel/events/core.c
8775
if (READ_ONCE(rb->aux_in_pause_resume))
kernel/events/core.c
894
return READ_ONCE(times[T_TOTAL].offset) - READ_ONCE(times[T_GUEST].offset);
kernel/events/core.c
897
return now + READ_ONCE(times[T_TOTAL].offset);
kernel/events/core.c
9079
cpu = READ_ONCE(iter->oncpu);
kernel/events/ring_buffer.c
202
tail = READ_ONCE(rb->user_page->data_tail);
kernel/events/ring_buffer.c
409
nest = READ_ONCE(rb->aux_nest);
kernel/events/ring_buffer.c
433
aux_tail = READ_ONCE(rb->user_page->aux_tail);
kernel/events/ring_buffer.c
64
nest = READ_ONCE(rb->nest);
kernel/events/uprobes.c
1805
area = READ_ONCE(mm->uprobes_state.xol_area); /* ^^^ */
kernel/events/uprobes.c
2229
area = READ_ONCE(current->mm->uprobes_state.xol_area); /* ^^^ */
kernel/events/uprobes.c
2452
vm_file = READ_ONCE(vma->vm_file);
kernel/events/uprobes.c
804
hstate = READ_ONCE(hprobe->state);
kernel/exit.c
1064
limit = READ_ONCE(oops_limit);
kernel/exit.c
1461
int exit_state = READ_ONCE(p->exit_state);
kernel/exit.c
473
struct mm_struct *t_mm = READ_ONCE(t->mm);
kernel/fork.c
1689
cpu_limit = READ_ONCE(sig->rlim[RLIMIT_CPU].rlim_cur);
kernel/fork.c
515
if (WARN_ON(READ_ONCE(tsk->__state) != TASK_DEAD))
kernel/freezer.c
111
unsigned int state = READ_ONCE(p->__state);
kernel/freezer.c
59
return READ_ONCE(p->__state) & TASK_FROZEN;
kernel/futex/core.c
1685
if (READ_ONCE(fph->state) == FR_PERCPU) {
kernel/futex/core.c
1699
if (READ_ONCE(fph->state) == FR_PERCPU) {
kernel/futex/core.c
345
mpol = READ_ONCE(vma->vm_policy);
kernel/futex/core.c
670
mapping = READ_ONCE(folio->mapping);
kernel/futex/core.c
746
if (READ_ONCE(folio->mapping) != mapping) {
kernel/futex/core.c
753
inode = READ_ONCE(mapping->host);
kernel/futex/core.c
936
lock_ptr = READ_ONCE(q->lock_ptr);
kernel/futex/core.c
976
lock_ptr = READ_ONCE(q->lock_ptr);
kernel/futex/requeue.c
247
task = READ_ONCE(q->task);
kernel/futex/waitwake.c
515
if (!READ_ONCE(vs->q.lock_ptr))
kernel/hung_task.c
106
unsigned int state = READ_ONCE(t->__state);
kernel/hung_task.c
145
blocker = READ_ONCE(task->blocker);
kernel/irq/irqdesc.c
781
unsigned int target_cpu = READ_ONCE(desc->redirect.target_cpu);
kernel/irq/spurious.c
122
state = READ_ONCE(desc->istate);
kernel/irq/spurious.c
216
action = READ_ONCE(desc->action);
kernel/kcov.c
1032
mode = READ_ONCE(t->kcov_mode);
kernel/kcov.c
185
mode = READ_ONCE(t->kcov_mode);
kernel/kcov.c
222
pos = READ_ONCE(area[0]) + 1;
kernel/kcov.c
258
count = READ_ONCE(area[0]);
kernel/kcov.c
574
READ_ONCE(area[offset]);
kernel/kcov.c
832
mode = READ_ONCE(t->kcov_mode);
kernel/kcov.c
883
mode = READ_ONCE(t->kcov_mode);
kernel/kcov.c
966
dst_len = READ_ONCE(*(unsigned long *)dst_area);
kernel/kcov.c
972
dst_len = READ_ONCE(*(u64 *)dst_area);
kernel/kcsan/core.c
317
return READ_ONCE(kcsan_enabled) && !ctx->disable_count;
kernel/kcsan/core.c
777
size = READ_ONCE(reorder_access->size);
kernel/kcsan/debugfs.c
197
seq_printf(file, "enabled: %i\n", READ_ONCE(kcsan_enabled));
kernel/kcsan/debugfs.c
60
const bool was_enabled = READ_ONCE(kcsan_enabled);
kernel/kcsan/kcsan_test.c
141
return READ_ONCE(observed.nlines) == ARRAY_SIZE(observed.lines);
kernel/kcsan/kcsan_test.c
326
sink_value(READ_ONCE(test_sink));
kernel/kcsan/kcsan_test.c
343
sink_value(READ_ONCE(test_var));
kernel/kcsan/kcsan_test.c
387
WRITE_ONCE(test_var, READ_ONCE(test_var) ^ TEST_CHANGE_BITS);
kernel/kcsan/kcsan_test.c
409
READ_ONCE(test_sink);
kernel/kcsan/kcsan_test.c
865
sink_value(READ_ONCE(test_var)); /* induce value-change */
kernel/ksysfs.c
148
return sysfs_emit(buf, "%d\n", READ_ONCE(rcu_expedited));
kernel/ksysfs.c
165
return sysfs_emit(buf, "%d\n", READ_ONCE(rcu_normal));
kernel/livepatch/transition.c
196
task->patch_state = READ_ONCE(klp_target_state);
kernel/locking/lockdep.c
3802
if (READ_ONCE(chain->chain_key) == chain_key) {
kernel/locking/lockdep.c
6834
int dl = READ_ONCE(debug_locks);
kernel/locking/lockdep.c
788
int i, depth = READ_ONCE(p->lockdep_depth);
kernel/locking/mcs_spinlock.h
105
while (!(next = READ_ONCE(node->next)))
kernel/locking/mcs_spinlock.h
96
struct mcs_spinlock *next = READ_ONCE(node->next);
kernel/locking/mutex.c
326
if (ww_ctx->acquired > 0 && READ_ONCE(ww->ctx))
kernel/locking/mutex.c
597
if (unlikely(ww_ctx == READ_ONCE(ww->ctx)))
kernel/locking/osq_lock.c
182
prev = READ_ONCE(node->prev);
kernel/locking/qspinlock.c
299
next = READ_ONCE(node->next);
kernel/locking/qspinlock_paravirt.h
124
return !READ_ONCE(lock->locked) &&
kernel/locking/qspinlock_paravirt.h
243
if (READ_ONCE(he->lock) == lock) {
kernel/locking/qspinlock_paravirt.h
244
node = READ_ONCE(he->node);
kernel/locking/qspinlock_paravirt.h
269
return READ_ONCE(prev->state) != VCPU_RUNNING;
kernel/locking/qspinlock_paravirt.h
299
if (READ_ONCE(node->locked))
kernel/locking/qspinlock_paravirt.h
319
if (!READ_ONCE(node->locked)) {
kernel/locking/qspinlock_paravirt.h
340
!READ_ONCE(node->locked));
kernel/locking/qspinlock_paravirt.h
410
if (READ_ONCE(pn->state) == VCPU_HASHED)
kernel/locking/rtmutex.c
194
owner = READ_ONCE(*p);
kernel/locking/rtmutex.c
246
owner = READ_ONCE(*p);
kernel/locking/semaphore.c
53
if (READ_ONCE((sem)->last_holder) == (unsigned long)current)
kernel/locking/semaphore.c
59
return READ_ONCE(sem->last_holder);
kernel/locking/spinlock_debug.c
187
RWLOCK_BUG_ON(READ_ONCE(lock->owner) == current, lock, "recursion");
kernel/locking/spinlock_debug.c
188
RWLOCK_BUG_ON(READ_ONCE(lock->owner_cpu) == raw_smp_processor_id(),
kernel/locking/spinlock_debug.c
57
struct task_struct *owner = READ_ONCE(lock->owner);
kernel/locking/spinlock_debug.c
66
lock, READ_ONCE(lock->magic),
kernel/locking/spinlock_debug.c
69
READ_ONCE(lock->owner_cpu));
kernel/locking/spinlock_debug.c
86
SPIN_BUG_ON(READ_ONCE(lock->magic) != SPINLOCK_MAGIC, lock, "bad magic");
kernel/locking/spinlock_debug.c
87
SPIN_BUG_ON(READ_ONCE(lock->owner) == current, lock, "recursion");
kernel/locking/spinlock_debug.c
88
SPIN_BUG_ON(READ_ONCE(lock->owner_cpu) == raw_smp_processor_id(),
kernel/locking/test-ww_mutex.c
202
WARN_ON(READ_ONCE(abba->b_mutex.ctx) != &ctx);
kernel/locking/test-ww_mutex.c
245
WARN_ON(READ_ONCE(abba.a_mutex.ctx) != &ctx);
kernel/locking/ww_mutex.h
469
struct ww_acquire_ctx *hold_ctx = READ_ONCE(ww->ctx);
kernel/locking/ww_rt_mutex.c
48
if (unlikely(ww_ctx == READ_ONCE(lock->ctx)))
kernel/panic.c
526
limit = READ_ONCE(warn_limit);
kernel/pid.c
210
pid_max[ns->level - i] = READ_ONCE(tmp->pid_max);
kernel/pid_namespace.c
48
kc = READ_ONCE(*pkc);
kernel/pid_namespace.c
61
return READ_ONCE(*pkc);
kernel/pid_sysctl.h
23
scope = max(READ_ONCE(ns->memfd_noexec_scope), parent_scope);
kernel/power/qos.c
55
return READ_ONCE(c->target_value);
kernel/printk/nbcon.c
1026
con_dropped = data_race(READ_ONCE(con->dropped));
kernel/printk/printk.c
1942
waiter = READ_ONCE(console_waiter);
kernel/printk/printk.c
2003
owner = READ_ONCE(console_owner);
kernel/printk/printk.c
2004
waiter = READ_ONCE(console_waiter);
kernel/printk/printk.c
2028
while (READ_ONCE(console_waiter))
kernel/rcu/rcu.h
130
s = (READ_ONCE(*sp) + 2 * RCU_SEQ_STATE_MASK + 1) & ~RCU_SEQ_STATE_MASK;
kernel/rcu/rcu.h
138
return READ_ONCE(*sp);
kernel/rcu/rcu.h
147
return ULONG_CMP_LT((s - 1) & ~RCU_SEQ_STATE_MASK, READ_ONCE(*sp));
kernel/rcu/rcu.h
156
return ULONG_CMP_GE(READ_ONCE(*sp), s);
kernel/rcu/rcu.h
175
unsigned long cur_s = READ_ONCE(*sp);
kernel/rcu/rcu_segcblist.c
270
&rsclp->head != READ_ONCE(rsclp->tails[RCU_DONE_TAIL]);
kernel/rcu/rcu_segcblist.c
94
return READ_ONCE(rsclp->seglen[seg]);
kernel/rcu/rcu_segcblist.h
105
return !READ_ONCE(*READ_ONCE(rsclp->tails[seg]));
kernel/rcu/rcu_segcblist.h
15
return READ_ONCE(rclp->len);
kernel/rcu/rcu_segcblist.h
45
return !READ_ONCE(rsclp->head);
kernel/rcu/rcu_segcblist.h
54
return READ_ONCE(rsclp->len);
kernel/rcu/rcu_segcblist.h
73
return READ_ONCE(rsclp->flags) & flags;
kernel/rcu/rcutorture.c
1258
lp = READ_ONCE(last_persist);
kernel/rcu/rcutorture.c
1321
oldstarttime = READ_ONCE(boost_starttime);
kernel/rcu/rcutorture.c
1365
while (oldstarttime == READ_ONCE(boost_starttime) && !kthread_should_stop()) {
kernel/rcu/rcutorture.c
2013
!READ_ONCE(rtp->rtort_chkp) &&
kernel/rcu/rcutorture.c
2015
rtrcp->rtc_chkloops = READ_ONCE(rtrcp_chked->rtc_myloops);
kernel/rcu/rcutorture.c
2025
rtrcp_assigner = READ_ONCE(rtrcp->rtc_assigner);
kernel/rcu/rcutorture.c
2032
loops = READ_ONCE(rtrcp_chked->rtc_myloops);
kernel/rcu/rcutorture.c
2371
pipe_count = READ_ONCE(rtorsp->p->rtort_pipe_count);
kernel/rcu/rcutorture.c
2413
WARN_ON_ONCE(leakpointer && READ_ONCE(rtorsp->p->rtort_pipe_count) > 1);
kernel/rcu/rcutorture.c
2737
pipesummary[i] += READ_ONCE(per_cpu(rcu_torture_count, cpu)[i]);
kernel/rcu/rcutorture.c
2738
batchsummary[i] += READ_ONCE(per_cpu(rcu_torture_batch, cpu)[i]);
kernel/rcu/rcutorture.c
2745
ndowns += READ_ONCE(updownreaders[i].rtorsu_ndowns);
kernel/rcu/rcutorture.c
2746
nups += READ_ONCE(updownreaders[i].rtorsu_nups);
kernel/rcu/rcutorture.c
2747
nunexpired += READ_ONCE(updownreaders[i].rtorsu_inuse);
kernel/rcu/rcutorture.c
2748
nmigrates += READ_ONCE(updownreaders[i].rtorsu_nmigrates);
kernel/rcu/rcutorture.c
2830
wtp = READ_ONCE(writer_task);
kernel/rcu/rcutorture.c
296
unsigned int i = READ_ONCE(rcu_torture_writer_state);
kernel/rcu/rcutorture.c
3139
if (READ_ONCE(fcsp->stop)) {
kernel/rcu/rcutorture.c
3306
cver = READ_ONCE(rcu_torture_current_version);
kernel/rcu/rcutorture.c
3315
!READ_ONCE(rcu_fwd_emergency_stop) && !torture_must_stop()) {
kernel/rcu/rcutorture.c
3325
!READ_ONCE(rcu_fwd_emergency_stop) && !torture_must_stop()) {
kernel/rcu/rcutorture.c
3327
cver = READ_ONCE(rcu_torture_current_version) - cver;
kernel/rcu/rcutorture.c
3341
WARN_ON(READ_ONCE(fcs.stop) != 2);
kernel/rcu/rcutorture.c
3366
if (READ_ONCE(rcu_fwd_emergency_stop))
kernel/rcu/rcutorture.c
3383
cver = READ_ONCE(rcu_torture_current_version);
kernel/rcu/rcutorture.c
3389
!READ_ONCE(rcu_fwd_emergency_stop) && !torture_must_stop()) {
kernel/rcu/rcutorture.c
3390
rfcp = READ_ONCE(rfp->rcu_fwd_cb_head);
kernel/rcu/rcutorture.c
3393
rfcpn = READ_ONCE(rfcp->rfc_next);
kernel/rcu/rcutorture.c
3424
n_launders_cb_snap = READ_ONCE(rfp->n_launders_cb);
kernel/rcu/rcutorture.c
3425
cver = READ_ONCE(rcu_torture_current_version) - cver;
kernel/rcu/rcutorture.c
3431
if (!torture_must_stop() && !READ_ONCE(rcu_fwd_emergency_stop) &&
kernel/rcu/rcutorture.c
3473
rcu_fwd_progress_check(1 + (jiffies - READ_ONCE(rfp[i].rcu_fwd_startat)) / 2);
kernel/rcu/rcutorture.c
3508
unsigned long oldseq = READ_ONCE(rcu_fwd_seq);
kernel/rcu/rcutorture.c
3530
while (READ_ONCE(rcu_fwd_seq) == oldseq && !torture_must_stop())
kernel/rcu/rcutorture.c
3532
oldseq = READ_ONCE(rcu_fwd_seq);
kernel/rcu/rcutorture.c
3851
if (READ_ONCE(read_exit_child_stop))
kernel/rcu/rcutorture.c
3869
for (; !errexit && !READ_ONCE(read_exit_child_stop) && i < read_exit_delay; i++)
kernel/rcu/rcutorture.c
3871
} while (!errexit && !READ_ONCE(read_exit_child_stop));
kernel/rcu/rcutorture.c
513
struct rcu_torture_reader_check *rtrcp = READ_ONCE(rp->rtort_chkp);
kernel/rcu/refscale.c
1066
a = READ_ONCE(rtsp->a);
kernel/rcu/refscale.c
1071
if (a != READ_ONCE(rtsp->a)) {
kernel/rcu/refscale.c
1077
b = READ_ONCE(rtsp->a);
kernel/rcu/refscale.c
438
WRITE_ONCE(*tap, READ_ONCE(*tap) + 1);
kernel/rcu/refscale.c
439
WRITE_ONCE(*tap, READ_ONCE(*tap) - 1);
kernel/rcu/refscale.c
450
WRITE_ONCE(*tap, READ_ONCE(*tap) + 1);
kernel/rcu/refscale.c
452
WRITE_ONCE(*tap, READ_ONCE(*tap) - 1);
kernel/rcu/refscale.c
472
WRITE_ONCE(*tap, READ_ONCE(*tap) + 1);
kernel/rcu/refscale.c
473
WRITE_ONCE(*tap, READ_ONCE(*tap) - 1);
kernel/rcu/refscale.c
487
WRITE_ONCE(*tap, READ_ONCE(*tap) + 1);
kernel/rcu/refscale.c
489
WRITE_ONCE(*tap, READ_ONCE(*tap) - 1);
kernel/rcu/refscale.c
510
WRITE_ONCE(*tap, READ_ONCE(*tap) + 1);
kernel/rcu/refscale.c
511
WRITE_ONCE(*tap, READ_ONCE(*tap) - 1);
kernel/rcu/refscale.c
525
WRITE_ONCE(*tap, READ_ONCE(*tap) + 1);
kernel/rcu/refscale.c
527
WRITE_ONCE(*tap, READ_ONCE(*tap) - 1);
kernel/rcu/refscale.c
550
WRITE_ONCE(*tap, READ_ONCE(*tap) + 1);
kernel/rcu/refscale.c
551
WRITE_ONCE(*tap, READ_ONCE(*tap) - 1);
kernel/rcu/refscale.c
566
WRITE_ONCE(*tap, READ_ONCE(*tap) + 1);
kernel/rcu/refscale.c
568
WRITE_ONCE(*tap, READ_ONCE(*tap) - 1);
kernel/rcu/srcutiny.c
109
newval = READ_ONCE(ssp->srcu_lock_nesting[idx]) - 1;
kernel/rcu/srcutiny.c
112
if (!newval && READ_ONCE(ssp->srcu_gp_waiting) && in_task() && !irqs_disabled())
kernel/rcu/srcutiny.c
131
if (ssp->srcu_gp_running || ULONG_CMP_GE(ssp->srcu_idx, READ_ONCE(ssp->srcu_idx_max))) {
kernel/rcu/srcutiny.c
151
!READ_ONCE(ssp->srcu_lock_nesting[idx]), HZ / 10);
kernel/rcu/srcutiny.c
152
} while (READ_ONCE(ssp->srcu_lock_nesting[idx]));
kernel/rcu/srcutiny.c
176
idx = ULONG_CMP_LT(ssp->srcu_idx, READ_ONCE(ssp->srcu_idx_max));
kernel/rcu/srcutiny.c
203
if (ULONG_CMP_GE(READ_ONCE(ssp->srcu_idx_max), cookie)) {
kernel/rcu/srcutiny.c
207
if (!READ_ONCE(ssp->srcu_gp_running)) {
kernel/rcu/srcutiny.c
271
ret = (READ_ONCE(ssp->srcu_idx) + 3) & ~0x1;
kernel/rcu/srcutiny.c
301
unsigned long cur_s = READ_ONCE(ssp->srcu_idx);
kernel/rcu/srcutree.c
1030
sgsne = READ_ONCE(snp->srcu_gp_seq_needed_exp);
kernel/rcu/srcutree.c
1279
tlast = READ_ONCE(ssp->srcu_sup->srcu_last_gp_end);
kernel/rcu/srcutree.c
1287
if (ULONG_CMP_LT(curseq, READ_ONCE(ssp->srcu_sup->srcu_gp_seq_needed)))
kernel/rcu/srcutree.c
1800
return READ_ONCE(ssp->srcu_sup->srcu_gp_seq);
kernel/rcu/srcutree.c
1834
idx = rcu_seq_state(READ_ONCE(ssp->srcu_sup->srcu_gp_seq));
kernel/rcu/srcutree.c
1844
if (rcu_seq_state(READ_ONCE(ssp->srcu_sup->srcu_gp_seq)) == SRCU_STATE_SCAN1) {
kernel/rcu/srcutree.c
1857
if (rcu_seq_state(READ_ONCE(ssp->srcu_sup->srcu_gp_seq)) == SRCU_STATE_SCAN2) {
kernel/rcu/srcutree.c
1981
if (READ_ONCE(sup->reschedule_jiffies) == j) {
kernel/rcu/srcutree.c
1983
WRITE_ONCE(sup->reschedule_count, READ_ONCE(sup->reschedule_count) + 1);
kernel/rcu/srcutree.c
1984
if (READ_ONCE(sup->reschedule_count) > srcu_max_nodelay)
kernel/rcu/srcutree.c
2037
int ss_state = READ_ONCE(ssp->srcu_sup->srcu_size_state);
kernel/rcu/srcutree.c
232
if (READ_ONCE(ssp->srcu_sup->srcu_size_state) == SRCU_SIZE_SMALL && SRCU_SIZING_IS_INIT()) {
kernel/rcu/srcutree.c
459
return ULONG_CMP_LT(READ_ONCE(sup->srcu_gp_seq), READ_ONCE(sup->srcu_gp_seq_needed_exp));
kernel/rcu/srcutree.c
479
mask = mask | READ_ONCE(sdp->srcu_reader_flavor);
kernel/rcu/srcutree.c
502
mask = mask | READ_ONCE(sdp->srcu_reader_flavor);
kernel/rcu/srcutree.c
687
if (rcu_seq_state(READ_ONCE(sup->srcu_gp_seq))) {
kernel/rcu/srcutree.c
689
gpstart = READ_ONCE(sup->srcu_gp_start);
kernel/rcu/srcutree.c
694
WRITE_ONCE(sup->srcu_n_exp_nodelay, READ_ONCE(sup->srcu_n_exp_nodelay) + 1);
kernel/rcu/srcutree.c
695
if (READ_ONCE(sup->srcu_n_exp_nodelay) > srcu_max_nodelay_phase)
kernel/rcu/srcutree.c
733
if (WARN_ON(rcu_seq_state(READ_ONCE(sup->srcu_gp_seq)) != SRCU_STATE_IDLE) ||
kernel/rcu/srcutree.c
737
__func__, ssp, rcu_seq_state(READ_ONCE(sup->srcu_gp_seq)),
kernel/rcu/srcutree.c
770
old_read_flavor = READ_ONCE(sdp->srcu_reader_flavor);
kernel/rcu/srcutree.c
792
struct srcu_ctr __percpu *scp = READ_ONCE(ssp->srcu_ctrp);
kernel/rcu/srcutree.c
821
struct srcu_ctr __percpu *scpp = READ_ONCE(ssp->srcu_ctrp);
kernel/rcu/sync.c
139
wait_event(rsp->gp_wait, READ_ONCE(rsp->gp_state) >= GP_PASSED);
kernel/rcu/sync.c
154
WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE);
kernel/rcu/sync.c
177
WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED);
kernel/rcu/sync.c
62
WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE);
kernel/rcu/sync.c
63
WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED);
kernel/rcu/tasks.h
1013
t->rcu_tasks_nvcsw = READ_ONCE(t->nvcsw);
kernel/rcu/tasks.h
1026
int rtsi = READ_ONCE(rcu_task_stall_info);
kernel/rcu/tasks.h
1092
if (!READ_ONCE(t->rcu_tasks_holdout) ||
kernel/rcu/tasks.h
1093
t->rcu_tasks_nvcsw != READ_ONCE(t->nvcsw) ||
kernel/rcu/tasks.h
1096
!is_idle_task(t) && READ_ONCE(t->rcu_tasks_idle_cpu) >= 0)) {
kernel/rcu/tasks.h
1160
rtsi = READ_ONCE(rcu_task_stall_info);
kernel/rcu/tasks.h
236
int j = READ_ONCE(i); // Prevent the compiler from reading twice
kernel/rcu/tasks.h
358
ideal_cpu = smp_processor_id() >> READ_ONCE(rtp->percpu_enqueue_shift);
kernel/rcu/tasks.h
370
READ_ONCE(rtp->percpu_enqueue_lim) != rcu_task_cpu_ids)
kernel/rcu/tasks.h
400
if (needwake && READ_ONCE(rtp->kthread_ptr))
kernel/rcu/tasks.h
670
if (READ_ONCE(rtp->kthread_ptr)) {
kernel/rcu/tasks.h
845
rtsi = READ_ONCE(rcu_task_stall_info);
kernel/rcu/tasks.h
869
rtst = READ_ONCE(rcu_task_stall_timeout);
kernel/rcu/tasks.h
979
if (!READ_ONCE(t->on_rq))
kernel/rcu/tiny.c
202
return READ_ONCE(rcu_ctrlblk.gp_seq);
kernel/rcu/tiny.c
229
return oldstate == RCU_GET_STATE_COMPLETED || READ_ONCE(rcu_ctrlblk.gp_seq) != oldstate;
kernel/rcu/tiny.c
236
return READ_ONCE(rcu_ctrlblk.gp_seq) & 0xffffULL;
kernel/rcu/tree.c
1070
if (!READ_ONCE(rcu_state.gp_kthread)) {
kernel/rcu/tree.c
1121
struct task_struct *t = READ_ONCE(rcu_state.gp_kthread);
kernel/rcu/tree.c
1124
!READ_ONCE(rcu_state.gp_flags) || !t)
kernel/rcu/tree.c
1127
WRITE_ONCE(rcu_state.gp_wake_seq, READ_ONCE(rcu_state.gp_seq));
kernel/rcu/tree.c
1197
if (!READ_ONCE(rdp->gpwrap) && ULONG_CMP_GE(rdp->gp_seq_needed, c)) {
kernel/rcu/tree.c
1331
!unlikely(READ_ONCE(rdp->gpwrap))) || /* w/out lock. */
kernel/rcu/tree.c
1365
atomic_t *rgssp = READ_ONCE(rcu_gp_slow_suppress);
kernel/rcu/tree.c
1770
first = READ_ONCE(rcu_state.srs_next.first);
kernel/rcu/tree.c
2014
*gfp = READ_ONCE(rcu_state.gp_flags);
kernel/rcu/tree.c
2019
if (!READ_ONCE(rnp->qsmask) && !rcu_preempt_blocked_readers_cgp(rnp))
kernel/rcu/tree.c
2030
int nr_fqs = READ_ONCE(rcu_state.nr_fqs_jiffies_stall);
kernel/rcu/tree.c
2054
if (READ_ONCE(rcu_state.gp_flags) & RCU_GP_FLAG_FQS) {
kernel/rcu/tree.c
2072
j = READ_ONCE(jiffies_till_first_fqs);
kernel/rcu/tree.c
2109
if (!READ_ONCE(rnp->qsmask) &&
kernel/rcu/tree.c
2128
j = READ_ONCE(jiffies_till_next_fqs);
kernel/rcu/tree.c
2282
READ_ONCE(rcu_state.gp_flags) &
kernel/rcu/tree.c
2385
oldmask = READ_ONCE(rnp_c->qsmask);
kernel/rcu/tree.c
2579
div = READ_ONCE(rcu_divisor);
kernel/rcu/tree.c
2585
long rrn = READ_ONCE(rcu_resched_ns);
kernel/rcu/tree.c
2668
rdp->n_force_qs_snap = READ_ONCE(rcu_state.n_force_qs);
kernel/rcu/tree.c
2803
ret = (READ_ONCE(rcu_state.gp_flags) & RCU_GP_FLAG_FQS) ||
kernel/rcu/tree.c
2816
if (READ_ONCE(rcu_state.gp_flags) & RCU_GP_FLAG_FQS) {
kernel/rcu/tree.c
2868
likely(READ_ONCE(rcu_scheduler_fully_active))) {
kernel/rcu/tree.c
2959
if (!READ_ONCE(*workp)) {
kernel/rcu/tree.c
3043
if (READ_ONCE(rcu_state.n_force_qs) == rdp->n_force_qs_snap &&
kernel/rcu/tree.c
3046
rdp->n_force_qs_snap = READ_ONCE(rcu_state.n_force_qs);
kernel/rcu/tree.c
3094
!!(READ_ONCE(rnp->cbovldmask) & rdp->grpmask)))
kernel/rcu/tree.c
3283
if (READ_ONCE(rcu_normal_wake_from_gp) < 1) {
kernel/rcu/tree.c
339
if (READ_ONCE(*vp))
kernel/rcu/tree.c
3689
time_before(jiffies, READ_ONCE(rcu_state.gp_start) +
kernel/rcu/tree.c
3711
unlikely(READ_ONCE(rdp->gpwrap))) /* outside lock */
kernel/rcu/tree.c
3756
unsigned long gseq = READ_ONCE(rcu_state.barrier_sequence);
kernel/rcu/tree.c
3757
unsigned long lseq = READ_ONCE(rdp->barrier_seq_snap);
kernel/rcu/tree.c
3874
WARN_ON_ONCE(READ_ONCE(rdp->barrier_seq_snap) != gseq);
kernel/rcu/tree.c
3884
WARN_ON_ONCE(READ_ONCE(rdp->barrier_seq_snap) != gseq);
kernel/rcu/tree.c
3937
unsigned long old = READ_ONCE(rcu_barrier_last_throttle);
kernel/rcu/tree.c
3948
old = READ_ONCE(rcu_barrier_last_throttle);
kernel/rcu/tree.c
3997
return READ_ONCE(rnp->qsmaskinitnext);
kernel/rcu/tree.c
4061
return !!READ_ONCE(rcu_state.n_online_cpus);
kernel/rcu/tree.c
4250
rdp->n_force_qs_snap = READ_ONCE(rcu_state.n_force_qs);
kernel/rcu/tree.c
4269
rdp->gp_seq = READ_ONCE(rnp->gp_seq);
kernel/rcu/tree.c
4362
rdp->rcu_onl_gp_seq = READ_ONCE(rcu_state.gp_seq);
kernel/rcu/tree.c
4363
rdp->rcu_onl_gp_state = READ_ONCE(rcu_state.gp_state);
kernel/rcu/tree.c
4426
rdp->rcu_ofl_gp_seq = READ_ONCE(rcu_state.gp_seq);
kernel/rcu/tree.c
4427
rdp->rcu_ofl_gp_state = READ_ONCE(rcu_state.gp_state);
kernel/rcu/tree.c
4523
blkd = !!(READ_ONCE(rnp->qsmask) & rdp->grpmask);
kernel/rcu/tree.c
4524
trace_rcu_grace_period(rcu_state.name, READ_ONCE(rnp->gp_seq),
kernel/rcu/tree.c
469
j = READ_ONCE(jiffies_till_first_fqs) +
kernel/rcu/tree.c
470
2 * READ_ONCE(jiffies_till_next_fqs);
kernel/rcu/tree.c
523
return READ_ONCE(rcu_state.gp_seq);
kernel/rcu/tree.c
552
*flags = READ_ONCE(rcu_state.gp_flags);
kernel/rcu/tree.c
560
return ((READ_ONCE(rcu_state.gp_seq) & 0xffffULL) << 40) |
kernel/rcu/tree.c
561
((READ_ONCE(rcu_state.expedited_sequence) & 0xffffffULL) << 16) |
kernel/rcu/tree.c
562
(READ_ONCE(rcu_state.gp_seq_polled) & 0xffffULL);
kernel/rcu/tree.c
674
!READ_ONCE(rdp->rcu_urgent_qs) ||
kernel/rcu/tree.c
675
READ_ONCE(rdp->rcu_forced_tick)) {
kernel/rcu/tree.c
688
if (READ_ONCE(rdp->rcu_urgent_qs) && !rdp->rcu_forced_tick) {
kernel/rcu/tree.c
808
WRITE_ONCE(rdp->gpwrap_count, READ_ONCE(rdp->gpwrap_count) + 1);
kernel/rcu/tree.c
88
return READ_ONCE(rdp->gpwrap_count);
kernel/rcu/tree.c
920
jtsq = READ_ONCE(jiffies_to_sched_qs);
kernel/rcu/tree.c
921
if (!READ_ONCE(rdp->rcu_need_heavy_qs) &&
kernel/rcu/tree.c
941
(time_after(jiffies, READ_ONCE(rdp->last_fqs_resched) + jtsq * 3) ||
kernel/rcu/tree.c
957
READ_ONCE(rdp->last_fqs_resched) + jtsq)) {
kernel/rcu/tree.c
995
trace_rcu_future_grace_period(rcu_state.name, READ_ONCE(rnp->gp_seq),
kernel/rcu/tree_exp.h
1004
pr_info("%s: i = %d s = %lx gp_seq_polled = %lx\n", __func__, i, s, READ_ONCE(rcu_state.gp_seq_polled));
kernel/rcu/tree_exp.h
164
return READ_ONCE(rnp->exp_tasks) == NULL &&
kernel/rcu/tree_exp.h
165
READ_ONCE(rnp->expmask) == 0;
kernel/rcu/tree_exp.h
308
if (ULONG_CMP_LT(READ_ONCE(rnp->exp_seq_rq), s) &&
kernel/rcu/tree_exp.h
310
ULONG_CMP_LT(READ_ONCE(rnp_root->exp_seq_rq), s)) &&
kernel/rcu/tree_exp.h
460
return !!READ_ONCE(rcu_exp_gp_kworker);
kernel/rcu/tree_exp.h
465
return !!READ_ONCE(rnp->exp_kworker);
kernel/rcu/tree_exp.h
476
kthread_queue_work(READ_ONCE(rnp->exp_kworker), &rnp->rew.rew_work);
kernel/rcu/tree_exp.h
516
if (!READ_ONCE(rnp->expmask))
kernel/rcu/tree_exp.h
566
if (READ_ONCE(csd_lock_suppress_rcu_stall) && csd_lock_is_stuck()) {
kernel/rcu/tree_exp.h
578
if (!(READ_ONCE(rnp->expmask) & mask))
kernel/rcu/tree_exp.h
613
if (!(READ_ONCE(rnp->expmask) & mask))
kernel/rcu/tree_exp.h
644
mask = READ_ONCE(rnp->expmask);
kernel/rcu/tree_exp.h
655
j = READ_ONCE(jiffies_till_first_fqs);
kernel/rcu/tree_exp.h
702
if (ULONG_CMP_LT(READ_ONCE(rnp->exp_seq_rq), s)) {
kernel/rcu/tree_exp.h
762
if (WARN_ON_ONCE(!(READ_ONCE(rnp->expmask) & rdp->grpmask) ||
kernel/rcu/tree_exp.h
763
READ_ONCE(rdp->cpu_no_qs.b.exp)))
kernel/rcu/tree_exp.h
845
if (!READ_ONCE(rnp->exp_tasks)) {
kernel/rcu/tree_exp.h
872
if (!(READ_ONCE(rnp->expmask) & rdp->grpmask) ||
kernel/rcu/tree_nocb.h
1194
count += READ_ONCE(rdp->lazy_len);
kernel/rcu/tree_nocb.h
1233
if (!READ_ONCE(rdp->lazy_len))
kernel/rcu/tree_nocb.h
1242
_count = READ_ONCE(rdp->lazy_len);
kernel/rcu/tree_nocb.h
1518
rnp->grplo, rnp->grphi, READ_ONCE(rdp->nocb_gp_loops),
kernel/rcu/tree_nocb.h
209
if (!READ_ONCE(rdp_gp->nocb_gp_kthread)) {
kernel/rcu/tree_nocb.h
218
if (READ_ONCE(rdp_gp->nocb_gp_sleep)) {
kernel/rcu/tree_nocb.h
407
bool bypass_is_lazy = (ncbs == READ_ONCE(rdp->lazy_len));
kernel/rcu/tree_nocb.h
459
if ((ncbs && !bypass_is_lazy && j != READ_ONCE(rdp->nocb_bypass_first)) ||
kernel/rcu/tree_nocb.h
461
(time_after(j, READ_ONCE(rdp->nocb_bypass_first) + rcu_get_jiffies_lazy_flush()))) ||
kernel/rcu/tree_nocb.h
540
t = READ_ONCE(rdp->nocb_gp_kthread);
kernel/rcu/tree_nocb.h
550
lazy_len = READ_ONCE(rdp->lazy_len);
kernel/rcu/tree_nocb.h
620
!READ_ONCE(my_rdp->nocb_gp_sleep));
kernel/rcu/tree_nocb.h
674
lazy_ncbs = READ_ONCE(rdp->lazy_len);
kernel/rcu/tree_nocb.h
677
(time_after(j, READ_ONCE(rdp->nocb_bypass_first) + rcu_get_jiffies_lazy_flush()) ||
kernel/rcu/tree_nocb.h
681
(time_after(j, READ_ONCE(rdp->nocb_bypass_first) + 1) ||
kernel/rcu/tree_nocb.h
693
lazy_ncbs = READ_ONCE(rdp->lazy_len);
kernel/rcu/tree_nocb.h
786
!READ_ONCE(my_rdp->nocb_gp_sleep));
kernel/rcu/tree_nocb.h
801
rdp_toggling = READ_ONCE(my_rdp->nocb_toggling_rdp);
kernel/rcu/tree_nocb.h
847
return !READ_ONCE(rdp->nocb_cb_sleep) || kthread_should_park();
kernel/rcu/tree_nocb.h
877
} else if (READ_ONCE(rdp->nocb_cb_sleep)) {
kernel/rcu/tree_nocb.h
937
return READ_ONCE(rdp->nocb_defer_wakeup) >= level;
kernel/rcu/tree_plugin.h
1048
if (READ_ONCE(rdp->cpu_no_qs.b.exp))
kernel/rcu/tree_plugin.h
1160
if (READ_ONCE(rnp->exp_tasks) == NULL &&
kernel/rcu/tree_plugin.h
1161
READ_ONCE(rnp->boost_tasks) == NULL)
kernel/rcu/tree_plugin.h
1210
return READ_ONCE(rnp->exp_tasks) != NULL ||
kernel/rcu/tree_plugin.h
1211
READ_ONCE(rnp->boost_tasks) != NULL;
kernel/rcu/tree_plugin.h
1227
rcu_wait(READ_ONCE(rnp->boost_tasks) ||
kernel/rcu/tree_plugin.h
1228
READ_ONCE(rnp->exp_tasks));
kernel/rcu/tree_plugin.h
1278
READ_ONCE(rnp->boost_kthread_status));
kernel/rcu/tree_plugin.h
1355
time_before(jiffies, READ_ONCE(rcu_state.gp_start) + HZ)))
kernel/rcu/tree_plugin.h
383
return READ_ONCE(rnp->gp_tasks) != NULL;
kernel/rcu/tree_plugin.h
391
WRITE_ONCE(current->rcu_read_lock_nesting, READ_ONCE(current->rcu_read_lock_nesting) + 1);
kernel/rcu/tree_plugin.h
396
int ret = READ_ONCE(current->rcu_read_lock_nesting) - 1;
kernel/rcu/tree_plugin.h
437
if (unlikely(READ_ONCE(t->rcu_read_unlock_special.s)))
kernel/rcu/tree_plugin.h
603
READ_ONCE(t->rcu_read_unlock_special.s)) &&
kernel/rcu/tree_plugin.h
681
if (t->rcu_blocked_node && READ_ONCE(t->rcu_blocked_node->exp_tasks))
kernel/rcu/tree_plugin.h
690
if (rdp->grpmask & READ_ONCE(rnp->expmask))
kernel/rcu/tree_plugin.h
702
((rdp->grpmask & READ_ONCE(rnp->qsmask)) || t->rcu_blocked_node))
kernel/rcu/tree_plugin.h
878
(long)READ_ONCE(rnp->gp_seq), (long)rnp->completedqs);
kernel/rcu/tree_plugin.h
883
__func__, READ_ONCE(rnp->gp_tasks), data_race(rnp->boost_tasks),
kernel/rcu/tree_plugin.h
884
READ_ONCE(rnp->exp_tasks));
kernel/rcu/tree_stall.h
1011
ULONG_CMP_GE(READ_ONCE(rnp_root->gp_seq),
kernel/rcu/tree_stall.h
1012
READ_ONCE(rnp_root->gp_seq_needed)) ||
kernel/rcu/tree_stall.h
1016
if (time_before(j, READ_ONCE(rcu_state.gp_req_activity) + gpssdelay) ||
kernel/rcu/tree_stall.h
1017
time_before(j, READ_ONCE(rcu_state.gp_activity) + gpssdelay) ||
kernel/rcu/tree_stall.h
1024
ULONG_CMP_GE(READ_ONCE(rnp_root->gp_seq),
kernel/rcu/tree_stall.h
1025
READ_ONCE(rnp_root->gp_seq_needed)) ||
kernel/rcu/tree_stall.h
1026
time_before(j, READ_ONCE(rcu_state.gp_req_activity) + gpssdelay) ||
kernel/rcu/tree_stall.h
1027
time_before(j, READ_ONCE(rcu_state.gp_activity) + gpssdelay) ||
kernel/rcu/tree_stall.h
1038
ULONG_CMP_GE(READ_ONCE(rnp_root->gp_seq),
kernel/rcu/tree_stall.h
1039
READ_ONCE(rnp_root->gp_seq_needed)) ||
kernel/rcu/tree_stall.h
1040
time_before(j, READ_ONCE(rcu_state.gp_req_activity) + gpssdelay) ||
kernel/rcu/tree_stall.h
1041
time_before(j, READ_ONCE(rcu_state.gp_activity) + gpssdelay) ||
kernel/rcu/tree_stall.h
1071
__func__, jiffies - data_race(READ_ONCE(rcu_state.gp_start)));
kernel/rcu/tree_stall.h
1075
__func__, jiffies - data_race(READ_ONCE(rcu_state.gp_end)));
kernel/rcu/tree_stall.h
113
int till_stall_check = READ_ONCE(rcu_cpu_stall_timeout);
kernel/rcu/tree_stall.h
212
rcu_state.n_force_qs_gpstart = READ_ONCE(rcu_state.n_force_qs);
kernel/rcu/tree_stall.h
231
if (!READ_ONCE(rcu_kick_kthreads))
kernel/rcu/tree_stall.h
233
j = READ_ONCE(rcu_state.jiffies_kick_kthreads);
kernel/rcu/tree_stall.h
235
(rcu_gp_in_progress() || READ_ONCE(rcu_state.gp_flags))) {
kernel/rcu/tree_stall.h
450
unsigned long j = jiffies - READ_ONCE(rcu_state.gp_activity);
kernel/rcu/tree_stall.h
471
j = jiffies - READ_ONCE(rdp->rcuc_activity);
kernel/rcu/tree_stall.h
580
data_race(READ_ONCE(rcu_state.gp_flags)),
kernel/rcu/tree_stall.h
582
data_race(READ_ONCE(rcu_state.gp_state)),
kernel/rcu/tree_stall.h
583
gpk ? data_race(READ_ONCE(gpk->__state)) : ~0, cpu);
kernel/rcu/tree_stall.h
592
} else if (!(data_race(READ_ONCE(rdp->mynode->qsmask)) & rdp->grpmask)) {
kernel/rcu/tree_stall.h
614
jiffies_fqs = READ_ONCE(rcu_state.jiffies_force_qs);
kernel/rcu/tree_stall.h
618
gpk && !READ_ONCE(gpk->on_rq)) {
kernel/rcu/tree_stall.h
623
data_race(READ_ONCE(rcu_state.gp_flags)), // Diagnostic read
kernel/rcu/tree_stall.h
625
data_race(READ_ONCE(gpk->__state)));
kernel/rcu/tree_stall.h
687
gpa = data_race(READ_ONCE(rcu_state.gp_activity));
kernel/rcu/tree_stall.h
690
data_race(READ_ONCE(jiffies_till_next_fqs)),
kernel/rcu/tree_stall.h
691
data_race(READ_ONCE(rcu_get_root()->qsmask)));
kernel/rcu/tree_stall.h
695
if (ULONG_CMP_GE(jiffies, READ_ONCE(rcu_state.jiffies_stall)))
kernel/rcu/tree_stall.h
750
if (ULONG_CMP_GE(jiffies, READ_ONCE(rcu_state.jiffies_stall)))
kernel/rcu/tree_stall.h
784
if ((rcu_stall_is_suppressed() && !READ_ONCE(rcu_kick_kthreads)) ||
kernel/rcu/tree_stall.h
795
if (READ_ONCE(rcu_state.nr_fqs_jiffies_stall) > 0)
kernel/rcu/tree_stall.h
818
gs1 = READ_ONCE(rcu_state.gp_seq);
kernel/rcu/tree_stall.h
820
js = READ_ONCE(rcu_state.jiffies_stall);
kernel/rcu/tree_stall.h
822
gps = READ_ONCE(rcu_state.gp_start);
kernel/rcu/tree_stall.h
824
gs2 = READ_ONCE(rcu_state.gp_seq);
kernel/rcu/tree_stall.h
832
self_detected = READ_ONCE(rnp->qsmask) & rdp->grpmask;
kernel/rcu/tree_stall.h
84
int cpu_stall_timeout = READ_ONCE(rcu_exp_cpu_stall_timeout);
kernel/rcu/tree_stall.h
849
if (READ_ONCE(csd_lock_suppress_rcu_stall) && csd_lock_is_stuck()) {
kernel/rcu/tree_stall.h
859
if (READ_ONCE(rcu_cpu_stall_ftrace_dump))
kernel/rcu/tree_stall.h
862
if (READ_ONCE(rcu_state.jiffies_stall) == jn) {
kernel/rcu/tree_stall.h
899
if (data_race(READ_ONCE(rnp->qsmask))) {
kernel/rcu/tree_stall.h
902
if (READ_ONCE(rnp->gp_tasks))
kernel/rcu/tree_stall.h
945
struct task_struct *t = READ_ONCE(rcu_state.gp_kthread);
kernel/rcu/tree_stall.h
948
ja = j - data_race(READ_ONCE(rcu_state.gp_activity));
kernel/rcu/tree_stall.h
949
jr = j - data_race(READ_ONCE(rcu_state.gp_req_activity));
kernel/rcu/tree_stall.h
950
js = j - data_race(READ_ONCE(rcu_state.gp_start));
kernel/rcu/tree_stall.h
951
jw = j - data_race(READ_ONCE(rcu_state.gp_wake_time));
kernel/rcu/tree_stall.h
954
data_race(READ_ONCE(rcu_state.gp_state)),
kernel/rcu/tree_stall.h
955
t ? data_race(READ_ONCE(t->__state)) : 0x1ffff, t ? t->rt_priority : 0xffU,
kernel/rcu/tree_stall.h
956
js, ja, jr, jw, (long)data_race(READ_ONCE(rcu_state.gp_wake_seq)),
kernel/rcu/tree_stall.h
957
(long)data_race(READ_ONCE(rcu_state.gp_seq)),
kernel/rcu/tree_stall.h
958
(long)data_race(READ_ONCE(rcu_get_root()->gp_seq_needed)),
kernel/rcu/tree_stall.h
959
data_race(READ_ONCE(rcu_state.gp_max)),
kernel/rcu/tree_stall.h
960
data_race(READ_ONCE(rcu_state.gp_flags)));
kernel/rcu/tree_stall.h
962
if (ULONG_CMP_GE(READ_ONCE(rcu_state.gp_seq), READ_ONCE(rnp->gp_seq_needed)) &&
kernel/rcu/tree_stall.h
963
!data_race(READ_ONCE(rnp->qsmask)) && !data_race(READ_ONCE(rnp->boost_tasks)) &&
kernel/rcu/tree_stall.h
964
!data_race(READ_ONCE(rnp->exp_tasks)) && !data_race(READ_ONCE(rnp->gp_tasks)))
kernel/rcu/tree_stall.h
968
(long)data_race(READ_ONCE(rnp->gp_seq)),
kernel/rcu/tree_stall.h
969
(long)data_race(READ_ONCE(rnp->gp_seq_needed)),
kernel/rcu/tree_stall.h
970
data_race(READ_ONCE(rnp->qsmask)),
kernel/rcu/tree_stall.h
971
".b"[!!data_race(READ_ONCE(rnp->boost_kthread_task))],
kernel/rcu/tree_stall.h
972
".B"[!!data_race(READ_ONCE(rnp->boost_tasks))],
kernel/rcu/tree_stall.h
973
".E"[!!data_race(READ_ONCE(rnp->exp_tasks))],
kernel/rcu/tree_stall.h
974
".G"[!!data_race(READ_ONCE(rnp->gp_tasks))],
kernel/rcu/tree_stall.h
975
data_race(READ_ONCE(rnp->n_boosts)));
kernel/rcu/tree_stall.h
980
if (READ_ONCE(rdp->gpwrap) ||
kernel/rcu/tree_stall.h
981
ULONG_CMP_GE(READ_ONCE(rcu_state.gp_seq),
kernel/rcu/tree_stall.h
982
READ_ONCE(rdp->gp_seq_needed)))
kernel/rcu/tree_stall.h
985
cpu, (long)data_race(READ_ONCE(rdp->gp_seq_needed)));
kernel/rcu/tree_stall.h
990
cbs += data_race(READ_ONCE(rdp->n_cbs_invoked));
kernel/rcu/update.c
143
return READ_ONCE(rcu_normal) &&
kernel/rcu/update.c
320
return rcu_scheduler_active != RCU_SCHEDULER_INACTIVE && READ_ONCE(debug_locks) &&
kernel/scftorture.c
186
bool isdone = READ_ONCE(scfdone);
kernel/scftorture.c
296
if (WARN_ON_ONCE(unlikely(!READ_ONCE(scfcp->scfc_in))))
kernel/sched/autogroup.h
36
int enabled = READ_ONCE(sysctl_sched_autogroup_enabled);
kernel/sched/completion.c
320
if (!READ_ONCE(x->done))
kernel/sched/completion.c
346
if (!READ_ONCE(x->done))
kernel/sched/core.c
10526
if (!mm || !READ_ONCE(mm->mm_cid.users))
kernel/sched/core.c
2086
state = READ_ONCE(p->__state);
kernel/sched/core.c
2209
if (READ_ONCE(p->__state) & state)
kernel/sched/core.c
2212
if (READ_ONCE(p->saved_state) & state)
kernel/sched/core.c
2954
if (task_on_cpu(rq, p) || READ_ONCE(p->__state) == TASK_WAKING) {
kernel/sched/core.c
3245
unsigned int state = READ_ONCE(p->__state);
kernel/sched/core.c
4160
if (READ_ONCE(p->on_rq) && ttwu_runnable(p, wake_flags))
kernel/sched/core.c
4253
unsigned int state = READ_ONCE(p->__state);
kernel/sched/core.c
5148
prev_state = READ_ONCE(prev->__state);
kernel/sched/core.c
5499
int latency_warn_ms = READ_ONCE(sysctl_resched_latency_warn_ms);
kernel/sched/core.c
5865
if (!preempt && READ_ONCE(prev->__state) && prev->non_block_count) {
kernel/sched/core.c
6548
unsigned long state = READ_ONCE(donor->__state);
kernel/sched/core.c
6631
if (!READ_ONCE(owner->on_rq) || owner->se.sched_delayed) {
kernel/sched/core.c
6831
prev_state = READ_ONCE(prev->__state);
kernel/sched/core.c
7873
unsigned int state = READ_ONCE(p->__state);
kernel/sched/core.c
9724
total += READ_ONCE(tg->cfs_rq[i]->throttled_clock_self_time);
kernel/sched/core.c
993
typeof(ti->flags) val = READ_ONCE(ti->flags);
kernel/sched/cpufreq_schedutil.c
86
if (unlikely(READ_ONCE(sg_policy->limits_changed))) {
kernel/sched/cputime.c
890
int state = READ_ONCE(vtime->state);
kernel/sched/deadline.c
1983
if (!dl_task(p) || READ_ONCE(p->__state) == TASK_DEAD) {
kernel/sched/deadline.c
1986
if (READ_ONCE(p->__state) == TASK_DEAD && dl_se->dl_non_contending) {
kernel/sched/deadline.c
2152
state = READ_ONCE(p->__state);
kernel/sched/deadline.c
2419
curr = READ_ONCE(rq->curr); /* unlocked access */
kernel/sched/deadline.c
2420
donor = READ_ONCE(rq->donor);
kernel/sched/deadline.c
2460
if (READ_ONCE(p->__state) != TASK_WAKING)
kernel/sched/deadline.c
3674
max = (u64)READ_ONCE(sysctl_sched_dl_period_max) * NSEC_PER_USEC;
kernel/sched/deadline.c
3675
min = (u64)READ_ONCE(sysctl_sched_dl_period_min) * NSEC_PER_USEC;
kernel/sched/deadline.c
441
if (!dl_task || READ_ONCE(p->__state) == TASK_DEAD) {
kernel/sched/deadline.c
444
if (READ_ONCE(p->__state) == TASK_DEAD)
kernel/sched/ext.c
1940
if (unlikely(READ_ONCE(scx_aborting)) && dsq->id != SCX_DSQ_BYPASS)
kernel/sched/ext.c
2499
while (READ_ONCE(cpu_rq(cpu)->scx.kick_sync) == ksyncs[cpu]) {
kernel/sched/ext.c
2813
last_runnable + READ_ONCE(scx_watchdog_timeout)))) {
kernel/sched/ext.c
2841
READ_ONCE(scx_watchdog_timeout) / 2);
kernel/sched/ext.c
2856
last_check = READ_ONCE(scx_watchdog_timestamp);
kernel/sched/ext.c
2858
last_check + READ_ONCE(scx_watchdog_timeout)))) {
kernel/sched/ext.c
3085
scx->slice = READ_ONCE(scx_slice_dfl);
kernel/sched/ext.c
3176
return unlikely(READ_ONCE(p->__state) == TASK_DEAD) &&
kernel/sched/ext.c
3263
if (scx_enabled() && READ_ONCE(p->scx.disallow) &&
kernel/sched/ext.c
3703
return sysfs_emit(buf, "%d\n", READ_ONCE(scx_switching_all));
kernel/sched/ext.c
3850
if (READ_ONCE(scx_switching_all))
kernel/sched/ext.c
3975
s32 delta = READ_ONCE(donor_dsq->nr) - nr_donor_target;
kernel/sched/ext.c
3984
min_delta_us = READ_ONCE(scx_bypass_lb_intv_us) / SCX_BYPASS_LB_MIN_DELTA_DIV;
kernel/sched/ext.c
3985
if (delta < DIV_ROUND_UP(min_delta_us, READ_ONCE(scx_slice_bypass_us)))
kernel/sched/ext.c
4042
if (READ_ONCE(donee_dsq->nr) >= nr_donee_target)
kernel/sched/ext.c
4077
u32 nr = READ_ONCE(cpu_rq(cpu)->scx.bypass_dsq.nr);
kernel/sched/ext.c
4099
if (READ_ONCE(cpu_rq(cpu)->scx.bypass_dsq.nr) < nr_target)
kernel/sched/ext.c
4113
if (READ_ONCE(donor_dsq->nr) <= nr_donor_target)
kernel/sched/ext.c
4124
u32 nr = READ_ONCE(cpu_rq(cpu)->scx.bypass_dsq.nr);
kernel/sched/ext.c
4151
if (unlikely(!sch) || !READ_ONCE(scx_bypass_depth))
kernel/sched/ext.c
4157
intv_us = READ_ONCE(scx_bypass_lb_intv_us);
kernel/sched/ext.c
4213
WRITE_ONCE(scx_slice_dfl, READ_ONCE(scx_slice_bypass_us) * NSEC_PER_USEC);
kernel/sched/ext.c
4218
intv_us = READ_ONCE(scx_bypass_lb_intv_us);
kernel/sched/ext.c
5100
if (WARN_ON_ONCE(READ_ONCE(scx_aborting)))
kernel/sched/ext.c
5168
READ_ONCE(scx_watchdog_timeout) / 2);
kernel/sched/ext.c
5268
p->scx.slice = READ_ONCE(scx_slice_dfl);
kernel/sched/ext.c
5335
if (!READ_ONCE(helper)) {
kernel/sched/ext.c
5353
kthread_queue_work(READ_ONCE(helper), &cmd.work);
kernel/sched/ext.c
5777
const char *all = READ_ONCE(scx_switching_all) ? "+all" : "";
kernel/sched/ext.c
6119
if (unlikely(READ_ONCE(scx_aborting)))
kernel/sched/ext.c
6668
ret = READ_ONCE(this_rq()->scx.local_dsq.nr);
kernel/sched/ext.c
6674
ret = READ_ONCE(cpu_rq(cpu)->scx.local_dsq.nr);
kernel/sched/ext.c
6680
ret = READ_ONCE(dsq->nr);
kernel/sched/ext.c
6751
READ_ONCE(kit->dsq->seq));
kernel/sched/ext.c
714
(dst_e)->kind += READ_ONCE((src_e)->kind); \
kernel/sched/ext.c
7360
clock = READ_ONCE(rq->scx.clock);
kernel/sched/ext.c
984
WRITE_ONCE(dsq->nr, READ_ONCE(dsq->nr) + delta);
kernel/sched/ext.c
989
p->scx.slice = READ_ONCE(scx_slice_dfl);
kernel/sched/fair.c
10373
return sched_asym(env->sd, env->dst_cpu, READ_ONCE(group->asym_prefer_cpu));
kernel/sched/fair.c
10610
return sched_asym_prefer(READ_ONCE(sds->busiest->asym_prefer_cpu),
kernel/sched/fair.c
10611
READ_ONCE(sg->asym_prefer_cpu));
kernel/sched/fair.c
10741
if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
kernel/sched/fair.c
12501
if (READ_ONCE(nohz.has_blocked_load) &&
kernel/sched/fair.c
12502
time_after(now, READ_ONCE(nohz.next_blocked)))
kernel/sched/fair.c
12602
if (READ_ONCE(nohz.needs_update))
kernel/sched/fair.c
12720
if (!time_after(jiffies, READ_ONCE(rq->last_blocked_load_update_tick)))
kernel/sched/fair.c
12891
if (!READ_ONCE(nohz.has_blocked_load) ||
kernel/sched/fair.c
12892
time_before(jiffies, READ_ONCE(nohz.next_blocked)))
kernel/sched/fair.c
13663
if (READ_ONCE(p->__state) == TASK_NEW)
kernel/sched/fair.c
1438
state = READ_ONCE(tsk->__state);
kernel/sched/fair.c
1554
(lockdep_is_held(__rq_lockp(task_rq(p))) && !READ_ONCE(p->on_cpu)));
kernel/sched/fair.c
1589
unsigned int scan_size = READ_ONCE(sysctl_numa_balancing_scan_size);
kernel/sched/fair.c
1772
max_dist = READ_ONCE(sched_max_numa_distance);
kernel/sched/fair.c
2210
if (READ_ONCE(rq->numa_migrate_on) ||
kernel/sched/fair.c
2331
if (READ_ONCE(dst_rq->numa_migrate_on))
kernel/sched/fair.c
2968
seq = READ_ONCE(p->mm->numa_scan_seq);
kernel/sched/fair.c
3111
tsk = READ_ONCE(cpu_rq(cpu)->curr);
kernel/sched/fair.c
3316
WRITE_ONCE(p->mm->numa_scan_seq, READ_ONCE(p->mm->numa_scan_seq) + 1);
kernel/sched/fair.c
3329
if ((READ_ONCE(current->mm->numa_scan_seq) - vma->numab_state->start_scan_seq) < 2)
kernel/sched/fair.c
3350
if (READ_ONCE(mm->numa_scan_seq) >
kernel/sched/fair.c
3528
vma->numab_state->pids_active[0] = READ_ONCE(vma->numab_state->pids_active[1]);
kernel/sched/fair.c
3788
typeof(*ptr) res, var = READ_ONCE(*ptr); \
kernel/sched/fair.c
3994
tg_shares = READ_ONCE(tg->shares);
kernel/sched/fair.c
4850
return READ_ONCE(p->se.avg.util_avg);
kernel/sched/fair.c
4855
return READ_ONCE(p->se.avg.runnable_avg);
kernel/sched/fair.c
4860
return READ_ONCE(p->se.avg.util_est) & ~UTIL_AVG_UNCHANGED;
kernel/sched/fair.c
4919
ewma = READ_ONCE(p->se.avg.util_est);
kernel/sched/fair.c
6869
return !sched_energy_enabled() || READ_ONCE(rd->overutilized);
kernel/sched/fair.c
7246
if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
kernel/sched/fair.c
7250
load = READ_ONCE(cfs_rq->avg.load_avg);
kernel/sched/fair.c
7269
if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
kernel/sched/fair.c
7273
runnable = READ_ONCE(cfs_rq->avg.runnable_avg);
kernel/sched/fair.c
7583
return READ_ONCE(sds->has_idle_cores);
kernel/sched/fair.c
7713
nr = READ_ONCE(sd_share->nr_idle_scan) + 1;
kernel/sched/fair.c
8008
unsigned long util = READ_ONCE(cfs_rq->avg.util_avg);
kernel/sched/fair.c
8012
runnable = READ_ONCE(cfs_rq->avg.runnable_avg);
kernel/sched/fair.c
8030
util_est = READ_ONCE(cfs_rq->avg.util_est);
kernel/sched/fair.c
8095
if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
kernel/sched/fair.c
9967
while ((se = READ_ONCE(cfs_rq->h_load_next)) != NULL) {
kernel/sched/idle.c
434
while (!READ_ONCE(it.done))
kernel/sched/isolation.c
115
READ_ONCE(housekeeping.flags) & BIT(type))
kernel/sched/isolation.c
33
return !!(READ_ONCE(housekeeping.flags) & BIT(type));
kernel/sched/isolation.c
69
if (READ_ONCE(housekeeping.flags) & BIT(type))
kernel/sched/loadavg.c
224
if (!time_before(jiffies, READ_ONCE(calc_load_update)))
kernel/sched/loadavg.c
272
this_rq->calc_load_update = READ_ONCE(calc_load_update);
kernel/sched/loadavg.c
310
sample_window = READ_ONCE(calc_load_update);
kernel/sched/loadavg.c
356
sample_window = READ_ONCE(calc_load_update);
kernel/sched/membarrier.c
245
if (READ_ONCE(rq->membarrier_state) == membarrier_state)
kernel/sched/membarrier.c
284
if (!(READ_ONCE(cpu_rq(cpu)->membarrier_state) &
kernel/sched/pelt.h
20
return READ_ONCE(rq->avg_hw.load_avg);
kernel/sched/rt.c
1315
state = READ_ONCE(p->__state);
kernel/sched/rt.c
1512
curr = READ_ONCE(rq->curr); /* unlocked access */
kernel/sched/rt.c
1513
donor = READ_ONCE(rq->donor);
kernel/sched/sched.h
1060
return READ_ONCE(rd->overloaded);
kernel/sched/sched.h
2395
return READ_ONCE(p->on_rq) == TASK_ON_RQ_QUEUED;
kernel/sched/sched.h
2400
return READ_ONCE(p->on_rq) == TASK_ON_RQ_MIGRATING;
kernel/sched/sched.h
3522
return READ_ONCE(rq->avg_dl.util_avg);
kernel/sched/sched.h
3531
return READ_ONCE(rq->avg_rt.util_avg);
kernel/sched/sched.h
3566
return READ_ONCE(rq->uclamp[clamp_id].value);
kernel/sched/sched.h
3590
max_util = READ_ONCE(rq->uclamp[UCLAMP_MAX].value);
kernel/sched/sched.h
3669
return READ_ONCE(rq->avg_irq.util_avg);
kernel/sched/sched.h
3736
if (READ_ONCE(rq->membarrier_state) == membarrier_state)
kernel/sched/sched.h
3849
unsigned int cid = __mm_get_cid(mm, READ_ONCE(mm->mm_cid.max_cids));
kernel/sched/sched.h
3896
max_cids = READ_ONCE(mm->mm_cid.max_cids);
kernel/sched/sched.h
3931
max_cids = READ_ONCE(mm->mm_cid.max_cids);
kernel/sched/sched.h
3967
mode = READ_ONCE(mm->mm_cid.mode);
kernel/sched/sched.h
3983
mode = READ_ONCE(mm->mm_cid.mode);
kernel/sched/sched.h
3990
if (!cid_in_transit(mode) && cid < READ_ONCE(mm->mm_cid.max_cids)) {
kernel/sched/wait_bit.c
260
unsigned long now = READ_ONCE(jiffies);
kernel/seccomp.c
2546
f = READ_ONCE(task->seccomp.filter);
kernel/seccomp.c
410
READ_ONCE(current->seccomp.filter);
kernel/seccomp.c
577
if (READ_ONCE(tsk->seccomp.filter) == NULL)
kernel/signal.c
2192
if (tsk->parent_exec_id != READ_ONCE(tsk->parent->self_exec_id))
kernel/smp.c
238
unsigned int flags = READ_ONCE(csd->node.u_flags);
kernel/smp.c
292
*bug_id, READ_ONCE(per_cpu(cur_csd_func, cpux)),
kernel/smp.c
293
READ_ONCE(per_cpu(cur_csd_info, cpux)));
kernel/stop_machine.c
228
newstate = READ_ONCE(msdata->state);
kernel/sysctl.c
1023
val = convdiv * READ_ONCE(*i) / convmul;
kernel/sysctl.c
393
uint val = READ_ONCE(*k_ptr);
kernel/sysctl.c
477
int val = READ_ONCE(*k_ptr);
kernel/sysctl.c
818
val = READ_ONCE(*data);
kernel/sysctl.c
956
val = READ_ONCE(*data);
kernel/task_work.c
133
work = READ_ONCE(*pprev);
kernel/task_work.c
137
work = READ_ONCE(*pprev);
kernel/task_work.c
210
work = READ_ONCE(task->task_works);
kernel/task_work.c
73
head = READ_ONCE(task->task_works);
kernel/time/clocksource-wdtest.c
155
WARN_ON_ONCE(READ_ONCE(wdtest_ktime_read_ndelays));
kernel/time/clocksource-wdtest.c
57
int wkrn = READ_ONCE(wdtest_ktime_read_ndelays);
kernel/time/clocksource-wdtest.c
66
if (READ_ONCE(wdtest_ktime_read_fuzz)) {
kernel/time/hrtimer.c
1442
struct hrtimer_clock_base *base = READ_ONCE(timer->base);
kernel/time/hrtimer.c
170
base = READ_ONCE(timer->base);
kernel/time/hrtimer.c
1710
base = READ_ONCE(timer->base);
kernel/time/hrtimer.c
1718
base != READ_ONCE(timer->base));
kernel/time/posix-cpu-timers.c
1113
if (READ_ONCE(pct->timers_active) && !READ_ONCE(pct->expiry_active)) {
kernel/time/posix-cpu-timers.c
1310
start = READ_ONCE(jiffies);
kernel/time/posix-cpu-timers.c
303
if (!READ_ONCE(pct->timers_active)) {
kernel/time/posix-cpu-timers.c
346
if (!READ_ONCE(pct->timers_active)) {
kernel/time/posix-cpu-timers.c
955
if (!READ_ONCE(pct->timers_active) || pct->expiry_active)
kernel/time/posix-timers.c
97
if ((READ_ONCE(timer->it_signal) == sig) && (timer->it_id == id))
kernel/time/tick-common.c
199
if (READ_ONCE(tick_do_timer_cpu) == TICK_DO_TIMER_BOOT) {
kernel/time/tick-common.c
88
if (READ_ONCE(tick_do_timer_cpu) == cpu) {
kernel/time/tick-sched.c
1044
tick_cpu = READ_ONCE(tick_do_timer_cpu);
kernel/time/tick-sched.c
1213
int tick_cpu = READ_ONCE(tick_do_timer_cpu);
kernel/time/tick-sched.c
241
tick_cpu = READ_ONCE(tick_do_timer_cpu);
kernel/time/tick-sched.c
261
ts->last_tick_jiffies = READ_ONCE(jiffies);
kernel/time/tick-sched.c
266
ts->last_tick_jiffies = READ_ONCE(jiffies);
kernel/time/tick-sched.c
639
if (tick_nohz_full_running && READ_ONCE(tick_do_timer_cpu) == cpu)
kernel/time/tick-sched.c
983
tick_cpu = READ_ONCE(tick_do_timer_cpu);
kernel/time/timekeeping.c
1217
base = READ_ONCE(cs->base);
kernel/time/timekeeping.c
1240
base = READ_ONCE(cs->base);
kernel/time/timekeeping.c
1422
struct clocksource_base *base = READ_ONCE(tk_core.timekeeper.tkr_mono.clock->base);
kernel/time/timekeeping.c
2680
struct clocksource *clock = READ_ONCE(tkr->clock);
kernel/time/timekeeping.c
2825
unsigned long active = READ_ONCE(aux_timekeepers);
kernel/time/timekeeping.c
2844
unsigned long active = READ_ONCE(aux_timekeepers);
kernel/time/timekeeping.c
293
struct clocksource *clock = READ_ONCE(tkr->clock);
kernel/time/timekeeping.c
3058
unsigned long active = READ_ONCE(aux_timekeepers);
kernel/time/timekeeping.c
914
return ktime_add(tmono, READ_ONCE(*offset));
kernel/time/timer.c
1000
tf = READ_ONCE(timer->flags);
kernel/time/timer.c
1543
tf = READ_ONCE(timer->flags);
kernel/time/timer.c
2380
if (time_before(jiffies, READ_ONCE(base->next_expiry)))
kernel/time/timer.c
2455
if (time_after_eq(jiffies, READ_ONCE(base->next_expiry)) ||
kernel/time/timer.c
973
__forward_timer_base(base, READ_ONCE(jiffies));
kernel/time/timer_migration.c
1118
if (READ_ONCE(tmc->wakeup) == KTIME_MAX)
kernel/time/timer_migration.c
1161
data->firstexp = READ_ONCE(group->next_expiry);
kernel/time/timer_migration.c
1220
if (data.now >= READ_ONCE(tmc->wakeup))
kernel/time/timer_migration.c
1255
ret = READ_ONCE(tmc->wakeup);
kernel/time/timer_migration.c
1445
nextevt = min_t(u64, nextevt, READ_ONCE(group->next_expiry));
kernel/time/timer_migration.c
573
group = READ_ONCE(group->parent);
kernel/time/timer_migration.c
610
if (!READ_ONCE(evt->ignore)) {
kernel/torture.c
193
return READ_ONCE(torture_online_cpus);
kernel/torture.c
622
while (READ_ONCE(fullstop) == FULLSTOP_SHUTDOWN) {
kernel/torture.c
692
if (READ_ONCE(fullstop) == FULLSTOP_DONTSTOP) {
kernel/torture.c
740
till_ns = READ_ONCE(stutter_till_abs_time);
kernel/torture.c
850
return READ_ONCE(torture_init_jiffies);
kernel/torture.c
872
if (READ_ONCE(fullstop) == FULLSTOP_SHUTDOWN) {
kernel/torture.c
911
return READ_ONCE(fullstop) != FULLSTOP_DONTSTOP;
kernel/trace/bpf_trace.c
552
ee = READ_ONCE(array->ptrs[index]);
kernel/trace/bpf_trace.c
630
ee = READ_ONCE(array->ptrs[index]);
kernel/trace/fgraph.c
597
offset = READ_ONCE(current->curr_ret_stack);
kernel/trace/fgraph.c
687
struct fgraph_ops *gops = READ_ONCE(fgraph_array[i]);
kernel/trace/fgraph.c
857
struct fgraph_ops *gops = READ_ONCE(fgraph_array[i]);
kernel/trace/fprobe.c
283
fp = READ_ONCE(node->fp);
kernel/trace/fprobe.c
394
fp = READ_ONCE(node->fp);
kernel/trace/fprobe.c
410
fp = READ_ONCE(node->fp);
kernel/trace/fprobe.c
430
fp = READ_ONCE(node->fp);
kernel/trace/fprobe.c
95
if (READ_ONCE(node->fp) != NULL) {
kernel/trace/ftrace.c
2639
addr = READ_ONCE(ops->direct_call);
kernel/trace/ring_buffer.c
1487
if (tail_page == READ_ONCE(cpu_buffer->tail_page)) {
kernel/trace/ring_buffer.c
3462
buffer_tail_page = READ_ONCE(cpu_buffer->tail_page);
kernel/trace/ring_buffer.c
3495
unsigned long bsize = READ_ONCE(cpu_buffer->buffer->subbuf_size);
kernel/trace/ring_buffer.c
3851
bpage = READ_ONCE(cpu_buffer->tail_page);
kernel/trace/ring_buffer.c
3926
while (cpu_buffer->commit_page != READ_ONCE(cpu_buffer->tail_page)) {
kernel/trace/ring_buffer.c
3967
if (unlikely(cpu_buffer->commit_page != READ_ONCE(cpu_buffer->tail_page)))
kernel/trace/ring_buffer.c
4471
tail_page = info->tail_page = READ_ONCE(cpu_buffer->tail_page);
kernel/trace/ring_buffer.c
4637
if (unlikely(READ_ONCE(cpu_buffer->buffer) != buffer)) {
kernel/trace/ring_buffer.c
5340
commit_page = READ_ONCE(cpu_buffer->commit_page);
kernel/trace/ring_buffer.c
5354
curr_commit_page = READ_ONCE(cpu_buffer->commit_page);
kernel/trace/ring_buffer.c
5355
curr_commit_ts = READ_ONCE(curr_commit_page->page->time_stamp);
kernel/trace/ring_buffer.c
5434
unsigned long bsize = READ_ONCE(cpu_buffer->buffer->subbuf_size);
kernel/trace/ring_buffer.c
666
struct buffer_page *tail_page = READ_ONCE(cpu_buffer->tail_page);
kernel/trace/ring_buffer_benchmark.c
194
while (!READ_ONCE(reader_finish)) {
kernel/trace/rv/rv.c
663
return READ_ONCE(monitoring_on);
kernel/trace/rv/rv_reactors.c
345
return READ_ONCE(reacting_on);
kernel/trace/trace.c
1245
struct trace_array *tr = READ_ONCE(printk_trace);
kernel/trace/trace.c
5393
module_delta = READ_ONCE(tr->module_delta);
kernel/trace/trace.c
5493
module_delta = READ_ONCE(tr->module_delta);
kernel/trace/trace.c
9272
module_delta = READ_ONCE(tr->module_delta);
kernel/trace/trace.h
1807
return READ_ONCE(file_inode(filp)->i_private);
kernel/trace/trace.h
1830
file = READ_ONCE(file_inode(filp)->i_private);
kernel/trace/trace_clock.c
115
prev_time = READ_ONCE(trace_clock_struct.prev_time);
kernel/trace/trace_clock.c
132
prev_time = READ_ONCE(trace_clock_struct.prev_time);
kernel/trace/trace_hwlat.c
206
u64 sample_width = READ_ONCE(hwlat_data.sample_width);
kernel/trace/trace_printk.c
455
struct trace_array *tr = READ_ONCE(printk_trace);
kernel/trace/trace_printk.c
611
struct trace_array *tr = READ_ONCE(printk_trace);
kernel/trace/trace_stack.c
159
int frame_size = READ_ONCE(tracer_frame);
kernel/trace/trace_syscalls.c
797
sbuf = READ_ONCE(syscall_buffer);
kernel/trace/trace_syscalls.c
886
trace_file = READ_ONCE(tr->enter_syscall_files[syscall_nr]);
kernel/trace/trace_syscalls.c
947
trace_file = READ_ONCE(tr->exit_syscall_files[syscall_nr]);
kernel/trace/trace_uprobe.c
1376
if (READ_ONCE(filter->nr_systemwide))
kernel/trace/tracing_map.c
534
val = READ_ONCE(entry->val);
kernel/tsacct.c
144
tsk->acct_vm_mem1 += delta * READ_ONCE(tsk->mm->total_vm) >> 10;
kernel/ucount.c
223
max = READ_ONCE(tns->ucount_max[type]);
kernel/unwind/deferred.c
255
bit = READ_ONCE(work->bit);
kernel/user_namespace.c
1217
unsigned long userns_flags = READ_ONCE(ns->flags);
kernel/watchdog.c
1037
watchdog_thresh = READ_ONCE(watchdog_thresh_next);
kernel/watchdog.c
1103
if (!err && old != READ_ONCE(*param))
kernel/watchdog.c
1155
watchdog_thresh_next = READ_ONCE(watchdog_thresh);
kernel/watchdog.c
1160
if (!err && write && old != READ_ONCE(watchdog_thresh_next))
kernel/watchdog.c
822
period_ts = READ_ONCE(*this_cpu_ptr(&watchdog_report_ts));
kernel/watchdog.c
993
watchdog_thresh = READ_ONCE(watchdog_thresh_next);
kernel/workqueue.c
1408
if (!READ_ONCE(worker->sleeping))
kernel/workqueue.c
1454
if (READ_ONCE(worker->sleeping))
kernel/workqueue.c
1510
if ((worker->flags & WORKER_NOT_RUNNING) || READ_ONCE(worker->sleeping) ||
kernel/workqueue.c
1595
int min_active = READ_ONCE(wq->min_active);
kernel/workqueue.c
1596
int max_active = READ_ONCE(wq->max_active);
kernel/workqueue.c
1707
int max = READ_ONCE(nna->max);
kernel/workqueue.c
1737
obtained = pwq->nr_active < READ_ONCE(wq->max_active);
kernel/workqueue.c
1982
if (atomic_dec_return(&nna->nr) >= READ_ONCE(nna->max))
kernel/workqueue.c
4092
if (READ_ONCE(wq->first_flusher) != &this_flusher)
kernel/workqueue.c
7665
unsigned long thresh = READ_ONCE(wq_watchdog_thresh) * HZ;
kernel/workqueue.c
7692
touched = READ_ONCE(per_cpu(wq_watchdog_touched_cpu, pool->cpu));
kernel/workqueue.c
7694
touched = READ_ONCE(wq_watchdog_touched);
kernel/workqueue.c
7695
pool_ts = READ_ONCE(pool->last_progress_ts);
kernel/workqueue.c
7753
unsigned long thresh = READ_ONCE(wq_watchdog_thresh) * HZ;
kernel/workqueue.c
7754
unsigned long touch_ts = READ_ONCE(wq_watchdog_touched);
lib/assoc_array.c
138
struct assoc_array_ptr *root = READ_ONCE(array->root); /* Address dependency. */
lib/assoc_array.c
185
cursor = READ_ONCE(array->root); /* Address dependency. */
lib/assoc_array.c
209
ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
lib/assoc_array.c
282
cursor = READ_ONCE(shortcut->next_node); /* Address dependency. */
lib/assoc_array.c
324
ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
lib/assoc_array.c
37
cursor = READ_ONCE(shortcut->next_node); /* Address dependency. */
lib/assoc_array.c
52
ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
lib/assoc_array.c
81
ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
lib/assoc_array.c
90
parent = READ_ONCE(node->back_pointer); /* Address dependency. */
lib/assoc_array.c
98
parent = READ_ONCE(shortcut->back_pointer); /* Address dependency. */
lib/buildid.c
172
name_sz = READ_ONCE(nhdr->n_namesz);
lib/buildid.c
173
desc_sz = READ_ONCE(nhdr->n_descsz);
lib/buildid.c
217
phnum = READ_ONCE(ehdr->e_phnum);
lib/buildid.c
218
phoff = READ_ONCE(ehdr->e_phoff);
lib/buildid.c
234
!parse_build_id(r, build_id, size, READ_ONCE(phdr->p_offset),
lib/buildid.c
235
READ_ONCE(phdr->p_filesz)))
lib/buildid.c
254
phnum = READ_ONCE(ehdr->e_phnum);
lib/buildid.c
255
phoff = READ_ONCE(ehdr->e_phoff);
lib/buildid.c
271
!parse_build_id(r, build_id, size, READ_ONCE(phdr->p_offset),
lib/buildid.c
272
READ_ONCE(phdr->p_filesz)))
lib/closure.c
129
p = READ_ONCE(s->task);
lib/cmpxchg-emu.c
33
ret = READ_ONCE(*p32);
lib/debugobjects.c
1137
if (!READ_ONCE(obj_freeing) && pool_count(&pool_to_free)) {
lib/debugobjects.c
1165
pool_used = READ_ONCE(pool_global.stats.cur_used);
lib/debugobjects.c
139
return READ_ONCE(pool->cnt);
lib/debugobjects.c
455
if (!time_after_eq(now, READ_ONCE(avg_period)))
lib/debugobjects.c
462
cur = READ_ONCE(pool_global.stats.cur_used) * ODEBUG_FREE_WORK_MAX;
lib/debugobjects.c
499
cur_used = READ_ONCE(avg_usage);
lib/debugobjects.c
550
if (!READ_ONCE(obj_freeing) && pool_count(&pool_to_free)) {
lib/dim/net_dim.c
177
dim->mode = READ_ONCE(irq_moder->dim_tx_mode);
lib/dim/net_dim.c
182
dim->mode = READ_ONCE(irq_moder->dim_rx_mode);
lib/dynamic_queue_limits.c
182
dql->prev_last_obj_cnt = READ_ONCE(dql->last_obj_cnt);
lib/dynamic_queue_limits.c
35
hist_head = READ_ONCE(dql->history_head);
lib/dynamic_queue_limits.c
68
if (hist_head != READ_ONCE(dql->history_head))
lib/dynamic_queue_limits.c
90
num_queued = READ_ONCE(dql->num_queued);
lib/dynamic_queue_limits.c
96
stall_thrs = READ_ONCE(dql->stall_thrs);
lib/errseq.c
126
errseq_t old = READ_ONCE(*eseq);
lib/errseq.c
148
errseq_t cur = READ_ONCE(*eseq);
lib/errseq.c
186
old = READ_ONCE(*eseq);
lib/errseq.c
73
old = READ_ONCE(*eseq);
lib/fault-inject.c
129
unsigned int fail_nth = READ_ONCE(current->fail_nth);
lib/genalloc.c
48
unsigned long val = READ_ONCE(*addr);
lib/genalloc.c
62
unsigned long val = READ_ONCE(*addr);
lib/generic-radix-tree.c
152
r = READ_ONCE(radix->root);
lib/generic-radix-tree.c
26
struct genradix_root *v = READ_ONCE(radix->root);
lib/generic-radix-tree.c
66
n = READ_ONCE(*p);
lib/generic-radix-tree.c
98
r = READ_ONCE(radix->root);
lib/llist.c
39
next = READ_ONCE(entry->next);
lib/llist.c
67
next = READ_ONCE(entry->next);
lib/lockref.c
15
old.lock_count = READ_ONCE(lockref->lock_count); \
lib/nmi_backtrace.c
105
if (!READ_ONCE(backtrace_idle) && regs && cpu_in_idle(instruction_pointer(regs))) {
lib/ratelimit.c
33
int interval = READ_ONCE(rs->interval);
lib/ratelimit.c
34
int burst = READ_ONCE(rs->burst);
lib/ratelimit.c
45
if (!(READ_ONCE(rs->flags) & RATELIMIT_INITIALIZED) || (!interval && !burst) ||
lib/ratelimit.c
61
if (READ_ONCE(rs->flags) & RATELIMIT_INITIALIZED &&
lib/sbitmap.c
289
depth = READ_ONCE(sb->depth);
lib/sbitmap.c
333
depth = READ_ONCE(sb->depth);
lib/sbitmap.c
410
unsigned long word = READ_ONCE(sb->map[i].word);
lib/sbitmap.c
411
unsigned long cleared = READ_ONCE(sb->map[i].cleared);
lib/sbitmap.c
526
depth = READ_ONCE(sb->depth);
lib/sbitmap.c
538
val = READ_ONCE(map->word);
lib/sbitmap.c
617
unsigned int wake_batch = READ_ONCE(sbq->wake_batch);
lib/sbitmap.c
85
return (READ_ONCE(map->word) & word_mask) != word_mask;
lib/stackdepot.c
496
const int pools_num_cached = READ_ONCE(pools_num);
lib/stackdepot.c
684
if (unlikely(can_alloc && !READ_ONCE(new_pool))) {
lib/sys_info.c
82
si_bits = READ_ONCE(*si_bits_global);
lib/syscall.c
77
state = READ_ONCE(target->__state);
lib/test_bitmap.c
1338
bitmap_write(NULL, 0, 0, READ_ONCE(zero_bits));
lib/test_bitmap.c
1340
bitmap_read(NULL, 0, READ_ONCE(zero_bits));
lib/test_bitmap.c
1346
bitmap_write(NULL, 0, 0, READ_ONCE(bits_per_long) + 1);
lib/test_bitmap.c
1348
bitmap_read(NULL, 0, READ_ONCE(bits_per_long) + 1);
lib/tests/test_ratelimit.c
89
while (!READ_ONCE(doneflag)) {
lib/vdso/getrandom.c
110
if (unlikely(!READ_ONCE(rng_info->is_ready)))
lib/vdso/getrandom.c
128
in_use = READ_ONCE(state->in_use);
lib/vdso/getrandom.c
139
current_generation = READ_ONCE(rng_info->generation);
lib/vdso/getrandom.c
214
if (unlikely(READ_ONCE(state->generation) != READ_ONCE(rng_info->generation))) {
lib/vdso/gettimeofday.c
172
while (unlikely((seq = READ_ONCE(vc->seq)) & 1)) {
lib/vdso/gettimeofday.c
230
while ((seq = READ_ONCE(vc->seq)) & 1) {
lib/vdso/gettimeofday.c
263
while ((seq = READ_ONCE(vc->seq)) & 1) {
lib/vdso/gettimeofday.c
409
t = READ_ONCE(vc[CS_HRES_COARSE].basetime[CLOCK_REALTIME].sec);
lib/vdso/gettimeofday.c
448
ns = READ_ONCE(vd->hrtimer_res);
lib/vsprintf.c
791
if (!READ_ONCE(filled_random_ptr_key))
lib/vsprintf.c
932
p = READ_ONCE(d->d_parent);
lib/vsprintf.c
933
array[i] = READ_ONCE(d->d_name.name);
mm/compaction.c
2790
*capture = READ_ONCE(capc.page);
mm/filemap.c
3341
mmap_miss = READ_ONCE(ra->mmap_miss);
mm/filemap.c
3434
mmap_miss = READ_ONCE(ra->mmap_miss);
mm/filemap.c
3946
mmap_miss_saved = READ_ONCE(file->f_ra.mmap_miss);
mm/filemap.c
736
errseq_t old = READ_ONCE(file->f_wb_err);
mm/gup.c
2784
mapping = READ_ONCE(folio->mapping);
mm/huge_memory.c
259
return READ_ONCE(huge_zero_folio);
mm/huge_memory.c
267
return READ_ONCE(huge_zero_folio);
mm/huge_memory.c
3240
entry = mk_pte(page, READ_ONCE(vma->vm_page_prot));
mm/huge_memory.c
4377
return READ_ONCE(ds_queue->split_queue_len);
mm/huge_memory.c
4930
pmde = folio_mk_pmd(folio, READ_ONCE(vma->vm_page_prot));
mm/hugetlb_cgroup.c
470
usage += READ_ONCE(h_cg->nodeinfo[nid]->usage[idx]);
mm/hugetlb_cgroup.c
476
READ_ONCE(h_cg->nodeinfo[nid]->usage[idx]) *
mm/hugetlb_cgroup.c
496
usage += READ_ONCE(hugetlb_cgroup_from_css(css)
mm/hugetlb_vmemmap.c
555
if (!READ_ONCE(vmemmap_optimize_enabled))
mm/hugetlb_vmemmap.c
755
if (!READ_ONCE(vmemmap_optimize_enabled))
mm/hugetlb_vmemmap.c
797
if (!READ_ONCE(vmemmap_optimize_enabled))
mm/hugetlb_vmemmap.c
842
if (!READ_ONCE(vmemmap_optimize_enabled))
mm/internal.h
702
#define buddy_order_unsafe(page) READ_ONCE(page_private(page))
mm/kasan/generic.c
210
shadow_byte = READ_ONCE(*(s8 *)kasan_mem_to_shadow(addr));
mm/kasan/kasan_test_c.c
1122
READ_ONCE(*p);
mm/kasan/kasan_test_c.c
1132
KUNIT_EXPECT_KASAN_FAIL_READ(test, READ_ONCE(*p));
mm/kasan/kasan_test_c.c
126
KUNIT_EXPECT_FALSE(test, READ_ONCE(test_status.report_found)); \
mm/kasan/kasan_test_c.c
132
if (READ_ONCE(test_status.report_found) != fail) { \
mm/kasan/kasan_test_c.c
142
if (READ_ONCE(test_status.report_found) && \
mm/kasan/kasan_test_c.c
143
!READ_ONCE(test_status.async_fault)) \
mm/kasan/kasan_test_c.c
772
KUNIT_EXPECT_KASAN_FAIL_READ(test, READ_ONCE(*i_unsafe));
mm/kasan/kasan_test_c.c
93
KUNIT_EXPECT_FALSE(test, READ_ONCE(test_status.report_found));
mm/kasan/quarantine.c
222
READ_ONCE(quarantine_batch_size)) {
mm/kasan/quarantine.c
246
if (likely(READ_ONCE(quarantine_size) <=
mm/kasan/quarantine.c
247
READ_ONCE(quarantine_max_size)))
mm/kasan/quarantine.c
336
if (READ_ONCE(q->offline))
mm/kasan/report.c
175
return READ_ONCE(kasan_kunit_executing);
mm/kasan/sw_tags.c
133
shadow_byte = READ_ONCE(*(u8 *)kasan_mem_to_shadow(untagged_addr));
mm/kasan/tags.c
120
old_ptr = READ_ONCE(entry->ptr);
mm/kfence/core.c
1098
if (READ_ONCE(meta->cache) != s || !kfence_obj_allocated(meta))
mm/kfence/core.c
1128
if (READ_ONCE(meta->cache) != s || READ_ONCE(meta->state) != KFENCE_OBJECT_FREED)
mm/kfence/core.c
1190
if (!READ_ONCE(kfence_enabled))
mm/kfence/core.c
1274
if (!READ_ONCE(kfence_enabled)) /* If disabled at runtime ... */
mm/kfence/core.c
282
enum kfence_object_state state = READ_ONCE(meta->state);
mm/kfence/core.c
302
if (READ_ONCE(meta->state) == KFENCE_OBJECT_RCU_FREEING)
mm/kfence/core.c
749
seq_printf(seq, "enabled: %i\n", READ_ONCE(kfence_enabled));
mm/kfence/core.c
75
if (!num && READ_ONCE(kfence_enabled)) {
mm/kfence/core.c
806
if (!READ_ONCE(kfence_enabled))
mm/kfence/core.c
82
if (num && !READ_ONCE(kfence_enabled) && system_state != SYSTEM_BOOTING)
mm/kfence/core.c
89
if (!READ_ONCE(kfence_enabled))
mm/kfence/core.c
894
if (!READ_ONCE(kfence_enabled))
mm/kfence/core.c
904
!READ_ONCE(kfence_enabled));
mm/kfence/kfence_test.c
342
READ_ONCE(*expect.addr);
mm/kfence/kfence_test.c
348
READ_ONCE(*expect.addr);
mm/kfence/kfence_test.c
383
READ_ONCE(*expect.addr);
mm/kfence/kfence_test.c
488
READ_ONCE(*(buf - 1));
mm/kfence/kfence_test.c
495
READ_ONCE(*(buf + size));
mm/kfence/kfence_test.c
500
READ_ONCE(*expect.addr);
mm/kfence/kfence_test.c
521
WRITE_ONCE(*expect.addr, READ_ONCE(*expect.addr) + 1);
mm/kfence/kfence_test.c
668
READ_ONCE(__kfence_pool[10]);
mm/kfence/kfence_test.c
744
READ_ONCE(*expect.addr); /* Ensure krealloc() actually freed earlier KFENCE object. */
mm/kfence/kfence_test.c
79
return READ_ONCE(observed.nlines) == ARRAY_SIZE(observed.lines);
mm/khugepaged.c
1712
if (READ_ONCE(vma->anon_vma))
mm/khugepaged.c
2547
unsigned int pages = READ_ONCE(khugepaged_pages_to_scan);
mm/kmemleak.c
1898
signed long timeout = READ_ONCE(jiffies_scan_wait);
mm/kmsan/kmsan_test.c
67
return READ_ONCE(observed.available);
mm/ksm.c
1000
if (READ_ONCE(folio->mapping) != expected_mapping) {
mm/ksm.c
1016
if (READ_ONCE(stable_node->kpfn) != kpfn)
mm/ksm.c
2262
get_kpfn_nid(READ_ONCE(stable_node->kpfn)) !=
mm/ksm.c
2816
sleep_ms = READ_ONCE(ksm_thread_sleep_millisecs);
mm/ksm.c
2818
sleep_ms != READ_ONCE(ksm_thread_sleep_millisecs),
mm/ksm.c
3658
if (READ_ONCE(ksm_max_page_sharing) == knob)
mm/ksm.c
956
kpfn = READ_ONCE(stable_node->kpfn); /* Address dependency. */
mm/ksm.c
958
if (READ_ONCE(folio->mapping) != expected_mapping)
mm/ksm.c
986
if (READ_ONCE(folio->mapping) != expected_mapping) {
mm/list_lru.c
260
count = l ? READ_ONCE(l->nr_items) : 0;
mm/list_lru.c
69
if (unlikely(READ_ONCE(l->nr_items) == LONG_MIN)) {
mm/mapping_dirty_helpers.c
221
vm_flags_t vm_flags = READ_ONCE(walk->vma->vm_flags);
mm/memcontrol-v1.c
1433
if (READ_ONCE(memcg->oom_kill_disable)) {
mm/memcontrol-v1.c
1481
limits_invariant = memsw ? max >= READ_ONCE(memcg->memory.max) :
mm/memcontrol-v1.c
1609
return (u64)READ_ONCE(memcg->soft_limit) * PAGE_SIZE;
mm/memcontrol-v1.c
170
unsigned long soft_limit = READ_ONCE(memcg->soft_limit);
mm/memcontrol-v1.c
1919
memory = min(memory, READ_ONCE(mi->memory.max));
mm/memcontrol-v1.c
1920
memsw = min(memsw, READ_ONCE(mi->memsw.max));
mm/memcontrol-v1.c
1994
seq_printf(sf, "oom_kill_disable %d\n", READ_ONCE(memcg->oom_kill_disable));
mm/memcontrol.c
1056
pos = READ_ONCE(iter->position);
mm/memcontrol.c
1332
limit = READ_ONCE(memcg->memory.max);
mm/memcontrol.c
1338
limit = READ_ONCE(memcg->memsw.max);
mm/memcontrol.c
1595
K((u64)READ_ONCE(memcg->memory.max)), memory_failcnt);
mm/memcontrol.c
1599
K((u64)READ_ONCE(memcg->swap.max)),
mm/memcontrol.c
1625
unsigned long max = READ_ONCE(memcg->memory.max);
mm/memcontrol.c
1630
unsigned long swap = READ_ONCE(memcg->memsw.max) - max;
mm/memcontrol.c
1636
max += min(READ_ONCE(memcg->swap.max),
mm/memcontrol.c
1767
if (READ_ONCE(memcg->oom_group))
mm/memcontrol.c
1855
if (memcg != READ_ONCE(stock->cached[i]))
mm/memcontrol.c
1858
stock_pages = READ_ONCE(stock->nr_pages[i]);
mm/memcontrol.c
1883
struct mem_cgroup *old = READ_ONCE(stock->cached[i]);
mm/memcontrol.c
1889
stock_pages = READ_ONCE(stock->nr_pages[i]);
mm/memcontrol.c
1969
cached = READ_ONCE(stock->cached[i]);
mm/memcontrol.c
1972
if (memcg == READ_ONCE(stock->cached[i])) {
mm/memcontrol.c
1973
stock_pages = READ_ONCE(stock->nr_pages[i]) + nr_pages;
mm/memcontrol.c
2005
memcg = READ_ONCE(stock->cached[i]);
mm/memcontrol.c
2009
if (READ_ONCE(stock->nr_pages[i]) &&
mm/memcontrol.c
2098
READ_ONCE(memcg->memory.high))
mm/memcontrol.c
2200
READ_ONCE(memcg->memory.high));
mm/memcontrol.c
2214
READ_ONCE(memcg->swap.high));
mm/memcontrol.c
2507
READ_ONCE(memcg->memory.high);
mm/memcontrol.c
2509
READ_ONCE(memcg->swap.high);
mm/memcontrol.c
2746
objcg = READ_ONCE(current->objcg);
mm/memcontrol.c
2988
if (objcg == READ_ONCE(stock->cached_objcg) && stock->nr_bytes >= nr_bytes) {
mm/memcontrol.c
3003
struct obj_cgroup *old = READ_ONCE(stock->cached_objcg);
mm/memcontrol.c
3065
struct obj_cgroup *objcg = READ_ONCE(stock->cached_objcg);
mm/memcontrol.c
3097
if (READ_ONCE(stock->cached_objcg) != objcg) { /* reset if necessary */
mm/memcontrol.c
3430
unsigned long ceiling = min(READ_ONCE(memcg->memory.max),
mm/memcontrol.c
3431
READ_ONCE(memcg->memory.high));
mm/memcontrol.c
3849
WRITE_ONCE(memcg->oom_kill_disable, READ_ONCE(parent->oom_kill_disable));
mm/memcontrol.c
4039
v = READ_ONCE(ac->cstat[i]);
mm/memcontrol.c
407
x = READ_ONCE(pn->lruvec_stats->state[i]);
mm/memcontrol.c
4210
if (task->mm && READ_ONCE(task->mm->owner) == task)
mm/memcontrol.c
4258
u64 fd_peak = READ_ONCE(ofp->value), peak;
mm/memcontrol.c
4264
peak = max(fd_peak, READ_ONCE(pc->local_watermark));
mm/memcontrol.c
430
x = READ_ONCE(pn->lruvec_stats->state_local[i]);
mm/memcontrol.c
4341
READ_ONCE(mem_cgroup_from_seq(m)->memory.min));
mm/memcontrol.c
4364
READ_ONCE(mem_cgroup_from_seq(m)->memory.low));
mm/memcontrol.c
4387
READ_ONCE(mem_cgroup_from_seq(m)->memory.high));
mm/memcontrol.c
4439
READ_ONCE(mem_cgroup_from_seq(m)->memory.max));
mm/memcontrol.c
4584
seq_printf(m, "%d\n", READ_ONCE(memcg->oom_group));
mm/memcontrol.c
5282
READ_ONCE(memcg->swap.max) -
mm/memcontrol.c
5305
if (usage * 2 >= READ_ONCE(memcg->swap.high) ||
mm/memcontrol.c
5306
usage * 2 >= READ_ONCE(memcg->swap.max))
mm/memcontrol.c
5353
READ_ONCE(mem_cgroup_from_seq(m)->swap.high));
mm/memcontrol.c
5376
READ_ONCE(mem_cgroup_from_seq(m)->swap.max));
mm/memcontrol.c
5469
unsigned long max = READ_ONCE(memcg->zswap_max);
mm/memcontrol.c
5549
if (!READ_ONCE(memcg->zswap_writeback))
mm/memcontrol.c
5567
READ_ONCE(mem_cgroup_from_seq(m)->zswap_max));
mm/memcontrol.c
5591
seq_printf(m, "%d\n", READ_ONCE(memcg->zswap_writeback));
mm/memcontrol.c
638
if (time_after64(jiffies_64, READ_ONCE(flush_last_time) + 2*FLUSH_TIME))
mm/memcontrol.c
660
x = READ_ONCE(memcg->vmstats->state[i]);
mm/memcontrol.c
730
x = READ_ONCE(memcg->vmstats->state_local[i]);
mm/memcontrol.c
870
return READ_ONCE(memcg->vmstats->events[i]);
mm/memcontrol.c
889
return READ_ONCE(memcg->vmstats->events_local[i]);
mm/memory.c
5735
pgoff_t nr_pages = READ_ONCE(fault_around_pages);
mm/memory.c
889
pte = pte_mkold(mk_pte(page, READ_ONCE(vma->vm_page_prot)));
mm/memory_hotplug.c
819
const int ratio = READ_ONCE(auto_movable_ratio);
mm/mempool.c
673
if (unlikely(READ_ONCE(pool->curr_nr) < pool->min_nr)) {
mm/mempool.c
681
READ_ONCE(pool->curr_nr) == 0)) {
mm/memtest.c
59
if (READ_ONCE(*p) == pattern)
mm/migrate.c
380
pte = mk_pte(new, READ_ONCE(vma->vm_page_prot));
mm/migrate.c
438
if (READ_ONCE(vma->vm_flags) & VM_LOCKED)
mm/mmap.c
1362
WRITE_ONCE(mm->total_vm, READ_ONCE(mm->total_vm)+npages);
mm/mmap_lock.c
226
if (READ_ONCE(vma->vm_lock_seq) == READ_ONCE(mm->mm_lock_seq.sequence)) {
mm/mmu_notifier.c
236
seq = READ_ONCE(interval_sub->invalidate_seq);
mm/mmu_notifier.c
251
READ_ONCE(subscriptions->invalidate_seq) != seq);
mm/mmzone.c
102
old_flags = READ_ONCE(folio->flags.f);
mm/mprotect.c
700
vm_flags_t oldflags = READ_ONCE(vma->vm_flags);
mm/oom_kill.c
499
const struct mm_struct *t_mm = READ_ONCE(t->mm);
mm/page-writeback.c
1066
unsigned long write_bw = READ_ONCE(wb->avg_write_bandwidth);
mm/page-writeback.c
1537
unsigned long elapsed = now - READ_ONCE(wb->bw_time_stamp);
mm/page-writeback.c
1569
unsigned long bw = READ_ONCE(wb->avg_write_bandwidth);
mm/page-writeback.c
159
unsigned long this_bw = READ_ONCE(wb->avg_write_bandwidth);
mm/page-writeback.c
1591
long hi = ilog2(READ_ONCE(wb->avg_write_bandwidth));
mm/page-writeback.c
1592
long lo = ilog2(READ_ONCE(wb->dirty_ratelimit));
mm/page-writeback.c
1909
if (time_is_before_jiffies(READ_ONCE(wb->bw_time_stamp) +
mm/page-writeback.c
1914
dirty_ratelimit = READ_ONCE(wb->dirty_ratelimit);
mm/page-writeback.c
2596
if (time_is_before_jiffies(READ_ONCE(wb->bw_time_stamp) +
mm/page_alloc.c
2594
high_min = READ_ONCE(pcp->high_min);
mm/page_alloc.c
2595
batch = READ_ONCE(pcp->batch);
mm/page_alloc.c
2633
batch = READ_ONCE(pcp->batch);
mm/page_alloc.c
2809
high_min = READ_ONCE(pcp->high_min);
mm/page_alloc.c
2810
high_max = READ_ONCE(pcp->high_max);
mm/page_alloc.c
2881
batch = READ_ONCE(pcp->batch);
mm/page_alloc.c
3289
base_batch = READ_ONCE(pcp->batch);
mm/page_alloc.c
3290
high_min = READ_ONCE(pcp->high_min);
mm/page_alloc.c
3291
high_max = READ_ONCE(pcp->high_max);
mm/page_alloc.c
3585
unusable_free += READ_ONCE(z->nr_free_highatomic);
mm/page_alloc.c
434
word = READ_ONCE(*bitmap_word);
mm/page_alloc.c
505
word = READ_ONCE(*bitmap_word);
mm/page_counter.c
101
if (new > READ_ONCE(c->local_watermark)) {
mm/page_counter.c
103
if (new > READ_ONCE(c->watermark))
mm/page_counter.c
159
if (new > READ_ONCE(c->local_watermark)) {
mm/page_counter.c
161
if (new > READ_ONCE(c->watermark))
mm/page_counter.c
30
protected = min(usage, READ_ONCE(c->min));
mm/page_counter.c
39
protected = min(usage, READ_ONCE(c->low));
mm/page_counter.c
446
counter->emin = READ_ONCE(counter->min);
mm/page_counter.c
447
counter->elow = READ_ONCE(counter->low);
mm/page_counter.c
454
READ_ONCE(counter->min),
mm/page_counter.c
455
READ_ONCE(parent->emin),
mm/page_counter.c
460
READ_ONCE(counter->low),
mm/page_counter.c
461
READ_ONCE(parent->elow),
mm/page_ext.c
256
struct page_ext *page_ext = READ_ONCE(section->page_ext);
mm/page_ext.c
350
base = READ_ONCE(ms->page_ext);
mm/page_owner.c
521
memcg_data = READ_ONCE(page->memcg_data);
mm/page_owner.c
638
handle = READ_ONCE(page_owner->handle);
mm/page_owner.c
644
handle = READ_ONCE(page_owner->free_handle);
mm/page_owner.c
739
handle = READ_ONCE(page_owner->handle);
mm/page_owner.c
960
*val = READ_ONCE(page_owner_pages_threshold);
mm/percpu.c
3375
return data_race(READ_ONCE(pcpu_nr_populated)) * pcpu_nr_units;
mm/rmap.c
595
anon_mapping = (unsigned long)READ_ONCE(folio->mapping);
mm/rmap.c
643
anon_mapping = (unsigned long)READ_ONCE(folio->mapping);
mm/rmap.c
650
root_anon_vma = READ_ONCE(anon_vma->root);
mm/rodata_test.c
24
if (unlikely(READ_ONCE(rodata_test_data) != TEST_VALUE)) {
mm/rodata_test.c
37
if (unlikely(READ_ONCE(rodata_test_data) != TEST_VALUE)) {
mm/shmem.c
1038
swapped = READ_ONCE(info->swapped);
mm/shmem.c
1837
unsigned long mask = READ_ONCE(huge_shmem_orders_always);
mm/shmem.c
1838
unsigned long within_size_orders = READ_ONCE(huge_shmem_orders_within_size);
mm/shmem.c
1863
return READ_ONCE(huge_shmem_orders_inherit);
mm/shmem.c
1869
mask |= READ_ONCE(huge_shmem_orders_madvise);
mm/shmem.c
1872
mask |= READ_ONCE(huge_shmem_orders_inherit);
mm/shmem.c
2017
READ_ONCE(mapping->nrpages);
mm/shmem.c
2840
hpage_orders = READ_ONCE(huge_shmem_orders_always);
mm/shmem.c
2841
hpage_orders |= READ_ONCE(huge_shmem_orders_within_size);
mm/shmem.c
2842
hpage_orders |= READ_ONCE(huge_shmem_orders_madvise);
mm/shmem.c
2844
hpage_orders |= READ_ONCE(huge_shmem_orders_inherit);
mm/shmem.c
447
READ_ONCE(inode->i_mapping->nrpages);
mm/shmem.c
839
if (!READ_ONCE(sbinfo->shrinklist_len))
mm/shmem.c
849
return READ_ONCE(sbinfo->shrinklist_len);
mm/slab.h
136
head = READ_ONCE(page->compound_head);
mm/slab.h
535
unsigned long obj_exts = READ_ONCE(slab->obj_exts);
mm/slab_common.c
1622
return !!READ_ONCE(krcp->head);
mm/slab_common.c
1805
for (i = READ_ONCE(krcp->nr_bkv_objs); i < nr_pages; i++) {
mm/slab_common.c
2136
count += READ_ONCE(krcp->nr_bkv_objs);
mm/slub.c
1089
handle = READ_ONCE(t->handle);
mm/slub.c
2190
old_exts = READ_ONCE(slab->obj_exts);
mm/slub.c
3768
flc.counters = data_race(READ_ONCE(slab->counters));
mm/slub.c
7945
handle = READ_ONCE(trackp->handle);
mm/slub.c
7953
handle = READ_ONCE(trackp->handle);
mm/slub.c
8692
handle = READ_ONCE(track->handle);
mm/slub.c
9655
handle = READ_ONCE(l->handle);
mm/swap.c
391
unsigned long new_flags, old_flags = READ_ONCE(folio->flags.f);
mm/swap.c
426
return gen == lru_gen_from_seq(READ_ONCE(lrugen->min_seq[type]));
mm/swap.h
73
si = READ_ONCE(swap_info[type]); /* rcu_dereference() */
mm/swap_state.c
431
return READ_ONCE(enable_vma_readahead) && !atomic_read(&nr_rotate_swap);
mm/swap_state.c
687
max_pages = 1 << READ_ONCE(page_cluster);
mm/swap_state.c
692
pages = __swapin_nr_pages(READ_ONCE(prev_offset), offset, hits,
mm/swap_state.c
782
max_win = 1 << min(READ_ONCE(page_cluster), SWAP_RA_ORDER_CEILING);
mm/swapfile.c
1004
if (!READ_ONCE(map[offset]) &&
mm/swapfile.c
1188
if (!(READ_ONCE(si->flags) & SWP_WRITEOK))
mm/swapfile.c
138
return READ_ONCE(swap_info[type]); /* rcu_dereference() */
mm/swapfile.c
2235
swp_count = READ_ONCE(si->swap_map[offset]);
mm/swapfile.c
2382
count = READ_ONCE(si->swap_map[i]);
mm/swapfile.c
777
if (READ_ONCE(map[offset]))
mm/vma.c
1330
WRITE_ONCE(mm->total_vm, READ_ONCE(mm->total_vm) - vms->nr_pages);
mm/vma.c
1987
struct anon_vma *anon_vma = READ_ONCE(old->anon_vma);
mm/vmalloc.c
2828
unsigned long free = READ_ONCE(vb->free);
mm/vmalloc.c
2829
unsigned long dirty = READ_ONCE(vb->dirty);
mm/vmalloc.c
2875
if (READ_ONCE(vb->free) < (1UL << order))
mm/vmalloc.c
5408
count += READ_ONCE(vn->pool[i].len);
mm/vmscan.c
268
return READ_ONCE(vm_swappiness);
mm/vmscan.c
2684
unsigned long max_seq = READ_ONCE((lruvec)->lrugen.max_seq)
mm/vmscan.c
2688
READ_ONCE((lruvec)->lrugen.min_seq[LRU_GEN_ANON]), \
mm/vmscan.c
2689
READ_ONCE((lruvec)->lrugen.min_seq[LRU_GEN_FILE]), \
mm/vmscan.c
2814
filter = READ_ONCE(mm_state->filters[gen]);
mm/vmscan.c
2830
filter = READ_ONCE(mm_state->filters[gen]);
mm/vmscan.c
3232
unsigned long new_flags, old_flags = READ_ONCE(folio->flags.f);
mm/vmscan.c
3260
unsigned long new_flags, old_flags = READ_ONCE(folio->flags.f);
mm/vmscan.c
3963
if (seq < READ_ONCE(lrugen->max_seq))
mm/vmscan.c
4030
VM_WARN_ON_ONCE(seq > READ_ONCE(lrugen->max_seq));
mm/vmscan.c
4036
if (seq <= READ_ONCE(mm_state->seq))
mm/vmscan.c
4122
total += max(READ_ONCE(lrugen->nr_pages[gen][type][zone]), 0L);
mm/vmscan.c
4146
birth = READ_ONCE(lruvec->lrugen.timestamps[gen]);
mm/vmscan.c
4157
unsigned long min_ttl = READ_ONCE(lru_gen_min_ttl);
mm/vmscan.c
4425
if (READ_ONCE(lruvec->lrugen.seg) != MEMCG_LRU_HEAD)
mm/vmscan.c
4797
size += max(READ_ONCE(lrugen->nr_pages[gen][type][zone]), 0L);
mm/vmscan.c
4920
if (READ_ONCE(lruvec->lrugen.seg) != MEMCG_LRU_TAIL)
mm/vmscan.c
4943
return READ_ONCE(lruvec->lrugen.seg) != MEMCG_LRU_TAIL ?
mm/vmscan.c
4958
gen = get_memcg_gen(READ_ONCE(pgdat->memcg_lru.seq));
mm/vmscan.c
4975
if (gen != READ_ONCE(lrugen->gen))
mm/vmscan.c
5224
return sysfs_emit(buf, "%u\n", jiffies_to_msecs(READ_ONCE(lru_gen_min_ttl)));
mm/vmscan.c
5372
n[0] = READ_ONCE(lrugen->avg_refaulted[type][tier]);
mm/vmscan.c
5373
n[1] = READ_ONCE(lrugen->avg_total[type][tier]);
mm/vmscan.c
5378
n[2] = READ_ONCE(lrugen->protected[hist][type][tier]);
mm/vmscan.c
5397
n = READ_ONCE(mm_state->stats[hist][i]);
mm/vmscan.c
5400
n = READ_ONCE(mm_state->stats[hist][i]);
mm/vmscan.c
5442
unsigned long birth = READ_ONCE(lruvec->lrugen.timestamps[gen]);
mm/vmscan.c
5451
size += max(READ_ONCE(lrugen->nr_pages[gen][type][zone]), 0L);
mm/vmscan.c
619
READ_ONCE(pgdat->nr_reclaim_start);
mm/vmscan.c
6468
if (READ_ONCE(pgdat->kswapd_highest_zoneidx) > ZONE_NORMAL)
mm/vmscan.c
7178
enum zone_type curr_idx = READ_ONCE(pgdat->kswapd_highest_zoneidx);
mm/vmscan.c
7228
if (READ_ONCE(pgdat->kswapd_order) < reclaim_order)
mm/vmscan.c
7308
alloc_order = reclaim_order = READ_ONCE(pgdat->kswapd_order);
mm/vmscan.c
7317
alloc_order = READ_ONCE(pgdat->kswapd_order);
mm/vmscan.c
7374
curr_idx = READ_ONCE(pgdat->kswapd_highest_zoneidx);
mm/vmscan.c
7379
if (READ_ONCE(pgdat->kswapd_order) < order)
mm/workingset.c
251
min_seq = READ_ONCE(lrugen->min_seq[type]);
mm/workingset.c
277
max_seq = READ_ONCE((*lruvec)->lrugen.max_seq);
mm/workingset.c
307
hist = lru_hist_from_seq(READ_ONCE(lrugen->min_seq[type]));
net/8021q/vlan_dev.c
691
rx_errors += READ_ONCE(p->rx_errors);
net/8021q/vlan_dev.c
692
tx_dropped += READ_ONCE(p->tx_dropped);
net/8021q/vlan_dev.c
745
return READ_ONCE(real_dev->ifindex);
net/9p/client.c
493
if (READ_ONCE(oldreq->status) == REQ_STATUS_SENT) {
net/9p/client.c
594
READ_ONCE(req->status) >= REQ_STATUS_RCVD);
net/9p/client.c
608
if (READ_ONCE(req->status) == REQ_STATUS_ERROR) {
net/9p/client.c
621
if (READ_ONCE(req->status) == REQ_STATUS_RCVD)
net/9p/client.c
693
if (READ_ONCE(req->status) == REQ_STATUS_ERROR) {
net/9p/client.c
706
if (READ_ONCE(req->status) == REQ_STATUS_RCVD)
net/9p/trans_fd.c
159
if (READ_ONCE(m->err)) {
net/9p/trans_fd.c
252
if (READ_ONCE(m->err) < 0)
net/9p/trans_fd.c
419
if (READ_ONCE(m->err) < 0) {
net/9p/trans_fd.c
591
if (READ_ONCE(m->err) < 0)
net/9p/trans_fd.c
642
err = READ_ONCE(m->err);
net/9p/trans_fd.c
917
err = READ_ONCE(csocket->ops)->connect(csocket,
net/9p/trans_fd.c
959
err = READ_ONCE(csocket->ops)->connect(csocket, (struct sockaddr_unsized *)&sun_server,
net/9p/trans_virtio.c
534
READ_ONCE(req->status) >= REQ_STATUS_RCVD);
net/9p/trans_virtio.c
536
if (READ_ONCE(req->status) == REQ_STATUS_RCVD &&
net/appletalk/aarp.c
654
skb->priority = READ_ONCE(skb->sk->sk_priority);
net/ax25/af_ax25.c
953
sk->sk_priority = READ_ONCE(osk->sk_priority);
net/batman-adv/hard-interface.c
714
hardif_mtu = READ_ONCE(hard_iface->net_dev->mtu);
net/batman-adv/hard-interface.c
715
required_mtu = READ_ONCE(mesh_iface->mtu) + max_header_len;
net/batman-adv/translation-table.c
3637
if (READ_ONCE(bat_priv->tt.local_changes) == 0) {
net/batman-adv/translation-table.c
443
changes = READ_ONCE(bat_priv->tt.local_changes);
net/batman-adv/translation-table.c
949
tt_diff_entries_num = READ_ONCE(bat_priv->tt.local_changes);
net/bluetooth/hci_conn.c
3104
if (conn->state == BT_CONNECT && READ_ONCE(hdev->req_status) == HCI_REQ_PEND) {
net/bluetooth/hci_core.c
4129
if (READ_ONCE(hdev->req_status) == HCI_REQ_PEND &&
net/bluetooth/hci_sync.c
177
READ_ONCE(hdev->req_status) != HCI_REQ_PEND,
net/bluetooth/hci_sync.c
183
switch (READ_ONCE(hdev->req_status)) {
net/bluetooth/hci_sync.c
2715
p[i].flags = READ_ONCE(params->flags);
net/bluetooth/hci_sync.c
2716
p[i].privacy_mode = READ_ONCE(params->privacy_mode);
net/bluetooth/hci_sync.c
28
if (READ_ONCE(hdev->req_status) != HCI_REQ_PEND)
net/bluetooth/hci_sync.c
668
if (READ_ONCE(hdev->req_status) == HCI_REQ_PEND) {
net/bluetooth/hci_sync.c
686
if (READ_ONCE(hdev->req_status) == HCI_REQ_PEND) {
net/bluetooth/iso.c
1980
if (sock_flag(sk, SOCK_LINGER) && READ_ONCE(sk->sk_lingertime) &&
net/bluetooth/iso.c
374
READ_ONCE(sk->sk_sndtimeo));
net/bluetooth/iso.c
384
READ_ONCE(sk->sk_sndtimeo));
net/bluetooth/iso.c
421
iso_sock_set_timer(sk, READ_ONCE(sk->sk_sndtimeo));
net/bluetooth/iso.c
478
READ_ONCE(sk->sk_sndtimeo));
net/bluetooth/iso.c
487
READ_ONCE(sk->sk_sndtimeo));
net/bluetooth/iso.c
520
iso_sock_set_timer(sk, READ_ONCE(sk->sk_sndtimeo));
net/bluetooth/l2cap_core.c
944
ident = ida_alloc_range(&conn->tx_ida, READ_ONCE(conn->tx_ident) + 1,
net/bluetooth/l2cap_sock.c
1689
skb->priority = READ_ONCE(sk->sk_priority);
net/bluetooth/l2cap_sock.c
1761
return READ_ONCE(sk->sk_sndtimeo);
net/bluetooth/l2cap_sock.c
258
READ_ONCE(sk->sk_sndtimeo));
net/bluetooth/sco.c
1353
if (sock_flag(sk, SOCK_LINGER) && READ_ONCE(sk->sk_lingertime) &&
net/bluetooth/sco.c
341
READ_ONCE(sk->sk_sndtimeo));
net/bluetooth/sco.c
381
sco_sock_set_timer(sk, READ_ONCE(sk->sk_sndtimeo));
net/bridge/br_arp_nd_proxy.c
198
if (!(READ_ONCE(n->nud_state) & NUD_VALID)) {
net/bridge/br_arp_nd_proxy.c
466
if (!(READ_ONCE(n->nud_state) & NUD_VALID)) {
net/bridge/br_device.c
403
dst = READ_ONCE(f->dst);
net/bridge/br_fdb.c
1009
if (now != READ_ONCE(fdb->updated)) {
net/bridge/br_fdb.c
1015
if (unlikely(source != READ_ONCE(fdb->dst) &&
net/bridge/br_fdb.c
1210
if (READ_ONCE(fdb->dst) != source) {
net/bridge/br_fdb.c
129
ci.ndm_used = jiffies_to_clock_t(now - READ_ONCE(fdb->used));
net/bridge/br_fdb.c
131
ci.ndm_updated = jiffies_to_clock_t(now - READ_ONCE(fdb->updated));
net/bridge/br_fdb.c
1402
if (!fdb || READ_ONCE(fdb->dst) != p)
net/bridge/br_fdb.c
1558
READ_ONCE(fdb->dst) != p)) {
net/bridge/br_fdb.c
1565
if (READ_ONCE(fdb->dst) != p) {
net/bridge/br_fdb.c
390
u32 max_learned = READ_ONCE(br->fdb_max_learned);
net/bridge/br_fdb.c
554
unsigned long this_timer = READ_ONCE(f->updated) + delay;
net/bridge/br_fdb.c
685
const struct net_bridge_port *dst = READ_ONCE(f->dst);
net/bridge/br_fdb.c
73
time_before_eq(READ_ONCE(fdb->updated) + hold_time(br), jiffies);
net/bridge/br_fdb.c
907
dst = READ_ONCE(fdb->dst);
net/bridge/br_fdb.c
93
const struct net_bridge_port *dst = READ_ONCE(fdb->dst);
net/bridge/br_fdb.c
958
delta = jiffies - READ_ONCE(f->updated);
net/bridge/br_forward.c
158
BR_INPUT_SKB_CB(skb)->backup_nhid = READ_ONCE(to->backup_nhid);
net/bridge/br_input.c
126
} else if (READ_ONCE(fdb_src->dst) != p ||
net/bridge/br_input.c
224
if (now != READ_ONCE(dst->used))
net/bridge/br_multicast.c
3152
int port_ifidx = READ_ONCE(querier->port_ifidx);
net/bridge/br_multicast.c
698
u32 max = READ_ONCE(pmctx->mdb_max_entries);
net/bridge/br_multicast.c
699
u32 n = READ_ONCE(pmctx->mdb_n_entries);
net/bridge/br_multicast.c
716
u32 n = READ_ONCE(pmctx->mdb_n_entries);
net/bridge/br_multicast.c
776
return READ_ONCE(pmctx->mdb_n_entries);
net/bridge/br_multicast.c
786
return READ_ONCE(pmctx->mdb_max_entries);
net/bridge/br_netfilter_hooks.c
289
if ((READ_ONCE(neigh->nud_state) & NUD_CONNECTED) &&
net/bridge/br_netfilter_hooks.c
290
READ_ONCE(neigh->hh.hh_len)) {
net/bridge/br_netfilter_hooks.c
315
ret = READ_ONCE(neigh->output)(neigh, skb);
net/bridge/br_netlink.c
458
u8 operstate = netif_running(dev) ? READ_ONCE(dev->operstate) :
net/bridge/br_private.h
1912
return READ_ONCE(v->state);
net/bridge/br_private.h
1923
return READ_ONCE(vg->pvid_state);
net/bridge/br_switchdev.c
133
const struct net_bridge_port *p = READ_ONCE(fdb->dst);
net/bridge/br_vlan_tunnel.c
196
tunnel_id = READ_ONCE(vlan->tinfo.tunnel_id);
net/can/af_can.c
224
if (unlikely(skb->len > READ_ONCE(skb->dev->mtu))) {
net/can/isotp.c
1328
if (READ_ONCE(dev->mtu) < so->ll.mtu) {
net/can/j1939/socket.c
1016
tsflags = READ_ONCE(sk->sk_tsflags);
net/can/j1939/socket.c
918
skcb->priority = j1939_prio(READ_ONCE(sk->sk_priority));
net/core/bpf_sk_storage.c
288
optmem_max = READ_ONCE(sock_net(sk)->core.sysctl_optmem_max);
net/core/datagram.c
103
if (READ_ONCE(queue->prev) != skb)
net/core/datagram.c
277
} while (READ_ONCE(queue->prev) != *last);
net/core/datagram.c
948
if (READ_ONCE(sk->sk_err) ||
net/core/datagram.c
953
shutdown = READ_ONCE(sk->sk_shutdown);
net/core/datagram.c
965
int state = READ_ONCE(sk->sk_state);
net/core/dev.c
10833
return current == READ_ONCE(cleanup_net_task);
net/core/dev.c
11651
READ_ONCE(netdev_unregister_timeout_secs) * HZ)) {
net/core/dev.c
11835
return READ_ONCE(dev->core_stats);
net/core/dev.c
11841
struct net_device_core_stats __percpu *p = READ_ONCE(dev->core_stats);
net/core/dev.c
11899
p = READ_ONCE(dev->core_stats);
net/core/dev.c
11906
storage->rx_dropped += READ_ONCE(core_stats->rx_dropped);
net/core/dev.c
11907
storage->tx_dropped += READ_ONCE(core_stats->tx_dropped);
net/core/dev.c
11908
storage->rx_nohandler += READ_ONCE(core_stats->rx_nohandler);
net/core/dev.c
11909
storage->rx_otherhost_dropped += READ_ONCE(core_stats->rx_otherhost_dropped);
net/core/dev.c
1307
if (((READ_ONCE(dev->flags) ^ if_flags) & mask) == 0) {
net/core/dev.c
2558
if (READ_ONCE(ptype->ignore_outgoing))
net/core/dev.c
3319
if (size < READ_ONCE(dev->gso_max_size))
net/core/dev.c
3321
if (size < READ_ONCE(dev->gso_ipv4_max_size))
net/core/dev.c
3338
if (segs < READ_ONCE(dev->gso_max_segs))
net/core/dev.c
3794
if (gso_segs > READ_ONCE(dev->gso_max_segs))
net/core/dev.c
4007
READ_ONCE(net_devmem_iov_binding(niov)->dev) != dev)
net/core/dev.c
4223
first_n = READ_ONCE(q->defer_list.first);
net/core/dev.c
4227
if (unlikely(defer_count > READ_ONCE(net_hotdata.qdisc_max_burst))) {
net/core/dev.c
4668
val = READ_ONCE(sk->sk_tx_queue_mapping);
net/core/dev.c
4676
resel = READ_ONCE(sock_net(sk)->core.sysctl_txq_reselection);
net/core/dev.c
4678
READ_ONCE(sk->sk_tx_queue_mapping_jiffies) + resel))
net/core/dev.c
4956
thread = READ_ONCE(napi->thread);
net/core/dev.c
5012
sd_input_head = READ_ONCE(per_cpu(softnet_data, cpu).input_queue_head);
net/core/dev.c
5013
flow_last_active = READ_ONCE(rflow->last_qtail);
net/core/dev.c
5051
tmp_cpu = READ_ONCE(tmp_rflow->cpu);
net/core/dev.c
5053
if (READ_ONCE(tmp_rflow->filter) != RPS_NO_FILTER) {
net/core/dev.c
5056
if (hash != READ_ONCE(tmp_rflow->hash) ||
net/core/dev.c
5076
head = READ_ONCE(per_cpu(softnet_data, next_cpu).input_queue_head);
net/core/dev.c
5134
ident = READ_ONCE(sock_flow_table->ents[hash & sock_flow_table->mask]);
net/core/dev.c
5159
((int)(READ_ONCE(per_cpu(softnet_data, tcpu).input_queue_head) -
net/core/dev.c
5213
cpu = READ_ONCE(rflow->cpu);
net/core/dev.c
5214
if (READ_ONCE(rflow->filter) == filter_id &&
net/core/dev.c
5362
max_backlog = READ_ONCE(net_hotdata.max_backlog);
net/core/dev.c
5668
net_timestamp_check(READ_ONCE(net_hotdata.tstamp_prequeue), skb);
net/core/dev.c
5963
net_timestamp_check(!READ_ONCE(net_hotdata.tstamp_prequeue), skb);
net/core/dev.c
6362
net_timestamp_check(READ_ONCE(net_hotdata.tstamp_prequeue), skb);
net/core/dev.c
6391
net_timestamp_check(READ_ONCE(net_hotdata.tstamp_prequeue),
net/core/dev.c
6484
if (READ_ONCE(skb->dev->reg_state) == NETREG_UNREGISTERING) {
net/core/dev.c
6494
if (READ_ONCE(skb->dev->reg_state) == NETREG_UNREGISTERING) {
net/core/dev.c
6637
napi->weight = READ_ONCE(net_hotdata.dev_rx_weight);
net/core/dev.c
6710
unsigned long new, val = READ_ONCE(n->state);
net/core/dev.c
6792
val = READ_ONCE(n->state);
net/core/dev.c
6938
unsigned long val = READ_ONCE(napi->state);
net/core/dev.c
713
return READ_ONCE(dev->ifindex);
net/core/dev.c
7137
val = READ_ONCE(napi->state);
net/core/dev.c
7565
napi_set_defer_hard_irqs(napi, READ_ONCE(dev->napi_defer_hard_irqs));
net/core/dev.c
7566
napi_set_gro_flush_timeout(napi, READ_ONCE(dev->gro_flush_timeout));
net/core/dev.c
7589
val = READ_ONCE(n->state);
net/core/dev.c
7593
val = READ_ONCE(n->state);
net/core/dev.c
7631
unsigned long new, val = READ_ONCE(n->state);
net/core/dev.c
7872
val = READ_ONCE(napi->state);
net/core/dev.c
7894
usecs_to_jiffies(READ_ONCE(net_hotdata.netdev_budget_usecs));
net/core/dev.c
7896
int budget = READ_ONCE(net_hotdata.netdev_budget);
net/core/dev.c
843
if (!napi || READ_ONCE(napi->dev->reg_state) != NETREG_REGISTERED) {
net/core/dev.c
9722
flags = (READ_ONCE(dev->flags) & ~(IFF_PROMISC |
net/core/dev.c
9727
(READ_ONCE(dev->gflags) & (IFF_PROMISC |
net/core/dev.h
217
return READ_ONCE(n->defer_hard_irqs);
net/core/dev.h
260
return READ_ONCE(n->gro_flush_timeout);
net/core/dev.h
306
return READ_ONCE(n->irq_suspend_timeout);
net/core/dev.h
364
WARN_ON(READ_ONCE(napi->list_owner) != -1);
net/core/devmem.c
400
unlikely(dst_dev != READ_ONCE(binding->dev))) {
net/core/dst_cache.c
54
READ_ONCE(dst_cache->reset_ts)) ||
net/core/dst_cache.c
55
(READ_ONCE(dst->obsolete) && !dst->ops->check(dst, idst->cookie)))) {
net/core/fib_rules.c
1181
if (READ_ONCE(rule->iifindex) == -1)
net/core/fib_rules.c
1188
if (READ_ONCE(rule->oifindex) == -1)
net/core/fib_rules.c
263
u8 iif_is_l3_master = READ_ONCE(rule->iif_is_l3_master);
net/core/fib_rules.c
272
u8 oif_is_l3_master = READ_ONCE(rule->oif_is_l3_master);
net/core/fib_rules.c
284
iifindex = READ_ONCE(rule->iifindex);
net/core/fib_rules.c
288
oifindex = READ_ONCE(rule->oifindex);
net/core/fib_rules.c
40
if (READ_ONCE(rule->iifindex) || READ_ONCE(rule->oifindex) ||
net/core/fib_rules.c
433
fib_rules_seq = READ_ONCE(ops->fib_rules_seq);
net/core/filter.c
12272
if (!READ_ONCE(net->ipv4.sysctl_tcp_window_scaling))
net/core/filter.c
12280
if (attrs->sack_ok && !READ_ONCE(net->ipv4.sysctl_tcp_sack))
net/core/filter.c
12284
if (!READ_ONCE(net->ipv4.sysctl_tcp_timestamps))
net/core/filter.c
1243
int optmem_max = READ_ONCE(sock_net(sk)->core.sysctl_optmem_max);
net/core/filter.c
1579
optmem_max = READ_ONCE(sock_net(sk)->core.sysctl_optmem_max);
net/core/filter.c
1647
optmem_max = READ_ONCE(sock_net(sk)->core.sysctl_optmem_max);
net/core/filter.c
4464
map = READ_ONCE(ri->map);
net/core/filter.c
4553
map = READ_ONCE(ri->map);
net/core/filter.c
4991
cgrp = READ_ONCE(array->ptrs[idx]);
net/core/filter.c
6227
if (!neigh || !(READ_ONCE(neigh->nud_state) & NUD_VALID))
net/core/filter.c
6261
if (unlikely(!idev || !READ_ONCE(idev->cnf.forwarding)))
net/core/filter.c
6369
if (!neigh || !(READ_ONCE(neigh->nud_state) & NUD_VALID))
net/core/filter.c
6500
mtu = READ_ONCE(dev->mtu);
net/core/filter.c
6546
mtu = READ_ONCE(dev->mtu);
net/core/filter.c
7571
if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies))
net/core/filter.c
7644
if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies))
net/core/flow_dissector.c
1160
prog = READ_ONCE(run_array->items[0].prog);
net/core/flow_dissector.c
355
key->ct_mark = READ_ONCE(ct->mark);
net/core/gro_cells.c
34
if (skb_queue_len(&cell->napi_skbs) > READ_ONCE(net_hotdata.max_backlog)) {
net/core/link_watch.c
78
if (operstate == READ_ONCE(dev->operstate))
net/core/neighbour.c
1001
if (atomic_read(&tbl->entries) < READ_ONCE(tbl->gc_thresh1))
net/core/neighbour.c
1305
if (READ_ONCE(hh->hh_len)) {
net/core/neighbour.c
1338
READ_ONCE(dst->obsolete) != DST_OBSOLETE_DEAD) {
net/core/neighbour.c
1343
READ_ONCE(n1->output)(n1, skb);
net/core/neighbour.c
1609
if (dev->header_ops->cache && !READ_ONCE(neigh->hh.hh_len))
net/core/neighbour.c
2160
state = READ_ONCE(neigh->nud_state) & ~NUD_PERMANENT;
net/core/neighbour.c
2207
nla_put_u32(skb, NDTPA_IFINDEX, READ_ONCE(parms->dev->ifindex))) ||
net/core/neighbour.c
2222
nla_put_msecs(skb, NDTPA_REACHABLE_TIME, READ_ONCE(parms->reachable_time),
net/core/neighbour.c
2264
nla_put_msecs(skb, NDTA_GC_INTERVAL, READ_ONCE(tbl->gc_interval),
net/core/neighbour.c
2266
nla_put_u32(skb, NDTA_THRESH1, READ_ONCE(tbl->gc_thresh1)) ||
net/core/neighbour.c
2267
nla_put_u32(skb, NDTA_THRESH2, READ_ONCE(tbl->gc_thresh2)) ||
net/core/neighbour.c
2268
nla_put_u32(skb, NDTA_THRESH3, READ_ONCE(tbl->gc_thresh3)))
net/core/neighbour.c
2272
long flush_delta = now - READ_ONCE(tbl->last_flush);
net/core/neighbour.c
2273
long rand_delta = now - READ_ONCE(tbl->last_rand);
net/core/neighbour.c
2281
.ndtc_proxy_qlen = READ_ONCE(tbl->proxy_queue.qlen),
net/core/neighbour.c
2302
ndst.ndts_allocs += READ_ONCE(st->allocs);
net/core/neighbour.c
2303
ndst.ndts_destroys += READ_ONCE(st->destroys);
net/core/neighbour.c
2304
ndst.ndts_hash_grows += READ_ONCE(st->hash_grows);
net/core/neighbour.c
2305
ndst.ndts_res_failed += READ_ONCE(st->res_failed);
net/core/neighbour.c
2306
ndst.ndts_lookups += READ_ONCE(st->lookups);
net/core/neighbour.c
2307
ndst.ndts_hits += READ_ONCE(st->hits);
net/core/neighbour.c
2308
ndst.ndts_rcv_probes_mcast += READ_ONCE(st->rcv_probes_mcast);
net/core/neighbour.c
2309
ndst.ndts_rcv_probes_ucast += READ_ONCE(st->rcv_probes_ucast);
net/core/neighbour.c
2310
ndst.ndts_periodic_gc_runs += READ_ONCE(st->periodic_gc_runs);
net/core/neighbour.c
2311
ndst.ndts_forced_gc_runs += READ_ONCE(st->forced_gc_runs);
net/core/neighbour.c
2312
ndst.ndts_table_fulls += READ_ONCE(st->table_fulls);
net/core/neighbour.c
256
READ_ONCE(tbl->gc_thresh2);
net/core/neighbour.c
2736
neigh_flags = READ_ONCE(pn->flags);
net/core/neighbour.c
2752
protocol = READ_ONCE(pn->protocol);
net/core/neighbour.c
3230
err = READ_ONCE(neigh->output)(neigh, skb);
net/core/neighbour.c
3275
if (READ_ONCE(n->nud_state) & ~NUD_NOARP)
net/core/neighbour.c
508
gc_thresh3 = READ_ONCE(tbl->gc_thresh3);
net/core/neighbour.c
510
(entries >= READ_ONCE(tbl->gc_thresh2) &&
net/core/neighbour.c
511
time_after(now, READ_ONCE(tbl->last_flush) + 5 * HZ))) {
net/core/net-procfs.c
137
flow_limit_count = READ_ONCE(fl->count);
net/core/net-procfs.c
148
READ_ONCE(sd->processed),
net/core/net-procfs.c
150
READ_ONCE(sd->time_squeeze), 0,
net/core/net-procfs.c
153
READ_ONCE(sd->received_rps), flow_limit_count,
net/core/net-procfs.c
243
nxt = READ_ONCE(pt->list.next);
net/core/net-procfs.c
250
nxt = READ_ONCE(dev->ptype_all.next);
net/core/net-procfs.c
257
nxt = READ_ONCE(net->ptype_all.next);
net/core/net-procfs.c
268
nxt = READ_ONCE(net->ptype_specific.next);
net/core/net-procfs.c
274
nxt = READ_ONCE(ptype_base[0].next);
net/core/net-procfs.c
281
nxt = READ_ONCE(ptype_base[hash].next);
net/core/net-sysfs.c
147
return sysfs_emit(buf, format_string, READ_ONCE(dev->field)); \
net/core/net-sysfs.c
1631
return sysfs_emit(buf, "%u\n", READ_ONCE(queue->dql.stall_max));
net/core/net-sysfs.c
242
return sysfs_emit(buf, fmt_dec, READ_ONCE(dev->name_assign_type));
net/core/net-sysfs.c
252
if (READ_ONCE(ndev->name_assign_type) != NET_NAME_UNKNOWN)
net/core/net-sysfs.c
43
return READ_ONCE(dev->reg_state) <= NETREG_REGISTERED;
net/core/net-sysfs.c
448
operstate = READ_ONCE(netdev->operstate);
net/core/net-sysfs.c
743
ret = sysfs_emit(buf, fmt_dec, READ_ONCE(netdev->threaded));
net/core/net_namespace.c
70
unsigned int gen_ptrs = READ_ONCE(max_gen_ptrs);
net/core/netpoll.c
234
while (READ_ONCE(skb_pool->qlen) < MAX_SKBS) {
net/core/netpoll.c
303
if (READ_ONCE(napi->poll_owner) == smp_processor_id())
net/core/page_pool.c
1220
netdev = READ_ONCE(pool->slow.netdev);
net/core/page_pool.c
1261
if (READ_ONCE(pool->p.napi) == napi)
net/core/page_pool.c
687
u32 hold_cnt = READ_ONCE(pool->pages_state_hold_cnt);
net/core/page_pool.c
891
if (READ_ONCE(pool->cpuid) == cpuid)
net/core/page_pool.c
894
napi = READ_ONCE(pool->p.napi);
net/core/page_pool.c
896
return napi && READ_ONCE(napi->list_owner) == cpuid;
net/core/page_pool_user.c
235
napi_id = pool->p.napi ? READ_ONCE(pool->p.napi->napi_id) : 0;
net/core/pktgen.c
3513
bool skb_shared = !!(READ_ONCE(pkt_dev->flags) & F_SHARED);
net/core/pktgen.c
3527
burst = READ_ONCE(pkt_dev->burst);
net/core/pktgen.c
3528
clone_skb = READ_ONCE(pkt_dev->clone_skb);
net/core/rtnetlink.c
1038
delta = jiffies - READ_ONCE(dst->lastuse);
net/core/rtnetlink.c
1056
unsigned int old = READ_ONCE(dev->operstate);
net/core/rtnetlink.c
1069
unsigned char operstate = READ_ONCE(dev->operstate);
net/core/rtnetlink.c
1702
map.mem_start = READ_ONCE(dev->mem_start);
net/core/rtnetlink.c
1703
map.mem_end = READ_ONCE(dev->mem_end);
net/core/rtnetlink.c
1704
map.base_addr = READ_ONCE(dev->base_addr);
net/core/rtnetlink.c
1705
map.irq = READ_ONCE(dev->irq);
net/core/rtnetlink.c
1706
map.dma = READ_ONCE(dev->dma);
net/core/rtnetlink.c
1707
map.port = READ_ONCE(dev->if_port);
net/core/rtnetlink.c
1847
READ_ONCE(upper_dev->ifindex));
net/core/rtnetlink.c
1858
if (force || READ_ONCE(dev->ifindex) != iflink)
net/core/rtnetlink.c
1980
if (nla_put_u8(skb, IFLA_PROTO_DOWN, READ_ONCE(dev->proto_down)))
net/core/rtnetlink.c
1983
preason = READ_ONCE(dev->proto_down_reason);
net/core/rtnetlink.c
2069
ifm->ifi_type = READ_ONCE(dev->type);
net/core/rtnetlink.c
2070
ifm->ifi_index = READ_ONCE(dev->ifindex);
net/core/rtnetlink.c
2081
if (nla_put_u32(skb, IFLA_TXQLEN, READ_ONCE(dev->tx_queue_len)) ||
net/core/rtnetlink.c
2083
netif_running(dev) ? READ_ONCE(dev->operstate) :
net/core/rtnetlink.c
2085
nla_put_u8(skb, IFLA_LINKMODE, READ_ONCE(dev->link_mode)) ||
net/core/rtnetlink.c
2087
nla_put_u32(skb, IFLA_MTU, READ_ONCE(dev->mtu)) ||
net/core/rtnetlink.c
2088
nla_put_u32(skb, IFLA_MIN_MTU, READ_ONCE(dev->min_mtu)) ||
net/core/rtnetlink.c
2089
nla_put_u32(skb, IFLA_MAX_MTU, READ_ONCE(dev->max_mtu)) ||
net/core/rtnetlink.c
2090
nla_put_u32(skb, IFLA_GROUP, READ_ONCE(dev->group)) ||
net/core/rtnetlink.c
2091
nla_put_u32(skb, IFLA_PROMISCUITY, READ_ONCE(dev->promiscuity)) ||
net/core/rtnetlink.c
2092
nla_put_u32(skb, IFLA_ALLMULTI, READ_ONCE(dev->allmulti)) ||
net/core/rtnetlink.c
2094
READ_ONCE(dev->num_tx_queues)) ||
net/core/rtnetlink.c
2096
READ_ONCE(dev->gso_max_segs)) ||
net/core/rtnetlink.c
2098
READ_ONCE(dev->gso_max_size)) ||
net/core/rtnetlink.c
2100
READ_ONCE(dev->gro_max_size)) ||
net/core/rtnetlink.c
2102
READ_ONCE(dev->gso_ipv4_max_size)) ||
net/core/rtnetlink.c
2104
READ_ONCE(dev->gro_ipv4_max_size)) ||
net/core/rtnetlink.c
2106
READ_ONCE(dev->tso_max_size)) ||
net/core/rtnetlink.c
2108
READ_ONCE(dev->tso_max_segs)) ||
net/core/rtnetlink.c
2110
READ_ONCE(dev->max_pacing_offload_horizon)) ||
net/core/rtnetlink.c
2113
READ_ONCE(dev->num_rx_queues)) ||
net/core/rtnetlink.c
2126
READ_ONCE(dev->needed_headroom)) ||
net/core/rtnetlink.c
2128
READ_ONCE(dev->needed_tailroom)))
net/core/scm.c
168
const struct proto_ops *ops = READ_ONCE(sock->ops);
net/core/secure_seq.c
119
if (READ_ONCE(net->ipv4.sysctl_tcp_timestamps) != 1)
net/core/secure_seq.c
72
if (READ_ONCE(net->ipv4.sysctl_tcp_timestamps) != 1)
net/core/skbuff.c
5474
(unsigned int)READ_ONCE(sk->sk_rcvbuf))
net/core/skbuff.c
5579
if (READ_ONCE(sk->sk_tsflags) & SOF_TIMESTAMPING_OPT_ID) {
net/core/skbuff.c
5597
if (likely(tsonly || READ_ONCE(sock_net(sk)->core.sysctl_tstamp_allow_data)))
net/core/skbuff.c
5606
sock = READ_ONCE(sk->sk_socket);
net/core/skbuff.c
5609
file = READ_ONCE(sock->file);
net/core/skbuff.c
5708
tsflags = READ_ONCE(sk->sk_tsflags);
net/core/skbuff.c
7304
defer_max = READ_ONCE(net_hotdata.sysctl_skb_defer_max);
net/core/skbuff.c
7351
size_t frag_limit = READ_ONCE(net_hotdata.sysctl_max_skb_frags);
net/core/skmsg.c
1015
prog = READ_ONCE(psock->progs.stream_verdict);
net/core/skmsg.c
1123
prog = READ_ONCE(psock->progs.stream_verdict);
net/core/skmsg.c
1150
prog = READ_ONCE(psock->progs.stream_parser);
net/core/skmsg.c
1251
prog = READ_ONCE(psock->progs.stream_verdict);
net/core/skmsg.c
1253
prog = READ_ONCE(psock->progs.skb_verdict);
net/core/skmsg.c
1277
sock = READ_ONCE(sk->sk_socket);
net/core/skmsg.c
1279
ops = READ_ONCE(sock->ops);
net/core/skmsg.c
772
prot = READ_ONCE(sk->sk_prot);
net/core/skmsg.c
926
prog = READ_ONCE(psock->progs.msg_parser);
net/core/sock.c
1244
if (val > READ_ONCE(sk->sk_busy_poll_budget) &&
net/core/sock.c
1268
pacing_rate = READ_ONCE(sk->sk_pacing_rate);
net/core/sock.c
1279
val = READ_ONCE(sock_net(sk)->core.sysctl_txrehash);
net/core/sock.c
1289
set_peek_off = READ_ONCE(sock->ops)->set_peek_off;
net/core/sock.c
1341
val = min_t(u32, val, READ_ONCE(sysctl_wmem_max));
net/core/sock.c
1373
__sock_set_rcvbuf(sk, min_t(u32, val, READ_ONCE(sysctl_rmem_max)));
net/core/sock.c
1456
set_rcvlowat = READ_ONCE(sock->ops)->set_rcvlowat;
net/core/sock.c
1766
v.val = READ_ONCE(sk->sk_sndbuf);
net/core/sock.c
1770
v.val = READ_ONCE(sk->sk_rcvbuf);
net/core/sock.c
1812
v.val = READ_ONCE(sk->sk_priority);
net/core/sock.c
1818
v.ling.l_linger = READ_ONCE(sk->sk_lingertime) / HZ;
net/core/sock.c
1850
v.timestamping.flags = READ_ONCE(sk->sk_tsflags);
net/core/sock.c
1851
v.timestamping.bind_phc = READ_ONCE(sk->sk_bind_phc);
net/core/sock.c
1857
lv = sock_get_timeout(READ_ONCE(sk->sk_rcvtimeo), &v,
net/core/sock.c
1863
lv = sock_get_timeout(READ_ONCE(sk->sk_sndtimeo), &v,
net/core/sock.c
1868
v.val = READ_ONCE(sk->sk_rcvlowat);
net/core/sock.c
1980
lv = READ_ONCE(sock->ops)->getname(sock, (struct sockaddr *)&address, 2);
net/core/sock.c
2009
v.val = READ_ONCE(sk->sk_mark);
net/core/sock.c
2029
if (!READ_ONCE(sock->ops)->set_peek_off)
net/core/sock.c
2032
v.val = READ_ONCE(sk->sk_peek_off);
net/core/sock.c
2062
v.val = READ_ONCE(sk->sk_ll_usec);
net/core/sock.c
2065
v.val = READ_ONCE(sk->sk_prefer_busy_poll);
net/core/sock.c
2073
v.ulval = READ_ONCE(sk->sk_max_pacing_rate);
net/core/sock.c
2077
READ_ONCE(sk->sk_max_pacing_rate));
net/core/sock.c
2082
v.val = READ_ONCE(sk->sk_incoming_cpu);
net/core/sock.c
2100
v.val = READ_ONCE(sk->sk_napi_id);
net/core/sock.c
2130
v.val = READ_ONCE(sk->sk_bound_dev_if);
net/core/sock.c
2145
v.val = READ_ONCE(sk->sk_reserved_mem);
net/core/sock.c
2153
v.val = READ_ONCE(sk->sk_txrehash);
net/core/sock.c
2205
const struct proto *prot = READ_ONCE(osk->sk_prot);
net/core/sock.c
2310
if (READ_ONCE(net->core.sysctl_bypass_prot_mem))
net/core/sock.c
2476
struct proto *prot = READ_ONCE(sk->sk_prot);
net/core/sock.c
2618
max_size = is_ipv6 ? READ_ONCE(dev->gso_max_size) :
net/core/sock.c
2619
READ_ONCE(dev->gso_ipv4_max_size);
net/core/sock.c
2651
max_segs = max_t(u32, READ_ONCE(dev->gso_max_segs), 1);
net/core/sock.c
2827
refcount_read(&sk->sk_wmem_alloc) < READ_ONCE(sk->sk_sndbuf)) {
net/core/sock.c
2853
READ_ONCE(sock_net(sk)->core.sysctl_optmem_max))
net/core/sock.c
2871
int optmem_max = READ_ONCE(sock_net(sk)->core.sysctl_optmem_max);
net/core/sock.c
2948
if (refcount_read(&sk->sk_wmem_alloc) < READ_ONCE(sk->sk_sndbuf))
net/core/sock.c
2950
if (READ_ONCE(sk->sk_shutdown) & SEND_SHUTDOWN)
net/core/sock.c
2952
if (READ_ONCE(sk->sk_err))
net/core/sock.c
2980
if (READ_ONCE(sk->sk_shutdown) & SEND_SHUTDOWN)
net/core/sock.c
2983
if (sk_wmem_alloc_get(sk) < READ_ONCE(sk->sk_sndbuf))
net/core/sock.c
3047
tsflags = READ_ONCE(sk->sk_tsflags);
net/core/sock.c
3113
if (memory_pressure && READ_ONCE(*memory_pressure))
net/core/sock.c
3704
sk->sk_rcvbuf = READ_ONCE(sysctl_rmem_default);
net/core/sock.c
3705
sk->sk_sndbuf = READ_ONCE(sysctl_wmem_default);
net/core/sock.c
3748
sk->sk_ll_usec = READ_ONCE(sysctl_net_busy_read);
net/core/sock.c
3946
return READ_ONCE(sk->sk_prot)->getsockopt(sk, level, optname, optval, optlen);
net/core/sock.c
3973
return READ_ONCE(sk->sk_prot)->setsockopt(sk, level, optname, optval, optlen);
net/core/sock.c
4017
mem[SK_MEMINFO_RCVBUF] = READ_ONCE(sk->sk_rcvbuf);
net/core/sock.c
4019
mem[SK_MEMINFO_SNDBUF] = READ_ONCE(sk->sk_sndbuf);
net/core/sock.c
4020
mem[SK_MEMINFO_FWD_ALLOC] = READ_ONCE(sk->sk_forward_alloc);
net/core/sock.c
4021
mem[SK_MEMINFO_WMEM_QUEUED] = READ_ONCE(sk->sk_wmem_queued);
net/core/sock.c
4023
mem[SK_MEMINFO_BACKLOG] = READ_ONCE(sk->sk_backlog.len);
net/core/sock.c
4438
ret = READ_ONCE(sk->sk_prot)->ioctl(sk, cmd, karg);
net/core/sock.c
4457
ret = READ_ONCE(sk->sk_prot)->ioctl(sk, cmd, &karg);
net/core/sock.c
493
if (atomic_read(&sk->sk_rmem_alloc) >= READ_ONCE(sk->sk_rcvbuf)) {
net/core/sock.c
564
if (sk_rcvqueues_full(sk, READ_ONCE(sk->sk_rcvbuf))) {
net/core/sock.c
582
} else if ((err = sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)))) {
net/core/sock.c
611
if (dst && READ_ONCE(dst->obsolete) &&
net/core/sock.c
629
if (dst && READ_ONCE(dst->obsolete) &&
net/core/sock.c
738
int bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
net/core/sock.c
781
switch (READ_ONCE(sk->sk_family)) {
net/core/sock_map.c
1637
saved_unhash = READ_ONCE(sk->sk_prot)->unhash;
net/core/sock_map.c
1659
saved_destroy = READ_ONCE(sk->sk_prot)->destroy;
net/core/sock_map.c
1694
saved_close = READ_ONCE(sk->sk_prot)->close;
net/core/sock_map.c
227
stream_verdict = READ_ONCE(progs->stream_verdict);
net/core/sock_map.c
234
stream_parser = READ_ONCE(progs->stream_parser);
net/core/sock_map.c
243
msg_parser = READ_ONCE(progs->msg_parser);
net/core/sock_map.c
252
skb_verdict = READ_ONCE(progs->skb_verdict);
net/core/sock_map.c
268
if ((msg_parser && READ_ONCE(psock->progs.msg_parser)) ||
net/core/sock_map.c
269
(stream_parser && READ_ONCE(psock->progs.stream_parser)) ||
net/core/sock_map.c
270
(skb_verdict && READ_ONCE(psock->progs.skb_verdict)) ||
net/core/sock_map.c
271
(skb_verdict && READ_ONCE(psock->progs.stream_verdict)) ||
net/core/sock_map.c
272
(stream_verdict && READ_ONCE(psock->progs.skb_verdict)) ||
net/core/sock_map.c
273
(stream_verdict && READ_ONCE(psock->progs.stream_verdict))) {
net/core/sock_map.c
385
return READ_ONCE(stab->sks[key]);
net/core/sock_reuseport.c
288
more_reuse->synq_overflow_ts = READ_ONCE(reuse->synq_overflow_ts);
net/core/sock_reuseport.c
475
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_migrate_req) ||
net/core/sock_reuseport.c
539
if (!READ_ONCE(reuse->incoming_cpu))
net/core/sock_reuseport.c
543
if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
net/core/sock_reuseport.c
586
socks = READ_ONCE(reuse->num_socks);
net/core/sock_reuseport.c
637
socks = READ_ONCE(reuse->num_socks);
net/core/sock_reuseport.c
647
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_migrate_req))
net/core/stream.c
149
ret = sk_wait_event(sk, &current_timeo, READ_ONCE(sk->sk_err) ||
net/core/stream.c
150
(READ_ONCE(sk->sk_shutdown) & SEND_SHUTDOWN) ||
net/core/stream.c
80
!READ_ONCE(sk->sk_err) &&
net/core/stream.c
81
!((1 << READ_ONCE(sk->sk_state)) &
net/core/stream.c
96
return (1 << READ_ONCE(sk->sk_state)) &
net/core/sysctl_net_core.c
317
weight = READ_ONCE(weight_p);
net/dsa/user.c
355
return READ_ONCE(dsa_user_to_conduit(dev)->ifindex);
net/handshake/request.c
249
if (READ_ONCE(hn->hn_pending) >= hn->hn_pending_max)
net/ieee802154/6lowpan/core.c
97
return READ_ONCE(lowpan_802154_dev(dev)->wdev->ifindex);
net/ipv4/af_inet.c
1306
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_ip_dynaddr) > 1) {
net/ipv4/af_inet.c
1346
if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_ip_dynaddr) ||
net/ipv4/af_inet.c
1588
unsigned int family = READ_ONCE(sk->sk_family);
net/ipv4/af_inet.c
218
tcp_fastopen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen);
net/ipv4/af_inet.c
356
if (READ_ONCE(net->ipv4.sysctl_ip_no_pmtu_disc))
net/ipv4/af_inet.c
368
sk->sk_txrehash = READ_ONCE(net->core.sysctl_txrehash);
net/ipv4/af_inet.c
587
prot = READ_ONCE(sk->sk_prot);
net/ipv4/af_inet.c
795
sk2 = READ_ONCE(sk1->sk_prot)->accept(sk1, arg);
net/ipv4/af_inet.c
879
prot = READ_ONCE(sk->sk_prot);
net/ipv4/arp.c
1187
if (READ_ONCE(neigh->nud_state) & NUD_NOARP) {
net/ipv4/arp.c
1213
if ((READ_ONCE(neigh->nud_state) & NUD_VALID) && !force) {
net/ipv4/arp.c
1218
if (READ_ONCE(neigh->nud_state) & ~NUD_NOARP)
net/ipv4/arp.c
379
if (!(READ_ONCE(neigh->nud_state) & NUD_VALID))
net/ipv4/cipso_ipv4.c
1199
if (READ_ONCE(cipso_v4_rbm_optfmt) && ret_val > 0 &&
net/ipv4/cipso_ipv4.c
1604
if (READ_ONCE(cipso_v4_rbm_strictvalid)) {
net/ipv4/cipso_ipv4.c
241
if (!READ_ONCE(cipso_v4_cache_enabled))
net/ipv4/cipso_ipv4.c
298
int bkt_size = READ_ONCE(cipso_v4_cache_bucketsize);
net/ipv4/cipso_ipv4.c
305
if (!READ_ONCE(cipso_v4_cache_enabled) || bkt_size <= 0)
net/ipv4/datagram.c
112
if (!dst || !READ_ONCE(dst->obsolete) || dst->ops->check(dst, 0)) {
net/ipv4/datagram.c
42
oif = READ_ONCE(inet->mc_index);
net/ipv4/datagram.c
44
saddr = READ_ONCE(inet->mc_addr);
net/ipv4/datagram.c
46
oif = READ_ONCE(inet->uc_index);
net/ipv4/devinet.c
1349
if (READ_ONCE(ifa->ifa_flags) & IFA_F_SECONDARY)
net/ipv4/devinet.c
1378
if (READ_ONCE(ifa->ifa_flags) & IFA_F_SECONDARY)
net/ipv4/devinet.c
1724
flags = READ_ONCE(ifa->ifa_flags);
net/ipv4/devinet.c
1737
tstamp = READ_ONCE(ifa->ifa_tstamp);
net/ipv4/devinet.c
1739
preferred = READ_ONCE(ifa->ifa_preferred_lft);
net/ipv4/devinet.c
1740
valid = READ_ONCE(ifa->ifa_valid_lft);
net/ipv4/devinet.c
1772
put_cacheinfo(skb, READ_ONCE(ifa->ifa_cstamp), tstamp,
net/ipv4/devinet.c
1918
READ_ONCE(net->dev_base_seq);
net/ipv4/devinet.c
2057
((u32 *) nla_data(nla))[i] = READ_ONCE(in_dev->cnf.data[i]);
net/ipv4/devinet.c
732
flags = READ_ONCE(ifa->ifa_flags);
net/ipv4/devinet.c
736
preferred_lft = READ_ONCE(ifa->ifa_preferred_lft);
net/ipv4/devinet.c
737
valid_lft = READ_ONCE(ifa->ifa_valid_lft);
net/ipv4/devinet.c
738
tstamp = READ_ONCE(ifa->ifa_tstamp);
net/ipv4/fib_lookup.h
31
u8 fa_state = READ_ONCE(fa->fa_state);
net/ipv4/fib_notifier.c
33
return READ_ONCE(net->ipv4.fib_seq) + fib4_rules_seq_read(net);
net/ipv4/fib_semantics.c
1316
if (READ_ONCE(nh->nh_saddr_genid) ==
net/ipv4/fib_semantics.c
1318
return READ_ONCE(nh->nh_saddr);
net/ipv4/fib_semantics.c
1769
if (!READ_ONCE(fi->fib_net->ipv4.sysctl_nexthop_compat_mode))
net/ipv4/fib_semantics.c
2156
state = READ_ONCE(n->nud_state);
net/ipv4/fib_semantics.c
2178
use_neigh = READ_ONCE(net->ipv4.sysctl_fib_multipath_use_neigh);
net/ipv4/fib_semantics.c
561
fri.offload = READ_ONCE(fa->offload);
net/ipv4/fib_semantics.c
562
fri.trap = READ_ONCE(fa->trap);
net/ipv4/fib_semantics.c
563
fri.offload_failed = READ_ONCE(fa->offload_failed);
net/ipv4/fib_semantics.c
595
state = READ_ONCE(n->nud_state);
net/ipv4/fib_trie.c
1054
if (READ_ONCE(fa_match->offload) == fri->offload &&
net/ipv4/fib_trie.c
1055
READ_ONCE(fa_match->trap) == fri->trap &&
net/ipv4/fib_trie.c
1056
READ_ONCE(fa_match->offload_failed) == fri->offload_failed)
net/ipv4/fib_trie.c
1062
fib_notify_on_flag_change = READ_ONCE(net->ipv4.sysctl_fib_notify_on_flag_change);
net/ipv4/fib_trie.c
1066
READ_ONCE(fa_match->offload_failed) == fri->offload_failed)
net/ipv4/fib_trie.c
1283
state = READ_ONCE(fa->fa_state);
net/ipv4/fib_trie.c
1561
if (READ_ONCE(fi->fib_dead))
net/ipv4/fib_trie.c
1748
if (READ_ONCE(fa_to_delete->fa_state) & FA_S_ACCESSED)
net/ipv4/fib_trie.c
2294
fri.offload = READ_ONCE(fa->offload);
net/ipv4/fib_trie.c
2295
fri.trap = READ_ONCE(fa->trap);
net/ipv4/fib_trie.c
2296
fri.offload_failed = READ_ONCE(fa->offload_failed);
net/ipv4/fib_trie.c
496
if (tnode_free_size >= READ_ONCE(sysctl_fib_sync_mem)) {
net/ipv4/icmp.c
1028
if (!ct || !(READ_ONCE(ct->status) & IPS_NAT_MASK)) {
net/ipv4/icmp.c
1137
switch (READ_ONCE(net->ipv4.sysctl_ip_no_pmtu_disc)) {
net/ipv4/icmp.c
1190
if (!READ_ONCE(net->ipv4.sysctl_icmp_ignore_bogus_error_responses) &&
net/ipv4/icmp.c
1251
if (READ_ONCE(net->ipv4.sysctl_icmp_echo_ignore_all))
net/ipv4/icmp.c
1288
if (!READ_ONCE(net->ipv4.sysctl_icmp_echo_enable_probe))
net/ipv4/icmp.c
1499
READ_ONCE(net->ipv4.sysctl_icmp_echo_ignore_broadcasts)) {
net/ipv4/icmp.c
248
oldstamp = READ_ONCE(net->ipv4.icmp_global_stamp);
net/ipv4/icmp.c
253
incr = READ_ONCE(net->ipv4.sysctl_icmp_msgs_per_sec);
net/ipv4/icmp.c
261
new = min(old + incr, READ_ONCE(net->ipv4.sysctl_icmp_msgs_burst));
net/ipv4/icmp.c
288
if (!((1 << type) & READ_ONCE(net->ipv4.sysctl_icmp_ratemask)))
net/ipv4/icmp.c
325
peer_timeout = READ_ONCE(net->ipv4.sysctl_icmp_ratelimit);
net/ipv4/icmp.c
658
if (READ_ONCE(ifa->ifa_flags) & IFA_F_SECONDARY)
net/ipv4/icmp.c
710
data = htonl(READ_ONCE(dev->mtu));
net/ipv4/icmp.c
745
ext_objs = READ_ONCE(net->ipv4.sysctl_icmp_errors_extension_mask);
net/ipv4/icmp.c
920
READ_ONCE(net->ipv4.sysctl_icmp_errors_use_inbound_ifaddr))
net/ipv4/igmp.c
1018
in_dev->mr_qrv = ih3->qrv ?: READ_ONCE(net->ipv4.sysctl_igmp_qrv);
net/ipv4/igmp.c
1057
!READ_ONCE(net->ipv4.sysctl_igmp_llm_reports))
net/ipv4/igmp.c
1198
pmc->crcount = in_dev->mr_qrv ?: READ_ONCE(net->ipv4.sysctl_igmp_qrv);
net/ipv4/igmp.c
1250
READ_ONCE(net->ipv4.sysctl_igmp_qrv);
net/ipv4/igmp.c
1253
READ_ONCE(net->ipv4.sysctl_igmp_qrv);
net/ipv4/igmp.c
1311
!READ_ONCE(net->ipv4.sysctl_igmp_llm_reports))
net/ipv4/igmp.c
1354
!READ_ONCE(net->ipv4.sysctl_igmp_llm_reports))
net/ipv4/igmp.c
1360
im->unsolicit_count = READ_ONCE(net->ipv4.sysctl_igmp_qrv);
net/ipv4/igmp.c
1374
im->crcount = in_dev->mr_qrv ?: READ_ONCE(net->ipv4.sysctl_igmp_qrv);
net/ipv4/igmp.c
1456
ci.cstamp = (READ_ONCE(im->mca_cstamp) - INITIAL_JIFFIES) * 100UL / HZ;
net/ipv4/igmp.c
1741
!READ_ONCE(net->ipv4.sysctl_igmp_llm_reports))
net/ipv4/igmp.c
1850
in_dev->mr_qrv = READ_ONCE(net->ipv4.sysctl_igmp_qrv);
net/ipv4/igmp.c
1985
psf->sf_crcount = in_dev->mr_qrv ?: READ_ONCE(net->ipv4.sysctl_igmp_qrv);
net/ipv4/igmp.c
2049
pmc->crcount = in_dev->mr_qrv ?: READ_ONCE(net->ipv4.sysctl_igmp_qrv);
net/ipv4/igmp.c
2228
pmc->crcount = in_dev->mr_qrv ?: READ_ONCE(net->ipv4.sysctl_igmp_qrv);
net/ipv4/igmp.c
2294
if (count >= READ_ONCE(net->ipv4.sysctl_igmp_max_memberships))
net/ipv4/igmp.c
230
int tv = get_random_u32_below(READ_ONCE(in_dev->mr_maxdelay));
net/ipv4/igmp.c
2481
if (psl && psl->sl_count >= READ_ONCE(net->ipv4.sysctl_igmp_max_msf)) {
net/ipv4/igmp.c
477
!READ_ONCE(net->ipv4.sysctl_igmp_llm_reports))
net/ipv4/igmp.c
480
mtu = READ_ONCE(dev->mtu);
net/ipv4/igmp.c
603
!READ_ONCE(net->ipv4.sysctl_igmp_llm_reports))
net/ipv4/igmp.c
747
!READ_ONCE(net->ipv4.sysctl_igmp_llm_reports))
net/ipv4/igmp.c
818
mr_ifc_count = READ_ONCE(in_dev->mr_ifc_count);
net/ipv4/igmp.c
836
WRITE_ONCE(in_dev->mr_ifc_count, in_dev->mr_qrv ?: READ_ONCE(net->ipv4.sysctl_igmp_qrv));
net/ipv4/igmp.c
932
!READ_ONCE(net->ipv4.sysctl_igmp_llm_reports))
net/ipv4/inet_connection_sock.c
1063
max_syn_ack_retries = READ_ONCE(icsk->icsk_syn_retries) ? :
net/ipv4/inet_connection_sock.c
1064
READ_ONCE(net->ipv4.sysctl_tcp_synack_retries);
net/ipv4/inet_connection_sock.c
1084
if ((qlen << 1) > max(8U, READ_ONCE(sk_listener->sk_max_ack_backlog))) {
net/ipv4/inet_connection_sock.c
1095
syn_ack_recalc(req, max_syn_ack_retries, READ_ONCE(queue->rskq_defer_accept),
net/ipv4/inet_connection_sock.c
139
sk_range = READ_ONCE(inet_sk(sk)->local_port_range);
net/ipv4/inet_connection_sock.c
166
bound_dev_if2 = READ_ONCE(sk2->sk_bound_dev_if);
net/ipv4/inet_connection_sock.c
240
reuseport_cb_ok = !reuseport_cb || READ_ONCE(reuseport_cb->num_closed_socks);
net/ipv4/inet_connection_sock.c
288
reuseport_cb_ok = !reuseport_cb || READ_ONCE(reuseport_cb->num_closed_socks);
net/ipv4/inet_connection_sock.c
398
if (READ_ONCE(net->ipv4.sysctl_ip_autobind_reuse) && !relax) {
net/ipv4/inet_diag.c
112
if (nla_put_u8(skb, INET_DIAG_TOS, READ_ONCE(inet->tos)) < 0)
net/ipv4/inet_diag.c
128
if (net_admin && nla_put_u32(skb, INET_DIAG_MARK, READ_ONCE(sk->sk_mark)))
net/ipv4/inet_diag.c
143
classid = READ_ONCE(sk->sk_priority);
net/ipv4/inet_diag.c
256
.idiag_wmem = READ_ONCE(sk->sk_wmem_queued),
net/ipv4/inet_diag.c
257
.idiag_fmem = READ_ONCE(sk->sk_forward_alloc),
net/ipv4/inet_diag.c
288
r->idiag_retrans = READ_ONCE(icsk->icsk_retransmits);
net/ipv4/inet_diag.c
293
r->idiag_retrans = READ_ONCE(icsk->icsk_probes_out);
net/ipv4/inet_diag.c
298
r->idiag_retrans = READ_ONCE(icsk->icsk_probes_out);
net/ipv4/inet_diag.c
317
ca_ops = READ_ONCE(icsk->icsk_ca_ops);
net/ipv4/inet_diag.c
337
ca_ops = READ_ONCE(icsk->icsk_ca_ops);
net/ipv4/inet_diag.c
55
if (!READ_ONCE(inet_diag_table[proto]))
net/ipv4/inet_diag.c
603
entry.family = READ_ONCE(sk->sk_family);
net/ipv4/inet_diag.c
605
entry.sport = READ_ONCE(inet->inet_num);
net/ipv4/inet_diag.c
606
entry.dport = ntohs(READ_ONCE(inet->inet_dport));
net/ipv4/inet_diag.c
607
entry.ifindex = READ_ONCE(sk->sk_bound_dev_if);
net/ipv4/inet_diag.c
609
entry.userlocks = sk_fullsock(sk) ? READ_ONCE(sk->sk_userlocks) : 0;
net/ipv4/inet_diag.c
612
entry.mark = READ_ONCE(sk->sk_mark);
net/ipv4/inet_diag.c
74
r->idiag_family = READ_ONCE(sk->sk_family);
net/ipv4/inet_diag.c
76
r->id.idiag_sport = htons(READ_ONCE(sk->sk_num));
net/ipv4/inet_diag.c
77
r->id.idiag_dport = READ_ONCE(sk->sk_dport);
net/ipv4/inet_diag.c
78
r->id.idiag_if = READ_ONCE(sk->sk_bound_dev_if);
net/ipv4/inet_diag.c
91
r->id.idiag_src[0] = READ_ONCE(sk->sk_rcv_saddr);
net/ipv4/inet_diag.c
92
r->id.idiag_dst[0] = READ_ONCE(sk->sk_daddr);
net/ipv4/inet_fragment.c
278
if (!READ_ONCE(fqdir->dead)) {
net/ipv4/inet_fragment.c
417
long high_thresh = READ_ONCE(fqdir->high_thresh);
net/ipv4/inet_hashtables.c
1078
offset = READ_ONCE(table_perturb[index]) + (port_offset >> 32);
net/ipv4/inet_hashtables.c
1176
WRITE_ONCE(table_perturb[index], READ_ONCE(table_perturb[index]) + i + step);
net/ipv4/inet_hashtables.c
355
if (net_eq(sock_net(sk), net) && READ_ONCE(sk->sk_num) == hnum &&
net/ipv4/inet_hashtables.c
366
if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
net/ipv4/inet_timewait_sock.c
175
READ_ONCE(dr->sysctl_max_tw_buckets))
net/ipv4/inetpeer.c
113
if (READ_ONCE(p->dtime) != now)
net/ipv4/inetpeer.c
143
peer_threshold = READ_ONCE(inet_peer_threshold);
net/ipv4/inetpeer.c
144
peer_maxttl = READ_ONCE(inet_peer_maxttl);
net/ipv4/inetpeer.c
145
peer_minttl = READ_ONCE(inet_peer_minttl);
net/ipv4/inetpeer.c
155
delta = (__u32)jiffies - READ_ONCE(p->dtime);
net/ipv4/inetpeer.c
255
token = otoken = READ_ONCE(peer->rate_tokens);
net/ipv4/inetpeer.c
257
delta = now - READ_ONCE(peer->rate_last);
net/ipv4/ip_forward.c
159
if (READ_ONCE(net->ipv4.sysctl_ip_fwd_update_priority))
net/ipv4/ip_fragment.c
146
if (READ_ONCE(qp->q.fqdir->dead)) {
net/ipv4/ip_input.c
337
if (READ_ONCE(net->ipv4.sysctl_ip_early_demux) &&
net/ipv4/ip_input.c
343
if (READ_ONCE(net->ipv4.sysctl_tcp_early_demux)) {
net/ipv4/ip_input.c
351
if (READ_ONCE(net->ipv4.sysctl_udp_early_demux)) {
net/ipv4/ip_output.c
1052
READ_ONCE(sk->sk_tsflags) & SOF_TIMESTAMPING_OPT_ID) {
net/ipv4/ip_output.c
1303
dst4_mtu(&rt->dst) : READ_ONCE(rt->dst.dev->mtu);
net/ipv4/ip_output.c
140
int ttl = READ_ONCE(inet->uc_ttl);
net/ipv4/ip_output.c
1439
pmtudisc = READ_ONCE(inet->pmtudisc);
net/ipv4/ip_output.c
1452
ttl = READ_ONCE(inet->mc_ttl);
net/ipv4/ip_output.c
1459
iph->tos = (cork->tos != -1) ? cork->tos : READ_ONCE(inet->tos);
net/ipv4/ip_output.c
1654
sk->sk_sndbuf = READ_ONCE(sysctl_wmem_default);
net/ipv4/ip_output.c
191
skb->priority = READ_ONCE(sk->sk_priority);
net/ipv4/ip_output.c
193
skb->mark = READ_ONCE(sk->sk_mark);
net/ipv4/ip_output.c
531
skb->priority = READ_ONCE(sk->sk_priority);
net/ipv4/ip_output.c
532
skb->mark = READ_ONCE(sk->sk_mark);
net/ipv4/ip_output.c
548
return __ip_queue_xmit(sk, skb, fl, READ_ONCE(inet_sk(sk)->tos));
net/ipv4/ip_sockglue.c
1259
if (optlen > READ_ONCE(net->core.sysctl_optmem_max)) {
net/ipv4/ip_sockglue.c
1270
msf->imsf_numsrc > READ_ONCE(net->ipv4.sysctl_igmp_max_msf)) {
net/ipv4/ip_sockglue.c
1586
val = READ_ONCE(inet->uc_ttl);
net/ipv4/ip_sockglue.c
1588
val = READ_ONCE(sock_net(sk)->ipv4.sysctl_ip_default_ttl);
net/ipv4/ip_sockglue.c
1591
val = READ_ONCE(inet->min_ttl);
net/ipv4/ip_sockglue.c
1594
val = READ_ONCE(inet->mc_ttl);
net/ipv4/ip_sockglue.c
1597
val = READ_ONCE(inet->pmtudisc);
net/ipv4/ip_sockglue.c
1600
val = READ_ONCE(inet->tos);
net/ipv4/ip_sockglue.c
1664
info.ipi_addr.s_addr = READ_ONCE(inet->inet_rcv_saddr);
net/ipv4/ip_sockglue.c
1665
info.ipi_spec_dst.s_addr = READ_ONCE(inet->inet_rcv_saddr);
net/ipv4/ip_sockglue.c
1666
info.ipi_ifindex = READ_ONCE(inet->mc_index);
net/ipv4/ip_sockglue.c
1670
int hlim = READ_ONCE(inet->mc_ttl);
net/ipv4/ip_sockglue.c
1675
int tos = READ_ONCE(inet->rcv_tos);
net/ipv4/ip_sockglue.c
1682
val = (__force int)htonl((__u32) READ_ONCE(inet->uc_index));
net/ipv4/ip_sockglue.c
1688
addr.s_addr = READ_ONCE(inet->mc_addr);
net/ipv4/ip_sockglue.c
1697
val = READ_ONCE(inet->local_port_range);
net/ipv4/ip_sockglue.c
512
if (!(READ_ONCE(sk->sk_tsflags) & SOF_TIMESTAMPING_OPT_CMSG) ||
net/ipv4/ip_sockglue.c
776
if (optlen > READ_ONCE(sock_net(sk)->core.sysctl_optmem_max))
net/ipv4/ip_sockglue.c
786
gsf->gf_numsrc > READ_ONCE(sock_net(sk)->ipv4.sysctl_igmp_max_msf))
net/ipv4/ip_sockglue.c
812
if (optlen > READ_ONCE(sock_net(sk)->core.sysctl_optmem_max) - 4)
net/ipv4/ip_sockglue.c
836
if (n > READ_ONCE(sock_net(sk)->ipv4.sysctl_igmp_max_msf))
net/ipv4/ip_tunnel.c
103
if (READ_ONCE(t->parms.link) == link)
net/ipv4/ip_tunnel.c
1110
return READ_ONCE(tunnel->net);
net/ipv4/ip_tunnel.c
1118
return READ_ONCE(tunnel->parms.link);
net/ipv4/ip_tunnel.c
117
if (READ_ONCE(t->parms.link) == link)
net/ipv4/ip_tunnel.c
137
if (READ_ONCE(t->parms.link) == link)
net/ipv4/ip_tunnel.c
151
if (READ_ONCE(t->parms.link) == link)
net/ipv4/ip_tunnel.c
164
ndev = READ_ONCE(itn->fb_tunnel_dev);
net/ipv4/ip_tunnel.c
225
link == READ_ONCE(t->parms.link) &&
net/ipv4/ip_tunnel.c
764
tunnel->net, READ_ONCE(tunnel->parms.link),
net/ipv4/ipmr.c
1076
int vif_num = READ_ONCE(mrt->mroute_reg_vif_num);
net/ipv4/ipmr.c
1659
vr->icount = READ_ONCE(vif->pkt_in);
net/ipv4/ipmr.c
1660
vr->ocount = READ_ONCE(vif->pkt_out);
net/ipv4/ipmr.c
1661
vr->ibytes = READ_ONCE(vif->bytes_in);
net/ipv4/ipmr.c
1662
vr->obytes = READ_ONCE(vif->bytes_out);
net/ipv4/ipmr.c
1728
vr.icount = READ_ONCE(vif->pkt_in);
net/ipv4/ipmr.c
1729
vr.ocount = READ_ONCE(vif->pkt_out);
net/ipv4/ipmr.c
1730
vr.ibytes = READ_ONCE(vif->bytes_in);
net/ipv4/ipmr.c
1731
vr.obytes = READ_ONCE(vif->bytes_out);
net/ipv4/ipmr.c
1987
for (ct = READ_ONCE(mrt->maxvif) - 1; ct >= 0; ct--) {
net/ipv4/ipmr.c
3188
return READ_ONCE(net->ipv4.ipmr_seq) + ipmr_rules_seq_read(net);
net/ipv4/ipmr.c
532
ipmr_cache_report(mrt, skb, READ_ONCE(mrt->mroute_reg_vif_num),
net/ipv4/ipmr.c
616
vif_num = READ_ONCE(mrt->mroute_reg_vif_num);
net/ipv4/ipmr_base.c
263
lastuse = READ_ONCE(c->mfc_un.res.lastuse);
net/ipv4/netfilter/arp_tables.c
207
private = READ_ONCE(table->private); /* Address dependency. */
net/ipv4/netfilter/ip_tables.c
260
private = READ_ONCE(table->private); /* Address dependency. */
net/ipv4/netfilter/nf_reject_ipv4.c
154
READ_ONCE(net->ipv4.sysctl_ip_default_ttl));
net/ipv4/netfilter/nf_reject_ipv4.c
74
READ_ONCE(net->ipv4.sysctl_ip_default_ttl));
net/ipv4/nexthop.c
1372
state = READ_ONCE(n->nud_state);
net/ipv4/nexthop.c
1389
state = READ_ONCE(n->nud_state);
net/ipv4/nexthop.c
2155
!READ_ONCE(net->ipv4.sysctl_nexthop_compat_mode));
net/ipv4/nexthop.c
2670
READ_ONCE(net->ipv4.sysctl_nexthop_compat_mode))
net/ipv4/ping.c
191
if (READ_ONCE(isk->inet_num) != ident)
net/ipv4/ping.c
194
bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
net/ipv4/ping.c
197
__be32 rcv_saddr = READ_ONCE(isk->inet_rcv_saddr);
net/ipv4/ping.c
555
if (READ_ONCE(inet_sock->pmtudisc) != IP_PMTUDISC_DONT) {
net/ipv4/ping.c
777
ipc.oif = READ_ONCE(inet->mc_index);
net/ipv4/ping.c
779
saddr = READ_ONCE(inet->mc_addr);
net/ipv4/ping.c
781
ipc.oif = READ_ONCE(inet->uc_index);
net/ipv4/ping.c
860
READ_ONCE(isk->inet_num));
net/ipv4/proc.c
400
READ_ONCE(net->ipv4.sysctl_ip_default_ttl));
net/ipv4/raw.c
180
READ_ONCE(sk->sk_rcvbuf)) {
net/ipv4/raw.c
249
harderr = READ_ONCE(inet->pmtudisc) != IP_PMTUDISC_DONT;
net/ipv4/raw.c
587
uc_index = READ_ONCE(inet->uc_index);
net/ipv4/raw.c
590
ipc.oif = READ_ONCE(inet->mc_index);
net/ipv4/raw.c
592
saddr = READ_ONCE(inet->mc_addr);
net/ipv4/route.c
1042
time_before(jiffies, READ_ONCE(dst->expires) -
net/ipv4/route.c
1146
if (READ_ONCE(odst->obsolete) && !odst->ops->check(odst, 0)) {
net/ipv4/route.c
1221
if (READ_ONCE(dst->obsolete) != DST_OBSOLETE_FORCE_CHK ||
net/ipv4/route.c
1422
if (READ_ONCE(dev_net(dev)->ipv4.sysctl_ip_fwd_use_pmtu) ||
net/ipv4/route.c
1435
mtu = min(READ_ONCE(dev->mtu), IP_MAX_MTU);
net/ipv4/route.c
1582
READ_ONCE(rt->dst.obsolete) == DST_OBSOLETE_FORCE_CHK &&
net/ipv4/route.c
1696
new_rt->dst.input = READ_ONCE(rt->dst.input);
net/ipv4/route.c
1697
new_rt->dst.output = READ_ONCE(rt->dst.output);
net/ipv4/route.c
1950
u32 hash_fields = READ_ONCE(net->ipv4.sysctl_fib_multipath_hash_fields);
net/ipv4/route.c
1979
u32 hash_fields = READ_ONCE(net->ipv4.sysctl_fib_multipath_hash_fields);
net/ipv4/route.c
2039
u32 hash_fields = READ_ONCE(net->ipv4.sysctl_fib_multipath_hash_fields);
net/ipv4/route.c
2073
switch (READ_ONCE(net->ipv4.sysctl_fib_multipath_hash_policy)) {
net/ipv4/route.c
3022
expires = READ_ONCE(rt->dst.expires);
net/ipv4/route.c
3430
fri.offload = READ_ONCE(fa->offload);
net/ipv4/route.c
3431
fri.trap = READ_ONCE(fa->trap);
net/ipv4/route.c
3433
READ_ONCE(fa->offload_failed);
net/ipv4/route.c
480
old = READ_ONCE(*p_tstamp);
net/ipv4/route.c
519
mark = READ_ONCE(sk->sk_mark);
net/ipv4/route.c
554
flowi4_init_output(fl4, sk->sk_bound_dev_if, READ_ONCE(sk->sk_mark),
net/ipv4/route.c
793
if (!(READ_ONCE(n->nud_state) & NUD_VALID)) {
net/ipv4/route.c
851
if ((READ_ONCE(dst->obsolete) > 0) ||
net/ipv4/route.c
853
READ_ONCE(rt->dst.expires))
net/ipv4/syncookies.c
247
if (!READ_ONCE(net->ipv4.sysctl_tcp_timestamps))
net/ipv4/syncookies.c
252
if (tcp_opt->sack_ok && !READ_ONCE(net->ipv4.sysctl_tcp_sack))
net/ipv4/syncookies.c
261
return READ_ONCE(net->ipv4.sysctl_tcp_window_scaling) != 0;
net/ipv4/syncookies.c
421
if (!READ_ONCE(net->ipv4.sysctl_tcp_syncookies) ||
net/ipv4/syncookies.c
475
req->rsk_window_clamp = READ_ONCE(tp->window_clamp) ? :
net/ipv4/sysctl_net_ipv4.c
120
pports = READ_ONCE(net->ipv4.sysctl_ip_prot_sock);
net/ipv4/sysctl_net_ipv4.c
504
user_seed = READ_ONCE(mphs->user_seed);
net/ipv4/sysctl_net_ipv4.c
94
(range[0] < READ_ONCE(net->ipv4.sysctl_ip_prot_sock)))
net/ipv4/tcp.c
1051
left = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_wmem[0]) - sk->sk_wmem_queued;
net/ipv4/tcp.c
1073
if (!(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen) &
net/ipv4/tcp.c
1146
sockc = (struct sockcm_cookie){ .tsflags = READ_ONCE(sk->sk_tsflags) };
net/ipv4/tcp.c
1305
if (i >= READ_ONCE(net_hotdata.sysctl_max_skb_frags)) {
net/ipv4/tcp.c
1450
READ_ONCE(sk->sk_write_space)(sk);
net/ipv4/tcp.c
1854
cap = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2]) >> 1;
net/ipv4/tcp.c
2379
u32 tsflags = READ_ONCE(sk->sk_tsflags);
net/ipv4/tcp.c
2448
u32 copied_seq = READ_ONCE(tp->copied_seq);
net/ipv4/tcp.c
2449
u32 rcv_nxt = READ_ONCE(tp->rcv_nxt);
net/ipv4/tcp.c
2453
if (unlikely(inq < 0 || copied_seq != READ_ONCE(tp->copied_seq))) {
net/ipv4/tcp.c
2783
if (copied >= target && !READ_ONCE(sk->sk_backlog.tail))
net/ipv4/tcp.c
3149
return READ_ONCE(tcp_orphan_cache) << shift >
net/ipv4/tcp.c
3150
READ_ONCE(sysctl_tcp_max_orphans);
net/ipv4/tcp.c
3296
if (READ_ONCE(tp->linger2) < 0) {
net/ipv4/tcp.c
335
if (READ_ONCE(tcp_memory_pressure))
net/ipv4/tcp.c
350
if (!READ_ONCE(tcp_memory_pressure))
net/ipv4/tcp.c
398
u32 rate = READ_ONCE(tp->rate_delivered);
net/ipv4/tcp.c
399
u32 intv = READ_ONCE(tp->rate_interval_us);
net/ipv4/tcp.c
4150
} else if (READ_ONCE(net->ipv4.sysctl_tcp_fastopen) &
net/ipv4/tcp.c
4185
READ_ONCE(sk->sk_write_space)(sk);
net/ipv4/tcp.c
4218
return READ_ONCE(icsk->icsk_af_ops)->setsockopt(sk, level, optname,
net/ipv4/tcp.c
4263
rate = READ_ONCE(sk->sk_pacing_rate);
net/ipv4/tcp.c
4267
rate = READ_ONCE(sk->sk_max_pacing_rate);
net/ipv4/tcp.c
4279
info->tcpi_unacked = READ_ONCE(sk->sk_ack_backlog);
net/ipv4/tcp.c
4280
info->tcpi_sacked = READ_ONCE(sk->sk_max_ack_backlog);
net/ipv4/tcp.c
4349
info->tcpi_segs_in = READ_ONCE(tp->segs_in);
net/ipv4/tcp.c
4350
info->tcpi_data_segs_in = READ_ONCE(tp->data_segs_in);
net/ipv4/tcp.c
443
rto_max_ms = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rto_max_ms);
net/ipv4/tcp.c
446
rto_min_us = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rto_min_us);
net/ipv4/tcp.c
4469
rate = READ_ONCE(sk->sk_pacing_rate);
net/ipv4/tcp.c
4481
READ_ONCE(inet_csk(sk)->icsk_retransmits));
net/ipv4/tcp.c
4530
user_mss = READ_ONCE(tp->rx_opt.user_mss);
net/ipv4/tcp.c
4553
val = READ_ONCE(icsk->icsk_syn_retries) ? :
net/ipv4/tcp.c
4554
READ_ONCE(net->ipv4.sysctl_tcp_syn_retries);
net/ipv4/tcp.c
4557
val = READ_ONCE(tp->linger2);
net/ipv4/tcp.c
4559
val = (val ? : READ_ONCE(net->ipv4.sysctl_tcp_fin_timeout)) / HZ;
net/ipv4/tcp.c
4562
val = READ_ONCE(icsk->icsk_accept_queue.rskq_defer_accept);
net/ipv4/tcp.c
4567
val = READ_ONCE(tp->window_clamp);
net/ipv4/tcp.c
470
tp->reordering = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_reordering);
net/ipv4/tcp.c
4701
val = READ_ONCE(icsk->icsk_user_timeout);
net/ipv4/tcp.c
4705
val = READ_ONCE(icsk->icsk_accept_queue.fastopenq.max_qlen);
net/ipv4/tcp.c
4717
val = READ_ONCE(tp->tcp_tx_delay);
net/ipv4/tcp.c
4721
val = tcp_clock_ts(tp->tcp_usec_ts) + READ_ONCE(tp->tsoffset);
net/ipv4/tcp.c
4728
val = READ_ONCE(tp->notsent_lowat);
net/ipv4/tcp.c
481
WRITE_ONCE(sk->sk_sndbuf, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_wmem[1]));
net/ipv4/tcp.c
482
WRITE_ONCE(sk->sk_rcvbuf, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[1]));
net/ipv4/tcp.c
4859
val = jiffies_to_usecs(READ_ONCE(inet_csk(sk)->icsk_rto_min));
net/ipv4/tcp.c
4862
val = jiffies_to_usecs(READ_ONCE(inet_csk(sk)->icsk_delack_max));
net/ipv4/tcp.c
4894
return READ_ONCE(icsk->icsk_af_ops)->getsockopt(sk, level, optname,
net/ipv4/tcp.c
4935
u8 keylen = READ_ONCE(key->keylen); /* paired with WRITE_ONCE() in tcp_md5_do_add */
net/ipv4/tcp.c
522
u32 notsent_bytes = READ_ONCE(tp->write_seq) - READ_ONCE(tp->snd_nxt);
net/ipv4/tcp.c
590
shutdown = READ_ONCE(sk->sk_shutdown);
net/ipv4/tcp.c
600
u16 urg_data = READ_ONCE(tp->urg_data);
net/ipv4/tcp.c
603
READ_ONCE(tp->urg_seq) == READ_ONCE(tp->copied_seq) &&
net/ipv4/tcp.c
641
if (READ_ONCE(sk->sk_err) ||
net/ipv4/tcp.c
665
answ = READ_ONCE(tp->urg_data) &&
net/ipv4/tcp.c
666
READ_ONCE(tp->urg_seq) == READ_ONCE(tp->copied_seq);
net/ipv4/tcp.c
675
answ = READ_ONCE(tp->write_seq) - tp->snd_una;
net/ipv4/tcp.c
684
answ = READ_ONCE(tp->write_seq) -
net/ipv4/tcp.c
685
READ_ONCE(tp->snd_nxt);
net/ipv4/tcp.c
745
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_autocorking) &&
net/ipv4/tcp_ao.c
1000
current_key = READ_ONCE(info->current_key);
net/ipv4/tcp_ao.c
1641
int bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
net/ipv4/tcp_ao.c
1798
if (unlikely(READ_ONCE(ao_info->current_key) == key ||
net/ipv4/tcp_ao.c
1799
READ_ONCE(ao_info->rnext_key) == key)) {
net/ipv4/tcp_ao.c
2235
current_key = READ_ONCE(ao_info->current_key);
net/ipv4/tcp_ao.c
2360
current_key = READ_ONCE(ao->current_key);
net/ipv4/tcp_ao.c
2436
opt.snd_sne = READ_ONCE(ao->snd_sne);
net/ipv4/tcp_ao.c
2437
opt.rcv_sne = READ_ONCE(ao->rcv_sne);
net/ipv4/tcp_ao.c
744
unsigned int family = READ_ONCE(sk->sk_family);
net/ipv4/tcp_ao.c
801
rnext_key = READ_ONCE(ao_info->rnext_key);
net/ipv4/tcp_ao.c
803
*sne = tcp_ao_compute_sne(READ_ONCE(ao_info->snd_sne),
net/ipv4/tcp_ao.c
838
sne = tcp_ao_compute_sne(READ_ONCE(ao->snd_sne), READ_ONCE(tp->snd_una),
net/ipv4/tcp_ao.c
971
state = READ_ONCE(sk->sk_state);
net/ipv4/tcp_ao.c
981
key = READ_ONCE(info->rnext_key);
net/ipv4/tcp_bbr.c
261
rate = min_t(u64, rate, READ_ONCE(sk->sk_max_pacing_rate));
net/ipv4/tcp_bbr.c
294
if (bbr_full_bw_reached(sk) || rate > READ_ONCE(sk->sk_pacing_rate))
net/ipv4/tcp_bbr.c
301
return READ_ONCE(sk->sk_pacing_rate) < (bbr_min_tso_rate >> 3) ? 1 : 2;
net/ipv4/tcp_bbr.c
313
READ_ONCE(sk->sk_pacing_rate) >> READ_ONCE(sk->sk_pacing_shift),
net/ipv4/tcp_cubic.c
379
rate = READ_ONCE(sk->sk_pacing_rate);
net/ipv4/tcp_diag.c
214
r->idiag_state = READ_ONCE(tw->tw_substate);
net/ipv4/tcp_diag.c
251
r->idiag_retrans = READ_ONCE(reqsk->num_retrans);
net/ipv4/tcp_diag.c
256
tmo = READ_ONCE(inet_reqsk(sk)->rsk_timer.expires) - jiffies;
net/ipv4/tcp_diag.c
27
r->idiag_rqueue = READ_ONCE(sk->sk_ack_backlog);
net/ipv4/tcp_diag.c
28
r->idiag_wqueue = READ_ONCE(sk->sk_max_ack_backlog);
net/ipv4/tcp_diag.c
32
r->idiag_rqueue = max_t(int, READ_ONCE(tp->rcv_nxt) -
net/ipv4/tcp_diag.c
33
READ_ONCE(tp->copied_seq), 0);
net/ipv4/tcp_diag.c
34
r->idiag_wqueue = READ_ONCE(tp->write_seq) - tp->snd_una;
net/ipv4/tcp_diag.c
506
READ_ONCE(inet_twsk(sk)->tw_substate) : sk->sk_state;
net/ipv4/tcp_diag.c
512
if (r->id.idiag_sport != htons(READ_ONCE(sk->sk_num)) &&
net/ipv4/tcp_fastopen.c
402
max_qlen = READ_ONCE(fastopenq->max_qlen);
net/ipv4/tcp_fastopen.c
428
return (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen) & flag) ||
net/ipv4/tcp_fastopen.c
443
int tcp_fastopen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen);
net/ipv4/tcp_fastopen.c
586
if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout))
net/ipv4/tcp_fastopen.c
608
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout);
net/ipv4/tcp_fastopen.c
627
timeout = READ_ONCE(sock_net(sk)->ipv4.tfo_active_disable_stamp) +
net/ipv4/tcp_input.c
1156
rate *= READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pacing_ss_ratio);
net/ipv4/tcp_input.c
1158
rate *= READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pacing_ca_ratio);
net/ipv4/tcp_input.c
1170
min_t(u64, rate, READ_ONCE(sk->sk_max_pacing_rate)));
net/ipv4/tcp_input.c
1298
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_max_reordering));
net/ipv4/tcp_input.c
218
if (!dev || len >= READ_ONCE(dev->mtu))
net/ipv4/tcp_input.c
2444
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_max_reordering));
net/ipv4/tcp_input.c
2580
reordering = READ_ONCE(net->ipv4.sysctl_tcp_reordering);
net/ipv4/tcp_input.c
2595
tp->frto = READ_ONCE(net->ipv4.sysctl_tcp_frto) &&
net/ipv4/tcp_input.c
3445
u32 wlen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_min_rtt_wlen) * HZ;
net/ipv4/tcp_input.c
3847
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_reordering))
net/ipv4/tcp_input.c
3993
u32 val = READ_ONCE(*last_oow_ack_time);
net/ipv4/tcp_input.c
3999
elapsed < READ_ONCE(net->ipv4.sysctl_tcp_invalid_ratelimit)) {
net/ipv4/tcp_input.c
4054
ack_limit = READ_ONCE(net->ipv4.sysctl_tcp_challenge_ack_limit);
net/ipv4/tcp_input.c
4060
if (now != READ_ONCE(net->ipv4.tcp_challenge_timestamp)) {
net/ipv4/tcp_input.c
4067
count = READ_ONCE(net->ipv4.tcp_challenge_count);
net/ipv4/tcp_input.c
4224
if ((READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_recovery) &
net/ipv4/tcp_input.c
4358
if (READ_ONCE(sk->sk_err_soft))
net/ipv4/tcp_input.c
4577
!estab && READ_ONCE(net->ipv4.sysctl_tcp_window_scaling)) {
net/ipv4/tcp_input.c
4593
(!estab && READ_ONCE(net->ipv4.sysctl_tcp_timestamps)))) {
net/ipv4/tcp_input.c
4601
!estab && READ_ONCE(net->ipv4.sysctl_tcp_sack)) {
net/ipv4/tcp_input.c
5032
if (tcp_is_sack(tp) && READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_dsack)) {
net/ipv4/tcp_input.c
5101
if (tcp_is_sack(tp) && READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_dsack)) {
net/ipv4/tcp_input.c
5414
READ_ONCE(sk->sk_data_ready)(sk);
net/ipv4/tcp_input.c
5624
READ_ONCE(sk->sk_data_ready)(sk);
net/ipv4/tcp_input.c
5680
READ_ONCE(sk->sk_data_ready)(sk);
net/ipv4/tcp_input.c
6103
INDIRECT_CALL_1(READ_ONCE(sk->sk_write_space),
net/ipv4/tcp_input.c
6158
READ_ONCE(net->ipv4.sysctl_tcp_backlog_ack_defer)) {
net/ipv4/tcp_input.c
6173
tp->compressed_ack >= READ_ONCE(net->ipv4.sysctl_tcp_comp_sack_nr))
net/ipv4/tcp_input.c
6201
READ_ONCE(net->ipv4.sysctl_tcp_comp_sack_rtt_percent);
net/ipv4/tcp_input.c
6203
delay = min(delay, READ_ONCE(net->ipv4.sysctl_tcp_comp_sack_delay_ns));
net/ipv4/tcp_input.c
6207
READ_ONCE(net->ipv4.sysctl_tcp_comp_sack_slack_ns),
net/ipv4/tcp_input.c
6235
if (ptr && !READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_stdurg))
net/ipv4/tcp_input.c
6316
READ_ONCE(sk->sk_data_ready)(sk);
net/ipv4/tcp_input.c
634
min(sndmem, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_wmem[2])));
net/ipv4/tcp_input.c
669
int window = tcp_win_from_space(sk, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])) >> 1;
net/ipv4/tcp_input.c
6792
if (mss == READ_ONCE(tp->rx_opt.user_mss)) {
net/ipv4/tcp_input.c
7006
READ_ONCE(icsk->icsk_accept_queue.rskq_defer_accept) ||
net/ipv4/tcp_input.c
7337
if (READ_ONCE(tp->linger2) < 0) {
net/ipv4/tcp_input.c
742
int tcp_app_win = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_app_win);
net/ipv4/tcp_input.c
7485
(READ_ONCE(net->ipv4.sysctl_tcp_ecn) >= 3 ||
net/ipv4/tcp_input.c
7499
ecn_ok = READ_ONCE(net->ipv4.sysctl_tcp_ecn) || ecn_ok_dst;
net/ipv4/tcp_input.c
7553
syncookies = READ_ONCE(net->ipv4.sysctl_tcp_syncookies);
net/ipv4/tcp_input.c
7564
if (syncookies != 2 && !READ_ONCE(queue->synflood_warned)) {
net/ipv4/tcp_input.c
7620
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies) != 2 &&
net/ipv4/tcp_input.c
7632
mss = tcp_parse_mss_option(th, READ_ONCE(tp->rx_opt.user_mss));
net/ipv4/tcp_input.c
7669
syncookies = READ_ONCE(net->ipv4.sysctl_tcp_syncookies);
net/ipv4/tcp_input.c
7698
tmp_opt.user_mss = READ_ONCE(tp->rx_opt.user_mss);
net/ipv4/tcp_input.c
7727
int max_syn_backlog = READ_ONCE(net->ipv4.sysctl_max_syn_backlog);
net/ipv4/tcp_input.c
7787
READ_ONCE(sk->sk_data_ready)(sk);
net/ipv4/tcp_input.c
786
rmem2 = READ_ONCE(net->ipv4.sysctl_tcp_rmem[2]);
net/ipv4/tcp_input.c
923
if (!READ_ONCE(net->ipv4.sysctl_tcp_moderate_rcvbuf) ||
net/ipv4/tcp_input.c
931
rtt_threshold = READ_ONCE(net->ipv4.sysctl_tcp_rcvbuf_low_rtt);
net/ipv4/tcp_input.c
947
cap = READ_ONCE(net->ipv4.sysctl_tcp_rmem[2]);
net/ipv4/tcp_ipv4.c
1019
inet_twsk(sk)->tw_mark : READ_ONCE(sk->sk_mark);
net/ipv4/tcp_ipv4.c
1021
inet_twsk(sk)->tw_priority : READ_ONCE(sk->sk_priority);
net/ipv4/tcp_ipv4.c
1073
key.sne = READ_ONCE(ao_info->snd_sne);
net/ipv4/tcp_ipv4.c
1074
rnext_key = READ_ONCE(ao_info->rnext_key);
net/ipv4/tcp_ipv4.c
1087
tcptw->tw_snd_nxt, READ_ONCE(tcptw->tw_rcv_nxt),
net/ipv4/tcp_ipv4.c
1090
READ_ONCE(tcptw->tw_ts_recent),
net/ipv4/tcp_ipv4.c
1172
READ_ONCE(tcp_rsk(req)->txhash));
net/ipv4/tcp_ipv4.c
120
int reuse = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_tw_reuse);
net/ipv4/tcp_ipv4.c
1205
tos = READ_ONCE(inet_sk(sk)->tos);
net/ipv4/tcp_ipv4.c
1207
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_reflect_tos))
net/ipv4/tcp_ipv4.c
127
if (READ_ONCE(tw->tw_substate) == TCP_FIN_WAIT2)
net/ipv4/tcp_ipv4.c
167
ts_recent_stamp = READ_ONCE(tcptw->tw_ts_recent_stamp);
net/ipv4/tcp_ipv4.c
168
reuse_thresh = READ_ONCE(tw->tw_entry_stamp) +
net/ipv4/tcp_ipv4.c
169
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_tw_reuse_delay);
net/ipv4/tcp_ipv4.c
1748
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_reflect_tos))
net/ipv4/tcp_ipv4.c
195
tp->rx_opt.ts_recent = READ_ONCE(tcptw->tw_ts_recent);
net/ipv4/tcp_ipv4.c
2088
limit = ((u64)READ_ONCE(sk->sk_rcvbuf)) << 1;
net/ipv4/tcp_ipv4.c
2090
limit += ((u32)READ_ONCE(sk->sk_sndbuf)) >> 1;
net/ipv4/tcp_ipv4.c
2280
if (unlikely(iph->ttl < READ_ONCE(inet_sk(sk)->min_ttl))) {
net/ipv4/tcp_ipv4.c
2881
rx_queue = READ_ONCE(sk->sk_ack_backlog);
net/ipv4/tcp_ipv4.c
2886
rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) -
net/ipv4/tcp_ipv4.c
2887
READ_ONCE(tp->copied_seq), 0);
net/ipv4/tcp_ipv4.c
2892
READ_ONCE(tp->write_seq) - tp->snd_una,
net/ipv4/tcp_ipv4.c
2896
READ_ONCE(icsk->icsk_retransmits),
net/ipv4/tcp_ipv4.c
2898
READ_ONCE(icsk->icsk_probes_out),
net/ipv4/tcp_ipv4.c
2924
i, src, srcp, dest, destp, READ_ONCE(tw->tw_substate), 0, 0,
net/ipv4/tcp_ipv4.c
3486
ehash_entries = READ_ONCE(old_net->ipv4.sysctl_tcp_child_ehash_entries);
net/ipv4/tcp_ipv4.c
380
mtu = READ_ONCE(tcp_sk(sk)->mtu_info);
net/ipv4/tcp_ipv4.c
551
if (unlikely(iph->ttl < READ_ONCE(inet_sk(sk)->min_ttl))) {
net/ipv4/tcp_ipv4.c
898
inet_twsk(sk)->tw_mark : READ_ONCE(sk->sk_mark);
net/ipv4/tcp_ipv4.c
900
inet_twsk(sk)->tw_priority : READ_ONCE(sk->sk_priority);
net/ipv4/tcp_metrics.c
152
limit = READ_ONCE(tm->tcpm_stamp) + TCP_METRICS_TIMEOUT;
net/ipv4/tcp_metrics.c
194
if (time_before(READ_ONCE(tm->tcpm_stamp),
net/ipv4/tcp_metrics.c
195
READ_ONCE(oldest->tcpm_stamp)))
net/ipv4/tcp_metrics.c
352
if (READ_ONCE(net->ipv4.sysctl_tcp_nometrics_save) || !dst)
net/ipv4/tcp_metrics.c
408
if (!READ_ONCE(net->ipv4.sysctl_tcp_no_ssthresh_metrics_save) &&
net/ipv4/tcp_metrics.c
424
if (!READ_ONCE(net->ipv4.sysctl_tcp_no_ssthresh_metrics_save) &&
net/ipv4/tcp_metrics.c
441
if (!READ_ONCE(net->ipv4.sysctl_tcp_no_ssthresh_metrics_save) &&
net/ipv4/tcp_metrics.c
452
READ_ONCE(net->ipv4.sysctl_tcp_reordering))
net/ipv4/tcp_metrics.c
490
val = READ_ONCE(net->ipv4.sysctl_tcp_no_ssthresh_metrics_save) ?
net/ipv4/tcp_metrics.c
57
return READ_ONCE(tm->tcpm_net);
net/ipv4/tcp_metrics.c
64
return READ_ONCE(tm->tcpm_lock) & (1 << idx);
net/ipv4/tcp_metrics.c
672
jiffies - READ_ONCE(tm->tcpm_stamp),
net/ipv4/tcp_metrics.c
71
return READ_ONCE(tm->tcpm_vals[idx]);
net/ipv4/tcp_minisocks.c
1007
READ_ONCE(parent->sk_data_ready)(parent);
net/ipv4/tcp_minisocks.c
106
u32 rcv_nxt = READ_ONCE(tcptw->tw_rcv_nxt);
net/ipv4/tcp_minisocks.c
119
ts_recent_stamp = READ_ONCE(tcptw->tw_ts_recent_stamp);
net/ipv4/tcp_minisocks.c
126
tmp_opt.ts_recent = READ_ONCE(tcptw->tw_ts_recent);
net/ipv4/tcp_minisocks.c
132
if (READ_ONCE(tw->tw_substate) == TCP_FIN_WAIT2) {
net/ipv4/tcp_minisocks.c
216
if (!READ_ONCE(twsk_net(tw)->ipv4.sysctl_tcp_rfc1337)) {
net/ipv4/tcp_minisocks.c
256
(s32)(READ_ONCE(tcptw->tw_ts_recent) - tmp_opt.rcv_tsval) < 0))) {
net/ipv4/tcp_minisocks.c
341
tw->tw_priority = READ_ONCE(sk->sk_priority);
net/ipv4/tcp_minisocks.c
446
window_clamp = READ_ONCE(tp->window_clamp);
net/ipv4/tcp_minisocks.c
593
newsk->sk_txhash = READ_ONCE(treq->txhash);
net/ipv4/tcp_minisocks.c
714
READ_ONCE(tcp_rsk(req)->snt_tsval_last));
net/ipv4/tcp_minisocks.c
914
if (req->num_timeout < READ_ONCE(inet_csk(sk)->icsk_accept_queue.rskq_defer_accept) &&
net/ipv4/tcp_minisocks.c
952
if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_abort_on_overflow)) {
net/ipv4/tcp_output.c
1004
if (likely(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_sack))) {
net/ipv4/tcp_output.c
1046
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn_option) &&
net/ipv4/tcp_output.c
1133
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn_option) &&
net/ipv4/tcp_output.c
1212
int ecn_opt = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn_option);
net/ipv4/tcp_output.c
1444
unsigned long rate = READ_ONCE(sk->sk_pacing_rate);
net/ipv4/tcp_output.c
1603
skb_set_dst_pending_confirm(skb, READ_ONCE(sk->sk_dst_pending_confirm));
net/ipv4/tcp_output.c
1993
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_min_snd_mss));
net/ipv4/tcp_output.c
2026
icsk->icsk_mtup.enabled = READ_ONCE(net->ipv4.sysctl_tcp_mtu_probing) > 1;
net/ipv4/tcp_output.c
2029
icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, READ_ONCE(net->ipv4.sysctl_tcp_base_mss));
net/ipv4/tcp_output.c
2165
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_slow_start_after_idle) &&
net/ipv4/tcp_output.c
2241
bytes = READ_ONCE(sk->sk_pacing_rate) >> READ_ONCE(sk->sk_pacing_shift);
net/ipv4/tcp_output.c
2243
r = tcp_min_rtt(tcp_sk(sk)) >> READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_tso_rtt_log);
net/ipv4/tcp_output.c
2262
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_min_tso_segs);
net/ipv4/tcp_output.c
235
u32 window_clamp = READ_ONCE(*__window_clamp);
net/ipv4/tcp_output.c
2476
win_divisor = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_tso_win_divisor);
net/ipv4/tcp_output.c
254
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_workaround_signed_windows))
net/ipv4/tcp_output.c
2552
interval = READ_ONCE(net->ipv4.sysctl_tcp_probe_interval);
net/ipv4/tcp_output.c
265
space = max_t(u32, space, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2]));
net/ipv4/tcp_output.c
266
space = max_t(u32, space, READ_ONCE(sysctl_rmem_max));
net/ipv4/tcp_output.c
2700
interval < READ_ONCE(net->ipv4.sysctl_tcp_probe_threshold)) {
net/ipv4/tcp_output.c
2842
READ_ONCE(sk->sk_pacing_rate) >> READ_ONCE(sk->sk_pacing_shift));
net/ipv4/tcp_output.c
2844
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_limit_output_bytes));
net/ipv4/tcp_output.c
2849
u64 extra_bytes = (u64)READ_ONCE(sk->sk_pacing_rate) *
net/ipv4/tcp_output.c
309
if (!READ_ONCE(net->ipv4.sysctl_tcp_shrink_window) || !tp->rx_opt.rcv_wscale) {
net/ipv4/tcp_output.c
3114
early_retrans = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_early_retrans);
net/ipv4/tcp_output.c
324
READ_ONCE(net->ipv4.sysctl_tcp_workaround_signed_windows))
net/ipv4/tcp_output.c
3346
if (READ_ONCE(net->ipv4.sysctl_tcp_shrink_window) && tp->rx_opt.rcv_wscale)
net/ipv4/tcp_output.c
3518
if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_retrans_collapse))
net/ipv4/tcp_output.c
4028
skb_set_hash(skb, READ_ONCE(tcp_rsk(req)->txhash), PKT_HASH_TYPE_L4);
net/ipv4/tcp_output.c
4116
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_timestamps))
net/ipv4/tcp_output.c
4122
user_mss = READ_ONCE(tp->rx_opt.user_mss);
net/ipv4/tcp_output.c
4150
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_window_scaling),
net/ipv4/tcp_output.c
4404
return min(READ_ONCE(inet_csk(sk)->icsk_delack_max), delack_from_rto_min);
net/ipv4/tcp_output.c
4622
if (icsk->icsk_backoff < READ_ONCE(net->ipv4.sysctl_tcp_retries2))
net/ipv4/tcp_output.c
4643
if (READ_ONCE(sk->sk_txrehash) == SOCK_TXREHASH_ENABLED)
net/ipv4/tcp_output.c
617
rnext_key = READ_ONCE(ao_info->rnext_key);
net/ipv4/tcp_output.c
974
timestamps = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_timestamps);
net/ipv4/tcp_output.c
999
if (likely(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_window_scaling))) {
net/ipv4/tcp_plb.c
30
if (!READ_ONCE(net->ipv4.sysctl_tcp_plb_enabled))
net/ipv4/tcp_plb.c
34
if (cong_ratio < READ_ONCE(net->ipv4.sysctl_tcp_plb_cong_thresh))
net/ipv4/tcp_plb.c
37
READ_ONCE(net->ipv4.sysctl_tcp_plb_rehash_rounds))
net/ipv4/tcp_plb.c
52
if (!READ_ONCE(net->ipv4.sysctl_tcp_plb_enabled))
net/ipv4/tcp_plb.c
56
READ_ONCE(net->ipv4.sysctl_tcp_plb_rehash_rounds);
net/ipv4/tcp_plb.c
58
idle_rehash = READ_ONCE(net->ipv4.sysctl_tcp_plb_idle_rehash_rounds) &&
net/ipv4/tcp_plb.c
61
READ_ONCE(net->ipv4.sysctl_tcp_plb_idle_rehash_rounds);
net/ipv4/tcp_plb.c
71
max_suspend = 2 * READ_ONCE(net->ipv4.sysctl_tcp_plb_suspend_rto_sec) * HZ;
net/ipv4/tcp_plb.c
96
if (!READ_ONCE(net->ipv4.sysctl_tcp_plb_enabled))
net/ipv4/tcp_plb.c
99
pause = READ_ONCE(net->ipv4.sysctl_tcp_plb_suspend_rto_sec) * HZ;
net/ipv4/tcp_recovery.c
17
!(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_recovery) &
net/ipv4/tcp_timer.c
117
if (READ_ONCE(sk->sk_err_soft))
net/ipv4/tcp_timer.c
151
int retries = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_orphan_retries); /* May be zero. */
net/ipv4/tcp_timer.c
154
if (READ_ONCE(sk->sk_err_soft) && !alive)
net/ipv4/tcp_timer.c
171
if (!READ_ONCE(net->ipv4.sysctl_tcp_mtu_probing))
net/ipv4/tcp_timer.c
179
mss = min(READ_ONCE(net->ipv4.sysctl_tcp_base_mss), mss);
net/ipv4/tcp_timer.c
180
mss = max(mss, READ_ONCE(net->ipv4.sysctl_tcp_mtu_probe_floor));
net/ipv4/tcp_timer.c
181
mss = max(mss, READ_ONCE(net->ipv4.sysctl_tcp_min_snd_mss));
net/ipv4/tcp_timer.c
254
retry_until = READ_ONCE(icsk->icsk_syn_retries) ? :
net/ipv4/tcp_timer.c
255
READ_ONCE(net->ipv4.sysctl_tcp_syn_retries);
net/ipv4/tcp_timer.c
259
max_retransmits += READ_ONCE(net->ipv4.sysctl_tcp_syn_linear_timeouts);
net/ipv4/tcp_timer.c
263
if (retransmits_timed_out(sk, READ_ONCE(net->ipv4.sysctl_tcp_retries1), 0)) {
net/ipv4/tcp_timer.c
270
retry_until = READ_ONCE(net->ipv4.sysctl_tcp_retries2);
net/ipv4/tcp_timer.c
284
READ_ONCE(icsk->icsk_user_timeout));
net/ipv4/tcp_timer.c
35
user_timeout = READ_ONCE(icsk->icsk_user_timeout);
net/ipv4/tcp_timer.c
371
!READ_ONCE(tcp_sk(sk)->compressed_ack))
net/ipv4/tcp_timer.c
412
u32 user_timeout = READ_ONCE(icsk->icsk_user_timeout);
net/ipv4/tcp_timer.c
419
max_probes = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_retries2);
net/ipv4/tcp_timer.c
467
max_retries = READ_ONCE(icsk->icsk_syn_retries) ? :
net/ipv4/tcp_timer.c
468
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_synack_retries) + 1;
net/ipv4/tcp_timer.c
498
u32 user_timeout = READ_ONCE(icsk->icsk_user_timeout);
net/ipv4/tcp_timer.c
56
user_timeout = READ_ONCE(icsk->icsk_user_timeout);
net/ipv4/tcp_timer.c
667
(tp->thin_lto || READ_ONCE(net->ipv4.sysctl_tcp_thin_linear_timeouts)) &&
net/ipv4/tcp_timer.c
676
READ_ONCE(net->ipv4.sysctl_tcp_syn_linear_timeouts)) {
net/ipv4/tcp_timer.c
685
if (retransmits_timed_out(sk, READ_ONCE(net->ipv4.sysctl_tcp_retries1) + 1, 0))
net/ipv4/tcp_timer.c
78
tcp_done_with_error(sk, READ_ONCE(sk->sk_err_soft) ? : ETIMEDOUT);
net/ipv4/tcp_timer.c
802
if (READ_ONCE(tp->linger2) >= 0) {
net/ipv4/tcp_timer.c
827
u32 user_timeout = READ_ONCE(icsk->icsk_user_timeout);
net/ipv4/udp.c
1306
if (READ_ONCE(up->pending)) {
net/ipv4/udp.c
1350
ipc.gso_size = READ_ONCE(up->gso_size);
net/ipv4/udp.c
1412
uc_index = READ_ONCE(inet->uc_index);
net/ipv4/udp.c
1415
ipc.oif = READ_ONCE(inet->mc_index);
net/ipv4/udp.c
1417
saddr = READ_ONCE(inet->mc_addr);
net/ipv4/udp.c
1557
if (!READ_ONCE(up->pending) || udp_test_bit(CORK, sk))
net/ipv4/udp.c
1640
if (size < READ_ONCE(up->forward_threshold) &&
net/ipv4/udp.c
1714
rcvbuf = READ_ONCE(sk->sk_rcvbuf);
net/ipv4/udp.c
1790
INDIRECT_CALL_1(READ_ONCE(sk->sk_data_ready),
net/ipv4/udp.c
1869
if (unlikely(READ_ONCE(udp_sk(sk)->peeking_with_offset)))
net/ipv4/udp.c
2418
READ_ONCE(up->encap_type)) {
net/ipv4/udp.c
2433
encap_rcv = READ_ONCE(up->encap_rcv);
net/ipv4/udp.c
2458
u16 pcrlen = READ_ONCE(up->pcrlen);
net/ipv4/udp.c
2951
encap_destroy = READ_ONCE(up->encap_destroy);
net/ipv4/udp.c
3159
val = READ_ONCE(up->encap_type);
net/ipv4/udp.c
3171
val = READ_ONCE(up->gso_size);
net/ipv4/udp.c
3181
val = READ_ONCE(up->pcslen);
net/ipv4/udp.c
3185
val = READ_ONCE(up->pcrlen);
net/ipv4/udp.c
3922
hash_entries = READ_ONCE(old_net->ipv4.sysctl_udp_child_hash_entries);
net/ipv4/udp.c
407
if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
net/ipv4/udp.c
892
lookup = READ_ONCE(up->encap_err_lookup);
net/ipv4/udp.c
905
lookup = READ_ONCE(up->encap_err_lookup);
net/ipv4/udp.c
948
if (!sk || READ_ONCE(udp_sk(sk)->encap_type)) {
net/ipv4/udp.c
984
if (READ_ONCE(inet->pmtudisc) != IP_PMTUDISC_DONT) {
net/ipv4/xfrm4_input.c
91
encap_type = READ_ONCE(up->encap_type);
net/ipv6/addrconf.c
1353
return READ_ONCE(idev->cnf.regen_min_advance) +
net/ipv6/addrconf.c
1354
READ_ONCE(idev->cnf.regen_max_retry) *
net/ipv6/addrconf.c
1355
READ_ONCE(idev->cnf.dad_transmits) *
net/ipv6/addrconf.c
1377
if (READ_ONCE(idev->cnf.use_tempaddr) <= 0) {
net/ipv6/addrconf.c
1385
if (ifp->regen_count++ >= READ_ONCE(idev->cnf.regen_max_retry)) {
net/ipv6/addrconf.c
1406
cnf_temp_preferred_lft = READ_ONCE(idev->cnf.temp_prefered_lft);
net/ipv6/addrconf.c
1408
READ_ONCE(idev->cnf.max_desync_factor),
net/ipv6/addrconf.c
1425
READ_ONCE(idev->cnf.temp_valid_lft) + age);
net/ipv6/addrconf.c
1555
if (!READ_ONCE(net->ipv6.devconf_all->optimistic_dad) &&
net/ipv6/addrconf.c
1556
!READ_ONCE(idev->cnf.optimistic_dad))
net/ipv6/addrconf.c
1558
if (!READ_ONCE(net->ipv6.devconf_all->use_optimistic) &&
net/ipv6/addrconf.c
1559
!READ_ONCE(idev->cnf.use_optimistic))
net/ipv6/addrconf.c
1574
if (!READ_ONCE(net->ipv6.devconf_all->optimistic_dad) &&
net/ipv6/addrconf.c
1575
!READ_ONCE(idev->cnf.optimistic_dad))
net/ipv6/addrconf.c
1681
READ_ONCE(score->ifa->idev->cnf.use_tempaddr) >= 2;
net/ipv6/addrconf.c
1857
(idev && READ_ONCE(idev->cnf.use_oif_addrs_only))) {
net/ipv6/addrconf.c
2204
max_addresses = READ_ONCE(idev->cnf.max_addresses);
net/ipv6/addrconf.c
2241
if (READ_ONCE(dev->flags) & (IFF_LOOPBACK | IFF_NOARP))
net/ipv6/addrconf.c
2252
if (READ_ONCE(idev->dev->flags) & (IFF_LOOPBACK | IFF_NOARP))
net/ipv6/addrconf.c
2617
max_valid = READ_ONCE(idev->cnf.temp_valid_lft) - age;
net/ipv6/addrconf.c
2621
max_prefered = READ_ONCE(idev->cnf.temp_prefered_lft) -
net/ipv6/addrconf.c
2654
if (create && READ_ONCE(idev->cnf.use_tempaddr) > 0) {
net/ipv6/addrconf.c
2682
int max_addresses = READ_ONCE(in6_dev->cnf.max_addresses);
net/ipv6/addrconf.c
2694
if ((READ_ONCE(net->ipv6.devconf_all->optimistic_dad) ||
net/ipv6/addrconf.c
2695
READ_ONCE(in6_dev->cnf.optimistic_dad)) &&
net/ipv6/addrconf.c
2744
if (update_lft && !READ_ONCE(in6_dev->cnf.ra_honor_pio_life)) {
net/ipv6/addrconf.c
2887
ignore_autoconf = READ_ONCE(in6_dev->cnf.ra_honor_pio_pflag) && pinfo->preferpd;
net/ipv6/addrconf.c
3304
if ((READ_ONCE(dev_net(idev->dev)->ipv6.devconf_all->optimistic_dad) ||
net/ipv6/addrconf.c
3305
READ_ONCE(idev->cnf.optimistic_dad)) &&
net/ipv6/addrconf.c
3883
int _keep_addr = READ_ONCE(net->ipv6.devconf_all->keep_addr_on_down);
net/ipv6/addrconf.c
3886
_keep_addr = READ_ONCE(idev->cnf.keep_addr_on_down);
net/ipv6/addrconf.c
4036
rtr_solicits = READ_ONCE(idev->cnf.rtr_solicits);
net/ipv6/addrconf.c
4049
READ_ONCE(idev->cnf.rtr_solicit_max_interval));
net/ipv6/addrconf.c
4052
READ_ONCE(idev->cnf.rtr_solicits)) ?
net/ipv6/addrconf.c
4053
READ_ONCE(idev->cnf.rtr_solicit_delay) :
net/ipv6/addrconf.c
4082
READ_ONCE(idev->cnf.rtr_solicit_delay) ? : 1);
net/ipv6/addrconf.c
4085
if (READ_ONCE(idev->cnf.enhanced_dad) ||
net/ipv6/addrconf.c
4086
READ_ONCE(dev_net(idev->dev)->ipv6.devconf_all->enhanced_dad)) {
net/ipv6/addrconf.c
4092
ifp->dad_probes = READ_ONCE(idev->cnf.dad_transmits);
net/ipv6/addrconf.c
4112
(READ_ONCE(net->ipv6.devconf_all->accept_dad) < 1 &&
net/ipv6/addrconf.c
4113
READ_ONCE(idev->cnf.accept_dad) < 1) ||
net/ipv6/addrconf.c
4208
if ((READ_ONCE(net->ipv6.devconf_all->accept_dad) > 1 ||
net/ipv6/addrconf.c
4209
READ_ONCE(idev->cnf.accept_dad) > 1) &&
net/ipv6/addrconf.c
4334
READ_ONCE(ifp->idev->cnf.rtr_solicits) != 0 &&
net/ipv6/addrconf.c
4348
(READ_ONCE(ifp->idev->cnf.ndisc_notify) ||
net/ipv6/addrconf.c
4349
READ_ONCE(dev_net(dev)->ipv6.devconf_all->ndisc_notify))) {
net/ipv6/addrconf.c
4369
READ_ONCE(ifp->idev->cnf.rtr_solicit_interval));
net/ipv6/addrconf.c
5147
flags = READ_ONCE(ifa->flags);
net/ipv6/addrconf.c
5155
preferred = READ_ONCE(ifa->prefered_lft);
net/ipv6/addrconf.c
5156
valid = READ_ONCE(ifa->valid_lft);
net/ipv6/addrconf.c
5161
long tval = (jiffies - READ_ONCE(ifa->tstamp)) / HZ;
net/ipv6/addrconf.c
5188
priority = READ_ONCE(ifa->rt_priority);
net/ipv6/addrconf.c
5192
if (put_cacheinfo(skb, ifa->cstamp, READ_ONCE(ifa->tstamp),
net/ipv6/addrconf.c
5199
proto = READ_ONCE(ifa->ifa_proto);
net/ipv6/addrconf.c
5236
put_cacheinfo(skb, ifmca->mca_cstamp, READ_ONCE(ifmca->mca_tstamp),
net/ipv6/addrconf.c
5271
put_cacheinfo(skb, ifaca->aca_cstamp, READ_ONCE(ifaca->aca_tstamp),
net/ipv6/addrconf.c
559
READ_ONCE(devconf->forwarding)) < 0)
net/ipv6/addrconf.c
5637
array[DEVCONF_FORWARDING] = READ_ONCE(cnf->forwarding);
net/ipv6/addrconf.c
5638
array[DEVCONF_HOPLIMIT] = READ_ONCE(cnf->hop_limit);
net/ipv6/addrconf.c
5639
array[DEVCONF_MTU6] = READ_ONCE(cnf->mtu6);
net/ipv6/addrconf.c
5640
array[DEVCONF_ACCEPT_RA] = READ_ONCE(cnf->accept_ra);
net/ipv6/addrconf.c
5641
array[DEVCONF_ACCEPT_REDIRECTS] = READ_ONCE(cnf->accept_redirects);
net/ipv6/addrconf.c
5642
array[DEVCONF_AUTOCONF] = READ_ONCE(cnf->autoconf);
net/ipv6/addrconf.c
5643
array[DEVCONF_DAD_TRANSMITS] = READ_ONCE(cnf->dad_transmits);
net/ipv6/addrconf.c
5644
array[DEVCONF_RTR_SOLICITS] = READ_ONCE(cnf->rtr_solicits);
net/ipv6/addrconf.c
5646
jiffies_to_msecs(READ_ONCE(cnf->rtr_solicit_interval));
net/ipv6/addrconf.c
5648
jiffies_to_msecs(READ_ONCE(cnf->rtr_solicit_max_interval));
net/ipv6/addrconf.c
5650
jiffies_to_msecs(READ_ONCE(cnf->rtr_solicit_delay));
net/ipv6/addrconf.c
5651
array[DEVCONF_FORCE_MLD_VERSION] = READ_ONCE(cnf->force_mld_version);
net/ipv6/addrconf.c
5653
jiffies_to_msecs(READ_ONCE(cnf->mldv1_unsolicited_report_interval));
net/ipv6/addrconf.c
5655
jiffies_to_msecs(READ_ONCE(cnf->mldv2_unsolicited_report_interval));
net/ipv6/addrconf.c
5656
array[DEVCONF_USE_TEMPADDR] = READ_ONCE(cnf->use_tempaddr);
net/ipv6/addrconf.c
5657
array[DEVCONF_TEMP_VALID_LFT] = READ_ONCE(cnf->temp_valid_lft);
net/ipv6/addrconf.c
5658
array[DEVCONF_TEMP_PREFERED_LFT] = READ_ONCE(cnf->temp_prefered_lft);
net/ipv6/addrconf.c
5659
array[DEVCONF_REGEN_MAX_RETRY] = READ_ONCE(cnf->regen_max_retry);
net/ipv6/addrconf.c
5660
array[DEVCONF_MAX_DESYNC_FACTOR] = READ_ONCE(cnf->max_desync_factor);
net/ipv6/addrconf.c
5661
array[DEVCONF_MAX_ADDRESSES] = READ_ONCE(cnf->max_addresses);
net/ipv6/addrconf.c
5662
array[DEVCONF_ACCEPT_RA_DEFRTR] = READ_ONCE(cnf->accept_ra_defrtr);
net/ipv6/addrconf.c
5663
array[DEVCONF_RA_DEFRTR_METRIC] = READ_ONCE(cnf->ra_defrtr_metric);
net/ipv6/addrconf.c
5665
READ_ONCE(cnf->accept_ra_min_hop_limit);
net/ipv6/addrconf.c
5666
array[DEVCONF_ACCEPT_RA_PINFO] = READ_ONCE(cnf->accept_ra_pinfo);
net/ipv6/addrconf.c
5668
array[DEVCONF_ACCEPT_RA_RTR_PREF] = READ_ONCE(cnf->accept_ra_rtr_pref);
net/ipv6/addrconf.c
5670
jiffies_to_msecs(READ_ONCE(cnf->rtr_probe_interval));
net/ipv6/addrconf.c
5673
READ_ONCE(cnf->accept_ra_rt_info_min_plen);
net/ipv6/addrconf.c
5675
READ_ONCE(cnf->accept_ra_rt_info_max_plen);
net/ipv6/addrconf.c
5678
array[DEVCONF_PROXY_NDP] = READ_ONCE(cnf->proxy_ndp);
net/ipv6/addrconf.c
5680
READ_ONCE(cnf->accept_source_route);
net/ipv6/addrconf.c
5682
array[DEVCONF_OPTIMISTIC_DAD] = READ_ONCE(cnf->optimistic_dad);
net/ipv6/addrconf.c
5683
array[DEVCONF_USE_OPTIMISTIC] = READ_ONCE(cnf->use_optimistic);
net/ipv6/addrconf.c
5688
array[DEVCONF_DISABLE_IPV6] = READ_ONCE(cnf->disable_ipv6);
net/ipv6/addrconf.c
5689
array[DEVCONF_ACCEPT_DAD] = READ_ONCE(cnf->accept_dad);
net/ipv6/addrconf.c
569
READ_ONCE(devconf->proxy_ndp)) < 0)
net/ipv6/addrconf.c
5690
array[DEVCONF_FORCE_TLLAO] = READ_ONCE(cnf->force_tllao);
net/ipv6/addrconf.c
5691
array[DEVCONF_NDISC_NOTIFY] = READ_ONCE(cnf->ndisc_notify);
net/ipv6/addrconf.c
5693
READ_ONCE(cnf->suppress_frag_ndisc);
net/ipv6/addrconf.c
5695
READ_ONCE(cnf->accept_ra_from_local);
net/ipv6/addrconf.c
5696
array[DEVCONF_ACCEPT_RA_MTU] = READ_ONCE(cnf->accept_ra_mtu);
net/ipv6/addrconf.c
5698
READ_ONCE(cnf->ignore_routes_with_linkdown);
net/ipv6/addrconf.c
5700
array[DEVCONF_USE_OIF_ADDRS_ONLY] = READ_ONCE(cnf->use_oif_addrs_only);
net/ipv6/addrconf.c
5702
READ_ONCE(cnf->drop_unicast_in_l2_multicast);
net/ipv6/addrconf.c
5703
array[DEVCONF_DROP_UNSOLICITED_NA] = READ_ONCE(cnf->drop_unsolicited_na);
net/ipv6/addrconf.c
5704
array[DEVCONF_KEEP_ADDR_ON_DOWN] = READ_ONCE(cnf->keep_addr_on_down);
net/ipv6/addrconf.c
5705
array[DEVCONF_SEG6_ENABLED] = READ_ONCE(cnf->seg6_enabled);
net/ipv6/addrconf.c
5707
array[DEVCONF_SEG6_REQUIRE_HMAC] = READ_ONCE(cnf->seg6_require_hmac);
net/ipv6/addrconf.c
5709
array[DEVCONF_ENHANCED_DAD] = READ_ONCE(cnf->enhanced_dad);
net/ipv6/addrconf.c
5710
array[DEVCONF_ADDR_GEN_MODE] = READ_ONCE(cnf->addr_gen_mode);
net/ipv6/addrconf.c
5711
array[DEVCONF_DISABLE_POLICY] = READ_ONCE(cnf->disable_policy);
net/ipv6/addrconf.c
5712
array[DEVCONF_NDISC_TCLASS] = READ_ONCE(cnf->ndisc_tclass);
net/ipv6/addrconf.c
5713
array[DEVCONF_RPL_SEG_ENABLED] = READ_ONCE(cnf->rpl_seg_enabled);
net/ipv6/addrconf.c
5714
array[DEVCONF_IOAM6_ENABLED] = READ_ONCE(cnf->ioam6_enabled);
net/ipv6/addrconf.c
5715
array[DEVCONF_IOAM6_ID] = READ_ONCE(cnf->ioam6_id);
net/ipv6/addrconf.c
5716
array[DEVCONF_IOAM6_ID_WIDE] = READ_ONCE(cnf->ioam6_id_wide);
net/ipv6/addrconf.c
5718
READ_ONCE(cnf->ndisc_evict_nocarrier);
net/ipv6/addrconf.c
5720
READ_ONCE(cnf->accept_untracked_na);
net/ipv6/addrconf.c
5721
array[DEVCONF_ACCEPT_RA_MIN_LFT] = READ_ONCE(cnf->accept_ra_min_lft);
net/ipv6/addrconf.c
5722
array[DEVCONF_FORCE_FORWARDING] = READ_ONCE(cnf->force_forwarding);
net/ipv6/addrconf.c
574
READ_ONCE(devconf->ignore_routes_with_linkdown)) < 0)
net/ipv6/addrconf.c
5827
if (nla_put_u32(skb, IFLA_INET6_FLAGS, READ_ONCE(idev->if_flags)))
net/ipv6/addrconf.c
5830
ci.tstamp = cstamp_delta(READ_ONCE(idev->tstamp));
net/ipv6/addrconf.c
5855
READ_ONCE(idev->cnf.addr_gen_mode)))
net/ipv6/addrconf.c
5858
ra_mtu = READ_ONCE(idev->ra_mtu);
net/ipv6/addrconf.c
5921
if (READ_ONCE(idev->cnf.rtr_solicits) == 0) {
net/ipv6/addrconf.c
5954
READ_ONCE(idev->cnf.rtr_solicit_interval));
net/ipv6/addrconf.c
6083
ifindex = READ_ONCE(dev->ifindex);
net/ipv6/addrconf.c
6092
nla_put_u32(skb, IFLA_MTU, READ_ONCE(dev->mtu)) ||
net/ipv6/addrconf.c
6096
netif_running(dev) ? READ_ONCE(dev->operstate) : IF_OPER_DOWN))
net/ipv6/addrconf.c
724
READ_ONCE(net->dev_base_seq);
net/ipv6/addrlabel.c
502
lseq = READ_ONCE(net->ipv6.ip6addrlbl_table.seq);
net/ipv6/addrlabel.c
610
lseq = READ_ONCE(net->ipv6.ip6addrlbl_table.seq);
net/ipv6/af_inet6.c
219
inet6_assign_bit(REPFLOW, sk, READ_ONCE(net->ipv6.sysctl.flowlabel_reflect) &
net/ipv6/af_inet6.c
222
sk->sk_txrehash = READ_ONCE(net->core.sysctl_txrehash);
net/ipv6/af_inet6.c
235
if (READ_ONCE(net->ipv4.sysctl_ip_no_pmtu_disc))
net/ipv6/af_inet6.c
440
prot = READ_ONCE(sk->sk_prot);
net/ipv6/af_inet6.c
579
prot = READ_ONCE(sk->sk_prot);
net/ipv6/af_inet6.c
650
prot = READ_ONCE(sk->sk_prot);
net/ipv6/af_inet6.c
669
prot = READ_ONCE(sk->sk_prot);
net/ipv6/anycast.c
137
ishost = !READ_ONCE(idev->cnf.forwarding);
net/ipv6/anycast.c
99
ishost = !READ_ONCE(net->ipv6.devconf_all->forwarding);
net/ipv6/datagram.c
131
if (!dst || !READ_ONCE(dst->obsolete) ||
net/ipv6/datagram.c
233
WRITE_ONCE(sk->sk_bound_dev_if, READ_ONCE(np->mcast_oif));
net/ipv6/datagram.c
63
oif = READ_ONCE(np->mcast_oif);
net/ipv6/datagram.c
65
oif = READ_ONCE(np->ucast_oif);
net/ipv6/datagram.c
809
(READ_ONCE(sk->sk_bound_dev_if) != fl6->flowi6_oif ||
net/ipv6/exthdrs.c
1062
if (extlen > READ_ONCE(net->ipv6.sysctl.max_hbh_opts_len))
net/ipv6/exthdrs.c
1067
READ_ONCE(net->ipv6.sysctl.max_hbh_opts_cnt))) {
net/ipv6/exthdrs.c
317
if (extlen > READ_ONCE(net->ipv6.sysctl.max_dst_opts_len))
net/ipv6/exthdrs.c
326
READ_ONCE(net->ipv6.sysctl.max_dst_opts_cnt))) {
net/ipv6/exthdrs.c
387
accept_seg6 = min(READ_ONCE(net->ipv6.devconf_all->seg6_enabled),
net/ipv6/exthdrs.c
388
READ_ONCE(idev->cnf.seg6_enabled));
net/ipv6/exthdrs.c
502
accept_rpl_seg = min(READ_ONCE(net->ipv6.devconf_all->rpl_seg_enabled),
net/ipv6/exthdrs.c
503
READ_ONCE(idev->cnf.rpl_seg_enabled));
net/ipv6/exthdrs.c
662
accept_source_route = READ_ONCE(net->ipv6.devconf_all->accept_source_route);
net/ipv6/exthdrs.c
666
READ_ONCE(idev->cnf.accept_source_route));
net/ipv6/exthdrs.c
919
if (!READ_ONCE(__in6_dev_get(skb->dev)->cnf.ioam6_enabled))
net/ipv6/icmp.c
1163
READ_ONCE(net->ipv4.sysctl_icmp_echo_enable_probe))
net/ipv6/icmp.c
220
int tmo = READ_ONCE(net->ipv6.sysctl.icmpv6_time);
net/ipv6/icmp.c
371
if (!READ_ONCE(net->ipv6.sysctl.icmpv6_error_anycast_as_unicast) &&
net/ipv6/icmp.c
545
data = htonl(READ_ONCE(dev->mtu));
net/ipv6/icmp.c
587
ext_objs = READ_ONCE(net->ipv6.sysctl.icmpv6_errors_extension_mask);
net/ipv6/icmp.c
785
fl6.flowi6_oif = READ_ONCE(np->mcast_oif);
net/ipv6/icmp.c
787
fl6.flowi6_oif = READ_ONCE(np->ucast_oif);
net/ipv6/icmp.c
963
if (READ_ONCE(net->ipv6.sysctl.flowlabel_reflect) &
net/ipv6/icmp.c
986
fl6.flowi6_oif = READ_ONCE(np->mcast_oif);
net/ipv6/icmp.c
988
fl6.flowi6_oif = READ_ONCE(np->ucast_oif);
net/ipv6/ila/ila_xlat.c
207
if (!READ_ONCE(ilan->xlat.hooks_registered)) {
net/ipv6/inet6_connection_sock.c
116
np->tclass, READ_ONCE(sk->sk_priority));
net/ipv6/inet6_hashtables.c
108
if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
net/ipv6/inet6_hashtables.c
99
READ_ONCE(inet_sk(sk)->inet_num) == hnum &&
net/ipv6/ioam6.c
741
raw16 = (__force u16)READ_ONCE(__in6_dev_get(skb->dev)->cnf.ioam6_id);
net/ipv6/ioam6.c
749
raw16 = (__force u16)READ_ONCE(__in6_dev_get(dev)->cnf.ioam6_id);
net/ipv6/ioam6.c
836
raw32 = READ_ONCE(__in6_dev_get(skb->dev)->cnf.ioam6_id_wide);
net/ipv6/ioam6.c
844
raw32 = READ_ONCE(__in6_dev_get(dev)->cnf.ioam6_id_wide);
net/ipv6/ip6_fib.c
1386
jiffies + READ_ONCE(net->ipv6.sysctl.ip6_rt_gc_interval));
net/ipv6/ip6_fib.c
1393
jiffies + READ_ONCE(net->ipv6.sysctl.ip6_rt_gc_interval));
net/ipv6/ip6_fib.c
2249
READ_ONCE(w->node->fn_sernum) != c->sernum)
net/ipv6/ip6_fib.c
2445
ip6_rt_gc_interval = READ_ONCE(net->ipv6.sysctl.ip6_rt_gc_interval);
net/ipv6/ip6_fib.c
2669
iter->sernum = READ_ONCE(iter->w.root->fn_sernum);
net/ipv6/ip6_fib.c
2697
int sernum = READ_ONCE(iter->w.root->fn_sernum);
net/ipv6/ip6_fib.c
363
fib_seq += READ_ONCE(tb->fib_seq);
net/ipv6/ip6_fib.c
454
nsiblings = READ_ONCE(rt->fib6_nsiblings);
net/ipv6/ip6_fib.c
602
cb->args[5] = READ_ONCE(w->root->fn_sernum);
net/ipv6/ip6_fib.c
605
int sernum = READ_ONCE(w->root->fn_sernum);
net/ipv6/ip6_fib.c
735
if (READ_ONCE(f6i->fib6_metrics) == &dst_default_metrics) {
net/ipv6/ip6_fib.c
750
m = READ_ONCE(f6i->fib6_metrics);
net/ipv6/ip6_fib.c
994
pcpu_rt = READ_ONCE(*ppcpu_rt);
net/ipv6/ip6_gre.c
1059
mtu = READ_ONCE(dst_dev(dst)->mtu);
net/ipv6/ip6_gre.c
211
ndev = READ_ONCE(ign->fb_tunnel_dev);
net/ipv6/ip6_icmp.c
62
if (!ct || !(READ_ONCE(ct->status) & IPS_NAT_MASK)) {
net/ipv6/ip6_input.c
170
!idev || unlikely(READ_ONCE(idev->cnf.disable_ipv6))) {
net/ipv6/ip6_input.c
172
if (idev && unlikely(READ_ONCE(idev->cnf.disable_ipv6)))
net/ipv6/ip6_input.c
240
READ_ONCE(idev->cnf.drop_unicast_in_l2_multicast)) {
net/ipv6/ip6_input.c
51
if (READ_ONCE(net->ipv4.sysctl_ip_early_demux) &&
net/ipv6/ip6_input.c
55
if (READ_ONCE(net->ipv4.sysctl_tcp_early_demux))
net/ipv6/ip6_input.c
59
if (READ_ONCE(net->ipv4.sysctl_udp_early_demux))
net/ipv6/ip6_output.c
1134
sk ? READ_ONCE(inet6_sk(sk)->srcprefs) : 0,
net/ipv6/ip6_output.c
1175
err = n && !(READ_ONCE(n->nud_state) & NUD_VALID) ? -EINVAL : 0;
net/ipv6/ip6_output.c
1392
mtu = READ_ONCE(np->pmtudisc) >= IPV6_PMTUDISC_PROBE ?
net/ipv6/ip6_output.c
1393
READ_ONCE(rt->dst.dev->mtu) : dst6_mtu(&rt->dst);
net/ipv6/ip6_output.c
1395
mtu = READ_ONCE(np->pmtudisc) >= IPV6_PMTUDISC_PROBE ?
net/ipv6/ip6_output.c
1396
READ_ONCE(rt->dst.dev->mtu) : dst6_mtu(xfrm_dst_path(&rt->dst));
net/ipv6/ip6_output.c
1398
frag_size = READ_ONCE(np->frag_size);
net/ipv6/ip6_output.c
1554
READ_ONCE(sk->sk_tsflags) & SOF_TIMESTAMPING_OPT_ID) {
net/ipv6/ip6_output.c
239
if (unlikely(!idev || READ_ONCE(idev->cnf.disable_ipv6))) {
net/ipv6/ip6_output.c
324
hlimit = READ_ONCE(np->hop_limit);
net/ipv6/ip6_output.c
509
if (!READ_ONCE(net->ipv6.devconf_all->forwarding) &&
net/ipv6/ip6_output.c
510
(!idev || !READ_ONCE(idev->cnf.force_forwarding)))
net/ipv6/ip6_output.c
522
if (!READ_ONCE(net->ipv6.devconf_all->disable_policy) &&
net/ipv6/ip6_output.c
523
(!idev || !READ_ONCE(idev->cnf.disable_policy)) &&
net/ipv6/ip6_output.c
561
if (READ_ONCE(net->ipv6.devconf_all->proxy_ndp) &&
net/ipv6/ip6_output.c
902
u32 frag_size = READ_ONCE(np->frag_size);
net/ipv6/ip6_tunnel.c
1311
tproto = READ_ONCE(t->parms.proto);
net/ipv6/ip6_tunnel.c
1779
return READ_ONCE(t->parms.link);
net/ipv6/ip6_tunnel.c
2221
return READ_ONCE(tunnel->net);
net/ipv6/ip6_tunnel.c
486
tproto = READ_ONCE(t->parms.proto);
net/ipv6/ip6_tunnel.c
943
u8 tproto = READ_ONCE(t->parms.proto);
net/ipv6/ip6mr.c
1104
msg->im6_mif = READ_ONCE(mrt->mroute_reg_vif_num);
net/ipv6/ip6mr.c
1283
return READ_ONCE(net->ipv6.ipmr_seq) + ip6mr_rules_seq_read(net);
net/ipv6/ip6mr.c
1922
vr->icount = READ_ONCE(vif->pkt_in);
net/ipv6/ip6mr.c
1923
vr->ocount = READ_ONCE(vif->pkt_out);
net/ipv6/ip6mr.c
1924
vr->ibytes = READ_ONCE(vif->bytes_in);
net/ipv6/ip6mr.c
1925
vr->obytes = READ_ONCE(vif->bytes_out);
net/ipv6/ip6mr.c
1991
vr.icount = READ_ONCE(vif->pkt_in);
net/ipv6/ip6mr.c
1992
vr.ocount = READ_ONCE(vif->pkt_out);
net/ipv6/ip6mr.c
1993
vr.ibytes = READ_ONCE(vif->bytes_in);
net/ipv6/ip6mr.c
1994
vr.obytes = READ_ONCE(vif->bytes_out);
net/ipv6/ip6mr.c
2141
for (ct = READ_ONCE(mrt->maxvif) - 1; ct >= 0; ct--) {
net/ipv6/ip6mr.c
583
reg_vif_num = READ_ONCE(mrt->mroute_reg_vif_num);
net/ipv6/ip6mr.c
632
ip6mr_cache_report(mrt, skb, READ_ONCE(mrt->mroute_reg_vif_num),
net/ipv6/ipv6_sockglue.c
1137
int mcast_oif = READ_ONCE(np->mcast_oif);
net/ipv6/ipv6_sockglue.c
1146
int hlim = READ_ONCE(np->mcast_hops);
net/ipv6/ipv6_sockglue.c
1156
int mcast_oif = READ_ONCE(np->mcast_oif);
net/ipv6/ipv6_sockglue.c
1166
int hlim = READ_ONCE(np->mcast_hops);
net/ipv6/ipv6_sockglue.c
1317
val = READ_ONCE(np->hop_limit);
net/ipv6/ipv6_sockglue.c
1319
val = READ_ONCE(np->mcast_hops);
net/ipv6/ipv6_sockglue.c
1330
val = READ_ONCE(sock_net(sk)->ipv6.devconf_all->hop_limit);
net/ipv6/ipv6_sockglue.c
1339
val = READ_ONCE(np->mcast_oif);
net/ipv6/ipv6_sockglue.c
1347
val = (__force int)htonl((__u32) READ_ONCE(np->ucast_oif));
net/ipv6/ipv6_sockglue.c
1351
val = READ_ONCE(np->pmtudisc);
net/ipv6/ipv6_sockglue.c
1395
u8 srcprefs = READ_ONCE(np->srcprefs);
net/ipv6/ipv6_sockglue.c
1414
val = READ_ONCE(np->min_hopcount);
net/ipv6/ipv6_sockglue.c
196
if (optlen > READ_ONCE(sock_net(sk)->core.sysctl_optmem_max))
net/ipv6/ipv6_sockglue.c
230
if (optlen > READ_ONCE(sock_net(sk)->core.sysctl_optmem_max) - 4)
net/ipv6/ipv6_sockglue.c
514
bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
net/ipv6/ipv6_sockglue.c
541
if (READ_ONCE(sk->sk_bound_dev_if))
net/ipv6/mcast.c
1063
rv = READ_ONCE(psf->sf_count[MCAST_INCLUDE]) ||
net/ipv6/mcast.c
1064
READ_ONCE(psf->sf_count[MCAST_EXCLUDE]) !=
net/ipv6/mcast.c
1065
READ_ONCE(mc->mca_sfcount[MCAST_EXCLUDE]);
net/ipv6/mcast.c
1067
rv = READ_ONCE(mc->mca_sfcount[MCAST_EXCLUDE]) != 0;
net/ipv6/mcast.c
1239
all_force = READ_ONCE(net->ipv6.devconf_all->force_mld_version);
net/ipv6/mcast.c
1244
return all_force ?: READ_ONCE(idev->cnf.force_mld_version);
net/ipv6/mcast.c
165
iv = READ_ONCE(idev->cnf.mldv1_unsolicited_report_interval);
net/ipv6/mcast.c
167
iv = READ_ONCE(idev->cnf.mldv2_unsolicited_report_interval);
net/ipv6/mcast.c
1750
hdr->hop_limit = READ_ONCE(inet6_sk(sk)->hop_limit);
net/ipv6/mcast.c
1921
mtu = READ_ONCE(dev->mtu);
net/ipv6/mcast.c
3118
READ_ONCE(psf->sf_count[MCAST_INCLUDE]),
net/ipv6/mcast.c
3119
READ_ONCE(psf->sf_count[MCAST_EXCLUDE]));
net/ipv6/mcast.c
945
if (READ_ONCE(idev->dead)) {
net/ipv6/ndisc.c
1024
READ_ONCE(idev->cnf.drop_unsolicited_na))
net/ipv6/ndisc.c
1079
if (!neigh && lladdr && idev && READ_ONCE(idev->cnf.forwarding)) {
net/ipv6/ndisc.c
1090
if (READ_ONCE(neigh->nud_state) & NUD_FAILED)
net/ipv6/ndisc.c
1099
READ_ONCE(net->ipv6.devconf_all->forwarding) &&
net/ipv6/ndisc.c
1100
READ_ONCE(net->ipv6.devconf_all->proxy_ndp) &&
net/ipv6/ndisc.c
1148
if (!READ_ONCE(idev->cnf.forwarding))
net/ipv6/ndisc.c
1316
if (!READ_ONCE(in6_dev->cnf.accept_ra_defrtr)) {
net/ipv6/ndisc.c
1324
lifetime < READ_ONCE(in6_dev->cnf.accept_ra_min_lft)) {
net/ipv6/ndisc.c
1334
if (!READ_ONCE(in6_dev->cnf.accept_ra_from_local) &&
net/ipv6/ndisc.c
1345
!READ_ONCE(in6_dev->cnf.accept_ra_rtr_pref))
net/ipv6/ndisc.c
1412
if (READ_ONCE(in6_dev->cnf.accept_ra_min_hop_limit) < 256 &&
net/ipv6/ndisc.c
1414
if (READ_ONCE(in6_dev->cnf.accept_ra_min_hop_limit) <=
net/ipv6/ndisc.c
1497
if (!READ_ONCE(in6_dev->cnf.accept_ra_from_local) &&
net/ipv6/ndisc.c
1505
if (READ_ONCE(in6_dev->cnf.accept_ra_rtr_pref) && ndopts.nd_opts_ri) {
net/ipv6/ndisc.c
1517
!READ_ONCE(in6_dev->cnf.accept_ra_defrtr))
net/ipv6/ndisc.c
1520
ntohl(ri->lifetime) < READ_ONCE(in6_dev->cnf.accept_ra_min_lft))
net/ipv6/ndisc.c
1522
if (ri->prefix_len < READ_ONCE(in6_dev->cnf.accept_ra_rt_info_min_plen))
net/ipv6/ndisc.c
1524
if (ri->prefix_len > READ_ONCE(in6_dev->cnf.accept_ra_rt_info_max_plen))
net/ipv6/ndisc.c
1543
if (READ_ONCE(in6_dev->cnf.accept_ra_pinfo) && ndopts.nd_opts_pi) {
net/ipv6/ndisc.c
1554
if (ndopts.nd_opts_mtu && READ_ONCE(in6_dev->cnf.accept_ra_mtu)) {
net/ipv6/ndisc.c
1561
if (READ_ONCE(in6_dev->ra_mtu) != mtu) {
net/ipv6/ndisc.c
1568
} else if (READ_ONCE(in6_dev->cnf.mtu6) != mtu) {
net/ipv6/ndisc.c
1797
READ_ONCE(idev->cnf.suppress_frag_ndisc)) {
net/ipv6/ndisc.c
1872
if (READ_ONCE(idev->cnf.ndisc_notify) ||
net/ipv6/ndisc.c
1873
READ_ONCE(net->ipv6.devconf_all->ndisc_notify))
net/ipv6/ndisc.c
1882
evict_nocarrier = READ_ONCE(idev->cnf.ndisc_evict_nocarrier) &&
net/ipv6/ndisc.c
1883
READ_ONCE(net->ipv6.devconf_all->ndisc_evict_nocarrier);
net/ipv6/ndisc.c
451
tclass = idev ? READ_ONCE(idev->cnf.ndisc_tclass) : 0;
net/ipv6/ndisc.c
506
ip6_nd_hdr(skb, saddr, daddr, READ_ONCE(inet6_sk(sk)->hop_limit), skb->len);
net/ipv6/ndisc.c
541
inc_opt |= READ_ONCE(ifp->idev->cnf.force_tllao);
net/ipv6/ndisc.c
752
if (!(READ_ONCE(neigh->nud_state) & NUD_VALID)) {
net/ipv6/ndisc.c
773
ret = !!(READ_ONCE(n->flags) & NTF_ROUTER);
net/ipv6/ndisc.c
902
(READ_ONCE(idev->cnf.forwarding) &&
net/ipv6/ndisc.c
903
(READ_ONCE(net->ipv6.devconf_all->proxy_ndp) ||
net/ipv6/ndisc.c
904
READ_ONCE(idev->cnf.proxy_ndp)) &&
net/ipv6/ndisc.c
929
is_router = READ_ONCE(idev->cnf.forwarding);
net/ipv6/ndisc.c
973
switch (READ_ONCE(idev->cnf.accept_untracked_na)) {
net/ipv6/netfilter/ip6_tables.c
282
private = READ_ONCE(table->private); /* Address dependency. */
net/ipv6/netfilter/nf_reject_ipv6.c
170
READ_ONCE(net->ipv6.devconf_all->hop_limit));
net/ipv6/netfilter/nf_reject_ipv6.c
99
READ_ONCE(net->ipv6.devconf_all->hop_limit));
net/ipv6/output_core.c
115
hoplimit = READ_ONCE(idev->cnf.hop_limit);
net/ipv6/output_core.c
117
hoplimit = READ_ONCE(dev_net(dev)->ipv6.devconf_all->hop_limit);
net/ipv6/ping.c
110
oif = READ_ONCE(np->mcast_oif);
net/ipv6/ping.c
112
oif = READ_ONCE(np->ucast_oif);
net/ipv6/ping.c
158
fl6.flowi6_oif = READ_ONCE(np->mcast_oif);
net/ipv6/ping.c
160
fl6.flowi6_oif = READ_ONCE(np->ucast_oif);
net/ipv6/raw.c
164
READ_ONCE(sk->sk_rcvbuf)) {
net/ipv6/raw.c
317
harderr = (READ_ONCE(np->pmtudisc) == IPV6_PMTUDISC_DO);
net/ipv6/raw.c
449
if (np->rxopt.bits.rxpmtu && READ_ONCE(np->rxpmtu))
net/ipv6/raw.c
884
fl6.flowi6_oif = READ_ONCE(np->mcast_oif);
net/ipv6/raw.c
886
fl6.flowi6_oif = READ_ONCE(np->ucast_oif);
net/ipv6/route.c
1648
mtu = READ_ONCE(idev->cnf.mtu6);
net/ipv6/route.c
2150
if (time_after_eq(now, READ_ONCE(rt->dst.lastuse) +
net/ipv6/route.c
2156
} else if (time_after(jiffies, READ_ONCE(rt->dst.expires))) {
net/ipv6/route.c
2278
if (READ_ONCE(net->ipv6.devconf_all->forwarding) == 0)
net/ipv6/route.c
2696
flags |= rt6_srcprefs2flags(READ_ONCE(inet6_sk(sk)->srcprefs));
net/ipv6/route.c
2794
READ_ONCE(rt->dst.obsolete) == DST_OBSOLETE_FORCE_CHK &&
net/ipv6/route.c
2901
rt6_update_expires(rt, READ_ONCE(net->ipv6.sysctl.ip6_rt_mtu_expires));
net/ipv6/route.c
3026
ip6_update_pmtu(skb, sock_net(sk), mtu, oif, READ_ONCE(sk->sk_mark),
net/ipv6/route.c
3030
if (!dst || !READ_ONCE(dst->obsolete) ||
net/ipv6/route.c
3248
READ_ONCE(sk->sk_mark), sk_uid(sk));
net/ipv6/route.c
3263
READ_ONCE(net->ipv6.sysctl.ip6_rt_min_advmss));
net/ipv6/route.c
3317
mtu = max_t(u32, mtu, READ_ONCE(idev->cnf.mtu6));
net/ipv6/route.c
3365
int rt_min_interval = READ_ONCE(net->ipv6.sysctl.ip6_rt_gc_min_interval);
net/ipv6/route.c
3366
int rt_elasticity = READ_ONCE(net->ipv6.sysctl.ip6_rt_gc_elasticity);
net/ipv6/route.c
3367
int rt_gc_timeout = READ_ONCE(net->ipv6.sysctl.ip6_rt_gc_timeout);
net/ipv6/route.c
3368
unsigned long rt_last_gc = READ_ONCE(net->ipv6.ip6_rt_last_gc);
net/ipv6/route.c
394
return time_after(jiffies, READ_ONCE(rt->dst.expires));
net/ipv6/route.c
405
if (time_after(jiffies, READ_ONCE(rt->dst.expires)))
net/ipv6/route.c
408
return READ_ONCE(rt->dst.obsolete) != DST_OBSOLETE_FORCE_CHK ||
net/ipv6/route.c
4253
if (READ_ONCE(in6_dev->cnf.forwarding) ||
net/ipv6/route.c
4254
!READ_ONCE(in6_dev->cnf.accept_redirects))
net/ipv6/route.c
4698
(READ_ONCE(net->ipv6.devconf_all->disable_policy) ||
net/ipv6/route.c
4699
READ_ONCE(idev->cnf.disable_policy)))
net/ipv6/route.c
5009
if (READ_ONCE(net->ipv6.sysctl.skip_notify_on_dev_down))
net/ipv6/route.c
5358
READ_ONCE(rt_last->fib6_nsiblings)) {
net/ipv6/route.c
5697
if (READ_ONCE(f6i->fib6_nsiblings)) {
net/ipv6/route.c
5703
if (!READ_ONCE(f6i->fib6_nsiblings))
net/ipv6/route.c
5870
} else if (READ_ONCE(rt->fib6_nsiblings)) {
net/ipv6/route.c
5906
if (READ_ONCE(net->ipv4.sysctl_nexthop_compat_mode) &&
net/ipv6/route.c
5920
expires = dst ? READ_ONCE(dst->expires) : rt->expires;
net/ipv6/route.c
5925
if (READ_ONCE(rt->offload))
net/ipv6/route.c
5927
if (READ_ONCE(rt->trap))
net/ipv6/route.c
5929
if (READ_ONCE(rt->offload_failed))
net/ipv6/route.c
5972
if (READ_ONCE(f6i->fib6_nsiblings)) {
net/ipv6/route.c
5982
if (!READ_ONCE(f6i->fib6_nsiblings))
net/ipv6/route.c
6413
if (READ_ONCE(f6i->offload) == offload &&
net/ipv6/route.c
6414
READ_ONCE(f6i->trap) == trap &&
net/ipv6/route.c
6415
READ_ONCE(f6i->offload_failed) == offload_failed)
net/ipv6/route.c
6421
fib_notify_on_flag_change = READ_ONCE(net->ipv6.sysctl.fib_notify_on_flag_change);
net/ipv6/route.c
6424
READ_ONCE(f6i->offload_failed) == offload_failed)
net/ipv6/route.c
6532
delay = READ_ONCE(net->ipv6.sysctl.flush_delay);
net/ipv6/route.c
673
last_probe = READ_ONCE(fib6_nh->last_probe);
net/ipv6/route.c
679
if (READ_ONCE(neigh->nud_state) & NUD_VALID)
net/ipv6/route.c
686
READ_ONCE(idev->cnf.rtr_probe_interval))) {
net/ipv6/route.c
693
READ_ONCE(idev->cnf.rtr_probe_interval))) {
net/ipv6/route.c
729
u8 nud_state = READ_ONCE(neigh->nud_state);
net/ipv6/seg6_hmac.c
194
require_hmac = READ_ONCE(idev->cnf.seg6_require_hmac);
net/ipv6/syncookies.c
188
if (!READ_ONCE(net->ipv4.sysctl_tcp_syncookies) ||
net/ipv6/syncookies.c
255
req->rsk_window_clamp = READ_ONCE(tp->window_clamp) ? :dst_metric(dst, RTAX_WINDOW);
net/ipv6/tcp_ipv6.c
1079
priority = READ_ONCE(sk->sk_priority);
net/ipv6/tcp_ipv6.c
1088
if (READ_ONCE(net->ipv6.sysctl.flowlabel_reflect) &
net/ipv6/tcp_ipv6.c
1150
rnext_key = READ_ONCE(ao_info->rnext_key);
net/ipv6/tcp_ipv6.c
1152
key.sne = READ_ONCE(ao_info->snd_sne);
net/ipv6/tcp_ipv6.c
1166
READ_ONCE(tcptw->tw_rcv_nxt),
net/ipv6/tcp_ipv6.c
1169
READ_ONCE(tcptw->tw_ts_recent), tw->tw_bound_dev_if,
net/ipv6/tcp_ipv6.c
1249
READ_ONCE(sk->sk_priority),
net/ipv6/tcp_ipv6.c
1250
READ_ONCE(tcp_rsk(req)->txhash));
net/ipv6/tcp_ipv6.c
1432
if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_reflect_tos))
net/ipv6/tcp_ipv6.c
1824
if (unlikely(hdr->hop_limit < READ_ONCE(tcp_inet6_sk(sk)->min_hopcount))) {
net/ipv6/tcp_ipv6.c
2166
rx_queue = READ_ONCE(sp->sk_ack_backlog);
net/ipv6/tcp_ipv6.c
2171
rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) -
net/ipv6/tcp_ipv6.c
2172
READ_ONCE(tp->copied_seq), 0);
net/ipv6/tcp_ipv6.c
2183
READ_ONCE(tp->write_seq) - tp->snd_una,
net/ipv6/tcp_ipv6.c
2187
READ_ONCE(icsk->icsk_retransmits),
net/ipv6/tcp_ipv6.c
2189
READ_ONCE(icsk->icsk_probes_out),
net/ipv6/tcp_ipv6.c
2222
READ_ONCE(tw->tw_substate), 0, 0,
net/ipv6/tcp_ipv6.c
359
mtu = READ_ONCE(tcp_sk(sk)->mtu_info);
net/ipv6/tcp_ipv6.c
429
if (ipv6_hdr(skb)->hop_limit < READ_ONCE(tcp_inet6_sk(sk)->min_hopcount)) {
net/ipv6/tcp_ipv6.c
556
tclass = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_reflect_tos) ?
net/ipv6/tcp_ipv6.c
569
err = ip6_xmit(sk, skb, fl6, skb->mark ? : READ_ONCE(sk->sk_mark),
net/ipv6/tcp_ipv6.c
570
opt, tclass, READ_ONCE(sk->sk_priority));
net/ipv6/tcp_ipv6.c
937
mark = READ_ONCE(sk->sk_mark);
net/ipv6/udp.c
1501
ipc6.gso_size = READ_ONCE(up->gso_size);
net/ipv6/udp.c
1528
} else if (!READ_ONCE(up->pending)) {
net/ipv6/udp.c
1559
if (READ_ONCE(up->pending)) {
net/ipv6/udp.c
1560
if (READ_ONCE(up->pending) == AF_INET)
net/ipv6/udp.c
162
bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
net/ipv6/udp.c
1620
fl6->flowi6_oif = READ_ONCE(sk->sk_bound_dev_if);
net/ipv6/udp.c
169
if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
net/ipv6/udp.c
1700
fl6->flowi6_oif = READ_ONCE(np->mcast_oif);
net/ipv6/udp.c
1703
fl6->flowi6_oif = READ_ONCE(np->ucast_oif);
net/ipv6/udp.c
1802
if (!READ_ONCE(up->pending) || udp_test_bit(CORK, sk))
net/ipv6/udp.c
1824
encap_destroy = READ_ONCE(up->encap_destroy);
net/ipv6/udp.c
483
if (np->rxopt.bits.rxpmtu && READ_ONCE(np->rxpmtu))
net/ipv6/udp.c
673
lookup = READ_ONCE(up->encap_err_lookup);
net/ipv6/udp.c
686
lookup = READ_ONCE(up->encap_err_lookup);
net/ipv6/udp.c
721
if (!sk || READ_ONCE(udp_sk(sk)->encap_type)) {
net/ipv6/udp.c
748
if (READ_ONCE(np->pmtudisc) != IPV6_PMTUDISC_DONT)
net/ipv6/udp.c
754
READ_ONCE(sk->sk_mark),
net/ipv6/udp.c
840
READ_ONCE(up->encap_type)) {
net/ipv6/udp.c
855
encap_rcv = READ_ONCE(up->encap_rcv);
net/ipv6/udp.c
880
u16 pcrlen = READ_ONCE(up->pcrlen);
net/ipv6/udp.c
955
!udp_sk_bound_dev_eq(net, READ_ONCE(sk->sk_bound_dev_if), dif, sdif) ||
net/ipv6/xfrm6_input.c
87
encap_type = READ_ONCE(up->encap_type);
net/kcm/kcmsock.c
184
if (!READ_ONCE(kcm->rx_wait) && !READ_ONCE(kcm->rx_psock) &&
net/l2tp/l2tp_core.c
281
struct l2tp_tunnel *tunnel = READ_ONCE(session->tunnel);
net/l2tp/l2tp_core.c
340
struct l2tp_tunnel *tunnel = READ_ONCE(session->tunnel);
net/l2tp/l2tp_core.c
374
struct l2tp_tunnel *tunnel = READ_ONCE(session->tunnel);
net/l2tp/l2tp_core.c
397
struct l2tp_tunnel *tunnel = READ_ONCE(session->tunnel);
net/l2tp/l2tp_ip.c
70
bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
net/l2tp/l2tp_ip6.c
475
lsa->l2tp_scope_id = READ_ONCE(sk->sk_bound_dev_if);
net/l2tp/l2tp_ip6.c
548
fl6.flowi6_mark = READ_ONCE(sk->sk_mark);
net/l2tp/l2tp_ip6.c
590
fl6.flowi6_oif = READ_ONCE(sk->sk_bound_dev_if);
net/l2tp/l2tp_ip6.c
632
fl6.flowi6_oif = READ_ONCE(np->mcast_oif);
net/l2tp/l2tp_ip6.c
634
fl6.flowi6_oif = READ_ONCE(np->ucast_oif);
net/l2tp/l2tp_ip6.c
79
bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
net/llc/af_llc.c
213
llc_ui_wait_for_disc(sk, READ_ONCE(sk->sk_rcvtimeo));
net/llc/af_llc.c
458
rc = llc_ui_wait_for_disc(sk, READ_ONCE(sk->sk_rcvtimeo));
net/llc/af_llc.c
589
READ_ONCE(sk->sk_state) == TCP_CLOSE, &wait))
net/llc/af_llc.c
610
READ_ONCE(sk->sk_state) != TCP_SYN_SENT, &wait))
net/llc/af_llc.c
629
(READ_ONCE(sk->sk_shutdown) & RCV_SHUTDOWN) ||
net/llc/af_llc.c
715
rc = llc_wait_data(sk, READ_ONCE(sk->sk_rcvtimeo));
net/llc/af_llc.c
814
if (copied >= target && !READ_ONCE(sk->sk_backlog.tail))
net/llc/af_llc.c
964
size = min_t(size_t, size, READ_ONCE(dev->mtu));
net/llc/af_llc.c
978
size > READ_ONCE(dev->mtu))
net/llc/llc_conn.c
828
if (sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)))
net/llc/llc_input.c
200
sap_handler = dest ? READ_ONCE(llc_type_handlers[dest - 1]) : NULL;
net/llc/llc_input.c
221
sta_handler = READ_ONCE(llc_station_handler);
net/mac80211/sta_info.c
2587
u32 rate = READ_ONCE(sta_get_last_rx_stats(sta, link_id)->last_rate);
net/mctp/route.c
1133
netid = READ_ONCE(dst->dev->net);
net/mctp/route.c
1366
cb->net = READ_ONCE(mdev->net);
net/mctp/route.c
1469
READ_ONCE(net->mctp.default_net);
net/mctp/route.c
694
return READ_ONCE(net->mctp.default_net);
net/mctp/route.c
863
READ_ONCE(rt->dev->net) : rt->gateway.net;
net/mctp/route.c
890
dst->mtu = READ_ONCE(dst->dev->dev->mtu);
net/mctp/route.c
923
dst->mtu = READ_ONCE(netdev->mtu);
net/mctp/test/route-test.c
444
net = READ_ONCE(dev->mdev->net);
net/mctp/test/utils.c
206
cb->net = READ_ONCE(dev->mdev->net);
net/mpls/af_mpls.c
1189
READ_ONCE(mdev->input_enabled)) < 0)
net/mpls/af_mpls.c
272
alive = READ_ONCE(rt->rt_nhn_alive);
net/mpls/af_mpls.c
281
unsigned int nh_flags = READ_ONCE(nh->nh_flags);
net/mptcp/ctrl.c
243
u8 pm_type = READ_ONCE(*(u8 *)ctl->data);
net/mptcp/ctrl.c
447
if (!READ_ONCE(pernet->blackhole_timeout))
net/mptcp/ctrl.c
476
blackhole_timeout = READ_ONCE(pernet->blackhole_timeout);
net/mptcp/ctrl.c
492
timeout = READ_ONCE(pernet->active_disable_stamp) +
net/mptcp/diag.c
50
if (READ_ONCE(sf->fully_established))
net/mptcp/mptcp_diag.c
199
READ_ONCE(mptcp_sk(sk)->backlog_len);
net/mptcp/mptcp_diag.c
203
struct sock *lsk = READ_ONCE(msk->first);
net/mptcp/mptcp_diag.c
209
r->idiag_rqueue = READ_ONCE(lsk->sk_ack_backlog);
net/mptcp/mptcp_diag.c
210
r->idiag_wqueue = READ_ONCE(lsk->sk_max_ack_backlog);
net/mptcp/options.c
1001
if (unlikely(!READ_ONCE(msk->pm.server_side)))
net/mptcp/options.c
1086
u64 new_wnd_end, new_snd_una, snd_nxt = READ_ONCE(msk->snd_nxt);
net/mptcp/options.c
1121
new_wnd_end, READ_ONCE(msk->wnd_end));
net/mptcp/options.c
1131
if (READ_ONCE(msk->rcv_data_fin))
net/mptcp/options.c
1135
mptcp_expand_seq(READ_ONCE(msk->ack_seq), data_fin_seq, use_64bit));
net/mptcp/options.c
1149
hmac = add_addr_generate_hmac(READ_ONCE(msk->remote_key),
net/mptcp/options.c
1150
READ_ONCE(msk->local_key),
net/mptcp/options.c
1182
__mptcp_snd_una_update(msk, READ_ONCE(msk->snd_nxt));
net/mptcp/options.c
1199
READ_ONCE(msk->local_key) == mp_opt.rcvr_key) {
net/mptcp/options.c
1311
ack_seq = READ_ONCE(msk->ack_seq);
net/mptcp/options.c
1350
READ_ONCE(sock_net(ssk)->ipv4.sysctl_tcp_workaround_signed_windows))
net/mptcp/options.c
470
if (READ_ONCE(subflow->fully_established) || snd_data_fin_enable ||
net/mptcp/options.c
487
opts->csum_reqd = READ_ONCE(msk->csum_enabled);
net/mptcp/options.c
540
u64 data_fin_tx_seq = READ_ONCE(mptcp_sk(subflow->conn)->write_seq) - 1;
net/mptcp/options.c
575
opts->csum_reqd = READ_ONCE(msk->csum_enabled);
net/mptcp/options.c
599
if (!READ_ONCE(msk->can_ack)) {
net/mptcp/options.c
604
ack_seq = READ_ONCE(msk->ack_seq);
net/mptcp/options.c
605
if (READ_ONCE(msk->use_64bit_ack)) {
net/mptcp/options.c
706
opts->ahmac = add_addr_generate_hmac(READ_ONCE(msk->local_key),
net/mptcp/options.c
707
READ_ONCE(msk->remote_key),
net/mptcp/options.c
809
opts->rcvr_key = READ_ONCE(msk->remote_key);
net/mptcp/options.c
950
if (likely(READ_ONCE(subflow->fully_established))) {
net/mptcp/pm.c
1011
u32 rcv_tstamp = READ_ONCE(tcp_sk(ssk)->rcv_tstamp);
net/mptcp/pm.c
465
u8 add_addr = READ_ONCE(msk->pm.addr_signal);
net/mptcp/pm.c
491
u8 rm_addr = READ_ONCE(msk->pm.addr_signal);
net/mptcp/pm.c
514
pr_debug("msk=%p, token=%u side=%d\n", msk, READ_ONCE(msk->token), server_side);
net/mptcp/pm.c
540
READ_ONCE(pm->accept_subflow));
net/mptcp/pm.c
543
if (!READ_ONCE(pm->accept_subflow))
net/mptcp/pm.c
547
if (READ_ONCE(pm->accept_subflow)) {
net/mptcp/pm.c
586
if (READ_ONCE(pm->work_pending) &&
net/mptcp/pm.c
614
if (!READ_ONCE(pm->work_pending))
net/mptcp/pm.c
619
if (READ_ONCE(pm->work_pending))
net/mptcp/pm.c
642
if (!READ_ONCE(pm->work_pending) && !update_subflows)
net/mptcp/pm.c
667
READ_ONCE(pm->accept_addr));
net/mptcp/pm.c
684
(addr->id > 0 && !READ_ONCE(pm->accept_addr) &&
net/mptcp/pm.c
704
if (!READ_ONCE(pm->work_pending))
net/mptcp/pm.c
709
if (mptcp_lookup_anno_list_by_saddr(msk, addr) && READ_ONCE(pm->work_pending))
net/mptcp/pm.c
751
u8 remote_id = READ_ONCE(subflow->remote_id);
net/mptcp/pm_kernel.c
1434
READ_ONCE(pernet->limit_add_addr_accepted)))
net/mptcp/pm_kernel.c
1438
READ_ONCE(pernet->limit_extra_subflows)))
net/mptcp/pm_kernel.c
183
bool deny_id0 = READ_ONCE(msk->pm.remote_deny_join_id0);
net/mptcp/pm_kernel.c
207
bool deny_id0 = READ_ONCE(msk->pm.remote_deny_join_id0);
net/mptcp/pm_kernel.c
218
if (READ_ONCE(subflow->local_id) == local->id)
net/mptcp/pm_kernel.c
224
addrs[i].id = READ_ONCE(subflow->remote_id);
net/mptcp/pm_kernel.c
54
return READ_ONCE(pernet->endp_signal_max);
net/mptcp/pm_kernel.c
62
return READ_ONCE(pernet->endp_subflow_max);
net/mptcp/pm_kernel.c
70
return READ_ONCE(pernet->endp_laminar_max);
net/mptcp/pm_kernel.c
78
return READ_ONCE(pernet->endp_fullmesh_max);
net/mptcp/pm_kernel.c
86
return READ_ONCE(pernet->limit_add_addr_accepted);
net/mptcp/pm_kernel.c
94
return READ_ONCE(pernet->limit_extra_subflows);
net/mptcp/pm_kernel.c
957
if (!READ_ONCE(msk->fully_established) ||
net/mptcp/pm_netlink.c
353
if (nla_put_u32(skb, MPTCP_ATTR_TOKEN, READ_ONCE(msk->token)))
net/mptcp/pm_netlink.c
370
sk_err = READ_ONCE(ssk->sk_err);
net/mptcp/pm_netlink.c
411
int err = nla_put_u32(skb, MPTCP_ATTR_TOKEN, READ_ONCE(msk->token));
net/mptcp/pm_netlink.c
417
if (READ_ONCE(msk->pm.server_side)) {
net/mptcp/pm_netlink.c
425
if (READ_ONCE(msk->pm.remote_deny_join_id0))
net/mptcp/pm_netlink.c
451
if (nla_put_u32(skb, MPTCP_ATTR_TOKEN, READ_ONCE(msk->token)))
net/mptcp/pm_netlink.c
486
if (nla_put_u32(skb, MPTCP_ATTR_TOKEN, READ_ONCE(msk->token)))
net/mptcp/pm_netlink.c
600
if (nla_put_u32(skb, MPTCP_ATTR_TOKEN, READ_ONCE(msk->token)) < 0)
net/mptcp/pm_userspace.c
256
if (READ_ONCE(subflow->local_id) == 0) {
net/mptcp/protocol.c
1318
if (!can_coalesce && i >= READ_ONCE(net_hotdata.sysctl_max_skb_frags)) {
net/mptcp/protocol.c
1338
u64 snd_una = READ_ONCE(msk->snd_una);
net/mptcp/protocol.c
1399
if (READ_ONCE(msk->csum_enabled))
net/mptcp/protocol.c
1405
if (READ_ONCE(msk->csum_enabled))
net/mptcp/protocol.c
1437
u32 rcv_tstamp = READ_ONCE(tcp_sk(mptcp_subflow_tcp_sock(subflow))->rcv_tstamp);
net/mptcp/protocol.c
1485
subflow->avg_pacing_rate = READ_ONCE(ssk->sk_pacing_rate);
net/mptcp/protocol.c
1491
linger_time = div_u64((u64)READ_ONCE(ssk->sk_wmem_queued) << 32, pace);
net/mptcp/protocol.c
1519
wmem = READ_ONCE(ssk->sk_wmem_queued);
net/mptcp/protocol.c
1525
READ_ONCE(ssk->sk_pacing_rate) * burst,
net/mptcp/protocol.c
160
int limit = READ_ONCE(sk->sk_rcvbuf);
net/mptcp/protocol.c
1637
if (READ_ONCE(subflow->scheduled)) {
net/mptcp/protocol.c
1716
if (READ_ONCE(subflow->scheduled)) {
net/mptcp/protocol.c
1725
if (READ_ONCE(subflow->scheduled)) {
net/mptcp/protocol.c
1834
if (sk->sk_wmem_queued >= READ_ONCE(sk->sk_sndbuf))
net/mptcp/protocol.c
2112
time = tcp_stamp_us_delta(mstamp, READ_ONCE(msk->rcvq_space.time));
net/mptcp/protocol.c
2126
sf_rtt_us = READ_ONCE(tp->rcv_rtt_est.rtt_us);
net/mptcp/protocol.c
2127
sf_advmss = READ_ONCE(tp->advmss);
net/mptcp/protocol.c
217
if (!READ_ONCE(net->ipv4.sysctl_tcp_moderate_rcvbuf) ||
net/mptcp/protocol.c
2250
u64 hint_val = READ_ONCE(msk->ack_seq) - MPTCP_SKB_CB(skb)->map_seq;
net/mptcp/protocol.c
229
cap = READ_ONCE(net->ipv4.sysctl_tcp_rmem[2]);
net/mptcp/protocol.c
2626
if (subflow->close_event_done || READ_ONCE(subflow->local_id) < 0)
net/mptcp/protocol.c
2699
if (likely(!READ_ONCE(msk->rcv_fastclose)))
net/mptcp/protocol.c
2780
if (READ_ONCE(subflow->scheduled)) {
net/mptcp/protocol.c
2791
info.limit = READ_ONCE(msk->csum_enabled) ? dfrag->data_len :
net/mptcp/protocol.c
2844
if (READ_ONCE(subflow->scheduled))
net/mptcp/protocol.c
2980
fail_tout = msk->first ? READ_ONCE(mptcp_subflow_ctx(msk->first)->fail_tout) : 0;
net/mptcp/protocol.c
3064
sk->sk_rcvbuf = READ_ONCE(net->ipv4.sysctl_tcp_rmem[1]);
net/mptcp/protocol.c
3065
sk->sk_sndbuf = READ_ONCE(net->ipv4.sysctl_tcp_wmem[1]);
net/mptcp/protocol.c
3689
if (READ_ONCE(mptcp_subflow_ctx(ssk)->fully_established))
net/mptcp/protocol.c
3900
delta += READ_ONCE(tp->write_seq) - tp->snd_una;
net/mptcp/protocol.c
3926
*karg = mptcp_ioctl_outq(msk, READ_ONCE(msk->snd_una));
net/mptcp/protocol.c
4193
ssk = READ_ONCE(msk->first);
net/mptcp/protocol.c
4290
struct sock *ssk = READ_ONCE(msk->first);
net/mptcp/protocol.c
4298
shutdown = READ_ONCE(sk->sk_shutdown);
net/mptcp/protocol.c
4318
if (READ_ONCE(sk->sk_err))
net/mptcp/protocol.c
448
msk->write_seq == READ_ONCE(msk->snd_una);
net/mptcp/protocol.c
479
if (READ_ONCE(msk->rcv_data_fin) &&
net/mptcp/protocol.c
482
u64 rcv_data_fin_seq = READ_ONCE(msk->rcv_data_fin_seq);
net/mptcp/protocol.c
484
if (READ_ONCE(msk->ack_seq) == rcv_data_fin_seq) {
net/mptcp/protocol.c
571
u8 ack_pending = READ_ONCE(icsk->icsk_ack.pending);
net/mptcp/protocol.c
575
((READ_ONCE(tp->rcv_nxt) - READ_ONCE(tp->rcv_wup) >
net/mptcp/protocol.c
576
READ_ONCE(icsk->icsk_ack.rcv_mss)) ||
net/mptcp/protocol.c
583
int old_space = READ_ONCE(msk->old_wspace);
net/mptcp/protocol.c
61
return READ_ONCE(msk->wnd_end);
net/mptcp/protocol.c
66
unsigned short family = READ_ONCE(sk->sk_family);
net/mptcp/protocol.c
907
mptcp_subflow_ctx(ssk)->map_seq = READ_ONCE(msk->ack_seq);
net/mptcp/protocol.h
1009
if (READ_ONCE(ssk->sk_sndbuf) != subflow->cached_sndbuf)
net/mptcp/protocol.h
1022
if (likely(READ_ONCE(ssk->sk_sndbuf) == subflow->cached_sndbuf))
net/mptcp/protocol.h
1143
return READ_ONCE(msk->pm.addr_signal) &
net/mptcp/protocol.h
1149
return READ_ONCE(msk->pm.addr_signal) & BIT(MPTCP_ADD_ADDR_SIGNAL);
net/mptcp/protocol.h
1154
return READ_ONCE(msk->pm.addr_signal) & BIT(MPTCP_ADD_ADDR_ECHO);
net/mptcp/protocol.h
1159
return READ_ONCE(msk->pm.addr_signal) & BIT(MPTCP_RM_ADDR_SIGNAL);
net/mptcp/protocol.h
1164
return READ_ONCE(msk->pm.pm_type) == MPTCP_PM_TYPE_USERSPACE;
net/mptcp/protocol.h
1169
return READ_ONCE(msk->pm.pm_type) == MPTCP_PM_TYPE_KERNEL;
net/mptcp/protocol.h
1220
int local_id = READ_ONCE(subflow->local_id);
net/mptcp/protocol.h
1255
return READ_ONCE(msk->pm.remote_deny_join_id0) &&
net/mptcp/protocol.h
1285
struct sock *ssk = READ_ONCE(msk->first);
net/mptcp/protocol.h
1303
if (READ_ONCE(msk->snd_data_fin_enable) && !(ssk->sk_shutdown & SEND_SHUTDOWN)) {
net/mptcp/protocol.h
413
return mptcp_win_from_space(sk, READ_ONCE(sk->sk_rcvbuf) -
net/mptcp/protocol.h
414
READ_ONCE(mptcp_sk(sk)->backlog_len) -
net/mptcp/protocol.h
824
return READ_ONCE(msk->bytes_received) - READ_ONCE(msk->bytes_consumed);
net/mptcp/protocol.h
852
if (subflow->request_join && !READ_ONCE(subflow->fully_established))
net/mptcp/protocol.h
915
READ_ONCE(mptcp_sk(sk)->fully_established);
net/mptcp/protocol.h
945
return READ_ONCE(msk->snd_data_fin_enable) &&
net/mptcp/protocol.h
946
READ_ONCE(msk->write_seq) == READ_ONCE(msk->snd_nxt);
net/mptcp/protocol.h
954
val = READ_ONCE(mptcp_sk(sk)->notsent_lowat);
net/mptcp/protocol.h
955
return val ?: READ_ONCE(net->ipv4.sysctl_tcp_notsent_lowat);
net/mptcp/protocol.h
963
notsent_bytes = READ_ONCE(msk->write_seq) - READ_ONCE(msk->snd_nxt);
net/mptcp/protocol.h
989
new_sndbuf = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_wmem[0]);
net/mptcp/protocol.h
991
ssk_sndbuf = READ_ONCE(mptcp_subflow_tcp_sock(subflow)->sk_sndbuf);
net/mptcp/sched.c
186
if (READ_ONCE(subflow->scheduled))
net/mptcp/sched.c
206
if (READ_ONCE(subflow->scheduled))
net/mptcp/sockopt.c
106
if (READ_ONCE(ssk->sk_mark) != sk->sk_mark) {
net/mptcp/sockopt.c
1422
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_keepalive_time) / HZ);
net/mptcp/sockopt.c
1426
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_keepalive_intvl) / HZ);
net/mptcp/sockopt.c
1430
READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_keepalive_probes));
net/mptcp/sockopt.c
1449
return mptcp_put_int_option(msk, optval, optlen, READ_ONCE(inet_sk(sk)->tos));
net/mptcp/sockopt.c
1461
READ_ONCE(inet_sk(sk)->local_port_range));
net/mptcp/sockopt.c
1588
WRITE_ONCE(inet_sk(ssk)->local_port_range, READ_ONCE(inet_sk(sk)->local_port_range));
net/mptcp/sockopt.c
1605
if (READ_ONCE(subflow->setsockopt_seq) != msk->setsockopt_seq) {
net/mptcp/sockopt.c
1627
cap = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2]) >> 1;
net/mptcp/sockopt.c
722
READ_ONCE(inet_sk(sk)->local_port_range));
net/mptcp/sockopt.c
749
val = READ_ONCE(inet_sk(sk)->tos);
net/mptcp/sockopt.c
965
info->mptcpi_extra_subflows = READ_ONCE(msk->pm.extra_subflows);
net/mptcp/sockopt.c
966
info->mptcpi_add_addr_signal = READ_ONCE(msk->pm.add_addr_signaled);
net/mptcp/sockopt.c
967
info->mptcpi_add_addr_accepted = READ_ONCE(msk->pm.add_addr_accepted);
net/mptcp/sockopt.c
968
info->mptcpi_local_addr_used = READ_ONCE(msk->pm.local_addr_used);
net/mptcp/sockopt.c
991
if (READ_ONCE(msk->can_ack))
net/mptcp/sockopt.c
996
info->mptcpi_csum_enabled = READ_ONCE(msk->csum_enabled);
net/mptcp/subflow.c
1113
bool csum_reqd = READ_ONCE(msk->csum_enabled);
net/mptcp/subflow.c
1197
map_seq = mptcp_expand_seq(READ_ONCE(msk->ack_seq), mpext->data_seq, mpext->dsn64);
net/mptcp/subflow.c
1330
if (WARN_ON_ONCE(ssk != READ_ONCE(msk->first)))
net/mptcp/subflow.c
1382
if (unlikely(!READ_ONCE(msk->can_ack)))
net/mptcp/subflow.c
1385
old_ack = READ_ONCE(msk->ack_seq);
net/mptcp/subflow.c
1484
*full_space = mptcp_win_from_space(sk, READ_ONCE(sk->sk_rcvbuf));
net/mptcp/subflow.c
1660
subflow->remote_key = READ_ONCE(msk->remote_key);
net/mptcp/subflow.c
1661
subflow->local_key = READ_ONCE(msk->local_key);
net/mptcp/subflow.c
66
READ_ONCE(msk->pm.accept_subflow));
net/mptcp/subflow.c
759
subflow_generate_hmac(READ_ONCE(msk->remote_key),
net/mptcp/subflow.c
760
READ_ONCE(msk->local_key),
net/mptcp/subflow.c
77
subflow_generate_hmac(READ_ONCE(msk->local_key),
net/mptcp/subflow.c
78
READ_ONCE(msk->remote_key),
net/mptcp/token.c
223
if (READ_ONCE(msk->token) == token)
net/mptcp/token.c
259
if (READ_ONCE(msk->token) != token ||
net/mptcp/token.c
266
if (READ_ONCE(msk->token) != token ||
net/netfilter/ipvs/ip_vs_conn.c
1441
if (!READ_ONCE(ipvs->enable))
net/netfilter/ipvs/ip_vs_conn.c
887
if (READ_ONCE(ipvs->enable))
net/netfilter/ipvs/ip_vs_ctl.c
108
amemthresh = max(READ_ONCE(ipvs->sysctl_amemthresh), 0);
net/netfilter/ipvs/ip_vs_ctl.c
1485
if (!READ_ONCE(ipvs->enable)) {
net/netfilter/ipvs/ip_vs_ctl.c
258
if (!READ_ONCE(ipvs->enable))
net/netfilter/ipvs/ip_vs_est.c
233
if (!READ_ONCE(ipvs->enable))
net/netfilter/ipvs/ip_vs_est.c
308
READ_ONCE(ipvs->enable) && ipvs->est_max_threads)
net/netfilter/ipvs/ip_vs_est.c
345
if (READ_ONCE(ipvs->enable) && !ip_vs_est_stopped(ipvs)) {
net/netfilter/ipvs/ip_vs_est.c
419
crow = READ_ONCE(kd->est_row);
net/netfilter/ipvs/ip_vs_est.c
488
if (!ipvs->est_max_threads && READ_ONCE(ipvs->enable))
net/netfilter/ipvs/ip_vs_est.c
665
if (!READ_ONCE(ipvs->enable) || kthread_should_stop())
net/netfilter/ipvs/ip_vs_est.c
683
if (!READ_ONCE(ipvs->enable) || kthread_should_stop())
net/netfilter/ipvs/ip_vs_est.c
759
if (!READ_ONCE(ipvs->enable))
net/netfilter/ipvs/ip_vs_est.c
789
if (!READ_ONCE(ipvs->enable) || kthread_should_stop())
net/netfilter/ipvs/ip_vs_sync.c
1282
READ_ONCE(sysctl_wmem_max));
net/netfilter/ipvs/ip_vs_sync.c
1287
READ_ONCE(sysctl_rmem_max));
net/netfilter/ipvs/ip_vs_sync.c
1302
if (READ_ONCE(sk->sk_family) == AF_INET6) {
net/netfilter/ipvs/ip_vs_sync.c
462
unsigned long orig = READ_ONCE(cp->sync_endtime);
net/netfilter/ipvs/ip_vs_xmit.c
99
if (READ_ONCE(dst->obsolete) &&
net/netfilter/nf_conncount.c
310
if ((u32)jiffies == READ_ONCE(list->last_gc))
net/netfilter/nf_conntrack_core.c
1828
ct->mark = READ_ONCE(exp->master->mark);
net/netfilter/nf_conntrack_core.c
1913
unsigned long status = READ_ONCE(ct->status);
net/netfilter/nf_conntrack_core.c
2110
if (READ_ONCE(ct->timeout) != extra_jiffies)
net/netfilter/nf_conntrack_core.c
654
s32 timeout = READ_ONCE(ct->timeout) - nfct_time_stamp;
net/netfilter/nf_conntrack_ecache.c
155
old = READ_ONCE(e->missed);
net/netfilter/nf_conntrack_ecache.c
336
if (!READ_ONCE(nf_ctnetlink_has_listener))
net/netfilter/nf_conntrack_extend.c
139
unsigned int this_id = READ_ONCE(ext->gen_id);
net/netfilter/nf_conntrack_netlink.c
1190
if ((READ_ONCE(ct->mark) & filter->mark.mask) != filter->mark.val)
net/netfilter/nf_conntrack_netlink.c
1193
status = (u32)READ_ONCE(ct->status);
net/netfilter/nf_conntrack_netlink.c
2009
newmark = (READ_ONCE(ct->mark) & mask) ^ mark;
net/netfilter/nf_conntrack_netlink.c
2010
if (newmark != READ_ONCE(ct->mark))
net/netfilter/nf_conntrack_netlink.c
341
u32 mark = READ_ONCE(ct->mark);
net/netfilter/nf_conntrack_proto_gre.c
224
status = READ_ONCE(ct->status);
net/netfilter/nf_conntrack_proto_tcp.c
765
timeout = READ_ONCE(timeouts[TCP_CONNTRACK_UNACK]);
net/netfilter/nf_conntrack_proto_udp.c
100
status = READ_ONCE(ct->status);
net/netfilter/nf_conntrack_sane.c
117
if (READ_ONCE(ct_sane_info->state) != SANE_STATE_START_REQUESTED)
net/netfilter/nf_conntrack_standalone.c
377
seq_printf(s, "mark=%u ", READ_ONCE(ct->mark));
net/netfilter/nf_flow_table_core.c
216
timeout = READ_ONCE(tn->timeouts[TCP_CONNTRACK_CLOSE]);
net/netfilter/nf_flow_table_core.c
219
tcp_state = READ_ONCE(ct->proto.tcp.state);
net/netfilter/nf_flow_table_core.c
221
timeout = READ_ONCE(tn->timeouts[tcp_state]);
net/netfilter/nf_flow_table_core.c
224
offload_timeout = READ_ONCE(tn->offload_timeout);
net/netfilter/nf_flow_table_core.c
232
timeout = READ_ONCE(tn->timeouts[state]);
net/netfilter/nf_flow_table_core.c
234
offload_timeout = READ_ONCE(tn->offload_timeout);
net/netfilter/nf_flow_table_core.c
246
nf_flow_timeout_delta(READ_ONCE(ct->timeout)) > (__s32)timeout)
net/netfilter/nf_flow_table_core.c
362
if (force || timeout - READ_ONCE(flow->timeout) > HZ)
net/netfilter/nf_flow_table_core.c
469
u8 state = READ_ONCE(ct->proto.tcp.state);
net/netfilter/nf_nat_core.c
279
if (READ_ONCE(ignored_ct->status) & uses_nat)
net/netfilter/nf_nat_core.c
298
if (READ_ONCE(ct->status) & uses_nat)
net/netfilter/nf_nat_core.c
316
old_state = READ_ONCE(ct->proto.tcp.state);
net/netfilter/nf_nat_core.c
376
flags = READ_ONCE(ct->status);
net/netfilter/nf_synproxy_core.c
409
iph->ttl = READ_ONCE(net->ipv4.sysctl_ip_default_ttl);
net/netfilter/nf_synproxy_core.c
804
iph->hop_limit = READ_ONCE(net->ipv6.devconf_all->hop_limit);
net/netfilter/nf_tables_api.c
10412
if (READ_ONCE(nft_net->gc_seq) != trans->seq || trans->set->dead) {
net/netfilter/nf_tables_api.c
10768
gc_seq = READ_ONCE(nft_net->gc_seq);
net/netfilter/nf_tables_api.c
1156
return READ_ONCE(net->nft.base_seq);
net/netfilter/nf_tables_api.c
4889
u64 timeout = READ_ONCE(set->timeout);
net/netfilter/nf_tables_api.c
4890
u32 gc_int = READ_ONCE(set->gc_int);
net/netfilter/nf_tables_api.c
6092
u64 timeout = READ_ONCE(nft_set_ext_timeout(ext)->timeout);
net/netfilter/nf_tables_api.c
6093
u64 set_timeout = READ_ONCE(set->timeout);
net/netfilter/nf_tables_api.c
6106
expires = READ_ONCE(nft_set_ext_timeout(ext)->expiration);
net/netfilter/nf_tables_core.c
187
pstats = READ_ONCE(base_chain->stats);
net/netfilter/nf_tables_core.c
259
bool genbit = READ_ONCE(net->nft.gencursor);
net/netfilter/nf_tables_trace.c
121
u32 status = READ_ONCE(ct->status);
net/netfilter/nfnetlink.c
698
v = READ_ONCE(nf_ctnetlink_has_listener);
net/netfilter/nfnetlink.c
757
u8 v = READ_ONCE(nf_ctnetlink_has_listener);
net/netfilter/nfnetlink_queue.c
716
switch ((enum nfqnl_config_mode)READ_ONCE(queue->copy_mode)) {
net/netfilter/nfnetlink_queue.c
727
data_len = READ_ONCE(queue->copy_range);
net/netfilter/nft_connlimit.c
45
count = READ_ONCE(priv->list->count);
net/netfilter/nft_connlimit.c
47
if ((count > READ_ONCE(priv->limit)) ^ READ_ONCE(priv->invert)) {
net/netfilter/nft_ct.c
295
if (READ_ONCE(ct->mark) != value) {
net/netfilter/nft_ct.c
93
*dest = READ_ONCE(ct->mark);
net/netfilter/nft_dynset.c
106
READ_ONCE(nft_set_ext_timeout(ext)->timeout) != 0) {
net/netfilter/nft_dynset.c
107
timeout = priv->timeout ? : READ_ONCE(set->timeout);
net/netfilter/nft_dynset.c
325
(timeout || READ_ONCE(set->timeout)))
net/netfilter/nft_dynset.c
67
timeout = priv->timeout ? : READ_ONCE(set->timeout);
net/netfilter/nft_last.c
62
if (READ_ONCE(last->jiffies) != jiffies)
net/netfilter/nft_last.c
64
if (READ_ONCE(last->set) == 0)
net/netfilter/nft_last.c
73
unsigned long last_jiffies = READ_ONCE(last->jiffies);
net/netfilter/nft_last.c
74
u32 last_set = READ_ONCE(last->set);
net/netfilter/nft_set_hash.c
425
gc_seq = READ_ONCE(nft_net->gc_seq);
net/netfilter/nft_set_hash.c
449
if (READ_ONCE(nft_net->gc_seq) != gc_seq) {
net/netfilter/nft_socket.c
133
*dest = READ_ONCE(sk->sk_mark);
net/netfilter/nft_synproxy.c
52
struct nf_synproxy_info info = READ_ONCE(priv->info);
net/netfilter/nft_synproxy.c
83
struct nf_synproxy_info info = READ_ONCE(priv->info);
net/netfilter/xt_connmark.c
131
return ((READ_ONCE(ct->mark) & info->mask) == info->mark) ^ info->invert;
net/netfilter/xt_connmark.c
41
oldmark = READ_ONCE(ct->mark);
net/netfilter/xt_connmark.c
48
if (READ_ONCE(ct->mark) != newmark) {
net/netfilter/xt_connmark.c
60
newmark = (READ_ONCE(ct->mark) & ~info->ctmask) ^
net/netfilter/xt_connmark.c
62
if (READ_ONCE(ct->mark) != newmark) {
net/netfilter/xt_connmark.c
68
new_targetmark = (READ_ONCE(ct->mark) & info->ctmask);
net/netfilter/xt_limit.c
72
if ((READ_ONCE(priv->credit) < r->cost) && (READ_ONCE(priv->prev) == jiffies))
net/netfilter/xt_limit.c
78
old_credit = READ_ONCE(priv->credit);
net/netfilter/xt_socket.c
141
pskb->mark = READ_ONCE(sk->sk_mark);
net/netfilter/xt_socket.c
80
pskb->mark = READ_ONCE(sk->sk_mark);
net/netlabel/netlabel_calipso.c
76
return READ_ONCE(calipso_ops);
net/netlink/af_netlink.c
1000
bound = READ_ONCE(nlk->bound);
net/netlink/af_netlink.c
1090
if (!READ_ONCE(nlk->bound))
net/netlink/af_netlink.c
1116
nladdr->nl_pid = READ_ONCE(nlk->dst_portid);
net/netlink/af_netlink.c
1117
nladdr->nl_groups = netlink_group_mask(READ_ONCE(nlk->dst_group));
net/netlink/af_netlink.c
1120
nladdr->nl_pid = READ_ONCE(nlk->portid);
net/netlink/af_netlink.c
1148
if (READ_ONCE(sock->sk_state) == NETLINK_CONNECTED &&
net/netlink/af_netlink.c
1149
READ_ONCE(nlk->dst_portid) != nlk_sk(ssk)->portid) {
net/netlink/af_netlink.c
1219
if ((rmem == skb->truesize || rmem <= READ_ONCE(sk->sk_rcvbuf)) &&
net/netlink/af_netlink.c
1239
if (((rmem && rmem + skb->truesize > READ_ONCE(sk->sk_rcvbuf)) ||
net/netlink/af_netlink.c
1394
rcvbuf = READ_ONCE(sk->sk_rcvbuf);
net/netlink/af_netlink.c
1851
dst_portid = READ_ONCE(nlk->dst_portid);
net/netlink/af_netlink.c
1852
dst_group = READ_ONCE(nlk->dst_group);
net/netlink/af_netlink.c
1856
if (!READ_ONCE(nlk->bound)) {
net/netlink/af_netlink.c
1940
max_recvmsg_len = max(READ_ONCE(nlk->max_recvmsg_len), len);
net/netlink/af_netlink.c
1974
if (READ_ONCE(nlk->cb_running) &&
net/netlink/af_netlink.c
2281
max_recvmsg_len = READ_ONCE(nlk->max_recvmsg_len);
net/netlink/af_netlink.c
2295
rcvbuf = READ_ONCE(sk->sk_rcvbuf);
net/netlink/af_netlink.c
2710
READ_ONCE(nlk->cb_running),
net/netrom/af_netrom.c
456
msecs_to_jiffies(READ_ONCE(sysctl_netrom_transport_timeout));
net/netrom/af_netrom.c
458
msecs_to_jiffies(READ_ONCE(sysctl_netrom_transport_acknowledge_delay));
net/netrom/af_netrom.c
460
msecs_to_jiffies(READ_ONCE(sysctl_netrom_transport_maximum_tries));
net/netrom/af_netrom.c
462
msecs_to_jiffies(READ_ONCE(sysctl_netrom_transport_busy_delay));
net/netrom/af_netrom.c
464
msecs_to_jiffies(READ_ONCE(sysctl_netrom_transport_no_activity_timeout));
net/netrom/af_netrom.c
465
nr->window = READ_ONCE(sysctl_netrom_transport_requested_window_size);
net/netrom/af_netrom.c
490
sk->sk_priority = READ_ONCE(osk->sk_priority);
net/netrom/af_netrom.c
957
if (READ_ONCE(sysctl_netrom_reset_circuit) &&
net/netrom/nr_dev.c
84
*buff++ = READ_ONCE(sysctl_netrom_network_ttl_initialiser);
net/netrom/nr_in.c
100
if (READ_ONCE(sysctl_netrom_reset_circuit))
net/netrom/nr_in.c
131
if (READ_ONCE(sysctl_netrom_reset_circuit))
net/netrom/nr_in.c
265
if (READ_ONCE(sysctl_netrom_reset_circuit))
net/netrom/nr_out.c
209
*dptr++ = READ_ONCE(sysctl_netrom_network_ttl_initialiser);
net/netrom/nr_route.c
156
nr_neigh->quality = READ_ONCE(sysctl_netrom_default_path_quality);
net/netrom/nr_route.c
726
if (++nr_neigh->failed < READ_ONCE(sysctl_netrom_link_fails_count)) {
net/netrom/nr_route.c
770
READ_ONCE(sysctl_netrom_obsolescence_count_initialiser));
net/netrom/nr_route.c
784
if (!READ_ONCE(sysctl_netrom_routing_control) && ax25 != NULL)
net/netrom/nr_subr.c
186
*dptr++ = READ_ONCE(sysctl_netrom_network_ttl_initialiser);
net/netrom/nr_subr.c
240
*dptr++ = READ_ONCE(sysctl_netrom_network_ttl_initialiser);
net/openvswitch/conntrack.c
157
return ct ? READ_ONCE(ct->mark) : 0;
net/openvswitch/conntrack.c
350
new_mark = ct_mark | (READ_ONCE(ct->mark) & ~(mask));
net/openvswitch/conntrack.c
351
if (READ_ONCE(ct->mark) != new_mark) {
net/openvswitch/flow_table.c
272
int err, ma_count = READ_ONCE(ma->count);
net/openvswitch/flow_table.c
300
int i, ma_count = READ_ONCE(ma->count);
net/openvswitch/flow_table.c
937
return READ_ONCE(ma->count);
net/openvswitch/flow_table.c
944
return READ_ONCE(mc->cache_size);
net/packet/af_packet.c
1247
len = READ_ONCE(po->rx_ring.frame_max) + 1;
net/packet/af_packet.c
1248
idx = READ_ONCE(po->rx_ring.head);
net/packet/af_packet.c
1260
len = READ_ONCE(po->rx_ring.prb_bdqc.knum_blocks);
net/packet/af_packet.c
1261
idx = READ_ONCE(po->rx_ring.prb_bdqc.kactive_blk_num);
net/packet/af_packet.c
1276
int rcvbuf = READ_ONCE(sk->sk_rcvbuf);
net/packet/af_packet.c
1345
if (READ_ONCE(history[i]) == rxhash)
net/packet/af_packet.c
1351
if (READ_ONCE(history[victim]) != rxhash)
net/packet/af_packet.c
1459
unsigned int num = READ_ONCE(f->num_members);
net/packet/af_packet.c
2303
vnet_hdr_sz = READ_ONCE(po->vnet_hdr_sz);
net/packet/af_packet.c
2314
if (READ_ONCE(po->copy_thresh) &&
net/packet/af_packet.c
2396
READ_ONCE(po->tp_tstamp) |
net/packet/af_packet.c
2727
int vnet_hdr_sz = READ_ONCE(po->vnet_hdr_sz);
net/packet/af_packet.c
2747
proto = READ_ONCE(po->num);
net/packet/af_packet.c
2952
int vnet_hdr_sz = READ_ONCE(po->vnet_hdr_sz);
net/packet/af_packet.c
2962
proto = READ_ONCE(po->num);
net/packet/af_packet.c
3416
int vnet_hdr_len = READ_ONCE(pkt_sk(sk)->vnet_hdr_sz);
net/packet/af_packet.c
3587
dev = dev_get_by_index_rcu(sock_net(sk), READ_ONCE(pkt_sk(sk)->ifindex));
net/packet/af_packet.c
3607
ifindex = READ_ONCE(po->ifindex);
net/packet/af_packet.c
3610
sll->sll_protocol = READ_ONCE(po->num);
net/packet/af_packet.c
4001
if (!READ_ONCE(po->fanout))
net/packet/af_packet.c
4102
val = !!READ_ONCE(po->vnet_hdr_sz);
net/packet/af_packet.c
4105
val = READ_ONCE(po->vnet_hdr_sz);
net/packet/af_packet.c
4108
val = READ_ONCE(pkt_sk(sk)->copy_thresh);
net/packet/af_packet.c
4141
val = READ_ONCE(po->tp_tstamp);
net/packet/af_packet.c
4151
val = READ_ONCE(po->prot_hook.ignore_outgoing);
net/packet/af_packet.c
439
return READ_ONCE(h.h1->tp_status);
net/packet/af_packet.c
442
return READ_ONCE(h.h2->tp_status);
net/packet/af_packet.c
445
return READ_ONCE(h.h3->tp_status);
net/packet/af_packet.c
4729
ntohs(READ_ONCE(po->num)),
net/packet/af_packet.c
4730
READ_ONCE(po->ifindex),
net/packet/af_packet.c
477
if (!(ts_status = tpacket_get_timestamp(skb, &ts, READ_ONCE(po->tp_tstamp))))
net/packet/diag.c
146
rp->pdiag_num = ntohs(READ_ONCE(po->num));
net/packet/diag.c
20
pinfo.pdi_copy_thresh = READ_ONCE(po->copy_thresh);
net/packet/diag.c
21
pinfo.pdi_tstamp = READ_ONCE(po->tp_tstamp);
net/packet/diag.c
30
if (READ_ONCE(po->vnet_hdr_sz))
net/phonet/pn_netlink.c
155
err = fill_addr(skb, READ_ONCE(pnd->netdev->ifindex),
net/phonet/pn_netlink.c
306
err = fill_route(skb, READ_ONCE(dev->ifindex), addr << 2,
net/rds/send.c
230
send_gen = READ_ONCE(cp->cp_send_gen) + 1;
net/rds/send.c
510
raced = send_gen != READ_ONCE(cp->cp_send_gen);
net/rds/tcp_listen.c
214
dev_if = READ_ONCE(inet6->mcast_oif);
net/rds/tcp_listen.c
313
if (READ_ONCE(sk->sk_state) == TCP_CLOSE_WAIT ||
net/rds/tcp_listen.c
314
READ_ONCE(sk->sk_state) == TCP_LAST_ACK ||
net/rds/tcp_listen.c
315
READ_ONCE(sk->sk_state) == TCP_CLOSE)
net/rds/tcp_listen.c
67
return ntohs(READ_ONCE(inet_sk(sk)->inet_dport));
net/rose/af_rose.c
612
sk->sk_priority = READ_ONCE(osk->sk_priority);
net/rxrpc/af_rxrpc.c
228
max = READ_ONCE(rxrpc_max_backlog);
net/rxrpc/ar-internal.h
1265
wake_up_process(READ_ONCE(local->io_thread));
net/rxrpc/call_accept.c
60
call_tail = READ_ONCE(b->call_backlog_tail);
net/rxrpc/call_accept.c
69
tail = READ_ONCE(b->peer_backlog_tail);
net/rxrpc/call_accept.c
82
tail = READ_ONCE(b->conn_backlog_tail);
net/rxrpc/call_event.c
142
req.tq = READ_ONCE(call->tx_qtail);
net/rxrpc/call_event.c
26
ktime_t delay = ms_to_ktime(READ_ONCE(rxrpc_idle_ack_delay));
net/rxrpc/call_event.c
264
if (call->tx_bottom == READ_ONCE(call->send_top))
net/rxrpc/call_event.c
273
if (call->tx_bottom == READ_ONCE(call->send_top)) {
net/rxrpc/call_event.c
367
t = ktime_sub(READ_ONCE(call->expect_term_by), now);
net/rxrpc/call_event.c
442
ktime_t next = READ_ONCE(call->expect_term_by), delay;
net/rxrpc/call_event.c
50
delay = ms_to_ktime(READ_ONCE(rxrpc_soft_ack_delay));
net/rxrpc/call_object.c
673
struct rxrpc_net *rxnet = READ_ONCE(call->rxnet);
net/rxrpc/conn_object.c
411
idle_timestamp = READ_ONCE(conn->idle_timestamp);
net/rxrpc/input.c
1029
rxrpc_seq_t base = READ_ONCE(call->acks_hard_ack);
net/rxrpc/input.c
1244
timo = READ_ONCE(call->next_rx_timo);
net/rxrpc/input.c
388
rxrpc_seq_t top = READ_ONCE(call->tx_top);
net/rxrpc/input.c
411
rxrpc_seq_t whigh = READ_ONCE(call->rx_highest_seq);
net/rxrpc/input.c
706
unsigned long timo = READ_ONCE(call->next_req_timo);
net/rxrpc/input.c
748
avail = READ_ONCE(call->rtt_avail);
net/rxrpc/io_thread.c
36
io_thread = READ_ONCE(local->io_thread);
net/rxrpc/output.c
480
else if (!last && !after(READ_ONCE(call->send_top), txb->seq))
net/rxrpc/output.c
625
ktime_t delay = ms_to_ktime(READ_ONCE(call->next_rx_timo));
net/rxrpc/output.c
68
ktime_t delay = ms_to_ktime(READ_ONCE(call->next_rx_timo) / 6);
net/rxrpc/peer_event.c
246
s32 last_tx_at = READ_ONCE(peer->last_tx_at);
net/rxrpc/peer_object.c
495
return READ_ONCE(peer->recent_srtt_us);
net/rxrpc/proc.c
305
(s32)now - (s32)READ_ONCE(peer->last_tx_at),
net/rxrpc/proc.c
306
READ_ONCE(peer->recent_srtt_us),
net/rxrpc/proc.c
307
READ_ONCE(peer->recent_rto_us));
net/rxrpc/proc.c
80
timeout = ktime_ms_delta(READ_ONCE(call->expect_rx_by), ktime_get_real());
net/rxrpc/proc.c
82
tx_bottom = READ_ONCE(call->tx_bottom);
net/rxrpc/proc.c
96
tx_bottom, READ_ONCE(call->tx_top) - tx_bottom,
net/rxrpc/rtt.c
188
u32 backoff = READ_ONCE(call->backoff);
net/rxrpc/rxkad.c
1218
u32 counter = READ_ONCE(conn->channels[i].call_counter);
net/rxrpc/sendmsg.c
137
rtt = READ_ONCE(call->srtt_us) >> 3;
net/rxrpc/sendmsg.c
143
tx_start = READ_ONCE(call->tx_bottom);
net/rxrpc/sendmsg.c
264
poke = (READ_ONCE(call->tx_bottom) == call->send_top);
net/rxrpc/sendmsg.c
97
rxrpc_seq_t tx_bottom = READ_ONCE(call->tx_bottom);
net/sched/act_connmark.c
68
skb->mark = READ_ONCE(c->mark);
net/sched/act_connmark.c
84
skb->mark = READ_ONCE(c->mark);
net/sched/act_ct.c
190
entry->ct_metadata.mark = READ_ONCE(ct->mark);
net/sched/act_ct.c
908
new_mark = mark | (READ_ONCE(ct->mark) & ~(mask));
net/sched/act_ct.c
909
if (READ_ONCE(ct->mark) != new_mark) {
net/sched/act_ctinfo.c
134
if (!cp->dscpstatemask || (READ_ONCE(ct->mark) & cp->dscpstatemask))
net/sched/act_ctinfo.c
37
newdscp = (((READ_ONCE(ct->mark) & cp->dscpmask) >> cp->dscpmaskshift) << 2) &
net/sched/act_ctinfo.c
77
skb->mark = READ_ONCE(ct->mark) & cp->cpmarkmask;
net/sched/act_gact.c
154
int action = READ_ONCE(gact->tcf_action);
net/sched/act_gact.c
158
u32 ptype = READ_ONCE(gact->tcfg_ptype);
net/sched/act_gact.c
177
int action = READ_ONCE(gact->tcf_action);
net/sched/act_gate.c
127
int action = READ_ONCE(gact->tcf_action);
net/sched/act_gate.c
445
old_tk_offset = READ_ONCE(gact->tk_offset);
net/sched/act_gate.c
621
opt.action = READ_ONCE(gact->tcf_action);
net/sched/act_mirred.c
399
m_eaction = READ_ONCE(m->tcfm_eaction);
net/sched/act_mirred.c
425
int retval = READ_ONCE(m->tcf_action);
net/sched/act_mirred.c
446
blockid = READ_ONCE(m->tcfm_blockid);
net/sched/act_mirred.c
467
m_mac_header_xmit = READ_ONCE(m->tcfm_mac_header_xmit);
net/sched/act_mirred.c
468
m_eaction = READ_ONCE(m->tcfm_eaction);
net/sched/act_sample.c
177
retval = READ_ONCE(s->tcf_action);
net/sched/cls_api.c
3624
bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
net/sched/cls_api.c
3660
bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
net/sched/cls_api.c
3712
bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
net/sched/cls_api.c
3764
bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
net/sched/em_meta.c
321
bound_dev_if = READ_ONCE(skb->sk->sk_bound_dev_if);
net/sched/em_meta.c
452
dst->value = READ_ONCE(sk->sk_wmem_queued);
net/sched/em_meta.c
463
dst->value = READ_ONCE(sk->sk_forward_alloc);
net/sched/em_meta.c
505
dst->value = READ_ONCE(sk->sk_lingertime) / HZ;
net/sched/em_meta.c
527
dst->value = READ_ONCE(sk->sk_ack_backlog);
net/sched/em_meta.c
538
dst->value = READ_ONCE(sk->sk_max_ack_backlog);
net/sched/em_meta.c
549
dst->value = READ_ONCE(sk->sk_priority);
net/sched/em_meta.c
560
dst->value = READ_ONCE(sk->sk_rcvlowat);
net/sched/em_meta.c
571
dst->value = READ_ONCE(sk->sk_rcvtimeo) / HZ;
net/sched/em_meta.c
582
dst->value = READ_ONCE(sk->sk_sndtimeo) / HZ;
net/sched/sch_cake.c
2033
other_qlen = READ_ONCE(other_sch->q.qlen);
net/sched/sch_cake.c
2034
other_last_active = READ_ONCE(other_priv->last_active);
net/sched/sch_cake.c
2905
READ_ONCE(q->rate_bps), TCA_CAKE_PAD))
net/sched/sch_cake.c
2908
flow_mode = READ_ONCE(q->flow_mode);
net/sched/sch_cake.c
2912
if (nla_put_u32(skb, TCA_CAKE_RTT, READ_ONCE(q->interval)))
net/sched/sch_cake.c
2915
if (nla_put_u32(skb, TCA_CAKE_TARGET, READ_ONCE(q->target)))
net/sched/sch_cake.c
2919
READ_ONCE(q->buffer_config_limit)))
net/sched/sch_cake.c
2922
rate_flags = READ_ONCE(q->rate_flags);
net/sched/sch_cake.c
2931
if (nla_put_u32(skb, TCA_CAKE_ACK_FILTER, READ_ONCE(q->ack_filter)))
net/sched/sch_cake.c
2938
if (nla_put_u32(skb, TCA_CAKE_DIFFSERV_MODE, READ_ONCE(q->tin_mode)))
net/sched/sch_cake.c
2945
if (nla_put_u32(skb, TCA_CAKE_OVERHEAD, READ_ONCE(q->rate_overhead)))
net/sched/sch_cake.c
2952
if (nla_put_u32(skb, TCA_CAKE_ATM, READ_ONCE(q->atm_mode)))
net/sched/sch_cake.c
2955
if (nla_put_u32(skb, TCA_CAKE_MPU, READ_ONCE(q->rate_mpu)))
net/sched/sch_cake.c
2962
if (nla_put_u32(skb, TCA_CAKE_FWMARK, READ_ONCE(q->fwmark_mask)))
net/sched/sch_cbs.c
462
opt.hicredit = READ_ONCE(q->hicredit);
net/sched/sch_cbs.c
463
opt.locredit = READ_ONCE(q->locredit);
net/sched/sch_cbs.c
464
opt.sendslope = div64_s64(READ_ONCE(q->sendslope), BYTES_PER_KBIT);
net/sched/sch_cbs.c
465
opt.idleslope = div64_s64(READ_ONCE(q->idleslope), BYTES_PER_KBIT);
net/sched/sch_cbs.c
466
opt.offload = READ_ONCE(q->offload);
net/sched/sch_choke.c
434
u8 Wlog = READ_ONCE(q->parms.Wlog);
net/sched/sch_choke.c
437
.limit = READ_ONCE(q->limit),
net/sched/sch_choke.c
438
.flags = READ_ONCE(q->flags),
net/sched/sch_choke.c
439
.qth_min = READ_ONCE(q->parms.qth_min) >> Wlog,
net/sched/sch_choke.c
440
.qth_max = READ_ONCE(q->parms.qth_max) >> Wlog,
net/sched/sch_choke.c
442
.Plog = READ_ONCE(q->parms.Plog),
net/sched/sch_choke.c
443
.Scell_log = READ_ONCE(q->parms.Scell_log),
net/sched/sch_choke.c
451
nla_put_u32(skb, TCA_CHOKE_MAX_P, READ_ONCE(q->parms.max_P)))
net/sched/sch_codel.c
201
codel_time_to_us(READ_ONCE(q->params.target))) ||
net/sched/sch_codel.c
203
READ_ONCE(sch->limit)) ||
net/sched/sch_codel.c
205
codel_time_to_us(READ_ONCE(q->params.interval))) ||
net/sched/sch_codel.c
207
READ_ONCE(q->params.ecn)))
net/sched/sch_codel.c
209
ce_threshold = READ_ONCE(q->params.ce_threshold);
net/sched/sch_dualpi2.c
1002
READ_ONCE(q->min_qlen_step)) ||
net/sched/sch_dualpi2.c
1004
READ_ONCE(q->coupling_factor)) ||
net/sched/sch_dualpi2.c
1006
READ_ONCE(q->drop_overload)) ||
net/sched/sch_dualpi2.c
1008
READ_ONCE(q->drop_early)) ||
net/sched/sch_dualpi2.c
1010
READ_ONCE(q->c_protection_wc)) ||
net/sched/sch_dualpi2.c
1011
nla_put_u8(skb, TCA_DUALPI2_ECN_MASK, READ_ONCE(q->ecn_mask)) ||
net/sched/sch_dualpi2.c
1012
nla_put_u8(skb, TCA_DUALPI2_SPLIT_GSO, READ_ONCE(q->split_gso))))
net/sched/sch_dualpi2.c
1026
.prob = READ_ONCE(q->pi2_prob),
net/sched/sch_dualpi2.c
292
prob = READ_ONCE(q->pi2_prob);
net/sched/sch_dualpi2.c
633
qc = READ_ONCE(q->c_head_ts);
net/sched/sch_dualpi2.c
634
ql = READ_ONCE(q->l_head_ts);
net/sched/sch_dualpi2.c
953
step_in_pkts = READ_ONCE(q->step_in_packets);
net/sched/sch_dualpi2.c
954
step_th = READ_ONCE(q->step_thresh);
net/sched/sch_dualpi2.c
961
(nla_put_u32(skb, TCA_DUALPI2_LIMIT, READ_ONCE(sch->limit)) ||
net/sched/sch_dualpi2.c
963
READ_ONCE(q->memory_limit)) ||
net/sched/sch_dualpi2.c
965
convert_ns_to_usec(READ_ONCE(q->pi2_target))) ||
net/sched/sch_dualpi2.c
967
convert_ns_to_usec(READ_ONCE(q->pi2_tupdate))) ||
net/sched/sch_dualpi2.c
969
dualpi2_unscale_alpha_beta(READ_ONCE(q->pi2_alpha))) ||
net/sched/sch_dualpi2.c
971
dualpi2_unscale_alpha_beta(READ_ONCE(q->pi2_beta))) ||
net/sched/sch_dualpi2.c
974
READ_ONCE(q->min_qlen_step)) ||
net/sched/sch_dualpi2.c
976
READ_ONCE(q->coupling_factor)) ||
net/sched/sch_dualpi2.c
978
READ_ONCE(q->drop_overload)) ||
net/sched/sch_dualpi2.c
980
READ_ONCE(q->drop_early)) ||
net/sched/sch_dualpi2.c
982
READ_ONCE(q->c_protection_wc)) ||
net/sched/sch_dualpi2.c
983
nla_put_u8(skb, TCA_DUALPI2_ECN_MASK, READ_ONCE(q->ecn_mask)) ||
net/sched/sch_dualpi2.c
984
nla_put_u8(skb, TCA_DUALPI2_SPLIT_GSO, READ_ONCE(q->split_gso))))
net/sched/sch_dualpi2.c
988
(nla_put_u32(skb, TCA_DUALPI2_LIMIT, READ_ONCE(sch->limit)) ||
net/sched/sch_dualpi2.c
990
READ_ONCE(q->memory_limit)) ||
net/sched/sch_dualpi2.c
992
convert_ns_to_usec(READ_ONCE(q->pi2_target))) ||
net/sched/sch_dualpi2.c
994
convert_ns_to_usec(READ_ONCE(q->pi2_tupdate))) ||
net/sched/sch_dualpi2.c
996
dualpi2_unscale_alpha_beta(READ_ONCE(q->pi2_alpha))) ||
net/sched/sch_dualpi2.c
998
dualpi2_unscale_alpha_beta(READ_ONCE(q->pi2_beta))) ||
net/sched/sch_etf.c
470
opt.delta = READ_ONCE(q->delta);
net/sched/sch_etf.c
471
opt.clockid = READ_ONCE(q->clockid);
net/sched/sch_etf.c
472
if (READ_ONCE(q->offload))
net/sched/sch_etf.c
475
if (READ_ONCE(q->deadline_mode))
net/sched/sch_etf.c
478
if (READ_ONCE(q->skip_sock_check))
net/sched/sch_ets.c
763
nbands = READ_ONCE(q->nbands);
net/sched/sch_ets.c
767
nstrict = READ_ONCE(q->nstrict);
net/sched/sch_ets.c
778
READ_ONCE(q->classes[band].quantum)))
net/sched/sch_ets.c
791
READ_ONCE(q->prio2band[prio])))
net/sched/sch_fifo.c
161
struct tc_fifo_qopt opt = { .limit = READ_ONCE(sch->limit) };
net/sched/sch_fifo.c
23
READ_ONCE(sch->limit)))
net/sched/sch_fifo.c
32
if (likely(sch->q.qlen < READ_ONCE(sch->limit)))
net/sched/sch_fifo.c
43
if (unlikely(READ_ONCE(sch->limit) == 0))
net/sched/sch_fifo.c
46
if (likely(sch->q.qlen < READ_ONCE(sch->limit)))
net/sched/sch_fq.c
1230
ce_threshold = READ_ONCE(q->ce_threshold);
net/sched/sch_fq.c
1233
horizon = READ_ONCE(q->horizon);
net/sched/sch_fq.c
1236
offload_horizon = READ_ONCE(q->offload_horizon);
net/sched/sch_fq.c
1240
READ_ONCE(sch->limit)) ||
net/sched/sch_fq.c
1242
READ_ONCE(q->flow_plimit)) ||
net/sched/sch_fq.c
1244
READ_ONCE(q->quantum)) ||
net/sched/sch_fq.c
1246
READ_ONCE(q->initial_quantum)) ||
net/sched/sch_fq.c
1248
READ_ONCE(q->rate_enable)) ||
net/sched/sch_fq.c
1251
READ_ONCE(q->flow_max_rate), ~0U)) ||
net/sched/sch_fq.c
1253
jiffies_to_usecs(READ_ONCE(q->flow_refill_delay))) ||
net/sched/sch_fq.c
1255
READ_ONCE(q->orphan_mask)) ||
net/sched/sch_fq.c
1257
READ_ONCE(q->low_rate_threshold)) ||
net/sched/sch_fq.c
1260
READ_ONCE(q->fq_trees_log)) ||
net/sched/sch_fq.c
1262
READ_ONCE(q->timer_slack)) ||
net/sched/sch_fq.c
1266
READ_ONCE(q->horizon_drop)))
net/sched/sch_fq.c
1273
weights[0] = READ_ONCE(q->band_flows[0].quantum);
net/sched/sch_fq.c
1274
weights[1] = READ_ONCE(q->band_flows[1].quantum);
net/sched/sch_fq.c
1275
weights[2] = READ_ONCE(q->band_flows[2].quantum);
net/sched/sch_fq.c
165
return (READ_ONCE(prio2band[prio / 4]) >> (2 * (prio & 0x3))) & 0x3;
net/sched/sch_fq.c
395
READ_ONCE(sk->sk_pacing_status) != SK_PACING_FQ)
net/sched/sch_fq.c
749
rate = min(READ_ONCE(skb->sk->sk_pacing_rate), rate);
net/sched/sch_fq_codel.c
544
codel_time_to_us(READ_ONCE(q->cparams.target))) ||
net/sched/sch_fq_codel.c
546
READ_ONCE(sch->limit)) ||
net/sched/sch_fq_codel.c
548
codel_time_to_us(READ_ONCE(q->cparams.interval))) ||
net/sched/sch_fq_codel.c
550
READ_ONCE(q->cparams.ecn)) ||
net/sched/sch_fq_codel.c
552
READ_ONCE(q->quantum)) ||
net/sched/sch_fq_codel.c
554
READ_ONCE(q->drop_batch_size)) ||
net/sched/sch_fq_codel.c
556
READ_ONCE(q->memory_limit)) ||
net/sched/sch_fq_codel.c
558
READ_ONCE(q->flows_cnt)))
net/sched/sch_fq_codel.c
561
ce_threshold = READ_ONCE(q->cparams.ce_threshold);
net/sched/sch_fq_codel.c
567
READ_ONCE(q->cparams.ce_threshold_selector)))
net/sched/sch_fq_codel.c
570
READ_ONCE(q->cparams.ce_threshold_mask)))
net/sched/sch_fq_pie.c
483
if (nla_put_u32(skb, TCA_FQ_PIE_LIMIT, READ_ONCE(sch->limit)) ||
net/sched/sch_fq_pie.c
484
nla_put_u32(skb, TCA_FQ_PIE_FLOWS, READ_ONCE(q->flows_cnt)) ||
net/sched/sch_fq_pie.c
486
((u32)PSCHED_TICKS2NS(READ_ONCE(q->p_params.target))) /
net/sched/sch_fq_pie.c
489
jiffies_to_usecs(READ_ONCE(q->p_params.tupdate))) ||
net/sched/sch_fq_pie.c
490
nla_put_u32(skb, TCA_FQ_PIE_ALPHA, READ_ONCE(q->p_params.alpha)) ||
net/sched/sch_fq_pie.c
491
nla_put_u32(skb, TCA_FQ_PIE_BETA, READ_ONCE(q->p_params.beta)) ||
net/sched/sch_fq_pie.c
492
nla_put_u32(skb, TCA_FQ_PIE_QUANTUM, READ_ONCE(q->quantum)) ||
net/sched/sch_fq_pie.c
494
READ_ONCE(q->memory_limit)) ||
net/sched/sch_fq_pie.c
495
nla_put_u32(skb, TCA_FQ_PIE_ECN_PROB, READ_ONCE(q->ecn_prob)) ||
net/sched/sch_fq_pie.c
496
nla_put_u32(skb, TCA_FQ_PIE_ECN, READ_ONCE(q->p_params.ecn)) ||
net/sched/sch_fq_pie.c
497
nla_put_u32(skb, TCA_FQ_PIE_BYTEMODE, READ_ONCE(q->p_params.bytemode)) ||
net/sched/sch_fq_pie.c
499
READ_ONCE(q->p_params.dq_rate_estimator)))
net/sched/sch_generic.c
417
int quota = READ_ONCE(net_hotdata.dev_tx_weight);
net/sched/sch_generic.c
435
unsigned long res = READ_ONCE(netdev_get_tx_queue(dev, 0)->trans_start);
net/sched/sch_generic.c
440
val = READ_ONCE(netdev_get_tx_queue(dev, i)->trans_start);
net/sched/sch_generic.c
526
trans_start = READ_ONCE(txq->trans_start);
net/sched/sch_generic.c
775
READ_ONCE(qdisc->state) & QDISC_STATE_NON_EMPTY) {
net/sched/sch_hfsc.c
1177
READ_ONCE(q->defcls)), sch);
net/sched/sch_hfsc.c
1527
qopt.defcls = READ_ONCE(q->defcls);
net/sched/sch_hhf.c
665
if (nla_put_u32(skb, TCA_HHF_BACKLOG_LIMIT, READ_ONCE(sch->limit)) ||
net/sched/sch_hhf.c
666
nla_put_u32(skb, TCA_HHF_QUANTUM, READ_ONCE(q->quantum)) ||
net/sched/sch_hhf.c
668
READ_ONCE(q->hh_flows_limit)) ||
net/sched/sch_hhf.c
670
jiffies_to_usecs(READ_ONCE(q->hhf_reset_timeout))) ||
net/sched/sch_hhf.c
672
READ_ONCE(q->hhf_admit_bytes)) ||
net/sched/sch_hhf.c
674
jiffies_to_usecs(READ_ONCE(q->hhf_evict_timeout))) ||
net/sched/sch_hhf.c
676
READ_ONCE(q->hhf_non_hh_weight)))
net/sched/sch_pie.c
479
((u32)PSCHED_TICKS2NS(READ_ONCE(q->params.target))) /
net/sched/sch_pie.c
481
nla_put_u32(skb, TCA_PIE_LIMIT, READ_ONCE(sch->limit)) ||
net/sched/sch_pie.c
483
jiffies_to_usecs(READ_ONCE(q->params.tupdate))) ||
net/sched/sch_pie.c
484
nla_put_u32(skb, TCA_PIE_ALPHA, READ_ONCE(q->params.alpha)) ||
net/sched/sch_pie.c
485
nla_put_u32(skb, TCA_PIE_BETA, READ_ONCE(q->params.beta)) ||
net/sched/sch_pie.c
488
READ_ONCE(q->params.bytemode)) ||
net/sched/sch_pie.c
490
READ_ONCE(q->params.dq_rate_estimator)))
net/sched/sch_sfq.c
621
period = READ_ONCE(q->perturb_period);
net/sched/sch_skbprio.c
202
opt.limit = READ_ONCE(sch->limit);
net/sched/sch_skbprio.c
84
if (sch->q.qlen < READ_ONCE(sch->limit)) {
net/sched/sch_taprio.c
181
enum tk_offsets tk_offset = READ_ONCE(q->tk_offset);
net/sched/sch_teql.c
81
if (q->q.qlen < READ_ONCE(dev->tx_queue_len)) {
net/sctp/diag.c
182
mem[SK_MEMINFO_BACKLOG] = READ_ONCE(sk->sk_backlog.len);
net/sctp/diag.c
424
r->idiag_rqueue = READ_ONCE(sk->sk_ack_backlog);
net/sctp/diag.c
425
r->idiag_wqueue = READ_ONCE(sk->sk_max_ack_backlog);
net/sctp/endpointola.c
253
int bound_dev_if = READ_ONCE(ep->base.sk->sk_bound_dev_if);
net/sctp/endpointola.c
305
int bound_dev_if = READ_ONCE(ep->base.sk->sk_bound_dev_if);
net/sctp/input.c
1017
bound_dev_if = READ_ONCE(t->asoc->base.sk->sk_bound_dev_if);
net/sctp/input.c
312
if (sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)))
net/sctp/input.c
327
if (!sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)))
net/sctp/input.c
354
ret = sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf));
net/sctp/input.c
989
l3mdev_accept = !!READ_ONCE(net->sctp.l3mdev_accept);
net/sctp/ipv6.c
250
tclass, READ_ONCE(sk->sk_priority));
net/sctp/ipv6.c
706
bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
net/sctp/output.c
137
packet->max_size = sk_can_gso(sk) ? min(READ_ONCE(tp->dst->dev->gso_max_size),
net/sctp/proc.c
285
READ_ONCE(sk->sk_wmem_queued),
net/sctp/protocol.c
1045
__u8 dscp = READ_ONCE(inet->tos);
net/sctp/protocol.c
364
!READ_ONCE(net->ipv4.sysctl_ip_nonlocal_bind))
net/sctp/socket.c
8479
int bound_dev_if2 = READ_ONCE(sk2->sk_bound_dev_if);
net/sctp/socket.c
9356
return READ_ONCE(sk->sk_sndbuf) > READ_ONCE(sk->sk_wmem_queued);
net/sctp/transport.c
234
if (!transport->dst || READ_ONCE(transport->dst->obsolete)) {
net/sctp/transport.c
501
rto_beta = READ_ONCE(net->sctp.rto_beta);
net/sctp/transport.c
505
rto_alpha = READ_ONCE(net->sctp.rto_alpha);
net/shaper/shaper.c
44
return READ_ONCE(binding->netdev->net_shaper_hierarchy);
net/shaper/shaper.c
58
READ_ONCE(binding->netdev->reg_state) <= NETREG_REGISTERED)
net/shaper/shaper.c
59
return READ_ONCE(binding->netdev->net_shaper_hierarchy);
net/smc/af_smc.c
143
if (READ_ONCE(sk->sk_ack_backlog) + atomic_read(&smc->queued_smc_hs) >
net/smc/af_smc.c
1611
long timeo = READ_ONCE(smc->sk.sk_sndtimeo);
net/smc/af_smc.c
400
WRITE_ONCE(sk->sk_sndbuf, 2 * READ_ONCE(net->smc.sysctl_wmem));
net/smc/af_smc.c
401
WRITE_ONCE(sk->sk_rcvbuf, 2 * READ_ONCE(net->smc.sysctl_rmem));
net/smc/af_smc.c
514
nsk->sk_sndtimeo = READ_ONCE(osk->sk_sndtimeo);
net/smc/af_smc.c
515
nsk->sk_rcvtimeo = READ_ONCE(osk->sk_rcvtimeo);
net/smc/af_smc.c
516
nsk->sk_mark = READ_ONCE(osk->sk_mark);
net/smc/af_smc.c
517
nsk->sk_priority = READ_ONCE(osk->sk_priority);
net/smc/smc_clc.c
692
long rcvtimeo = READ_ONCE(smc->clcsock->sk->sk_rcvtimeo);
net/smc/smc_close.c
70
READ_ONCE(sk->sk_err) == ECONNABORTED ||
net/smc/smc_close.c
71
READ_ONCE(sk->sk_err) == ECONNRESET ||
net/smc/smc_llc.c
2159
lgr->llc_testlink_time = READ_ONCE(net->smc.sysctl_smcr_testlink_time);
net/smc/smc_llc.c
2160
lgr->max_send_wr = (u16)(READ_ONCE(net->smc.sysctl_smcr_max_send_wr));
net/smc/smc_llc.c
2161
lgr->max_recv_wr = (u16)(READ_ONCE(net->smc.sysctl_smcr_max_recv_wr));
net/smc/smc_rx.c
278
READ_ONCE(sk->sk_err) ||
net/smc/smc_rx.c
280
READ_ONCE(sk->sk_shutdown) & RCV_SHUTDOWN ||
net/smc/smc_tx.c
116
READ_ONCE(sk->sk_err) ||
net/smc/smc_tx.c
117
(READ_ONCE(sk->sk_shutdown) & SEND_SHUTDOWN) ||
net/socket.c
1078
int ret = INDIRECT_CALL_INET(READ_ONCE(sock->ops)->recvmsg,
net/socket.c
1136
ops = READ_ONCE(sock->ops);
net/socket.c
1148
ops = READ_ONCE(sock->ops);
net/socket.c
1247
const struct proto_ops *ops = READ_ONCE(sock->ops);
net/socket.c
1291
ops = READ_ONCE(sock->ops);
net/socket.c
142
const struct proto_ops *ops = READ_ONCE(sock->ops);
net/socket.c
1428
const struct proto_ops *ops = READ_ONCE(sock->ops);
net/socket.c
1450
return READ_ONCE(sock->ops)->mmap(file, sock, vma);
net/socket.c
1827
err = READ_ONCE(sock1->ops)->socketpair(sock1, sock2);
net/socket.c
1874
err = READ_ONCE(sock->ops)->bind(sock,
net/socket.c
1922
somaxconn = READ_ONCE(sock_net(sock->sk)->core.sysctl_somaxconn);
net/socket.c
1928
err = READ_ONCE(sock->ops)->listen(sock, backlog);
net/socket.c
1968
ops = READ_ONCE(sock->ops);
net/socket.c
2089
err = READ_ONCE(sock->ops)->connect(sock, (struct sockaddr_unsized *)address,
net/socket.c
2129
err = READ_ONCE(sock->ops)->getname(sock, (struct sockaddr *)&address, peer);
net/socket.c
2316
ops = READ_ONCE(sock->ops);
net/socket.c
2373
ops = READ_ONCE(sock->ops);
net/socket.c
2432
err = READ_ONCE(sock->ops)->shutdown(sock, how);
net/socket.c
3432
ops = READ_ONCE(sock->ops);
net/socket.c
3517
const struct proto_ops *ops = READ_ONCE(sock->ops);
net/socket.c
3554
return READ_ONCE(sock->ops)->bind(sock, (struct sockaddr_unsized *)&address,
net/socket.c
3569
return READ_ONCE(sock->ops)->listen(sock, backlog);
net/socket.c
3587
const struct proto_ops *ops = READ_ONCE(sock->ops);
net/socket.c
3634
return READ_ONCE(sock->ops)->connect(sock, (struct sockaddr_unsized *)&address,
net/socket.c
3650
return READ_ONCE(sock->ops)->getname(sock, addr, 0);
net/socket.c
3665
return READ_ONCE(sock->ops)->getname(sock, addr, 1);
net/socket.c
3679
return READ_ONCE(sock->ops)->shutdown(sock, how);
net/socket.c
655
const struct proto_ops *ops = READ_ONCE(sock->ops);
net/socket.c
727
int ret = INDIRECT_CALL_INET(READ_ONCE(sock->ops)->sendmsg, inet6_sendmsg,
net/socket.c
818
bool cycles = READ_ONCE(sk->sk_tsflags) & SOF_TIMESTAMPING_BIND_PHC;
net/socket.c
864
u32 tsflags = READ_ONCE(sk->sk_tsflags);
net/socket.c
881
u32 tsflags = READ_ONCE(sk->sk_tsflags);
net/socket.c
900
READ_ONCE(sk->sk_bind_phc));
net/socket.c
963
tsflags = READ_ONCE(sk->sk_tsflags);
net/socket.c
984
READ_ONCE(sk->sk_bind_phc));
net/strparser/strparser.c
336
strp->sk->sk_rcvbuf, READ_ONCE(strp->sk->sk_rcvtimeo));
net/sunrpc/auth_gss/auth_gss.c
2126
if (gss_seq_is_newer(*req->rq_seqnos, READ_ONCE(ctx->gc_seq)))
net/sunrpc/auth_gss/auth_gss.c
2129
seq_xmit = READ_ONCE(ctx->gc_seq_xmit);
net/sunrpc/sched.c
78
unsigned long timeout = READ_ONCE(task->tk_timeout);
net/sunrpc/sched.c
867
queue = READ_ONCE(task->tk_waitqueue);
net/sunrpc/sched.c
878
queue = READ_ONCE(task->tk_waitqueue);
net/sunrpc/sched.c
931
(status = READ_ONCE(task->tk_rpc_status)) != 0) {
net/sunrpc/sched.c
973
if (READ_ONCE(task->tk_rpc_status) != 0) {
net/sunrpc/svc.c
746
ln = READ_ONCE(pool->sp_idle_threads.first);
net/sunrpc/svc_xprt.c
451
xpt_flags = READ_ONCE(xprt->xpt_flags);
net/sunrpc/xprt.c
1130
READ_ONCE(task->tk_rqstp->rq_reply_bytes_recvd) != 0;
net/sunrpc/xprt.c
1137
READ_ONCE(task->tk_rqstp->rq_reply_bytes_recvd) == 0;
net/sunrpc/xprt.c
802
return READ_ONCE(xprt->connect_cookie);
net/sunrpc/xprtmultipath.c
232
if (READ_ONCE(xps->xps_iter_ops) != &rpc_xprt_iter_roundrobin)
net/sunrpc/xprtmultipath.c
467
nactive = READ_ONCE(xps->xps_nactive);
net/sunrpc/xprtrdma/verbs.c
691
if (next_head == READ_ONCE(buf->rb_sc_tail))
net/sunrpc/xprtsock.c
2313
READ_ONCE(net->ipv4.sysctl_tcp_syn_retries), 1);
net/tipc/socket.c
2312
return READ_ONCE(sk->sk_rcvbuf);
net/tipc/socket.c
2315
return READ_ONCE(sk->sk_rcvbuf) << msg_importance(hdr);
net/tipc/socket.c
2318
return READ_ONCE(sk->sk_rcvbuf);
net/tipc/socket.c
316
return READ_ONCE(sk->sk_state) == TIPC_ESTABLISHED;
net/tipc/socket.c
3987
i += scnprintf(buf + i, sz - i, " | %d\n", READ_ONCE(sk->sk_backlog.len));
net/tipc/socket.c
518
sk->sk_rcvbuf = READ_ONCE(sysctl_tipc_rmem[1]);
net/tls/tls.h
221
return READ_ONCE(ctx->strp.msg_ready);
net/tls/tls_device.c
377
READ_ONCE(sk->sk_prot)->enter_memory_pressure(sk);
net/tls/tls_main.c
159
!READ_ONCE(sk->sk_write_pending), &wait);
net/tls/tls_main.c
415
shutdown = READ_ONCE(sk->sk_shutdown);
net/tls/tls_main.c
426
READ_ONCE(ctx->key_update_pending))
net/tls/tls_main.c
923
ctx->sk_proto = READ_ONCE(sk->sk_prot);
net/tls/tls_main.c
974
struct proto *prot = READ_ONCE(sk->sk_prot);
net/tls/tls_sw.c
1984
!READ_ONCE(ctx->reader_present), &wait);
net/tls/tls_sw.c
2665
return READ_ONCE(rec->tx_ready);
net/tls/tls_sw.c
439
if (READ_ONCE(rec->tx_ready)) {
net/unix/af_unix.c
1110
sk->sk_max_ack_backlog = READ_ONCE(net->unx.sysctl_max_dgram_qlen);
net/unix/af_unix.c
1526
if (unix_may_passcred(sk) && !READ_ONCE(unix_sk(sk)->addr)) {
net/unix/af_unix.c
1645
if (unix_may_passcred(sk) && !READ_ONCE(u->addr)) {
net/unix/af_unix.c
1713
state = READ_ONCE(sk->sk_state);
net/unix/af_unix.c
1788
READ_ONCE(other->sk_data_ready)(other);
net/unix/af_unix.c
1847
if (READ_ONCE(sk->sk_state) != TCP_LISTEN)
net/unix/af_unix.c
1931
if (unlikely(READ_ONCE(user->unix_inflight) > task_rlimit(p, RLIMIT_NOFILE)))
net/unix/af_unix.c
2107
if (unix_may_passcred(sk) && !READ_ONCE(u->addr)) {
net/unix/af_unix.c
2113
if (len > READ_ONCE(sk->sk_sndbuf) - 32) {
net/unix/af_unix.c
2283
READ_ONCE(other->sk_data_ready)(other);
net/unix/af_unix.c
2356
READ_ONCE(other->sk_data_ready)(other);
net/unix/af_unix.c
2393
err = READ_ONCE(sk->sk_state) == TCP_ESTABLISHED ? -EISCONN : -EOPNOTSUPP;
net/unix/af_unix.c
2405
if (READ_ONCE(sk->sk_shutdown) & SEND_SHUTDOWN)
net/unix/af_unix.c
2418
size = min_t(int, size, (READ_ONCE(sk->sk_sndbuf) >> 1) - 64);
net/unix/af_unix.c
2482
READ_ONCE(other->sk_data_ready)(other);
net/unix/af_unix.c
2522
if (READ_ONCE(sk->sk_state) != TCP_ESTABLISHED)
net/unix/af_unix.c
2536
if (READ_ONCE(sk->sk_state) != TCP_ESTABLISHED)
net/unix/af_unix.c
2592
(READ_ONCE(sk->sk_shutdown) & RCV_SHUTDOWN))
net/unix/af_unix.c
2667
const struct proto *prot = READ_ONCE(sk->sk_prot);
net/unix/af_unix.c
2800
if (likely(unix_skb_len(skb) && skb != READ_ONCE(u->oob_skb)))
net/unix/af_unix.c
2854
if (unlikely(READ_ONCE(sk->sk_state) != TCP_ESTABLISHED))
net/unix/af_unix.c
2909
if (unlikely(READ_ONCE(sk->sk_state) != TCP_ESTABLISHED)) {
net/unix/af_unix.c
3086
bool do_cmsg = READ_ONCE(u->recvmsg_inq);
net/unix/af_unix.c
3091
msg->msg_inq = READ_ONCE(u->inq_len);
net/unix/af_unix.c
3141
const struct proto *prot = READ_ONCE(sk->sk_prot);
net/unix/af_unix.c
3206
const struct proto *prot = READ_ONCE(other->sk_prot);
net/unix/af_unix.c
3234
if (READ_ONCE(sk->sk_state) == TCP_LISTEN)
net/unix/af_unix.c
3238
return READ_ONCE(unix_sk(sk)->inq_len);
net/unix/af_unix.c
3307
struct sk_buff *oob_skb = READ_ONCE(u->oob_skb);
net/unix/af_unix.c
3347
shutdown = READ_ONCE(sk->sk_shutdown);
net/unix/af_unix.c
3348
state = READ_ONCE(sk->sk_state);
net/unix/af_unix.c
3351
if (READ_ONCE(sk->sk_err))
net/unix/af_unix.c
3364
if (READ_ONCE(unix_sk(sk)->oob_skb))
net/unix/af_unix.c
3394
shutdown = READ_ONCE(sk->sk_shutdown);
net/unix/af_unix.c
3395
state = READ_ONCE(sk->sk_state);
net/unix/af_unix.c
3398
if (READ_ONCE(sk->sk_err) ||
net/unix/af_unix.c
594
(refcount_read(&sk->sk_wmem_alloc) << 2) <= READ_ONCE(sk->sk_sndbuf);
net/unix/af_unix.c
602
if (unix_writable(sk, READ_ONCE(sk->sk_state))) {
net/unix/af_unix.c
820
if (!READ_ONCE(u->addr))
net/unix/af_unix.c
901
s_state = READ_ONCE(sk->sk_state);
net/unix/diag.c
128
rep->udiag_state = READ_ONCE(sk->sk_state);
net/unix/diag.c
157
if (nla_put_u8(skb, UNIX_DIAG_SHUTDOWN, READ_ONCE(sk->sk_shutdown)))
net/unix/diag.c
194
if (!(req->udiag_states & (1 << READ_ONCE(sk->sk_state))))
net/unix/diag.c
68
if (READ_ONCE(sk->sk_state) == TCP_LISTEN) {
net/unix/diag.c
95
if (READ_ONCE(sk->sk_state) == TCP_LISTEN) {
net/unix/garbage.c
103
ops = READ_ONCE(sock->ops);
net/unix/garbage.c
331
if (!READ_ONCE(gc_in_progress))
net/unix/garbage.c
642
if (READ_ONCE(unix_graph_state) == UNIX_GRAPH_NOT_CYCLIC)
net/unix/garbage.c
649
READ_ONCE(user->unix_inflight) < UNIX_INFLIGHT_SANE_USER)
net/unix/garbage.c
652
if (!READ_ONCE(gc_in_progress)) {
net/unix/garbage.c
657
if (user && READ_ONCE(unix_graph_cyclic_sccs))
net/vmw_vsock/af_vsock.c
1503
prot = READ_ONCE(sk->sk_prot);
net/vmw_vsock/af_vsock.c
2576
prot = READ_ONCE(sk->sk_prot);
net/wireless/nl80211.c
20463
u32 nlportid = READ_ONCE(wdev->ap_unexpected_nlportid);
net/wireless/nl80211.c
20679
u32 nlportid = READ_ONCE(wdev->conn_owner_nlportid);
net/x25/af_x25.c
601
sk->sk_priority = READ_ONCE(osk->sk_priority);
net/x25/af_x25.c
894
rc = x25_wait_for_data(sk, READ_ONCE(sk->sk_rcvtimeo));
net/x25/x25_dev.c
58
queued = !sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf));
net/xdp/xsk.c
1014
return READ_ONCE(sk->sk_prefer_busy_poll) && READ_ONCE(sk->sk_ll_usec) &&
net/xdp/xsk.c
1015
napi_id_valid(READ_ONCE(sk->sk_napi_id));
net/xdp/xsk.c
1301
bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
net/xdp/xsk.c
1755
int state = READ_ONCE(xs->state);
net/xdp/xsk.c
1762
q = READ_ONCE(xs->rx);
net/xdp/xsk.c
1764
q = READ_ONCE(xs->tx);
net/xdp/xsk.c
1769
q = state == XSK_READY ? READ_ONCE(xs->fq_tmp) :
net/xdp/xsk.c
1770
READ_ONCE(xs->pool->fq);
net/xdp/xsk.c
1772
q = state == XSK_READY ? READ_ONCE(xs->cq_tmp) :
net/xdp/xsk.c
1773
READ_ONCE(xs->pool->cq);
net/xdp/xsk.c
325
if (READ_ONCE(xs->state) == XSK_BOUND) {
net/xdp/xsk.c
651
skb->priority = READ_ONCE(xs->sk.sk_priority);
net/xdp/xsk.c
652
skb->mark = READ_ONCE(xs->sk.sk_mark);
net/xdp/xsk.c
928
max_batch = READ_ONCE(xs->max_tx_budget);
net/xdp/xsk_diag.c
114
if (READ_ONCE(xs->state) == XSK_UNBOUND)
net/xdp/xsk_queue.h
372
return READ_ONCE(q->ring->producer) - READ_ONCE(q->ring->consumer);
net/xdp/xsk_queue.h
379
return READ_ONCE(q->ring->producer);
net/xdp/xsk_queue.h
390
q->cached_cons = READ_ONCE(q->ring->consumer);
net/xdp/xsk_queue.h
484
return READ_ONCE(q->ring->consumer) == READ_ONCE(q->ring->producer);
net/xfrm/espintcp.c
174
READ_ONCE(net_hotdata.max_backlog)) {
net/xfrm/xfrm_input.c
804
if (skb_queue_len(&trans->queue) >= READ_ONCE(net_hotdata.max_backlog))
net/xfrm/xfrm_interface_core.c
728
return READ_ONCE(xi->p.link);
net/xfrm/xfrm_interface_core.c
925
return READ_ONCE(xi->net);
net/xfrm/xfrm_policy.c
2260
if ((READ_ONCE(sk->sk_mark) & pol->mark.m) != pol->mark.v ||
net/xfrm/xfrm_policy.c
379
(READ_ONCE(xp->curlft.use_time) ? : xp->curlft.add_time) - now;
net/xfrm/xfrm_policy.c
3941
if (READ_ONCE(dst->obsolete) < 0 && !stale_bundle(dst))
net/xfrm/xfrm_policy.c
3969
if (READ_ONCE(dst->obsolete))
net/xfrm/xfrm_policy.c
397
(READ_ONCE(xp->curlft.use_time) ? : xp->curlft.add_time) - now;
net/xfrm/xfrm_state.c
2249
if (READ_ONCE(x1->curlft.use_time))
net/xfrm/xfrm_state.c
2289
if (!READ_ONCE(x->curlft.use_time))
net/xfrm/xfrm_state.c
3114
const struct xfrm_type *type = READ_ONCE(x->type);
net/xfrm/xfrm_state.c
3157
READ_ONCE(xs_net(x)->ipv4.sysctl_ip_no_pmtu_disc))
net/xfrm/xfrm_state.c
674
(READ_ONCE(x->curlft.use_time) ? : now) - now;
net/xfrm/xfrm_state.c
696
(READ_ONCE(x->curlft.use_time) ? : now) - now;
net/xfrm/xfrm_state.c
769
struct net_device *dev = READ_ONCE(xso->dev);
net/xfrm/xfrm_state.c
783
struct net_device *dev = READ_ONCE(xso->dev);
rust/helpers/atomic_ext.c
19
return READ_ONCE(*ptr);
rust/helpers/atomic_ext.c
9
return READ_ONCE(*ptr);
samples/bpf/xdp_sample.bpf.h
116
#define NO_TEAR_ADD(x, val) WRITE_ONCE((x), READ_ONCE(x) + (val))
samples/bpf/xdp_sample_user.c
357
rec->cpu[i].processed = READ_ONCE(values[i].processed);
samples/bpf/xdp_sample_user.c
358
rec->cpu[i].dropped = READ_ONCE(values[i].dropped);
samples/bpf/xdp_sample_user.c
359
rec->cpu[i].issue = READ_ONCE(values[i].issue);
samples/bpf/xdp_sample_user.c
360
rec->cpu[i].xdp_pass = READ_ONCE(values[i].xdp_pass);
samples/bpf/xdp_sample_user.c
361
rec->cpu[i].xdp_drop = READ_ONCE(values[i].xdp_drop);
samples/bpf/xdp_sample_user.c
362
rec->cpu[i].xdp_redirect = READ_ONCE(values[i].xdp_redirect);
security/apparmor/apparmorfs.c
1921
ns = __aa_find_or_create_ns(parent, READ_ONCE(dentry->d_name.name),
security/apparmor/apparmorfs.c
635
READ_ONCE(rev->ns->revision)))
security/apparmor/apparmorfs.c
683
WRITE_ONCE(ns->revision, READ_ONCE(ns->revision) + 1);
security/apparmor/domain.c
437
long rev = READ_ONCE(ns->revision);
security/apparmor/domain.c
447
READ_ONCE(ns->revision))
security/ipe/audit.c
137
if (act != IPE_ACTION_DENY && !READ_ONCE(success_audit))
security/ipe/audit.c
146
op, audit_hook_names[ctx->hook], READ_ONCE(enforce),
security/ipe/eval.c
378
if (!READ_ONCE(enforce))
security/ipe/fs.c
113
result = ((READ_ONCE(enforce)) ? "1" : "0");
security/ipe/fs.c
61
result = ((READ_ONCE(success_audit)) ? "1" : "0");
security/ipe/fs.c
86
old_value = READ_ONCE(enforce);
security/keys/keyring.c
578
unsigned long kflags = READ_ONCE(key->flags);
security/keys/keyring.c
579
short state = READ_ONCE(key->state);
security/keys/keyring.c
591
time64_t expiry = READ_ONCE(key->expiry);
security/keys/keyring.c
732
ptr = READ_ONCE(keyring->keys.root);
security/keys/keyring.c
745
ptr = READ_ONCE(shortcut->next_node);
security/keys/keyring.c
762
ptr = READ_ONCE(shortcut->next_node);
security/keys/keyring.c
773
ptr = READ_ONCE(node->slots[slot]);
security/keys/keyring.c
814
ptr = READ_ONCE(node->back_pointer);
security/keys/keyring.c
819
ptr = READ_ONCE(shortcut->back_pointer);
security/keys/permission.c
104
unsigned long flags = READ_ONCE(key->flags);
security/keys/permission.c
105
time64_t expiry = READ_ONCE(key->expiry);
security/keys/proc.c
200
expiry = READ_ONCE(key->expiry);
security/keys/proc.c
225
flags = READ_ONCE(key->flags);
security/keys/process_keys.c
187
struct key *reg_keyring = READ_ONCE(cred->user_ns->user_keyring_register);
security/keys/process_keys.c
40
struct key *reg_keyring = READ_ONCE(user_ns->user_keyring_register);
security/landlock/audit.c
109
if (READ_ONCE(hierarchy->log_status) == LANDLOCK_LOG_RECORDED)
security/landlock/audit.c
394
if (READ_ONCE(youngest_denied->log_status) == LANDLOCK_LOG_DISABLED)
security/landlock/audit.c
456
if (READ_ONCE(hierarchy->log_status) != LANDLOCK_LOG_RECORDED)
security/selinux/include/security.h
127
return READ_ONCE(selinux_state.enforcing);
security/selinux/include/security.h
153
return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_NETPEER]);
security/selinux/include/security.h
158
return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_OPENPERM]);
security/selinux/include/security.h
163
return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_EXTSOCKCLASS]);
security/selinux/include/security.h
168
return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_ALWAYSNETWORK]);
security/selinux/include/security.h
173
return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_CGROUPSECLABEL]);
security/selinux/include/security.h
178
return READ_ONCE(
security/selinux/include/security.h
184
return READ_ONCE(
security/selinux/include/security.h
190
return READ_ONCE(
security/selinux/include/security.h
196
return READ_ONCE(
security/selinux/include/security.h
202
return READ_ONCE(
security/selinux/include/security.h
208
return READ_ONCE(
security/selinux/include/security.h
214
return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_MEMFD_CLASS]);
security/selinux/include/security.h
219
return READ_ONCE(
security/tomoyo/file.c
369
u16 perm = READ_ONCE(*a_perm);
security/tomoyo/file.c
444
u8 perm = READ_ONCE(*a_perm);
security/tomoyo/file.c
524
u8 perm = READ_ONCE(*a_perm);
security/tomoyo/file.c
663
u8 perm = READ_ONCE(*a_perm);
security/tomoyo/network.c
236
u8 perm = READ_ONCE(*a_perm);
security/tomoyo/network.c
262
u8 perm = READ_ONCE(*a_perm);
security/tomoyo/util.c
1054
if (READ_ONCE(domain->flags[TOMOYO_DIF_QUOTA_WARNED]))
sound/core/pcm_lib.c
2364
appl_ptr = READ_ONCE(runtime->control->appl_ptr);
sound/core/pcm_lib.c
72
snd_pcm_uframes_t appl_ptr = READ_ONCE(runtime->control->appl_ptr);
sound/core/seq/seq_virmidi.c
150
while (READ_ONCE(vmidi->trigger)) {
sound/core/seq/seq_virmidi.c
85
if (!READ_ONCE(vmidi->trigger))
sound/core/ump.c
646
if (!READ_ONCE(ump->stream_finished)) {
sound/firewire/amdtp-stream.c
1139
pcm = READ_ONCE(s->pcm);
sound/firewire/amdtp-stream.c
1190
struct snd_pcm_substream *pcm = READ_ONCE(s->pcm);
sound/firewire/amdtp-stream.c
1872
return READ_ONCE(s->pcm_buffer_pointer);
sound/firewire/amdtp-stream.c
1949
pcm = READ_ONCE(s->pcm);
sound/firewire/amdtp-stream.c
631
struct snd_pcm_substream *pcm = READ_ONCE(s->pcm);
sound/firewire/amdtp-stream.c
680
cip_header[0] = cpu_to_be32(READ_ONCE(s->source_node_id_field) |
sound/firewire/fireface/ff-protocol-former.c
414
substream = READ_ONCE(ff->tx_midi_substreams[0]);
sound/firewire/fireface/ff-protocol-former.c
539
struct snd_rawmidi_substream *substream = READ_ONCE(ff->tx_midi_substreams[port]);
sound/firewire/fireface/ff-protocol-latter.c
431
substream = READ_ONCE(ff->tx_midi_substreams[index]);
sound/firewire/fireface/ff-transaction.c
14
READ_ONCE(ff->rx_midi_substreams[port]);
sound/firewire/fireface/ff-transaction.c
56
READ_ONCE(ff->rx_midi_substreams[port]);
sound/firewire/isight.c
114
if (!READ_ONCE(isight->pcm_running))
sound/firewire/isight.c
134
if (READ_ONCE(isight->pcm_active))
sound/firewire/isight.c
144
if (!READ_ONCE(isight->pcm_running))
sound/firewire/isight.c
431
return READ_ONCE(isight->buffer_pointer);
sound/firewire/motu/amdtp-motu.c
215
struct snd_rawmidi_substream *midi = READ_ONCE(p->midi);
sound/firewire/motu/amdtp-motu.c
247
midi = READ_ONCE(p->midi);
sound/firewire/oxfw/oxfw-scs1x.c
114
stream = READ_ONCE(scs->input);
sound/firewire/oxfw/oxfw-scs1x.c
185
stream = READ_ONCE(scs->output);
sound/firewire/tascam/amdtp-tascam.c
133
bool used = READ_ONCE(tscm->hwdep->used);
sound/firewire/tascam/tascam-transaction.c
150
struct snd_rawmidi_substream *substream = READ_ONCE(port->substream);
sound/firewire/tascam/tascam-transaction.c
175
struct snd_rawmidi_substream *substream = READ_ONCE(port->substream);
sound/firewire/tascam/tascam-transaction.c
284
substream = READ_ONCE(tscm->tx_midi_substreams[port]);
sound/soc/dwc/dwc-pcm.c
111
ptr = READ_ONCE(dev->tx_ptr);
sound/soc/dwc/dwc-pcm.c
116
ptr = READ_ONCE(dev->rx_ptr);
sound/soc/dwc/dwc-pcm.c
235
pos = READ_ONCE(dev->tx_ptr);
sound/soc/dwc/dwc-pcm.c
237
pos = READ_ONCE(dev->rx_ptr);
sound/soc/renesas/rcar/ssi.c
942
*pointer = bytes_to_frames(runtime, READ_ONCE(ssi->byte_pos));
sound/soc/sof/ipc3-dtrace.c
252
u32 host_offset = READ_ONCE(priv->host_offset);
sound/soc/sof/ipc3-dtrace.c
269
loff_t host_offset = READ_ONCE(priv->host_offset);
sound/soc/xtensa/xtfpga-i2s.c
165
unsigned tx_ptr = READ_ONCE(i2s->tx_ptr);
sound/soc/xtensa/xtfpga-i2s.c
461
snd_pcm_uframes_t pos = READ_ONCE(i2s->tx_ptr);
sound/usb/bcd2000/bcd2000.c
102
midi_receive_substream = READ_ONCE(bcd2k->midi_receive_substream);
sound/usb/bcd2000/bcd2000.c
133
midi_out_substream = READ_ONCE(bcd2k->midi_out_substream);
sound/usb/endpoint.c
241
data_subs = READ_ONCE(ep->data_subs);
sound/usb/endpoint.c
267
sync_sink = READ_ONCE(ep->sync_sink);
sound/usb/endpoint.c
338
data_subs = READ_ONCE(ep->data_subs);
sound/usb/endpoint.c
411
data_subs = READ_ONCE(ep->data_subs);
tools/arch/powerpc/include/asm/barrier.h
41
typeof(*p) ___p1 = READ_ONCE(*p); \
tools/arch/riscv/include/asm/barrier.h
34
typeof(*p) ___p1 = READ_ONCE(*p); \
tools/arch/s390/include/asm/barrier.h
39
typeof(*p) ___p1 = READ_ONCE(*p); \
tools/arch/sparc/include/asm/barrier_64.h
51
typeof(*p) ___p1 = READ_ONCE(*p); \
tools/arch/x86/include/asm/atomic.h
29
return READ_ONCE((v)->counter);
tools/arch/x86/include/asm/barrier.h
41
typeof(*p) ___p1 = READ_ONCE(*p); \
tools/include/asm-generic/atomic-gcc.h
26
return READ_ONCE((v)->counter);
tools/include/asm/barrier.h
61
typeof(*p) ___p1 = READ_ONCE(*p); \
tools/include/linux/rbtree.h
39
#define RB_EMPTY_ROOT(root) (READ_ONCE((root)->rb_node) == NULL)
tools/include/linux/ring_buffer.h
61
u64 head = READ_ONCE(base->data_head);
tools/lib/bpf/features.c
660
if (READ_ONCE(cache->res[feat_id]) == FEAT_UNKNOWN) {
tools/lib/bpf/features.c
673
return READ_ONCE(cache->res[feat_id]) == FEAT_SUPPORTED;
tools/lib/bpf/libbpf.c
14284
tmp_cpus = READ_ONCE(cpus);
tools/lib/perf/mmap.c
490
seq = READ_ONCE(pc->lock);
tools/lib/perf/mmap.c
493
count->ena = READ_ONCE(pc->time_enabled);
tools/lib/perf/mmap.c
494
count->run = READ_ONCE(pc->time_running);
tools/lib/perf/mmap.c
498
time_mult = READ_ONCE(pc->time_mult);
tools/lib/perf/mmap.c
499
time_shift = READ_ONCE(pc->time_shift);
tools/lib/perf/mmap.c
500
time_offset = READ_ONCE(pc->time_offset);
tools/lib/perf/mmap.c
503
time_cycles = READ_ONCE(pc->time_cycles);
tools/lib/perf/mmap.c
504
time_mask = READ_ONCE(pc->time_mask);
tools/lib/perf/mmap.c
508
idx = READ_ONCE(pc->index);
tools/lib/perf/mmap.c
509
cnt = READ_ONCE(pc->offset);
tools/lib/perf/mmap.c
512
u16 width = READ_ONCE(pc->pmc_width);
tools/lib/perf/mmap.c
521
} while (READ_ONCE(pc->lock) != seq);
tools/perf/util/auxtrace.c
1886
first = READ_ONCE(pc->aux_head);
tools/perf/util/auxtrace.c
1889
second = READ_ONCE(pc->aux_head);
tools/perf/util/auxtrace.c
1892
last = READ_ONCE(pc->aux_head);
tools/perf/util/auxtrace.h
470
head = READ_ONCE(pc->aux_head);
tools/perf/util/session.h
194
#define session_done() READ_ONCE(session_done)
tools/testing/selftests/bpf/bpf_atomic.h
109
VAL = (__unqual_typeof(*(p)))READ_ONCE(*__ptr); \
tools/testing/selftests/bpf/bpf_atomic.h
126
#define atomic_read(p) READ_ONCE((p)->counter)
tools/testing/selftests/bpf/bpf_atomic.h
89
__unqual_typeof(*(p)) __v = READ_ONCE(*(p)); \
tools/testing/selftests/bpf/map_tests/sk_storage_map.c
103
while (READ_ONCE(sk_storage_map) != -1 && !is_stopped())
tools/testing/selftests/bpf/map_tests/sk_storage_map.c
181
map_fd = READ_ONCE(sk_storage_map);
tools/testing/selftests/bpf/map_tests/sk_storage_map.c
283
int map_fd = READ_ONCE(sk_storage_map);
tools/testing/selftests/bpf/map_tests/sk_storage_map.c
307
int map_fd = READ_ONCE(sk_storage_map);
tools/testing/selftests/bpf/map_tests/sk_storage_map.c
40
return READ_ONCE(stop);
tools/testing/selftests/bpf/map_tests/sk_storage_map.c
45
return READ_ONCE(nr_sk_threads_err);
tools/testing/selftests/bpf/map_tests/sk_storage_map.c
63
return READ_ONCE(nr_sk_threads_done);
tools/testing/selftests/bpf/map_tests/sk_storage_map.c
95
while (READ_ONCE(sk_storage_map) == -1 && !is_stopped())
tools/testing/selftests/bpf/network_helpers.c
698
while (bytes < a->bytes && !READ_ONCE(a->stop)) {
tools/testing/selftests/bpf/network_helpers.c
746
while (bytes < total_bytes && !READ_ONCE(arg.stop)) {
tools/testing/selftests/bpf/prog_tests/fexit_sleep.c
51
while (READ_ONCE(fexit_skel->bss->fentry_cnt) != 2);
tools/testing/selftests/bpf/prog_tests/fexit_sleep.c
52
fexit_cnt = READ_ONCE(fexit_skel->bss->fexit_cnt);
tools/testing/selftests/bpf/prog_tests/fexit_sleep.c
81
fexit_cnt = READ_ONCE(fexit_skel->bss->fexit_cnt);
tools/testing/selftests/bpf/prog_tests/map_kptr.c
130
while (!READ_ONCE(rcu->bss->done))
tools/testing/selftests/bpf/prog_tests/map_kptr_race.c
133
while (!READ_ONCE(watcher->bss->map_freed))
tools/testing/selftests/bpf/prog_tests/map_kptr_race.c
194
while (!READ_ONCE(watcher->bss->map_freed))
tools/testing/selftests/bpf/prog_tests/map_kptr_race.c
72
while (!READ_ONCE(watcher->bss->map_freed))
tools/testing/selftests/bpf/prog_tests/res_spin_lock.c
26
while (!READ_ONCE(skip)) {
tools/testing/selftests/bpf/prog_tests/res_spin_lock.c
82
while (!topts.retval && !err && !READ_ONCE(skel->bss->err)) {
tools/testing/selftests/bpf/prog_tests/res_spin_lock.c
95
ASSERT_EQ(READ_ONCE(skel->bss->err), -EDEADLK, "timeout err");
tools/testing/selftests/bpf/prog_tests/sk_assign.c
258
for (i = 0; i < ARRAY_SIZE(tests) && !READ_ONCE(stop); i++) {
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
482
return READ_ONCE(ctx->done);
tools/testing/selftests/bpf/prog_tests/test_struct_ops_assoc.c
131
while (!READ_ONCE(skel->bss->timer_cb_run))
tools/testing/selftests/bpf/prog_tests/test_struct_ops_assoc.c
173
while (!READ_ONCE(skel->bss->timer_cb_run))
tools/testing/selftests/bpf/prog_tests/timer_lockup.c
35
for (i = 0; !READ_ONCE(*timer1_err) && !READ_ONCE(*timer2_err); i++) {
tools/testing/selftests/bpf/prog_tests/timer_mim.c
24
cnt1 = READ_ONCE(timer_skel->bss->cnt);
tools/testing/selftests/bpf/prog_tests/timer_mim.c
26
cnt2 = READ_ONCE(timer_skel->bss->cnt);
tools/testing/selftests/bpf/prog_tests/timer_mim.c
42
cnt1 = READ_ONCE(timer_skel->bss->cnt);
tools/testing/selftests/bpf/prog_tests/timer_mim.c
45
cnt2 = READ_ONCE(timer_skel->bss->cnt);
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
390
next = READ_ONCE(node->next);
tools/testing/selftests/bpf/progs/lpm_trie_bench.c
105
blackbox = READ_ONCE(key.data);
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
155
if (READ_ONCE(pause)) {
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
175
if (ready == expected && !READ_ONCE(pause)) {
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
199
if (!cpu || READ_ONCE(pause))
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
93
int mode = READ_ONCE(test_mode);
tools/testing/selftests/kvm/access_tracking_perf_test.c
316
if (READ_ONCE(memstress_args.stop_vcpus))
tools/testing/selftests/kvm/access_tracking_perf_test.c
319
*current_iteration = READ_ONCE(iteration);
tools/testing/selftests/kvm/access_tracking_perf_test.c
333
switch (READ_ONCE(iteration_work)) {
tools/testing/selftests/kvm/access_tracking_perf_test.c
349
while (READ_ONCE(vcpu_last_completed_iteration[vcpu_idx]) !=
tools/testing/selftests/kvm/arm64/arch_timer.c
132
irq_iter = READ_ONCE(shared_data->nr_iter);
tools/testing/selftests/kvm/arm64/page_fault_test.c
119
val = READ_ONCE(*guest_test_memory);
tools/testing/selftests/kvm/arm64/page_fault_test.c
127
val = READ_ONCE(*guest_test_memory);
tools/testing/selftests/kvm/arm64/page_fault_test.c
155
val = READ_ONCE(*guest_test_memory);
tools/testing/selftests/kvm/arm64/page_fault_test.c
189
val = READ_ONCE(*guest_test_memory);
tools/testing/selftests/kvm/arm64/page_fault_test.c
88
val = READ_ONCE(*guest_test_memory);
tools/testing/selftests/kvm/arm64/vgic_irq.c
832
} while (READ_ONCE(args->shared_data) == MIN_SPI);
tools/testing/selftests/kvm/arm64/vgic_irq.c
837
} while (READ_ONCE(args->shared_data) != MIN_SPI);
tools/testing/selftests/kvm/arm64/vgic_lpi_stress.c
138
while (!READ_ONCE(test_data.request_vcpus_stop))
tools/testing/selftests/kvm/dirty_log_perf_test.c
159
while (READ_ONCE(vcpu_last_completed_iteration[i]) !=
tools/testing/selftests/kvm/dirty_log_perf_test.c
188
while (READ_ONCE(vcpu_last_completed_iteration[i])
tools/testing/selftests/kvm/dirty_log_perf_test.c
50
while (!READ_ONCE(host_quit)) {
tools/testing/selftests/kvm/dirty_log_perf_test.c
51
int current_iteration = READ_ONCE(iteration);
tools/testing/selftests/kvm/dirty_log_perf_test.c
84
while (current_iteration == READ_ONCE(iteration) &&
tools/testing/selftests/kvm/dirty_log_perf_test.c
85
READ_ONCE(iteration) >= 0 && !READ_ONCE(host_quit)) {}
tools/testing/selftests/kvm/dirty_log_test.c
116
vcpu_arch_put_guest(*(uint64_t *)addr, READ_ONCE(iteration));
tools/testing/selftests/kvm/dirty_log_test.c
122
while (!READ_ONCE(vcpu_stop)) {
tools/testing/selftests/kvm/dirty_log_test.c
128
vcpu_arch_put_guest(*(uint64_t *)addr, READ_ONCE(iteration));
tools/testing/selftests/kvm/dirty_log_test.c
275
if (READ_ONCE(vcpu_stop)) {
tools/testing/selftests/kvm/dirty_log_test.c
486
while (!READ_ONCE(host_quit)) {
tools/testing/selftests/kvm/dirty_log_test.c
751
if (i && !READ_ONCE(dirty_ring_vcpu_ring_full))
tools/testing/selftests/kvm/guest_memfd_test.c
183
TEST_EXPECT_SIGBUS((void)READ_ONCE(mem[accessible_size]));
tools/testing/selftests/kvm/guest_memfd_test.c
186
TEST_ASSERT_EQ(READ_ONCE(mem[i]), val);
tools/testing/selftests/kvm/guest_memfd_test.c
61
TEST_ASSERT_EQ(READ_ONCE(mem[i]), val);
tools/testing/selftests/kvm/guest_memfd_test.c
68
TEST_ASSERT_EQ(READ_ONCE(mem[i]), 0x00);
tools/testing/selftests/kvm/guest_memfd_test.c
70
TEST_ASSERT_EQ(READ_ONCE(mem[i]), val);
tools/testing/selftests/kvm/guest_memfd_test.c
74
TEST_ASSERT_EQ(READ_ONCE(mem[i]), val);
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
65
uint64_t __old = READ_ONCE(mem); \
tools/testing/selftests/kvm/irqfd_test.c
123
kvm_irqfd(vm1, 11, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN);
tools/testing/selftests/kvm/irqfd_test.c
124
kvm_irqfd(vm1, 12, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN);
tools/testing/selftests/kvm/irqfd_test.c
125
kvm_irqfd(vm1, 13, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN);
tools/testing/selftests/kvm/irqfd_test.c
126
kvm_irqfd(vm1, 14, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN);
tools/testing/selftests/kvm/irqfd_test.c
127
kvm_irqfd(vm1, 10, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN);
tools/testing/selftests/kvm/irqfd_test.c
55
while (!READ_ONCE(done)) {
tools/testing/selftests/kvm/irqfd_test.c
56
juggle_eventfd_secondary(vm1, READ_ONCE(__eventfd));
tools/testing/selftests/kvm/irqfd_test.c
57
juggle_eventfd_secondary(vm2, READ_ONCE(__eventfd));
tools/testing/selftests/kvm/kvm_page_table_test.c
118
READ_ONCE(*(uint64_t *)addr);
tools/testing/selftests/kvm/kvm_page_table_test.c
155
READ_ONCE(*(uint64_t *)addr);
tools/testing/selftests/kvm/kvm_page_table_test.c
170
READ_ONCE(*(uint64_t *)addr);
tools/testing/selftests/kvm/kvm_page_table_test.c
194
while (!READ_ONCE(host_quit)) {
tools/testing/selftests/kvm/kvm_page_table_test.c
198
if (READ_ONCE(host_quit))
tools/testing/selftests/kvm/kvm_page_table_test.c
211
stage = READ_ONCE(*current_stage);
tools/testing/selftests/kvm/kvm_page_table_test.c
98
switch (READ_ONCE(*current_stage)) {
tools/testing/selftests/kvm/lib/memstress.c
278
while (!READ_ONCE(all_vcpu_threads_running))
tools/testing/selftests/kvm/lib/memstress.c
305
while (!READ_ONCE(vcpu_threads[i].running))
tools/testing/selftests/kvm/lib/memstress.c
81
READ_ONCE(*(uint64_t *)addr);
tools/testing/selftests/kvm/loongarch/arch_timer.c
106
irq_iter = READ_ONCE(shared_data->nr_iter);
tools/testing/selftests/kvm/loongarch/arch_timer.c
81
irq_iter = READ_ONCE(shared_data->nr_iter);
tools/testing/selftests/kvm/memslot_modification_stress_test.c
44
while (!READ_ONCE(memstress_args.stop_vcpus)) {
tools/testing/selftests/kvm/memslot_perf_test.c
451
uint32_t page_size = (typeof(page_size))READ_ONCE(sync->guest_page_size);
tools/testing/selftests/kvm/memslot_perf_test.c
452
uintptr_t base = (typeof(base))READ_ONCE(sync->move_area_ptr);
tools/testing/selftests/kvm/memslot_perf_test.c
480
uint32_t page_size = (typeof(page_size))READ_ONCE(sync->guest_page_size);
tools/testing/selftests/kvm/memslot_perf_test.c
547
uint32_t page_size = (typeof(page_size))READ_ONCE(sync->guest_page_size);
tools/testing/selftests/kvm/mmu_stress_test.c
60
} while (!READ_ONCE(mprotect_ro_done) || !READ_ONCE(all_vcpus_hit_ro_fault));
tools/testing/selftests/kvm/pre_fault_memory_test.c
50
while (!READ_ONCE(data->prefault_ready))
tools/testing/selftests/kvm/pre_fault_memory_test.c
55
while (!READ_ONCE(data->recreate_slot))
tools/testing/selftests/kvm/pre_fault_memory_test.c
88
while (!READ_ONCE(data.worker_ready))
tools/testing/selftests/kvm/riscv/arch_timer.c
58
irq_iter = READ_ONCE(shared_data->nr_iter);
tools/testing/selftests/kvm/riscv/ebreak_test.c
25
GUEST_ASSERT_EQ(READ_ONCE(sw_bp_addr), LABEL_ADDRESS(sw_bp_2));
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
173
overflown_mask = READ_ONCE(snapshot_data->ctr_overflow_mask);
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
340
counter_value_post = READ_ONCE(snapshot_data->ctr_values[0]);
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
353
counter_value_pre = READ_ONCE(snapshot_data->ctr_values[0]);
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
358
counter_value_post = READ_ONCE(snapshot_data->ctr_values[0]);
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
369
counter_value_post = READ_ONCE(snapshot_data->ctr_values[0]);
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
395
counter_value_post = READ_ONCE(snapshot_data->ctr_values[0]);
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
487
GUEST_ASSERT_EQ(READ_ONCE(snapshot_data->ctr_overflow_mask), 0);
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
490
GUEST_ASSERT_EQ(READ_ONCE(snapshot_data->ctr_values[i]), 0);
tools/testing/selftests/kvm/set_memory_region_test.c
45
val = READ_ONCE(*((uint64_t *)MEM_REGION_GPA));
tools/testing/selftests/kvm/steal_time.c
127
GUEST_ASSERT_EQ(READ_ONCE(st->rev), 0);
tools/testing/selftests/kvm/steal_time.c
128
GUEST_ASSERT_EQ(READ_ONCE(st->attr), 0);
tools/testing/selftests/kvm/steal_time.c
238
GUEST_ASSERT(!(READ_ONCE(st->sequence) & 1));
tools/testing/selftests/kvm/steal_time.c
239
GUEST_ASSERT(READ_ONCE(st->flags) == 0);
tools/testing/selftests/kvm/steal_time.c
240
GUEST_ASSERT(READ_ONCE(st->preempted) == 0);
tools/testing/selftests/kvm/steal_time.c
258
sequence = READ_ONCE(st->sequence);
tools/testing/selftests/kvm/steal_time.c
263
GUEST_ASSERT(sequence < READ_ONCE(st->sequence));
tools/testing/selftests/kvm/steal_time.c
320
GUEST_ASSERT(!(READ_ONCE(st->version) & 1));
tools/testing/selftests/kvm/steal_time.c
321
GUEST_ASSERT_EQ(READ_ONCE(st->flags), 0);
tools/testing/selftests/kvm/steal_time.c
322
GUEST_ASSERT_EQ(READ_ONCE(st->preempted), 0);
tools/testing/selftests/kvm/steal_time.c
335
version = READ_ONCE(st->version);
tools/testing/selftests/kvm/steal_time.c
340
GUEST_ASSERT(version < READ_ONCE(st->version));
tools/testing/selftests/kvm/steal_time.c
37
GUEST_ASSERT(!(READ_ONCE(st->version) & 1));
tools/testing/selftests/kvm/steal_time.c
38
GUEST_ASSERT_EQ(READ_ONCE(st->flags), 0);
tools/testing/selftests/kvm/steal_time.c
39
GUEST_ASSERT_EQ(READ_ONCE(st->preempted), 0);
tools/testing/selftests/kvm/steal_time.c
54
version = READ_ONCE(st->version);
tools/testing/selftests/kvm/steal_time.c
59
GUEST_ASSERT(version < READ_ONCE(st->version));
tools/testing/selftests/kvm/x86/dirty_log_page_splitting_test.c
60
while (READ_ONCE(vcpu_last_completed_iteration[i]) !=
tools/testing/selftests/kvm/x86/dirty_log_page_splitting_test.c
71
while (!READ_ONCE(host_quit)) {
tools/testing/selftests/kvm/x86/dirty_log_page_splitting_test.c
72
int current_iteration = READ_ONCE(iteration);
tools/testing/selftests/kvm/x86/dirty_log_page_splitting_test.c
81
while (current_iteration == READ_ONCE(iteration) &&
tools/testing/selftests/kvm/x86/dirty_log_page_splitting_test.c
82
READ_ONCE(iteration) >= 0 &&
tools/testing/selftests/kvm/x86/dirty_log_page_splitting_test.c
83
!READ_ONCE(host_quit))
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
114
if (expected != READ_ONCE(*this_cpu))
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
85
expected = READ_ONCE(*this_cpu);
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
94
val = READ_ONCE(*(u64 *)data->test_pages);
tools/testing/selftests/kvm/x86/msrs_test.c
194
const struct kvm_msr *msr = &msrs[READ_ONCE(idx)];
tools/testing/selftests/kvm/x86/nested_dirty_log_test.c
55
READ_ONCE(*(u64 *)page0);
tools/testing/selftests/kvm/x86/nested_dirty_log_test.c
59
READ_ONCE(*(u64 *)page0);
tools/testing/selftests/kvm/x86/nested_dirty_log_test.c
66
READ_ONCE(*(u64 *)page1);
tools/testing/selftests/kvm/x86/nx_huge_pages_test.c
53
READ_ONCE(*(uint64_t *)hpage_1);
tools/testing/selftests/kvm/x86/nx_huge_pages_test.c
56
READ_ONCE(*(uint64_t *)hpage_2);
tools/testing/selftests/kvm/x86/nx_huge_pages_test.c
65
READ_ONCE(*(uint64_t *)hpage_1);
tools/testing/selftests/kvm/x86/nx_huge_pages_test.c
68
READ_ONCE(*(uint64_t *)hpage_3);
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
321
while (!READ_ONCE(run_vcpus))
tools/testing/selftests/kvm/x86/tsc_scaling_sync.c
31
tmp = READ_ONCE(tsc_sync);
tools/testing/selftests/livepatch/test_modules/test_klp_callbacks_busy.c
26
while (READ_ONCE(block_transition)) {
tools/testing/selftests/vfio/lib/drivers/dsa/dsa.c
249
status = READ_ONCE(completion->status);
tools/testing/vma/include/dup.h
1024
WRITE_ONCE(mm->total_vm, READ_ONCE(mm->total_vm)+npages);
tools/virtio/linux/uaccess.h
21
x = READ_ONCE(*(__pu_ptr)); \
virt/kvm/coalesced_mmio.c
62
insert = READ_ONCE(ring->last);
virt/kvm/coalesced_mmio.c
64
(insert + 1) % KVM_COALESCED_MMIO_MAX == READ_ONCE(ring->first)) {
virt/kvm/dirty_ring.c
142
next_slot = READ_ONCE(entry->slot);
virt/kvm/dirty_ring.c
143
next_offset = READ_ONCE(entry->offset);
virt/kvm/dirty_ring.c
40
return READ_ONCE(ring->dirty_index) - READ_ONCE(ring->reset_index);
virt/kvm/guest_memfd.c
757
struct file *slot_file = READ_ONCE(slot->gmem.file);
virt/kvm/kvm_main.c
239
cpu = READ_ONCE(vcpu->cpu);
virt/kvm/kvm_main.c
3591
grow_start = READ_ONCE(halt_poll_ns_grow_start);
virt/kvm/kvm_main.c
3592
grow = READ_ONCE(halt_poll_ns_grow);
virt/kvm/kvm_main.c
3610
shrink = READ_ONCE(halt_poll_ns_shrink);
virt/kvm/kvm_main.c
3611
grow_start = READ_ONCE(halt_poll_ns_grow_start);
virt/kvm/kvm_main.c
3715
return READ_ONCE(kvm->max_halt_poll_ns);
virt/kvm/kvm_main.c
3718
return READ_ONCE(halt_poll_ns);
virt/kvm/kvm_main.c
3844
cpu = READ_ONCE(vcpu->cpu);
virt/kvm/kvm_main.c
4001
start = READ_ONCE(kvm->last_boosted_vcpu) + 1;
virt/kvm/kvm_main.c
4008
if (!READ_ONCE(vcpu->ready))
virt/kvm/kvm_main.c
4019
if (READ_ONCE(vcpu->preempted) && yield_to_kernel_mode &&
virt/kvm/kvm_main.c
4475
vcpu->wants_to_run = !READ_ONCE(vcpu->run->immediate_exit__unsafe);