Symbol: data_race
arch/powerpc/kernel/interrupt.c
172
if (!data_race(warned)) {
arch/powerpc/kernel/interrupt.c
173
data_race(warned = true);
arch/x86/mm/pat/set_memory.c
163
data_race(cpa_4k_install++);
drivers/block/loop.c
1860
if (data_race(READ_ONCE(lo->lo_state)) != Lo_bound)
drivers/block/loop.c
2204
data_race(READ_ONCE(lo->lo_state)) == Lo_unbound)
drivers/block/zloop.c
711
if (data_race(READ_ONCE(zlo->state)) == Zlo_deleting)
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1145
if (!data_race(priv->rx.initialized))
drivers/net/wireless/ath/ath9k/wmi.c
159
if (!data_race(priv->initialized)) {
drivers/virtio/virtio_ring.c
3194
data_race(vq->event_triggered = false);
drivers/virtio/virtio_ring.c
3250
data_race(vq->event_triggered = true);
fs/btrfs/block-group.c
1858
return data_race(bg1->used > bg2->used);
fs/btrfs/block-rsv.h
110
return data_race(rsv->full);
fs/btrfs/extent-tree.c
2102
max_count = data_race(delayed_refs->num_heads_ready);
fs/btrfs/file.c
181
if (data_race(args->start >= inode->disk_i_size) && !args->replace_extent)
fs/btrfs/space-info.c
1979
if (data_race(list_empty(&sinfo->ro_bgs)))
fs/btrfs/tree-log.c
3751
if (data_race(inode->logged_trans) == trans->transid)
fs/btrfs/tree-log.c
4634
ASSERT(data_race(BTRFS_I(inode)->logged_trans) > 0);
fs/btrfs/tree-log.c
4635
if (data_race(BTRFS_I(inode)->logged_trans) < trans->transid)
fs/btrfs/tree-log.c
5479
if (data_race(inode->logged_trans) == trans->transid) {
fs/btrfs/zoned.c
3222
data_reloc_bg = data_race(fs_info->data_reloc_bg);
fs/btrfs/zoned.c
3226
treelog_bg = data_race(fs_info->treelog_bg);
fs/inode.c
118
sum += data_race(per_cpu(mg_ctime_updates, i));
fs/inode.c
128
sum += data_race(per_cpu(mg_fine_stamps, i));
fs/inode.c
138
sum += data_race(per_cpu(mg_ctime_swaps, i));
fs/iomap/direct-io.c
532
if (unlikely(data_race(dio->error)))
fs/jbd2/transaction.c
1515
if (data_race(jh->b_transaction != transaction &&
fs/jbd2/transaction.c
1522
if (data_race(jh->b_modified == 1)) {
fs/jbd2/transaction.c
1524
if (data_race(jh->b_transaction == transaction &&
fs/jbd2/transaction.c
344
if (!data_race(journal->j_running_transaction)) {
fs/xfs/xfs_trace.h
1357
__entry->nrefs = data_race(dqp->q_lockref.count);
include/linux/bpf.h
526
data_race(*ldst++ = *lsrc++);
include/linux/console.h
530
return data_race(READ_ONCE(con->flags));
include/linux/mm.h
3139
if (data_race(mm->hiwater_rss) < _rss)
include/linux/mm.h
3140
data_race(mm->hiwater_rss = _rss);
include/linux/page-flags.h
954
return page_type_has_type(data_race(page->page_type));
include/linux/page-flags.h
960
return data_race(folio->page.page_type >> 24) == PGTY_##lname; \
include/linux/page-flags.h
966
VM_BUG_ON_FOLIO(data_race(folio->page.page_type) != UINT_MAX, \
include/linux/page-flags.h
982
return data_race(page->page_type >> 24) == PGTY_##lname; \
include/linux/page-flags.h
988
VM_BUG_ON_PAGE(data_race(page->page_type) != UINT_MAX, page); \
include/linux/rseq_entry.h
163
curr->rseq.slice.expires = data_race(rseq_slice_ext_nsecs) + ktime_get_mono_fast_ns();
include/linux/srcutiny.h
148
idx = ((data_race(READ_ONCE(ssp->srcu_idx)) + 1) & 0x2) >> 1;
include/linux/srcutiny.h
151
data_race(READ_ONCE(ssp->srcu_lock_nesting[!idx])),
include/linux/srcutiny.h
152
data_race(READ_ONCE(ssp->srcu_lock_nesting[idx])),
include/linux/srcutiny.h
153
data_race(READ_ONCE(ssp->srcu_idx)),
include/linux/srcutiny.h
154
data_race(READ_ONCE(ssp->srcu_idx_max)));
include/net/sock.h
2524
if (likely(data_race(!sk->sk_err)))
io_uring/fdinfo.c
87
seq_printf(m, "CachedSqHead:\t%u\n", data_race(ctx->cached_sq_head));
io_uring/fdinfo.c
91
seq_printf(m, "CachedCqTail:\t%u\n", data_race(ctx->cached_cq_tail));
io_uring/refs.h
23
WARN_ON_ONCE(!(data_race(req->flags) & REQ_F_REFCOUNT));
io_uring/timeout.c
615
tail = data_race(ctx->cached_cq_tail) - atomic_read(&ctx->cq_timeouts);
kernel/fork.c
2102
if (data_race(nr_threads >= max_threads))
kernel/irq/irqdesc.c
1010
return data_race(desc->tot_count);
kernel/irq/irqdesc.c
1013
sum += data_race(per_cpu(desc->kstat_irqs->cnt, cpu));
kernel/kcsan/debugfs.c
157
if (data_race(report_filterlist.used == report_filterlist.size)) {
kernel/kcsan/kcsan_test.c
360
static noinline void test_kernel_data_race(void) { data_race(test_var++); }
kernel/locking/locktorture.c
1006
long max = 0, min = statp ? data_race(statp[0].n_lock_acquired) : 0;
kernel/locking/locktorture.c
1011
if (data_race(statp[i].n_lock_fail))
kernel/locking/locktorture.c
1013
cur = data_race(statp[i].n_lock_acquired);
kernel/locking/osq_lock.c
164
if (data_race(prev->next) == node &&
kernel/printk/nbcon.c
1026
con_dropped = data_race(READ_ONCE(con->dropped));
kernel/rcu/rcutorture.c
2779
data_race(n_barrier_successes),
kernel/rcu/rcutorture.c
2780
data_race(n_barrier_attempts),
kernel/rcu/rcutorture.c
2781
data_race(n_rcu_torture_barrier_error));
kernel/rcu/rcutorture.c
2782
pr_cont("read-exits: %ld ", data_race(n_read_exits)); // Statistic.
kernel/rcu/srcutree.c
2058
u0 = data_race(atomic_long_read(&sdp->srcu_ctrs[!idx].srcu_unlocks));
kernel/rcu/srcutree.c
2059
u1 = data_race(atomic_long_read(&sdp->srcu_ctrs[idx].srcu_unlocks));
kernel/rcu/srcutree.c
2067
l0 = data_race(atomic_long_read(&sdp->srcu_ctrs[!idx].srcu_locks));
kernel/rcu/srcutree.c
2068
l1 = data_race(atomic_long_read(&sdp->srcu_ctrs[idx].srcu_locks));
kernel/rcu/tasks.h
1114
data_race(t->rcu_tasks_idle_cpu), cpu);
kernel/rcu/tasks.h
235
int i = data_race(rtp->gp_state); // Let KCSAN detect update races
kernel/rcu/tasks.h
299
rtp->name, data_race(rtp->percpu_enqueue_shift), data_race(rtp->percpu_enqueue_lim),
kernel/rcu/tasks.h
728
if (!data_race(rcu_segcblist_empty(&rtpcp->cblist)))
kernel/rcu/tasks.h
730
if (data_race(rtpcp->urgent_gp))
kernel/rcu/tasks.h
732
if (!data_race(rcu_segcblist_empty(&rtpcp->cblist)) && data_race(rtpcp->urgent_gp))
kernel/rcu/tasks.h
739
tasks_gp_state_getname(rtp), data_race(rtp->gp_state),
kernel/rcu/tasks.h
740
jiffies - data_race(rtp->gp_jiffies),
kernel/rcu/tasks.h
741
data_race(rcu_seq_current(&rtp->tasks_gp_seq)),
kernel/rcu/tasks.h
742
data_race(rtp->n_ipis_fails), data_race(rtp->n_ipis),
kernel/rcu/tasks.h
743
".k"[!!data_race(rtp->kthread_ptr)],
kernel/rcu/tasks.h
761
tt, tf, tst, data_race(rtp->tasks_gp_seq),
kernel/rcu/tasks.h
762
j - data_race(rtp->gp_start), j - data_race(rtp->gp_jiffies),
kernel/rcu/tasks.h
763
data_race(rtp->gp_state), tasks_gp_state_getname(rtp));
kernel/rcu/tasks.h
765
data_race(rtp->percpu_enqueue_shift),
kernel/rcu/tasks.h
766
data_race(rtp->percpu_enqueue_lim),
kernel/rcu/tasks.h
767
data_race(rtp->percpu_dequeue_lim),
kernel/rcu/tasks.h
768
data_race(rtp->percpu_dequeue_gpseq));
kernel/rcu/tasks.h
788
data_race(rtp->barrier_q_seq), j - data_race(rtp->barrier_q_start),
kernel/rcu/tree.c
1074
trace_rcu_grace_period(rcu_state.name, data_race(rcu_state.gp_seq), TPS("newreq"));
kernel/rcu/tree_exp.h
586
"D."[!!data_race(rdp->cpu_no_qs.b.exp)]);
kernel/rcu/tree_exp.h
590
j - jiffies_start, rcu_state.expedited_sequence, data_race(rnp_root->expmask),
kernel/rcu/tree_exp.h
591
".T"[!!data_race(rnp_root->exp_tasks)]);
kernel/rcu/tree_exp.h
605
rnp->level, rnp->grplo, rnp->grphi, data_race(rnp->expmask),
kernel/rcu/tree_exp.h
606
".T"[!!data_race(rnp->exp_tasks)]);
kernel/rcu/tree_plugin.h
883
__func__, READ_ONCE(rnp->gp_tasks), data_race(rnp->boost_tasks),
kernel/rcu/tree_stall.h
1071
__func__, jiffies - data_race(READ_ONCE(rcu_state.gp_start)));
kernel/rcu/tree_stall.h
1075
__func__, jiffies - data_race(READ_ONCE(rcu_state.gp_end)));
kernel/rcu/tree_stall.h
405
if (gp_seq != data_race(rcu_state.gp_seq)) {
kernel/rcu/tree_stall.h
410
if (!(data_race(rnp->qsmask) & leaf_node_cpu_bit(rnp, cpu)))
kernel/rcu/tree_stall.h
561
data_race(rcu_state.n_force_qs) - rcu_state.n_force_qs_gpstart,
kernel/rcu/tree_stall.h
580
data_race(READ_ONCE(rcu_state.gp_flags)),
kernel/rcu/tree_stall.h
582
data_race(READ_ONCE(rcu_state.gp_state)),
kernel/rcu/tree_stall.h
583
gpk ? data_race(READ_ONCE(gpk->__state)) : ~0, cpu);
kernel/rcu/tree_stall.h
592
} else if (!(data_race(READ_ONCE(rdp->mynode->qsmask)) & rdp->grpmask)) {
kernel/rcu/tree_stall.h
623
data_race(READ_ONCE(rcu_state.gp_flags)), // Diagnostic read
kernel/rcu/tree_stall.h
625
data_race(READ_ONCE(gpk->__state)));
kernel/rcu/tree_stall.h
675
data_race(rcu_state.n_online_cpus)); // Diagnostic read
kernel/rcu/tree_stall.h
687
gpa = data_race(READ_ONCE(rcu_state.gp_activity));
kernel/rcu/tree_stall.h
690
data_race(READ_ONCE(jiffies_till_next_fqs)),
kernel/rcu/tree_stall.h
691
data_race(READ_ONCE(rcu_get_root()->qsmask)));
kernel/rcu/tree_stall.h
741
data_race(rcu_state.n_online_cpus)); // Diagnostic read
kernel/rcu/tree_stall.h
899
if (data_race(READ_ONCE(rnp->qsmask))) {
kernel/rcu/tree_stall.h
948
ja = j - data_race(READ_ONCE(rcu_state.gp_activity));
kernel/rcu/tree_stall.h
949
jr = j - data_race(READ_ONCE(rcu_state.gp_req_activity));
kernel/rcu/tree_stall.h
950
js = j - data_race(READ_ONCE(rcu_state.gp_start));
kernel/rcu/tree_stall.h
951
jw = j - data_race(READ_ONCE(rcu_state.gp_wake_time));
kernel/rcu/tree_stall.h
954
data_race(READ_ONCE(rcu_state.gp_state)),
kernel/rcu/tree_stall.h
955
t ? data_race(READ_ONCE(t->__state)) : 0x1ffff, t ? t->rt_priority : 0xffU,
kernel/rcu/tree_stall.h
956
js, ja, jr, jw, (long)data_race(READ_ONCE(rcu_state.gp_wake_seq)),
kernel/rcu/tree_stall.h
957
(long)data_race(READ_ONCE(rcu_state.gp_seq)),
kernel/rcu/tree_stall.h
958
(long)data_race(READ_ONCE(rcu_get_root()->gp_seq_needed)),
kernel/rcu/tree_stall.h
959
data_race(READ_ONCE(rcu_state.gp_max)),
kernel/rcu/tree_stall.h
960
data_race(READ_ONCE(rcu_state.gp_flags)));
kernel/rcu/tree_stall.h
963
!data_race(READ_ONCE(rnp->qsmask)) && !data_race(READ_ONCE(rnp->boost_tasks)) &&
kernel/rcu/tree_stall.h
964
!data_race(READ_ONCE(rnp->exp_tasks)) && !data_race(READ_ONCE(rnp->gp_tasks)))
kernel/rcu/tree_stall.h
968
(long)data_race(READ_ONCE(rnp->gp_seq)),
kernel/rcu/tree_stall.h
969
(long)data_race(READ_ONCE(rnp->gp_seq_needed)),
kernel/rcu/tree_stall.h
970
data_race(READ_ONCE(rnp->qsmask)),
kernel/rcu/tree_stall.h
971
".b"[!!data_race(READ_ONCE(rnp->boost_kthread_task))],
kernel/rcu/tree_stall.h
972
".B"[!!data_race(READ_ONCE(rnp->boost_tasks))],
kernel/rcu/tree_stall.h
973
".E"[!!data_race(READ_ONCE(rnp->exp_tasks))],
kernel/rcu/tree_stall.h
974
".G"[!!data_race(READ_ONCE(rnp->gp_tasks))],
kernel/rcu/tree_stall.h
975
data_race(READ_ONCE(rnp->n_boosts)));
kernel/rcu/tree_stall.h
985
cpu, (long)data_race(READ_ONCE(rdp->gp_seq_needed)));
kernel/rcu/tree_stall.h
990
cbs += data_race(READ_ONCE(rdp->n_cbs_invoked));
kernel/rseq.c
136
stats.exit += data_race(per_cpu(rseq_stats.exit, cpu));
kernel/rseq.c
137
stats.signal += data_race(per_cpu(rseq_stats.signal, cpu));
kernel/rseq.c
138
stats.slowpath += data_race(per_cpu(rseq_stats.slowpath, cpu));
kernel/rseq.c
139
stats.fastpath += data_race(per_cpu(rseq_stats.fastpath, cpu));
kernel/rseq.c
140
stats.ids += data_race(per_cpu(rseq_stats.ids, cpu));
kernel/rseq.c
141
stats.cs += data_race(per_cpu(rseq_stats.cs, cpu));
kernel/rseq.c
142
stats.clear += data_race(per_cpu(rseq_stats.clear, cpu));
kernel/rseq.c
143
stats.fixup += data_race(per_cpu(rseq_stats.fixup, cpu));
kernel/rseq.c
145
stats.s_granted += data_race(per_cpu(rseq_stats.s_granted, cpu));
kernel/rseq.c
146
stats.s_expired += data_race(per_cpu(rseq_stats.s_expired, cpu));
kernel/rseq.c
147
stats.s_revoked += data_race(per_cpu(rseq_stats.s_revoked, cpu));
kernel/rseq.c
148
stats.s_yielded += data_race(per_cpu(rseq_stats.s_yielded, cpu));
kernel/rseq.c
149
stats.s_aborted += data_race(per_cpu(rseq_stats.s_aborted, cpu));
kernel/scftorture.c
190
invoked_count += data_race(per_cpu(scf_invoked_count, cpu));
kernel/sched/core.c
2745
if (data_race(!src->user_cpus_ptr))
kernel/time/timekeeping.c
536
return (ktime_get_mono_fast_ns() + ktime_to_ns(data_race(tk->offs_boot)));
kernel/time/timekeeping.c
553
return (ktime_get_mono_fast_ns() + ktime_to_ns(data_race(tk->offs_tai)));
kernel/time/timekeeping_debug.c
65
sum += data_race(per_cpu(timekeeping_mg_floor_swaps, cpu));
lib/cmpxchg-emu.c
41
ret = data_race(cmpxchg(p32, old32.w, new32.w)); // Overridden above.
lib/debugobjects.c
1175
seq_printf(m, "pool_min_free : %u\n", data_race(pool_global.stats.min_fill));
lib/debugobjects.c
1177
seq_printf(m, "pool_max_used : %u\n", data_race(pool_global.stats.max_used));
lib/group_cpus.c
527
cpumask_copy(npresmsk, data_race(cpu_present_mask));
lib/sbitmap.c
639
data_race(*per_cpu_ptr(sb->alloc_hint, cpu) = tag);
lib/stackdepot.c
618
if (data_race(stackdepot_memcmp(entries, stack->entries, size)))
lib/stackdepot.c
859
seq_printf(seq, "pools: %d\n", data_race(pools_num));
lib/stackdepot.c
861
seq_printf(seq, "%s: %ld\n", counter_names[i], data_race(counters[i]));
lib/test_context-analysis.c
95
data_race(d->counter++); /* no warning */
mm/internal.h
845
if (data_race(list_empty(&folio->_deferred_list)))
mm/kfence/core.c
1288
distance = addr - data_race(meta->addr + meta->size);
mm/kfence/core.c
1294
if (!to_report || distance > data_race(meta->addr) - addr)
mm/memory.c
4437
if (data_race(si->flags & SWP_SYNCHRONOUS_IO))
mm/memory.c
4789
if (data_race(si->flags & SWP_SYNCHRONOUS_IO)) {
mm/memory.c
4972
data_race(si->flags & SWP_STABLE_WRITES)) {
mm/page_counter.c
151
data_race(c->failcnt++);
mm/page_io.c
457
if (data_race(sis->flags & SWP_FS_OPS))
mm/page_io.c
464
else if (data_race(sis->flags & SWP_SYNCHRONOUS_IO))
mm/page_io.c
643
if (data_race(sis->flags & SWP_FS_OPS)) {
mm/percpu.c
3375
return data_race(READ_ONCE(pcpu_nr_populated)) * pcpu_nr_units;
mm/shmem.c
2338
if (data_race(si->flags & SWP_SYNCHRONOUS_IO)) {
mm/slab.h
139
if (data_race(page->page_type >> 24) != PGTY_slab)
mm/slub.c
3092
if (!data_race(barn->nr_empty))
mm/slub.c
3147
if (!data_race(barn->nr_full) && !data_race(barn->nr_empty))
mm/slub.c
3181
if (!data_race(barn->nr_full))
mm/slub.c
3215
if (data_race(barn->nr_full) >= MAX_FULL_SHEAVES)
mm/slub.c
3217
if (!data_race(barn->nr_empty))
mm/slub.c
3744
if (!n || data_race(!n->nr_partial))
mm/slub.c
3768
flc.counters = data_race(READ_ONCE(slab->counters));
mm/slub.c
5079
if (!barn || data_race(barn->nr_full) >= MAX_FULL_SHEAVES ||
mm/slub.c
5830
if (data_race(barn->nr_full) < MAX_FULL_SHEAVES) {
mm/slub.c
5841
if (barn && data_race(barn->nr_empty) < MAX_EMPTY_SHEAVES) {
mm/swap.c
652
if (data_race(folio_batch_count(fbatch))) {
mm/vma_init.c
59
data_race(memcpy(&dest->shared, &src->shared, sizeof(dest->shared)));
mm/vmscan.c
1077
return !data_race(folio_swap_flags(folio) & SWP_FS_OPS);
mm/vmscan.c
1301
if (data_race(!list_empty(&folio->_deferred_list) &&
mm/vmstat.c
1082
blocks = data_race(zone->free_area[order].nr_free);
mm/vmstat.c
1558
seq_printf(m, "%6lu ", data_race(zone->free_area[order].nr_free));
net/9p/trans_fd.c
739
data_race(ts->rd->f_flags |= O_NONBLOCK);
net/9p/trans_fd.c
745
data_race(ts->wr->f_flags |= O_NONBLOCK);
net/core/sock_map.c
1589
id = data_race(prog->aux->id);
net/ipv4/af_inet.c
598
if (data_race(!inet_sk(sk)->inet_num) && inet_autobind(sk))
net/ipv4/af_inet.c
850
if (data_race(!inet_sk(sk)->inet_num) && !sk->sk_prot->no_autobind &&
net/ipv4/inet_diag.c
83
data_race(*(struct in6_addr *)r->id.idiag_src = sk->sk_v6_rcv_saddr);
net/ipv4/inet_diag.c
84
data_race(*(struct in6_addr *)r->id.idiag_dst = sk->sk_v6_daddr);
net/ipv4/tcp.c
4940
data_race(({ md5_update(ctx, key->key, keylen), 0; }));
net/ipv4/tcp_ipv4.c
1384
data_race(memcpy(key->key, newkey, newkeylen));
net/packet/af_packet.c
3105
if (data_race(po->tx_ring.pg_vec))
security/selinux/hooks.c
1708
if (data_race(unlikely(isec->initialized != LABEL_INITIALIZED)))
security/selinux/hooks.c
1727
if (data_race(unlikely(isec->initialized != LABEL_INITIALIZED)))
security/selinux/hooks.c
314
if (data_race(likely(isec->initialized == LABEL_INITIALIZED)))
security/selinux/hooks.c
330
if (data_race(likely(isec->initialized == LABEL_INITIALIZED)))
security/selinux/hooks.c
350
if (data_race(likely(isec->initialized == LABEL_INITIALIZED)))
security/tomoyo/gc.c
668
is_write = data_race(head->write_buf != NULL);
security/tomoyo/util.c
1069
perm = data_race(container_of(ptr, struct tomoyo_path_acl, head)->perm);
security/tomoyo/util.c
1072
perm = data_race(container_of(ptr, struct tomoyo_path2_acl, head)->perm);
security/tomoyo/util.c
1075
perm = data_race(container_of(ptr, struct tomoyo_path_number_acl, head)
security/tomoyo/util.c
1079
perm = data_race(container_of(ptr, struct tomoyo_mkdev_acl, head)->perm);
security/tomoyo/util.c
1082
perm = data_race(container_of(ptr, struct tomoyo_inet_acl, head)->perm);
security/tomoyo/util.c
1085
perm = data_race(container_of(ptr, struct tomoyo_unix_acl, head)->perm);