smp_load_acquire
while (!(smp_load_acquire(lock))) \
VAL = smp_load_acquire(__PTR); \
VAL = smp_load_acquire(__PTR); \
!smp_load_acquire(&has_version_negotiated)) {
if (likely(smp_load_acquire(&dist->ready)))
ldt = smp_load_acquire(¤t->mm->context.ldt);
asid = smp_load_acquire(&mm->context.global_asid);
return smp_load_acquire(&kvm->arch.shadow_root_allocated);
struct hlist_head *page_hash = smp_load_acquire(&kvm->arch.mmu_page_hash);
return smp_load_acquire(&kvm->arch.external_write_tracking_enabled);
bool in_guest_mode = (smp_load_acquire(&vcpu->mode) == IN_GUEST_MODE);
if (likely(smp_load_acquire(&tfms_inited[mode_num])))
return smp_load_acquire(&alloc->pages[index]);
return smp_load_acquire(&alloc->mapped);
if (smp_load_acquire(&vi->data_avail)) {
return !!smp_load_acquire(&__scm);
parent = smp_load_acquire(&fence->parent);
int head = smp_load_acquire(&bg->tx_circ_buf.head);
struct ib_ucontext *ucontext = smp_load_acquire(&ufile->ucontext);
if (!smp_load_acquire(&tx->complete))
head = smp_load_acquire(&tx_ring->head);
slast = smp_load_acquire(&qp->s_last);
prod = smp_load_acquire(&q->buf->producer_index);
prod = smp_load_acquire(&q->buf->producer_index);
cons = smp_load_acquire(&q->buf->consumer_index);
cons = smp_load_acquire(&q->buf->consumer_index);
input = smp_load_acquire(&applespi->touchpad_input_dev);
if (!smp_load_acquire(&b->state)) /* fast case */
if (likely(!smp_load_acquire(&b->state))) {
recalc_sector = le64_to_cpu(smp_load_acquire(&ic->sb->recalc_sector));
if (!did_inc || smp_load_acquire(&conf->quiesce) != 0) {
smp_store_release(&rbuf->pread, smp_load_acquire(&rbuf->pwrite));
return (rbuf->pread == smp_load_acquire(&rbuf->pwrite));
avail = smp_load_acquire(&rbuf->pwrite) - rbuf->pread;
if (smp_load_acquire(&dev->flush_status))
u32 ntc = smp_load_acquire(&ring->ntc);
u32 ntu = smp_load_acquire(&ring->ntu);
int ltu = smp_load_acquire(&ring->last_to_use);
ntc = smp_load_acquire(&tx_spare->last_to_clean);
int begin = smp_load_acquire(&ring->next_to_clean);
unsigned int clean = smp_load_acquire(&(R)->next_to_clean); \
u16 htb_maj_id = smp_load_acquire(&pf->qos.maj_id);
return NUM_QOS_SQ_STATS * smp_load_acquire(&priv->htb_max_qos_sqs);
u16 max_qos_sqs = smp_load_acquire(&priv->htb_max_qos_sqs);
max_qos_sqs = smp_load_acquire(&priv->htb_max_qos_sqs);
max_qos_sqs = smp_load_acquire(&priv->htb_max_qos_sqs);
sign = smp_load_acquire(&clock_info->sign);
if (smp_load_acquire(&comp_done->done) == 1) { /* ^^^ */
if (!smp_load_acquire(&nsim_bus_dev->init))
if (!smp_load_acquire(&nsim_bus_enable)) {
if (!smp_load_acquire(&nsim_bus_enable)) {
if (!smp_load_acquire(&nsim_bus_dev->init))
struct sk_buff *tail = queue->tail, *next = smp_load_acquire(&NEXT(tail));
next = smp_load_acquire(&NEXT(next));
next = smp_load_acquire(&NEXT(tail));
queue_state = smp_load_acquire(&queue->rcv_state);
size_t tail = smp_load_acquire(&ldata->read_tail);
size_t head = smp_load_acquire(&ldata->commit_head);
canon_head = smp_load_acquire(&ldata->canon_head);
canon_head = smp_load_acquire(&ldata->canon_head);
while ((next = smp_load_acquire(&buf->head->next)) != NULL) {
next = smp_load_acquire(&head->next);
count = smp_load_acquire(&head->commit) - head->lookahead;
next = smp_load_acquire(&head->next);
count = smp_load_acquire(&head->commit) - head->read;
if (!smp_load_acquire(&waiter.task))
epfile = smp_load_acquire(&inode->i_private);
bh && smp_load_acquire(&bh->state) >=
bh && smp_load_acquire(&bh->state) >=
if (!smp_load_acquire(&df->access_granted))
if (!smp_load_acquire(&df->access_granted))
if (!smp_load_acquire(&df->access_granted))
if (!smp_load_acquire(&df->access_granted))
if (smp_load_acquire(&req->processed) ==
smp_load_acquire(&alist->addr_pref_version) == preflist->version)
smp_load_acquire(&net->address_pref_version) == alist->addr_pref_version)
state = smp_load_acquire(&cell->state); /* vs error */
state = smp_load_acquire(&cell->state); /* vs error */
state = smp_load_acquire(&cell->state); /* vs error */
if (smp_load_acquire(&myself.have_lock) || /* The lock barrier */
if (smp_load_acquire(&myself.have_lock)) /* The lock barrier */
dns_lookup_count = smp_load_acquire(&cell->dns_lookup_count);
smp_load_acquire(&cell->dns_lookup_count)
head = smp_load_acquire(&req->head);
misc = smp_load_acquire(&user_ns->binfmt_misc);
if (likely(smp_load_acquire(&fscrypt_bounce_page_pool)))
return smp_load_acquire(&prep_key->blk_key) != NULL;
return smp_load_acquire(&prep_key->tfm) != NULL;
return smp_load_acquire(&prep_key->tfm) != NULL;
keyring = smp_load_acquire(&sb->s_master_keys);
if (!smp_load_acquire(&mk->mk_ino_hash_key_initialized)) {
const char *dname = smp_load_acquire(&name->name); /* ^^^ */
seq = smp_load_acquire(&parent->d_inode->i_dir_seq);
whead = smp_load_acquire(&pwq->whead);
if (!smp_load_acquire(&v->next))
__set_task_comm(me, smp_load_acquire(&bprm->file->f_path.dentry->d_name.name),
struct fuse_ring *ring = smp_load_acquire(&fc->ring);
cns = smp_load_acquire(&inode->i_ctime_nsec);
cns = smp_load_acquire(&inode->i_ctime_nsec);
cur = smp_load_acquire(&inode->i_ctime_nsec);
if (!smp_load_acquire(&waiter->flc_blocker) &&
flags = smp_load_acquire(&path->dentry->d_flags);
flags = smp_load_acquire(&path->dentry->d_flags);
unsigned flags = smp_load_acquire(&path->dentry->d_flags);
if (!IS_ERR(ret) && d_flags_negative(smp_load_acquire(&ret->d_flags))) {
if (!IS_ERR(ret) && d_flags_negative(smp_load_acquire(&ret->d_flags))) {
if (!IS_ERR(ret) && d_flags_negative(smp_load_acquire(&ret->d_flags))) {
return smp_load_acquire(&cache->state);
if (!smp_load_acquire(&stream->active))
smp_load_acquire(&dentry->d_fsdata)
if (!(smp_load_acquire(&nfsi->cache_validity) & NFS_INO_INVALID_DATA))
smp_load_acquire(&rqstp->rq_status_counter);
if (smp_load_acquire(&rqstp->rq_status_counter) !=
if (!IS_ERR(ret) && d_flags_negative(smp_load_acquire(&ret->d_flags))) {
unsigned int head = smp_load_acquire(&pipe->head);
return smp_load_acquire(&sb->s_flags) & flags;
return smp_load_acquire(&ifp->if_needextents) != 0;
#ifndef smp_load_acquire
#ifndef smp_load_acquire
return 1UL & (smp_load_acquire(p) >> (nr & (BITS_PER_LONG-1)));
ret = smp_load_acquire(&(v)->counter);
ret = smp_load_acquire(&(v)->counter);
if (likely(!(smp_load_acquire(&inode->i_opflags) & IOP_FLCTX)))
return smp_load_acquire(&inode->i_size);
return smp_load_acquire(&cookie->state);
return smp_load_acquire(fscrypt_inode_info_addr(inode));
return smp_load_acquire(&key->state);
struct list_head *next = smp_load_acquire(&head->next);
smp_load_acquire(&q->ready) == NULL;
smp_load_acquire(&q->ready) == NULL;
return smp_load_acquire(&q->ready) == NULL && llist_empty(&q->new);
uint32_t head = smp_load_acquire(&slot->head);
unsigned seq = smp_load_acquire(&s->seqcount.sequence); \
seq = smp_load_acquire(&s->seqcount.sequence); \
return smp_load_acquire(&s->sequence);
return smp_load_acquire(&sk->sk_state);
return smp_load_acquire(&sk->sk_pacing_status) == SK_PACING_NEEDED;
(smp_load_acquire(&sk->sk_validate_xmit_skb) ==
smp_load_acquire(&sk->sk_destruct) != tls_device_sk_destruct)
#define RDMA_READ_UAPI_ATOMIC(member) smp_load_acquire(&(member).val)
if (unlikely(smp_load_acquire(&ctx->flags) & IORING_SETUP_R_DISABLED))
entries = smp_load_acquire(&rings->sq.tail) - ctx->cached_sq_head;
tail = smp_load_acquire(&br->tail);
tail = smp_load_acquire(&br->tail);
if (smp_load_acquire(&target_ctx->flags) & IORING_SETUP_R_DISABLED)
if (smp_load_acquire(&target_ctx->flags) & IORING_SETUP_R_DISABLED)
struct wait_queue_head *head = smp_load_acquire(&poll->head);
head = smp_load_acquire(&poll->head);
if (!smp_load_acquire(&req->iopoll_completed))
head = smp_load_acquire(&iw->head);
entries = smp_load_acquire(&ifq->rq_ring->tail) - ifq->cached_rq_head;
if (!smp_load_acquire(&sma->use_global_lock)) {
smp_load_acquire(&st_map->kvalue.common.state) == BPF_STRUCT_OPS_STATE_READY;
state = smp_load_acquire(&kvalue->common.state);
DST = (SIZE)smp_load_acquire( \
cons_pos = smp_load_acquire(&rb->consumer_pos);
over_pos = smp_load_acquire(&rb->overwrite_pos);
prod_pos = smp_load_acquire(&rb->producer_pos);
prod_pos = smp_load_acquire(&rb->producer_pos);
cons_pos = smp_load_acquire(&rb->consumer_pos);
cons_pos = smp_load_acquire(&rb->consumer_pos) & rb->mask;
return smp_load_acquire(&rb->consumer_pos);
return smp_load_acquire(&rb->producer_pos);
return smp_load_acquire(&rb->overwrite_pos);
prod_pos = smp_load_acquire(&rb->producer_pos);
cons_pos = smp_load_acquire(&rb->consumer_pos);
hdr_len = smp_load_acquire(hdr);
if (!smp_load_acquire(&event->ctx))
if (smp_load_acquire(&fph->state) == FR_PERCPU)
func = smp_load_acquire(&access_kernels[idx]);
func = smp_load_acquire(&access_kernels[idx]);
KCSAN_EXPECT_READ_BARRIER(smp_load_acquire(&test_var), false);
KCSAN_EXPECT_WRITE_BARRIER(smp_load_acquire(&test_var), false);
KCSAN_EXPECT_RW_BARRIER(smp_load_acquire(&test_var), false);
if (!smp_load_acquire(&crcp->crc_stop)) {
if (smp_load_acquire(&node->locked))
if (!smp_load_acquire(&wq_entry.private))
if (!smp_load_acquire(&waiter.task)) {
smp_load_acquire(&rtrcp->rtc_chkrdr) < 0 && // Pairs with smp_store_release below.
!smp_load_acquire(&rtrcp_chker->rtc_assigner)) { // Pairs with smp_store_release below.
if (!rtrcp_assigner || !smp_load_acquire(&rtrcp_assigner->rtc_ready))
if (!smp_load_acquire(&rtorsup->rtorsu_inuse))
j = smp_load_acquire(&jiffies); // Time before ->rtorsu_inuse.
if (smp_load_acquire(&rtorsup->rtorsu_inuse)) {
smp_load_acquire(&barrier_phase)) != lastphase ||
wait_event(read_exit_wq, smp_load_acquire(&read_exit_child_stopped));
wait_event(rt->wq, (atomic_read(&nreaders_exp) && smp_load_acquire(&rt->start_reader)) ||
x = smp_load_acquire(this_cpu_ptr(&test_acqrel));
x = smp_load_acquire(this_cpu_ptr(&test_acqrel));
if (smp_load_acquire(&sup->srcu_size_state) < SRCU_SIZE_WAIT_BARRIER)
ss_state = smp_load_acquire(&ssp->srcu_sup->srcu_size_state);
if (smp_load_acquire(&ssp->srcu_sup->srcu_size_state) < SRCU_SIZE_WAIT_BARRIER)
idx = rcu_seq_state(smp_load_acquire(&ssp->srcu_sup->srcu_gp_seq)); /* ^^^ */
if (!rcu_seq_state(smp_load_acquire(&ssp->srcu_sup->srcu_gp_seq_needed)) &&
if (smp_load_acquire(&ssp->srcu_sup->srcu_size_state) != SRCU_SIZE_SMALL)
if (smp_load_acquire(&ssp->srcu_sup->srcu_size_state) != SRCU_SIZE_SMALL) {
if (!rcu_seq_state(smp_load_acquire(&ssp->srcu_sup->srcu_gp_seq_needed))) /*^^^*/
ss_state = smp_load_acquire(&sup->srcu_size_state);
bool havekthread = smp_load_acquire(&rtp->kthread_ptr);
if (cpu >= smp_load_acquire(&rtp->percpu_dequeue_lim))
dequeue_limit = smp_load_acquire(&rtp->percpu_dequeue_lim);
if (rtpcp_next->cpu < smp_load_acquire(&rtp->percpu_dequeue_lim)) {
if (rtpcp_next->cpu < smp_load_acquire(&rtp->percpu_dequeue_lim)) {
done = smp_load_acquire(&rcu_state.srs_done_tail);
if (smp_load_acquire(this_cpu_ptr(&rcu_data.rcu_urgent_qs))) {
if (smp_load_acquire(&rdp->barrier_seq_snap) == gseq)
return smp_load_acquire(&rdp->beenonline);
int ncpus = smp_load_acquire(&rcu_state.ncpus); /* Order vs. locking. */
if (!smp_load_acquire(this_cpu_ptr(&rcu_data.rcu_urgent_qs)))
if (!smp_load_acquire(this_cpu_ptr(&rcu_data.rcu_urgent_qs))) {
!smp_load_acquire(&rcu_state.gp_kthread)) // Get stable kthread.
gp_state = smp_load_acquire(&rcu_state.gp_state);
struct inode *inode = smp_load_acquire(&iomem_inode);
return smp_load_acquire(&iomem_inode)->i_mapping;
if (smp_load_acquire(&p->on_cpu) &&
smp_load_acquire(&cpu_rq(cpu)->scx.kick_sync) != ksyncs[cpu]) {
if (smp_load_acquire(&rq->scx.flags) & SCX_RQ_CLK_VALID) {
t = smp_load_acquire(trigger_ptr);
cpu_cur_csd = smp_load_acquire(&per_cpu(cur_csd, cpux)); /* Before func and info. */
stats = smp_load_acquire(&sig->stats);
if (ktime_before(now, smp_load_acquire(&tick_next_period)))
return smp_load_acquire(&tp->event->flags);
int *map = smp_load_acquire(&tgid_map);
entry = smp_load_acquire(&head->first);
entry = smp_load_acquire(&head->first);
flags = smp_load_acquire(&wait->flags);
flags = smp_load_acquire(&wait->flags);
KUNIT_EXPECT_KASAN_FAIL_READ(test, smp_load_acquire(i_unsafe));
if (!smp_load_acquire(&kfence_metadata))
subscriptions = smp_load_acquire(&mm->notifier_subscriptions);
stack = smp_load_acquire(&stack_list);
this_gen = smp_load_acquire(&lru_drain_gen);
icsk_pending = smp_load_acquire(&icsk->icsk_pending);
if (unlikely(ops != smp_load_acquire(&tcpv6_prot_saved))) {
icsk_pending = smp_load_acquire(&icsk->icsk_pending);
unsigned long flags = smp_load_acquire(&sk->sk_tsq_flags);
oval = smp_load_acquire(&sk->sk_tsq_flags);
if (!(smp_load_acquire(&icsk->icsk_ack.pending) & ICSK_ACK_TIMER) &&
if (!smp_load_acquire(&inet_csk(sk)->icsk_pending))
if (unlikely(ops != smp_load_acquire(&udpv6_prot_saved))) {
icsk_pending = smp_load_acquire(&icsk->icsk_pending);
if (likely(smp_load_acquire(&nf_conntrack_locks_all) == false))
return smp_load_acquire(&net->nft.base_seq);
return smp_load_acquire(&call->_state);
return smp_load_acquire(&conn->state) == RXRPC_CONN_ABORTED;
call_head = smp_load_acquire(&b->call_backlog_head);
conn_head = smp_load_acquire(&b->conn_backlog_head);
peer_head = smp_load_acquire(&b->peer_backlog_head);
send_top = smp_load_acquire(&call->send_top);
abort_code = smp_load_acquire(&call->send_abort);
old = smp_load_acquire(cursor);
unlikely(prot != smp_load_acquire(&saved_tcpv6_prot))) {
unlikely(prot != smp_load_acquire(&saved_tcpv4_prot))) {
addr = smp_load_acquire(&unix_sk(sk)->addr);
struct unix_address *addr = smp_load_acquire(&unix_sk(sk)->addr);
if (!smp_load_acquire(&unix_sk(sk)->addr))
struct unix_address *addr = smp_load_acquire(&unix_sk(sk)->addr);
if (unlikely(ops != smp_load_acquire(&unix_dgram_prot_saved))) {
if (unlikely(ops != smp_load_acquire(&unix_stream_prot_saved))) {
if (unlikely(ops != smp_load_acquire(&vsock_prot_saved))) {
q->cached_prod = smp_load_acquire(&q->ring->producer); /* C, matches B */
return smp_load_acquire(ptr);
return smp_load_acquire(ptr);
addr = smp_load_acquire(&u->addr);
addr = smp_load_acquire(&u->addr);
return smp_load_acquire(&selinux_state.initialized);
sid = smp_load_acquire(&c->sid[index]);
u32 count = smp_load_acquire(&s->count);
#ifndef smp_load_acquire
return smp_load_acquire(&base->data_head);
cons_pos = smp_load_acquire(r->consumer_pos);
prod_pos = smp_load_acquire(r->producer_pos);
len = smp_load_acquire(len_ptr);
return smp_load_acquire(r->consumer_pos);
return smp_load_acquire(r->producer_pos);
cons_pos = smp_load_acquire(rb->consumer_pos);
prod_pos = smp_load_acquire(rb->producer_pos);
cons_pos = smp_load_acquire(r->consumer_pos);
prod_pos = smp_load_acquire(r->producer_pos);
len = smp_load_acquire(len_ptr);
return smp_load_acquire(&gfn->flags) == KVM_DIRTY_GFN_F_DIRTY;
return smp_load_acquire(&gfn->flags) & KVM_DIRTY_GFN_F_RESET;