Symbol: READ_ONCE
sys/arch/amd64/amd64/vmm_machdep.c
508
ci = READ_ONCE(vcpu->vc_curcpu);
sys/arch/arm64/arm64/pmap.c
2409
(pm->pm_asid & ~PMAP_ASID_MASK) != READ_ONCE(pmap_asid_gen))
sys/arch/octeon/dev/if_ogx.c
1228
txfree = READ_ONCE(sc->sc_txfree);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1633
if (!READ_ONCE(pinfo->block_mmu_notifications)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2491
if (READ_ONCE(process_info->block_mmu_notifications))
sys/dev/pci/drm/amd/amdgpu/amdgpu_fence.c
313
uint64_t seq = READ_ONCE(ring->fence_drv.sync_seq);
sys/dev/pci/drm/amd/amdgpu/amdgpu_fence.c
373
emitted += READ_ONCE(ring->fence_drv.sync_seq);
sys/dev/pci/drm/amd/amdgpu/amdgpu_fence.c
391
sync_seq = READ_ONCE(ring->fence_drv.sync_seq);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ih.c
183
if (wptr != READ_ONCE(ih->rptr)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_object.c
1680
pin_count = READ_ONCE(bo->tbo.pin_count);
sys/dev/pci/drm/amd/amdgpu/amdgpu_object.c
1684
dma_buf = READ_ONCE(bo->tbo.base.dma_buf);
sys/dev/pci/drm/amd/amdgpu/amdgpu_object.c
1685
attachment = READ_ONCE(bo->tbo.base.import_attach);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ring_mux.c
561
chunk->sync_seq = READ_ONCE(ring->fence_drv.sync_seq);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
2649
!READ_ONCE(adev->barrier_has_auto_waitcnt));
sys/dev/pci/drm/amd/amdgpu/sdma_v4_0.c
678
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
sys/dev/pci/drm/amd/amdgpu/sdma_v4_0.c
747
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
sys/dev/pci/drm/amd/amdgpu/sdma_v4_4_2.c
225
rptr = READ_ONCE(*((u64 *)&ring->adev->wb.wb[ring->rptr_offs]));
sys/dev/pci/drm/amd/amdgpu/sdma_v4_4_2.c
245
wptr = READ_ONCE(*((u64 *)&adev->wb.wb[ring->wptr_offs]));
sys/dev/pci/drm/amd/amdgpu/sdma_v4_4_2.c
314
wptr = READ_ONCE(*((u64 *)&adev->wb.wb[ring->wptr_offs]));
sys/dev/pci/drm/amd/amdgpu/sdma_v5_0.c
352
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
sys/dev/pci/drm/amd/amdgpu/sdma_v5_2.c
192
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
sys/dev/pci/drm/amd/amdgpu/sdma_v6_0.c
192
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
sys/dev/pci/drm/amd/amdgpu/sdma_v7_0.c
192
wptr = READ_ONCE(*((u64 *)ring->wptr_cpu_addr));
sys/dev/pci/drm/amd/amdkfd/kfd_events.c
779
if (READ_ONCE(slots[id]) != UNSIGNALED_EVENT_SLOT)
sys/dev/pci/drm/amd/amdkfd/kfd_events.c
788
if (READ_ONCE(slots[id]) != UNSIGNALED_EVENT_SLOT) {
sys/dev/pci/drm/amd/amdkfd/kfd_events.c
862
if (!READ_ONCE(event_waiters[i].event))
sys/dev/pci/drm/amd/amdkfd/kfd_events.c
865
if (READ_ONCE(event_waiters[i].activated)) {
sys/dev/pci/drm/amd/amdkfd/kfd_migrate.c
965
if (READ_ONCE(p->svms.faulting_task) == current) {
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2191
!READ_ONCE(pdd->process->irq_drain_is_open));
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
456
return sysfs_emit(buf, "%llu\n", READ_ONCE(pdd->faults));
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
461
return sysfs_emit(buf, "%llu\n", READ_ONCE(pdd->page_in));
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
466
return sysfs_emit(buf, "%llu\n", READ_ONCE(pdd->page_out));
sys/dev/pci/drm/amd/amdkfd/kfd_smi_events.c
166
uint64_t events = READ_ONCE(client->events);
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
2091
max_pages = READ_ONCE(max_svm_range_pages);
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
2229
READ_ONCE(max_svm_range_pages),
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
2242
READ_ONCE(max_svm_range_pages),
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
496
if (READ_ONCE(prange->svm_bo->evicting)) {
sys/dev/pci/drm/amd/display/dmub/inc/dmub_cmd.h
6729
(void)READ_ONCE(*data++);
sys/dev/pci/drm/display/drm_dp_helper.c
720
return !aux->is_remote && !READ_ONCE(aux->dpcd_probe_disabled);
sys/dev/pci/drm/display/drm_dp_mst_topology.c
1260
state = READ_ONCE(txmsg->state);
sys/dev/pci/drm/drm_fb_helper.c
226
if (READ_ONCE(fb_helper->deferred_setup))
sys/dev/pci/drm/drm_gem.c
483
if (drm_WARN_ON(dev, READ_ONCE(obj->handle_count) == 0))
sys/dev/pci/drm/drm_vblank.c
1479
drm_WARN_ON(dev, !READ_ONCE(vblank->inmodeset));
sys/dev/pci/drm/drm_vblank.c
1660
if (!READ_ONCE(vblank->enabled)) {
sys/dev/pci/drm/drm_vblank.c
1810
READ_ONCE(vblank->enabled)) {
sys/dev/pci/drm/drm_vblank.c
1859
!READ_ONCE(vblank->enabled),
sys/dev/pci/drm/drm_vblank.c
2046
vblank_enabled = READ_ONCE(vblank->config.disable_immediate) &&
sys/dev/pci/drm/drm_vblank.c
2047
READ_ONCE(vblank->enabled);
sys/dev/pci/drm/drm_vblank.c
2147
if (!READ_ONCE(vblank->enabled)) {
sys/dev/pci/drm/drm_vblank.c
507
drm_WARN_ON(dev, READ_ONCE(vblank->enabled) &&
sys/dev/pci/drm/i915/display/intel_display_irq.c
1709
int vblank_enable_count = READ_ONCE(display->irq.vblank_enable_count);
sys/dev/pci/drm/i915/display/intel_dp.c
3620
if (READ_ONCE(intel_dp->oui_valid))
sys/dev/pci/drm/i915/display/intel_fbc.c
2020
if (READ_ONCE(fbc->underrun_detected))
sys/dev/pci/drm/i915/display/intel_hdmi.c
1840
READ_ONCE(to_intel_digital_connector_state(conn_state)->force_audio) != HDMI_AUDIO_OFF_DVI;
sys/dev/pci/drm/i915/display/intel_hotplug.c
793
enabled = READ_ONCE(display->hotplug.poll_enabled);
sys/dev/pci/drm/i915/display/intel_hotplug.c
804
READ_ONCE(display->hotplug.poll_enabled));
sys/dev/pci/drm/i915/display/intel_psr.c
3306
if (READ_ONCE(intel_dp->psr.irq_aux_error)) {
sys/dev/pci/drm/i915/display/intel_psr.c
4237
*val = READ_ONCE(intel_dp->psr.debug);
sys/dev/pci/drm/i915/display/intel_psr.c
907
READ_ONCE(vblank->enabled);
sys/dev/pci/drm/i915/display/intel_sdvo.c
1322
READ_ONCE(to_intel_digital_connector_state(conn_state)->force_audio) != HDMI_AUDIO_OFF_DVI;
sys/dev/pci/drm/i915/gem/i915_gem_domain.c
633
if (READ_ONCE(obj->write_domain) == read_domains)
sys/dev/pci/drm/i915/gem/i915_gem_lmem.c
44
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
sys/dev/pci/drm/i915/gem/i915_gem_object.c
634
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
sys/dev/pci/drm/i915/gem/i915_gem_object.h
311
return READ_ONCE(obj->frontbuffer) || obj->is_dpt;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
661
return !IS_ERR_OR_NULL(READ_ONCE(obj->mm.pages));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
670
if (n < READ_ONCE(iter->sg_idx))
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
294
count = READ_ONCE(i915->mm.shrink_memory) >> PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
295
num_objects = READ_ONCE(i915->mm.shrink_count);
sys/dev/pci/drm/i915/gem/i915_gem_tiling.c
440
READ_ONCE(obj->tiling_and_stride) & TILING_MASK;
sys/dev/pci/drm/i915/gt/intel_breadcrumbs.c
208
if (!signal && READ_ONCE(b->irq_armed) && list_empty(&b->signalers))
sys/dev/pci/drm/i915/gt/intel_breadcrumbs.c
269
if (!READ_ONCE(b->irq_armed) && !list_empty(&b->signalers))
sys/dev/pci/drm/i915/gt/intel_breadcrumbs.c
273
if (READ_ONCE(b->irq_armed) && !atomic_read(&b->active))
sys/dev/pci/drm/i915/gt/intel_breadcrumbs.c
321
if (!READ_ONCE(b->irq_armed))
sys/dev/pci/drm/i915/gt/intel_breadcrumbs.c
352
struct intel_breadcrumbs *b = READ_ONCE(rq->engine)->breadcrumbs;
sys/dev/pci/drm/i915/gt/intel_breadcrumbs.c
408
if (!READ_ONCE(b->irq_armed) || __i915_request_is_complete(rq))
sys/dev/pci/drm/i915/gt/intel_breadcrumbs.c
439
struct intel_breadcrumbs *b = READ_ONCE(rq->engine)->breadcrumbs;
sys/dev/pci/drm/i915/gt/intel_context.c
625
active = READ_ONCE(ce->stats.active);
sys/dev/pci/drm/i915/gt/intel_context_types.h
94
__intel_context_inflight(READ_ONCE((ce)->inflight))
sys/dev/pci/drm/i915/gt/intel_context_types.h
96
__intel_context_inflight_count(READ_ONCE((ce)->inflight))
sys/dev/pci/drm/i915/gt/intel_engine.h
138
cur = READ_ONCE(execlists->active);
sys/dev/pci/drm/i915/gt/intel_engine.h
143
active = READ_ONCE(*cur);
sys/dev/pci/drm/i915/gt/intel_engine.h
144
cur = READ_ONCE(execlists->active);
sys/dev/pci/drm/i915/gt/intel_engine.h
156
return READ_ONCE(engine->status_page.addr[reg]);
sys/dev/pci/drm/i915/gt/intel_engine.h
346
return READ_ONCE(engine->props.heartbeat_interval_ms);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1626
return READ_ONCE(engine->props.stop_timeout_ms);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
1932
if (!READ_ONCE(gt->awake))
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
2075
if (!READ_ONCE(t->to_time))
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
2164
write = READ_ONCE(*execlists->csb_write);
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
2411
engine->fw_domain, READ_ONCE(engine->fw_active));
sys/dev/pci/drm/i915/gt/intel_engine_cs.c
2414
rq = READ_ONCE(engine->heartbeat.systole);
sys/dev/pci/drm/i915/gt/intel_engine_heartbeat.c
166
long delay = READ_ONCE(engine->props.heartbeat_interval_ms);
sys/dev/pci/drm/i915/gt/intel_engine_heartbeat.c
209
serial = READ_ONCE(engine->serial);
sys/dev/pci/drm/i915/gt/intel_engine_heartbeat.c
28
delay = READ_ONCE(engine->props.heartbeat_interval_ms);
sys/dev/pci/drm/i915/gt/intel_engine_heartbeat.c
50
longer = READ_ONCE(engine->props.preempt_timeout_ms) * 2;
sys/dev/pci/drm/i915/gt/intel_engine_heartbeat.c
81
engine->wakeref_serial = READ_ONCE(engine->serial) + 1;
sys/dev/pci/drm/i915/gt/intel_engine_pm.c
85
READ_ONCE(*ce->timeline->hwsp_seqno),
sys/dev/pci/drm/i915/gt/intel_engine_pm.c
88
READ_ONCE(*ce->timeline->hwsp_seqno));
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1031
struct i915_request *rq = READ_ONCE(ve->request);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1055
GEM_BUG_ON(READ_ONCE(ve->context.inflight));
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1149
return rq->context->lrc.ccid == READ_ONCE(el->yield);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1163
if (READ_ONCE(engine->execlists.pending[0]))
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1203
return READ_ONCE(engine->props.timeslice_duration_ms);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1250
return READ_ONCE(engine->props.preempt_timeout_ms);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1764
if (wait_for_atomic_us((entry = READ_ONCE(*csb)) != -1, 10)) {
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1785
u64 entry = READ_ONCE(*csb);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1842
tail = READ_ONCE(*execlists->csb_write);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
2461
if (unlikely(READ_ONCE(engine->execlists.error_interrupt))) {
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
262
return READ_ONCE(rq->sched.attr.priority);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
2846
GEM_BUG_ON(READ_ONCE(*execlists->csb_write) != reset_value);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3296
locked = READ_ONCE(rq->engine);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3298
while (unlikely(locked != (engine = READ_ONCE(rq->engine)))) {
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3408
if (READ_ONCE(stats->active))
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3797
rq = READ_ONCE(ve->request);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3822
const int prio = READ_ONCE(sched_engine->queue_priority_hint);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3833
struct intel_engine_cs *sibling = READ_ONCE(ve->siblings[n]);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3838
if (!READ_ONCE(ve->request))
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
4119
READ_ONCE(sched_engine->queue_priority_hint));
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
4147
struct i915_request *rq = READ_ONCE(ve->request);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
496
unsigned int tag = ffs(READ_ONCE(engine->context_tag));
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
583
if (READ_ONCE(ve->request))
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
839
!READ_ONCE(prev->fence.error)) {
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
962
return READ_ONCE(rq->fence.flags);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
193
age = READ_ONCE(node->age);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
60
age = READ_ONCE(node->age);
sys/dev/pci/drm/i915/gt/intel_gt_pm_debugfs.c
495
seq_printf(m, "Interactive? %d\n", READ_ONCE(rps->power.interactive));
sys/dev/pci/drm/i915/gt/intel_gt_pm_debugfs.c
509
seq_printf(m, "Wait boosts: %d\n", READ_ONCE(rps->boosts));
sys/dev/pci/drm/i915/gt/intel_gt_pm_debugfs.c
84
READ_ONCE(fw_domain->wake_count));
sys/dev/pci/drm/i915/gt/intel_gt_requests.c
104
first = READ_ONCE(engine->retire);
sys/dev/pci/drm/i915/gt/intel_gtt.h
454
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
sys/dev/pci/drm/i915/gt/intel_lrc.c
1956
return READ_ONCE(ce->lrc_reg_state[CTX_TIMESTAMP]);
sys/dev/pci/drm/i915/gt/intel_reset.c
851
vma = READ_ONCE(gt->ggtt->fence_regs[i].vma);
sys/dev/pci/drm/i915/gt/intel_ring.h
92
unsigned int head = READ_ONCE(ring->head);
sys/dev/pci/drm/i915/gt/intel_rps.c
1034
struct intel_rps *rps = &READ_ONCE(rq->engine)->gt->rps;
sys/dev/pci/drm/i915/gt/intel_rps.c
1070
if (READ_ONCE(rps->cur_freq) < rps->boost_freq)
sys/dev/pci/drm/i915/gt/mock_engine.c
273
locked = READ_ONCE(rq->engine);
sys/dev/pci/drm/i915/gt/mock_engine.c
275
while (unlikely(locked != (engine = READ_ONCE(rq->engine)))) {
sys/dev/pci/drm/i915/gt/selftest_engine_heartbeat.c
134
GEM_BUG_ON(READ_ONCE(engine->serial) != engine->wakeref_serial);
sys/dev/pci/drm/i915/gt/selftest_engine_heartbeat.c
224
if (READ_ONCE(engine->heartbeat.systole)) {
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1217
while (!READ_ONCE(slot[i]) &&
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1378
} while (READ_ONCE(engine->execlists.pending[0]));
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1639
if (wait_for(READ_ONCE(*map), 10)) {
sys/dev/pci/drm/i915/gt/selftest_execlists.c
1691
GEM_BUG_ON(READ_ONCE(*map));
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3318
if (READ_ONCE(result[0]) != NUM_GPR) {
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3425
while (READ_ONCE(engine->execlists.pending[0]))
sys/dev/pci/drm/i915/gt/selftest_execlists.c
58
if (!READ_ONCE(engine->execlists.pending[0]) && is_active(rq))
sys/dev/pci/drm/i915/gt/selftest_execlists.c
78
if (READ_ONCE(engine->execlists.pending[0]))
sys/dev/pci/drm/i915/gt/selftest_execlists.c
84
if (READ_ONCE(rq->fence.error))
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1196
ret = READ_ONCE(threads[tmp].result);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
255
return READ_ONCE(h->seqno[rq->fence.context % (PAGE_SIZE/sizeof(u32))]);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
914
while (!READ_ONCE(arg->stop)) {
sys/dev/pci/drm/i915/gt/selftest_lrc.c
194
u32 lri = READ_ONCE(hw[dw]);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
240
u32 offset = READ_ONCE(hw[dw]);
sys/dev/pci/drm/i915/gt/selftest_lrc.c
69
if (!READ_ONCE(engine->execlists.pending[0]) && is_active(rq))
sys/dev/pci/drm/i915/gt/selftest_lrc.c
836
timestamp = READ_ONCE(arg->ce[0]->lrc_reg_state[CTX_TIMESTAMP]);
sys/dev/pci/drm/i915/gt/selftest_rc6.c
269
engine->name, READ_ONCE(*res));
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
112
if (err || READ_ONCE(*result)) {
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
123
if (READ_ONCE(*result) != STACK_MAGIC) {
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
134
if (READ_ONCE(*result) != STACK_MAGIC) {
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
166
if (READ_ONCE(*result)) {
sys/dev/pci/drm/i915/gt/selftest_ring_submission.c
197
if (READ_ONCE(*result)) {
sys/dev/pci/drm/i915/gt/selftest_rps.c
544
dc = READ_ONCE(*cntr);
sys/dev/pci/drm/i915/gt/selftest_rps.c
547
dc = READ_ONCE(*cntr) - dc;
sys/dev/pci/drm/i915/gt/selftest_rps.c
807
if (wait_for(READ_ONCE(*cntr), 10)) {
sys/dev/pci/drm/i915/gt/selftest_slpc.c
526
status = READ_ONCE(threads[i].result);
sys/dev/pci/drm/i915/gt/selftest_timeline.c
1393
if (READ_ONCE(*tl->hwsp_seqno) != count) {
sys/dev/pci/drm/i915/gt/selftest_timeline.c
496
if (READ_ONCE(*tl->hwsp_seqno) != tl->seqno) {
sys/dev/pci/drm/i915/gt/selftest_timeline.c
584
if (!err && READ_ONCE(*tl->hwsp_seqno) != n) {
sys/dev/pci/drm/i915/gt/selftest_timeline.c
656
if (!err && READ_ONCE(*tl->hwsp_seqno) != n) {
sys/dev/pci/drm/i915/gt/selftest_timeline.c
756
if (READ_ONCE(*hwsp_seqno[0]) != seqno[0] ||
sys/dev/pci/drm/i915/gt/selftest_timeline.c
757
READ_ONCE(*hwsp_seqno[1]) != seqno[1]) {
sys/dev/pci/drm/i915/gt/sysfs_engines.c
335
if (READ_ONCE(engine->execlists.pending[0]))
sys/dev/pci/drm/i915/gt/uc/intel_guc_ct.c
452
if (unlikely(tail != READ_ONCE(desc->tail))) {
sys/dev/pci/drm/i915/gt/uc/intel_guc_ct.c
458
if (unlikely(READ_ONCE(desc->head) >= size)) {
sys/dev/pci/drm/i915/gt/uc/intel_guc_ct.c
556
FIELD_GET(GUC_HXG_MSG_0_ORIGIN, READ_ONCE(req->status)) == \
sys/dev/pci/drm/i915/gt/uc/intel_guc_ct.c
634
head = READ_ONCE(desc->head);
sys/dev/pci/drm/i915/gt/uc/intel_guc_ct.c
892
u32 tail = READ_ONCE(desc->tail);
sys/dev/pci/drm/i915/gt/uc/intel_guc_ct.c
923
if (unlikely(head != READ_ONCE(desc->head))) {
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
4998
!READ_ONCE(wq->busy),
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
506
ce->parallel.guc.wqi_head = READ_ONCE(*ce->parallel.guc.wq_head);
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
5590
READ_ONCE(*ce->parallel.guc.wq_head));
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
5592
READ_ONCE(*ce->parallel.guc.wq_tail));
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
5594
READ_ONCE(*ce->parallel.guc.wq_status));
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
6008
if (READ_ONCE(engine->props.heartbeat_interval_ms))
sys/dev/pci/drm/i915/i915_active.c
249
it = READ_ONCE(ref->cache);
sys/dev/pci/drm/i915/i915_active.c
251
u64 cached = READ_ONCE(it->timeline);
sys/dev/pci/drm/i915/i915_active.c
33
#define fetch_node(x) rb_entry(READ_ONCE(x), typeof(struct active_node), node)
sys/dev/pci/drm/i915/i915_active.c
57
return (struct intel_engine_cs *)READ_ONCE(node->base.cb.node.prev);
sys/dev/pci/drm/i915/i915_active.c
803
p = READ_ONCE(p->rb_right);
sys/dev/pci/drm/i915/i915_active.c
805
p = READ_ONCE(p->rb_left);
sys/dev/pci/drm/i915/i915_debugfs.c
386
seq_printf(m, "Interactive? %d\n", READ_ONCE(rps->power.interactive));
sys/dev/pci/drm/i915/i915_debugfs.c
400
seq_printf(m, "Wait boosts: %d\n", READ_ONCE(rps->boosts));
sys/dev/pci/drm/i915/i915_debugfs.c
92
return READ_ONCE(obj->userfault_count) ? 'g' : ' ';
sys/dev/pci/drm/i915/i915_gpu_error.c
1016
if (READ_ONCE(error->sgl))
sys/dev/pci/drm/i915/i915_gpu_error.c
1064
sg = READ_ONCE(error->fit);
sys/dev/pci/drm/i915/i915_gpu_error.c
2235
error = READ_ONCE(i915->gpu_error.first_error);
sys/dev/pci/drm/i915/i915_gpu_error.c
2402
error = READ_ONCE(i915->gpu_error.first_error);
sys/dev/pci/drm/i915/i915_list_util.h
20
return READ_ONCE(list->next) == head;
sys/dev/pci/drm/i915/i915_mitigations.c
102
unsigned long local = READ_ONCE(mitigations);
sys/dev/pci/drm/i915/i915_mitigations.c
27
return READ_ONCE(mitigations) & BIT(CLEAR_RESIDUALS);
sys/dev/pci/drm/i915/i915_perf.c
3438
stream = READ_ONCE(engine->oa_group->exclusive_stream);
sys/dev/pci/drm/i915/i915_pmu.c
512
if (!READ_ONCE(pmu->timer_enabled))
sys/dev/pci/drm/i915/i915_pmu.c
700
val = READ_ONCE(pmu->irq_count);
sys/dev/pci/drm/i915/i915_request.c
1165
struct list_head *pos = READ_ONCE(signal->link.prev);
sys/dev/pci/drm/i915/i915_request.c
1187
if (unlikely(READ_ONCE(prev->link.next) != &signal->link)) {
sys/dev/pci/drm/i915/i915_request.c
1223
return rq->sched.semaphores | READ_ONCE(rq->engine->saturated);
sys/dev/pci/drm/i915/i915_request.c
1288
const intel_engine_mask_t mask = READ_ONCE(from->engine)->mask;
sys/dev/pci/drm/i915/i915_request.c
1543
if (to->engine == READ_ONCE(from->engine))
sys/dev/pci/drm/i915/i915_request.c
1573
is_power_of_2(to->execution_mask | READ_ONCE(from->execution_mask)))
sys/dev/pci/drm/i915/i915_request.c
1766
bool pow2 = is_power_of_2(READ_ONCE(prev->engine)->mask |
sys/dev/pci/drm/i915/i915_request.c
2015
timeout_ns = READ_ONCE(rq->engine->props.max_busywait_duration_ns);
sys/dev/pci/drm/i915/i915_request.c
2183
if (READ_ONCE(wait.tsk))
sys/dev/pci/drm/i915/i915_request.c
276
locked = READ_ONCE(rq->engine);
sys/dev/pci/drm/i915/i915_request.c
278
while (unlikely(locked != (engine = READ_ONCE(rq->engine)))) {
sys/dev/pci/drm/i915/i915_request.c
482
return READ_ONCE(engine->execlists.active);
sys/dev/pci/drm/i915/i915_request.c
534
(rq = READ_ONCE(*port)); /* may race with promotion of pending[] */
sys/dev/pci/drm/i915/i915_request.c
631
old = READ_ONCE(rq->fence.error);
sys/dev/pci/drm/i915/i915_request.h
491
const u32 *hwsp = READ_ONCE(rq->hwsp_seqno);
sys/dev/pci/drm/i915/i915_request.h
493
return READ_ONCE(*hwsp);
sys/dev/pci/drm/i915/i915_scheduler.c
152
while (locked != (sched_engine = READ_ONCE(rq->engine)->sched_engine)) {
sys/dev/pci/drm/i915/i915_scheduler.c
220
if (prio > READ_ONCE(p->signaler->attr.priority))
sys/dev/pci/drm/i915/i915_timer_util.h
16
return READ_ONCE(t->expires);
sys/dev/pci/drm/i915/i915_timer_util.h
18
return READ_ONCE(t->to_time);
sys/dev/pci/drm/i915/i915_vma.c
603
ptr = READ_ONCE(vma->iomap);
sys/dev/pci/drm/i915/i915_vma.h
344
return READ_ONCE(vma->iomap);
sys/dev/pci/drm/i915/intel_uncore.c
931
unsigned int actual = READ_ONCE(domain->wake_count);
sys/dev/pci/drm/i915/intel_wakeref.h
241
return READ_ONCE(wf->wakeref);
sys/dev/pci/drm/i915/selftests/i915_active.c
124
if (READ_ONCE(active->retired) && count) {
sys/dev/pci/drm/i915/selftests/i915_active.c
158
if (!READ_ONCE(active->retired)) {
sys/dev/pci/drm/i915/selftests/i915_active.c
191
if (!READ_ONCE(active->retired)) {
sys/dev/pci/drm/i915/selftests/i915_active.c
236
if (!READ_ONCE(active->retired)) {
sys/dev/pci/drm/i915/selftests/i915_request.c
1671
status = READ_ONCE(threads[idx].result);
sys/dev/pci/drm/i915/selftests/i915_request.c
1841
err = READ_ONCE(threads[i].result);
sys/dev/pci/drm/i915/selftests/i915_request.c
2058
if (wait_for(READ_ONCE(*sema) == 0, 50)) {
sys/dev/pci/drm/i915/selftests/i915_request.c
2069
if (wait_for(READ_ONCE(*sema) == 0, 50)) {
sys/dev/pci/drm/i915/selftests/i915_request.c
2203
if (i > 1 && wait_for(READ_ONCE(sema[i - 1]), 500)) {
sys/dev/pci/drm/i915/selftests/i915_request.c
2217
wait_for(READ_ONCE(sema[i - 1]), 500);
sys/dev/pci/drm/i915/selftests/i915_request.c
2500
if (wait_for(READ_ONCE(sema[2 * i]) == -1, 500)) {
sys/dev/pci/drm/i915/selftests/i915_request.c
2528
if (wait_for(READ_ONCE(sema[2 * i - 2]) != -1, 500)) {
sys/dev/pci/drm/i915/selftests/i915_request.c
2615
if (wait_for(READ_ONCE(sema[i]) == -1, 50)) {
sys/dev/pci/drm/i915/selftests/i915_request.c
2622
while (!READ_ONCE(cb.seen))
sys/dev/pci/drm/i915/selftests/i915_request.c
3246
status = READ_ONCE(engines[idx].result);
sys/dev/pci/drm/i915/selftests/i915_request.c
342
while (!READ_ONCE(thread->stop)) {
sys/dev/pci/drm/i915/selftests/i915_request.c
518
err = READ_ONCE(threads[n].result);
sys/dev/pci/drm/i915/selftests/i915_sw_fence.c
555
if (READ_ONCE(ipc.value)) {
sys/dev/pci/drm/i915/selftests/i915_sw_fence.c
563
if (!READ_ONCE(ipc.value)) {
sys/dev/pci/drm/i915/selftests/igt_spinner.c
225
return READ_ONCE(*seqno);
sys/dev/pci/drm/include/drm/drm_connector.h
2421
return READ_ONCE(connector->registration_state) ==
sys/dev/pci/drm/include/drm/drm_mm.h
277
return READ_ONCE(mm->hole_stack.next);
sys/dev/pci/drm/include/drm/spsc_queue.h
103
next = READ_ONCE(node->next);
sys/dev/pci/drm/include/drm/spsc_queue.h
114
} while (unlikely(!(queue->head = READ_ONCE(node->next))));
sys/dev/pci/drm/include/drm/spsc_queue.h
98
node = READ_ONCE(queue->head);
sys/dev/pci/drm/include/linux/atomic.h
124
#define atomic_long_read(p) READ_ONCE(*(p))
sys/dev/pci/drm/include/linux/atomic.h
134
#define atomic64_read(p) READ_ONCE(*(p))
sys/dev/pci/drm/include/linux/atomic.h
45
#define atomic_read(p) READ_ONCE(*(p))
sys/dev/pci/drm/include/linux/atomic.h
501
__typeof(*x) _v = READ_ONCE(*x); \
sys/dev/pci/drm/include/linux/dma-resv.h
453
return READ_ONCE(obj->lock.ctx);
sys/dev/pci/drm/include/linux/iosys-map.h
110
v = READ_ONCE(*(_t *)((_ism)->vaddr + (_o))); \
sys/dev/pci/drm/include/linux/seqlock.h
142
return READ_ONCE(sl->seq);
sys/dev/pci/drm/include/linux/seqlock.h
161
return READ_ONCE(sm->seq.sequence);
sys/dev/pci/drm/radeon/radeon_gem.c
608
cur_placement = READ_ONCE(robj->tbo.resource->mem_type);
sys/dev/pci/drm/radeon/radeon_gem.c
639
cur_placement = READ_ONCE(robj->tbo.resource->mem_type);
sys/dev/pci/drm/scheduler/sched_internal.h
85
if (READ_ONCE(entity->dependency))
sys/dev/pci/drm/scheduler/sched_main.c
349
if (!READ_ONCE(sched->pause_submit))
sys/dev/pci/drm/scheduler/sched_main.c
359
if (!READ_ONCE(sched->pause_submit))
sys/dev/pci/if_rge.c
478
if_bpf = READ_ONCE(ifp->if_bpf);
sys/kern/kern_pledge.c
542
p->p_pledge = READ_ONCE(p->p_p->ps_pledge); /* pledge checks are per-thread */
sys/kern/kern_sig.c
1163
sigmask = READ_ONCE(p->p_sigmask);
sys/kern/kern_sig.c
1330
ps_siglist = READ_ONCE(pr->ps_siglist);
sys/kern/kern_sig.c
949
sigmask = READ_ONCE(p->p_sigmask);
sys/kern/kern_sig.c
965
tmpmask = READ_ONCE(q->p_sigmask);
sys/kern/kern_sig.c
987
tmpmask = READ_ONCE(q->p_sigmask);
sys/kern/kern_smr.c
153
smrgp = READ_ONCE(smr_grace_period) + 1;
sys/kern/kern_smr.c
161
if (READ_ONCE(ci->ci_schedstate.spc_smrgp) == smrgp)
sys/kern/kern_smr.c
242
smrgp = READ_ONCE(smr_grace_period);
sys/kern/sched_bsd.c
262
cpt = READ_ONCE(p->p_cpticks);
sys/kern/sys_generic.c
611
nfiles = READ_ONCE(p->p_fd->fd_nfiles);
sys/kern/uipc_mbuf.c
1853
oldval = newval = READ_ONCE(mq->mq_maxlen);
sys/kern/uipc_socket.c
1519
if ((error = READ_ONCE(so->so_error)))
sys/kern/uipc_socket.c
1526
error = READ_ONCE(sosp->so_error);
sys/kern/uipc_socket.c
2267
u_int state = READ_ONCE(so->so_state);
sys/kern/uipc_socket.c
2268
u_int error = READ_ONCE(so->so_error);
sys/kern/uipc_socket.c
2274
short qlen = READ_ONCE(so->so_qlen);
sys/kern/uipc_socket.c
2330
u_int state = READ_ONCE(so->so_state);
sys/kern/uipc_socket.c
2331
u_int error = READ_ONCE(so->so_error);
sys/kern/uipc_socket.c
2381
u_int state = READ_ONCE(so->so_state);
sys/kern/uipc_socket.c
646
if ((error = READ_ONCE(so->so_error))) {
sys/kern/uipc_socket.c
923
if ((error2 = READ_ONCE(so->so_error))) {
sys/net/if.c
2677
u_short rtlabelid = READ_ONCE(ifp->if_rtlabelid);
sys/net/if_pfsync.c
2002
if (READ_ONCE(s->s_deferred) > 0)
sys/net/if_pfsync.c
2858
if (READ_ONCE(st->sync_defer) != NULL)
sys/net/if_pfsync.c
2921
if (READ_ONCE(st->sync_defer) != NULL)
sys/net/if_rport.c
271
caddr_t if_bpf = READ_ONCE(ifp->if_bpf);
sys/net/if_rport.c
304
if_bpf = READ_ONCE(ifp0->if_bpf);
sys/net/if_tpmr.c
352
iff = READ_ONCE(ifp->if_flags);
sys/net/if_tpmr.c
374
if_bpf = READ_ONCE(ifp->if_bpf);
sys/net/if_veb.c
1057
unsigned int bif_flags = READ_ONCE(tp->p_bif_flags);
sys/net/if_veb.c
1303
bif_flags = READ_ONCE(p->p_bif_flags);
sys/net/if_veb.c
1340
if_bpf = READ_ONCE(ifp->if_bpf);
sys/net/if_veb.c
3713
if_bpf = READ_ONCE(ifp->if_bpf);
sys/net/if_veb.c
984
bif_flags = READ_ONCE(tp->p_bif_flags);
sys/net/if_vxlan.c
587
caddr_t if_bpf = READ_ONCE(ifp->if_bpf);
sys/net/ifq.c
727
ifiq_bpfp = READ_ONCE(ifiq->ifiq_bpfp);
sys/net/ifq.c
729
ifiq_bpf = READ_ONCE(*ifiq_bpfp);
sys/net/ifq.c
730
if_bpf = READ_ONCE(ifp->if_bpf);
sys/net/ifq.h
454
#define ifq_len(_ifq) READ_ONCE((_ifq)->ifq_len)
sys/net/ifq.h
496
#define ifiq_len(_ifiq) READ_ONCE(ml_len(&(_ifiq)->ifiq_ml))
sys/net/pf.c
1634
if (READ_ONCE(st->sync_defer) != NULL)
sys/net/pf.c
2179
if (READ_ONCE(inp->inp_pf_sk) == NULL)
sys/net/pf.c
2319
if (ISSET(READ_ONCE(curcpu()->ci_schedstate.spc_schedflags),
sys/net/pf.c
4412
return (digest.words[0] + READ_ONCE(tcp_iss) + pf_tcp_iss_off);
sys/net/pf.c
8816
if (READ_ONCE(sk->sk_inp) != NULL)
sys/net/pf.c
8842
else if (READ_ONCE(sk->sk_inp) != NULL) {
sys/net/pf.c
8877
if (READ_ONCE(inp->inp_pf_sk) == NULL)
sys/net/pf.c
9005
if (inp == NULL || READ_ONCE(sk->sk_inp) != NULL)
sys/net/pf.c
9023
if (READ_ONCE(sk->sk_inp) == NULL)
sys/net/pfkeyv2.c
543
kcb_reg = READ_ONCE(kp->kcb_reg);
sys/net/pfkeyv2.c
571
int flags = READ_ONCE(kp->kcb_flags);
sys/netinet/raw_ip.c
175
if (ISSET(READ_ONCE(inp->inp_socket->so_rcv.sb_state),
sys/netinet/tcp_input.c
3654
if (READ_ONCE(sc->sc_rxtshift) > 0)
sys/netinet/tcp_timer.c
284
if (!READ_ONCE(so->so_snd.sb_cc))
sys/netinet/tcp_usrreq.c
209
if ((error = READ_ONCE(so->so_error)))
sys/netinet6/raw_ip6.c
187
if (ISSET(READ_ONCE(inp->inp_socket->so_rcv.sb_state),
sys/sys/mbuf.h
565
#define mq_len(_mq) READ_ONCE((_mq)->mq_list.ml_len)
sys/sys/mbuf.h
567
#define mq_full(_mq) (mq_len((_mq)) >= READ_ONCE((_mq)->mq_maxlen))
sys/sys/mbuf.h
568
#define mq_drops(_mq) READ_ONCE((_mq)->mq_drops)
sys/sys/smr.h
72
#define SMR_PTR_GET(pptr) READ_ONCE(*pptr)