Symbol: xa_load
arch/arm64/kvm/vgic/vgic-its.c
113
oldirq = xa_load(&dist->lpi_xa, intid);
arch/arm64/kvm/vgic/vgic-its.c
546
irq = xa_load(&its->translation_cache, cache_key);
arch/arm64/kvm/vgic/vgic.c
73
irq = xa_load(&dist->lpi_xa, intid);
arch/arm64/mm/mteswap.c
52
void *tags = xa_load(&mte_pages, entry.val);
arch/powerpc/kernel/rtas.c
659
return xa_load(&rtas_token_to_function_xarray, token);
arch/x86/kernel/cpu/sgx/encl.c
287
entry = xa_load(&encl->page_array, PFN_DOWN(addr));
arch/x86/kernel/cpu/sgx/encl.c
307
entry = xa_load(&encl->page_array, PFN_DOWN(addr));
arch/x86/kernel/cpu/sgx/encl.c
457
(!xa_load(&encl->page_array, PFN_DOWN(addr))))
arch/x86/kernel/cpu/sgx/encl.c
86
entry = xa_load(&encl->page_array, PFN_DOWN(addr));
arch/x86/kernel/cpu/sgx/main.c
659
return !!xa_load(&sgx_epc_address_space, paddr);
arch/x86/kernel/cpu/sgx/main.c
667
section = xa_load(&sgx_epc_address_space, paddr);
arch/x86/kernel/cpu/sgx/virt.c
46
epc_page = xa_load(&vepc->page_array, index);
block/bio.c
123
bslab = xa_load(&bio_slabs, size);
block/bio.c
142
bslab = xa_load(&bio_slabs, slab_size);
block/blk-mq-sched.c
462
ctx = xa_load(elv_tbl, q->id);
block/blk-mq-sched.c
587
ctx = xa_load(elv_tbl, q->id);
block/blk-mq-sched.c
604
ctx = xa_load(elv_tbl, q->id);
block/blk-mq.c
5059
struct elv_change_ctx *ctx = xa_load(elv_tbl, q->id);
block/blk-mq.c
5090
ctx = xa_load(elv_tbl, q->id);
block/genhd.c
1438
part = xa_load(&disk->part_tbl, partno);
block/partitions/core.c
320
if (xa_load(&disk->part_tbl, partno))
block/partitions/core.c
464
part = xa_load(&disk->part_tbl, partno);
block/partitions/core.c
497
part = xa_load(&disk->part_tbl, partno);
drivers/accel/amdxdna/amdxdna_ctx.c
345
hwctx = xa_load(&client->hwctx_xa, args->handle);
drivers/accel/amdxdna/amdxdna_ctx.c
376
hwctx = xa_load(&client->hwctx_xa, abo->assigned_hwctx);
drivers/accel/amdxdna/amdxdna_ctx.c
494
hwctx = xa_load(&client->hwctx_xa, hwctx_hdl);
drivers/accel/ivpu/ivpu_job.c
1046
cmdq = xa_load(&file_priv->cmdq_xa, args->cmdq_id);
drivers/accel/ivpu/ivpu_job.c
349
cmdq = xa_load(&file_priv->cmdq_xa, cmdq_id);
drivers/accel/ivpu/ivpu_job.c
575
struct ivpu_job *job = xa_load(&vdev->submitted_jobs_xa, job_id);
drivers/accel/ivpu/ivpu_job.c
611
job = xa_load(&vdev->submitted_jobs_xa, job_id);
drivers/accel/ivpu/ivpu_mmu.c
964
file_priv = xa_load(&vdev->context_xa, ssid);
drivers/base/arch_topology.c
893
entry = xa_load(&hetero_cpu, hetero_id);
drivers/base/auxiliary_sysfs.c
102
struct auxiliary_irq_info *info __free(kfree) = xa_load(&auxdev->sysfs.irqs, irq);
drivers/base/memory.c
1178
group = xa_load(&memory_groups, mgid);
drivers/base/memory.c
1197
return xa_load(&memory_groups, mgid);
drivers/base/memory.c
656
mem = xa_load(&memory_blocks, block_id);
drivers/block/rnbd/rnbd-srv.c
94
sess_dev = xa_load(&srv_sess->index_idr, dev_id);
drivers/block/ublk_drv.c
462
return xa_load(&ublk_zoned_report_descs, (unsigned long)req);
drivers/cxl/core/cdat.c
154
dent = xa_load(dsmas_xa, dslbis->handle);
drivers/cxl/core/cdat.c
669
perf_ctx = xa_load(usp_xa, index);
drivers/cxl/core/cdat.c
795
us_ctx = xa_load(res_xa, us_index);
drivers/cxl/core/cdat.c
879
hb_ctx = xa_load(hb_xa, hb_index);
drivers/cxl/core/cdat.c
928
mw_ctx = xa_load(mw_xa, mw_index);
drivers/cxl/core/edac.c
925
rec = xa_load(&array_rec->rec_gen_media, attrbs->dpa);
drivers/cxl/core/edac.c
946
rec = xa_load(&array_rec->rec_dram, attrbs->dpa);
drivers/cxl/core/port.c
974
return xa_load(&cxl_root_buses, (unsigned long)port->uport_dev);
drivers/cxl/core/region.c
222
return xa_load(&port->regions, (unsigned long)cxlr);
drivers/cxl/cxl.h
692
return xa_load(&port->dports, (unsigned long)dport_dev);
drivers/cxl/cxlmem.h
135
return xa_load(&port->endpoints, (unsigned long)&cxlmd->dev);
drivers/dax/super.c
99
dax_dev = xa_load(&dax_hosts, (unsigned long)bdev->bd_disk);
drivers/dma-buf/dma-heap.c
88
heap = xa_load(&dma_heap_minors, iminor(inode));
drivers/dma/idxd/cdev.c
205
ctx = xa_load(&wq->upasid_xa, pasid);
drivers/dma/idxd/cdev.c
741
ctx = xa_load(&wq->upasid_xa, pasid);
drivers/dpll/dpll_core.c
1101
ref = xa_load(&dpll->pin_refs, pin->pin_idx);
drivers/dpll/dpll_core.c
1114
ref = xa_load(&pin->parent_refs, parent->pin_idx);
drivers/dpll/dpll_core.c
166
return xa_load(&dpll_device_xa, id);
drivers/dpll/dpll_core.c
900
ref_sync_pin = xa_load(&pin->ref_sync_pins, ref_sync_pin_id);
drivers/dpll/dpll_netlink.c
1212
parent = xa_load(&dpll_pin_xa, parent_idx);
drivers/dpll/dpll_netlink.c
1215
parent_ref = xa_load(&pin->parent_refs, parent->pin_idx);
drivers/dpll/dpll_netlink.c
1248
ref = xa_load(&pin->dpll_refs, dpll->id);
drivers/dpll/dpll_netlink.c
1275
ref = xa_load(&pin->dpll_refs, dpll->id);
drivers/dpll/dpll_netlink.c
1303
ref = xa_load(&pin->dpll_refs, dpll->id);
drivers/dpll/dpll_netlink.c
1414
dpll = xa_load(&dpll_device_xa, pdpll_idx);
drivers/dpll/dpll_netlink.c
1419
ref = xa_load(&pin->dpll_refs, dpll->id);
drivers/dpll/dpll_netlink.c
1981
info->user_ptr[0] = xa_load(&dpll_pin_xa,
drivers/firewire/core-cdev.c
548
resource = xa_load(&client->resource_xa, index);
drivers/firewire/core-device.c
898
device = xa_load(&fw_device_xa, MINOR(devt));
drivers/firmware/arm_ffa/driver.c
1199
phead = xa_load(&drv_info->partition_info, dev->vm_id);
drivers/firmware/arm_ffa/driver.c
1638
phead = xa_load(&drv_info->partition_info, dev->vm_id);
drivers/firmware/arm_ffa/driver.c
1782
if (xa_load(&drv_info->partition_info, drv_info->vm_id))
drivers/firmware/arm_ffa/driver.c
971
phead = xa_load(&drv_info->partition_info, part_id);
drivers/firmware/arm_scmi/driver.c
235
proto = xa_load(&scmi_protocols, key);
drivers/firmware/arm_scmi/driver.c
311
proto = xa_load(&scmi_protocols, protocol_id);
drivers/firmware/arm_scmi/perf.c
1125
opp = xa_load(&dom->opps_by_lvl, index);
drivers/firmware/arm_scmi/perf.c
1131
opp = xa_load(&dom->opps_by_idx, index);
drivers/firmware/arm_scmi/perf.c
584
opp = xa_load(&dom->opps_by_lvl, min_perf);
drivers/firmware/arm_scmi/perf.c
592
opp = xa_load(&dom->opps_by_lvl, max_perf);
drivers/firmware/arm_scmi/perf.c
663
opp = xa_load(&dom->opps_by_idx, *min_perf);
drivers/firmware/arm_scmi/perf.c
669
opp = xa_load(&dom->opps_by_idx, *max_perf);
drivers/firmware/arm_scmi/perf.c
733
opp = xa_load(&dom->opps_by_lvl, level);
drivers/firmware/arm_scmi/perf.c
796
opp = xa_load(&dom->opps_by_idx, *level);
drivers/firmware/arm_scmi/perf.c
979
opp = xa_load(&dom->opps_by_idx, level);
drivers/firmware/arm_scmi/raw_mode.c
255
return xa_load(&raw->chans_q, chan_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_acpi.c
895
numa_info = xa_load(&numa_info_xa, pxm);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1206
range = xa_load(&xa, key);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
690
queue = xa_load(&uq_mgr->userq_xa, qid);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2481
vm = xa_load(&adev->vm_manager.pasids, pasid);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2965
vm = xa_load(&adev->vm_manager.pasids, pasid);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2991
vm = xa_load(&adev->vm_manager.pasids, pasid);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
3161
vm = xa_load(&adev->vm_manager.pasids, pasid);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
6463
fence_drv = xa_load(xa, doorbell_offset);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
4829
fence_drv = xa_load(xa, doorbell_offset);
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
3619
fence_drv = xa_load(xa, doorbell_offset);
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
1666
fence_drv = xa_load(xa, doorbell_offset);
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
1598
fence_drv = xa_load(xa, doorbell_offset);
drivers/gpu/drm/drm_drv.c
244
minor = xa_load(minor_xa, minor_id);
drivers/gpu/drm/drm_syncobj.c
256
syncobj = xa_load(&file_private->syncobj_xa, handle);
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1290
fence = xa_load(&gpu->user_fences, id);
drivers/gpu/drm/i915/gem/i915_gem_context.c
2300
ctx = xa_load(&file_priv->context_xa, id);
drivers/gpu/drm/i915/gem/i915_gem_context.c
2353
pc = xa_load(&file_priv->proto_context_xa, id);
drivers/gpu/drm/i915/gem/i915_gem_context.c
2612
pc = xa_load(&file_priv->proto_context_xa, args->ctx_id);
drivers/gpu/drm/i915/gem/i915_gem_context.c
358
vm = xa_load(&file_priv->vm_xa, id);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2137
wait = xa_load(&guc->tlb_lookup, guc->serial_slot);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4909
wait = xa_load(&guc->tlb_lookup, seqno);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4994
wq = xa_load(&guc->tlb_lookup, guc->serial_slot);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
518
struct intel_context *ce = xa_load(&guc->context_lookup, id);
drivers/gpu/drm/imagination/pvr_context.h
165
ctx = pvr_context_get(xa_load(&pvr_file->ctx_handles, handle));
drivers/gpu/drm/imagination/pvr_context.h
195
ctx = xa_load(&pvr_dev->ctx_ids, id);
drivers/gpu/drm/imagination/pvr_free_list.h
143
free_list = pvr_free_list_get(xa_load(&pvr_file->free_list_handles, handle));
drivers/gpu/drm/imagination/pvr_free_list.h
172
free_list = xa_load(&pvr_dev->free_list_ids, id);
drivers/gpu/drm/imagination/pvr_hwrt.h
105
hwrt = xa_load(&pvr_file->hwrt_handles, handle);
drivers/gpu/drm/imagination/pvr_vm.c
650
vm_ctx = xa_load(&pvr_file->vm_ctx_handles, handle);
drivers/gpu/drm/lima/lima_ctx.c
73
ctx = xa_load(&mgr->handles, id);
drivers/gpu/drm/panfrost/panfrost_job.c
1019
jm_ctx = panfrost_jm_ctx_get(xa_load(&priv->jm_ctxs, handle));
drivers/gpu/drm/panthor/panthor_heap.c
384
heap = xa_load(&pool->xa, heap_id);
drivers/gpu/drm/panthor/panthor_heap.c
448
heap = xa_load(&pool->xa, heap_id);
drivers/gpu/drm/panthor/panthor_mmu.c
1550
vm = panthor_vm_get(xa_load(&pool->xa, handle));
drivers/gpu/drm/scheduler/sched_entity.c
452
f = xa_load(&job->dependencies, job->last_dependency);
drivers/gpu/drm/tegra/submit.c
152
mapping = xa_load(&context->mappings, id);
drivers/gpu/drm/tegra/submit.c
338
sp = xa_load(syncpoints, args->syncpt.id);
drivers/gpu/drm/tegra/submit.c
524
context = xa_load(&fpriv->contexts, args->context);
drivers/gpu/drm/tegra/uapi.c
173
context = xa_load(&fpriv->contexts, args->context);
drivers/gpu/drm/tegra/uapi.c
203
context = xa_load(&fpriv->contexts, args->context);
drivers/gpu/drm/tegra/uapi.c
284
context = xa_load(&fpriv->contexts, args->context);
drivers/gpu/drm/xe/xe_exec_queue.c
548
q = xa_load(&xef->exec_queue.xa, id);
drivers/gpu/drm/xe/xe_guc_submit.c
2751
q = xa_load(&guc->submission_state.exec_queue_lookup, guc_id);
drivers/gpu/drm/xe/xe_irq.c
997
irq_buf = xa_load(&xe->irq.msix.indexes, msix);
drivers/gpu/drm/xe/xe_pagefault.c
147
vm = xa_load(&xe->usm.asid_to_vm, asid);
drivers/gpu/drm/xe/xe_reg_sr.c
76
struct xe_reg_sr_entry *pentry = xa_load(&sr->xa, idx);
drivers/gpu/drm/xe/xe_tlb_inval_job.c
212
xe_assert(gt_to_xe(job->q->gt), !xa_load(&job->dep.drm.dependencies, 0));
drivers/gpu/drm/xe/xe_tlb_inval_job.c
249
xa_load(&job->dep.drm.dependencies, 0) ==
drivers/gpu/drm/xe/xe_vm.c
1886
vm = xa_load(&xef->vm.xa, id);
drivers/gpu/drm/xe/xe_vm.c
2006
vm = xa_load(&xef->vm.xa, args->vm_id);
drivers/infiniband/core/cm.c
592
cm_id_priv = xa_load(&cm.local_id_table, cm_local_id(local_id));
drivers/infiniband/core/cma.c
232
return xa_load(xa, snum);
drivers/infiniband/core/device.c
1807
if (xa_load(&clients, highest_client_id - 1))
drivers/infiniband/core/device.c
1942
struct ib_client *client = xa_load(&clients, index);
drivers/infiniband/core/device.c
2231
if (xa_load(&devices, ib_dev->index) != ib_dev)
drivers/infiniband/core/device.c
2667
struct ib_client *client = xa_load(&clients, index);
drivers/infiniband/core/device.c
324
device = xa_load(&devices, index);
drivers/infiniband/core/device.c
415
struct ib_client *client = xa_load(&clients, index);
drivers/infiniband/core/device.c
675
if (xa_load(&devices, device->index) == device)
drivers/infiniband/core/device.c
770
client_data = xa_load(&device->client_data, client_id);
drivers/infiniband/core/device.c
772
client = xa_load(&clients, client_id);
drivers/infiniband/core/device.c
952
cdev = xa_load(&device->compat_devs, rnet->id);
drivers/infiniband/core/ib_core_uverbs.c
131
entry = xa_load(&ucontext->mmap_xa, pgoff);
drivers/infiniband/core/mad.c
1748
mad_agent = xa_load(&ib_mad_clients, hi_tid);
drivers/infiniband/core/rdma_core.c
326
uobj = xa_load(&ufile->idr, id);
drivers/infiniband/core/restrack.c
252
res = xa_load(&rt->xa, id);
drivers/infiniband/core/sa_query.c
1219
if (xa_load(&queries, id) != query) {
drivers/infiniband/core/ucma.c
133
ctx = xa_load(&ctx_table, id);
drivers/infiniband/core/ucma.c
1643
mc = xa_load(&multicast_table, cmd.id);
drivers/infiniband/core/ucma.c
368
if (xa_load(&ctx_table, ctx->id) == ctx)
drivers/infiniband/core/verbs.c
1224
real_qp = xa_load(&xrcd->tgt_qps, qp_open_attr->qp_num);
drivers/infiniband/hw/cxgb4/cm.c
362
ep = xa_load(&dev->hwtids, tid);
drivers/infiniband/hw/cxgb4/cm.c
379
ep = xa_load(&dev->stids, stid);
drivers/infiniband/hw/cxgb4/ev.c
127
qhp = xa_load(&dev->qps, CQE_QPID(err_cqe));
drivers/infiniband/hw/cxgb4/ev.c
228
chp = xa_load(&dev->cqs, qid);
drivers/infiniband/hw/cxgb4/iw_cxgb4.h
345
return xa_load(&rhp->cqs, cqid);
drivers/infiniband/hw/cxgb4/iw_cxgb4.h
350
return xa_load(&rhp->qps, qpid);
drivers/infiniband/hw/cxgb4/mem.c
739
mhp = xa_load(&rhp->mrs, rkey >> 8);
drivers/infiniband/hw/efa/efa_main.c
83
cq = xa_load(&dev->cqs_xa, cqn);
drivers/infiniband/hw/erdma/erdma_verbs.h
355
return (struct erdma_qp *)xa_load(&dev->qp_xa, id);
drivers/infiniband/hw/erdma/erdma_verbs.h
360
return (struct erdma_cq *)xa_load(&dev->cq_xa, id);
drivers/infiniband/hw/hfi1/init.c
973
return xa_load(&hfi1_dev_table, unit);
drivers/infiniband/hw/hfi1/netdev_rx.c
469
return xa_load(&rx->dev_tbl, id);
drivers/infiniband/hw/hns/hns_roce_bond.c
281
struct hns_roce_die_info *die_info = xa_load(&roce_bond_xa, bus_num);
drivers/infiniband/hw/hns/hns_roce_bond.c
306
struct hns_roce_die_info *die_info = xa_load(&roce_bond_xa, bus_num);
drivers/infiniband/hw/hns/hns_roce_bond.c
52
struct hns_roce_die_info *die_info = xa_load(&roce_bond_xa, bus_num);
drivers/infiniband/hw/hns/hns_roce_bond.c
854
if (xa_load(&roce_bond_xa, bus_num))
drivers/infiniband/hw/hns/hns_roce_bond.c
956
die_info = xa_load(&roce_bond_xa, bus_num);
drivers/infiniband/hw/hns/hns_roce_bond.c
989
die_info = xa_load(&roce_bond_xa, bus_num);
drivers/infiniband/hw/hns/hns_roce_cq.c
514
hr_cq = xa_load(&hr_dev->cq_table.array,
drivers/infiniband/hw/hns/hns_roce_cq.c
544
hr_cq = xa_load(&hr_dev->cq_table.array,
drivers/infiniband/hw/hns/hns_roce_device.h
1111
return xa_load(&hr_dev->qp_table_xa, qpn);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
5020
hr_dip = xa_load(dip_xa, qpn);
drivers/infiniband/hw/hns/hns_roce_srq.c
18
srq = xa_load(&srq_table->xa, srqn & (hr_dev->caps.num_srqs - 1));
drivers/infiniband/hw/ionic/ionic_admin.c
796
cq = xa_load(&dev->cq_tbl, cqid);
drivers/infiniband/hw/ionic/ionic_admin.c
841
qp = xa_load(&dev->qp_tbl, qpid);
drivers/infiniband/hw/ionic/ionic_datapath.c
520
qp = xa_load(&dev->qp_tbl, qid);
drivers/infiniband/hw/ionic/ionic_hw_stats.c
294
cntr = xa_load(&dev->counter_stats->xa_counters, counter->id);
drivers/infiniband/hw/ionic/ionic_hw_stats.c
328
cntr = xa_load(&cs->xa_counters, counter_id);
drivers/infiniband/hw/mana/mana_ib.h
576
qp = xa_load(&mdev->qp_table_wq, qid);
drivers/infiniband/hw/mlx4/cm.c
276
ent = xa_load(&sriov->pv_id_table, *pv_cm_id);
drivers/infiniband/hw/mlx4/cm.c
367
item = xa_load(&sriov->xa_rej_tmout, (unsigned long)rem_pv_cm_id);
drivers/infiniband/hw/mlx4/cm.c
412
item = xa_load(&sriov->xa_rej_tmout, (unsigned long)rem_pv_cm_id);
drivers/infiniband/hw/mlx5/cq.c
563
sig = xa_load(&dev->sig_mrs,
drivers/infiniband/hw/mlx5/devx.c
1435
event = xa_load(&dev->devx_event_table.event_xa,
drivers/infiniband/hw/mlx5/devx.c
1439
xa_val_level2 = xa_load(&event->object_ids, sub->xa_key_level2);
drivers/infiniband/hw/mlx5/devx.c
1502
event = xa_load(&table->event_xa, MLX5_EVENT_TYPE_COMP);
drivers/infiniband/hw/mlx5/devx.c
1506
obj_event = xa_load(&event->object_ids, obj_id);
drivers/infiniband/hw/mlx5/devx.c
1929
event = xa_load(&devx_event_table->event_xa, key_level1);
drivers/infiniband/hw/mlx5/devx.c
1932
xa_val_level2 = xa_load(&event->object_ids,
drivers/infiniband/hw/mlx5/devx.c
1951
event = xa_load(&devx_event_table->event_xa, key_level1);
drivers/infiniband/hw/mlx5/devx.c
1973
obj_event = xa_load(&event->object_ids, key_level2);
drivers/infiniband/hw/mlx5/devx.c
2203
event = xa_load(&devx_event_table->event_xa,
drivers/infiniband/hw/mlx5/devx.c
2213
obj_event = xa_load(&event->object_ids, obj_id);
drivers/infiniband/hw/mlx5/devx.c
2594
event = xa_load(&table->event_xa, event_type | (obj_type << 16));
drivers/infiniband/hw/mlx5/devx.c
2607
obj_event = xa_load(&event->object_ids, obj_id);
drivers/infiniband/hw/mlx5/dm.c
197
op_entry = xa_load(&dm->ops, op);
drivers/infiniband/hw/mlx5/fs.c
1021
per_qp_opfc = xa_load(qpn_opfc_xa, qp_num);
drivers/infiniband/hw/mlx5/fs.c
1372
per_qp_opfc = xa_load(qpn_opfc_xa, qp->qp_num);
drivers/infiniband/hw/mlx5/mr.c
133
WARN_ON(xa_load(&dev->odp_mkeys, mlx5_base_mkey(mr->mmkey.key)));
drivers/infiniband/hw/mlx5/odp.c
146
struct mlx5_ib_mr *mtt = xa_load(&imr->implicit_children, idx);
drivers/infiniband/hw/mlx5/odp.c
1976
mmkey = xa_load(&dev->odp_mkeys, mlx5_base_mkey(lkey));
drivers/infiniband/hw/mlx5/odp.c
769
mtt = xa_load(&imr->implicit_children, idx);
drivers/infiniband/hw/mlx5/odp.c
974
mmkey = xa_load(&dev->odp_mkeys, mlx5_base_mkey(key));
drivers/infiniband/hw/mlx5/qpc.c
101
dct = xa_load(&dev->qp_table.dct_xa, qpn);
drivers/infiniband/hw/mlx5/srq_cmd.c
741
srq = xa_load(&table->array, srqn);
drivers/infiniband/hw/mlx5/srq_cmd.c
87
srq = xa_load(&table->array, srqn);
drivers/infiniband/hw/qedr/main.c
756
srq = xa_load(&dev->srqs, srq_id);
drivers/infiniband/hw/qedr/qedr_iw_cm.c
524
qp = xa_load(&dev->qps, qpn);
drivers/infiniband/hw/qedr/qedr_iw_cm.c
818
return xa_load(&dev->qps, qpn);
drivers/infiniband/sw/rxe/rxe_pool.c
160
elem = xa_load(xa, index);
drivers/infiniband/sw/siw/siw.h
600
qp = xa_load(&sdev->qp_xa, id);
drivers/infiniband/sw/siw/siw_mem.c
32
mem = xa_load(&sdev->mem_xa, stag_index);
drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c
151
return xa_load(&port->vports, vport_num);
drivers/infiniband/ulp/srpt/ib_srpt.c
124
e = xa_load(&srpt_memory_caches, object_size);
drivers/iommu/amd/iommu.c
2274
pdom_iommu_info = xa_load(&pdom->iommu_array, iommu->index);
drivers/iommu/amd/iommu.c
2310
pdom_iommu_info = xa_load(&pdom->iommu_array, iommu->index);
drivers/iommu/amd/nested.c
66
elm = xa_load(xa, index);
drivers/iommu/intel/iommu.c
1038
curr = xa_load(&domain->iommu_array, iommu->seq_id);
drivers/iommu/intel/iommu.c
1079
info = xa_load(&domain->iommu_array, iommu->seq_id);
drivers/iommu/intel/iommu.h
817
xa_load(&domain->iommu_array, iommu->seq_id);
drivers/iommu/iommu.c
3797
entry = xa_load(&group->pasid_array, pasid);
drivers/iommu/iommufd/device.c
362
attach = xa_load(&igroup->pasid_attach, pasid);
drivers/iommu/iommufd/device.c
409
return !xa_load(&igroup->pasid_attach, pasid);
drivers/iommu/iommufd/device.c
445
attach = xa_load(&idev->igroup->pasid_attach, pasid);
drivers/iommu/iommufd/device.c
446
return xa_load(&attach->device_array, idev->obj.id);
drivers/iommu/iommufd/device.c
470
attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID);
drivers/iommu/iommufd/device.c
695
attach = xa_load(&igroup->pasid_attach, pasid);
drivers/iommu/iommufd/device.c
742
attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID);
drivers/iommu/iommufd/device.c
759
attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID);
drivers/iommu/iommufd/device.c
795
attach = xa_load(&igroup->pasid_attach, pasid);
drivers/iommu/iommufd/device.c
82
igroup = xa_load(&ictx->groups, id);
drivers/iommu/iommufd/driver.c
100
vdev = xa_load(&viommu->vdevs, vdev_id);
drivers/iommu/iommufd/io_pagetable.c
987
xa_load(&iopt->domains, 0);
drivers/iommu/iommufd/main.c
175
obj = xa_load(&ictx->objects, id);
drivers/iommu/iommufd/pages.c
1945
area->storage_domain = xa_load(&area->iopt->domains, 0);
drivers/irqchip/irq-gic-v5-its.c
740
struct gicv5_its_dev *dev = xa_load(&its->its_devices, device_id);
drivers/md/dm-zoned-metadata.c
1452
struct dm_zone *zone = xa_load(&zmd->zones, idx);
drivers/md/dm-zoned-metadata.c
300
return xa_load(&zmd->zones, zone_id);
drivers/md/md.c
4321
head = xa_load(&md_submodule, id);
drivers/md/md.c
691
head = xa_load(&md_submodule, mddev->bitmap_id);
drivers/md/md.c
7892
mddev->cluster_ops = xa_load(&md_submodule, ID_CLUSTER);
drivers/misc/uacce/uacce.c
157
uacce = xa_load(&uacce_xa, iminor(inode));
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6286
ctx = xa_load(&bp->dev->ethtool->rss_ctx,
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1236
ctx = xa_load(&bp->dev->ethtool->rss_ctx, index);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1860
ctx = xa_load(&bp->dev->ethtool->rss_ctx, rxfh->rss_context);
drivers/net/ethernet/fungible/funeth/funeth_main.c
368
err = funeth_txq_create(dev, i, depth, xa_load(&fp->irqs, i),
drivers/net/ethernet/fungible/funeth/funeth_main.c
397
xa_load(&fp->irqs, i + fp->rx_irq_ofst),
drivers/net/ethernet/fungible/funeth/funeth_main.c
537
xa_load(&fp->irqs,
drivers/net/ethernet/fungible/funeth/funeth_main.c
544
err = fun_txq_create_dev(qset->txqs[i], xa_load(&fp->irqs, i));
drivers/net/ethernet/intel/ice/ice_adapter.c
102
adapter = xa_load(&ice_adapters, index);
drivers/net/ethernet/intel/ice/ice_adapter.c
138
adapter = xa_load(&ice_adapters, index);
drivers/net/ethernet/intel/ice/ice_eswitch.c
185
repr = xa_load(&pf->eswitch.reprs, *repr_id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
588
struct ice_repr *repr = xa_load(&pf->eswitch.reprs, vf->repr_id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
606
struct ice_repr *repr = xa_load(&pf->eswitch.reprs, sf->repr_id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
630
repr = xa_load(&eswitch->reprs, le16_to_cpu(desc->src_vsi));
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
312
port = xa_load(&bridge->ports, vsi_idx);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
318
vlan = xa_load(&port->vlans, vid);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
742
port = xa_load(&bridge->ports, vsi_idx);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
753
vlan = xa_load(&port->vlans, vid);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
777
port = xa_load(&bridge->ports, vsi_idx);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
781
vlan = xa_load(&port->vlans, vid);
drivers/net/ethernet/intel/ice/ice_irq.c
223
entry = xa_load(&pf->irq_tracker.entries, map.index);
drivers/net/ethernet/intel/ice/ice_repr.c
529
return xa_load(&pf->eswitch.reprs, id);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1317
stats = xa_load(&cmd->stats, ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1797
stats = xa_load(&cmd->stats, ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1945
return !!xa_load(&dev->cmd.vars.privileged_uids, uid);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
2053
stats = xa_load(&dev->cmd.stats, opcode);
drivers/net/ethernet/mellanox/mlx5/core/en/mapping.c
118
mi = xa_load(&ctx->xarray, index);
drivers/net/ethernet/mellanox/mlx5/core/en/mapping.c
141
mi = xa_load(&ctx->xarray, index);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
494
sa_entry = xa_load(&ipsec->sadb, be32_to_cpu(object->obj_id));
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c
326
sa_entry = xa_load(&ipsec->sadb, sa_handle);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
1697
sc_xarray_element = xa_load(&macsec->sc_xarray, fs_id);
drivers/net/ethernet/mellanox/mlx5/core/en_rep.c
1693
struct mlx5e_rep_sq_peer *sq_peer = xa_load(&rep_sq->sq_peer, i);
drivers/net/ethernet/mellanox/mlx5/core/en_rep.c
1717
sq_peer = xa_load(&rep_sq->sq_peer, i);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
1018
irq = xa_load(&table->comp_irqs, vecidx);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
1064
eq = xa_load(&table->comp_eqs, vecidx);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
1095
eq = xa_load(&table->comp_eqs, vector);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
1112
eq = xa_load(&table->comp_eqs, vector);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
815
irq = xa_load(&table->comp_irqs, vecidx);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
863
irq = xa_load(&table->comp_irqs, vecidx);
drivers/net/ethernet/mellanox/mlx5/core/esw/acl/egress_ofld.c
21
xa_load(&vport->egress.offloads.bounce_rules, rule_index);
drivers/net/ethernet/mellanox/mlx5/core/esw/bridge.c
1021
return xa_load(&port->vlans, vid);
drivers/net/ethernet/mellanox/mlx5/core/esw/bridge.c
953
return xa_load(&br_offloads->ports, mlx5_esw_bridge_port_key_from_data(vport_num,
drivers/net/ethernet/mellanox/mlx5/core/esw/bridge_mcast.c
32
return xa_load(&entry->ports, mlx5_esw_bridge_port_key(port));
drivers/net/ethernet/mellanox/mlx5/core/esw/ipsec_fs.c
116
val = xa_load(&ipsec->ipsec_obj_id_map, id);
drivers/net/ethernet/mellanox/mlx5/core/eswitch.c
114
vport = xa_load(&esw->vports, vport_num);
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
2509
struct mlx5_eswitch_rep *rep = xa_load(&esw->offloads.vport_reps,
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
2999
if (xa_load(&vport->egress.offloads.bounce_rules, slave_index))
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
3165
esw_paired = !!xa_load(&esw->paired, peer_esw_i);
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
4507
u16 *res = xa_load(&esw->offloads.vhca_map, vhca_id);
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
80
return xa_load(&esw->offloads.vport_reps, vport_num);
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
2810
vport_ns = xa_load(&steering->esw_egress_root_ns, vport_idx);
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
2816
vport_ns = xa_load(&steering->esw_ingress_root_ns, vport_idx);
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
3670
if (xa_load(esw_acl_root_ns, vport_idx))
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
670
fs_id_found = xa_load(fs_id_xa, fs_id);
drivers/net/ethernet/mellanox/mlx5/core/lib/st.c
170
idx_data = xa_load(&st->idx_xa, st_index);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
106
root = xa_load(&dev->priv.page_root_xa, function);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
177
root = xa_load(&dev->priv.page_root_xa, function);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
259
root = xa_load(&dev->priv.page_root_xa, fwp->function);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
443
root = xa_load(&dev->priv.page_root_xa, function);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
504
root = xa_load(&dev->priv.page_root_xa, get_function(func_id, ec_function));
drivers/net/ethernet/mellanox/mlx5/core/pci_irq.c
395
irq = xa_load(&pool->irqs, vecidx);
drivers/net/ethernet/mellanox/mlx5/core/sf/dev/dev.c
179
sf_dev = xa_load(&table->devices, sf_index);
drivers/net/ethernet/mellanox/mlx5/core/sf/dev/dev.c
234
if (!xa_load(&work_ctx->table->devices, work_ctx->sf_index))
drivers/net/ethernet/mellanox/mlx5/core/sf/devlink.c
39
return xa_load(&table->function_ids, fn_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
528
peer_ctx = xa_load(&fc->ctx->peer_ctx_xa, id);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
1235
pr_pool = xa_load(pr_pools, index);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
387
return xa_load(&fs_ctx->hws_pool.table_dests, dst->dest_attr.ft->id);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
396
return xa_load(&fs_ctx->hws_pool.table_dests, table_num);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
436
dest = xa_load(dests_xa, idx);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
479
fs_hws_data = xa_load(cache_xa, index);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
530
meter_hws_data = xa_load(meters_xa, exe_aso->base_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
565
sampler_hws_data = xa_load(sampler_xa, sampler_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/vport.c
75
entry = xa_load(&ctx->vports.vport_gvmi_xa, vport);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_action.c
2116
vport_dmn = peer_vport ? xa_load(&dmn->peer_dmn_xa, vhca_id) : dmn;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
953
vport_caps = xa_load(&caps->vports.vports_caps_xa, i);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_definer.c
143
definer_obj = xa_load(&dmn->definers_xa, definer_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
329
vport_caps = xa_load(&caps->vports.vports_caps_xa, vport);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
546
peer = xa_load(&dmn->peer_dmn_xa, peer_vhca_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
552
peer = xa_load(&dmn->peer_dmn_xa, peer_vhca_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
81
recalc_cs_ft = xa_load(&dmn->csum_fts_xa, vport_num);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1663
peer = xa_load(&dmn->peer_dmn_xa, id);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1856
peer = xa_load(&dmn->peer_dmn_xa, id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_port_range.c
143
prr = xa_load(&pr_core->prr_xa, prr_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
3205
nhct = xa_load(&nh_grp->nhgi->nexthop_counters, nh->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
3233
nhct = xa_load(&nh_grp->nhgi->nexthop_counters, nh->id);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1689
gic = xa_load(&gc->irq_contexts, i);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1778
gic = xa_load(&gc->irq_contexts, i);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1870
gic = xa_load(&gc->irq_contexts, i);
drivers/net/ethernet/microsoft/mana/gdma_main.c
734
gic = xa_load(&gc->irq_contexts, msi_index);
drivers/net/ethernet/microsoft/mana/gdma_main.c
761
gic = xa_load(&gc->irq_contexts, msix_index);
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
627
x = xa_load(&nn->xa_ipsec, saidx);
drivers/net/ethernet/sfc/rx_common.c
565
ctx = xa_load(&efx->net_dev->ethtool->rss_ctx, id);
drivers/net/pse-pd/pse_core.c
905
pw_d = xa_load(&pse_pw_d_map, pcdev->pi[i].pw_d->id);
drivers/nvme/host/core.c
3318
struct nvme_effects_log *old, *cel = xa_load(&ctrl->cels, csi);
drivers/nvme/target/core.c
442
req->ns = xa_load(&subsys->namespaces, nsid);
drivers/nvme/target/pr.c
993
pc_ref = xa_load(&req->ns->pr_per_ctrl_refs,
drivers/pci/p2pdma.c
1103
type = xa_to_value(xa_load(&p2pdma->map_types,
drivers/scsi/elx/efct/efct_unsol.c
32
node = xa_load(&efct->lookup, id);
drivers/scsi/elx/libefc/efc_node.c
28
return xa_load(&nport->lookup, port_id);
drivers/scsi/elx/libefc/efc_nport.c
169
nport = xa_load(&domain->lookup, d_id);
drivers/scsi/scsi_debug.c
1329
lsip = xa_load(per_store_ap, 0); /* should never be NULL */
drivers/scsi/scsi_debug.c
4012
return xa_load(per_store_ap, devip->sdbg_host->si_idx);
drivers/scsi/scsi_debug.c
8253
struct sdeb_store_info *sip = xa_load(per_store_ap, 0);
drivers/scsi/scsi_debug.c
8752
sip = xa_load(per_store_ap, idx);
drivers/target/target_core_user.c
1484
cmd = xa_load(&udev->commands, entry->hdr.cmd_id);
drivers/target/target_core_user.c
1844
page = xa_load(&udev->data_pages, dpi);
drivers/target/target_core_user.c
736
page = xa_load(&udev->data_pages, dpi);
drivers/tee/qcomtee/core.c
274
object = xa_load(&qcomtee->xa_local_objects, object_id);
drivers/tee/tee_heap.c
258
if (!xa_load(&tee_dma_heap, id))
drivers/tee/tee_heap.c
295
h = xa_load(&tee_dma_heap, id);
drivers/tee/tstee/core.c
176
sess = xa_load(&ctxdata->sess_list, arg->session);
drivers/tty/serial/liteuart.c
380
uart = (struct liteuart_port *)xa_load(&liteuart_array, co->index);
drivers/tty/serial/liteuart.c
397
uart = (struct liteuart_port *)xa_load(&liteuart_array, co->index);
drivers/vfio/pci/vfio_pci_intrs.c
56
return xa_load(&vdev->ctx, index);
drivers/vfio/vfio_main.c
80
dev_set = xa_load(&vfio_device_set_xa, idx);
drivers/xen/grant-dma-ops.c
49
data = xa_load(&xen_grant_dma_devices, (unsigned long)dev);
fs/btrfs/delayed-inode.c
166
ptr = xa_load(&root->delayed_nodes, ino);
fs/btrfs/delayed-inode.c
84
node = xa_load(&root->delayed_nodes, ino);
fs/btrfs/delayed-ref.c
1203
return xa_load(&delayed_refs->head_refs, index);
fs/btrfs/delayed-ref.c
869
existing = xa_load(&delayed_refs->head_refs, index);
fs/btrfs/extent_io.c
2191
eb = xa_load(&fs_info->buffer_tree, index);
fs/btrfs/qgroup.c
2030
existing = xa_load(&delayed_refs->dirty_extents, index);
fs/dlm/lock.c
1548
lkb = xa_load(&ls->ls_lkbxa, lkid);
fs/dlm/recover.c
354
r = xa_load(&ls->ls_recover_xa, (int)id);
fs/erofs/zdata.c
831
pcl = xa_load(&EROFS_SB(sb)->managed_pslots, map->m_pa);
fs/ext4/mballoc.c
1261
if (!xa_empty(xa) && xa_load(xa, grp->bb_group))
fs/f2fs/node.c
1502
afolio = xa_load(&NODE_MAPPING(sbi)->i_pages, nid);
fs/nfsd/nfs4state.c
1961
struct nfsd4_slot *slot = xa_load(&ses->se_slots, i);
fs/nfsd/nfs4state.c
2014
struct nfsd4_slot *slot = xa_load(&ses->se_slots, i);
fs/nfsd/nfs4state.c
4427
slot = xa_load(&session->se_slots, seq->slotid);
fs/nfsd/nfs4state.c
4510
prev_slot = xa_load(&session->se_slots, s);
fs/smb/server/connection.c
224
if (conn->binding || xa_load(&conn->sessions, sess_id))
fs/smb/server/connection.c
247
if (conn->binding || xa_load(&conn->sessions, sess_id)) {
fs/smb/server/mgmt/tree_connect.c
142
tcon = xa_load(&sess->tree_conns, id);
fs/smb/server/mgmt/user_session.c
374
entry = xa_load(&sess->rpc_handle_list, id);
fs/smb/server/mgmt/user_session.c
508
sess = xa_load(&conn->sessions, id);
fs/smb/server/mgmt/user_session.c
524
sess = xa_load(&conn->sessions, id);
fs/smb/server/smb2pdu.c
87
chann = xa_load(&sess->ksmbd_chann_list, (long)conn);
fs/xfs/libxfs/xfs_group.c
40
xg = xa_load(&mp->m_groups[type].xa, index);
fs/xfs/libxfs/xfs_group.c
81
xg = xa_load(&mp->m_groups[type].xa, index);
include/linux/kvm_host.h
1007
return xa_load(&kvm->vcpu_array, i);
include/linux/kvm_host.h
2519
return xa_to_value(xa_load(&kvm->mem_attr_array, gfn));
include/linux/pagemap.h
1435
folio = xa_load(&ractl->mapping->i_pages, ractl->_index);
include/linux/phy_link_topology.h
55
pdn = xa_load(&topo->phys, phyindex);
include/linux/xarray.h
355
void *xa_load(struct xarray *, unsigned long index);
include/rdma/ib_verbs.h
3051
return xa_load(&device->client_data, client->client_id);
io_uring/io_uring.c
1816
req->creds = xa_load(&ctx->personalities, personality);
io_uring/kbuf.c
753
bl = xa_load(&ctx->io_bl_xa, bgid);
io_uring/kbuf.c
83
return xa_load(&ctx->io_bl_xa, bgid);
io_uring/net.c
1259
zc->ifq = xa_load(&req->ctx->zcrx_ctxs, ifq_idx);
io_uring/tctx.c
141
if (!xa_load(&tctx->xa, (unsigned long)ctx)) {
io_uring/zcrx.c
1179
zcrx = xa_load(&ctx->zcrx_ctxs, ctrl.zcrx_id);
io_uring/zcrx.c
639
struct io_zcrx_ifq *ifq = xa_load(&ctx->zcrx_ctxs, id);
kernel/irq/msi.c
471
desc = xa_load(xa, pcimsi ? 0 : index);
kernel/liveupdate/kexec_handover.c
124
void *res = xa_load(xa, index);
kernel/liveupdate/kexec_handover.c
153
physxa = xa_load(&track->orders, order);
kernel/liveupdate/kexec_handover.c
157
bits = xa_load(&physxa->phys_bits, pfn_high / PRESERVE_BITS);
kernel/liveupdate/kexec_handover.c
186
physxa = xa_load(&track->orders, order);
kernel/trace/trace_syscalls.c
108
return xa_load(&syscalls_metadata_sparse, (unsigned long)nr);
lib/ref_tracker.c
388
if (!xa_load(&debugfs_dentries, index) ||
lib/ref_tracker.c
432
dentry = xa_load(&debugfs_dentries, (unsigned long)dir);
lib/ref_tracker.c
465
symlink = xa_load(&debugfs_symlinks, (unsigned long)dir);
lib/ref_tracker.c
466
dentry = xa_load(&debugfs_dentries, (unsigned long)dir);
lib/test_hmm.c
375
entry = xa_load(&dmirror->pt, pfn);
lib/test_hmm.c
441
entry = xa_load(&dmirror->pt, pfn);
lib/test_hmm.c
805
entry = xa_load(&dmirror->pt, pfn);
lib/test_xarray.c
1136
XA_BUG_ON(xa, xa_load(xa, start) != xa_mk_index(start));
lib/test_xarray.c
1137
XA_BUG_ON(xa, xa_load(xa, start + (1UL << order) - 1) !=
lib/test_xarray.c
1411
XA_BUG_ON(xa, xa_load(xa, index) !=
lib/test_xarray.c
153
void *entry = xa_load(xa, j);
lib/test_xarray.c
164
void *entry = xa_load(xa, j);
lib/test_xarray.c
1808
XA_BUG_ON(xa, xa_load(xa, first) != xa_mk_index(first));
lib/test_xarray.c
1809
XA_BUG_ON(xa, xa_load(xa, last) != xa_mk_index(first));
lib/test_xarray.c
1810
XA_BUG_ON(xa, xa_load(xa, first - 1) != NULL);
lib/test_xarray.c
1811
XA_BUG_ON(xa, xa_load(xa, last + 1) != NULL);
lib/test_xarray.c
1855
XA_BUG_ON(xa, xa_load(xa, index + i) != xa_mk_index(val));
lib/test_xarray.c
1902
XA_BUG_ON(xa, xa_load(xa, index + i) != xa_mk_index(val));
lib/test_xarray.c
344
XA_BUG_ON(xa, xa_load(xa, 1) != NULL);
lib/test_xarray.c
349
XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0));
lib/test_xarray.c
356
XA_BUG_ON(xa, xa_load(xa, max) != xa_mk_value(0));
lib/test_xarray.c
357
XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL);
lib/test_xarray.c
366
XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL);
lib/test_xarray.c
379
XA_BUG_ON(xa, xa_load(xa, i - 1) != NULL);
lib/test_xarray.c
380
XA_BUG_ON(xa, xa_load(xa, i + 1) != NULL);
lib/test_xarray.c
386
XA_BUG_ON(xa, xa_load(xa, (1UL << i) - 1) != NULL);
lib/test_xarray.c
387
XA_BUG_ON(xa, xa_load(xa, (1UL << i) + 1) != NULL);
lib/test_xarray.c
391
XA_BUG_ON(xa, xa_load(xa, (1UL << i) - 2) != NULL);
lib/test_xarray.c
392
XA_BUG_ON(xa, xa_load(xa, 1UL << i) != NULL);
lib/test_xarray.c
397
XA_BUG_ON(xa, xa_load(xa, 0UL) != NULL);
lib/test_xarray.c
398
XA_BUG_ON(xa, xa_load(xa, ~1UL) != NULL);
lib/test_xarray.c
439
XA_BUG_ON(xa, xa_load(xa, i) != FIVE);
lib/test_xarray.c
444
XA_BUG_ON(xa, xa_load(xa, 1 << order) != NULL);
lib/test_xarray.c
452
XA_BUG_ON(xa, xa_load(xa, i) != FIVE);
lib/test_xarray.c
464
XA_BUG_ON(xa, xa_load(xa, i) != NULL);
lib/test_xarray.c
470
XA_BUG_ON(xa, xa_load(xa, i) != FIVE);
lib/test_xarray.c
489
XA_BUG_ON(xa, xa_load(xa, 12345678));
lib/test_xarray.c
592
XA_BUG_ON(xa, xa_load(xa, min) != xa_mk_index(index));
lib/test_xarray.c
593
XA_BUG_ON(xa, xa_load(xa, max - 1) != xa_mk_index(index));
lib/test_xarray.c
594
XA_BUG_ON(xa, xa_load(xa, max) != NULL);
lib/test_xarray.c
595
XA_BUG_ON(xa, xa_load(xa, min - 1) != NULL);
lib/test_xarray.c
600
XA_BUG_ON(xa, xa_load(xa, min) != xa_mk_index(min));
lib/test_xarray.c
601
XA_BUG_ON(xa, xa_load(xa, max - 1) != xa_mk_index(min));
lib/test_xarray.c
602
XA_BUG_ON(xa, xa_load(xa, max) != NULL);
lib/test_xarray.c
603
XA_BUG_ON(xa, xa_load(xa, min - 1) != NULL);
lib/test_xarray.c
63
XA_BUG_ON(xa, xa_load(xa, index) != NULL);
lib/test_xarray.c
658
XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0));
lib/test_xarray.c
659
XA_BUG_ON(xa, xa_load(xa, 1) != xa_mk_value(0));
lib/test_xarray.c
660
XA_BUG_ON(xa, xa_load(xa, 2) != NULL);
lib/test_xarray.c
668
XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0));
lib/test_xarray.c
669
XA_BUG_ON(xa, xa_load(xa, 1) != xa_mk_value(0));
lib/test_xarray.c
670
XA_BUG_ON(xa, xa_load(xa, 2) != NULL);
lib/test_xarray.c
678
XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(1));
lib/test_xarray.c
679
XA_BUG_ON(xa, xa_load(xa, 1) != xa_mk_value(1));
lib/test_xarray.c
680
XA_BUG_ON(xa, xa_load(xa, 2) != xa_mk_value(1));
lib/test_xarray.c
681
XA_BUG_ON(xa, xa_load(xa, 3) != xa_mk_value(1));
lib/test_xarray.c
682
XA_BUG_ON(xa, xa_load(xa, 4) != NULL);
lib/test_xarray.c
704
void *entry = xa_load(xa, (1UL << k) - 1);
lib/test_xarray.c
760
XA_BUG_ON(xa, xa_load(xa, index) != p);
lib/xarray.c
1625
EXPORT_SYMBOL(xa_load);
mm/khugepaged.c
2034
VM_BUG_ON_FOLIO(folio != xa_load(xas.xa, index), folio);
mm/list_lru.c
510
return idx < 0 || xa_load(&lru->xa, idx);
mm/list_lru.c
56
struct list_lru_memcg *mlru = xa_load(&lru->xa, idx);
mm/memcontrol.c
3663
return xa_load(&mem_cgroup_private_ids, id);
mm/memremap.c
407
pgmap = xa_load(&pgmap_array, PHYS_PFN(phys));
mm/readahead.c
259
struct folio *folio = xa_load(&mapping->i_pages, index + i);
mm/readahead.c
786
struct folio *folio = xa_load(&mapping->i_pages, index);
mm/readahead.c
815
struct folio *folio = xa_load(&mapping->i_pages, index);
mm/vmalloc.c
2923
vb = xa_load(xa, addr_to_vb_idx(addr));
mm/vmalloc.c
4516
vb = xa_load(xa, addr_to_vb_idx((unsigned long)addr));
mm/zswap.c
1039
if (entry != xa_load(tree, offset)) {
mm/zswap.c
1617
entry = xa_load(tree, offset);
net/core/devmem.c
346
binding = xa_load(&net_devmem_dmabuf_bindings, id);
net/core/page_pool_user.c
47
pool = xa_load(&page_pools, id);
net/devlink/param.c
165
return xa_load(params, param_id);
net/devlink/port.c
30
return xa_load(&devlink->ports, port_index);
net/devlink/region.c
284
p = xa_load(&devlink->snapshot_ids, id);
net/devlink/region.c
326
p = xa_load(&devlink->snapshot_ids, id);
net/devlink/region.c
368
if (xa_load(&devlink->snapshot_ids, id)) {
net/ethtool/common.c
718
ctx = xa_load(&dev->ethtool->rss_ctx,
net/ethtool/ioctl.c
1196
!xa_load(&dev->ethtool->rss_ctx, info.rss_context))
net/ethtool/ioctl.c
1474
ctx = xa_load(&dev->ethtool->rss_ctx, rxfh.rss_context);
net/ethtool/ioctl.c
1679
ctx = xa_load(&dev->ethtool->rss_ctx, rxfh.rss_context);
net/ethtool/phy.c
86
pdn = xa_load(&topo->phys, phydev->phyindex);
net/ethtool/rss.c
1176
ctx = xa_load(&dev->ethtool->rss_ctx, rss_context);
net/ethtool/rss.c
211
ctx = xa_load(&dev->ethtool->rss_ctx, request->rss_context);
net/ethtool/rss.c
869
ctx = xa_load(&dev->ethtool->rss_ctx, request->rss_context);
net/hsr/hsr_framereg.c
338
block = xa_load(&node->seq_blocks, block_idx);
net/hsr/prp_dup_discard_test.c
48
block = xa_load(&data->node.seq_blocks, block_idx);
net/netlink/genetlink.c
727
return xa_load(family->sock_privs, (unsigned long) sk);
net/psp/psp_nl.c
50
psd = xa_load(&psp_devs, nla_get_u32(dev_id));
net/qrtr/af_qrtr.c
230
flow = xa_load(&node->qrtr_tx_flow, key);
net/qrtr/af_qrtr.c
269
flow = xa_load(&node->qrtr_tx_flow, key);
net/qrtr/af_qrtr.c
327
flow = xa_load(&node->qrtr_tx_flow, key);
net/qrtr/af_qrtr.c
670
ipc = xa_load(&qrtr_ports, port);
net/qrtr/ns.c
273
srv = xa_load(&node->servers, port);
net/qrtr/ns.c
76
node = xa_load(&nodes, node_id);
net/sched/cls_api.c
129
return xa_load(&tcf_exts_miss_cookies_xa, mc.miss_cookie_base);
net/shaper/shaper.c
296
return xa_load(&hierarchy->shapers, index);
net/shaper/shaper.c
342
cur = xa_load(&hierarchy->shapers, index);
net/shaper/shaper.c
409
cur = xa_load(&hierarchy->shapers, index);
security/apparmor/secid.c
47
return xa_load(&aa_secids, secid);
sound/core/control.c
434
matched = xa_load(&card->ctl_hash, h);
sound/core/control.c
822
return xa_load(&card->ctl_numids, numid);
sound/core/control.c
851
kctl = xa_load(&card->ctl_hash, get_ctl_id_hash(id));
tools/testing/radix-tree/multiorder.c
247
struct item *item = xa_load(ptr, index);
virt/kvm/guest_memfd.c
766
if (xa_load(&f->bindings, index) != slot) {
virt/kvm/guest_memfd.c
767
WARN_ON_ONCE(xa_load(&f->bindings, index));
virt/kvm/kvm_main.c
4007
vcpu = xa_load(&kvm->vcpu_array, idx);
virt/kvm/kvm_main.c
498
WARN_ON_ONCE(xa_load(&kvm->vcpu_array, i) || kvm_get_vcpu(kvm, i));
virt/lib/irqbypass.c
111
consumer = xa_load(&consumers, index);
virt/lib/irqbypass.c
177
producer = xa_load(&producers, index);