Symbol: xa_store
arch/arm64/kernel/hibernate.c
230
ret = xa_store(&mte_pages, pfn, tag_storage, GFP_KERNEL);
arch/arm64/kvm/emulate-nested.c
2267
prev = xa_store(&sr_forward_xa, enc,
arch/arm64/kvm/emulate-nested.c
2316
prev = xa_store(&sr_forward_xa, enc,
arch/arm64/kvm/emulate-nested.c
2384
ret = xa_store(&sr_forward_xa, encoding,
arch/arm64/kvm/vgic/vgic-its.c
573
old = xa_store(&its->translation_cache, cache_key, irq, GFP_KERNEL_ACCOUNT);
arch/arm64/mm/mteswap.c
37
ret = xa_store(&mte_pages, page_swap_entry(page).val, tag_storage,
arch/powerpc/kernel/rtas.c
643
err = xa_err(xa_store(&rtas_token_to_function_xarray,
arch/x86/kernel/cpu/sgx/virt.c
54
ret = xa_err(xa_store(&vepc->page_array, index, epc_page, GFP_KERNEL));
block/bio.c
102
if (!xa_err(xa_store(&bio_slabs, size, bslab, GFP_KERNEL)))
drivers/base/arch_topology.c
901
xa_store(&hetero_cpu, hetero_id,
drivers/base/auxiliary_sysfs.c
83
xa_store(&auxdev->sysfs.irqs, irq, no_free_ptr(info), GFP_KERNEL);
drivers/base/memory.c
707
ret = xa_err(xa_store(&memory_blocks, memory->dev.id, memory,
drivers/cxl/core/cdat.c
676
ptr = xa_store(usp_xa, index, c, GFP_KERNEL);
drivers/cxl/core/cdat.c
803
ptr = xa_store(res_xa, us_index, n, GFP_KERNEL);
drivers/cxl/core/cdat.c
886
ptr = xa_store(hb_xa, hb_index, n, GFP_KERNEL);
drivers/cxl/core/cdat.c
935
ptr = xa_store(mw_xa, mw_index, n, GFP_KERNEL);
drivers/cxl/core/edac.c
1101
old_rec = xa_store(&array_rec->rec_gen_media,
drivers/cxl/core/edac.c
1131
old_rec = xa_store(&array_rec->rec_dram,
drivers/gpu/drm/amd/amdgpu/amdgpu_acpi.c
914
xa_store(&numa_info_xa, numa_info->pxm, numa_info, GFP_KERNEL);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1220
xa_store(&xa, key, range, GFP_KERNEL);
drivers/gpu/drm/drm_drv.c
199
entry = xa_store(drm_minor_get_xa(type), minor->index, minor, GFP_KERNEL);
drivers/gpu/drm/drm_drv.c
223
xa_store(drm_minor_get_xa(type), minor->index, NULL, GFP_KERNEL);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1738
old = xa_store(&fpriv->context_xa, id, ctx, GFP_KERNEL);
drivers/gpu/drm/i915/gem/i915_gem_context.c
329
old = xa_store(&fpriv->proto_context_xa, *id, pc, GFP_KERNEL);
drivers/gpu/drm/scheduler/sched_main.c
905
xa_store(&job->dependencies, index, fence, GFP_KERNEL);
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
849
return xa_err(xa_store(&vbo->detached_resources, (unsigned long)res, res, GFP_KERNEL));
drivers/gpu/drm/xe/xe_guc_ct.c
1057
ret = xa_err(xa_store(&ct->fence_lookup,
drivers/gpu/drm/xe/xe_guc_ct.c
1286
ret = xa_err(xa_store(&ct->fence_lookup, g2h_fence.seqno,
drivers/gpu/drm/xe/xe_guc_submit.c
454
ret = xa_err(xa_store(&guc->submission_state.exec_queue_lookup,
drivers/gpu/drm/xe/xe_reg_sr.c
99
ret = xa_err(xa_store(&sr->xa, idx, pentry, GFP_KERNEL));
drivers/gpu/drm/xe/xe_tlb_inval_job.c
253
ptr = xa_store(&job->dep.drm.dependencies, 0, fence,
drivers/infiniband/core/cm.c
867
xa_store(&cm.local_id_table, cm_local_id(cm_id_priv->id.local_id),
drivers/infiniband/core/device.c
1135
ret = xa_err(xa_store(&rdma_nets, rnet->id, NULL, GFP_KERNEL));
drivers/infiniband/core/device.c
2022
rc = xa_store(&device->client_data, client->client_id, data,
drivers/infiniband/core/device.c
724
ret = xa_err(xa_store(&device->client_data, client->client_id, NULL,
drivers/infiniband/core/device.c
981
ret = xa_err(xa_store(&device->compat_devs, rnet->id,
drivers/infiniband/core/rdma_core.c
599
old = xa_store(&ufile->idr, uobj->id, uobj, GFP_KERNEL);
drivers/infiniband/core/ucma.c
1567
xa_store(&multicast_table, mc->id, mc, 0);
drivers/infiniband/core/ucma.c
228
xa_store(&ctx_table, ctx->id, ctx, GFP_KERNEL);
drivers/infiniband/core/verbs.c
1256
err = xa_err(xa_store(&qp_init_attr->xrcd->tgt_qps, real_qp->qp_num,
drivers/infiniband/hw/efa/efa_verbs.c
1272
err = xa_err(xa_store(&dev->cqs_xa, cq->cq_idx, cq, GFP_KERNEL));
drivers/infiniband/hw/erdma/erdma_verbs.c
1021
old_entry = xa_store(&dev->qp_xa, 1, qp, GFP_KERNEL);
drivers/infiniband/hw/hns/hns_roce_bond.c
260
ret = xa_err(xa_store(&roce_bond_xa, bus_num, die_info, GFP_KERNEL));
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
5028
ret = xa_err(xa_store(dip_xa, qpn, hr_dip, GFP_KERNEL));
drivers/infiniband/hw/mlx5/fs.c
1151
err = xa_err(xa_store(qpn_opfc_xa, qp_num, per_qp_opfc, GFP_KERNEL));
drivers/infiniband/hw/mlx5/mlx5_ib.h
1668
return xa_err(xa_store(&dev->odp_mkeys, mlx5_base_mkey(mmkey->key),
drivers/infiniband/hw/mlx5/mr.c
2443
err = xa_err(xa_store(&dev->sig_mrs, mlx5_base_mkey(mr->mmkey.key),
drivers/infiniband/hw/mlx5/odp.c
532
ret = xa_store(&dev->odp_mkeys, mlx5_base_mkey(mr->mmkey.key),
drivers/infiniband/sw/rxe/rxe_pool.c
253
xa_ret = xa_store(&elem->pool->xa, elem->index, elem, GFP_KERNEL);
drivers/infiniband/ulp/srpt/ib_srpt.c
137
res = xa_store(&srpt_memory_caches, object_size, e, GFP_KERNEL);
drivers/iommu/iommu.c
3617
WARN_ON(xa_is_err(xa_store(&group->pasid_array,
drivers/iommu/iommu.c
3718
WARN_ON(xa_is_err(xa_store(&group->pasid_array,
drivers/iommu/iommu.c
3855
WARN_ON(xa_is_err(xa_store(&group->pasid_array,
drivers/iommu/iommu.c
3918
curr = xa_store(&group->pasid_array, IOMMU_NO_PASID, entry, GFP_KERNEL);
drivers/iommu/iommufd/device.c
662
WARN_ON(xa_is_err(xa_store(&igroup->pasid_attach, pasid, attach,
drivers/iommu/iommufd/device.c
666
WARN_ON(xa_is_err(xa_store(&attach->device_array, idev->obj.id,
drivers/iommu/iommufd/io_pagetable.c
1194
xa_store(&iopt->domains, iopt->next_domain_id, domain, GFP_KERNEL);
drivers/iommu/iommufd/io_pagetable.c
1267
xa_store(&iopt->domains, index, iter_domain, GFP_KERNEL);
drivers/iommu/iommufd/ioas.c
428
rc = xa_err(xa_store(ioas_list, index, ioas, GFP_KERNEL));
drivers/irqchip/irq-gic-v5-its.c
781
entry = xa_store(&its->its_devices, dev_id, its_dev, GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_adapter.c
117
xa_store(&ice_adapters, index, adapter, GFP_KERNEL);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
1035
err = xa_err(xa_store(&table->comp_eqs, vecidx, eq, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/eq.c
854
return xa_err(xa_store(&table->comp_irqs, vecidx, irq, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/eq.c
905
return xa_err(xa_store(&table->comp_irqs, vecidx, irq, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
4487
old_entry = xa_store(&esw->offloads.vhca_map, vhca_id, vhca_map_entry, GFP_KERNEL);
drivers/net/ethernet/mellanox/mlx5/core/fs_counters.c
295
err = xa_err(xa_store(&fc_stats->counters, id, counter, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
744
err = xa_err(xa_store(fs_id_xa, fs_id, fs_id_iter, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/pci_irq.c
314
err = xa_err(xa_store(&pool->irqs, irq->pool_index, irq, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/context.c
253
if (xa_err(xa_store(&ctx->peer_ctx_xa, peer_vhca_id, peer_ctx, GFP_KERNEL)))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
550
WARN_ON(xa_err(xa_store(&dmn->peer_dmn_xa, peer_vhca_id, peer_dmn, GFP_KERNEL)));
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
88
ret = xa_err(xa_store(&dmn->csum_fts_xa, vport_num,
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
3215
err = xa_err(xa_store(&nh_grp->nhgi->nexthop_counters, nh->id, nhct,
drivers/net/ethernet/microsoft/mana/gdma_main.c
1662
xa_store(&gc->irq_contexts, i, gic, GFP_KERNEL);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1744
xa_store(&gc->irq_contexts, i, gic, GFP_KERNEL);
drivers/nvme/host/core.c
3335
old = xa_store(&ctrl->cels, csi, cel, GFP_KERNEL);
drivers/nvme/host/core.c
3410
old = xa_store(&ctrl->cels, csi, effects, GFP_KERNEL);
drivers/pci/p2pdma.c
769
xa_store(&p2pdma->map_types, map_types_idx(client),
drivers/scsi/elx/efct/efct_lio.c
1205
if (xa_err(xa_store(&efct->lookup, id, tgt_node, GFP_KERNEL)))
drivers/scsi/elx/libefc/efc_domain.c
478
rc = xa_err(xa_store(&domain->lookup, fc_id, domain->nport,
drivers/scsi/elx/libefc/efc_node.c
97
rc = xa_err(xa_store(&nport->lookup, port_id, node, GFP_ATOMIC));
drivers/scsi/elx/libefc/efc_nport.c
185
rc = xa_err(xa_store(&nport->domain->lookup, fc_id, nport, GFP_ATOMIC));
drivers/target/target_core_user.c
533
if (xa_store(&udev->data_pages, dpi + i, page, GFP_NOIO)) {
drivers/tee/tee_heap.c
282
return WARN_ON(xa_is_err(xa_store(&tee_dma_heap, id, h, GFP_KERNEL)));
fs/btrfs/delayed-ref.c
879
existing = xa_store(&delayed_refs->head_refs, index, head_ref, GFP_ATOMIC);
fs/btrfs/inode.c
4027
existing = xa_store(&root->inodes, ino, inode, GFP_ATOMIC);
fs/nfsd/nfs4state.c
1968
xa_store(&ses->se_slots, i, xa_mk_value(slot->sl_seqid), 0);
fs/nfsd/nfs4state.c
2056
if (!slot || xa_is_err(xa_store(&new->se_slots, 0, slot, GFP_KERNEL)))
fs/nfsd/nfs4state.c
2064
if (xa_is_err(xa_store(&new->se_slots, i, slot, gfp))) {
fs/nfsd/nfs4state.c
4516
!xa_is_err(xa_store(&session->se_slots, s, slot,
fs/smb/server/mgmt/tree_connect.c
83
ret = xa_err(xa_store(&sess->tree_conns, tree_conn->id, tree_conn,
fs/smb/server/mgmt/user_session.c
335
old = xa_store(&sess->rpc_handle_list, id, entry, KSMBD_DEFAULT_GFP);
fs/smb/server/mgmt/user_session.c
440
return xa_err(xa_store(&conn->sessions, sess->id, sess, KSMBD_DEFAULT_GFP));
fs/smb/server/smb2pdu.c
1572
old = xa_store(&sess->ksmbd_chann_list, (long)conn, chann,
fs/smb/server/smb2pdu.c
1667
old = xa_store(&sess->ksmbd_chann_list, (long)conn,
include/linux/xarray.h
356
void *xa_store(struct xarray *, unsigned long index, void *entry, gfp_t);
io_uring/kbuf.c
96
return xa_err(xa_store(&ctx->io_bl_xa, bgid, bl, GFP_KERNEL));
io_uring/tctx.c
148
ret = xa_err(xa_store(&tctx->xa, (unsigned long)ctx,
io_uring/zcrx.c
741
if (xa_store(&ctx->zcrx_ctxs, id, ifq, GFP_KERNEL))
io_uring/zcrx.c
855
if (xa_store(&ctx->zcrx_ctxs, id, ifq, GFP_KERNEL))
kernel/trace/trace_syscalls.c
1359
ret = xa_store(&syscalls_metadata_sparse, i, meta,
lib/test_hmm.c
236
entry = xa_store(&dmirror->pt, pfn, entry, GFP_ATOMIC);
lib/test_hmm.c
822
entry = xa_store(&dmirror->pt, addr >> PAGE_SHIFT, entry, GFP_ATOMIC);
lib/test_hmm.c
873
entry = xa_store(&dmirror->pt, pfn + i, entry, GFP_ATOMIC);
lib/test_xarray.c
1013
XA_BUG_ON(xa, xa_store(xa, 3, xa_mk_index(3), GFP_KERNEL) != NULL);
lib/test_xarray.c
1014
XA_BUG_ON(xa, xa_store(xa, 4, xa_mk_index(4), GFP_KERNEL) != NULL);
lib/test_xarray.c
1015
XA_BUG_ON(xa, xa_store(xa, 4, NULL, GFP_KERNEL) != xa_mk_index(4));
lib/test_xarray.c
1972
XA_BUG_ON(xa, xa_store(xa, 0, name + i, GFP_KERNEL) != NULL);
lib/test_xarray.c
1978
XA_BUG_ON(xa, xa_store(xa, 0, name + i, 0) != NULL);
lib/test_xarray.c
2215
xa_store(xa, ULONG_MAX, xa, GFP_KERNEL);
lib/test_xarray.c
42
return xa_store(xa, index, xa_mk_index(index), gfp);
lib/test_xarray.c
667
xa_store(xa, 3, xa, GFP_KERNEL);
lib/test_xarray.c
96
XA_BUG_ON(xa, xa_err(xa_store(xa, 1, xa_mk_value(0), GFP_KERNEL)) != 0);
lib/tests/kunit_iov_iter.c
556
void *x = xa_store(xarray, i, pages[i], GFP_KERNEL);
lib/xarray.c
1739
EXPORT_SYMBOL(xa_store);
mm/memcontrol.c
3909
xa_store(&mem_cgroup_private_ids, memcg->id.id, memcg, GFP_KERNEL);
mm/zswap.c
1425
old = xa_store(swap_zswap_tree(page_swpentry),
net/core/dev.c
423
WARN_ON(xa_store(&net->dev_by_index, dev->ifindex, dev, GFP_KERNEL));
net/hsr/hsr_framereg.c
354
res = xa_store(&node->seq_blocks, block_idx, block, GFP_ATOMIC);
net/psp/psp_main.c
130
xa_store(&psp_devs, psd->id, NULL, GFP_KERNEL);
net/qrtr/af_qrtr.c
274
if (xa_err(xa_store(&node->qrtr_tx_flow, key, flow,
net/qrtr/ns.c
246
old = xa_store(&node->servers, port, srv, GFP_KERNEL);
net/qrtr/ns.c
88
if (xa_store(&nodes, node_id, node, GFP_KERNEL)) {
tools/testing/radix-tree/iteration_check_2.c
43
xa_store(xa, i, xa_mk_value(i), GFP_KERNEL);
tools/testing/radix-tree/iteration_check_2.c
64
xa_store(&array, 100, xa_mk_value(100), GFP_KERNEL);
virt/kvm/kvm_main.c
2593
r = xa_err(xa_store(&kvm->mem_attr_array, i, entry,