Symbol: xa_err
arch/arm64/kernel/hibernate.c
233
return xa_err(ret);
arch/arm64/kvm/emulate-nested.c
2275
ret = xa_err(prev);
arch/arm64/kvm/emulate-nested.c
2320
ret = xa_err(prev);
arch/arm64/kvm/emulate-nested.c
2387
return xa_err(ret);
arch/arm64/kvm/vgic/vgic-its.c
119
ret = xa_err(__xa_store(&dist->lpi_xa, intid, irq, 0));
arch/arm64/mm/mteswap.c
41
return xa_err(ret);
arch/powerpc/kernel/rtas.c
643
err = xa_err(xa_store(&rtas_token_to_function_xarray,
arch/x86/kernel/cpu/sgx/virt.c
54
ret = xa_err(xa_store(&vepc->page_array, index, epc_page, GFP_KERNEL));
block/bio.c
102
if (!xa_err(xa_store(&bio_slabs, size, bslab, GFP_KERNEL)))
drivers/base/memory.c
707
ret = xa_err(xa_store(&memory_blocks, memory->dev.id, memory,
drivers/block/brd.c
114
return ERR_PTR(xa_err(ret));
drivers/cxl/core/cdat.c
678
return xa_err(ptr);
drivers/cxl/core/cdat.c
805
return ERR_PTR(xa_err(ptr));
drivers/cxl/core/cdat.c
888
return ERR_PTR(xa_err(ptr));
drivers/cxl/core/cdat.c
937
return ERR_PTR(xa_err(ptr));
drivers/cxl/core/edac.c
1106
return xa_err(old_rec);
drivers/cxl/core/edac.c
1136
return xa_err(old_rec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
863
r = xa_err(xa_store_irq(&adev->userq_doorbell_xa, index, queue, GFP_KERNEL));
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
108
r = xa_err(__xa_store(&adev->userq_xa, userq->doorbell_index,
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2660
r = xa_err(xa_store_irq(&adev->vm_manager.pasids, pasid, vm, GFP_KERNEL));
drivers/gpu/drm/drm_drv.c
201
ret = xa_err(entry);
drivers/gpu/drm/i915/gem/i915_gem_context.c
332
return xa_err(old);
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
849
return xa_err(xa_store(&vbo->detached_resources, (unsigned long)res, res, GFP_KERNEL));
drivers/gpu/drm/xe/xe_guc_ct.c
1057
ret = xa_err(xa_store(&ct->fence_lookup,
drivers/gpu/drm/xe/xe_guc_ct.c
1286
ret = xa_err(xa_store(&ct->fence_lookup, g2h_fence.seqno,
drivers/gpu/drm/xe/xe_guc_submit.c
454
ret = xa_err(xa_store(&guc->submission_state.exec_queue_lookup,
drivers/gpu/drm/xe/xe_reg_sr.c
99
ret = xa_err(xa_store(&sr->xa, idx, pentry, GFP_KERNEL));
drivers/infiniband/core/device.c
1135
ret = xa_err(xa_store(&rdma_nets, rnet->id, NULL, GFP_KERNEL));
drivers/infiniband/core/device.c
724
ret = xa_err(xa_store(&device->client_data, client->client_id, NULL,
drivers/infiniband/core/device.c
981
ret = xa_err(xa_store(&device->compat_devs, rnet->id,
drivers/infiniband/core/verbs.c
1256
err = xa_err(xa_store(&qp_init_attr->xrcd->tgt_qps, real_qp->qp_num,
drivers/infiniband/hw/efa/efa_verbs.c
1272
err = xa_err(xa_store(&dev->cqs_xa, cq->cq_idx, cq, GFP_KERNEL));
drivers/infiniband/hw/erdma/erdma_verbs.c
1023
ret = xa_err(old_entry);
drivers/infiniband/hw/hns/hns_roce_bond.c
260
ret = xa_err(xa_store(&roce_bond_xa, bus_num, die_info, GFP_KERNEL));
drivers/infiniband/hw/hns/hns_roce_cq.c
207
ret = xa_err(xa_store_irq(&cq_table->array, hr_cq->cqn, hr_cq, GFP_KERNEL));
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
5028
ret = xa_err(xa_store(dip_xa, qpn, hr_dip, GFP_KERNEL));
drivers/infiniband/hw/hns/hns_roce_qp.c
343
ret = xa_err(xa_store_irq(xa, hr_qp->qpn, hr_qp, GFP_KERNEL));
drivers/infiniband/hw/hns/hns_roce_srq.c
125
ret = xa_err(xa_store_irq(&srq_table->xa, srq->srqn, srq, GFP_KERNEL));
drivers/infiniband/hw/ionic/ionic_controlpath.c
151
rc = xa_err(entry);
drivers/infiniband/hw/ionic/ionic_controlpath.c
2336
rc = xa_err(entry);
drivers/infiniband/hw/mlx4/cm.c
370
if (xa_err(item))
drivers/infiniband/hw/mlx4/cm.c
371
ret = xa_err(item);
drivers/infiniband/hw/mlx4/cm.c
392
old, xa_err(old));
drivers/infiniband/hw/mlx4/cm.c
394
return xa_err(old);
drivers/infiniband/hw/mlx4/cm.c
414
if (!item || xa_err(item)) {
drivers/infiniband/hw/mlx4/cm.c
416
rem_pv_cm_id, xa_err(item));
drivers/infiniband/hw/mlx4/cm.c
417
slave = !item ? -ENOENT : xa_err(item);
drivers/infiniband/hw/mlx5/fs.c
1151
err = xa_err(xa_store(qpn_opfc_xa, qp_num, per_qp_opfc, GFP_KERNEL));
drivers/infiniband/hw/mlx5/mlx5_ib.h
1668
return xa_err(xa_store(&dev->odp_mkeys, mlx5_base_mkey(mmkey->key),
drivers/infiniband/hw/mlx5/mr.c
2443
err = xa_err(xa_store(&dev->sig_mrs, mlx5_base_mkey(mr->mmkey.key),
drivers/infiniband/hw/mlx5/odp.c
520
ret = ERR_PTR(xa_err(ret));
drivers/infiniband/hw/mlx5/odp.c
535
ret = ERR_PTR(xa_err(ret));
drivers/infiniband/hw/mlx5/qpc.c
236
err = xa_err(xa_store_irq(&dev->qp_table.dct_xa, qp->qpn, dct, GFP_KERNEL));
drivers/infiniband/hw/mlx5/qpc.c
310
return xa_err(tmp) ?: -EINVAL;
drivers/infiniband/hw/mlx5/srq_cmd.c
653
err = xa_err(xa_store_irq(&table->array, srq->srqn, srq, GFP_KERNEL));
drivers/infiniband/hw/mlx5/srq_cmd.c
674
return xa_err(tmp) ?: -EINVAL;
drivers/infiniband/sw/rxe/rxe_pool.c
191
WARN_ON(xa_err(xa_ret));
drivers/infiniband/sw/rxe/rxe_pool.c
254
WARN_ON(xa_err(xa_ret));
drivers/iommu/amd/nested.c
279
if (WARN_ON(!curr || xa_err(curr)))
drivers/iommu/amd/nested.c
78
res = ERR_PTR(xa_err(res));
drivers/iommu/intel/iommu.c
1058
ret = xa_err(curr) ? : -EBUSY;
drivers/iommu/iommu.c
3679
ret = xa_err(curr);
drivers/iommu/iommufd/device.c
121
return ERR_PTR(xa_err(igroup));
drivers/iommu/iommufd/device.c
617
rc = xa_err(attach);
drivers/iommu/iommufd/ioas.c
428
rc = xa_err(xa_store(ioas_list, index, ioas, GFP_KERNEL));
drivers/iommu/iommufd/viommu.c
223
rc = xa_err(curr) ?: -EEXIST;
drivers/irqchip/irq-gic-v5-its.c
783
ret = xa_err(entry);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
1035
err = xa_err(xa_store(&table->comp_eqs, vecidx, eq, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/eq.c
854
return xa_err(xa_store(&table->comp_irqs, vecidx, irq, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/eq.c
905
return xa_err(xa_store(&table->comp_irqs, vecidx, irq, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
4490
return xa_err(old_entry);
drivers/net/ethernet/mellanox/mlx5/core/fs_counters.c
295
err = xa_err(xa_store(&fc_stats->counters, id, counter, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
744
err = xa_err(xa_store(fs_id_xa, fs_id, fs_id_iter, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/pci_irq.c
314
err = xa_err(xa_store(&pool->irqs, irq->pool_index, irq, GFP_KERNEL));
drivers/net/ethernet/mellanox/mlx5/core/sf/dev/dev.c
137
goto xa_err;
drivers/net/ethernet/mellanox/mlx5/core/sf/dev/dev.c
140
xa_err:
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/context.c
253
if (xa_err(xa_store(&ctx->peer_ctx_xa, peer_vhca_id, peer_ctx, GFP_KERNEL)))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
550
WARN_ON(xa_err(xa_store(&dmn->peer_dmn_xa, peer_vhca_id, peer_dmn, GFP_KERNEL)));
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
88
ret = xa_err(xa_store(&dmn->csum_fts_xa, vport_num,
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
3215
err = xa_err(xa_store(&nh_grp->nhgi->nexthop_counters, nh->id, nhct,
drivers/nvme/host/core.c
3338
return xa_err(old);
drivers/nvme/host/core.c
3413
return xa_err(old);
drivers/nvme/target/passthru.c
613
ret = xa_err(old);
drivers/scsi/elx/efct/efct_lio.c
1205
if (xa_err(xa_store(&efct->lookup, id, tgt_node, GFP_KERNEL)))
drivers/scsi/elx/libefc/efc_domain.c
478
rc = xa_err(xa_store(&domain->lookup, fc_id, domain->nport,
drivers/scsi/elx/libefc/efc_node.c
97
rc = xa_err(xa_store(&nport->lookup, port_id, node, GFP_ATOMIC));
drivers/scsi/elx/libefc/efc_nport.c
185
rc = xa_err(xa_store(&nport->domain->lookup, fc_id, nport, GFP_ATOMIC));
drivers/vfio/vfio_main.c
103
return xa_err(dev_set);
drivers/xen/grant-dma-ops.c
62
ret = xa_err(__xa_store(&xen_grant_dma_devices, (unsigned long)dev, data,
fs/btrfs/delayed-inode.c
173
ASSERT(xa_err(ptr) != -EINVAL);
fs/btrfs/delayed-inode.c
174
ASSERT(xa_err(ptr) != -ENOMEM);
fs/btrfs/delayed-ref.c
882
ASSERT(xa_err(existing) != -ENOMEM);
fs/btrfs/delayed-ref.c
883
return ERR_PTR(xa_err(existing));
fs/btrfs/extent_io.c
3227
ret = xa_err(exists);
fs/btrfs/extent_io.c
3544
ret = xa_err(existing_eb);
fs/btrfs/inode.c
4030
ret = xa_err(existing);
fs/btrfs/qgroup.c
2043
qgroup_mark_inconsistent(fs_info, "xarray insert error: %d", xa_err(ret));
fs/btrfs/qgroup.c
2044
return xa_err(ret);
fs/erofs/zdata.c
787
err = xa_err(pre);
fs/smb/server/mgmt/tree_connect.c
83
ret = xa_err(xa_store(&sess->tree_conns, tree_conn->id, tree_conn,
fs/smb/server/mgmt/user_session.c
440
return xa_err(xa_store(&conn->sessions, sess->id, sess, KSMBD_DEFAULT_GFP));
fs/smb/server/smb2pdu.c
1577
return xa_err(old);
fs/smb/server/smb2pdu.c
1672
return xa_err(old);
include/linux/xarray.h
1088
return xa_err(xa_cmpxchg(xa, index, NULL, XA_ZERO_ENTRY, gfp));
include/linux/xarray.h
1106
return xa_err(xa_cmpxchg_bh(xa, index, NULL, XA_ZERO_ENTRY, gfp));
include/linux/xarray.h
1124
return xa_err(xa_cmpxchg_irq(xa, index, NULL, XA_ZERO_ENTRY, gfp));
include/linux/xarray.h
1437
return xa_err(xas->xa_node);
io_uring/kbuf.c
96
return xa_err(xa_store(&ctx->io_bl_xa, bgid, bl, GFP_KERNEL));
io_uring/tctx.c
148
ret = xa_err(xa_store(&tctx->xa, (unsigned long)ctx,
kernel/liveupdate/kexec_handover.c
139
return ERR_PTR(xa_err(res));
kernel/liveupdate/kexec_handover.c
198
err = xa_err(physxa);
lib/alloc_tag.c
519
ret = ERR_PTR(xa_err(mas.node));
lib/alloc_tag.c
528
ret = ERR_PTR(xa_err(mas.node));
lib/alloc_tag.c
534
ret = ERR_PTR(xa_err(mas.node));
lib/maple_tree.c
4051
return xa_err(mas->node);
lib/maple_tree.c
4765
return xa_err(mas->node);
lib/maple_tree.c
4826
return xa_err(mas->node);
lib/maple_tree.c
5134
ret = xa_err(mas->node);
lib/maple_tree.c
5196
int ret = xa_err(mas->node);
lib/maple_tree.c
5993
ret = xa_err(ms.node);
lib/maple_tree.c
6045
ret = xa_err(mas.node);
lib/maple_tree.c
6127
ret = xa_err(mas.node);
lib/maple_tree.c
6384
ret = xa_err(mas.node);
lib/maple_tree.c
6424
ret = xa_err(mas.node);
lib/test_hmm.c
238
return xa_err(entry);
lib/test_hmm.c
825
return xa_err(entry);
lib/test_hmm.c
876
return xa_err(entry);
lib/test_xarray.c
88
XA_BUG_ON(xa, xa_err(xa_store_index(xa, 0, GFP_NOWAIT)) != 0);
lib/test_xarray.c
89
XA_BUG_ON(xa, xa_err(xa_erase(xa, 0)) != 0);
lib/test_xarray.c
92
XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_NOWAIT)) != -ENOMEM);
lib/test_xarray.c
93
XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_NOWAIT)) != -ENOMEM);
lib/test_xarray.c
95
XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_KERNEL)) != 0);
lib/test_xarray.c
96
XA_BUG_ON(xa, xa_err(xa_store(xa, 1, xa_mk_value(0), GFP_KERNEL)) != 0);
lib/test_xarray.c
97
XA_BUG_ON(xa, xa_err(xa_erase(xa, 1)) != 0);
lib/xarray.c
1815
errno = xa_err(curr);
mm/memremap.c
180
error = xa_err(xa_store_range(&pgmap_array, PHYS_PFN(range->start),
mm/zswap.c
1429
int err = xa_err(old);
net/devlink/region.c
298
err = xa_err(__xa_store(&devlink->snapshot_ids, id, xa_mk_value(count),
net/devlink/region.c
372
err = xa_err(__xa_store(&devlink->snapshot_ids, id, xa_mk_value(0),
net/netlink/genetlink.c
761
return ERR_PTR(xa_err(old_priv));
net/qrtr/af_qrtr.c
274
if (xa_err(xa_store(&node->qrtr_tx_flow, key, flow,
net/qrtr/ns.c
250
srv->service, srv->instance, xa_err(old));
net/shaper/shaper.c
380
if (xa_err(prev)) {
net/shaper/shaper.c
383
ret = xa_err(prev);
rust/helpers/xarray.c
7
return xa_err(entry);
virt/kvm/kvm_main.c
2593
r = xa_err(xa_store(&kvm->mem_attr_array, i, entry,