Symbol: xa_alloc
drivers/base/memory.c
1087
ret = xa_alloc(&memory_groups, &mgid, new_group, xa_limit_31b,
drivers/block/rnbd/rnbd-srv.c
429
error = xa_alloc(&srv_sess->index_idr, &sess_dev->device_id, sess_dev,
drivers/dma-buf/dma-heap.c
256
ret = xa_alloc(&dma_heap_minors, &minor, heap,
drivers/firewire/core-cdev.c
519
ret = xa_alloc(&client->resource_xa, &index, resource, xa_limit_32b,
drivers/firewire/core-device.c
1162
ret = xa_alloc(&fw_device_xa, &minor, device, XA_LIMIT(0, MINORMASK), GFP_KERNEL);
drivers/fpga/intel-m10-bmc-sec-update.c
703
ret = xa_alloc(&fw_upload_xa, &sec->fw_name_id, sec,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
870
r = xa_alloc(&uq_mgr->userq_xa, &qid, queue,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
966
r = xa_alloc(&waitq->fence_drv_xa, &index, fence_drv,
drivers/gpu/drm/drm_drv.c
154
r = xa_alloc(drm_minor_get_xa(type), &minor->index,
drivers/gpu/drm/drm_drv.c
157
r = xa_alloc(&drm_minors_xa, &minor->index,
drivers/gpu/drm/drm_syncobj.c
604
ret = xa_alloc(&file_private->syncobj_xa, handle, syncobj, xa_limit_32b,
drivers/gpu/drm/drm_syncobj.c
718
ret = xa_alloc(&file_private->syncobj_xa, handle, syncobj, xa_limit_32b,
drivers/gpu/drm/i915/gem/i915_gem_context.c
1834
err = xa_alloc(&file_priv->vm_xa, &id, &ppgtt->vm,
drivers/gpu/drm/i915/gem/i915_gem_context.c
1892
err = xa_alloc(&file_priv->vm_xa, &id, vm, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/i915/gem/i915_gem_context.c
2409
ret = xa_alloc(&ext_data.fpriv->context_xa, &id, NULL,
drivers/gpu/drm/i915/gem/i915_gem_context.c
325
ret = xa_alloc(&fpriv->context_xa, id, NULL, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/i915/gem/selftests/mock_context.c
97
err = xa_alloc(&fpriv->context_xa, &id, NULL, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/imagination/pvr_context.c
334
err = xa_alloc(&pvr_dev->ctx_ids, &ctx->ctx_id, ctx, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/imagination/pvr_context.c
338
err = xa_alloc(&pvr_file->ctx_handles, &args->handle, ctx, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/imagination/pvr_drv.c
758
err = xa_alloc(&pvr_file->free_list_handles,
drivers/gpu/drm/imagination/pvr_drv.c
844
err = xa_alloc(&pvr_file->hwrt_handles,
drivers/gpu/drm/imagination/pvr_drv.c
935
err = xa_alloc(&pvr_file->vm_ctx_handles,
drivers/gpu/drm/imagination/pvr_free_list.c
433
err = xa_alloc(&pvr_file->pvr_dev->free_list_ids,
drivers/gpu/drm/imagination/pvr_job.c
426
err = xa_alloc(&pvr_dev->job_ids, &job->id, job, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/imagination/pvr_sync.c
95
err = xa_alloc(array, &id, sig_sync, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/lima/lima_ctx.c
27
err = xa_alloc(&mgr->handles, id, ctx, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/panfrost/panfrost_job.c
1075
ret = xa_alloc(&priv->jm_ctxs, &args->handle, jm_ctx,
drivers/gpu/drm/panthor/panthor_heap.c
332
ret = xa_alloc(&pool->xa, &id, heap,
drivers/gpu/drm/panthor/panthor_mmu.c
1477
ret = xa_alloc(&pool->xa, &id, vm,
drivers/gpu/drm/panthor/panthor_sched.c
3730
ret = xa_alloc(&gpool->xa, &gid, group, XA_LIMIT(1, MAX_GROUPS_PER_POOL), GFP_KERNEL);
drivers/gpu/drm/scheduler/sched_main.c
912
ret = xa_alloc(&job->dependencies, &id, fence, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/tegra/uapi.c
138
err = xa_alloc(&fpriv->contexts, &args->context, context, XA_LIMIT(1, U32_MAX),
drivers/gpu/drm/tegra/uapi.c
255
err = xa_alloc(&context->mappings, &args->mapping, mapping, XA_LIMIT(1, U32_MAX),
drivers/gpu/drm/vkms/vkms_config.c
429
return xa_alloc(&plane_cfg->possible_crtcs, &crtc_idx, crtc_cfg,
drivers/gpu/drm/vkms/vkms_config.c
572
return xa_alloc(&encoder_cfg->possible_crtcs, &crtc_idx, crtc_cfg,
drivers/gpu/drm/vkms/vkms_config.c
632
return xa_alloc(&connector_cfg->possible_encoders, &encoder_idx,
drivers/gpu/drm/xe/xe_exec_queue.c
1237
err = xa_alloc(&xef->exec_queue.xa, &id, q, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/xe/xe_exec_queue.c
766
err = xa_alloc(&group->xa, &pos, xe_lrc_get(q->lrc[0]),
drivers/gpu/drm/xe/xe_irq.c
940
ret = xa_alloc(&xe->irq.msix.indexes, &id, irq_buf, limit, GFP_KERNEL);
drivers/gpu/drm/xe/xe_vm.c
1978
err = xa_alloc(&xef->vm.xa, &id, vm, xa_limit_32b, GFP_KERNEL);
drivers/gpu/drm/xe/xe_vm.c
2396
err = xa_alloc(&op->prefetch_range.range,
drivers/infiniband/core/device.c
1181
ret = xa_alloc(&rdma_nets, &rnet->id, rnet, xa_limit_32b, GFP_KERNEL);
drivers/infiniband/core/rdma_core.c
304
return xa_alloc(&uobj->ufile->idr, &uobj->id, NULL, xa_limit_32b,
drivers/infiniband/core/ucma.c
210
if (xa_alloc(&ctx_table, &ctx->id, NULL, xa_limit_32b, GFP_KERNEL)) {
drivers/infiniband/hw/ionic/ionic_hw_stats.c
251
err = xa_alloc(&dev->counter_stats->xa_counters, &counter->id,
drivers/infiniband/sw/siw/siw_qp.c
1284
int rv = xa_alloc(&sdev->qp_xa, &qp->base_qp.qp_num, qp, xa_limit_32b,
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
351
ret = xa_alloc(&arm_smmu_asid_xa, &asid, smmu_domain,
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
2567
ret = xa_alloc(&arm_smmu_asid_xa, &asid, smmu_domain,
drivers/iommu/iommufd/eventq.c
148
rc = xa_alloc(&fault->response, &group->cookie, group,
drivers/iommu/iommufd/io_pagetable.c
1461
rc = xa_alloc(&iopt->access_list, &new_id, access, xa_limit_16b,
drivers/iommu/iommufd/main.c
56
rc = xa_alloc(&ictx->objects, &obj->id, XA_ZERO_ENTRY, xa_limit_31b,
drivers/misc/uacce/uacce.c
517
ret = xa_alloc(&uacce_xa, &uacce->dev_id, uacce, xa_limit_32b,
drivers/net/ethernet/intel/ice/devlink/port.c
835
return xa_alloc(&pf->sf_nums, sfnum, NULL, xa_limit_32b,
drivers/net/ethernet/intel/ice/ice_idc.c
425
ret = xa_alloc(&ice_aux_id, &pf->aux_idx, NULL, XA_LIMIT(1, INT_MAX),
drivers/net/ethernet/intel/ice/ice_irq.c
92
ret = xa_alloc(&pf->irq_tracker.entries, &index, entry, limit,
drivers/net/ethernet/intel/ice/ice_sched.c
980
status = xa_alloc(&pi->sched_node_ids, &new_node->id, NULL, XA_LIMIT(0, UINT_MAX),
drivers/net/ethernet/intel/ice/ice_sf_eth.c
269
err = xa_alloc(&ice_sf_aux_id, &id, NULL, xa_limit_32b,
drivers/net/ethernet/mellanox/mlx5/core/en/mapping.c
66
err = xa_alloc(&ctx->xarray, &mi->id, mi, XA_LIMIT(1, ctx->max_id),
drivers/net/ethernet/mellanox/mlx5/core/en/tc/post_act.c
137
err = xa_alloc(&post_act->ids, &handle->id, post_attr,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
717
err = xa_alloc(&macsec->sc_xarray, &sc_xarray_element->fs_id, sc_xarray_element,
drivers/net/ethernet/mellanox/mlx5/core/irq_affinity.c
59
err = xa_alloc(&pool->irqs, &irq_index, NULL, pool->xa_num_irqs, GFP_KERNEL);
drivers/net/ethernet/mellanox/mlx5/core/lib/st.c
136
ret = xa_alloc(&st->idx_xa, &xa_id, idx_data, st->index_limit, GFP_KERNEL);
drivers/net/ethernet/mellanox/mlxsw/spectrum_port_range.c
62
err = xa_alloc(&pr_core->prr_xa, &prr->index, prr, pr_core->prr_ids,
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
526
err = xa_alloc(&nn->xa_ipsec, &saidx, x,
drivers/net/pse-pd/pse_core.c
929
ret = xa_alloc(&pse_pw_d_map, &index, pw_d, XA_LIMIT(1, PSE_PW_D_LIMIT),
drivers/opp/core.c
2633
ret = xa_alloc(&opp_configs, &id, data, XA_LIMIT(1, INT_MAX),
drivers/platform/x86/intel/pmt/class.c
303
ret = xa_alloc(ns->xa, &entry->devid, entry, PMT_XA_LIMIT, GFP_KERNEL);
drivers/platform/x86/intel/vsec.c
217
ret = xa_alloc(&auxdev_array, &intel_vsec_dev->id, intel_vsec_dev,
drivers/ptp/ptp_clock.c
337
err = xa_alloc(&ptp_clocks_map, &index, ptp, xa_limit_31b,
drivers/target/target_core_tpg.c
458
ret = xa_alloc(&tpg_xa, &val, se_tpg,
drivers/target/target_core_user.c
1075
if (xa_alloc(&udev->commands, &cmd_id, tcmu_cmd, XA_LIMIT(1, 0xffff),
drivers/tee/tstee/core.c
131
rc = xa_alloc(&ctxdata->sess_list, &sess_id, sess, xa_limit_32b,
drivers/tty/serial/liteuart.c
311
ret = xa_alloc(&liteuart_array, &dev_id, uart, limit, GFP_KERNEL);
drivers/vhost/vhost.c
775
ret = xa_alloc(&dev->worker_xa, &id, worker, xa_limit_32b, GFP_KERNEL);
drivers/vhost/vhost.c
797
ret = xa_alloc(&dev->worker_xa, &id, worker, xa_limit_32b, GFP_KERNEL);
fs/dlm/lock.c
1525
rv = xa_alloc(&ls->ls_lkbxa, &lkb->lkb_id, lkb, limit, GFP_ATOMIC);
fs/dlm/recover.c
323
rv = xa_alloc(&ls->ls_recover_xa, &id, r, limit, GFP_ATOMIC);
include/linux/xarray.h
1362
struct xa_node *xa_alloc;
include/linux/xarray.h
1383
.xa_alloc = NULL, \
io_uring/zcrx.c
727
ret = xa_alloc(&ctx->zcrx_ctxs, &id, NULL, xa_limit_31b, GFP_KERNEL);
io_uring/zcrx.c
813
ret = xa_alloc(&ctx->zcrx_ctxs, &id, NULL, xa_limit_31b, GFP_KERNEL);
kernel/irq/msi.c
117
ret = xa_alloc(xa, &index, desc, limit, GFP_KERNEL);
lib/test_xarray.c
1001
XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b, GFP_KERNEL) != 0);
lib/test_xarray.c
1008
XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b,
lib/test_xarray.c
1017
XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b, GFP_KERNEL) != 0);
lib/test_xarray.c
1952
XA_BUG_ON(xa, xa_alloc(xa, &id, name + i, xa_limit_32b,
lib/test_xarray.c
532
XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_value(8),
lib/test_xarray.c
537
XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_value(6),
lib/test_xarray.c
55
XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(index), xa_limit_32b,
lib/test_xarray.c
961
XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(UINT_MAX - 1),
lib/test_xarray.c
965
XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(UINT_MAX),
lib/test_xarray.c
970
XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(0),
lib/test_xarray.c
976
XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(10), XA_LIMIT(10, 5),
lib/test_xarray.c
979
XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(10), XA_LIMIT(10, 5),
lib/test_xarray.c
993
XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b, GFP_KERNEL) != 0);
lib/xarray.c
1064
RCU_INIT_POINTER(node->parent, xas->xa_alloc);
lib/xarray.c
1065
xas->xa_alloc = node;
lib/xarray.c
1103
struct xa_node *child = xas->xa_alloc;
lib/xarray.c
1105
xas->xa_alloc = rcu_dereference_raw(child->parent);
lib/xarray.c
1196
struct xa_node *child = xas->xa_alloc;
lib/xarray.c
1221
RCU_INIT_POINTER(child->parent, xas->xa_alloc);
lib/xarray.c
1225
xas->xa_alloc = rcu_dereference_raw(child->parent);
lib/xarray.c
272
struct xa_node *next, *node = xas->xa_alloc;
lib/xarray.c
278
xas->xa_alloc = node = next;
lib/xarray.c
309
xas->xa_alloc = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
lib/xarray.c
310
if (!xas->xa_alloc)
lib/xarray.c
312
xas->xa_alloc->parent = NULL;
lib/xarray.c
313
XA_NODE_BUG_ON(xas->xa_alloc, !list_empty(&xas->xa_alloc->private_list));
lib/xarray.c
341
xas->xa_alloc = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
lib/xarray.c
344
xas->xa_alloc = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
lib/xarray.c
346
if (!xas->xa_alloc)
lib/xarray.c
348
xas->xa_alloc->parent = NULL;
lib/xarray.c
349
XA_NODE_BUG_ON(xas->xa_alloc, !list_empty(&xas->xa_alloc->private_list));
lib/xarray.c
365
struct xa_node *node = xas->xa_alloc;
lib/xarray.c
371
xas->xa_alloc = NULL;
mm/memcontrol.c
3758
error = xa_alloc(&mem_cgroup_private_ids, &memcg->id.id, NULL,
net/core/devmem.c
171
err = xa_alloc(&binding->bound_rxqs, &xa_idx, rxq, xa_limit_32b,
net/core/page_pool.c
485
err = xa_alloc(&pool->dma_mapped, &id, netmem_to_page(netmem),
net/devlink/region.c
396
return xa_alloc(&devlink->snapshot_ids, id, xa_mk_value(1),
net/ethtool/ioctl.c
1670
ret = xa_alloc(&dev->ethtool->rss_ctx, &ctx_id, ctx,
net/ethtool/rss.c
1059
ret = xa_alloc(&dev->ethtool->rss_ctx, &req.rss_context, ctx,
net/qrtr/af_qrtr.c
734
rc = xa_alloc(&qrtr_ports, port, ipc, QRTR_EPH_PORT_RANGE,
net/shaper/shaper.c
356
ret = xa_alloc(&hierarchy->shapers, &index, NULL,
net/sunrpc/xprtrdma/ib_client.c
65
if (xa_alloc(&rd->rd_xa, &rn->rn_index, rn, xa_limit_32b, GFP_KERNEL) < 0)