xa_limit_32b
xa_limit_32b, GFP_NOWAIT);
ret = xa_alloc_cyclic(&dpll_device_xa, &dpll->id, dpll, xa_limit_32b,
ret = xa_alloc_cyclic(&dpll_pin_xa, &pin->id, pin, xa_limit_32b,
ret = xa_alloc(&client->resource_xa, &index, resource, xa_limit_32b,
xa_limit_32b, GFP_NOWAIT);
xa_limit_32b, GFP_KERNEL);
xa_limit_32b, GFP_KERNEL);
ret = xa_alloc(&file_private->syncobj_xa, handle, syncobj, xa_limit_32b,
ret = xa_alloc(&file_private->syncobj_xa, handle, syncobj, xa_limit_32b,
xa_limit_32b, &priv->next_context_id, GFP_KERNEL);
submit->out_fence, xa_limit_32b,
xa_limit_32b, GFP_KERNEL);
err = xa_alloc(&file_priv->vm_xa, &id, vm, xa_limit_32b, GFP_KERNEL);
xa_limit_32b, GFP_KERNEL);
ret = xa_alloc(&fpriv->context_xa, id, NULL, xa_limit_32b, GFP_KERNEL);
err = xa_alloc(&fpriv->context_xa, &id, NULL, xa_limit_32b, GFP_KERNEL);
xa_limit_32b, &guc->next_seqno, GFP_KERNEL);
xa_limit_32b, &guc->next_seqno,
err = xa_alloc(&pvr_dev->ctx_ids, &ctx->ctx_id, ctx, xa_limit_32b, GFP_KERNEL);
err = xa_alloc(&pvr_file->ctx_handles, &args->handle, ctx, xa_limit_32b, GFP_KERNEL);
xa_limit_32b,
xa_limit_32b,
xa_limit_32b,
xa_limit_32b,
err = xa_alloc(&pvr_dev->job_ids, &job->id, job, xa_limit_32b, GFP_KERNEL);
err = xa_alloc(array, &id, sig_sync, xa_limit_32b, GFP_KERNEL);
err = xa_alloc(&mgr->handles, id, ctx, xa_limit_32b, GFP_KERNEL);
ret = xa_alloc(&job->dependencies, &id, fence, xa_limit_32b, GFP_KERNEL);
xa_limit_32b, GFP_KERNEL);
xa_limit_32b, GFP_KERNEL);
encoder_cfg, xa_limit_32b, GFP_KERNEL);
err = xa_alloc(&xef->exec_queue.xa, &id, q, xa_limit_32b, GFP_KERNEL);
err = xa_alloc(&xef->vm.xa, &id, vm, xa_limit_32b, GFP_KERNEL);
&i, svm_range, xa_limit_32b,
ret = xa_alloc_cyclic(&cm.local_id_table, &id, NULL, xa_limit_32b,
ret = xa_alloc(&rdma_nets, &rnet->id, rnet, xa_limit_32b, GFP_KERNEL);
return xa_alloc(&uobj->ufile->idr, &uobj->id, NULL, xa_limit_32b,
ret = xa_alloc_cyclic(&rt->xa, &res->id, res, xa_limit_32b,
ret = __xa_alloc(&queries, &id, query, xa_limit_32b, gfp_mask);
if (__xa_alloc(&multicast_table, &mc->id, NULL, xa_limit_32b,
if (xa_alloc(&ctx_table, &ctx->id, NULL, xa_limit_32b, GFP_KERNEL)) {
ret = xa_alloc_irq(&hfi1_dev_table, &dd->unit, dd, xa_limit_32b,
xa_limit_32b, &sriov->pv_id_next, GFP_KERNEL);
int rv = xa_alloc(&sdev->qp_xa, &qp->base_qp.qp_num, qp, xa_limit_32b,
xa_limit_32b, GFP_KERNEL);
ret = xa_alloc(&uacce_xa, &uacce->dev_id, uacce, xa_limit_32b,
return xa_alloc(&pf->sf_nums, sfnum, NULL, xa_limit_32b,
err = xa_alloc(&ice_sf_aux_id, &id, NULL, xa_limit_32b,
xa_limit_32b, &topo->next_phy_index,
rc = xa_alloc(&ctxdata->sess_list, &sess_id, sess, xa_limit_32b,
ret = xa_alloc(&dev->worker_xa, &id, worker, xa_limit_32b, GFP_KERNEL);
ret = xa_alloc(&dev->worker_xa, &id, worker, xa_limit_32b, GFP_KERNEL);
XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b, GFP_KERNEL) != 0);
XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b,
XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b, GFP_KERNEL) != 0);
xa_limit_32b, &next, GFP_KERNEL) != 0);
xa_limit_32b, &next, GFP_KERNEL) != 0);
xa_limit_32b, &next, GFP_KERNEL) != 0);
xa_limit_32b, &next, GFP_KERNEL) != 0);
ret = __xa_alloc_cyclic(xa, &id, xa_mk_index(base), xa_limit_32b,
XA_BUG_ON(xa, xa_alloc(xa, &id, name + i, xa_limit_32b,
XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(index), xa_limit_32b,
XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b, GFP_KERNEL) != 0);
err = xa_alloc(&binding->bound_rxqs, &xa_idx, rxq, xa_limit_32b,
binding, xa_limit_32b, &id_alloc_next,
err = xa_alloc_cyclic(&page_pools, &pool->user.id, pool, xa_limit_32b,
xa_limit_32b, &next, GFP_KERNEL);
xa_limit_32b, GFP_KERNEL);
n, xa_limit_32b, &next, GFP_KERNEL);
if (xa_alloc(&rd->rd_xa, &rn->rn_index, rn, xa_limit_32b, GFP_KERNEL) < 0)