idr_alloc
if (idr_alloc(&kvm->arch.kvm_nested_guest_idr,
ret = idr_alloc(&hv->conn_to_evt, eventfd, conn_id, conn_id + 1,
ret = idr_alloc(&kvm->arch.xen.evtchn_ports, evtchnfd, port, port + 1,
hdl_id = idr_alloc(&mgr->handles, handle, 1, 0, GFP_ATOMIC);
rc = idr_alloc(&ctx_mgr->handles, ctx, 1, 0, GFP_KERNEL);
id = idr_alloc(&hl_devs_idr, hdev, 0, HL_MAX_MINORS, GFP_KERNEL);
handle = idr_alloc(&vm->phys_pg_pack_handles, phys_pg_pack, 1, 0,
rc = idr_alloc(&mmg->handles, buf, 1, 0, GFP_ATOMIC);
error = idr_alloc(&protocols, pr, 0, 0, GFP_KERNEL);
id = idr_alloc(&parport_list, port, port->number, port->number,
id1 = idr_alloc(&card->idr, handle1, 0, 0, GFP_ATOMIC);
id2 = idr_alloc(&card->idr, handle2, 0, 0, GFP_ATOMIC);
id = idr_alloc(&drbd_devices, device, minor, minor + 1, GFP_KERNEL);
id = idr_alloc(&resource->devices, device, vnr, vnr + 1, GFP_KERNEL);
id = idr_alloc(&connection->peer_devices, peer_device, vnr, vnr + 1, GFP_KERNEL);
err = idr_alloc(&loop_index_idr, lo, i, i + 1, GFP_KERNEL);
err = idr_alloc(&loop_index_idr, lo, 0, 0, GFP_KERNEL);
err = idr_alloc(&nbd_index_idr, nbd, index, index + 1,
err = idr_alloc(&nbd_index_idr, nbd, 0,
err = idr_alloc(&ublk_index_idr, ub, i, i + 1, GFP_NOWAIT);
err = idr_alloc(&ublk_index_idr, ub, 0, UBLK_MAX_UBLKS,
ret = idr_alloc(&zloop_index_idr, zlo,
ret = idr_alloc(&zloop_index_idr, zlo, 0, 0, GFP_KERNEL);
ret = idr_alloc(&zram_index_idr, zram, 0, 0, GFP_KERNEL);
rc = idr_alloc(&dev_nums_idr, NULL, 0, TPM_NUM_DEVICES, GFP_KERNEL);
crypto_async_idr = idr_alloc(&eip93->ring->crypto_async_idr, async, 0,
crypto_async_idr = idr_alloc(&eip93->ring->crypto_async_idr, async, 0,
ret = idr_alloc(&dca_idr, dca, 0, 0, GFP_NOWAIT);
ret = idr_alloc(&scmi_requested_devices, (void *)phead,
ret = idr_alloc(&info->protocols, pi, proto->id, proto->id + 1,
goto idr_alloc;
idr_alloc:
ret = idr_alloc(idr, cinfo, prot_id, prot_id + 1, GFP_KERNEL);
ret = idr_alloc(&info->active_protocols, child,
id = idr_alloc(&dfl_devs[type].id, dev, 0, 0, GFP_KERNEL);
ret = idr_alloc(&gpio_aggregator_idr, new, 0, 0, GFP_KERNEL);
r = idr_alloc(&fpriv->bo_list_handles, list, 1, 0, GFP_KERNEL);
r = idr_alloc(&mgr->ctx_handles, ctx, 1, AMDGPU_VM_MAX_NUM_CTX, GFP_KERNEL);
idr_handle = idr_alloc(&pdd->alloc_idr, *kgd_mem, bo_priv->idr_handle,
id = idr_alloc(&p->event_idr, ev, *restore_id, *restore_id + 1,
id = idr_alloc(&p->event_idr, ev, 0, p->signal_mapped_size / 8,
id = idr_alloc(&p->event_idr, ev, *restore_id, *restore_id + 1,
id = idr_alloc(&p->event_idr, ev, KFD_FIRST_NONSIGNAL_EVENT_ID,
id = idr_alloc(&p->event_idr, NULL, 0, 1, GFP_KERNEL);
return idr_alloc(&pdd->alloc_idr, mem, 0, 0, GFP_KERNEL);
index = idr_alloc(&aux_idr, aux_dev, 0, DRM_AUX_MINORS, GFP_KERNEL);
ret = idr_alloc(&file_priv->master->magic_map, file_priv,
ret = idr_alloc(&dev->mode_config.tile_idr, tg, 1, 0, GFP_KERNEL);
ret = idr_alloc(&file_priv->object_idr, obj, handle, handle + 1,
ret = idr_alloc(&file_priv->object_idr, NULL, 1, 0, GFP_NOWAIT);
ret = idr_alloc(&dev->object_name_idr, obj, 1, 0, GFP_KERNEL);
id = idr_alloc(&(drm_lease_owner(lessor)->lessee_idr), lessee, 1, 0, GFP_KERNEL);
ret = idr_alloc(leases, &drm_lease_idr_object , object_id, object_id + 1, GFP_KERNEL);
ret = idr_alloc(leases, &drm_lease_idr_object, crtc->primary->base.id, crtc->primary->base.id + 1, GFP_KERNEL);
ret = idr_alloc(leases, &drm_lease_idr_object, crtc->cursor->base.id, crtc->cursor->base.id + 1, GFP_KERNEL);
ret = idr_alloc(&dev->mode_config.object_idr, register_obj ? obj : NULL,
err = idr_alloc(&to_drm_file(file)->object_idr,
ret = idr_alloc(&vgpu->object_idr, dmabuf_obj, 1, 0, GFP_NOWAIT);
ret = idr_alloc(&gvt->vgpu_idr, vgpu, IDLE_VGPU_IDR + 1, GVT_MAX_VGPU,
oa_config->id = idr_alloc(&perf->metrics_idr,
oa_config->id = idr_alloc(&perf->metrics_idr, oa_config, 2, 0, GFP_KERNEL);
id = idr_alloc(&gsp->client_id.idr, client, 0, NVKM_RM_CLIENT_MASK + 1, GFP_KERNEL);
idr_ret = idr_alloc(&qdev->surf_id_idr, NULL, 1, 0, GFP_NOWAIT);
handle = idr_alloc(&qdev->release_idr, release, 1, 0, GFP_NOWAIT);
err = idr_alloc(&fpriv->legacy_contexts, context, 1, 0, GFP_KERNEL);
ret = idr_alloc(&v3d_priv->perfmon.idr, perfmon, V3D_PERFMONID_MIN,
ret = idr_alloc(&vc4file->perfmon.idr, perfmon, VC4_PERFMONID_MIN,
ret = idr_alloc(&vfile->fence_idr, fence, 1, 0, GFP_KERNEL);
ret = idr_alloc(&tdev->idr, base, 1, 0, GFP_NOWAIT);
ret = idr_alloc(idr, res, 1, 0, GFP_NOWAIT);
oa_config->id = idr_alloc(&oa->metrics_idr, oa_config, 1, 0, GFP_KERNEL);
ret = idr_alloc(&port_table_idr, info, PORTID_MIN,
ret = idr_alloc(&drvdata->idr, etr_buf, pid, pid + 1, GFP_KERNEL);
id = idr_alloc(&i2c_adapter_idr, adap, adap->nr, adap->nr + 1, GFP_KERNEL);
id = idr_alloc(&i2c_adapter_idr, adapter,
ret = idr_alloc(&i3c_bus_idr, i3cbus, start, end, GFP_KERNEL);
id = idr_alloc(&icc_idr, node, ICC_DYN_ID_START, 0, GFP_KERNEL);
id = idr_alloc(&icc_idr, node, id, id + 1, GFP_KERNEL);
r = idr_alloc(&_minor_idr, MINOR_ALLOCED, minor, minor + 1, GFP_NOWAIT);
r = idr_alloc(&_minor_idr, MINOR_ALLOCED, 0, 1 << MINORBITS, GFP_NOWAIT);
rc = idr_alloc(&memstick_host_idr, host, 0, 0, GFP_NOWAIT);
msb->disk_id = idr_alloc(&msb_disk_idr, card, 0, 256, GFP_KERNEL);
disk_id = idr_alloc(&mspro_block_disk_idr, card, 0, 256, GFP_KERNEL);
ret = idr_alloc(&c2port_idr, c2dev, 0, 0, GFP_NOWAIT);
ret = idr_alloc(&rtsx_pci_idr, pcr, 0, 0, GFP_NOWAIT);
ret = idr_alloc(&mei_idr, dev, 0, MEI_MAX_DEVS, GFP_KERNEL);
irq->id = idr_alloc(&ctx->irq_idr, irq, 0, MAX_IRQ_PER_CONTEXT,
pasid = idr_alloc(&afu->contexts_idr, ctx, afu->pasid_base,
minor = idr_alloc(&minors_idr, info, 0, OCXL_NUM_MINORS, GFP_KERNEL);
rc = idr_alloc(&tifm_adapter_idr, fm, 0, 0, GFP_NOWAIT);
i = idr_alloc(&mtd_idr, mtd, ofidx, ofidx + 1, GFP_KERNEL);
i = idr_alloc(&mtd_idr, mtd, 0, 0, GFP_KERNEL);
gd->first_minor = idr_alloc(&ubiblock_minor_idr, dev, 0, 0, GFP_KERNEL);
i = idr_alloc(&vf->fdir.fdir_rule_idr, conf, 0,
index = idr_alloc(&mlxsw_sp->pgt->pgt_idr, NULL, 0,
id = idr_alloc(&family->policer_idr, policer, family->start_index,
id = idr_alloc(&router->rif_mac_profiles_idr, profile, 0,
id = idr_alloc(&priv->internal_ports.port_ids, netdev,
unit = idr_alloc(p, ptr, n, n + 1, GFP_KERNEL);
return idr_alloc(p, ptr, min, 0, GFP_KERNEL);
retval = idr_alloc(&tap_major->minor_idr, tap, 1, TAP_NUM_DEVS, GFP_ATOMIC);
ret = idr_alloc(&htt->pending_tx, skb, 0,
ret = idr_alloc(&wmi->mgmt_pending_tx, pkt_addr, 0,
buf_id = idr_alloc(&ring->bufs_idr, buff, 0, ring->bufs_max, GFP_ATOMIC);
*buf_id = idr_alloc(&rx_ring->bufs_idr, skb, 0,
buf_id = idr_alloc(&rx_refill_ring->bufs_idr, defrag_skb, 0,
buf_id = idr_alloc(&rx_ring->bufs_idr, skb, 1,
ret = idr_alloc(&tx_ring->txbuf_idr, skb, 0,
buf_id = idr_alloc(&ar->txmgmt_idr, skb, 0,
buf_id = idr_alloc(&ring->bufs_idr, buff, 0, ring->bufs_max, gfp);
*buf_id = idr_alloc(&rx_ring->bufs_idr, skb, 0,
buf_id = idr_alloc(&buf_ring->bufs_idr, skb, 0,
buf_id = idr_alloc(&rx_ring->bufs_idr, skb, 0,
buf_id = idr_alloc(&ar->txmgmt_idr, skb, 0,
id = idr_alloc(&priv->ack_status_frames, orig_skb,
token = idr_alloc(&dev->token, *ptxwi, 0, dev->token_size, GFP_ATOMIC);
pid = idr_alloc(&wcid->pktid, skb, MT_PACKET_ID_FIRST,
token = idr_alloc(&dev->token, *ptxwi, dev->token_start,
token = idr_alloc(&dev->rx_token, t, 0, dev->rx_token_size,
ovcs->id = idr_alloc(&ovcs_idr, ovcs, 1, 0, GFP_KERNEL);
index = idr_alloc(&functions_idr, epf_group, 0, 0, GFP_KERNEL);
handle = idr_alloc(&instance->context_map, msg_context,
num = idr_alloc(&bq2415x_id, client, 0, 0, GFP_KERNEL);
bq->id = idr_alloc(&bq25890_id, bq, 0, 0, GFP_KERNEL);
result = idr_alloc(power_zone->parent_idr, NULL, 0, 0, GFP_KERNEL);
err = idr_alloc(&pps_idr, pps, 0, PPS_MAX_SOURCES, GFP_KERNEL);
err = idr_alloc(&ptp_ocp_idr, bp, 0, 0, GFP_KERNEL);
ret = idr_alloc(&pwm_chips, chip, 0, 0, GFP_KERNEL);
ret = idr_alloc(&rproc->notifyids, rvring, 0, 0, GFP_KERNEL);
return idr_alloc(&data->idr, control, 0, 0, GFP_KERNEL);
ret = idr_alloc(&channel->riids, intent,
ret = idr_alloc(&glink->rcids, channel, rcid, rcid + 1, GFP_ATOMIC);
id = idr_alloc(&vrp->endpoints, ept, id_min, id_max, GFP_KERNEL);
error = idr_alloc(&bfad_im_port_index, im_port, 0, 0, GFP_KERNEL);
ret = idr_alloc(&ch_index_idr, ch, 0, CH_MAX_DEVS + 1, GFP_NOWAIT);
ret = idr_alloc(&lpfc_hba_index, NULL, 0, 0, GFP_KERNEL);
id = idr_alloc(&iscsi_ep_idr, ep, 1, -1, GFP_NOIO);
error = idr_alloc(&sg_index_idr, sdp, 0, SG_MAX_DEVS, GFP_NOWAIT);
error = idr_alloc(&st_index_idr, tpnt, 0, ST_MAX_TAPES + 1, GFP_NOWAIT);
ret = idr_alloc(&apr->svcs_idr, svc, svc_id, svc_id + 1, GFP_ATOMIC);
id = idr_alloc(&spi_controller_idr, ctlr, start, end, GFP_KERNEL);
minor = idr_alloc(&tty_minors, gb_tty, 0, GB_NUM_MINORS, GFP_KERNEL);
ret = idr_alloc(&tiqn_idr, NULL, 0, 0, GFP_NOWAIT);
*id = idr_alloc(&supp->idr, req, 1, 0, GFP_KERNEL);
ret = idr_alloc(&ctxdata->qtee_objects_idr, object, 0, 0,
ret = idr_alloc(&ctxdata->reqs_idr, ureq, 0, 0, GFP_KERNEL);
id = idr_alloc(&teedev->idr, NULL, 1, 0, GFP_KERNEL);
ref->shm.id = idr_alloc(&ref->shm.ctx->teedev->idr, &ref->shm,
id = idr_alloc(&teedev->idr, NULL, 1, 0, GFP_KERNEL);
ret = idr_alloc(&tty_idr, cport, 0, MAX_TTY_RPMSG, GFP_KERNEL);
id = idr_alloc(&ports_idr, (void *)mps_port, id, MPS2_MAX_PORTS, GFP_KERNEL);
retval = idr_alloc(&uio_idr, idev, 0, UIO_MAX_DEVICES, GFP_KERNEL);
minor = idr_alloc(&acm_minors, acm, 0, ACM_TTY_MINORS, GFP_KERNEL);
busnum = idr_alloc(&usb_bus_idr, bus, 1, USB_MAXBUS, GFP_KERNEL);
port->minor = idr_alloc(&dbc_tty_minors, port, 0, 64, GFP_KERNEL);
minor = idr_alloc(&serial_minors, port, 0,
ret = idr_alloc(&vduse_idr, dev, 1, VDUSE_DEV_MAX, GFP_KERNEL);
err = idr_alloc(&sbi->devs->tree, dif, 0, 0, GFP_KERNEL);
ret = idr_alloc(&sbi->devs->tree, dif, 0, 0, GFP_KERNEL);
ret = idr_alloc(&nn->cb_ident_idr, clp, 1, 0, GFP_NOWAIT);
ret = idr_alloc(&nn->nn_status_idr, nsw, 0, 0, GFP_ATOMIC);
reg->id = idr_alloc(&cifs_swnreg_idr, reg, 1, 0, GFP_ATOMIC);
int idr_alloc(struct idr *, void *ptr, int start, int end, gfp_t);
DEFINE_CLASS(idr_alloc, struct __class_idr,
.id = idr_alloc(idr, ptr, start, end, gfp),
idx = idr_alloc(&ids->ipcs_idr, new, ipcid_to_idx(next_id),
ret = idr_alloc(idr, ptr, start, end, gfp_mask & ~__GFP_DIRECT_RECLAIM);
CLASS(idr_alloc, pmu_type)(&pmu_idr, NULL, max, 0, GFP_KERNEL);
nr = idr_alloc(&tmp->idr, NULL, tid,
ret = idr_alloc(&worker_pool_idr, pool, 0, WORK_OFFQ_POOL_NONE,
mod_id = idr_alloc(&cttype->mod_idr, cmod, 0, 0, GFP_KERNEL);
EXPORT_SYMBOL_GPL(idr_alloc);
id = idr_alloc(&shrinker_idr, shrinker, 0, 0, GFP_KERNEL);
tag = idr_alloc(&c->reqs, req, P9_NOTAG, P9_NOTAG + 1,
tag = idr_alloc(&c->reqs, req, 0, P9_NOTAG, GFP_NOWAIT);
handle = idr_alloc(&hdev->adv_monitors_idr, monitor, min, max,
return idr_alloc(&net->netns_ids, peer, min, max, GFP_ATOMIC);
id = idr_alloc(&local->ack_status_frames, ack_skb,
ret = idr_alloc(&sdata->u.nan.function_inst_ids,
id = idr_alloc(&local->ack_status_frames, ack_skb,
ret = idr_alloc(&s->conn_idr, con, 0, 0, GFP_ATOMIC);
ret = idr_alloc(&ac97_adapter_idr, ac97_ctrl, 0, 0, GFP_KERNEL);
id = idr_alloc(&apm->graph_idr, graph, graph_id, graph_id + 1, GFP_KERNEL);
assert(idr_alloc(&idr, NULL, 0, 0, GFP_KERNEL) == 0);
assert(idr_alloc(&idr, NULL, 0, 0, GFP_KERNEL) == 0);
assert(idr_alloc(&idr, NULL, 0, 0, GFP_KERNEL) == i);
assert(idr_alloc(&idr, NULL, 0, 0, GFP_KERNEL) == 5);
assert(idr_alloc(&idr, NULL, 0, 0, GFP_KERNEL) == 0);
assert(idr_alloc(&idr, NULL, 1, 0, GFP_KERNEL) == i);
assert(idr_alloc(&idr, item, i, i + 1, GFP_NOWAIT) == i);
assert(idr_alloc(&idr, item, indices[i], indices[i+1],
BUG_ON(idr_alloc(idr, &name[i], 0, 0, GFP_KERNEL) != i);
BUG_ON(idr_alloc(idr, &name[i], 0, 0, GFP_KERNEL) != i - 1);
BUG_ON(idr_alloc(idr, &name[i], 0, 0, GFP_KERNEL) != i - 2);
BUG_ON(idr_alloc(idr, &name[i], 0, 0, GFP_KERNEL) != i - 3);
BUG_ON(idr_alloc(idr, &name[i], 0, 0, GFP_KERNEL) != 0);
BUG_ON(idr_alloc(idr, &name[i + 1], 0, 0, GFP_KERNEL) != 1);
BUG_ON(idr_alloc(idr, NULL, 0, 0, GFP_KERNEL) != 0);
BUG_ON(idr_alloc(idr, &name[i], 0, 0, GFP_KERNEL) != 0);
BUG_ON(idr_alloc(idr, NULL, 0, 0, GFP_KERNEL) != 1);
idr_alloc(&find_idr, xa_mk_value(id), id, id + 1, GFP_KERNEL);
BUG_ON(idr_alloc(&find_idr, xa_mk_value(anchor_id), anchor_id,
assert(idr_alloc(&idr, item, 0, 20000, GFP_KERNEL) == i);
assert(idr_alloc(&idr, DUMMY_PTR, 5, 30, GFP_KERNEL) < 0);
assert(idr_alloc(&idr, DUMMY_PTR, 0, 0, GFP_KERNEL) == 0);
assert(idr_alloc(&idr, DUMMY_PTR, 0, 0, GFP_KERNEL) == i);
assert(idr_alloc(&idr, item, i, i + 10, GFP_KERNEL) == i);
assert(idr_alloc(&idr, DUMMY_PTR, i - 2, i, GFP_KERNEL) == -ENOSPC);
assert(idr_alloc(&idr, DUMMY_PTR, i - 2, i + 10, GFP_KERNEL) == -ENOSPC);
assert(idr_alloc(&idr, item, 1, 20000, GFP_KERNEL) == i);
id = idr_alloc(&idr, idr_alloc2_test, 0, 1, GFP_KERNEL);
id = idr_alloc(&idr, idr_alloc2_test, 1, 2, GFP_KERNEL);
id = idr_alloc(&idr, idr_alloc2_test, 0, 1, GFP_KERNEL);
id = idr_alloc(&idr, idr_alloc2_test, 0, 2, GFP_KERNEL);
idr_alloc(&idr, (void *)-1, 10, 11, GFP_KERNEL);