bitmap_zalloc
vmid_map = bitmap_zalloc(NUM_USER_VMIDS, GFP_KERNEL);
asid_map = bitmap_zalloc(NUM_USER_ASIDS, GFP_KERNEL);
pinned_asid_map = bitmap_zalloc(NUM_USER_ASIDS, GFP_KERNEL);
info->map = bitmap_zalloc(NUM_CTXT_ASIDS(info), GFP_KERNEL);
mm_ctx->bd_emupage_allocmap = bitmap_zalloc(emupage_frame_count,
mmid_map = bitmap_zalloc(num_mmids, GFP_KERNEL);
kvmppc_uvmem_bitmap = bitmap_zalloc(pfn_last - pfn_first, GFP_KERNEL);
mpic->protected = bitmap_zalloc(intvec_top + 1, GFP_KERNEL);
xibm->bitmap = bitmap_zalloc(xibm->count, GFP_KERNEL);
context_asid_map = bitmap_zalloc(num_asids, GFP_KERNEL);
sq_bitmap = bitmap_zalloc(nr_pages, GFP_KERNEL);
sev_asid_bitmap = bitmap_zalloc(nr_asids, GFP_KERNEL);
sev_reclaim_asid_bitmap = bitmap_zalloc(nr_asids, GFP_KERNEL);
hdev->asid_bitmap = bitmap_zalloc(hdev->asic_prop.max_asid, GFP_KERNEL);
valid = bitmap_zalloc(numprops, GFP_KERNEL);
new = bitmap_zalloc(nbits, GFP_KERNEL);
new = bitmap_zalloc(nbits, GFP_KERNEL);
dmap->map = bitmap_zalloc(NBITS_MIN, GFP_KERNEL);
he_dev->rbpl_table = bitmap_zalloc(RBPL_TABLE_SIZE, GFP_KERNEL);
cache->valid = bitmap_zalloc(cache_size, map->alloc_flags);
req_q->req_bitmap = bitmap_zalloc(req_q->size, GFP_KERNEL);
bmap = bitmap_zalloc(size, GFP_KERNEL);
wq->opcap_bmap = bitmap_zalloc(IDXD_MAX_OPCAP_BITS, GFP_KERNEL);
opmask = bitmap_zalloc(IDXD_MAX_OPCAP_BITS, GFP_KERNEL);
shdma_slave_used = bitmap_zalloc(slave_num, GFP_KERNEL);
unsigned long *values __free(bitmap) = bitmap_zalloc(descs->ndescs,
unsigned long *values __free(bitmap) = bitmap_zalloc(descs->ndescs,
cdev->watched_lines = bitmap_zalloc(gdev->ngpio, GFP_KERNEL);
mes->doorbell_bitmap = bitmap_zalloc(PAGE_SIZE / sizeof(u32), GFP_KERNEL);
kfd->gtt_sa_bitmap = bitmap_zalloc(kfd->gtt_sa_num_of_chunks,
qpd->doorbell_bitmap = bitmap_zalloc(KFD_MAX_NUM_OF_QUEUES_PER_PROCESS,
kfd->doorbell_bitmap = bitmap_zalloc(size / sizeof(u32), GFP_KERNEL);
pqm->queue_slot_bitmap = bitmap_zalloc(KFD_MAX_NUM_OF_QUEUES_PER_PROCESS,
obj->bit_17 = bitmap_zalloc(obj->base.size >> PAGE_SHIFT,
obj->bit_17 = bitmap_zalloc(page_count, GFP_KERNEL);
bitmap_zalloc(NUMBER_MULTI_LRC_GUC_ID(guc), GFP_KERNEL);
jmp = bitmap_zalloc(DIV_ROUND_UP(batch_length, sizeof(u32)),
valid = bitmap_zalloc(FW_RANGE, GFP_KERNEL);
dbm->bitmap = bitmap_zalloc(count, GFP_KERNEL);
idm->bitmap = bitmap_zalloc(limit, GFP_KERNEL);
chlist->allocated_channels = bitmap_zalloc(num_channels, GFP_KERNEL);
bitmap = bitmap_zalloc(size, GFP_KERNEL);
alias_pool->use_mask = bitmap_zalloc(num_aliases, GFP_KERNEL);
cb_buff->buffer.scan_mask = bitmap_zalloc(iio_get_masklength(cb_buff->indio_dev),
buffer->scan_mask = bitmap_zalloc(masklength,
compound_mask = bitmap_zalloc(masklength, GFP_KERNEL);
alloc->table = bitmap_zalloc(num, GFP_KERNEL);
bitmap_zalloc(dev->res_cb[i].max_cap, GFP_KERNEL);
pq->req_in_use = bitmap_zalloc(hfi1_sdma_comp_ring_size, GFP_KERNEL);
idx_que->bitmap = bitmap_zalloc(srq->wqe_cnt, GFP_KERNEL);
rf->allocated_ws_nodes = bitmap_zalloc(IRDMA_MAX_WS_NODES,
pchunk->bitmapbuf = bitmap_zalloc(sizeofbitmap, GFP_KERNEL);
var_table->bitmap = bitmap_zalloc(var_table->num_var_hw_entries,
alloc->table = bitmap_zalloc(num, GFP_KERNEL);
buddy->bits[i] = bitmap_zalloc(1 << (buddy->max_order - i),
dev->pd_mgr->pd_dpp_bitmap = bitmap_zalloc(rsp->pd_count,
dev->pd_mgr->pd_norm_bitmap = bitmap_zalloc(rsp->pd_count,
tbl->table = bitmap_zalloc(num, GFP_KERNEL);
clt->permits_map = bitmap_zalloc(clt->queue_depth, GFP_KERNEL);
mask = bitmap_zalloc(cnt, GFP_KERNEL);
bits = bitmap_zalloc(n_events, GFP_KERNEL);
iommu->copied_tables = bitmap_zalloc(BIT_ULL(16), GFP_KERNEL);
bitmap = bitmap_zalloc(INTR_REMAP_TABLE_ENTRIES, GFP_KERNEL);
unsigned long *msi_map __free(kfree) = bitmap_zalloc(priv->num_spis, GFP_KERNEL);
mip->bitmap = bitmap_zalloc(mip->num_msis, GFP_KERNEL);
v2m->bm = bitmap_zalloc(v2m->nr_spis, GFP_KERNEL);
bitmap = bitmap_zalloc(nr_irqs, GFP_ATOMIC);
mbi_ranges[n].bm = bitmap_zalloc(mbi_ranges[n].nr_spis, GFP_KERNEL);
its_dev->event_map = (unsigned long *)bitmap_zalloc(its_dev->num_events, GFP_KERNEL);
priv->msi_map = bitmap_zalloc(priv->num_irqs, GFP_KERNEL);
odmis_bm = bitmap_zalloc(odmis_count * NODMIS_PER_FRAME, GFP_KERNEL);
lpriv->dirty_bitmap = bitmap_zalloc(global->nr_ids + 1, GFP_KERNEL);
priv->prio_save = bitmap_zalloc(nr_irqs, GFP_KERNEL);
need_reset = bitmap_zalloc(nr_zones, GFP_NOIO);
q->bufs_bitmap = bitmap_zalloc(q->max_num_buffers, GFP_KERNEL);
ent_enum->bmap = bitmap_zalloc(idx_max, GFP_KERNEL);
mfc_dev->mem_bitmap = bitmap_zalloc(mem_size >> PAGE_SHIFT, GFP_KERNEL);
msb->used_blocks_bitmap = bitmap_zalloc(msb->block_count, GFP_KERNEL);
msb->erased_blocks_bitmap = bitmap_zalloc(msb->block_count, GFP_KERNEL);
xpnet_broadcast_partitions = bitmap_zalloc(xp_max_npartitions,
chip_map = bitmap_zalloc(max_chips, GFP_KERNEL);
nand->bbt.cache = bitmap_zalloc(nblocks * bits_per_block, GFP_KERNEL);
vol->checkmap = bitmap_zalloc(leb_count, GFP_KERNEL);
ret = bitmap_zalloc(ubi->peb_count, GFP_NOFS);
bp->ntp_fltr_bmap = bitmap_zalloc(bp->max_fltr, GFP_KERNEL);
id_tbl->table = bitmap_zalloc(size, GFP_KERNEL);
t = bitmap_zalloc(adap->sge.egr_sz, GFP_KERNEL);
eth_filter->port[i].bmap = bitmap_zalloc(nentries, GFP_KERNEL);
adap->sge.starving_fl = bitmap_zalloc(adap->sge.egr_sz, GFP_KERNEL);
adap->sge.txq_maperr = bitmap_zalloc(adap->sge.egr_sz, GFP_KERNEL);
adap->sge.blocked_fl = bitmap_zalloc(adap->sge.egr_sz, GFP_KERNEL);
adap->msix_bmap.msix_bmap = bitmap_zalloc(num_vec, GFP_KERNEL);
epsfp.psfp_sfi_bitmap = bitmap_zalloc(priv->psfp_cap.max_psfp_filter,
fdev->irq_map = bitmap_zalloc(fdev->num_irqs, GFP_KERNEL);
priv->xsk_pools = bitmap_zalloc(priv->rx_cfg.max_queues, GFP_KERNEL);
db_area->db_bitmap_array = bitmap_zalloc(db_max_areas, GFP_KERNEL);
vsi->af_xdp_zc_qps = bitmap_zalloc(pf->num_lan_qps, GFP_KERNEL);
pf->virt_irq_tracker.bm = bitmap_zalloc(num_entries, GFP_KERNEL);
pf->avail_txqs = bitmap_zalloc(pf->max_pf_txqs, GFP_KERNEL);
pf->avail_rxqs = bitmap_zalloc(pf->max_pf_rxqs, GFP_KERNEL);
pf->txtime_txqs = bitmap_zalloc(pf->max_pf_txqs, GFP_KERNEL);
in_use = bitmap_zalloc(tx->len, GFP_KERNEL);
stale = bitmap_zalloc(tx->len, GFP_KERNEL);
adapter->af_xdp_zc_qps = bitmap_zalloc(IXGBE_MAX_XDP_QS, GFP_KERNEL);
pf_bmap = bitmap_zalloc(num, GFP_KERNEL);
mcam->bmap = bitmap_zalloc(mcam->bmap_entries, GFP_KERNEL);
mcam->bmap_reverse = bitmap_zalloc(mcam->bmap_entries, GFP_KERNEL);
pf->af_xdp_zc_qidx = bitmap_zalloc(qcount, GFP_KERNEL);
vf->af_xdp_zc_qidx = bitmap_zalloc(qcount, GFP_KERNEL);
bitmap->table = bitmap_zalloc(bitmap->max, GFP_KERNEL);
bulk->need_sync = bitmap_zalloc(num_deks, GFP_KERNEL);
bulk->in_use = bitmap_zalloc(num_deks, GFP_KERNEL);
bitmap_zalloc(steering_icm_blocks, GFP_KERNEL);
bitmap_zalloc(header_modify_icm_blocks, GFP_KERNEL);
bitmap_zalloc(header_sw_encap_icm_blocks, GFP_KERNEL);
bitmap_zalloc(header_modify_pattern_icm_blocks, GFP_KERNEL);
mpfs->bitmap = bitmap_zalloc(l2table_size, GFP_KERNEL);
buddy->bitmap[i] = bitmap_zalloc(s, GFP_KERNEL);
bitmap = bitmap_zalloc(1 << log_range, GFP_KERNEL);
buddy->bitmap[i] = bitmap_zalloc(size, GFP_KERNEL);
chosen_blocks_bm = bitmap_zalloc(mlxsw_afk->blocks_count, GFP_KERNEL);
ports_bm->bitmap = bitmap_zalloc(nbits, GFP_KERNEL);
region_12kb->used_lkey_id = bitmap_zalloc(max_lkey_id, GFP_KERNEL);
pool->usage = bitmap_zalloc(pool->pool_size, GFP_KERNEL);
fid_family->fids_bitmap = bitmap_zalloc(nr_fids, GFP_KERNEL);
activity = bitmap_zalloc(nh_grp->nhgi->count, GFP_KERNEL);
r->map = bitmap_zalloc(res_avail, GFP_KERNEL);
abm->threshold_undef = bitmap_zalloc(abm->num_thresholds, GFP_KERNEL);
p_map->cid_map = bitmap_zalloc(cid_count, GFP_KERNEL);
bmap->bitmap = bitmap_zalloc(max_count, GFP_KERNEL);
table->used_bitmap = bitmap_zalloc(table->size, GFP_KERNEL);
table->used_bitmap = bitmap_zalloc(table->size, GFP_KERNEL);
priv->af_xdp_zc_qps = bitmap_zalloc(MTL_MAX_TX_QUEUES, GFP_KERNEL);
net_device->send_section_map = bitmap_zalloc(net_device->send_section_cnt,
ipa->available = bitmap_zalloc(IPA_ENDPOINT_MAX, GFP_KERNEL);
ipa->available = bitmap_zalloc(limit, GFP_KERNEL);
ipa->defined = bitmap_zalloc(ipa->endpoint_count, GFP_KERNEL);
ipa->set_up = bitmap_zalloc(ipa->endpoint_count, GFP_KERNEL);
ipa->enabled = bitmap_zalloc(ipa->endpoint_count, GFP_KERNEL);
interrupt->suspend_enabled = bitmap_zalloc(ipa->endpoint_count,
activity = bitmap_zalloc(nexthop->occ, GFP_KERNEL);
priv->used_rxkeys = bitmap_zalloc(priv->rx_keycache_size,
bat_req->bat_bitmap = bitmap_zalloc(bat_req->bat_size_cnt, GFP_KERNEL);
idmap = bitmap_zalloc(max_ports, GFP_KERNEL);
dev->dma_alias_mask = bitmap_zalloc(MAX_NR_DEVFNS, GFP_KERNEL);
bitmap_zalloc(zynqmp_desc.npins, GFP_KERNEL);
sli4->ext[i].use_map = bitmap_zalloc(total, GFP_KERNEL);
h->cmd_pool_bits = bitmap_zalloc(h->nr_cmds, GFP_KERNEL);
vport->vmid_priority_range = bitmap_zalloc
removepend_bitmap = bitmap_zalloc(mrioc->facts.max_devhandle,
mrioc->removepend_bitmap = bitmap_zalloc(mrioc->dev_handle_bitmap_bits,
mrioc->devrem_bitmap = bitmap_zalloc(MPI3MR_NUM_DEVRMCMD, GFP_KERNEL);
mrioc->evtack_cmds_bitmap = bitmap_zalloc(MPI3MR_NUM_EVTACKCMD,
mrioc->chain_bitmap = bitmap_zalloc(num_chains, GFP_KERNEL);
mvi->rsvd_tags = bitmap_zalloc(MVS_RSVD_SLOTS, GFP_KERNEL);
pm8001_ha->rsvd_tags = bitmap_zalloc(PM8001_RESERVE_SLOT, GFP_KERNEL);
pmc->wake_type_level_map = bitmap_zalloc(pmc->soc->max_wake_events, GFP_KERNEL);
pmc->wake_type_dual_edge_map = bitmap_zalloc(pmc->soc->max_wake_events, GFP_KERNEL);
pmc->wake_sw_status_map = bitmap_zalloc(pmc->soc->max_wake_events, GFP_KERNEL);
pmc->wake_cntrl_level_map = bitmap_zalloc(pmc->soc->max_wake_events, GFP_KERNEL);
udev->data_bitmap = bitmap_zalloc(udev->max_blocks, GFP_KERNEL);
optee->notif.bitmap = bitmap_zalloc(max_key, GFP_KERNEL);
gnttab_free_bitmap = bitmap_zalloc(max_nr_grefs, GFP_KERNEL);
acpi_id_present = bitmap_zalloc(nr_acpi_bits, GFP_KERNEL);
acpi_id_cst_present = bitmap_zalloc(nr_acpi_bits, GFP_KERNEL);
acpi_ids_done = bitmap_zalloc(nr_acpi_bits, GFP_KERNEL);
rbio->error_bitmap = bitmap_zalloc(num_sectors, GFP_NOFS);
rbio->stripe_uptodate_bitmap = bitmap_zalloc(num_sectors, GFP_NOFS);
rbio->csum_bitmap = bitmap_zalloc(rbio->nr_data * rbio->stripe_nsectors,
active = bitmap_zalloc(map->num_stripes, GFP_NOFS);
zone_info->seq_zones = bitmap_zalloc(zone_info->nr_zones, GFP_KERNEL);
zone_info->empty_zones = bitmap_zalloc(zone_info->nr_zones, GFP_KERNEL);
zone_info->active_zones = bitmap_zalloc(zone_info->nr_zones, GFP_KERNEL);
zone_info->seq_zones = bitmap_zalloc(zone_info->nr_zones, GFP_KERNEL);
zone_info->empty_zones = bitmap_zalloc(zone_info->nr_zones, GFP_KERNEL);
zone_info->active_zones = bitmap_zalloc(zone_info->nr_zones, GFP_KERNEL);
bufmap->buffer_index_array = bitmap_zalloc(bufmap->desc_count, GFP_KERNEL);
d->rmid_busy_llc = bitmap_zalloc(idx_limit, GFP_KERNEL);
unsigned long *bitmap_zalloc(unsigned int nbits, gfp_t flags);
fq->flows_bitmap = bitmap_zalloc(fq->flows_cnt, GFP_KERNEL);
table->bitmap = bitmap_zalloc(nr_files, GFP_KERNEL_ACCOUNT);
dma_mem->bitmap = bitmap_zalloc(pages, GFP_KERNEL);
unsigned long *pending __free(bitmap) = bitmap_zalloc(num_irqs, GFP_KERNEL);
tmp_bitmap = bitmap_zalloc(bitmap_len, GFP_KERNEL);
EXPORT_SYMBOL(bitmap_zalloc);
graph->edges = bitmap_zalloc(nodes_count * nodes_count, GFP_KERNEL);
record->component_bitmap = bitmap_zalloc(record->component_bitmap_len,
unsigned long *bits = bitmap_zalloc(nbits, GFP_KERNEL);
cmr->bitmap = bitmap_zalloc(cma_bitmap_maxno(cma, cmr),
filter = bitmap_zalloc(BIT(BLOOM_FILTER_SHIFT),
inuse = bitmap_zalloc(BR_MAX_PORTS, GFP_KERNEL);
changed = bitmap_zalloc(BR_MAX_PORTS, GFP_KERNEL);
inuse = bitmap_zalloc(max_netdevices, GFP_ATOMIC);
mask = bitmap_zalloc(nr_ids, GFP_NOWAIT);
mask = bitmap_zalloc(dev->num_rx_queues, GFP_KERNEL);
rxq_bitmap = bitmap_zalloc(netdev->real_num_rx_queues, GFP_KERNEL);
aid_map = bitmap_zalloc(IEEE80211_MAX_AID + 1, GFP_KERNEL);
map->members = bitmap_zalloc(elements, GFP_KERNEL | __GFP_NOWARN);
map->members = bitmap_zalloc(elements, GFP_KERNEL | __GFP_NOWARN);
map->members = bitmap_zalloc(map->elements, GFP_KERNEL | __GFP_NOWARN);
table = bitmap_zalloc(IP_VS_MH_TAB_SIZE, GFP_KERNEL);
link->wr_tx_mask = bitmap_zalloc(link->max_send_wr, GFP_KERNEL);
unsigned long *to_test = bitmap_zalloc(num_bits);
c2c_he->cpuset = bitmap_zalloc(c2c.cpus_cnt);
c2c_he->nodeset = bitmap_zalloc(c2c.nodes_cnt);
set = bitmap_zalloc(c2c.cpus_cnt);
mask->bits = bitmap_zalloc(mask->nbits);
bm = bitmap_zalloc(nbits);
bm = bitmap_zalloc(nbits);
a->orig_cpus = bitmap_zalloc(cpu_set_size * 8);
a->sched_cpus = bitmap_zalloc(cpu_set_size * 8);
unsigned long *cpuset = bitmap_zalloc(cpu_set_size * 8);
n->set = bitmap_zalloc(size);
set = bitmap_zalloc(size);
node_mask = bitmap_zalloc(node_index + 1);
map->affinity_mask.bits = bitmap_zalloc(map->affinity_mask.nbits);
vcpu_done_map = bitmap_zalloc(test_args.nr_vcpus);
bmap = bitmap_zalloc(size / getpagesize());
bmap[0] = bitmap_zalloc(host_num_pages);
bmap[1] = bitmap_zalloc(host_num_pages);
bitmaps[i] = bitmap_zalloc(pages_per_slot);
bmap = bitmap_zalloc(TEST_MEM_PAGES);