kvmalloc_objs
sops = kvmalloc_objs(*sops, nsops);
*relocation_hashtable = kvmalloc_objs(**relocation_hashtable,
cpump->save = kvmalloc_objs(struct pai_userdata,
e2 = kvmalloc_objs(*e2, cpuid->nent, GFP_KERNEL_ACCOUNT);
addrs = kvmalloc_objs(*addrs, prog->len + 1);
kvmalloc_objs(profile->slot_hashtable[0], slot_hashtable_size);
ubuf->pages = kvmalloc_objs(*ubuf->pages, ubuf->nr_pages);
kvmalloc_objs(*jobs, args->job_count);
userptr->pages = kvmalloc_objs(struct page *, npages);
pages = kvmalloc_objs(*pages, nr_pages);
jobs = kvmalloc_objs(*jobs, args->job_count);
rjob->tasks = kvmalloc_objs(*rjob->tasks, job->task_count);
dev->zones = kvmalloc_objs(struct nullb_zone, dev->nr_zones,
fc->chains = kvmalloc_objs(*fc->chains, count, GFP_KERNEL | __GFP_ZERO);
fc->fences = kvmalloc_objs(*fc->fences, count, GFP_KERNEL | __GFP_ZERO);
pages = kvmalloc_objs(*pages, ubuf->pagecount);
ubuf->folios = kvmalloc_objs(*ubuf->folios, pgcnt);
ubuf->pinned_folios = kvmalloc_objs(*ubuf->pinned_folios, pgcnt);
p->chunks = kvmalloc_objs(struct amdgpu_cs_chunk, p->nchunks);
kvmalloc_objs(struct amdgpu_userq_fence_driver *, count,
bo_buckets = kvmalloc_objs(*bo_buckets, args->num_bos);
bo_privs = kvmalloc_objs(*bo_privs, args->num_bos);
pages = kvmalloc_objs(struct page *, npages);
objs = kvmalloc_objs(struct drm_gem_object *, count,
kvmalloc_objs(*svm_pages->dma_addr, npages);
pvec = kvmalloc_objs(struct page *, npages);
etnaviv_obj->pages = kvmalloc_objs(struct page *, npages);
bos = kvmalloc_objs(*bos, args->nr_bos);
relocs = kvmalloc_objs(*relocs, args->nr_relocs);
pmrs = kvmalloc_objs(*pmrs, args->nr_pmrs);
g2d_userptr->pages = kvmalloc_objs(*g2d_userptr->pages, npages);
pages = kvmalloc_objs(*pages, n_pages);
pvec = kvmalloc_objs(struct page *, num_pages);
timelines = kvmalloc_objs(*timelines, NUM_TIMELINES * I915_NUM_ENGINES);
timelines = kvmalloc_objs(*timelines, NUM_TIMELINES * I915_NUM_ENGINES);
pages = kvmalloc_objs(*pages, n_pages);
job_data = kvmalloc_objs(*job_data, args->jobs.count,
pages = kvmalloc_objs(*pages, bo->base.base.size >> PAGE_SHIFT,
msm_obj->pages = kvmalloc_objs(struct page *, npages);
p->pages = kvmalloc_objs(*p->pages, p->count);
in = kvmalloc_objs(*in, 1024);
if (!(mem->mem = kvmalloc_objs(*mem->mem, size)))
if (!(mem->dma = kvmalloc_objs(*mem->dma, size)))
job->mappings = kvmalloc_objs(struct panfrost_gem_mapping *,
ops = kvmalloc_objs(*ops, args->op_count);
bo->sgts = kvmalloc_objs(struct sg_table,
pages = kvmalloc_objs(struct page *, nr_pages,
ctx->jobs = kvmalloc_objs(*ctx->jobs, job_count,
list = kvmalloc_objs(struct radeon_bo_list, vm->max_pde_used + 2);
kvmalloc_objs(struct v3d_submit_outsync, count);
query_info->queries = kvmalloc_objs(struct v3d_timestamp_query,
query_info->queries = kvmalloc_objs(struct v3d_timestamp_query,
query_info->queries = kvmalloc_objs(struct v3d_timestamp_query,
kvmalloc_objs(struct v3d_performance_query, reset.count);
kvmalloc_objs(struct v3d_performance_query, copy.count);
*ents = kvmalloc_objs(struct virtio_gpu_mem_entry, *nents);
*ents = kvmalloc_objs(struct virtio_gpu_mem_entry, sgt->nents);
src_pages = kvmalloc_objs(struct page *, src->ttm->num_pages);
dst_pages = kvmalloc_objs(struct page *, dst->ttm->num_pages);
*bind_ops = kvmalloc_objs(struct drm_xe_vm_bind_op,
xen_obj->pages = kvmalloc_objs(struct page *, xen_obj->num_pages);
qp->sq.w_list = kvmalloc_objs(*qp->sq.w_list, qp->sq.wqe_cnt);
ring = kvmalloc_objs(ring[0], ring_size);
pci_seg->alias_table = kvmalloc_objs(*pci_seg->alias_table,
clone->ht = kvmalloc_objs(struct hash_table_bucket, sz);
pl = kvmalloc_objs(struct page_list, n_pages + 1,
sl = kvmalloc_objs(struct scatterlist *, ic->journal_sections,
s = kvmalloc_objs(struct scatterlist, n_pages);
sg = kvmalloc_objs(struct scatterlist,
ic->sk_requests = kvmalloc_objs(struct skcipher_request *,
ic->bbs = kvmalloc_objs(struct bitmap_block_status,
et->table = kvmalloc_objs(struct dm_hlist_head, size);
helpers = kvmalloc_objs(helper[0], cs->count);
helpers = kvmalloc_objs(helper[0], cs->count);
indir->table = kvmalloc_objs(*indir->table, max_table_size);
priv->stats = kvmalloc_objs(struct nfp_fl_stats, stats_size);
temp_ring = kvmalloc_objs(struct wx_ring, i);
temp_ring = kvmalloc_objs(struct wx_ring, i);
pcpu_sum = kvmalloc_objs(struct netvsc_ethtool_pcpu_stats, nr_cpu_ids);
pages = kvmalloc_objs(*pages, count);
pages = kvmalloc_objs(struct page *, n_pages);
pages = kvmalloc_objs(struct page *, n_pages);
new_ring = kvmalloc_objs(*new_ring, new_size);
add->grants = kvmalloc_objs(add->grants[0], count);
add->map_ops = kvmalloc_objs(add->map_ops[0], count);
add->unmap_ops = kvmalloc_objs(add->unmap_ops[0], count);
add->kmap_ops = kvmalloc_objs(add->kmap_ops[0], count);
add->kunmap_ops = kvmalloc_objs(add->kunmap_ops[0], count);
bv = kvmalloc_objs(*bv, npages, GFP_KERNEL | __GFP_ZERO);
cprm->vma_meta = kvmalloc_objs(*cprm->vma_meta, cprm->vma_count);
sbi->vol_amap = kvmalloc_objs(struct buffer_head *, sbi->map_sectors);
sbi->s_group_desc = kvmalloc_objs(struct buffer_head *, db_count);
oi->of_binfo = kvmalloc_objs(struct ext4_orphan_block, oi->of_blocks);
kvmalloc_objs(struct buffer_head *, db_count));
data = kvmalloc_objs(struct file *, nr, GFP_KERNEL_ACCOUNT);
bufs = kvmalloc_objs(struct pipe_buffer, pipe->max_usage);
bufs = kvmalloc_objs(struct pipe_buffer, count);
kvmalloc_objs(struct list_head, hash_size);
table->hbs = kvmalloc_objs(table->hbs[0], hash_buckets,
pages = kvmalloc_objs(*pages, mr->nr_pages, gfp);
pages = kvmalloc_objs(struct page *, nr_pages, GFP_KERNEL_ACCOUNT);
data->nodes = kvmalloc_objs(struct io_rsrc_node *, nr,
new_array = kvmalloc_objs(struct page *, nr_folios);
area->nia.niovs = kvmalloc_objs(area->nia.niovs[0], nr_iovs,
area->user_refs = kvmalloc_objs(area->user_refs[0], nr_iovs,
sops = kvmalloc_objs(*sops, nsops);
array = kvmalloc_objs(pid_t, length);
pages = kvmalloc_objs(struct page *, count);
pages = kvmalloc_objs(*pages, total_pages);
tlocks = kvmalloc_objs(spinlock_t, size, gfp);
*res = kvmalloc_objs(struct page *, count);
folios = kvmalloc_objs(*state->folios, max_nr);
folios = kvmalloc_objs(*folios, max_folios);
binding->tx_vec = kvmalloc_objs(struct net_iov *,
owner->area.niovs = kvmalloc_objs(*owner->area.niovs,
tokens = kvmalloc_objs(*tokens, num_tokens);
new_batch = kvmalloc_objs(*new_batch, new_batch_sz,
ip_vs_conn_tab = kvmalloc_objs(*ip_vs_conn_tab, ip_vs_conn_tab_size);
expr_info = kvmalloc_objs(struct nft_expr_info,
dst->mt = kvmalloc_objs(*src->mt, src->rules_alloc,
new_mt = kvmalloc_objs(*new_mt, rules_alloc, GFP_KERNEL_ACCOUNT);
ti->buckets = kvmalloc_objs(struct hlist_head, new_size);
tc_u_common_hash = kvmalloc_objs(struct hlist_head, U32_HASH_SIZE);
h = kvmalloc_objs(struct hlist_head, n);
net->unx.table.locks = kvmalloc_objs(spinlock_t, UNIX_HASH_SIZE);
net->unx.table.buckets = kvmalloc_objs(struct hlist_head,
fpl->edges = kvmalloc_objs(*fpl->edges, fpl->count_unix,
cellptr = kvmalloc_objs(struct snd_seq_event_cell, pool->size);