Symbol: kvmalloc_objs
arch/arm/kernel/sys_oabi-compat.c
352
sops = kvmalloc_objs(*sops, nsops);
arch/riscv/kernel/module.c
756
*relocation_hashtable = kvmalloc_objs(**relocation_hashtable,
arch/s390/kernel/perf_pai.c
284
cpump->save = kvmalloc_objs(struct pai_userdata,
arch/x86/kvm/cpuid.c
605
e2 = kvmalloc_objs(*e2, cpuid->nent, GFP_KERNEL_ACCOUNT);
arch/x86/net/bpf_jit_comp.c
3788
addrs = kvmalloc_objs(*addrs, prog->len + 1);
block/blk-crypto-profile.c
123
kvmalloc_objs(profile->slot_hashtable[0], slot_hashtable_size);
drivers/accel/amdxdna/amdxdna_ubuf.c
196
ubuf->pages = kvmalloc_objs(*ubuf->pages, ubuf->nr_pages);
drivers/accel/ethosu/ethosu_job.c
489
kvmalloc_objs(*jobs, args->job_count);
drivers/accel/habanalabs/common/memory.c
2326
userptr->pages = kvmalloc_objs(struct page *, npages);
drivers/accel/ivpu/ivpu_gem_userptr.c
80
pages = kvmalloc_objs(*pages, nr_pages);
drivers/accel/rocket/rocket_job.c
612
jobs = kvmalloc_objs(*jobs, args->job_count);
drivers/accel/rocket/rocket_job.c
74
rjob->tasks = kvmalloc_objs(*rjob->tasks, job->task_count);
drivers/block/null_blk/zoned.c
94
dev->zones = kvmalloc_objs(struct nullb_zone, dev->nr_zones,
drivers/dma-buf/st-dma-fence-chain.c
119
fc->chains = kvmalloc_objs(*fc->chains, count, GFP_KERNEL | __GFP_ZERO);
drivers/dma-buf/st-dma-fence-chain.c
123
fc->fences = kvmalloc_objs(*fc->fences, count, GFP_KERNEL | __GFP_ZERO);
drivers/dma-buf/udmabuf.c
118
pages = kvmalloc_objs(*pages, ubuf->pagecount);
drivers/dma-buf/udmabuf.c
210
ubuf->folios = kvmalloc_objs(*ubuf->folios, pgcnt);
drivers/dma-buf/udmabuf.c
218
ubuf->pinned_folios = kvmalloc_objs(*ubuf->pinned_folios, pgcnt);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
195
p->chunks = kvmalloc_objs(struct amdgpu_cs_chunk, p->nchunks);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
271
kvmalloc_objs(struct amdgpu_userq_fence_driver *, count,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
2470
bo_buckets = kvmalloc_objs(*bo_buckets, args->num_bos);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
2488
bo_privs = kvmalloc_objs(*bo_privs, args->num_bos);
drivers/gpu/drm/drm_gem.c
687
pages = kvmalloc_objs(struct page *, npages);
drivers/gpu/drm/drm_gem.c
835
objs = kvmalloc_objs(struct drm_gem_object *, count,
drivers/gpu/drm/drm_gpusvm.c
1474
kvmalloc_objs(*svm_pages->dma_addr, npages);
drivers/gpu/drm/etnaviv/etnaviv_gem.c
678
pvec = kvmalloc_objs(struct page *, npages);
drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c
129
etnaviv_obj->pages = kvmalloc_objs(struct page *, npages);
drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
470
bos = kvmalloc_objs(*bos, args->nr_bos);
drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
471
relocs = kvmalloc_objs(*relocs, args->nr_relocs);
drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
472
pmrs = kvmalloc_objs(*pmrs, args->nr_pmrs);
drivers/gpu/drm/exynos/exynos_drm_g2d.c
473
g2d_userptr->pages = kvmalloc_objs(*g2d_userptr->pages, npages);
drivers/gpu/drm/i915/gem/i915_gem_pages.c
317
pages = kvmalloc_objs(*pages, n_pages);
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
261
pvec = kvmalloc_objs(struct page *, num_pages);
drivers/gpu/drm/i915/gt/selftest_timeline.c
539
timelines = kvmalloc_objs(*timelines, NUM_TIMELINES * I915_NUM_ENGINES);
drivers/gpu/drm/i915/gt/selftest_timeline.c
612
timelines = kvmalloc_objs(*timelines, NUM_TIMELINES * I915_NUM_ENGINES);
drivers/gpu/drm/i915/gt/shmem_utils.c
66
pages = kvmalloc_objs(*pages, n_pages);
drivers/gpu/drm/imagination/pvr_job.c
721
job_data = kvmalloc_objs(*job_data, args->jobs.count,
drivers/gpu/drm/lima/lima_gem.c
42
pages = kvmalloc_objs(*pages, bo->base.base.size >> PAGE_SHIFT,
drivers/gpu/drm/msm/msm_gem.c
1304
msm_obj->pages = kvmalloc_objs(struct page *, npages);
drivers/gpu/drm/msm/msm_iommu.c
335
p->pages = kvmalloc_objs(*p->pages, p->count);
drivers/gpu/drm/nouveau/dispnv50/lut.c
41
in = kvmalloc_objs(*in, 1024);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
194
if (!(mem->mem = kvmalloc_objs(*mem->mem, size)))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
196
if (!(mem->dma = kvmalloc_objs(*mem->dma, size)))
drivers/gpu/drm/panfrost/panfrost_drv.c
210
job->mappings = kvmalloc_objs(struct panfrost_gem_mapping *,
drivers/gpu/drm/panfrost/panfrost_drv.c
600
ops = kvmalloc_objs(*ops, args->op_count);
drivers/gpu/drm/panfrost/panfrost_mmu.c
623
bo->sgts = kvmalloc_objs(struct sg_table,
drivers/gpu/drm/panfrost/panfrost_mmu.c
631
pages = kvmalloc_objs(struct page *, nr_pages,
drivers/gpu/drm/panthor/panthor_drv.c
726
ctx->jobs = kvmalloc_objs(*ctx->jobs, job_count,
drivers/gpu/drm/radeon/radeon_vm.c
136
list = kvmalloc_objs(struct radeon_bo_list, vm->max_pde_used + 2);
drivers/gpu/drm/v3d/v3d_submit.c
338
kvmalloc_objs(struct v3d_submit_outsync, count);
drivers/gpu/drm/v3d/v3d_submit.c
487
query_info->queries = kvmalloc_objs(struct v3d_timestamp_query,
drivers/gpu/drm/v3d/v3d_submit.c
545
query_info->queries = kvmalloc_objs(struct v3d_timestamp_query,
drivers/gpu/drm/v3d/v3d_submit.c
601
query_info->queries = kvmalloc_objs(struct v3d_timestamp_query,
drivers/gpu/drm/v3d/v3d_submit.c
727
kvmalloc_objs(struct v3d_performance_query, reset.count);
drivers/gpu/drm/v3d/v3d_submit.c
767
kvmalloc_objs(struct v3d_performance_query, copy.count);
drivers/gpu/drm/virtio/virtgpu_object.c
180
*ents = kvmalloc_objs(struct virtio_gpu_mem_entry, *nents);
drivers/gpu/drm/virtio/virtgpu_prime.c
167
*ents = kvmalloc_objs(struct virtio_gpu_mem_entry, sgt->nents);
drivers/gpu/drm/vmwgfx/vmwgfx_blit.c
589
src_pages = kvmalloc_objs(struct page *, src->ttm->num_pages);
drivers/gpu/drm/vmwgfx/vmwgfx_blit.c
598
dst_pages = kvmalloc_objs(struct page *, dst->ttm->num_pages);
drivers/gpu/drm/xe/xe_vm.c
3390
*bind_ops = kvmalloc_objs(struct drm_xe_vm_bind_op,
drivers/gpu/drm/xen/xen_drm_front_gem.c
50
xen_obj->pages = kvmalloc_objs(struct page *, xen_obj->num_pages);
drivers/infiniband/hw/mlx5/qp.c
1188
qp->sq.w_list = kvmalloc_objs(*qp->sq.w_list, qp->sq.wqe_cnt);
drivers/infiniband/ulp/srpt/ib_srpt.c
793
ring = kvmalloc_objs(ring[0], ring_size);
drivers/iommu/amd/init.c
696
pci_seg->alias_table = kvmalloc_objs(*pci_seg->alias_table,
drivers/md/dm-clone-target.c
583
clone->ht = kvmalloc_objs(struct hash_table_bucket, sz);
drivers/md/dm-integrity.c
4246
pl = kvmalloc_objs(struct page_list, n_pages + 1,
drivers/md/dm-integrity.c
4281
sl = kvmalloc_objs(struct scatterlist *, ic->journal_sections,
drivers/md/dm-integrity.c
4299
s = kvmalloc_objs(struct scatterlist, n_pages);
drivers/md/dm-integrity.c
4502
sg = kvmalloc_objs(struct scatterlist,
drivers/md/dm-integrity.c
4570
ic->sk_requests = kvmalloc_objs(struct skcipher_request *,
drivers/md/dm-integrity.c
5273
ic->bbs = kvmalloc_objs(struct bitmap_block_status,
drivers/md/dm-snap.c
666
et->table = kvmalloc_objs(struct dm_hlist_head, size);
drivers/media/v4l2-core/v4l2-ctrls-api.c
434
helpers = kvmalloc_objs(helper[0], cs->count);
drivers/media/v4l2-core/v4l2-ctrls-api.c
619
helpers = kvmalloc_objs(helper[0], cs->count);
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
96
indir->table = kvmalloc_objs(*indir->table, max_table_size);
drivers/net/ethernet/netronome/nfp/flower/metadata.c
576
priv->stats = kvmalloc_objs(struct nfp_fl_stats, stats_size);
drivers/net/ethernet/wangxun/ngbe/ngbe_ethtool.c
84
temp_ring = kvmalloc_objs(struct wx_ring, i);
drivers/net/ethernet/wangxun/txgbe/txgbe_ethtool.c
76
temp_ring = kvmalloc_objs(struct wx_ring, i);
drivers/net/hyperv/netvsc_drv.c
1527
pcpu_sum = kvmalloc_objs(struct netvsc_ethtool_pcpu_stats, nr_cpu_ids);
drivers/staging/media/ipu3/ipu3-dmamap.c
42
pages = kvmalloc_objs(*pages, count);
drivers/usb/gadget/function/f_fs.c
817
pages = kvmalloc_objs(struct page *, n_pages);
drivers/usb/host/xhci-sideband.c
31
pages = kvmalloc_objs(struct page *, n_pages);
drivers/xen/evtchn.c
335
new_ring = kvmalloc_objs(*new_ring, new_size);
drivers/xen/gntdev.c
148
add->grants = kvmalloc_objs(add->grants[0], count);
drivers/xen/gntdev.c
149
add->map_ops = kvmalloc_objs(add->map_ops[0], count);
drivers/xen/gntdev.c
150
add->unmap_ops = kvmalloc_objs(add->unmap_ops[0], count);
drivers/xen/gntdev.c
161
add->kmap_ops = kvmalloc_objs(add->kmap_ops[0], count);
drivers/xen/gntdev.c
162
add->kunmap_ops = kvmalloc_objs(add->kunmap_ops[0], count);
fs/ceph/file.c
143
bv = kvmalloc_objs(*bv, npages, GFP_KERNEL | __GFP_ZERO);
fs/coredump.c
1739
cprm->vma_meta = kvmalloc_objs(*cprm->vma_meta, cprm->vma_count);
fs/exfat/balloc.c
99
sbi->vol_amap = kvmalloc_objs(struct buffer_head *, sbi->map_sectors);
fs/ext2/super.c
1125
sbi->s_group_desc = kvmalloc_objs(struct buffer_head *, db_count);
fs/ext4/orphan.c
601
oi->of_binfo = kvmalloc_objs(struct ext4_orphan_block, oi->of_blocks);
fs/ext4/super.c
4906
kvmalloc_objs(struct buffer_head *, db_count));
fs/file.c
219
data = kvmalloc_objs(struct file *, nr, GFP_KERNEL_ACCOUNT);
fs/fuse/dev.c
1601
bufs = kvmalloc_objs(struct pipe_buffer, pipe->max_usage);
fs/fuse/dev.c
2313
bufs = kvmalloc_objs(struct pipe_buffer, count);
fs/jbd2/revoke.c
234
kvmalloc_objs(struct list_head, hash_size);
io_uring/io_uring.c
198
table->hbs = kvmalloc_objs(table->hbs[0], hash_buckets,
io_uring/memmap.c
161
pages = kvmalloc_objs(*pages, mr->nr_pages, gfp);
io_uring/memmap.c
59
pages = kvmalloc_objs(struct page *, nr_pages, GFP_KERNEL_ACCOUNT);
io_uring/rsrc.c
202
data->nodes = kvmalloc_objs(struct io_rsrc_node *, nr,
io_uring/rsrc.c
686
new_array = kvmalloc_objs(struct page *, nr_folios);
io_uring/zcrx.c
474
area->nia.niovs = kvmalloc_objs(area->nia.niovs[0], nr_iovs,
io_uring/zcrx.c
484
area->user_refs = kvmalloc_objs(area->user_refs[0], nr_iovs,
ipc/sem.c
2237
sops = kvmalloc_objs(*sops, nsops);
kernel/cgroup/cgroup-v1.c
355
array = kvmalloc_objs(pid_t, length);
kernel/dma/remap.c
48
pages = kvmalloc_objs(struct page *, count);
kernel/liveupdate/kexec_handover.c
1093
pages = kvmalloc_objs(*pages, total_pages);
lib/bucket_locks.c
34
tlocks = kvmalloc_objs(spinlock_t, size, gfp);
lib/iov_iter.c
906
*res = kvmalloc_objs(struct page *, count);
lib/test_kho.c
214
folios = kvmalloc_objs(*state->folios, max_nr);
mm/memfd_luo.c
115
folios = kvmalloc_objs(*folios, max_folios);
net/core/devmem.c
244
binding->tx_vec = kvmalloc_objs(struct net_iov *,
net/core/devmem.c
291
owner->area.niovs = kvmalloc_objs(*owner->area.niovs,
net/core/sock.c
1100
tokens = kvmalloc_objs(*tokens, num_tokens);
net/ipv4/udp.c
3980
new_batch = kvmalloc_objs(*new_batch, new_batch_sz,
net/netfilter/ipvs/ip_vs_conn.c
1513
ip_vs_conn_tab = kvmalloc_objs(*ip_vs_conn_tab, ip_vs_conn_tab_size);
net/netfilter/nf_tables_api.c
4322
expr_info = kvmalloc_objs(struct nft_expr_info,
net/netfilter/nft_set_pipapo.c
1463
dst->mt = kvmalloc_objs(*src->mt, src->rules_alloc,
net/netfilter/nft_set_pipapo.c
655
new_mt = kvmalloc_objs(*new_mt, rules_alloc, GFP_KERNEL_ACCOUNT);
net/openvswitch/flow_table.c
159
ti->buckets = kvmalloc_objs(struct hlist_head, new_size);
net/sched/cls_u32.c
1483
tc_u_common_hash = kvmalloc_objs(struct hlist_head, U32_HASH_SIZE);
net/sched/sch_api.c
671
h = kvmalloc_objs(struct hlist_head, n);
net/unix/af_unix.c
3803
net->unx.table.locks = kvmalloc_objs(spinlock_t, UNIX_HASH_SIZE);
net/unix/af_unix.c
3807
net->unx.table.buckets = kvmalloc_objs(struct hlist_head,
net/unix/garbage.c
298
fpl->edges = kvmalloc_objs(*fpl->edges, fpl->count_unix,
sound/core/seq/seq_memory.c
444
cellptr = kvmalloc_objs(struct snd_seq_event_cell, pool->size);