Symbol: __GFP_ZERO
arch/alpha/include/asm/page.h
16
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
arch/alpha/kernel/pci_iommu.c
416
cpu_addr = (void *)__get_free_pages(gfp | __GFP_ZERO, order);
arch/arc/kernel/perf_event.c
676
sizeof(*arc_pmu->attr), GFP_KERNEL | __GFP_ZERO);
arch/arc/kernel/perf_event.c
681
sizeof(*arc_pmu->attrs), GFP_KERNEL | __GFP_ZERO);
arch/arc/kernel/perf_event.c
686
sizeof(*arc_pmu->raw_entry), GFP_KERNEL | __GFP_ZERO);
arch/arm/mm/pgd.c
20
#define _pgd_alloc(mm) kmalloc_objs(pgd_t, PTRS_PER_PGD, GFP_KERNEL | __GFP_ZERO)
arch/arm64/kvm/arm.c
524
vcpu->arch.mmu_page_cache.gfp_zero = __GFP_ZERO;
arch/arm64/kvm/mmu.c
1005
mmu->split_page_cache.gfp_zero = __GFP_ZERO;
arch/arm64/kvm/mmu.c
1182
struct kvm_mmu_memory_cache cache = { .gfp_zero = __GFP_ZERO };
arch/arm64/kvm/mmu.c
211
return alloc_pages_exact(size, GFP_KERNEL_ACCOUNT | __GFP_ZERO);
arch/arm64/kvm/nested.c
80
__GFP_ZERO);
arch/arm64/kvm/nested.c
94
GFP_KERNEL_ACCOUNT | __GFP_ZERO);
arch/arm64/mm/fault.c
981
gfp_t flags = GFP_HIGHUSER_MOVABLE | __GFP_ZERO;
arch/arm64/mm/mmu.c
1277
alloc = __get_free_pages(GFP_ATOMIC | __GFP_ZERO, order);
arch/arm64/mm/mmu.c
527
struct ptdesc *ptdesc = pagetable_alloc(gfp & ~__GFP_ZERO, 0);
arch/m68k/include/asm/mcf_pgalloc.h
17
struct ptdesc *ptdesc = pagetable_alloc((GFP_DMA | __GFP_ZERO) &
arch/m68k/include/asm/mcf_pgalloc.h
50
struct ptdesc *ptdesc = pagetable_alloc(GFP_DMA | __GFP_ZERO, 0);
arch/m68k/include/asm/page_no.h
16
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
arch/m68k/mm/motorola.c
159
ptdesc = pagetable_alloc(GFP_KERNEL | __GFP_ZERO, 0);
arch/parisc/kernel/pci-dma.c
414
paddr = __get_free_pages(gfp | __GFP_ZERO, order);
arch/powerpc/include/asm/pgalloc.h
21
#define PGALLOC_GFP (GFP_KERNEL | __GFP_ZERO)
arch/powerpc/kernel/fadump.c
782
vaddr = alloc_pages_exact(size, GFP_KERNEL | __GFP_ZERO);
arch/powerpc/kvm/book3s_64_mmu_hv.c
91
hpt = __get_free_pages(GFP_KERNEL|__GFP_ZERO|__GFP_RETRY_MAYFAIL
arch/powerpc/kvm/book3s_64_vio.c
219
page = alloc_page(GFP_KERNEL | __GFP_ZERO);
arch/powerpc/kvm/book3s_pr.c
1746
p = __get_free_page(GFP_KERNEL|__GFP_ZERO);
arch/powerpc/kvm/booke.c
2196
kvmppc_booke_handlers = __get_free_pages(GFP_KERNEL | __GFP_ZERO,
arch/powerpc/kvm/e500.c
451
vcpu->arch.shared = (void*)__get_free_page(GFP_KERNEL|__GFP_ZERO);
arch/powerpc/kvm/e500mc.c
324
vcpu->arch.shared = (void *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
arch/powerpc/mm/book3s64/pgtable.c
396
gfp_t gfp = GFP_KERNEL_ACCOUNT | __GFP_ZERO;
arch/powerpc/perf/imc-pmu.c
1215
GFP_KERNEL | __GFP_ZERO | __GFP_THISNODE |
arch/powerpc/perf/imc-pmu.c
615
GFP_KERNEL | __GFP_ZERO | __GFP_THISNODE |
arch/powerpc/perf/imc-pmu.c
932
GFP_KERNEL | __GFP_ZERO | __GFP_THISNODE |
arch/powerpc/platforms/powernv/memtrace.c
106
__GFP_NOWARN | __GFP_ZERO, nid, NULL);
arch/powerpc/platforms/powernv/opal-core.c
324
GFP_KERNEL | __GFP_ZERO);
arch/powerpc/platforms/ps3/system-bus.c
514
flag |= __GFP_ZERO;
arch/riscv/kvm/aia_imsic.c
1115
swfile_page = alloc_pages(GFP_KERNEL | __GFP_ZERO,
arch/riscv/kvm/mmu.c
47
.gfp_zero = __GFP_ZERO,
arch/riscv/kvm/mmu.c
572
pgd_page = alloc_pages(GFP_KERNEL | __GFP_ZERO,
arch/riscv/kvm/nacl.c
141
shmem_page = alloc_pages(GFP_KERNEL | __GFP_ZERO,
arch/riscv/kvm/vcpu.c
137
vcpu->arch.mmu_page_cache.gfp_zero = __GFP_ZERO;
arch/s390/include/asm/page.h
71
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
arch/s390/kernel/diag/diag310.c
264
buf = __vmalloc_node(data_size, PAGE_SIZE, GFP_KERNEL | __GFP_ZERO | __GFP_ACCOUNT,
arch/s390/kernel/perf_pai.c
1091
attrs = kmalloc_objs(*attrs, min_attr + 1, GFP_KERNEL | __GFP_ZERO);
arch/s390/kvm/kvm-s390.c
3172
gfp_t alloc_flags = GFP_KERNEL_ACCOUNT | __GFP_ZERO;
arch/s390/kvm/pci.c
69
page = alloc_pages(GFP_KERNEL | __GFP_ZERO, size);
arch/s390/kvm/pv.c
286
sida_addr = (void *)__get_free_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);
arch/s390/kvm/vsie.c
1443
vsie_page = (void *)__get_free_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO | GFP_DMA);
arch/sh/mm/pgtable.c
52
return kmem_cache_alloc(pmd_cachep, GFP_KERNEL | __GFP_ZERO);
arch/sparc/kernel/irq_64.c
1022
p = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order);
arch/sparc/kernel/irq_64.c
1106
__get_free_pages(GFP_KERNEL | __GFP_ZERO, order);
arch/sparc/kernel/pci_sun4v.c
769
table = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, order);
arch/sparc/mm/init_64.c
2849
struct ptdesc *ptdesc = pagetable_alloc(GFP_KERNEL | __GFP_ZERO, 0);
arch/sparc/mm/io-unit.c
231
va = __get_free_pages(gfp | __GFP_ZERO, get_order(len));
arch/sparc/mm/iommu.c
329
va = __get_free_pages(gfp | __GFP_ZERO, get_order(len));
arch/um/kernel/skas/mmu.c
52
stack = __get_free_pages(GFP_KERNEL | __GFP_ZERO, ilog2(STUB_DATA_PAGES));
arch/x86/coco/sev/core.c
749
p = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL_ACCOUNT | __GFP_ZERO, 1);
arch/x86/events/intel/ds.c
837
page = __alloc_pages_node(node, flags | __GFP_ZERO, order);
arch/x86/events/intel/lbr.c
655
GFP_KERNEL | __GFP_ZERO,
arch/x86/events/intel/pt.c
686
p = alloc_pages_node(node, gfp | __GFP_ZERO, 0);
arch/x86/hyperv/hv_crash.c
546
gfp_t flags32 = GFP_KERNEL | GFP_DMA32 | __GFP_ZERO;
arch/x86/hyperv/hv_init.c
150
*hvp = __vmalloc(PAGE_SIZE, GFP_KERNEL | __GFP_ZERO);
arch/x86/hyperv/ivm.c
298
__get_free_page(GFP_KERNEL | __GFP_ZERO);
arch/x86/include/asm/page.h
32
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
arch/x86/kernel/amd_gart_64.c
649
gatt = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
arch/x86/kernel/amd_gart_64.c
758
iommu_gart_bitmap = (void *) __get_free_pages(GFP_KERNEL | __GFP_ZERO,
arch/x86/kernel/espfix_64.c
52
#define PGALLOC_GFP (GFP_KERNEL | __GFP_ZERO)
arch/x86/kernel/ldt.c
171
new_ldt->entries = __vmalloc(alloc_size, GFP_KERNEL_ACCOUNT | __GFP_ZERO);
arch/x86/kernel/machine_kexec_32.c
61
image->arch.pgd = (pgd_t *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
arch/x86/kvm/mmu/mmu.c
6691
vcpu->arch.mmu_pte_list_desc_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6694
vcpu->arch.mmu_page_header_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6699
vcpu->arch.mmu_shadow_page_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6862
kvm->arch.split_page_header_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6864
kvm->arch.split_shadow_page_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6867
kvm->arch.split_desc_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/svm/avic.c
295
kvm_svm->avic_physical_id_table = (void *)__get_free_pages(GFP_KERNEL_ACCOUNT | __GFP_ZERO,
arch/x86/kvm/svm/sev.c
1188
tpage = (void *)alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);
arch/x86/kvm/svm/sev.c
2981
status = snp_alloc_firmware_page(GFP_KERNEL | __GFP_ZERO);
arch/x86/kvm/svm/sev.c
411
req_page = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);
arch/x86/kvm/svm/sev.c
415
sev->guest_resp_buf = snp_alloc_firmware_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);
arch/x86/kvm/svm/sev.c
4817
return alloc_pages_node(node, gfp | __GFP_ZERO, 0);
arch/x86/kvm/svm/sev.c
4828
p = alloc_pages_node(node, gfp | __GFP_ZERO, 1);
arch/x86/kvm/svm/sev.c
5144
vmsa = snp_alloc_firmware_page(__GFP_ZERO);
arch/x86/kvm/svm/svm.h
905
return alloc_pages_node(node, gfp | __GFP_ZERO, 0);
arch/x86/kvm/vmx/vmx.c
4919
pages = alloc_pages(GFP_KERNEL_ACCOUNT | __GFP_ZERO,
arch/x86/kvm/vmx/vmx.c
7792
vmx->pml_pg = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);
arch/x86/kvm/vmx/vmx.c
7848
page = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);
arch/x86/kvm/x86.c
12762
page = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);
arch/x86/mm/init.c
135
return (void *)__get_free_pages(GFP_ATOMIC | __GFP_ZERO, order);
arch/x86/mm/pti.c
181
gfp_t gfp = (GFP_KERNEL | __GFP_NOTRACK | __GFP_ZERO);
arch/x86/mm/pti.c
208
gfp_t gfp = (GFP_KERNEL | __GFP_NOTRACK | __GFP_ZERO);
arch/x86/mm/pti.c
253
gfp_t gfp = (GFP_KERNEL | __GFP_NOTRACK | __GFP_ZERO);
arch/x86/platform/efi/efi_64.c
76
gfp_mask = GFP_KERNEL | __GFP_ZERO;
arch/x86/virt/svm/sev.c
351
page = alloc_page(__GFP_ZERO);
arch/x86/virt/vmx/tdx/tdx.c
325
GFP_KERNEL | __GFP_ZERO);
arch/x86/xen/mmu_pv.c
1468
user_pgd = (pgd_t *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
block/bfq-cgroup.c
1431
bfqg = kmalloc_node(sizeof(*bfqg), GFP_KERNEL | __GFP_ZERO, node);
block/bfq-iosched.c
5850
bfqq = kmem_cache_alloc_node(bfq_pool, GFP_NOWAIT | __GFP_ZERO,
block/bio-integrity.c
72
gfp_t gfp = GFP_NOIO | (zero_buffer ? __GFP_ZERO : 0);
block/blk-core.c
398
q = kmem_cache_alloc_node(blk_requestq_cachep, GFP_KERNEL | __GFP_ZERO,
block/blk-ioc.c
226
ioc = kmem_cache_alloc_node(iocontext_cachep, gfp_flags | __GFP_ZERO,
block/blk-ioc.c
362
icq = kmem_cache_alloc_node(et->icq_cache, GFP_ATOMIC | __GFP_ZERO,
block/blk-map.c
394
page = alloc_page(GFP_NOIO | __GFP_ZERO | gfp_mask);
block/blk-mq-sched.c
510
gfp_t gfp = GFP_NOIO | __GFP_ZERO | __GFP_NOWARN | __GFP_NORETRY;
block/blk-mq.c
3647
GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY | __GFP_ZERO,
block/kyber-iosched.c
364
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/common/command_buffer.c
143
GFP_USER | __GFP_ZERO);
drivers/accel/habanalabs/common/hw_queue.c
834
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/common/hw_queue.c
900
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/common/irq.c
601
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/common/memory.c
1606
rc = sg_alloc_table(sgt, nents, GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/common/mmu/mmu.c
1088
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/common/mmu/mmu.c
856
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/gaudi/gaudi.c
1071
cpu_addr = hl_asic_dma_alloc_coherent(hdev, fw_size, &dma_handle, GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/gaudi/gaudi.c
1711
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/gaudi/gaudi.c
1792
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/gaudi/gaudi.c
5946
kernel_addr = hl_asic_dma_alloc_coherent(hdev, SZ_2M, &dma_addr, GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/gaudi/gaudi.c
8316
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/gaudi2/gaudi2.c
11045
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/gaudi2/gaudi2.c
11138
GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/gaudi2/gaudi2.c
3808
&dma_addr_arr[i], GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/goya/goya.c
996
GFP_KERNEL | __GFP_ZERO);
drivers/accel/ivpu/ivpu_mmu_context.c
53
page = alloc_page(GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO);
drivers/accel/qaic/qaic_data.c
489
__GFP_NOWARN | __GFP_ZERO |
drivers/acpi/acpi_mrrm.c
85
GFP_KERNEL | __GFP_ZERO);
drivers/android/binder_alloc.c
287
page = alloc_page(GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO);
drivers/base/devres.c
146
if (!(gfp & __GFP_ZERO))
drivers/base/devres.c
189
dr = alloc_dr(release, size, gfp | __GFP_ZERO, nid);
drivers/block/brd.c
91
page = alloc_page(gfp | __GFP_ZERO | __GFP_HIGHMEM);
drivers/block/drbd/drbd_bitmap.c
399
new_pages = __vmalloc(bytes, GFP_NOIO | __GFP_ZERO);
drivers/block/null_blk/zoned.c
95
GFP_KERNEL | __GFP_ZERO);
drivers/block/ublk_drv.c
4167
gfp_t gfp_flags = GFP_KERNEL | __GFP_ZERO;
drivers/block/xen-blkfront.c
2201
__GFP_ZERO);
drivers/block/xen-blkfront.c
322
granted_page = alloc_page(GFP_NOIO | __GFP_ZERO);
drivers/char/agp/generic.c
1193
page = alloc_page(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO);
drivers/char/agp/generic.c
1221
page = alloc_page(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO);
drivers/char/agp/intel-gtt.c
301
page = alloc_page(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO);
drivers/char/misc_minor_kunit.c
403
GFP_KERNEL | __GFP_ZERO);
drivers/char/misc_minor_kunit.c
446
GFP_KERNEL | __GFP_ZERO);
drivers/char/misc_minor_kunit.c
487
GFP_KERNEL | __GFP_ZERO);
drivers/char/xillybus/xillybus_core.c
386
GFP_KERNEL | __GFP_DMA32 | __GFP_ZERO,
drivers/clocksource/arm_arch_timer_mmio.c
383
at = devm_kmalloc(&pdev->dev, sizeof(*at), GFP_KERNEL | __GFP_ZERO);
drivers/clocksource/arm_arch_timer_mmio.c
389
GFP_KERNEL | __GFP_ZERO);
drivers/crypto/ccp/dbc.c
203
dbc_dev->mbox = (void *)devm_get_free_pages(dev, GFP_KERNEL | __GFP_ZERO, 0);
drivers/crypto/ccp/sev-dev-tio.c
444
pg = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/crypto/ccp/sev-dev-tio.c
451
pg = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/crypto/ccp/sev-dev-tio.c
459
pg = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/crypto/ccp/sev-dev.c
1136
page = alloc_pages(GFP_KERNEL | __GFP_ZERO, order);
drivers/crypto/ccp/sev-dev.c
1498
GFP_KERNEL_ACCOUNT | __GFP_ZERO, 0, true));
drivers/crypto/intel/qat/qat_common/qat_comp_algs.c
70
scratch->workspace = kvmalloc(cctx_size, GFP_KERNEL | __GFP_ZERO);
drivers/dibs/dibs_loopback.c
78
__GFP_NORETRY | __GFP_ZERO,
drivers/dma-buf/dma-buf-mapping.c
136
ret = sg_alloc_table(&dma->sgt, nents, GFP_KERNEL | __GFP_ZERO);
drivers/dma-buf/heaps/system_heap.c
41
#define LOW_ORDER_GFP (GFP_HIGHUSER | __GFP_ZERO)
drivers/dma-buf/heaps/system_heap.c
42
#define HIGH_ORDER_GFP (((GFP_HIGHUSER | __GFP_ZERO | __GFP_NOWARN \
drivers/dma-buf/st-dma-fence-chain.c
119
fc->chains = kvmalloc_objs(*fc->chains, count, GFP_KERNEL | __GFP_ZERO);
drivers/dma-buf/st-dma-fence-chain.c
123
fc->fences = kvmalloc_objs(*fc->fences, count, GFP_KERNEL | __GFP_ZERO);
drivers/dma/ioat/prep.c
83
gfp_t flags = __GFP_ZERO | GFP_ATOMIC;
drivers/firewire/core-iso.c
42
unsigned long nr_populated = alloc_pages_bulk(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO,
drivers/firmware/efi/capsule-loader.c
63
GFP_KERNEL | __GFP_ZERO);
drivers/firmware/efi/capsule-loader.c
71
GFP_KERNEL | __GFP_ZERO);
drivers/firmware/efi/stmm/tee_stmm_efi.c
177
GFP_KERNEL | __GFP_ZERO);
drivers/firmware/stratix10-svc.c
1939
sizeof(*chans), GFP_KERNEL | __GFP_ZERO);
drivers/gpio/gpiolib.c
5200
GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
121
gfp_t gfp_flags = GFP_KERNEL | __GFP_ZERO;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
2246
qpd->cwsr_kaddr = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
drivers/gpu/drm/drm_managed.c
146
GFP_KERNEL | __GFP_ZERO,
drivers/gpu/drm/gma500/psb_drv.c
326
dev_priv->scratch_page = alloc_page(GFP_DMA32 | __GFP_ZERO);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
119
page = alloc_pages(GFP | __GFP_ZERO, order);
drivers/gpu/drm/i915/gvt/gtt.c
771
spt = alloc_spt(GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/i915/gvt/opregion.c
229
__GFP_ZERO,
drivers/gpu/drm/imagination/pvr_job.c
722
GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/imagination/pvr_mmu.c
269
raw_page = alloc_page(__GFP_ZERO | GFP_KERNEL);
drivers/gpu/drm/imagination/pvr_vm_mips.c
51
mips_data->pt_pages[page_nr] = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/lima/lima_gem.c
43
GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/lima/lima_vm.c
213
GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO);
drivers/gpu/drm/lima/lima_vm.c
59
&vm->bts[pbe].dma, GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO);
drivers/gpu/drm/msm/adreno/a2xx_gpummu.c
103
GFP_KERNEL | __GFP_ZERO, DMA_ATTR_FORCE_CONTIGUOUS);
drivers/gpu/drm/msm/msm_gem_vma.c
882
GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/msm/msm_iommu.c
392
gfp | __GFP_ZERO, get_order(size));
drivers/gpu/drm/msm/msm_iommu.c
604
iommu->prr_page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/nouveau/nouveau_dmem.c
242
dpage = folio_page(vma_alloc_folio(GFP_HIGHUSER | __GFP_ZERO,
drivers/gpu/drm/nouveau/nouveau_dmem.c
245
dpage = alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO, vmf->vma,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
282
fb->sysmem.flush_page = alloc_page(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
155
gfp_t gfp = GFP_USER | __GFP_ZERO;
drivers/gpu/drm/panfrost/panfrost_drv.c
211
job->bo_count, GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/panfrost/panfrost_mmu.c
625
GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/panfrost/panfrost_mmu.c
632
GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/panthor/panthor_device.c
214
p = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/panthor/panthor_drv.c
729
GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/panthor/panthor_mmu.c
444
gfp | __GFP_ZERO, get_order(size));
drivers/gpu/drm/radeon/radeon_device.c
786
rdev->dummy_page.page = alloc_page(GFP_DMA32 | GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/scheduler/sched_main.c
1357
GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/tegra/drm.c
1024
gfp = GFP_KERNEL | __GFP_ZERO;
drivers/gpu/drm/tilcdc/tilcdc_crtc.c
992
GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/ttm/ttm_bo_vm.c
301
page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/ttm/ttm_device.c
102
glob->dummy_read_page = alloc_page(__GFP_ZERO | GFP_DMA32 |
drivers/gpu/drm/ttm/ttm_device.c
107
glob->dummy_read_page = alloc_page(__GFP_ZERO);
drivers/gpu/drm/ttm/ttm_pool.c
774
gfp_flags |= __GFP_ZERO;
drivers/gpu/drm/v3d/v3d_drv.c
427
GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO);
drivers/gpu/drm/v3d/v3d_gem.c
323
GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO);
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
700
struct page *page = alloc_pages(GFP_KERNEL | __GFP_ZERO, MKSSTAT_KERNEL_PAGES_ORDER);
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
994
page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
112
struct page *page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/xe/xe_device.c
297
page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/xe/xe_mmio_gem.c
174
page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/gpu/drm/xe/xe_page_reclaim.c
132
page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/hid/hid-playstation.c
872
GFP_KERNEL | __GFP_ZERO);
drivers/hv/channel.c
180
GFP_KERNEL|__GFP_ZERO, order);
drivers/hv/channel.c
183
page = alloc_pages(GFP_KERNEL|__GFP_ZERO, order);
drivers/hv/connection.c
226
(void *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
drivers/hv/mshv_vtl_main.c
212
reg_page = alloc_page(GFP_KERNEL | __GFP_ZERO | __GFP_RETRY_MAYFAIL);
drivers/hv/mshv_vtl_main.c
330
per_cpu->run = (struct mshv_vtl_run *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
drivers/hwtracing/coresight/coresight-tmc-etr.c
224
GFP_KERNEL | __GFP_ZERO, 0);
drivers/hwtracing/intel_th/msu.c
936
page = alloc_pages(GFP_KERNEL | __GFP_ZERO | GFP_DMA32, order);
drivers/infiniband/core/uverbs_main.c
807
alloc_pages(vmf->gfp_mask | __GFP_ZERO, 0);
drivers/infiniband/hw/efa/efa_verbs.c
190
addr = alloc_pages_exact(size, GFP_KERNEL | __GFP_ZERO);
drivers/infiniband/hw/mlx5/umr.c
509
gfp_mask |= __GFP_ZERO | __GFP_NORETRY;
drivers/infiniband/hw/mthca/mthca_memfree.c
115
page = alloc_pages(gfp_mask | __GFP_ZERO, order);
drivers/input/misc/xen-kbdfront.c
219
info->page = (void *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
drivers/iommu/amd/ppr.c
22
iommu->ppr_log = iommu_alloc_4k_pages(iommu, GFP_KERNEL | __GFP_ZERO,
drivers/iommu/dma-iommu.c
1654
gfp |= __GFP_ZERO;
drivers/iommu/fsl_pamu.c
831
p = alloc_pages(GFP_KERNEL | __GFP_ZERO, order);
drivers/iommu/io-pgtable-arm-v7s.c
257
table = (void *)__get_free_pages(gfp_l1 | __GFP_ZERO, get_order(size));
drivers/iommu/iommu-pages.c
66
folio = __folio_alloc_node(gfp | __GFP_ZERO, order, nid);
drivers/iommu/iommufd/selftest.c
840
(u32 *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, 1);
drivers/irqchip/irq-gic-v3-its.c
2402
page = its_alloc_pages_node(its->numa_node, GFP_KERNEL | __GFP_ZERO, order);
drivers/irqchip/irq-gic-v3-its.c
261
page = its_alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, get_order(size));
drivers/irqchip/irq-gic-v3-its.c
271
page = its_alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0);
drivers/irqchip/irq-gic-v3-its.c
2878
page = its_alloc_pages(GFP_KERNEL | __GFP_ZERO, get_order(psz));
drivers/irqchip/irq-gic-v3-its.c
2997
page = its_alloc_pages(GFP_ATOMIC | __GFP_ZERO, get_order(np * PAGE_SIZE));
drivers/irqchip/irq-gic-v3-its.c
3042
pend_page = its_alloc_pages(gfp_flags | __GFP_ZERO, get_order(LPI_PENDBASE_SZ));
drivers/irqchip/irq-gic-v3-its.c
3389
page = its_alloc_pages_node(its->numa_node, GFP_KERNEL | __GFP_ZERO,
drivers/irqchip/irq-gic-v3-its.c
5275
GFP_KERNEL | __GFP_ZERO,
drivers/mailbox/riscv-sbi-mpxy-mbox.c
349
shmem_page = alloc_pages(GFP_KERNEL | __GFP_ZERO, get_order(mpxy_shmem_size));
drivers/md/bcache/super.c
1868
((void *) __get_free_pages(__GFP_ZERO|__GFP_COMP|gfp, ilog2(meta_bucket_pages(sb))))
drivers/md/dm-integrity.c
4243
GFP_KERNEL | __GFP_ZERO);
drivers/md/dm-integrity.c
4278
GFP_KERNEL | __GFP_ZERO);
drivers/md/dm-integrity.c
4568
GFP_KERNEL | __GFP_ZERO);
drivers/md/dm-vdo/memory-alloc.c
213
const gfp_t gfp_flags = GFP_KERNEL | __GFP_ZERO | __GFP_RETRY_MAYFAIL;
drivers/md/dm-vdo/memory-alloc.c
307
void *p = kmalloc(size, GFP_NOWAIT | __GFP_ZERO);
drivers/md/md-bitmap.c
1033
store->sb_page = alloc_page(GFP_KERNEL|__GFP_ZERO);
drivers/md/md-bitmap.c
1046
store->filemap[pnum] = alloc_page(GFP_KERNEL|__GFP_ZERO);
drivers/md/md-bitmap.c
763
bitmap->storage.sb_page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/md/md-llbitmap.c
524
page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/md/md-llbitmap.c
619
GFP_KERNEL | __GFP_ZERO);
drivers/md/md-llbitmap.c
624
pctl = kmalloc_array(nr_pages, size, GFP_KERNEL | __GFP_ZERO);
drivers/md/raid5-ppl.c
1045
page = alloc_page(GFP_NOIO | __GFP_ZERO);
drivers/media/common/videobuf2/videobuf2-dma-sg.c
78
pages = alloc_pages(GFP_KERNEL | __GFP_ZERO |
drivers/media/pci/cobalt/cobalt-alsa-pcm.c
468
s->q.gfp_flags |= __GFP_ZERO;
drivers/media/platform/amphion/vdec.c
1941
GFP_KERNEL | __GFP_ZERO);
drivers/media/platform/amphion/vdec.c
331
GFP_KERNEL | __GFP_ZERO);
drivers/misc/lkdtm/usercopy.c
412
folio = folio_alloc(GFP_KERNEL | __GFP_ZERO, 1);
drivers/misc/ocxl/link.c
358
__get_free_pages(GFP_KERNEL | __GFP_ZERO, spa->spa_order);
drivers/misc/sgi-xp/xpc_uv.c
174
GFP_KERNEL | __GFP_ZERO | __GFP_THISNODE,
drivers/misc/vmw_balloon.c
1316
page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
267
__GFP_ZERO | __GFP_NORETRY);
drivers/net/dsa/lantiq/lantiq_gswip.c
282
GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1239
GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1249
GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1413
GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
31
req_addr = dma_pool_alloc(bd->hwrm_dma_pool, GFP_KERNEL | __GFP_ZERO,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
2717
__get_free_page(GFP_ATOMIC | __GFP_ZERO);
drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c
340
hwrm_req_alloc_flags(bp, get, GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
70
req_addr = dma_pool_alloc(bp->hwrm_dma_pool, GFP_KERNEL | __GFP_ZERO,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4080
__GFP_ZERO, page_order);
drivers/net/ethernet/ibm/ibmvnic.c
4052
(union sub_crq *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, 2);
drivers/net/ethernet/intel/ice/ice_controlq.c
117
GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/intel/ice/ice_controlq.c
168
GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/intel/ice/ice_controlq.c
237
GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/intel/ice/ice_controlq.c
97
GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/intel/ice/ice_lib.c
3046
GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/intel/ice/ice_lib.c
3065
GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/mediatek/mtk_wed.c
548
__GFP_ZERO | __GFP_COMP |
drivers/net/ethernet/mellanox/mlx4/mr.c
118
buddy->bits[i] = kvmalloc_array(s, sizeof(long), GFP_KERNEL | __GFP_ZERO);
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
137
gfp = GFP_KERNEL | __GFP_ZERO;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/bwc.c
788
__GFP_ZERO | GFP_KERNEL);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/matcher.c
1024
__GFP_ZERO | GFP_KERNEL);
drivers/net/wireless/ath/ath12k/dp.c
1420
GFP_KERNEL | __GFP_ZERO);
drivers/net/wireless/ti/wl18xx/main.c
1583
GFP_KERNEL | __GFP_ZERO);
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_rx.c
1023
&rxq->pit_bus_addr, GFP_KERNEL | __GFP_ZERO);
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_rx.c
965
&bat_req->bat_bus_addr, GFP_KERNEL | __GFP_ZERO);
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_tx.c
546
&txq->drb_bus_addr, GFP_KERNEL | __GFP_ZERO);
drivers/net/xen-netfront.c
283
GFP_ATOMIC | __GFP_NOWARN | __GFP_ZERO);
drivers/net/xen-netfront.c
682
struct sk_buff *n = alloc_skb(size, GFP_ATOMIC | __GFP_ZERO);
drivers/nvme/host/tcp.c
1429
GFP_KERNEL | __GFP_ZERO);
drivers/nvme/host/tcp.c
562
GFP_KERNEL | __GFP_ZERO);
drivers/nvme/target/pci-epf.c
710
GFP_KERNEL | __GFP_ZERO);
drivers/nvme/target/tcp.c
1463
sizeof(*c->cmd_pdu) + hdgst, GFP_KERNEL | __GFP_ZERO);
drivers/nvme/target/tcp.c
1469
sizeof(*c->rsp_pdu) + hdgst, GFP_KERNEL | __GFP_ZERO);
drivers/nvme/target/tcp.c
1475
sizeof(*c->data_pdu) + hdgst, GFP_KERNEL | __GFP_ZERO);
drivers/nvme/target/tcp.c
1480
sizeof(*c->r2t_pdu) + hdgst, GFP_KERNEL | __GFP_ZERO);
drivers/perf/riscv_pmu_sbi.c
680
snapshot_page = alloc_page(GFP_ATOMIC | __GFP_ZERO);
drivers/pinctrl/mediatek/mtk-eint.c
521
GFP_KERNEL | __GFP_ZERO);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
392
sizeof(*hw->eint->base), GFP_KERNEL | __GFP_ZERO);
drivers/ptp/ptp_kvm_x86.c
32
p = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/s390/char/sclp_cmd.c
105
sccb = (void *)__get_free_pages(GFP_KERNEL | GFP_DMA | __GFP_ZERO, get_order(length));
drivers/s390/cio/css.c
1028
#define CIO_DMA_GFP (GFP_KERNEL | __GFP_ZERO)
drivers/scsi/cxgbi/libcxgbi.c
3082
rsvd_page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/scsi/fnic/fnic_fcs.c
523
GFP_ATOMIC | __GFP_ZERO);
drivers/scsi/fnic/fnic_fcs.c
708
GFP_ATOMIC | __GFP_ZERO);
drivers/scsi/megaraid/megaraid_sas_fusion.c
1648
__get_free_pages(__GFP_ZERO | GFP_KERNEL,
drivers/scsi/megaraid/megaraid_sas_fusion.c
5311
(PLD_SPAN_INFO)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
drivers/scsi/megaraid/megaraid_sas_fusion.c
5327
(struct LD_LOAD_BALANCE_INFO *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
drivers/scsi/qedi/qedi_main.c
248
__GFP_ZERO, 2);
drivers/scsi/sg.c
1870
gfp_t gfp_mask = GFP_ATOMIC | __GFP_COMP | __GFP_NOWARN | __GFP_ZERO;
drivers/scsi/st.c
4014
priority |= __GFP_ZERO;
drivers/sh/maple/maple.c
742
(void *) __get_free_pages(GFP_KERNEL | __GFP_ZERO,
drivers/spi/spi-imx.c
1671
GFP_KERNEL | __GFP_ZERO);
drivers/spi/spi-qcom-qspi.c
311
virt_cmd_desc = dma_pool_alloc(ctrl->dma_cmd_pool, GFP_ATOMIC | __GFP_ZERO, &dma_cmd_desc);
drivers/spi/spi.c
1410
GFP_KERNEL | GFP_DMA | __GFP_ZERO);
drivers/staging/media/ipu3/ipu3-dmamap.c
47
gfp |= __GFP_HIGHMEM | __GFP_ZERO;
drivers/target/iscsi/cxgbit/cxgbit_target.c
380
page = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/target/target_core_transport.c
2779
gfp_t gfp = GFP_KERNEL | (zero_page ? __GFP_ZERO : 0);
drivers/target/target_core_user.c
529
page = alloc_page(GFP_NOIO | __GFP_ZERO);
drivers/tee/amdtee/shm_pool.c
22
va = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order);
drivers/tee/tee_shm.c
351
GFP_KERNEL | __GFP_ZERO);
drivers/thermal/intel/intel_hfi.c
459
GFP_KERNEL | __GFP_ZERO);
drivers/thunderbolt/nhi.c
615
&ring->descriptors_dma, GFP_KERNEL | __GFP_ZERO);
drivers/tty/hvc/hvc_xen.c
526
info->intf = (void *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
drivers/usb/gadget/function/uvc_configfs.c
881
GFP_KERNEL | __GFP_ZERO);
drivers/usb/gadget/function/uvc_configfs.c
934
GFP_KERNEL | __GFP_ZERO);
drivers/vhost/scsi.c
2294
GFP_KERNEL | __GFP_ZERO);
drivers/vhost/scsi.c
2298
vs->vqs = kmalloc_objs(*vs->vqs, nvqs, GFP_KERNEL | __GFP_ZERO);
drivers/video/fbdev/fsl-diu-fb.c
1693
&dma_addr, GFP_DMA | __GFP_ZERO);
drivers/video/fbdev/fsl-diu-fb.c
873
p = alloc_pages_exact(smem_len, GFP_DMA | __GFP_ZERO);
drivers/video/fbdev/pxafb.c
1699
fbi->video_mem = alloc_pages_exact(size, GFP_KERNEL | __GFP_ZERO);
drivers/video/fbdev/pxafb.c
915
GFP_KERNEL | __GFP_ZERO);
drivers/video/fbdev/ssd1307fb.c
670
vmem = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
drivers/video/fbdev/xen-fbfront.c
398
info->page = (void *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
drivers/virt/coco/sev-guest/sev-guest.c
214
page = alloc_pages(GFP_KERNEL_ACCOUNT | __GFP_ZERO,
drivers/virt/coco/sev-guest/sev-guest.c
387
buffer = alloc_pages_exact(size, __GFP_ZERO);
drivers/virt/coco/tdx-guest/tdx-guest.c
232
addr = alloc_pages_exact(len, GFP_KERNEL | __GFP_ZERO);
drivers/virtio/virtio_ring.c
1265
GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO,
drivers/virtio/virtio_ring.c
1279
&dma_addr, GFP_KERNEL | __GFP_ZERO,
drivers/virtio/virtio_ring.c
2405
GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO,
drivers/virtio/virtio_ring.c
2418
GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO,
drivers/virtio/virtio_ring.c
2429
GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO,
drivers/xen/events/events_base.c
2323
pirq_eoi_map = (void *)__get_free_page(GFP_KERNEL|__GFP_ZERO);
drivers/xen/gntalloc.c
140
gref->page = alloc_page(GFP_KERNEL|__GFP_ZERO);
drivers/xen/privcmd-buf.c
149
vma_priv->pages[i] = alloc_page(GFP_KERNEL | __GFP_ZERO);
drivers/xen/pvcalls-front.c
1198
__GFP_ZERO);
drivers/xen/pvcalls-front.c
367
GFP_KERNEL | __GFP_ZERO);
drivers/xen/xenbus/xenbus_client.c
404
addr = *vaddr = alloc_pages_exact(ring_size, gfp | __GFP_ZERO);
fs/aio.c
579
GFP_USER | __GFP_ZERO);
fs/ceph/file.c
143
bv = kvmalloc_objs(*bv, npages, GFP_KERNEL | __GFP_ZERO);
fs/ceph/mds_client.c
2554
__GFP_ZERO,
fs/ceph/mdsmap.c
324
GFP_NOFS | __GFP_ZERO);
fs/dlm/lowcomms.c
1173
entry->page = alloc_page(GFP_ATOMIC | __GFP_ZERO);
fs/ecryptfs/crypto.c
1027
page = alloc_pages(gfp_mask | __GFP_ZERO, order);
fs/exec.c
209
page = alloc_page(GFP_HIGHUSER|__GFP_ZERO);
fs/ext4/super.c
4232
char *buf = kvmalloc(sb->s_blocksize, GFP_NOFS | __GFP_ZERO);
fs/f2fs/data.c
2440
GFP_NOIO | __GFP_ZERO, true, NULL);
fs/f2fs/f2fs.h
3698
return f2fs_kmalloc(sbi, size, flags | __GFP_ZERO);
fs/f2fs/f2fs.h
3713
return f2fs_kvmalloc(sbi, size, flags | __GFP_ZERO);
fs/fuse/cuse.c
442
folio = folio_alloc(GFP_KERNEL | __GFP_ZERO, 0);
fs/gfs2/dir.c
1995
ht = __vmalloc(size, GFP_NOFS | __GFP_NOWARN | __GFP_ZERO);
fs/gfs2/quota.c
1413
__GFP_ZERO);
fs/jfs/jfs_logmgr.c
1815
struct page *page = alloc_page(GFP_KERNEL | __GFP_ZERO);
fs/ntfs3/bitmap.c
680
kvmalloc_array(wnd->nwnd, sizeof(u16), GFP_KERNEL | __GFP_ZERO);
fs/ntfs3/file.c
950
gfp_mask | __GFP_ZERO);
fs/ocfs2/cluster/tcp.c
2104
folio = folio_alloc(GFP_KERNEL | __GFP_ZERO, 0);
fs/proc/vmcore.c
1244
elfcorebuf = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
fs/proc/vmcore.c
1299
elfcorebuf = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
fs/proc/vmcore.c
1554
elfcorebuf_new = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
fs/smb/server/ksmbd_work.c
125
KSMBD_DEFAULT_GFP | __GFP_ZERO);
fs/squashfs/block.c
92
GFP_KERNEL | __GFP_ZERO);
fs/xfs/xfs_buf.c
185
gfp_mask |= __GFP_ZERO;
fs/xfs/xfs_zone_alloc.c
1192
sizeof(unsigned long), GFP_KERNEL | __GFP_ZERO);
include/asm-generic/pgalloc.h
7
#define GFP_PGTABLE_KERNEL (GFP_KERNEL | __GFP_ZERO)
include/drm/drm_managed.h
105
return drmm_kmalloc_array(dev, n, size, flags | __GFP_ZERO);
include/drm/drm_managed.h
66
return drmm_kmalloc(dev, size, gfp | __GFP_ZERO);
include/kunit/test.h
521
return kunit_kmalloc(test, size, gfp | __GFP_ZERO);
include/kunit/test.h
535
return kunit_kmalloc_array(test, n, size, gfp | __GFP_ZERO);
include/linux/cpumask.h
1034
return alloc_cpumask_var_node(mask, flags | __GFP_ZERO, node);
include/linux/cpumask.h
1058
return alloc_cpumask_var(mask, flags | __GFP_ZERO);
include/linux/device/devres.h
50
return devm_kmalloc(dev, size, gfp | __GFP_ZERO);
include/linux/device/devres.h
63
return devm_kmalloc_array(dev, n, size, flags | __GFP_ZERO);
include/linux/dmapool.h
74
return dma_pool_alloc(pool, mem_flags | __GFP_ZERO, handle);
include/linux/f2fs_fs.h
46
#define GFP_F2FS_ZERO (GFP_NOFS | __GFP_ZERO)
include/linux/mm.h
4662
return flags & __GFP_ZERO;
include/linux/ptr_ring.h
481
return kvmalloc_array_noprof(size, sizeof(void *), gfp | __GFP_ZERO);
include/linux/scs.h
20
#define GFP_SCS (GFP_KERNEL | __GFP_ZERO)
include/linux/slab.h
1138
#define kcalloc(n, size, flags) kmalloc_array(n, size, (flags) | __GFP_ZERO)
include/linux/slab.h
1174
kmalloc_array_node(_n, _size, (_flags) | __GFP_ZERO, _node)
include/linux/slab.h
1179
#define kmem_cache_zalloc(_k, _flags) kmem_cache_alloc(_k, (_flags)|__GFP_ZERO)
include/linux/slab.h
1188
return kmalloc_noprof(size, flags | __GFP_ZERO);
include/linux/slab.h
1191
#define kzalloc_node(_size, _flags, _node) kmalloc_node(_size, (_flags)|__GFP_ZERO, _node)
include/linux/slab.h
1201
#define kvzalloc(_size, _flags) kvmalloc(_size, (_flags)|__GFP_ZERO)
include/linux/slab.h
1203
#define kvzalloc_node(_size, _flags, _node) kvmalloc_node(_size, (_flags)|__GFP_ZERO, _node)
include/linux/slab.h
1220
#define kvcalloc_node_noprof(_n,_s,_f,_node) kvmalloc_array_node_noprof(_n,_s,(_f)|__GFP_ZERO,_node)
include/linux/thread_info.h
95
#define THREADINFO_GFP (GFP_KERNEL_ACCOUNT | __GFP_ZERO)
include/rdma/uverbs_ioctl.h
876
return _uverbs_alloc(bundle, size, GFP_KERNEL | __GFP_ZERO);
include/rv/da_monitor.h
344
__GFP_ZERO, NUMA_NO_NODE);
io_uring/io_uring.c
967
gfp_t gfp = GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO;
io_uring/memmap.c
156
gfp_t gfp = GFP_KERNEL_ACCOUNT | __GFP_ZERO | __GFP_NOWARN;
io_uring/msg_ring.c
102
target = kmem_cache_alloc(req_cachep, GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO) ;
io_uring/rsrc.c
203
GFP_KERNEL_ACCOUNT | __GFP_ZERO);
io_uring/zcrx.c
480
GFP_KERNEL_ACCOUNT | __GFP_ZERO);
io_uring/zcrx.c
485
GFP_KERNEL_ACCOUNT | __GFP_ZERO);
io_uring/zcrx.c
490
GFP_KERNEL_ACCOUNT | __GFP_ZERO);
kernel/bpf/bpf_local_storage.c
491
__GFP_ZERO, NUMA_NO_NODE);
kernel/bpf/bpf_local_storage.c
79
__GFP_ZERO, NUMA_NO_NODE);
kernel/bpf/core.c
101
gfp_t gfp_flags = bpf_memcg_flags(GFP_KERNEL | __GFP_ZERO | gfp_extra_flags);
kernel/bpf/core.c
1453
gfp_t gfp_flags = GFP_KERNEL | __GFP_ZERO | gfp_extra_flags;
kernel/bpf/core.c
151
gfp_t gfp_flags = bpf_memcg_flags(GFP_KERNEL | __GFP_ZERO | gfp_extra_flags);
kernel/bpf/core.c
259
gfp_t gfp_flags = bpf_memcg_flags(GFP_KERNEL | __GFP_ZERO | gfp_extra_flags);
kernel/bpf/cpumap.c
446
rcpu = bpf_map_kmalloc_node(map, sizeof(*rcpu), gfp | __GFP_ZERO, numa);
kernel/bpf/fixups.c
245
GFP_KERNEL_ACCOUNT | __GFP_ZERO);
kernel/bpf/local_storage.c
168
__GFP_ZERO | GFP_NOWAIT,
kernel/bpf/local_storage.c
506
const gfp_t gfp = __GFP_ZERO | GFP_USER;
kernel/bpf/memalloc.c
157
return kmalloc_node(c->unit_size, flags | __GFP_ZERO, node);
kernel/bpf/ringbuf.c
96
__GFP_NOWARN | __GFP_ZERO;
kernel/bpf/stream.c
33
elem = kmalloc_nolock(alloc_size, __GFP_ZERO, -1);
kernel/bpf/syscall.c
384
gfp_t gfp = bpf_memcg_flags(__GFP_NOWARN | __GFP_ZERO);
kernel/bpf/syscall.c
611
GFP_KERNEL | __GFP_ZERO | __GFP_ACCOUNT
kernel/cpu.c
1986
if (!alloc_cpumask_var(&frozen_cpus, GFP_KERNEL|__GFP_ZERO))
kernel/dma/direct.c
190
page = __dma_direct_alloc_pages(dev, size, gfp & ~__GFP_ZERO, true);
kernel/dma/direct.c
254
page = __dma_direct_alloc_pages(dev, size, gfp & ~__GFP_ZERO, true);
kernel/dma/swiotlb.c
499
__get_free_pages(GFP_KERNEL | __GFP_ZERO, area_order);
kernel/dma/swiotlb.c
503
mem->slots = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
kernel/events/core.c
12802
GFP_KERNEL | __GFP_ZERO,
kernel/events/core.c
13289
kmem_cache_alloc_node(perf_event_cache, GFP_KERNEL | __GFP_ZERO, node);
kernel/events/ring_buffer.c
614
#define PERF_AUX_GFP (GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN | __GFP_NORETRY)
kernel/events/ring_buffer.c
822
page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0);
kernel/events/uprobes.c
1767
area->page = alloc_page(GFP_HIGHUSER | __GFP_ZERO);
kernel/fork.c
207
#define GFP_VMAP_STACK (GFP_KERNEL | __GFP_ZERO)
kernel/kexec_core.c
285
pages = alloc_pages(gfp_mask & ~__GFP_ZERO, order);
kernel/kexec_core.c
298
if (gfp_mask & __GFP_ZERO)
kernel/liveupdate/kexec_handover.c
1200
folio = folio_alloc(GFP_KERNEL | __GFP_ZERO, order);
kernel/profile.c
106
GFP_KERNEL|__GFP_ZERO|__GFP_NOWARN);
kernel/relay.c
121
buf->page_array[i] = alloc_page(GFP_KERNEL | __GFP_ZERO);
kernel/sched/core.c
9340
tg = kmem_cache_alloc(task_group_cache, GFP_KERNEL | __GFP_ZERO);
kernel/time/namespace_vdso.c
150
ns->vvar_page = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);
kernel/trace/ftrace.c
3863
pg->records = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, order);
kernel/trace/ring_buffer.c
402
mflags = GFP_KERNEL | __GFP_RETRY_MAYFAIL | __GFP_COMP | __GFP_ZERO;
kernel/trace/ring_buffer.c
7270
page = alloc_page(GFP_USER | __GFP_ZERO);
kernel/trace/trace_events.c
51
#define GFP_TRACE (GFP_KERNEL | __GFP_ZERO)
kernel/vmcore_info.c
146
vmcoreinfo_data = (unsigned char *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, order);
kernel/vmcore_info.c
153
GFP_KERNEL | __GFP_ZERO);
lib/alloc_tag.c
674
GFP_KERNEL | __GFP_ZERO);
lib/bitmap.c
741
return bitmap_alloc(nbits, flags | __GFP_ZERO);
lib/bitmap.c
754
return bitmap_alloc_node(nbits, flags | __GFP_ZERO, node);
lib/bitmap.c
792
return devm_bitmap_alloc(dev, nbits, flags | __GFP_ZERO);
lib/objpool.c
142
pool->gfp = gfp & ~__GFP_ZERO;
lib/rhashtable.c
134
kmalloc_noprof(PAGE_SIZE, GFP_ATOMIC|__GFP_ZERO));
lib/rhashtable.c
162
kmalloc_noprof(size, gfp|__GFP_ZERO));
lib/rhashtable.c
188
1, gfp|__GFP_ZERO, NUMA_NO_NODE));
lib/test_hmm.c
1561
page = alloc_page(GFP_KERNEL | __GFP_ZERO);
lib/test_meminit.c
223
gfp_t alloc_mask = GFP_KERNEL | (want_zero ? __GFP_ZERO : 0);
lib/test_vmalloc.c
102
ptr = __vmalloc_node(size, align, GFP_KERNEL | __GFP_ZERO, 0,
lib/test_vmalloc.c
125
ptr = __vmalloc_node(PAGE_SIZE, align, GFP_KERNEL|__GFP_ZERO, 0,
lib/test_vmalloc.c
143
GFP_KERNEL | __GFP_ZERO, 0,
lib/tests/slub_kunit.c
268
p = krealloc(p, 40, GFP_KERNEL | __GFP_ZERO);
lib/tests/slub_kunit.c
273
p = krealloc(p, 56, GFP_KERNEL | __GFP_ZERO);
lib/tests/slub_kunit.c
284
p = krealloc(p, 112, GFP_KERNEL | __GFP_ZERO);
mm/dmapool.c
425
page = pool_alloc_page(pool, mem_flags & (~__GFP_ZERO));
mm/huge_memory.c
228
zero_folio = folio_alloc((GFP_TRANSHUGE | __GFP_ZERO | __GFP_ZEROTAGS) &
mm/hugetlb_vmemmap.c
506
tail = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0);
mm/kasan/kasan_test_c.c
1630
ptr = kmalloc(size, GFP_KERNEL | __GFP_ZERO);
mm/kasan/kasan_test_c.c
1656
ptr = kmalloc(size, GFP_KERNEL | __GFP_ZERO);
mm/kasan/kasan_test_c.c
1679
ptr = kmalloc(size, GFP_KERNEL | __GFP_ZERO);
mm/kasan/kasan_test_c.c
1683
src = kmalloc(KASAN_GRANULE_SIZE, GFP_KERNEL | __GFP_ZERO);
mm/kasan/shadow.c
364
data.pages = (struct page **)__get_free_page(gfp_mask | __GFP_ZERO);
mm/kfence/kfence_test.c
639
buf2 = test_alloc(test, size, GFP_KERNEL | __GFP_ZERO, ALLOCATE_ANY);
mm/kmsan/hooks.c
100
if (flags & __GFP_ZERO)
mm/kmsan/hooks.c
153
gfp_t gfp_mask = GFP_KERNEL | __GFP_ZERO;
mm/kmsan/hooks.c
62
if (flags & __GFP_ZERO)
mm/kmsan/kmsan_test.c
401
page = alloc_pages(GFP_KERNEL | __GFP_ZERO, order);
mm/kmsan/shadow.c
172
bool initialized = (flags & __GFP_ZERO) || !kmsan_enabled;
mm/memcontrol.c
3759
pn = kmem_cache_alloc_node(memcg_pn_cachep, GFP_KERNEL | __GFP_ZERO,
mm/mempool.c
319
pool = kmalloc_node_noprof(sizeof(*pool), gfp_mask | __GFP_ZERO, node_id);
mm/mempool.c
557
VM_WARN_ON_ONCE(gfp_mask & __GFP_ZERO);
mm/mm_init.c
2451
gfp_flags = (flags & HASH_ZERO) ? GFP_ATOMIC | __GFP_ZERO : GFP_ATOMIC;
mm/nommu.c
162
return __vmalloc_user_flags(size, GFP_KERNEL | __GFP_ZERO);
mm/nommu.c
236
return __vmalloc_noprof(size, GFP_KERNEL | __GFP_ZERO);
mm/page_alloc.c
5296
return get_free_pages_noprof(gfp_mask | __GFP_ZERO, 0);
mm/page_alloc.c
6908
__GFP_ZERO | __GFP_ZEROTAGS | __GFP_SKIP_ZERO |
mm/page_alloc.c
7758
gfp_t alloc_gfp = __GFP_NOWARN | __GFP_ZERO | __GFP_NOMEMALLOC | __GFP_COMP
mm/page_ext.c
272
gfp_t flags = GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN;
mm/percpu.c
512
return __vmalloc(size, gfp | __GFP_ZERO);
mm/secretmem.c
69
folio = folio_alloc(gfp | __GFP_ZERO, 0);
mm/slab.h
683
return flags & __GFP_ZERO;
mm/slab.h
686
return flags & __GFP_ZERO;
mm/slab_common.c
447
b = kmem_cache_alloc(kmem_buckets_cache, GFP_KERNEL|__GFP_ZERO);
mm/slub.c
2168
vec = kmalloc_nolock(sz, __GFP_ZERO | __GFP_NO_OBJ_EXT,
mm/slub.c
2171
vec = kmalloc_node(sz, gfp | __GFP_ZERO, slab_nid(slab));
mm/slub.c
3523
WARN_ON_ONCE(s->ctor && (flags & __GFP_ZERO));
mm/slub.c
5329
VM_WARN_ON_ONCE(gfp_flags & ~(__GFP_ACCOUNT | __GFP_ZERO |
mm/swapfile.c
3550
GFP_KERNEL | __GFP_ZERO);
mm/swapfile.c
421
folio = folio_alloc(gfp | __GFP_ZERO, 0);
mm/util.c
685
return __vmalloc_array_noprof(n, size, flags | __GFP_ZERO);
mm/util.c
696
return __vmalloc_array_noprof(n, size, GFP_KERNEL | __GFP_ZERO);
mm/vmalloc.c
3833
const gfp_t nested_gfp = (gfp_mask & GFP_RECLAIM_MASK) | __GFP_ZERO;
mm/vmalloc.c
3939
__GFP_NOFAIL | __GFP_ZERO |\
mm/vmalloc.c
4202
return __vmalloc_node_noprof(size, 1, GFP_KERNEL | __GFP_ZERO, NUMA_NO_NODE,
mm/vmalloc.c
4219
GFP_KERNEL | __GFP_ZERO, PAGE_KERNEL,
mm/vmalloc.c
4258
return __vmalloc_node_noprof(size, 1, GFP_KERNEL | __GFP_ZERO, node,
mm/vmalloc.c
4412
GFP_VMALLOC32 | __GFP_ZERO, PAGE_KERNEL,
net/9p/trans_xen.c
350
GFP_KERNEL | __GFP_ZERO);
net/bpf/test_run.c
245
gfp_t gfp = __GFP_ZERO | GFP_ATOMIC;
net/core/dst_cache.c
168
gfp | __GFP_ZERO);
net/core/pktgen.c
2853
pkt_dev->page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0);
net/core/skbuff.c
364
GFP_ATOMIC | __GFP_ZERO | __GFP_NOWARN,
net/core/sock.c
2241
sk = kmem_cache_alloc(slab, priority & ~__GFP_ZERO);
net/core/sock.c
2303
sk = sk_prot_alloc(prot, priority | __GFP_ZERO, family);
net/ipv4/nexthop.c
171
info->nh_res_table = __vmalloc(size, GFP_KERNEL | __GFP_ZERO |
net/ipv4/nexthop.c
568
res_table = __vmalloc(size, GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN);
net/ipv4/tcp_ipv4.c
1373
key = sock_kmalloc(sk, sizeof(*key), gfp | __GFP_ZERO);
net/ipv6/seg6_local.c
183
((__gfp) | __GFP_ZERO))
net/key/af_key.c
1719
supp_skb = compose_sadb_supported(hdr, GFP_KERNEL | __GFP_ZERO);
net/openvswitch/flow_table.c
86
GFP_KERNEL | __GFP_ZERO,
net/packet/af_packet.c
4365
__GFP_ZERO | __GFP_NOWARN | __GFP_NORETRY;
net/packet/af_packet.c
4504
GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO);
net/smc/smc_core.c
2333
__GFP_NORETRY | __GFP_ZERO,
net/xfrm/xfrm_hash.c
26
__get_free_pages(GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO,
rust/bindings/bindings_helper.h
116
const gfp_t RUST_CONST_HELPER___GFP_ZERO = __GFP_ZERO;
samples/vfio-mdev/mbochs.c
732
alloc_pages(GFP_HIGHUSER | __GFP_ZERO, 0);
security/selinux/status.c
49
selinux_state.status_page = alloc_page(GFP_KERNEL|__GFP_ZERO);
security/tomoyo/common.c
2166
buffer = kmalloc(len, GFP_NOFS | __GFP_ZERO);
sound/soc/meson/meson-card-utils.c
51
GFP_KERNEL | __GFP_ZERO);
sound/soc/meson/meson-card-utils.c
57
GFP_KERNEL | __GFP_ZERO);
sound/usb/usx2y/usb_stream.c
196
GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN);
sound/usb/usx2y/usb_stream.c
215
GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN);
tools/include/linux/slab.h
140
return kmalloc(size, gfp | __GFP_ZERO);
tools/lib/slab.c
25
if (gfp & __GFP_ZERO)
tools/lib/slab.c
51
if (gfp & __GFP_ZERO)
tools/testing/shared/linux.c
203
else if (gfp & __GFP_ZERO)
tools/testing/shared/linux.c
94
else if (gfp & __GFP_ZERO)
tools/virtio/ringtest/ptr_ring.c
33
if (gfp & __GFP_ZERO)
tools/virtio/ringtest/ptr_ring.c
40
return kmalloc(size, flags | __GFP_ZERO);
tools/virtio/ringtest/ptr_ring.c
52
return kmalloc_array(n, size, flags | __GFP_ZERO);
virt/kvm/coalesced_mmio.c
99
page = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);
virt/kvm/kvm_main.c
4188
page = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);