arch/alpha/kernel/pci_iommu.c
406
dma_addr_t *dma_addrp, gfp_t gfp,
arch/alpha/kernel/pci_iommu.c
413
gfp &= ~GFP_DMA;
arch/alpha/kernel/pci_iommu.c
416
cpu_addr = (void *)__get_free_pages(gfp | __GFP_ZERO, order);
arch/alpha/kernel/pci_iommu.c
430
if (alpha_mv.mv_pci_tbi || (gfp & GFP_DMA))
arch/alpha/kernel/pci_iommu.c
434
gfp |= GFP_DMA;
arch/arm/mm/dma-mapping.c
1038
static void *__iommu_alloc_simple(struct device *dev, size_t size, gfp_t gfp,
arch/arm/mm/dma-mapping.c
1046
addr = __alloc_simple_buffer(dev, size, gfp, &page);
arch/arm/mm/dma-mapping.c
1074
dma_addr_t *handle, gfp_t gfp, unsigned long attrs)
arch/arm/mm/dma-mapping.c
1084
if (coherent_flag == COHERENT || !gfpflags_allow_blocking(gfp))
arch/arm/mm/dma-mapping.c
1085
return __iommu_alloc_simple(dev, size, gfp, handle,
arch/arm/mm/dma-mapping.c
1088
pages = __iommu_alloc_buffer(dev, size, gfp, attrs, coherent_flag);
arch/arm/mm/dma-mapping.c
142
gfp_t gfp, int coherent_flag)
arch/arm/mm/dma-mapping.c
147
page = alloc_pages(gfp, order);
arch/arm/mm/dma-mapping.c
1738
gfp_t gfp, unsigned long attrs)
arch/arm/mm/dma-mapping.c
1740
return __dma_alloc(dev, size, dma_handle, gfp,
arch/arm/mm/dma-mapping.c
179
int coherent_flag, gfp_t gfp);
arch/arm/mm/dma-mapping.c
181
static void *__alloc_remap_buffer(struct device *dev, size_t size, gfp_t gfp,
arch/arm/mm/dma-mapping.c
203
gfp_t gfp = GFP_KERNEL | GFP_DMA;
arch/arm/mm/dma-mapping.c
219
ptr = __alloc_remap_buffer(NULL, atomic_pool_size, gfp, prot,
arch/arm/mm/dma-mapping.c
328
static void *__alloc_remap_buffer(struct device *dev, size_t size, gfp_t gfp,
arch/arm/mm/dma-mapping.c
338
page = __dma_alloc_buffer(dev, size, gfp, NORMAL);
arch/arm/mm/dma-mapping.c
394
int coherent_flag, gfp_t gfp)
arch/arm/mm/dma-mapping.c
401
page = dma_alloc_from_contiguous(dev, count, order, gfp & __GFP_NOWARN);
arch/arm/mm/dma-mapping.c
44
gfp_t gfp;
arch/arm/mm/dma-mapping.c
446
static void *__alloc_simple_buffer(struct device *dev, size_t size, gfp_t gfp,
arch/arm/mm/dma-mapping.c
451
page = __dma_alloc_buffer(dev, size, gfp, COHERENT);
arch/arm/mm/dma-mapping.c
462
return __alloc_simple_buffer(args->dev, args->size, args->gfp,
arch/arm/mm/dma-mapping.c
482
args->gfp);
arch/arm/mm/dma-mapping.c
515
return __alloc_remap_buffer(args->dev, args->size, args->gfp,
arch/arm/mm/dma-mapping.c
534
gfp_t gfp, pgprot_t prot, bool is_coherent,
arch/arm/mm/dma-mapping.c
545
.gfp = gfp,
arch/arm/mm/dma-mapping.c
562
gfp & ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM));
arch/arm/mm/dma-mapping.c
567
gfp |= GFP_DMA;
arch/arm/mm/dma-mapping.c
569
args.gfp = gfp;
arch/arm/mm/dma-mapping.c
572
allowblock = gfpflags_allow_blocking(gfp);
arch/arm/mm/dma-mapping.c
851
gfp_t gfp, unsigned long attrs,
arch/arm/mm/dma-mapping.c
870
gfp & __GFP_NOWARN);
arch/arm/mm/dma-mapping.c
889
gfp |= __GFP_NOWARN | __GFP_HIGHMEM;
arch/arm/mm/dma-mapping.c
904
pages[i] = alloc_pages(gfp | __GFP_NORETRY, order);
arch/arm/mm/dma-mapping.c
912
pages[i] = alloc_pages(gfp, 0);
arch/arm64/mm/mmu.c
530
static phys_addr_t __pgd_pgtable_alloc(struct mm_struct *mm, gfp_t gfp,
arch/arm64/mm/mmu.c
534
struct ptdesc *ptdesc = pagetable_alloc(gfp & ~__GFP_ZERO, 0);
arch/arm64/mm/mmu.c
561
pgd_pgtable_alloc_init_mm_gfp(enum pgtable_type pgtable_type, gfp_t gfp)
arch/arm64/mm/mmu.c
563
return __pgd_pgtable_alloc(&init_mm, gfp, pgtable_type);
arch/arm64/mm/mmu.c
587
static int split_pmd(pmd_t *pmdp, pmd_t pmd, gfp_t gfp, bool to_cont)
arch/arm64/mm/mmu.c
596
pte_phys = pgd_pgtable_alloc_init_mm_gfp(TABLE_PTE, gfp);
arch/arm64/mm/mmu.c
631
static int split_pud(pud_t *pudp, pud_t pud, gfp_t gfp, bool to_cont)
arch/arm64/mm/mmu.c
641
pmd_phys = pgd_pgtable_alloc_init_mm_gfp(TABLE_PMD, gfp);
arch/arm64/mm/mmu.c
833
gfp_t gfp = *(gfp_t *)walk->private;
arch/arm64/mm/mmu.c
838
ret = split_pud(pudp, pud, gfp, false);
arch/arm64/mm/mmu.c
846
gfp_t gfp = *(gfp_t *)walk->private;
arch/arm64/mm/mmu.c
853
ret = split_pmd(pmdp, pmd, gfp, false);
arch/arm64/mm/mmu.c
882
static int range_split_to_ptes(unsigned long start, unsigned long end, gfp_t gfp)
arch/arm64/mm/mmu.c
888
&split_to_ptes_ops, NULL, &gfp);
arch/arm64/mm/pgd.c
33
gfp_t gfp = GFP_PGTABLE_USER;
arch/arm64/mm/pgd.c
38
return kmem_cache_alloc(pgd_cache, gfp);
arch/mips/jazz/jazzdma.c
492
dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
arch/mips/jazz/jazzdma.c
498
gfp |= __GFP_NOWARN;
arch/mips/jazz/jazzdma.c
501
page = alloc_pages(gfp, get_order(size));
arch/parisc/include/asm/pgalloc.h
35
gfp_t gfp = GFP_PGTABLE_USER;
arch/parisc/include/asm/pgalloc.h
38
gfp = GFP_PGTABLE_KERNEL;
arch/parisc/include/asm/pgalloc.h
39
ptdesc = pagetable_alloc(gfp, PMD_TABLE_ORDER);
arch/parisc/kernel/pci-dma.c
402
dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
arch/parisc/kernel/pci-dma.c
414
paddr = __get_free_pages(gfp | __GFP_ZERO, order);
arch/powerpc/include/asm/pgalloc.h
11
return gfp;
arch/powerpc/include/asm/pgalloc.h
12
return gfp | __GFP_ACCOUNT;
arch/powerpc/include/asm/pgalloc.h
15
static inline gfp_t pgtable_gfp_flags(struct mm_struct *mm, gfp_t gfp)
arch/powerpc/include/asm/pgalloc.h
17
return gfp | __GFP_ACCOUNT;
arch/powerpc/include/asm/pgalloc.h
8
static inline gfp_t pgtable_gfp_flags(struct mm_struct *mm, gfp_t gfp)
arch/powerpc/mm/book3s64/pgtable.c
418
gfp_t gfp = GFP_KERNEL_ACCOUNT | __GFP_ZERO;
arch/powerpc/mm/book3s64/pgtable.c
421
gfp &= ~__GFP_ACCOUNT;
arch/powerpc/mm/book3s64/pgtable.c
422
ptdesc = pagetable_alloc(gfp, 0);
arch/powerpc/mm/pgtable-frag.c
59
gfp_t gfp = PGALLOC_GFP;
arch/powerpc/mm/pgtable-frag.c
62
gfp |= __GFP_ACCOUNT;
arch/powerpc/mm/pgtable-frag.c
64
ptdesc = pagetable_alloc(gfp, 0);
arch/powerpc/perf/hv-24x7.c
423
static char *memdup_to_str(char *maybe_str, int max_len, gfp_t gfp)
arch/powerpc/perf/hv-24x7.c
425
return kasprintf(gfp, "%.*s", max_len, maybe_str);
arch/s390/mm/pgalloc.c
119
gfp_t gfp = GFP_KERNEL_ACCOUNT;
arch/s390/mm/pgalloc.c
124
gfp &= ~__GFP_ACCOUNT;
arch/s390/mm/pgalloc.c
125
ptdesc = pagetable_alloc_noprof(gfp, 0);
arch/s390/mm/pgalloc.c
19
gfp_t gfp = GFP_KERNEL_ACCOUNT;
arch/s390/mm/pgalloc.c
24
gfp &= ~__GFP_ACCOUNT;
arch/s390/mm/pgalloc.c
25
ptdesc = pagetable_alloc_noprof(gfp, CRST_ALLOC_ORDER);
arch/sparc/kernel/iommu.c
198
dma_addr_t *dma_addrp, gfp_t gfp,
arch/sparc/kernel/iommu.c
214
page = alloc_pages_node(nid, gfp, order);
arch/sparc/kernel/pci_sun4v.c
182
dma_addr_t *dma_addrp, gfp_t gfp,
arch/sparc/kernel/pci_sun4v.c
206
page = alloc_pages_node(nid, gfp, order);
arch/sparc/mm/io-unit.c
219
dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
arch/sparc/mm/io-unit.c
231
va = __get_free_pages(gfp | __GFP_ZERO, get_order(len));
arch/sparc/mm/iommu.c
316
dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
arch/sparc/mm/iommu.c
329
va = __get_free_pages(gfp | __GFP_ZERO, get_order(len));
arch/x86/events/intel/pt.c
1246
unsigned long nr_pages, gfp_t gfp)
arch/x86/events/intel/pt.c
1251
topa = topa_alloc(cpu, gfp);
arch/x86/events/intel/pt.c
1258
err = topa_insert_pages(buf, cpu, gfp);
arch/x86/events/intel/pt.c
680
static struct topa *topa_alloc(int cpu, gfp_t gfp)
arch/x86/events/intel/pt.c
686
p = alloc_pages_node(node, gfp | __GFP_ZERO, 0);
arch/x86/events/intel/pt.c
770
static int topa_insert_pages(struct pt_buffer *buf, int cpu, gfp_t gfp)
arch/x86/events/intel/pt.c
781
topa = topa_alloc(cpu, gfp);
arch/x86/include/asm/kexec.h
197
gfp_t gfp);
arch/x86/kernel/machine_kexec_64.c
702
int arch_kexec_post_alloc_pages(void *vaddr, unsigned int pages, gfp_t gfp)
arch/x86/kvm/svm/sev.c
4811
struct page *snp_safe_alloc_page_node(int node, gfp_t gfp)
arch/x86/kvm/svm/sev.c
4817
return alloc_pages_node(node, gfp | __GFP_ZERO, 0);
arch/x86/kvm/svm/sev.c
4828
p = alloc_pages_node(node, gfp | __GFP_ZERO, 1);
arch/x86/kvm/svm/svm.h
881
struct page *snp_safe_alloc_page_node(int node, gfp_t gfp);
arch/x86/kvm/svm/svm.h
903
static inline struct page *snp_safe_alloc_page_node(int node, gfp_t gfp)
arch/x86/kvm/svm/svm.h
905
return alloc_pages_node(node, gfp | __GFP_ZERO, 0);
arch/x86/mm/pgtable.c
180
gfp_t gfp = GFP_PGTABLE_USER;
arch/x86/mm/pgtable.c
183
gfp &= ~__GFP_ACCOUNT;
arch/x86/mm/pgtable.c
184
gfp &= ~__GFP_HIGHMEM;
arch/x86/mm/pgtable.c
188
struct ptdesc *ptdesc = pagetable_alloc(gfp, 0);
arch/x86/mm/pti.c
181
gfp_t gfp = (GFP_KERNEL | __GFP_NOTRACK | __GFP_ZERO);
arch/x86/mm/pti.c
189
unsigned long new_p4d_page = __get_free_page(gfp);
arch/x86/mm/pti.c
208
gfp_t gfp = (GFP_KERNEL | __GFP_NOTRACK | __GFP_ZERO);
arch/x86/mm/pti.c
218
unsigned long new_pud_page = __get_free_page(gfp);
arch/x86/mm/pti.c
232
unsigned long new_pmd_page = __get_free_page(gfp);
arch/x86/mm/pti.c
253
gfp_t gfp = (GFP_KERNEL | __GFP_NOTRACK | __GFP_ZERO);
arch/x86/mm/pti.c
273
unsigned long new_pte_page = __get_free_page(gfp);
block/bfq-cgroup.c
19
static int bfq_stat_init(struct bfq_stat *stat, gfp_t gfp)
block/bfq-cgroup.c
23
ret = percpu_counter_init(&stat->cpu_cnt, 0, gfp);
block/bfq-cgroup.c
455
static int bfqg_stats_init(struct bfqg_stats *stats, gfp_t gfp)
block/bfq-cgroup.c
457
if (blkg_rwstat_init(&stats->bytes, gfp) ||
block/bfq-cgroup.c
458
blkg_rwstat_init(&stats->ios, gfp))
block/bfq-cgroup.c
462
if (blkg_rwstat_init(&stats->merged, gfp) ||
block/bfq-cgroup.c
463
blkg_rwstat_init(&stats->service_time, gfp) ||
block/bfq-cgroup.c
464
blkg_rwstat_init(&stats->wait_time, gfp) ||
block/bfq-cgroup.c
465
blkg_rwstat_init(&stats->queued, gfp) ||
block/bfq-cgroup.c
466
bfq_stat_init(&stats->time, gfp) ||
block/bfq-cgroup.c
467
bfq_stat_init(&stats->avg_queue_size_sum, gfp) ||
block/bfq-cgroup.c
468
bfq_stat_init(&stats->avg_queue_size_samples, gfp) ||
block/bfq-cgroup.c
469
bfq_stat_init(&stats->dequeue, gfp) ||
block/bfq-cgroup.c
470
bfq_stat_init(&stats->group_wait_time, gfp) ||
block/bfq-cgroup.c
471
bfq_stat_init(&stats->idle_time, gfp) ||
block/bfq-cgroup.c
472
bfq_stat_init(&stats->empty_time, gfp))
block/bfq-cgroup.c
493
static struct blkcg_policy_data *bfq_cpd_alloc(gfp_t gfp)
block/bfq-cgroup.c
497
bgd = kzalloc_obj(*bgd, gfp);
block/bfq-cgroup.c
511
struct blkcg *blkcg, gfp_t gfp)
block/bfq-cgroup.c
515
bfqg = kzalloc_node(sizeof(*bfqg), gfp, disk->node_id);
block/bfq-cgroup.c
519
if (bfqg_stats_init(&bfqg->stats, gfp)) {
block/bio-integrity-auto.c
100
gfp_t gfp = GFP_NOIO;
block/bio-integrity-auto.c
130
gfp |= __GFP_ZERO;
block/bio-integrity-auto.c
132
gfp |= __GFP_ZERO;
block/bio-integrity-auto.c
145
bio_integrity_alloc_buf(bio, gfp & __GFP_ZERO);
block/bio-integrity.c
24
gfp_t gfp = GFP_NOIO | (zero_buffer ? __GFP_ZERO : 0);
block/bio-integrity.c
27
buf = kmalloc(len, (gfp & ~__GFP_DIRECT_RECLAIM) |
block/bio.c
1303
static struct folio *folio_alloc_greedy(gfp_t gfp, size_t *size)
block/bio.c
1308
folio = folio_alloc(gfp | __GFP_NORETRY, get_order(*size));
block/bio.c
1314
return folio_alloc(gfp, get_order(*size));
block/bio.c
176
static inline gfp_t bvec_alloc_gfp(gfp_t gfp)
block/bio.c
178
return (gfp & ~(__GFP_DIRECT_RECLAIM | __GFP_IO)) |
block/bio.c
1811
gfp_t gfp, struct bio_set *bs)
block/bio.c
1828
split = bio_alloc_clone(bio->bi_bdev, bio, gfp, bs);
block/bio.c
410
unsigned int nr_pages, blk_opf_t opf, gfp_t gfp)
block/bio.c
412
return bio_chain_and_submit(bio, bio_alloc(bdev, nr_pages, opf, gfp));
block/bio.c
487
unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp,
block/bio.c
870
static int __bio_clone(struct bio *bio, struct bio *bio_src, gfp_t gfp)
block/bio.c
885
if (bio_crypt_clone(bio, bio_src, gfp) < 0)
block/bio.c
888
bio_integrity_clone(bio, bio_src, gfp) < 0)
block/bio.c
906
gfp_t gfp, struct bio_set *bs)
block/bio.c
910
bio = bio_alloc_bioset(bdev, 0, bio_src->bi_opf, gfp, bs);
block/bio.c
914
if (__bio_clone(bio, bio_src, gfp) < 0) {
block/bio.c
937
struct bio *bio_src, gfp_t gfp)
block/bio.c
942
ret = __bio_clone(bio, bio_src, gfp);
block/blk-cgroup-rwstat.c
12
ret = percpu_counter_init_many(rwstat->cpu_cnt, 0, gfp, BLKG_RWSTAT_NR);
block/blk-cgroup-rwstat.c
8
int blkg_rwstat_init(struct blkg_rwstat *rwstat, gfp_t gfp)
block/blk-cgroup-rwstat.h
42
int blkg_rwstat_init(struct blkg_rwstat *rwstat, gfp_t gfp);
block/blk-iocost.c
2945
static struct blkcg_policy_data *ioc_cpd_alloc(gfp_t gfp)
block/blk-iocost.c
2949
iocc = kzalloc_obj(struct ioc_cgrp, gfp);
block/blk-iocost.c
2963
struct blkcg *blkcg, gfp_t gfp)
block/blk-iocost.c
2968
iocg = kzalloc_node(struct_size(iocg, ancestors, levels), gfp,
block/blk-iocost.c
2973
iocg->pcpu_stat = alloc_percpu_gfp(struct iocg_pcpu_stat, gfp);
block/blk-iolatency.c
966
struct blkcg *blkcg, gfp_t gfp)
block/blk-iolatency.c
970
iolat = kzalloc_node(sizeof(*iolat), gfp, disk->node_id);
block/blk-iolatency.c
974
__alignof__(struct latency_stat), gfp);
block/blk-ioprio.c
102
blkcg = kzalloc_obj(*blkcg, gfp);
block/blk-ioprio.c
98
static struct blkcg_policy_data *ioprio_alloc_cpd(gfp_t gfp)
block/blk-lib.c
147
sector_t nr_sects, gfp_t gfp, unsigned flags)
block/blk-lib.c
155
__blkdev_issue_write_zeroes(bdev, sector, nr_sects, gfp, &bio,
block/blk-lib.c
227
sector_t nr_sects, gfp_t gfp, unsigned flags)
block/blk-lib.c
237
__blkdev_issue_zero_pages(bdev, sector, nr_sects, gfp, &bio, flags);
block/blk-lib.c
329
sector_t nr_sects, gfp_t gfp)
block/blk-lib.c
353
bio = blk_next_bio(bio, bdev, 0, REQ_OP_SECURE_ERASE, gfp);
block/blk-mq-sched.c
510
gfp_t gfp = GFP_NOIO | __GFP_ZERO | __GFP_NOWARN | __GFP_NORETRY;
block/blk-mq-sched.c
517
et = kmalloc_flex(*et, tags, nr_tags, gfp);
block/blk-mq.c
3996
gfp_t gfp = GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY;
block/blk-mq.c
3998
hctx->fq = blk_alloc_flush_queue(hctx->numa_node, set->cmd_size, gfp);
block/blk-mq.c
4031
gfp_t gfp = GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY;
block/blk-mq.c
4033
hctx = kzalloc_node(sizeof(struct blk_mq_hw_ctx), gfp, node);
block/blk-mq.c
4037
if (!zalloc_cpumask_var_node(&hctx->cpumask, gfp, node))
block/blk-mq.c
4060
gfp, node);
block/blk-mq.c
4065
gfp, node, false, false))
block/blk-throttle.c
262
struct blkcg *blkcg, gfp_t gfp)
block/blk-throttle.c
267
tg = kzalloc_node(sizeof(*tg), gfp, disk->node_id);
block/blk-throttle.c
271
if (blkg_rwstat_init(&tg->stat_bytes, gfp))
block/blk-throttle.c
274
if (blkg_rwstat_init(&tg->stat_ios, gfp))
block/ioctl.c
900
gfp_t gfp = nowait ? GFP_NOWAIT : GFP_KERNEL;
block/ioctl.c
922
bio = blk_alloc_discard_bio(bdev, §or, &nr_sects, gfp);
crypto/acompress.c
564
size_t total, gfp_t gfp)
crypto/acompress.c
568
nreq = container_of(crypto_request_clone(&req->base, total, gfp),
crypto/api.c
409
u32 mask, gfp_t gfp)
crypto/api.c
416
tfm = kzalloc(tfm_size, gfp);
crypto/api.c
505
gfp_t gfp)
crypto/api.c
515
mem = kzalloc_node(total, gfp, node);
crypto/api.c
726
struct crypto_async_request *req, size_t total, gfp_t gfp)
crypto/api.c
731
nreq = kmemdup(req, total, gfp);
crypto/internal.h
124
u32 mask, gfp_t gfp);
crypto/krb5/internal.h
153
gfp_t gfp);
crypto/krb5/internal.h
156
gfp_t gfp);
crypto/krb5/internal.h
162
u32 usage, struct krb5_buffer *key, gfp_t gfp);
crypto/krb5/internal.h
164
u32 usage, struct krb5_buffer *key, gfp_t gfp);
crypto/krb5/internal.h
166
u32 usage, struct krb5_buffer *key, gfp_t gfp);
crypto/krb5/internal.h
179
gfp_t gfp);
crypto/krb5/internal.h
184
gfp_t gfp);
crypto/krb5/internal.h
189
gfp_t gfp);
crypto/krb5/internal.h
193
gfp_t gfp);
crypto/krb5/internal.h
22
gfp_t gfp);
crypto/krb5/internal.h
29
gfp_t gfp);
crypto/krb5/internal.h
36
gfp_t gfp);
crypto/krb5/internal.h
43
gfp_t gfp);
crypto/krb5/internal.h
50
gfp_t gfp);
crypto/krb5/internal.h
57
gfp_t gfp);
crypto/krb5/internal.h
64
gfp_t gfp);
crypto/krb5/internal.h
70
gfp_t gfp);
crypto/krb5/krb5_api.c
163
gfp_t gfp)
crypto/krb5/krb5_api.c
208
u32 usage, gfp_t gfp)
crypto/krb5/krb5_api.c
214
ret = krb5->profile->derive_encrypt_keys(krb5, TK, usage, &keys, gfp);
crypto/krb5/krb5_api.c
218
ci = krb5_prepare_encryption(krb5, &keys, gfp);
crypto/krb5/krb5_api.c
237
gfp_t gfp)
crypto/krb5/krb5_api.c
276
u32 usage, gfp_t gfp)
crypto/krb5/krb5_api.c
282
ret = krb5->profile->derive_checksum_key(krb5, TK, usage, &keys, gfp);
crypto/krb5/krb5_api.c
288
ci = krb5_prepare_checksum(krb5, &keys, gfp);
crypto/krb5/krb5_kdf.c
111
u32 usage, struct krb5_buffer *key, gfp_t gfp)
crypto/krb5/krb5_kdf.c
120
return krb5->profile->calc_Ke(krb5, TK, &usage_constant, key, gfp);
crypto/krb5/krb5_kdf.c
135
u32 usage, struct krb5_buffer *key, gfp_t gfp)
crypto/krb5/krb5_kdf.c
144
return krb5->profile->calc_Ki(krb5, TK, &usage_constant, key, gfp);
crypto/krb5/krb5_kdf.c
37
gfp_t gfp)
crypto/krb5/krb5_kdf.c
47
buffer = kzalloc(round16(L + Tn.len) + round16(n_S.len), gfp);
crypto/krb5/krb5_kdf.c
58
ret = krb5->profile->calc_PRF(krb5, K, &n_S, &Tn, gfp);
crypto/krb5/krb5_kdf.c
87
u32 usage, struct krb5_buffer *key, gfp_t gfp)
crypto/krb5/krb5_kdf.c
96
return krb5->profile->calc_Kc(krb5, TK, &usage_constant, key, gfp);
crypto/krb5/rfc3961_simplified.c
141
gfp_t gfp)
crypto/krb5/rfc3961_simplified.c
154
desc = kzalloc(desc_size, gfp);
crypto/krb5/rfc3961_simplified.c
159
digest->data = kzalloc(digest->len, gfp);
crypto/krb5/rfc3961_simplified.c
268
gfp_t gfp)
crypto/krb5/rfc3961_simplified.c
294
inblock.data = kzalloc(blocksize * 2 + keybytes, gfp);
crypto/krb5/rfc3961_simplified.c
351
gfp_t gfp)
crypto/krb5/rfc3961_simplified.c
389
gfp_t gfp)
crypto/krb5/rfc3961_simplified.c
403
buffer = kzalloc(round16(tmp1.len) + round16(derived_key.len), gfp);
crypto/krb5/rfc3961_simplified.c
410
ret = rfc3961_calc_H(krb5, octet_string, &tmp1, gfp);
crypto/krb5/rfc3961_simplified.c
417
ret = rfc3961_calc_DK(krb5, protocol_key, &prfconstant, &derived_key, gfp);
crypto/krb5/rfc3961_simplified.c
421
ret = rfc3961_calc_E(krb5, &derived_key, &tmp2, result, gfp);
crypto/krb5/rfc3961_simplified.c
436
gfp_t gfp)
crypto/krb5/rfc3961_simplified.c
459
ret = krb5_derive_Ke(krb5, TK, usage, &Ke, gfp);
crypto/krb5/rfc3961_simplified.c
464
ret = krb5_derive_Ki(krb5, TK, usage, &Ki, gfp);
crypto/krb5/rfc3961_simplified.c
478
gfp_t gfp)
crypto/krb5/rfc3961_simplified.c
506
gfp_t gfp)
crypto/krb5/rfc3961_simplified.c
515
ret = krb5_derive_Kc(krb5, TK, usage, setkey, gfp);
crypto/krb5/rfc3961_simplified.c
528
gfp_t gfp)
crypto/krb5/rfc6803_camellia.c
119
gfp_t gfp)
crypto/krb5/rfc6803_camellia.c
152
&Kp, gfp);
crypto/krb5/rfc6803_camellia.c
28
gfp_t gfp)
crypto/krb5/rfc8009_aes2.c
123
gfp_t gfp)
crypto/krb5/rfc8009_aes2.c
129
result, gfp);
crypto/krb5/rfc8009_aes2.c
142
gfp_t gfp)
crypto/krb5/rfc8009_aes2.c
146
result, gfp);
crypto/krb5/rfc8009_aes2.c
161
gfp_t gfp)
crypto/krb5/rfc8009_aes2.c
165
result, gfp);
crypto/krb5/rfc8009_aes2.c
34
gfp_t gfp)
crypto/skcipher.c
231
gfp_t gfp = skcipher_walk_gfp(walk);
crypto/skcipher.c
233
walk->page = (void *)__get_free_page(gfp);
drivers/accel/habanalabs/common/command_buffer.c
182
static int hl_cb_mmap_mem_alloc(struct hl_mmap_mem_buf *buf, gfp_t gfp, void *args)
drivers/accel/habanalabs/common/habanalabs.h
4118
struct hl_mmap_mem_buf_behavior *behavior, gfp_t gfp,
drivers/accel/habanalabs/common/habanalabs.h
937
int (*alloc)(struct hl_mmap_mem_buf *buf, gfp_t gfp, void *args);
drivers/accel/habanalabs/common/memory.c
2110
static int hl_ts_alloc_buf(struct hl_mmap_mem_buf *buf, gfp_t gfp, void *args)
drivers/accel/habanalabs/common/memory.c
2119
ts_buff = kzalloc_obj(*ts_buff, gfp);
drivers/accel/habanalabs/common/memory_mgr.c
149
struct hl_mmap_mem_buf_behavior *behavior, gfp_t gfp,
drivers/accel/habanalabs/common/memory_mgr.c
155
buf = kzalloc_obj(*buf, gfp);
drivers/accel/habanalabs/common/memory_mgr.c
174
rc = buf->behavior->alloc(buf, gfp, args);
drivers/base/devcoredump.c
266
gfp_t gfp)
drivers/base/devcoredump.c
268
dev_coredumpm(dev, NULL, data, datalen, gfp, devcd_readv, devcd_freev);
drivers/base/devcoredump.c
360
void *data, size_t datalen, gfp_t gfp,
drivers/base/devcoredump.c
384
devcd = kzalloc_obj(*devcd, gfp);
drivers/base/devcoredump.c
462
size_t datalen, gfp_t gfp)
drivers/base/devcoredump.c
464
dev_coredumpm(dev, NULL, table, datalen, gfp, devcd_read_from_sgtable,
drivers/base/devres.c
1010
const char *devm_kstrdup_const(struct device *dev, const char *s, gfp_t gfp)
drivers/base/devres.c
1015
return devm_kstrdup(dev, s, gfp);
drivers/base/devres.c
1030
char *devm_kvasprintf(struct device *dev, gfp_t gfp, const char *fmt,
drivers/base/devres.c
1041
p = devm_kmalloc(dev, len+1, gfp);
drivers/base/devres.c
1062
char *devm_kasprintf(struct device *dev, gfp_t gfp, const char *fmt, ...)
drivers/base/devres.c
1068
p = devm_kvasprintf(dev, gfp, fmt, ap);
drivers/base/devres.c
1108
void *devm_kmemdup(struct device *dev, const void *src, size_t len, gfp_t gfp)
drivers/base/devres.c
111
size_t size, gfp_t gfp, int nid)
drivers/base/devres.c
1112
p = devm_kmalloc(dev, len, gfp);
drivers/base/devres.c
1132
devm_kmemdup_const(struct device *dev, const void *src, size_t len, gfp_t gfp)
drivers/base/devres.c
1137
return devm_kmemdup(dev, src, len, gfp);
drivers/base/devres.c
119
dr = kmalloc_node_track_caller(tot_size, gfp, nid);
drivers/base/devres.c
124
if (!(gfp & __GFP_ZERO))
drivers/base/devres.c
162
void *__devres_alloc_node(dr_release_t release, size_t size, gfp_t gfp, int nid,
drivers/base/devres.c
167
dr = alloc_dr(release, size, gfp | __GFP_ZERO, nid);
drivers/base/devres.c
552
void *devres_open_group(struct device *dev, void *id, gfp_t gfp)
drivers/base/devres.c
557
grp = kmalloc_obj(*grp, gfp);
drivers/base/devres.c
856
void *devm_kmalloc(struct device *dev, size_t size, gfp_t gfp)
drivers/base/devres.c
864
dr = alloc_dr(devm_kmalloc_release, size, gfp, dev_to_node(dev));
drivers/base/devres.c
894
void *devm_krealloc(struct device *dev, void *ptr, size_t new_size, gfp_t gfp)
drivers/base/devres.c
906
return devm_kmalloc(dev, new_size, gfp);
drivers/base/devres.c
939
new_size, gfp, dev_to_node(dev));
drivers/base/devres.c
988
char *devm_kstrdup(struct device *dev, const char *s, gfp_t gfp)
drivers/base/devres.c
993
return devm_kmemdup(dev, s, strlen(s) + 1, gfp);
drivers/base/firmware_loader/main.c
1163
const char *name, struct device *device, gfp_t gfp, void *context,
drivers/base/firmware_loader/main.c
1168
fw_work = kzalloc_obj(struct firmware_work, gfp);
drivers/base/firmware_loader/main.c
1173
fw_work->name = kstrdup_const(name, gfp);
drivers/base/firmware_loader/main.c
1228
const char *name, struct device *device, gfp_t gfp, void *context,
drivers/base/firmware_loader/main.c
1231
return _request_firmware_nowait(module, uevent, name, device, gfp,
drivers/base/firmware_loader/main.c
1254
struct device *device, gfp_t gfp, void *context,
drivers/base/firmware_loader/main.c
1258
gfp, context, cont, true);
drivers/block/brd.c
88
gfp_t gfp = (opf & REQ_NOWAIT) ? GFP_NOWAIT : GFP_NOIO;
drivers/block/brd.c
91
page = alloc_page(gfp | __GFP_ZERO | __GFP_HIGHMEM);
drivers/block/brd.c
97
page, gfp);
drivers/block/drbd/drbd_state.c
58
static struct drbd_state_change *alloc_state_change(unsigned int n_devices, unsigned int n_connections, gfp_t gfp)
drivers/block/drbd/drbd_state.c
67
state_change = kmalloc(size, gfp);
drivers/block/drbd/drbd_state.c
83
struct drbd_state_change *remember_old_state(struct drbd_resource *resource, gfp_t gfp)
drivers/block/drbd/drbd_state.c
99
state_change = alloc_state_change(n_devices, n_connections, gfp);
drivers/block/loop.c
1119
gfp_t gfp = lo->old_gfp_mask;
drivers/block/loop.c
1147
mapping_set_gfp_mask(filp->f_mapping, gfp);
drivers/crypto/atmel-ecc.c
170
gfp_t gfp;
drivers/crypto/atmel-ecc.c
182
gfp = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL :
drivers/crypto/atmel-ecc.c
185
work_data = kmalloc_obj(*work_data, gfp);
drivers/crypto/bcm/cipher.c
1076
mssg->spu.dst = kmalloc_objs(struct scatterlist, rx_frag_num, rctx->gfp);
drivers/crypto/bcm/cipher.c
1177
mssg->spu.src = kmalloc_objs(struct scatterlist, tx_frag_num, rctx->gfp);
drivers/crypto/bcm/cipher.c
144
mssg->spu.dst = kmalloc_objs(struct scatterlist, rx_frag_num, rctx->gfp);
drivers/crypto/bcm/cipher.c
1716
rctx->gfp = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
drivers/crypto/bcm/cipher.c
1907
rctx->gfp = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
drivers/crypto/bcm/cipher.c
2018
gfp_t gfp;
drivers/crypto/bcm/cipher.c
2033
gfp = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
drivers/crypto/bcm/cipher.c
2036
crypto_shash_descsize(hash), gfp);
drivers/crypto/bcm/cipher.c
207
mssg->spu.src = kmalloc_objs(struct scatterlist, tx_frag_num, rctx->gfp);
drivers/crypto/bcm/cipher.c
2091
gfp_t gfp;
drivers/crypto/bcm/cipher.c
2105
gfp = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
drivers/crypto/bcm/cipher.c
2107
tmpbuf = kmalloc(req->nbytes, gfp);
drivers/crypto/bcm/cipher.c
2185
gfp_t gfp;
drivers/crypto/bcm/cipher.c
2201
gfp = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
drivers/crypto/bcm/cipher.c
2203
tmpbuf = kmalloc(req->nbytes, gfp);
drivers/crypto/bcm/cipher.c
2612
rctx->gfp = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
drivers/crypto/bcm/cipher.c
533
mssg->spu.dst = kmalloc_objs(struct scatterlist, rx_frag_num, rctx->gfp);
drivers/crypto/bcm/cipher.c
587
mssg->spu.src = kmalloc_objs(struct scatterlist, tx_frag_num, rctx->gfp);
drivers/crypto/bcm/cipher.h
342
gfp_t gfp;
drivers/crypto/cavium/nitrox/nitrox_aead.c
158
creq->gfp = (rctx->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL :
drivers/crypto/cavium/nitrox/nitrox_req.h
517
gfp_t gfp;
drivers/crypto/cavium/nitrox/nitrox_req.h
555
static inline void *alloc_req_buf(int nents, int extralen, gfp_t gfp)
drivers/crypto/cavium/nitrox/nitrox_req.h
562
return kzalloc(size, gfp);
drivers/crypto/cavium/nitrox/nitrox_req.h
625
nkreq->src = alloc_req_buf(nents, ivsize, creq->gfp);
drivers/crypto/cavium/nitrox/nitrox_req.h
672
nkreq->dst = alloc_req_buf(nents, extralen, creq->gfp);
drivers/crypto/cavium/nitrox/nitrox_req.h
70
gfp_t gfp;
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
121
sgcomp = kzalloc(sz_comp, sr->gfp);
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
393
sr = kzalloc_obj(*sr, req->gfp);
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
399
sr->gfp = req->gfp;
drivers/crypto/cavium/nitrox/nitrox_skcipher.c
258
creq->gfp = (skreq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
drivers/crypto/ccp/ccp-crypto-aes-cmac.c
108
gfp = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ?
drivers/crypto/ccp/ccp-crypto-aes-cmac.c
110
ret = sg_alloc_table(&rctx->data_sg, sg_count, gfp);
drivers/crypto/ccp/ccp-crypto-aes-cmac.c
65
gfp_t gfp;
drivers/crypto/ccp/ccp-crypto-main.c
274
gfp_t gfp;
drivers/crypto/ccp/ccp-crypto-main.c
276
gfp = req->flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL : GFP_ATOMIC;
drivers/crypto/ccp/ccp-crypto-main.c
278
crypto_cmd = kzalloc_obj(*crypto_cmd, gfp);
drivers/crypto/ccp/ccp-crypto-sha.c
102
gfp = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ?
drivers/crypto/ccp/ccp-crypto-sha.c
105
ret = sg_alloc_table(&rctx->data_sg, sg_count, gfp);
drivers/crypto/ccp/ccp-crypto-sha.c
68
gfp_t gfp;
drivers/crypto/hisilicon/sec/sec_algs.c
179
gfp_t gfp)
drivers/crypto/hisilicon/sec/sec_algs.c
194
gfp, &sgl_next_dma);
drivers/crypto/hisilicon/sec/sec_algs.c
549
int *steps, gfp_t gfp)
drivers/crypto/hisilicon/sec/sec_algs.c
556
sizes = kzalloc_objs(*sizes, *steps, gfp);
drivers/crypto/hisilicon/sec/sec_algs.c
572
struct device *dev, gfp_t gfp)
drivers/crypto/hisilicon/sec/sec_algs.c
580
*splits = kzalloc_objs(struct scatterlist *, steps, gfp);
drivers/crypto/hisilicon/sec/sec_algs.c
585
*splits_nents = kzalloc_objs(int, steps, gfp);
drivers/crypto/hisilicon/sec/sec_algs.c
593
*splits, *splits_nents, gfp);
drivers/crypto/hisilicon/sec/sec_algs.c
634
struct sec_dev_info *info, gfp_t gfp)
drivers/crypto/hisilicon/sec/sec_algs.c
640
el = kzalloc_obj(*el, gfp);
drivers/crypto/hisilicon/sec/sec_algs.c
672
n_ents_in, info, gfp);
drivers/crypto/hisilicon/sec/sec_algs.c
683
n_ents_out, info, gfp);
drivers/crypto/hisilicon/sec/sec_algs.c
724
gfp_t gfp = skreq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL : GFP_ATOMIC;
drivers/crypto/hisilicon/sec/sec_algs.c
733
&steps, gfp);
drivers/crypto/hisilicon/sec/sec_algs.c
739
info->dev, gfp);
drivers/crypto/hisilicon/sec/sec_algs.c
747
sec_req->len_out, info->dev, gfp);
drivers/crypto/hisilicon/sec/sec_algs.c
780
info, gfp);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
171
struct otx_cpt_req_info *req, gfp_t gfp)
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
196
info = kzalloc(total_mem_len, gfp);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
310
gfp_t gfp;
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
312
gfp = (req->areq->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL :
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
314
ret = setup_sgio_list(pdev, &info, req, gfp);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
345
if (gfp == GFP_KERNEL &&
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
351
gfp_t gfp)
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
396
info = kzalloc(total_mem_len, gfp);
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
450
gfp_t gfp)
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
504
info = kzalloc(total_mem_len, gfp);
drivers/crypto/marvell/octeontx2/otx2_cptlf.h
105
gfp_t gfp);
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
104
info = lf->lfs->ops->cpt_sg_info_create(pdev, req, gfp);
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
134
if (gfp == GFP_KERNEL &&
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
97
gfp_t gfp;
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
99
gfp = (req->areq->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL :
drivers/crypto/qce/aead.c
119
gfp_t gfp;
drivers/crypto/qce/aead.c
135
gfp = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
drivers/crypto/qce/aead.c
137
ret = sg_alloc_table(&rctx->dst_tbl, rctx->dst_nents, gfp);
drivers/crypto/qce/aead.c
193
gfp_t gfp;
drivers/crypto/qce/aead.c
243
gfp = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL : GFP_ATOMIC;
drivers/crypto/qce/aead.c
244
ret = sg_alloc_table(&rctx->src_tbl, rctx->src_nents, gfp);
drivers/crypto/qce/skcipher.c
101
gfp = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
drivers/crypto/qce/skcipher.c
104
ret = sg_alloc_table(&rctx->dst_tbl, rctx->dst_nents, gfp);
drivers/crypto/qce/skcipher.c
74
gfp_t gfp;
drivers/dma/bcm2835-dma.c
287
size_t period_len, gfp_t gfp)
drivers/dma/bcm2835-dma.c
300
d = kzalloc_flex(*d, cb_list, frames, gfp);
drivers/dma/bcm2835-dma.c
314
cb_entry->cb = dma_pool_alloc(c->cb_pool, gfp,
drivers/dma/sh/rcar-dmac.c
530
static int rcar_dmac_desc_alloc(struct rcar_dmac_chan *chan, gfp_t gfp)
drivers/dma/sh/rcar-dmac.c
537
page = (void *)get_zeroed_page(gfp);
drivers/dma/sh/rcar-dmac.c
661
static int rcar_dmac_xfer_chunk_alloc(struct rcar_dmac_chan *chan, gfp_t gfp)
drivers/dma/sh/rcar-dmac.c
668
page = (void *)get_zeroed_page(gfp);
drivers/dma/sh/usb-dmac.c
264
gfp_t gfp)
drivers/dma/sh/usb-dmac.c
269
desc = kzalloc_flex(*desc, sg, sg_len, gfp);
drivers/dma/sh/usb-dmac.c
299
unsigned int sg_len, gfp_t gfp)
drivers/dma/sh/usb-dmac.c
316
if (!usb_dmac_desc_alloc(chan, sg_len, gfp)) {
drivers/firmware/qcom/qcom_tzmem.c
195
size_t size, gfp_t gfp)
drivers/firmware/qcom/qcom_tzmem.c
200
gfp);
drivers/firmware/qcom/qcom_tzmem.c
207
&area->paddr, gfp);
drivers/firmware/qcom/qcom_tzmem.c
375
size_t requested, gfp_t gfp)
drivers/firmware/qcom/qcom_tzmem.c
392
return !qcom_tzmem_pool_add_memory(pool, requested, gfp);
drivers/firmware/qcom/qcom_tzmem.c
407
void *qcom_tzmem_alloc(struct qcom_tzmem_pool *pool, size_t size, gfp_t gfp)
drivers/firmware/qcom/qcom_tzmem.c
418
gfp);
drivers/firmware/qcom/qcom_tzmem.c
425
if (qcom_tzmem_try_grow_pool(pool, size, gfp))
drivers/gpu/drm/drm_managed.c
229
void *drmm_kmalloc(struct drm_device *dev, size_t size, gfp_t gfp)
drivers/gpu/drm/drm_managed.c
233
dr = alloc_dr(NULL, size, gfp, dev_to_node(dev->dev));
drivers/gpu/drm/drm_managed.c
236
size, gfp);
drivers/gpu/drm/drm_managed.c
239
dr->node.name = kstrdup_const("kmalloc", gfp);
drivers/gpu/drm/drm_managed.c
257
char *drmm_kstrdup(struct drm_device *dev, const char *s, gfp_t gfp)
drivers/gpu/drm/drm_managed.c
266
buf = drmm_kmalloc(dev, size, gfp);
drivers/gpu/drm/drm_managed.c
88
size_t size, gfp_t gfp, int nid)
drivers/gpu/drm/drm_managed.c
97
dr = kmalloc_node_track_caller(tot_size, gfp, nid);
drivers/gpu/drm/drm_suballoc.c
316
gfp_t gfp, bool intr, size_t align)
drivers/gpu/drm/drm_suballoc.c
332
sa = kmalloc_obj(*sa, gfp);
drivers/gpu/drm/i915/gem/i915_gem_internal.c
40
gfp_t gfp;
drivers/gpu/drm/i915/gem/i915_gem_internal.c
49
gfp = GFP_KERNEL | __GFP_HIGHMEM | __GFP_RECLAIMABLE;
drivers/gpu/drm/i915/gem/i915_gem_internal.c
52
gfp &= ~__GFP_HIGHMEM;
drivers/gpu/drm/i915/gem/i915_gem_internal.c
53
gfp |= __GFP_DMA32;
drivers/gpu/drm/i915/gem/i915_gem_internal.c
74
page = alloc_pages(gfp | (order ? QUIET : MAYFAIL),
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
110
gfp_t gfp = noreclaim;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
114
folio = shmem_read_folio_gfp(mapping, i, gfp);
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
136
gfp = mapping_gfp_mask(mapping);
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
152
gfp |= __GFP_RETRY_MAYFAIL | __GFP_NOWARN;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
180
GEM_BUG_ON(gfp & __GFP_DMA32 && next_pfn >= 0x00100000UL);
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
71
heartbeat_create(struct intel_context *ce, gfp_t gfp)
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
76
rq = __i915_request_create(ce, gfp);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2222
const gfp_t gfp = __GFP_KSWAPD_RECLAIM | __GFP_RETRY_MAYFAIL |
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2229
vma = intel_engine_coredump_add_request(gt->engine, cap->rq, gfp);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2254
const gfp_t gfp = GFP_ATOMIC | __GFP_NOWARN;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2257
cap = kmalloc_obj(*cap, gfp);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2261
cap->error = i915_gpu_coredump_alloc(engine->i915, gfp);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2265
cap->error->gt = intel_gt_coredump_alloc(engine->gt, gfp, CORE_DUMP_FLAG_NONE);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2269
cap->error->gt->engine = intel_engine_coredump_alloc(engine, gfp, CORE_DUMP_FLAG_NONE);
drivers/gpu/drm/i915/i915_deps.c
50
void i915_deps_init(struct i915_deps *deps, gfp_t gfp)
drivers/gpu/drm/i915/i915_deps.c
53
deps->gfp = gfp;
drivers/gpu/drm/i915/i915_deps.c
85
new_fences = kmalloc_objs(*new_fences, new_size, deps->gfp);
drivers/gpu/drm/i915/i915_deps.h
29
gfp_t gfp;
drivers/gpu/drm/i915/i915_deps.h
32
void i915_deps_init(struct i915_deps *deps, gfp_t gfp);
drivers/gpu/drm/i915/i915_gpu_error.c
1498
gfp_t gfp, const char *name)
drivers/gpu/drm/i915/i915_gpu_error.c
1505
c = kmalloc_obj(*c, gfp);
drivers/gpu/drm/i915/i915_gpu_error.c
1525
gfp_t gfp)
drivers/gpu/drm/i915/i915_gpu_error.c
1538
next = capture_vma_snapshot(next, vma->resource, gfp, name);
drivers/gpu/drm/i915/i915_gpu_error.c
1546
gfp_t gfp)
drivers/gpu/drm/i915/i915_gpu_error.c
1551
capture = capture_vma_snapshot(capture, c->vma_res, gfp,
drivers/gpu/drm/i915/i915_gpu_error.c
1597
intel_engine_coredump_alloc(struct intel_engine_cs *engine, gfp_t gfp, u32 dump_flags)
drivers/gpu/drm/i915/i915_gpu_error.c
1601
ee = kzalloc_obj(*ee, gfp);
drivers/gpu/drm/i915/i915_gpu_error.c
1618
gfp_t gfp)
drivers/gpu/drm/i915/i915_gpu_error.c
1631
vma = capture_vma(vma, ce->ring->vma, "ring", gfp);
drivers/gpu/drm/i915/i915_gpu_error.c
1632
vma = capture_vma(vma, ce->state, "HW context", gfp);
drivers/gpu/drm/i915/i915_gpu_error.c
1640
gfp_t gfp)
drivers/gpu/drm/i915/i915_gpu_error.c
1644
vma = engine_coredump_add_context(ee, rq->context, gfp);
drivers/gpu/drm/i915/i915_gpu_error.c
1653
vma = capture_vma_snapshot(vma, rq->batch_res, gfp, "batch");
drivers/gpu/drm/i915/i915_gpu_error.c
1654
vma = capture_user(vma, rq, gfp);
drivers/gpu/drm/i915/i915_gpu_error.c
196
static int pool_refill(struct folio_batch *fbatch, gfp_t gfp)
drivers/gpu/drm/i915/i915_gpu_error.c
201
folio = folio_alloc(gfp, 0);
drivers/gpu/drm/i915/i915_gpu_error.c
2084
i915_gpu_coredump_alloc(struct drm_i915_private *i915, gfp_t gfp)
drivers/gpu/drm/i915/i915_gpu_error.c
2091
error = kzalloc_obj(*error, gfp);
drivers/gpu/drm/i915/i915_gpu_error.c
211
static int pool_init(struct folio_batch *fbatch, gfp_t gfp)
drivers/gpu/drm/i915/i915_gpu_error.c
2111
intel_gt_coredump_alloc(struct intel_gt *gt, gfp_t gfp, u32 dump_flags)
drivers/gpu/drm/i915/i915_gpu_error.c
2115
gc = kzalloc_obj(*gc, gfp);
drivers/gpu/drm/i915/i915_gpu_error.c
217
err = pool_refill(fbatch, gfp);
drivers/gpu/drm/i915/i915_gpu_error.c
224
static void *pool_alloc(struct folio_batch *fbatch, gfp_t gfp)
drivers/gpu/drm/i915/i915_gpu_error.c
228
folio = folio_alloc(gfp, 0);
drivers/gpu/drm/i915/i915_gpu_error.h
284
i915_gpu_coredump_alloc(struct drm_i915_private *i915, gfp_t gfp);
drivers/gpu/drm/i915/i915_gpu_error.h
287
intel_gt_coredump_alloc(struct intel_gt *gt, gfp_t gfp, u32 dump_flags);
drivers/gpu/drm/i915/i915_gpu_error.h
290
intel_engine_coredump_alloc(struct intel_engine_cs *engine, gfp_t gfp, u32 dump_flags);
drivers/gpu/drm/i915/i915_gpu_error.h
295
gfp_t gfp);
drivers/gpu/drm/i915/i915_gpu_error.h
348
i915_gpu_coredump_alloc(struct drm_i915_private *i915, gfp_t gfp)
drivers/gpu/drm/i915/i915_gpu_error.h
354
intel_gt_coredump_alloc(struct intel_gt *gt, gfp_t gfp, u32 dump_flags)
drivers/gpu/drm/i915/i915_gpu_error.h
360
intel_engine_coredump_alloc(struct intel_engine_cs *engine, gfp_t gfp, u32 dump_flags)
drivers/gpu/drm/i915/i915_gpu_error.h
368
gfp_t gfp)
drivers/gpu/drm/i915/i915_request.c
1204
gfp_t gfp)
drivers/gpu/drm/i915/i915_request.c
1235
if (__await_execution(to, from, gfp))
drivers/gpu/drm/i915/i915_request.c
505
gfp_t gfp)
drivers/gpu/drm/i915/i915_request.c
512
cb = kmem_cache_alloc(slab_execute_cbs, gfp);
drivers/gpu/drm/i915/i915_request.c
838
gfp_t gfp)
drivers/gpu/drm/i915/i915_request.c
843
if (!gfpflags_allow_blocking(gfp)) {
drivers/gpu/drm/i915/i915_request.c
859
gfp | __GFP_RETRY_MAYFAIL | __GFP_NOWARN);
drivers/gpu/drm/i915/i915_request.c
871
return kmem_cache_alloc(slab_requests, gfp);
drivers/gpu/drm/i915/i915_request.c
896
__i915_request_create(struct intel_context *ce, gfp_t gfp)
drivers/gpu/drm/i915/i915_request.c
903
might_alloc(gfp);
drivers/gpu/drm/i915/i915_request.c
938
gfp | __GFP_RETRY_MAYFAIL | __GFP_NOWARN);
drivers/gpu/drm/i915/i915_request.c
940
rq = request_alloc_slow(tl, &ce->engine->request_pool, gfp);
drivers/gpu/drm/i915/i915_request.h
375
__i915_request_create(struct intel_context *ce, gfp_t gfp);
drivers/gpu/drm/i915/i915_sw_fence.c
342
wait_queue_entry_t *wq, gfp_t gfp)
drivers/gpu/drm/i915/i915_sw_fence.c
348
might_sleep_if(gfpflags_allow_blocking(gfp));
drivers/gpu/drm/i915/i915_sw_fence.c
363
wq = kmalloc_obj(*wq, gfp);
drivers/gpu/drm/i915/i915_sw_fence.c
365
if (!gfpflags_allow_blocking(gfp))
drivers/gpu/drm/i915/i915_sw_fence.c
405
gfp_t gfp)
drivers/gpu/drm/i915/i915_sw_fence.c
407
return __i915_sw_fence_await_sw_fence(fence, signaler, NULL, gfp);
drivers/gpu/drm/i915/i915_sw_fence.c
484
gfp_t gfp)
drivers/gpu/drm/i915/i915_sw_fence.c
491
might_sleep_if(gfpflags_allow_blocking(gfp));
drivers/gpu/drm/i915/i915_sw_fence.c
501
gfp);
drivers/gpu/drm/i915/i915_sw_fence.c
503
if (!gfpflags_allow_blocking(gfp))
drivers/gpu/drm/i915/i915_sw_fence.c
583
gfp_t gfp)
drivers/gpu/drm/i915/i915_sw_fence.c
590
might_sleep_if(gfpflags_allow_blocking(gfp));
drivers/gpu/drm/i915/i915_sw_fence.c
595
gfp);
drivers/gpu/drm/i915/i915_sw_fence.h
77
gfp_t gfp);
drivers/gpu/drm/i915/i915_sw_fence.h
90
gfp_t gfp);
drivers/gpu/drm/i915/i915_sw_fence.h
96
gfp_t gfp);
drivers/gpu/drm/i915/i915_vma_resource.c
384
gfp_t gfp)
drivers/gpu/drm/i915/i915_vma_resource.c
390
might_alloc(gfp);
drivers/gpu/drm/i915/i915_vma_resource.c
400
0, gfp);
drivers/gpu/drm/i915/i915_vma_resource.h
253
gfp_t gfp);
drivers/gpu/drm/i915/selftests/lib_sw_fence.c
108
struct i915_sw_fence *heap_fence_create(gfp_t gfp)
drivers/gpu/drm/i915/selftests/lib_sw_fence.c
112
h = kmalloc_obj(*h, gfp);
drivers/gpu/drm/i915/selftests/lib_sw_fence.h
40
struct i915_sw_fence *heap_fence_create(gfp_t gfp);
drivers/gpu/drm/msm/msm_iommu.c
381
msm_iommu_pagetable_alloc_pt(void *cookie, size_t size, gfp_t gfp)
drivers/gpu/drm/msm/msm_iommu.c
392
gfp | __GFP_ZERO, get_order(size));
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
155
gfp_t gfp = GFP_USER | __GFP_ZERO;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
200
gfp |= GFP_HIGHUSER;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
202
gfp |= GFP_DMA32;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
205
struct page *p = alloc_page(gfp);
drivers/gpu/drm/panthor/panthor_mmu.c
433
static void *alloc_pt(void *cookie, size_t size, gfp_t gfp)
drivers/gpu/drm/panthor/panthor_mmu.c
444
gfp | __GFP_ZERO, get_order(size));
drivers/gpu/drm/tegra/drm.c
1016
gfp_t gfp;
drivers/gpu/drm/tegra/drm.c
1024
gfp = GFP_KERNEL | __GFP_ZERO;
drivers/gpu/drm/tegra/drm.c
1032
gfp |= GFP_DMA;
drivers/gpu/drm/tegra/drm.c
1035
virt = (void *)__get_free_pages(gfp, get_order(size));
drivers/gpu/drm/ttm/ttm_pool.c
1029
gfp = GFP_DMA32;
drivers/gpu/drm/ttm/ttm_pool.c
1031
gfp = GFP_HIGHUSER;
drivers/gpu/drm/ttm/ttm_pool.c
1051
gfp, alloc_gfp);
drivers/gpu/drm/ttm/ttm_pool.c
855
gfp_t gfp = GFP_KERNEL | __GFP_NOWARN;
drivers/gpu/drm/ttm/ttm_pool.c
859
gfp |= __GFP_RETRY_MAYFAIL;
drivers/gpu/drm/ttm/ttm_pool.c
861
restore = kzalloc_obj(*restore, gfp);
drivers/gpu/drm/ttm/ttm_pool.c
982
gfp_t gfp;
drivers/gpu/drm/xe/xe_sa.c
166
struct drm_suballoc *__xe_sa_bo_new(struct xe_sa_manager *sa_manager, u32 size, gfp_t gfp)
drivers/gpu/drm/xe/xe_sa.c
175
return drm_suballoc_new(&sa_manager->base, size, gfp, true, 0);
drivers/gpu/drm/xe/xe_sa.h
20
struct drm_suballoc *__xe_sa_bo_new(struct xe_sa_manager *sa_manager, u32 size, gfp_t gfp);
drivers/greybus/operation.c
215
static int gb_message_send(struct gb_message *message, gfp_t gfp)
drivers/greybus/operation.c
223
gfp);
drivers/greybus/operation.c
470
size_t response_size, gfp_t gfp)
drivers/greybus/operation.c
478
response = gb_operation_message_alloc(hd, type, response_size, gfp);
drivers/greybus/operation.c
579
gfp_t gfp)
drivers/greybus/operation.c
593
flags, gfp);
drivers/greybus/operation.c
605
gfp_t gfp)
drivers/greybus/operation.c
613
flags, gfp);
drivers/greybus/operation.c
724
gfp_t gfp)
drivers/greybus/operation.c
769
ret = gb_message_send(operation->request, gfp);
drivers/infiniband/core/iwpm_util.c
308
u8 nl_client, gfp_t gfp)
drivers/infiniband/core/iwpm_util.c
313
nlmsg_request = kzalloc_obj(struct iwpm_nlmsg_request, gfp);
drivers/infiniband/core/iwpm_util.h
106
u8 nl_client, gfp_t gfp);
drivers/infiniband/hw/cxgb4/cm.c
142
static struct sk_buff *get_skb(struct sk_buff *skb, int len, gfp_t gfp);
drivers/infiniband/hw/cxgb4/cm.c
309
static void *alloc_ep(int size, gfp_t gfp)
drivers/infiniband/hw/cxgb4/cm.c
313
epc = kzalloc(size, gfp);
drivers/infiniband/hw/cxgb4/cm.c
315
epc->wr_waitp = c4iw_alloc_wr_wait(gfp);
drivers/infiniband/hw/cxgb4/cm.c
3642
int c4iw_ep_disconnect(struct c4iw_ep *ep, int abrupt, gfp_t gfp)
drivers/infiniband/hw/cxgb4/cm.c
455
static struct sk_buff *get_skb(struct sk_buff *skb, int len, gfp_t gfp)
drivers/infiniband/hw/cxgb4/cm.c
462
skb = alloc_skb(len, gfp);
drivers/infiniband/hw/cxgb4/device.c
1520
struct c4iw_wr_wait *c4iw_alloc_wr_wait(gfp_t gfp)
drivers/infiniband/hw/cxgb4/device.c
1524
wr_waitp = kzalloc_obj(*wr_waitp, gfp);
drivers/infiniband/hw/cxgb4/iw_cxgb4.h
1040
int c4iw_ep_disconnect(struct c4iw_ep *ep, int abrupt, gfp_t gfp);
drivers/infiniband/hw/cxgb4/iw_cxgb4.h
1070
struct c4iw_wr_wait *c4iw_alloc_wr_wait(gfp_t gfp);
drivers/infiniband/hw/cxgb4/qp.c
1558
gfp_t gfp)
drivers/infiniband/hw/hfi1/tid_rdma.c
113
gfp_t gfp);
drivers/infiniband/hw/hfi1/tid_rdma.c
1632
gfp_t gfp)
drivers/infiniband/hw/hfi1/tid_rdma.c
1639
flows = kmalloc_node(MAX_FLOWS * sizeof(*flows), gfp,
drivers/infiniband/hw/ionic/ionic_controlpath.c
639
gfp_t gfp = GFP_ATOMIC;
drivers/infiniband/hw/ionic/ionic_controlpath.c
646
gfp = GFP_KERNEL;
drivers/infiniband/hw/ionic/ionic_controlpath.c
672
hdr_buf = kmalloc(PAGE_SIZE, gfp);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
145
gfp_t gfp)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
169
struct page *page = alloc_page(gfp);
drivers/iommu/amd/amd_iommu.h
176
int amd_iommu_pdom_id_reserve(u16 id, gfp_t gfp);
drivers/iommu/amd/amd_iommu.h
27
gfp_t gfp, size_t size);
drivers/iommu/amd/init.c
848
void *__init iommu_alloc_4k_pages(struct amd_iommu *iommu, gfp_t gfp,
drivers/iommu/amd/init.c
854
buf = iommu_alloc_pages_sz(gfp, size);
drivers/iommu/amd/iommu.c
1878
int amd_iommu_pdom_id_reserve(u16 id, gfp_t gfp)
drivers/iommu/amd/iommu.c
1880
return ida_alloc_range(&pdom_ids, id, id, gfp);
drivers/iommu/apple-dart.c
545
size_t pgcount, int prot, gfp_t gfp,
drivers/iommu/apple-dart.c
554
return ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot, gfp,
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3426
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3433
return ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot, gfp, mapped);
drivers/iommu/arm/arm-smmu/arm-smmu.c
1271
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/arm/arm-smmu/arm-smmu.c
1281
ret = ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot, gfp, mapped);
drivers/iommu/arm/arm-smmu/qcom_iommu.c
431
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/dma-iommu.c
1004
dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
drivers/iommu/dma-iommu.c
1011
pages = __iommu_dma_alloc_noncontiguous(dev, size, &sgt, gfp, attrs);
drivers/iommu/dma-iommu.c
1044
enum dma_data_direction dir, gfp_t gfp, unsigned long attrs)
drivers/iommu/dma-iommu.c
1048
sh = kmalloc_obj(*sh, gfp);
drivers/iommu/dma-iommu.c
1052
sh->pages = __iommu_dma_alloc_noncontiguous(dev, size, &sh->sgt, gfp, attrs);
drivers/iommu/dma-iommu.c
1599
struct page **pagep, gfp_t gfp, unsigned long attrs)
drivers/iommu/dma-iommu.c
1607
page = dma_alloc_contiguous(dev, alloc_size, gfp);
drivers/iommu/dma-iommu.c
1609
page = alloc_pages_node(node, gfp, get_order(alloc_size));
drivers/iommu/dma-iommu.c
1636
gfp_t gfp, unsigned long attrs)
drivers/iommu/dma-iommu.c
1643
gfp |= __GFP_ZERO;
drivers/iommu/dma-iommu.c
1645
if (gfpflags_allow_blocking(gfp) &&
drivers/iommu/dma-iommu.c
1647
return iommu_dma_alloc_remap(dev, size, handle, gfp, attrs);
drivers/iommu/dma-iommu.c
1651
!gfpflags_allow_blocking(gfp) && !coherent)
drivers/iommu/dma-iommu.c
1653
gfp, NULL);
drivers/iommu/dma-iommu.c
1655
cpu_addr = iommu_dma_alloc_pages(dev, size, &page, gfp, attrs);
drivers/iommu/dma-iommu.c
874
unsigned int count, unsigned long order_mask, gfp_t gfp)
drivers/iommu/dma-iommu.c
888
gfp |= __GFP_NOWARN | __GFP_HIGHMEM;
drivers/iommu/dma-iommu.c
902
gfp_t alloc_flags = gfp;
drivers/iommu/dma-iommu.c
930
size_t size, struct sg_table *sgt, gfp_t gfp, unsigned long attrs)
drivers/iommu/dma-iommu.c
958
gfp);
drivers/iommu/dma-iommu.c
972
gfp &= ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM | __GFP_COMP);
drivers/iommu/dma-iommu.c
974
if (sg_alloc_table_from_pages(sgt, pages, count, 0, size, gfp))
drivers/iommu/dma-iommu.c
986
gfp);
drivers/iommu/exynos-iommu.c
1239
size_t count, int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/generic_pt/fmt/defs_amdv1.h
17
gfp_t gfp;
drivers/iommu/generic_pt/fmt/defs_vtdss.h
17
gfp_t gfp;
drivers/iommu/generic_pt/fmt/defs_x86_64.h
17
gfp_t gfp;
drivers/iommu/generic_pt/iommu_pt.h
1222
const struct pt_iommu_table_cfg *cfg, gfp_t gfp)
drivers/iommu/generic_pt/iommu_pt.h
1267
table_mem = table_alloc_top(common, common->top_of_table, gfp,
drivers/iommu/generic_pt/iommu_pt.h
366
size_t lg2sz, gfp_t gfp,
drivers/iommu/generic_pt/iommu_pt.h
372
table_mem = iommu_alloc_pages_node_sz(iommu_table->nid, gfp,
drivers/iommu/generic_pt/iommu_pt.h
391
gfp_t gfp,
drivers/iommu/generic_pt/iommu_pt.h
400
gfp, mode);
drivers/iommu/generic_pt/iommu_pt.h
405
gfp_t gfp, enum alloc_mode mode)
drivers/iommu/generic_pt/iommu_pt.h
413
gfp, mode);
drivers/iommu/generic_pt/iommu_pt.h
426
table_mem = table_alloc(pts, attrs->gfp, ALLOC_NORMAL);
drivers/iommu/generic_pt/iommu_pt.h
703
map->attrs.gfp, ALLOC_DEFER_COHERENT_FLUSH);
drivers/iommu/generic_pt/iommu_pt.h
845
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/generic_pt/iommu_pt.h
877
map.attrs.gfp = gfp;
drivers/iommu/generic_pt/kunit_iommu.h
17
const struct pt_iommu_table_cfg *cfg, gfp_t gfp);
drivers/iommu/io-pgtable-arm-v7s.c
238
static void *__arm_v7s_alloc_table(int lvl, gfp_t gfp,
drivers/iommu/io-pgtable-arm-v7s.c
259
table = kmem_cache_zalloc(data->l2_tables, gfp);
drivers/iommu/io-pgtable-arm-v7s.c
462
int lvl, arm_v7s_iopte *ptep, gfp_t gfp)
drivers/iommu/io-pgtable-arm-v7s.c
483
cptep = __arm_v7s_alloc_table(lvl + 1, gfp, data);
drivers/iommu/io-pgtable-arm-v7s.c
504
return __arm_v7s_map(data, iova, paddr, size, prot, lvl + 1, cptep, gfp);
drivers/iommu/io-pgtable-arm-v7s.c
509
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/io-pgtable-arm-v7s.c
523
gfp);
drivers/iommu/io-pgtable-arm.c
251
static void *__arm_lpae_alloc_pages(size_t size, gfp_t gfp,
drivers/iommu/io-pgtable-arm.c
266
pages = cfg->alloc(cookie, alloc_size, gfp);
drivers/iommu/io-pgtable-arm.c
268
pages = iommu_alloc_pages_node_sz(dev_to_node(dev), gfp,
drivers/iommu/io-pgtable-arm.c
425
gfp_t gfp, size_t *mapped)
drivers/iommu/io-pgtable-arm.c
455
cptep = __arm_lpae_alloc_pages(tblsz, gfp, cfg, data->iop.cookie);
drivers/iommu/io-pgtable-arm.c
476
cptep, gfp, mapped);
drivers/iommu/io-pgtable-arm.c
551
int iommu_prot, gfp_t gfp, size_t *mapped)
drivers/iommu/io-pgtable-arm.c
573
ptep, gfp, mapped);
drivers/iommu/io-pgtable-dart.c
228
int iommu_prot, gfp_t gfp, size_t *mapped)
drivers/iommu/io-pgtable-dart.c
259
cptep = iommu_alloc_pages_sz(gfp, tblsz);
drivers/iommu/iommu-pages.c
43
void *iommu_alloc_pages_node_sz(int nid, gfp_t gfp, size_t size)
drivers/iommu/iommu-pages.c
51
if (WARN_ON(gfp & __GFP_HIGHMEM))
drivers/iommu/iommu-pages.c
66
folio = __folio_alloc_node(gfp | __GFP_ZERO, order, nid);
drivers/iommu/iommu-pages.h
101
return iommu_alloc_pages_node_sz(NUMA_NO_NODE, gfp, size);
drivers/iommu/iommu-pages.h
52
void *iommu_alloc_pages_node_sz(int nid, gfp_t gfp, size_t size);
drivers/iommu/iommu-pages.h
99
static inline void *iommu_alloc_pages_sz(gfp_t gfp, size_t size)
drivers/iommu/iommu.c
2576
phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
drivers/iommu/iommu.c
2585
might_sleep_if(gfpflags_allow_blocking(gfp));
drivers/iommu/iommu.c
2594
if (WARN_ON_ONCE(gfp & (__GFP_COMP | __GFP_DMA | __GFP_DMA32 |
drivers/iommu/iommu.c
2622
gfp, &mapped);
drivers/iommu/iommu.c
2657
phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
drivers/iommu/iommu.c
2661
ret = iommu_map_nosync(domain, iova, paddr, size, prot, gfp);
drivers/iommu/iommu.c
2787
gfp_t gfp)
drivers/iommu/iommu.c
2799
len, prot, gfp);
drivers/iommu/iommu.c
2942
gfp_t gfp)
drivers/iommu/iommu.c
2946
region = kzalloc_obj(*region, gfp);
drivers/iommu/ipmmu-vmsa.c
671
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/ipmmu-vmsa.c
676
prot, gfp, mapped);
drivers/iommu/msm_iommu.c
488
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/mtk_iommu.c
813
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/mtk_iommu.c
822
return dom->iop->map_pages(dom->iop, iova, paddr, pgsize, pgcount, prot, gfp, mapped);
drivers/iommu/mtk_iommu_v1.c
354
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/omap-iommu.c
1299
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/riscv/iommu.c
1117
gfp_t gfp)
drivers/iommu/riscv/iommu.c
1149
addr = iommu_alloc_pages_node_sz(domain->numa_node, gfp,
drivers/iommu/riscv/iommu.c
1193
gfp_t gfp, size_t *mapped)
drivers/iommu/riscv/iommu.c
1210
ptr = riscv_iommu_pte_alloc(domain, iova, pgsize, gfp);
drivers/iommu/rockchip-iommu.c
824
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/s390-iommu.c
216
static unsigned long *dma_alloc_cpu_table(gfp_t gfp)
drivers/iommu/s390-iommu.c
220
table = kmem_cache_alloc(dma_region_table_cache, gfp);
drivers/iommu/s390-iommu.c
306
static unsigned long *dma_alloc_page_table(gfp_t gfp)
drivers/iommu/s390-iommu.c
310
table = kmem_cache_alloc(dma_page_table_cache, gfp);
drivers/iommu/s390-iommu.c
320
dma_addr_t dma_addr, gfp_t gfp)
drivers/iommu/s390-iommu.c
331
rto = dma_alloc_cpu_table(gfp);
drivers/iommu/s390-iommu.c
350
dma_addr_t dma_addr, gfp_t gfp)
drivers/iommu/s390-iommu.c
361
rso = dma_alloc_cpu_table(gfp);
drivers/iommu/s390-iommu.c
380
return dma_walk_rs_table(rso, dma_addr, gfp);
drivers/iommu/s390-iommu.c
383
static unsigned long *dma_get_seg_table_origin(unsigned long *rtep, gfp_t gfp)
drivers/iommu/s390-iommu.c
392
sto = dma_alloc_cpu_table(gfp);
drivers/iommu/s390-iommu.c
410
static unsigned long *dma_get_page_table_origin(unsigned long *step, gfp_t gfp)
drivers/iommu/s390-iommu.c
419
pto = dma_alloc_page_table(gfp);
drivers/iommu/s390-iommu.c
437
dma_addr_t dma_addr, gfp_t gfp)
drivers/iommu/s390-iommu.c
441
return dma_walk_rf_table(domain->dma_table, dma_addr, gfp);
drivers/iommu/s390-iommu.c
443
return dma_walk_rs_table(domain->dma_table, dma_addr, gfp);
drivers/iommu/s390-iommu.c
452
dma_addr_t dma_addr, gfp_t gfp)
drivers/iommu/s390-iommu.c
457
rto = dma_walk_region_tables(domain, dma_addr, gfp);
drivers/iommu/s390-iommu.c
462
sto = dma_get_seg_table_origin(&rto[rtx], gfp);
drivers/iommu/s390-iommu.c
467
pto = dma_get_page_table_origin(&sto[sx], gfp);
drivers/iommu/s390-iommu.c
860
gfp_t gfp)
drivers/iommu/s390-iommu.c
868
entry = dma_walk_cpu_trans(s390_domain, dma_addr, gfp);
drivers/iommu/s390-iommu.c
883
entry = dma_walk_cpu_trans(s390_domain, dma_addr, gfp);
drivers/iommu/s390-iommu.c
915
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/s390-iommu.c
935
pgcount, flags, gfp);
drivers/iommu/sprd-iommu.c
289
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/sun50i-iommu.c
525
gfp_t gfp)
drivers/iommu/sun50i-iommu.c
530
page_table = kmem_cache_zalloc(iommu->pt_pool, gfp);
drivers/iommu/sun50i-iommu.c
557
dma_addr_t iova, gfp_t gfp)
drivers/iommu/sun50i-iommu.c
572
page_table = sun50i_iommu_alloc_page_table(iommu, gfp);
drivers/iommu/sun50i-iommu.c
597
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/sun50i-iommu.c
613
page_table = sun50i_dte_get_page_table(sun50i_domain, iova, gfp);
drivers/iommu/tegra-smmu.c
682
unsigned long iova, gfp_t gfp,
drivers/iommu/tegra-smmu.c
697
if (gfpflags_allow_blocking(gfp))
drivers/iommu/tegra-smmu.c
700
pt = iommu_alloc_pages_sz(gfp | __GFP_DMA, SMMU_SIZE_PT);
drivers/iommu/tegra-smmu.c
702
if (gfpflags_allow_blocking(gfp))
drivers/iommu/tegra-smmu.c
722
phys_addr_t paddr, size_t size, int prot, gfp_t gfp,
drivers/iommu/tegra-smmu.c
731
pt = as_get_pde_page(as, iova, gfp, flags);
drivers/iommu/tegra-smmu.c
777
int prot, gfp_t gfp, size_t *mapped)
drivers/iommu/tegra-smmu.c
784
ret = __tegra_smmu_map(domain, iova, paddr, size, prot, gfp, &flags);
drivers/iommu/virtio-iommu.c
844
int prot, gfp_t gfp, size_t *mapped)
drivers/irqchip/irq-gic-v3-its.c
213
static struct page *its_alloc_pages_node(int node, gfp_t gfp,
drivers/irqchip/irq-gic-v3-its.c
219
page = alloc_pages_node(node, gfp | gfp_flags_quirk, order);
drivers/irqchip/irq-gic-v3-its.c
237
static struct page *its_alloc_pages(gfp_t gfp, unsigned int order)
drivers/irqchip/irq-gic-v3-its.c
239
return its_alloc_pages_node(NUMA_NO_NODE, gfp, order);
drivers/md/bcache/bset.c
317
gfp_t gfp)
drivers/md/bcache/bset.c
325
t->data = (void *) __get_free_pages(__GFP_COMP|gfp, b->page_order);
drivers/md/bcache/bset.c
330
? kmalloc(bset_tree_bytes(b), gfp)
drivers/md/bcache/bset.c
331
: (void *) __get_free_pages(gfp, get_order(bset_tree_bytes(b)));
drivers/md/bcache/bset.c
336
? kmalloc(bset_prev_bytes(b), gfp)
drivers/md/bcache/bset.c
337
: (void *) __get_free_pages(gfp, get_order(bset_prev_bytes(b)));
drivers/md/bcache/bset.h
294
gfp_t gfp);
drivers/md/bcache/btree.c
548
static void mca_data_alloc(struct btree *b, struct bkey *k, gfp_t gfp)
drivers/md/bcache/btree.c
554
gfp)) {
drivers/md/bcache/btree.c
582
struct bkey *k, gfp_t gfp)
drivers/md/bcache/btree.c
588
struct btree *b = kzalloc_obj(struct btree, gfp);
drivers/md/bcache/btree.c
602
mca_data_alloc(b, k, gfp);
drivers/md/bcache/super.c
1859
#define alloc_meta_bucket_pages(gfp, sb) \
drivers/md/bcache/super.c
1860
((void *) __get_free_pages(__GFP_ZERO|__GFP_COMP|gfp, ilog2(meta_bucket_pages(sb))))
drivers/md/bcache/util.h
124
#define __init_fifo(fifo, gfp) \
drivers/md/bcache/util.h
135
(fifo)->data = kvmalloc(_bytes, (gfp) & GFP_KERNEL); \
drivers/md/bcache/util.h
139
#define init_fifo_exact(fifo, _size, gfp) \
drivers/md/bcache/util.h
142
__init_fifo(fifo, gfp); \
drivers/md/bcache/util.h
145
#define init_fifo(fifo, _size, gfp) \
drivers/md/bcache/util.h
150
__init_fifo(fifo, gfp); \
drivers/md/bcache/util.h
39
#define init_heap(heap, _size, gfp) \
drivers/md/bcache/util.h
45
(heap)->data = kvmalloc(_bytes, (gfp) & GFP_KERNEL); \
drivers/md/dm-bio-prison-v1.c
74
struct dm_bio_prison_cell *dm_bio_prison_alloc_cell(struct dm_bio_prison *prison, gfp_t gfp)
drivers/md/dm-bio-prison-v1.c
76
return mempool_alloc(&prison->cell_pool, gfp);
drivers/md/dm-bio-prison-v1.h
71
gfp_t gfp);
drivers/md/dm-bio-prison-v2.c
67
struct dm_bio_prison_cell_v2 *dm_bio_prison_alloc_cell_v2(struct dm_bio_prison_v2 *prison, gfp_t gfp)
drivers/md/dm-bio-prison-v2.c
69
return mempool_alloc(&prison->cell_pool, gfp);
drivers/md/dm-bio-prison-v2.h
68
gfp_t gfp);
drivers/md/dm-crypt.c
1860
static int kcryptd_io_read(struct dm_crypt_io *io, gfp_t gfp)
drivers/md/dm-crypt.c
1866
if (!(gfp & __GFP_DIRECT_RECLAIM))
drivers/md/dm-crypt.c
1886
clone = bio_alloc_clone(cc->dev->bdev, io->base_bio, gfp, &cc->bs);
drivers/md/dm-kcopyd.c
218
static struct page_list *alloc_pl(gfp_t gfp)
drivers/md/dm-kcopyd.c
222
pl = kmalloc_obj(*pl, gfp);
drivers/md/dm-kcopyd.c
226
pl->page = alloc_page(gfp | __GFP_HIGHMEM);
drivers/md/dm-snap.c
729
static struct dm_exception *alloc_completed_exception(gfp_t gfp)
drivers/md/dm-snap.c
733
e = kmem_cache_alloc(exception_cache, gfp);
drivers/md/dm-snap.c
734
if (!e && gfp == GFP_NOIO)
drivers/md/dm-table.c
507
gfp_t gfp;
drivers/md/dm-table.c
511
gfp = GFP_KERNEL;
drivers/md/dm-table.c
514
gfp = GFP_NOIO;
drivers/md/dm-table.c
516
argv = kmalloc_array(new_size, sizeof(*argv), gfp);
drivers/md/raid1.c
1327
gfp_t gfp = r1_bio ? (GFP_NOIO | __GFP_HIGH) : GFP_NOIO;
drivers/md/raid1.c
1395
read_bio = bio_alloc_clone(mirror->rdev->bdev, bio, gfp,
drivers/md/raid10.c
1158
gfp_t gfp = GFP_NOIO;
drivers/md/raid10.c
1173
gfp = GFP_NOIO | __GFP_HIGH;
drivers/md/raid10.c
1225
read_bio = bio_alloc_clone(rdev->bdev, bio, gfp, &mddev->bio_set);
drivers/md/raid5.c
2349
static struct stripe_head *alloc_stripe(struct kmem_cache *sc, gfp_t gfp,
drivers/md/raid5.c
2354
sh = kmem_cache_zalloc(sc, gfp);
drivers/md/raid5.c
2367
sh->ppl_page = alloc_page(gfp);
drivers/md/raid5.c
2382
static int grow_one_stripe(struct r5conf *conf, gfp_t gfp)
drivers/md/raid5.c
2386
sh = alloc_stripe(conf->slab_cache, gfp, conf->pool_size, conf);
drivers/md/raid5.c
2390
if (grow_buffers(sh, gfp)) {
drivers/md/raid5.c
484
static int alloc_stripe_pages(struct stripe_head *sh, gfp_t gfp)
drivers/md/raid5.c
494
p = alloc_page(gfp);
drivers/md/raid5.c
548
static int grow_buffers(struct stripe_head *sh, gfp_t gfp)
drivers/md/raid5.c
557
if (!(page = alloc_page(gfp))) {
drivers/md/raid5.c
565
if (alloc_stripe_pages(sh, gfp))
drivers/media/pci/intel/ipu6/ipu6-dma.c
153
dma_addr_t *dma_handle, gfp_t gfp,
drivers/media/pci/intel/ipu6/ipu6-dma.c
179
pages = __alloc_buffer(size, gfp, attrs);
drivers/media/pci/intel/ipu6/ipu6-dma.c
58
static struct page **__alloc_buffer(size_t size, gfp_t gfp, unsigned long attrs)
drivers/media/pci/intel/ipu6/ipu6-dma.c
69
gfp |= __GFP_NOWARN;
drivers/media/pci/intel/ipu6/ipu6-dma.c
74
pages[i] = alloc_pages(gfp, order);
drivers/media/pci/intel/ipu6/ipu6-dma.c
76
pages[i] = alloc_pages(gfp, --order);
drivers/media/pci/intel/ipu6/ipu6-dma.h
26
dma_addr_t *dma_handle, gfp_t gfp,
drivers/media/platform/amphion/vpu_core.c
144
gfp_t gfp = GFP_KERNEL | GFP_DMA32;
drivers/media/platform/amphion/vpu_core.c
149
buf->virt = dma_alloc_coherent(dev, buf->length, &buf->phys, gfp);
drivers/mfd/dln2.c
303
int *obuf_len, gfp_t gfp)
drivers/mfd/dln2.c
310
buf = kmalloc(len, gfp);
drivers/mfd/dln2.c
630
static int dln2_start_rx_urbs(struct dln2_dev *dln2, gfp_t gfp)
drivers/mfd/dln2.c
637
ret = usb_submit_urb(dln2->rx_urb[i], gfp);
drivers/mmc/core/queue.c
166
static struct scatterlist *mmc_alloc_sg(unsigned short sg_len, gfp_t gfp)
drivers/mmc/core/queue.c
170
sg = kmalloc_objs(*sg, sg_len, gfp);
drivers/net/ethernet/airoha/airoha_npu.c
463
void *data, int data_len, gfp_t gfp)
drivers/net/ethernet/airoha/airoha_npu.c
469
wlan_data = kzalloc(len, gfp);
drivers/net/ethernet/airoha/airoha_npu.c
486
void *data, int data_len, gfp_t gfp)
drivers/net/ethernet/airoha/airoha_npu.c
492
wlan_data = kzalloc(len, gfp);
drivers/net/ethernet/alacritech/slicoss.c
392
static void slic_refill_rx_queue(struct slic_device *sdev, gfp_t gfp)
drivers/net/ethernet/alacritech/slicoss.c
406
skb = alloc_skb(maplen + ALIGN_MASK, gfp);
drivers/net/ethernet/amd/lance.c
867
lance_init_ring(struct net_device *dev, gfp_t gfp)
drivers/net/ethernet/amd/lance.c
879
skb = alloc_skb(PKT_BUF_SZ, GFP_DMA | gfp);
drivers/net/ethernet/amd/lance.c
884
rx_buff = kmalloc(PKT_BUF_SZ, GFP_DMA | gfp);
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
182
gfp_t gfp;
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
189
gfp = GFP_ATOMIC | __GFP_COMP | __GFP_NOWARN;
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
191
pages = alloc_pages_node(node, gfp, order);
drivers/net/ethernet/atheros/alx/main.c
71
static int alx_refill_rx_ring(struct alx_priv *alx, gfp_t gfp)
drivers/net/ethernet/atheros/alx/main.c
96
skb = __netdev_alloc_skb(alx->dev, alx->rxbuf_size + 64, gfp);
drivers/net/ethernet/broadcom/b44.c
1165
static int b44_alloc_consistent(struct b44 *bp, gfp_t gfp)
drivers/net/ethernet/broadcom/b44.c
1170
bp->rx_buffers = kzalloc(size, gfp);
drivers/net/ethernet/broadcom/b44.c
1175
bp->tx_buffers = kzalloc(size, gfp);
drivers/net/ethernet/broadcom/b44.c
1181
&bp->rx_ring_dma, gfp);
drivers/net/ethernet/broadcom/b44.c
1189
rx_ring = kzalloc(size, gfp);
drivers/net/ethernet/broadcom/b44.c
1209
&bp->tx_ring_dma, gfp);
drivers/net/ethernet/broadcom/b44.c
1217
tx_ring = kzalloc(size, gfp);
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
121
ctx->gfp = GFP_KERNEL;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
46
ctx->gfp = GFP_KERNEL;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
512
addr = dma_alloc_coherent(bd->dev, size, dma_handle, ctx->gfp);
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
93
void bnge_hwrm_req_alloc_flags(struct bnge_dev *bd, void *req, gfp_t gfp)
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
98
ctx->gfp = gfp;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.h
30
gfp_t gfp;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1047
gfp_t gfp)
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1053
BNGE_RX_PAGE_SIZE, gfp);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1055
netmem = page_pool_alloc_netmems(rxr->page_pool, gfp);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1067
gfp_t gfp)
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1073
bn->rx_buf_size, gfp);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1082
u16 prod, gfp_t gfp)
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1090
data = __bnge_alloc_rx_frag(bn, &mapping, rxr, gfp);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1146
u16 prod, gfp_t gfp)
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1156
netmem = __bnge_alloc_rx_netmem(bn, &mapping, rxr, &offset, gfp);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1760
gfp_t gfp)
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1771
fltr = kzalloc_obj(*fltr, gfp);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.h
563
u16 prod, gfp_t gfp);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.h
566
struct bnge_rx_ring_info *rxr, gfp_t gfp);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.h
568
u16 prod, gfp_t gfp);
drivers/net/ethernet/broadcom/bnx2.c
2722
bnx2_alloc_rx_page(struct bnx2 *bp, struct bnx2_rx_ring_info *rxr, u16 index, gfp_t gfp)
drivers/net/ethernet/broadcom/bnx2.c
2728
struct page *page = alloc_page(gfp);
drivers/net/ethernet/broadcom/bnx2.c
2763
bnx2_alloc_rx_data(struct bnx2 *bp, struct bnx2_rx_ring_info *rxr, u16 index, gfp_t gfp)
drivers/net/ethernet/broadcom/bnx2.c
2771
data = kmalloc(bp->rx_buf_size, gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1031
u16 prod, gfp_t gfp)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1041
netmem = __bnxt_alloc_rx_netmem(bp, &mapping, rxr, &offset, gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6081
gfp_t gfp)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6093
fltr = kzalloc_obj(*fltr, gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
911
gfp_t gfp)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
932
gfp_t gfp)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
938
rxr->rx_page_size, gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
940
netmem = page_pool_alloc_netmems(rxr->page_pool, gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
952
gfp_t gfp)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
958
bp->rx_buf_size, gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
967
u16 prod, gfp_t gfp)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
976
__bnxt_alloc_rx_page(bp, &mapping, rxr, &offset, gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
985
u8 *data = __bnxt_alloc_rx_frag(bp, &mapping, rxr, gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt.h
2907
u16 prod, gfp_t gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
158
void hwrm_req_alloc_flags(struct bnxt *bp, void *req, gfp_t gfp)
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
163
ctx->gfp = gfp;
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
211
ctx->gfp = GFP_KERNEL;
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
809
addr = dma_alloc_coherent(&bp->pdev->dev, size, dma_handle, ctx->gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
85
ctx->gfp = GFP_KERNEL;
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.h
37
gfp_t gfp;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
122
page = alloc_pages(gfp | __GFP_COMP | __GFP_NOWARN, 0);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
177
gfp_t gfp, u32 buf_len, u64 *rbuf)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
194
pgcache = nicvf_alloc_page(nic, rbdr, gfp);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
379
static void nicvf_refill_rbdr(struct nicvf *nic, gfp_t gfp)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
417
if (nicvf_alloc_rcv_buffer(nic, rbdr, gfp, RCV_FRAG_LEN, &rbuf))
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
94
struct rbdr *rbdr, gfp_t gfp)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
791
gfp_t gfp)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
798
skb = alloc_skb(len, gfp);
drivers/net/ethernet/chelsio/cxgb3/sge.c
439
struct rx_sw_desc *sd, gfp_t gfp,
drivers/net/ethernet/chelsio/cxgb3/sge.c
445
q->pg_chunk.page = alloc_pages(gfp, order);
drivers/net/ethernet/chelsio/cxgb3/sge.c
501
static int refill_fl(struct adapter *adap, struct sge_fl *q, int n, gfp_t gfp)
drivers/net/ethernet/chelsio/cxgb3/sge.c
512
if (unlikely(alloc_pg_chunk(adap, q, sd, gfp,
drivers/net/ethernet/chelsio/cxgb3/sge.c
527
struct sk_buff *skb = alloc_skb(q->buf_size, gfp);
drivers/net/ethernet/chelsio/cxgb4/sge.c
536
gfp_t gfp)
drivers/net/ethernet/chelsio/cxgb4/sge.c
551
gfp |= __GFP_NOWARN;
drivers/net/ethernet/chelsio/cxgb4/sge.c
561
pg = alloc_pages_node(node, gfp | __GFP_COMP, s->fl_pg_order);
drivers/net/ethernet/chelsio/cxgb4/sge.c
592
pg = alloc_pages_node(node, gfp, 0);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
605
int n, gfp_t gfp)
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
621
gfp |= __GFP_NOWARN;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
633
page = __dev_alloc_pages(gfp, s->fl_pg_order);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
677
page = __dev_alloc_page(gfp);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
1966
size_t len, gfp_t gfp)
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
1973
skb = alloc_skb(len, gfp);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1135
gfp_t gfp = sk->sk_allocation;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1139
page = alloc_pages(gfp | __GFP_COMP |
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1147
page = alloc_page(gfp);
drivers/net/ethernet/emulex/benet/be_main.c
2580
static inline struct page *be_alloc_pages(u32 size, gfp_t gfp)
drivers/net/ethernet/emulex/benet/be_main.c
2585
gfp |= __GFP_COMP;
drivers/net/ethernet/emulex/benet/be_main.c
2586
return alloc_pages(gfp, order);
drivers/net/ethernet/emulex/benet/be_main.c
2593
static void be_post_rx_frags(struct be_rx_obj *rxo, gfp_t gfp, u32 frags_needed)
drivers/net/ethernet/emulex/benet/be_main.c
2607
pagep = be_alloc_pages(adapter->big_page_size, gfp);
drivers/net/ethernet/faraday/ftgmac100.c
431
struct ftgmac100_rxdes *rxdes, gfp_t gfp)
drivers/net/ethernet/faraday/ftmac100.c
697
struct ftmac100_rxdes *rxdes, gfp_t gfp)
drivers/net/ethernet/faraday/ftmac100.c
703
page = alloc_page(gfp);
drivers/net/ethernet/faraday/ftmac100.c
74
struct ftmac100_rxdes *rxdes, gfp_t gfp);
drivers/net/ethernet/freescale/fec_main.c
389
gfp_t gfp)
drivers/net/ethernet/freescale/fec_main.c
391
return dma_alloc_noncoherent(dev, size, handle, DMA_BIDIRECTIONAL, gfp);
drivers/net/ethernet/freescale/fec_main.c
401
gfp_t gfp)
drivers/net/ethernet/freescale/fec_main.c
403
return dma_alloc_coherent(dev, size, handle, gfp);
drivers/net/ethernet/freescale/fec_main.c
427
gfp_t gfp)
drivers/net/ethernet/freescale/fec_main.c
432
dr = devres_alloc(fec_dmam_release, sizeof(*dr), gfp);
drivers/net/ethernet/freescale/fec_main.c
435
vaddr = fec_dma_alloc(dev, size, handle, gfp);
drivers/net/ethernet/fungible/funeth/funeth_rx.c
106
p = __alloc_pages_node(node, gfp | __GFP_NOWARN, 0);
drivers/net/ethernet/fungible/funeth/funeth_rx.c
99
int node, gfp_t gfp)
drivers/net/ethernet/intel/e1000e/e1000.h
259
gfp_t gfp);
drivers/net/ethernet/intel/e1000e/netdev.c
648
int cleaned_count, gfp_t gfp)
drivers/net/ethernet/intel/e1000e/netdev.c
669
skb = __netdev_alloc_skb_ip_align(netdev, bufsz, gfp);
drivers/net/ethernet/intel/e1000e/netdev.c
718
int cleaned_count, gfp_t gfp)
drivers/net/ethernet/intel/e1000e/netdev.c
744
ps_page->page = alloc_page(gfp);
drivers/net/ethernet/intel/e1000e/netdev.c
770
gfp);
drivers/net/ethernet/intel/e1000e/netdev.c
823
int cleaned_count, gfp_t gfp)
drivers/net/ethernet/intel/e1000e/netdev.c
844
skb = __netdev_alloc_skb_ip_align(netdev, bufsz, gfp);
drivers/net/ethernet/intel/e1000e/netdev.c
855
buffer_info->page = alloc_page(gfp);
drivers/net/ethernet/marvell/sky2.c
1441
static struct sk_buff *sky2_rx_alloc(struct sky2_port *sky2, gfp_t gfp)
drivers/net/ethernet/marvell/sky2.c
1448
gfp);
drivers/net/ethernet/marvell/sky2.c
1466
struct page *page = alloc_page(gfp);
drivers/net/ethernet/mellanox/mlx4/en_rx.c
121
gfp_t gfp)
drivers/net/ethernet/mellanox/mlx4/en_rx.c
128
return mlx4_en_alloc_frags(priv, ring, rx_desc, frags, gfp);
drivers/net/ethernet/mellanox/mlx4/en_rx.c
59
gfp_t gfp)
drivers/net/ethernet/mellanox/mlx4/en_rx.c
66
frags->page = page_pool_alloc_pages(ring->pp, gfp);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1895
gfp_t gfp)
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1928
msg = mlx5_alloc_cmd_msg(dev, gfp, in_size, 0);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1962
gfp_t gfp;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1985
gfp = callback ? GFP_ATOMIC : GFP_KERNEL;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1987
inb = alloc_msg(dev, in_size, gfp);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
2001
outb = mlx5_alloc_cmd_msg(dev, gfp, out_size, token);
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
132
gfp_t gfp;
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
137
gfp = GFP_KERNEL | __GFP_ZERO;
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
138
buff = (void *)__get_free_pages(gfp,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec.c
779
gfp_t gfp;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec.c
787
gfp = (x->xso.flags & XFRM_DEV_OFFLOAD_FLAG_ACQ) ? GFP_ATOMIC : GFP_KERNEL;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec.c
788
sa_entry = kzalloc_obj(*sa_entry, gfp);
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h
71
#define l2addr_hash_add(hash, mac, type, gfp) ({ \
drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h
75
ptr = kzalloc_obj(type, gfp); \
drivers/net/ethernet/microchip/lan743x_main.c
2318
gfp_t gfp)
drivers/net/ethernet/microchip/lan743x_main.c
2332
skb = __netdev_alloc_skb(netdev, buffer_length, gfp);
drivers/net/ethernet/natsemi/ns83820.c
539
static inline int rx_refill(struct net_device *ndev, gfp_t gfp)
drivers/net/ethernet/natsemi/ns83820.c
549
if (gfp == GFP_ATOMIC)
drivers/net/ethernet/natsemi/ns83820.c
556
skb = __netdev_alloc_skb(ndev, REAL_RX_BUF_SIZE+16, gfp);
drivers/net/ethernet/natsemi/ns83820.c
561
if (gfp != GFP_ATOMIC)
drivers/net/ethernet/natsemi/ns83820.c
564
if (gfp != GFP_ATOMIC)
drivers/net/ethernet/natsemi/ns83820.c
571
if (gfp == GFP_ATOMIC)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1607
gfp_t gfp = (GFP_ATOMIC | __GFP_NOWARN);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1610
gfp |= GFP_DMA32;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1613
buf->page = page_pool_alloc_pages(rx_q->page_pool, gfp);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1620
buf->sec_page = page_pool_alloc_pages(rx_q->page_pool, gfp);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4942
gfp_t gfp = (GFP_ATOMIC | __GFP_NOWARN);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4945
gfp |= GFP_DMA32;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4958
buf->page = page_pool_alloc_pages(rx_q->page_pool, gfp);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4964
buf->sec_page = page_pool_alloc_pages(rx_q->page_pool, gfp);
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
328
gfp_t gfp, int order)
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
334
gfp |= __GFP_COMP | __GFP_NOWARN;
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
336
pages = alloc_pages(gfp, order);
drivers/net/mctp/mctp-usb.c
123
static int mctp_usb_rx_queue(struct mctp_usb *mctp_usb, gfp_t gfp)
drivers/net/mctp/mctp-usb.c
128
skb = __netdev_alloc_skb(mctp_usb->netdev, MCTP_USB_XFER_SIZE, gfp);
drivers/net/mctp/mctp-usb.c
139
rc = usb_submit_urb(mctp_usb->rx_urb, gfp);
drivers/net/usb/hso.c
298
struct usb_device *usb, gfp_t gfp);
drivers/net/usb/hso.c
3178
struct usb_device *usb, gfp_t gfp)
drivers/net/usb/hso.c
3190
result = usb_submit_urb(shared_int->shared_intr_urb, gfp);
drivers/net/virtio_net.c
1444
struct xsk_buff_pool *pool, gfp_t gfp)
drivers/net/virtio_net.c
1471
xsk_buffs[i], NULL, gfp);
drivers/net/virtio_net.c
2665
gfp_t gfp)
drivers/net/virtio_net.c
2676
if (unlikely(!skb_page_frag_refill(len, &rq->alloc_frag, gfp)))
drivers/net/virtio_net.c
2679
buf = virtnet_rq_alloc(rq, len, gfp);
drivers/net/virtio_net.c
2687
err = virtqueue_add_inbuf_premapped(rq->vq, rq->sg, 1, buf, ctx, gfp);
drivers/net/virtio_net.c
2697
gfp_t gfp)
drivers/net/virtio_net.c
2707
first = get_a_page(rq, gfp);
drivers/net/virtio_net.c
2720
first = get_a_page(rq, gfp);
drivers/net/virtio_net.c
2738
first, gfp);
drivers/net/virtio_net.c
2763
struct receive_queue *rq, gfp_t gfp)
drivers/net/virtio_net.c
2780
if (unlikely(!skb_page_frag_refill(len + room, alloc_frag, gfp)))
drivers/net/virtio_net.c
2786
buf = virtnet_rq_alloc(rq, len + room, gfp);
drivers/net/virtio_net.c
2807
err = virtqueue_add_inbuf_premapped(rq->vq, rq->sg, 1, buf, ctx, gfp);
drivers/net/virtio_net.c
2824
gfp_t gfp)
drivers/net/virtio_net.c
2829
err = virtnet_add_recvbuf_xsk(vi, rq, rq->xsk_pool, gfp);
drivers/net/virtio_net.c
2835
err = add_recvbuf_mergeable(vi, rq, gfp);
drivers/net/virtio_net.c
2837
err = add_recvbuf_big(vi, rq, gfp);
drivers/net/virtio_net.c
2839
err = add_recvbuf_small(vi, rq, gfp);
drivers/net/virtio_net.c
999
static void *virtnet_rq_alloc(struct receive_queue *rq, u32 size, gfp_t gfp)
drivers/net/wireless/ath/ath10k/wmi-tlv.c
143
size_t len, gfp_t gfp)
drivers/net/wireless/ath/ath10k/wmi-tlv.c
148
tb = kzalloc_objs(*tb, WMI_TLV_TAG_MAX, gfp);
drivers/net/wireless/ath/ath11k/wmi.c
247
struct sk_buff *skb, gfp_t gfp)
drivers/net/wireless/ath/ath11k/wmi.c
252
tb = kzalloc_objs(*tb, WMI_TAG_MAX, gfp);
drivers/net/wireless/ath/ath11k/wmi.h
6433
struct sk_buff *skb, gfp_t gfp);
drivers/net/wireless/ath/ath12k/dbring.c
102
ret = ath12k_dbring_bufs_replenish(ar, ring, buff, gfp);
drivers/net/wireless/ath/ath12k/dbring.c
14
gfp_t gfp)
drivers/net/wireless/ath/ath12k/dbring.c
40
buf_id = idr_alloc(&ring->bufs_idr, buff, 0, ring->bufs_max, gfp);
drivers/net/wireless/ath/ath12k/dbring.c
78
gfp_t gfp)
drivers/net/wireless/ath/ath12k/dbring.c
98
buff = kzalloc(size, gfp);
drivers/net/wireless/ath/ath12k/wmi.c
301
struct sk_buff *skb, gfp_t gfp)
drivers/net/wireless/ath/ath12k/wmi.c
306
tb = kzalloc_objs(*tb, WMI_TAG_MAX, gfp);
drivers/net/wireless/ath/carl9170/usb.c
316
static int carl9170_usb_submit_rx_urb(struct ar9170 *ar, gfp_t gfp)
drivers/net/wireless/ath/carl9170/usb.c
327
err = usb_submit_urb(urb, gfp);
drivers/net/wireless/ath/carl9170/usb.c
466
static struct urb *carl9170_usb_alloc_rx_urb(struct ar9170 *ar, gfp_t gfp)
drivers/net/wireless/ath/carl9170/usb.c
471
buf = kmalloc(ar->fw.rx_size, gfp);
drivers/net/wireless/ath/carl9170/usb.c
475
urb = usb_alloc_urb(0, gfp);
drivers/net/wireless/ath/wcn36xx/dxe.c
293
gfp_t gfp)
drivers/net/wireless/ath/wcn36xx/dxe.c
298
skb = alloc_skb(WCN36XX_PKT_SIZE, gfp);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c
519
gfp_t gfp = inirq ? GFP_ATOMIC : GFP_KERNEL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c
522
BCMILCP_SUBTYPE_VENDOR_LONG, gfp);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fweh.c
477
u32 packet_len, gfp_t gfp)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fweh.c
500
event = kzalloc_flex(*event, data, datalen, gfp);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fweh.h
362
u32 packet_len, gfp_t gfp);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fweh.h
367
gfp_t gfp)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fweh.h
398
brcmf_fweh_process_event(drvr, event_packet, skb->len + ETH_HLEN, gfp);
drivers/net/wireless/intel/iwlegacy/3945-rs.c
390
il3945_rs_alloc_sta(void *il_priv, struct ieee80211_sta *sta, gfp_t gfp)
drivers/net/wireless/intel/iwlegacy/4965-rs.c
2247
il4965_rs_alloc_sta(void *il_rate, struct ieee80211_sta *sta, gfp_t gfp)
drivers/net/wireless/intel/iwlwifi/dvm/rs.c
2725
gfp_t gfp)
drivers/net/wireless/intel/iwlwifi/mvm/rs.c
2724
gfp_t gfp)
drivers/net/wireless/mediatek/mt76/mcu.c
10
int len, int data_len, gfp_t gfp)
drivers/net/wireless/mediatek/mt76/mcu.c
18
skb = alloc_skb(len, gfp);
drivers/net/wireless/mediatek/mt76/mt76.h
1724
u32 val, gfp_t gfp)
drivers/net/wireless/mediatek/mt76/mt76.h
1727
gfp);
drivers/net/wireless/mediatek/mt76/mt76.h
1732
u32 *val, gfp_t gfp)
drivers/net/wireless/mediatek/mt76/mt76.h
1735
gfp);
drivers/net/wireless/mediatek/mt76/mt76.h
1868
int len, int data_len, gfp_t gfp);
drivers/net/wireless/mediatek/mt7601u/dma.c
12
struct mt7601u_dma_buf_rx *e, gfp_t gfp);
drivers/net/wireless/mediatek/mt7601u/dma.c
399
struct mt7601u_dma_buf_rx *e, gfp_t gfp)
drivers/net/wireless/mediatek/mt7601u/dma.c
412
ret = usb_submit_urb(e->urb, gfp);
drivers/net/wireless/mediatek/mt7601u/usb.c
56
struct mt7601u_dma_buf *buf, gfp_t gfp,
drivers/net/wireless/mediatek/mt7601u/usb.c
74
ret = usb_submit_urb(buf->urb, gfp);
drivers/net/wireless/mediatek/mt7601u/usb.h
60
struct mt7601u_dma_buf *buf, gfp_t gfp,
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1444
void rt2x00lib_rxdone(struct queue_entry *entry, gfp_t gfp);
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
737
void rt2x00lib_rxdone(struct queue_entry *entry, gfp_t gfp)
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
757
skb = rt2x00queue_alloc_rxskb(entry, gfp);
drivers/net/wireless/ralink/rt2x00/rt2x00lib.h
94
struct sk_buff *rt2x00queue_alloc_rxskb(struct queue_entry *entry, gfp_t gfp);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
23
struct sk_buff *rt2x00queue_alloc_rxskb(struct queue_entry *entry, gfp_t gfp)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
59
skb = __dev_alloc_skb(frame_size + head_size + tail_size, gfp);
drivers/net/wireless/realtek/rtlwifi/rc.c
276
struct ieee80211_sta *sta, gfp_t gfp)
drivers/net/wireless/realtek/rtlwifi/rc.c
281
rate_priv = kzalloc_obj(*rate_priv, gfp);
drivers/net/wireless/realtek/rtw88/usb.c
681
gfp_t gfp)
drivers/net/wireless/realtek/rtw88/usb.c
689
rx_skb = alloc_skb(RTW_USB_MAX_RECVBUF_SZ, gfp);
drivers/net/wireless/realtek/rtw88/usb.c
704
error = usb_submit_urb(rxcb->rx_urb, gfp);
drivers/net/wireless/realtek/rtw89/usb.c
463
gfp_t gfp)
drivers/net/wireless/realtek/rtw89/usb.c
471
rx_skb = alloc_skb(RTW89_USB_RECVBUF_SZ, gfp);
drivers/net/wireless/realtek/rtw89/usb.c
486
ret = usb_submit_urb(rxcb->rx_urb, gfp);
drivers/net/wwan/qcom_bam_dmux.c
487
static bool bam_dmux_skb_dma_queue_rx(struct bam_dmux_skb_dma *skb_dma, gfp_t gfp)
drivers/net/wwan/qcom_bam_dmux.c
490
skb_dma->skb = __netdev_alloc_skb(NULL, BAM_DMUX_BUFFER_SIZE, gfp);
drivers/parisc/sba_iommu.c
880
gfp_t gfp, unsigned long attrs)
drivers/parisc/sba_iommu.c
890
ret = (void *) __get_free_pages(gfp, get_order(size));
drivers/s390/net/qeth_core_main.c
2566
gfp_t gfp)
drivers/s390/net/qeth_core_main.c
2570
newbuf = kmem_cache_zalloc(qeth_qdio_outbuf_cache, gfp);
drivers/s390/net/qeth_core_main.c
3955
gfp_t gfp = GFP_ATOMIC | (skb_pfmemalloc(skb) ? __GFP_MEMALLOC : 0);
drivers/s390/net/qeth_core_main.c
4015
*hdr = kmem_cache_alloc(qeth_core_header_cache, gfp);
drivers/scsi/cxgbi/libcxgbi.h
365
static inline struct sk_buff *alloc_wr(int wrlen, int dlen, gfp_t gfp)
drivers/scsi/cxgbi/libcxgbi.h
367
struct sk_buff *skb = alloc_skb(wrlen + dlen, gfp);
drivers/scsi/libfc/fc_fcp.c
131
static struct fc_fcp_pkt *fc_fcp_pkt_alloc(struct fc_lport *lport, gfp_t gfp)
drivers/scsi/libfc/fc_fcp.c
136
fsp = mempool_alloc(si->scsi_pkt_pool, gfp);
drivers/scsi/scsi_dh.c
360
const char *scsi_dh_attached_handler_name(struct request_queue *q, gfp_t gfp)
drivers/scsi/scsi_dh.c
370
handler_name = kstrdup(sdev->handler->name, gfp) ? :
drivers/scsi/scsi_transport_iscsi.c
2532
iscsi_multicast_skb(struct sk_buff *skb, uint32_t group, gfp_t gfp)
drivers/scsi/scsi_transport_iscsi.c
2534
return nlmsg_multicast(nls, skb, 0, group, gfp);
drivers/spi/spi.c
3709
gfp_t gfp)
drivers/spi/spi.c
3719
gfp);
drivers/spi/spi.c
958
size_t size, gfp_t gfp)
drivers/spi/spi.c
962
sres = kzalloc(sizeof(*sres) + size, gfp);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
623
const gfp_t gfp = __GFP_NOWARN | __GFP_RECLAIM | __GFP_FS;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
626
ret = alloc_pages_bulk(gfp, bo->pgnr, bo->pages);
drivers/staging/media/ipu3/ipu3-dmamap.c
34
static struct page **imgu_dmamap_alloc_buffer(size_t size, gfp_t gfp)
drivers/staging/media/ipu3/ipu3-dmamap.c
47
gfp |= __GFP_HIGHMEM | __GFP_ZERO;
drivers/staging/media/ipu3/ipu3-dmamap.c
59
gfp | high_order_gfp : gfp, order);
drivers/staging/media/ipu7/ipu7-dma.c
153
dma_addr_t *dma_handle, gfp_t gfp,
drivers/staging/media/ipu7/ipu7-dma.c
179
pages = __alloc_buffer(size, gfp, attrs);
drivers/staging/media/ipu7/ipu7-dma.c
58
static struct page **__alloc_buffer(size_t size, gfp_t gfp, unsigned long attrs)
drivers/staging/media/ipu7/ipu7-dma.c
69
gfp |= __GFP_NOWARN;
drivers/staging/media/ipu7/ipu7-dma.c
74
pages[i] = alloc_pages(gfp, order);
drivers/staging/media/ipu7/ipu7-dma.c
76
pages[i] = alloc_pages(gfp, --order);
drivers/staging/media/ipu7/ipu7-dma.h
29
dma_addr_t *dma_handle, gfp_t gfp,
drivers/staging/rtl8723bs/include/ioctl_cfg80211.h
112
#define rtw_cfg80211_rx_mgmt(adapter, freq, sig_dbm, buf, len, gfp) cfg80211_rx_mgmt((adapter)->rtw_wdev, freq, sig_dbm, buf, len, 0)
drivers/staging/rtl8723bs/include/ioctl_cfg80211.h
114
#define rtw_cfg80211_mgmt_tx_status(adapter, cookie, buf, len, ack, gfp) cfg80211_mgmt_tx_status((adapter)->rtw_wdev, cookie, buf, len, ack, gfp)
drivers/staging/rtl8723bs/include/ioctl_cfg80211.h
115
#define rtw_cfg80211_ready_on_channel(adapter, cookie, chan, channel_type, duration, gfp) cfg80211_ready_on_channel((adapter)->rtw_wdev, cookie, chan, duration, gfp)
drivers/staging/rtl8723bs/include/ioctl_cfg80211.h
116
#define rtw_cfg80211_remain_on_channel_expired(adapter, cookie, chan, chan_type, gfp) cfg80211_remain_on_channel_expired((adapter)->rtw_wdev, cookie, chan, gfp)
drivers/target/target_core_transport.c
1504
target_cmd_init_cdb(struct se_cmd *cmd, unsigned char *cdb, gfp_t gfp)
drivers/target/target_core_transport.c
1525
cmd->t_task_cdb = kzalloc(scsi_command_size(cdb), gfp);
drivers/target/target_core_transport.c
1755
gfp_t gfp)
drivers/target/target_core_transport.c
1759
rc = target_cmd_init_cdb(se_cmd, cdb, gfp);
drivers/target/target_core_transport.c
1988
gfp_t gfp, u64 tag, int flags)
drivers/target/target_core_transport.c
2003
ret = core_tmr_alloc_req(se_cmd, fabric_tmr_ptr, tm_type, gfp);
drivers/target/target_core_transport.c
2779
gfp_t gfp = GFP_KERNEL | (zero_page ? __GFP_ZERO : 0);
drivers/target/target_core_transport.c
2781
*sgl = sgl_alloc_order(length, 0, chainable, gfp, nents);
drivers/usb/gadget/udc/pch_udc.c
1560
gfp_t gfp)
drivers/usb/gadget/udc/pch_udc.c
1565
retval = pch_udc_create_dma_chain(ep, req, ep->ep.maxpacket, gfp);
drivers/usb/gadget/udc/pch_udc.c
1711
gfp_t gfp)
drivers/usb/gadget/udc/pch_udc.c
1720
req = kzalloc_obj(*req, gfp);
drivers/usb/gadget/udc/pch_udc.c
1728
dma_desc = dma_pool_alloc(ep->dev->data_requests, gfp,
drivers/usb/gadget/udc/pch_udc.c
1785
gfp_t gfp)
drivers/usb/gadget/udc/snps_udc_core.c
1048
udc_queue(struct usb_ep *usbep, struct usb_request *usbreq, gfp_t gfp)
drivers/usb/gadget/udc/snps_udc_core.c
514
udc_alloc_request(struct usb_ep *usbep, gfp_t gfp)
drivers/usb/gadget/udc/snps_udc_core.c
526
req = kzalloc_obj(struct udc_request, gfp);
drivers/usb/gadget/udc/snps_udc_core.c
535
dma_desc = dma_pool_alloc(ep->dev->data_requests, gfp,
drivers/usb/gadget/udc/snps_udc_core.c
871
static int prep_dma(struct udc_ep *ep, struct udc_request *req, gfp_t gfp)
drivers/usb/gadget/udc/snps_udc_core.c
889
retval = udc_create_dma_chain(ep, req, ep->ep.maxpacket, gfp);
drivers/usb/gadget/udc/tegra-xudc.c
1311
gfp_t gfp)
drivers/usb/gadget/udc/tegra-xudc.c
1905
tegra_xudc_ep_alloc_request(struct usb_ep *usb_ep, gfp_t gfp)
drivers/usb/gadget/udc/tegra-xudc.c
1909
req = kzalloc_obj(*req, gfp);
drivers/usb/mtu3/mtu3_gadget_ep0.c
837
struct usb_request *req, gfp_t gfp)
drivers/usb/storage/uas.c
449
static struct urb *uas_alloc_data_urb(struct uas_dev_info *devinfo, gfp_t gfp,
drivers/usb/storage/uas.c
455
struct urb *urb = usb_alloc_urb(0, gfp);
drivers/usb/storage/uas.c
472
static struct urb *uas_alloc_sense_urb(struct uas_dev_info *devinfo, gfp_t gfp,
drivers/usb/storage/uas.c
477
struct urb *urb = usb_alloc_urb(0, gfp);
drivers/usb/storage/uas.c
483
iu = kzalloc_obj(*iu, gfp);
drivers/usb/storage/uas.c
499
static struct urb *uas_alloc_cmd_urb(struct uas_dev_info *devinfo, gfp_t gfp,
drivers/usb/storage/uas.c
505
struct urb *urb = usb_alloc_urb(0, gfp);
drivers/usb/storage/uas.c
516
iu = kzalloc(sizeof(*iu) + len, gfp);
drivers/usb/storage/uas.c
543
static int uas_submit_sense_urb(struct scsi_cmnd *cmnd, gfp_t gfp)
drivers/usb/storage/uas.c
549
urb = uas_alloc_sense_urb(devinfo, gfp, cmnd);
drivers/usb/storage/uas.c
553
err = usb_submit_urb(urb, gfp);
drivers/vhost/vringh.c
1408
gfp_t gfp)
drivers/vhost/vringh.c
1422
gfp, copydesc_iotlb);
drivers/vhost/vringh.c
218
static int resize_iovec(struct vringh_kiov *iov, gfp_t gfp)
drivers/vhost/vringh.c
228
new = krealloc_array(iov->iov, new_num, sizeof(*new), gfp);
drivers/vhost/vringh.c
230
new = kmalloc_objs(*new, new_num, gfp);
drivers/vhost/vringh.c
300
gfp_t gfp,
drivers/vhost/vringh.c
403
err = resize_iovec(iov, gfp);
drivers/vhost/vringh.c
949
gfp_t gfp)
drivers/vhost/vringh.c
963
gfp, copydesc_kern);
drivers/virtio/virtio_ring.c
1459
gfp_t gfp)
drivers/virtio/virtio_ring.c
1470
gfp &= ~__GFP_HIGHMEM;
drivers/virtio/virtio_ring.c
1474
desc = kmalloc(size, gfp);
drivers/virtio/virtio_ring.c
1493
gfp_t gfp,
drivers/virtio/virtio_ring.c
1505
desc = alloc_indirect_packed(total_sg, gfp);
drivers/virtio/virtio_ring.c
1623
gfp_t gfp,
drivers/virtio/virtio_ring.c
1651
in_sgs, data, premapped, gfp,
drivers/virtio/virtio_ring.c
177
void *ctx, bool premapped, gfp_t gfp,
drivers/virtio/virtio_ring.c
1781
gfp_t gfp,
drivers/virtio/virtio_ring.c
1808
in_sgs, data, premapped, gfp,
drivers/virtio/virtio_ring.c
2791
gfp_t gfp,
drivers/virtio/virtio_ring.c
2798
ctx, premapped, gfp, attr);
drivers/virtio/virtio_ring.c
2824
gfp_t gfp)
drivers/virtio/virtio_ring.c
2836
data, NULL, false, gfp, 0);
drivers/virtio/virtio_ring.c
2856
gfp_t gfp)
drivers/virtio/virtio_ring.c
2858
return virtqueue_add(vq, &sg, num, 1, 0, data, NULL, false, gfp, 0);
drivers/virtio/virtio_ring.c
2879
gfp_t gfp)
drivers/virtio/virtio_ring.c
2881
return virtqueue_add(vq, &sg, num, 1, 0, data, NULL, true, gfp, 0);
drivers/virtio/virtio_ring.c
2901
gfp_t gfp)
drivers/virtio/virtio_ring.c
2903
return virtqueue_add(vq, &sg, num, 0, 1, data, NULL, false, gfp, 0);
drivers/virtio/virtio_ring.c
2928
gfp_t gfp)
drivers/virtio/virtio_ring.c
2930
return virtqueue_add(vq, &sg, num, 0, 1, data, NULL, false, gfp,
drivers/virtio/virtio_ring.c
2953
gfp_t gfp)
drivers/virtio/virtio_ring.c
2955
return virtqueue_add(vq, &sg, num, 0, 1, data, ctx, false, gfp, 0);
drivers/virtio/virtio_ring.c
2978
gfp_t gfp)
drivers/virtio/virtio_ring.c
2980
return virtqueue_add(vq, &sg, num, 0, 1, data, ctx, true, gfp, 0);
drivers/virtio/virtio_ring.c
3686
gfp_t gfp)
drivers/virtio/virtio_ring.c
3690
map_handle, gfp);
drivers/virtio/virtio_ring.c
3693
map_handle, gfp);
drivers/virtio/virtio_ring.c
546
gfp_t gfp)
drivers/virtio/virtio_ring.c
557
gfp &= ~__GFP_HIGHMEM;
drivers/virtio/virtio_ring.c
561
desc = kmalloc(size, gfp);
drivers/virtio/virtio_ring.c
607
gfp_t gfp,
drivers/virtio/virtio_ring.c
636
desc = alloc_indirect_split(vq, total_sg, gfp);
drivers/xen/balloon.c
433
static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp)
drivers/xen/balloon.c
445
page = alloc_page(gfp);
drivers/xen/grant-dma-ops.c
148
gfp_t gfp)
drivers/xen/grant-dma-ops.c
152
vaddr = xen_grant_dma_alloc(dev, size, dma_handle, gfp, 0);
drivers/xen/grant-dma-ops.c
79
dma_addr_t *dma_handle, gfp_t gfp,
drivers/xen/grant-dma-ops.c
95
ret = alloc_pages_exact(n_pages * XEN_PAGE_SIZE, gfp);
drivers/xen/grant-table.c
556
gfp_t gfp = (in_atomic() || irqs_disabled()) ? GFP_ATOMIC : GFP_KERNEL;
drivers/xen/grant-table.c
559
entry = kmalloc_obj(*entry, gfp);
drivers/xen/xenbus/xenbus_client.c
395
int xenbus_setup_ring(struct xenbus_device *dev, gfp_t gfp, void **vaddr,
drivers/xen/xenbus/xenbus_client.c
404
addr = *vaddr = alloc_pages_exact(ring_size, gfp | __GFP_ZERO);
fs/afs/internal.h
1383
void afs_make_call(struct afs_call *call, gfp_t gfp);
fs/afs/internal.h
1418
gfp_t gfp)
fs/afs/internal.h
1429
afs_make_call(call, gfp);
fs/afs/rxrpc.c
158
gfp_t gfp)
fs/afs/rxrpc.c
163
call = kzalloc_obj(*call, gfp);
fs/afs/rxrpc.c
337
void afs_make_call(struct afs_call *call, gfp_t gfp)
fs/afs/rxrpc.c
378
gfp,
fs/btrfs/delayed-inode.h
220
gfp_t gfp)
fs/btrfs/delayed-inode.h
225
return ref_tracker_alloc(&node->ref_dir.dir, &tracker->tracker, gfp);
fs/btrfs/delayed-inode.h
245
gfp_t gfp)
fs/btrfs/extent_io.c
675
const gfp_t gfp = nofail ? (GFP_NOFS | __GFP_NOFAIL) : GFP_NOFS;
fs/btrfs/extent_io.c
681
allocated = alloc_pages_bulk(gfp, nr_pages, page_array);
fs/btrfs/file.c
840
gfp_t gfp;
fs/btrfs/file.c
842
gfp = btrfs_alloc_write_mask(inode->i_mapping);
fs/btrfs/file.c
844
gfp &= ~__GFP_DIRECT_RECLAIM;
fs/btrfs/file.c
845
gfp |= GFP_NOWAIT;
fs/btrfs/file.c
848
return gfp;
fs/btrfs/lru_cache.c
108
gfp_t gfp)
fs/btrfs/lru_cache.c
114
head = kmalloc_obj(*head, gfp);
fs/btrfs/lru_cache.c
118
ret = mtree_insert(&cache->entries, key, head, gfp);
fs/btrfs/lru_cache.h
66
gfp_t gfp);
fs/btrfs/volumes.c
5789
struct btrfs_chunk_map *btrfs_alloc_chunk_map(int num_stripes, gfp_t gfp)
fs/btrfs/volumes.c
5793
map = kmalloc(btrfs_chunk_map_size(num_stripes), gfp);
fs/btrfs/volumes.h
785
struct btrfs_chunk_map *btrfs_alloc_chunk_map(int num_stripes, gfp_t gfp);
fs/buffer.c
1040
pgoff_t index, unsigned size, gfp_t gfp)
fs/buffer.c
1048
FGP_LOCK | FGP_ACCESSED | FGP_CREAT, gfp);
fs/buffer.c
1072
bh = folio_alloc_buffers(folio, size, gfp | __GFP_ACCOUNT);
fs/buffer.c
1097
unsigned size, gfp_t gfp)
fs/buffer.c
1113
return grow_dev_folio(bdev, block, pos / PAGE_SIZE, size, gfp);
fs/buffer.c
1118
unsigned size, gfp_t gfp)
fs/buffer.c
1120
bool blocking = gfpflags_allow_blocking(gfp);
fs/buffer.c
1131
if (!grow_buffers(bdev, block, size, gfp))
fs/buffer.c
1445
unsigned size, gfp_t gfp)
fs/buffer.c
1449
if (gfpflags_allow_blocking(gfp))
fs/buffer.c
1454
might_alloc(gfp);
fs/buffer.c
1458
return __getblk_slow(bdev, block, size, gfp);
fs/buffer.c
1500
unsigned size, gfp_t gfp)
fs/buffer.c
1504
gfp |= mapping_gfp_constraint(bdev->bd_mapping, ~__GFP_FS);
fs/buffer.c
1510
gfp |= __GFP_NOFAIL;
fs/buffer.c
1512
bh = bdev_getblk(bdev, block, size, gfp);
fs/buffer.c
1686
gfp_t gfp = GFP_NOFS | __GFP_ACCOUNT | __GFP_NOFAIL;
fs/buffer.c
1688
head = folio_alloc_buffers(folio, blocksize, gfp);
fs/buffer.c
919
gfp_t gfp)
fs/buffer.c
932
bh = alloc_buffer_head(gfp);
fs/buffer.c
966
gfp_t gfp = GFP_NOFS | __GFP_ACCOUNT;
fs/buffer.c
968
return folio_alloc_buffers(page_folio(page), size, gfp);
fs/erofs/compress.h
20
gfp_t gfp; /* allocation flags for extra temporary buffers */
fs/erofs/decompressor.c
101
victim = __erofs_allocpage(pagepool, rq->gfp, true);
fs/erofs/decompressor.c
359
*pgo = erofs_allocpage(pgpl, rq->gfp);
fs/erofs/decompressor.c
410
tmppage = erofs_allocpage(pgpl, rq->gfp);
fs/erofs/decompressor_crypto.c
110
victim = __erofs_allocpage(pgpl, rq->gfp, true);
fs/erofs/internal.h
499
struct page *__erofs_allocpage(struct page **pagepool, gfp_t gfp, bool tryrsv);
fs/erofs/internal.h
500
static inline struct page *erofs_allocpage(struct page **pagepool, gfp_t gfp)
fs/erofs/internal.h
502
return __erofs_allocpage(pagepool, gfp, false);
fs/erofs/zdata.c
1311
.gfp = pcl->besteffort ? GFP_KERNEL :
fs/erofs/zdata.c
1492
gfp_t gfp = mapping_gfp_mask(mc);
fs/erofs/zdata.c
1563
page = __erofs_allocpage(&f->pagepool, gfp, true);
fs/erofs/zdata.c
1580
filemap_add_folio(mc, folio, (pcl->pos >> PAGE_SHIFT) + nr, gfp)) {
fs/erofs/zdata.c
542
gfp_t gfp = (mapping_gfp_mask(mc) & ~__GFP_DIRECT_RECLAIM) |
fs/erofs/zdata.c
566
newfolio = filemap_alloc_folio(gfp, 0, NULL);
fs/erofs/zdata.c
618
static bool z_erofs_cache_release_folio(struct folio *folio, gfp_t gfp)
fs/erofs/zutil.c
176
struct page *__erofs_allocpage(struct page **pagepool, gfp_t gfp, bool tryrsv)
fs/erofs/zutil.c
189
page = alloc_page(gfp);
fs/ext4/mballoc.c
1373
static int ext4_mb_init_cache(struct folio *folio, char *incore, gfp_t gfp)
fs/ext4/mballoc.c
1404
bh = kzalloc(i, gfp);
fs/ext4/mballoc.c
1543
ext4_group_t group, struct ext4_buddy *e4b, gfp_t gfp)
fs/ext4/mballoc.c
1560
FGP_LOCK | FGP_ACCESSED | FGP_CREAT, gfp);
fs/ext4/mballoc.c
1578
FGP_LOCK | FGP_ACCESSED | FGP_CREAT, gfp);
fs/ext4/mballoc.c
1605
int ext4_mb_init_group(struct super_block *sb, ext4_group_t group, gfp_t gfp)
fs/ext4/mballoc.c
1628
ret = ext4_mb_get_buddy_folio_lock(sb, group, &e4b, gfp);
fs/ext4/mballoc.c
1638
ret = ext4_mb_init_cache(folio, NULL, gfp);
fs/ext4/mballoc.c
1657
ret = ext4_mb_init_cache(folio, e4b.bd_bitmap, gfp);
fs/ext4/mballoc.c
1676
struct ext4_buddy *e4b, gfp_t gfp)
fs/ext4/mballoc.c
1705
ret = ext4_mb_init_group(sb, group, gfp);
fs/ext4/mballoc.c
1733
FGP_LOCK | FGP_ACCESSED | FGP_CREAT, gfp);
fs/ext4/mballoc.c
1743
ret = ext4_mb_init_cache(folio, NULL, gfp);
fs/ext4/mballoc.c
1783
FGP_LOCK | FGP_ACCESSED | FGP_CREAT, gfp);
fs/ext4/mballoc.c
1794
gfp);
fs/ext4/super.c
232
blk_opf_t op_flags, gfp_t gfp)
fs/ext4/super.c
237
bh = sb_getblk_gfp(sb, block, gfp);
fs/ext4/super.c
254
gfp_t gfp = mapping_gfp_constraint(sb->s_bdev->bd_mapping,
fs/ext4/super.c
257
return __ext4_sb_bread_gfp(sb, block, op_flags, gfp);
fs/ext4/super.c
263
gfp_t gfp = mapping_gfp_constraint(sb->s_bdev->bd_mapping,
fs/ext4/super.c
266
return __ext4_sb_bread_gfp(sb, block, 0, gfp);
fs/ext4/super.c
272
gfp_t gfp = mapping_gfp_constraint(sb->s_bdev->bd_mapping,
fs/ext4/super.c
275
return __ext4_sb_bread_gfp(sb, block, 0, gfp);
fs/fserror.c
135
gfp_t gfp)
fs/fserror.c
145
event = fserror_alloc_event(sb, gfp);
fs/fuse/virtio_fs.c
101
gfp_t gfp);
fs/fuse/virtio_fs.c
1372
gfp_t gfp)
fs/fuse/virtio_fs.c
1393
sgs = kmalloc_objs(sgs[0], total_sgs, gfp);
fs/fuse/virtio_fs.c
1394
sg = kmalloc_objs(sg[0], total_sgs, gfp);
fs/fuse/virtio_fs.c
1402
ret = copy_args_to_argbuf(req, gfp);
fs/fuse/virtio_fs.c
695
static int copy_args_to_argbuf(struct fuse_req *req, gfp_t gfp)
fs/fuse/virtio_fs.c
709
req->argbuf = kmalloc(len, gfp);
fs/inode.c
228
int inode_init_always_gfp(struct super_block *sb, struct inode *inode, gfp_t gfp)
fs/inode.c
307
if (unlikely(security_inode_alloc(inode, gfp)))
fs/iomap/bio.c
100
gfp |= __GFP_NORETRY | __GFP_NOWARN;
fs/iomap/bio.c
102
gfp);
fs/iomap/bio.c
92
gfp_t gfp = mapping_gfp_constraint(folio->mapping, GFP_KERNEL);
fs/iomap/bio.c
93
gfp_t orig_gfp = gfp;
fs/iomap/buffered-io.c
225
gfp_t gfp;
fs/iomap/buffered-io.c
231
gfp = GFP_NOWAIT;
fs/iomap/buffered-io.c
233
gfp = GFP_NOFS | __GFP_NOFAIL;
fs/iomap/buffered-io.c
241
ifs = kzalloc_flex(*ifs, state, BITS_TO_LONGS(2 * nr_blocks), gfp);
fs/mpage.c
171
gfp_t gfp = mapping_gfp_constraint(folio->mapping, GFP_KERNEL);
fs/mpage.c
175
gfp |= __GFP_NORETRY | __GFP_NOWARN;
fs/mpage.c
288
gfp);
fs/netfs/misc.c
21
size_t *_cur_size, ssize_t size, gfp_t gfp)
fs/netfs/misc.c
293
bool netfs_release_folio(struct folio *folio, gfp_t gfp)
fs/netfs/misc.c
308
if (current_is_kswapd() || !(gfp & __GFP_FS))
fs/netfs/misc.c
54
folio = folio_alloc(gfp, order);
fs/netfs/misc.c
56
folio = folio_alloc(gfp, 0);
fs/netfs/rolling_buffer.c
25
struct folio_queue *netfs_folioq_alloc(unsigned int rreq_id, gfp_t gfp,
fs/netfs/rolling_buffer.c
30
fq = kmalloc_obj(*fq, gfp);
fs/nfs/file.c
505
static bool nfs_release_folio(struct folio *folio, gfp_t gfp)
fs/nfs/file.c
511
if ((current_gfp_context(gfp) & GFP_KERNEL) != GFP_KERNEL ||
fs/nfs/file.c
518
return nfs_fscache_release_folio(folio, gfp);
fs/nfs/fscache.h
102
static inline bool nfs_fscache_release_folio(struct folio *folio, gfp_t gfp)
fs/nfs/fscache.h
105
if (current_is_kswapd() || !(gfp & __GFP_FS))
fs/nfs/fscache.h
186
static inline bool nfs_fscache_release_folio(struct folio *folio, gfp_t gfp)
fs/nfsd/nfs4state.c
2023
int index, gfp_t gfp)
fs/nfsd/nfs4state.c
2035
slot = kzalloc_flex(*slot, sl_data, size, gfp);
fs/nfsd/nfs4state.c
2060
const gfp_t gfp = GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN;
fs/nfsd/nfs4state.c
2061
slot = nfsd4_alloc_slot(fattrs, i, gfp);
fs/nfsd/nfs4state.c
2064
if (xa_is_err(xa_store(&new->se_slots, i, slot, gfp))) {
fs/notify/fanotify/fanotify.c
414
gfp_t gfp)
fs/notify/fanotify/fanotify.c
442
if (gfp && fh_len > FANOTIFY_INLINE_FH_LEN) {
fs/notify/fanotify/fanotify.c
445
ext_buf = kmalloc(fh_len, gfp);
fs/notify/fanotify/fanotify.c
555
gfp_t gfp)
fs/notify/fanotify/fanotify.c
559
pevent = kmem_cache_alloc(fanotify_path_event_cachep, gfp);
fs/notify/fanotify/fanotify.c
571
static struct fanotify_event *fanotify_alloc_mnt_event(u64 mnt_id, gfp_t gfp)
fs/notify/fanotify/fanotify.c
575
pevent = kmem_cache_alloc(fanotify_mnt_event_cachep, gfp);
fs/notify/fanotify/fanotify.c
587
gfp_t gfp)
fs/notify/fanotify/fanotify.c
594
pevent = kmem_cache_alloc(fanotify_perm_event_cachep, gfp);
fs/notify/fanotify/fanotify.c
616
gfp_t gfp)
fs/notify/fanotify/fanotify.c
620
ffe = kmem_cache_alloc(fanotify_fid_event_cachep, gfp);
fs/notify/fanotify/fanotify.c
628
hash, gfp);
fs/notify/fanotify/fanotify.c
639
gfp_t gfp)
fs/notify/fanotify/fanotify.c
661
fne = kmalloc(size, gfp);
fs/notify/fanotify/fanotify.c
753
gfp_t gfp = GFP_KERNEL_ACCOUNT;
fs/notify/fanotify/fanotify.c
838
gfp |= __GFP_NOFAIL;
fs/notify/fanotify/fanotify.c
840
gfp |= __GFP_RETRY_MAYFAIL;
fs/notify/fanotify/fanotify.c
846
event = fanotify_alloc_perm_event(data, data_type, gfp);
fs/notify/fanotify/fanotify.c
852
moved, &hash, gfp);
fs/notify/fanotify/fanotify.c
854
event = fanotify_alloc_fid_event(id, fsid, &hash, gfp);
fs/notify/fanotify/fanotify.c
856
event = fanotify_alloc_path_event(path, &hash, gfp);
fs/notify/fanotify/fanotify.c
858
event = fanotify_alloc_mnt_event(mnt_id, gfp);
fs/notify/group.c
116
int flags, gfp_t gfp)
fs/notify/group.c
120
group = kzalloc_obj(struct fsnotify_group, gfp);
fs/notify/group.c
148
gfp_t gfp = (flags & FSNOTIFY_GROUP_USER) ? GFP_KERNEL_ACCOUNT :
fs/notify/group.c
151
return __fsnotify_alloc_group(ops, flags, gfp);
fs/ntfs3/frecord.c
1856
pgoff_t index, gfp_t gfp)
fs/ntfs3/frecord.c
1859
mapping, index, FGP_LOCK | FGP_ACCESSED | FGP_CREAT, gfp);
fs/ntfs3/frecord.c
1871
page = alloc_page(gfp);
fs/ntfs3/inode.c
626
gfp_t gfp = mapping_gfp_constraint(folio->mapping, GFP_KERNEL);
fs/ntfs3/inode.c
627
gfp_t orig_gfp = gfp;
fs/ntfs3/inode.c
634
gfp |= __GFP_NORETRY | __GFP_NOWARN;
fs/ntfs3/inode.c
636
gfp);
fs/posix_acl.c
555
__posix_acl_create(struct posix_acl **acl, gfp_t gfp, umode_t *mode_p)
fs/posix_acl.c
557
struct posix_acl *clone = posix_acl_clone(*acl, gfp);
fs/posix_acl.c
573
__posix_acl_chmod(struct posix_acl **acl, gfp_t gfp, umode_t mode)
fs/posix_acl.c
575
struct posix_acl *clone = posix_acl_clone(*acl, gfp);
fs/posix_acl.c
834
size_t *sizep, gfp_t gfp)
fs/posix_acl.c
842
ext_acl = kmalloc(size, gfp);
fs/ramfs/file-nommu.c
108
gfp);
fs/ramfs/file-nommu.c
69
gfp_t gfp = mapping_gfp_mask(inode->i_mapping);
fs/ramfs/file-nommu.c
84
pages = alloc_pages(gfp, order);
fs/smb/client/fs_context.c
582
char *cifs_sanitize_prepath(char *prepath, gfp_t gfp)
fs/smb/client/fs_context.c
613
s = kstrdup(prepath, gfp);
fs/smb/client/fs_context.h
385
char *cifs_sanitize_prepath(char *prepath, gfp_t gfp);
include/asm-generic/pgalloc.h
139
gfp_t gfp = GFP_PGTABLE_USER;
include/asm-generic/pgalloc.h
142
gfp = GFP_PGTABLE_KERNEL;
include/asm-generic/pgalloc.h
143
ptdesc = pagetable_alloc_noprof(gfp, 0);
include/asm-generic/pgalloc.h
175
gfp_t gfp = GFP_PGTABLE_USER;
include/asm-generic/pgalloc.h
179
gfp = GFP_PGTABLE_KERNEL;
include/asm-generic/pgalloc.h
181
ptdesc = pagetable_alloc_noprof(gfp, 0);
include/asm-generic/pgalloc.h
232
gfp_t gfp = GFP_PGTABLE_USER;
include/asm-generic/pgalloc.h
236
gfp = GFP_PGTABLE_KERNEL;
include/asm-generic/pgalloc.h
238
ptdesc = pagetable_alloc_noprof(gfp, 0);
include/asm-generic/pgalloc.h
279
gfp_t gfp = GFP_PGTABLE_USER;
include/asm-generic/pgalloc.h
283
gfp = GFP_PGTABLE_KERNEL;
include/asm-generic/pgalloc.h
285
ptdesc = pagetable_alloc_noprof(gfp, order);
include/asm-generic/pgalloc.h
72
static inline pgtable_t __pte_alloc_one_noprof(struct mm_struct *mm, gfp_t gfp)
include/asm-generic/pgalloc.h
76
ptdesc = pagetable_alloc_noprof(gfp, 0);
include/crypto/acompress.h
251
struct crypto_acomp *tfm, size_t extra, gfp_t gfp)
include/crypto/acompress.h
260
req = kzalloc_noprof(len, gfp);
include/crypto/acompress.h
270
#define acomp_request_alloc_noprof_1(tfm, gfp) \
include/crypto/acompress.h
271
acomp_request_alloc_extra_noprof(tfm, 0, gfp)
include/crypto/acompress.h
50
#define ACOMP_REQUEST_CLONE(name, gfp) \
include/crypto/acompress.h
51
acomp_request_clone(name, sizeof(__##name##_req), gfp)
include/crypto/acompress.h
555
size_t total, gfp_t gfp);
include/crypto/aead.h
519
gfp_t gfp)
include/crypto/aead.h
523
req = kmalloc(sizeof(*req) + crypto_aead_reqsize(tfm), gfp);
include/crypto/akcipher.h
183
struct crypto_akcipher *tfm, gfp_t gfp)
include/crypto/akcipher.h
187
req = kmalloc(sizeof(*req) + crypto_akcipher_reqsize(tfm), gfp);
include/crypto/hash.h
1049
struct ahash_request *req, size_t total, gfp_t gfp)
include/crypto/hash.h
1051
return container_of(crypto_request_clone(&req->base, total, gfp),
include/crypto/hash.h
214
#define HASH_REQUEST_CLONE(name, gfp) \
include/crypto/hash.h
215
hash_request_clone(name, sizeof(__##name##_req), gfp)
include/crypto/hash.h
619
struct crypto_ahash *tfm, gfp_t gfp)
include/crypto/hash.h
624
crypto_ahash_reqsize(tfm), gfp);
include/crypto/kpp.h
181
gfp_t gfp)
include/crypto/kpp.h
185
req = kmalloc(sizeof(*req) + crypto_kpp_reqsize(tfm), gfp);
include/crypto/krb5.h
129
u32 usage, gfp_t gfp);
include/crypto/krb5.h
132
u32 usage, gfp_t gfp);
include/crypto/krb5.h
163
gfp_t gfp);
include/crypto/skcipher.h
847
struct crypto_skcipher *tfm, gfp_t gfp)
include/crypto/skcipher.h
852
crypto_skcipher_reqsize(tfm), gfp);
include/drm/drm_managed.h
108
char *drmm_kstrdup(struct drm_device *dev, const char *s, gfp_t gfp);
include/drm/drm_managed.h
52
void *drmm_kmalloc(struct drm_device *dev, size_t size, gfp_t gfp) __malloc;
include/drm/drm_managed.h
64
static inline void *drmm_kzalloc(struct drm_device *dev, size_t size, gfp_t gfp)
include/drm/drm_managed.h
66
return drmm_kmalloc(dev, size, gfp | __GFP_ZERO);
include/drm/drm_suballoc.h
58
gfp_t gfp, bool intr, size_t align);
include/kunit/skbuff.h
31
gfp_t gfp)
include/kunit/skbuff.h
33
struct sk_buff *res = alloc_skb(len, gfp);
include/kunit/test.h
486
void *kunit_kmalloc_array(struct kunit *test, size_t n, size_t size, gfp_t gfp);
include/kunit/test.h
499
static inline void *kunit_kmalloc(struct kunit *test, size_t size, gfp_t gfp)
include/kunit/test.h
501
return kunit_kmalloc_array(test, 1, size, gfp);
include/kunit/test.h
519
static inline void *kunit_kzalloc(struct kunit *test, size_t size, gfp_t gfp)
include/kunit/test.h
521
return kunit_kmalloc(test, size, gfp | __GFP_ZERO);
include/kunit/test.h
533
static inline void *kunit_kcalloc(struct kunit *test, size_t n, size_t size, gfp_t gfp)
include/kunit/test.h
535
return kunit_kmalloc_array(test, n, size, gfp | __GFP_ZERO);
include/kunit/test.h
558
static inline char *kunit_kstrdup(struct kunit *test, const char *str, gfp_t gfp)
include/kunit/test.h
567
buf = kunit_kmalloc(test, len, gfp);
include/kunit/test.h
584
const char *kunit_kstrdup_const(struct kunit *test, const char *str, gfp_t gfp);
include/linux/audit.h
442
enum audit_nfcfgop op, gfp_t gfp);
include/linux/audit.h
586
enum audit_nfcfgop op, gfp_t gfp)
include/linux/audit.h
589
__audit_log_nfcfg(name, af, nentries, op, gfp);
include/linux/audit.h
721
enum audit_nfcfgop op, gfp_t gfp)
include/linux/backing-dev.h
145
gfp_t gfp);
include/linux/backing-dev.h
208
wb_get_create_current(struct backing_dev_info *bdi, gfp_t gfp)
include/linux/backing-dev.h
222
wb = wb_get_create(bdi, memcg_css, gfp);
include/linux/backing-dev.h
323
wb_get_create_current(struct backing_dev_info *bdi, gfp_t gfp)
include/linux/bio-integrity.h
133
bio_integrity_alloc(struct bio *bio, gfp_t gfp, unsigned int nr)
include/linux/bio-integrity.h
74
struct bio_integrity_payload *bio_integrity_alloc(struct bio *bio, gfp_t gfp,
include/linux/bio.h
318
gfp_t gfp, struct bio_set *bs);
include/linux/bio.h
335
gfp_t gfp, struct bio_set *bs)
include/linux/bio.h
340
return bio_split(bio, sectors, gfp, bs);
include/linux/bio.h
359
gfp_t gfp, struct bio_set *bs);
include/linux/bio.h
361
struct bio *bio_src, gfp_t gfp);
include/linux/bio.h
741
unsigned int nr_pages, blk_opf_t opf, gfp_t gfp);
include/linux/blk-mq.h
1077
gfp_t gfp);
include/linux/blkdev.h
1265
sector_t nr_sects, gfp_t gfp);
include/linux/btree-128.h
42
void *val, gfp_t gfp)
include/linux/btree-128.h
46
(unsigned long *)&key, val, gfp);
include/linux/btree-128.h
79
gfp_t gfp)
include/linux/btree-128.h
81
return btree_merge(&target->h, &victim->h, &btree_geo128, gfp);
include/linux/btree-type.h
32
gfp_t gfp)
include/linux/btree-type.h
34
return btree_merge(&target->h, &victim->h, BTREE_TYPE_GEO, gfp);
include/linux/btree-type.h
45
void *val, gfp_t gfp)
include/linux/btree-type.h
48
return btree_insert(&head->h, BTREE_TYPE_GEO, &_key, val, gfp);
include/linux/btree-type.h
88
void *val, gfp_t gfp)
include/linux/btree-type.h
91
val, gfp);
include/linux/btree.h
116
unsigned long *key, void *val, gfp_t gfp);
include/linux/btree.h
159
struct btree_geo *geo, gfp_t gfp);
include/linux/buffer_head.h
201
gfp_t gfp);
include/linux/buffer_head.h
229
unsigned size, gfp_t gfp);
include/linux/buffer_head.h
234
sector_t block, unsigned size, gfp_t gfp);
include/linux/buffer_head.h
364
gfp_t gfp;
include/linux/buffer_head.h
366
gfp = mapping_gfp_constraint(bdev->bd_mapping, ~__GFP_FS);
include/linux/buffer_head.h
367
gfp |= __GFP_NOFAIL;
include/linux/buffer_head.h
369
return bdev_getblk(bdev, block, size, gfp);
include/linux/buffer_head.h
375
gfp_t gfp;
include/linux/buffer_head.h
377
gfp = mapping_gfp_constraint(bdev->bd_mapping, ~__GFP_FS);
include/linux/buffer_head.h
378
gfp |= __GFP_MOVABLE | __GFP_NOFAIL;
include/linux/buffer_head.h
380
return bdev_getblk(bdev, block, size, gfp);
include/linux/buffer_head.h
390
sector_t block, gfp_t gfp)
include/linux/buffer_head.h
392
return bdev_getblk(sb->s_bdev, block, sb->s_blocksize, gfp);
include/linux/ceph/buffer.h
22
extern struct ceph_buffer *ceph_buffer_new(size_t len, gfp_t gfp);
include/linux/ceph/decode.h
108
size_t *lenp, gfp_t gfp)
include/linux/ceph/decode.h
118
buf = kmalloc(len + 1, gfp);
include/linux/ceph/osd_client.h
547
int ceph_osdc_alloc_messages(struct ceph_osd_request *req, gfp_t gfp);
include/linux/ceph/osdmap.h
136
int ceph_oid_aprintf(struct ceph_object_id *oid, gfp_t gfp,
include/linux/crypto.h
518
struct crypto_async_request *req, size_t total, gfp_t gfp);
include/linux/devcoredump.h
117
void *data, size_t datalen, gfp_t gfp,
include/linux/devcoredump.h
122
dev_coredumpm_timeout(dev, owner, data, datalen, gfp, read, free,
include/linux/devcoredump.h
58
gfp_t gfp);
include/linux/devcoredump.h
61
void *data, size_t datalen, gfp_t gfp,
include/linux/devcoredump.h
69
size_t datalen, gfp_t gfp);
include/linux/devcoredump.h
74
size_t datalen, gfp_t gfp)
include/linux/devcoredump.h
81
void *data, size_t datalen, gfp_t gfp,
include/linux/devcoredump.h
92
size_t datalen, gfp_t gfp)
include/linux/device/devres.h
23
__devres_alloc_node(dr_release_t release, size_t size, gfp_t gfp, int nid, const char *name);
include/linux/device/devres.h
24
#define devres_alloc(release, size, gfp) \
include/linux/device/devres.h
25
__devres_alloc_node(release, size, gfp, NUMA_NO_NODE, #release)
include/linux/device/devres.h
26
#define devres_alloc_node(release, size, gfp, nid) \
include/linux/device/devres.h
27
__devres_alloc_node(release, size, gfp, nid, #release)
include/linux/device/devres.h
38
void * __must_check devres_open_group(struct device *dev, void *id, gfp_t gfp);
include/linux/device/devres.h
45
devm_kmalloc(struct device *dev, size_t size, gfp_t gfp);
include/linux/device/devres.h
47
devm_krealloc(struct device *dev, void *ptr, size_t size, gfp_t gfp);
include/linux/device/devres.h
48
static inline void *devm_kzalloc(struct device *dev, size_t size, gfp_t gfp)
include/linux/device/devres.h
50
return devm_kmalloc(dev, size, gfp | __GFP_ZERO);
include/linux/device/devres.h
79
devm_kmemdup(struct device *dev, const void *src, size_t len, gfp_t gfp);
include/linux/device/devres.h
81
devm_kmemdup_const(struct device *dev, const void *src, size_t len, gfp_t gfp);
include/linux/device/devres.h
89
devm_kstrdup(struct device *dev, const char *s, gfp_t gfp);
include/linux/device/devres.h
90
const char *devm_kstrdup_const(struct device *dev, const char *s, gfp_t gfp);
include/linux/device/devres.h
92
devm_kvasprintf(struct device *dev, gfp_t gfp, const char *fmt, va_list ap);
include/linux/device/devres.h
94
devm_kasprintf(struct device *dev, gfp_t gfp, const char *fmt, ...);
include/linux/dma-direct.h
143
gfp_t gfp, unsigned long attrs);
include/linux/dma-direct.h
147
dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp);
include/linux/dma-map-ops.h
111
struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp);
include/linux/dma-map-ops.h
141
gfp_t gfp)
include/linux/dma-map-ops.h
18
dma_addr_t *dma_handle, gfp_t gfp,
include/linux/dma-map-ops.h
208
dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp);
include/linux/dma-map-ops.h
24
gfp_t gfp);
include/linux/dma-map-ops.h
314
gfp_t gfp, unsigned long attrs);
include/linux/dma-mapping.h
176
gfp_t gfp, unsigned long attrs);
include/linux/dma-mapping.h
195
enum dma_data_direction dir, gfp_t gfp, unsigned long attrs);
include/linux/dma-mapping.h
263
dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
include/linux/dma-mapping.h
320
size_t size, enum dma_data_direction dir, gfp_t gfp,
include/linux/dma-mapping.h
490
dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp);
include/linux/dma-mapping.h
497
dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp)
include/linux/dma-mapping.h
499
struct page *page = dma_alloc_pages(dev, size, dma_handle, dir, gfp);
include/linux/dma-mapping.h
605
dma_addr_t *dma_handle, gfp_t gfp)
include/linux/dma-mapping.h
607
return dma_alloc_attrs(dev, size, dma_handle, gfp,
include/linux/dma-mapping.h
608
(gfp & __GFP_NOWARN) ? DMA_ATTR_NO_WARN : 0);
include/linux/dma-mapping.h
734
dma_addr_t *dma_handle, gfp_t gfp)
include/linux/dma-mapping.h
736
return dmam_alloc_attrs(dev, size, dma_handle, gfp,
include/linux/dma-mapping.h
737
(gfp & __GFP_NOWARN) ? DMA_ATTR_NO_WARN : 0);
include/linux/dma-mapping.h
741
dma_addr_t *dma_addr, gfp_t gfp)
include/linux/dma-mapping.h
745
if (gfp & __GFP_NOWARN)
include/linux/dma-mapping.h
748
return dma_alloc_attrs(dev, size, dma_addr, gfp, attrs);
include/linux/firmware.h
103
struct device *device, gfp_t gfp, void *context,
include/linux/firmware.h
111
const char *name, struct device *device, gfp_t gfp, void *context,
include/linux/firmware.h
132
struct device *device, gfp_t gfp, void *context,
include/linux/firmware.h
154
const char *name, struct device *device, gfp_t gfp, void *context,
include/linux/firmware/qcom/qcom_tzmem.h
58
void *qcom_tzmem_alloc(struct qcom_tzmem_pool *pool, size_t size, gfp_t gfp);
include/linux/flex_proportions.h
37
int fprop_global_init(struct fprop_global *p, gfp_t gfp);
include/linux/flex_proportions.h
52
int fprop_local_init_percpu(struct fprop_local_percpu *pl, gfp_t gfp);
include/linux/fortify-string.h
752
extern void *__real_kmemdup(const void *src, size_t len, gfp_t gfp) __RENAME(kmemdup_noprof)
include/linux/fortify-string.h
754
__FORTIFY_INLINE void *kmemdup_noprof(const void * const POS0 p, size_t size, gfp_t gfp)
include/linux/fortify-string.h
762
__real_kmemdup(p, 0, gfp));
include/linux/fortify-string.h
763
return __real_kmemdup(p, size, gfp);
include/linux/fserror.h
42
gfp_t gfp);
include/linux/fserror.h
46
u64 len, int error, gfp_t gfp)
include/linux/fserror.h
48
fserror_report(inode->i_sb, inode, type, pos, len, error, gfp);
include/linux/fserror.h
52
u64 len, gfp_t gfp)
include/linux/fserror.h
55
gfp);
include/linux/fserror.h
59
gfp_t gfp)
include/linux/fserror.h
61
fserror_report(inode->i_sb, inode, FSERR_METADATA, 0, 0, error, gfp);
include/linux/fserror.h
65
gfp_t gfp)
include/linux/fserror.h
67
fserror_report(sb, NULL, FSERR_METADATA, 0, 0, error, gfp);
include/linux/fserror.h
70
static inline void fserror_report_shutdown(struct super_block *sb, gfp_t gfp)
include/linux/fserror.h
72
fserror_report(sb, NULL, FSERR_METADATA, 0, 0, -ESHUTDOWN, gfp);
include/linux/generic_pt/iommu.h
200
int prot, gfp_t gfp, size_t *mapped); \
include/linux/generic_pt/iommu.h
210
gfp_t gfp); \
include/linux/gfp.h
229
struct page *__alloc_pages_noprof(gfp_t gfp, unsigned int order, int preferred_nid,
include/linux/gfp.h
233
struct folio *__folio_alloc_noprof(gfp_t gfp, unsigned int order, int preferred_nid,
include/linux/gfp.h
237
unsigned long alloc_pages_bulk_noprof(gfp_t gfp, int preferred_nid,
include/linux/gfp.h
242
unsigned long alloc_pages_bulk_mempolicy_noprof(gfp_t gfp,
include/linux/gfp.h
253
alloc_pages_bulk_node_noprof(gfp_t gfp, int nid, unsigned long nr_pages,
include/linux/gfp.h
259
return alloc_pages_bulk_noprof(gfp, nid, NULL, nr_pages, page_array);
include/linux/gfp.h
295
struct folio *__folio_alloc_node_noprof(gfp_t gfp, unsigned int order, int nid)
include/linux/gfp.h
298
warn_if_node_offline(nid, gfp);
include/linux/gfp.h
300
return __folio_alloc_noprof(gfp, order, nid, NULL);
include/linux/gfp.h
322
struct page *alloc_pages_noprof(gfp_t gfp, unsigned int order);
include/linux/gfp.h
323
struct folio *folio_alloc_noprof(gfp_t gfp, unsigned int order);
include/linux/gfp.h
324
struct folio *folio_alloc_mpol_noprof(gfp_t gfp, unsigned int order,
include/linux/gfp.h
326
struct folio *vma_alloc_folio_noprof(gfp_t gfp, int order, struct vm_area_struct *vma,
include/linux/gfp.h
333
static inline struct folio *folio_alloc_noprof(gfp_t gfp, unsigned int order)
include/linux/gfp.h
335
return __folio_alloc_node_noprof(gfp, order, numa_node_id());
include/linux/gfp.h
337
static inline struct folio *folio_alloc_mpol_noprof(gfp_t gfp, unsigned int order,
include/linux/gfp.h
340
return folio_alloc_noprof(gfp, order);
include/linux/gfp.h
342
static inline struct folio *vma_alloc_folio_noprof(gfp_t gfp, int order,
include/linux/gfp.h
345
return folio_alloc_noprof(gfp, order);
include/linux/gfp.h
356
static inline struct page *alloc_page_vma_noprof(gfp_t gfp,
include/linux/gfp.h
359
struct folio *folio = vma_alloc_folio_noprof(gfp, 0, vma, addr);
include/linux/gfp.h
418
static inline bool gfp_has_flags(gfp_t gfp, gfp_t flags)
include/linux/gfp.h
420
return (gfp & flags) == flags;
include/linux/gfp.h
423
static inline bool gfp_has_io_fs(gfp_t gfp)
include/linux/gfp.h
425
return gfp_has_flags(gfp, __GFP_IO | __GFP_FS);
include/linux/greybus/operation.h
149
gfp_t gfp);
include/linux/greybus/operation.h
154
size_t response_size, gfp_t gfp)
include/linux/greybus/operation.h
157
response_size, 0, gfp);
include/linux/greybus/operation.h
164
gfp_t gfp);
include/linux/greybus/operation.h
170
size_t response_size, gfp_t gfp);
include/linux/greybus/operation.h
175
gfp_t gfp);
include/linux/hugetlb.h
940
gfp_t gfp = __GFP_COMP | __GFP_NOWARN;
include/linux/hugetlb.h
942
gfp |= hugepage_movable_supported(h) ? GFP_HIGHUSER_MOVABLE : GFP_HIGHUSER;
include/linux/hugetlb.h
944
return gfp;
include/linux/idr.h
140
.id = idr_alloc(idr, ptr, start, end, gfp),
include/linux/idr.h
142
struct idr *idr, void *ptr, int start, int end, gfp_t gfp);
include/linux/idr.h
291
static inline int ida_alloc(struct ida *ida, gfp_t gfp)
include/linux/idr.h
293
return ida_alloc_range(ida, 0, ~0, gfp);
include/linux/idr.h
309
static inline int ida_alloc_min(struct ida *ida, unsigned int min, gfp_t gfp)
include/linux/idr.h
311
return ida_alloc_range(ida, min, ~0, gfp);
include/linux/idr.h
327
static inline int ida_alloc_max(struct ida *ida, unsigned int max, gfp_t gfp)
include/linux/idr.h
329
return ida_alloc_range(ida, 0, max, gfp);
include/linux/igmp.h
135
extern void __ip_mc_dec_group(struct in_device *in_dev, __be32 addr, gfp_t gfp);
include/linux/igmp.h
141
gfp_t gfp);
include/linux/io-pgtable.h
127
void *(*alloc)(void *cookie, size_t size, gfp_t gfp);
include/linux/io-pgtable.h
217
int prot, gfp_t gfp, size_t *mapped);
include/linux/iommu-dma.h
33
gfp_t gfp, unsigned long attrs);
include/linux/iommu-dma.h
46
enum dma_data_direction dir, gfp_t gfp, unsigned long attrs);
include/linux/iommu.h
1244
phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
include/linux/iommu.h
1264
unsigned int nents, int prot, gfp_t gfp)
include/linux/iommu.h
761
int prot, gfp_t gfp, size_t *mapped);
include/linux/iommu.h
916
phys_addr_t paddr, size_t size, int prot, gfp_t gfp);
include/linux/iommu.h
918
phys_addr_t paddr, size_t size, int prot, gfp_t gfp);
include/linux/iommu.h
928
int prot, gfp_t gfp);
include/linux/iommu.h
940
enum iommu_resv_type type, gfp_t gfp);
include/linux/kexec.h
521
static inline int arch_kexec_post_alloc_pages(void *vaddr, unsigned int pages, gfp_t gfp) { return 0; }
include/linux/kmemleak.h
107
static inline void kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp)
include/linux/kmemleak.h
117
gfp_t gfp)
include/linux/kmemleak.h
19
gfp_t gfp) __ref;
include/linux/kmemleak.h
21
gfp_t gfp) __ref;
include/linux/kmemleak.h
23
gfp_t gfp) __ref;
include/linux/kmemleak.h
32
extern void kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) __ref;
include/linux/kmemleak.h
35
gfp_t gfp) __ref;
include/linux/kmemleak.h
41
gfp_t gfp)
include/linux/kmemleak.h
44
kmemleak_alloc(ptr, size, min_count, gfp);
include/linux/kmemleak.h
64
gfp_t gfp)
include/linux/kmemleak.h
69
gfp_t gfp)
include/linux/kmemleak.h
73
gfp_t gfp)
include/linux/kmemleak.h
77
gfp_t gfp)
include/linux/list_lru.h
83
gfp_t gfp);
include/linux/lsm_hook_defs.h
216
LSM_HOOK(int, 0, cred_alloc_blank, struct cred *cred, gfp_t gfp)
include/linux/lsm_hook_defs.h
219
gfp_t gfp)
include/linux/lsm_hook_defs.h
390
struct xfrm_user_sec_ctx *sec_ctx, gfp_t gfp)
include/linux/lsm_hook_defs.h
423
void **lsmrule, gfp_t gfp)
include/linux/maple_tree.h
321
void *entry, gfp_t gfp);
include/linux/maple_tree.h
323
unsigned long last, void *entry, gfp_t gfp);
include/linux/maple_tree.h
326
unsigned long max, gfp_t gfp);
include/linux/maple_tree.h
329
unsigned long *next, gfp_t gfp);
include/linux/maple_tree.h
332
unsigned long max, gfp_t gfp);
include/linux/maple_tree.h
335
unsigned long last, void *entry, gfp_t gfp);
include/linux/maple_tree.h
337
void *entry, gfp_t gfp);
include/linux/maple_tree.h
340
int mtree_dup(struct maple_tree *mt, struct maple_tree *new, gfp_t gfp);
include/linux/maple_tree.h
341
int __mt_dup(struct maple_tree *mt, struct maple_tree *new, gfp_t gfp);
include/linux/maple_tree.h
516
int mas_store_gfp(struct ma_state *mas, void *entry, gfp_t gfp);
include/linux/maple_tree.h
522
int mas_preallocate(struct ma_state *mas, void *entry, gfp_t gfp);
include/linux/maple_tree.h
525
unsigned long *next, gfp_t gfp);
include/linux/maple_tree.h
527
bool mas_nomem(struct ma_state *mas, gfp_t gfp);
include/linux/memcontrol.h
1137
struct mm_struct *mm, gfp_t gfp)
include/linux/memcontrol.h
1142
static inline int mem_cgroup_charge_hugetlb(struct folio* folio, gfp_t gfp)
include/linux/memcontrol.h
1148
struct mm_struct *mm, gfp_t gfp, swp_entry_t entry)
include/linux/memcontrol.h
1681
int __memcg_kmem_charge_page(struct page *page, gfp_t gfp, int order);
include/linux/memcontrol.h
1705
int obj_cgroup_charge(struct obj_cgroup *objcg, gfp_t gfp, size_t size);
include/linux/memcontrol.h
1721
static inline int memcg_kmem_charge_page(struct page *page, gfp_t gfp,
include/linux/memcontrol.h
1725
return __memcg_kmem_charge_page(page, gfp, order);
include/linux/memcontrol.h
1776
static inline int memcg_kmem_charge_page(struct page *page, gfp_t gfp,
include/linux/memcontrol.h
1786
static inline int __memcg_kmem_charge_page(struct page *page, gfp_t gfp,
include/linux/memcontrol.h
643
int __mem_cgroup_charge(struct folio *folio, struct mm_struct *mm, gfp_t gfp);
include/linux/memcontrol.h
660
gfp_t gfp)
include/linux/memcontrol.h
664
return __mem_cgroup_charge(folio, mm, gfp);
include/linux/memcontrol.h
667
int mem_cgroup_charge_hugetlb(struct folio* folio, gfp_t gfp);
include/linux/memcontrol.h
670
gfp_t gfp, swp_entry_t entry);
include/linux/memregion.h
15
int memregion_alloc(gfp_t gfp);
include/linux/memregion.h
18
static inline int memregion_alloc(gfp_t gfp)
include/linux/mm.h
1339
unsigned long start, unsigned long end, gfp_t gfp)
include/linux/mm.h
1342
mas_store_gfp(&vmi->mas, NULL, gfp);
include/linux/mm.h
3402
static inline struct ptdesc *pagetable_alloc_noprof(gfp_t gfp, unsigned int order)
include/linux/mm.h
3404
struct page *page = alloc_pages_noprof(gfp | __GFP_COMP, order);
include/linux/netdevice.h
3078
#define __netdev_alloc_pcpu_stats(type, gfp) \
include/linux/netdevice.h
3080
typeof(type) __percpu *pcpu_stats = alloc_percpu_gfp(type, gfp);\
include/linux/netdevice.h
3423
netdevice_tracker *tracker, gfp_t gfp);
include/linux/netdevice.h
3426
netdevice_tracker *tracker, gfp_t gfp);
include/linux/netdevice.h
4433
gfp_t gfp)
include/linux/netdevice.h
4436
ref_tracker_alloc(&dev->refcnt_tracker, tracker, gfp);
include/linux/netdevice.h
4444
netdevice_tracker *tracker, gfp_t gfp)
include/linux/netdevice.h
4448
__netdev_tracker_alloc(dev, tracker, gfp);
include/linux/netdevice.h
4461
netdevice_tracker *tracker, gfp_t gfp)
include/linux/netdevice.h
4465
__netdev_tracker_alloc(dev, tracker, gfp);
include/linux/netdevice.h
4507
gfp_t gfp)
include/linux/netdevice.h
4516
__netdev_tracker_alloc(ndev, tracker, gfp);
include/linux/netfs.h
426
bool netfs_release_folio(struct folio *folio, gfp_t gfp);
include/linux/netfs.h
454
struct folio_queue *netfs_folioq_alloc(unsigned int rreq_id, gfp_t gfp,
include/linux/netfs.h
462
size_t *_cur_size, ssize_t size, gfp_t gfp);
include/linux/objpool.h
122
gfp_t gfp, void *context, objpool_init_obj_cb objinit,
include/linux/objpool.h
90
gfp_t gfp;
include/linux/pagemap.h
1286
pgoff_t index, gfp_t gfp);
include/linux/pagemap.h
1288
pgoff_t index, gfp_t gfp);
include/linux/pagemap.h
1294
bool filemap_release_folio(struct folio *folio, gfp_t gfp);
include/linux/pagemap.h
1300
pgoff_t index, gfp_t gfp, void **shadowp);
include/linux/pagemap.h
654
struct folio *filemap_alloc_folio_noprof(gfp_t gfp, unsigned int order,
include/linux/pagemap.h
657
static inline struct folio *filemap_alloc_folio_noprof(gfp_t gfp, unsigned int order,
include/linux/pagemap.h
660
return folio_alloc_noprof(gfp, order);
include/linux/pagemap.h
667
static inline struct page *__page_cache_alloc(gfp_t gfp)
include/linux/pagemap.h
669
return &filemap_alloc_folio(gfp, 0, NULL)->page;
include/linux/pagemap.h
756
pgoff_t index, fgf_t fgf_flags, gfp_t gfp, struct mempolicy *policy);
include/linux/pagemap.h
758
fgf_t fgp_flags, gfp_t gfp);
include/linux/pagemap.h
761
pgoff_t index, fgf_t fgf_flags, gfp_t gfp)
include/linux/pagemap.h
763
return __filemap_get_folio_mpol(mapping, index, fgf_flags, gfp, NULL);
include/linux/percpu-refcount.h
123
gfp_t gfp);
include/linux/percpu.h
137
gfp_t gfp) __alloc_size(1);
include/linux/percpu.h
146
#define alloc_percpu_gfp(type, gfp) \
include/linux/percpu.h
148
__alignof__(type), gfp)
include/linux/percpu_counter.h
140
s64 amount, gfp_t gfp,
include/linux/percpu_counter.h
152
gfp_t gfp)
include/linux/percpu_counter.h
154
return percpu_counter_init_many(fbc, amount, gfp, 1);
include/linux/percpu_counter.h
34
gfp_t gfp, u32 nr_counters,
include/linux/percpu_counter.h
37
#define percpu_counter_init_many(fbc, value, gfp, nr_counters) \
include/linux/percpu_counter.h
41
__percpu_counter_init_many(fbc, value, gfp, nr_counters,\
include/linux/percpu_counter.h
46
#define percpu_counter_init(fbc, value, gfp) \
include/linux/percpu_counter.h
47
percpu_counter_init_many(fbc, value, gfp, 1)
include/linux/posix_acl_xattr.h
48
size_t *sizep, gfp_t gfp);
include/linux/ptr_ring.h
477
static inline void **__ptr_ring_init_queue_alloc_noprof(unsigned int size, gfp_t gfp)
include/linux/ptr_ring.h
481
return kvmalloc_array_noprof(size, sizeof(void *), gfp | __GFP_ZERO);
include/linux/ptr_ring.h
497
static inline int ptr_ring_init_noprof(struct ptr_ring *r, int size, gfp_t gfp)
include/linux/ptr_ring.h
499
r->queue = __ptr_ring_init_queue_alloc_noprof(size, gfp);
include/linux/ptr_ring.h
566
int size, gfp_t gfp,
include/linux/ptr_ring.h
597
static inline int ptr_ring_resize_noprof(struct ptr_ring *r, int size, gfp_t gfp,
include/linux/ptr_ring.h
601
void **queue = __ptr_ring_init_queue_alloc_noprof(size, gfp);
include/linux/ptr_ring.h
610
old = __ptr_ring_swap_queue(r, queue, size, gfp, destroy);
include/linux/ptr_ring.h
629
int size, gfp_t gfp,
include/linux/ptr_ring.h
635
queues = kmalloc_array_noprof(nrings, sizeof(*queues), gfp);
include/linux/ptr_ring.h
640
queues[i] = __ptr_ring_init_queue_alloc_noprof(size, gfp);
include/linux/ptr_ring.h
649
size, gfp, destroy);
include/linux/radix-tree.h
264
struct radix_tree_iter *iter, gfp_t gfp,
include/linux/ref_tracker.h
129
gfp_t gfp)
include/linux/ref_tracker.h
85
struct ref_tracker **trackerp, gfp_t gfp);
include/linux/rslib.h
82
int nroots, gfp_t gfp);
include/linux/scatterlist.h
508
gfp_t gfp, unsigned int *nent_p);
include/linux/scatterlist.h
509
struct scatterlist *sgl_alloc(unsigned long long length, gfp_t gfp,
include/linux/security.h
1227
static inline int security_cred_alloc_blank(struct cred *cred, gfp_t gfp)
include/linux/security.h
1237
gfp_t gfp)
include/linux/security.h
1964
struct xfrm_user_sec_ctx *sec_ctx, gfp_t gfp);
include/linux/security.h
1984
gfp_t gfp)
include/linux/security.h
2188
gfp_t gfp);
include/linux/security.h
2197
void **lsmrule, gfp_t gfp)
include/linux/security.h
399
int security_inode_alloc(struct inode *inode, gfp_t gfp);
include/linux/security.h
494
int security_cred_alloc_blank(struct cred *cred, gfp_t gfp);
include/linux/security.h
496
int security_prepare_creds(struct cred *new, const struct cred *old, gfp_t gfp);
include/linux/security.h
857
static inline int security_inode_alloc(struct inode *inode, gfp_t gfp)
include/linux/shmem_fs.h
177
pgoff_t index, gfp_t gfp);
include/linux/skb_array.h
180
static inline int skb_array_init_noprof(struct skb_array *a, int size, gfp_t gfp)
include/linux/skb_array.h
182
return ptr_ring_init_noprof(&a->ring, size, gfp);
include/linux/skb_array.h
197
static inline int skb_array_resize(struct skb_array *a, int size, gfp_t gfp)
include/linux/skb_array.h
199
return ptr_ring_resize(&a->ring, size, gfp, __skb_array_destroy_skb);
include/linux/skb_array.h
205
gfp_t gfp)
include/linux/skb_array.h
209
nrings, size, gfp,
include/linux/skbuff.h
3503
unsigned int length, gfp_t gfp)
include/linux/skbuff.h
3505
struct sk_buff *skb = __netdev_alloc_skb(dev, length + NET_IP_ALIGN, gfp);
include/linux/skbuff.h
4269
gfp_t gfp);
include/linux/slab.h
831
kmem_cache_prefill_sheaf(struct kmem_cache *s, gfp_t gfp, unsigned int size);
include/linux/slab.h
833
int kmem_cache_refill_sheaf(struct kmem_cache *s, gfp_t gfp,
include/linux/slab.h
836
void kmem_cache_return_sheaf(struct kmem_cache *s, gfp_t gfp,
include/linux/slab.h
839
void *kmem_cache_alloc_from_sheaf_noprof(struct kmem_cache *cachep, gfp_t gfp,
include/linux/soc/airoha/airoha_offload.h
197
void *data, int data_len, gfp_t gfp);
include/linux/soc/airoha/airoha_offload.h
200
void *data, int data_len, gfp_t gfp);
include/linux/soc/airoha/airoha_offload.h
223
void *data, int data_len, gfp_t gfp)
include/linux/soc/airoha/airoha_offload.h
225
return npu->ops.wlan_send_msg(npu, ifindex, cmd, data, data_len, gfp);
include/linux/soc/airoha/airoha_offload.h
230
void *data, int data_len, gfp_t gfp)
include/linux/soc/airoha/airoha_offload.h
232
return npu->ops.wlan_get_msg(npu, ifindex, cmd, data, data_len, gfp);
include/linux/soc/airoha/airoha_offload.h
279
void *data, int data_len, gfp_t gfp)
include/linux/soc/airoha/airoha_offload.h
286
void *data, int data_len, gfp_t gfp)
include/linux/spinlock.h
519
gfp_t gfp, const char *name,
include/linux/spinlock.h
522
#define alloc_bucket_spinlocks(locks, lock_mask, max_size, cpu_mult, gfp) \
include/linux/spinlock.h
528
cpu_mult, gfp, #locks, &key); \
include/linux/sprintf.h
17
__printf(2, 3) __malloc char *kasprintf(gfp_t gfp, const char *fmt, ...);
include/linux/sprintf.h
18
__printf(2, 0) __malloc char *kvasprintf(gfp_t gfp, const char *fmt, va_list args);
include/linux/sprintf.h
19
__printf(2, 0) const char *kvasprintf_const(gfp_t gfp, const char *fmt, va_list args);
include/linux/string.h
301
extern char *kstrdup(const char *s, gfp_t gfp) __malloc;
include/linux/string.h
302
extern const char *kstrdup_const(const char *s, gfp_t gfp);
include/linux/string.h
303
extern char *kstrndup(const char *s, size_t len, gfp_t gfp);
include/linux/string.h
304
extern void *kmemdup_noprof(const void *src, size_t len, gfp_t gfp) __realloc_size(2);
include/linux/string.h
307
extern void *kvmemdup(const void *src, size_t len, gfp_t gfp) __realloc_size(2);
include/linux/string.h
308
extern char *kmemdup_nul(const char *s, size_t len, gfp_t gfp);
include/linux/string.h
309
extern void *kmemdup_array(const void *src, size_t count, size_t element_size, gfp_t gfp)
include/linux/string.h
313
extern char **argv_split(gfp_t gfp, const char *str, int *argcp);
include/linux/string_helpers.h
113
char *kstrdup_quotable(const char *src, gfp_t gfp);
include/linux/string_helpers.h
114
char *kstrdup_quotable_cmdline(struct task_struct *task, gfp_t gfp);
include/linux/string_helpers.h
115
char *kstrdup_quotable_file(struct file *file, gfp_t gfp);
include/linux/string_helpers.h
117
char *kstrdup_and_replace(const char *src, char old, char new, gfp_t gfp);
include/linux/string_helpers.h
119
char **kasprintf_strarray(gfp_t gfp, const char *prefix, size_t n);
include/linux/sunrpc/xdr.h
139
int xdr_alloc_bvec(struct xdr_buf *buf, gfp_t gfp);
include/linux/swap.h
582
void __folio_throttle_swaprate(struct folio *folio, gfp_t gfp);
include/linux/swap.h
583
static inline void folio_throttle_swaprate(struct folio *folio, gfp_t gfp)
include/linux/swap.h
587
__folio_throttle_swaprate(folio, gfp);
include/linux/swap.h
590
static inline void folio_throttle_swaprate(struct folio *folio, gfp_t gfp)
include/linux/virtio.h
287
gfp_t gfp);
include/linux/virtio.h
58
gfp_t gfp);
include/linux/virtio.h
63
gfp_t gfp);
include/linux/virtio.h
68
gfp_t gfp);
include/linux/virtio.h
74
gfp_t gfp);
include/linux/virtio.h
80
gfp_t gfp);
include/linux/virtio.h
85
gfp_t gfp);
include/linux/virtio.h
92
gfp_t gfp);
include/linux/virtio_config.h
206
dma_addr_t *map_handle, gfp_t gfp);
include/linux/vmpressure.h
33
extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree,
include/linux/vmpressure.h
35
extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio);
include/linux/vmpressure.h
47
static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree,
include/linux/vmpressure.h
49
static inline void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg,
include/linux/vringh.h
233
gfp_t gfp);
include/linux/vringh.h
306
gfp_t gfp);
include/linux/writeback.h
332
int wb_domain_init(struct wb_domain *dom, gfp_t gfp);
include/linux/xarray.h
1016
struct xa_limit limit, u32 *next, gfp_t gfp)
include/linux/xarray.h
1020
might_alloc(gfp);
include/linux/xarray.h
1022
err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
include/linux/xarray.h
1055
struct xa_limit limit, u32 *next, gfp_t gfp)
include/linux/xarray.h
1059
might_alloc(gfp);
include/linux/xarray.h
1061
err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
include/linux/xarray.h
1086
int xa_reserve(struct xarray *xa, unsigned long index, gfp_t gfp)
include/linux/xarray.h
1088
return xa_err(xa_cmpxchg(xa, index, NULL, XA_ZERO_ENTRY, gfp));
include/linux/xarray.h
1104
int xa_reserve_bh(struct xarray *xa, unsigned long index, gfp_t gfp)
include/linux/xarray.h
1106
return xa_err(xa_cmpxchg_bh(xa, index, NULL, XA_ZERO_ENTRY, gfp));
include/linux/xarray.h
1122
int xa_reserve_irq(struct xarray *xa, unsigned long index, gfp_t gfp)
include/linux/xarray.h
1124
return xa_err(xa_cmpxchg_irq(xa, index, NULL, XA_ZERO_ENTRY, gfp));
include/linux/xarray.h
1584
unsigned int order, gfp_t gfp)
include/linux/xarray.h
590
void *entry, gfp_t gfp)
include/linux/xarray.h
594
might_alloc(gfp);
include/linux/xarray.h
596
curr = __xa_store(xa, index, entry, gfp);
include/linux/xarray.h
617
void *entry, gfp_t gfp)
include/linux/xarray.h
621
might_alloc(gfp);
include/linux/xarray.h
623
curr = __xa_store(xa, index, entry, gfp);
include/linux/xarray.h
693
void *old, void *entry, gfp_t gfp)
include/linux/xarray.h
697
might_alloc(gfp);
include/linux/xarray.h
699
curr = __xa_cmpxchg(xa, index, old, entry, gfp);
include/linux/xarray.h
721
void *old, void *entry, gfp_t gfp)
include/linux/xarray.h
725
might_alloc(gfp);
include/linux/xarray.h
727
curr = __xa_cmpxchg(xa, index, old, entry, gfp);
include/linux/xarray.h
749
void *old, void *entry, gfp_t gfp)
include/linux/xarray.h
753
might_alloc(gfp);
include/linux/xarray.h
755
curr = __xa_cmpxchg(xa, index, old, entry, gfp);
include/linux/xarray.h
779
unsigned long index, void *entry, gfp_t gfp)
include/linux/xarray.h
783
might_alloc(gfp);
include/linux/xarray.h
785
err = __xa_insert(xa, index, entry, gfp);
include/linux/xarray.h
809
unsigned long index, void *entry, gfp_t gfp)
include/linux/xarray.h
813
might_alloc(gfp);
include/linux/xarray.h
815
err = __xa_insert(xa, index, entry, gfp);
include/linux/xarray.h
839
unsigned long index, void *entry, gfp_t gfp)
include/linux/xarray.h
843
might_alloc(gfp);
include/linux/xarray.h
845
err = __xa_insert(xa, index, entry, gfp);
include/linux/xarray.h
872
void *entry, struct xa_limit limit, gfp_t gfp)
include/linux/xarray.h
876
might_alloc(gfp);
include/linux/xarray.h
878
err = __xa_alloc(xa, id, entry, limit, gfp);
include/linux/xarray.h
905
void *entry, struct xa_limit limit, gfp_t gfp)
include/linux/xarray.h
909
might_alloc(gfp);
include/linux/xarray.h
911
err = __xa_alloc(xa, id, entry, limit, gfp);
include/linux/xarray.h
938
void *entry, struct xa_limit limit, gfp_t gfp)
include/linux/xarray.h
942
might_alloc(gfp);
include/linux/xarray.h
944
err = __xa_alloc(xa, id, entry, limit, gfp);
include/linux/xarray.h
977
struct xa_limit limit, u32 *next, gfp_t gfp)
include/linux/xarray.h
981
might_alloc(gfp);
include/linux/xarray.h
983
err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp);
include/net/af_rxrpc.h
61
gfp_t gfp,
include/net/af_rxrpc.h
77
struct sockaddr_rxrpc *srx, gfp_t gfp);
include/net/af_rxrpc.h
87
unsigned long user_call_ID, gfp_t gfp,
include/net/cfg80211.h
10015
u64 cookie, gfp_t gfp);
include/net/cfg80211.h
10029
gfp_t gfp);
include/net/cfg80211.h
10041
gfp_t gfp);
include/net/cfg80211.h
10054
gfp_t gfp);
include/net/cfg80211.h
10146
gfp_t gfp);
include/net/cfg80211.h
10341
struct ieee80211_channel *chan, gfp_t gfp);
include/net/cfg80211.h
10355
gfp_t gfp);
include/net/cfg80211.h
7801
gfp_t gfp);
include/net/cfg80211.h
7807
s32 signal, gfp_t gfp)
include/net/cfg80211.h
7814
return cfg80211_inform_bss_frame_data(wiphy, &data, mgmt, len, gfp);
include/net/cfg80211.h
7933
gfp_t gfp);
include/net/cfg80211.h
7941
s32 signal, gfp_t gfp)
include/net/cfg80211.h
7950
gfp);
include/net/cfg80211.h
8205
const u8 *tsc, gfp_t gfp);
include/net/cfg80211.h
8223
struct ieee80211_channel *channel, gfp_t gfp);
include/net/cfg80211.h
8242
int sig_dbm, gfp_t gfp);
include/net/cfg80211.h
8319
int approxlen, gfp_t gfp);
include/net/cfg80211.h
8321
void __cfg80211_send_event_skb(struct sk_buff *skb, gfp_t gfp);
include/net/cfg80211.h
8400
int approxlen, int event_idx, gfp_t gfp)
include/net/cfg80211.h
8404
0, event_idx, approxlen, gfp);
include/net/cfg80211.h
8435
int event_idx, gfp_t gfp)
include/net/cfg80211.h
8439
portid, event_idx, approxlen, gfp);
include/net/cfg80211.h
8450
static inline void cfg80211_vendor_event(struct sk_buff *skb, gfp_t gfp)
include/net/cfg80211.h
8452
__cfg80211_send_event_skb(skb, gfp);
include/net/cfg80211.h
8536
cfg80211_testmode_alloc_event_skb(struct wiphy *wiphy, int approxlen, gfp_t gfp)
include/net/cfg80211.h
8540
approxlen, gfp);
include/net/cfg80211.h
8553
static inline void cfg80211_testmode_event(struct sk_buff *skb, gfp_t gfp)
include/net/cfg80211.h
8555
__cfg80211_send_event_skb(skb, gfp);
include/net/cfg80211.h
8667
gfp_t gfp);
include/net/cfg80211.h
8711
size_t resp_ie_len, int status, gfp_t gfp,
include/net/cfg80211.h
8726
cfg80211_connect_done(dev, ¶ms, gfp);
include/net/cfg80211.h
8753
u16 status, gfp_t gfp)
include/net/cfg80211.h
8756
resp_ie_len, status, gfp,
include/net/cfg80211.h
8780
const u8 *req_ie, size_t req_ie_len, gfp_t gfp,
include/net/cfg80211.h
8784
gfp, timeout_reason);
include/net/cfg80211.h
8845
gfp_t gfp);
include/net/cfg80211.h
8868
const u8* td_bitmap, u8 td_bitmap_len, gfp_t gfp);
include/net/cfg80211.h
8885
bool locally_generated, gfp_t gfp);
include/net/cfg80211.h
8898
unsigned int duration, gfp_t gfp);
include/net/cfg80211.h
8909
gfp_t gfp);
include/net/cfg80211.h
8919
struct ieee80211_channel *chan, gfp_t gfp);
include/net/cfg80211.h
8929
int cfg80211_sinfo_alloc_tid_stats(struct station_info *sinfo, gfp_t gfp);
include/net/cfg80211.h
8940
gfp_t gfp);
include/net/cfg80211.h
8971
struct station_info *sinfo, gfp_t gfp);
include/net/cfg80211.h
8981
struct station_info *sinfo, gfp_t gfp);
include/net/cfg80211.h
8991
const u8 *mac_addr, gfp_t gfp)
include/net/cfg80211.h
8993
cfg80211_del_sta_sinfo(dev, mac_addr, NULL, gfp);
include/net/cfg80211.h
9013
gfp_t gfp);
include/net/cfg80211.h
9151
struct cfg80211_tx_status *status, gfp_t gfp);
include/net/cfg80211.h
9168
size_t len, bool ack, gfp_t gfp)
include/net/cfg80211.h
9177
cfg80211_mgmt_tx_status_ext(wdev, &status, gfp);
include/net/cfg80211.h
9196
gfp_t gfp);
include/net/cfg80211.h
9233
s32 rssi_level, gfp_t gfp);
include/net/cfg80211.h
9245
const u8 *peer, u32 num_packets, gfp_t gfp);
include/net/cfg80211.h
9260
u32 num_packets, u32 rate, u32 intvl, gfp_t gfp);
include/net/cfg80211.h
9269
void cfg80211_cqm_beacon_loss_notify(struct net_device *dev, gfp_t gfp);
include/net/cfg80211.h
9282
bool offchan, gfp_t gfp);
include/net/cfg80211.h
9287
gfp_t gfp)
include/net/cfg80211.h
9289
__cfg80211_radar_event(wiphy, chandef, false, gfp);
include/net/cfg80211.h
9295
gfp_t gfp)
include/net/cfg80211.h
9297
__cfg80211_radar_event(wiphy, chandef, true, gfp);
include/net/cfg80211.h
9312
gfp_t gfp);
include/net/cfg80211.h
9328
enum nl80211_radar_event event, gfp_t gfp,
include/net/cfg80211.h
9348
const u8 *replay_ctr, gfp_t gfp);
include/net/cfg80211.h
9359
const u8 *bssid, bool preauth, gfp_t gfp);
include/net/cfg80211.h
9375
int link_id, gfp_t gfp);
include/net/cfg80211.h
9392
int link_id, gfp_t gfp);
include/net/cfg80211.h
9406
bool is_valid_ack_signal, gfp_t gfp);
include/net/cfg80211.h
9613
u16 reason_code, gfp_t gfp);
include/net/cfg80211.h
9802
gfp_t gfp);
include/net/cfg80211.h
9814
void cfg80211_crit_proto_stopped(struct wireless_dev *wdev, gfp_t gfp);
include/net/cfg80211.h
9883
int link_id, gfp_t gfp);
include/net/cfg80211.h
9898
cfg80211_stop_iface(struct wiphy *wiphy, struct wireless_dev *wdev, gfp_t gfp)
include/net/cfg80211.h
9900
cfg80211_stop_link(wiphy, wdev, -1, gfp);
include/net/cfg80211.h
9999
struct cfg80211_nan_match_params *match, gfp_t gfp);
include/net/dst_cache.h
98
int dst_cache_init(struct dst_cache *dst_cache, gfp_t gfp);
include/net/flow_offload.h
231
gfp_t gfp);
include/net/hwbm.h
23
int hwbm_pool_refill(struct hwbm_pool *bm_pool, gfp_t gfp);
include/net/hwbm.h
28
static inline int hwbm_pool_refill(struct hwbm_pool *bm_pool, gfp_t gfp)
include/net/ioam6.h
71
void ioam6_event(enum ioam6_event_type type, struct net *net, gfp_t gfp,
include/net/iucv/iucv.h
210
static inline struct iucv_path *iucv_path_alloc(u16 msglim, u8 flags, gfp_t gfp)
include/net/iucv/iucv.h
214
path = kzalloc_obj(struct iucv_path, gfp);
include/net/mac80211.h
6136
const u8 *replay_ctr, gfp_t gfp);
include/net/mac80211.h
6915
gfp_t gfp);
include/net/mac80211.h
6923
void ieee80211_cqm_beacon_loss_notify(struct ieee80211_vif *vif, gfp_t gfp);
include/net/mac80211.h
7160
void *(*alloc_sta)(void *priv, struct ieee80211_sta *sta, gfp_t gfp);
include/net/mac80211.h
7400
gfp_t gfp);
include/net/mac80211.h
7495
u16 reason_code, gfp_t gfp);
include/net/mac80211.h
7718
gfp_t gfp);
include/net/mac80211.h
7733
gfp_t gfp);
include/net/net_namespace.h
364
gfp_t gfp)
include/net/net_namespace.h
369
tracker, gfp);
include/net/net_namespace.h
374
gfp_t gfp)
include/net/net_namespace.h
376
__netns_tracker_alloc(net, tracker, true, gfp);
include/net/net_namespace.h
390
netns_tracker *tracker, gfp_t gfp)
include/net/net_namespace.h
393
netns_tracker_alloc(net, tracker, gfp);
include/net/net_namespace.h
454
int peernet2id_alloc(struct net *net, struct net *peer, gfp_t gfp);
include/net/netfilter/nf_conntrack.h
256
gfp_t gfp);
include/net/netfilter/nf_conntrack_acct.h
30
struct nf_conn_acct *nf_ct_acct_ext_add(struct nf_conn *ct, gfp_t gfp)
include/net/netfilter/nf_conntrack_acct.h
39
acct = nf_ct_ext_add(ct, NF_CT_EXT_ACCT, gfp);
include/net/netfilter/nf_conntrack_ecache.h
80
bool nf_ct_ecache_ext_add(struct nf_conn *ct, u16 ctmask, u16 expmask, gfp_t gfp);
include/net/netfilter/nf_conntrack_ecache.h
95
static inline bool nf_ct_ecache_ext_add(struct nf_conn *ct, u16 ctmask, u16 expmask, gfp_t gfp)
include/net/netfilter/nf_conntrack_extend.h
71
void *nf_ct_ext_add(struct nf_conn *ct, enum nf_ct_ext_id id, gfp_t gfp);
include/net/netfilter/nf_conntrack_helper.h
113
struct nf_conn_help *nf_ct_helper_ext_add(struct nf_conn *ct, gfp_t gfp);
include/net/netfilter/nf_conntrack_timeout.h
53
gfp_t gfp)
include/net/netfilter/nf_conntrack_timeout.h
58
timeout_ext = nf_ct_ext_add(ct, NF_CT_EXT_TIMEOUT, gfp);
include/net/netfilter/nf_conntrack_timestamp.h
27
struct nf_conn_tstamp *nf_ct_tstamp_ext_add(struct nf_conn *ct, gfp_t gfp)
include/net/netfilter/nf_conntrack_timestamp.h
35
return nf_ct_ext_add(ct, NF_CT_EXT_TSTAMP, gfp);
include/net/netfilter/nf_tables.h
1411
int event, u16 flags, int family, int report, gfp_t gfp);
include/net/netfilter/nf_tables.h
1897
unsigned int gc_seq, gfp_t gfp);
include/net/netfilter/nf_tables.h
1901
unsigned int gc_seq, gfp_t gfp);
include/net/netfilter/nf_tables.h
1904
struct nft_trans_gc *nft_trans_gc_queue_sync(struct nft_trans_gc *gc, gfp_t gfp);
include/net/netfilter/nf_tables.h
424
int nft_expr_clone(struct nft_expr *dst, struct nft_expr *src, gfp_t gfp);
include/net/netfilter/nf_tables.h
874
u64 timeout, u64 expiration, gfp_t gfp);
include/net/netfilter/nf_tables.h
958
const struct nft_expr *src, gfp_t gfp);
include/net/netlink.h
2208
static inline void *nla_memdup_noprof(const struct nlattr *src, gfp_t gfp)
include/net/netlink.h
2210
return kmemdup_noprof(nla_data(src), nla_len(src), gfp);
include/net/nfc/nfc.h
272
struct sk_buff *nfc_alloc_recv_skb(unsigned int size, gfp_t gfp);
include/net/page_pool/helpers.h
113
gfp_t gfp = (GFP_ATOMIC | __GFP_NOWARN);
include/net/page_pool/helpers.h
115
return page_pool_alloc_frag(pool, offset, size, gfp);
include/net/page_pool/helpers.h
120
unsigned int *size, gfp_t gfp)
include/net/page_pool/helpers.h
128
return page_pool_alloc_netmems(pool, gfp);
include/net/page_pool/helpers.h
131
netmem = page_pool_alloc_frag_netmem(pool, offset, *size, gfp);
include/net/page_pool/helpers.h
151
gfp_t gfp = GFP_ATOMIC | __GFP_NOWARN;
include/net/page_pool/helpers.h
153
return page_pool_alloc_netmem(pool, offset, size, gfp);
include/net/page_pool/helpers.h
158
gfp_t gfp = GFP_ATOMIC | __GFP_NOWARN;
include/net/page_pool/helpers.h
160
return page_pool_alloc_netmems(pool, gfp);
include/net/page_pool/helpers.h
165
unsigned int *size, gfp_t gfp)
include/net/page_pool/helpers.h
167
return netmem_to_page(page_pool_alloc_netmem(pool, offset, size, gfp));
include/net/page_pool/helpers.h
186
gfp_t gfp = (GFP_ATOMIC | __GFP_NOWARN);
include/net/page_pool/helpers.h
188
return page_pool_alloc(pool, offset, size, gfp);
include/net/page_pool/helpers.h
192
unsigned int *size, gfp_t gfp)
include/net/page_pool/helpers.h
198
page = page_pool_alloc(pool, &offset, size, gfp & ~__GFP_HIGHMEM);
include/net/page_pool/helpers.h
219
gfp_t gfp = (GFP_ATOMIC | __GFP_NOWARN);
include/net/page_pool/helpers.h
221
return page_pool_alloc_va(pool, size, gfp);
include/net/page_pool/helpers.h
94
gfp_t gfp = (GFP_ATOMIC | __GFP_NOWARN);
include/net/page_pool/helpers.h
96
return page_pool_alloc_pages(pool, gfp);
include/net/page_pool/memory_provider.h
13
netmem_ref (*alloc_netmems)(struct page_pool *pool, gfp_t gfp);
include/net/page_pool/types.h
255
struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp);
include/net/page_pool/types.h
256
netmem_ref page_pool_alloc_netmems(struct page_pool *pool, gfp_t gfp);
include/net/page_pool/types.h
258
unsigned int size, gfp_t gfp);
include/net/page_pool/types.h
261
gfp_t gfp);
include/net/sctp/auth.h
104
int sctp_auth_init(struct sctp_endpoint *ep, gfp_t gfp);
include/net/sctp/auth.h
66
struct sctp_shared_key *sctp_auth_shkey_create(__u16 key_id, gfp_t gfp);
include/net/sctp/auth.h
68
int sctp_auth_asoc_init_active_key(struct sctp_association *asoc, gfp_t gfp);
include/net/sctp/auth.h
74
gfp_t gfp);
include/net/sctp/auth.h
88
struct sctp_shared_key *ep_key, gfp_t gfp);
include/net/sctp/sctp.h
78
enum sctp_scope, gfp_t gfp, int flags);
include/net/sctp/stream_interleave.h
25
int len, __u8 flags, gfp_t gfp);
include/net/sctp/stream_interleave.h
29
struct sctp_chunk *chunk, gfp_t gfp);
include/net/sctp/stream_interleave.h
33
struct sctp_chunk *chunk, gfp_t gfp);
include/net/sctp/stream_interleave.h
34
void (*start_pd)(struct sctp_ulpq *ulpq, gfp_t gfp);
include/net/sctp/stream_interleave.h
35
void (*abort_pd)(struct sctp_ulpq *ulpq, gfp_t gfp);
include/net/sctp/stream_sched.h
22
gfp_t gfp);
include/net/sctp/stream_sched.h
28
int (*init_sid)(struct sctp_stream *stream, __u16 sid, gfp_t gfp);
include/net/sctp/stream_sched.h
48
__u16 value, gfp_t gfp);
include/net/sctp/stream_sched.h
54
int sctp_sched_init_sid(struct sctp_stream *stream, __u16 sid, gfp_t gfp);
include/net/sctp/structs.h
1116
void sctp_outq_uncork(struct sctp_outq *, gfp_t gfp);
include/net/sctp/structs.h
1177
enum sctp_scope scope, gfp_t gfp,
include/net/sctp/structs.h
1181
gfp_t gfp);
include/net/sctp/structs.h
1183
int new_size, __u8 addr_state, gfp_t gfp);
include/net/sctp/structs.h
1199
gfp_t gfp);
include/net/sctp/structs.h
1201
__u16 port, gfp_t gfp);
include/net/sctp/structs.h
1363
struct sctp_init_chunk *init, gfp_t gfp);
include/net/sctp/structs.h
2111
enum sctp_scope scope, gfp_t gfp);
include/net/sctp/structs.h
2125
const gfp_t gfp,
include/net/sctp/structs.h
2150
enum sctp_scope scope, gfp_t gfp);
include/net/sctp/structs.h
2153
gfp_t gfp);
include/net/sctp/structs.h
383
gfp_t gfp);
include/net/sctp/structs.h
663
struct sock *, gfp_t gfp);
include/net/sctp/structs.h
730
int one_packet, gfp_t gfp);
include/net/sctp/tsnmap.h
85
__u32 initial_tsn, gfp_t gfp);
include/net/sctp/ulpevent.h
103
gfp_t gfp);
include/net/sctp/ulpevent.h
108
gfp_t gfp);
include/net/sctp/ulpevent.h
113
__u32 flags, gfp_t gfp);
include/net/sctp/ulpevent.h
116
const struct sctp_association *asoc, gfp_t gfp);
include/net/sctp/ulpevent.h
120
gfp_t gfp);
include/net/sctp/ulpevent.h
124
__u32 indication, gfp_t gfp);
include/net/sctp/ulpevent.h
127
const struct sctp_association *asoc, gfp_t gfp);
include/net/sctp/ulpevent.h
131
__u16 stream_num, __be16 *stream_list, gfp_t gfp);
include/net/sctp/ulpevent.h
135
__u32 local_tsn, __u32 remote_tsn, gfp_t gfp);
include/net/sctp/ulpevent.h
139
__u32 strchange_instrms, __u32 strchange_outstrms, gfp_t gfp);
include/net/sctp/ulpevent.h
81
gfp_t gfp);
include/net/sctp/ulpevent.h
90
gfp_t gfp);
include/net/sctp/ulpevent.h
96
gfp_t gfp);
include/net/sock.h
1648
void __sk_charge(struct sock *sk, gfp_t gfp);
include/net/tc_act/tc_ife.h
53
int ife_alloc_meta_u32(struct tcf_meta_info *mi, void *metaval, gfp_t gfp);
include/net/tc_act/tc_ife.h
54
int ife_alloc_meta_u16(struct tcf_meta_info *mi, void *metaval, gfp_t gfp);
include/net/tcp.h
396
struct sk_buff *tcp_stream_alloc_skb(struct sock *sk, gfp_t gfp,
include/net/tcp.h
746
unsigned int mss_now, gfp_t gfp);
include/net/xfrm.h
1860
struct xfrm_policy *xfrm_policy_alloc(struct net *net, gfp_t gfp);
include/rdma/ib_verbs.h
2349
#define rdma_zalloc_drv_obj_gfp(ib_dev, ib_type, gfp) \
include/rdma/ib_verbs.h
2351
gfp, false))
include/rdma/ib_verbs.h
2913
gfp_t gfp, bool is_numa_aware)
include/rdma/ib_verbs.h
2916
return kzalloc_node(size, gfp, dev->ops.get_numa_node(dev));
include/rdma/ib_verbs.h
2918
return kzalloc(size, gfp);
include/scsi/scsi_dh.h
82
gfp_t gfp)
include/target/target_core_fabric.h
177
struct scatterlist *sgl_prot, u32 sgl_prot_count, gfp_t gfp);
include/target/target_core_fabric.h
181
gfp_t gfp);
include/xen/xenbus.h
220
int xenbus_setup_ring(struct xenbus_device *dev, gfp_t gfp, void **vaddr,
io_uring/alloc_cache.c
36
void *io_cache_alloc_new(struct io_alloc_cache *cache, gfp_t gfp)
io_uring/alloc_cache.c
40
obj = kmalloc(cache->elem_size, gfp);
io_uring/alloc_cache.h
19
void *io_cache_alloc_new(struct io_alloc_cache *cache, gfp_t gfp);
io_uring/alloc_cache.h
54
static inline void *io_cache_alloc(struct io_alloc_cache *cache, gfp_t gfp)
io_uring/alloc_cache.h
61
return io_cache_alloc_new(cache, gfp);
io_uring/io_uring.c
656
struct io_big_cqe *big_cqe, gfp_t gfp)
io_uring/io_uring.c
667
ocqe = kzalloc(ocq_size, gfp | __GFP_ACCOUNT);
io_uring/io_uring.c
961
gfp_t gfp = GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO;
io_uring/io_uring.c
965
ret = kmem_cache_alloc_bulk(req_cachep, gfp, ARRAY_SIZE(reqs), reqs);
io_uring/io_uring.c
972
reqs[0] = kmem_cache_alloc(req_cachep, gfp);
io_uring/memmap.c
156
gfp_t gfp = GFP_KERNEL_ACCOUNT | __GFP_ZERO | __GFP_NOWARN;
io_uring/memmap.c
161
pages = kvmalloc_objs(*pages, mr->nr_pages, gfp);
io_uring/memmap.c
165
if (io_mem_alloc_compound(pages, mr->nr_pages, size, gfp)) {
io_uring/memmap.c
170
nr_allocated = alloc_pages_bulk_node(gfp, NUMA_NO_NODE,
io_uring/memmap.c
19
size_t size, gfp_t gfp)
io_uring/memmap.c
28
gfp |= __GFP_COMP;
io_uring/memmap.c
30
page = alloc_pages(gfp, order);
io_uring/rsrc.c
1313
gfp_t gfp = GFP_KERNEL_ACCOUNT | __GFP_NOWARN;
io_uring/rsrc.c
1316
iov = kmalloc_objs(iov[0], nr_entries, gfp);
io_uring/zcrx.c
1004
static netmem_ref io_pp_zc_alloc_netmems(struct page_pool *pp, gfp_t gfp)
kernel/auditsc.c
2866
enum audit_nfcfgop op, gfp_t gfp)
kernel/auditsc.c
2871
ab = audit_log_start(audit_context(), gfp, AUDIT_NETFILTER_CFG);
kernel/bpf/cpumap.c
436
gfp_t gfp = GFP_KERNEL | __GFP_NOWARN;
kernel/bpf/cpumap.c
443
rcpu = bpf_map_kmalloc_node(map, sizeof(*rcpu), gfp | __GFP_ZERO, numa);
kernel/bpf/cpumap.c
449
sizeof(void *), gfp);
kernel/bpf/cpumap.c
460
rcpu->queue = bpf_map_kmalloc_node(map, sizeof(*rcpu->queue), gfp,
kernel/bpf/cpumap.c
465
err = ptr_ring_init(rcpu->queue, value->qsize, gfp);
kernel/bpf/local_storage.c
506
const gfp_t gfp = __GFP_ZERO | GFP_USER;
kernel/bpf/local_storage.c
519
gfp, map->numa_node);
kernel/bpf/local_storage.c
524
storage->buf = bpf_map_kmalloc_node(map, size, gfp,
kernel/bpf/local_storage.c
530
storage->percpu_buf = bpf_map_alloc_percpu(map, size, 8, gfp);
kernel/bpf/memalloc.c
210
gfp_t gfp;
kernel/bpf/memalloc.c
214
gfp = __GFP_NOWARN | __GFP_ACCOUNT;
kernel/bpf/memalloc.c
215
gfp |= atomic ? GFP_NOWAIT : GFP_KERNEL;
kernel/bpf/memalloc.c
248
obj = __alloc(c, node, gfp);
kernel/bpf/syscall.c
384
gfp_t gfp = bpf_memcg_flags(__GFP_NOWARN | __GFP_ZERO);
kernel/bpf/syscall.c
398
area = kmalloc_node(size, gfp | GFP_USER | __GFP_NORETRY,
kernel/bpf/syscall.c
405
gfp | GFP_KERNEL | __GFP_RETRY_MAYFAIL, PAGE_KERNEL,
kernel/dma/contiguous.c
350
static struct page *cma_alloc_aligned(struct cma *cma, size_t size, gfp_t gfp)
kernel/dma/contiguous.c
354
return cma_alloc(cma, size >> PAGE_SHIFT, align, gfp & __GFP_NOWARN);
kernel/dma/contiguous.c
372
struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp)
kernel/dma/contiguous.c
379
if (!gfpflags_allow_blocking(gfp))
kernel/dma/contiguous.c
382
return cma_alloc_aligned(dev->cma_area, size, gfp);
kernel/dma/contiguous.c
387
if (nid != NUMA_NO_NODE && !(gfp & (GFP_DMA | GFP_DMA32))) {
kernel/dma/contiguous.c
392
page = cma_alloc_aligned(cma, size, gfp);
kernel/dma/contiguous.c
399
page = cma_alloc_aligned(cma, size, gfp);
kernel/dma/contiguous.c
408
return cma_alloc_aligned(dma_contiguous_default_area, size, gfp);
kernel/dma/debug.c
625
static int dma_debug_create_entries(gfp_t gfp)
kernel/dma/debug.c
630
entry = (void *)get_zeroed_page(gfp);
kernel/dma/direct.c
120
gfp_t gfp, bool allow_highmem)
kernel/dma/direct.c
131
gfp |= dma_direct_optimal_gfp_mask(dev, &phys_limit);
kernel/dma/direct.c
132
page = dma_alloc_contiguous(dev, size, gfp);
kernel/dma/direct.c
141
while ((page = alloc_pages_node(node, gfp, get_order(size)))
kernel/dma/direct.c
147
!(gfp & (GFP_DMA32 | GFP_DMA)))
kernel/dma/direct.c
148
gfp |= GFP_DMA32;
kernel/dma/direct.c
149
else if (IS_ENABLED(CONFIG_ZONE_DMA) && !(gfp & GFP_DMA))
kernel/dma/direct.c
150
gfp = (gfp & ~GFP_DMA32) | GFP_DMA;
kernel/dma/direct.c
162
static bool dma_direct_use_pool(struct device *dev, gfp_t gfp)
kernel/dma/direct.c
164
return !gfpflags_allow_blocking(gfp) && !is_swiotlb_for_alloc(dev);
kernel/dma/direct.c
168
dma_addr_t *dma_handle, gfp_t gfp)
kernel/dma/direct.c
177
gfp |= dma_direct_optimal_gfp_mask(dev, &phys_limit);
kernel/dma/direct.c
178
page = dma_alloc_from_pool(dev, size, &ret, gfp, dma_coherent_ok);
kernel/dma/direct.c
186
dma_addr_t *dma_handle, gfp_t gfp)
kernel/dma/direct.c
190
page = __dma_direct_alloc_pages(dev, size, gfp & ~__GFP_ZERO, true);
kernel/dma/direct.c
204
dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
kernel/dma/direct.c
212
gfp |= __GFP_NOWARN;
kernel/dma/direct.c
216
return dma_direct_alloc_no_mapping(dev, size, dma_handle, gfp);
kernel/dma/direct.c
221
return arch_dma_alloc(dev, size, dma_handle, gfp,
kernel/dma/direct.c
250
dma_direct_use_pool(dev, gfp))
kernel/dma/direct.c
251
return dma_direct_alloc_from_pool(dev, size, dma_handle, gfp);
kernel/dma/direct.c
254
page = __dma_direct_alloc_pages(dev, size, gfp & ~__GFP_ZERO, true);
kernel/dma/direct.c
354
dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp)
kernel/dma/direct.c
359
if (force_dma_unencrypted(dev) && dma_direct_use_pool(dev, gfp))
kernel/dma/direct.c
360
return dma_direct_alloc_from_pool(dev, size, dma_handle, gfp);
kernel/dma/direct.c
362
page = __dma_direct_alloc_pages(dev, size, gfp, false);
kernel/dma/mapping.c
103
vaddr = dma_alloc_attrs(dev, size, dma_handle, gfp, attrs);
kernel/dma/mapping.c
698
dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp)
kernel/dma/mapping.c
704
if (WARN_ON_ONCE(gfp & (__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM)))
kernel/dma/mapping.c
706
if (WARN_ON_ONCE(gfp & __GFP_COMP))
kernel/dma/mapping.c
711
return dma_direct_alloc_pages(dev, size, dma_handle, dir, gfp);
kernel/dma/mapping.c
713
return dma_common_alloc_pages(dev, size, dma_handle, dir, gfp);
kernel/dma/mapping.c
716
return ops->alloc_pages_op(dev, size, dma_handle, dir, gfp);
kernel/dma/mapping.c
720
dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp)
kernel/dma/mapping.c
722
struct page *page = __dma_alloc_pages(dev, size, dma_handle, dir, gfp);
kernel/dma/mapping.c
726
size, dir, gfp, 0);
kernel/dma/mapping.c
729
trace_dma_alloc_pages(dev, NULL, 0, size, dir, gfp, 0);
kernel/dma/mapping.c
772
enum dma_data_direction dir, gfp_t gfp)
kernel/dma/mapping.c
777
sgt = kmalloc_obj(*sgt, gfp);
kernel/dma/mapping.c
780
if (sg_alloc_table(sgt, 1, gfp))
kernel/dma/mapping.c
782
page = __dma_alloc_pages(dev, size, &sgt->sgl->dma_address, dir, gfp);
kernel/dma/mapping.c
796
enum dma_data_direction dir, gfp_t gfp, unsigned long attrs)
kernel/dma/mapping.c
802
if (WARN_ON_ONCE(gfp & __GFP_COMP))
kernel/dma/mapping.c
806
sgt = iommu_dma_alloc_noncontiguous(dev, size, dir, gfp, attrs);
kernel/dma/mapping.c
808
sgt = alloc_single_sgt(dev, size, dir, gfp);
kernel/dma/mapping.c
812
trace_dma_alloc_sgt(dev, sgt, size, dir, gfp, attrs);
kernel/dma/mapping.c
815
trace_dma_alloc_sgt_err(dev, NULL, 0, size, dir, gfp, attrs);
kernel/dma/mapping.c
94
gfp_t gfp, unsigned long attrs)
kernel/dma/mapping.c
99
dr = devres_alloc(dmam_release, sizeof(*dr), gfp);
kernel/dma/ops_helpers.c
63
dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp)
kernel/dma/ops_helpers.c
69
page = dma_alloc_contiguous(dev, size, gfp);
kernel/dma/ops_helpers.c
71
page = alloc_pages_node(dev_to_node(dev), gfp, get_order(size));
kernel/dma/pool.c
125
dma_atomic_pool_size_add(gfp, pool_size);
kernel/dma/pool.c
145
static void atomic_pool_resize(struct gen_pool *pool, gfp_t gfp)
kernel/dma/pool.c
148
atomic_pool_expand(pool, gen_pool_size(pool), gfp);
kernel/dma/pool.c
163
gfp_t gfp)
kernel/dma/pool.c
174
ret = atomic_pool_expand(pool, pool_size, gfp);
kernel/dma/pool.c
178
pool_size >> 10, &gfp);
kernel/dma/pool.c
183
gen_pool_size(pool) >> 10, &gfp);
kernel/dma/pool.c
233
static inline struct gen_pool *dma_guess_pool(struct gen_pool *prev, gfp_t gfp)
kernel/dma/pool.c
236
if (gfp & GFP_DMA)
kernel/dma/pool.c
238
if (gfp & GFP_DMA32)
kernel/dma/pool.c
275
void **cpu_addr, gfp_t gfp,
kernel/dma/pool.c
282
while ((pool = dma_guess_pool(pool, gfp))) {
kernel/dma/pool.c
291
WARN(!(gfp & __GFP_NOWARN), "DMA pool exhausted for %s\n", dev_name(dev));
kernel/dma/pool.c
46
static void dma_atomic_pool_size_add(gfp_t gfp, size_t size)
kernel/dma/pool.c
48
if (gfp & __GFP_DMA)
kernel/dma/pool.c
50
else if (gfp & __GFP_DMA32)
kernel/dma/pool.c
56
static bool cma_in_zone(gfp_t gfp)
kernel/dma/pool.c
72
if (IS_ENABLED(CONFIG_ZONE_DMA) && (gfp & GFP_DMA))
kernel/dma/pool.c
74
if (IS_ENABLED(CONFIG_ZONE_DMA32) && (gfp & GFP_DMA32))
kernel/dma/pool.c
80
gfp_t gfp)
kernel/dma/pool.c
92
if (cma_in_zone(gfp))
kernel/dma/pool.c
96
page = alloc_pages(gfp | __GFP_NOWARN, order);
kernel/dma/swiotlb.c
573
static struct page *alloc_dma_pages(gfp_t gfp, size_t bytes, u64 phys_limit)
kernel/dma/swiotlb.c
580
page = alloc_pages(gfp, order);
kernel/dma/swiotlb.c
612
u64 phys_limit, gfp_t gfp)
kernel/dma/swiotlb.c
620
if (!gfpflags_allow_blocking(gfp) && dev && force_dma_unencrypted(dev)) {
kernel/dma/swiotlb.c
626
return dma_alloc_from_pool(dev, bytes, &vaddr, gfp,
kernel/dma/swiotlb.c
630
gfp &= ~GFP_ZONEMASK;
kernel/dma/swiotlb.c
632
gfp |= __GFP_DMA;
kernel/dma/swiotlb.c
634
gfp |= __GFP_DMA32;
kernel/dma/swiotlb.c
636
while (IS_ERR(page = alloc_dma_pages(gfp, bytes, phys_limit))) {
kernel/dma/swiotlb.c
639
!(gfp & (__GFP_DMA32 | __GFP_DMA)))
kernel/dma/swiotlb.c
640
gfp |= __GFP_DMA32;
kernel/dma/swiotlb.c
642
!(gfp & __GFP_DMA))
kernel/dma/swiotlb.c
643
gfp = (gfp & ~__GFP_DMA32) | __GFP_DMA;
kernel/dma/swiotlb.c
684
unsigned int nareas, u64 phys_limit, gfp_t gfp)
kernel/dma/swiotlb.c
698
pool = kzalloc(pool_size, gfp);
kernel/dma/swiotlb.c
704
while (!(tlb = swiotlb_alloc_tlb(dev, tlb_size, phys_limit, gfp))) {
kernel/dma/swiotlb.c
714
__get_free_pages(gfp, slot_order);
kernel/power/swap.c
361
gfp_t gfp = GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY;
kernel/power/swap.c
371
src = (void *)__get_free_page(gfp);
kernel/power/swap.c
376
src = (void *)__get_free_page(gfp);
kernel/rcu/tree.c
2005
static bool rcu_gp_fqs_check_wake(int *gfp)
kernel/rcu/tree.c
2010
if (*gfp & RCU_GP_FLAG_OVLD)
kernel/rcu/tree.c
2014
*gfp = READ_ONCE(rcu_state.gp_flags);
kernel/rcu/tree.c
2015
if (*gfp & RCU_GP_FLAG_FQS)
lib/argv_split.c
60
char **argv_split(gfp_t gfp, const char *str, int *argcp)
lib/argv_split.c
67
argv_str = kstrndup(str, KMALLOC_MAX_SIZE - 1, gfp);
lib/argv_split.c
72
argv = kmalloc_array(argc + 2, sizeof(*argv), gfp);
lib/btree.c
409
gfp_t gfp)
lib/btree.c
414
node = btree_node_alloc(head, gfp);
lib/btree.c
445
gfp_t gfp)
lib/btree.c
452
err = btree_grow(head, geo, gfp);
lib/btree.c
468
new = btree_node_alloc(head, gfp);
lib/btree.c
473
new, level + 1, gfp);
lib/btree.c
506
unsigned long *key, void *val, gfp_t gfp)
lib/btree.c
509
return btree_insert_level(head, geo, key, val, 1, gfp);
lib/btree.c
635
struct btree_geo *geo, gfp_t gfp)
lib/btree.c
659
err = btree_insert(target, geo, key, val, gfp);
lib/btree.c
92
static unsigned long *btree_node_alloc(struct btree_head *head, gfp_t gfp)
lib/btree.c
96
node = mempool_alloc(head->mempool, gfp);
lib/bucket_locks.c
15
size_t max_size, unsigned int cpu_mult, gfp_t gfp,
lib/bucket_locks.c
34
tlocks = kvmalloc_objs(spinlock_t, size, gfp);
lib/debugobjects.c
365
static bool kmem_alloc_batch(struct hlist_head *head, struct kmem_cache *cache, gfp_t gfp)
lib/debugobjects.c
371
obj = kmem_cache_zalloc(cache, gfp);
lib/debugobjects.c
401
gfp_t gfp = __GFP_HIGH | __GFP_NOWARN;
lib/debugobjects.c
418
gfp |= __GFP_KSWAPD_RECLAIM;
lib/debugobjects.c
420
if (!kmem_alloc_batch(&head, obj_cache, gfp))
lib/flex_proportions.c
38
int fprop_global_init(struct fprop_global *p, gfp_t gfp)
lib/flex_proportions.c
44
err = percpu_counter_init(&p->events, 1, gfp);
lib/flex_proportions.c
92
int fprop_local_init_percpu(struct fprop_local_percpu *pl, gfp_t gfp)
lib/flex_proportions.c
96
err = percpu_counter_init(&pl->events, 0, gfp);
lib/idr.c
119
int idr_alloc_cyclic(struct idr *idr, void *ptr, int start, int end, gfp_t gfp)
lib/idr.c
127
err = idr_alloc_u32(idr, ptr, &id, max, gfp);
lib/idr.c
130
err = idr_alloc_u32(idr, ptr, &id, max, gfp);
lib/idr.c
34
unsigned long max, gfp_t gfp)
lib/idr.c
383
gfp_t gfp)
lib/idr.c
456
if (xas_nomem(&xas, gfp)) {
lib/idr.c
468
alloc = kzalloc_obj(*bitmap, gfp);
lib/idr.c
48
slot = idr_get_free(&idr->idr_rt, &iter, gfp, max - base);
lib/idr.c
81
int idr_alloc(struct idr *idr, void *ptr, int start, int end, gfp_t gfp)
lib/idr.c
89
ret = idr_alloc_u32(idr, ptr, &id, end > 0 ? end - 1 : INT_MAX, gfp);
lib/kasprintf.c
15
char *kvasprintf(gfp_t gfp, const char *fmt, va_list ap)
lib/kasprintf.c
25
p = kmalloc_track_caller(first+1, gfp);
lib/kasprintf.c
43
const char *kvasprintf_const(gfp_t gfp, const char *fmt, va_list ap)
lib/kasprintf.c
46
return kstrdup_const(fmt, gfp);
lib/kasprintf.c
48
return kstrdup_const(va_arg(ap, const char*), gfp);
lib/kasprintf.c
49
return kvasprintf(gfp, fmt, ap);
lib/kasprintf.c
53
char *kasprintf(gfp_t gfp, const char *fmt, ...)
lib/kasprintf.c
59
p = kvasprintf(gfp, fmt, ap);
lib/kunit/string-stream-test.c
52
KUNIT_EXPECT_TRUE(test, (stream->gfp == GFP_KERNEL));
lib/kunit/string-stream-test.c
68
KUNIT_EXPECT_TRUE(test, (stream->gfp == GFP_KERNEL));
lib/kunit/string-stream.c
123
buf = kzalloc(buf_len, stream->gfp);
lib/kunit/string-stream.c
157
struct string_stream *alloc_string_stream(gfp_t gfp)
lib/kunit/string-stream.c
161
stream = kzalloc_obj(*stream, gfp);
lib/kunit/string-stream.c
165
stream->gfp = gfp;
lib/kunit/string-stream.c
17
static struct string_stream_fragment *alloc_string_stream_fragment(int len, gfp_t gfp)
lib/kunit/string-stream.c
190
struct string_stream *kunit_alloc_string_stream(struct kunit *test, gfp_t gfp)
lib/kunit/string-stream.c
194
stream = alloc_string_stream(gfp);
lib/kunit/string-stream.c
21
frag = kzalloc_obj(*frag, gfp);
lib/kunit/string-stream.c
25
frag->fragment = kmalloc(len, gfp);
lib/kunit/string-stream.c
67
frag_container = alloc_string_stream_fragment(buf_len, stream->gfp);
lib/kunit/string-stream.h
26
gfp_t gfp;
lib/kunit/string-stream.h
32
struct string_stream *kunit_alloc_string_stream(struct kunit *test, gfp_t gfp);
lib/kunit/string-stream.h
35
struct string_stream *alloc_string_stream(gfp_t gfp);
lib/kunit/test.c
1013
const char *kunit_kstrdup_const(struct kunit *test, const char *str, gfp_t gfp)
lib/kunit/test.c
1019
return kunit_kstrdup(test, str, gfp);
lib/kunit/test.c
979
void *kunit_kmalloc_array(struct kunit *test, size_t n, size_t size, gfp_t gfp)
lib/kunit/test.c
983
data = kmalloc_array(n, size, gfp);
lib/maple_tree.c
1098
static inline void mas_alloc_nodes(struct ma_state *mas, gfp_t gfp)
lib/maple_tree.c
1110
mas->alloc = mt_alloc_one(gfp);
lib/maple_tree.c
1133
if (mt_refill_sheaf(gfp, &mas->sheaf, refill))
lib/maple_tree.c
1140
mas->sheaf = mt_get_sheaf(gfp, mas->node_request);
lib/maple_tree.c
172
static inline struct maple_node *mt_alloc_one(gfp_t gfp)
lib/maple_tree.c
174
return kmem_cache_alloc(maple_node_cache, gfp);
lib/maple_tree.c
187
static struct slab_sheaf *mt_get_sheaf(gfp_t gfp, int count)
lib/maple_tree.c
189
return kmem_cache_prefill_sheaf(maple_node_cache, gfp, count);
lib/maple_tree.c
192
static int mt_refill_sheaf(gfp_t gfp, struct slab_sheaf **sheaf,
lib/maple_tree.c
195
return kmem_cache_refill_sheaf(maple_node_cache, gfp, sheaf, size);
lib/maple_tree.c
4027
unsigned long *next, gfp_t gfp)
lib/maple_tree.c
4049
} while (mas_nomem(mas, gfp));
lib/maple_tree.c
5118
int mas_store_gfp(struct ma_state *mas, void *entry, gfp_t gfp)
lib/maple_tree.c
5127
if (unlikely(mas_nomem(mas, gfp))) {
lib/maple_tree.c
5183
int mas_preallocate(struct ma_state *mas, void *entry, gfp_t gfp)
lib/maple_tree.c
5194
mas_alloc_nodes(mas, gfp);
lib/maple_tree.c
5842
bool mas_nomem(struct ma_state *mas, gfp_t gfp)
lib/maple_tree.c
5848
if (gfpflags_allow_blocking(gfp) && !mt_external_lock(mas->tree)) {
lib/maple_tree.c
5850
mas_alloc_nodes(mas, gfp);
lib/maple_tree.c
5853
mas_alloc_nodes(mas, gfp);
lib/maple_tree.c
5925
unsigned long last, void *entry, gfp_t gfp)
lib/maple_tree.c
5938
ret = mas_store_gfp(&mas, entry, gfp);
lib/maple_tree.c
5956
gfp_t gfp)
lib/maple_tree.c
5958
return mtree_store_range(mt, index, index, entry, gfp);
lib/maple_tree.c
5974
unsigned long last, void *entry, gfp_t gfp)
lib/maple_tree.c
5988
if (mas_nomem(&ms, gfp))
lib/maple_tree.c
6011
gfp_t gfp)
lib/maple_tree.c
6013
return mtree_insert_range(mt, index, index, entry, gfp);
lib/maple_tree.c
6019
unsigned long max, gfp_t gfp)
lib/maple_tree.c
6041
if (mas_nomem(&mas, gfp))
lib/maple_tree.c
6081
unsigned long *next, gfp_t gfp)
lib/maple_tree.c
6093
next, gfp);
lib/maple_tree.c
6101
unsigned long max, gfp_t gfp)
lib/maple_tree.c
6123
if (mas_nomem(&mas, gfp))
lib/maple_tree.c
6241
gfp_t gfp)
lib/maple_tree.c
6255
mas_alloc_nodes(mas, gfp);
lib/maple_tree.c
6282
gfp_t gfp)
lib/maple_tree.c
6299
node = mt_alloc_one(gfp);
lib/maple_tree.c
6316
mas_dup_alloc(mas, new_mas, gfp);
lib/maple_tree.c
6376
int __mt_dup(struct maple_tree *mt, struct maple_tree *new, gfp_t gfp)
lib/maple_tree.c
6382
mas_dup_build(&mas, &new_mas, gfp);
lib/maple_tree.c
6413
int mtree_dup(struct maple_tree *mt, struct maple_tree *new, gfp_t gfp)
lib/maple_tree.c
6421
mas_dup_build(&mas, &new_mas, gfp);
lib/memregion.c
11
return ida_alloc(&memregion_ids, gfp);
lib/memregion.c
9
int memregion_alloc(gfp_t gfp)
lib/objpool.c
119
gfp_t gfp, void *context, objpool_init_obj_cb objinit,
lib/objpool.c
142
pool->gfp = gfp & ~__GFP_ZERO;
lib/objpool.c
146
pool->cpu_slots = kzalloc(slot_size, pool->gfp);
lib/objpool.c
83
if ((pool->gfp & (GFP_ATOMIC | GFP_KERNEL)) != GFP_ATOMIC)
lib/objpool.c
84
slot = __vmalloc_node(size, sizeof(void *), pool->gfp,
lib/objpool.c
88
slot = kmalloc_node(size, pool->gfp, cpu_to_node(i));
lib/percpu-refcount.c
64
unsigned int flags, gfp_t gfp)
lib/percpu-refcount.c
72
__alloc_percpu_gfp(sizeof(unsigned long), align, gfp);
lib/percpu-refcount.c
76
data = kzalloc_obj(*ref->data, gfp);
lib/percpu_counter.c
189
gfp_t gfp, u32 nr_counters,
lib/percpu_counter.c
199
__alignof__(*counters), gfp);
lib/radix-tree.c
1477
struct radix_tree_iter *iter, gfp_t gfp,
lib/radix-tree.c
1493
int error = radix_tree_extend(root, gfp, start, shift);
lib/radix-tree.c
1506
child = radix_tree_node_alloc(gfp, node, root, shift,
lib/radix-tree.c
408
static int radix_tree_extend(struct radix_tree_root *root, gfp_t gfp,
lib/radix-tree.c
425
struct radix_tree_node *node = radix_tree_node_alloc(gfp, NULL,
lib/radix-tree.c
607
gfp_t gfp = root_gfp_mask(root);
lib/radix-tree.c
613
int error = radix_tree_extend(root, gfp, max, shift);
lib/radix-tree.c
624
child = radix_tree_node_alloc(gfp, node, root, shift,
lib/reed_solomon/reed_solomon.c
215
int prim, int nroots, gfp_t gfp)
lib/reed_solomon/reed_solomon.c
237
rs = kzalloc(sizeof(*rs) + bsize, gfp);
lib/reed_solomon/reed_solomon.c
266
rs->codec = codec_init(symsize, gfpoly, gffunc, fcr, prim, nroots, gfp);
lib/reed_solomon/reed_solomon.c
289
int nroots, gfp_t gfp)
lib/reed_solomon/reed_solomon.c
291
return init_rs_internal(symsize, gfpoly, NULL, fcr, prim, nroots, gfp);
lib/reed_solomon/reed_solomon.c
71
int fcr, int prim, int nroots, gfp_t gfp)
lib/reed_solomon/reed_solomon.c
76
rs = kzalloc_obj(*rs, gfp);
lib/reed_solomon/reed_solomon.c
91
rs->alpha_to = kmalloc_array(rs->nn + 1, sizeof(uint16_t), gfp);
lib/reed_solomon/reed_solomon.c
95
rs->index_of = kmalloc_array(rs->nn + 1, sizeof(uint16_t), gfp);
lib/reed_solomon/reed_solomon.c
99
rs->genpoly = kmalloc_array(rs->nroots + 1, sizeof(uint16_t), gfp);
lib/ref_tracker.c
254
gfp_t gfp)
lib/ref_tracker.c
259
gfp_t gfp_mask = gfp | __GFP_NOWARN;
lib/ref_tracker.c
268
if (gfp & __GFP_DIRECT_RECLAIM)
lib/ref_tracker.c
277
tracker->alloc_stack_handle = stack_depot_save(entries, nr_entries, gfp);
lib/rhashtable.c
150
gfp_t gfp)
lib/rhashtable.c
162
kmalloc_noprof(size, gfp|__GFP_ZERO));
lib/rhashtable.c
179
gfp_t gfp)
lib/rhashtable.c
188
1, gfp|__GFP_ZERO, NUMA_NO_NODE));
lib/rhashtable.c
192
if (tbl == NULL && !gfpflags_allow_blocking(gfp)) {
lib/rhashtable.c
193
tbl = nested_bucket_table_alloc(ht, nbuckets, gfp);
lib/scatterlist.c
616
gfp_t gfp, unsigned int *nent_p)
lib/scatterlist.c
634
sgl = kmalloc_objs(struct scatterlist, nalloc, gfp & ~GFP_DMA);
lib/scatterlist.c
642
page = alloc_pages(gfp, order);
lib/scatterlist.c
667
struct scatterlist *sgl_alloc(unsigned long long length, gfp_t gfp,
lib/scatterlist.c
670
return sgl_alloc_order(length, 0, false, gfp, nent_p);
lib/string_helpers.c
652
char *kstrdup_quotable(const char *src, gfp_t gfp)
lib/string_helpers.c
664
dst = kmalloc(dlen + 1, gfp);
lib/string_helpers.c
680
char *kstrdup_quotable_cmdline(struct task_struct *task, gfp_t gfp)
lib/string_helpers.c
702
quoted = kstrdup_quotable(buffer, gfp);
lib/string_helpers.c
713
char *kstrdup_quotable_file(struct file *file, gfp_t gfp)
lib/string_helpers.c
718
return kstrdup("<unknown>", gfp);
lib/string_helpers.c
723
return kstrdup("<no_memory>", gfp);
lib/string_helpers.c
727
pathname = kstrdup("<too_long>", gfp);
lib/string_helpers.c
729
pathname = kstrdup_quotable(pathname, gfp);
lib/string_helpers.c
739
char *kstrdup_and_replace(const char *src, char old, char new, gfp_t gfp)
lib/string_helpers.c
743
dst = kstrdup(src, gfp);
lib/string_helpers.c
763
char **kasprintf_strarray(gfp_t gfp, const char *prefix, size_t n)
lib/string_helpers.c
768
names = kcalloc(n + 1, sizeof(char *), gfp);
lib/string_helpers.c
773
names[i] = kasprintf(gfp, "%s-%zu", prefix, i);
lib/test_firmware.c
214
static int __kstrncpy(char **dst, const char *name, size_t count, gfp_t gfp)
lib/test_firmware.c
216
*dst = kstrndup(name, count, gfp);
lib/test_free_pages.c
14
static void test_free_pages(gfp_t gfp)
lib/test_free_pages.c
19
unsigned long addr = __get_free_pages(gfp, 3);
lib/test_kmod.c
608
static int __kstrncpy(char **dst, const char *name, size_t count, gfp_t gfp)
lib/test_kmod.c
610
*dst = kstrndup(name, count, gfp);
lib/test_maple_tree.c
64
unsigned long index, gfp_t gfp)
lib/test_maple_tree.c
66
return mtree_insert(mt, index, xa_mk_value(index & LONG_MAX), gfp);
lib/test_objpool.c
288
gfp_t gfp = GFP_KERNEL;
lib/test_objpool.c
295
gfp = GFP_ATOMIC;
lib/test_objpool.c
298
gfp, sop, ot_init_node, NULL)) {
lib/test_objpool.c
454
gfp_t gfp = GFP_KERNEL;
lib/test_objpool.c
461
gfp = GFP_ATOMIC;
lib/test_objpool.c
463
if (objpool_init(&sop->pool, max, test->objsz, gfp, sop,
lib/test_vmalloc.c
297
gfp_t gfp = use_atomic ? GFP_ATOMIC : GFP_NOWAIT;
lib/test_vmalloc.c
301
ptr = __vmalloc(size, gfp);
lib/test_xarray.c
40
static void *xa_store_index(struct xarray *xa, unsigned long index, gfp_t gfp)
lib/test_xarray.c
42
return xa_store(xa, index, xa_mk_index(index), gfp);
lib/test_xarray.c
51
static void xa_alloc_index(struct xarray *xa, unsigned long index, gfp_t gfp)
lib/test_xarray.c
56
gfp) != 0);
lib/test_xarray.c
72
unsigned order, void *entry, gfp_t gfp)
lib/test_xarray.c
81
} while (xas_nomem(&xas, gfp));
lib/tests/fortify_kunit.c
220
gfp_t gfp = GFP_KERNEL | __GFP_NOWARN; \
lib/tests/fortify_kunit.c
224
checker(expected_size, kmalloc(alloc_size, gfp), \
lib/tests/fortify_kunit.c
227
kmalloc_node(alloc_size, gfp, NUMA_NO_NODE), \
lib/tests/fortify_kunit.c
229
checker(expected_size, kzalloc(alloc_size, gfp), \
lib/tests/fortify_kunit.c
232
kzalloc_node(alloc_size, gfp, NUMA_NO_NODE), \
lib/tests/fortify_kunit.c
234
checker(expected_size, kcalloc(1, alloc_size, gfp), \
lib/tests/fortify_kunit.c
236
checker(expected_size, kcalloc(alloc_size, 1, gfp), \
lib/tests/fortify_kunit.c
239
kcalloc_node(1, alloc_size, gfp, NUMA_NO_NODE), \
lib/tests/fortify_kunit.c
242
kcalloc_node(alloc_size, 1, gfp, NUMA_NO_NODE), \
lib/tests/fortify_kunit.c
244
checker(expected_size, kmalloc_array(1, alloc_size, gfp), \
lib/tests/fortify_kunit.c
246
checker(expected_size, kmalloc_array(alloc_size, 1, gfp), \
lib/tests/fortify_kunit.c
249
kmalloc_array_node(1, alloc_size, gfp, NUMA_NO_NODE), \
lib/tests/fortify_kunit.c
252
kmalloc_array_node(alloc_size, 1, gfp, NUMA_NO_NODE), \
lib/tests/fortify_kunit.c
255
orig = kmalloc(alloc_size, gfp); \
lib/tests/fortify_kunit.c
258
krealloc(orig, (alloc_size) * 2, gfp), \
lib/tests/fortify_kunit.c
260
orig = kmalloc(alloc_size, gfp); \
lib/tests/fortify_kunit.c
263
krealloc_array(orig, 1, (alloc_size) * 2, gfp), \
lib/tests/fortify_kunit.c
265
orig = kmalloc(alloc_size, gfp); \
lib/tests/fortify_kunit.c
268
krealloc_array(orig, (alloc_size) * 2, 1, gfp), \
lib/tests/fortify_kunit.c
275
checker(len, kmemdup("hello there", len, gfp), kfree(p)); \
lib/tests/fortify_kunit.c
281
gfp_t gfp = GFP_KERNEL | __GFP_NOWARN; \
lib/tests/fortify_kunit.c
287
__vmalloc((alloc_pages) * PAGE_SIZE, gfp), vfree(p)); \
lib/tests/fortify_kunit.c
293
gfp_t gfp = GFP_KERNEL | __GFP_NOWARN; \
lib/tests/fortify_kunit.c
298
kvmalloc((alloc_pages) * PAGE_SIZE, gfp), \
lib/tests/fortify_kunit.c
301
kvmalloc_node((alloc_pages) * PAGE_SIZE, gfp, NUMA_NO_NODE), \
lib/tests/fortify_kunit.c
304
kvzalloc((alloc_pages) * PAGE_SIZE, gfp), \
lib/tests/fortify_kunit.c
307
kvzalloc_node((alloc_pages) * PAGE_SIZE, gfp, NUMA_NO_NODE), \
lib/tests/fortify_kunit.c
310
kvcalloc(1, (alloc_pages) * PAGE_SIZE, gfp), \
lib/tests/fortify_kunit.c
313
kvcalloc((alloc_pages) * PAGE_SIZE, 1, gfp), \
lib/tests/fortify_kunit.c
316
kvmalloc_array(1, (alloc_pages) * PAGE_SIZE, gfp), \
lib/tests/fortify_kunit.c
319
kvmalloc_array((alloc_pages) * PAGE_SIZE, 1, gfp), \
lib/tests/fortify_kunit.c
323
orig = kvmalloc(prev_size, gfp); \
lib/tests/fortify_kunit.c
326
kvrealloc(orig, ((alloc_pages) * PAGE_SIZE) * 2, gfp), \
lib/tests/fortify_kunit.c
332
gfp_t gfp = GFP_KERNEL | __GFP_NOWARN; \
lib/tests/fortify_kunit.c
343
checker(expected_size, devm_kmalloc(dev, alloc_size, gfp), \
lib/tests/fortify_kunit.c
345
checker(expected_size, devm_kzalloc(dev, alloc_size, gfp), \
lib/tests/fortify_kunit.c
348
devm_kmalloc_array(dev, 1, alloc_size, gfp), \
lib/tests/fortify_kunit.c
351
devm_kmalloc_array(dev, alloc_size, 1, gfp), \
lib/tests/fortify_kunit.c
354
devm_kcalloc(dev, 1, alloc_size, gfp), \
lib/tests/fortify_kunit.c
357
devm_kcalloc(dev, alloc_size, 1, gfp), \
lib/tests/fortify_kunit.c
360
orig = devm_kmalloc(dev, alloc_size, gfp); \
lib/tests/fortify_kunit.c
363
devm_krealloc(dev, orig, (alloc_size) * 2, gfp), \
lib/tests/fortify_kunit.c
370
checker(len, devm_kmemdup(dev, "Ohai", len, gfp), \
lib/tests/fortify_kunit.c
384
gfp_t gfp = GFP_KERNEL; \
lib/tests/fortify_kunit.c
391
checker(len, kmemdup_array(test_strs[i], 1, len, gfp), \
lib/tests/fortify_kunit.c
393
checker(len, kmemdup(test_strs[i], len, gfp), \
lib/tests/printf_kunit.c
644
gfp_t gfp;
lib/tests/printf_kunit.c
664
gfp = GFP_TRANSHUGE;
lib/tests/printf_kunit.c
665
test("GFP_TRANSHUGE", "%pGg", &gfp);
lib/tests/printf_kunit.c
667
gfp = GFP_ATOMIC|__GFP_DMA;
lib/tests/printf_kunit.c
668
test("GFP_ATOMIC|GFP_DMA", "%pGg", &gfp);
lib/tests/printf_kunit.c
670
gfp = __GFP_HIGH;
lib/tests/printf_kunit.c
671
test("__GFP_HIGH", "%pGg", &gfp);
lib/tests/printf_kunit.c
674
gfp = ~__GFP_BITS_MASK;
lib/tests/printf_kunit.c
675
snprintf(cmp_buffer, BUF_SIZE, "%#lx", (unsigned long) gfp);
lib/tests/printf_kunit.c
676
test(cmp_buffer, "%pGg", &gfp);
lib/tests/printf_kunit.c
679
(unsigned long) gfp);
lib/tests/printf_kunit.c
680
gfp |= __GFP_HIGH;
lib/tests/printf_kunit.c
681
test(cmp_buffer, "%pGg", &gfp);
lib/xarray.c
1046
gfp_t gfp)
lib/xarray.c
1059
node = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
lib/xarray.c
1182
gfp_t gfp = GFP_NOWAIT;
lib/xarray.c
1189
gfp |= __GFP_ACCOUNT;
lib/xarray.c
1215
xas->xa_lru, gfp);
lib/xarray.c
1692
void *__xa_store(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp)
lib/xarray.c
1706
} while (__xas_nomem(&xas, gfp));
lib/xarray.c
1729
void *xa_store(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp)
lib/xarray.c
1734
curr = __xa_store(xa, index, entry, gfp);
lib/xarray.c
1742
void *old, void *entry, gfp_t gfp);
lib/xarray.c
1764
void *old, void *entry, gfp_t gfp)
lib/xarray.c
1766
return xa_zero_to_null(__xa_cmpxchg_raw(xa, index, old, entry, gfp));
lib/xarray.c
1771
void *old, void *entry, gfp_t gfp)
lib/xarray.c
1786
} while (__xas_nomem(&xas, gfp));
lib/xarray.c
1807
int __xa_insert(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp)
lib/xarray.c
1814
curr = __xa_cmpxchg_raw(xa, index, NULL, entry, gfp);
lib/xarray.c
1873
unsigned long last, void *entry, gfp_t gfp)
lib/xarray.c
1902
} while (xas_nomem(&xas, gfp));
lib/xarray.c
1986
struct xa_limit limit, gfp_t gfp)
lib/xarray.c
2007
} while (__xas_nomem(&xas, gfp));
lib/xarray.c
2038
struct xa_limit limit, u32 *next, gfp_t gfp)
lib/xarray.c
2044
ret = __xa_alloc(xa, id, entry, limit, gfp);
lib/xarray.c
2052
ret = __xa_alloc(xa, id, entry, limit, gfp);
lib/xarray.c
301
bool xas_nomem(struct xa_state *xas, gfp_t gfp)
lib/xarray.c
308
gfp |= __GFP_ACCOUNT;
lib/xarray.c
309
xas->xa_alloc = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
lib/xarray.c
328
static bool __xas_nomem(struct xa_state *xas, gfp_t gfp)
lib/xarray.c
338
gfp |= __GFP_ACCOUNT;
lib/xarray.c
339
if (gfpflags_allow_blocking(gfp)) {
lib/xarray.c
341
xas->xa_alloc = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
lib/xarray.c
344
xas->xa_alloc = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
lib/xarray.c
373
gfp_t gfp = GFP_NOWAIT;
lib/xarray.c
376
gfp |= __GFP_ACCOUNT;
lib/xarray.c
378
node = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
mm/backing-dev.c
516
gfp_t gfp)
mm/backing-dev.c
542
err = fprop_local_init_percpu(&wb->completions, gfp);
mm/backing-dev.c
546
err = percpu_counter_init_many(wb->stat, 0, gfp, NR_WB_STAT_ITEMS);
mm/backing-dev.c
666
struct cgroup_subsys_state *memcg_css, gfp_t gfp)
mm/backing-dev.c
692
wb = kmalloc_obj(*wb, gfp);
mm/backing-dev.c
698
ret = wb_init(wb, bdi, gfp);
mm/backing-dev.c
702
ret = percpu_ref_init(&wb->refcnt, cgwb_release, 0, gfp);
mm/backing-dev.c
706
ret = fprop_local_init_percpu(&wb->memcg_completions, gfp);
mm/backing-dev.c
820
gfp_t gfp)
mm/backing-dev.c
824
might_alloc(gfp);
mm/backing-dev.c
828
} while (!wb && !cgwb_create(bdi, memcg_css, gfp));
mm/cma.c
782
struct page **pagep, gfp_t gfp)
mm/cma.c
840
ret = alloc_contig_frozen_range(pfn, pfn + count, ACR_FLAGS_CMA, gfp);
mm/cma.c
861
unsigned long count, unsigned int align, gfp_t gfp)
mm/cma.c
883
&page, gfp);
mm/cma.c
898
if (ret && !(gfp & __GFP_NOWARN)) {
mm/cma.c
921
gfp_t gfp = GFP_KERNEL | (no_warn ? __GFP_NOWARN : 0);
mm/cma.c
923
return __cma_alloc_frozen(cma, count, align, gfp);
mm/cma.c
928
gfp_t gfp = GFP_KERNEL | __GFP_COMP | __GFP_NOWARN;
mm/cma.c
930
return __cma_alloc_frozen(cma, 1 << order, order, gfp);
mm/filemap.c
1000
return folio_alloc_mpol_noprof(gfp, order, policy,
mm/filemap.c
1008
folio = __folio_alloc_node_noprof(gfp, order, n);
mm/filemap.c
1013
return folio_alloc_noprof(gfp, order);
mm/filemap.c
1941
pgoff_t index, fgf_t fgp_flags, gfp_t gfp, struct mempolicy *policy)
mm/filemap.c
1989
gfp |= __GFP_WRITE;
mm/filemap.c
1991
gfp &= ~__GFP_FS;
mm/filemap.c
1993
gfp &= ~GFP_KERNEL;
mm/filemap.c
1994
gfp |= GFP_NOWAIT;
mm/filemap.c
2006
gfp_t alloc_gfp = gfp;
mm/filemap.c
2021
err = filemap_add_folio(mapping, folio, index, gfp);
mm/filemap.c
4059
pgoff_t index, filler_t filler, struct file *file, gfp_t gfp)
mm/filemap.c
4069
folio = filemap_alloc_folio(gfp, mapping_min_folio_order(mapping), NULL);
mm/filemap.c
4073
err = filemap_add_folio(mapping, folio, index, gfp);
mm/filemap.c
4161
pgoff_t index, gfp_t gfp)
mm/filemap.c
4163
return do_read_cache_folio(mapping, index, NULL, NULL, gfp);
mm/filemap.c
4168
pgoff_t index, filler_t *filler, struct file *file, gfp_t gfp)
mm/filemap.c
4172
folio = do_read_cache_folio(mapping, index, filler, file, gfp);
mm/filemap.c
4203
gfp_t gfp)
mm/filemap.c
4205
return do_read_cache_page(mapping, index, NULL, NULL, gfp);
mm/filemap.c
4493
bool filemap_release_folio(struct folio *folio, gfp_t gfp)
mm/filemap.c
4504
return mapping->a_ops->release_folio(folio, gfp);
mm/filemap.c
849
struct folio *folio, pgoff_t index, gfp_t gfp, void **shadowp)
mm/filemap.c
866
gfp &= GFP_RECLAIM_MASK;
mm/filemap.c
932
if (!xas_nomem(&xas, gfp))
mm/filemap.c
950
pgoff_t index, gfp_t gfp)
mm/filemap.c
959
ret = mem_cgroup_charge(folio, NULL, gfp);
mm/filemap.c
966
ret = __filemap_add_folio(mapping, folio, index, gfp, &shadow);
mm/filemap.c
980
if (!(gfp & __GFP_WRITE) && shadow)
mm/filemap.c
993
struct folio *filemap_alloc_folio_noprof(gfp_t gfp, unsigned int order,
mm/folio-compat.c
70
pgoff_t index, gfp_t gfp)
mm/folio-compat.c
72
return filemap_add_folio(mapping, page_folio(page), index, gfp);
mm/folio-compat.c
78
fgf_t fgp_flags, gfp_t gfp)
mm/folio-compat.c
82
folio = __filemap_get_folio(mapping, index, fgp_flags, gfp);
mm/huge_memory.c
1259
gfp_t gfp = vma_thp_gfp_mask(vma);
mm/huge_memory.c
1263
folio = vma_alloc_folio(gfp, order, vma, addr & HPAGE_PMD_MASK);
mm/huge_memory.c
1272
if (mem_cgroup_charge(folio, vma->vm_mm, gfp)) {
mm/huge_memory.c
1280
folio_throttle_swaprate(folio, gfp);
mm/huge_memory.c
3997
gfp_t gfp;
mm/huge_memory.c
4006
gfp = current_gfp_context(mapping_gfp_mask(mapping) &
mm/huge_memory.c
4009
if (!filemap_release_folio(folio, gfp)) {
mm/huge_memory.c
4016
xas_split_alloc(&xas, folio, old_order, gfp);
mm/hugetlb.c
2219
gfp_t gfp = gfp_mask & ~(__GFP_DIRECT_RECLAIM | __GFP_NOFAIL);
mm/hugetlb.c
2221
folio = alloc_surplus_hugetlb_folio(h, gfp, nid, nodemask);
mm/hugetlb.c
2277
static nodemask_t *policy_mbind_nodemask(gfp_t gfp)
mm/hugetlb.c
2287
(apply_policy_zone(mpol, gfp_zone(gfp)) &&
mm/hugetlb.c
2918
gfp_t gfp = htlb_alloc_mask(h) | __GFP_RETRY_MAYFAIL;
mm/hugetlb.c
3042
ret = mem_cgroup_charge_hugetlb(folio, gfp);
mm/internal.h
1667
pgoff_t index, void *expected, gfp_t gfp);
mm/internal.h
528
gfp_t gfp);
mm/internal.h
905
static inline struct page *alloc_frozen_pages_noprof(gfp_t gfp, unsigned int order)
mm/internal.h
907
return __alloc_frozen_pages_noprof(gfp, order, numa_node_id(), NULL);
mm/internal.h
92
#define WARN_ON_ONCE_GFP(cond, gfp) ({ \
mm/internal.h
96
if (unlikely(!(gfp & __GFP_NOWARN) && __ret_warn_once && !__warned)) { \
mm/kfence/core.c
418
static void *kfence_guarded_alloc(struct kmem_cache *cache, size_t size, gfp_t gfp,
mm/kfence/core.c
504
if (unlikely(slab_want_init_on_alloc(gfp, cache)))
mm/kfence/kfence_test.c
245
static void *test_alloc(struct kunit *test, size_t size, gfp_t gfp, enum allocation_policy policy)
mm/kfence/kfence_test.c
266
kunit_info(test, "%s: size=%zu, gfp=%x, policy=%s, cache=%i\n", __func__, size, gfp,
mm/kfence/kfence_test.c
282
alloc = kmem_cache_alloc(test_cache, gfp);
mm/kfence/kfence_test.c
284
alloc = kmalloc(size, gfp);
mm/khugepaged.c
1053
gfp_t gfp = (cc->is_khugepaged ? alloc_hugepage_khugepaged_gfpmask() :
mm/khugepaged.c
1058
folio = __folio_alloc(gfp, HPAGE_PMD_ORDER, node, &cc->alloc_nmask);
mm/khugepaged.c
1066
if (unlikely(mem_cgroup_charge(folio, mm, gfp))) {
mm/kmemleak.c
1002
gfp_nested_mask(gfp));
mm/kmemleak.c
1091
gfp_t gfp)
mm/kmemleak.c
1096
create_object((unsigned long)ptr, size, min_count, gfp);
mm/kmemleak.c
1110
gfp_t gfp)
mm/kmemleak.c
1115
create_object_percpu((__force unsigned long)ptr, size, 1, gfp);
mm/kmemleak.c
1128
void __ref kmemleak_vmalloc(const struct vm_struct *area, size_t size, gfp_t gfp)
mm/kmemleak.c
1137
create_object((unsigned long)area->addr, size, 2, gfp);
mm/kmemleak.c
1306
void __ref kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp)
mm/kmemleak.c
1311
add_scan_area((unsigned long)ptr, size, gfp);
mm/kmemleak.c
1340
void __ref kmemleak_alloc_phys(phys_addr_t phys, size_t size, gfp_t gfp)
mm/kmemleak.c
1349
create_object_phys((unsigned long)phys, size, 0, gfp);
mm/kmemleak.c
475
static struct kmemleak_object *mem_pool_alloc(gfp_t gfp)
mm/kmemleak.c
484
gfp_nested_mask(gfp));
mm/kmemleak.c
661
static struct kmemleak_object *__alloc_object(gfp_t gfp)
mm/kmemleak.c
665
object = mem_pool_alloc(gfp);
mm/kmemleak.c
773
int min_count, gfp_t gfp, unsigned int objflags)
mm/kmemleak.c
779
object = __alloc_object(gfp);
mm/kmemleak.c
792
int min_count, gfp_t gfp)
mm/kmemleak.c
794
__create_object(ptr, size, min_count, gfp, 0);
mm/kmemleak.c
799
int min_count, gfp_t gfp)
mm/kmemleak.c
801
__create_object(ptr, size, min_count, gfp, OBJECT_PHYS);
mm/kmemleak.c
806
int min_count, gfp_t gfp)
mm/kmemleak.c
808
__create_object(ptr, size, min_count, gfp, OBJECT_PERCPU);
mm/kmemleak.c
982
static void add_scan_area(unsigned long ptr, size_t size, gfp_t gfp)
mm/list_lru.c
405
static struct list_lru_memcg *memcg_init_list_lru_one(struct list_lru *lru, gfp_t gfp)
mm/list_lru.c
410
mlru = kmalloc_flex(*mlru, node, nr_node_ids, gfp);
mm/list_lru.c
514
gfp_t gfp)
mm/list_lru.c
524
gfp &= GFP_RECLAIM_MASK;
mm/list_lru.c
543
mlru = memcg_init_list_lru_one(lru, gfp);
mm/list_lru.c
556
} while (xas_nomem(&xas, gfp));
mm/memcontrol.c
2849
static int obj_cgroup_charge_pages(struct obj_cgroup *objcg, gfp_t gfp,
mm/memcontrol.c
2857
ret = try_charge_memcg(memcg, gfp, nr_pages);
mm/memcontrol.c
2894
int __memcg_kmem_charge_page(struct page *page, gfp_t gfp, int order)
mm/memcontrol.c
2901
ret = obj_cgroup_charge_pages(objcg, gfp, 1 << order);
mm/memcontrol.c
3122
static int obj_cgroup_charge_account(struct obj_cgroup *objcg, gfp_t gfp, size_t size,
mm/memcontrol.c
3160
ret = obj_cgroup_charge_pages(objcg, gfp, nr_pages);
mm/memcontrol.c
3168
int obj_cgroup_charge(struct obj_cgroup *objcg, gfp_t gfp, size_t size)
mm/memcontrol.c
3170
return obj_cgroup_charge_account(objcg, gfp, size, NULL, 0);
mm/memcontrol.c
3371
static int memcg_wb_domain_init(struct mem_cgroup *memcg, gfp_t gfp)
mm/memcontrol.c
3373
return wb_domain_init(&memcg->cgwb_domain, gfp);
mm/memcontrol.c
3565
static int memcg_wb_domain_init(struct mem_cgroup *memcg, gfp_t gfp)
mm/memcontrol.c
4740
gfp_t gfp)
mm/memcontrol.c
4744
ret = try_charge(memcg, gfp, folio_nr_pages(folio));
mm/memcontrol.c
4755
int __mem_cgroup_charge(struct folio *folio, struct mm_struct *mm, gfp_t gfp)
mm/memcontrol.c
4761
ret = charge_memcg(folio, memcg, gfp);
mm/memcontrol.c
4779
int mem_cgroup_charge_hugetlb(struct folio *folio, gfp_t gfp)
mm/memcontrol.c
4793
if (charge_memcg(folio, memcg, gfp))
mm/memcontrol.c
4814
gfp_t gfp, swp_entry_t entry)
mm/memcontrol.c
4830
ret = charge_memcg(folio, memcg, gfp);
mm/memory.c
4610
gfp_t gfp;
mm/memory.c
4662
gfp = vma_thp_gfp_mask(vma);
mm/memory.c
4665
folio = vma_alloc_folio(gfp, order, vma, addr);
mm/memory.c
4668
gfp, entry))
mm/memory.c
5135
gfp_t gfp;
mm/memory.c
5180
gfp = vma_thp_gfp_mask(vma);
mm/memory.c
5183
folio = vma_alloc_folio(gfp, order, vma, addr);
mm/memory.c
5185
if (mem_cgroup_charge(folio, vma->vm_mm, gfp)) {
mm/memory.c
5190
folio_throttle_swaprate(folio, gfp);
mm/mempolicy.c
1442
gfp_t gfp;
mm/mempolicy.c
1452
gfp = htlb_alloc_mask(h);
mm/mempolicy.c
1453
nodemask = policy_nodemask(gfp, pol, ilx, &nid);
mm/mempolicy.c
1454
return alloc_hugetlb_folio_nodemask(h, nid, nodemask, gfp,
mm/mempolicy.c
1459
gfp = GFP_TRANSHUGE;
mm/mempolicy.c
1461
gfp = GFP_HIGHUSER_MOVABLE | __GFP_RETRY_MAYFAIL | __GFP_COMP;
mm/mempolicy.c
1463
return folio_alloc_mpol(gfp, order, pol, ilx, nid);
mm/mempolicy.c
2269
static nodemask_t *policy_nodemask(gfp_t gfp, struct mempolicy *pol,
mm/mempolicy.c
2286
if (apply_policy_zone(pol, gfp_zone(gfp)) &&
mm/mempolicy.c
2296
WARN_ON_ONCE(gfp & __GFP_THISNODE);
mm/mempolicy.c
2414
static struct page *alloc_pages_preferred_many(gfp_t gfp, unsigned int order,
mm/mempolicy.c
2426
preferred_gfp = gfp | __GFP_NOWARN;
mm/mempolicy.c
2430
page = __alloc_frozen_pages_noprof(gfp, order, nid, NULL);
mm/mempolicy.c
2445
static struct page *alloc_pages_mpol(gfp_t gfp, unsigned int order,
mm/mempolicy.c
2451
nodemask = policy_nodemask(gfp, pol, ilx, &nid);
mm/mempolicy.c
2454
return alloc_pages_preferred_many(gfp, order, nid, nodemask);
mm/mempolicy.c
2477
gfp | __GFP_THISNODE | __GFP_NORETRY, order,
mm/mempolicy.c
2479
if (page || !(gfp & __GFP_DIRECT_RECLAIM))
mm/mempolicy.c
2490
page = __alloc_frozen_pages_noprof(gfp, order, nid, nodemask);
mm/mempolicy.c
2506
struct folio *folio_alloc_mpol_noprof(gfp_t gfp, unsigned int order,
mm/mempolicy.c
2509
struct page *page = alloc_pages_mpol(gfp | __GFP_COMP, order, pol,
mm/mempolicy.c
2533
struct folio *vma_alloc_folio_noprof(gfp_t gfp, int order, struct vm_area_struct *vma,
mm/mempolicy.c
2541
gfp |= __GFP_NOWARN;
mm/mempolicy.c
2544
folio = folio_alloc_mpol_noprof(gfp, order, pol, ilx, numa_node_id());
mm/mempolicy.c
2550
struct page *alloc_frozen_pages_noprof(gfp_t gfp, unsigned order)
mm/mempolicy.c
2558
if (!in_interrupt() && !(gfp & __GFP_THISNODE))
mm/mempolicy.c
2561
return alloc_pages_mpol(gfp, order, pol, NO_INTERLEAVE_INDEX,
mm/mempolicy.c
2579
struct page *alloc_pages_noprof(gfp_t gfp, unsigned int order)
mm/mempolicy.c
2581
struct page *page = alloc_frozen_pages_noprof(gfp, order);
mm/mempolicy.c
2589
struct folio *folio_alloc_noprof(gfp_t gfp, unsigned int order)
mm/mempolicy.c
2591
return page_rmappable_folio(alloc_pages_noprof(gfp | __GFP_COMP, order));
mm/mempolicy.c
2595
static unsigned long alloc_pages_bulk_interleave(gfp_t gfp,
mm/mempolicy.c
2612
nr_allocated = alloc_pages_bulk_noprof(gfp,
mm/mempolicy.c
2618
nr_allocated = alloc_pages_bulk_noprof(gfp,
mm/mempolicy.c
2630
static unsigned long alloc_pages_bulk_weighted_interleave(gfp_t gfp,
mm/mempolicy.c
2669
nr_allocated = __alloc_pages_bulk(gfp, node, NULL, node_pages,
mm/mempolicy.c
2734
nr_allocated = __alloc_pages_bulk(gfp, node, NULL, node_pages,
mm/mempolicy.c
2748
static unsigned long alloc_pages_bulk_preferred_many(gfp_t gfp, int nid,
mm/mempolicy.c
2755
preferred_gfp = gfp | __GFP_NOWARN;
mm/mempolicy.c
2762
nr_allocated += alloc_pages_bulk_noprof(gfp, numa_node_id(), NULL,
mm/mempolicy.c
2774
unsigned long alloc_pages_bulk_mempolicy_noprof(gfp_t gfp,
mm/mempolicy.c
2781
if (!in_interrupt() && !(gfp & __GFP_THISNODE))
mm/mempolicy.c
2785
return alloc_pages_bulk_interleave(gfp, pol,
mm/mempolicy.c
2790
gfp, pol, nr_pages, page_array);
mm/mempolicy.c
2793
return alloc_pages_bulk_preferred_many(gfp,
mm/mempolicy.c
2797
nodemask = policy_nodemask(gfp, pol, NO_INTERLEAVE_INDEX, &nid);
mm/mempolicy.c
2798
return alloc_pages_bulk_noprof(gfp, nid, nodemask,
mm/mempolicy.c
608
static nodemask_t *policy_nodemask(gfp_t gfp, struct mempolicy *pol,
mm/migrate.c
2641
gfp_t gfp = __GFP_THISNODE;
mm/migrate.c
2644
gfp |= GFP_TRANSHUGE_LIGHT;
mm/migrate.c
2646
gfp |= GFP_HIGHUSER_MOVABLE | __GFP_NOMEMALLOC | __GFP_NORETRY |
mm/migrate.c
2648
gfp &= ~__GFP_RECLAIM;
mm/migrate.c
2650
return __folio_alloc_node(gfp, order, nid);
mm/migrate_device.c
800
gfp_t gfp = vma_thp_gfp_mask(vma);
mm/migrate_device.c
823
if (mem_cgroup_charge(folio, migrate->vma->vm_mm, gfp)) {
mm/page-writeback.c
620
int wb_domain_init(struct wb_domain *dom, gfp_t gfp)
mm/page-writeback.c
630
return fprop_global_init(&dom->completions, gfp);
mm/page_alloc.c
5065
unsigned long alloc_pages_bulk_noprof(gfp_t gfp, int preferred_nid,
mm/page_alloc.c
5096
if (memcg_kmem_online() && (gfp & __GFP_ACCOUNT))
mm/page_alloc.c
5116
gfp &= gfp_allowed_mask;
mm/page_alloc.c
5117
alloc_gfp = gfp;
mm/page_alloc.c
5118
if (!prepare_alloc_pages(gfp, 0, preferred_nid, nodemask, &ac, &alloc_gfp, &alloc_flags))
mm/page_alloc.c
5120
gfp = alloc_gfp;
mm/page_alloc.c
5128
!__cpuset_zone_allowed(zone, gfp)) {
mm/page_alloc.c
5142
alloc_flags, gfp)) {
mm/page_alloc.c
5190
prep_new_page(page, 0, gfp, 0);
mm/page_alloc.c
5204
page = __alloc_pages_noprof(gfp, 0, preferred_nid, nodemask);
mm/page_alloc.c
5214
struct page *__alloc_frozen_pages_noprof(gfp_t gfp, unsigned int order,
mm/page_alloc.c
5226
if (WARN_ON_ONCE_GFP(order > MAX_PAGE_ORDER, gfp))
mm/page_alloc.c
5229
gfp &= gfp_allowed_mask;
mm/page_alloc.c
5237
gfp = current_gfp_context(gfp);
mm/page_alloc.c
5238
alloc_gfp = gfp;
mm/page_alloc.c
5239
if (!prepare_alloc_pages(gfp, order, preferred_nid, nodemask, &ac,
mm/page_alloc.c
5247
alloc_flags |= alloc_flags_nofragment(zonelist_zone(ac.preferred_zoneref), gfp);
mm/page_alloc.c
5254
alloc_gfp = gfp;
mm/page_alloc.c
5266
if (memcg_kmem_online() && (gfp & __GFP_ACCOUNT) && page &&
mm/page_alloc.c
5267
unlikely(__memcg_kmem_charge_page(page, gfp, order) != 0)) {
mm/page_alloc.c
5279
struct page *__alloc_pages_noprof(gfp_t gfp, unsigned int order,
mm/page_alloc.c
5284
page = __alloc_frozen_pages_noprof(gfp, order, preferred_nid, nodemask);
mm/page_alloc.c
5291
struct folio *__folio_alloc_noprof(gfp_t gfp, unsigned int order, int preferred_nid,
mm/page_alloc.c
5294
struct page *page = __alloc_pages_noprof(gfp | __GFP_COMP, order,
mm/page_frag_cache.c
54
gfp_t gfp = gfp_mask;
mm/page_frag_cache.c
63
page = __alloc_pages(gfp, 0, numa_mem_id(), NULL);
mm/percpu-km.c
42
int page_start, int page_end, gfp_t gfp)
mm/percpu-km.c
53
static struct pcpu_chunk *pcpu_create_chunk(gfp_t gfp)
mm/percpu-km.c
61
chunk = pcpu_alloc_chunk(gfp);
mm/percpu-km.c
65
pages = alloc_pages(gfp, order_base_2(nr_pages));
mm/percpu-vm.c
277
int page_start, int page_end, gfp_t gfp)
mm/percpu-vm.c
285
if (pcpu_alloc_pages(chunk, pages, page_start, page_end, gfp))
mm/percpu-vm.c
333
static struct pcpu_chunk *pcpu_create_chunk(gfp_t gfp)
mm/percpu-vm.c
338
chunk = pcpu_alloc_chunk(gfp);
mm/percpu-vm.c
84
gfp_t gfp)
mm/percpu-vm.c
89
gfp |= __GFP_HIGHMEM;
mm/percpu-vm.c
95
*pagep = alloc_pages_node(cpu_to_node(cpu), gfp, 0);
mm/percpu.c
1431
static struct pcpu_chunk *pcpu_alloc_chunk(gfp_t gfp)
mm/percpu.c
1436
chunk = pcpu_mem_zalloc(pcpu_chunk_struct_size, gfp);
mm/percpu.c
1445
sizeof(chunk->alloc_map[0]), gfp);
mm/percpu.c
1450
sizeof(chunk->bound_map[0]), gfp);
mm/percpu.c
1455
sizeof(chunk->md_blocks[0]), gfp);
mm/percpu.c
1463
sizeof(struct pcpuobj_ext), gfp);
mm/percpu.c
1568
int page_start, int page_end, gfp_t gfp);
mm/percpu.c
1573
static struct pcpu_chunk *pcpu_create_chunk(gfp_t gfp);
mm/percpu.c
1616
static bool pcpu_memcg_pre_alloc_hook(size_t size, gfp_t gfp,
mm/percpu.c
1621
if (!memcg_kmem_online() || !(gfp & __GFP_ACCOUNT))
mm/percpu.c
1628
if (obj_cgroup_charge(objcg, gfp, pcpu_obj_full_size(size)))
mm/percpu.c
1679
pcpu_memcg_pre_alloc_hook(size_t size, gfp_t gfp, struct obj_cgroup **objcgp)
mm/percpu.c
1737
gfp_t gfp)
mm/percpu.c
1751
gfp = current_gfp_context(gfp);
mm/percpu.c
1753
pcpu_gfp = gfp & (GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN);
mm/percpu.c
1754
is_atomic = !gfpflags_allow_blocking(gfp);
mm/percpu.c
1755
do_warn = !(gfp & __GFP_NOWARN);
mm/percpu.c
1777
if (unlikely(!pcpu_memcg_pre_alloc_hook(size, gfp, &objcg)))
mm/percpu.c
1786
if (gfp & __GFP_NOFAIL) {
mm/percpu.c
1896
kmemleak_alloc_percpu(ptr, size, gfp);
mm/percpu.c
1900
pcpu_obj_full_size(size), gfp);
mm/percpu.c
2008
const gfp_t gfp = GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN;
mm/percpu.c
2055
ret = pcpu_populate_chunk(chunk, rs, rs + nr, gfp);
mm/percpu.c
2073
chunk = pcpu_create_chunk(gfp);
mm/percpu.c
504
static void *pcpu_mem_zalloc(size_t size, gfp_t gfp)
mm/percpu.c
510
return kzalloc(size, gfp);
mm/percpu.c
512
return __vmalloc(size, gfp | __GFP_ZERO);
mm/readahead.c
446
pgoff_t mark, unsigned int order, gfp_t gfp)
mm/readahead.c
449
struct folio *folio = ractl_alloc_folio(ractl, gfp, order);
mm/readahead.c
456
err = filemap_add_folio(ractl->mapping, folio, index, gfp);
mm/readahead.c
478
gfp_t gfp = readahead_gfp_mask(mapping);
mm/readahead.c
515
err = ra_alloc_folio(ractl, index, mark, order, gfp);
mm/rmap.c
140
static inline struct anon_vma_chain *anon_vma_chain_alloc(gfp_t gfp)
mm/rmap.c
142
return kmem_cache_alloc(anon_vma_chain_cachep, gfp);
mm/secretmem.c
55
gfp_t gfp = vmf->gfp_mask;
mm/secretmem.c
69
folio = folio_alloc(gfp | __GFP_ZERO, 0);
mm/secretmem.c
83
err = filemap_add_folio(mapping, folio, offset, gfp);
mm/shmem.c
161
struct folio **foliop, enum sgp_type sgp, gfp_t gfp,
mm/shmem.c
1777
static struct folio *shmem_swapin_cluster(swp_entry_t swap, gfp_t gfp,
mm/shmem.c
1785
folio = swap_cluster_readahead(swap, gfp, mpol, ilx);
mm/shmem.c
1922
static struct folio *shmem_alloc_folio(gfp_t gfp, int order,
mm/shmem.c
1930
folio = folio_alloc_mpol(gfp, order, mpol, ilx, numa_node_id());
mm/shmem.c
1937
gfp_t gfp, struct inode *inode, pgoff_t index,
mm/shmem.c
1959
folio = shmem_alloc_folio(gfp, order, info, aligned_index);
mm/shmem.c
1972
folio = shmem_alloc_folio(gfp, 0, info, index);
mm/shmem.c
1981
gfp &= GFP_RECLAIM_MASK;
mm/shmem.c
1982
error = mem_cgroup_charge(folio, fault_mm, gfp);
mm/shmem.c
1998
error = shmem_add_to_page_cache(folio, mapping, index, NULL, gfp);
mm/shmem.c
2042
swp_entry_t entry, int order, gfp_t gfp)
mm/shmem.c
2053
gfp &= ~GFP_CONSTRAINT_MASK;
mm/shmem.c
2054
alloc_gfp = gfp;
mm/shmem.c
2071
alloc_gfp = limit_gfp_mask(vma_thp_gfp_mask(vma), gfp);
mm/shmem.c
2105
alloc_gfp = gfp;
mm/shmem.c
2123
static bool shmem_should_replace_folio(struct folio *folio, gfp_t gfp)
mm/shmem.c
2125
return folio_zonenum(folio) > gfp_zone(gfp);
mm/shmem.c
2128
static int shmem_replace_folio(struct folio **foliop, gfp_t gfp,
mm/shmem.c
2142
gfp &= ~GFP_CONSTRAINT_MASK;
mm/shmem.c
2147
gfp = limit_gfp_mask(huge_gfp, gfp);
mm/shmem.c
2151
new = shmem_alloc_folio(gfp, folio_order(old), info, index);
mm/shmem.c
2216
swp_entry_t swap, gfp_t gfp)
mm/shmem.c
2224
gfp &= GFP_RECLAIM_MASK;
mm/shmem.c
2277
if (!xas_nomem(&xas, gfp))
mm/shmem.c
2295
gfp_t gfp, struct vm_area_struct *vma,
mm/shmem.c
2341
index_entry, order, gfp);
mm/shmem.c
2349
folio = shmem_swapin_cluster(swap, gfp, info, index);
mm/shmem.c
2373
error = shmem_split_large_entry(inode, index, index_entry, gfp);
mm/shmem.c
2419
if (shmem_should_replace_folio(folio, gfp)) {
mm/shmem.c
2420
error = shmem_replace_folio(&folio, gfp, info, index, vma);
mm/shmem.c
2426
swp_to_radix_entry(swap), gfp);
mm/shmem.c
2469
gfp_t gfp, struct vm_fault *vmf, vm_fault_t *fault_type)
mm/shmem.c
2501
sgp, gfp, vma, fault_type);
mm/shmem.c
2554
huge_gfp = limit_gfp_mask(huge_gfp, gfp);
mm/shmem.c
2567
folio = shmem_alloc_and_add_folio(vmf, gfp, inode, index, fault_mm, 0);
mm/shmem.c
2752
gfp_t gfp = mapping_gfp_mask(inode->i_mapping);
mm/shmem.c
2769
gfp, vmf, &ret);
mm/shmem.c
3194
gfp_t gfp = mapping_gfp_mask(mapping);
mm/shmem.c
3216
folio = shmem_alloc_folio(gfp, 0, info, pgoff);
mm/shmem.c
3273
ret = mem_cgroup_charge(folio, dst_vma->vm_mm, gfp);
mm/shmem.c
3276
ret = shmem_add_to_page_cache(folio, mapping, pgoff, NULL, gfp);
mm/shmem.c
5982
pgoff_t index, gfp_t gfp)
mm/shmem.c
5990
&folio, SGP_CACHE, gfp, NULL, NULL);
mm/shmem.c
6000
return mapping_read_folio_gfp(mapping, index, gfp);
mm/shmem.c
6006
pgoff_t index, gfp_t gfp)
mm/shmem.c
6008
struct folio *folio = shmem_read_folio_gfp(mapping, index, gfp);
mm/shmem.c
883
pgoff_t index, void *expected, gfp_t gfp)
mm/shmem.c
898
gfp &= GFP_RECLAIM_MASK;
mm/shmem.c
899
folio_throttle_swaprate(folio, gfp);
mm/shmem.c
929
} while (xas_nomem(&xas, gfp));
mm/slab.h
605
gfp_t gfp, bool new_slab);
mm/slab.h
664
gfp_t gfp);
mm/slab.h
668
unsigned int count, gfp_t gfp)
mm/slab_common.c
1058
gfp_t gfp)
mm/slab_common.c
1064
cachep->random_seq = kcalloc(count, sizeof(unsigned int), gfp);
mm/slub.c
2117
struct slab *slab, gfp_t gfp)
mm/slub.c
2128
obj_exts_cache = kmalloc_slab(sz, NULL, gfp, 0);
mm/slub.c
2142
gfp_t gfp, bool new_slab)
mm/slub.c
2144
bool allow_spin = gfpflags_allow_spinning(gfp);
mm/slub.c
2151
gfp &= ~OBJCGS_CLEAR_MASK;
mm/slub.c
2153
gfp |= __GFP_NO_OBJ_EXT;
mm/slub.c
2155
sz = obj_exts_alloc_size(s, slab, gfp);
mm/slub.c
2167
vec = kmalloc_node(sz, gfp | __GFP_ZERO, slab_nid(slab));
mm/slub.c
2318
gfp_t gfp, bool new_slab)
mm/slub.c
2744
static struct slab_sheaf *__alloc_empty_sheaf(struct kmem_cache *s, gfp_t gfp,
mm/slub.c
2750
if (gfp & __GFP_NO_OBJ_EXT)
mm/slub.c
2753
gfp &= ~OBJCGS_CLEAR_MASK;
mm/slub.c
2761
gfp |= __GFP_NO_OBJ_EXT;
mm/slub.c
2764
sheaf = kzalloc(sheaf_size, gfp);
mm/slub.c
2777
gfp_t gfp)
mm/slub.c
2779
return __alloc_empty_sheaf(s, gfp, s->sheaf_capacity);
mm/slub.c
2800
refill_objects(struct kmem_cache *s, void **p, gfp_t gfp, unsigned int min,
mm/slub.c
2804
gfp_t gfp)
mm/slub.c
2812
filled = refill_objects(s, &sheaf->objects[sheaf->size], gfp, to_fill,
mm/slub.c
2827
static struct slab_sheaf *alloc_full_sheaf(struct kmem_cache *s, gfp_t gfp)
mm/slub.c
2829
struct slab_sheaf *sheaf = alloc_empty_sheaf(s, gfp);
mm/slub.c
2834
if (refill_sheaf(s, sheaf, gfp | __GFP_NOMEMALLOC | __GFP_NOWARN)) {
mm/slub.c
3430
struct kmem_cache *s, gfp_t gfp)
mm/slub.c
3435
alloc_slab_obj_exts(slab, s, gfp, true);
mm/slub.c
4561
__pcs_replace_empty_main(struct kmem_cache *s, struct slub_percpu_sheaves *pcs, gfp_t gfp)
mm/slub.c
4587
allow_spin = gfpflags_allow_spinning(gfp);
mm/slub.c
4615
if (!refill_sheaf(s, empty, gfp | __GFP_NOMEMALLOC | __GFP_NOWARN)) {
mm/slub.c
4626
full = alloc_full_sheaf(s, gfp);
mm/slub.c
4672
void *alloc_from_pcs(struct kmem_cache *s, gfp_t gfp, int node)
mm/slub.c
4717
pcs = __pcs_replace_empty_main(s, pcs, gfp);
mm/slub.c
4747
unsigned int alloc_from_pcs_bulk(struct kmem_cache *s, gfp_t gfp, size_t size,
mm/slub.c
4783
gfpflags_allow_spinning(gfp));
mm/slub.c
4927
struct slab_sheaf *sheaf, gfp_t gfp)
mm/slub.c
4932
gfp_nomemalloc = gfp | __GFP_NOMEMALLOC;
mm/slub.c
4933
if (gfp_pfmemalloc_allowed(gfp))
mm/slub.c
4938
if (likely(!ret || !gfp_pfmemalloc_allowed(gfp)))
mm/slub.c
4945
ret = refill_sheaf(s, sheaf, gfp);
mm/slub.c
4961
kmem_cache_prefill_sheaf(struct kmem_cache *s, gfp_t gfp, unsigned int size)
mm/slub.c
4972
sheaf = kzalloc_flex(*sheaf, objects, size, gfp);
mm/slub.c
4984
if (!__kmem_cache_alloc_bulk(s, gfp, size,
mm/slub.c
5018
sheaf = alloc_empty_sheaf(s, gfp);
mm/slub.c
5025
__prefill_sheaf_pfmemalloc(s, sheaf, gfp)) {
mm/slub.c
5045
void kmem_cache_return_sheaf(struct kmem_cache *s, gfp_t gfp,
mm/slub.c
5080
refill_sheaf(s, sheaf, gfp)) {
mm/slub.c
5100
int kmem_cache_refill_sheaf(struct kmem_cache *s, gfp_t gfp,
mm/slub.c
5118
return __prefill_sheaf_pfmemalloc(s, sheaf, gfp);
mm/slub.c
5120
if (!__kmem_cache_alloc_bulk(s, gfp, sheaf->capacity - sheaf->size,
mm/slub.c
5134
sheaf = kmem_cache_prefill_sheaf(s, gfp, size);
mm/slub.c
5138
kmem_cache_return_sheaf(s, gfp, *sheafp);
mm/slub.c
5156
kmem_cache_alloc_from_sheaf_noprof(struct kmem_cache *s, gfp_t gfp,
mm/slub.c
5165
ret = kfence_alloc(s, s->object_size, gfp);
mm/slub.c
5170
init = slab_want_init_on_alloc(gfp, s);
mm/slub.c
5173
slab_post_alloc_hook(s, NULL, gfp | __GFP_NOFAIL, 1, &ret, init, s->object_size);
mm/slub.c
5175
trace_kmem_cache_alloc(_RET_IP_, ret, s, gfp, NUMA_NO_NODE);
mm/slub.c
7013
__refill_objects_node(struct kmem_cache *s, void **p, gfp_t gfp, unsigned int min,
mm/slub.c
7023
pc.flags = gfp;
mm/slub.c
7093
__refill_objects_any(struct kmem_cache *s, void **p, gfp_t gfp, unsigned int min,
mm/slub.c
7099
enum zone_type highest_zoneidx = gfp_zone(gfp);
mm/slub.c
7110
zonelist = node_zonelist(mempolicy_slab_node(), gfp);
mm/slub.c
7117
if (!n || !cpuset_zone_allowed(zone, gfp) ||
mm/slub.c
7121
r = __refill_objects_node(s, p, gfp, min, max, n,
mm/slub.c
7144
__refill_objects_any(struct kmem_cache *s, void **p, gfp_t gfp, unsigned int min,
mm/slub.c
7152
refill_objects(struct kmem_cache *s, void **p, gfp_t gfp, unsigned int min,
mm/slub.c
7159
if (WARN_ON_ONCE(!gfpflags_allow_spinning(gfp)))
mm/slub.c
7162
refilled = __refill_objects_node(s, p, gfp, min, max,
mm/slub.c
7168
refilled += __refill_objects_any(s, p + refilled, gfp, min - refilled,
mm/slub.c
7175
slab = new_slab(s, gfp, local_node);
mm/swap_state.c
489
gfp_t gfp, bool charged)
mm/swap_state.c
498
if (!charged && mem_cgroup_swapin_charge_folio(folio, NULL, gfp, entry))
mm/swapfile.c
3914
void __folio_throttle_swaprate(struct folio *folio, gfp_t gfp)
mm/swapfile.c
3918
if (!(gfp & __GFP_IO))
mm/swapfile.c
417
static struct swap_table *swap_table_alloc(gfp_t gfp)
mm/swapfile.c
422
return kmem_cache_zalloc(swap_table_cachep, gfp);
mm/swapfile.c
424
folio = folio_alloc(gfp | __GFP_ZERO, 0);
mm/truncate.c
623
gfp_t gfp)
mm/truncate.c
638
if (!filemap_release_folio(folio, gfp))
mm/util.c
104
return kstrdup(s, gfp);
mm/util.c
118
char *kstrndup(const char *s, size_t max, gfp_t gfp)
mm/util.c
120
return s ? __kmemdup_nul(s, strnlen(s, max), gfp) : NULL;
mm/util.c
134
void *kmemdup_noprof(const void *src, size_t len, gfp_t gfp)
mm/util.c
138
p = kmalloc_node_track_caller_noprof(len, gfp, NUMA_NO_NODE, _RET_IP_);
mm/util.c
156
void *kmemdup_array(const void *src, size_t count, size_t element_size, gfp_t gfp)
mm/util.c
158
return kmemdup(src, size_mul(element_size, count), gfp);
mm/util.c
172
void *kvmemdup(const void *src, size_t len, gfp_t gfp)
mm/util.c
176
p = kvmalloc(len, gfp);
mm/util.c
192
char *kmemdup_nul(const char *s, size_t len, gfp_t gfp)
mm/util.c
194
return s ? __kmemdup_nul(s, len, gfp) : NULL;
mm/util.c
59
static __always_inline char *__kmemdup_nul(const char *s, size_t len, gfp_t gfp)
mm/util.c
64
buf = kmalloc_track_caller(len + 1, gfp);
mm/util.c
82
char *kstrdup(const char *s, gfp_t gfp)
mm/util.c
84
return s ? __kmemdup_nul(s, strlen(s), gfp) : NULL;
mm/util.c
99
const char *kstrdup_const(const char *s, gfp_t gfp)
mm/vma.h
278
struct vm_area_struct *vma, gfp_t gfp)
mm/vma.h
286
mas_store_gfp(&vmi->mas, vma, gfp);
mm/vmalloc.c
3641
vm_area_alloc_pages(gfp_t gfp, int nid,
mm/vmalloc.c
3650
gfp_t large_gfp = vmalloc_gfp_adjust(gfp, large_order) & ~__GFP_DIRECT_RECLAIM;
mm/vmalloc.c
3706
nr = alloc_pages_bulk_mempolicy_noprof(gfp,
mm/vmalloc.c
3710
nr = alloc_pages_bulk_node_noprof(gfp, nid,
mm/vmalloc.c
3727
if (!(gfp & __GFP_NOFAIL) && fatal_signal_pending(current))
mm/vmalloc.c
3731
page = alloc_pages_noprof(gfp, order);
mm/vmalloc.c
3733
page = alloc_pages_node_noprof(nid, gfp, order);
mm/vmpressure.c
239
void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree,
mm/vmpressure.c
268
if (!(gfp & (__GFP_HIGHMEM | __GFP_MOVABLE | __GFP_IO | __GFP_FS)))
mm/vmpressure.c
335
void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio)
mm/vmpressure.c
351
vmpressure(gfp, memcg, true, vmpressure_win, 0);
mm/zsmalloc.c
1297
unsigned long zs_malloc(struct zs_pool *pool, size_t size, gfp_t gfp,
mm/zsmalloc.c
1311
handle = cache_alloc_handle(gfp);
mm/zsmalloc.c
1333
zspage = alloc_zspage(pool, class, gfp, nid);
mm/zsmalloc.c
237
static inline struct zpdesc *alloc_zpdesc(gfp_t gfp, const int nid)
mm/zsmalloc.c
239
struct page *page = alloc_pages_node(nid, gfp, 0);
mm/zsmalloc.c
375
static unsigned long cache_alloc_handle(gfp_t gfp)
mm/zsmalloc.c
377
gfp = gfp & ~(__GFP_HIGHMEM | __GFP_MOVABLE);
mm/zsmalloc.c
379
return (unsigned long)kmem_cache_alloc(handle_cachep, gfp);
mm/zsmalloc.c
387
static struct zspage *cache_alloc_zspage(gfp_t gfp)
mm/zsmalloc.c
389
gfp = gfp & ~(__GFP_HIGHMEM | __GFP_MOVABLE);
mm/zsmalloc.c
391
return kmem_cache_zalloc(zspage_cachep, gfp);
mm/zsmalloc.c
934
gfp_t gfp, const int nid)
mm/zsmalloc.c
938
struct zspage *zspage = cache_alloc_zspage(gfp);
mm/zsmalloc.c
944
gfp &= ~__GFP_MOVABLE;
mm/zsmalloc.c
954
zpdesc = alloc_zpdesc(gfp, nid);
mm/zswap.c
699
static struct zswap_entry *zswap_entry_cache_alloc(gfp_t gfp, int nid)
mm/zswap.c
702
entry = kmem_cache_alloc_node(zswap_entry_cache, gfp, nid);
mm/zswap.c
861
gfp_t gfp;
mm/zswap.c
907
gfp = GFP_NOWAIT | __GFP_NORETRY | __GFP_HIGHMEM | __GFP_MOVABLE;
mm/zswap.c
908
handle = zs_malloc(pool->zs_pool, dlen, gfp, page_to_nid(page));
net/bpf/test_run.c
245
gfp_t gfp = __GFP_ZERO | GFP_ATOMIC;
net/bpf/test_run.c
249
n = kmem_cache_alloc_bulk(net_hotdata.skbuff_cache, gfp, nframes,
net/ceph/buffer.c
12
struct ceph_buffer *ceph_buffer_new(size_t len, gfp_t gfp)
net/ceph/buffer.c
16
b = kmalloc_obj(*b, gfp);
net/ceph/buffer.c
20
b->vec.iov_base = kvmalloc(len, gfp);
net/ceph/mon_client.c
610
alloc_generic_request(struct ceph_mon_client *monc, gfp_t gfp)
net/ceph/mon_client.c
614
req = kzalloc_obj(*req, gfp);
net/ceph/osd_client.c
581
static int __ceph_osdc_alloc_messages(struct ceph_osd_request *req, gfp_t gfp,
net/ceph/osd_client.c
615
num_request_data_items, gfp, true);
net/ceph/osd_client.c
632
num_reply_data_items, gfp, true);
net/ceph/osd_client.c
702
int ceph_osdc_alloc_messages(struct ceph_osd_request *req, gfp_t gfp)
net/ceph/osd_client.c
707
return __ceph_osdc_alloc_messages(req, gfp, num_request_data_items,
net/ceph/osdmap.c
2180
int oid_aprintf_vargs(struct ceph_object_id *oid, gfp_t gfp,
net/ceph/osdmap.c
2193
external_name = kmalloc(len + 1, gfp);
net/ceph/osdmap.c
2208
int ceph_oid_aprintf(struct ceph_object_id *oid, gfp_t gfp,
net/ceph/osdmap.c
2215
ret = oid_aprintf_vargs(oid, gfp, fmt, ap);
net/core/dev.c
1021
netdevice_tracker *tracker, gfp_t gfp)
net/core/dev.c
1027
netdev_tracker_alloc(dev, tracker, gfp);
net/core/dev.c
936
netdevice_tracker *tracker, gfp_t gfp)
net/core/dev.c
942
netdev_tracker_alloc(dev, tracker, gfp);
net/core/devmem.c
452
netmem_ref mp_dmabuf_devmem_alloc_netmems(struct page_pool *pool, gfp_t gfp)
net/core/dst_cache.c
163
int dst_cache_init(struct dst_cache *dst_cache, gfp_t gfp)
net/core/dst_cache.c
168
gfp | __GFP_ZERO);
net/core/flow_offload.c
219
gfp_t gfp)
net/core/flow_offload.c
223
cookie = kmalloc(sizeof(*cookie) + len, gfp);
net/core/hwbm.c
23
int hwbm_pool_refill(struct hwbm_pool *bm_pool, gfp_t gfp)
net/core/hwbm.c
31
buf = kmalloc(frag_size, gfp);
net/core/mp_dmabuf_devmem.h
16
netmem_ref mp_dmabuf_devmem_alloc_netmems(struct page_pool *pool, gfp_t gfp);
net/core/mp_dmabuf_devmem.h
28
mp_dmabuf_devmem_alloc_netmems(struct page_pool *pool, gfp_t gfp)
net/core/net_namespace.c
1182
struct nlmsghdr *nlh, gfp_t gfp)
net/core/net_namespace.c
1193
msg = nlmsg_new(rtnl_net_get_size(), gfp);
net/core/net_namespace.c
1201
rtnl_notify(msg, net, portid, RTNLGRP_NSID, nlh, gfp);
net/core/net_namespace.c
310
struct nlmsghdr *nlh, gfp_t gfp);
net/core/net_namespace.c
314
int peernet2id_alloc(struct net *net, struct net *peer, gfp_t gfp)
net/core/net_namespace.c
345
rtnl_net_notifyid(net, RTM_NEWNSID, id, 0, NULL, gfp);
net/core/page_pool.c
1055
gfp_t gfp)
net/core/page_pool.c
1076
netmem = page_pool_alloc_netmems(pool, gfp);
net/core/page_pool.c
1099
unsigned int size, gfp_t gfp)
net/core/page_pool.c
1102
gfp));
net/core/page_pool.c
476
netmem_ref netmem, gfp_t gfp)
net/core/page_pool.c
486
PP_DMA_INDEX_LIMIT, gfp);
net/core/page_pool.c
489
PP_DMA_INDEX_LIMIT, gfp);
net/core/page_pool.c
525
static bool page_pool_dma_map(struct page_pool *pool, netmem_ref netmem, gfp_t gfp)
net/core/page_pool.c
547
err = page_pool_register_dma_index(pool, netmem, gfp);
net/core/page_pool.c
565
gfp_t gfp)
net/core/page_pool.c
569
gfp |= __GFP_COMP;
net/core/page_pool.c
570
page = alloc_pages_node(pool->p.nid, gfp, pool->p.order);
net/core/page_pool.c
574
if (pool->dma_map && unlikely(!page_pool_dma_map(pool, page_to_netmem(page), gfp))) {
net/core/page_pool.c
591
gfp_t gfp)
net/core/page_pool.c
602
if ((gfp & GFP_ATOMIC) == GFP_ATOMIC)
net/core/page_pool.c
603
gfp |= __GFP_NOWARN;
net/core/page_pool.c
607
return page_to_netmem(__page_pool_alloc_page_order(pool, gfp));
net/core/page_pool.c
616
nr_pages = alloc_pages_bulk_node(gfp, pool->p.nid, bulk,
net/core/page_pool.c
626
if (dma_map && unlikely(!page_pool_dma_map(pool, netmem, gfp))) {
net/core/page_pool.c
654
netmem_ref page_pool_alloc_netmems(struct page_pool *pool, gfp_t gfp)
net/core/page_pool.c
665
netmem = pool->mp_ops->alloc_netmems(pool, gfp);
net/core/page_pool.c
667
netmem = __page_pool_alloc_netmems_slow(pool, gfp);
net/core/page_pool.c
673
struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp)
net/core/page_pool.c
675
return netmem_to_page(page_pool_alloc_netmems(pool, gfp));
net/core/rtnetlink.c
1876
struct net *src_net, gfp_t gfp)
net/core/rtnetlink.c
1884
int id = peernet2id_alloc(src_net, link_net, gfp);
net/core/rtnetlink.c
2054
int tgt_netnsid, gfp_t gfp)
net/core/skbuff.c
6859
static int pskb_carve(struct sk_buff *skb, const u32 off, gfp_t gfp);
net/core/skbuff.c
6994
static int pskb_carve(struct sk_buff *skb, const u32 len, gfp_t gfp)
net/core/skbuff.c
6999
return pskb_carve_inside_header(skb, len, headlen, gfp);
net/core/skbuff.c
7001
return pskb_carve_inside_nonlinear(skb, len, headlen, gfp);
net/core/skbuff.c
7008
int to_copy, gfp_t gfp)
net/core/skbuff.c
7010
struct sk_buff *clone = skb_clone(skb, gfp);
net/core/skbuff.c
7015
if (pskb_carve(clone, off, gfp) < 0 ||
net/core/skmsg.c
521
static struct sk_msg *alloc_sk_msg(gfp_t gfp)
net/core/skmsg.c
525
msg = kzalloc_obj(*msg, gfp | __GFP_NOWARN);
net/core/sock.c
3130
bool skb_page_frag_refill(unsigned int sz, struct page_frag *pfrag, gfp_t gfp)
net/core/sock.c
3146
pfrag->page = alloc_pages((gfp & ~__GFP_DIRECT_RECLAIM) |
net/core/sock.c
3155
pfrag->page = alloc_page(gfp);
net/core/sock.c
3451
void __sk_charge(struct sock *sk, gfp_t gfp)
net/core/sock.c
3455
gfp |= __GFP_NOFAIL;
net/core/sock.c
3463
mem_cgroup_sk_charge(sk, amt, gfp);
net/core/sock.c
3466
kmem_cache_charge(sk, gfp);
net/core/xdp.c
238
static int __mem_id_cyclic_get(gfp_t gfp)
net/core/xdp.c
244
id = ida_alloc_range(&mem_id_pool, mem_id_next, MEM_ID_MAX - 1, gfp);
net/core/xdp.c
276
gfp_t gfp = GFP_KERNEL;
net/core/xdp.c
300
xdp_alloc = kzalloc_obj(*xdp_alloc, gfp);
net/core/xdp.c
305
id = __mem_id_cyclic_get(gfp);
net/handshake/handshake-test.c
105
.gfp = GFP_KERNEL | __GFP_NOWARN,
net/handshake/handshake-test.c
111
.gfp = GFP_KERNEL,
net/handshake/handshake-test.c
135
result = handshake_req_alloc(param->proto, param->gfp);
net/handshake/handshake-test.c
36
gfp_t gfp;
net/handshake/handshake-test.c
75
.gfp = GFP_KERNEL,
net/handshake/handshake-test.c
81
.gfp = GFP_KERNEL,
net/handshake/handshake-test.c
87
.gfp = GFP_KERNEL,
net/handshake/handshake-test.c
93
.gfp = GFP_KERNEL,
net/handshake/handshake-test.c
99
.gfp = GFP_KERNEL,
net/ipv4/igmp.c
1179
gfp_t gfp)
net/ipv4/igmp.c
1190
pmc = kzalloc_obj(*pmc, gfp);
net/ipv4/igmp.c
1294
static void __igmp_group_dropped(struct ip_mc_list *im, gfp_t gfp)
net/ipv4/igmp.c
1326
igmpv3_add_delrec(in_dev, im, gfp);
net/ipv4/igmp.c
1505
unsigned int mode, gfp_t gfp)
net/ipv4/igmp.c
1535
im = kzalloc_obj(*im, gfp);
net/ipv4/igmp.c
1571
void __ip_mc_inc_group(struct in_device *in_dev, __be32 addr, gfp_t gfp)
net/ipv4/igmp.c
1573
____ip_mc_inc_group(in_dev, addr, MCAST_EXCLUDE, gfp);
net/ipv4/igmp.c
160
gfp_t gfp);
net/ipv4/igmp.c
1762
void __ip_mc_dec_group(struct in_device *in_dev, __be32 addr, gfp_t gfp)
net/ipv4/igmp.c
1777
__igmp_group_dropped(i, gfp);
net/ipv4/tcp.c
940
struct sk_buff *tcp_stream_alloc_skb(struct sock *sk, gfp_t gfp,
net/ipv4/tcp.c
945
skb = alloc_skb_fclone(MAX_TCP_HEADER, gfp);
net/ipv4/tcp_ipv4.c
1351
static int tcp_md5sig_info_add(struct sock *sk, gfp_t gfp)
net/ipv4/tcp_ipv4.c
1356
md5sig = kmalloc_obj(*md5sig, gfp);
net/ipv4/tcp_ipv4.c
1369
const u8 *newkey, u8 newkeylen, gfp_t gfp)
net/ipv4/tcp_ipv4.c
1399
key = sock_kmalloc(sk, sizeof(*key), gfp | __GFP_ZERO);
net/ipv4/tcp_output.c
1823
unsigned int mss_now, gfp_t gfp)
net/ipv4/tcp_output.c
1851
if (skb_unclone_keeptruesize(skb, gfp))
net/ipv4/tcp_output.c
1855
buff = tcp_stream_alloc_skb(sk, gfp, true);
net/ipv4/tcp_output.c
2380
unsigned int mss_now, gfp_t gfp)
net/ipv4/tcp_output.c
2389
buff = tcp_stream_alloc_skb(sk, gfp, true);
net/ipv4/tcp_output.c
2967
int push_one, gfp_t gfp)
net/ipv4/tcp_output.c
3050
unlikely(tso_fragment(sk, skb, limit, mss_now, gfp)))
net/ipv4/tcp_output.c
3064
if (unlikely(tcp_transmit_skb(sk, skb, 1, gfp)))
net/ipv4/tcp_output.c
67
int push_one, gfp_t gfp);
net/ipv6/ioam6.c
635
void ioam6_event(enum ioam6_event_type type, struct net *net, gfp_t gfp,
net/ipv6/ioam6.c
645
skb = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/ipv6/ioam6.c
666
IOAM6_GENL_EV_GRP_OFFSET, gfp);
net/ipv6/ip6_output.c
1306
gfp_t gfp)
net/ipv6/ip6_output.c
1308
return src ? kmemdup(src, (src->hdrlen + 1) * 8, gfp) : NULL;
net/ipv6/ip6_output.c
1312
gfp_t gfp)
net/ipv6/ip6_output.c
1314
return src ? kmemdup(src, (src->hdrlen + 1) * 8, gfp) : NULL;
net/key/af_key.c
464
gfp_t gfp)
net/key/af_key.c
469
uctx = kmalloc((sizeof(*uctx)+ctx_size), gfp);
net/mac80211/cfg.c
4496
u64 *cookie, gfp_t gfp)
net/mac80211/cfg.c
4502
ack_skb = skb_copy(skb, gfp);
net/mac80211/cfg.c
4855
gfp_t gfp)
net/mac80211/cfg.c
4880
reason, cookie, gfp);
net/mac80211/cfg.c
4886
gfp_t gfp)
net/mac80211/cfg.c
4905
cfg80211_nan_match(ieee80211_vif_to_wdev(vif), match, gfp);
net/mac80211/ieee80211_i.h
1943
u64 *cookie, gfp_t gfp);
net/mac80211/key.c
1244
const u8 *replay_ctr, gfp_t gfp)
net/mac80211/key.c
1250
cfg80211_gtk_rekey_notify(sdata->dev, bssid, replay_ctr, gfp);
net/mac80211/mlme.c
10284
gfp_t gfp)
net/mac80211/mlme.c
10290
cfg80211_cqm_rssi_notify(sdata->dev, rssi_event, rssi_level, gfp);
net/mac80211/mlme.c
10294
void ieee80211_cqm_beacon_loss_notify(struct ieee80211_vif *vif, gfp_t gfp)
net/mac80211/mlme.c
10300
cfg80211_cqm_beacon_loss_notify(sdata->dev, gfp);
net/mac80211/pm.c
194
gfp_t gfp)
net/mac80211/pm.c
198
cfg80211_report_wowlan_wakeup(&sdata->wdev, wakeup, gfp);
net/mac80211/rate.h
40
struct sta_info *sta, gfp_t gfp)
net/mac80211/rate.h
43
return ref->ops->alloc_sta(ref->priv, &sta->sta, gfp);
net/mac80211/rc80211_minstrel_ht.c
1850
minstrel_ht_alloc_sta(void *priv, struct ieee80211_sta *sta, gfp_t gfp)
net/mac80211/rc80211_minstrel_ht.c
1865
return kzalloc_obj(*mi, gfp);
net/mac80211/sta_info.c
535
struct sta_info *sta, gfp_t gfp)
net/mac80211/sta_info.c
542
sta, gfp);
net/mac80211/sta_info.c
551
gfp_t gfp)
net/mac80211/sta_info.c
558
alloc_percpu_gfp(struct ieee80211_sta_rx_stats, gfp);
net/mac80211/sta_info.c
612
gfp_t gfp)
net/mac80211/sta_info.c
621
sta = kzalloc(sizeof(*sta) + hw->sta_data_size, gfp);
net/mac80211/sta_info.c
628
if (sta_info_alloc_link(local, &sta->deflink, gfp))
net/mac80211/sta_info.c
647
sta->mesh = kzalloc_obj(*sta->mesh, gfp);
net/mac80211/sta_info.c
693
txq_data = kcalloc(ARRAY_SIZE(sta->sta.txq), size, gfp);
net/mac80211/sta_info.c
704
if (sta_prepare_rate_control(local, sta, gfp))
net/mac80211/sta_info.c
781
const u8 *addr, gfp_t gfp)
net/mac80211/sta_info.c
783
return __sta_info_alloc(sdata, addr, -1, addr, gfp);
net/mac80211/sta_info.c
790
gfp_t gfp)
net/mac80211/sta_info.c
792
return __sta_info_alloc(sdata, mld_addr, link_id, link_addr, gfp);
net/mac80211/sta_info.h
911
const u8 *addr, gfp_t gfp);
net/mac80211/sta_info.h
916
gfp_t gfp);
net/mac80211/tdls.c
1504
u16 reason_code, gfp_t gfp)
net/mac80211/tdls.c
1514
cfg80211_tdls_oper_request(sdata->dev, peer, oper, reason_code, gfp);
net/mctp/route.c
226
u8 tag, gfp_t gfp)
net/mctp/route.c
230
key = kzalloc_obj(*key, gfp);
net/mptcp/pm_netlink.c
285
static void mptcp_nl_mcast_send(struct net *net, struct sk_buff *nlskb, gfp_t gfp)
net/mptcp/pm_netlink.c
288
nlskb, 0, MPTCP_PM_EV_GRP_OFFSET, gfp);
net/mptcp/pm_netlink.c
573
const struct sock *ssk, gfp_t gfp)
net/mptcp/pm_netlink.c
582
skb = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/mptcp/pm_netlink.c
623
mptcp_nl_mcast_send(net, skb, gfp);
net/mptcp/protocol.c
1184
static bool __mptcp_add_ext(struct sk_buff *skb, gfp_t gfp)
net/mptcp/protocol.c
1186
struct skb_ext *mpext = __skb_ext_alloc(gfp);
net/mptcp/protocol.c
1194
static struct sk_buff *__mptcp_do_alloc_tx_skb(struct sock *sk, gfp_t gfp)
net/mptcp/protocol.c
1198
skb = alloc_skb_fclone(MAX_TCP_HEADER, gfp);
net/mptcp/protocol.c
1200
if (likely(__mptcp_add_ext(skb, gfp))) {
net/mptcp/protocol.c
1213
static struct sk_buff *__mptcp_alloc_tx_skb(struct sock *sk, struct sock *ssk, gfp_t gfp)
net/mptcp/protocol.c
1217
skb = __mptcp_do_alloc_tx_skb(sk, gfp);
net/mptcp/protocol.c
1232
gfp_t gfp = data_lock_held ? GFP_ATOMIC : sk->sk_allocation;
net/mptcp/protocol.c
1234
return __mptcp_alloc_tx_skb(sk, ssk, gfp);
net/mptcp/protocol.c
4162
gfp_t gfp = GFP_KERNEL | __GFP_NOFAIL;
net/mptcp/protocol.c
4178
mem_cgroup_sk_charge(sk, amt, gfp);
net/mptcp/protocol.h
1128
const struct sock *ssk, gfp_t gfp);
net/mptcp/protocol.h
753
void __mptcp_inherit_memcg(struct sock *sk, struct sock *ssk, gfp_t gfp);
net/mptcp/subflow.c
1723
void __mptcp_inherit_memcg(struct sock *sk, struct sock *ssk, gfp_t gfp)
net/mptcp/subflow.c
1730
__sk_charge(ssk, gfp);
net/netfilter/nf_conntrack_core.c
1662
gfp_t gfp, u32 hash)
net/netfilter/nf_conntrack_core.c
1689
ct = kmem_cache_alloc(nf_conntrack_cachep, gfp);
net/netfilter/nf_conntrack_core.c
1720
gfp_t gfp)
net/netfilter/nf_conntrack_core.c
1722
return __nf_conntrack_alloc(net, zone, orig, repl, gfp, 0);
net/netfilter/nf_conntrack_ecache.c
324
bool nf_ct_ecache_ext_add(struct nf_conn *ct, u16 ctmask, u16 expmask, gfp_t gfp)
net/netfilter/nf_conntrack_ecache.c
351
e = nf_ct_ext_add(ct, NF_CT_EXT_ECACHE, gfp);
net/netfilter/nf_conntrack_extend.c
117
new = krealloc(ct->ext, alloc, gfp);
net/netfilter/nf_conntrack_extend.c
92
void *nf_ct_ext_add(struct nf_conn *ct, enum nf_ct_ext_id id, gfp_t gfp)
net/netfilter/nf_conntrack_helper.c
178
nf_ct_helper_ext_add(struct nf_conn *ct, gfp_t gfp)
net/netfilter/nf_conntrack_helper.c
182
help = nf_ct_ext_add(ct, NF_CT_EXT_HELPER, gfp);
net/netfilter/nf_tables_api.c
10446
unsigned int gc_seq, gfp_t gfp)
net/netfilter/nf_tables_api.c
10451
trans = kzalloc_obj(*trans, gfp);
net/netfilter/nf_tables_api.c
10483
unsigned int gc_seq, gfp_t gfp)
net/netfilter/nf_tables_api.c
10493
return nft_trans_gc_alloc(set, gc_seq, gfp);
net/netfilter/nf_tables_api.c
10506
struct nft_trans_gc *nft_trans_gc_queue_sync(struct nft_trans_gc *gc, gfp_t gfp)
net/netfilter/nf_tables_api.c
10519
return nft_trans_gc_alloc(set, 0, gfp);
net/netfilter/nf_tables_api.c
3615
int nft_expr_clone(struct nft_expr *dst, struct nft_expr *src, gfp_t gfp)
net/netfilter/nf_tables_api.c
3623
err = src->ops->clone(dst, src, gfp);
net/netfilter/nf_tables_api.c
6692
u64 timeout, u64 expiration, gfp_t gfp)
net/netfilter/nf_tables_api.c
6697
elem = kzalloc(set->ops->elemsize + tmpl->len, gfp);
net/netfilter/nf_tables_api.c
8630
u16 flags, int family, int report, gfp_t gfp)
net/netfilter/nf_tables_api.c
8640
skb = nlmsg_new(NLMSG_GOODSIZE, gfp);
net/netfilter/nf_tables_api.c
8660
u16 flags, int family, int report, gfp_t gfp)
net/netfilter/nf_tables_api.c
8662
char *buf = kasprintf(gfp, "%s:%u",
net/netfilter/nf_tables_api.c
8671
gfp);
net/netfilter/nf_tables_api.c
8675
flags, family, report, gfp);
net/netfilter/nft_connlimit.c
218
static int nft_connlimit_clone(struct nft_expr *dst, const struct nft_expr *src, gfp_t gfp)
net/netfilter/nft_connlimit.c
223
priv_dst->list = kmalloc_obj(*priv_dst->list, gfp);
net/netfilter/nft_counter.c
250
static int nft_counter_clone(struct nft_expr *dst, const struct nft_expr *src, gfp_t gfp)
net/netfilter/nft_counter.c
260
cpu_stats = alloc_percpu_gfp(struct nft_counter, gfp);
net/netfilter/nft_last.c
105
static int nft_last_clone(struct nft_expr *dst, const struct nft_expr *src, gfp_t gfp)
net/netfilter/nft_last.c
110
priv_dst->last = kzalloc_obj(*priv_dst->last, gfp);
net/netfilter/nft_limit.c
153
const struct nft_limit_priv *priv_src, gfp_t gfp)
net/netfilter/nft_limit.c
161
priv_dst->limit = kmalloc_obj(*priv_dst->limit, gfp);
net/netfilter/nft_limit.c
227
gfp_t gfp)
net/netfilter/nft_limit.c
234
return nft_limit_clone(&priv_dst->limit, &priv_src->limit, gfp);
net/netfilter/nft_limit.c
286
gfp_t gfp)
net/netfilter/nft_limit.c
291
return nft_limit_clone(priv_dst, priv_src, gfp);
net/netfilter/nft_quota.c
243
static int nft_quota_clone(struct nft_expr *dst, const struct nft_expr *src, gfp_t gfp)
net/netfilter/nft_quota.c
251
priv_dst->consumed = kmalloc_obj(*priv_dst->consumed, gfp);
net/nfc/core.c
737
struct sk_buff *nfc_alloc_recv_skb(unsigned int size, gfp_t gfp)
net/nfc/core.c
743
skb = alloc_skb(total_size, gfp);
net/nfc/llcp.h
214
struct sock *nfc_llcp_sock_alloc(struct socket *sock, int type, gfp_t gfp, int kern);
net/nfc/llcp_sock.c
974
struct sock *nfc_llcp_sock_alloc(struct socket *sock, int type, gfp_t gfp, int kern)
net/nfc/llcp_sock.c
979
sk = sk_alloc(&init_net, PF_NFC, gfp, &llcp_sock_proto, kern);
net/nfc/netlink.c
1631
u32 oui, u32 subcmd, gfp_t gfp)
net/nfc/netlink.c
1636
skb = nlmsg_new(approxlen + 100, gfp);
net/openvswitch/datapath.c
2128
u32 flags, u8 cmd, gfp_t gfp)
net/rds/connection.c
165
gfp_t gfp, u8 tos,
net/rds/connection.c
195
conn = kmem_cache_zalloc(rds_conn_slab, gfp);
net/rds/connection.c
200
conn->c_path = kzalloc_objs(struct rds_conn_path, npaths, gfp);
net/rds/connection.c
365
gfp_t gfp, int dev_if)
net/rds/connection.c
367
return __rds_conn_create(net, laddr, faddr, trans, gfp, tos, 0, dev_if);
net/rds/connection.c
375
u8 tos, gfp_t gfp, int dev_if)
net/rds/connection.c
377
return __rds_conn_create(net, laddr, faddr, trans, gfp, tos, 1, dev_if);
net/rds/ib.h
368
int rds_ib_conn_alloc(struct rds_connection *conn, gfp_t gfp);
net/rds/ib.h
395
int rds_ib_recv_alloc_caches(struct rds_ib_connection *ic, gfp_t gfp);
net/rds/ib.h
397
void rds_ib_recv_refill(struct rds_connection *conn, int prefill, gfp_t gfp);
net/rds/ib_cm.c
1199
int rds_ib_conn_alloc(struct rds_connection *conn, gfp_t gfp)
net/rds/ib_cm.c
1206
ic = kzalloc_obj(struct rds_ib_connection, gfp);
net/rds/ib_cm.c
1210
ret = rds_ib_recv_alloc_caches(ic, gfp);
net/rds/ib_recv.c
102
static int rds_ib_recv_alloc_cache(struct rds_ib_refill_cache *cache, gfp_t gfp)
net/rds/ib_recv.c
107
cache->percpu = alloc_percpu_gfp(struct rds_ib_cache_head, gfp);
net/rds/ib_recv.c
122
int rds_ib_recv_alloc_caches(struct rds_ib_connection *ic, gfp_t gfp)
net/rds/ib_recv.c
126
ret = rds_ib_recv_alloc_cache(&ic->i_cache_incs, gfp);
net/rds/ib_recv.c
128
ret = rds_ib_recv_alloc_cache(&ic->i_cache_frags, gfp);
net/rds/ib_recv.c
309
struct rds_ib_recv_work *recv, gfp_t gfp)
net/rds/ib_recv.c
314
gfp_t slab_mask = gfp;
net/rds/ib_recv.c
315
gfp_t page_mask = gfp;
net/rds/ib_recv.c
317
if (gfp & __GFP_DIRECT_RECLAIM) {
net/rds/ib_recv.c
383
void rds_ib_recv_refill(struct rds_connection *conn, int prefill, gfp_t gfp)
net/rds/ib_recv.c
389
bool can_wait = !!(gfp & __GFP_DIRECT_RECLAIM);
net/rds/ib_recv.c
409
ret = rds_ib_recv_refill_one(conn, recv, gfp);
net/rds/loop.c
135
static int rds_loop_conn_alloc(struct rds_connection *conn, gfp_t gfp)
net/rds/loop.c
140
lc = kzalloc_obj(struct rds_loop_connection, gfp);
net/rds/message.c
324
struct rds_message *rds_message_alloc(unsigned int extra_len, gfp_t gfp)
net/rds/message.c
331
rm = kzalloc(sizeof(struct rds_message) + extra_len, gfp);
net/rds/page.c
124
page = alloc_page(gfp);
net/rds/page.c
71
gfp_t gfp)
net/rds/page.c
77
gfp |= __GFP_HIGHMEM;
net/rds/page.c
81
page = alloc_page(gfp);
net/rds/rds.h
539
int (*conn_alloc)(struct rds_connection *conn, gfp_t gfp);
net/rds/rds.h
802
u8 tos, gfp_t gfp,
net/rds/rds.h
808
u8 tos, gfp_t gfp, int dev_if);
net/rds/rds.h
887
struct rds_message *rds_message_alloc(unsigned int nents, gfp_t gfp);
net/rds/rds.h
920
gfp_t gfp);
net/rds/rds.h
931
struct rds_incoming *inc, gfp_t gfp);
net/rds/recv.c
318
struct rds_incoming *inc, gfp_t gfp)
net/rds/tcp.c
369
static int rds_tcp_conn_alloc(struct rds_connection *conn, gfp_t gfp)
net/rds/tcp.c
376
tc = kmem_cache_zalloc(rds_tcp_conn_slab, gfp);
net/rds/tcp_recv.c
152
gfp_t gfp;
net/rds/tcp_recv.c
175
arg->gfp);
net/rds/tcp_recv.c
219
clone = pskb_extract(skb, offset, to_copy, arg->gfp);
net/rds/tcp_recv.c
246
arg->gfp);
net/rds/tcp_recv.c
263
static int rds_tcp_read_sock(struct rds_conn_path *cp, gfp_t gfp)
net/rds/tcp_recv.c
272
arg.gfp = gfp;
net/rds/tcp_recv.c
278
rdsdebug("tcp_read_sock for tc %p gfp 0x%x returned %d\n", tc, gfp,
net/rxrpc/af_rxrpc.c
274
struct sockaddr_rxrpc *srx, gfp_t gfp)
net/rxrpc/af_rxrpc.c
284
peer = rxrpc_lookup_peer(rx->local, srx, gfp);
net/rxrpc/af_rxrpc.c
345
gfp_t gfp,
net/rxrpc/af_rxrpc.c
384
call = rxrpc_new_client_call(rx, &cp, &p, gfp, debug_id);
net/rxrpc/ar-internal.h
1141
int rxrpc_look_up_bundle(struct rxrpc_call *call, gfp_t gfp);
net/rxrpc/ar-internal.h
1395
struct sockaddr_rxrpc *srx, gfp_t gfp);
net/rxrpc/ar-internal.h
1519
size_t data_align, gfp_t gfp);
net/rxrpc/ar-internal.h
278
struct rxrpc_txbuf *(*alloc_txbuf)(struct rxrpc_call *call, size_t remaining, gfp_t gfp);
net/rxrpc/call_accept.c
162
int rxrpc_service_prealloc(struct rxrpc_sock *rx, gfp_t gfp)
net/rxrpc/call_accept.c
167
b = kzalloc_obj(struct rxrpc_backlog, gfp);
net/rxrpc/call_accept.c
37
unsigned long user_call_ID, gfp_t gfp,
net/rxrpc/call_accept.c
471
unsigned long user_call_ID, gfp_t gfp,
net/rxrpc/call_accept.c
481
gfp, debug_id);
net/rxrpc/call_accept.c
73
peer = rxrpc_alloc_peer(rx->local, gfp, rxrpc_peer_new_prealloc);
net/rxrpc/call_accept.c
86
conn = rxrpc_prealloc_service_connection(rxnet, gfp);
net/rxrpc/call_accept.c
97
call = rxrpc_alloc_call(rx, gfp, debug_id);
net/rxrpc/call_object.c
121
struct rxrpc_call *rxrpc_alloc_call(struct rxrpc_sock *rx, gfp_t gfp,
net/rxrpc/call_object.c
127
call = kmem_cache_zalloc(rxrpc_call_jar, gfp);
net/rxrpc/call_object.c
191
gfp_t gfp,
net/rxrpc/call_object.c
200
call = rxrpc_alloc_call(rx, gfp, debug_id);
net/rxrpc/call_object.c
261
static struct semaphore *rxrpc_get_call_slot(struct rxrpc_call_params *p, gfp_t gfp)
net/rxrpc/call_object.c
291
static int rxrpc_connect_call(struct rxrpc_call *call, gfp_t gfp)
net/rxrpc/call_object.c
298
ret = rxrpc_look_up_bundle(call, gfp);
net/rxrpc/call_object.c
323
gfp_t gfp,
net/rxrpc/call_object.c
341
limiter = rxrpc_get_call_slot(p, gfp);
net/rxrpc/call_object.c
347
call = rxrpc_alloc_client_call(rx, cp, p, gfp, debug_id);
net/rxrpc/call_object.c
398
ret = rxrpc_connect_call(call, gfp);
net/rxrpc/conn_client.c
250
int rxrpc_look_up_bundle(struct rxrpc_call *call, gfp_t gfp)
net/rxrpc/conn_client.c
263
call->bundle = rxrpc_alloc_bundle(call, gfp);
net/rxrpc/conn_client.c
291
candidate = rxrpc_alloc_bundle(call, gfp);
net/rxrpc/conn_client.c
74
gfp_t gfp)
net/rxrpc/conn_client.c
79
bundle = kzalloc_obj(*bundle, gfp);
net/rxrpc/conn_object.c
56
gfp_t gfp)
net/rxrpc/conn_object.c
62
conn = kzalloc_obj(struct rxrpc_connection, gfp);
net/rxrpc/conn_service.c
121
gfp_t gfp)
net/rxrpc/conn_service.c
123
struct rxrpc_connection *conn = rxrpc_alloc_connection(rxnet, gfp);
net/rxrpc/insecure.c
20
static struct rxrpc_txbuf *none_alloc_txbuf(struct rxrpc_call *call, size_t remain, gfp_t gfp)
net/rxrpc/insecure.c
22
return rxrpc_alloc_data_txbuf(call, umin(remain, RXRPC_JUMBO_DATALEN), 1, gfp);
net/rxrpc/output.c
83
gfp_t gfp = rcu_read_lock_held() ? GFP_ATOMIC | __GFP_NOWARN : GFP_NOFS;
net/rxrpc/output.c
88
sizeof(*whdr) + sizeof(*ack) + 1 + 3 + sizeof(*trailer), gfp);
net/rxrpc/output.c
93
buf2 = page_frag_alloc(&call->local->tx_alloc, sack_size, gfp);
net/rxrpc/peer_object.c
222
struct rxrpc_peer *rxrpc_alloc_peer(struct rxrpc_local *local, gfp_t gfp,
net/rxrpc/peer_object.c
229
peer = kzalloc_obj(struct rxrpc_peer, gfp);
net/rxrpc/peer_object.c
287
gfp_t gfp)
net/rxrpc/peer_object.c
293
peer = rxrpc_alloc_peer(local, gfp, rxrpc_peer_new_client);
net/rxrpc/peer_object.c
335
struct sockaddr_rxrpc *srx, gfp_t gfp)
net/rxrpc/peer_object.c
354
candidate = rxrpc_create_peer(local, srx, hash_key, gfp);
net/rxrpc/rxgk.c
283
static struct rxrpc_txbuf *rxgk_alloc_txbuf(struct rxrpc_call *call, size_t remain, gfp_t gfp)
net/rxrpc/rxgk.c
293
return rxrpc_alloc_data_txbuf(call, alloc, 1, gfp);
net/rxrpc/rxgk.c
325
txb = rxrpc_alloc_data_txbuf(call, alloc, 16, gfp);
net/rxrpc/rxgk.c
620
gfp_t gfp = GFP_NOFS;
net/rxrpc/rxgk.c
625
gfp |= __GFP_COMP;
net/rxrpc/rxgk.c
626
return alloc_pages(gfp, order);
net/rxrpc/rxgk_common.h
56
gfp_t gfp);
net/rxrpc/rxgk_common.h
61
gfp_t gfp);
net/rxrpc/rxgk_kdf.c
108
buffer = kzalloc(krb5->key_bytes, gfp);
net/rxrpc/rxgk_kdf.c
115
ret = rxgk_derive_transport_key(conn, gk, rxgk, &TK, gfp);
net/rxrpc/rxgk_kdf.c
119
aead = crypto_krb5_prepare_encryption(krb5, &TK, RXGK_CLIENT_ENC_RESPONSE, gfp);
net/rxrpc/rxgk_kdf.c
137
krb5, &TK, RXGK_SERVER_MIC_PACKET, gfp);
net/rxrpc/rxgk_kdf.c
142
krb5, &TK, RXGK_CLIENT_MIC_PACKET, gfp);
net/rxrpc/rxgk_kdf.c
149
krb5, &TK, RXGK_SERVER_ENC_PACKET, gfp);
net/rxrpc/rxgk_kdf.c
154
krb5, &TK, RXGK_CLIENT_ENC_PACKET, gfp);
net/rxrpc/rxgk_kdf.c
164
krb5, &TK, RXGK_CLIENT_MIC_PACKET, gfp);
net/rxrpc/rxgk_kdf.c
169
krb5, &TK, RXGK_SERVER_MIC_PACKET, gfp);
net/rxrpc/rxgk_kdf.c
176
krb5, &TK, RXGK_CLIENT_ENC_PACKET, gfp);
net/rxrpc/rxgk_kdf.c
181
krb5, &TK, RXGK_SERVER_ENC_PACKET, gfp);
net/rxrpc/rxgk_kdf.c
208
gfp_t gfp)
net/rxrpc/rxgk_kdf.c
227
ret = rxgk_set_up_ciphers(conn, gk, key, gfp);
net/rxrpc/rxgk_kdf.c
272
gfp_t gfp)
net/rxrpc/rxgk_kdf.c
281
aead = crypto_krb5_prepare_encryption(krb5, server_key, RXGK_SERVER_ENC_TOKEN, gfp);
net/rxrpc/rxgk_kdf.c
60
gfp_t gfp)
net/rxrpc/rxgk_kdf.c
73
buffer = kzalloc(round16(conn_info.len), gfp);
net/rxrpc/rxgk_kdf.c
86
ret = crypto_krb5_calc_PRFplus(krb5, &rxgk->key, L, &conn_info, TK, gfp);
net/rxrpc/rxgk_kdf.c
98
gfp_t gfp)
net/rxrpc/rxkad.c
148
static struct rxrpc_txbuf *rxkad_alloc_txbuf(struct rxrpc_call *call, size_t remain, gfp_t gfp)
net/rxrpc/rxkad.c
158
return rxrpc_alloc_data_txbuf(call, alloc, 1, gfp);
net/rxrpc/rxkad.c
176
txb = rxrpc_alloc_data_txbuf(call, alloc, RXKAD_ALIGN, gfp);
net/rxrpc/txbuf.c
20
size_t data_align, gfp_t gfp)
net/rxrpc/txbuf.c
26
txb = kzalloc_obj(*txb, gfp);
net/rxrpc/txbuf.c
38
buf = page_frag_alloc_align(&call->conn->tx_data_alloc, total, gfp,
net/sched/act_ife.c
114
int ife_alloc_meta_u32(struct tcf_meta_info *mi, void *metaval, gfp_t gfp)
net/sched/act_ife.c
116
mi->metaval = kmemdup(metaval, sizeof(u32), gfp);
net/sched/act_ife.c
124
int ife_alloc_meta_u16(struct tcf_meta_info *mi, void *metaval, gfp_t gfp)
net/sched/act_ife.c
126
mi->metaval = kmemdup(metaval, sizeof(u16), gfp);
net/sctp/associola.c
1555
enum sctp_scope scope, gfp_t gfp)
net/sctp/associola.c
1574
scope, gfp, flags);
net/sctp/associola.c
1580
gfp_t gfp)
net/sctp/associola.c
1588
asoc->ep->base.bind_addr.port, gfp);
net/sctp/associola.c
1606
int sctp_assoc_set_id(struct sctp_association *asoc, gfp_t gfp)
net/sctp/associola.c
1608
bool preload = gfpflags_allow_blocking(gfp);
net/sctp/associola.c
1616
idr_preload(gfp);
net/sctp/associola.c
232
if (sctp_stream_init(&asoc->stream, asoc->c.sinit_num_ostreams, 0, gfp))
net/sctp/associola.c
256
if (sctp_auth_asoc_copy_shkeys(ep, asoc, gfp))
net/sctp/associola.c
288
enum sctp_scope scope, gfp_t gfp)
net/sctp/associola.c
292
asoc = kzalloc_obj(*asoc, gfp);
net/sctp/associola.c
296
if (!sctp_association_init(asoc, ep, sk, scope, gfp))
net/sctp/associola.c
55
enum sctp_scope scope, gfp_t gfp)
net/sctp/associola.c
585
const gfp_t gfp,
net/sctp/associola.c
617
peer = sctp_transport_new(asoc->base.net, addr, gfp);
net/sctp/auth.c
189
gfp_t gfp)
net/sctp/auth.c
203
new = sctp_auth_create_key(len, gfp);
net/sctp/auth.c
224
gfp_t gfp)
net/sctp/auth.c
229
(struct sctp_hmac_algo_param *)asoc->c.auth_hmacs, gfp);
net/sctp/auth.c
235
gfp_t gfp)
net/sctp/auth.c
240
gfp);
net/sctp/auth.c
257
gfp_t gfp)
net/sctp/auth.c
267
secret = sctp_auth_create_key(auth_len, gfp);
net/sctp/auth.c
290
gfp_t gfp)
net/sctp/auth.c
312
local_key_vector = sctp_auth_make_local_vector(asoc, gfp);
net/sctp/auth.c
313
peer_key_vector = sctp_auth_make_peer_vector(asoc, gfp);
net/sctp/auth.c
342
gfp);
net/sctp/auth.c
356
gfp_t gfp)
net/sctp/auth.c
364
new = sctp_auth_shkey_create(sh_key->key_id, gfp);
net/sctp/auth.c
384
int sctp_auth_asoc_init_active_key(struct sctp_association *asoc, gfp_t gfp)
net/sctp/auth.c
404
secret = sctp_auth_asoc_create_secret(asoc, ep_key, gfp);
net/sctp/auth.c
59
static struct sctp_auth_bytes *sctp_auth_create_key(__u32 key_len, gfp_t gfp)
net/sctp/auth.c
618
struct sctp_shared_key *ep_key, gfp_t gfp)
net/sctp/auth.c
637
asoc_key = sctp_auth_asoc_create_secret(asoc, ep_key, gfp);
net/sctp/auth.c
68
key = kmalloc(sizeof(struct sctp_auth_bytes) + key_len, gfp);
net/sctp/auth.c
80
struct sctp_shared_key *sctp_auth_shkey_create(__u16 key_id, gfp_t gfp)
net/sctp/auth.c
85
new = kzalloc_obj(struct sctp_shared_key, gfp);
net/sctp/auth.c
925
int sctp_auth_init(struct sctp_endpoint *ep, gfp_t gfp)
net/sctp/auth.c
935
SCTP_AUTH_NUM_HMACS, gfp);
net/sctp/auth.c
954
SCTP_NUM_CHUNK_TYPES, gfp);
net/sctp/bind_addr.c
106
1, gfp);
net/sctp/bind_addr.c
145
int new_size, __u8 addr_state, gfp_t gfp)
net/sctp/bind_addr.c
150
addr = kzalloc_obj(*addr, gfp);
net/sctp/bind_addr.c
213
gfp_t gfp)
net/sctp/bind_addr.c
240
retval.v = kmalloc(len, gfp);
net/sctp/bind_addr.c
264
int addrs_len, __u16 port, gfp_t gfp)
net/sctp/bind_addr.c
288
SCTP_ADDR_SRC, gfp);
net/sctp/bind_addr.c
35
gfp_t gfp, int flags);
net/sctp/bind_addr.c
45
enum sctp_scope scope, gfp_t gfp,
net/sctp/bind_addr.c
461
gfp_t gfp, int flags)
net/sctp/bind_addr.c
466
error = sctp_copy_local_addr_list(net, dest, scope, gfp, flags);
net/sctp/bind_addr.c
479
SCTP_ADDR_SRC, gfp);
net/sctp/bind_addr.c
57
gfp, flags);
net/sctp/bind_addr.c
69
SCTP_SCOPE_LINK, gfp,
net/sctp/bind_addr.c
96
gfp_t gfp)
net/sctp/chunk.c
47
static struct sctp_datamsg *sctp_datamsg_new(gfp_t gfp)
net/sctp/chunk.c
50
msg = kmalloc_obj(struct sctp_datamsg, gfp);
net/sctp/endpointola.c
103
null_key = sctp_auth_shkey_create(0, gfp);
net/sctp/endpointola.c
133
struct sctp_endpoint *sctp_endpoint_new(struct sock *sk, gfp_t gfp)
net/sctp/endpointola.c
138
ep = kzalloc_obj(*ep, gfp);
net/sctp/endpointola.c
142
if (!sctp_endpoint_init(ep, sk, gfp))
net/sctp/endpointola.c
52
gfp_t gfp)
net/sctp/endpointola.c
60
if (sctp_auth_init(ep, gfp))
net/sctp/output.c
182
int one_packet, gfp_t gfp)
net/sctp/output.c
194
error = sctp_packet_transmit(packet, gfp);
net/sctp/output.c
437
struct sk_buff *head, int gso, gfp_t gfp)
net/sctp/output.c
472
nskb = alloc_skb(pkt_size + MAX_HEADER, gfp);
net/sctp/output.c
521
packet->auth->shkey, gfp);
net/sctp/output.c
569
int sctp_packet_transmit(struct sctp_packet *packet, gfp_t gfp)
net/sctp/output.c
599
MAX_HEADER, gfp);
net/sctp/output.c
621
pkt_count = sctp_packet_pack(packet, head, gso, gfp);
net/sctp/outqueue.c
1013
&start_timer, ctx->gfp);
net/sctp/outqueue.c
1112
ctx->gfp);
net/sctp/outqueue.c
1173
error = sctp_packet_transmit(packet, ctx->gfp);
net/sctp/outqueue.c
1192
static void sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp)
net/sctp/outqueue.c
1200
.gfp = gfp,
net/sctp/outqueue.c
281
void sctp_outq_tail(struct sctp_outq *q, struct sctp_chunk *chunk, gfp_t gfp)
net/sctp/outqueue.c
313
sctp_outq_flush(q, 0, gfp);
net/sctp/outqueue.c
56
static void sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp);
net/sctp/outqueue.c
599
int rtx_timeout, int *start_timer, gfp_t gfp)
net/sctp/outqueue.c
675
sctp_packet_transmit(pkt, gfp);
net/sctp/outqueue.c
680
error = sctp_packet_transmit(pkt, gfp);
net/sctp/outqueue.c
696
error = sctp_packet_transmit(pkt, gfp);
net/sctp/outqueue.c
706
error = sctp_packet_transmit(pkt, gfp);
net/sctp/outqueue.c
759
void sctp_outq_uncork(struct sctp_outq *q, gfp_t gfp)
net/sctp/outqueue.c
764
sctp_outq_flush(q, 0, gfp);
net/sctp/outqueue.c
768
struct sctp_chunk *chunk, gfp_t gfp)
net/sctp/outqueue.c
783
return sctp_packet_transmit(&singleton, gfp);
net/sctp/outqueue.c
796
gfp_t gfp;
net/sctp/outqueue.c
915
ctx->gfp);
net/sctp/outqueue.c
946
chunk, ctx->gfp);
net/sctp/outqueue.c
960
one_packet, ctx->gfp);
net/sctp/protocol.c
136
enum sctp_scope scope, gfp_t gfp, int copy_flags)
net/sctp/sm_make_chunk.c
1369
struct sock *sk, gfp_t gfp)
net/sctp/sm_make_chunk.c
1373
retval = kmem_cache_zalloc(sctp_chunk_cachep, gfp);
net/sctp/sm_make_chunk.c
1422
gfp_t gfp)
net/sctp/sm_make_chunk.c
1435
skb = alloc_skb(chunklen, gfp);
net/sctp/sm_make_chunk.c
1446
retval = sctp_chunkify(skb, asoc, sk, gfp);
net/sctp/sm_make_chunk.c
1465
__u8 flags, int paylen, gfp_t gfp)
net/sctp/sm_make_chunk.c
1467
return _sctp_make_chunk(asoc, SCTP_CID_DATA, flags, paylen, gfp);
net/sctp/sm_make_chunk.c
1471
__u8 flags, int paylen, gfp_t gfp)
net/sctp/sm_make_chunk.c
1473
return _sctp_make_chunk(asoc, SCTP_CID_I_DATA, flags, paylen, gfp);
net/sctp/sm_make_chunk.c
1478
gfp_t gfp)
net/sctp/sm_make_chunk.c
1482
chunk = _sctp_make_chunk(asoc, type, flags, paylen, gfp);
net/sctp/sm_make_chunk.c
1622
gfp_t gfp)
net/sctp/sm_make_chunk.c
1630
asoc = sctp_association_new(ep, ep->base.sk, scope, gfp);
net/sctp/sm_make_chunk.c
1725
struct sctp_chunk *chunk, gfp_t gfp,
net/sctp/sm_make_chunk.c
1831
retval = sctp_association_new(ep, ep->base.sk, scope, gfp);
net/sctp/sm_make_chunk.c
200
gfp_t gfp, int vparam_len)
net/sctp/sm_make_chunk.c
225
addrs = sctp_bind_addrs_to_raw(bp, &addrs_len, gfp);
net/sctp/sm_make_chunk.c
2327
struct sctp_init_chunk *peer_init, gfp_t gfp)
net/sctp/sm_make_chunk.c
2347
if (!sctp_assoc_add_peer(asoc, peer_addr, gfp, SCTP_ACTIVE))
net/sctp/sm_make_chunk.c
2366
if (!sctp_process_param(asoc, param, peer_addr, gfp))
net/sctp/sm_make_chunk.c
2452
asoc->peer.i.initial_tsn, gfp))
net/sctp/sm_make_chunk.c
2464
asoc->c.sinit_max_instreams, gfp))
net/sctp/sm_make_chunk.c
2470
if (!asoc->temp && sctp_assoc_set_id(asoc, gfp))
net/sctp/sm_make_chunk.c
2513
gfp_t gfp)
net/sctp/sm_make_chunk.c
2546
if (!sctp_assoc_add_peer(asoc, &addr, gfp, SCTP_UNCONFIRMED))
net/sctp/sm_make_chunk.c
2603
asoc->peer.cookie = kmemdup(param.cookie->body, asoc->peer.cookie_len, gfp);
net/sctp/sm_make_chunk.c
2672
ntohs(param.p->length), gfp);
net/sctp/sm_make_chunk.c
2686
ntohs(param.p->length), gfp);
net/sctp/sm_make_chunk.c
2702
ntohs(param.p->length), gfp);
net/sctp/sm_make_chunk.c
311
retval = sctp_make_control(asoc, SCTP_CID_INIT, 0, chunksize, gfp);
net/sctp/sm_make_chunk.c
373
gfp_t gfp, int unkparam_len)
net/sctp/sm_make_chunk.c
392
addrs = sctp_bind_addrs_to_raw(&asoc->base.bind_addr, &addrs_len, gfp);
net/sctp/sm_make_chunk.c
464
retval = sctp_make_control(asoc, SCTP_CID_INIT_ACK, 0, chunksize, gfp);
net/sctp/sm_make_chunk.c
51
gfp_t gfp);
net/sctp/sm_make_chunk.c
53
__u8 flags, int paylen, gfp_t gfp);
net/sctp/sm_make_chunk.c
56
gfp_t gfp);
net/sctp/sm_make_chunk.c
66
gfp_t gfp);
net/sctp/sm_make_chunk.c
716
int len, __u8 flags, gfp_t gfp)
net/sctp/sm_make_chunk.c
732
retval = sctp_make_data(asoc, flags, sizeof(dp) + len, gfp);
net/sctp/sm_sideeffect.c
1114
struct sctp_datamsg *msg, gfp_t gfp)
net/sctp/sm_sideeffect.c
1119
sctp_outq_tail(&asoc->outqueue, chunk, gfp);
net/sctp/sm_sideeffect.c
1152
void *event_arg, gfp_t gfp)
net/sctp/sm_sideeffect.c
1177
&commands, gfp);
net/sctp/sm_sideeffect.c
1194
gfp_t gfp)
net/sctp/sm_sideeffect.c
1207
commands, gfp)))
net/sctp/sm_sideeffect.c
1283
gfp_t gfp)
net/sctp/sm_sideeffect.c
1316
sctp_outq_uncork(&asoc->outqueue, gfp);
net/sctp/sm_sideeffect.c
1332
sctp_outq_uncork(&asoc->outqueue, gfp);
net/sctp/sm_sideeffect.c
1396
cmd->obj.init, gfp);
net/sctp/sm_sideeffect.c
1487
sctp_outq_tail(&asoc->outqueue, cmd->obj.chunk, gfp);
net/sctp/sm_sideeffect.c
1493
sctp_packet_transmit(packet, gfp);
net/sctp/sm_sideeffect.c
1716
sctp_outq_uncork(&asoc->outqueue, gfp);
net/sctp/sm_sideeffect.c
1755
sctp_outq_uncork(&asoc->outqueue, gfp);
net/sctp/sm_sideeffect.c
1785
sctp_cmd_send_msg(asoc, cmd->obj.msg, gfp);
net/sctp/sm_sideeffect.c
1793
sctp_outq_uncork(&asoc->outqueue, gfp);
net/sctp/sm_sideeffect.c
1822
sctp_outq_uncork(&asoc->outqueue, gfp);
net/sctp/sm_sideeffect.c
1824
sctp_outq_uncork(&asoc->outqueue, gfp);
net/sctp/sm_sideeffect.c
48
gfp_t gfp);
net/sctp/sm_sideeffect.c
57
gfp_t gfp);
net/sctp/sm_sideeffect.c
703
gfp_t gfp)
net/sctp/sm_sideeffect.c
712
if (!sctp_process_init(asoc, chunk, sctp_source(chunk), peer_init, gfp))
net/sctp/stream.c
104
ret = genradix_prealloc(&stream->out, outcnt, gfp);
net/sctp/stream.c
114
gfp_t gfp)
net/sctp/stream.c
121
ret = genradix_prealloc(&stream->in, incnt, gfp);
net/sctp/stream.c
131
gfp_t gfp)
net/sctp/stream.c
136
gfp |= __GFP_NOWARN;
net/sctp/stream.c
149
ret = sctp_stream_alloc_out(stream, outcnt, gfp);
net/sctp/stream.c
161
return sctp_stream_alloc_in(stream, incnt, gfp);
net/sctp/stream.c
97
gfp_t gfp)
net/sctp/stream_interleave.c
1053
static void sctp_intl_abort_pd(struct sctp_ulpq *ulpq, gfp_t gfp)
net/sctp/stream_interleave.c
1066
sctp_intl_stream_abort_pd(ulpq, sid, mid, 0x1, gfp);
net/sctp/stream_interleave.c
1073
sctp_intl_stream_abort_pd(ulpq, sid, mid, 0, gfp);
net/sctp/stream_interleave.c
27
int len, __u8 flags, gfp_t gfp)
net/sctp/stream_interleave.c
38
retval = sctp_make_idata(asoc, flags, sizeof(dp) + len, gfp);
net/sctp/stream_interleave.c
820
struct sctp_chunk *chunk, gfp_t gfp)
net/sctp/stream_interleave.c
826
event = sctp_ulpevent_make_rcvmsg(chunk->asoc, chunk, gfp);
net/sctp/stream_interleave.c
926
static void sctp_intl_start_pd(struct sctp_ulpq *ulpq, gfp_t gfp)
net/sctp/stream_interleave.c
955
gfp_t gfp)
net/sctp/stream_interleave.c
974
if (freed >= needed && sctp_ulpevent_idata(ulpq, chunk, gfp) <= 0)
net/sctp/stream_interleave.c
975
sctp_intl_start_pd(ulpq, gfp);
net/sctp/stream_interleave.c
979
__u32 mid, __u16 flags, gfp_t gfp)
net/sctp/stream_interleave.c
989
sid, mid, flags, gfp);
net/sctp/stream_sched.c
207
__u16 value, gfp_t gfp)
net/sctp/stream_sched.c
220
return asoc->outqueue.sched->set(&asoc->stream, sid, value, gfp);
net/sctp/stream_sched.c
26
__u16 value, gfp_t gfp)
net/sctp/stream_sched.c
264
int sctp_sched_init_sid(struct sctp_stream *stream, __u16 sid, gfp_t gfp)
net/sctp/stream_sched.c
270
return sched->init_sid(stream, sid, gfp);
net/sctp/stream_sched.c
44
gfp_t gfp)
net/sctp/stream_sched_fc.c
28
__u16 weight, gfp_t gfp)
net/sctp/stream_sched_fc.c
49
__u16 weight, gfp_t gfp)
net/sctp/stream_sched_fc.c
68
gfp_t gfp)
net/sctp/stream_sched_prio.c
164
__u16 prio, gfp_t gfp)
net/sctp/stream_sched_prio.c
175
prio_head = sctp_sched_prio_get_head(stream, prio, gfp);
net/sctp/stream_sched_prio.c
203
gfp_t gfp)
net/sctp/stream_sched_prio.c
206
return sctp_sched_prio_set(stream, sid, 0, gfp);
net/sctp/stream_sched_prio.c
41
struct sctp_stream *stream, int prio, gfp_t gfp)
net/sctp/stream_sched_prio.c
45
p = kmalloc_obj(*p, gfp);
net/sctp/stream_sched_prio.c
59
struct sctp_stream *stream, int prio, gfp_t gfp)
net/sctp/stream_sched_prio.c
90
return sctp_sched_prio_new_head(stream, prio, gfp);
net/sctp/stream_sched_rr.c
66
__u16 prio, gfp_t gfp)
net/sctp/stream_sched_rr.c
86
gfp_t gfp)
net/sctp/transport.c
43
gfp_t gfp)
net/sctp/transport.c
91
gfp_t gfp)
net/sctp/transport.c
95
transport = kzalloc_obj(*transport, gfp);
net/sctp/transport.c
99
sctp_transport_init(net, transport, addr, gfp);
net/sctp/tsnmap.c
36
__u32 initial_tsn, gfp_t gfp)
net/sctp/tsnmap.c
39
map->tsn_map = kzalloc(len>>3, gfp);
net/sctp/ulpevent.c
117
__u16 inbound, struct sctp_chunk *chunk, gfp_t gfp)
net/sctp/ulpevent.c
131
sizeof(struct sctp_assoc_change), 0, gfp);
net/sctp/ulpevent.c
149
MSG_NOTIFICATION, gfp);
net/sctp/ulpevent.c
244
int flags, int state, int error, gfp_t gfp)
net/sctp/ulpevent.c
251
MSG_NOTIFICATION, gfp);
net/sctp/ulpevent.c
376
gfp_t gfp)
net/sctp/ulpevent.c
395
skb = skb_copy_expand(chunk->skb, sizeof(*sre), 0, gfp);
net/sctp/ulpevent.c
432
__u16 flags, __u32 error, gfp_t gfp)
net/sctp/ulpevent.c
445
gfp);
net/sctp/ulpevent.c
49
gfp_t gfp)
net/sctp/ulpevent.c
535
__u16 flags, __u32 error, gfp_t gfp)
net/sctp/ulpevent.c
54
skb = alloc_skb(size, gfp);
net/sctp/ulpevent.c
542
skb = skb_copy_expand(chunk->skb, sizeof(*ssf), 0, gfp);
net/sctp/ulpevent.c
579
__u16 flags, gfp_t gfp)
net/sctp/ulpevent.c
586
MSG_NOTIFICATION, gfp);
net/sctp/ulpevent.c
641
const struct sctp_association *asoc, gfp_t gfp)
net/sctp/ulpevent.c
648
MSG_NOTIFICATION, gfp);
net/sctp/ulpevent.c
677
gfp_t gfp)
net/sctp/ulpevent.c
701
skb = skb_clone(chunk->skb, gfp);
net/sctp/ulpevent.c
774
__u32 flags, gfp_t gfp)
net/sctp/ulpevent.c
781
MSG_NOTIFICATION, gfp);
net/sctp/ulpevent.c
827
__u32 indication, gfp_t gfp)
net/sctp/ulpevent.c
834
MSG_NOTIFICATION, gfp);
net/sctp/ulpevent.c
865
const struct sctp_association *asoc, gfp_t gfp)
net/sctp/ulpevent.c
872
MSG_NOTIFICATION, gfp);
net/sctp/ulpevent.c
890
__be16 *stream_list, gfp_t gfp)
net/sctp/ulpevent.c
898
event = sctp_ulpevent_new(length, MSG_NOTIFICATION, gfp);
net/sctp/ulpevent.c
919
__u32 remote_tsn, gfp_t gfp)
net/sctp/ulpevent.c
926
MSG_NOTIFICATION, gfp);
net/sctp/ulpevent.c
946
__u32 strchange_instrms, __u32 strchange_outstrms, gfp_t gfp)
net/sctp/ulpevent.c
953
MSG_NOTIFICATION, gfp);
net/sctp/ulpqueue.c
1020
gfp_t gfp)
net/sctp/ulpqueue.c
1070
gfp_t gfp)
net/sctp/ulpqueue.c
1087
int retval = sctp_ulpq_tail_data(ulpq, chunk, gfp);
net/sctp/ulpqueue.c
1093
sctp_ulpq_partial_delivery(ulpq, gfp);
net/sctp/ulpqueue.c
1102
void sctp_ulpq_abort_pd(struct sctp_ulpq *ulpq, gfp_t gfp)
net/sctp/ulpqueue.c
1117
0, 0, 0, gfp);
net/sctp/ulpqueue.c
83
gfp_t gfp)
net/sctp/ulpqueue.c
90
event = sctp_ulpevent_make_rcvmsg(chunk->asoc, chunk, gfp);
net/sunrpc/auth.c
515
int flags, gfp_t gfp)
net/sunrpc/auth.c
538
new = auth->au_ops->crcreate(auth, acred, flags, gfp);
net/sunrpc/auth_gss/auth_gss.c
1377
gss_create_cred(struct rpc_auth *auth, struct auth_cred *acred, int flags, gfp_t gfp)
net/sunrpc/auth_gss/auth_gss.c
1383
if (!(cred = kzalloc_obj(*cred, gfp)))
net/sunrpc/sched.c
1058
gfp_t gfp = rpc_task_gfp_mask();
net/sunrpc/sched.c
1062
buf = kmem_cache_alloc(rpc_buffer_slabp, gfp);
net/sunrpc/sched.c
1067
buf = kmalloc(size, gfp);
net/sunrpc/xdr.c
116
xdr_alloc_bvec(struct xdr_buf *buf, gfp_t gfp)
net/sunrpc/xdr.c
121
buf->bvec = kmalloc_objs(buf->bvec[0], n, gfp);
net/sunrpc/xdr.c
504
unsigned int buflen, gfp_t gfp)
net/sunrpc/xdr.c
519
buf->pages[i] = alloc_page(gfp);
net/sunrpc/xprtsock.c
340
xs_alloc_sparse_pages(struct xdr_buf *buf, size_t want, gfp_t gfp)
net/sunrpc/xprtsock.c
350
buf->bvec[i].bv_page = buf->pages[i] = alloc_page(gfp);
net/tipc/msg.c
68
struct sk_buff *tipc_buf_acquire(u32 size, gfp_t gfp)
net/tipc/msg.c
72
skb = alloc_skb_fclone(BUF_OVERHEAD + size, gfp);
net/tipc/msg.h
1155
struct sk_buff *tipc_buf_acquire(u32 size, gfp_t gfp);
net/vmw_vsock/virtio_transport.c
102
struct virtio_vsock *vsock, gfp_t gfp)
net/vmw_vsock/virtio_transport.c
146
ret = virtqueue_add_sgs(vq, sgs, out_sg, in_sg, skb, gfp);
net/wireless/core.c
1435
int link_id, gfp_t gfp)
net/wireless/core.c
1449
ev = kzalloc_obj(*ev, gfp);
net/wireless/ibss.c
60
struct ieee80211_channel *channel, gfp_t gfp)
net/wireless/ibss.c
72
ev = kzalloc_obj(*ev, gfp);
net/wireless/mlme.c
1106
bool offchan, gfp_t gfp)
net/wireless/mlme.c
1123
nl80211_radar_notify(rdev, chandef, NL80211_RADAR_DETECTED, NULL, gfp);
net/wireless/mlme.c
1132
enum nl80211_radar_event event, gfp_t gfp,
net/wireless/mlme.c
1172
nl80211_radar_notify(rdev, chandef, event, netdev, gfp);
net/wireless/mlme.c
235
const u8 *tsc, gfp_t gfp)
net/wireless/mlme.c
241
char *buf = kmalloc(128, gfp);
net/wireless/mlme.c
256
nl80211_michael_mic_failure(rdev, dev, addr, key_type, key_id, tsc, gfp);
net/wireless/nl80211.c
12939
gfp_t gfp)
net/wireless/nl80211.c
12945
skb = nlmsg_new(approxlen + 100, gfp);
net/wireless/nl80211.c
12998
int approxlen, gfp_t gfp)
net/wireless/nl80211.c
13021
cmd, attr, info, gfp);
net/wireless/nl80211.c
13025
void __cfg80211_send_event_skb(struct sk_buff *skb, gfp_t gfp)
net/wireless/nl80211.c
13047
skb, 0, mcgrp, gfp);
net/wireless/nl80211.c
16215
struct cfg80211_nan_match_params *match, gfp_t gfp)
net/wireless/nl80211.c
16229
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
16285
msg, 0, NL80211_MCGRP_NAN, gfp);
net/wireless/nl80211.c
16300
u64 cookie, gfp_t gfp)
net/wireless/nl80211.c
16314
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
16348
msg, 0, NL80211_MCGRP_NAN, gfp);
net/wireless/nl80211.c
19507
gfp_t gfp)
net/wireless/nl80211.c
19512
msg = nlmsg_new(100 + event->buf_len + event->req_ies_len, gfp);
net/wireless/nl80211.c
19550
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
19559
size_t len, gfp_t gfp)
net/wireless/nl80211.c
19568
nl80211_send_mlme_event(rdev, netdev, &event, gfp);
net/wireless/nl80211.c
19589
size_t len, bool reconnect, gfp_t gfp)
net/wireless/nl80211.c
19599
nl80211_send_mlme_event(rdev, netdev, &event, gfp);
net/wireless/nl80211.c
19604
size_t len, bool reconnect, gfp_t gfp)
net/wireless/nl80211.c
19614
nl80211_send_mlme_event(rdev, netdev, &event, gfp);
net/wireless/nl80211.c
19654
const u8 *addr, gfp_t gfp)
net/wireless/nl80211.c
19659
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
19678
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
19687
gfp_t gfp)
net/wireless/nl80211.c
19690
addr, gfp);
net/wireless/nl80211.c
19695
gfp_t gfp)
net/wireless/nl80211.c
19698
addr, gfp);
net/wireless/nl80211.c
19704
gfp_t gfp)
net/wireless/nl80211.c
19731
gfp);
net/wireless/nl80211.c
19807
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
19816
struct cfg80211_roam_info *info, gfp_t gfp)
net/wireless/nl80211.c
19845
link_info_size, gfp);
net/wireless/nl80211.c
19911
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20079
gfp_t gfp)
net/wireless/nl80211.c
20084
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
20102
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20111
int sig_dbm, gfp_t gfp)
net/wireless/nl80211.c
20123
msg = nlmsg_new(100 + ie_len, gfp);
net/wireless/nl80211.c
20145
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20156
const u8 *tsc, gfp_t gfp)
net/wireless/nl80211.c
20161
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
20183
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20248
unsigned int duration, gfp_t gfp)
net/wireless/nl80211.c
20253
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
20282
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20329
unsigned int duration, gfp_t gfp)
net/wireless/nl80211.c
20337
duration, gfp);
net/wireless/nl80211.c
20343
gfp_t gfp)
net/wireless/nl80211.c
20350
rdev, wdev, cookie, chan, 0, gfp);
net/wireless/nl80211.c
20356
gfp_t gfp)
net/wireless/nl80211.c
20363
rdev, wdev, cookie, chan, 0, gfp);
net/wireless/nl80211.c
20368
struct station_info *sinfo, gfp_t gfp)
net/wireless/nl80211.c
20376
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
20387
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20392
struct station_info *sinfo, gfp_t gfp)
net/wireless/nl80211.c
20404
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
20417
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20423
gfp_t gfp)
net/wireless/nl80211.c
20430
msg = nlmsg_new(NLMSG_GOODSIZE, gfp);
net/wireless/nl80211.c
20448
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20457
const u8 *addr, int link_id, gfp_t gfp)
net/wireless/nl80211.c
20468
msg = nlmsg_new(100, gfp);
net/wireless/nl80211.c
20495
int link_id, gfp_t gfp)
net/wireless/nl80211.c
20508
addr, link_id, gfp);
net/wireless/nl80211.c
20515
int link_id, gfp_t gfp)
net/wireless/nl80211.c
20530
addr, link_id, gfp);
net/wireless/nl80211.c
20538
struct cfg80211_rx_info *info, gfp_t gfp)
net/wireless/nl80211.c
20544
msg = nlmsg_new(100 + info->len, gfp);
net/wireless/nl80211.c
20589
gfp_t gfp, enum nl80211_commands command)
net/wireless/nl80211.c
20604
msg = nlmsg_new(100 + status->len, gfp);
net/wireless/nl80211.c
20634
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20643
gfp_t gfp)
net/wireless/nl80211.c
20652
nl80211_frame_tx_status(wdev, &status, gfp,
net/wireless/nl80211.c
20658
struct cfg80211_tx_status *status, gfp_t gfp)
net/wireless/nl80211.c
20660
nl80211_frame_tx_status(wdev, status, gfp, NL80211_CMD_FRAME_TX_STATUS);
net/wireless/nl80211.c
20668
gfp_t gfp)
net/wireless/nl80211.c
20684
msg = nlmsg_new(100 + skb->len, gfp);
net/wireless/nl80211.c
20734
const char *mac, gfp_t gfp)
net/wireless/nl80211.c
20738
struct sk_buff *msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
20771
static void cfg80211_send_cqm(struct sk_buff *msg, gfp_t gfp)
net/wireless/nl80211.c
20782
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20787
s32 rssi_level, gfp_t gfp)
net/wireless/nl80211.c
20851
u32 rate, u32 intvl, gfp_t gfp)
net/wireless/nl80211.c
20855
msg = cfg80211_prepare_cqm(dev, peer, gfp);
net/wireless/nl80211.c
20868
cfg80211_send_cqm(msg, gfp);
net/wireless/nl80211.c
20877
const u8 *peer, u32 num_packets, gfp_t gfp)
net/wireless/nl80211.c
20883
msg = cfg80211_prepare_cqm(dev, peer, gfp);
net/wireless/nl80211.c
20890
cfg80211_send_cqm(msg, gfp);
net/wireless/nl80211.c
20898
void cfg80211_cqm_beacon_loss_notify(struct net_device *dev, gfp_t gfp)
net/wireless/nl80211.c
20902
msg = cfg80211_prepare_cqm(dev, NULL, gfp);
net/wireless/nl80211.c
20909
cfg80211_send_cqm(msg, gfp);
net/wireless/nl80211.c
20919
const u8 *replay_ctr, gfp_t gfp)
net/wireless/nl80211.c
20925
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
20953
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
20961
const u8 *replay_ctr, gfp_t gfp)
net/wireless/nl80211.c
20968
nl80211_gtk_rekey_notify(rdev, dev, bssid, replay_ctr, gfp);
net/wireless/nl80211.c
20975
const u8 *bssid, bool preauth, gfp_t gfp)
net/wireless/nl80211.c
20981
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
21010
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
21018
const u8 *bssid, bool preauth, gfp_t gfp)
net/wireless/nl80211.c
21025
nl80211_pmksa_candidate_notify(rdev, dev, index, bssid, preauth, gfp);
net/wireless/nl80211.c
21033
gfp_t gfp,
net/wireless/nl80211.c
21041
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
21072
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
21197
struct net_device *netdev, gfp_t gfp)
net/wireless/nl80211.c
21202
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
21234
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
21243
gfp_t gfp)
net/wireless/nl80211.c
21253
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
21287
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
21298
bool is_valid_ack_signal, gfp_t gfp)
net/wireless/nl80211.c
21307
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
21331
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
21447
gfp_t gfp)
net/wireless/nl80211.c
21459
msg = nlmsg_new(size, gfp);
net/wireless/nl80211.c
21556
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
21567
u16 reason_code, gfp_t gfp)
net/wireless/nl80211.c
21577
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
21598
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
21717
void cfg80211_crit_proto_stopped(struct wireless_dev *wdev, gfp_t gfp)
net/wireless/nl80211.c
21731
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
21788
gfp_t gfp)
net/wireless/nl80211.c
21798
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
21848
gfp_t gfp)
net/wireless/nl80211.c
21857
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
21888
NL80211_MCGRP_MLME, gfp);
net/wireless/nl80211.c
21945
struct ieee80211_channel *chan, gfp_t gfp)
net/wireless/nl80211.c
21957
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
21985
gfp_t gfp)
net/wireless/nl80211.c
21996
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/nl80211.c
22015
msg, 0, NL80211_MCGRP_NAN, gfp);
net/wireless/nl80211.h
108
gfp_t gfp);
net/wireless/nl80211.h
112
struct cfg80211_rx_info *info, gfp_t gfp);
net/wireless/nl80211.h
118
struct net_device *netdev, gfp_t gfp);
net/wireless/nl80211.h
60
const u8 *buf, size_t len, gfp_t gfp);
net/wireless/nl80211.h
67
bool reconnect, gfp_t gfp);
net/wireless/nl80211.h
71
bool reconnect, gfp_t gfp);
net/wireless/nl80211.h
74
const u8 *addr, gfp_t gfp);
net/wireless/nl80211.h
77
const u8 *addr, gfp_t gfp);
net/wireless/nl80211.h
81
gfp_t gfp);
net/wireless/nl80211.h
84
struct cfg80211_roam_info *info, gfp_t gfp);
net/wireless/nl80211.h
99
int key_id, const u8 *tsc, gfp_t gfp);
net/wireless/pmsr.c
367
gfp_t gfp)
net/wireless/pmsr.c
376
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/pmsr.c
576
gfp_t gfp)
net/wireless/pmsr.c
591
msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
net/wireless/reg.c
3649
gfp_t gfp)
net/wireless/reg.c
3667
reg_beacon = kzalloc_obj(struct reg_beacon, gfp);
net/wireless/reg.h
87
gfp_t gfp);
net/wireless/scan.c
2255
gfp_t gfp)
net/wireless/scan.c
2327
gfp);
net/wireless/scan.c
2331
gfp);
net/wireless/scan.c
2344
ies = kzalloc(sizeof(*ies) + data->ielen, gfp);
net/wireless/scan.c
2475
gfp_t gfp)
net/wireless/scan.c
2505
new_ie = kmalloc(IEEE80211_MAX_DATA_LEN, gfp);
net/wireless/scan.c
2509
profile = kmalloc(tx_data->ielen, gfp);
net/wireless/scan.c
2583
bss = cfg80211_inform_single_bss_data(wiphy, &data, gfp);
net/wireless/scan.c
2675
gfp_t gfp)
net/wireless/scan.c
2696
res = kzalloc_flex(*res, data, buf_len, gfp);
net/wireless/scan.c
2837
gfp_t gfp)
net/wireless/scan.c
2913
gfp);
net/wireless/scan.c
2935
gfp_t gfp)
net/wireless/scan.c
2984
mle = cfg80211_defrag_mle(elem, tx_data->ie, tx_data->ielen, gfp);
net/wireless/scan.c
2992
new_ie = kmalloc(IEEE80211_MAX_DATA_LEN, gfp);
net/wireless/scan.c
3001
gfp);
net/wireless/scan.c
3170
bss = cfg80211_inform_single_bss_data(wiphy, &data, gfp);
net/wireless/scan.c
3185
gfp_t gfp)
net/wireless/scan.c
3198
elem, gfp);
net/wireless/scan.c
3207
gfp_t gfp)
net/wireless/scan.c
3226
res = cfg80211_inform_single_bss_data(wiphy, &inform_data, gfp);
net/wireless/scan.c
3234
cfg80211_parse_mbssid_data(wiphy, &inform_data, res, gfp);
net/wireless/scan.c
3236
cfg80211_parse_ml_sta_data(wiphy, &inform_data, res, gfp);
net/wireless/scan.c
3246
gfp_t gfp)
net/wireless/scan.c
3320
gfp);
net/wireless/sme.c
1167
gfp_t gfp)
net/wireless/sme.c
1206
(info->ap_mld_addr ? ETH_ALEN : 0) + link_info_size, gfp);
net/wireless/sme.c
1311
const u8 *td_bitmap, u8 td_bitmap_len, gfp_t gfp)
net/wireless/sme.c
1321
ev = kzalloc(sizeof(*ev) + td_bitmap_len, gfp);
net/wireless/sme.c
1408
bool locally_generated, gfp_t gfp)
net/wireless/sme.c
1415
ev = kzalloc(sizeof(*ev) + ie_len, gfp);
net/wireless/sme.c
970
gfp_t gfp)
net/wireless/sme.c
990
gfp);
net/wireless/util.c
2714
gfp_t gfp)
net/wireless/util.c
2717
IEEE80211_NUM_TIDS + 1, gfp);
net/wireless/util.c
2725
int cfg80211_sinfo_alloc_tid_stats(struct station_info *sinfo, gfp_t gfp)
net/wireless/util.c
2728
gfp);
net/xfrm/xfrm_policy.c
428
struct xfrm_policy *xfrm_policy_alloc(struct net *net, gfp_t gfp)
net/xfrm/xfrm_policy.c
432
policy = kzalloc_obj(struct xfrm_policy, gfp);
security/apparmor/audit.c
220
int aa_audit_rule_init(u32 field, u32 op, char *rulestr, void **vrule, gfp_t gfp)
security/apparmor/audit.c
233
rule = kzalloc_obj(struct aa_audit_rule, gfp);
security/apparmor/audit.c
240
gfp, true, false);
security/apparmor/include/audit.h
208
int aa_audit_rule_init(u32 field, u32 op, char *rulestr, void **vrule, gfp_t gfp);
security/apparmor/include/label.h
268
bool aa_label_init(struct aa_label *label, int size, gfp_t gfp);
security/apparmor/include/label.h
269
struct aa_label *aa_label_alloc(int size, struct aa_proxy *proxy, gfp_t gfp);
security/apparmor/include/label.h
287
gfp_t gfp);
security/apparmor/include/label.h
290
bool aa_update_label_name(struct aa_ns *ns, struct aa_label *label, gfp_t gfp);
security/apparmor/include/label.h
300
int flags, gfp_t gfp);
security/apparmor/include/label.h
302
struct aa_label *label, int flags, gfp_t gfp);
security/apparmor/include/label.h
304
struct aa_label *label, int flags, gfp_t gfp);
security/apparmor/include/label.h
306
struct aa_label *label, int flags, gfp_t gfp);
security/apparmor/include/label.h
308
gfp_t gfp);
security/apparmor/include/label.h
309
void aa_label_printk(struct aa_label *label, gfp_t gfp);
security/apparmor/include/label.h
312
size_t n, gfp_t gfp, bool create,
security/apparmor/include/label.h
315
gfp_t gfp, bool create, bool force_stack);
security/apparmor/include/label.h
440
struct aa_proxy *aa_alloc_proxy(struct aa_label *l, gfp_t gfp);
security/apparmor/include/label.h
61
gfp_t gfp);
security/apparmor/include/lib.h
162
bool aa_resize_str_table(struct aa_str_table *t, int newsize, gfp_t gfp);
security/apparmor/include/lib.h
176
char *aa_str_alloc(int size, gfp_t gfp);
security/apparmor/include/lib.h
272
const char *name, gfp_t gfp);
security/apparmor/include/policy.h
126
struct aa_policydb *aa_alloc_pdb(gfp_t gfp);
security/apparmor/include/policy.h
293
struct aa_ruleset *aa_alloc_ruleset(gfp_t gfp);
security/apparmor/include/policy.h
295
gfp_t gfp);
security/apparmor/include/policy.h
297
gfp_t gfp);
security/apparmor/include/policy.h
299
const char *base, gfp_t gfp);
security/apparmor/include/secid.h
34
int aa_alloc_secid(struct aa_label *label, gfp_t gfp);
security/apparmor/label.c
1213
gfp_t gfp)
security/apparmor/label.c
1238
new = aa_label_alloc(a->size + b->size, NULL, gfp);
security/apparmor/label.c
1433
bool aa_update_label_name(struct aa_ns *ns, struct aa_label *label, gfp_t gfp)
security/apparmor/label.c
1446
if (aa_label_acntsxprint(&name, ns, label, FLAGS_NONE, gfp) < 0)
security/apparmor/label.c
1677
int flags, gfp_t gfp)
security/apparmor/label.c
1688
*strp = kmalloc(size + 1, gfp);
security/apparmor/label.c
1706
struct aa_label *label, int flags, gfp_t gfp)
security/apparmor/label.c
1717
*strp = aa_str_alloc(size + 1, gfp);
security/apparmor/label.c
1725
struct aa_label *label, int flags, gfp_t gfp)
security/apparmor/label.c
1736
len = aa_label_asxprint(&name, ns, label, flags, gfp);
security/apparmor/label.c
1755
struct aa_label *label, int flags, gfp_t gfp)
security/apparmor/label.c
1764
len = aa_label_asxprint(&str, ns, label, flags, gfp);
security/apparmor/label.c
1779
gfp_t gfp)
security/apparmor/label.c
1787
len = aa_label_asxprint(&str, ns, label, flags, gfp);
security/apparmor/label.c
1801
void aa_label_printk(struct aa_label *label, gfp_t gfp)
security/apparmor/label.c
1805
aa_label_xprintk(ns, label, FLAG_VIEW_SUBNS, gfp);
security/apparmor/label.c
1859
size_t n, gfp_t gfp, bool create,
security/apparmor/label.c
1885
error = vec_setup(profile, vec, len, gfp);
security/apparmor/label.c
1924
label = aa_vec_find_or_create_label(vec, len, gfp);
security/apparmor/label.c
1941
gfp_t gfp, bool create, bool force_stack)
security/apparmor/label.c
1943
return aa_label_strn_parse(base, str, strlen(str), gfp, create,
security/apparmor/label.c
408
bool aa_label_init(struct aa_label *label, int size, gfp_t gfp)
security/apparmor/label.c
413
if (aa_alloc_secid(label, gfp) < 0)
security/apparmor/label.c
434
struct aa_label *aa_label_alloc(int size, struct aa_proxy *proxy, gfp_t gfp)
security/apparmor/label.c
441
new = kzalloc_flex(*new, vec, size + 1, gfp);
security/apparmor/label.c
446
if (!aa_label_init(new, size, gfp))
security/apparmor/label.c
450
proxy = aa_alloc_proxy(new, gfp);
security/apparmor/label.c
61
struct aa_proxy *aa_alloc_proxy(struct aa_label *label, gfp_t gfp)
security/apparmor/label.c
65
new = kzalloc_obj(struct aa_proxy, gfp);
security/apparmor/label.c
868
int len, gfp_t gfp)
security/apparmor/label.c
886
new = aa_label_alloc(len, NULL, gfp);
security/apparmor/label.c
902
gfp_t gfp)
security/apparmor/label.c
909
return vec_create_and_insert_label(vec, len, gfp);
security/apparmor/lib.c
121
bool aa_resize_str_table(struct aa_str_table *t, int newsize, gfp_t gfp)
security/apparmor/lib.c
128
n = kzalloc_objs(*n, newsize, gfp);
security/apparmor/lib.c
234
__counted char *aa_str_alloc(int size, gfp_t gfp)
security/apparmor/lib.c
238
str = kmalloc_flex(*str, name, size, gfp);
security/apparmor/lib.c
480
const char *name, gfp_t gfp)
security/apparmor/lib.c
487
hname = aa_str_alloc(hname_sz, gfp);
security/apparmor/lsm.c
1069
static int apparmor_sk_alloc_security(struct sock *sk, int family, gfp_t gfp)
security/apparmor/lsm.c
85
static int apparmor_cred_alloc_blank(struct cred *cred, gfp_t gfp)
security/apparmor/lsm.c
95
gfp_t gfp)
security/apparmor/policy.c
132
struct aa_policydb *aa_alloc_pdb(gfp_t gfp)
security/apparmor/policy.c
134
struct aa_policydb *pdb = kzalloc_obj(struct aa_policydb, gfp);
security/apparmor/policy.c
298
struct aa_ruleset *aa_alloc_ruleset(gfp_t gfp)
security/apparmor/policy.c
302
rules = kzalloc_obj(*rules, gfp);
security/apparmor/policy.c
368
gfp_t gfp)
security/apparmor/policy.c
376
profile = kzalloc_flex(*profile, label.rules, 1, gfp);
security/apparmor/policy.c
380
if (!aa_policy_init(&profile->base, NULL, hname, gfp))
security/apparmor/policy.c
382
if (!aa_label_init(&profile->label, 1, gfp))
security/apparmor/policy.c
386
profile->label.rules[0] = aa_alloc_ruleset(gfp);
security/apparmor/policy.c
393
proxy = aa_alloc_proxy(&profile->label, gfp);
security/apparmor/policy.c
551
gfp_t gfp)
security/apparmor/policy.c
568
gfp);
security/apparmor/policy.c
571
profile = aa_alloc_null(parent, name, gfp);
security/apparmor/policy.c
685
gfp_t gfp)
security/apparmor/policy.c
690
profile = aa_alloc_profile(name, NULL, gfp);
security/apparmor/policy.c
733
const char *base, gfp_t gfp)
security/apparmor/policy.c
744
name = kmalloc(name_sz, gfp);
security/apparmor/policy.c
754
name = kmalloc(name_sz, gfp);
security/apparmor/policy.c
767
profile = aa_alloc_null(parent, name, gfp);
security/apparmor/secid.c
126
int aa_alloc_secid(struct aa_label *label, gfp_t gfp)
security/apparmor/secid.c
133
XA_LIMIT(AA_FIRST_SECID, INT_MAX), gfp);
security/integrity/ima/ima.h
562
void **lsmrule, gfp_t gfp)
security/integrity/ima/ima_policy.c
411
gfp_t gfp)
security/integrity/ima/ima_policy.c
420
nentry = kmemdup(entry, sizeof(*nentry), gfp);
security/integrity/ima/ima_policy.c
436
gfp);
security/landlock/cred.c
32
const struct cred *const old, const gfp_t gfp)
security/lsm.h
35
int lsm_cred_alloc(struct cred *cred, gfp_t gfp);
security/security.c
1188
int security_inode_alloc(struct inode *inode, gfp_t gfp)
security/security.c
1190
int rc = lsm_inode_alloc(inode, gfp);
security/security.c
186
static int lsm_blob_alloc(void **dest, size_t size, gfp_t gfp)
security/security.c
193
*dest = kzalloc(size, gfp);
security/security.c
208
int lsm_cred_alloc(struct cred *cred, gfp_t gfp)
security/security.c
210
return lsm_blob_alloc(&cred->security, blob_sizes.lbs_cred, gfp);
security/security.c
222
static int lsm_inode_alloc(struct inode *inode, gfp_t gfp)
security/security.c
229
inode->i_security = kmem_cache_zalloc(lsm_inode_cache, gfp);
security/security.c
2718
int security_cred_alloc_blank(struct cred *cred, gfp_t gfp)
security/security.c
2720
int rc = lsm_cred_alloc(cred, gfp);
security/security.c
2725
rc = call_int_hook(cred_alloc_blank, cred, gfp);
security/security.c
2762
int security_prepare_creds(struct cred *new, const struct cred *old, gfp_t gfp)
security/security.c
2764
int rc = lsm_cred_alloc(new, gfp);
security/security.c
2769
rc = call_int_hook(cred_prepare, new, old, gfp);
security/security.c
4374
static int lsm_sock_alloc(struct sock *sock, gfp_t gfp)
security/security.c
4376
return lsm_blob_alloc(&sock->sk_security, blob_sizes.lbs_sock, gfp);
security/security.c
4821
gfp_t gfp)
security/security.c
4823
return call_int_hook(xfrm_policy_alloc_security, ctxp, sec_ctx, gfp);
security/security.c
5110
gfp_t gfp)
security/security.c
5112
return call_int_hook(audit_rule_init, field, op, rulestr, lsmrule, gfp);
security/selinux/hooks.c
4188
gfp_t gfp)
security/selinux/include/audit.h
40
gfp_t gfp);
security/selinux/include/security.h
305
u32 *out_sid, gfp_t gfp);
security/selinux/include/security.h
307
int security_context_str_to_sid(const char *scontext, u32 *out_sid, gfp_t gfp);
security/selinux/include/xfrm.h
17
struct xfrm_user_sec_ctx *uctx, gfp_t gfp);
security/selinux/ss/services.c
1640
gfp_t gfp)
security/selinux/ss/services.c
1643
sid, SECSID_NULL, gfp, 0);
security/selinux/ss/services.c
1646
int security_context_str_to_sid(const char *scontext, u32 *sid, gfp_t gfp)
security/selinux/ss/services.c
1649
sid, gfp);
security/selinux/ss/services.c
3591
gfp_t gfp)
security/selinux/ss/services.c
3632
tmprule = kzalloc_obj(struct selinux_audit_rule, gfp);
security/selinux/xfrm.c
101
&ctx->ctx_sid, gfp);
security/selinux/xfrm.c
281
gfp_t gfp)
security/selinux/xfrm.c
283
return selinux_xfrm_alloc_user(ctxp, uctx, gfp);
security/selinux/xfrm.c
76
gfp_t gfp)
security/selinux/xfrm.c
91
ctx = kmalloc_flex(*ctx, ctx_str, str_len + 1, gfp);
security/smack/smack.h
308
gfp_t gfp);
security/smack/smack_access.c
584
smk_import_allocated_label(char *smack, gfp_t gfp)
security/smack/smack_access.c
595
skp = kzalloc_obj(*skp, gfp);
security/smack/smack_access.c
653
smk_import_valid_label(const char *label, int label_len, gfp_t gfp)
security/smack/smack_access.c
655
char *smack = kstrndup(label, label_len, gfp);
security/smack/smack_access.c
660
return smk_import_allocated_label(smack, gfp);
security/smack/smack_lsm.c
2093
static int smack_cred_alloc_blank(struct cred *cred, gfp_t gfp)
security/smack/smack_lsm.c
2130
gfp_t gfp)
security/smack/smack_lsm.c
2138
rc = smk_copy_rules(&new_tsp->smk_rules, &old_tsp->smk_rules, gfp);
security/smack/smack_lsm.c
2143
gfp);
security/smack/smack_lsm.c
342
gfp_t gfp)
security/smack/smack_lsm.c
349
nrp = kmem_cache_zalloc(smack_rule_cache, gfp);
security/smack/smack_lsm.c
369
gfp_t gfp)
security/smack/smack_lsm.c
375
nklep = kzalloc_obj(struct smack_known_list_elem, gfp);
security/smack/smack_lsm.c
4778
gfp_t gfp)
security/tomoyo/tomoyo.c
38
gfp_t gfp)
sound/core/memalloc.c
290
gfp_t gfp = GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN;
sound/core/memalloc.c
293
p = alloc_pages_exact(size, gfp);
sound/core/memalloc.c
300
if (IS_ENABLED(CONFIG_ZONE_DMA32) && !(gfp & GFP_DMA32)) {
sound/core/memalloc.c
301
gfp |= GFP_DMA32;
sound/core/memalloc.c
304
if (IS_ENABLED(CONFIG_ZONE_DMA) && !(gfp & GFP_DMA)) {
sound/core/memalloc.c
305
gfp = (gfp & ~GFP_DMA32) | GFP_DMA;
sound/virtio/virtio_card.c
31
gfp_t gfp)
sound/virtio/virtio_card.c
41
if (virtqueue_add_sgs(vqueue, psgs, 0, 1, event, gfp) || !notify)
sound/virtio/virtio_ctl_msg.c
86
size_t response_size, gfp_t gfp)
sound/virtio/virtio_ctl_msg.c
93
msg = kzalloc(sizeof(*msg) + request_size + response_size, gfp);
sound/virtio/virtio_ctl_msg.h
24
size_t response_size, gfp_t gfp);
sound/virtio/virtio_pcm.h
115
unsigned int command, gfp_t gfp);
sound/virtio/virtio_pcm_msg.c
394
unsigned int command, gfp_t gfp)
sound/virtio/virtio_pcm_msg.c
406
msg = virtsnd_ctl_msg_alloc(request_size, response_size, gfp);
tools/include/linux/sched/mm.h
4
#define might_alloc(gfp) do { } while (0)
tools/include/linux/slab.h
138
static inline void *kzalloc(size_t size, gfp_t gfp)
tools/include/linux/slab.h
140
return kmalloc(size, gfp | __GFP_ZERO);
tools/include/linux/slab.h
186
int kmem_cache_alloc_bulk(struct kmem_cache *cachep, gfp_t gfp, size_t size,
tools/include/linux/slab.h
189
kmem_cache_prefill_sheaf(struct kmem_cache *s, gfp_t gfp, unsigned int size);
tools/include/linux/slab.h
192
kmem_cache_alloc_from_sheaf(struct kmem_cache *s, gfp_t gfp,
tools/include/linux/slab.h
195
void kmem_cache_return_sheaf(struct kmem_cache *s, gfp_t gfp,
tools/include/linux/slab.h
197
int kmem_cache_refill_sheaf(struct kmem_cache *s, gfp_t gfp,
tools/include/linux/slab.h
33
void *kmalloc(size_t size, gfp_t gfp);
tools/include/linux/slab.h
35
void *kmalloc_array(size_t n, size_t size, gfp_t gfp);
tools/lib/slab.c
14
void *kmalloc(size_t size, gfp_t gfp)
tools/lib/slab.c
18
if (!(gfp & __GFP_DIRECT_RECLAIM))
tools/lib/slab.c
25
if (gfp & __GFP_ZERO)
tools/lib/slab.c
40
void *kmalloc_array(size_t n, size_t size, gfp_t gfp)
tools/lib/slab.c
44
if (!(gfp & __GFP_DIRECT_RECLAIM))
tools/lib/slab.c
51
if (gfp & __GFP_ZERO)
tools/perf/builtin-kmem.c
745
struct gfp_flag *gfp;
tools/perf/builtin-kmem.c
747
gfp = bsearch(&key, gfps, nr_gfps, sizeof(*gfps), gfpcmp);
tools/perf/builtin-kmem.c
748
if (gfp)
tools/perf/builtin-kmem.c
749
return gfp->compact_str;
tools/testing/memblock/linux/kmemleak.h
10
gfp_t gfp)
tools/testing/shared/linux.c
157
int kmem_cache_alloc_bulk(struct kmem_cache *cachep, gfp_t gfp, size_t size,
tools/testing/shared/linux.c
170
if (!(gfp & __GFP_DIRECT_RECLAIM)) {
tools/testing/shared/linux.c
186
if (!(gfp & __GFP_DIRECT_RECLAIM)) {
tools/testing/shared/linux.c
203
else if (gfp & __GFP_ZERO)
tools/testing/shared/linux.c
255
kmem_cache_prefill_sheaf(struct kmem_cache *s, gfp_t gfp, unsigned int size)
tools/testing/shared/linux.c
274
sheaf->size = kmem_cache_alloc_bulk(s, gfp, size, sheaf->objects);
tools/testing/shared/linux.c
283
int kmem_cache_refill_sheaf(struct kmem_cache *s, gfp_t gfp,
tools/testing/shared/linux.c
293
sheaf = kmem_cache_prefill_sheaf(s, gfp, size);
tools/testing/shared/linux.c
297
kmem_cache_return_sheaf(s, gfp, *sheafp);
tools/testing/shared/linux.c
302
refill = kmem_cache_alloc_bulk(s, gfp, size - sheaf->size,
tools/testing/shared/linux.c
311
void kmem_cache_return_sheaf(struct kmem_cache *s, gfp_t gfp,
tools/testing/shared/linux.c
321
kmem_cache_alloc_from_sheaf(struct kmem_cache *s, gfp_t gfp,
tools/testing/shared/linux.c
55
int gfp)
tools/testing/shared/linux.c
65
if (!(gfp & __GFP_DIRECT_RECLAIM)) {
tools/testing/shared/linux.c
94
else if (gfp & __GFP_ZERO)
tools/testing/vma/include/dup.h
1055
unsigned long start, unsigned long end, gfp_t gfp)
tools/testing/vma/include/dup.h
1058
mas_store_gfp(&vmi->mas, NULL, gfp);
tools/virtio/linux/kernel.h
119
static inline void *krealloc_array(void *p, size_t new_n, size_t new_size, gfp_t gfp)
tools/virtio/linux/kernel.h
126
return krealloc(p, bytes, gfp);
tools/virtio/linux/kernel.h
57
static inline void *kmalloc(size_t s, gfp_t gfp)
tools/virtio/linux/kernel.h
63
static inline void *kmalloc_array(unsigned n, size_t s, gfp_t gfp)
tools/virtio/linux/kernel.h
65
return kmalloc(n * s, gfp);
tools/virtio/linux/kernel.h
68
static inline void *kzalloc(size_t s, gfp_t gfp)
tools/virtio/linux/kernel.h
70
void *p = kmalloc(s, gfp);
tools/virtio/linux/kernel.h
76
static inline void *alloc_pages_exact(size_t s, gfp_t gfp)
tools/virtio/linux/kernel.h
78
return kmalloc(s, gfp);
tools/virtio/linux/kernel.h
93
static inline void *krealloc(void *p, size_t s, gfp_t gfp)
tools/virtio/linux/kernel.h
99
static inline unsigned long __get_free_page(gfp_t gfp)
tools/virtio/ringtest/ptr_ring.c
27
static void *kmalloc(unsigned size, gfp_t gfp)
tools/virtio/ringtest/ptr_ring.c
33
if (gfp & __GFP_ZERO)
virt/kvm/kvm_main.c
367
gfp_t gfp = mc->gfp_custom ? mc->gfp_custom : GFP_KERNEL_ACCOUNT;
virt/kvm/kvm_main.c
384
mc->objects = kvmalloc_array(capacity, sizeof(void *), gfp);
virt/kvm/kvm_main.c
396
obj = mmu_memory_cache_alloc_obj(mc, gfp);