arch/arm/kernel/topology.c
95
GFP_NOWAIT);
arch/arm/xen/p2m.c
179
p2m_entry = kzalloc_obj(*p2m_entry, GFP_NOWAIT);
arch/s390/pci/pci_clp.c
458
rrb = clp_alloc_block(GFP_NOWAIT);
arch/sparc/kernel/adi_64.c
205
mm->context.tag_store = kzalloc(size, GFP_NOWAIT);
arch/sparc/kernel/adi_64.c
284
tags = kzalloc(size, GFP_NOWAIT);
arch/um/drivers/mconsole_kern.c
90
new = kmalloc_obj(*new, GFP_NOWAIT);
block/bfq-iosched.c
5850
bfqq = kmem_cache_alloc_node(bfq_pool, GFP_NOWAIT | __GFP_ZERO,
block/blk-cgroup.c
1452
INIT_RADIX_TREE(&blkcg->blkg_tree, GFP_NOWAIT);
block/blk-cgroup.c
1615
GFP_NOWAIT);
block/blk-cgroup.c
391
new_blkg = blkg_alloc(blkcg, disk, GFP_NOWAIT);
block/blk-zoned.c
1439
gfp_mask = GFP_NOWAIT;
block/ioctl.c
900
gfp_t gfp = nowait ? GFP_NOWAIT : GFP_KERNEL;
crypto/async_tx/async_memcpy.c
43
unmap = dmaengine_get_unmap_data(device->dev, 2, GFP_NOWAIT);
crypto/async_tx/async_pq.c
190
unmap = dmaengine_get_unmap_data(device->dev, disks, GFP_NOWAIT);
crypto/async_tx/async_pq.c
312
unmap = dmaengine_get_unmap_data(device->dev, disks, GFP_NOWAIT);
crypto/async_tx/async_raid6_recov.c
31
unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT);
crypto/async_tx/async_raid6_recov.c
99
unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT);
crypto/async_tx/async_xor.c
195
unmap = dmaengine_get_unmap_data(device->dev, src_cnt+1, GFP_NOWAIT);
crypto/async_tx/async_xor.c
328
unmap = dmaengine_get_unmap_data(device->dev, src_cnt, GFP_NOWAIT);
drivers/accel/amdxdna/amdxdna_mailbox.c
159
&mb_chann->next_msgid, GFP_NOWAIT);
drivers/block/brd.c
88
gfp_t gfp = (opf & REQ_NOWAIT) ? GFP_NOWAIT : GFP_NOIO;
drivers/block/drbd/drbd_nl.c
4549
err = drbd_genl_multicast_events(msg, GFP_NOWAIT);
drivers/block/drbd/drbd_nl.c
4611
err = drbd_genl_multicast_events(skb, GFP_NOWAIT);
drivers/block/drbd/drbd_nl.c
4661
err = drbd_genl_multicast_events(skb, GFP_NOWAIT);
drivers/block/drbd/drbd_nl.c
4711
err = drbd_genl_multicast_events(skb, GFP_NOWAIT);
drivers/block/drbd/drbd_nl.c
4762
err = drbd_genl_multicast_events(skb, GFP_NOWAIT);
drivers/block/drbd/drbd_nl.c
4809
err = drbd_genl_multicast_events(skb, GFP_NOWAIT);
drivers/block/loop.c
825
worker = kzalloc_obj(struct loop_worker, GFP_NOWAIT);
drivers/block/rnbd/rnbd-srv.c
430
xa_limit_32b, GFP_NOWAIT);
drivers/block/ublk_drv.c
4176
err = idr_alloc(&ublk_index_idr, ub, i, i + 1, GFP_NOWAIT);
drivers/block/ublk_drv.c
4181
GFP_NOWAIT);
drivers/connector/cn_proc.c
111
cn_netlink_send_mult(msg, msg->len, 0, CN_IDX_PROC, GFP_NOWAIT,
drivers/crypto/ccp/ccp-dmaengine.c
324
cmd = kmem_cache_alloc(chan->ccp->dma_cmd_cache, GFP_NOWAIT);
drivers/crypto/ccp/ccp-dmaengine.c
336
desc = kmem_cache_zalloc(chan->ccp->dma_desc_cache, GFP_NOWAIT);
drivers/dca/dca-core.c
47
domain = kzalloc_obj(*domain, GFP_NOWAIT);
drivers/dca/dca-sysfs.c
45
ret = idr_alloc(&dca_idr, dca, 0, 0, GFP_NOWAIT);
drivers/dma/altera-msgdma.c
660
mdev->sw_desq = kzalloc_objs(*desc, MSGDMA_DESC_NUM, GFP_NOWAIT);
drivers/dma/altera-msgdma.c
816
mdev = devm_kzalloc(&pdev->dev, sizeof(*mdev), GFP_NOWAIT);
drivers/dma/amba-pl08x.c
1268
txd->llis_va = dma_pool_alloc(pl08x->pool, GFP_NOWAIT, &txd->llis_bus);
drivers/dma/amba-pl08x.c
1746
struct pl08x_txd *txd = kzalloc_obj(*txd, GFP_NOWAIT);
drivers/dma/amba-pl08x.c
1898
dsg = kzalloc_obj(struct pl08x_sg, GFP_NOWAIT);
drivers/dma/amba-pl08x.c
2023
dsg = kzalloc_obj(struct pl08x_sg, GFP_NOWAIT);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
322
desc = kmem_cache_zalloc(chan->pt->dma_desc_cache, GFP_NOWAIT);
drivers/dma/amd/qdma/qdma.c
772
vdesc = kzalloc_obj(*vdesc, GFP_NOWAIT);
drivers/dma/apple-admac.c
263
adtx = kzalloc_obj(*adtx, GFP_NOWAIT);
drivers/dma/arm-dma350.c
222
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/arm-dma350.c
260
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/at_hdmac.c
1019
atdma_sg->lli = dma_pool_alloc(atdma->lli_pool, GFP_NOWAIT,
drivers/dma/at_hdmac.c
1069
atdma_sg->lli = dma_pool_alloc(atdma->lli_pool, GFP_NOWAIT,
drivers/dma/at_hdmac.c
1119
vaddr = dma_pool_alloc(atdma->memset_pool, GFP_NOWAIT, &paddr);
drivers/dma/at_hdmac.c
1186
vaddr = dma_pool_alloc(atdma->memset_pool, GFP_NOWAIT, &paddr);
drivers/dma/at_hdmac.c
1304
GFP_NOWAIT,
drivers/dma/at_hdmac.c
1353
GFP_NOWAIT,
drivers/dma/at_hdmac.c
938
atdma_sg->lli = dma_pool_alloc(atdma->lli_pool, GFP_NOWAIT,
drivers/dma/at_xdmac.c
594
desc = at_xdmac_alloc_desc(&atchan->chan, GFP_NOWAIT);
drivers/dma/bcm2835-dma.c
664
GFP_NOWAIT);
drivers/dma/bcm2835-dma.c
751
period_len, GFP_NOWAIT);
drivers/dma/dma-axi-dmac.c
543
desc = kzalloc_flex(*desc, sg, num_sgs, GFP_NOWAIT);
drivers/dma/dma-jz4780.c
240
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/dma-jz4780.c
244
desc->desc = dma_pool_alloc(jzchan->desc_pool, GFP_NOWAIT,
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
296
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
300
desc->hw_desc = kzalloc_objs(*desc->hw_desc, num, GFP_NOWAIT);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
316
lli = dma_pool_zalloc(chan->desc_pool, GFP_NOWAIT, &phys);
drivers/dma/dw-edma/dw-edma-core.c
114
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/dw-edma/dw-edma-core.c
47
burst = kzalloc_obj(*burst, GFP_NOWAIT);
drivers/dma/dw-edma/dw-edma-core.c
71
chunk = kzalloc_obj(*chunk, GFP_NOWAIT);
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
102
comp_temp = kzalloc_obj(*comp_temp, GFP_NOWAIT);
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
106
dma_pool_alloc(dpaa2_chan->fd_pool, GFP_NOWAIT,
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
112
dma_pool_alloc(dpaa2_chan->fl_pool, GFP_NOWAIT,
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
118
dma_pool_alloc(dpaa2_chan->sdd_pool, GFP_NOWAIT,
drivers/dma/fsl-edma-common.c
568
fsl_desc = kzalloc_flex(*fsl_desc, tcd, sg_len, GFP_NOWAIT);
drivers/dma/fsl-edma-common.c
576
GFP_NOWAIT, &fsl_desc->tcd[i].ptcd);
drivers/dma/fsl_raid.c
293
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/fsl_raid.c
297
cf = dma_pool_alloc(re_chan->re_dev->cf_desc_pool, GFP_NOWAIT,
drivers/dma/hisi_dma.c
488
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/hsu/hsu.c
248
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/hsu/hsu.c
252
desc->sg = kzalloc_objs(*desc->sg, nents, GFP_NOWAIT);
drivers/dma/idma64.c
199
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/idma64.c
203
desc->hw = kzalloc_objs(*desc->hw, ndesc, GFP_NOWAIT);
drivers/dma/idma64.c
318
hw->lli = dma_pool_alloc(idma64c->pool, GFP_NOWAIT, &hw->llp);
drivers/dma/img-mdc-dma.c
297
mdesc = kzalloc_obj(*mdesc, GFP_NOWAIT);
drivers/dma/img-mdc-dma.c
306
curr = dma_pool_alloc(mdma->desc_pool, GFP_NOWAIT, &curr_phys);
drivers/dma/img-mdc-dma.c
385
mdesc = kzalloc_obj(*mdesc, GFP_NOWAIT);
drivers/dma/img-mdc-dma.c
400
curr = dma_pool_alloc(mdma->desc_pool, GFP_NOWAIT,
drivers/dma/img-mdc-dma.c
468
mdesc = kzalloc_obj(*mdesc, GFP_NOWAIT);
drivers/dma/img-mdc-dma.c
480
curr = dma_pool_alloc(mdma->desc_pool, GFP_NOWAIT,
drivers/dma/imx-sdma.c
1395
&sdma->bd0_phys, GFP_NOWAIT);
drivers/dma/imx-sdma.c
1421
desc->bd = dma_alloc_coherent(sdma->dev, bd_size, &desc->bd_phys, GFP_NOWAIT);
drivers/dma/imx-sdma.c
1547
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/ioat/init.c
687
GFP_NOWAIT, &ioat_chan->completion_dma);
drivers/dma/ioat/init.c
697
ring = ioat_alloc_ring(c, order, GFP_NOWAIT);
drivers/dma/k3dma.c
474
ds = kzalloc_obj(*ds, GFP_NOWAIT);
drivers/dma/k3dma.c
478
ds->desc_hw = dma_pool_zalloc(d->pool, GFP_NOWAIT, &ds->desc_hw_lli);
drivers/dma/lgm/lgm-dma.c
684
ds = kzalloc_obj(*ds, GFP_NOWAIT);
drivers/dma/lgm/lgm-dma.c
985
ds = kzalloc_obj(*ds, GFP_NOWAIT);
drivers/dma/loongson1-apb-dma.c
207
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/loongson1-apb-dma.c
255
lli = dma_pool_zalloc(chan->lli_pool, GFP_NOWAIT, &phys);
drivers/dma/loongson1-apb-dma.c
338
sgl = kmalloc_objs(*sgl, sg_len, GFP_NOWAIT);
drivers/dma/loongson2-apb-dma.c
338
desc = kzalloc_flex(*desc, sg, sg_len, GFP_NOWAIT);
drivers/dma/loongson2-apb-dma.c
350
ldma_sg->hw = dma_pool_alloc(lchan->pool, GFP_NOWAIT, &ldma_sg->llp);
drivers/dma/loongson2-apb-dma.c
403
desc = kzalloc_flex(*desc, sg, num_periods, GFP_NOWAIT);
drivers/dma/loongson2-apb-dma.c
416
ldma_sg->hw = dma_pool_alloc(lchan->pool, GFP_NOWAIT, &ldma_sg->llp);
drivers/dma/mediatek/mtk-cqdma.c
504
cvd = kzalloc_objs(*cvd, nr_vd, GFP_NOWAIT);
drivers/dma/mediatek/mtk-cqdma.c
509
cvd[i] = kzalloc_obj(*cvd[i], GFP_NOWAIT);
drivers/dma/mediatek/mtk-hsdma.c
328
&ring->tphys, GFP_NOWAIT);
drivers/dma/mediatek/mtk-hsdma.c
337
ring->cb = kzalloc_objs(*ring->cb, MTK_DMA_SIZE, GFP_NOWAIT);
drivers/dma/mediatek/mtk-hsdma.c
725
hvd = kzalloc_obj(*hvd, GFP_NOWAIT);
drivers/dma/mediatek/mtk-uart-apdma.c
354
d = kzalloc_obj(*d, GFP_NOWAIT);
drivers/dma/milbeaut-hdmac.c
268
md = kzalloc_obj(*md, GFP_NOWAIT);
drivers/dma/milbeaut-hdmac.c
272
md->sgl = kzalloc_objs(*sgl, sg_len, GFP_NOWAIT);
drivers/dma/milbeaut-xdmac.c
195
md = kzalloc_obj(*md, GFP_NOWAIT);
drivers/dma/owl-dma.c
354
lli = dma_pool_alloc(od->lli_pool, GFP_NOWAIT, &phys);
drivers/dma/owl-dma.c
881
txd = kzalloc_obj(*txd, GFP_NOWAIT);
drivers/dma/owl-dma.c
932
txd = kzalloc_obj(*txd, GFP_NOWAIT);
drivers/dma/owl-dma.c
996
txd = kzalloc_obj(*txd, GFP_NOWAIT);
drivers/dma/pxa_dma.c
744
sw_desc = kzalloc_flex(*sw_desc, hw_desc, nb_hw_desc, GFP_NOWAIT);
drivers/dma/pxa_dma.c
750
desc = dma_pool_alloc(sw_desc->desc_pool, GFP_NOWAIT, &dma);
drivers/dma/qcom/bam_dma.c
665
async_desc = kzalloc_flex(*async_desc, desc, num_alloc, GFP_NOWAIT);
drivers/dma/qcom/gpi.c
1613
new_config = krealloc(gchan->config, config->peripheral_size, GFP_NOWAIT);
drivers/dma/qcom/gpi.c
1839
gpi_desc = kzalloc_obj(*gpi_desc, GFP_NOWAIT);
drivers/dma/qcom/hidma.c
355
mdesc = kzalloc_obj(struct hidma_desc, GFP_NOWAIT);
drivers/dma/qcom/qcom_adm.c
404
async_desc = kzalloc_obj(*async_desc, GFP_NOWAIT);
drivers/dma/qcom/qcom_adm.c
417
async_desc->cpl = kzalloc(async_desc->dma_len, GFP_NOWAIT);
drivers/dma/sf-pdma/sf-pdma.c
59
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/sh/rcar-dmac.c
1257
sgl = kmalloc_objs(*sgl, sg_len, GFP_NOWAIT);
drivers/dma/sh/rcar-dmac.c
642
ret = rcar_dmac_desc_alloc(chan, GFP_NOWAIT);
drivers/dma/sh/rcar-dmac.c
712
ret = rcar_dmac_xfer_chunk_alloc(chan, GFP_NOWAIT);
drivers/dma/sh/rcar-dmac.c
752
&desc->hwdescs.dma, GFP_NOWAIT);
drivers/dma/sh/usb-dmac.c
431
desc = usb_dmac_desc_get(uchan, sg_len, GFP_NOWAIT);
drivers/dma/sprd-dma.c
904
sdesc = kzalloc_obj(*sdesc, GFP_NOWAIT);
drivers/dma/sprd-dma.c
989
sdesc = kzalloc_obj(*sdesc, GFP_NOWAIT);
drivers/dma/st_fdma.c
244
fdesc = kzalloc_flex(*fdesc, node, sg_len, GFP_NOWAIT);
drivers/dma/st_fdma.c
252
GFP_NOWAIT, &fdesc->node[i].pdesc);
drivers/dma/ste_dma40.c
2532
sg = kzalloc_objs(struct scatterlist, periods + 1, GFP_NOWAIT);
drivers/dma/ste_dma40.c
676
base = kmalloc(d40d->lli_pool.size + align, GFP_NOWAIT);
drivers/dma/ste_dma40.c
808
desc = kmem_cache_zalloc(d40c->base->desc_slab, GFP_NOWAIT);
drivers/dma/stm32/stm32-dma.c
1104
desc = kzalloc_flex(*desc, sg_req, sg_len, GFP_NOWAIT);
drivers/dma/stm32/stm32-dma.c
1216
desc = kzalloc_flex(*desc, sg_req, num_periods, GFP_NOWAIT);
drivers/dma/stm32/stm32-dma.c
1253
desc = kzalloc_flex(*desc, sg_req, num_sgs, GFP_NOWAIT);
drivers/dma/stm32/stm32-dma3.c
418
swdesc = kzalloc_flex(*swdesc, lli, count, GFP_NOWAIT);
drivers/dma/stm32/stm32-dma3.c
424
swdesc->lli[i].hwdesc = dma_pool_zalloc(chan->lli_pool, GFP_NOWAIT,
drivers/dma/stm32/stm32-mdma.c
321
desc = kzalloc_flex(*desc, node, count, GFP_NOWAIT);
drivers/dma/stm32/stm32-mdma.c
328
dma_pool_alloc(chan->desc_pool, GFP_NOWAIT,
drivers/dma/sun4i-dma.c
535
promise = kzalloc_obj(*promise, GFP_NOWAIT);
drivers/dma/sun4i-dma.c
598
promise = kzalloc_obj(*promise, GFP_NOWAIT);
drivers/dma/sun4i-dma.c
651
contract = kzalloc_obj(*contract, GFP_NOWAIT);
drivers/dma/sun6i-dma.c
681
txd = kzalloc_obj(*txd, GFP_NOWAIT);
drivers/dma/sun6i-dma.c
685
v_lli = dma_pool_alloc(sdev->pool, GFP_DMA32 | GFP_NOWAIT, &p_lli);
drivers/dma/sun6i-dma.c
739
txd = kzalloc_obj(*txd, GFP_NOWAIT);
drivers/dma/sun6i-dma.c
744
v_lli = dma_pool_alloc(sdev->pool, GFP_DMA32 | GFP_NOWAIT, &p_lli);
drivers/dma/sun6i-dma.c
822
txd = kzalloc_obj(*txd, GFP_NOWAIT);
drivers/dma/sun6i-dma.c
827
v_lli = dma_pool_alloc(sdev->pool, GFP_DMA32 | GFP_NOWAIT, &p_lli);
drivers/dma/tegra186-gpc-dma.c
1073
dma_desc = kzalloc_flex(*dma_desc, sg_req, sg_len, GFP_NOWAIT);
drivers/dma/tegra186-gpc-dma.c
1208
dma_desc = kzalloc_flex(*dma_desc, sg_req, period_count, GFP_NOWAIT);
drivers/dma/tegra186-gpc-dma.c
911
dma_desc = kzalloc_flex(*dma_desc, sg_req, 1, GFP_NOWAIT);
drivers/dma/tegra186-gpc-dma.c
980
dma_desc = kzalloc_flex(*dma_desc, sg_req, 1, GFP_NOWAIT);
drivers/dma/tegra20-apb-dma.c
285
dma_desc = kzalloc_obj(*dma_desc, GFP_NOWAIT);
drivers/dma/tegra20-apb-dma.c
324
sg_req = kzalloc_obj(*sg_req, GFP_NOWAIT);
drivers/dma/tegra210-adma.c
758
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/ti/k3-udma.c
2783
d = kzalloc(sizeof(*d) + sizeof(d->hwdesc[0]), GFP_NOWAIT);
drivers/dma/ti/k3-udma.c
2796
GFP_NOWAIT,
drivers/dma/ti/k3-udma.c
2806
GFP_NOWAIT);
drivers/dma/ti/k3-udma.c
3244
d = kzalloc_flex(*d, hwdesc, sglen, GFP_NOWAIT);
drivers/dma/ti/k3-udma.c
3268
GFP_NOWAIT,
drivers/dma/ti/k3-udma.c
3591
d = kzalloc_flex(*d, hwdesc, periods, GFP_NOWAIT);
drivers/dma/ti/k3-udma.c
3612
GFP_NOWAIT,
drivers/dma/uniphier-mdmac.c
241
md = kzalloc_obj(*md, GFP_NOWAIT);
drivers/dma/uniphier-xdmac.c
295
xd = kzalloc_flex(*xd, nodes, nr, GFP_NOWAIT);
drivers/dma/uniphier-xdmac.c
351
xd = kzalloc_flex(*xd, nodes, sg_len, GFP_NOWAIT);
drivers/dma/xgene-dma.c
495
desc = dma_pool_zalloc(chan->desc_pool, GFP_NOWAIT, &phys);
drivers/dma/xilinx/xdma.c
278
sw_desc = kzalloc_obj(*sw_desc, GFP_NOWAIT);
drivers/dma/xilinx/xdma.c
288
GFP_NOWAIT);
drivers/dma/xilinx/xdma.c
299
addr = dma_pool_alloc(chan->desc_pool, GFP_NOWAIT, &dma_addr);
drivers/dma/xilinx/xilinx_dma.c
839
desc = kzalloc_obj(*desc, GFP_NOWAIT);
drivers/dma/xilinx/xilinx_dpdma.c
638
tx_desc = kzalloc_obj(*tx_desc, GFP_NOWAIT);
drivers/firewire/core-cdev.c
520
GFP_NOWAIT);
drivers/firewire/core-cdev.c
523
xa_limit_32b, GFP_NOWAIT);
drivers/gpio/gpio-mpsse.c
521
worker = kzalloc(sizeof(*worker), GFP_NOWAIT);
drivers/gpu/drm/amd/amdgpu/amdgpu_dev_coredump.c
335
coredump = kzalloc_obj(*coredump, GFP_NOWAIT);
drivers/gpu/drm/amd/amdgpu/amdgpu_dev_coredump.c
361
dev_coredumpm(dev->dev, THIS_MODULE, coredump, 0, GFP_NOWAIT,
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
7563
r = amdgpu_sync_fence(&isolation->active, &f->finished, GFP_NOWAIT);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
549
GFP_NOWAIT);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
556
GFP_NOWAIT);
drivers/gpu/drm/drm_gem.c
1028
GFP_NOWAIT);
drivers/gpu/drm/drm_gem.c
499
ret = idr_alloc(&file_priv->object_idr, NULL, 1, 0, GFP_NOWAIT);
drivers/gpu/drm/drm_mm.c
116
node->stack = stack_depot_save(entries, n, GFP_NOWAIT);
drivers/gpu/drm/drm_modeset_lock.c
101
buf = kmalloc(PAGE_SIZE, GFP_NOWAIT | __GFP_NOWARN);
drivers/gpu/drm/drm_modeset_lock.c
91
return stack_depot_save(entries, n, GFP_NOWAIT | __GFP_NOWARN);
drivers/gpu/drm/etnaviv/etnaviv_dump.c
162
iter.start = __vmalloc(file_size, GFP_NOWAIT);
drivers/gpu/drm/etnaviv/etnaviv_dump.c
232
dev_coredumpv(gpu->dev, iter.start, iter.data - iter.start, GFP_NOWAIT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
680
resv = kzalloc_obj(*resv, GFP_NOWAIT);
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
226
rq = heartbeat_create(ce, GFP_NOWAIT | __GFP_NOWARN);
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c
288
rq = heartbeat_create(ce, GFP_NOWAIT | __GFP_NOWARN);
drivers/gpu/drm/i915/gt/intel_engine_pm.c
212
rq = __i915_request_create(ce, GFP_NOWAIT);
drivers/gpu/drm/i915/gt/intel_ggtt.c
376
rq = __i915_request_create(ce, GFP_NOWAIT | GFP_ATOMIC);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
1015
buf = kmalloc(SZ_4K, GFP_NOWAIT);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
415
ct->requests.lost_and_found[lost].stack = stack_depot_save(entries, n, GFP_NOWAIT);
drivers/gpu/drm/i915/gvt/dmabuf.c
469
ret = idr_alloc(&vgpu->object_idr, dmabuf_obj, 1, 0, GFP_NOWAIT);
drivers/gpu/drm/i915/i915_active.c
729
GFP_NOWAIT | __GFP_NOWARN);
drivers/gpu/drm/i915/i915_vma.c
521
GFP_NOWAIT |
drivers/gpu/drm/i915/intel_wakeref.c
210
buf = kmalloc(buf_size, GFP_NOWAIT);
drivers/gpu/drm/i915/intel_wakeref.h
275
ref_tracker_alloc(dir, &user, GFP_NOWAIT);
drivers/gpu/drm/i915/selftests/i915_sw_fence.c
659
err = i915_sw_fence_await_dma_fence(fence, dma, delay, GFP_NOWAIT);
drivers/gpu/drm/msm/msm_gem_submit.c
753
GFP_NOWAIT);
drivers/gpu/drm/msm/msm_gem_submit.c
766
INT_MAX, GFP_NOWAIT);
drivers/gpu/drm/qxl/qxl_cmd.c
428
idr_ret = idr_alloc(&qdev->surf_id_idr, NULL, 1, 0, GFP_NOWAIT);
drivers/gpu/drm/qxl/qxl_release.c
106
handle = idr_alloc(&qdev->release_idr, release, 1, 0, GFP_NOWAIT);
drivers/gpu/drm/vmwgfx/ttm_object.c
203
ret = idr_alloc(&tdev->idr, base, 1, 0, GFP_NOWAIT);
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
196
ret = idr_alloc(idr, res, 1, 0, GFP_NOWAIT);
drivers/gpu/drm/xe/xe_guc_ct.c
109
ct->fast_req[slot].stack = stack_depot_save(entries, n, GFP_NOWAIT);
drivers/gpu/drm/xe/xe_guc_ct.c
117
char *buf __cleanup(kfree) = kmalloc(SZ_4K, GFP_NOWAIT);
drivers/gpu/drm/xe/xe_guc_relay.c
237
txn = mempool_alloc(&relay->pool, incoming ? GFP_ATOMIC : GFP_NOWAIT);
drivers/gpu/drm/xe/xe_guc_submit.c
455
q->guc->id + i, q, GFP_NOWAIT));
drivers/gpu/drm/xe/xe_lrc.c
2272
struct xe_lrc_snapshot *snapshot = kmalloc_obj(*snapshot, GFP_NOWAIT);
drivers/gpu/drm/xe/xe_vm.c
4141
snap = kvzalloc(offsetof(struct xe_vm_snapshot, snap[num_snaps]), GFP_NOWAIT);
drivers/gpu/host1x/job.c
554
GFP_NOWAIT);
drivers/infiniband/core/sa_query.c
883
GFP_NOWAIT;
drivers/infiniband/hw/hfi1/netdev_rx.c
442
return xa_insert(&rx->dev_tbl, id, data, GFP_NOWAIT);
drivers/irqchip/irq-gic-v3-its.c
2337
page = its_allocate_prop_table(GFP_NOWAIT);
drivers/irqchip/irq-gic-v3-its.c
3101
pend_page = its_allocate_pending_table(GFP_NOWAIT);
drivers/md/bcache/bset.c
1241
struct bset *out = (void *) __get_free_pages(__GFP_NOWARN|GFP_NOWAIT,
drivers/md/bcache/btree.c
375
if (!bch_bio_alloc_pages(b->bio, GFP_NOWAIT)) {
drivers/md/bcache/request.c
931
0, GFP_NOWAIT, &dc->disk.bio_split);
drivers/md/dm-bufio.c
1345
bio = bio_kmalloc(1, GFP_NOWAIT);
drivers/md/dm-bufio.c
1622
b = alloc_buffer(c, GFP_NOWAIT | __GFP_NORETRY | __GFP_NOMEMALLOC);
drivers/md/dm-cache-background-tracker.c
164
return kmem_cache_alloc(btracker_work_cache, GFP_NOWAIT);
drivers/md/dm-crypt.c
1670
gfp_t gfp_mask = GFP_NOWAIT | __GFP_HIGHMEM;
drivers/md/dm-crypt.c
1858
#define CRYPT_MAP_READ_GFP GFP_NOWAIT
drivers/md/dm-pcache/cache_key.c
406
return cache_key_alloc(ctx->cache_tree, GFP_NOWAIT);
drivers/md/dm-pcache/cache_req.c
301
return cache_miss_req_alloc(cache, pcache_req, GFP_NOWAIT);
drivers/md/dm-vdo/flush.c
105
if ((gfp_mask & GFP_NOWAIT) == GFP_NOWAIT) {
drivers/md/dm-vdo/flush.c
398
struct vdo_flush *flush = mempool_alloc(vdo->flusher->flush_pool, GFP_NOWAIT);
drivers/md/dm-vdo/memory-alloc.c
307
void *p = kmalloc(size, GFP_NOWAIT | __GFP_ZERO);
drivers/md/dm-verity-fec.c
318
fio->bufs[n] = kmem_cache_alloc(f->cache, GFP_NOWAIT);
drivers/md/dm.c
1483
try ? GFP_NOIO : GFP_NOWAIT);
drivers/md/dm.c
1754
clone = alloc_tio(ci, ti, 0, &len, GFP_NOWAIT);
drivers/md/dm.c
1993
io = alloc_io(md, bio, GFP_NOWAIT);
drivers/md/dm.c
2182
r = idr_alloc(&_minor_idr, MINOR_ALLOCED, minor, minor + 1, GFP_NOWAIT);
drivers/md/dm.c
2198
r = idr_alloc(&_minor_idr, MINOR_ALLOCED, 0, 1 << MINORBITS, GFP_NOWAIT);
drivers/md/raid5-cache.c
3107
INIT_RADIX_TREE(&log->big_stripe_tree, GFP_NOWAIT);
drivers/md/raid5-ppl.c
240
io = mempool_alloc(&ppl_conf->io_pool, GFP_NOWAIT);
drivers/memstick/core/memstick.c
530
rc = idr_alloc(&memstick_host_idr, host, 0, 0, GFP_NOWAIT);
drivers/misc/c2port/core.c
921
ret = idr_alloc(&c2port_idr, c2dev, 0, 0, GFP_NOWAIT);
drivers/misc/cardreader/rtsx_pcr.c
1511
ret = idr_alloc(&rtsx_pci_idr, pcr, 0, 0, GFP_NOWAIT);
drivers/misc/tifm_core.c
197
rc = idr_alloc(&tifm_adapter_idr, fm, 0, 0, GFP_NOWAIT);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
689
GFP_NOWAIT);
drivers/ntb/ntb_transport.c
1533
unmap = dmaengine_get_unmap_data(device->dev, 2, GFP_NOWAIT);
drivers/ntb/ntb_transport.c
1867
unmap = dmaengine_get_unmap_data(device->dev, 1, GFP_NOWAIT);
drivers/ntb/test/ntb_perf.c
812
unmap = dmaengine_get_unmap_data(dma_dev, 1, GFP_NOWAIT);
drivers/nvdimm/badrange.c
155
new_len, GFP_NOWAIT);
drivers/pci/controller/pci-hyperv.c
2842
dr_wrk = kzalloc_obj(*dr_wrk, GFP_NOWAIT);
drivers/pci/controller/pci-hyperv.c
2882
dr = kzalloc_flex(*dr, func, relations->device_count, GFP_NOWAIT);
drivers/pci/controller/pci-hyperv.c
2916
dr = kzalloc_flex(*dr, func, relations->device_count, GFP_NOWAIT);
drivers/rapidio/rio_cm.c
1302
id = idr_alloc_cyclic(&ch_idr, ch, start, end, GFP_NOWAIT);
drivers/scsi/ch.c
912
ret = idr_alloc(&ch_index_idr, ch, 0, CH_MAX_DEVS + 1, GFP_NOWAIT);
drivers/scsi/sg.c
1449
error = idr_alloc(&sg_index_idr, sdp, 0, SG_MAX_DEVS, GFP_NOWAIT);
drivers/scsi/st.c
4450
error = idr_alloc(&st_index_idr, tpnt, 0, ST_MAX_TAPES + 1, GFP_NOWAIT);
drivers/sh/intc/core.c
207
d = kzalloc_obj(*d, GFP_NOWAIT);
drivers/sh/intc/core.c
220
d->window = kzalloc_objs(*d->window, d->nr_windows, GFP_NOWAIT);
drivers/sh/intc/core.c
246
d->reg = kcalloc(d->nr_reg, sizeof(*d->reg), GFP_NOWAIT);
drivers/sh/intc/core.c
251
d->smp = kcalloc(d->nr_reg, sizeof(*d->smp), GFP_NOWAIT);
drivers/sh/intc/core.c
269
d->prio = kzalloc_objs(*d->prio, hw->nr_vectors, GFP_NOWAIT);
drivers/sh/intc/core.c
284
d->sense = kzalloc_objs(*d->sense, hw->nr_vectors, GFP_NOWAIT);
drivers/sh/intc/virq.c
171
entry = kmalloc_obj(*entry, GFP_NOWAIT);
drivers/staging/nvec/nvec_power.c
380
power = devm_kzalloc(&pdev->dev, sizeof(struct nvec_power), GFP_NOWAIT);
drivers/target/iscsi/iscsi_target.c
142
ret = idr_alloc(&tiqn_idr, NULL, 0, 0, GFP_NOWAIT);
drivers/target/target_core_user.c
1076
GFP_NOWAIT) < 0) {
drivers/tty/serial/cpm_uart.c
898
mem_addr = kzalloc(memsz, GFP_NOWAIT);
drivers/tty/vt/vt.c
3807
GFP_NOWAIT);
drivers/tty/vt/vt.c
3812
vc->vc_screenbuf = kzalloc(vc->vc_screenbuf_size, GFP_NOWAIT);
drivers/usb/host/xhci-hub.c
489
command = xhci_alloc_command(xhci, false, GFP_NOWAIT);
drivers/usb/host/xhci.c
3373
stop_cmd = xhci_alloc_command(xhci, true, GFP_NOWAIT);
drivers/usb/host/xhci.c
3377
cfg_cmd = xhci_alloc_command_with_ctx(xhci, true, GFP_NOWAIT);
drivers/virtio/virtio_balloon.c
208
err = virtqueue_add_inbuf(vq, sg, nents, vb, GFP_NOWAIT);
fs/btrfs/extent-io-tree.c
596
*mask = (*bits & EXTENT_NOWAIT ? GFP_NOWAIT : GFP_NOFS);
fs/btrfs/file.c
845
gfp |= GFP_NOWAIT;
fs/buffer.c
1468
GFP_NOWAIT | __GFP_MOVABLE);
fs/ceph/addr.c
1254
gfp_t gfp_flags = ceph_wbc->locked_pages ? GFP_NOWAIT : GFP_NOFS;
fs/crypto/bio.c
172
GFP_NOWAIT);
fs/erofs/zdata.c
1312
GFP_NOWAIT | __GFP_NORETRY
fs/ext4/page-io.c
557
gfp_flags = GFP_NOWAIT;
fs/ext4/super.c
281
sb->s_blocksize, GFP_NOWAIT);
fs/fs-writeback.c
1176
work = kzalloc_obj(*work, GFP_NOWAIT);
fs/iomap/buffered-io.c
231
gfp = GFP_NOWAIT;
fs/nfs/nfs4client.c
39
ret = idr_alloc(&nn->cb_ident_idr, clp, 1, 0, GFP_NOWAIT);
fs/nfs/nfs4session.c
175
return nfs4_find_or_create_slot(tbl, slotid, 0, GFP_NOWAIT);
fs/nfs/nfs4session.c
252
ret = nfs4_find_or_create_slot(tbl, slotid, 1, GFP_NOWAIT);
fs/nfs/write.c
100
p = mempool_alloc(nfs_wdata_mempool, GFP_NOWAIT);
fs/nfs/write.c
78
p = mempool_alloc(nfs_commit_mempool, GFP_NOWAIT);
fs/nfsd/nfs4state.c
4509
GFP_NOWAIT);
fs/nfsd/nfs4state.c
4517
GFP_NOWAIT))) {
fs/nfsd/nfs4state.c
922
new_id = idr_alloc_cyclic(&cl->cl_stateids, stid, 1, 0, GFP_NOWAIT);
fs/nfsd/nfs4state.c
956
new_id = idr_alloc_cyclic(&nn->s2s_cp_stateids, stid, 0, 0, GFP_NOWAIT);
fs/notify/inotify/inotify_user.c
402
ret = idr_alloc_cyclic(idr, i_mark, 1, 0, GFP_NOWAIT);
fs/ntfs3/namei.c
409
uni = kmalloc(PATH_MAX, GFP_NOWAIT);
fs/ntfs3/namei.c
470
uni1 = kmalloc(PATH_MAX, GFP_NOWAIT);
fs/smb/server/vfs_cache.c
711
ret = idr_alloc_cyclic(ft->idr, fp, 0, INT_MAX - 1, GFP_NOWAIT);
include/trace/events/mmflags.h
99
gfpflag_string(GFP_NOWAIT), \
io_uring/futex.c
287
ifd = io_cache_alloc(&ctx->futex_cache, GFP_NOWAIT);
io_uring/io_uring.c
807
ocqe = io_alloc_ocqe(ctx, cqe, big_cqe, GFP_NOWAIT);
io_uring/napi.c
60
e = kmalloc(sizeof(*e), GFP_NOWAIT);
ipc/util.c
231
GFP_NOWAIT);
ipc/util.c
256
0, GFP_NOWAIT);
kernel/bpf/devmap.c
892
GFP_NOWAIT,
kernel/bpf/local_storage.c
168
__GFP_ZERO | GFP_NOWAIT,
kernel/bpf/memalloc.c
215
gfp |= atomic ? GFP_NOWAIT : GFP_KERNEL;
kernel/cgroup/dmem.c
362
pool = kzalloc_obj(*pool, GFP_NOWAIT);
kernel/dma/swiotlb.c
1227
GFP_NOWAIT);
kernel/events/uprobes.c
1223
GFP_NOWAIT | __GFP_NOMEMALLOC);
kernel/irq/irqdesc.c
44
zalloc_cpumask_var(&irq_default_affinity, GFP_NOWAIT);
kernel/rcu/rcutorture.c
2465
struct rcu_head *rhp = kmalloc_obj(*rhp, GFP_NOWAIT);
kernel/sched/core.c
8623
ptr = (unsigned long)kzalloc(ptr, GFP_NOWAIT);
kernel/sched/fair.c
14054
zalloc_cpumask_var(&nohz.idle_cpus_mask, GFP_NOWAIT);
kernel/time/tick-broadcast.c
1233
zalloc_cpumask_var(&tick_broadcast_mask, GFP_NOWAIT);
kernel/time/tick-broadcast.c
1234
zalloc_cpumask_var(&tick_broadcast_on, GFP_NOWAIT);
kernel/time/tick-broadcast.c
1235
zalloc_cpumask_var(&tmpmask, GFP_NOWAIT);
kernel/time/tick-broadcast.c
1237
zalloc_cpumask_var(&tick_broadcast_oneshot_mask, GFP_NOWAIT);
kernel/time/tick-broadcast.c
1238
zalloc_cpumask_var(&tick_broadcast_pending_mask, GFP_NOWAIT);
kernel/time/tick-broadcast.c
1239
zalloc_cpumask_var(&tick_broadcast_force_mask, GFP_NOWAIT);
kernel/trace/pid_list.c
362
chunk = kzalloc_obj(*chunk, GFP_NOWAIT);
kernel/trace/pid_list.c
373
chunk = kzalloc_obj(*chunk, GFP_NOWAIT);
kernel/trace/trace_events_user.c
499
fault = kmem_cache_zalloc(fault_cache, GFP_NOWAIT);
kernel/trace/trace_events_user.c
640
enabler = kzalloc_obj(*enabler, GFP_NOWAIT | __GFP_ACCOUNT);
lib/idr.c
420
bitmap = kzalloc_obj(*bitmap, GFP_NOWAIT);
lib/idr.c
447
bitmap = kzalloc_obj(*bitmap, GFP_NOWAIT);
lib/maple_tree.c
1086
ret = kmem_cache_alloc_from_sheaf(maple_node_cache, GFP_NOWAIT, mas->sheaf);
lib/maple_tree.c
184
kmem_cache_return_sheaf(maple_node_cache, GFP_NOWAIT, sheaf);
lib/maple_tree.c
3952
mas_alloc_nodes(mas, GFP_NOWAIT);
lib/maple_tree.c
5098
mas_alloc_nodes(mas, GFP_NOWAIT);
lib/ref_tracker.c
161
sbuf = kmalloc(STACK_BUF_SIZE, GFP_NOWAIT);
lib/ref_tracker.c
308
GFP_NOWAIT);
lib/ref_tracker.c
77
stats = kmalloc_flex(*stats, stacks, limit, GFP_NOWAIT);
lib/test_vmalloc.c
297
gfp_t gfp = use_atomic ? GFP_ATOMIC : GFP_NOWAIT;
lib/test_xarray.c
495
XA_BUG_ON(xa, xa_store_index(xa, 12345678, GFP_NOWAIT) != NULL);
lib/test_xarray.c
503
xa_mk_value(12345678), GFP_NOWAIT) != NULL);
lib/test_xarray.c
88
XA_BUG_ON(xa, xa_err(xa_store_index(xa, 0, GFP_NOWAIT)) != 0);
lib/test_xarray.c
92
XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_NOWAIT)) != -ENOMEM);
lib/test_xarray.c
93
XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_NOWAIT)) != -ENOMEM);
lib/xarray.c
1182
gfp_t gfp = GFP_NOWAIT;
lib/xarray.c
373
gfp_t gfp = GFP_NOWAIT;
mm/damon/ops-common.c
312
__GFP_NOMEMALLOC | GFP_NOWAIT,
mm/filemap.c
1994
gfp |= GFP_NOWAIT;
mm/kmemleak.c
656
trace_handle = stack_depot_save(entries, nr_entries, GFP_NOWAIT);
mm/memblock.c
1488
void *vaddr = kzalloc_node(size, GFP_NOWAIT, nid);
mm/mempool.c
608
mempool_alloc_from_pool(pool, &element, 1, 0, GFP_NOWAIT);
mm/mmu_gather.c
368
*batch = (struct mmu_table_batch *)__get_free_page(GFP_NOWAIT);
mm/mmu_gather.c
37
batch = (void *)__get_free_page(GFP_NOWAIT);
mm/slab_common.c
730
struct kmem_cache *s = kmem_cache_zalloc(kmem_cache, GFP_NOWAIT);
mm/slub.c
2637
delayed_free = kmalloc_obj(*delayed_free, GFP_NOWAIT);
mm/slub.c
4405
pc.flags = GFP_NOWAIT | __GFP_THISNODE;
mm/slub.c
5722
empty = alloc_empty_sheaf(s, GFP_NOWAIT);
mm/slub.c
5909
empty = alloc_empty_sheaf(s, GFP_NOWAIT);
mm/slub.c
7521
slab = new_slab(kmem_cache_node, GFP_NOWAIT, node);
mm/slub.c
8274
struct kmem_cache *s = kmem_cache_zalloc(kmem_cache, GFP_NOWAIT);
mm/vmalloc.c
1777
lva = kmem_cache_alloc(vmap_area_cachep, GFP_NOWAIT);
mm/vmalloc.c
3942
#define GFP_VMALLOC_SUPPORTED (GFP_KERNEL | GFP_ATOMIC | GFP_NOWAIT |\
mm/vmalloc.c
5319
free = kmem_cache_zalloc(vmap_area_cachep, GFP_NOWAIT);
mm/vmalloc.c
5334
free = kmem_cache_zalloc(vmap_area_cachep, GFP_NOWAIT);
mm/vmalloc.c
5369
vn = kmalloc_objs(*vn, n, GFP_NOWAIT);
mm/vmalloc.c
5458
va = kmem_cache_zalloc(vmap_area_cachep, GFP_NOWAIT);
mm/vmscan.c
1036
__GFP_NOMEMALLOC | GFP_NOWAIT,
mm/zswap.c
907
gfp = GFP_NOWAIT | __GFP_NORETRY | __GFP_HIGHMEM | __GFP_MOVABLE;
net/9p/client.c
196
GFP_NOWAIT);
net/9p/client.c
198
tag = idr_alloc(&c->reqs, req, 0, P9_NOTAG, GFP_NOWAIT);
net/9p/client.c
745
GFP_NOWAIT);
net/core/net-sysfs.c
1730
mask = bitmap_zalloc(nr_ids, GFP_NOWAIT);
net/core/pktgen.c
2911
skb = __alloc_skb(NET_SKB_PAD + size, GFP_NOWAIT, 0, node);
net/core/pktgen.c
2917
skb = __netdev_alloc_skb(dev, size, GFP_NOWAIT);
net/ipv4/tcp_cdg.c
382
GFP_NOWAIT);
net/ipv4/tcp_ipv4.c
3240
err = bpf_iter_tcp_realloc_batch(iter, expected, GFP_NOWAIT);
net/ipv4/udp.c
3590
GFP_NOWAIT);
net/ipv4/udp.c
3985
if (flags != GFP_NOWAIT)
net/openvswitch/flow.c
96
GFP_NOWAIT |
net/rds/ib_recv.c
1037
rds_ib_recv_refill(conn, 0, GFP_NOWAIT);
net/rds/message.c
381
rm = rds_message_alloc(extra_bytes, GFP_NOWAIT);
net/sched/cls_api.c
997
GFP_NOWAIT);
net/sctp/associola.c
1622
GFP_NOWAIT);
net/sunrpc/auth_unix.c
52
ret = mempool_alloc(unix_pool, GFP_NOWAIT);
net/sunrpc/sched.c
1065
buf = mempool_alloc(rpc_buffer_mempool, GFP_NOWAIT);
net/sunrpc/sched.c
1134
return mempool_alloc(rpc_task_mempool, GFP_NOWAIT);
net/sunrpc/socklib.c
89
*ppage = alloc_page(GFP_NOWAIT);
net/sunrpc/xprtrdma/rpc_rdma.c
193
*ppages = alloc_page(GFP_NOWAIT);
rust/bindings/bindings_helper.h
111
const gfp_t RUST_CONST_HELPER_GFP_NOWAIT = GFP_NOWAIT;
samples/livepatch/livepatch-shadow-fix2.c
51
sizeof(*shadow_count), GFP_NOWAIT,
security/selinux/avc.c
296
xpd_node = kmem_cache_zalloc(avc_xperms_decision_cachep, GFP_NOWAIT);
security/selinux/avc.c
303
GFP_NOWAIT);
security/selinux/avc.c
309
GFP_NOWAIT);
security/selinux/avc.c
315
GFP_NOWAIT);
security/selinux/avc.c
343
xp_node = kmem_cache_zalloc(avc_xperms_cachep, GFP_NOWAIT);
security/selinux/avc.c
498
node = kmem_cache_zalloc(avc_node_cachep, GFP_NOWAIT);
sound/soc/bcm/bcm63xx-pcm-whistler.c
51
dma_desc = kzalloc_obj(*dma_desc, GFP_NOWAIT);
tools/testing/radix-tree/idr-test.c
159
assert(idr_alloc(&idr, item, i, i + 1, GFP_NOWAIT) == i);
tools/testing/radix-tree/idr-test.c
466
id = ida_alloc_min(&ida, 256, GFP_NOWAIT);
tools/testing/radix-tree/idr-test.c
468
id = ida_alloc_min(&ida, 1UL << 30, GFP_NOWAIT);
tools/testing/radix-tree/idr-test.c
482
int id = ida_alloc(&ida, GFP_NOWAIT);
tools/testing/radix-tree/maple.c
35206
MT_BUG_ON(mt, mas_preallocate(&mas, ptr, GFP_KERNEL & GFP_NOWAIT) == 0);
tools/testing/radix-tree/maple.c
35229
MT_BUG_ON(mt, mas_preallocate(&mas, ptr, GFP_KERNEL & GFP_NOWAIT) == 0);
tools/testing/radix-tree/maple.c
35837
ret = mtree_dup(mt, &new, GFP_NOWAIT);
tools/testing/radix-tree/maple.c
35865
ret = mtree_dup(mt, &new, GFP_NOWAIT);
virt/kvm/async_pf.c
195
work = kmem_cache_zalloc(async_pf_cache, GFP_NOWAIT);