GFP_DMA32
gfp_t flags32 = GFP_KERNEL | GFP_DMA32 | __GFP_ZERO;
page = alloc_page(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO);
page = alloc_page(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO);
page = alloc_pages(GFP_KERNEL | GFP_DMA32, 2);
page = alloc_page(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO);
v_lli = dma_pool_alloc(sdev->pool, GFP_DMA32 | GFP_NOWAIT, &p_lli);
v_lli = dma_pool_alloc(sdev->pool, GFP_DMA32 | GFP_NOWAIT, &p_lli);
v_lli = dma_pool_alloc(sdev->pool, GFP_DMA32 | GFP_NOWAIT, &p_lli);
unsigned long nr_populated = alloc_pages_bulk(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO,
unsigned long nr_populated = alloc_pages_bulk(GFP_KERNEL | GFP_DMA32, AR_BUFFERS, pages);
GFP_KERNEL | GFP_DMA32);
GFP_KERNEL | GFP_DMA32);
GFP_KERNEL | GFP_DMA32);
GFP_KERNEL | GFP_DMA32);
GFP_KERNEL | GFP_DMA32);
GFP_KERNEL | GFP_DMA32);
priv->shm_gfp_mask |= GFP_DMA32;
pd->p = alloc_page(GFP_DMA32);
pd->dummy_pt = alloc_page(GFP_DMA32);
pd->dummy_page = alloc_page(GFP_DMA32);
pt->p = alloc_page(GFP_DMA32);
dev_priv->scratch_page = alloc_page(GFP_DMA32 | __GFP_ZERO);
fb->sysmem.flush_page = alloc_page(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO);
gfp |= GFP_DMA32;
rdev->dummy_page.page = alloc_page(GFP_DMA32 | GFP_KERNEL | __GFP_ZERO);
glob->dummy_read_page = alloc_page(__GFP_ZERO | GFP_DMA32 |
gfp = GFP_DMA32;
gfp_flags |= GFP_DMA32;
page = alloc_pages(GFP_KERNEL | __GFP_ZERO | GFP_DMA32, order);
pci_seg->dev_table = iommu_alloc_pages_sz(GFP_KERNEL | GFP_DMA32,
#define ARM_V7S_TABLE_GFP_DMA GFP_DMA32
.gfp_flags = GFP_DMA32,
iommu_alloc_pages_sz(GFP_KERNEL | GFP_DMA32, DT_SIZE);
gfp_flags_quirk |= GFP_DMA32;
s->vidq.gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
pd->vidq.gfp_flags = GFP_DMA32;
void *pt = (void *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
vindev->queue.gfp_flags = GFP_DMA32;
voutdev->queue.gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
q->gfp_flags = GFP_DMA32;
GFP_KERNEL | GFP_DMA32);
GFP_KERNEL | GFP_DMA32);
GFP_DMA32 : 0;
vq->gfp_flags = GFP_DMA32;
gfp_t gfp = GFP_KERNEL | GFP_DMA32;
GFP_DMA32,
page = __dev_alloc_page(GFP_KERNEL | GFP_DMA32);
page = __dev_alloc_page(GFP_KERNEL | GFP_DMA32);
GFP_ATOMIC | GFP_DMA32);
gfp |= GFP_DMA32;
gfp |= GFP_DMA32;
GFP_ATOMIC | __GFP_NOWARN | GFP_DMA32);
(unsigned char *)__get_free_pages(GFP_DMA32, ordernum);
GFP_DMA32);
sfc->buffer = (u8 *)__get_free_pages(GFP_KERNEL | GFP_DMA32,
virt = (void *)__get_free_page(GFP_KERNEL | GFP_DMA32);
void *pt = (void *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
return dma_pool_zalloc(pdev->device_pool, GFP_DMA32 | GFP_ATOMIC,
GFP_DMA32 | GFP_ATOMIC,
req = (void *)__get_free_pages(GFP_KERNEL | GFP_DMA32, order);
gfpflag_string(GFP_DMA32), \
if (nid != NUMA_NO_NODE && !(gfp & (GFP_DMA | GFP_DMA32))) {
!(gfp & (GFP_DMA32 | GFP_DMA)))
gfp |= GFP_DMA32;
gfp = (gfp & ~GFP_DMA32) | GFP_DMA;
return GFP_DMA32;
GFP_KERNEL | GFP_DMA32);
GFP_KERNEL | GFP_DMA32);
if (gfp & GFP_DMA32)
if (IS_ENABLED(CONFIG_ZONE_DMA32) && (gfp & GFP_DMA32))
s->allocflags |= GFP_DMA32;
if (!(gfp_mask & (GFP_DMA | GFP_DMA32)))
#define GFP_VMALLOC32 (GFP_DMA32 | GFP_KERNEL)
#define GFP_VMALLOC32 (GFP_DMA32 | GFP_KERNEL)
q->gfp_flags = GFP_DMA32;
if (IS_ENABLED(CONFIG_ZONE_DMA32) && !(gfp & GFP_DMA32)) {
gfp |= GFP_DMA32;
gfp = (gfp & ~GFP_DMA32) | GFP_DMA;