__GFP_DIRECT_RECLAIM
buf = kmalloc(len, (gfp & ~__GFP_DIRECT_RECLAIM) |
return (gfp & ~(__GFP_DIRECT_RECLAIM | __GFP_IO)) |
if (likely(bvl) || !(gfp_mask & __GFP_DIRECT_RECLAIM))
gfp_mask &= ~__GFP_DIRECT_RECLAIM;
WARN_ON_ONCE(!(gfp_mask & __GFP_DIRECT_RECLAIM));
if (!(sc->gfp_mask & __GFP_DIRECT_RECLAIM))
gfp_flags &= ~__GFP_DIRECT_RECLAIM;
if (unlikely(gfp_mask & __GFP_DIRECT_RECLAIM))
gfp_mask |= __GFP_DIRECT_RECLAIM;
if (unlikely(gfp_mask & __GFP_DIRECT_RECLAIM))
if (!(gfp & __GFP_DIRECT_RECLAIM))
gfp_t flags = __GFP_NOWARN | __GFP_DIRECT_RECLAIM | __GFP_NORETRY;
mask &= ~__GFP_DIRECT_RECLAIM;
gfp &= ~__GFP_DIRECT_RECLAIM;
gfp_t gfp = (mapping_gfp_mask(mc) & ~__GFP_DIRECT_RECLAIM) |
gfp_mask &= ~__GFP_DIRECT_RECLAIM;
flags &= ~__GFP_DIRECT_RECLAIM;
if ((sc->gfp_mask & (__GFP_FS|__GFP_DIRECT_RECLAIM)) != (__GFP_FS|__GFP_DIRECT_RECLAIM))
return !!(gfp_flags & __GFP_DIRECT_RECLAIM);
#define GFP_TRANSHUGE (GFP_TRANSHUGE_LIGHT | __GFP_DIRECT_RECLAIM)
ret = idr_alloc(idr, ptr, start, end, gfp_mask & ~__GFP_DIRECT_RECLAIM);
if (gfp & __GFP_DIRECT_RECLAIM)
(gfp_mask & __GFP_DIRECT_RECLAIM))
(gfpflags & __GFP_DIRECT_RECLAIM))
(vma_madvised ? __GFP_DIRECT_RECLAIM :
(vma_madvised ? __GFP_DIRECT_RECLAIM : 0);
gfp_t gfp = gfp_mask & ~(__GFP_DIRECT_RECLAIM | __GFP_NOFAIL);
preferred_gfp &= ~(__GFP_DIRECT_RECLAIM | __GFP_NOFAIL);
if (page || !(gfp & __GFP_DIRECT_RECLAIM))
preferred_gfp &= ~(__GFP_DIRECT_RECLAIM | __GFP_NOFAIL);
if (gfp_mask & __GFP_DIRECT_RECLAIM) {
return *gfp_mask & ~(__GFP_DIRECT_RECLAIM | __GFP_IO);
if (gfp_mask & __GFP_DIRECT_RECLAIM) {
if (!in_task() || !(gfp_mask & __GFP_DIRECT_RECLAIM))
~__GFP_DIRECT_RECLAIM, order,
if (!(gfp_mask & __GFP_DIRECT_RECLAIM))
if (!(gfp_mask & __GFP_DIRECT_RECLAIM)) {
bool can_direct_reclaim = gfp_mask & __GFP_DIRECT_RECLAIM;
gfp_mask = (gfp_mask & ~__GFP_DIRECT_RECLAIM) | __GFP_COMP |
if ((alloc_gfp & __GFP_DIRECT_RECLAIM) && oo_order(oo) > oo_order(s->min))
flags &= ~__GFP_DIRECT_RECLAIM;
gfp_t large_gfp = vmalloc_gfp_adjust(gfp, large_order) & ~__GFP_DIRECT_RECLAIM;
if (!(gfp_flags & __GFP_DIRECT_RECLAIM))
page = alloc_pages((gfp_mask & ~__GFP_DIRECT_RECLAIM) |
(gfp_mask & (__GFP_DIRECT_RECLAIM | GFP_DMA))) {
(gfp_mask & (__GFP_DIRECT_RECLAIM | GFP_DMA))) {
pfrag->page = alloc_pages((gfp & ~__GFP_DIRECT_RECLAIM) |
(allocation & ~__GFP_DIRECT_RECLAIM) |
(GFP_KERNEL & ~__GFP_DIRECT_RECLAIM) |
if (gfp & __GFP_DIRECT_RECLAIM) {
bool can_wait = !!(gfp & __GFP_DIRECT_RECLAIM);
return !!(gfp_flags & __GFP_DIRECT_RECLAIM);
#define GFP_TRANSHUGE (GFP_TRANSHUGE_LIGHT | __GFP_DIRECT_RECLAIM)
if (!(gfp & __GFP_DIRECT_RECLAIM))
if (!(gfp & __GFP_DIRECT_RECLAIM))
if (!(gfp & __GFP_DIRECT_RECLAIM)) {
if (!(gfp & __GFP_DIRECT_RECLAIM)) {
list[i] = kmem_cache_alloc(test_cache, __GFP_DIRECT_RECLAIM);
kmem_cache_alloc_bulk(test_cache, __GFP_DIRECT_RECLAIM, 5, list);
list[i] = kmem_cache_alloc(test_cache, __GFP_DIRECT_RECLAIM);
kmem_cache_alloc_bulk(test_cache2, __GFP_DIRECT_RECLAIM, 10, list);
if (!(gfp & __GFP_DIRECT_RECLAIM)) {