MAX_PAGE_ORDER
pool->max_order = min(MAX_PAGE_ORDER,
chunk = (1UL << (PAGE_SHIFT + MAX_PAGE_ORDER)) /
const u64 maxblock = 1UL << (PAGE_SHIFT + MAX_PAGE_ORDER);
if (unlikely(order > MAX_PAGE_ORDER))
if (max_tsb_size > PAGE_SIZE << MAX_PAGE_ORDER)
max_tsb_size = PAGE_SIZE << MAX_PAGE_ORDER;
if (order > MAX_PAGE_ORDER)
max_order = min(MAX_PAGE_ORDER, get_order(size));
if (count > (PAGE_SIZE << MAX_PAGE_ORDER))
count = PAGE_SIZE << MAX_PAGE_ORDER;
if (count > (PAGE_SIZE << MAX_PAGE_ORDER))
count = PAGE_SIZE << MAX_PAGE_ORDER;
#define MAX_LEN (1UL << MAX_PAGE_ORDER << PAGE_SHIFT)
block_size = 1 << (PAGE_SHIFT + MAX_PAGE_ORDER < 32 ?
PAGE_SHIFT + MAX_PAGE_ORDER : 31);
int max_order = MAX_PAGE_ORDER;
GEM_BUG_ON(order > MAX_PAGE_ORDER);
.order = MAX_PAGE_ORDER + 1,
.order = MAX_PAGE_ORDER + 1,
if (params->order <= MAX_PAGE_ORDER) {
min_t(unsigned int, MAX_PAGE_ORDER,
#define TTM_SHRINKER_BATCH ((1 << (MAX_PAGE_ORDER / 2)) * NR_PAGE_ORDERS)
for (order = ttm_pool_alloc_find_order(MAX_PAGE_ORDER, alloc);
aeq_size = min(aeq_size, (u32)((PAGE_SIZE << MAX_PAGE_ORDER) /
#define Q_MAX_SZ_SHIFT (PAGE_SHIFT + MAX_PAGE_ORDER)
order_mask &= GENMASK(MAX_PAGE_ORDER, 0);
if (new_order > MAX_PAGE_ORDER) {
new_order = MAX_PAGE_ORDER;
unsigned int order = MAX_PAGE_ORDER;
order = MAX_PAGE_ORDER;
if (get_order(vsize) > MAX_PAGE_ORDER)
if (get_order(size) > MAX_PAGE_ORDER)
if (get_order(sgl->sgl_size) > MAX_PAGE_ORDER) {
if (order > MAX_PAGE_ORDER) {
#define IBMVNIC_ONE_LTB_MAX ((u32)((1 << MAX_PAGE_ORDER) * PAGE_SIZE))
if ((!mrioc->seg_tb_support && (get_order(trace_size) > MAX_PAGE_ORDER)) ||
if (get_order(fw_size) <= MAX_PAGE_ORDER) {
#define VIRTIO_BALLOON_HINT_BLOCK_ORDER MAX_PAGE_ORDER
unsigned long order = MAX_PAGE_ORDER;
if (unlikely(order > MAX_PAGE_ORDER))
return order > MAX_PAGE_ORDER;
#define MAX_FOLIO_ORDER MAX_PAGE_ORDER
#if (MAX_PAGE_ORDER + PAGE_SHIFT) > SECTION_SIZE_BITS
#define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER)
#define NR_PAGE_ORDERS (MAX_PAGE_ORDER + 1)
#define PAGE_BLOCK_MAX_ORDER MAX_PAGE_ORDER
#if (PAGE_BLOCK_MAX_ORDER > MAX_PAGE_ORDER)
#define KMALLOC_SHIFT_MAX (MAX_PAGE_ORDER + PAGE_SHIFT)
if (order > MAX_PAGE_ORDER)
order = min(get_order(pool_size), MAX_PAGE_ORDER);
if (nslabs > SLABS_PER_PAGE << MAX_PAGE_ORDER) {
nslabs = SLABS_PER_PAGE << MAX_PAGE_ORDER;
if (order > MAX_PAGE_ORDER)
order = MAX_PAGE_ORDER;
if (get_order((unsigned long)nr_pages * sizeof(void *)) > MAX_PAGE_ORDER)
if (order_base_2(size) > PAGE_SHIFT+MAX_PAGE_ORDER)
if (order > MAX_PAGE_ORDER)
order > MAX_PAGE_ORDER ? "vmalloc" : "linear");
if (order > MAX_PAGE_ORDER)
if (WARN_ON_ONCE(info.magic != KHO_PAGE_MAGIC || info.order > MAX_PAGE_ORDER))
static long max_mem = (PAGE_SIZE << MAX_PAGE_ORDER) * 2;
if (freepage_order > 0 && freepage_order <= MAX_PAGE_ORDER) {
if (order <= MAX_PAGE_ORDER) {
if ((order <= MAX_PAGE_ORDER) &&
if (order <= MAX_PAGE_ORDER) {
if (kstrtoul(buf, 10, &res) < 0 || res > MAX_PAGE_ORDER / 2) {
if (order > MAX_PAGE_ORDER) {
if (order <= MAX_PAGE_ORDER)
MAYBE_BUILD_BUG_ON(HPAGE_PMD_ORDER > MAX_PAGE_ORDER);
if (order <= MAX_PAGE_ORDER)
VM_WARN_ON(order <= MAX_PAGE_ORDER);
.order = MAX_PAGE_ORDER,
collect.order = MAX_PAGE_ORDER;
for (int i = MAX_PAGE_ORDER; i >= 0; i--) {
order = min_t(int, MAX_PAGE_ORDER, __ffs(start));
order = MAX_PAGE_ORDER;
order = min_t(int, MAX_PAGE_ORDER, __ffs(pfn));
order = MAX_PAGE_ORDER;
__free_pages_core(page, MAX_PAGE_ORDER, MEMINIT_EARLY);
} else if (get_order(size) > MAX_PAGE_ORDER || hashdist) {
if (order == MAX_PAGE_ORDER && __free_unaccepted(page))
int order = start_pfn ? __ffs(start_pfn) : MAX_PAGE_ORDER;
if (++order > MAX_PAGE_ORDER)
if (pageblock_order == MAX_PAGE_ORDER)
for (current_order = MAX_PAGE_ORDER; current_order >= min_order;
if (WARN_ON_ONCE_GFP(order > MAX_PAGE_ORDER, gfp))
return order <= MAX_PAGE_ORDER;
accept_memory(page_to_phys(page), PAGE_SIZE << MAX_PAGE_ORDER);
__free_pages_ok(page, MAX_PAGE_ORDER, FPI_TO_TAIL);
if (order >= MAX_PAGE_ORDER - 1)
while (order < MAX_PAGE_ORDER) {
if (order >= pageblock_order && order < MAX_PAGE_ORDER) {
if (freepage_order <= MAX_PAGE_ORDER)
if (freepage_order <= MAX_PAGE_ORDER)
if (order > 0 && order <= MAX_PAGE_ORDER)
return param_set_uint_minmax(val, kp, 0, MAX_PAGE_ORDER);
if (prdev->order > 0 && prdev->order <= MAX_PAGE_ORDER)
#define SHUFFLE_ORDER MAX_PAGE_ORDER
if (order <= MAX_PAGE_ORDER)
slub_max_order = min_t(unsigned int, slub_max_order, MAX_PAGE_ORDER);
unsigned int max_attempt_order = MAX_PAGE_ORDER;
BUILD_BUG_ON(MAX_PAGE_ORDER >= S8_MAX);
if (WARN_ON_ONCE(order > MAX_PAGE_ORDER))
} else if (pool->p.order > MAX_PAGE_ORDER) {
smc_order = MAX_PAGE_ORDER - cqe_size_order;
if (order > MAX_PAGE_ORDER)
for (o = 0; o < MAX_PAGE_ORDER; o++) {
static int order_stats[MAX_PAGE_ORDER][MAX_MIGRATE_TYPES];
#define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER)
#define pageblock_order MAX_PAGE_ORDER
#define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER)
#define pageblock_order MAX_PAGE_ORDER
if (ctl_len > PAGE_SIZE << MAX_PAGE_ORDER) {