page_order
unsigned long page_order;
kvm_loongarch_ops->page_order = order;
free_pages(addr, kvm_loongarch_ops->page_order);
unsigned long page_order = get_order(page_size);
free_pages((unsigned long)p, page_order);
unsigned long page_order = get_order(page_size);
nr_pages = 1 << page_order;
free_pages((unsigned long)(__va(addr)), page_order);
int page_order = IOMMU_PAGE_SHIFT_4K;
page_order = simple_strtol(value, NULL, 10);
CMO_PageSize = 1 << page_order;
unsigned long page_order = PAGE_SHIFT;
rc = gstage_level_to_page_order(level, &page_order);
*out_pgsize = BIT(page_order);
size_t page_order;
static inline struct idal_buffer *idal_buffer_alloc(size_t size, int page_order)
nr_chunks = (PAGE_SIZE << page_order) >> IDA_SIZE_SHIFT;
ib->page_order = page_order;
vaddr = (void *)__get_free_pages(GFP_KERNEL, page_order);
free_pages((unsigned long)vaddr, ib->page_order);
nr_chunks = (PAGE_SIZE << ib->page_order) >> IDA_SIZE_SHIFT;
free_pages((unsigned long)vaddr, ib->page_order);
static inline struct idal_buffer **idal_buffer_array_alloc(size_t size, int page_order)
ibs[i] = idal_buffer_alloc(ib_size, page_order);
if (ib->size > (PAGE_SIZE << ib->page_order))
nr_pages = 1U << map_data->page_order;
int page_order;
int page_order;
int page_order;
int page_order;
page_count = 1 << A_SIZE_32(agp_bridge->current_size)->page_order;
aper_size->page_order = __ffs(aper_size->num_entries / 1024);
set_memory_wb((unsigned long)bridge->gatt_table, 1 << page_order);
table_end = table + ((PAGE_SIZE * (1 << page_order)) - 1);
free_gatt_pages(bridge->gatt_table_real, page_order);
int page_order;
page_order = num_entries = 0;
page_order =
A_SIZE_8(temp)->page_order;
page_order = A_SIZE_16(temp)->page_order;
page_order = A_SIZE_32(temp)->page_order;
page_order = num_entries = 0;
table = alloc_gatt_pages(page_order);
page_order = ((struct aper_size_info_fixed *) temp)->page_order;
table = alloc_gatt_pages(page_order);
table_end = table + ((PAGE_SIZE * (1 << page_order)) - 1);
if (set_memory_uc((unsigned long)table, 1 << page_order))
(PAGE_SIZE * (1 << page_order)));
free_gatt_pages(table, page_order);
int page_order;
page_order = A_SIZE_8(temp)->page_order;
page_order = A_SIZE_16(temp)->page_order;
page_order = A_SIZE_32(temp)->page_order;
page_order = A_SIZE_FIX(temp)->page_order;
page_order = 0;
int page_order;
size = page_order = num_entries = 0;
page_order = A_SIZE_32(temp)->page_order;
table = (char *) __get_free_pages(GFP_KERNEL, page_order);
uninorth_priv.pages_arr = kmalloc_objs(struct page *, 1 << page_order);
table_end = table + ((PAGE_SIZE * (1 << page_order)) - 1);
bridge->gatt_table = vmap(uninorth_priv.pages_arr, (1 << page_order), 0, PAGE_KERNEL_NCG);
free_pages((unsigned long)table, page_order);
int page_order;
page_order = A_SIZE_32(temp)->page_order;
table_end = table + ((PAGE_SIZE * (1 << page_order)) - 1);
free_pages((unsigned long) bridge->gatt_table_real, page_order);
unsigned int page_order;
page_order = folio_order(page_folio(page));
if (page_order != PMD_ORDER)
return 1 << page_order;
unsigned int page_order)
state->page_order = page_order;
state->crit_factor = int_sqrt(1 << page_order);
return mempool_init_page_pool(&state->pool, 1, page_order);
BUG_ON(order > state->page_order);
order = state->page_order;
if (!start && order == b->page_order) {
size_t order = b->page_order, keys = 0;
__btree_sort(b, iter, 0, b->page_order, true, state);
return PAGE_SIZE << b->page_order;
free_pages((unsigned long) t->data, b->page_order);
unsigned int page_order,
b->page_order = page_order;
t->data = (void *) __get_free_pages(__GFP_COMP|gfp, b->page_order);
PAGE_SIZE << b->page_order);
uint8_t page_order;
BUG_ON((PAGE_SIZE << b->page_order) <
return ((PAGE_SIZE << b->page_order) -
int bch_btree_keys_alloc(struct btree_keys *b, unsigned int page_order,
unsigned int page_order;
unsigned int page_order);
if (b->keys.page_order < min_order)
ret += 1 << (b->keys.page_order + PAGE_SHIFT);
self->page_order = AQ_CFG_XDP_PAGEORDER;
self->page_order = fls(self->frame_max / PAGE_SIZE +
if (aq_nic_cfg->rxpageorder > self->page_order)
self->page_order = aq_nic_cfg->rxpageorder;
unsigned int order = rx_ring->page_order;
unsigned int order = self->page_order;
u16 page_order;
u32 page_order;
page_order = HMA_PAGE_ORDER;
sgt->orig_nents = (hma_size << 20) / (page_size << page_order);
__GFP_ZERO, page_order);
sg_set_page(iter, newpage, page_size << page_order, 0);
((page_size << page_order) >> 12));
.order = rx->page_order,
lan966x->rx.page_order = round_up(new_mtu, PAGE_SIZE) / PAGE_SIZE - 1;
lan966x->rx.fdma.db_size = PAGE_SIZE << lan966x->rx.page_order;
lan966x->tx.fdma.db_size = PAGE_SIZE << lan966x->rx.page_order;
u8 page_order;
xdp_init_buff(&xdp, PAGE_SIZE << lan966x->rx.page_order,
u8 page_order;
cp->page_order = 0;
cp->page_order = CAS_JUMBO_PAGE_SHIFT - PAGE_SHIFT;
cp->page_size = (PAGE_SIZE << cp->page_order);
__free_pages(page->buffer, cp->page_order);
page->buffer = alloc_pages(flags, cp->page_order);
int page_order;
int page_order;
length = 1 << (PAGE_SHIFT + rsv_schp->page_order);
length = 1 << (PAGE_SHIFT + rsv_schp->page_order);
md->page_order = req_schp->page_order;
schp->page_order = order;
__free_pages(schp->pages[k], schp->page_order);
num = 1 << (PAGE_SHIFT + schp->page_order);
num = 1 << (PAGE_SHIFT + rsv_schp->page_order);
req_schp->page_order = rsv_schp->page_order;
req_schp->page_order = 0;
mdata->page_order = 0;
mdata->page_order = STp->buffer->reserved_page_order;
DIV_ROUND_UP(bytes, PAGE_SIZE << mdata->page_order);
unsigned short page_order;
return find_vm_area(addr)->page_order > 0;
unsigned int page_order;
unsigned int page_order = get_order(size);
if (!dma_release_from_global_coherent(page_order, cpu_addr))
return rb->page_order;
return rb->nr_pages << page_order(rb);
return rb->nr_pages << (PAGE_SHIFT + page_order(rb));
handle->size = PAGE_SIZE << page_order(rb); \
int page_order; /* allocation order */
page_shift = PAGE_SHIFT + page_order(rb);
rb->page_order = ilog2(nr_pages);
int page_order = buddy_order(page_head);
if (PageBuddy(page_head) && page_order >= order) {
del_page_from_free_list(page_head, zone, page_order,
page_order, migratetype);
return vm->page_order;
vm->page_order = order;
unsigned int page_order = vm_area_page_order(area);
for (i = 0; i < area->nr_pages; i += 1U << page_order) {
page_size = PAGE_SIZE << page_order;
unsigned int page_order;
page_order = vm_area_page_order(area);
vmalloc_gfp_adjust(gfp_mask, page_order), node,
page_order, nr_small_pages, area->pages);
if (!fatal_signal_pending(current) && page_order == 0)