MMU_PAGE_4K
#define mmu_virtual_psize MMU_PAGE_4K
#define pte_pagesize_index(mm, addr, pte) MMU_PAGE_4K
unsigned __split = (psize == MMU_PAGE_4K || \
(((pte) & H_PAGE_COMBO)? MMU_PAGE_4K: MMU_PAGE_64K)
if (actual_psize != MMU_PAGE_4K)
if (actual_psize == MMU_PAGE_4K)
BUG_ON(psize != MMU_PAGE_4K);
#define mmu_virtual_psize MMU_PAGE_4K
if (psize == MMU_PAGE_4K)
#define mmu_virtual_psize MMU_PAGE_4K
#define mmu_virtual_psize MMU_PAGE_4K
#define mmu_virtual_psize MMU_PAGE_4K
if (mmu_linear_psize == MMU_PAGE_4K)
pte->page_size = MMU_PAGE_4K;
gpte->page_size = MMU_PAGE_4K;
pgsize = slbe->large ? MMU_PAGE_16M : MMU_PAGE_4K;
slbe->base_page_size = MMU_PAGE_4K;
int hpsize = MMU_PAGE_4K;
add_rmmu_ap_encoding(info, MMU_PAGE_4K, &i);
for (ps = MMU_PAGE_4K; ps < MMU_PAGE_COUNT; ++ps)
MMU_PAGE_4K, MMU_PAGE_4K, old_pte);
unsigned long shift = mmu_psize_defs[MMU_PAGE_4K].shift;
if (mmu_hash_ops.hpte_updatepp(gslot, rflags, vpn, MMU_PAGE_4K,
MMU_PAGE_4K, ssize, flags) == -1)
MMU_PAGE_4K, MMU_PAGE_4K, ssize);
MMU_PAGE_4K,
MMU_PAGE_4K, ssize);
MMU_PAGE_4K, MMU_PAGE_4K,
MMU_PAGE_4K, MMU_PAGE_4K, ssize);
MMU_PAGE_4K, MMU_PAGE_4K,
MMU_PAGE_4K, MMU_PAGE_4K,
MMU_PAGE_4K, MMU_PAGE_4K, old_pte);
unsigned long shift = mmu_psize_defs[MMU_PAGE_4K].shift;
if (psize == MMU_PAGE_4K)
if (psize == MMU_PAGE_4K) {
case MMU_PAGE_4K:
size = MMU_PAGE_4K;
a_size = MMU_PAGE_4K;
case MMU_PAGE_4K:
psize = MMU_PAGE_4K;
if (mmu_psize_defs[MMU_PAGE_4K].penc[MMU_PAGE_16M] == -1)
if (mmu_linear_psize == MMU_PAGE_4K)
int mmu_linear_psize = MMU_PAGE_4K;
int mmu_virtual_psize = MMU_PAGE_4K;
int mmu_vmalloc_psize = MMU_PAGE_4K;
int mmu_io_psize = MMU_PAGE_4K;
[MMU_PAGE_4K] = {
.penc = {[MMU_PAGE_4K] = 0, [1 ... MMU_PAGE_COUNT - 1] = -1},
[MMU_PAGE_4K] = {
.penc = {[MMU_PAGE_4K] = 0, [1 ... MMU_PAGE_COUNT - 1] = -1},
if (get_slice_psize(mm, addr) == MMU_PAGE_4K)
slice_set_range_psize(mm, addr, 1, MMU_PAGE_4K);
if ((get_paca_psize(addr) != MMU_PAGE_4K) && (current->mm == mm)) {
if (psize != MMU_PAGE_4K)
psize = MMU_PAGE_4K;
psize = MMU_PAGE_4K;
psize = mmu_vmalloc_psize = MMU_PAGE_4K;
if (unlikely((psize == MMU_PAGE_4K) && subpage_protection(mm, ea)))
idx = MMU_PAGE_4K;
if (base_idx == MMU_PAGE_4K || base_idx == MMU_PAGE_64K)
int mmu_vmemmap_psize = MMU_PAGE_4K;
atomic_long_read(&direct_pages_count[MMU_PAGE_4K]) << 2);
idx = MMU_PAGE_4K;
mmu_psize_defs[MMU_PAGE_4K].shift = 12;
mmu_psize_defs[MMU_PAGE_4K].ap = 0x0;
mmu_psize_defs[MMU_PAGE_4K].h_rpt_pgsize =
psize_to_rpti_pgsize(MMU_PAGE_4K);
mmu_virtual_psize = MMU_PAGE_4K;
#define MMU_PAGE_BASE MMU_PAGE_4K
compat_maskp = slice_mask_for_size(&mm->context, MMU_PAGE_4K);
compat_maskp = slice_mask_for_size(&mm->context, MMU_PAGE_4K);
[MMU_PAGE_4K] = {
[MMU_PAGE_4K] = {
mmu_vmemmap_psize = MMU_PAGE_4K;
set_hblkrm_bloc_size(MMU_PAGE_4K, MMU_PAGE_4K, block_size);