PGDIR_SIZE
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
#define VMALLOC_START (-2*PGDIR_SIZE)
#define VMALLOC_END (-PGDIR_SIZE)
#define FIXMAP_SIZE PGDIR_SIZE
#define PKMAP_SIZE PGDIR_SIZE
num = USER_PTRS_PER_PGD + USER_KERNEL_GUTTER / PGDIR_SIZE;
num2 = VMALLOC_SIZE / PGDIR_SIZE;
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD (PAGE_OFFSET / PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
memcpy(dst_pgd, src_pgd, sizeof(pgd_t) * (IO_SIZE / PGDIR_SIZE));
memcpy(dst_pgd, src_pgd, sizeof(pgd_t) * (EASI_SIZE / PGDIR_SIZE));
addr = start + i * PGDIR_SIZE;
end = ALIGN(end, PGDIR_SIZE);
#define PGDIR_MASK (~(PGDIR_SIZE-1))
return PGDIR_SIZE - PUD_SIZE;
if (ALIGN_DOWN(addr, PGDIR_SIZE) == addr)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD (PAGE_OFFSET/PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define XKVRANGE_VC_SIZE round_up(KFENCE_AREA_END - MODULES_VADDR + 1, PGDIR_SIZE)
#define KASAN_SHADOW_START round_up(KFENCE_AREA_END, PGDIR_SIZE)
#define KASAN_SHADOW_END (round_up(KASAN_SHADOW_START + KASAN_SHADOW_SIZE, PGDIR_SIZE) - 1)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
#define USER_PTRS_PER_PGD ((TASK_SIZE64 / PGDIR_SIZE)?(TASK_SIZE64 / PGDIR_SIZE):1)
BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_START, PGDIR_SIZE));
BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_END + 1, PGDIR_SIZE));
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
if (!(virtaddr & (PGDIR_SIZE-1)) &&
size >= PGDIR_SIZE) {
size -= PGDIR_SIZE;
virtaddr += PGDIR_SIZE;
physaddr += PGDIR_SIZE;
end2 = (vaddr + (PGDIR_SIZE-1)) & PGDIR_MASK;
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD (0x80000000UL/PGDIR_SIZE)
#define USER_PTRS_PER_PGD ((TASK_SIZE64 / PGDIR_SIZE)?(TASK_SIZE64 / PGDIR_SIZE):1)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
(CONFIG_NIOS2_KERNEL_MMU_REGION_BASE / PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
address = (address + PGDIR_SIZE) & PGDIR_MASK;
if (end > PGDIR_SIZE)
end = PGDIR_SIZE;
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
if (end > PGDIR_SIZE)
end = PGDIR_SIZE;
vaddr = vaddr + PGDIR_SIZE;
if (end > PGDIR_SIZE)
end = PGDIR_SIZE;
vaddr = vaddr + PGDIR_SIZE;
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define FIXMAP_ALIGNED_SIZE (ALIGN(FIXADDR_TOP, PGDIR_SIZE) - \
ALIGN_DOWN(FIXADDR_START, PGDIR_SIZE))
#define FIXMAP_PTE_SIZE (FIXMAP_ALIGNED_SIZE / PGDIR_SIZE * PTE_TABLE_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
else if (sz < PGDIR_SIZE)
pdsize = PGDIR_SIZE;
pmd_end = ((start + PGDIR_SIZE) & PGDIR_MASK) - 1;
for (addr = KASAN_SHADOW_START; addr != KASAN_SHADOW_END; addr += PGDIR_SIZE)
BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_START, PGDIR_SIZE));
BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_END, PGDIR_SIZE));
for (va = ALIGN_DOWN(MODULES_VADDR, PGDIR_SIZE); va < MODULES_END; va += PGDIR_SIZE)
else if (sz < PGDIR_SIZE)
pdsize = PGDIR_SIZE;
unsigned long addr = ALIGN_DOWN(FIXADDR_START, PGDIR_SIZE);
addr += PGDIR_SIZE, ptep += PTRS_PER_PTE, pmdp++)
addr = KERN_VIRT_START + i * PGDIR_SIZE;
ptdump_range[0].end = PAGE_OFFSET + (PGDIR_SIZE * PTRS_PER_PGD);
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
#define MAX_FDT_SIZE PGDIR_SIZE
#define FIXADDR_SIZE (PGDIR_SIZE + FIX_FDT_SIZE)
#define TASK_SIZE_64 (PGDIR_SIZE * PTRS_PER_PGD / 2)
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
#define KERN_VIRT_SIZE ((PTRS_PER_PGD / 2 * PGDIR_SIZE) / 2)
if (sz >= PGDIR_SIZE)
BUG_ON((PAGE_OFFSET % PGDIR_SIZE) != 0);
fixmap_pgd_next, PGDIR_SIZE, PAGE_TABLE);
trampoline_pgd_next, PGDIR_SIZE, PAGE_TABLE);
kernel_map.phys_addr, PGDIR_SIZE, PAGE_KERNEL_EXEC);
PGDIR_SIZE, PAGE_TABLE);
for (addr = start; addr < end && addr >= start; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
if (sz == PGDIR_SIZE) {
PGDIR_SIZE, PAGE_TABLE);
if (pgd_none(pgdp_get(pgdp)) && IS_ALIGNED(vaddr, PGDIR_SIZE) &&
(next - vaddr) >= PGDIR_SIZE) {
phys_addr = memblock_phys_alloc(PGDIR_SIZE, PGDIR_SIZE);
memset(__va(phys_addr), KASAN_SHADOW_INIT, PGDIR_SIZE);
if (pgtable_l5_enabled && IS_ALIGNED(vaddr, PGDIR_SIZE) &&
(next - vaddr) >= PGDIR_SIZE) {
if (pgd_none(pgdp_get(pgdp)) && IS_ALIGNED(vaddr, PGDIR_SIZE) &&
(next - vaddr) >= PGDIR_SIZE) {
if (stride_size >= PGDIR_SIZE)
stride_size = PGDIR_SIZE;
IS_ALIGNED(addr, PGDIR_SIZE) && end - addr >= PGDIR_SIZE) {
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD PAGE_OFFSET / PGDIR_SIZE
#define PGDIR_MASK (~(PGDIR_SIZE-1))
if (srmmu_probe(addr + PGDIR_SIZE) == probed)
start += PGDIR_SIZE;
vstart += PGDIR_SIZE; pstart += PGDIR_SIZE;
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE)
if (task_size > (unsigned long long) PTRS_PER_PGD * PGDIR_SIZE)
task_size = PTRS_PER_PGD * PGDIR_SIZE;
pgd_size = (((pgd_end - pgd_start) / PGDIR_SIZE) + 1) * sizeof(pgd_t);
entries += (DIV_ROUND_UP(len, PGDIR_SIZE) + 1) * sizeof(p4d_t) * PTRS_PER_P4D;
tables += DIV_ROUND_UP(entries, PGDIR_SIZE) * sizeof(p4d_t) * PTRS_PER_P4D;
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
#define LDT_END_ADDR (LDT_BASE_ADDR + PGDIR_SIZE)
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
for (addr = VMALLOC_START; addr <= VMEMORY_END; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
for (addr = start; addr <= end; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
for (addr = start; addr <= end; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
vaddr_next = (vaddr & PGDIR_MASK) + PGDIR_SIZE;
for (; start < pgd_end; start += PGDIR_SIZE) {
unpin = (vaddr == 2 * PGDIR_SIZE);
#define USER_LIMIT ((STACK_TOP_MAX + PGDIR_SIZE - 1) & PGDIR_MASK)
new_p2m = (unsigned long *)(2 * PGDIR_SIZE);
#define PGDIR_MASK (~(PGDIR_SIZE-1))
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
({ unsigned long __boundary = ((addr) + PGDIR_SIZE) & PGDIR_MASK; \
if (IS_ALIGNED(addr, PGDIR_SIZE) && end - addr >= PGDIR_SIZE) {
if (IS_ALIGNED(addr, PGDIR_SIZE) &&
IS_ALIGNED(next, PGDIR_SIZE)) {
step_forward(pvmw, PGDIR_SIZE);