pgd_index
long index = pgd_index(address);
index = pgd_index(addr);
pgd += pgd_index(addr);
sizeof(pgd_t) * (pgd_index(end) - pgd_index(start)));
BUILD_BUG_ON(pgd_index(KASAN_SHADOW_START) !=
pgd_index(KASAN_SHADOW_END));
set_pgd(&tmp_pgd_table[pgd_index(KASAN_SHADOW_START)],
pgd = pgd_base + pgd_index(0);
new_p4d = p4d_alloc(mm, new_pgd + pgd_index(MODULES_VADDR),
BUILD_BUG_ON(pgd_index(direct_map_end - 1) == pgd_index(direct_map_end) &&
pgd_index(_PAGE_OFFSET(VA_BITS_MIN)) != PTRS_PER_PGD - 1);
offset = pgd_index(addr);
pgd = swapper_pg_dir + pgd_index(vaddr);
i = pgd_index(vaddr);
pgd += pgd_index(addr);
int i_min = pgd_index(start_gpa);
int i_max = pgd_index(end_gpa);
int i_min = pgd_index(start); \
int i_max = pgd_index(end); \
int offset = pgd_index(address);
i = pgd_index(vaddr);
pgd = swapper_pg_dir + pgd_index(vaddr);
int offset = pgd_index(address);
int offset = pgd_index(address);
pgdp = pgt + pgd_index(gpa);
pgd = pgtable + pgd_index(gpa);
pgd = mm->pgd + pgd_index(addr);
pgdp = pgdir + pgd_index(ea);
index = pgd_index(addr);
unsigned long idx = pgd_index(__fix_to_virt(FIX_FDT));
uintptr_t pgd_idx = pgd_index(va);
kasan_early_populate_pgd(early_pg_dir + pgd_index(KASAN_SHADOW_START),
set_pgd(&tmp_pg_dir[pgd_index(KASAN_SHADOW_END)],
pgd_k = swapper_pg_dir + pgd_index(address);
unsigned index = pgd_index(address);
pgd += pgd_index(addr);
i = pgd_index(vaddr);
int offset = pgd_index(address);
pgd[pgd_index(__START_KERNEL_map)] += load_delta;
pgd[pgd_index(__START_KERNEL_map)] = (pgdval_t)p4d | _PAGE_TABLE;
pgd_p = ppd->pgd + pgd_index(ppd->vaddr);
pgd = ppd->pgd + pgd_index(ppd->vaddr);
decrypted_base = (pgd_index(workarea_end) + 1) & (PTRS_PER_PGD - 1);
check_base = (pgd_index(initrd_end) + 1) & (PTRS_PER_PGD - 1);
pgd = &pgd[pgd_index(va)];
p4d = hv_crash_ptpgs[0] + pgd_index(addr) * sizeof(p4d);
#define KERNEL_PGD_BOUNDARY pgd_index(PAGE_OFFSET)
pgd = &init_top_pgt[pgd_index(ESPFIX_BASE_ADDR)];
pgd_p = &early_top_pgt[pgd_index(address)].pgd;
pgd += pgd_index(vaddr);
pgd += pgd_index(vaddr);
pgd = init_mm.pgd + pgd_index(address);
unsigned index = pgd_index(address);
pgd_t *pgd = &base[pgd_index(address)];
pgd_t *pgd = base + pgd_index(address);
pgd += pgd_index(address);
pgd_t *pgd = pgd_page + pgd_index(addr);
trampoline_pgd_entry = init_top_pgt[pgd_index(__PAGE_OFFSET)];
int pgd_idx = pgd_index(vaddr);
pgd_idx = pgd_index(vaddr);
pgd_idx = pgd_index(vaddr);
pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
pgd = base + pgd_index(va);
pgd = (pgd_t *)page_address(page) + pgd_index(addr);
pgd = (pgd_t *)page_address(page) + pgd_index(addr);
pgd_t *pgd = &base[pgd_index(addr)];
pgd += pgd_index(addr);
set_pgd(&early_top_pgt[pgd_index(KASAN_SHADOW_END)],
pgd_t *pgd = (pgd_t *)page_address(page) + pgd_index(addr);
pgd_entry = cpa->pgd + pgd_index(addr);
pgd = cpa->pgd + pgd_index(address);
pgd = (pgd_t *)page_address(page) + pgd_index(address);
pgd = swapper_pg_dir + pgd_index(vaddr);
pgd_efi = efi_pgd + pgd_index(PAGE_OFFSET);
num_entries = pgd_index(EFI_VA_END) - pgd_index(PAGE_OFFSET);
pgd_efi = efi_pgd + pgd_index(EFI_VA_END);
pgd = efi_pgd + pgd_index(EFI_VA_END);
pgd_index(relocated_restore_code);
pgd = pgd_base + pgd_index(restore_jump_address);
pgd_idx = pgd_index(PAGE_OFFSET);
set_pgd(pgd + pgd_index(restore_jump_address), new_pgd);
set_pgd(pgd + pgd_index(restore_jump_address), new_pgd);
for (i = pgd_index(__PAGE_OFFSET); i < PTRS_PER_PGD; i++)
pgd += pgd_index(hva);
user_pgd[pgd_index(VSYSCALL_ADDR)] =
init_top_pgt[pgd_index(__PAGE_OFFSET_BASE_L4)].pgd =
init_top_pgt[pgd_index(__START_KERNEL_map)].pgd =
l3 = m2v(pgd[pgd_index(__START_KERNEL_map)].pgd);
i = pgd_index(xen_start_info->mfn_list);
if (i && i < pgd_index(__START_KERNEL_map))
pgd = native_make_pgd(xen_read_phys_ulong(pa + pgd_index(vaddr) *
if (offset < pgd_index(USER_LIMIT)) {
hole_low = pgd_index(GUARD_HOLE_BASE_ADDR);
hole_high = pgd_index(GUARD_HOLE_END_ADDR);
nr = pgd_index(limit) + 1;
int index = pgd_index(address);
return (pgd + pgd_index(address));
#ifndef pgd_index
pgd = walk->pgd + pgd_index(addr);
ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pgd_index(vm, vaddr) * 8;
ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pgd_index(vm, gva) * 8;