Symbol: pgd_index
arch/alpha/mm/fault.c
236
long index = pgd_index(address);
arch/arm/mm/fault.c
519
index = pgd_index(addr);
arch/arm/mm/idmap.c
103
pgd += pgd_index(addr);
arch/arm/mm/ioremap.c
136
sizeof(pgd_t) * (pgd_index(end) - pgd_index(start)));
arch/arm/mm/kasan_init.c
231
BUILD_BUG_ON(pgd_index(KASAN_SHADOW_START) !=
arch/arm/mm/kasan_init.c
232
pgd_index(KASAN_SHADOW_END));
arch/arm/mm/kasan_init.c
236
set_pgd(&tmp_pgd_table[pgd_index(KASAN_SHADOW_START)],
arch/arm/mm/pgd.c
151
pgd = pgd_base + pgd_index(0);
arch/arm/mm/pgd.c
55
new_p4d = p4d_alloc(mm, new_pgd + pgd_index(MODULES_VADDR),
arch/arm64/mm/mmu.c
1137
BUILD_BUG_ON(pgd_index(direct_map_end - 1) == pgd_index(direct_map_end) &&
arch/arm64/mm/mmu.c
1138
pgd_index(_PAGE_OFFSET(VA_BITS_MIN)) != PTRS_PER_PGD - 1);
arch/csky/mm/fault.c
136
offset = pgd_index(addr);
arch/csky/mm/highmem.c
31
pgd = swapper_pg_dir + pgd_index(vaddr);
arch/csky/mm/init.c
98
i = pgd_index(vaddr);
arch/mips/kvm/mmu.c
111
pgd += pgd_index(addr);
arch/mips/kvm/mmu.c
233
int i_min = pgd_index(start_gpa);
arch/mips/kvm/mmu.c
234
int i_max = pgd_index(end_gpa);
arch/mips/kvm/mmu.c
355
int i_min = pgd_index(start); \
arch/mips/kvm/mmu.c
356
int i_max = pgd_index(end); \
arch/mips/mm/fault.c
285
int offset = pgd_index(address);
arch/mips/mm/init.c
239
i = pgd_index(vaddr);
arch/mips/mm/pgtable-32.c
72
pgd = swapper_pg_dir + pgd_index(vaddr);
arch/nios2/mm/fault.c
229
int offset = pgd_index(address);
arch/openrisc/mm/fault.c
294
int offset = pgd_index(address);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1363
pgdp = pgt + pgd_index(gpa);
arch/powerpc/kvm/book3s_64_mmu_radix.c
628
pgd = pgtable + pgd_index(gpa);
arch/powerpc/mm/pgtable.c
398
pgd = mm->pgd + pgd_index(addr);
arch/powerpc/mm/pgtable.c
468
pgdp = pgdir + pgd_index(ea);
arch/riscv/mm/fault.c
190
index = pgd_index(addr);
arch/riscv/mm/init.c
1343
unsigned long idx = pgd_index(__fix_to_virt(FIX_FDT));
arch/riscv/mm/init.c
721
uintptr_t pgd_idx = pgd_index(va);
arch/riscv/mm/kasan_init.c
355
kasan_early_populate_pgd(early_pg_dir + pgd_index(KASAN_SHADOW_START),
arch/riscv/mm/kasan_init.c
467
set_pgd(&tmp_pg_dir[pgd_index(KASAN_SHADOW_END)],
arch/sh/kernel/io_trapped.c
143
pgd_k = swapper_pg_dir + pgd_index(address);
arch/sh/mm/fault.c
122
unsigned index = pgd_index(address);
arch/sh/mm/fault.c
51
pgd += pgd_index(addr);
arch/sh/mm/init.c
178
i = pgd_index(vaddr);
arch/sparc/mm/fault_32.c
272
int offset = pgd_index(address);
arch/x86/boot/startup/map_kernel.c
127
pgd[pgd_index(__START_KERNEL_map)] += load_delta;
arch/x86/boot/startup/map_kernel.c
133
pgd[pgd_index(__START_KERNEL_map)] = (pgdval_t)p4d | _PAGE_TABLE;
arch/x86/boot/startup/sme.c
105
pgd_p = ppd->pgd + pgd_index(ppd->vaddr);
arch/x86/boot/startup/sme.c
117
pgd = ppd->pgd + pgd_index(ppd->vaddr);
arch/x86/boot/startup/sme.c
410
decrypted_base = (pgd_index(workarea_end) + 1) & (PTRS_PER_PGD - 1);
arch/x86/boot/startup/sme.c
414
check_base = (pgd_index(initrd_end) + 1) & (PTRS_PER_PGD - 1);
arch/x86/coco/sev/vc-handle.c
49
pgd = &pgd[pgd_index(va)];
arch/x86/hyperv/hv_crash.c
515
p4d = hv_crash_ptpgs[0] + pgd_index(addr) * sizeof(p4d);
arch/x86/include/asm/pgtable.h
1172
#define KERNEL_PGD_BOUNDARY pgd_index(PAGE_OFFSET)
arch/x86/kernel/espfix_64.c
114
pgd = &init_top_pgt[pgd_index(ESPFIX_BASE_ADDR)];
arch/x86/kernel/head64.c
92
pgd_p = &early_top_pgt[pgd_index(address)].pgd;
arch/x86/kernel/machine_kexec_32.c
86
pgd += pgd_index(vaddr);
arch/x86/kernel/machine_kexec_64.c
181
pgd += pgd_index(vaddr);
arch/x86/mm/fault.c
1002
pgd = init_mm.pgd + pgd_index(address);
arch/x86/mm/fault.c
177
unsigned index = pgd_index(address);
arch/x86/mm/fault.c
297
pgd_t *pgd = &base[pgd_index(address)];
arch/x86/mm/fault.c
353
pgd_t *pgd = base + pgd_index(address);
arch/x86/mm/fault.c
521
pgd += pgd_index(address);
arch/x86/mm/ident_map.c
194
pgd_t *pgd = pgd_page + pgd_index(addr);
arch/x86/mm/init.c
752
trampoline_pgd_entry = init_top_pgt[pgd_index(__PAGE_OFFSET)];
arch/x86/mm/init_32.c
109
int pgd_idx = pgd_index(vaddr);
arch/x86/mm/init_32.c
138
pgd_idx = pgd_index(vaddr);
arch/x86/mm/init_32.c
218
pgd_idx = pgd_index(vaddr);
arch/x86/mm/init_32.c
291
pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
arch/x86/mm/init_32.c
438
pgd = base + pgd_index(va);
arch/x86/mm/init_64.c
153
pgd = (pgd_t *)page_address(page) + pgd_index(addr);
arch/x86/mm/init_64.c
195
pgd = (pgd_t *)page_address(page) + pgd_index(addr);
arch/x86/mm/ioremap.c
830
pgd_t *pgd = &base[pgd_index(addr)];
arch/x86/mm/kasan_init_64.c
237
pgd += pgd_index(addr);
arch/x86/mm/kasan_init_64.c
366
set_pgd(&early_top_pgt[pgd_index(KASAN_SHADOW_END)],
arch/x86/mm/pat/set_memory.c
1298
pgd_t *pgd = (pgd_t *)page_address(page) + pgd_index(addr);
arch/x86/mm/pat/set_memory.c
1743
pgd_entry = cpa->pgd + pgd_index(addr);
arch/x86/mm/pat/set_memory.c
799
pgd = cpa->pgd + pgd_index(address);
arch/x86/mm/pat/set_memory.c
895
pgd = (pgd_t *)page_address(page) + pgd_index(address);
arch/x86/mm/pgtable_32.c
35
pgd = swapper_pg_dir + pgd_index(vaddr);
arch/x86/platform/efi/efi_64.c
116
pgd_efi = efi_pgd + pgd_index(PAGE_OFFSET);
arch/x86/platform/efi/efi_64.c
119
num_entries = pgd_index(EFI_VA_END) - pgd_index(PAGE_OFFSET);
arch/x86/platform/efi/efi_64.c
122
pgd_efi = efi_pgd + pgd_index(EFI_VA_END);
arch/x86/platform/efi/efi_64.c
80
pgd = efi_pgd + pgd_index(EFI_VA_END);
arch/x86/power/hibernate.c
170
pgd_index(relocated_restore_code);
arch/x86/power/hibernate_32.c
151
pgd = pgd_base + pgd_index(restore_jump_address);
arch/x86/power/hibernate_32.c
89
pgd_idx = pgd_index(PAGE_OFFSET);
arch/x86/power/hibernate_64.c
77
set_pgd(pgd + pgd_index(restore_jump_address), new_pgd);
arch/x86/power/hibernate_64.c
81
set_pgd(pgd + pgd_index(restore_jump_address), new_pgd);
arch/x86/realmode/init.c
173
for (i = pgd_index(__PAGE_OFFSET); i < PTRS_PER_PGD; i++)
arch/x86/virt/svm/sev.c
847
pgd += pgd_index(hva);
arch/x86/xen/mmu_pv.c
1473
user_pgd[pgd_index(VSYSCALL_ADDR)] =
arch/x86/xen/mmu_pv.c
1783
init_top_pgt[pgd_index(__PAGE_OFFSET_BASE_L4)].pgd =
arch/x86/xen/mmu_pv.c
1785
init_top_pgt[pgd_index(__START_KERNEL_map)].pgd =
arch/x86/xen/mmu_pv.c
1804
l3 = m2v(pgd[pgd_index(__START_KERNEL_map)].pgd);
arch/x86/xen/mmu_pv.c
1833
i = pgd_index(xen_start_info->mfn_list);
arch/x86/xen/mmu_pv.c
1834
if (i && i < pgd_index(__START_KERNEL_map))
arch/x86/xen/mmu_pv.c
1918
pgd = native_make_pgd(xen_read_phys_ulong(pa + pgd_index(vaddr) *
arch/x86/xen/mmu_pv.c
518
if (offset < pgd_index(USER_LIMIT)) {
arch/x86/xen/mmu_pv.c
681
hole_low = pgd_index(GUARD_HOLE_BASE_ADDR);
arch/x86/xen/mmu_pv.c
682
hole_high = pgd_index(GUARD_HOLE_END_ADDR);
arch/x86/xen/mmu_pv.c
684
nr = pgd_index(limit) + 1;
arch/xtensa/mm/fault.c
34
int index = pgd_index(address);
include/linux/pgtable.h
142
return (pgd + pgd_index(address));
include/linux/pgtable.h
69
#ifndef pgd_index
mm/pagewalk.c
301
pgd = walk->pgd + pgd_index(addr);
tools/testing/selftests/kvm/lib/arm64/processor.c
145
ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pgd_index(vm, vaddr) * 8;
tools/testing/selftests/kvm/lib/arm64/processor.c
191
ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pgd_index(vm, gva) * 8;