Symbol: pud_index
arch/arm64/include/asm/pgtable.h
1000
return (pud_t *)PTR_ALIGN_DOWN(p4dp, PAGE_SIZE) + pud_index(addr);
arch/arm64/include/asm/pgtable.h
1012
return p4d_page_paddr(READ_ONCE(*p4dp)) + pud_index(addr) * sizeof(pud_t);
arch/arm64/include/asm/pgtable.h
1020
return (pud_t *)__va(p4d_page_paddr(p4d)) + pud_index(addr);
arch/arm64/mm/mmu.c
365
pudp += pud_index(addr);
arch/csky/mm/init.c
96
j = pud_index(vaddr);
arch/mips/kvm/mmu.c
204
int i_min = pud_index(start_gpa);
arch/mips/kvm/mmu.c
205
int i_max = pud_index(end_gpa);
arch/mips/kvm/mmu.c
331
int i_min = pud_index(start); \
arch/mips/kvm/mmu.c
332
int i_max = pud_index(end); \
arch/mips/mm/init.c
237
j = pud_index(vaddr);
arch/powerpc/mm/book3s64/radix_pgtable.c
868
pud = pud_start + pud_index(addr);
arch/riscv/mm/init.c
1705
pudp = pud_base + pud_index(addr);
arch/riscv/mm/init.c
635
uintptr_t pud_index = pud_index(va);
arch/riscv/mm/init.c
638
if (pud_val(pudp[pud_index]) == 0)
arch/riscv/mm/init.c
639
pudp[pud_index] = pfn_pud(PFN_DOWN(pa), prot);
arch/riscv/mm/init.c
643
if (pud_val(pudp[pud_index]) == 0) {
arch/riscv/mm/init.c
645
pudp[pud_index] = pfn_pud(PFN_DOWN(next_phys), PAGE_TABLE);
arch/riscv/mm/init.c
649
next_phys = PFN_PHYS(_pud_pfn(pudp[pud_index]));
arch/riscv/mm/kasan_init.c
175
pudp = base_pud + pud_index(vaddr);
arch/riscv/mm/kasan_init.c
246
pudp = base_pud + pud_index(vaddr);
arch/riscv/mm/pgtable.c
45
return p4d_pgtable(p4dp_get(p4d)) + pud_index(address);
arch/s390/include/asm/pgtable.h
1432
return (pud_t *) p4d_deref(p4d) + pud_index(address);
arch/s390/include/asm/pgtable.h
1635
r3o = __pa(pudp) - pud_index(addr) * sizeof(pud_t);
arch/sh/mm/init.c
179
j = pud_index(vaddr);
arch/x86/hyperv/hv_crash.c
520
pud = hv_crash_ptpgs[1] + pud_index(addr) * sizeof(pud);
arch/x86/include/asm/pgtable_64.h
277
L3_START_KERNEL = pud_index(__START_KERNEL_map)
arch/x86/kernel/espfix_64.c
162
pud_p = &espfix_pud_page[pud_index(addr)];
arch/x86/kernel/head64.c
129
pud_p += pud_index(address);
arch/x86/mm/ident_map.c
100
pud_t *pud = pud_page + pud_index(addr);
arch/x86/mm/init_64.c
1191
pud = pud_start + pud_index(addr);
arch/x86/mm/init_64.c
338
pud_t *pud = pud_page + pud_index(vaddr);
arch/x86/mm/init_64.c
613
int i = pud_index(vaddr);
arch/x86/mm/init_64.c
621
pud = pud_page + pud_index(vaddr);
arch/x86/mm/kaslr.c
194
pud_tramp = pud_page_tramp + pud_index(paddr);
arch/x86/platform/efi/efi_64.c
143
num_entries = pud_index(EFI_VA_END);
arch/x86/platform/efi/efi_64.c
149
num_entries = PTRS_PER_PUD - pud_index(EFI_VA_START);
arch/x86/power/hibernate_64.c
68
set_pud(pud + pud_index(restore_jump_address),
arch/x86/xen/mmu_pv.c
1805
l2 = m2v(l3[pud_index(__START_KERNEL_map)].pud);
arch/x86/xen/mmu_pv.c
1924
pud = native_make_pud(xen_read_phys_ulong(pa + pud_index(vaddr) *
arch/x86/xen/mmu_pv.c
622
nr = last ? pud_index(limit) + 1 : PTRS_PER_PUD;
include/linux/pgtable.h
135
return p4d_pgtable(*p4d) + pud_index(address);
include/linux/pgtable.h
61
#ifndef pud_index
include/linux/pgtable.h
66
#define pud_index pud_index
mm/vmscan.c
3748
for (i = pud_index(start), addr = start; addr != end; i++, addr = next) {
tools/testing/selftests/kvm/lib/arm64/processor.c
152
ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pud_index(vm, vaddr) * 8;
tools/testing/selftests/kvm/lib/arm64/processor.c
199
ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pud_index(vm, gva) * 8;