Symbol: P4D_SIZE
arch/arm/mm/dump.c
389
addr = start + i * P4D_SIZE;
arch/arm64/include/asm/pgtable-hwdef.h
74
#define P4D_MASK (~(P4D_SIZE-1))
arch/arm64/mm/kasan_init.c
182
#define SHADOW_ALIGN P4D_SIZE
arch/arm64/mm/mmu.c
691
if (ALIGN_DOWN(addr, P4D_SIZE) == addr)
arch/powerpc/include/asm/nohash/pgtable.h
80
else if (sz < P4D_SIZE)
arch/powerpc/include/asm/nohash/pgtable.h
83
pdsize = P4D_SIZE;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1367
gpa = (gpa & P4D_MASK) + P4D_SIZE;
arch/powerpc/mm/book3s64/radix_pgtable.c
914
if (!IS_ALIGNED(addr, P4D_SIZE) ||
arch/powerpc/mm/book3s64/radix_pgtable.c
915
!IS_ALIGNED(next, P4D_SIZE)) {
arch/powerpc/mm/hugetlbpage.c
53
if (!mm_pud_folded(mm) && sz >= P4D_SIZE)
arch/powerpc/mm/pgtable.c
371
else if (sz < P4D_SIZE)
arch/powerpc/mm/pgtable.c
374
pdsize = P4D_SIZE;
arch/powerpc/mm/ptdump/hashpagetable.c
449
addr = start + i * P4D_SIZE;
arch/riscv/include/asm/pgtable-64.h
32
#define P4D_MASK (~(P4D_SIZE - 1))
arch/riscv/mm/hugetlbpage.c
135
return P4D_SIZE - PUD_SIZE;
arch/riscv/mm/hugetlbpage.c
224
else if (sz >= P4D_SIZE)
arch/riscv/mm/init.c
1190
(uintptr_t)fixmap_pud, P4D_SIZE, PAGE_TABLE);
arch/riscv/mm/init.c
1202
(uintptr_t)trampoline_pud, P4D_SIZE, PAGE_TABLE);
arch/riscv/mm/init.c
1748
free_vmemmap_storage(p4d_page(p4d), P4D_SIZE, altmap);
arch/riscv/mm/init.c
666
if (sz == P4D_SIZE) {
arch/riscv/mm/init.c
748
!(pa & (P4D_SIZE - 1)) && !(va & (P4D_SIZE - 1)) && size >= P4D_SIZE)
arch/riscv/mm/init.c
749
return P4D_SIZE;
arch/riscv/mm/init.c
877
P4D_SIZE, PAGE_TABLE);
arch/riscv/mm/kasan_init.c
128
if (p4d_none(p4dp_get(p4dp)) && IS_ALIGNED(vaddr, P4D_SIZE) &&
arch/riscv/mm/kasan_init.c
129
(next - vaddr) >= P4D_SIZE) {
arch/riscv/mm/kasan_init.c
130
phys_addr = memblock_phys_alloc(P4D_SIZE, P4D_SIZE);
arch/riscv/mm/kasan_init.c
133
memset(__va(phys_addr), KASAN_SHADOW_INIT, P4D_SIZE);
arch/riscv/mm/kasan_init.c
206
if (pgtable_l4_enabled && IS_ALIGNED(vaddr, P4D_SIZE) &&
arch/riscv/mm/kasan_init.c
207
(next - vaddr) >= P4D_SIZE) {
arch/riscv/mm/kasan_init.c
290
if (p4d_none(p4dp_get(p4dp)) && IS_ALIGNED(vaddr, P4D_SIZE) &&
arch/riscv/mm/kasan_init.c
291
(next - vaddr) >= P4D_SIZE) {
arch/riscv/mm/pageattr.c
196
if (next - vaddr >= P4D_SIZE &&
arch/riscv/mm/tlbflush.c
188
else if (stride_size >= P4D_SIZE)
arch/riscv/mm/tlbflush.c
189
stride_size = P4D_SIZE;
arch/s390/boot/vmem.c
152
IS_ALIGNED(addr, P4D_SIZE) && end - addr >= P4D_SIZE) {
arch/x86/boot/startup/sme.c
268
entries += (DIV_ROUND_UP(len, P4D_SIZE) + 1) * sizeof(pud_t) * PTRS_PER_PUD;
arch/x86/boot/startup/sme.c
279
tables += DIV_ROUND_UP(entries, P4D_SIZE) * sizeof(pud_t) * PTRS_PER_PUD;
arch/x86/include/asm/pgtable_64_types.h
60
#define P4D_MASK (~(P4D_SIZE - 1))
arch/x86/include/asm/pgtable_areas.h
19
#define CPU_ENTRY_AREA_MAP_SIZE P4D_SIZE
arch/x86/mm/init_64.c
707
vaddr_next = (vaddr & P4D_MASK) + P4D_SIZE;
arch/x86/mm/kasan_init_64.c
187
for (; start < end; start += P4D_SIZE)
arch/x86/xen/mmu_pv.c
1165
xen_free_ro_pages(pa, P4D_SIZE);
arch/x86/xen/mmu_pv.c
1968
n_pud = roundup(size, P4D_SIZE) >> P4D_SHIFT;
include/asm-generic/pgtable-nop4d.h
14
#define P4D_MASK (~(P4D_SIZE-1))
include/asm-generic/tlb.h
684
if (_sz >= P4D_SIZE) \
include/linux/pgtable.h
1408
({ unsigned long __boundary = ((addr) + P4D_SIZE) & P4D_MASK; \
include/linux/pgtable.h
2281
#define p4d_leaf_size(x) P4D_SIZE
mm/hugetlb.c
7098
return P4D_SIZE - PUD_SIZE;
mm/kasan/init.c
190
if (IS_ALIGNED(addr, P4D_SIZE) && end - addr >= P4D_SIZE) {
mm/kasan/init.c
433
if (IS_ALIGNED(addr, P4D_SIZE) &&
mm/kasan/init.c
434
IS_ALIGNED(next, P4D_SIZE)) {
mm/page_vma_mapped.c
228
step_forward(pvmw, P4D_SIZE);
mm/vmalloc.c
256
if ((end - addr) != P4D_SIZE)
mm/vmalloc.c
259
if (!IS_ALIGNED(addr, P4D_SIZE))
mm/vmalloc.c
262
if (!IS_ALIGNED(phys_addr, P4D_SIZE))
mm/vmscan.c
3765
end = round_up(end, P4D_SIZE);