Symbol: PGDIR_SIZE
arch/alpha/include/asm/pgtable.h
39
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/alpha/include/asm/pgtable.h
48
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/alpha/include/asm/pgtable.h
56
#define VMALLOC_START (-2*PGDIR_SIZE)
arch/alpha/include/asm/pgtable.h
58
#define VMALLOC_END (-PGDIR_SIZE)
arch/arc/include/asm/highmem.h
14
#define FIXMAP_SIZE PGDIR_SIZE
arch/arc/include/asm/highmem.h
15
#define PKMAP_SIZE PGDIR_SIZE
arch/arc/include/asm/pgalloc.h
61
num = USER_PTRS_PER_PGD + USER_KERNEL_GUTTER / PGDIR_SIZE;
arch/arc/include/asm/pgalloc.h
62
num2 = VMALLOC_SIZE / PGDIR_SIZE;
arch/arc/include/asm/pgtable-levels.h
71
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
arch/arc/include/asm/pgtable.h
20
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/arm/include/asm/pgtable-2level.h
106
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/arm/include/asm/pgtable-2level.h
90
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/arm/include/asm/pgtable-3level.h
52
#define USER_PTRS_PER_PGD (PAGE_OFFSET / PGDIR_SIZE)
arch/arm/include/asm/pgtable-nommu.h
31
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/arm/mach-rpc/ecard.c
234
memcpy(dst_pgd, src_pgd, sizeof(pgd_t) * (IO_SIZE / PGDIR_SIZE));
arch/arm/mach-rpc/ecard.c
239
memcpy(dst_pgd, src_pgd, sizeof(pgd_t) * (EASI_SIZE / PGDIR_SIZE));
arch/arm/mm/dump.c
406
addr = start + i * PGDIR_SIZE;
arch/arm/mm/ioremap.c
134
end = ALIGN(end, PGDIR_SIZE);
arch/arm64/include/asm/pgtable-hwdef.h
84
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/arm64/mm/hugetlbpage.c
328
return PGDIR_SIZE - PUD_SIZE;
arch/arm64/mm/mmu.c
680
if (ALIGN_DOWN(addr, PGDIR_SIZE) == addr)
arch/csky/include/asm/pgtable.h
14
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/csky/include/asm/pgtable.h
16
#define USER_PTRS_PER_PGD (PAGE_OFFSET/PGDIR_SIZE)
arch/hexagon/include/asm/pgtable.h
77
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/loongarch/include/asm/kasan.h
54
#define XKVRANGE_VC_SIZE round_up(KFENCE_AREA_END - MODULES_VADDR + 1, PGDIR_SIZE)
arch/loongarch/include/asm/kasan.h
60
#define KASAN_SHADOW_START round_up(KFENCE_AREA_END, PGDIR_SIZE)
arch/loongarch/include/asm/kasan.h
62
#define KASAN_SHADOW_END (round_up(KASAN_SHADOW_START + KASAN_SHADOW_SIZE, PGDIR_SIZE) - 1)
arch/loongarch/include/asm/pgtable.h
44
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/loongarch/include/asm/pgtable.h
62
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/loongarch/include/asm/pgtable.h
64
#define USER_PTRS_PER_PGD ((TASK_SIZE64 / PGDIR_SIZE)?(TASK_SIZE64 / PGDIR_SIZE):1)
arch/loongarch/mm/kasan_init.c
247
BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_START, PGDIR_SIZE));
arch/loongarch/mm/kasan_init.c
248
BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_END + 1, PGDIR_SIZE));
arch/m68k/include/asm/pgtable_mm.h
51
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/m68k/include/asm/pgtable_mm.h
72
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
arch/m68k/mm/motorola.c
326
if (!(virtaddr & (PGDIR_SIZE-1)) &&
arch/m68k/mm/motorola.c
327
size >= PGDIR_SIZE) {
arch/m68k/mm/motorola.c
332
size -= PGDIR_SIZE;
arch/m68k/mm/motorola.c
333
virtaddr += PGDIR_SIZE;
arch/m68k/mm/motorola.c
334
physaddr += PGDIR_SIZE;
arch/m68k/sun3x/dvma.c
107
end2 = (vaddr + (PGDIR_SIZE-1)) & PGDIR_MASK;
arch/microblaze/include/asm/pgtable.h
101
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/microblaze/include/asm/pgtable.h
91
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/mips/include/asm/pgtable-32.h
71
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/mips/include/asm/pgtable-32.h
94
#define USER_PTRS_PER_PGD (0x80000000UL/PGDIR_SIZE)
arch/mips/include/asm/pgtable-64.h
134
#define USER_PTRS_PER_PGD ((TASK_SIZE64 / PGDIR_SIZE)?(TASK_SIZE64 / PGDIR_SIZE):1)
arch/mips/include/asm/pgtable-64.h
66
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/nios2/include/asm/pgtable.h
62
(CONFIG_NIOS2_KERNEL_MMU_REGION_BASE / PGDIR_SIZE)
arch/nios2/include/asm/pgtable.h
66
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/nios2/mm/ioremap.c
107
address = (address + PGDIR_SIZE) & PGDIR_MASK;
arch/nios2/mm/ioremap.c
58
if (end > PGDIR_SIZE)
arch/nios2/mm/ioremap.c
59
end = PGDIR_SIZE;
arch/openrisc/include/asm/pgtable.h
58
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/openrisc/include/asm/pgtable.h
75
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
arch/parisc/include/asm/pgtable.h
129
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/parisc/kernel/pci-dma.c
112
if (end > PGDIR_SIZE)
arch/parisc/kernel/pci-dma.c
113
end = PGDIR_SIZE;
arch/parisc/kernel/pci-dma.c
147
vaddr = vaddr + PGDIR_SIZE;
arch/parisc/kernel/pci-dma.c
206
if (end > PGDIR_SIZE)
arch/parisc/kernel/pci-dma.c
207
end = PGDIR_SIZE;
arch/parisc/kernel/pci-dma.c
224
vaddr = vaddr + PGDIR_SIZE;
arch/powerpc/include/asm/book3s/32/pgtable.h
131
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/powerpc/include/asm/book3s/32/pgtable.h
133
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/powerpc/include/asm/book3s/64/pgtable.h
220
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/powerpc/include/asm/fixmap.h
84
#define FIXMAP_ALIGNED_SIZE (ALIGN(FIXADDR_TOP, PGDIR_SIZE) - \
arch/powerpc/include/asm/fixmap.h
85
ALIGN_DOWN(FIXADDR_START, PGDIR_SIZE))
arch/powerpc/include/asm/fixmap.h
86
#define FIXMAP_PTE_SIZE (FIXMAP_ALIGNED_SIZE / PGDIR_SIZE * PTE_TABLE_SIZE)
arch/powerpc/include/asm/nohash/32/pgtable.h
47
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/powerpc/include/asm/nohash/32/pgtable.h
52
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/powerpc/include/asm/nohash/64/pgtable-4k.h
42
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/powerpc/include/asm/nohash/pgtable.h
82
else if (sz < PGDIR_SIZE)
arch/powerpc/include/asm/nohash/pgtable.h
85
pdsize = PGDIR_SIZE;
arch/powerpc/mm/book3s32/tlb.c
63
pmd_end = ((start + PGDIR_SIZE) & PGDIR_MASK) - 1;
arch/powerpc/mm/kasan/init_book3e_64.c
104
for (addr = KASAN_SHADOW_START; addr != KASAN_SHADOW_END; addr += PGDIR_SIZE)
arch/powerpc/mm/kasan/init_book3e_64.c
89
BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_START, PGDIR_SIZE));
arch/powerpc/mm/kasan/init_book3e_64.c
90
BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_END, PGDIR_SIZE));
arch/powerpc/mm/mem.c
390
for (va = ALIGN_DOWN(MODULES_VADDR, PGDIR_SIZE); va < MODULES_END; va += PGDIR_SIZE)
arch/powerpc/mm/pgtable.c
373
else if (sz < PGDIR_SIZE)
arch/powerpc/mm/pgtable.c
376
pdsize = PGDIR_SIZE;
arch/powerpc/mm/pgtable_32.c
40
unsigned long addr = ALIGN_DOWN(FIXADDR_START, PGDIR_SIZE);
arch/powerpc/mm/pgtable_32.c
45
addr += PGDIR_SIZE, ptep += PTRS_PER_PTE, pmdp++)
arch/powerpc/mm/ptdump/hashpagetable.c
467
addr = KERN_VIRT_START + i * PGDIR_SIZE;
arch/powerpc/mm/ptdump/ptdump.c
414
ptdump_range[0].end = PAGE_OFFSET + (PGDIR_SIZE * PTRS_PER_PGD);
arch/riscv/include/asm/pgtable-32.h
16
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
arch/riscv/include/asm/pgtable-64.h
23
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
arch/riscv/include/asm/pgtable.h
106
#define MAX_FDT_SIZE PGDIR_SIZE
arch/riscv/include/asm/pgtable.h
108
#define FIXADDR_SIZE (PGDIR_SIZE + FIX_FDT_SIZE)
arch/riscv/include/asm/pgtable.h
1247
#define TASK_SIZE_64 (PGDIR_SIZE * PTRS_PER_PGD / 2)
arch/riscv/include/asm/pgtable.h
169
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/riscv/include/asm/pgtable.h
41
#define KERN_VIRT_SIZE ((PTRS_PER_PGD / 2 * PGDIR_SIZE) / 2)
arch/riscv/mm/hugetlbpage.c
222
if (sz >= PGDIR_SIZE)
arch/riscv/mm/init.c
1156
BUG_ON((PAGE_OFFSET % PGDIR_SIZE) != 0);
arch/riscv/mm/init.c
1184
fixmap_pgd_next, PGDIR_SIZE, PAGE_TABLE);
arch/riscv/mm/init.c
1199
trampoline_pgd_next, PGDIR_SIZE, PAGE_TABLE);
arch/riscv/mm/init.c
1216
kernel_map.phys_addr, PGDIR_SIZE, PAGE_KERNEL_EXEC);
arch/riscv/mm/init.c
1349
PGDIR_SIZE, PAGE_TABLE);
arch/riscv/mm/init.c
1479
for (addr = start; addr < end && addr >= start; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
arch/riscv/mm/init.c
723
if (sz == PGDIR_SIZE) {
arch/riscv/mm/init.c
894
PGDIR_SIZE, PAGE_TABLE);
arch/riscv/mm/kasan_init.c
151
if (pgd_none(pgdp_get(pgdp)) && IS_ALIGNED(vaddr, PGDIR_SIZE) &&
arch/riscv/mm/kasan_init.c
152
(next - vaddr) >= PGDIR_SIZE) {
arch/riscv/mm/kasan_init.c
153
phys_addr = memblock_phys_alloc(PGDIR_SIZE, PGDIR_SIZE);
arch/riscv/mm/kasan_init.c
156
memset(__va(phys_addr), KASAN_SHADOW_INIT, PGDIR_SIZE);
arch/riscv/mm/kasan_init.c
224
if (pgtable_l5_enabled && IS_ALIGNED(vaddr, PGDIR_SIZE) &&
arch/riscv/mm/kasan_init.c
225
(next - vaddr) >= PGDIR_SIZE) {
arch/riscv/mm/kasan_init.c
311
if (pgd_none(pgdp_get(pgdp)) && IS_ALIGNED(vaddr, PGDIR_SIZE) &&
arch/riscv/mm/kasan_init.c
312
(next - vaddr) >= PGDIR_SIZE) {
arch/riscv/mm/tlbflush.c
186
if (stride_size >= PGDIR_SIZE)
arch/riscv/mm/tlbflush.c
187
stride_size = PGDIR_SIZE;
arch/s390/boot/vmem.c
141
IS_ALIGNED(addr, PGDIR_SIZE) && end - addr >= PGDIR_SIZE) {
arch/sh/include/asm/pgtable-2level.h
22
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
arch/sh/include/asm/pgtable.h
57
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/sparc/include/asm/pgtable_32.h
21
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/sparc/include/asm/pgtable_32.h
50
#define USER_PTRS_PER_PGD PAGE_OFFSET / PGDIR_SIZE
arch/sparc/include/asm/pgtable_64.h
67
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/sparc/mm/srmmu.c
798
if (srmmu_probe(addr + PGDIR_SIZE) == probed)
arch/sparc/mm/srmmu.c
807
start += PGDIR_SIZE;
arch/sparc/mm/srmmu.c
867
vstart += PGDIR_SIZE; pstart += PGDIR_SIZE;
arch/um/include/asm/pgtable-2level.h
17
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/um/include/asm/pgtable-2level.h
24
#define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE)
arch/um/include/asm/pgtable-4level.h
16
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/um/include/asm/pgtable-4level.h
43
#define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE)
arch/um/kernel/um_arch.c
336
if (task_size > (unsigned long long) PTRS_PER_PGD * PGDIR_SIZE)
arch/um/kernel/um_arch.c
337
task_size = PTRS_PER_PGD * PGDIR_SIZE;
arch/x86/boot/startup/sme.c
103
pgd_size = (((pgd_end - pgd_start) / PGDIR_SIZE) + 1) * sizeof(pgd_t);
arch/x86/boot/startup/sme.c
267
entries += (DIV_ROUND_UP(len, PGDIR_SIZE) + 1) * sizeof(p4d_t) * PTRS_PER_P4D;
arch/x86/boot/startup/sme.c
278
tables += DIV_ROUND_UP(entries, PGDIR_SIZE) * sizeof(p4d_t) * PTRS_PER_P4D;
arch/x86/include/asm/pgtable_32_types.h
21
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
arch/x86/include/asm/pgtable_64_types.h
105
#define LDT_END_ADDR (LDT_BASE_ADDR + PGDIR_SIZE)
arch/x86/include/asm/pgtable_64_types.h
87
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
arch/x86/mm/init_64.c
1323
for (addr = VMALLOC_START; addr <= VMEMORY_END; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
arch/x86/mm/init_64.c
137
for (addr = start; addr <= end; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
arch/x86/mm/init_64.c
174
for (addr = start; addr <= end; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
arch/x86/mm/init_64.c
758
vaddr_next = (vaddr & PGDIR_MASK) + PGDIR_SIZE;
arch/x86/mm/kasan_init_64.c
174
for (; start < pgd_end; start += PGDIR_SIZE) {
arch/x86/xen/mmu_pv.c
1189
unpin = (vaddr == 2 * PGDIR_SIZE);
arch/x86/xen/mmu_pv.c
187
#define USER_LIMIT ((STACK_TOP_MAX + PGDIR_SIZE - 1) & PGDIR_MASK)
arch/x86/xen/mmu_pv.c
1991
new_p2m = (unsigned long *)(2 * PGDIR_SIZE);
arch/xtensa/include/asm/pgtable.h
51
#define PGDIR_MASK (~(PGDIR_SIZE-1))
arch/xtensa/include/asm/pgtable.h
60
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
include/linux/pgtable.h
1402
({ unsigned long __boundary = ((addr) + PGDIR_SIZE) & PGDIR_MASK; \
mm/kasan/init.c
241
if (IS_ALIGNED(addr, PGDIR_SIZE) && end - addr >= PGDIR_SIZE) {
mm/kasan/init.c
467
if (IS_ALIGNED(addr, PGDIR_SIZE) &&
mm/kasan/init.c
468
IS_ALIGNED(next, PGDIR_SIZE)) {
mm/page_vma_mapped.c
223
step_forward(pvmw, PGDIR_SIZE);