Symbol: PTRS_PER_PMD
arch/arc/mm/init.c
169
BUILD_BUG_ON((PTRS_PER_PMD * sizeof(pmd_t)) > PAGE_SIZE);
arch/arm/mm/dump.c
349
for (i = 0; i < PTRS_PER_PMD; i++, pmd++) {
arch/arm/mm/fault.c
79
if (PTRS_PER_PMD != 1)
arch/arm/mm/idmap.c
42
PTRS_PER_PMD * sizeof(pmd_t));
arch/arm/mm/kasan_init.c
29
pmd_t tmp_pmd_table[PTRS_PER_PMD] __page_aligned_bss;
arch/arm/mm/mmu.c
1334
PTRS_PER_PGD * PTRS_PER_PMD * sizeof(pmd_t))
arch/arm/mm/pgd.c
79
clean_dcache_area(new_pmd, PTRS_PER_PMD * sizeof(pmd_t));
arch/arm64/kvm/mmu.c
1423
pfn &= ~(PTRS_PER_PMD - 1);
arch/arm64/mm/fixmap.c
35
static pmd_t bm_pmd[PTRS_PER_PMD] __page_aligned_bss __maybe_unused;
arch/arm64/mm/mmu.c
1646
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/arm64/mm/mmu.c
654
for (i = 0; i < PTRS_PER_PMD; i++, pmdp++, pfn += step)
arch/csky/mm/init.c
107
for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) {
arch/loongarch/include/asm/pgtable.h
109
min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, (1UL << cpu_vabits)) - PMD_SIZE - VMEMMAP_SIZE - KFENCE_AREA_SIZE)
arch/loongarch/include/asm/pgtable.h
113
min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, (1UL << cpu_vabits) / 2) - PMD_SIZE - VMEMMAP_SIZE - KFENCE_AREA_SIZE)
arch/loongarch/include/asm/pgtable.h
194
extern pmd_t invalid_pmd_table[PTRS_PER_PMD];
arch/loongarch/kernel/asm-offsets.c
211
DEFINE(_PTRS_PER_PMD, PTRS_PER_PMD);
arch/loongarch/mm/init.c
217
pmd_t invalid_pmd_table[PTRS_PER_PMD] __page_aligned_bss;
arch/loongarch/mm/pgtable.c
77
end = p + PTRS_PER_PMD;
arch/m68k/mm/init.c
110
for (j = 0; j < PTRS_PER_PMD; j++) {
arch/m68k/mm/motorola.c
287
last_pmd_table += PTRS_PER_PMD;
arch/mips/include/asm/pgtable-64.h
144
min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, \
arch/mips/include/asm/pgtable-64.h
234
extern pmd_t invalid_pmd_table[PTRS_PER_PMD];
arch/mips/kernel/asm-offsets.c
219
DEFINE(_PTRS_PER_PMD, PTRS_PER_PMD);
arch/mips/kvm/mmu.c
178
bool safe_to_remove = (i_min == 0 && i_max == PTRS_PER_PMD - 1);
arch/mips/mm/init.c
248
for (; (k < PTRS_PER_PMD) && (vaddr < end); pmd++, k++) {
arch/mips/mm/init.c
559
pmd_t invalid_pmd_table[PTRS_PER_PMD] __page_aligned_bss;
arch/mips/mm/pgtable-64.c
52
end = p + PTRS_PER_PMD;
arch/mips/mm/tlbex.c
1164
uasm_i_andi(p, scratch, scratch, (PTRS_PER_PMD - 1) << 3);
arch/mips/mm/tlbex.c
2445
pwsize |= ilog2(PTRS_PER_PMD) << MIPS_PWSIZE_MDW_SHIFT;
arch/mips/mm/tlbex.c
840
uasm_i_andi(p, tmp, tmp, (PTRS_PER_PMD - 1)<<3);
arch/openrisc/include/asm/pgtable.h
339
(((address) >> PMD_SHIFT) & (PTRS_PER_PMD-1))
arch/parisc/mm/init.c
368
start_pmd = ((start_vaddr >> PMD_SHIFT) & (PTRS_PER_PMD - 1));
arch/parisc/mm/init.c
387
for (tmp1 = start_pmd; tmp1 < PTRS_PER_PMD; tmp1++, pmd++) {
arch/parisc/mm/init.c
43
pmd_t pmd0[PTRS_PER_PMD] __section(".data..vm0.pmd") __attribute__ ((aligned(PAGE_SIZE)));
arch/parisc/mm/init.c
554
BUILD_BUG_ON(PT_INITIAL > PTRS_PER_PMD);
arch/powerpc/include/asm/book3s/64/hash-64k.h
211
return *(char **)(pmdp + PTRS_PER_PMD);
arch/powerpc/kvm/book3s_64_mmu_radix.c
503
for (im = 0; im < PTRS_PER_PMD; ++im, ++p) {
arch/powerpc/mm/book3s64/hash_pgtable.c
279
pgtable_slot = (pgtable_t *)pmdp + PTRS_PER_PMD;
arch/powerpc/mm/book3s64/hash_pgtable.c
297
pgtable_slot = (pgtable_t *)pmdp + PTRS_PER_PMD;
arch/powerpc/mm/book3s64/hash_pgtable.c
363
pgtable_slot = (pgtable_t *)pmdp + PTRS_PER_PMD;
arch/powerpc/mm/book3s64/hash_tlb.c
70
offset = PTRS_PER_PMD;
arch/powerpc/mm/book3s64/hugetlbpage.c
84
offset = PTRS_PER_PMD;
arch/powerpc/mm/book3s64/radix_pgtable.c
1648
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/powerpc/mm/book3s64/radix_pgtable.c
713
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/powerpc/mm/kasan/init_book3e_64.c
96
for (i = 0; i < PTRS_PER_PMD; i++)
arch/powerpc/mm/kasan/init_book3s_64.c
69
for (i = 0; i < PTRS_PER_PMD; i++)
arch/powerpc/mm/ptdump/hashpagetable.c
420
for (i = 0; i < PTRS_PER_PMD; i++, pmd++) {
arch/riscv/include/asm/pgtable-64.h
251
#define pmd_index(addr) (((addr) >> PMD_SHIFT) & (PTRS_PER_PMD - 1))
arch/riscv/kvm/mmu.c
439
hfn &= ~(PTRS_PER_PMD - 1);
arch/riscv/mm/init.c
1595
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/riscv/mm/init.c
462
static pmd_t trampoline_pmd[PTRS_PER_PMD] __page_aligned_bss;
arch/riscv/mm/init.c
463
static pmd_t fixmap_pmd[PTRS_PER_PMD] __page_aligned_bss;
arch/riscv/mm/init.c
464
static pmd_t early_pmd[PTRS_PER_PMD] __initdata __aligned(PAGE_SIZE);
arch/riscv/mm/kasan_init.c
333
for (i = 0; i < PTRS_PER_PMD; ++i)
arch/riscv/mm/kasan_init.c
57
p = memblock_alloc_or_panic(PTRS_PER_PMD * sizeof(pmd_t), PAGE_SIZE);
arch/riscv/mm/pageattr.c
163
for (i = 0; i < PTRS_PER_PMD; ++i, ++pmdp_new)
arch/riscv/mm/pgtable.c
97
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/s390/include/asm/pgtable.h
1355
#define pmd_index(address) (((address) >> PMD_SHIFT) & (PTRS_PER_PMD-1))
arch/s390/mm/pageattr.c
217
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/s390/mm/pageattr.c
224
update_page_count(PG_DIRECT_MAP_1M, PTRS_PER_PMD);
arch/s390/mm/pageattr.c
66
mask = ~(PTRS_PER_PMD * sizeof(pmd_t) - 1);
arch/s390/mm/vmem.c
308
for (i = 0; i < PTRS_PER_PMD; i++, pmd++)
arch/sh/mm/fault.c
96
if (PTRS_PER_PMD != 1)
arch/sh/mm/init.c
190
for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) {
arch/sh/mm/pgtable.c
29
PTRS_PER_PMD * (1<<PTE_MAGNITUDE),
arch/sparc/include/asm/pgtsrmmu.h
21
#define SRMMU_PMD_TABLE_SIZE (PTRS_PER_PMD*4)
arch/x86/boot/compressed/ident_map_64.c
210
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/x86/boot/startup/map_kernel.c
14
extern pmd_t early_dynamic_pgts[EARLY_DYNAMIC_PAGE_TABLES][PTRS_PER_PMD];
arch/x86/boot/startup/map_kernel.c
182
pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE;
arch/x86/boot/startup/map_kernel.c
213
for (; i < PTRS_PER_PMD; i++)
arch/x86/boot/startup/map_kernel.c
64
early_snp_set_memory_shared(paddr, paddr, PTRS_PER_PMD);
arch/x86/boot/startup/map_kernel.c
90
pmd_t (*early_pgts)[PTRS_PER_PMD] = rip_rel_ptr(early_dynamic_pgts);
arch/x86/boot/startup/sme.c
136
memset(pmd, 0, sizeof(*pmd) * PTRS_PER_PMD);
arch/x86/boot/startup/sme.c
137
ppd->pgtable_area += sizeof(*pmd) * PTRS_PER_PMD;
arch/x86/boot/startup/sme.c
269
entries += (DIV_ROUND_UP(len, PUD_SIZE) + 1) * sizeof(pmd_t) * PTRS_PER_PMD;
arch/x86/boot/startup/sme.c
280
tables += DIV_ROUND_UP(entries, PUD_SIZE) * sizeof(pmd_t) * PTRS_PER_PMD;
arch/x86/include/asm/pgtable_32.h
58
#if PTRS_PER_PMD > 1
arch/x86/include/asm/pgtable_32.h
59
#define PAGE_TABLE_SIZE(pages) (((pages) / PTRS_PER_PMD) + PTRS_PER_PGD)
arch/x86/kernel/espfix_64.c
90
#define ESPFIX_PMD_CLONES PTRS_PER_PMD
arch/x86/kernel/head64.c
141
memset(pmd_p, 0, sizeof(*pmd_p) * PTRS_PER_PMD);
arch/x86/kernel/head64.c
49
extern pmd_t early_dynamic_pgts[EARLY_DYNAMIC_PAGE_TABLES][PTRS_PER_PMD];
arch/x86/kvm/svm/sev.c
4835
if (IS_ALIGNED(pfn, PTRS_PER_PMD))
arch/x86/kvm/svm/sev.c
4929
kvm_zap_gfn_range(kvm, gfn, gfn + PTRS_PER_PMD);
arch/x86/kvm/svm/sev.c
4973
kvm_pfn_t pfn_aligned = ALIGN_DOWN(pfn, PTRS_PER_PMD);
arch/x86/kvm/svm/sev.c
4981
is_pfn_range_shared(pfn_aligned, pfn_aligned + PTRS_PER_PMD))
arch/x86/kvm/svm/sev.c
5013
pfn_aligned = ALIGN_DOWN(pfn, PTRS_PER_PMD);
arch/x86/kvm/svm/sev.c
5014
gfn_aligned = ALIGN_DOWN(gfn, PTRS_PER_PMD);
arch/x86/kvm/svm/sev.c
5052
use_2m_update = IS_ALIGNED(pfn, PTRS_PER_PMD) &&
arch/x86/kvm/svm/sev.c
5053
end >= (pfn + PTRS_PER_PMD) &&
arch/x86/kvm/svm/sev.c
5092
pfn += use_2m_update ? PTRS_PER_PMD : 1;
arch/x86/mm/dump_pagetables.c
154
#define PUD_LEVEL_MULT (PTRS_PER_PMD * PMD_LEVEL_MULT)
arch/x86/mm/ident_map.c
19
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/x86/mm/init_32.c
142
for (; (pmd_idx < PTRS_PER_PMD) && (vaddr != end);
arch/x86/mm/init_32.c
225
for (; (pmd_idx < PTRS_PER_PMD) && (vaddr != end);
arch/x86/mm/init_32.c
304
for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn;
arch/x86/mm/init_64.c
1063
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/x86/mm/init_64.c
531
for (; i < PTRS_PER_PMD; i++, paddr = paddr_next) {
arch/x86/mm/kasan_init_64.c
305
for (i = 0; i < PTRS_PER_PMD; i++)
arch/x86/mm/pat/set_memory.c
1341
for (i = 1, pmd++; i < PTRS_PER_PMD; i++, pmd++) {
arch/x86/mm/pat/set_memory.c
1419
for (i = 0; i < PTRS_PER_PMD; i++)
arch/x86/mm/pgtable.c
267
sizeof(pmd_t) * PTRS_PER_PMD);
arch/x86/mm/pgtable.c
293
sizeof(pmd_t) * PTRS_PER_PMD);
arch/x86/mm/pgtable.c
740
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/x86/mm/pgtable.c
751
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/x86/power/hibernate_32.c
101
for (pmd_idx = 0; pmd_idx < PTRS_PER_PMD; pmd++, pmd_idx++) {
arch/x86/virt/svm/sev.c
818
pfn_i = ALIGN_DOWN(pfn, PTRS_PER_PMD);
arch/x86/virt/svm/sev.c
819
pfn_end = pfn_i + PTRS_PER_PMD;
arch/x86/virt/svm/sev.c
931
(!IS_ALIGNED(pfn, PTRS_PER_PMD) || !pfn_valid(pfn + PTRS_PER_PMD - 1)))
arch/x86/xen/mmu_pv.c
1075
for (; vaddr <= vaddr_end && (pmd < (level2_kernel_pgt + PTRS_PER_PMD));
arch/x86/xen/mmu_pv.c
109
static pmd_t level2_ident_pgt[PTRS_PER_PMD] __page_aligned_bss;
arch/x86/xen/mmu_pv.c
1148
for (i = 0; i < PTRS_PER_PMD; i++) {
arch/x86/xen/mmu_pv.c
1825
for (i = 0; i < PTRS_PER_PMD; ++i) {
arch/x86/xen/mmu_pv.c
1999
for (idx_pt = 0; idx_pt < min(n_pt, PTRS_PER_PMD);
arch/x86/xen/mmu_pv.c
2018
n_pt -= PTRS_PER_PMD;
arch/x86/xen/mmu_pv.c
608
nr = last ? pmd_index(limit) + 1 : PTRS_PER_PMD;
arch/x86/xen/mmu_pv.c
630
if (PTRS_PER_PMD > 1)
drivers/hv/mshv_regions.c
20
#define MSHV_MAP_FAULT_IN_PAGES PTRS_PER_PMD
drivers/hv/mshv_regions.c
45
!IS_ALIGNED(gfn, PTRS_PER_PMD) ||
drivers/hv/mshv_regions.c
46
!IS_ALIGNED(page_count, PTRS_PER_PMD))
drivers/iommu/amd/init.c
3981
if (!(pfn % PTRS_PER_PMD)) {
include/linux/mm.h
3256
atomic_long_add(PTRS_PER_PMD * sizeof(pmd_t), &mm->pgtables_bytes);
include/linux/mm.h
3263
atomic_long_sub(PTRS_PER_PMD * sizeof(pmd_t), &mm->pgtables_bytes);
include/linux/mm.h
3587
unsigned long mask = ~(PTRS_PER_PMD * sizeof(pmd_t) - 1);
include/linux/pgtable.h
2321
#define MAX_PTRS_PER_PMD PTRS_PER_PMD
include/linux/pgtable.h
56
return (address >> PMD_SHIFT) & (PTRS_PER_PMD - 1);
mm/kasan/init.c
304
for (i = 0; i < PTRS_PER_PMD; i++) {
mm/pagewalk.c
21
if (depth == 3 && PTRS_PER_PMD == 1)
mm/vmscan.c
3729
if (i < PTRS_PER_PMD && get_next_vma(PUD_MASK, PMD_SIZE, args, &start, &end))