Symbol: PTRS_PER_PTE
arch/arc/mm/highmem.c
68
BUILD_BUG_ON(LAST_PKMAP > PTRS_PER_PTE);
arch/arc/mm/highmem.c
69
BUILD_BUG_ON(FIX_KMAP_SLOTS > PTRS_PER_PTE);
arch/arc/mm/init.c
170
BUILD_BUG_ON((PTRS_PER_PTE * sizeof(pte_t)) > PAGE_SIZE);
arch/arm/include/asm/highmem.h
9
#define LAST_PKMAP PTRS_PER_PTE
arch/arm/include/asm/kfence.h
20
for (i = 0; i < PTRS_PER_PTE; i++)
arch/arm/include/asm/pgtable-2level.h
74
#define PTE_HWTABLE_PTRS (PTRS_PER_PTE)
arch/arm/include/asm/pgtable-2level.h
76
#define PTE_HWTABLE_SIZE (PTRS_PER_PTE * sizeof(u32))
arch/arm/include/asm/pgtable-3level.h
26
#define PTE_HWTABLE_SIZE (PTRS_PER_PTE * sizeof(u64))
arch/arm/mm/dump.c
317
for (i = 0; i < PTRS_PER_PTE; i++, pte++) {
arch/arm/mm/kasan_init.c
292
for (i = 0; i < PTRS_PER_PTE; i++)
arch/arm/mm/mmu.c
357
static pte_t bm_pte[PTRS_PER_PTE + PTE_HWTABLE_PTRS]
arch/arm64/include/asm/kvm_pkvm.h
115
nr_pages = DIV_ROUND_UP(nr_pages, PTRS_PER_PTE);
arch/arm64/include/asm/tlbflush.h
395
#define MAX_DVM_OPS PTRS_PER_PTE
arch/arm64/kernel/pi/map_range.c
42
tbl += (start >> (lshift + PAGE_SHIFT)) % PTRS_PER_PTE;
arch/arm64/kernel/pi/map_range.c
62
*pte += PTRS_PER_PTE * sizeof(pte_t);
arch/arm64/kvm/hyp/pgtable.c
1468
return PTRS_PER_PTE + 1;
arch/arm64/kvm/hyp/pgtable.c
231
for (idx = kvm_pgtable_idx(data, level); idx < PTRS_PER_PTE; ++idx) {
arch/arm64/kvm/hyp/pgtable.c
258
kvm_pteref_t pteref = &pgt->pgd[idx * PTRS_PER_PTE];
arch/arm64/mm/fixmap.c
34
static pte_t bm_pte[NR_BM_PTE_TABLES][PTRS_PER_PTE] __page_aligned_bss;
arch/arm64/mm/kasan_init.c
217
static pte_t tbl[PTRS_PER_PTE] __page_aligned_bss;
arch/arm64/mm/kasan_init.c
26
static pgd_t tmp_pg_dir[PTRS_PER_PTE] __initdata __aligned(PAGE_SIZE);
arch/arm64/mm/kasan_init.c
274
return (addr >> (shift + PAGE_SHIFT)) % PTRS_PER_PTE;
arch/arm64/mm/kasan_init.c
294
clear_next_level(l++, next_level_idx(start), PTRS_PER_PTE);
arch/arm64/mm/kasan_init.c
368
for (i = 0; i < PTRS_PER_PTE; i++)
arch/arm64/mm/mmu.c
1606
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/arm64/mm/mmu.c
609
for (i = 0; i < PTRS_PER_PTE; i++, ptep++, pfn++)
arch/csky/include/asm/pgtable.h
83
extern pte_t invalid_pte_table[PTRS_PER_PTE];
arch/csky/mm/init.c
34
((PTRS_PER_PGD - USER_PTRS_PER_PGD) * PTRS_PER_PTE)
arch/csky/mm/init.c
37
pte_t invalid_pte_table[PTRS_PER_PTE] __page_aligned_bss;
arch/csky/mm/init.c
70
__pa(kernel_pte_tables + (PTRS_PER_PTE * (i - USER_PTRS_PER_PGD)));
arch/hexagon/include/asm/mem-layout.h
84
#define LAST_PKMAP PTRS_PER_PTE
arch/loongarch/include/asm/kvm_mmu.h
111
return table + ((addr >> ctx->pgtable_shift) & (PTRS_PER_PTE - 1));
arch/loongarch/include/asm/pgtable.h
109
min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, (1UL << cpu_vabits)) - PMD_SIZE - VMEMMAP_SIZE - KFENCE_AREA_SIZE)
arch/loongarch/include/asm/pgtable.h
113
min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, (1UL << cpu_vabits) / 2) - PMD_SIZE - VMEMMAP_SIZE - KFENCE_AREA_SIZE)
arch/loongarch/include/asm/pgtable.h
140
extern pte_t invalid_pte_table[PTRS_PER_PTE];
arch/loongarch/kernel/asm-offsets.c
212
DEFINE(_PTRS_PER_PTE, PTRS_PER_PTE);
arch/loongarch/kvm/mmu.c
128
end = p + PTRS_PER_PTE;
arch/loongarch/kvm/mmu.c
738
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/loongarch/kvm/mmu.c
748
kvm->stat.pages += PTRS_PER_PTE;
arch/loongarch/kvm/mmu.c
750
return child + (gfn & (PTRS_PER_PTE - 1));
arch/loongarch/kvm/mmu.c
881
gfn = gfn & ~(PTRS_PER_PTE - 1);
arch/loongarch/kvm/mmu.c
882
pfn = pfn & ~(PTRS_PER_PTE - 1);
arch/loongarch/mm/init.c
220
pte_t invalid_pte_table[PTRS_PER_PTE] __page_aligned_bss;
arch/loongarch/mm/kasan_init.c
322
for (i = 0; i < PTRS_PER_PTE; i++)
arch/loongarch/mm/pgtable.c
123
end = p + PTRS_PER_PTE;
arch/m68k/mm/mcfmmu.c
58
next_pgtable += PTRS_PER_PTE * sizeof(pte_t);
arch/m68k/mm/mcfmmu.c
63
for (i = 0; i < PTRS_PER_PTE; ++i, ++pg_table) {
arch/m68k/mm/motorola.c
253
last_pte_table += PTRS_PER_PTE;
arch/m68k/mm/motorola.c
366
for (i = 1; i < PTRS_PER_PTE; physaddr += PAGE_SIZE, i++)
arch/m68k/mm/sun3mmu.c
64
next_pgtable += PTRS_PER_PTE * sizeof (pte_t);
arch/m68k/mm/sun3mmu.c
70
for (i=0; i<PTRS_PER_PTE; ++i, ++pg_table) {
arch/mips/include/asm/pgtable-32.h
119
extern pte_t invalid_pte_table[PTRS_PER_PTE];
arch/mips/include/asm/pgtable-64.h
144
min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, \
arch/mips/include/asm/pgtable-64.h
167
extern pte_t invalid_pte_table[PTRS_PER_PTE];
arch/mips/kernel/asm-offsets.c
220
DEFINE(_PTRS_PER_PTE, PTRS_PER_PTE);
arch/mips/kvm/mmu.c
159
bool safe_to_remove = (i_min == 0 && i_max == PTRS_PER_PTE - 1);
arch/mips/mm/init.c
562
pte_t invalid_pte_table[PTRS_PER_PTE] __page_aligned_bss;
arch/mips/mm/tlbex.c
2016
uasm_i_andi(p, wr.r1, wr.r1, (PTRS_PER_PTE - 1) << PTE_T_LOG2);
arch/mips/mm/tlbex.c
2443
pwsize |= ilog2(PTRS_PER_PTE) << MIPS_PWSIZE_PTW_SHIFT;
arch/mips/mm/tlbex.c
958
unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1);
arch/nios2/mm/init.c
67
pte_t invalid_pte_table[PTRS_PER_PTE] __aligned(PAGE_SIZE);
arch/openrisc/mm/init.c
102
for (j = 0; p < e && j < PTRS_PER_PTE;
arch/parisc/mm/init.c
369
start_pte = ((start_vaddr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1));
arch/parisc/mm/init.c
394
for (tmp2 = start_pte; tmp2 < PTRS_PER_PTE; tmp2++, pg_table++) {
arch/parisc/mm/init.c
47
pte_t pg0[PT_INITIAL * PTRS_PER_PTE] __section(".data..vm0.pte") __attribute__ ((aligned(PAGE_SIZE)));
arch/powerpc/kvm/book3s_64_mmu_radix.c
485
for (it = 0; it < PTRS_PER_PTE; ++it, ++p) {
arch/powerpc/mm/book3s64/hash_4k.c
122
new_pte |= pte_set_hidx(ptep, rpte, 0, slot, PTRS_PER_PTE);
arch/powerpc/mm/book3s64/hash_4k.c
60
rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE);
arch/powerpc/mm/book3s64/hash_64k.c
218
new_pte |= pte_set_hidx(ptep, rpte, subpg_index, slot, PTRS_PER_PTE);
arch/powerpc/mm/book3s64/hash_64k.c
269
rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE);
arch/powerpc/mm/book3s64/hash_64k.c
334
new_pte |= pte_set_hidx(ptep, rpte, 0, slot, PTRS_PER_PTE);
arch/powerpc/mm/book3s64/hash_64k.c
91
rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE);
arch/powerpc/mm/book3s64/hash_tlb.c
246
for (pte = start_pte; pte < start_pte + PTRS_PER_PTE; pte++) {
arch/powerpc/mm/book3s64/hash_tlb.c
84
offset = PTRS_PER_PTE;
arch/powerpc/mm/book3s64/hash_utils.c
1865
pte_val(*(ptep + PTRS_PER_PTE)));
arch/powerpc/mm/book3s64/hash_utils.c
1930
pte_val(*(ptep + PTRS_PER_PTE)));
arch/powerpc/mm/book3s64/radix_pgtable.c
698
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/powerpc/mm/book3s64/subpage_prot.c
122
i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1);
arch/powerpc/mm/book3s64/subpage_prot.c
123
nw = PTRS_PER_PTE - i;
arch/powerpc/mm/book3s64/subpage_prot.c
261
i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1);
arch/powerpc/mm/book3s64/subpage_prot.c
262
nw = PTRS_PER_PTE - i;
arch/powerpc/mm/kasan/8xx.c
29
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/powerpc/mm/kasan/init_32.c
27
for (i = 0; i < PTRS_PER_PTE; i++, ptep++)
arch/powerpc/mm/kasan/init_book3e_64.c
120
for (i = 0; i < PTRS_PER_PTE; i++)
arch/powerpc/mm/kasan/init_book3e_64.c
92
for (i = 0; i < PTRS_PER_PTE; i++)
arch/powerpc/mm/kasan/init_book3s_64.c
65
for (i = 0; i < PTRS_PER_PTE; i++)
arch/powerpc/mm/kasan/init_book3s_64.c
82
for (i = 0; i < PTRS_PER_PTE; i++)
arch/powerpc/mm/pgtable_32.c
45
addr += PGDIR_SIZE, ptep += PTRS_PER_PTE, pmdp++)
arch/powerpc/mm/ptdump/hashpagetable.c
386
for (i = 0; i < PTRS_PER_PTE; i++, pte++) {
arch/riscv/kvm/gstage.c
238
for (i = 0; i < PTRS_PER_PTE; i++)
arch/riscv/kvm/gstage.c
31
mask = (PTRS_PER_PTE * (1UL << kvm_riscv_gstage_pgd_xbits)) - 1;
arch/riscv/kvm/gstage.c
33
mask = PTRS_PER_PTE - 1;
arch/riscv/kvm/tlb.c
128
if (PTRS_PER_PTE < (gvsz >> order)) {
arch/riscv/kvm/tlb.c
29
if (PTRS_PER_PTE < (gpsz >> order)) {
arch/riscv/kvm/tlb.c
57
if (PTRS_PER_PTE < (gpsz >> order)) {
arch/riscv/kvm/tlb.c
88
if (PTRS_PER_PTE < (gvsz >> order)) {
arch/riscv/mm/init.c
1574
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/riscv/mm/init.c
359
static pte_t fixmap_pte[PTRS_PER_PTE] __page_aligned_bss;
arch/riscv/mm/kasan_init.c
329
for (i = 0; i < PTRS_PER_PTE; ++i)
arch/riscv/mm/kasan_init.c
35
p = memblock_alloc_or_panic(PTRS_PER_PTE * sizeof(pte_t), PAGE_SIZE);
arch/riscv/mm/kasan_init.c
525
for (i = 0; i < PTRS_PER_PTE; i++)
arch/riscv/mm/pageattr.c
123
for (i = 0; i < PTRS_PER_PTE; ++i, ++ptep_new)
arch/s390/boot/vmem.c
109
memset64((u64 *)kasan_early_shadow_pte, pte_val(pte_z), PTRS_PER_PTE);
arch/s390/boot/vmem.c
251
memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE);
arch/s390/include/asm/pgtable.h
1106
pto = __pa(ptep) & ~(PTRS_PER_PTE * sizeof(pte_t) - 1);
arch/s390/include/asm/pgtable.h
1934
((pgtable_t)__va(pmd_val(pmd) & -sizeof(pte_t)*PTRS_PER_PTE))
arch/s390/kvm/dat.h
628
memset64((void *)pt->pgstes, val, PTRS_PER_PTE);
arch/s390/kvm/dat.h
634
memset64((void *)pt->ptes, ptes, PTRS_PER_PTE);
arch/s390/mm/pageattr.c
130
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/s390/mm/pageattr.c
137
update_page_count(PG_DIRECT_MAP_4K, PTRS_PER_PTE);
arch/s390/mm/pageattr.c
450
nr = PTRS_PER_PTE - (nr & (PTRS_PER_PTE - 1));
arch/s390/mm/pageattr.c
69
mask = ~(PTRS_PER_PTE * sizeof(pte_t) - 1);
arch/s390/mm/pgalloc.c
134
memset64((u64 *)table, _PAGE_INVALID, PTRS_PER_PTE);
arch/s390/mm/pgalloc.c
135
memset64((u64 *)table + PTRS_PER_PTE, 0, PTRS_PER_PTE);
arch/s390/mm/pgalloc.c
177
memset64((u64 *)table, _PAGE_INVALID, PTRS_PER_PTE);
arch/s390/mm/vmem.c
214
for (i = 0; i < PTRS_PER_PTE; i++, pte++) {
arch/s390/mm/vmem.c
82
memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE);
arch/sparc/include/asm/pgtable_32.h
51
#define PTE_SIZE (PTRS_PER_PTE*4)
arch/sparc/include/asm/pgtsrmmu.h
20
#define SRMMU_PTE_TABLE_SIZE (PTRS_PER_PTE*4)
arch/sparc/vdso/vma.c
358
offset = get_random_u32_below(PTRS_PER_PTE);
arch/x86/boot/startup/sme.c
176
memset(pte, 0, sizeof(*pte) * PTRS_PER_PTE);
arch/x86/boot/startup/sme.c
177
ppd->pgtable_area += sizeof(*pte) * PTRS_PER_PTE;
arch/x86/boot/startup/sme.c
270
entries += 2 * sizeof(pte_t) * PTRS_PER_PTE;
arch/x86/include/asm/fixmap.h
133
-PTRS_PER_PTE
arch/x86/include/asm/pgtable.h
1478
#define PTE_SHIFT ilog2(PTRS_PER_PTE)
arch/x86/kernel/espfix_64.c
89
#define ESPFIX_PTE_CLONES (PTRS_PER_PTE/PTE_STRIDE)
arch/x86/kernel/head32.c
118
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/x86/mm/dump_pagetables.c
153
#define PMD_LEVEL_MULT (PTRS_PER_PTE * PTE_LEVEL_MULT)
arch/x86/mm/init_32.c
177
for (i = 0; i < PTRS_PER_PTE; i++)
arch/x86/mm/init_32.c
189
&& lastpte && lastpte + PTRS_PER_PTE != pte);
arch/x86/mm/init_32.c
324
addr2 = (pfn + PTRS_PER_PTE-1) * PAGE_SIZE +
arch/x86/mm/init_32.c
337
pfn += PTRS_PER_PTE;
arch/x86/mm/init_32.c
344
for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn;
arch/x86/mm/init_64.c
1045
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/x86/mm/init_64.c
483
for (; i < PTRS_PER_PTE; i++, paddr = paddr_next, pte++) {
arch/x86/mm/kasan_init_64.c
302
for (i = 0; i < PTRS_PER_PTE; i++)
arch/x86/mm/kasan_init_64.c
440
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/x86/mm/pat/set_memory.c
106
direct_pages_count[level - 1] += PTRS_PER_PTE;
arch/x86/mm/pat/set_memory.c
118
direct_pages_count[level - 1] -= PTRS_PER_PTE;
arch/x86/mm/pat/set_memory.c
1183
for (i = 0; i < PTRS_PER_PTE; i++, pfn += pfninc, lpaddr += lpinc)
arch/x86/mm/pat/set_memory.c
1271
for (i = 1, pte++; i < PTRS_PER_PTE; i++, pte++) {
arch/x86/mm/pat/set_memory.c
1348
if (pmd_pfn(entry) != pmd_pfn(first) + i * PTRS_PER_PTE)
arch/x86/mm/pat/set_memory.c
1407
for (i = 0; i < PTRS_PER_PTE; i++)
arch/x86/mm/pgtable.c
574
(FIXMAP_PMD_NUM * PTRS_PER_PTE));
arch/x86/power/hibernate_32.c
111
pfn += PTRS_PER_PTE;
arch/x86/power/hibernate_32.c
119
max_pte = pte + PTRS_PER_PTE;
arch/x86/xen/mmu_pv.c
1125
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/x86/xen/mmu_pv.c
1733
for (i = 0; i < PTRS_PER_PTE; i++)
arch/x86/xen/mmu_pv.c
1846
set_page_prot(level1_fixmap_pgt + i * PTRS_PER_PTE,
arch/x86/xen/mmu_pv.c
2004
idx_pte < min(n_pte, PTRS_PER_PTE);
arch/x86/xen/mmu_pv.c
2010
n_pte -= PTRS_PER_PTE;
arch/x86/xen/p2m.c
339
for (i = 0; i < PTRS_PER_PTE; i++) {
arch/x86/xen/p2m.c
93
#define PMDS_PER_MID_PAGE (P2M_MID_PER_PAGE / PTRS_PER_PTE)
arch/xtensa/include/asm/highmem.h
23
#define LAST_PKMAP (PTRS_PER_PTE * DCACHE_N_COLORS)
arch/xtensa/include/asm/pgalloc.h
39
for (i = 0; i < PTRS_PER_PTE; i++)
arch/xtensa/mm/kasan_init.c
24
for (i = 0; i < PTRS_PER_PTE; ++i)
arch/xtensa/mm/kasan_init.c
38
unsigned long n_pmds = n_pages / PTRS_PER_PTE;
arch/xtensa/mm/kasan_init.c
49
for (k = 0; k < PTRS_PER_PTE; ++k, ++j) {
arch/xtensa/mm/kasan_init.c
62
for (i = 0; i < n_pmds ; ++i, pte += PTRS_PER_PTE)
arch/xtensa/mm/kasan_init.c
87
for (i = 0; i < PTRS_PER_PTE; ++i)
arch/xtensa/mm/mmu.c
30
n_pages = ALIGN(n_pages, PTRS_PER_PTE);
arch/xtensa/mm/mmu.c
43
for (i = 0; i < n_pages; i += PTRS_PER_PTE, ++pmd) {
drivers/iommu/riscv/iommu.c
1103
for (i = 0; i < PTRS_PER_PTE; i++) {
drivers/iommu/riscv/iommu.c
1127
ptr += ((iova >> shift) & (PTRS_PER_PTE - 1));
drivers/iommu/riscv/iommu.c
1176
ptr += ((iova >> shift) & (PTRS_PER_PTE - 1));
include/linux/mm.h
3280
atomic_long_add(PTRS_PER_PTE * sizeof(pte_t), &mm->pgtables_bytes);
include/linux/mm.h
3285
atomic_long_sub(PTRS_PER_PTE * sizeof(pte_t), &mm->pgtables_bytes);
include/linux/pgtable.h
2317
#define MAX_PTRS_PER_PTE PTRS_PER_PTE
include/linux/pgtable.h
50
return (address >> PAGE_SHIFT) & (PTRS_PER_PTE - 1);
lib/test_hmm.c
1052
src_pfns = kvcalloc(PTRS_PER_PTE, sizeof(*src_pfns), GFP_KERNEL | __GFP_NOFAIL);
lib/test_hmm.c
1053
dst_pfns = kvcalloc(PTRS_PER_PTE, sizeof(*dst_pfns), GFP_KERNEL | __GFP_NOFAIL);
lib/test_hmm.c
1072
next = min(end, addr + (PTRS_PER_PTE << PAGE_SHIFT));
lib/test_hmm.c
1128
src_pfns = kvcalloc(PTRS_PER_PTE, sizeof(*src_pfns),
lib/test_hmm.c
1133
dst_pfns = kvcalloc(PTRS_PER_PTE, sizeof(*dst_pfns),
lib/test_hmm.c
1146
next = min(end, addr + (PTRS_PER_PTE << PAGE_SHIFT));
mm/hugetlb_vmemmap.c
66
for (i = 0; i < PTRS_PER_PTE; i++, addr += PAGE_SIZE) {
mm/kasan/init.c
289
for (i = 0; i < PTRS_PER_PTE; i++) {
mm/memory.c
1875
for (i = 0, pte = start_pte; i < PTRS_PER_PTE; i++, pte++) {
mm/memory.c
2435
remaining_pages_total, PTRS_PER_PTE - pte_index(addr));
mm/memory.c
5637
pte_off + (nr_pages - idx) > PTRS_PER_PTE)) {
mm/memory.c
5689
if (val / PAGE_SIZE > PTRS_PER_PTE)
mm/memory.c
5747
to_pte = min3(from_pte + nr_pages, (pgoff_t)PTRS_PER_PTE,
mm/page_table_check.c
278
for (i = 0; i < PTRS_PER_PTE; i++) {
mm/page_table_check.c
283
pte_unmap(ptep - PTRS_PER_PTE);
mm/vmscan.c
3563
if (i < PTRS_PER_PTE && get_next_vma(PMD_MASK, PAGE_SIZE, args, &start, &end))