Symbol: pte_page
arch/arc/include/asm/pgalloc.h
49
static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, pgtable_t pte_page)
arch/arc/include/asm/pgalloc.h
51
set_pmd(pmd, __pmd((unsigned long)page_address(pte_page)));
arch/arm64/kernel/mte.c
40
struct page *page = pte_page(pte);
arch/arm64/mm/contpte.c
268
page = pte_page(pte);
arch/arm64/mm/flush.c
54
struct folio *folio = page_folio(pte_page(pte));
arch/arm64/mm/mmu.c
1463
free_hotplug_page_range(pte_page(pte),
arch/hexagon/mm/init.c
51
page = pte_page(pte);
arch/microblaze/kernel/signal.c
198
address = (unsigned long) page_address(pte_page(*ptep));
arch/nios2/include/asm/pgtable.h
189
unsigned long paddr = (unsigned long)page_to_virt(pte_page(pte));
arch/parisc/include/asm/pgalloc.h
57
#define pmd_populate(mm, pmd, pte_page) \
arch/parisc/include/asm/pgalloc.h
58
pmd_populate_kernel(mm, pmd, page_address(pte_page))
arch/powerpc/include/asm/book3s/32/pgalloc.h
41
pgtable_t pte_page)
arch/powerpc/include/asm/book3s/32/pgalloc.h
43
*pmdp = __pmd(__pa(pte_page) | _PMD_PRESENT);
arch/powerpc/include/asm/book3s/64/pgalloc.h
165
pgtable_t pte_page)
arch/powerpc/include/asm/book3s/64/pgalloc.h
167
*pmd = __pmd(__pgtable_ptr_val(pte_page) | PMD_VAL_BITS);
arch/powerpc/include/asm/nohash/32/pgalloc.h
27
pgtable_t pte_page)
arch/powerpc/include/asm/nohash/32/pgalloc.h
30
*pmdp = __pmd((unsigned long)pte_page | _PMD_PRESENT);
arch/powerpc/include/asm/nohash/32/pgalloc.h
32
*pmdp = __pmd(__pa(pte_page) | _PMD_USER | _PMD_PRESENT);
arch/powerpc/include/asm/nohash/64/pgalloc.h
46
pgtable_t pte_page)
arch/powerpc/include/asm/nohash/64/pgalloc.h
48
pmd_set(pmd, (unsigned long)pte_page);
arch/powerpc/mm/book3s64/hash_utils.c
1581
folio = page_folio(pte_page(pte));
arch/powerpc/mm/book3s64/radix_pgtable.c
1400
pte = radix__vmemmap_pte_populate(pmd, addr, node, NULL, pte_page(*tail_page_pte));
arch/powerpc/mm/book3s64/radix_pgtable.c
805
free_vmemmap_pages(pte_page(*pte), altmap, 0);
arch/powerpc/mm/book3s64/radix_pgtable.c
811
free_vmemmap_pages(pte_page(*pte), altmap, 0);
arch/powerpc/mm/kasan/init_32.c
92
if (pte_page(*ptep) != virt_to_page(lm_alias(kasan_early_shadow_page)))
arch/powerpc/mm/pgtable_64.c
106
return pte_page(p4d_pte(p4d));
arch/powerpc/mm/pgtable_64.c
117
return pte_page(pud_pte(pud));
arch/powerpc/mm/pgtable_64.c
136
return pte_page(pmd_pte(pmd));
arch/riscv/mm/cacheflush.c
102
struct folio *folio = page_folio(pte_page(pte));
arch/riscv/mm/init.c
1668
free_vmemmap_storage(pte_page(pte), PAGE_SIZE, altmap);
arch/riscv/mm/pageattr.c
112
struct page *pte_page;
arch/riscv/mm/pageattr.c
118
pte_page = alloc_page(GFP_KERNEL);
arch/riscv/mm/pageattr.c
119
if (!pte_page)
arch/riscv/mm/pageattr.c
122
ptep_new = (pte_t *)page_address(pte_page);
arch/riscv/mm/pageattr.c
128
set_pmd(pmdp, pfn_pmd(page_to_pfn(pte_page), PAGE_TABLE));
arch/sparc/kernel/signal32.c
337
paddr = (unsigned long) page_address(pte_page(pte));
arch/um/kernel/skas/uaccess.c
266
page = pte_page(*pte);
arch/um/kernel/skas/uaccess.c
345
page = pte_page(*pte);
arch/um/kernel/skas/uaccess.c
75
page = pte_page(*pte);
arch/x86/mm/init_64.c
1121
free_pagetable(pte_page(*pte), 0);
arch/x86/mm/init_64.c
1633
get_page_bootmem(section_nr, pte_page(*pte),
arch/x86/mm/init_64.c
472
phys_pte_init(pte_t *pte_page, unsigned long paddr, unsigned long paddr_end,
arch/x86/mm/init_64.c
480
pte = pte_page + pte_index(paddr);
fs/proc/task_mmu.c
1246
folio = page_folio(pte_page(ptent));
fs/proc/task_mmu.c
2140
struct folio *folio = page_folio(pte_page(pte));
fs/proc/task_mmu.c
2491
if (!PageAnon(pte_page(pte)))
fs/proc/task_mmu.c
3271
page = pte_page(huge_pte);
include/asm-generic/pgalloc.h
114
static inline void pte_free(struct mm_struct *mm, struct page *pte_page)
include/asm-generic/pgalloc.h
116
struct ptdesc *ptdesc = page_ptdesc(pte_page);
kernel/bpf/arena.c
157
page = pte_page(old_pte);
kernel/bpf/arena.c
253
page = pte_page(pte);
mm/gup.c
1070
*page = pte_page(entry);
mm/gup.c
2862
page = pte_page(pte);
mm/gup.c
840
page = pte_page(pte);
mm/highmem.c
175
return pte_page(ptep_get(&pkmap_page_table[PKMAP_NR(addr)]));
mm/highmem.c
189
return pte_page(pteval);
mm/highmem.c
229
page = pte_page(ptent);
mm/highmem.c
528
kunmap_high(pte_page(ptep_get(&pkmap_page_table[PKMAP_NR(vaddr)])));
mm/hugetlb.c
4981
pte_folio = page_folio(pte_page(entry));
mm/hugetlb.c
5248
if (folio != page_folio(pte_page(pte))) {
mm/hugetlb.c
5259
folio = page_folio(pte_page(pte));
mm/hugetlb.c
5480
old_folio = page_folio(pte_page(pte));
mm/hugetlb.c
6114
folio = page_folio(pte_page(vmf.orig_pte));
mm/hugetlb_vmemmap.c
127
pte_page(ptep_get(pte_offset_kernel(pmd, addr)));
mm/hugetlb_vmemmap.c
149
vmemmap_walk->reuse_page = pte_page(ptep_get(pte));
mm/hugetlb_vmemmap.c
216
struct page *page = pte_page(ptep_get(pte));
mm/hugetlb_vmemmap.c
263
BUG_ON(pte_page(ptep_get(pte)) != walk->reuse_page);
mm/kasan/init.c
77
return pte_page(pte) == virt_to_page(lm_alias(kasan_early_shadow_page));
mm/khugepaged.c
698
struct page *src_page = pte_page(pteval);
mm/khugepaged.c
794
src_page = pte_page(pteval);
mm/vmalloc.c
849
page = pte_page(pte);