pte_page
static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, pgtable_t pte_page)
set_pmd(pmd, __pmd((unsigned long)page_address(pte_page)));
struct page *page = pte_page(pte);
page = pte_page(pte);
struct folio *folio = page_folio(pte_page(pte));
free_hotplug_page_range(pte_page(pte),
page = pte_page(pte);
address = (unsigned long) page_address(pte_page(*ptep));
unsigned long paddr = (unsigned long)page_to_virt(pte_page(pte));
#define pmd_populate(mm, pmd, pte_page) \
pmd_populate_kernel(mm, pmd, page_address(pte_page))
pgtable_t pte_page)
*pmdp = __pmd(__pa(pte_page) | _PMD_PRESENT);
pgtable_t pte_page)
*pmd = __pmd(__pgtable_ptr_val(pte_page) | PMD_VAL_BITS);
pgtable_t pte_page)
*pmdp = __pmd((unsigned long)pte_page | _PMD_PRESENT);
*pmdp = __pmd(__pa(pte_page) | _PMD_USER | _PMD_PRESENT);
pgtable_t pte_page)
pmd_set(pmd, (unsigned long)pte_page);
folio = page_folio(pte_page(pte));
pte = radix__vmemmap_pte_populate(pmd, addr, node, NULL, pte_page(*tail_page_pte));
free_vmemmap_pages(pte_page(*pte), altmap, 0);
free_vmemmap_pages(pte_page(*pte), altmap, 0);
if (pte_page(*ptep) != virt_to_page(lm_alias(kasan_early_shadow_page)))
return pte_page(p4d_pte(p4d));
return pte_page(pud_pte(pud));
return pte_page(pmd_pte(pmd));
struct folio *folio = page_folio(pte_page(pte));
free_vmemmap_storage(pte_page(pte), PAGE_SIZE, altmap);
struct page *pte_page;
pte_page = alloc_page(GFP_KERNEL);
if (!pte_page)
ptep_new = (pte_t *)page_address(pte_page);
set_pmd(pmdp, pfn_pmd(page_to_pfn(pte_page), PAGE_TABLE));
paddr = (unsigned long) page_address(pte_page(pte));
page = pte_page(*pte);
page = pte_page(*pte);
page = pte_page(*pte);
free_pagetable(pte_page(*pte), 0);
get_page_bootmem(section_nr, pte_page(*pte),
phys_pte_init(pte_t *pte_page, unsigned long paddr, unsigned long paddr_end,
pte = pte_page + pte_index(paddr);
folio = page_folio(pte_page(ptent));
struct folio *folio = page_folio(pte_page(pte));
if (!PageAnon(pte_page(pte)))
page = pte_page(huge_pte);
static inline void pte_free(struct mm_struct *mm, struct page *pte_page)
struct ptdesc *ptdesc = page_ptdesc(pte_page);
page = pte_page(old_pte);
page = pte_page(pte);
*page = pte_page(entry);
page = pte_page(pte);
page = pte_page(pte);
return pte_page(ptep_get(&pkmap_page_table[PKMAP_NR(addr)]));
return pte_page(pteval);
page = pte_page(ptent);
kunmap_high(pte_page(ptep_get(&pkmap_page_table[PKMAP_NR(vaddr)])));
pte_folio = page_folio(pte_page(entry));
if (folio != page_folio(pte_page(pte))) {
folio = page_folio(pte_page(pte));
old_folio = page_folio(pte_page(pte));
folio = page_folio(pte_page(vmf.orig_pte));
pte_page(ptep_get(pte_offset_kernel(pmd, addr)));
vmemmap_walk->reuse_page = pte_page(ptep_get(pte));
struct page *page = pte_page(ptep_get(pte));
BUG_ON(pte_page(ptep_get(pte)) != walk->reuse_page);
return pte_page(pte) == virt_to_page(lm_alias(kasan_early_shadow_page));
struct page *src_page = pte_page(pteval);
src_page = pte_page(pteval);
page = pte_page(pte);