pte_pfn
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
struct page *page = pfn_to_page(pte_pfn(*ptep));
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
unsigned long pfn = pte_pfn(*ptep);
pfn = pte_pfn(pteval);
map.pfn = pte_pfn(*pte);
#define pte_page(pte) (pfn_to_page(pte_pfn(pte)))
bool palign = (pte_pfn(pte) & contmask) == contmask;
unsigned long pfn = pte_pfn(pte);
return pfn_pte(pte_pfn(pte) + nr, pte_pgprot(pte));
pfn = ALIGN_DOWN(pte_pfn(pte), CONT_PTES);
return pte_valid_cont(pte) && pte_pfn(pte) == pfn &&
pfn = pte_pfn(orig_pte) - (orig_ptep - ptep);
pfn = pte_pfn(pte);
pte = pfn_pte(ALIGN_DOWN(pte_pfn(pte), CONT_PTES), pte_pgprot(pte));
if (pte_pfn(__pte(old)) != pte_pfn(__pte(new)))
unsigned long pfn = pte_pfn(*ptep);
unsigned long pfn = pte_pfn(*pte);
#define pte_page(x) pfn_to_page(pte_pfn(x))
#define pte_page(x) pfn_to_page(pte_pfn(x))
#define pte_page(x) pfn_to_page(pte_pfn(x))
return phys_to_page(pfn_to_phys(pte_pfn(*virt_to_kpte(kaddr))));
#define pte_page(x) pfn_to_page(pte_pfn(x))
#define pte_page(x) pfn_to_page(pte_pfn(x))
(pte_pfn(ptep[i]) == pte_pfn(pte)))
pfn = pte_pfn(pte);
#define pte_page(pte) (pfn_to_page(pte_pfn(pte)))
unsigned long pfn = pte_pfn(pte);
#define pte_page(pte) (pfn_to_page(pte_pfn(pte)))
unsigned long pfn = pte_pfn(pte);
__flush_cache_page(vma, vmaddr, PFN_PHYS(pte_pfn(pte)));
__flush_cache_page(vma, addr, PFN_PHYS(pte_pfn(pte)));
unsigned long pfn = pte_pfn(pte);
#define pmd_pfn(pmd) pte_pfn(pmd_pte(pmd))
#define pud_pfn(pud) pte_pfn(pud_pte(pud))
#define pte_page(x) pfn_to_page(pte_pfn(x))
pa = PFN_PHYS(pte_pfn(*ptep));
pfn = pte_pfn(pte);
pfn = pte_pfn(__pte(pte_val(pte) | (addr & rpnmask)));
hpa = pte_pfn(pte) << PAGE_SHIFT;
pte_pfn(*p) << PAGE_SHIFT,
pte_pfn(*(pte_t *)p) << PAGE_SHIFT,
pa = pte_pfn(pte) << PAGE_SHIFT;
pa = pte_pfn(pte) << PAGE_SHIFT;
pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
pa = pte_pfn(__pte(old_pte)) << HW_PAGE_SHIFT;
pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
if (!pfn_valid(pte_pfn(pte)))
pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
mas7_3 = (u64)pte_pfn(pte) << PAGE_SHIFT;
unsigned long pfn = pte_pfn(pte);
#define pte_page(x) pfn_to_page(pte_pfn(x))
unsigned long pfn = pte_pfn(pte);
new_pte = pfn_pte(pte_pfn(ptep_get(ptep)), prot);
} else if (ALIGN_DOWN(PFN_PHYS(pte_pfn(ptep_get(ptep))), page_size) == hpa) {
u64 pa = PFN_PHYS(pte_pfn(__pte(val)));
#define pte_page(x) pfn_to_page(pte_pfn(x))
return uv_destroy_folio(pfn_folio(pte_pfn(pte)));
return uv_convert_from_secure_folio(pfn_folio(pte_pfn(pte)));
if (is_zero_pfn(pte_pfn(*pte))) {
vmem_free_pages((unsigned long)pfn_to_virt(pte_pfn(*pte)), get_order(PAGE_SIZE), altmap);
#define pte_page(x) pfn_to_page(pte_pfn(x))
return pfn_to_kaddr(pte_pfn(entry));
unsigned long pfn = pte_pfn(pte);
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
#define pte_page(x) pfn_to_page(pte_pfn(x))
return pte_pfn(pte);
pfn = pte_pfn(pte);
pfn = pte_pfn(pte);
return pte_pfn(pte);
unsigned long this_pfn = pte_pfn(pte); \
pa = (pte_pfn(pte) << PAGE_SHIFT);
return pfn_valid(pte_pfn(*pte));
unsigned long pfn = pte_pfn(pte);
unsigned long paddr, pfn = pte_pfn(orig);
#define pte_page(x) pfn_to_page(pte_pfn(x))
pa = (phys_addr_t)pte_pfn(*pte) << PAGE_SHIFT;
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
if (pte_pfn(oldpte) != pte_pfn(newpte))
return pte_pfn(pfn_pte(pfn, __pgprot(0))) == pfn;
pfn = pte_pfn(*kpte);
pfn = pte_pfn(first);
if (pte_pfn(entry) != pte_pfn(first) + i)
unsigned long pfn = pte_pfn(old_pte);
phys_addr = (phys_addr_t)pte_pfn(*pte) << PAGE_SHIFT;
paddr = PFN_PHYS(pte_pfn(*pte)) | (hva & ~page_level_mask(level));
pfn = pte_pfn(*ptep);
pfn = pte_pfn(*ptep);
pa = pte_pfn(pte_tbl[i]) << PAGE_SHIFT;
pa = pte_pfn(pte) << PAGE_SHIFT;
if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity)))
p2m_pfn = pte_pfn(READ_ONCE(*ptep));
if (pte_pfn(*ptep) == p2m_pfn) {
if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_missing)))
if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity)))
#define pte_page(x) pfn_to_page(pte_pfn(x))
unsigned long pfn = pte_pfn(*ptep);
*paddr = pte_pfn(pte) << PAGE_SHIFT;
if (pte_pfn(pte) == pfn)
frame = pte_pfn(pte);
frame = pte_pfn(pte) +
if (is_zero_pfn(pte_pfn(pte)))
if (is_zero_pfn(pte_pfn(pte)))
#define is_ksm_zero_pte(pte) (is_zero_pfn(pte_pfn(pte)) && pte_dirty(pte))
flush_cache_page(vma, vaddr, pte_pfn(fw->pte));
pfn = pte_pfn(pteval);
struct folio *folio = pfn_folio(pte_pfn(entry));
folio = pfn_folio(pte_pfn(entry));
args->pte_pfn = ULONG_MAX;
if (args->pte_pfn != ULONG_MAX) {
__free_page(pfn_to_page(args->pte_pfn));
args->pte_pfn = ULONG_MAX;
args->pte_pfn = ULONG_MAX;
args->pte_pfn = args->pud_pfn;
args->pte_pfn = args->pmd_pfn;
args->pte_pfn = page_to_pfn(page);
page = (args->pte_pfn != ULONG_MAX) ? pfn_to_page(args->pte_pfn) : NULL;
pte = pfn_pte(args->pte_pfn, args->page_prot);
pte = pfn_pte(args->pte_pfn, args->page_prot);
pte = pfn_pte(args->pte_pfn, args->page_prot);
pte_t pte = pfn_pte(args->pte_pfn, args->page_prot);
page = (args->pte_pfn != ULONG_MAX) ? pfn_to_page(args->pte_pfn) : NULL;
unsigned long pte_pfn;
page = (args->pte_pfn != ULONG_MAX) ? pfn_to_page(args->pte_pfn) : NULL;
args->pte_pfn = ULONG_MAX;
if ((gup_flags & FOLL_DUMP) || !is_zero_pfn(pte_pfn(entry)))
VM_WARN_ON_ONCE(!pfn_valid(pte_pfn(pte)));
if (is_zero_pfn(pte_pfn(pte))) {
idx = arch_kmap_local_map_idx(i, pte_pfn(pteval));
idx = arch_kmap_local_map_idx(i, pte_pfn(pteval));
idx = arch_kmap_local_map_idx(i, pte_pfn(pteval));
!is_zero_pfn(pte_pfn(pte))) {
new_pfn_flags = pte_pfn(pte) | cpu_flags;
pfn = pte_pfn(entry) + ((start & ~hmask) >> PAGE_SHIFT);
VM_WARN_ON_FOLIO(page_folio(pfn_to_page(pte_pfn(pte))) != folio, folio);
folio_pfn(folio) + folio_nr_pages(folio) - pte_pfn(pte));
__free_page(pfn_to_page(pte_pfn(pte)));
return pte_present(pte) && is_zero_pfn(pte_pfn(pte));
pfn = pte_pfn(pteval);
flush_cache_page(vma, addr, pte_pfn(ptep_get(ptep)));
pfn = pte_pfn(pte);
if (pte_pfn(pteval) != page_to_pfn(page)) {
WARN_ON_ONCE(!is_zero_pfn(pte_pfn(pteval)));
if (pte_pfn(entry) != pfn) {
WARN_ON_ONCE(!is_zero_pfn(pte_pfn(entry)));
VM_WARN_ON(is_zero_pfn(pte_pfn(vmf->orig_pte)));
flush_cache_page(vma, vmf->address, pte_pfn(vmf->orig_pte));
pfn_is_zero = is_zero_pfn(pte_pfn(vmf->orig_pte));
flush_cache_page(vma, vmf->address, pte_pfn(vmf->orig_pte));
int nr = pte_pfn(fault_pte) - folio_pfn(folio);
if (pfn_folio(pte_pfn(ptent)) != folio)
pte_pfn(pte), PAGE_MASK, pte_write(pte),
return __vm_normal_page(vma, addr, pte_pfn(pte), pte_special(pte),
folio = pfn_folio(pte_pfn(ptep));
unsigned long pfn = pte_pfn(orig_pte);
flush_cache_page(vma, addr, pte_pfn(orig_pte));
pfn = pte_pfn(pte);
flush_cache_page(vma, addr, pte_pfn(pte));
return pfn_modify_allowed(pte_pfn(ptep_get(pte)),
return pfn_modify_allowed(pte_pfn(ptep_get(pte)),
VM_WARN_ON_ONCE(is_zero_pfn(pte_pfn(pte)) && pte_dirty(pte));
page_table_check_clear(pte_pfn(pte), PAGE_SIZE >> PAGE_SHIFT);
page_table_check_set(pte_pfn(pte), nr, pte_write(pte));
pfn = pte_pfn(ptent);
is_zero_pfn(pte_pfn(pte))) {
page = pfn_to_page(pte_pfn(pte));
flush_cache_page(vma, address, pte_pfn(entry));
pfn = pte_pfn(pteval);
pfn = pte_pfn(pteval);
unsigned long pfn = pte_pfn(ptep_get(pte));
pte_pfn(ptep_get(pte)),
pte_pfn(ptep_get(pte)),
if (!pte_present(orig_src_pte) || is_zero_pfn(pte_pfn(orig_src_pte)))
if (is_zero_pfn(pte_pfn(orig_src_pte))) {
unsigned long pfn = pte_pfn(pte);