pfn_valid
#define virt_addr_valid(kaddr) pfn_valid((__pa(kaddr) >> PAGE_SHIFT))
#define pfn_valid pfn_valid
#define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr))
extern int pfn_valid(unsigned long pfn);
EXPORT_SYMBOL(pfn_valid);
&& pfn_valid(virt_to_pfn(kaddr)))
extern int pfn_valid(unsigned long);
#define pfn_valid pfn_valid
if (!pfn_valid(pfn))
if (!pfn_valid(pfn))
EXPORT_SYMBOL(pfn_valid);
if (!pfn_valid(pfn))
BUG_ON(!pfn_valid(pte_pfn(pte)));
if (!pfn_valid(pfn))
if (!pfn_valid(pfn))
if (!pfn_valid(pfn)) {
if (pfn_valid(PFN_DOWN(CONFIG_ITCM_RAM_BASE)))
if (pfn_valid(PFN_DOWN(CONFIG_DTCM_RAM_BASE)))
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
if (pfn_valid(pfn))
return pfn_valid(PFN_DOWN(PHYSADDR(kaddr)));
#define virt_addr_valid(vaddr) (pfn_valid(virt_to_pfn(vaddr)))
if (unlikely(!pfn_valid(pfn)))
if (pfn_valid(start_pfn + i) &&
return pfn_valid(PFN_DOWN(virt_to_phys(kaddr)));
# define virt_addr_valid(vaddr) pfn_valid(PFN_DOWN(virt_to_phys(vaddr)))
if (!pfn_valid(pfn))
#define virt_addr_valid(kaddr) (pfn_valid(virt_to_pfn(kaddr)))
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
if (!pfn_valid(pfn))
if (pfn_valid(pfn))
if (!pfn_valid(pfn))
pfn_valid(virt_to_pfn((void *)_addr)); \
if (!pfn_valid(pte_pfn(pte)))
if (pfn_valid(page_to_pfn(start)))
if (unlikely(!pfn_valid(pfn)))
if (!pfn_valid(start_pfn))
(unsigned long)(_addr) >= PAGE_OFFSET && pfn_valid(virt_to_pfn(_addr)); \
#define virt_addr_valid(kaddr) pfn_valid(phys_to_pfn(__pa_nodebug((unsigned long)(kaddr))))
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
if (pfn_valid(pfn)) {
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
if (pfn_valid(this_pfn) && \
if (pfn_valid(afar >> PAGE_SHIFT))
if (pfn_valid(addr >> PAGE_SHIFT))
return pfn_valid(pa >> PAGE_SHIFT);
return pfn_valid(pud_pfn(*pud));
return pfn_valid(pmd_pfn(*pmd));
return pfn_valid(pte_pfn(*pte));
if (pfn_valid(pfn))
if (!pfn_valid(pfn))
#define virt_addr_valid(v) pfn_valid(phys_to_pfn(__pa(v)))
ret = pfn_valid(x >> PAGE_SHIFT);
if (pfn_valid(pfn))
if (pfn_valid(pfn))
return pfn_valid(x >> PAGE_SHIFT);
return pfn_valid((x - PAGE_OFFSET) >> PAGE_SHIFT);
if (!pfn_valid(pfn))
if (!pfn_valid(pfn))
(!IS_ALIGNED(pfn, PTRS_PER_PMD) || !pfn_valid(pfn + PTRS_PER_PMD - 1)))
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
if (!pfn_valid(pfn))
WARN_ON(pfn_valid(pfn));
if (!pfn_valid(pfn))
return pfn_valid(__phys_to_pfn(addr));
if (!pfn_valid(page))
if (!pfn_valid(pfn))
if (!pfn_valid(pfn))
if (!pfn_valid(pfn)) {
if (!pfn_valid(pfn)) {
if (!pfn_valid(pfn)) {
if (!pfn_valid(pfn)) {
if (!pfn_valid(nums[i]))
if (!pfn_valid(pfn)) {
if (!pfn_valid(PHYS_PFN(paddr)))
if (pfn_valid(pfn))
if (!pfn_valid(phys_pfn)) {
if (pfn_valid(cur_pfn) &&
if (pfn_valid(PFN_DOWN(paddr)))
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dev_addr))))
if (pfn_valid(PFN_DOWN(dma_to_phys(hwdev, dev_addr))))
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr))))
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr))))
if (!pfn_valid(pfn))
if (pfn_valid(start >> PAGE_SHIFT))
if (pfn_valid(prz->paddr >> PAGE_SHIFT)) {
#ifndef pfn_valid
#define pfn_valid pfn_valid
WARN_ON_ONCE(!pfn_valid(__pfn)); \
if (pfn_valid(pfn)) \
#define pfn_in_present_section pfn_valid
if (pfn_valid(_pfn))
if (WARN_ON_ONCE(!pfn_valid(page_to_pfn(page))))
if (!pfn_valid(pfn))
WARN_ON_ONCE(pfn_valid(PHYS_PFN(phys_addr))))
if (pfn_valid(pfn) && !PageHighMem(pfn_to_page(pfn)) &&
if (pfn_valid(pfn))
if (!pfn_valid(pfn))
if (!pfn_valid(pfn))
if (fr_pfn != BM_END_OF_MAP && pfn_valid(fr_pfn)) {
if (pfn_valid(decoded_pfn) && memory_bm_pfn_present(bm, decoded_pfn)) {
if (!pfn_valid(decoded_pfn))
if (pfn_valid(pfn) && (early_pfn_to_nid(pfn) == node))
WARN_ON_ONCE(!pfn_valid(base_pfn));
if (cc->direct_compaction && pfn_valid(min_pfn)) {
WARN_ON(!pfn_valid(args->fixed_pte_pfn));
VM_WARN_ON_ONCE(!pfn_valid(pte_pfn(pte)));
if (!pfn_valid(pfn))
if (pfn_valid(page_to_pfn(page)))
if (!pfn_valid(pfn) && !arch_is_platform_page(PFN_PHYS(pfn))) {
if (pfn_valid(pfn)) {
if (!pfn_valid(pfn)) {
BUG_ON((vma->vm_flags & VM_MIXEDMAP) && pfn_valid(pfn));
if (!IS_ENABLED(CONFIG_ARCH_HAS_PTE_SPECIAL) && pfn_valid(pfn)) {
if (!pfn_valid(pfn))
if (IS_ENABLED(CONFIG_HAVE_ARCH_PFN_VALID) && !pfn_valid(pfn))
if (!pfn_valid(end_pfn))
if (!pfn_valid(pfn))
while (!pfn_valid(pfn) && (pfn & (MAX_ORDER_NR_PAGES - 1)) != 0)
if ((pfn & (MAX_ORDER_NR_PAGES - 1)) == 0 && !pfn_valid(pfn)) {
if (!pfn_valid(pfn)) {
if (!pfn_valid(pfn))
if (!pfn_valid(pfn))
if (pfn_valid(pfn)) {
if (WARN_ON_ONCE(pfn_valid(pfn)))
if (WARN_ON(!pfn_valid(page_to_pfn(page)))) {
if (WARN_ON_ONCE(!pfn_valid(pfn)))
if (WARN_ON_ONCE(!pfn_valid(pfn)))
if (pfn_valid(map->pfn)) {
if (pfn_valid(pfn))
if (pfn_valid(pfn)) {