Symbol: PFN_DOWN
arch/alpha/kernel/setup.c
239
#define PFN_MAX PFN_DOWN(0x80000000)
arch/arc/mm/init.c
111
max_low_pfn = max_pfn = PFN_DOWN(low_mem_start + low_mem_sz);
arch/arc/mm/init.c
153
min_high_pfn = PFN_DOWN(high_mem_start);
arch/arc/mm/init.c
154
max_high_pfn = PFN_DOWN(high_mem_start + high_mem_sz);
arch/arm/include/asm/kfence.h
14
unsigned long pfn = PFN_DOWN(__pa(addr));
arch/arm/mm/init.c
78
*max_low = PFN_DOWN(memblock_get_current_limit());
arch/arm/mm/init.c
80
*max_high = PFN_DOWN(memblock_end_of_DRAM());
arch/arm/mm/mmu.c
1665
__pv_phys_pfn_offset += PFN_DOWN(offset);
arch/arm64/kernel/setup.c
260
start = max(PFN_PHYS(PFN_DOWN(r_start)), mem->start);
arch/arm64/mm/init.c
127
max_zone_pfns[ZONE_DMA] = PFN_DOWN(max_zone_phys(zone_dma_limit));
arch/arm64/mm/init.c
130
max_zone_pfns[ZONE_DMA32] = PFN_DOWN(dma32_phys_limit);
arch/arm64/mm/init.c
305
max = PFN_DOWN(memblock_end_of_DRAM());
arch/arm64/mm/init.c
334
bool swiotlb = max_pfn > PFN_DOWN(arm64_dma_phys_limit);
arch/csky/include/asm/page.h
66
#define ARCH_PFN_OFFSET PFN_DOWN(va_pa_offset + PHYS_OFFSET_OFFSET)
arch/csky/include/asm/page.h
78
#define MAP_NR(x) PFN_DOWN((unsigned long)(x) - PAGE_OFFSET - \
arch/csky/kernel/setup.c
64
unsigned long lowmem_size = PFN_DOWN(LOWMEM_LIMIT - PHYS_OFFSET_OFFSET);
arch/csky/kernel/setup.c
65
unsigned long sseg_size = PFN_DOWN(SSEG_SIZE - PHYS_OFFSET_OFFSET);
arch/csky/kernel/setup.c
76
max_low_pfn = max_pfn = PFN_DOWN(memblock_end_of_DRAM());
arch/csky/mm/init.c
76
set_pte(&kernel_pte_tables[i - PFN_DOWN(va_pa_offset)], pfn_pte(i, PAGE_KERNEL));
arch/csky/mm/tcm.c
34
if (pfn_valid(PFN_DOWN(CONFIG_ITCM_RAM_BASE)))
arch/csky/mm/tcm.c
55
if (pfn_valid(PFN_DOWN(CONFIG_DTCM_RAM_BASE)))
arch/hexagon/include/asm/page.h
90
#define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr)))
arch/hexagon/mm/init.c
121
bootmem_lastpg = PFN_DOWN((bootmem_lastpg << PAGE_SHIFT) &
arch/hexagon/mm/init.c
134
max_low_pfn = bootmem_lastpg - PFN_DOWN(DMA_RESERVED_BYTES);
arch/hexagon/mm/init.c
96
bootmem_lastpg = PFN_DOWN(size);
arch/loongarch/kernel/mem.c
49
max_pfn = PFN_DOWN(memblock_end_of_DRAM());
arch/loongarch/kernel/mem.c
50
max_low_pfn = min(PFN_DOWN(HIGHMEM_START), max_pfn);
arch/loongarch/kernel/numa.c
251
max_pfn = PFN_DOWN(memblock_end_of_DRAM());
arch/loongarch/kernel/numa.c
252
max_low_pfn = min(PFN_DOWN(HIGHMEM_START), max_pfn);
arch/loongarch/kernel/paravirt.c
224
if (PFN_DOWN(addr) != PFN_DOWN(addr + sizeof(*st))) {
arch/loongarch/kernel/setup.c
397
max_pfn = PFN_DOWN(memblock_end_of_DRAM());
arch/loongarch/kernel/setup.c
398
max_low_pfn = min(PFN_DOWN(HIGHMEM_START), max_pfn);
arch/loongarch/kernel/setup.c
412
register_nosave_region(PFN_DOWN(__pa_symbol(&__nosave_begin)),
arch/loongarch/kernel/setup.c
555
start = max(PFN_PHYS(PFN_DOWN(r_start)), mem->start);
arch/loongarch/mm/mmap.c
125
return pfn_valid(PFN_DOWN(PHYSADDR(kaddr)));
arch/loongarch/power/hibernate.c
44
unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin));
arch/m68k/kernel/setup_no.c
154
min_low_pfn = PFN_DOWN(memory_start);
arch/m68k/kernel/setup_no.c
155
max_pfn = max_low_pfn = PFN_DOWN(memory_end);
arch/m68k/mm/init.c
45
max_zone_pfns[ZONE_DMA] = PFN_DOWN(memblock_end_of_DRAM());
arch/m68k/mm/mcfmmu.c
162
num_pages = PFN_DOWN(_ramend - _rambase);
arch/m68k/mm/mcfmmu.c
166
min_low_pfn = PFN_DOWN(_rambase);
arch/m68k/mm/mcfmmu.c
167
max_pfn = max_low_pfn = PFN_DOWN(_ramend);
arch/microblaze/include/asm/page.h
95
# define phys_to_pfn(phys) (PFN_DOWN(phys))
arch/mips/include/asm/dma.h
93
#define MAX_DMA_PFN PFN_DOWN(virt_to_phys((void *)MAX_DMA_ADDRESS))
arch/mips/include/asm/page.h
209
#define virt_to_pfn(kaddr) PFN_DOWN(virt_to_phys((void *)(kaddr)))
arch/mips/kernel/setup.c
306
max_pfn = PFN_DOWN(ramend);
arch/mips/kernel/setup.c
314
if (start >= PFN_DOWN(HIGHMEM_START))
arch/mips/kernel/setup.c
316
if (end > PFN_DOWN(HIGHMEM_START))
arch/mips/kernel/setup.c
317
end = PFN_DOWN(HIGHMEM_START);
arch/mips/kernel/setup.c
325
if (max_pfn > PFN_DOWN(HIGHMEM_START)) {
arch/mips/kernel/setup.c
326
max_low_pfn = PFN_DOWN(HIGHMEM_START);
arch/mips/kernel/vdso.c
148
gic_pfn = PFN_DOWN(__pa(gic_base));
arch/mips/loongson64/numa.c
102
unsigned long kernel_start_pfn = PFN_DOWN(__pa_symbol(&_text));
arch/mips/mm/init.c
475
for (pfn = PFN_UP(begin); pfn < PFN_DOWN(end); pfn++) {
arch/mips/mm/ioremap.c
83
pfn = PFN_DOWN(phys_addr);
arch/mips/mm/ioremap.c
84
last_pfn = PFN_DOWN(last_addr);
arch/mips/mm/mmap.c
127
return pfn_valid(PFN_DOWN(virt_to_phys(kaddr)));
arch/mips/power/cpu.c
39
unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin));
arch/mips/sgi-ip22/ip28-berr.c
318
if (page_is_ram(PFN_DOWN(pte))) {
arch/nios2/include/asm/page.h
85
# define virt_to_page(vaddr) pfn_to_page(PFN_DOWN(virt_to_phys(vaddr)))
arch/nios2/include/asm/page.h
86
# define virt_addr_valid(vaddr) pfn_valid(PFN_DOWN(virt_to_phys(vaddr)))
arch/nios2/kernel/setup.c
140
*max_low = PFN_DOWN(memblock_get_current_limit());
arch/nios2/kernel/setup.c
142
*max_high = PFN_DOWN(memblock_end_of_DRAM());
arch/nios2/mm/ioremap.c
37
pfn = PFN_DOWN(phys_addr);
arch/openrisc/kernel/setup.c
63
ram_end_pfn = PFN_DOWN(memblock_end_of_DRAM());
arch/parisc/mm/init.c
698
max_zone_pfns[ZONE_NORMAL] = PFN_DOWN(memblock_end_of_DRAM());
arch/powerpc/mm/pgtable_32.c
131
PFN_DOWN((unsigned long)_sinittext);
arch/powerpc/mm/pgtable_32.c
170
PFN_DOWN((unsigned long)_stext);
arch/powerpc/platforms/512x/mpc512x_shared.c
185
end = PFN_DOWN(addr + size);
arch/powerpc/platforms/pseries/hotplug-memory.c
170
section_nr = pfn_to_section_nr(PFN_DOWN(lmb->base_addr));
arch/riscv/include/asm/page.h
182
#define phys_to_pfn(phys) (PFN_DOWN(phys))
arch/riscv/include/asm/page.h
97
#define ARCH_PFN_OFFSET (PFN_DOWN((unsigned long)phys_ram_base))
arch/riscv/kernel/hibernate.c
398
hibernate_restore_image(resume_hdr.saved_satp, (PFN_DOWN(__pa(resume_pg_dir)) | satp_mode),
arch/riscv/kernel/setup.c
248
start = max(PFN_PHYS(PFN_DOWN(r_start)), mem->start);
arch/riscv/kvm/gstage.c
145
set_pte(ptep, pfn_pte(PFN_DOWN(__pa(next_ptep)),
arch/riscv/kvm/gstage.c
206
out_map->pte = pfn_pte(PFN_DOWN(hpa), prot);
arch/riscv/mm/init.c
1369
csr_write(CSR_SATP, PFN_DOWN(__pa_symbol(swapper_pg_dir)) | satp_mode);
arch/riscv/mm/init.c
171
bool swiotlb = max_pfn > PFN_DOWN(dma32_phys_limit);
arch/riscv/mm/init.c
291
max_low_pfn = max_pfn = PFN_DOWN(phys_ram_end);
arch/riscv/mm/init.c
457
ptep[pte_idx] = pfn_pte(PFN_DOWN(pa), prot);
arch/riscv/mm/init.c
540
pmdp[pmd_idx] = pfn_pmd(PFN_DOWN(pa), prot);
arch/riscv/mm/init.c
546
pmdp[pmd_idx] = pfn_pmd(PFN_DOWN(pte_phys), PAGE_TABLE);
arch/riscv/mm/init.c
642
pudp[pud_index] = pfn_pud(PFN_DOWN(pa), prot);
arch/riscv/mm/init.c
648
pudp[pud_index] = pfn_pud(PFN_DOWN(next_phys), PAGE_TABLE);
arch/riscv/mm/init.c
668
p4dp[p4d_index] = pfn_p4d(PFN_DOWN(pa), prot);
arch/riscv/mm/init.c
674
p4dp[p4d_index] = pfn_p4d(PFN_DOWN(next_phys), PAGE_TABLE);
arch/riscv/mm/init.c
725
pgdp[pgd_idx] = pfn_pgd(PFN_DOWN(pa), prot);
arch/riscv/mm/init.c
731
pgdp[pgd_idx] = pfn_pgd(PFN_DOWN(next_phys), PAGE_TABLE);
arch/riscv/mm/init.c
85
max_zone_pfns[ZONE_DMA32] = PFN_DOWN(dma32_phys_limit);
arch/riscv/mm/init.c
896
identity_satp = PFN_DOWN((uintptr_t)&early_pg_dir) | satp_mode;
arch/riscv/mm/kasan_init.c
101
set_pud(pudp, pfn_pud(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/kasan_init.c
120
set_pgd(pgd, pfn_pgd(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
132
set_p4d(p4dp, pfn_p4d(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/kasan_init.c
155
set_pgd(pgdp, pfn_pgd(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/kasan_init.c
255
set_pud(pudp, pfn_pud(PFN_DOWN(phys_addr), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
293
set_p4d(p4dp, pfn_p4d(PFN_DOWN(phys_addr), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
314
set_pgd(pgdp, pfn_pgd(PFN_DOWN(phys_addr), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
335
pfn_pmd(PFN_DOWN
arch/riscv/mm/kasan_init.c
342
pfn_pud(PFN_DOWN
arch/riscv/mm/kasan_init.c
350
pfn_p4d(PFN_DOWN
arch/riscv/mm/kasan_init.c
36
set_pmd(pmd, pfn_pmd(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
389
set_pud(pud_k, pfn_pud(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
409
set_p4d(p4d_k, pfn_p4d(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
428
set_pgd(pgd_k, pfn_pgd(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
44
set_pte(ptep, pfn_pte(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/kasan_init.c
468
pfn_pgd(PFN_DOWN(__pa(tmp_p4d)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
479
pfn_p4d(PFN_DOWN(__pa(tmp_pud)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
489
csr_write(CSR_SATP, PFN_DOWN(__pa(tmp_pg_dir)) | satp_mode);
arch/riscv/mm/kasan_init.c
534
csr_write(CSR_SATP, PFN_DOWN(__pa(swapper_pg_dir)) | satp_mode);
arch/riscv/mm/kasan_init.c
58
set_pud(pud, pfn_pud(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
70
set_pmd(pmdp, pfn_pmd(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/kasan_init.c
89
set_p4d(p4d, pfn_p4d(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/s390/kernel/machine_kexec.c
184
diag10_range(PFN_DOWN(crashk_res.start),
arch/s390/kernel/machine_kexec.c
185
PFN_DOWN(crashk_res.end - crashk_res.start + 1));
arch/s390/kernel/setup.c
553
max_pfn = max_low_pfn = PFN_DOWN(ident_map_size);
arch/s390/kernel/setup.c
572
if (arg->start_pfn < PFN_DOWN(resource_size(&crashk_res)))
arch/s390/kernel/setup.c
652
diag10_range(PFN_DOWN(crash_base), PFN_DOWN(crash_size));
arch/s390/mm/init.c
263
unsigned long start_pfn = PFN_DOWN(start);
arch/s390/mm/init.c
264
unsigned long size_pages = PFN_DOWN(size);
arch/sh/mm/numa.c
30
start_pfn = PFN_DOWN(start);
arch/sh/mm/numa.c
31
end_pfn = PFN_DOWN(end);
arch/sparc/power/hibernate.c
25
unsigned long nosave_begin_pfn = PFN_DOWN((unsigned long)&__nosave_begin);
arch/sparc/power/hibernate.c
26
unsigned long nosave_end_pfn = PFN_DOWN((unsigned long)&__nosave_end);
arch/x86/events/intel/pt.c
1041
cur_pg = PFN_DOWN(topa->offset);
arch/x86/events/intel/pt.c
635
return PFN_DOWN(virt_to_phys(topa_to_page(topa)));
arch/x86/include/asm/setup.h
19
#define MAXMEM_PFN PFN_DOWN(MAXMEM)
arch/x86/include/asm/xen/page.h
233
return XMADDR(PFN_PHYS(pfn_to_mfn(PFN_DOWN(phys.paddr))) | offset);
arch/x86/include/asm/xen/page.h
239
return XPADDR(PFN_PHYS(mfn_to_pfn(PFN_DOWN(machine.maddr))) | offset);
arch/x86/include/asm/xen/page.h
300
return PFN_DOWN(__pa(v));
arch/x86/kernel/amd_gart_64.c
749
start_pfn = PFN_DOWN(aper_base);
arch/x86/kernel/cpu/sgx/encl.c
147
page_index = PFN_DOWN(encl_page->desc - encl_page->encl->base);
arch/x86/kernel/cpu/sgx/encl.c
149
page_index = PFN_DOWN(encl->size);
arch/x86/kernel/cpu/sgx/encl.c
199
sgx_encl_truncate_backing_page(encl, PFN_DOWN(page_pcmd_off));
arch/x86/kernel/cpu/sgx/encl.c
287
entry = xa_load(&encl->page_array, PFN_DOWN(addr));
arch/x86/kernel/cpu/sgx/encl.c
307
entry = xa_load(&encl->page_array, PFN_DOWN(addr));
arch/x86/kernel/cpu/sgx/encl.c
379
ret = xa_insert(&encl->page_array, PFN_DOWN(encl_page->desc),
arch/x86/kernel/cpu/sgx/encl.c
408
vmret = vmf_insert_pfn(vma, addr, PFN_DOWN(phys_addr));
arch/x86/kernel/cpu/sgx/encl.c
417
xa_erase(&encl->page_array, PFN_DOWN(encl_page->desc));
arch/x86/kernel/cpu/sgx/encl.c
457
(!xa_load(&encl->page_array, PFN_DOWN(addr))))
arch/x86/kernel/cpu/sgx/encl.c
474
ret = vmf_insert_pfn(vma, addr, PFN_DOWN(phys_addr));
arch/x86/kernel/cpu/sgx/encl.c
530
XA_STATE(xas, &encl->page_array, PFN_DOWN(start));
arch/x86/kernel/cpu/sgx/encl.c
546
xas_for_each(&xas, page, PFN_DOWN(end - 1)) {
arch/x86/kernel/cpu/sgx/encl.c
703
unsigned long max_page_index = PFN_DOWN(encl->base + encl->size - 1);
arch/x86/kernel/cpu/sgx/encl.c
708
XA_STATE(xas, &encl->page_array, PFN_DOWN(encl->base));
arch/x86/kernel/cpu/sgx/encl.c
86
entry = xa_load(&encl->page_array, PFN_DOWN(addr));
arch/x86/kernel/cpu/sgx/encl.c
983
pcmd = sgx_encl_get_backing_page(encl, PFN_DOWN(page_pcmd_off));
arch/x86/kernel/cpu/sgx/ioctl.c
1135
xa_erase(&encl->page_array, PFN_DOWN(entry->desc));
arch/x86/kernel/cpu/sgx/ioctl.c
307
ret = xa_insert(&encl->page_array, PFN_DOWN(encl_page->desc),
arch/x86/kernel/cpu/sgx/ioctl.c
339
xa_erase(&encl->page_array, PFN_DOWN(encl_page->desc));
arch/x86/kernel/cpu/sgx/main.c
270
ret = sgx_encl_alloc_backing(encl, PFN_DOWN(encl->size),
arch/x86/kernel/cpu/sgx/main.c
338
page_index = PFN_DOWN(encl_page->desc - encl_page->encl->base);
arch/x86/kernel/cpu/sgx/main.c
671
return &section->pages[PFN_DOWN(paddr - section->phys_addr)];
arch/x86/kernel/cpu/sgx/virt.c
44
index = vma->vm_pgoff + PFN_DOWN(addr - vma->vm_start);
arch/x86/kernel/cpu/sgx/virt.c
58
pfn = PFN_DOWN(sgx_get_epc_phys_addr(epc_page));
arch/x86/kernel/dumpstack_64.c
72
[PFN_DOWN(CEA_ESTACK_OFFS(st)) ... \
arch/x86/kernel/dumpstack_64.c
73
PFN_DOWN(CEA_ESTACK_OFFS(st) + CEA_ESTACK_SIZE(st) - 1)] = { \
arch/x86/kernel/e820.c
792
register_nosave_region(PFN_DOWN(last_addr), PFN_UP(entry->addr));
arch/x86/kernel/e820.c
797
register_nosave_region(PFN_DOWN(last_addr), limit_pfn);
arch/x86/kernel/setup.c
378
if (pfn_range_is_mapped(PFN_DOWN(ramdisk_image),
arch/x86/kernel/setup.c
379
PFN_DOWN(ramdisk_end))) {
arch/x86/kernel/tboot.c
172
map_base = PFN_DOWN(tboot->tboot_base);
arch/x86/kvm/vmx/sgx.c
94
*hva = kvm_vcpu_gfn_to_hva(vcpu, PFN_DOWN(gpa));
arch/x86/mm/init.c
409
limit_pfn = PFN_DOWN(end);
arch/x86/mm/init.c
412
pfn = start_pfn = PFN_DOWN(start);
arch/x86/mm/init.c
421
end_pfn = PFN_DOWN(PMD_SIZE);
arch/x86/mm/init.c
423
end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
arch/x86/mm/init.c
425
end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
arch/x86/mm/init.c
435
start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
arch/x86/mm/init.c
437
end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE));
arch/x86/mm/init.c
439
end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE));
arch/x86/mm/init.c
440
if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE)))
arch/x86/mm/init.c
441
end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE));
arch/x86/mm/init.c
452
start_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE));
arch/x86/mm/init.c
453
end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE));
arch/x86/mm/init.c
462
start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
arch/x86/mm/init.c
463
end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE));
arch/x86/mm/kasan_init_64.c
67
entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL);
arch/x86/mm/pat/set_memory.c
1088
unsigned int npg = PFN_DOWN(size);
arch/x86/mm/pat/set_memory.c
1187
unsigned long pfn = PFN_DOWN(__pa(address));
arch/x86/mm/pat/set_memory.c
490
#define BIOS_PFN PFN_DOWN(BIOS_BEGIN)
arch/x86/mm/pat/set_memory.c
491
#define BIOS_PFN_END PFN_DOWN(BIOS_END - 1)
arch/x86/mm/pat/set_memory.c
513
unsigned long epfn_ro, spfn_ro = PFN_DOWN(__pa_symbol(__start_rodata));
arch/x86/mm/pat/set_memory.c
519
epfn_ro = PFN_DOWN(__pa_symbol(__end_rodata)) - 1;
arch/x86/platform/efi/efi_32.c
42
start_pfn = PFN_DOWN(md->phys_addr);
arch/x86/xen/enlighten_pvh.c
100
if (pages != (PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr))) {
arch/x86/xen/enlighten_pvh.c
80
ram_pages += PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr);
arch/x86/xen/enlighten_pvh.c
98
PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr));
arch/x86/xen/mmu.c
14
return PFN_DOWN(maddr.maddr);
arch/x86/xen/mmu_pv.c
1106
pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(pa));
arch/x86/xen/mmu_pv.c
1382
mfn = pfn_to_mfn(PFN_DOWN(cr3));
arch/x86/xen/mmu_pv.c
1739
if (*pt_base == PFN_DOWN(__pa(addr))) {
arch/x86/xen/mmu_pv.c
1744
if (*pt_end == PFN_DOWN(__pa(addr))) {
arch/x86/xen/mmu_pv.c
1775
max_pfn_mapped = PFN_DOWN(__pa(xen_start_info->mfn_list));
arch/x86/xen/mmu_pv.c
1777
pt_base = PFN_DOWN(__pa(xen_start_info->pt_base));
arch/x86/xen/mmu_pv.c
1852
PFN_DOWN(__pa_symbol(init_top_pgt)));
arch/x86/xen/mmu_pv.c
1855
pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
arch/x86/xen/mmu_pv.c
1861
PFN_DOWN(__pa_symbol(level3_user_vsyscall)));
arch/x86/xen/mmu_pv.c
1988
p2m_pfn = PFN_DOWN(pt_phys) + n_pt;
arch/x86/xen/mmu_pv.c
2014
PFN_DOWN(pt_phys));
arch/x86/xen/mmu_pv.c
2022
PFN_DOWN(pmd_phys));
arch/x86/xen/mmu_pv.c
2029
pin_pagetable_pfn(MMUEXT_PIN_L3_TABLE, PFN_DOWN(pud_phys));
arch/x86/xen/mmu_pv.c
2039
p2m_pfn = PFN_DOWN(xen_early_virt_to_phys(xen_start_info->mfn_list));
arch/x86/xen/mmu_pv.c
2041
p2m_pfn_end = p2m_pfn + PFN_DOWN(size);
arch/x86/xen/mmu_pv.c
2064
xen_start_info->first_p2m_pfn = PFN_DOWN(new_area);
arch/x86/xen/mmu_pv.c
2108
pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
arch/x86/xen/mmu_pv.c
2118
pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
arch/x86/xen/mmu_pv.c
803
xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(pgd)));
arch/x86/xen/mmu_pv.c
808
PFN_DOWN(__pa(user_pgd)));
arch/x86/xen/mmu_pv.c
918
xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
arch/x86/xen/mmu_pv.c
922
PFN_DOWN(__pa(user_pgd)));
arch/x86/xen/p2m.c
341
pfn_pte(PFN_DOWN(__pa(p2m_missing)), PAGE_KERNEL_RO));
arch/x86/xen/p2m.c
343
pfn_pte(PFN_DOWN(__pa(p2m_identity)), PAGE_KERNEL_RO));
arch/x86/xen/p2m.c
376
pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL));
arch/x86/xen/p2m.c
386
pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO));
arch/x86/xen/p2m.c
445
if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity)))
arch/x86/xen/p2m.c
575
if (p2m_pfn == PFN_DOWN(__pa(p2m_identity)) ||
arch/x86/xen/p2m.c
576
p2m_pfn == PFN_DOWN(__pa(p2m_missing))) {
arch/x86/xen/p2m.c
584
if (p2m_pfn == PFN_DOWN(__pa(p2m_missing)))
arch/x86/xen/p2m.c
595
pfn_pte(PFN_DOWN(__pa(p2m)), PAGE_KERNEL));
arch/x86/xen/p2m.c
658
if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_missing)))
arch/x86/xen/p2m.c
661
if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity)))
arch/x86/xen/p2m.c
710
pte = (pte_t *)(mfn_to_virt(PFN_DOWN(map_ops[i].host_addr)) +
arch/x86/xen/p2m.c
714
mfn = PFN_DOWN(map_ops[i].dev_bus_addr);
arch/x86/xen/p2m.c
814
pfn = PFN_DOWN(remap->paddr);
arch/x86/xen/p2m.c
815
mfn = PFN_DOWN(remap->maddr);
arch/x86/xen/setup.c
173
e_pfn = PFN_DOWN(entry->addr + entry->size);
arch/x86/xen/setup.c
431
unsigned long start_pfn = PFN_DOWN(start);
arch/x86/xen/setup.c
891
maxmem_pages = EXTRA_MEM_RATIO * min(ini_nr_pages, PFN_DOWN(MAXMEM));
arch/x86/xen/setup.c
911
n_pfns = PFN_DOWN(addr + chunk_size) - pfn_s;
arch/xtensa/kernel/hibernate.c
9
unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin));
arch/xtensa/kernel/pci-dma.c
28
unsigned long pfn = PFN_DOWN(paddr);
arch/xtensa/mm/init.c
57
max_pfn = PFN_DOWN(memblock_end_of_DRAM());
drivers/acpi/pfr_telemetry.c
333
if (io_remap_pfn_range(vma, vma->vm_start, PFN_DOWN(base_addr),
drivers/base/memory.c
864
const unsigned long start_block_id = pfn_to_block_id(PFN_DOWN(start));
drivers/base/memory.c
865
unsigned long end_block_id = pfn_to_block_id(PFN_DOWN(start + size));
drivers/base/memory.c
901
const unsigned long start_block_id = pfn_to_block_id(PFN_DOWN(start));
drivers/base/memory.c
902
const unsigned long end_block_id = pfn_to_block_id(PFN_DOWN(start + size));
drivers/dma/loongson1-apb-dma.c
344
sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(buf_addr)),
drivers/dma/sh/rcar-dmac.c
1137
sg_set_page(&sgl, pfn_to_page(PFN_DOWN(dma_src)), len,
drivers/dma/sh/rcar-dmac.c
1266
sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(src)), period_len,
drivers/dma/sh/shdma-base.c
664
sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_src)), len,
drivers/dma/sh/shdma-base.c
752
sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(src)), period_len,
drivers/edac/bluefield_edac.c
242
PFN_DOWN(ecc_dimm_addr),
drivers/edac/thunderx_edac.c
28
#define phys_to_pfn(phys) (PFN_DOWN(phys))
drivers/gpu/drm/amd/amdkfd/kfd_process.c
2257
PFN_DOWN(__pa(qpd->cwsr_kaddr)),
drivers/hv/mshv_root_main.c
833
unsigned long user_pfn = PFN_DOWN(args.buf_ptr);
drivers/hv/mshv_root_main.c
834
size_t page_count = PFN_DOWN(args.buf_sz);
drivers/hwtracing/intel_th/msu.c
1050
PFN_DOWN(sg_ptr->length));
drivers/hwtracing/intel_th/msu.c
1070
PFN_DOWN(sg_ptr->length));
drivers/hwtracing/intel_th/msu.c
1238
bdesc->next_blk = PFN_DOWN(addr);
drivers/hwtracing/intel_th/msu.c
1426
size_t pgsz = PFN_DOWN(sg->length);
drivers/hwtracing/intel_th/msu.c
305
return PFN_DOWN(msc_win_base_dma(win));
drivers/infiniband/hw/erdma/erdma_verbs.c
1433
err = rdma_user_mmap_io(ctx, vma, PFN_DOWN(entry->address), PAGE_SIZE,
drivers/infiniband/hw/hfi1/file_ops.c
565
vma->vm_pgoff = PFN_DOWN(memaddr);
drivers/infiniband/hw/hfi1/file_ops.c
573
PFN_DOWN(memaddr),
drivers/infiniband/hw/hfi1/file_ops.c
578
PFN_DOWN(__pa(memvirt)),
drivers/infiniband/hw/hfi1/file_ops.c
583
PFN_DOWN(memaddr),
drivers/infiniband/hw/hfi1/pin_system.c
187
ret = pin_system_pages(req, start, len, node, PFN_DOWN(len));
drivers/infiniband/hw/hfi1/pin_system.c
301
page_index = PFN_DOWN(start - cache_entry->rb.addr);
drivers/media/pci/intel/ipu3/ipu3-cio2.c
148
cio2->dummy_lop[i] = PFN_DOWN(cio2->dummy_page_bus_addr);
drivers/media/pci/intel/ipu3/ipu3-cio2.c
184
entry[i].lop_page_addr = PFN_DOWN(cio2->dummy_lop_bus_addr);
drivers/media/pci/intel/ipu3/ipu3-cio2.c
216
entry->lop_page_addr = PFN_DOWN(b->lop_bus_addr[i]);
drivers/media/pci/intel/ipu3/ipu3-cio2.c
225
entry->lop_page_addr = PFN_DOWN(cio2->dummy_lop_bus_addr);
drivers/media/pci/intel/ipu3/ipu3-cio2.c
489
writel(PFN_DOWN(q->fbpt_bus_addr), base + CIO2_REG_CDMABA(CIO2_DMA_CHAN));
drivers/media/pci/intel/ipu3/ipu3-cio2.c
885
b->lop[i][j] = PFN_DOWN(sg_page_iter_dma_address(&sg_iter));
drivers/media/pci/intel/ipu3/ipu3-cio2.c
893
b->lop[i][j] = PFN_DOWN(cio2->dummy_page_bus_addr);
drivers/platform/x86/intel/pmt/class.c
118
unsigned long pfn = PFN_DOWN(phys);
drivers/s390/block/dcssblk.c
944
*pfn = PFN_DOWN(dev_info->start + offset);
drivers/s390/char/sclp_mem.c
234
mem = find_memory_block(pfn_to_section_nr(PFN_DOWN(addr)));
drivers/s390/char/sclp_mem.c
240
mem = find_memory_block(pfn_to_section_nr(PFN_DOWN(addr)));
drivers/s390/char/sclp_mem.c
297
mem = find_memory_block(pfn_to_section_nr(PFN_DOWN(sclp_mem->id * block_size)));
drivers/s390/net/qeth_core.h
940
return PFN_UP(end) - PFN_DOWN(start);
drivers/staging/media/ipu7/ipu7-mmu.c
669
PFN_DOWN(dmap->mmu_info->aperture_start));
drivers/usb/host/xen-hcd.c
582
return PFN_UP(vaddr + length) - PFN_DOWN(vaddr);
drivers/usb/musb/ux500_dma.c
94
sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_addr)), len,
drivers/vdpa/vdpa_user/iova_domain.c
163
unsigned long pfn = PFN_DOWN(orig);
drivers/vdpa/vdpa_user/vduse_dev.c
1241
lock_limit = PFN_DOWN(rlimit(RLIMIT_MEMLOCK));
drivers/vhost/vdpa.c
1022
atomic64_add(PFN_DOWN(size), &dev->mm->pinned_vm);
drivers/vhost/vdpa.c
1125
lock_limit = PFN_DOWN(rlimit(RLIMIT_MEMLOCK));
drivers/vhost/vdpa.c
1496
return vmf_insert_pfn(vma, vmf->address & PAGE_MASK, PFN_DOWN(notify.addr));
drivers/vhost/vdpa.c
929
pinned = PFN_DOWN(map->size);
drivers/vhost/vdpa.c
930
for (pfn = PFN_DOWN(map->addr);
drivers/vhost/vdpa.c
937
atomic64_sub(PFN_DOWN(map->size), &dev->mm->pinned_vm);
drivers/virtio/virtio_mem.c
1350
generic_online_page(pfn_to_page(PFN_DOWN(addr)), order);
drivers/virtio/virtio_mem.c
1352
virtio_mem_set_fake_offline(PFN_DOWN(addr), 1 << order,
drivers/virtio/virtio_mem.c
1748
pfn = PFN_DOWN(virtio_mem_mb_id_to_phys(mb_id) +
drivers/virtio/virtio_mem.c
1750
nr_pages = PFN_DOWN(count * vm->sbm.sb_size);
drivers/virtio/virtio_mem.c
1980
const unsigned long nr_pages = PFN_DOWN(vm->sbm.sb_size) * count;
drivers/virtio/virtio_mem.c
1985
start_pfn = PFN_DOWN(virtio_mem_mb_id_to_phys(mb_id) +
drivers/virtio/virtio_mem.c
2161
const unsigned long start_pfn = PFN_DOWN(virtio_mem_bb_id_to_phys(vm, bb_id));
drivers/virtio/virtio_mem.c
2162
const unsigned long nr_pages = PFN_DOWN(vm->bbm.bb_size);
drivers/virtio/virtio_mem.c
2227
const unsigned long start_pfn = PFN_DOWN(virtio_mem_bb_id_to_phys(vm, bb_id));
drivers/virtio/virtio_mem.c
2228
const unsigned long nr_pages = PFN_DOWN(vm->bbm.bb_size);
drivers/virtio/virtio_mem.c
2246
const unsigned long start_pfn = PFN_DOWN(virtio_mem_bb_id_to_phys(vm, bb_id));
drivers/virtio/virtio_mem.c
2247
const unsigned long nr_pages = PFN_DOWN(vm->bbm.bb_size);
drivers/virtio/virtio_mem.c
930
const unsigned long nr_pages = PFN_DOWN(vm->sbm.sb_size);
drivers/virtio/virtio_mem.c
937
pfn = PFN_DOWN(virtio_mem_mb_id_to_phys(mb_id) +
drivers/virtio/virtio_mem.c
946
const unsigned long nr_pages = PFN_DOWN(vm->sbm.sb_size);
drivers/virtio/virtio_mem.c
953
pfn = PFN_DOWN(virtio_mem_mb_id_to_phys(mb_id) +
drivers/xen/balloon.c
308
pfn = PFN_DOWN(resource->start);
drivers/xen/biomerge.c
15
return bfn1 + PFN_DOWN(vec1->bv_offset + vec1->bv_len) == bfn2;
drivers/xen/swiotlb-xen.c
110
if (pfn_valid(PFN_DOWN(paddr)))
drivers/xen/swiotlb-xen.c
265
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dev_addr))))
drivers/xen/swiotlb-xen.c
290
if (pfn_valid(PFN_DOWN(dma_to_phys(hwdev, dev_addr))))
drivers/xen/swiotlb-xen.c
311
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr))))
drivers/xen/swiotlb-xen.c
334
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr))))
drivers/xen/unpopulated-alloc.c
112
xen_pfn_t pfn = PFN_DOWN(res->start);
include/linux/efi.h
1039
*npages = PFN_UP(*addr + (*npages<<EFI_PAGE_SHIFT)) - PFN_DOWN(*addr);
include/linux/memblock.h
540
return PFN_DOWN(reg->base + reg->size);
include/linux/memblock.h
551
return PFN_DOWN(reg->base);
include/linux/memory.h
194
return pfn_to_block_id(PFN_DOWN(phys));
kernel/dma/coherent.c
61
dma_mem->pfn_base = PFN_DOWN(phys_addr);
kernel/dma/direct.c
569
for (m = dev->dma_range_map; PFN_DOWN(m->size); m++) {
kernel/dma/direct.c
570
unsigned long cpu_start_pfn = PFN_DOWN(m->cpu_start);
kernel/dma/direct.c
573
start_pfn - cpu_start_pfn < PFN_DOWN(m->size))
kernel/dma/direct.c
598
start_pfn = PFN_DOWN(bdr->cpu_start) + PFN_DOWN(bdr->size);
kernel/dma/direct.c
608
return !walk_system_ram_range(0, PFN_DOWN(ULONG_MAX) + 1, dev,
kernel/dma/swiotlb.c
1790
return pfn_to_page(PFN_DOWN(tlb_addr));
kernel/dma/swiotlb.c
864
unsigned long pfn = PFN_DOWN(orig_addr);
kernel/kexec_core.c
977
npages = PFN_UP(eaddr) - PFN_DOWN(addr);
kernel/liveupdate/kexec_handover.c
1438
ulong start_pfn = pageblock_start_pfn(PFN_DOWN(start));
kernel/resource.c
1849
if (PFN_DOWN(p->start) <= PFN_DOWN(addr) &&
kernel/resource.c
1850
PFN_DOWN(p->end) >= PFN_DOWN(end))
kernel/resource.c
575
end_pfn = PFN_DOWN(res.end + 1);
mm/cma.c
296
cma->ranges[0].base_pfn = PFN_DOWN(base);
mm/cma.c
297
cma->ranges[0].early_pfn = PFN_DOWN(base);
mm/kasan/init.c
98
zero_pte = pfn_pte(PFN_DOWN(__pa_symbol(kasan_early_shadow_page)),
mm/kasan/shadow.c
310
index = PFN_DOWN(addr - data->start);
mm/memblock.c
1399
if (PFN_UP(r->base) >= PFN_DOWN(r->base + r->size))
mm/memblock.c
1412
*out_end_pfn = PFN_DOWN(r->base + r->size);
mm/memblock.c
1772
end = PFN_DOWN(base + size);
mm/memblock.c
1983
*start_pfn = PFN_DOWN(type->regions[mid].base);
mm/memblock.c
1984
*end_pfn = PFN_DOWN(type->regions[mid].base + type->regions[mid].size);
mm/memblock.c
2230
unsigned long end_pfn = PFN_DOWN(end);
mm/memblock.c
2285
nid = early_pfn_to_nid(PFN_DOWN(start));
mm/memblock.c
980
for (pfn = PFN_UP(start); pfn < PFN_DOWN(end); pfn++)
mm/memory_hotplug.c
1427
mem = find_memory_block(pfn_to_section_nr(PFN_DOWN(cur_start)));
mm/memory_hotplug.c
1570
register_memory_blocks_under_node_hotplug(nid, PFN_DOWN(start),
mm/mm_init.c
1323
end_pfn = PFN_DOWN(end_addr);
mm/mm_init.c
2069
unsigned long epfn = PFN_DOWN(end);
mm/mm_init.c
786
for_each_valid_pfn(pfn, PFN_DOWN(start), PFN_UP(end)) {
mm/numa_emulation.c
38
unsigned long end_pfn = PFN_DOWN(end);
mm/percpu.c
1078
start = PFN_DOWN(bit_off * PCPU_MIN_ALLOC_SIZE);
mm/percpu.c
1870
rs = PFN_DOWN(off);
mm/percpu.c
2726
pcpu_nr_populated += PFN_DOWN(size_sum);
mm/percpu.c
3108
PFN_DOWN(size_sum), ai->static_size, ai->reserved_size,
mm/swap_state.c
791
win = __swapin_nr_pages(PFN_DOWN(prev_faddr), PFN_DOWN(faddr), hits,
mm/swap_state.c
844
ilx = targ_ilx - PFN_DOWN(vmf->address - start);
mm/vmalloc.c
129
pfn += PFN_DOWN(size);
mm/vmalloc.c
135
} while (pte += PFN_DOWN(size), addr += size, addr != end);
sound/soc/intel/catpt/pcm.c
162
pfn = PFN_DOWN(snd_sgbuf_get_addr(databuf, i * PAGE_SIZE));
sound/soc/intel/catpt/pcm.c
400
rinfo.ring_first_page_pfn = PFN_DOWN(snd_sgbuf_get_addr(dmab, 0));
sound/soc/renesas/siu_pcm.c
116
sg_set_page(&sg, pfn_to_page(PFN_DOWN(buff)),
sound/soc/renesas/siu_pcm.c
166
sg_set_page(&sg, pfn_to_page(PFN_DOWN(buff)),