Symbol: __pa
arch/alpha/include/asm/page.h
81
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/alpha/include/asm/page.h
82
#define virt_addr_valid(kaddr) pfn_valid((__pa(kaddr) >> PAGE_SHIFT))
arch/alpha/kernel/core_irongate.c
231
if (initrd_end && __pa(initrd_end) > pci_mem) {
arch/alpha/kernel/pci_iommu.c
472
#define SG_ENT_PHYS_ADDRESS(SG) __pa(SG_ENT_VIRT_ADDRESS(SG))
arch/alpha/kernel/setup.c
273
if (!start || __pa(start) + size > mem_limit) {
arch/arc/include/asm/page.h
127
return __pa(kaddr) >> PAGE_SHIFT;
arch/arc/kernel/devtree.c
65
if (!early_init_dt_scan(dt, __pa(dt)))
arch/arc/mm/init.c
127
__pa(_end) - CONFIG_LINUX_LINK_BASE);
arch/arm/include/asm/cacheflush.h
393
outer_clean_range(__pa(_p), __pa(_p + size));
arch/arm/include/asm/cacheflush.h
415
outer_flush_range(__pa(_p), __pa(_p + size));
arch/arm/include/asm/kfence.h
14
unsigned long pfn = PFN_DOWN(__pa(addr));
arch/arm/include/asm/pgalloc.h
130
__pmd_populate(pmdp, __pa(ptep), _PAGE_KERNEL_TABLE);
arch/arm/include/asm/pgalloc.h
28
set_pud(pud, __pud(__pa(pmd) | PMD_TYPE_TABLE));
arch/arm/kernel/devtree.c
203
if (!dt_virt || !early_init_dt_verify(dt_virt, __pa(dt_virt)))
arch/arm/kernel/setup.c
1014
unsigned long long lowmem_max = __pa(high_memory - 1) + 1;
arch/arm/mach-highbank/pm.c
24
return psci_ops.cpu_suspend(HIGHBANK_SUSPEND_PARAM, __pa(cpu_resume));
arch/arm/mach-omap2/control.c
264
(u32) __pa(omap3_secure_ram_storage);
arch/arm/mach-omap2/omap-secure.c
131
param[1] = __pa(addr); /* Physical address for saving */
arch/arm/mach-omap2/omap-secure.c
136
ret = save_secure_ram_context(__pa(param));
arch/arm/mach-omap2/omap-secure.c
174
outer_clean_range(__pa(param), __pa(param + 5));
arch/arm/mach-omap2/omap-secure.c
175
ret = omap_smc3(idx, process, flag, __pa(param));
arch/arm/mach-omap2/omap-secure.c
81
outer_clean_range(__pa(param), __pa(param + 5));
arch/arm/mach-omap2/omap-secure.c
82
ret = omap_smc2(idx, flag, __pa(param));
arch/arm/mach-omap2/omap-smp.c
295
if ((addr >= __pa(PAGE_OFFSET)) && (addr <= __pa(__bss_start)))
arch/arm/mach-shmobile/smp-emev2.c
37
iowrite32(__pa(shmobile_boot_vector), smu + SMU_GENERAL_REG0);
arch/arm/mach-shmobile/smp-r8a7779.c
51
writel(__pa(shmobile_boot_vector), base + AVECR);
arch/arm/mach-shmobile/smp-sh73a0.c
58
writel(__pa(shmobile_boot_vector), sysc + SBAR);
arch/arm/mach-versatile/integrator.c
93
memblock_reserve(PHYS_OFFSET, __pa(swapper_pg_dir) - PHYS_OFFSET);
arch/arm/mach-zynq/common.c
54
if (!__pa(PAGE_OFFSET))
arch/arm/mach-zynq/common.c
55
memblock_reserve(__pa(PAGE_OFFSET), 0x80000);
arch/arm/mach-zynq/platsmp.c
46
if (__pa(PAGE_OFFSET)) {
arch/arm/mach-zynq/platsmp.c
70
if (__pa(PAGE_OFFSET))
arch/arm/mm/dma-mapping.c
132
outer_flush_range(__pa(ptr), __pa(ptr) + size);
arch/arm/mm/init.c
183
memblock_reserve(__pa(KERNEL_START), KERNEL_END - KERNEL_START);
arch/arm/mm/init.c
231
memblock_phys_free(PHYS_OFFSET, __pa(swapper_pg_dir) - PHYS_OFFSET);
arch/arm/mm/kasan_init.c
237
__pgd(__pa(tmp_pmd_table) | PMD_TYPE_TABLE | L_PGD_SWAPPER));
arch/arm/mm/kasan_init.c
33
return memblock_alloc_try_nid_raw(size, size, __pa(MAX_DMA_ADDRESS),
arch/arm/mm/kasan_init.c
39
return memblock_alloc_try_nid(size, size, __pa(MAX_DMA_ADDRESS),
arch/arm/mm/mmap.c
155
if (addr + size > __pa(high_memory - 1) + 1)
arch/arm/mm/mmu.c
1348
memblock_reserve(__pa(swapper_pg_dir), SWAPPER_PG_DIR_SIZE);
arch/arm/mm/mmu.c
1355
memblock_reserve(PHYS_OFFSET, __pa(swapper_pg_dir) - PHYS_OFFSET);
arch/arm/mm/mmu.c
1587
phys_addr_t kernel_x_end = round_up(__pa(__init_end), SECTION_SIZE);
arch/arm/mm/mmu.c
1656
lpae_pgtables_remap = (pgtables_remap *)(unsigned long)__pa(lpae_pgtables_remap_asm);
arch/arm/mm/mmu.c
1657
pa_pgd = __pa(swapper_pg_dir);
arch/arm/mm/mmu.c
1760
kernel_sec_end = round_up(__pa(_end), SECTION_SIZE);
arch/arm/mm/mmu.c
749
__pmd_populate(pmd, __pa(pte), prot);
arch/arm/mm/pmsa-v7.c
260
num = allocate_region(CONFIG_XIP_PHYS_ADDR, __pa(_exiprom) - CONFIG_XIP_PHYS_ADDR,
arch/arm/mm/pmsa-v8.c
255
subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END));
arch/arm/mm/pmsa-v8.c
256
subtract_range(io, ARRAY_SIZE(io), __pa(KERNEL_START), __pa(KERNEL_END));
arch/arm/mm/pmsa-v8.c
260
subtract_range(mem, ARRAY_SIZE(mem), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom));
arch/arm/mm/pmsa-v8.c
261
subtract_range(io, ARRAY_SIZE(io), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom));
arch/arm/mm/pmsa-v8.c
277
err |= pmsav8_setup_fixed(PMSAv8_XIP_REGION, CONFIG_XIP_PHYS_ADDR, __pa(_exiprom));
arch/arm/mm/pmsa-v8.c
280
err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END));
arch/arm64/include/asm/memory.h
113
#define DIRECT_MAP_PHYSMEM_END __pa(PAGE_END - 1)
arch/arm64/include/asm/pgalloc.h
112
__pmd_populate(pmdp, __pa(ptep),
arch/arm64/include/asm/pgalloc.h
34
__pud_populate(pudp, __pa(pmdp), pudval);
arch/arm64/include/asm/pgalloc.h
56
__p4d_populate(p4dp, __pa(pudp), p4dval);
arch/arm64/include/asm/pgalloc.h
85
__pgd_populate(pgdp, __pa(p4dp), pgdval);
arch/arm64/kernel/machine_kexec.c
138
kimage->arch.ttbr1 = __pa(trans_pgd);
arch/arm64/kernel/machine_kexec.c
143
kimage->arch.kern_reloc = __pa(reloc_code);
arch/arm64/kvm/arm.c
2734
err = create_hyp_stack(__pa(stack_base), &params->stack_hyp_va);
arch/arm64/kvm/arm.c
2746
params->stack_pa = __pa(stack_base);
arch/arm64/kvm/mmu.c
1007
mmu->pgd_phys = __pa(pgt->pgd);
arch/arm64/kvm/mmu.c
2267
return __pa(hyp_pgtable->pgd);
arch/arm64/kvm/mmu.c
274
return __pa(addr);
arch/arm64/kvm/mmu.c
440
return __pa(kaddr);
arch/arm64/kvm/mmu.c
553
start = ALIGN_DOWN(__pa(from), PAGE_SIZE);
arch/arm64/kvm/mmu.c
554
end = PAGE_ALIGN(__pa(to));
arch/arm64/kvm/mmu.c
573
start = ALIGN_DOWN(__pa(from), PAGE_SIZE);
arch/arm64/kvm/mmu.c
574
end = PAGE_ALIGN(__pa(to));
arch/arm64/kvm/va_layout.c
46
hyp_physvirt_offset = (s64)__pa(kern_va) - (s64)hyp_va;
arch/arm64/mm/kasan_init.c
38
__pa(MAX_DMA_ADDRESS),
arch/arm64/mm/kasan_init.c
43
__pa(MAX_DMA_ADDRESS));
arch/arm64/mm/kasan_init.c
45
return __pa(p);
arch/arm64/mm/kasan_init.c
51
__pa(MAX_DMA_ADDRESS),
arch/arm64/mm/kasan_init.c
57
__pa(MAX_DMA_ADDRESS));
arch/arm64/mm/kasan_init.c
59
return __pa(p);
arch/arm64/mm/mmu.c
1260
kpti_ng_temp_alloc = kpti_ng_temp_pgd_pa = __pa(kpti_ng_temp_pgd);
arch/arm64/mm/mmu.c
1278
ret = __create_pgd_mapping_locked(kpti_ng_temp_pgd, __pa(alloc),
arch/arm64/mm/mmu.c
1758
pmd_set_huge(pmdp, __pa(p), __pgprot(PROT_SECT_NORMAL));
arch/arm64/mm/mmu.c
1940
phys_addr_t start_linear_pa = __pa(_PAGE_OFFSET(vabits_actual));
arch/arm64/mm/mmu.c
1941
phys_addr_t end_linear_pa = __pa(PAGE_END - 1);
arch/csky/abiv1/inc/abi/ckmmu.h
93
cpwcr("cpcr29", __pa(pgd) | BIT(0));
arch/csky/abiv2/inc/abi/ckmmu.h
131
:"r"(asid), "r"(__pa(pgd) | BIT(0))
arch/csky/include/asm/page.h
71
#define __pa_symbol(x) __pa(RELOC_HIDE((unsigned long)(x), 0))
arch/csky/include/asm/page.h
75
return __pa(kaddr) >> PAGE_SHIFT;
arch/csky/include/asm/pgalloc.h
16
set_pmd(pmd, __pmd(__pa(pte)));
arch/csky/include/asm/pgalloc.h
22
set_pmd(pmd, __pmd(__pa(page_address(pte))));
arch/csky/include/asm/pgtable.h
119
return pmd_val(pmd) == __pa(invalid_pte_table);
arch/csky/include/asm/pgtable.h
126
return (pmd_val(pmd) != __pa(invalid_pte_table));
arch/csky/include/asm/pgtable.h
131
pmd_val(*p) = (__pa(invalid_pte_table));
arch/csky/kernel/setup.c
158
early_init_dt_scan(__dtb_start, __pa(dtb_start));
arch/csky/kernel/setup.c
160
early_init_dt_scan(dtb_start, __pa(dtb_start));
arch/csky/kernel/setup.c
25
if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) {
arch/csky/kernel/setup.c
32
if (memblock_is_region_reserved(__pa(initrd_start), size)) {
arch/csky/kernel/setup.c
34
__pa(initrd_start), size);
arch/csky/kernel/setup.c
38
memblock_reserve(__pa(initrd_start), size);
arch/csky/kernel/setup.c
68
memblock_reserve(__pa(_start), _end - _start);
arch/csky/mm/init.c
115
set_pmd(pmd, __pmd(__pa(pte)));
arch/csky/mm/init.c
55
p[i] = __pa(invalid_pte_table);
arch/csky/mm/init.c
66
swapper_pg_dir[i].pgd = __pa(invalid_pte_table);
arch/csky/mm/init.c
70
__pa(kernel_pte_tables + (PTRS_PER_PTE * (i - USER_PTRS_PER_PGD)));
arch/hexagon/include/asm/io.h
24
return __pa(address);
arch/hexagon/include/asm/page.h
120
return __pa(kaddr) >> PAGE_SHIFT;
arch/hexagon/include/asm/page.h
90
#define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr)))
arch/hexagon/include/asm/page.h
95
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
arch/hexagon/include/asm/pgalloc.h
39
mm->context.ptbase = __pa(pgd);
arch/hexagon/include/asm/pgalloc.h
77
set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
arch/hexagon/include/asm/pgalloc.h
85
set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
arch/hexagon/mm/init.c
75
init_mm.context.ptbase = __pa(init_mm.pgd);
arch/loongarch/include/asm/io.h
73
(likely((unsigned long)kaddr < vm_map_base)) ? __pa((unsigned long)kaddr) : \
arch/loongarch/include/asm/page.h
81
#define virt_to_page(kaddr) phys_to_page(__pa(kaddr))
arch/loongarch/kernel/kexec_elf.c
42
*text_offset = __pa(phdr->p_paddr);
arch/loongarch/kernel/kexec_elf.c
92
image->start = kernel_segment->mem + __pa(ehdr.e_entry) - text_offset;
arch/loongarch/kernel/setup.c
301
early_init_dt_scan(fdt_pointer, __pa(fdt_pointer));
arch/loongarch/kernel/traps.c
1142
csr_write(__pa(eentry), LOONGARCH_CSR_MERRENTRY);
arch/loongarch/kernel/traps.c
1143
csr_write(__pa(tlbrentry), LOONGARCH_CSR_TLBRENTRY);
arch/loongarch/kernel/traps.c
1188
unsigned long uncached_eentry = TO_UNCACHE(__pa(eentry));
arch/loongarch/kvm/mmu.c
167
kvm_set_pte(entry, __pa(child));
arch/loongarch/kvm/mmu.c
745
kvm_set_pte(ptep, __pa(child));
arch/loongarch/kvm/vcpu.c
1543
vcpu->arch.kvm_pgd = __pa(vcpu->kvm->arch.pgd);
arch/loongarch/mm/kasan_init.c
112
__pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, node);
arch/loongarch/mm/kasan_init.c
115
__func__, PAGE_SIZE, PAGE_SIZE, node, __pa(MAX_DMA_ADDRESS));
arch/loongarch/mm/kasan_init.c
117
return __pa(p);
arch/loongarch/mm/kasan_init.c
20
(__pa(pgd_val(pgd)) == (unsigned long)__pa(kasan_early_shadow_p4d)))
arch/loongarch/mm/kasan_init.c
27
(__pa(p4d_val(p4d)) == (unsigned long)__pa(kasan_early_shadow_pud)))
arch/loongarch/mm/kasan_init.c
34
(__pa(pud_val(pud)) == (unsigned long)__pa(kasan_early_shadow_pmd)))
arch/loongarch/mm/kasan_init.c
38
(__pa(pmd_val(pmd)) == (unsigned long)__pa(kasan_early_shadow_pte)))
arch/loongarch/mm/kasan_init.c
41
((pte_val(pte) & _PFN_MASK) == (unsigned long)__pa(kasan_early_shadow_page)))
arch/loongarch/mm/pageattr.c
171
return memblock_is_memory(__pa(addr));
arch/loongarch/mm/pgtable.c
14
return phys_to_page(__pa(kaddr));
arch/loongarch/power/hibernate.c
44
unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin));
arch/loongarch/power/hibernate.c
45
unsigned long nosave_end_pfn = PFN_UP(__pa(&__nosave_end));
arch/m68k/include/asm/cacheflush_mm.h
237
unsigned long paddr = __pa(vaddr);
arch/m68k/include/asm/motorola_pgtable.h
97
pud_val(*pudp) = _PAGE_TABLE | _PAGE_ACCESSED | __pa(pmdp);
arch/m68k/include/asm/page_mm.h
125
return __pa(kaddr) >> PAGE_SHIFT;
arch/m68k/include/asm/page_no.h
23
return __pa(kaddr) >> PAGE_SHIFT;
arch/m68k/include/asm/sun3_pgalloc.h
25
pmd_val(*pmd) = __pa((unsigned long)pte);
arch/m68k/include/asm/sun3_pgalloc.h
30
pmd_val(*pmd) = __pa((unsigned long)page_address(page));
arch/m68k/include/asm/sun3_pgtable.h
18
#define VTOP(addr) __pa(addr)
arch/m68k/include/asm/virtconvert.h
22
return __pa(address);
arch/m68k/mm/sun3mmu.c
63
pg_table = (pte_t *) __pa (next_pgtable);
arch/m68k/sun3/config.c
116
max_pfn = num_pages = __pa(memory_end) >> PAGE_SHIFT;
arch/m68k/sun3/mmu_emu.c
143
for (i=0; i < __pa(bootmem_end) / SUN3_PMEG_SIZE ; ++i)
arch/m68k/sun3x/dvma.c
127
__pa(kaddr), vaddr);
arch/m68k/sun3x/dvma.c
162
dvma_entry_set(index, __pa(kaddr));
arch/microblaze/include/asm/page.h
122
return phys_to_pfn(__pa(vaddr));
arch/microblaze/include/asm/page.h
98
# define virt_to_page(kaddr) (pfn_to_page(__pa(kaddr) >> PAGE_SHIFT))
arch/microblaze/include/asm/pgtable.h
374
#define pmd_pfn(pmd) (__pa(pmd_val(pmd)) >> PAGE_SHIFT)
arch/microblaze/include/asm/pgtable.h
377
#define pmd_page(pmd) (pfn_to_page(__pa(pmd_val(pmd)) >> PAGE_SHIFT))
arch/microblaze/kernel/prom.c
21
early_init_dt_scan(params, __pa(params));
arch/microblaze/mm/init.c
196
kstart = __pa(CONFIG_KERNEL_START); /* kernel start */
arch/mips/bcm47xx/prom.c
164
off = EXTVBASE + __pa(off);
arch/mips/dec/prom/memory.c
111
end = __pa(&_text) - 0x00020000;
arch/mips/dec/prom/memory.c
114
end = __pa(&_text);
arch/mips/include/asm/io.h
94
return __pa(address);
arch/mips/include/asm/page.h
195
#define __pa_symbol_nodebug(x) __pa(RELOC_HIDE((unsigned long)(x), 0))
arch/mips/jazz/jazzdma.c
510
return (void *)(UNCAC_BASE + __pa(ret));
arch/mips/kernel/prom.c
44
if (!early_init_dt_scan(bph, __pa(bph)))
arch/mips/kernel/relocate.c
340
early_init_dt_scan(fdt, __pa(fdt));
arch/mips/kernel/setup.c
168
end = __pa(initrd_end);
arch/mips/kernel/setup.c
170
initrd_start = (unsigned long)__va(__pa(initrd_start));
arch/mips/kernel/setup.c
222
if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) {
arch/mips/kernel/setup.c
229
memblock_reserve(__pa(initrd_start), size);
arch/mips/kernel/traps.c
2303
unsigned long uncached_ebase = CKSEG1ADDR_OR_64BIT(__pa(ebase));
arch/mips/kernel/vdso.c
148
gic_pfn = PFN_DOWN(__pa(gic_base));
arch/mips/mm/dma-noncoherent.c
54
return (void *)(__pa(addr) + UNCAC_BASE);
arch/mips/power/cpu.c
39
unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin));
arch/mips/power/cpu.c
40
unsigned long nosave_end_pfn = PFN_UP(__pa(&__nosave_end));
arch/nios2/kernel/prom.c
30
early_init_dt_scan(params, __pa(params));
arch/nios2/kernel/prom.c
40
early_init_dt_scan(params, __pa(params));
arch/openrisc/include/asm/page.h
67
return __pa(kaddr) >> PAGE_SHIFT;
arch/openrisc/include/asm/pgalloc.h
29
set_pmd(pmd, __pmd(_KERNPG_TABLE + __pa(pte)))
arch/openrisc/include/asm/pgtable.h
281
pte_val(pte) = __pa(page) | pgprot_val(pgprot);
arch/openrisc/kernel/dma.c
37
local_dcache_range_flush(__pa(addr), __pa(next));
arch/openrisc/kernel/prom.c
25
early_init_dt_scan(params, __pa(params));
arch/openrisc/kernel/setup.c
76
memblock_reserve(__pa(_stext), _end - _stext);
arch/openrisc/kernel/setup.c
84
memblock_reserve(__pa(aligned_start), aligned_end - aligned_start);
arch/openrisc/mm/init.c
99
set_pmd(pme, __pmd(_KERNPG_TABLE + __pa(pte)));
arch/parisc/boot/compressed/misc.c
315
kernel_len = __pa(SZ_end) - __pa(SZparisc_kernel_start);
arch/parisc/include/asm/io.h
8
#define virt_to_phys(a) ((unsigned long)__pa(a))
arch/parisc/include/asm/mmu_context.h
58
mtctl(__pa(__ldcw_align(&pgd_lock->rlock.raw_lock)), 28);
arch/parisc/include/asm/mmu_context.h
60
mtctl(__pa(next->pgd), 25);
arch/parisc/include/asm/page.h
166
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
arch/parisc/include/asm/page.h
168
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/parisc/include/asm/pgalloc.h
29
(__u32)(__pa((unsigned long)pmd) >> PxD_VALUE_SHIFT)));
arch/parisc/include/asm/pgalloc.h
54
+ (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT)));
arch/parisc/kernel/cache.c
367
flush_dcache_page_asm(__pa(vaddr), vaddr);
arch/parisc/kernel/cache.c
383
flush_icache_page_asm(__pa(vaddr), vaddr);
arch/parisc/kernel/firmware.c
1026
__pa(pdc_result), hpa, __pa(tbl));
arch/parisc/kernel/firmware.c
1051
__pa(pdc_result), hpa, cfg_addr&~3UL, 4UL);
arch/parisc/kernel/firmware.c
1074
__pa(pdc_result), hpa,
arch/parisc/kernel/firmware.c
1094
retval = mem_pdc_call(PDC_TOD, PDC_TOD_READ, __pa(pdc_result), 0);
arch/parisc/kernel/firmware.c
1109
retval = mem_pdc_call(PDC_MEM, PDC_MEM_MEMINFO, __pa(pdc_result), 0);
arch/parisc/kernel/firmware.c
1124
retval = mem_pdc_call(PDC_MEM, PDC_MEM_READ_PDT, __pa(pdc_result),
arch/parisc/kernel/firmware.c
1125
__pa(pdt_entries_ptr));
arch/parisc/kernel/firmware.c
1154
retval = mem_pdc_call(PDC_PIM, PDC_PIM_TOC, __pa(pdc_result),
arch/parisc/kernel/firmware.c
1155
__pa(ret), sizeof(*ret));
arch/parisc/kernel/firmware.c
1170
retval = mem_pdc_call(PDC_PIM, PDC_PIM_TOC, __pa(pdc_result),
arch/parisc/kernel/firmware.c
1171
__pa(ret), sizeof(*ret));
arch/parisc/kernel/firmware.c
1204
retval = mem_pdc_call(PDC_MEM, PDC_MEM_TABLE, __pa(pdc_result), __pa(pdc_result2), entries);
arch/parisc/kernel/firmware.c
1262
retval = mem_pdc_call(PDC_SOFT_POWER, PDC_SOFT_POWER_INFO, __pa(pdc_result), 0);
arch/parisc/kernel/firmware.c
1292
retval = mem_pdc_call(PDC_SOFT_POWER, PDC_SOFT_POWER_ENABLE, __pa(pdc_result), sw_control);
arch/parisc/kernel/firmware.c
1308
retval = mem_pdc_call(PDC_SOFT_POWER, PDC_SOFT_POWER_ENABLE, __pa(pdc_result), sw_control);
arch/parisc/kernel/firmware.c
1387
PAGE0->mem_cons.spa, __pa(PAGE0->mem_cons.dp.layers),
arch/parisc/kernel/firmware.c
1388
__pa(pdc_result), 0, __pa(iodc_dbuf), i, 0);
arch/parisc/kernel/firmware.c
1415
PAGE0->mem_kbd.spa, __pa(PAGE0->mem_kbd.dp.layers),
arch/parisc/kernel/firmware.c
1416
__pa(pdc_result), 0, __pa(iodc_dbuf), 1, 0);
arch/parisc/kernel/firmware.c
1466
retval = mem_pdc_call(PDC_PAT_CELL, PDC_PAT_CELL_GET_NUMBER, __pa(pdc_result));
arch/parisc/kernel/firmware.c
1492
retval = mem_pdc_call(PDC_PAT_CELL, PDC_PAT_CELL_MODULE, __pa(pdc_result),
arch/parisc/kernel/firmware.c
1493
ploc, mod, view_type, __pa(&result));
arch/parisc/kernel/firmware.c
1522
__pa(pdc_result), __pa(&result), *actcnt,
arch/parisc/kernel/firmware.c
1547
__pa(&pdc_result), hpa);
arch/parisc/kernel/firmware.c
1569
__pa(pdc_result), cell_num);
arch/parisc/kernel/firmware.c
1590
__pa(r_addr), cell_num);
arch/parisc/kernel/firmware.c
1611
retval = mem_pdc_call(PDC_PAT_PD, PDC_PAT_PD_GET_ADDR_MAP, __pa(pdc_result),
arch/parisc/kernel/firmware.c
1612
__pa(pdc_result2), count, offset);
arch/parisc/kernel/firmware.c
162
__pa(pdc_result), 0);
arch/parisc/kernel/firmware.c
1635
__pa(pdc_result));
arch/parisc/kernel/firmware.c
1661
__pa(pdc_result));
arch/parisc/kernel/firmware.c
1686
__pa(pdc_result), pci_addr, pci_size);
arch/parisc/kernel/firmware.c
1730
__pa(&pdc_result));
arch/parisc/kernel/firmware.c
1753
__pa(&pdc_result), cell);
arch/parisc/kernel/firmware.c
1777
__pa(&pdc_result), parisc_cell_num,
arch/parisc/kernel/firmware.c
1778
__pa(pdt_entries_ptr));
arch/parisc/kernel/firmware.c
1809
__pa(&pdc_result), __pa(pdt_entries_ptr),
arch/parisc/kernel/firmware.c
1838
__pa(&pdc_result), phys_addr);
arch/parisc/kernel/firmware.c
250
retval = mem_pdc_call(PDC_INSTR, 0UL, __pa(pdc_result));
arch/parisc/kernel/firmware.c
275
__pa(pdc_result), __pa(pdc_result2), len);
arch/parisc/kernel/firmware.c
300
retval = mem_pdc_call(PDC_PAT_CHASSIS_LOG, PDC_PAT_CHASSIS_WRITE_LOG, __pa(&state), __pa(&data));
arch/parisc/kernel/firmware.c
366
__pa(pdc_result));
arch/parisc/kernel/firmware.c
382
retval = mem_pdc_call(PDC_CHASSIS, PDC_CHASSIS_WARN, __pa(pdc_result));
arch/parisc/kernel/firmware.c
393
ret = mem_pdc_call(PDC_COPROC, PDC_COPROC_CFG, __pa(pdc_result));
arch/parisc/kernel/firmware.c
440
retval = mem_pdc_call(PDC_IODC, PDC_IODC_READ, __pa(pdc_result), hpa,
arch/parisc/kernel/firmware.c
441
index, __pa(pdc_result2), iodc_data_size);
arch/parisc/kernel/firmware.c
468
retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_MODULE, __pa(pdc_result),
arch/parisc/kernel/firmware.c
469
__pa(pdc_result2), mod_index);
arch/parisc/kernel/firmware.c
495
retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_ADDRESS, __pa(pdc_result),
arch/parisc/kernel/firmware.c
517
retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_INFO, __pa(pdc_result), 0);
arch/parisc/kernel/firmware.c
540
retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_SYSMODEL, __pa(pdc_result),
arch/parisc/kernel/firmware.c
541
os_id, __pa(name));
arch/parisc/kernel/firmware.c
570
retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_VERSIONS, __pa(pdc_result), id);
arch/parisc/kernel/firmware.c
592
retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_CPU_ID, __pa(pdc_result), 0);
arch/parisc/kernel/firmware.c
614
retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_CAPABILITIES, __pa(pdc_result), 0);
arch/parisc/kernel/firmware.c
643
__pa(orig_prod_num), __pa(current_prod_num), __pa(serial_no));
arch/parisc/kernel/firmware.c
662
retval = mem_pdc_call(PDC_CACHE, PDC_CACHE_INFO, __pa(pdc_result), 0);
arch/parisc/kernel/firmware.c
683
retval = mem_pdc_call(PDC_CACHE, PDC_CACHE_RET_SPID, __pa(pdc_result), 0);
arch/parisc/kernel/firmware.c
706
retval = mem_pdc_call(PDC_BLOCK_TLB, PDC_BTLB_INFO, __pa(pdc_result), 0);
arch/parisc/kernel/firmware.c
769
retval = mem_pdc_call(PDC_MEM_MAP, PDC_MEM_MAP_HPA, __pa(pdc_result),
arch/parisc/kernel/firmware.c
770
__pa(pdc_result2));
arch/parisc/kernel/firmware.c
791
__pa(pdc_result), hpa);
arch/parisc/kernel/firmware.c
821
__pa(pdc_result), count);
arch/parisc/kernel/firmware.c
849
__pa(pdc_result), count);
arch/parisc/kernel/firmware.c
871
retval = mem_pdc_call(PDC_STABLE, PDC_STABLE_RETURN_SIZE, __pa(pdc_result));
arch/parisc/kernel/firmware.c
943
__pa(pdc_result), __pa(hwpath));
arch/parisc/kernel/firmware.c
999
__pa(pdc_result), hpa);
arch/parisc/kernel/kexec_file.c
36
image->start = __pa(elf_info.ehdr->e_entry);
arch/parisc/kernel/kexec_file.c
39
image->segment[i].mem = __pa(image->segment[i].mem);
arch/parisc/kernel/pci-dma.c
416
paddr = __pa(paddr);
arch/parisc/kernel/pdt.c
196
pdt_status.first_dbe_loc <= __pa((unsigned long)&_end))
arch/parisc/kernel/setup.c
283
if (__pa((unsigned long) &_end) >= KERNEL_INITIAL_SIZE)
arch/parisc/kernel/toc.c
114
PAGE0->vec_toc = __pa(toc_code) & 0xffffffff;
arch/parisc/kernel/toc.c
116
PAGE0->vec_toc_hi = __pa(toc_code) >> 32;
arch/parisc/kernel/traps.c
841
ivap[6] = (u32)__pa(os_hpmc);
arch/parisc/mm/init.c
287
memblock_reserve(__pa(KERNEL_BINARY_TEXT_START),
arch/parisc/mm/init.c
303
if (__pa(initrd_start) < mem_max) {
arch/parisc/mm/init.c
306
if (__pa(initrd_end) > mem_max) {
arch/parisc/mm/init.c
307
initrd_reserve = mem_max - __pa(initrd_start);
arch/parisc/mm/init.c
312
printk(KERN_INFO "initrd: reserving %08lx-%08lx (mem_max %08lx)\n", __pa(initrd_start), __pa(initrd_start) + initrd_reserve, mem_max);
arch/parisc/mm/init.c
314
memblock_reserve(__pa(initrd_start), initrd_reserve);
arch/parisc/mm/init.c
360
ro_start = __pa((unsigned long)_text);
arch/parisc/mm/init.c
361
ro_end = __pa((unsigned long)&data_start);
arch/parisc/mm/init.c
362
kernel_start = __pa((unsigned long)&__init_begin);
arch/parisc/mm/init.c
363
kernel_end = __pa((unsigned long)&_end);
arch/parisc/mm/init.c
443
map_pages(start, __pa(start), end-start,
arch/parisc/mm/init.c
458
map_pages(init_end, __pa(init_end), kernel_end - init_end,
arch/parisc/mm/init.c
467
map_pages(init_begin, __pa(init_begin), init_end - init_begin,
arch/parisc/mm/init.c
471
map_pages(init_begin, __pa(init_begin), init_end - init_begin,
arch/parisc/mm/init.c
498
map_pages(start, __pa(start), end - start, PAGE_KERNEL, 0);
arch/parisc/mm/init.c
638
map_pages(initrd_start, __pa(initrd_start),
arch/parisc/mm/init.c
663
map_pages(linux_gateway_page_addr, __pa(&linux_gateway_page),
arch/parisc/mm/init.c
738
pdc_btlb_insert(start >> PAGE_SHIFT, __pa(start) >> PAGE_SHIFT,
arch/powerpc/include/asm/book3s/32/pgalloc.h
37
*pmdp = __pmd(__pa(pte) | _PMD_PRESENT);
arch/powerpc/include/asm/book3s/32/pgalloc.h
43
*pmdp = __pmd(__pa(pte_page) | _PMD_PRESENT);
arch/powerpc/include/asm/book3s/32/pgtable.h
240
unsigned long ptephys = __pa(ptep) & PAGE_MASK;
arch/powerpc/include/asm/book3s/64/pgtable.h
982
#define __pgtable_ptr_val(ptr) __pa(ptr)
arch/powerpc/include/asm/guest-state-buffer.h
397
return __pa(kvmppc_gsb_header(gsb));
arch/powerpc/include/asm/io.h
881
return __pa((unsigned long)address);
arch/powerpc/include/asm/io.h
915
return __pa(address) + PCI_DRAM_OFFSET;
arch/powerpc/include/asm/nohash/32/pgalloc.h
23
*pmdp = __pmd(__pa(pte) | _PMD_PRESENT);
arch/powerpc/include/asm/nohash/32/pgalloc.h
32
*pmdp = __pmd(__pa(pte_page) | _PMD_USER | _PMD_PRESENT);
arch/powerpc/include/asm/nohash/32/pgtable.h
172
#define pmd_pfn(pmd) (__pa(pmd_val(pmd)) >> PAGE_SHIFT)
arch/powerpc/include/asm/page.h
222
return __pa(kaddr) >> PAGE_SHIFT;
arch/powerpc/include/asm/rtas-work-area.h
82
return __pa(area->buf);
arch/powerpc/kernel/crash_dump.c
64
create_trampoline(__pa(system_reset_fwnmi) - PHYSICAL_START);
arch/powerpc/kernel/crash_dump.c
65
create_trampoline(__pa(machine_check_fwnmi) - PHYSICAL_START);
arch/powerpc/kernel/dt_cpu_ftrs.c
870
if (!early_init_dt_verify(fdt, __pa(fdt)))
arch/powerpc/kernel/fadump.c
768
phdr->p_paddr = __pa(fw_dump.cpu_notes_buf_vaddr);
arch/powerpc/kernel/paca.c
293
memblock_phys_free(__pa(paca_ptrs) + new_ptrs_size,
arch/powerpc/kernel/paca.c
302
memblock_phys_free(__pa(paca_ptrs[boot_cpuid]->slb_shadow_ptr),
arch/powerpc/kernel/paca.c
85
uv_share_page(PHYS_PFN(__pa(shared_lppaca)),
arch/powerpc/kernel/proc_powerpc.c
40
__pa(pde_data(file_inode(file))) >> PAGE_SHIFT,
arch/powerpc/kernel/prom.c
124
start = __pa(initial_boot_params);
arch/powerpc/kernel/prom.c
690
memblock_reserve(ALIGN_DOWN(__pa(initrd_start), PAGE_SIZE),
arch/powerpc/kernel/prom.c
794
if (!early_init_dt_verify(params, __pa(params)))
arch/powerpc/kernel/prom.c
844
memblock_reserve(PHYSICAL_START, __pa(_end) - PHYSICAL_START);
arch/powerpc/kernel/prom_init.c
1829
os_term_args.args[0] = cpu_to_be32(__pa(str));
arch/powerpc/kernel/prom_init.c
3161
prom_initrd_start = is_kernel_addr(r3) ? __pa(r3) : r3;
arch/powerpc/kernel/rtas.c
1031
err_args.args[0] = cpu_to_be32(__pa(rtas_err_buf));
arch/powerpc/kernel/rtas.c
1668
rtas_call_unlocked(&args, token, 1, 1, NULL, __pa(rtas_os_term_buf));
arch/powerpc/kernel/rtas.c
698
enter_rtas(__pa(args));
arch/powerpc/kernel/rtas_flash.c
440
(u32) __pa(rtas_data_buf), args_buf->buf_size);
arch/powerpc/kernel/rtas_flash.c
572
rtas_block_list = __pa(flist);
arch/powerpc/kernel/rtas_flash.c
586
f->blocks[i].data = (char *)cpu_to_be64(__pa(f->blocks[i].data));
arch/powerpc/kernel/rtas_flash.c
593
f->next = (struct flash_block_list *)cpu_to_be64(__pa(f->next));
arch/powerpc/kernel/rtasd.c
401
__pa(logdata), rtas_error_log_max);
arch/powerpc/kernel/suspend.c
20
unsigned long nosave_begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT;
arch/powerpc/kernel/suspend.c
21
unsigned long nosave_end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT;
arch/powerpc/kexec/core.c
126
kernel_start = __pa(_stext);
arch/powerpc/kexec/core.c
218
kernel_end = cpu_to_be_ulong(__pa(_end));
arch/powerpc/kexec/core_64.c
444
htab_base = cpu_to_be64(__pa(htab_address));
arch/powerpc/kexec/core_64.c
52
if (image->segment[i].mem < __pa(_end))
arch/powerpc/kexec/ranges.c
320
return add_mem_range(mem_ranges, __pa(htab_address), htab_size_bytes);
arch/powerpc/kexec/ranges.c
335
return add_mem_range(mem_ranges, 0, __pa(_end));
arch/powerpc/kvm/book3s_64_mmu_hv.c
122
kvm->arch.sdr1 = __pa(info->virt) | (info->order - 18);
arch/powerpc/kvm/book3s_64_mmu_radix.c
49
(to != NULL) ? __pa(to): 0,
arch/powerpc/kvm/book3s_64_mmu_radix.c
50
(from != NULL) ? __pa(from): 0, n);
arch/powerpc/kvm/book3s_hv.c
4384
trap = plpar_hcall_norets(H_ENTER_NESTED, __pa(&hvregs),
arch/powerpc/kvm/book3s_hv.c
4385
__pa(&vcpu->arch.regs));
arch/powerpc/kvm/book3s_hv.c
5460
__pa(kvm->arch.pgtable) | RADIX_PGD_INDEX_SIZE;
arch/powerpc/kvm/book3s_hv.c
751
kvmhv_nestedv2_set_vpa(vcpu, __pa(vcpu->arch.vpa.pinned_addr));
arch/powerpc/kvm/book3s_hv_nested.c
487
ptcr = __pa(pseries_partition_tb) | (ptb_order - 12);
arch/powerpc/kvm/book3s_hv_nested.c
559
__pa(gp->shadow_pgtable) | RADIX_PGD_INDEX_SIZE;
arch/powerpc/kvm/guest-state-buffer.c
630
__pa(gsb->hdr), gsb->capacity, &i);
arch/powerpc/kvm/guest-state-buffer.c
657
__pa(gsb->hdr), gsb->capacity, &i);
arch/powerpc/mm/book3s32/mmu.c
379
_SDR1 = __pa(Hash) | SDR1_LOW_BITS;
arch/powerpc/mm/book3s32/mmu_context.c
129
mtspr(SPRN_SDR1, rol32(__pa(next->pgd), 4) & 0xffff01ff);
arch/powerpc/mm/book3s64/hash_utils.c
1249
rc = htab_bolt_mapping(start, end, __pa(start),
arch/powerpc/mm/book3s64/hash_utils.c
1418
BUG_ON(htab_bolt_mapping(base, base + size, __pa(base),
arch/powerpc/mm/book3s64/hash_utils.c
1441
__pa(tce_alloc_start), prot,
arch/powerpc/mm/book3s64/hash_utils.c
1554
set_ptcr_when_no_uv(__pa(partition_tb) |
arch/powerpc/mm/book3s64/hash_utils.c
296
ret = hpte_insert_repeating(hash, vpn, __pa(vaddr), mode,
arch/powerpc/mm/book3s64/hash_utils.c
382
lmi = __pa(vaddr) >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/mmu_context.c
174
process_tb[index].prtb0 = cpu_to_be64(rts_field | __pa(mm->pgd) | RADIX_PGD_INDEX_SIZE);
arch/powerpc/mm/book3s64/pgtable.c
333
ptcr = __pa(partition_tb) | (PATB_SIZE_SHIFT - 12);
arch/powerpc/mm/book3s64/radix_pgtable.c
1008
entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL);
arch/powerpc/mm/book3s64/radix_pgtable.c
1056
entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL);
arch/powerpc/mm/book3s64/radix_pgtable.c
472
process_tb->prtb0 = cpu_to_be64(rts_field | __pa(init_mm.pgd) | RADIX_PGD_INDEX_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
497
dw0 = rts_field | __pa(init_mm.pgd) | RADIX_PGD_INDEX_SIZE | PATB_HR;
arch/powerpc/mm/book3s64/radix_pgtable.c
498
dw1 = __pa(process_tb) | (PRTB_SIZE_SHIFT - 12) | PATB_GR;
arch/powerpc/mm/book3s64/radix_pgtable.c
667
set_ptcr_when_no_uv(__pa(partition_tb) |
arch/powerpc/mm/book3s64/radix_pgtable.c
942
return create_physical_mapping(__pa(start), __pa(end),
arch/powerpc/mm/init_64.c
246
if (vmemmap_list_populate(__pa(p), start, node)) {
arch/powerpc/mm/init_64.c
266
rc = vmemmap_create_mapping(start, page_size, __pa(p));
arch/powerpc/mm/kasan/8xx.c
30
pte_t pte = pte_mkhuge(pfn_pte(PHYS_PFN(__pa(block + i * PAGE_SIZE)), PAGE_KERNEL));
arch/powerpc/mm/kasan/8xx.c
68
pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
arch/powerpc/mm/kasan/init_32.c
104
phys_addr_t pa = __pa(kasan_early_shadow_page);
arch/powerpc/mm/kasan/init_32.c
24
phys_addr_t pa = __pa(kasan_early_shadow_page);
arch/powerpc/mm/kasan/init_32.c
75
pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
arch/powerpc/mm/kasan/init_book3e_64.c
79
kasan_map_kernel_page(k_cur, __pa(va), PAGE_KERNEL);
arch/powerpc/mm/kasan/init_book3s_64.c
35
map_kernel_page(k_cur, __pa(va), PAGE_KERNEL);
arch/powerpc/mm/nohash/8xx.c
145
unsigned long etext8 = ALIGN(__pa(_etext), SZ_8M);
arch/powerpc/mm/nohash/8xx.c
146
unsigned long sinittext = __pa(_sinittext);
arch/powerpc/mm/nohash/8xx.c
149
unsigned long einittext8 = ALIGN(__pa(_einittext), SZ_8M);
arch/powerpc/mm/nohash/8xx.c
173
unsigned long etext8 = ALIGN(__pa(_etext), SZ_8M);
arch/powerpc/mm/nohash/8xx.c
174
unsigned long sinittext = __pa(_sinittext);
arch/powerpc/mm/nohash/8xx.c
176
unsigned long einittext8 = ALIGN(__pa(_einittext), SZ_8M);
arch/powerpc/mm/nohash/8xx.c
191
unsigned long sinittext = __pa(_sinittext);
arch/powerpc/mm/nohash/8xx.c
33
return __pa(va);
arch/powerpc/mm/nohash/book3e_pgtable.c
60
__pa(MAX_DMA_ADDRESS), NUMA_NO_NODE);
arch/powerpc/mm/nohash/book3e_pgtable.c
64
__func__, size, size, __pa(MAX_DMA_ADDRESS));
arch/powerpc/mm/nohash/kaslr_booke.c
156
if (regions_overlap(start, end, __pa(_stext), __pa(_end)))
arch/powerpc/mm/nohash/kaslr_booke.c
327
regions.dtb_start = __pa(dt_ptr);
arch/powerpc/mm/nohash/kaslr_booke.c
328
regions.dtb_end = __pa(dt_ptr) + fdt_totalsize(dt_ptr);
arch/powerpc/mm/nohash/mmu_context.c
206
mtspr(SPRN_M_TWB, __pa(pgd));
arch/powerpc/mm/pgtable.c
425
return __pa(pfn_to_kaddr(pfn)) + offset_in_page(va);
arch/powerpc/perf/imc-pmu.c
1223
rc = opal_imc_counters_init(OPAL_IMC_COUNTERS_TRACE, __pa((void *)local_mem),
arch/powerpc/perf/imc-pmu.c
625
__pa((void *)mem_info->vbase),
arch/powerpc/perf/vpa-dtl.c
448
addr = __pa(dtl->buf);
arch/powerpc/platforms/44x/iss4xx.c
113
spin_table[1] = __pa(start_secondary_47x);
arch/powerpc/platforms/44x/ppc476.c
186
spin_table[1] = __pa(start_secondary_47x);
arch/powerpc/platforms/85xx/smp.c
253
__pa(ppc_function_entry(generic_secondary_smp_init)));
arch/powerpc/platforms/85xx/smp.c
262
out_be32(&spin_table->addr_h, __pa(__early_start) >> 32);
arch/powerpc/platforms/85xx/smp.c
264
out_be32(&spin_table->addr_l, __pa(__early_start));
arch/powerpc/platforms/chrp/nvram.c
35
__pa(nvram_buf), 1) != 0) || 1 != done)
arch/powerpc/platforms/chrp/nvram.c
57
__pa(nvram_buf), 1) != 0) || 1 != done)
arch/powerpc/platforms/pasemi/iommu.c
227
regword = IOBMAP_L1E_V | (__pa(iob_l2_base + i*0x2000) >> 12);
arch/powerpc/platforms/pasemi/iommu.c
93
rpn = __pa(uaddr) >> IOBMAP_PAGE_SHIFT;
arch/powerpc/platforms/powermac/smp.c
333
unsigned long start = __pa(__secondary_start_pmac_0) + nr * 8;
arch/powerpc/platforms/powernv/opal-dump.c
265
addr = __pa(list);
arch/powerpc/platforms/powernv/opal-elog.c
172
opal_rc = opal_read_elog(__pa(elog->buffer),
arch/powerpc/platforms/powernv/opal-elog.c
215
rc = opal_read_elog(__pa(elog->buffer),
arch/powerpc/platforms/powernv/opal-flash.c
137
ret = opal_validate_flash(__pa(buf), &size, &result);
arch/powerpc/platforms/powernv/opal-flash.c
293
addr = __pa(list);
arch/powerpc/platforms/powernv/opal-hmi.c
306
while (opal_get_msg(__pa(&msg), sizeof(msg)) == OPAL_SUCCESS) {
arch/powerpc/platforms/powernv/opal-nvram.c
36
rc = opal_read_nvram(__pa(buf), count, off);
arch/powerpc/platforms/powernv/opal-nvram.c
59
rc = opal_write_nvram(__pa(buf), count, off);
arch/powerpc/platforms/powernv/opal-powercap.c
49
ret = opal_get_powercap(pcap_attr->handle, token, (u32 *)__pa(&pcap));
arch/powerpc/platforms/powernv/opal-psr.c
43
(u32 *)__pa(&psr));
arch/powerpc/platforms/powernv/opal-xscom.c
61
rc = opal_xscom_read(chip, reg, (__be64 *)__pa(&v));
arch/powerpc/platforms/powernv/opal.c
1161
sg->next = cpu_to_be64(__pa(next));
arch/powerpc/platforms/powernv/opal.c
357
ret = opal_get_msg(__pa(opal_msg), opal_msg_size);
arch/powerpc/platforms/powernv/opal.c
923
__pa(addr), size);
arch/powerpc/platforms/powernv/pci-ioda-tce.c
105
tce = __pa(tmp2) | TCE_PCI_READ | TCE_PCI_WRITE;
arch/powerpc/platforms/powernv/pci-ioda-tce.c
129
u64 rpn = __pa(uaddr) >> tbl->it_page_shift;
arch/powerpc/platforms/powernv/pci-ioda-tce.c
280
addr[i] = cpu_to_be64(__pa(tmp) |
arch/powerpc/platforms/powernv/pci-ioda.c
1104
__pa(tces),
arch/powerpc/platforms/powernv/pci-ioda.c
1299
__pa(tbl->it_base),
arch/powerpc/platforms/powernv/smp.c
69
__pa(ppc_function_entry(generic_secondary_smp_init));
arch/powerpc/platforms/powernv/vas-fault.c
219
attr.rx_fifo = __pa(vinst->fault_fifo);
arch/powerpc/platforms/ps3/device-init.c
790
lpar = ps3_mm_phys_to_lpar(__pa(&local->buf));
arch/powerpc/platforms/ps3/interrupt.c
532
lpar_addr = ps3_mm_phys_to_lpar(__pa(virt_addr_bmp));
arch/powerpc/platforms/ps3/interrupt.c
758
ps3_mm_phys_to_lpar(__pa(&pd->bmp)));
arch/powerpc/platforms/ps3/interrupt.c
761
pd->thread_id, ps3_mm_phys_to_lpar(__pa(&pd->bmp)));
arch/powerpc/platforms/ps3/mm.c
1076
unsigned long phys_addr = is_kernel_addr(virt_addr) ? __pa(virt_addr)
arch/powerpc/platforms/ps3/mm.c
1125
lpar_addr = addr ? ps3_mm_phys_to_lpar(__pa(addr)) : 0;
arch/powerpc/platforms/ps3/mm.c
780
unsigned long phys_addr = is_kernel_addr(virt_addr) ? __pa(virt_addr)
arch/powerpc/platforms/ps3/mm.c
835
unsigned long phys_addr = is_kernel_addr(virt_addr) ? __pa(virt_addr)
arch/powerpc/platforms/pseries/dtl.c
144
addr = __pa(dtl->buf);
arch/powerpc/platforms/pseries/io_event_irq.c
127
__pa(ioei_rtas_buf),
arch/powerpc/platforms/pseries/iommu.c
142
rpn = __pa(uaddr) >> tceshift;
arch/powerpc/platforms/pseries/iommu.c
218
rpn = __pa(uaddr) >> tceshift;
arch/powerpc/platforms/pseries/iommu.c
290
rpn = __pa(uaddr) >> tceshift;
arch/powerpc/platforms/pseries/iommu.c
310
(u64)__pa(tcep),
arch/powerpc/platforms/pseries/iommu.c
545
(u64)__pa(tcep),
arch/powerpc/platforms/pseries/lpar.c
134
ret = register_dtl(hwcpu, __pa(dtl));
arch/powerpc/platforms/pseries/lpar.c
1734
pseries_lpar_register_process_table(__pa(process_tb),
arch/powerpc/platforms/pseries/lpar.c
1772
addr = __pa((unsigned long)page_address(page));
arch/powerpc/platforms/pseries/lpar.c
695
addr = __pa(&lppaca_of(cpu));
arch/powerpc/platforms/pseries/lpar.c
710
addr = __pa(paca_ptrs[cpu]->slb_shadow_ptr);
arch/powerpc/platforms/pseries/nvram.c
108
if ((rtas_call(nvram_store, 3, 2, &done, i, __pa(nvram_buf),
arch/powerpc/platforms/pseries/nvram.c
63
if ((rtas_call(nvram_fetch, 3, 2, &done, i, __pa(nvram_buf),
arch/powerpc/platforms/pseries/papr-hvpipe.c
594
RTAS_HVPIPE_MSG_EVENTS, 1, __pa(&hvpipe_ras_buf),
arch/powerpc/platforms/pseries/papr-vpd.c
73
__pa(loc_code),
arch/powerpc/platforms/pseries/plpks.c
1354
early_init_dt_verify(fdt, __pa(fdt));
arch/powerpc/platforms/pseries/pseries_energy.c
204
rc = plpar_hcall9(H_BEST_ENERGY, retbuf, flags, 0, __pa(buf_page),
arch/powerpc/platforms/pseries/ras.c
301
RTAS_HOTPLUG_EVENTS, 0, __pa(&ras_log_buf),
arch/powerpc/platforms/pseries/ras.c
339
virq_to_hw(irq), RTAS_EPOW_WARNING, critical, __pa(&ras_log_buf),
arch/powerpc/platforms/pseries/ras.c
370
__pa(&ras_log_buf),
arch/powerpc/platforms/pseries/rtas-fadump.c
618
rtas_fadump_get_config(fadump_conf, (void *)__pa(fdm_active));
arch/powerpc/platforms/pseries/rtas-work-area.c
137
const phys_addr_t pa_start = __pa(rwa_state.arena);
arch/powerpc/platforms/pseries/setup.c
147
system_reset_addr = __pa(system_reset_fwnmi) - PHYSICAL_START;
arch/powerpc/platforms/pseries/setup.c
148
machine_check_addr = __pa(machine_check_fwnmi) - PHYSICAL_START;
arch/powerpc/platforms/pseries/smp.c
91
__pa(ppc_function_entry(generic_secondary_smp_init));
arch/powerpc/platforms/pseries/svm.c
49
uv_unshare_page(PHYS_PFN(__pa(addr)), numpages);
arch/powerpc/platforms/pseries/svm.c
62
uv_share_page(PHYS_PFN(__pa(addr)), numpages);
arch/powerpc/platforms/pseries/svm.c
86
unsigned long pfn = PHYS_PFN(__pa(addr));
arch/powerpc/sysdev/dart_iommu.c
189
rpn = __pa(uaddr) >> DART_PAGE_SHIFT;
arch/powerpc/sysdev/xive/native.c
143
qpage_phys = __pa(qpage);
arch/powerpc/sysdev/xive/native.c
662
opal_xive_donate_page(chip, __pa(p));
arch/powerpc/sysdev/xive/spapr.c
491
qpage_phys = __pa(qpage);
arch/powerpc/sysdev/xive/spapr.c
560
uv_unshare_page(PHYS_PFN(__pa(q->qpage)), 1 << alloc_order);
arch/riscv/include/asm/page.h
185
#define virt_to_pfn(vaddr) (phys_to_pfn(__pa(vaddr)))
arch/riscv/kernel/cpu_ops_sbi.c
78
hsm_data = __pa(bdata);
arch/riscv/kernel/hibernate.c
398
hibernate_restore_image(resume_hdr.saved_satp, (PFN_DOWN(__pa(resume_pg_dir)) | satp_mode),
arch/riscv/kernel/paravirt.c
69
phys_addr_t pa = __pa(st);
arch/riscv/kernel/sbi.c
603
base_addr = __pa(bytes);
arch/riscv/kernel/sbi.c
632
base_addr = __pa(bytes);
arch/riscv/kvm/gstage.c
145
set_pte(ptep, pfn_pte(PFN_DOWN(__pa(next_ptep)),
arch/riscv/mm/init.c
1312
if (start <= __pa(PAGE_OFFSET) &&
arch/riscv/mm/init.c
1313
__pa(PAGE_OFFSET) < end)
arch/riscv/mm/init.c
1314
start = __pa(PAGE_OFFSET);
arch/riscv/mm/init.c
1797
mhp_range.start = __pa(PAGE_OFFSET);
arch/riscv/mm/init.c
1798
mhp_range.end = __pa(PAGE_END - 1);
arch/riscv/mm/init.c
264
max_mapped_addr = __pa(PAGE_OFFSET) + KERN_VIRT_SIZE;
arch/riscv/mm/init.c
446
return __pa((pte_t *)ptdesc_address(ptdesc));
arch/riscv/mm/init.c
527
return __pa((pmd_t *)ptdesc_address(ptdesc));
arch/riscv/mm/init.c
592
return __pa((pud_t *)ptdesc_address(ptdesc));
arch/riscv/mm/init.c
630
return __pa((p4d_t *)ptdesc_address(ptdesc));
arch/riscv/mm/init.c
689
#define get_pgd_next_virt(__pa) (pgtable_l5_enabled ? \
arch/riscv/mm/init.c
690
pt_ops.get_p4d_virt(__pa) : (pgd_next_t *)(pgtable_l4_enabled ? \
arch/riscv/mm/init.c
691
pt_ops.get_pud_virt(__pa) : (pud_t *)pt_ops.get_pmd_virt(__pa)))
arch/riscv/mm/init.c
692
#define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \
arch/riscv/mm/init.c
694
create_p4d_mapping(__nextp, __va, __pa, __sz, __prot) : \
arch/riscv/mm/init.c
696
create_pud_mapping((pud_t *)__nextp, __va, __pa, __sz, __prot) : \
arch/riscv/mm/init.c
697
create_pmd_mapping((pmd_t *)__nextp, __va, __pa, __sz, __prot)))
arch/riscv/mm/init.c
707
#define get_pgd_next_virt(__pa) pt_ops.get_pte_virt(__pa)
arch/riscv/mm/init.c
708
#define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \
arch/riscv/mm/init.c
709
create_pte_mapping(__nextp, __va, __pa, __sz, __prot)
arch/riscv/mm/init.c
711
#define create_p4d_mapping(__pmdp, __va, __pa, __sz, __prot) do {} while(0)
arch/riscv/mm/init.c
712
#define create_pud_mapping(__pmdp, __va, __pa, __sz, __prot) do {} while(0)
arch/riscv/mm/init.c
713
#define create_pmd_mapping(__pmdp, __va, __pa, __sz, __prot) do {} while(0)
arch/riscv/mm/kasan_init.c
120
set_pgd(pgd, pfn_pgd(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
254
phys_addr = __pa((uintptr_t)kasan_early_shadow_pmd);
arch/riscv/mm/kasan_init.c
292
phys_addr = __pa((uintptr_t)kasan_early_shadow_pud);
arch/riscv/mm/kasan_init.c
313
phys_addr = __pa((uintptr_t)kasan_early_shadow_p4d);
arch/riscv/mm/kasan_init.c
336
(__pa((uintptr_t)kasan_early_shadow_pte)),
arch/riscv/mm/kasan_init.c
343
(__pa(((uintptr_t)kasan_early_shadow_pmd))),
arch/riscv/mm/kasan_init.c
351
(__pa(((uintptr_t)kasan_early_shadow_pud))),
arch/riscv/mm/kasan_init.c
36
set_pmd(pmd, pfn_pmd(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
389
set_pud(pud_k, pfn_pud(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
409
set_p4d(p4d_k, pfn_p4d(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
428
set_pgd(pgd_k, pfn_pgd(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
468
pfn_pgd(PFN_DOWN(__pa(tmp_p4d)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
479
pfn_p4d(PFN_DOWN(__pa(tmp_pud)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
489
csr_write(CSR_SATP, PFN_DOWN(__pa(tmp_pg_dir)) | satp_mode);
arch/riscv/mm/kasan_init.c
534
csr_write(CSR_SATP, PFN_DOWN(__pa(swapper_pg_dir)) | satp_mode);
arch/riscv/mm/kasan_init.c
58
set_pud(pud, pfn_pud(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
89
set_p4d(p4d, pfn_p4d(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/s390/boot/vmem.c
100
p4d_t p4d_z = __p4d(__pa(kasan_early_shadow_pud) | _REGION2_ENTRY);
arch/s390/boot/vmem.c
105
pte_z = __pte(__pa(kasan_early_shadow_page) | pgprot_val(PAGE_KERNEL_RO));
arch/s390/boot/vmem.c
98
pmd_t pmd_z = __pmd(__pa(kasan_early_shadow_pte) | _SEGMENT_ENTRY);
arch/s390/boot/vmem.c
99
pud_t pud_z = __pud(__pa(kasan_early_shadow_pmd) | _REGION3_ENTRY);
arch/s390/include/asm/dma-types.h
75
return (__force dma64_t)__pa(ptr);
arch/s390/include/asm/mmu_context.h
61
mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
arch/s390/include/asm/page-states.h
39
unsigned long paddr = __pa(addr) & PAGE_MASK;
arch/s390/include/asm/page.h
226
#define __pa32(x) __pa(x)
arch/s390/include/asm/page.h
270
return phys_to_pfn(__pa(kaddr));
arch/s390/include/asm/pgalloc.h
120
set_pgd(pgd, __pgd(_REGION1_ENTRY | __pa(p4d)));
arch/s390/include/asm/pgalloc.h
125
set_p4d(p4d, __p4d(_REGION2_ENTRY | __pa(pud)));
arch/s390/include/asm/pgalloc.h
130
set_pud(pud, __pud(_REGION3_ENTRY | __pa(pmd)));
arch/s390/include/asm/pgalloc.h
154
set_pmd(pmd, __pmd(_SEGMENT_ENTRY | __pa(pte)));
arch/s390/include/asm/pgtable.h
1106
pto = __pa(ptep) & ~(PTRS_PER_PTE * sizeof(pte_t) - 1);
arch/s390/include/asm/pgtable.h
1117
unsigned long pto = __pa(ptep);
arch/s390/include/asm/pgtable.h
1139
unsigned long pto = __pa(ptep);
arch/s390/include/asm/pgtable.h
1372
return __pa(pmd_deref(pmd)) >> PAGE_SHIFT;
arch/s390/include/asm/pgtable.h
1388
return __pa(pud_deref(pud)) >> PAGE_SHIFT;
arch/s390/include/asm/pgtable.h
1609
sto = __pa(pmdp) - pmd_index(addr) * sizeof(pmd_t);
arch/s390/include/asm/pgtable.h
1635
r3o = __pa(pudp) - pud_index(addr) * sizeof(pud_t);
arch/s390/include/asm/pgtable.h
644
union register_pair r2 = { .even = __pa(table) | dtt, .odd = address, };
arch/s390/kernel/abs_lowcore.c
12
unsigned long phys = __pa(lc);
arch/s390/kernel/cpcmd.c
32
: [rx] "d" (__pa(cpcmd_buf))
arch/s390/kernel/cpcmd.c
42
rx.even = __pa(cpcmd_buf);
arch/s390/kernel/cpcmd.c
43
rx.odd = __pa(response);
arch/s390/kernel/diag/diag.c
285
unsigned long addr = __pa(ptr);
arch/s390/kernel/early.c
80
end_pfn = PFN_UP(__pa(_end));
arch/s390/kernel/ipl.c
1947
abs_lc->ipib = __pa(reipl_block_actual);
arch/s390/kernel/nmi.c
64
*mcesad = __pa(&boot_mcesa);
arch/s390/kernel/nmi.c
83
*mcesad = __pa(origin);
arch/s390/kernel/os_info.c
52
os_info.entry[nr].addr = __pa(ptr);
arch/s390/kernel/os_info.c
91
abs_lc->os_info = __pa(&os_info);
arch/s390/kernel/setup.c
401
lc->restart_psw.addr = __pa(restart_int_handler);
arch/s390/kernel/setup.c
463
set_prefix(__pa(lc));
arch/s390/kernel/setup.c
737
memblock_reserve(__pa(start), __pa(end));
arch/s390/kernel/setup.c
750
memblock_reserve(__pa(sclp_early_sccb), EXT_SCCB_READ_SCP);
arch/s390/kernel/setup.c
751
memblock_reserve(__pa(_stext), _end - _stext);
arch/s390/kernel/smp.c
210
pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, __pa(lc));
arch/s390/kernel/smp.c
553
pa = __pa(&lc->floating_pt_save_area);
arch/s390/kernel/smp.c
640
__pcpu_sigp_relax(addr, SIGP_STORE_STATUS_AT_ADDRESS, __pa(page));
arch/s390/kernel/smp.c
643
__pcpu_sigp_relax(addr, SIGP_STORE_ADDITIONAL_STATUS, __pa(page));
arch/s390/kernel/uv.c
74
if (uv_init(__pa(uv_stor_base), uv_info.uv_base_stor_len)) {
arch/s390/kvm/dat.c
131
asce->rsto = __pa(table) >> PAGE_SHIFT;
arch/s390/kvm/dat.h
606
unsigned long table_origin = __pa(crste_table_start(crstep));
arch/s390/kvm/gmap.c
77
gmap->asce.val = __pa(table);
arch/s390/mm/pageattr.c
135
new = __pmd(__pa(pt_dir) | _SEGMENT_ENTRY);
arch/s390/mm/pageattr.c
222
new = __pud(__pa(pm_dir) | _REGION3_ENTRY);
arch/s390/mm/pgalloc.c
100
mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
arch/s390/mm/pgalloc.c
263
*ste = __pa(table) | _SEGMENT_ENTRY;
arch/s390/mm/pgalloc.c
292
*rtte = __pa(table) | _REGION3_ENTRY;
arch/s390/mm/pgalloc.c
320
*rste = __pa(table) | _REGION2_ENTRY;
arch/s390/mm/pgalloc.c
348
*rfte = __pa(table) | _REGION1_ENTRY;
arch/s390/mm/pgalloc.c
432
asce = __pa(table) | _ASCE_TYPE_SEGMENT | _ASCE_TABLE_LENGTH;
arch/s390/mm/pgalloc.c
438
asce = __pa(table) | _ASCE_TYPE_REGION3 | _ASCE_TABLE_LENGTH;
arch/s390/mm/pgalloc.c
444
asce = __pa(table) | _ASCE_TYPE_REGION2 | _ASCE_TABLE_LENGTH;
arch/s390/mm/pgalloc.c
450
asce = __pa(table) | _ASCE_TYPE_REGION1 | _ASCE_TABLE_LENGTH;
arch/s390/mm/pgalloc.c
91
mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
arch/s390/mm/vmem.c
191
set_pte(pte, __pte(__pa(new_page) | prot));
arch/s390/mm/vmem.c
193
set_pte(pte, __pte(__pa(addr) | prot));
arch/s390/mm/vmem.c
257
set_pmd(pmd, __pmd(__pa(addr) | prot));
arch/s390/mm/vmem.c
272
set_pmd(pmd, __pmd(__pa(new_page) | prot));
arch/s390/mm/vmem.c
347
set_pud(pud, __pud(__pa(addr) | prot));
arch/sh/boot/compressed/misc.c
129
output_addr = __pa((unsigned long)&_text+PAGE_SIZE);
arch/sh/drivers/pci/pci-sh7780.c
299
memphys = __pa(memory_start);
arch/sh/drivers/pci/pcie-sh7786.c
360
memstart = __pa(memory_start);
arch/sh/drivers/pci/pcie-sh7786.c
361
memend = __pa(memory_end);
arch/sh/drivers/pci/pcie-sh7786.c
489
dma_direct_set_offset(&pdev->dev, __pa(memory_start),
arch/sh/drivers/pci/pcie-sh7786.c
490
__pa(memory_start) - memstart, memsize);
arch/sh/include/asm/io.h
268
#define virt_to_phys(address) (__pa(address))
arch/sh/include/asm/page.h
157
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/sh/include/asm/page.h
158
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
arch/sh/include/asm/pgtable_32.h
400
#define pmd_pfn(pmd) (__pa(pmd_val(pmd)) >> PAGE_SHIFT)
arch/sh/kernel/setup.c
165
memblock_reserve(__pa(initrd_start), INITRD_SIZE);
arch/sh/kernel/setup.c
258
if (!dt_virt || !early_init_dt_scan(dt_virt, __pa(dt_virt))) {
arch/sh/kernel/swsusp.c
21
unsigned long begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT;
arch/sh/kernel/swsusp.c
22
unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT;
arch/sh/mm/cache-sh3.c
47
p = __pa(v);
arch/sh/mm/init.c
242
start_pfn = PFN_UP(__pa(_end));
arch/sh/mm/mmap.c
175
if (addr + count > __pa(high_memory))
arch/sh/mm/pmb.c
89
return ppn >= __pa(memory_start) && ppn < __pa(memory_end);
arch/sparc/include/asm/mmu_context_64.h
40
__tsb_context_switch(__pa(mm->pgd),
arch/sparc/include/asm/mmu_context_64.h
49
, __pa(&mm->context.tsb_descr[MM_TSB_BASE]),
arch/sparc/include/asm/page_32.h
125
#define virt_to_phys __pa
arch/sparc/include/asm/page_32.h
129
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/sparc/include/asm/page_64.h
150
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr)>>PAGE_SHIFT)
arch/sparc/include/asm/page_64.h
152
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
arch/sparc/include/asm/page_64.h
154
#define virt_to_phys __pa
arch/sparc/include/asm/pgtable_64.h
829
unsigned long val = __pa((unsigned long) (ptep));
arch/sparc/include/asm/pgtable_64.h
835
(pud_val(*(pudp)) = (__pa((unsigned long) (pmdp))))
arch/sparc/include/asm/pgtable_64.h
889
(p4d_val(*(p4dp)) = (__pa((unsigned long) (pudp))))
arch/sparc/include/asm/pgtsrmmu.h
106
#define __nocache_pa(VADDR) (((unsigned long)VADDR) - SRMMU_NOCACHE_VADDR + __pa((unsigned long)srmmu_nocache_pool))
arch/sparc/kernel/iommu.c
129
iommu->dummy_page_pa = (unsigned long) __pa(iommu->dummy_page);
arch/sparc/kernel/iommu.c
234
first_page = __pa(first_page);
arch/sparc/kernel/iommu_common.h
36
#define SG_ENT_PHYS_ADDRESS(SG) (__pa(sg_virt((SG))))
arch/sparc/kernel/irq_64.c
1028
*pa_ptr = __pa(p);
arch/sparc/kernel/irq_64.c
1046
tb->cpu_mondo_block_pa = __pa(mondo);
arch/sparc/kernel/irq_64.c
1054
tb->cpu_list_pa = __pa(page);
arch/sparc/kernel/irq_64.c
1114
ivector_table_pa = __pa(ivector_table);
arch/sparc/kernel/irq_64.c
282
irq = bucket_get_irq(__pa(bucket));
arch/sparc/kernel/irq_64.c
619
irq = bucket_get_irq(__pa(bucket));
arch/sparc/kernel/irq_64.c
622
bucket_set_irq(__pa(bucket), irq);
arch/sparc/kernel/irq_64.c
684
cookie = ~__pa(&ihd->bucket);
arch/sparc/kernel/irq_64.c
740
bucket_set_irq(__pa(bucket), irq);
arch/sparc/kernel/ldc.c
1081
hv_err = sun4v_ldc_set_map_table(lp->id, __pa(table),
arch/sparc/kernel/ldc.c
2206
pa = __pa(buf);
arch/sparc/kernel/ldc.c
2277
ra = __pa(buf);
arch/sparc/kernel/ldc.c
994
*ra = __pa(q);
arch/sparc/kernel/leon_pci_grpci1.c
431
ahbadr = 0xf0000000 & (u32)__pa(PAGE_ALIGN((unsigned long) &_end));
arch/sparc/kernel/leon_pci_grpci2.c
625
ahbadr = 0xf0000000 & (u32)__pa(PAGE_ALIGN(
arch/sparc/kernel/mdesc.c
1340
status = sun4v_mach_desc(__pa(&hp->mdesc), len, &real_len);
arch/sparc/kernel/mdesc.c
192
start = __pa(hp);
arch/sparc/kernel/mdesc.c
517
status = sun4v_mach_desc(__pa(&hp->mdesc), len, &real_len);
arch/sparc/kernel/pci_fire.c
247
__pa(pbm->msi_queues)),
arch/sparc/kernel/pci_fire.c
65
upa_writeq(__pa(iommu->page_table) | 0x7UL, iommu->iommu_tsbbase);
arch/sparc/kernel/pci_msi.c
435
__pa(pbm->msi_queues));
arch/sparc/kernel/pci_psycho.c
455
__pa(pbm->stc.strbuf_flushflag);
arch/sparc/kernel/pci_schizo.c
1115
__pa(pbm->stc.strbuf_flushflag);
arch/sparc/kernel/pci_schizo.c
1208
upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase);
arch/sparc/kernel/pci_sun4v.c
107
__pa(pglist));
arch/sparc/kernel/pci_sun4v.c
1082
unsigned long err, base = __pa(pages + (i * q_size));
arch/sparc/kernel/pci_sun4v.c
113
npages, prot, __pa(pglist),
arch/sparc/kernel/pci_sun4v.c
124
__pa(pglist),
arch/sparc/kernel/pci_sun4v.c
131
__pa(pglist), ret);
arch/sparc/kernel/pci_sun4v.c
228
first_page = __pa(first_page);
arch/sparc/kernel/pci_sun4v.c
775
iotsb->ra = __pa(table);
arch/sparc/kernel/psycho_common.c
430
upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase);
arch/sparc/kernel/ptrace_64.c
124
unsigned long start = __pa(kaddr);
arch/sparc/kernel/sbus.c
587
__pa(strbuf->strbuf_flushflag);
arch/sparc/kernel/sbus.c
626
upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase);
arch/sparc/kernel/smp_64.c
1013
xcall_deliver(data0, __pa(pg_addr),
arch/sparc/kernel/smp_64.c
332
__pa(hdesc));
arch/sparc/kernel/smp_64.c
895
if (tp->pgd_paddr == __pa(mm->pgd))
arch/sparc/kernel/smp_64.c
972
xcall_deliver(data0, __pa(pg_addr),
arch/sparc/kernel/sysfs.c
116
ra = __pa(&per_cpu(mmu_stats, smp_processor_id()));
arch/sparc/mm/init_32.c
171
start_pfn = (unsigned long)__pa(PAGE_ALIGN((unsigned long) &_end));
arch/sparc/mm/init_64.c
1629
unsigned long pa = __pa(addr);
arch/sparc/mm/init_64.c
1772
unsigned long this_end, paddr = __pa(vstart);
arch/sparc/mm/init_64.c
2475
if (paddr >= __pa(initrd_start) &&
arch/sparc/mm/init_64.c
2476
paddr < __pa(PAGE_ALIGN(initrd_end)))
arch/sparc/mm/init_64.c
2613
pmd_val(*pmd) = pte_base | __pa(block);
arch/sparc/mm/init_64.c
279
tsb_addr = __pa(tsb_addr);
arch/sparc/mm/init_64.c
740
start = __pa(start);
arch/sparc/mm/init_64.c
741
end = __pa(end);
arch/sparc/mm/io-unit.c
257
sbus_writel(iopte_val(MKIOPTE(__pa(page))), iopte);
arch/sparc/mm/iommu.c
109
base = __pa((unsigned long)iommu->page_table) >> 4;
arch/sparc/mm/srmmu.c
295
paddr = __pa((unsigned long)srmmu_nocache_pool);
arch/sparc/mm/tsb.c
130
base = __pa(base);
arch/sparc/mm/tsb.c
144
base = __pa(base);
arch/sparc/mm/tsb.c
163
base = __pa(base);
arch/sparc/mm/tsb.c
178
base = __pa(base);
arch/sparc/mm/tsb.c
216
tsb_paddr = __pa(mm->context.tsb_block[tsb_idx].tsb);
arch/sparc/mm/tsb.c
506
old_tsb_base = __pa(old_tsb_base);
arch/sparc/mm/tsb.c
507
new_tsb_base = __pa(new_tsb_base);
arch/um/include/asm/page.h
91
#define virt_addr_valid(v) pfn_valid(phys_to_pfn(__pa(v)))
arch/um/include/asm/pgalloc.h
16
set_pmd(pmd, __pmd(_PAGE_TABLE + (unsigned long) __pa(pte)))
arch/um/include/asm/pgtable-4level.h
62
set_pud(pud, __pud(_PAGE_TABLE + __pa(pmd)))
arch/um/include/asm/pgtable-4level.h
70
set_p4d(p4d, __p4d(_PAGE_TABLE + __pa(pud)))
arch/um/include/asm/pgtable.h
267
#define __virt_to_page(virt) phys_to_page(__pa(virt))
arch/um/kernel/dtb.c
20
if (!early_init_dt_scan(area, __pa(area))) {
arch/um/kernel/mem.c
75
map_memory(brk_end, __pa(brk_end), uml_reserved - brk_end, 1, 1, 0);
arch/um/kernel/mem.c
78
min_low_pfn = PFN_UP(__pa(uml_reserved));
arch/um/kernel/physmem.c
90
os_seek_file(physmem_fd, __pa(__syscall_stub_start));
arch/um/kernel/physmem.c
93
memblock_add(__pa(start), len);
arch/um/kernel/physmem.c
94
memblock_reserve(__pa(start), reserve);
arch/um/kernel/physmem.c
96
min_low_pfn = PFN_UP(__pa(reserve_end));
arch/x86/boot/compressed/ident_map_64.c
307
snp_set_page_shared(__pa(address & PAGE_MASK));
arch/x86/boot/compressed/ident_map_64.c
322
snp_set_page_private(__pa(address & PAGE_MASK));
arch/x86/boot/compressed/misc.c
518
accept_memory(__pa(output), needed_size);
arch/x86/boot/compressed/sev.c
156
sev_es_wr_ghcb_msr(__pa(ghcb));
arch/x86/boot/compressed/sev.c
94
snp_register_ghcb_early(__pa(&boot_ghcb_page));
arch/x86/boot/startup/sme.c
122
set_pgd(pgd, __pgd(PGD_FLAGS | __pa(p4d)));
arch/x86/boot/startup/sme.c
130
set_p4d(p4d, __p4d(P4D_FLAGS | __pa(pud)));
arch/x86/boot/startup/sme.c
138
set_pud(pud, __pud(PUD_FLAGS | __pa(pmd)));
arch/x86/boot/startup/sme.c
178
set_pmd(pmd, __pmd(PMD_FLAGS | __pa(pte)));
arch/x86/coco/sev/core.c
1086
snp_register_ghcb_early(__pa(ghcb));
arch/x86/coco/sev/core.c
1128
snp_register_ghcb_early(__pa(&boot_ghcb_page));
arch/x86/coco/sev/core.c
1145
sev_es_wr_ghcb_msr(__pa(ghcb));
arch/x86/coco/sev/core.c
1208
per_cpu(svsm_caa_pa, cpu) = __pa(caa);
arch/x86/coco/sev/core.c
174
sev_es_wr_ghcb_msr(__pa(ghcb));
arch/x86/coco/sev/core.c
1913
req->input.req_gpa = __pa(mdesc->request);
arch/x86/coco/sev/core.c
1914
req->input.resp_gpa = __pa(mdesc->response);
arch/x86/coco/sev/core.c
1915
req->input.data_gpa = req->certs_data ? __pa(req->certs_data) : 0;
arch/x86/coco/sev/core.c
281
ghcb_set_sw_scratch(ghcb, (u64)__pa(data));
arch/x86/coco/sev/core.c
344
pfn = __pa(vaddr) >> PAGE_SHIFT;
arch/x86/coco/sev/core.c
400
return early_set_pages_state(vaddr, __pa(vaddr), npages, &d);
arch/x86/coco/sev/core.c
461
ghcb_set_sw_exit_info_2(ghcb, __pa(vmsa));
arch/x86/coco/sev/core.c
463
sev_es_wr_ghcb_msr(__pa(ghcb));
arch/x86/coco/sev/core.c
490
call.rcx = __pa(va);
arch/x86/coco/sev/core.c
495
call.rdx = __pa(caa);
arch/x86/coco/sev/core.c
665
pa = __pa(vmsa);
arch/x86/coco/sev/core.c
959
address = __pa(&data->ghcb_page);
arch/x86/coco/sev/svsm.c
332
call.rcx = __pa(buffer);
arch/x86/coco/sev/svsm.c
48
sev_es_wr_ghcb_msr(__pa(ghcb));
arch/x86/coco/sev/vc-handle.c
484
unsigned long ghcb_pa = __pa(ghcb);
arch/x86/coco/sev/vc-shared.c
358
sw_scratch = __pa(ghcb) + offsetof(struct ghcb, shared_buffer);
arch/x86/coco/sev/vc-shared.c
457
sev_es_wr_ghcb_msr(__pa(ghcb));
arch/x86/coco/tdx/tdx.c
966
phys_addr_t start = __pa(vaddr);
arch/x86/coco/tdx/tdx.c
967
phys_addr_t end = __pa(vaddr + numpages * PAGE_SIZE);
arch/x86/entry/vdso/vma.c
99
__pa(pvti) >> PAGE_SHIFT,
arch/x86/hyperv/ivm.c
363
*(u64 *)&start_vp_input->vp_context = __pa(vmsa) | 1;
arch/x86/include/asm/io.h
131
return __pa(address);
arch/x86/include/asm/mem_encrypt.h
112
#define __sme_pa(x) (__pa(x) | sme_me_mask)
arch/x86/include/asm/page.h
34
#ifndef __pa
arch/x86/include/asm/page.h
56
#define __boot_pa(x) __pa(x)
arch/x86/include/asm/page.h
62
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/x86/include/asm/pgalloc.h
100
set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd)));
arch/x86/include/asm/pgalloc.h
105
paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
106
set_pud_safe(pud, __pud(_PAGE_TABLE | __pa(pmd)));
arch/x86/include/asm/pgalloc.h
113
paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
114
set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud)));
arch/x86/include/asm/pgalloc.h
119
paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
120
set_p4d_safe(p4d, __p4d(_PAGE_TABLE | __pa(pud)));
arch/x86/include/asm/pgalloc.h
136
paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
137
set_pgd(pgd, __pgd(_PAGE_TABLE | __pa(p4d)));
arch/x86/include/asm/pgalloc.h
144
paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
145
set_pgd_safe(pgd, __pgd(_PAGE_TABLE | __pa(p4d)));
arch/x86/include/asm/pgalloc.h
65
paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
66
set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
arch/x86/include/asm/pgalloc.h
72
paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
73
set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
arch/x86/include/asm/pgalloc.h
99
paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
arch/x86/include/asm/uv/uv_hub.h
429
return uv_soc_phys_ram_to_nasid(__pa(v));
arch/x86/include/asm/uv/uv_hub.h
456
return uv_soc_phys_ram_to_gpa(__pa(v));
arch/x86/include/asm/xen/page.h
297
#define virt_to_machine(v) (phys_to_machine(XPADDR(__pa(v))))
arch/x86/include/asm/xen/page.h
300
return PFN_DOWN(__pa(v));
arch/x86/kernel/acpi/madt_wakeup.c
112
mstart = PAGE_ALIGN_DOWN(__pa(asm_acpi_mp_play_dead));
arch/x86/kernel/acpi/madt_wakeup.c
124
acpi_mp_pgd = __pa(pgd);
arch/x86/kernel/amd_gart_64.c
549
enable_gart_translation(dev, __pa(agp_gatt_table));
arch/x86/kernel/amd_gart_64.c
809
gart_unmapped_entry = GPTE_ENCODE(__pa(scratch));
arch/x86/kernel/apic/io_apic.c
2576
ioapic_phys = __pa(ioapic_phys);
arch/x86/kernel/apic/x2apic_savic.c
352
gpa = __pa(ap);
arch/x86/kernel/check.c
155
pr_err("Corrupted low memory at %p (%lx phys) = %08lx\n", addr, __pa(addr), *addr);
arch/x86/kernel/devicetree.c
309
early_init_dt_verify(dt, __pa(dt));
arch/x86/kernel/espfix_64.c
168
pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask));
arch/x86/kernel/espfix_64.c
169
paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT);
arch/x86/kernel/espfix_64.c
180
pmd = __pmd(__pa(pte_p) | (PGTABLE_PROT & ptemask));
arch/x86/kernel/espfix_64.c
181
paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT);
arch/x86/kernel/espfix_64.c
192
pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask));
arch/x86/kernel/machine_kexec_32.c
111
(unsigned long)control_page, __pa(control_page));
arch/x86/kernel/machine_kexec_32.c
117
__pa(control_page), __pa(control_page));
arch/x86/kernel/machine_kexec_32.c
196
page_list[PA_CONTROL_PAGE] = __pa(control_page);
arch/x86/kernel/machine_kexec_32.c
198
page_list[PA_PGD] = __pa(image->arch.pgd);
arch/x86/kernel/machine_kexec_32.c
89
set_pgd(pgd, __pgd(__pa(pmd) | _PAGE_PRESENT));
arch/x86/kernel/machine_kexec_32.c
95
set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
arch/x86/kernel/machine_kexec_64.c
187
set_pgd(pgd, __pgd(__pa(p4d) | _KERNPG_TABLE));
arch/x86/kernel/machine_kexec_64.c
195
set_p4d(p4d, __p4d(__pa(pud) | _KERNPG_TABLE));
arch/x86/kernel/machine_kexec_64.c
203
set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE));
arch/x86/kernel/machine_kexec_64.c
211
set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
arch/x86/kernel/machine_kexec_64.c
367
result = init_pgtable(image, __pa(control_page));
arch/x86/kernel/machine_kexec_64.c
371
kexec_pa_table_page = (unsigned long)__pa(image->arch.pgd);
arch/x86/kernel/machine_kexec_64.c
376
prepare_debug_idt((unsigned long)__pa(control_page),
arch/x86/kvm/kvm_onhyperv.h
31
return __pa(*p_hv_pa_pg);
arch/x86/kvm/mmu/mmu.c
3915
return __pa(sp->spt);
arch/x86/kvm/mmu/mmu.c
3956
mmu->root.hpa = __pa(mmu->pae_root);
arch/x86/kvm/mmu/mmu.c
4126
mmu->pml4_root[0] = __pa(mmu->pae_root) | pm_mask;
arch/x86/kvm/mmu/mmu.c
4133
mmu->pml5_root[0] = __pa(mmu->pml4_root) | pm_mask;
arch/x86/kvm/mmu/mmu.c
4161
mmu->root.hpa = __pa(mmu->pml5_root);
arch/x86/kvm/mmu/mmu.c
4163
mmu->root.hpa = __pa(mmu->pml4_root);
arch/x86/kvm/mmu/mmu.c
4165
mmu->root.hpa = __pa(mmu->pae_root);
arch/x86/kvm/mmu/spte.c
391
spte |= __pa(child_pt) | shadow_present_mask | PT_WRITABLE_MASK |
arch/x86/kvm/mmu/spte.h
273
return to_shadow_page(__pa(sptep));
arch/x86/kvm/mmu/tdp_mmu.c
316
mmu->mirror_root_hpa = __pa(root->spt);
arch/x86/kvm/mmu/tdp_mmu.c
318
mmu->root.hpa = __pa(root->spt);
arch/x86/kvm/svm/avic.c
362
return __sme_set(__pa(svm->vcpu.arch.apic->regs));
arch/x86/kvm/svm/avic.c
370
vmcb->control.avic_logical_id = __sme_set(__pa(kvm_svm->avic_logical_id_table));
arch/x86/kvm/svm/avic.c
371
vmcb->control.avic_physical_id = __sme_set(__pa(kvm_svm->avic_physical_id_table));
arch/x86/kvm/svm/nested.c
326
svm->vmcb->control.msrpm_base_pa = __sme_set(__pa(svm->nested.msrpm));
arch/x86/kvm/svm/sev.c
2455
u64 pfn = __pa(svm->sev_es.vmsa) >> PAGE_SHIFT;
arch/x86/kvm/svm/sev.c
3269
u64 pfn = __pa(svm->sev_es.vmsa) >> PAGE_SHIFT;
arch/x86/kvm/svm/sev.c
4613
svm->vmcb->control.vmsa_pa = __pa(svm->sev_es.vmsa);
arch/x86/kvm/svm/sev.c
5159
if (snp_page_reclaim(vcpu->kvm, PHYS_PFN(__pa(vmsa))))
arch/x86/kvm/svm/sev.c
626
start.dh_cert_address = __sme_set(__pa(dh_blob));
arch/x86/kvm/svm/sev.c
638
start.session_address = __sme_set(__pa(session_blob));
arch/x86/kvm/svm/svm.c
1139
control->msrpm_base_pa = __sme_set(__pa(svm->msrpm));
arch/x86/kvm/svm/svm.c
5505
iopm_base = __sme_set(__pa(iopm_va));
arch/x86/kvm/vmx/nested.c
2300
vmcs_write64(VE_INFORMATION_ADDRESS, __pa(vmx->ve_info));
arch/x86/kvm/vmx/nested.c
2310
vmcs_write64(MSR_BITMAP, __pa(vmx->nested.vmcs02.msr_bitmap));
arch/x86/kvm/vmx/nested.c
2333
vmcs_write64(VM_EXIT_MSR_STORE_ADDR, __pa(vmx->msr_autostore.val));
arch/x86/kvm/vmx/nested.c
2334
vmcs_write64(VM_EXIT_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.host.val));
arch/x86/kvm/vmx/nested.c
2335
vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest.val));
arch/x86/kvm/vmx/nested.c
5850
__pa(vmx->vmcs01.shadow_vmcs));
arch/x86/kvm/vmx/nested.c
7064
vmcs_write64(VMREAD_BITMAP, __pa(vmx_vmread_bitmap));
arch/x86/kvm/vmx/nested.c
7065
vmcs_write64(VMWRITE_BITMAP, __pa(vmx_vmwrite_bitmap));
arch/x86/kvm/vmx/posted_intr.c
311
.pi_desc_addr = __pa(vcpu_to_pi_desc(vcpu)),
arch/x86/kvm/vmx/tdx.c
2506
err = tdh_mng_init(&kvm_tdx->td, __pa(td_params), &rcx);
arch/x86/kvm/vmx/tdx.c
3093
td_vmcs_write64(tdx, POSTED_INTR_DESC_ADDR, __pa(&tdx->vt.pi_desc));
arch/x86/kvm/vmx/vmx.c
3010
u64 phys_addr = __pa(per_cpu(vmxarea, cpu));
arch/x86/kvm/vmx/vmx.c
4944
vmcs_write64(MSR_BITMAP, __pa(vmx->vmcs01.msr_bitmap));
arch/x86/kvm/vmx/vmx.c
4957
__pa(vmx->ve_info));
arch/x86/kvm/vmx/vmx.c
4972
vmcs_write64(POSTED_INTR_DESC_ADDR, __pa((&vmx->vt.pi_desc)));
arch/x86/kvm/vmx/vmx.c
4976
vmcs_write64(PID_POINTER_TABLE, __pa(kvm_vmx->pid_table));
arch/x86/kvm/vmx/vmx.c
5003
vmcs_write64(VM_EXIT_MSR_STORE_ADDR, __pa(vmx->msr_autostore.val));
arch/x86/kvm/vmx/vmx.c
5005
vmcs_write64(VM_EXIT_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.host.val));
arch/x86/kvm/vmx/vmx.c
5007
vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest.val));
arch/x86/kvm/vmx/vmx.c
5052
__pa(vmx->vcpu.arch.apic->regs));
arch/x86/kvm/vmx/vmx.c
6768
ve_info_pa == __pa(ve_info) ? "" : "(corrupted!)");
arch/x86/kvm/vmx/vmx.c
7857
__pa(&vmx->vt.pi_desc) | PID_TABLE_ENTRY_VALID);
arch/x86/kvm/vmx/vmx_ops.h
291
u64 phys_addr = __pa(vmcs);
arch/x86/kvm/vmx/vmx_ops.h
298
u64 phys_addr = __pa(vmcs);
arch/x86/mm/ident_map.c
139
set_pud(pud, __pud(__pa(pmd) | info->kernpg_flag));
arch/x86/mm/ident_map.c
172
set_p4d(p4d, __p4d(__pa(pud) | info->kernpg_flag | _PAGE_NOPTISHADOW));
arch/x86/mm/ident_map.c
213
set_pgd(pgd, __pgd(__pa(p4d) | info->kernpg_flag | _PAGE_NOPTISHADOW));
arch/x86/mm/ident_map.c
220
set_pgd(pgd, __pgd(__pa(pud) | info->kernpg_flag | _PAGE_NOPTISHADOW));
arch/x86/mm/init.c
148
ret = __pa(extend_brk(PAGE_SIZE * num, PAGE_SIZE));
arch/x86/mm/init.c
192
base = __pa(extend_brk(tables, PAGE_SIZE));
arch/x86/mm/init_32.c
100
set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE));
arch/x86/mm/init_32.c
181
set_pmd(pmd, __pmd(__pa(newpte)|_PAGE_TABLE));
arch/x86/mm/init_32.c
451
pfn, pmd, __pa(pmd));
arch/x86/mm/init_32.c
460
pfn, pmd, __pa(pmd), pte, __pa(pte));
arch/x86/mm/init_32.c
76
set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT));
arch/x86/mm/init_64.c
1523
entry = pfn_pte(__pa(p) >> PAGE_SHIFT,
arch/x86/mm/init_64.c
400
set_pgd(pgd, __pgd(__pa(p4d) | _KERNPG_TABLE |
arch/x86/mm/init_64.c
406
set_p4d(p4d, __p4d(__pa(pud) | _KERNPG_TABLE |
arch/x86/mm/init_64.c
412
set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE |
arch/x86/mm/init_64.c
708
paddr = __pa(vaddr);
arch/x86/mm/init_64.c
711
paddr_next = __pa(vaddr_next);
arch/x86/mm/init_64.c
723
paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end),
arch/x86/mm/init_64.c
729
paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end),
arch/x86/mm/init_64.c
762
paddr_last = phys_p4d_init(p4d, __pa(vaddr),
arch/x86/mm/init_64.c
763
__pa(vaddr_end),
arch/x86/mm/init_64.c
770
paddr_last = phys_p4d_init(p4d, __pa(vaddr), __pa(vaddr_end),
arch/x86/mm/kasan_init_64.c
28
__pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, nid);
arch/x86/mm/kasan_init_64.c
32
(void *)_RET_IP_, nid, __pa(MAX_DMA_ADDRESS));
arch/x86/mm/kasan_init_64.c
367
__pgd(__pa(tmp_p4d_table) | _KERNPG_TABLE));
arch/x86/mm/kasan_init_64.c
426
early_pfn_to_nid(__pa(_stext)));
arch/x86/mm/kasan_init_64.c
447
pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot));
arch/x86/mm/kasan_init_64.c
49
if (p && pmd_set_huge(pmd, __pa(p), PAGE_KERNEL))
arch/x86/mm/kasan_init_64.c
67
entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL);
arch/x86/mm/kasan_init_64.c
85
if (p && pud_set_huge(pud, __pa(p), PAGE_KERNEL))
arch/x86/mm/kaslr.c
203
__p4d(_KERNPG_TABLE | __pa(pud_page_tramp)));
arch/x86/mm/kaslr.c
206
__pgd(_KERNPG_TABLE | __pa(p4d_page_tramp));
arch/x86/mm/kaslr.c
209
__pgd(_KERNPG_TABLE | __pa(pud_page_tramp));
arch/x86/mm/mem_encrypt_amd.c
449
kernel_physical_mapping_change(__pa(vaddr & pmask),
arch/x86/mm/mem_encrypt_amd.c
450
__pa((vaddr_end & pmask) + psize),
arch/x86/mm/mmap.c
208
return addr + count - 1 <= __pa(high_memory - 1);
arch/x86/mm/pat/memtype.c
811
if (base > __pa(high_memory-1))
arch/x86/mm/pat/memtype.c
821
id_sz = (__pa(high_memory-1) <= base + size) ?
arch/x86/mm/pat/memtype.c
822
__pa(high_memory) - base : size;
arch/x86/mm/pat/set_memory.c
1187
unsigned long pfn = PFN_DOWN(__pa(address));
arch/x86/mm/pat/set_memory.c
1546
set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
arch/x86/mm/pat/set_memory.c
1556
set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE));
arch/x86/mm/pat/set_memory.c
1750
set_pgd(pgd_entry, __pgd(__pa(p4d) | _KERNPG_TABLE));
arch/x86/mm/pat/set_memory.c
1762
set_p4d(p4d, __p4d(__pa(pud) | _KERNPG_TABLE));
arch/x86/mm/pat/set_memory.c
1814
cpa->pfn = __pa(vaddr) >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
2168
ret = memtype_reserve(__pa(addr), __pa(addr) + numpages * PAGE_SIZE,
arch/x86/mm/pat/set_memory.c
2180
memtype_free(__pa(addr), __pa(addr) + numpages * PAGE_SIZE);
arch/x86/mm/pat/set_memory.c
2206
ret = memtype_reserve(__pa(addr), __pa(addr) + numpages * PAGE_SIZE,
arch/x86/mm/pat/set_memory.c
2213
memtype_free(__pa(addr), __pa(addr) + numpages * PAGE_SIZE);
arch/x86/mm/pat/set_memory.c
2240
memtype_free(__pa(addr), __pa(addr) + numpages * PAGE_SIZE);
arch/x86/mm/pgtable.c
139
paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
arch/x86/mm/pgtable.c
143
set_pud(pudp, __pud(__pa(pmd) | _PAGE_PRESENT));
arch/x86/mm/pgtable.c
30
paravirt_release_pmd(__pa(pmd) >> PAGE_SHIFT);
arch/x86/mm/pgtable.c
44
paravirt_release_pud(__pa(pud) >> PAGE_SHIFT);
arch/x86/mm/pgtable.c
51
paravirt_release_p4d(__pa(p4d) >> PAGE_SHIFT);
arch/x86/mm/pti.c
193
set_pgd(pgd, __pgd(_KERNPG_TABLE | __pa(new_p4d_page)));
arch/x86/mm/pti.c
222
set_p4d(p4d, __p4d(_KERNPG_TABLE | __pa(new_pud_page)));
arch/x86/mm/pti.c
236
set_pud(pud, __pud(_KERNPG_TABLE | __pa(new_pmd_page)));
arch/x86/mm/pti.c
277
set_pmd(pmd, __pmd(_KERNPG_TABLE | __pa(new_pte_page)));
arch/x86/mm/tlb.c
1088
WARN_ON((cr3 & CR3_ADDR_MASK) != __pa(mm->pgd));
arch/x86/mm/tlb.c
1787
VM_WARN_ON_ONCE(__pa(current_mm->pgd) != read_cr3_pa());
arch/x86/pci/fixup.c
844
phys_addr_t top_of_dram = __pa(high_memory - 1) + 1;
arch/x86/pci/irq.c
144
__pa(ir));
arch/x86/pci/irq.c
92
__pa(rt));
arch/x86/platform/efi/efi.c
803
pa = __pa(new_memmap);
arch/x86/platform/efi/efi_32.c
119
__pa(&efi.runtime));
arch/x86/platform/efi/efi_64.c
242
text = __pa(_text);
arch/x86/platform/efi/efi_64.c
250
rodata = __pa(__start_rodata);
arch/x86/platform/efi/efi_64.c
259
tramp = __pa(__efi64_thunk_ret_tramp);
arch/x86/power/hibernate_32.c
141
__pgd(__pa(empty_zero_page) | _PAGE_PRESENT));
arch/x86/power/hibernate_32.c
189
temp_pgt = __pa(resume_pg_dir);
arch/x86/power/hibernate_32.c
41
set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT));
arch/x86/power/hibernate_32.c
66
set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE));
arch/x86/power/hibernate_64.c
123
temp_pgt = __pa(pgd);
arch/x86/power/hibernate_64.c
71
__pud(__pa(pmd) | pgprot_val(pgtable_prot)));
arch/x86/power/hibernate_64.c
73
p4d_t new_p4d = __p4d(__pa(pud) | pgprot_val(pgtable_prot));
arch/x86/power/hibernate_64.c
74
pgd_t new_pgd = __pgd(__pa(p4d) | pgprot_val(pgtable_prot));
arch/x86/power/hibernate_64.c
80
pgd_t new_pgd = __pgd(__pa(pud) | pgprot_val(pgtable_prot));
arch/x86/realmode/init.c
119
phys_base = __pa(base);
arch/x86/realmode/init.c
195
__pa(base);
arch/x86/virt/vmx/tdx/tdx.c
958
tdmr_pa_array[i] = __pa(tdmr_entry(tdmr_list, i));
arch/x86/virt/vmx/tdx/tdx.c
960
args.rcx = __pa(tdmr_pa_array);
arch/x86/xen/efi.c
143
boot_params->efi_info.efi_systab = (__u32)__pa(efi_systab_xen);
arch/x86/xen/efi.c
144
boot_params->efi_info.efi_systab_hi = (__u32)(__pa(efi_systab_xen) >> 32);
arch/x86/xen/enlighten_pv.c
1503
initrd_start = __pa(xen_start_info->mod_start);
arch/x86/xen/enlighten_pv.c
1510
boot_params.hdr.cmd_line_ptr = __pa(xen_start_info->cmd_line);
arch/x86/xen/mmu_hvm.c
42
a.gpa = __pa(mm->pgd);
arch/x86/xen/mmu_pv.c
1012
if (per_cpu(xen_current_cr3, cpu) == __pa(mm->pgd))
arch/x86/xen/mmu_pv.c
1103
unsigned long pa = __pa(pgtbl) & PHYSICAL_PAGE_MASK;
arch/x86/xen/mmu_pv.c
1249
xen_start_info->pt_base = (unsigned long)__va(__pa(xen_start_info->pt_base));
arch/x86/xen/mmu_pv.c
1416
__xen_write_cr3(false, __pa(user_pgd));
arch/x86/xen/mmu_pv.c
1474
__pgd(__pa(level3_user_vsyscall) | _PAGE_TABLE);
arch/x86/xen/mmu_pv.c
1703
unsigned long pfn = __pa(addr) >> PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
1739
if (*pt_base == PFN_DOWN(__pa(addr))) {
arch/x86/xen/mmu_pv.c
1744
if (*pt_end == PFN_DOWN(__pa(addr))) {
arch/x86/xen/mmu_pv.c
1775
max_pfn_mapped = PFN_DOWN(__pa(xen_start_info->mfn_list));
arch/x86/xen/mmu_pv.c
1777
pt_base = PFN_DOWN(__pa(xen_start_info->pt_base));
arch/x86/xen/mmu_pv.c
1855
pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
arch/x86/xen/mmu_pv.c
1869
__xen_write_cr3(true, __pa(init_top_pgt));
arch/x86/xen/mmu_pv.c
1887
xen_start_info = (struct start_info *)__va(__pa(xen_start_info));
arch/x86/xen/mmu_pv.c
2072
memblock_reserve(__pa(xen_start_info), PAGE_SIZE);
arch/x86/xen/mmu_pv.c
2108
pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
arch/x86/xen/mmu_pv.c
2118
pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
arch/x86/xen/mmu_pv.c
2565
return __pa(vmcoreinfo_note);
arch/x86/xen/mmu_pv.c
803
xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(pgd)));
arch/x86/xen/mmu_pv.c
808
PFN_DOWN(__pa(user_pgd)));
arch/x86/xen/mmu_pv.c
918
xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
arch/x86/xen/mmu_pv.c
922
PFN_DOWN(__pa(user_pgd)));
arch/x86/xen/mmu_pv.c
977
if (this_cpu_read(xen_current_cr3) == __pa(mm->pgd))
arch/x86/xen/mmu_pv.c
996
if (per_cpu(xen_current_cr3, cpu) != __pa(mm->pgd))
arch/x86/xen/p2m.c
336
paravirt_alloc_pte(&init_mm, __pa(p2m_missing_pte) >> PAGE_SHIFT);
arch/x86/xen/p2m.c
338
paravirt_alloc_pte(&init_mm, __pa(p2m_identity_pte) >> PAGE_SHIFT);
arch/x86/xen/p2m.c
341
pfn_pte(PFN_DOWN(__pa(p2m_missing)), PAGE_KERNEL_RO));
arch/x86/xen/p2m.c
343
pfn_pte(PFN_DOWN(__pa(p2m_identity)), PAGE_KERNEL_RO));
arch/x86/xen/p2m.c
376
pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL));
arch/x86/xen/p2m.c
386
pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO));
arch/x86/xen/p2m.c
396
set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE));
arch/x86/xen/p2m.c
445
if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity)))
arch/x86/xen/p2m.c
483
paravirt_alloc_pte(&init_mm, __pa(pte_newpg[i]) >> PAGE_SHIFT);
arch/x86/xen/p2m.c
495
__pmd(__pa(pte_newpg[i]) | _KERNPG_TABLE));
arch/x86/xen/p2m.c
504
paravirt_release_pte(__pa(pte_newpg[i]) >> PAGE_SHIFT);
arch/x86/xen/p2m.c
575
if (p2m_pfn == PFN_DOWN(__pa(p2m_identity)) ||
arch/x86/xen/p2m.c
576
p2m_pfn == PFN_DOWN(__pa(p2m_missing))) {
arch/x86/xen/p2m.c
584
if (p2m_pfn == PFN_DOWN(__pa(p2m_missing)))
arch/x86/xen/p2m.c
595
pfn_pte(PFN_DOWN(__pa(p2m)), PAGE_KERNEL));
arch/x86/xen/p2m.c
658
if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_missing)))
arch/x86/xen/p2m.c
661
if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity)))
arch/x86/xen/setup.c
770
start = __pa(xen_start_info->mfn_list);
arch/x86/xen/setup.c
867
xen_chk_is_e820_usable(__pa(xen_start_info), sizeof(*xen_start_info),
arch/x86/xen/smp_pv.c
279
per_cpu(xen_cr3, cpu) = __pa(swapper_pg_dir);
arch/xtensa/include/asm/page.h
170
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/xtensa/include/asm/page.h
172
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
arch/xtensa/include/asm/pgtable.h
228
#define pmd_pfn(pmd) (__pa(pmd_val(pmd)) >> PAGE_SHIFT)
arch/xtensa/kernel/hibernate.c
10
unsigned long nosave_end_pfn = PFN_UP(__pa(&__nosave_end));
arch/xtensa/kernel/hibernate.c
9
unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin));
arch/xtensa/kernel/setup.c
219
early_init_dt_scan(params, __pa(params));
arch/xtensa/kernel/setup.c
294
!mem_reserve(__pa(initrd_start), __pa(initrd_end)))
arch/xtensa/kernel/setup.c
300
mem_reserve(__pa(_stext), __pa(_end));
arch/xtensa/kernel/setup.c
303
mem_reserve(__pa(_xip_text_start), __pa(_xip_text_end));
arch/xtensa/kernel/setup.c
305
mem_reserve(__pa(_xip_start), __pa(_xip_end));
arch/xtensa/kernel/setup.c
310
mem_reserve(__pa(_WindowVectors_text_start),
arch/xtensa/kernel/setup.c
311
__pa(_WindowVectors_text_end));
arch/xtensa/kernel/setup.c
314
mem_reserve(__pa(_DebugInterruptVector_text_start),
arch/xtensa/kernel/setup.c
315
__pa(_DebugInterruptVector_text_end));
arch/xtensa/kernel/setup.c
317
mem_reserve(__pa(_KernelExceptionVector_text_start),
arch/xtensa/kernel/setup.c
318
__pa(_KernelExceptionVector_text_end));
arch/xtensa/kernel/setup.c
320
mem_reserve(__pa(_UserExceptionVector_text_start),
arch/xtensa/kernel/setup.c
321
__pa(_UserExceptionVector_text_end));
arch/xtensa/kernel/setup.c
323
mem_reserve(__pa(_DoubleExceptionVector_text_start),
arch/xtensa/kernel/setup.c
324
__pa(_DoubleExceptionVector_text_end));
arch/xtensa/kernel/setup.c
326
mem_reserve(__pa(_exception_text_start),
arch/xtensa/kernel/setup.c
327
__pa(_exception_text_end));
arch/xtensa/kernel/setup.c
329
mem_reserve(__pa(_Level2InterruptVector_text_start),
arch/xtensa/kernel/setup.c
330
__pa(_Level2InterruptVector_text_end));
arch/xtensa/kernel/setup.c
333
mem_reserve(__pa(_Level3InterruptVector_text_start),
arch/xtensa/kernel/setup.c
334
__pa(_Level3InterruptVector_text_end));
arch/xtensa/kernel/setup.c
337
mem_reserve(__pa(_Level4InterruptVector_text_start),
arch/xtensa/kernel/setup.c
338
__pa(_Level4InterruptVector_text_end));
arch/xtensa/kernel/setup.c
341
mem_reserve(__pa(_Level5InterruptVector_text_start),
arch/xtensa/kernel/setup.c
342
__pa(_Level5InterruptVector_text_end));
arch/xtensa/kernel/setup.c
345
mem_reserve(__pa(_Level6InterruptVector_text_start),
arch/xtensa/kernel/setup.c
346
__pa(_Level6InterruptVector_text_end));
arch/xtensa/kernel/setup.c
352
mem_reserve(__pa(_SecondaryResetVector_text_start),
arch/xtensa/kernel/setup.c
353
__pa(_SecondaryResetVector_text_end));
arch/xtensa/kernel/setup.c
408
unsigned long paddr = __pa(vaddr);
drivers/block/ps3disk.c
358
ps3_mm_phys_to_lpar(__pa(&ata_cmnd)),
drivers/block/ps3vram.c
681
xdr_lpar = ps3_mm_phys_to_lpar(__pa(priv->xdr_buf));
drivers/block/ps3vram.c
794
ps3_mm_phys_to_lpar(__pa(priv->xdr_buf)),
drivers/char/hw_random/n2-drv.c
150
unsigned long ra = __pa(&np->scratch_control[0]);
drivers/char/hw_random/n2-drv.c
423
unsigned long ra = __pa(&np->test_data);
drivers/char/hw_random/n2-drv.c
458
unsigned long ra = __pa(&np->test_data);
drivers/char/hw_random/n2-drv.c
468
unsigned long post_ctl_ra = __pa(post_control);
drivers/char/hw_random/n2-drv.c
469
unsigned long pre_ctl_ra = __pa(pre_control);
drivers/char/hw_random/n2-drv.c
470
unsigned long buffer_ra = __pa(buffer);
drivers/char/hw_random/n2-drv.c
622
unsigned long ctl_ra = __pa(&up->control[0]);
drivers/char/mem.c
265
return addr >= __pa(high_memory);
drivers/char/mem.c
50
return addr + count <= __pa(high_memory);
drivers/char/powernv-op-panel.c
179
oppanel_lines[i].line = cpu_to_be64(__pa(&oppanel_data[i *
drivers/cpuidle/cpuidle-calxeda.c
31
return psci_ops.cpu_suspend(CALXEDA_IDLE_PARAM, __pa(cpu_resume));
drivers/cpuidle/cpuidle-pseries.c
176
NULL, CEDE_LATENCY_TOKEN, __pa(&xcede_latency_parameter),
drivers/crypto/ccp/sev-dev-tio.c
46
#define sla_to_pfn(sla) (__pa(sla_to_va(sla)) >> PAGE_SHIFT)
drivers/crypto/ccp/sev-dev.c
1039
data.tmr_address = __pa(sev_es_tmr);
drivers/crypto/ccp/sev-dev.c
1062
data.tmr_address = __pa(sev_es_tmr);
drivers/crypto/ccp/sev-dev.c
1560
if (rmp_mark_pages_firmware(__pa(sev_init_ex_buffer), npages, false)) {
drivers/crypto/ccp/sev-dev.c
2396
if (rmp_mark_pages_firmware(__pa(data), 1, true)) {
drivers/crypto/ccp/sev-dev.c
2411
if (snp_reclaim_pages(__pa(data), 1, true))
drivers/crypto/ccp/sev-dev.c
462
paddr = __pa((unsigned long)page_address(page));
drivers/crypto/ccp/sev-dev.c
487
paddr = __pa((unsigned long)page_address(page));
drivers/crypto/ccp/sev-dev.c
826
if (rmp_mark_pages_firmware(__pa(cmd_buf), 1, true))
drivers/crypto/ccp/sev-dev.c
843
if (snp_reclaim_pages(__pa(cmd_buf), 1, true))
drivers/crypto/nx/nx-842.h
108
return __pa(addr);
drivers/crypto/nx/nx.c
289
nx_ctx->op.csbcpb = __pa(nx_ctx->csbcpb);
drivers/crypto/nx/nx.c
290
nx_ctx->op.in = __pa(nx_ctx->in_sg);
drivers/crypto/nx/nx.c
291
nx_ctx->op.out = __pa(nx_ctx->out_sg);
drivers/crypto/nx/nx.c
297
nx_ctx->op_aead.csbcpb = __pa(nx_ctx->csbcpb_aead);
drivers/crypto/nx/nx.c
298
nx_ctx->op_aead.in = __pa(nx_ctx->in_sg);
drivers/crypto/nx/nx.c
299
nx_ctx->op_aead.out = __pa(nx_ctx->out_sg);
drivers/crypto/nx/nx.c
94
sg_addr = __pa(sg_addr);
drivers/dma/ppc4xx/adma.c
4478
iowrite32(((u32)__pa(ppc440spe_dma_fifo_buf)), &i2o_reg->ifbal);
drivers/firmware/efi/efi.c
1143
rc = efi_mem_reserve_iomem(__pa(rsv), SZ_4K);
drivers/firmware/efi/efi.c
1162
efi_memreserve_root->next = __pa(rsv);
drivers/gpu/drm/amd/amdkfd/kfd_events.c
1080
pfn = __pa(page->kernel_address);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
2257
PFN_DOWN(__pa(qpd->cwsr_kaddr)),
drivers/i2c/busses/i2c-opal.c
101
req.buffer_ra = cpu_to_be64(__pa(msgs[1].buf));
drivers/i2c/busses/i2c-opal.c
126
req.buffer_ra = cpu_to_be64(__pa(&data->byte));
drivers/i2c/busses/i2c-opal.c
134
req.buffer_ra = cpu_to_be64(__pa(&data->byte));
drivers/i2c/busses/i2c-opal.c
146
req.buffer_ra = cpu_to_be64(__pa(local));
drivers/i2c/busses/i2c-opal.c
154
req.buffer_ra = cpu_to_be64(__pa(&data->block[1]));
drivers/i2c/busses/i2c-opal.c
90
req.buffer_ra = cpu_to_be64(__pa(msgs[0].buf));
drivers/infiniband/hw/hfi1/file_ops.c
578
PFN_DOWN(__pa(memvirt)),
drivers/iommu/intel/irq_remapping.c
1130
return __pa(per_cpu_ptr(&posted_msi_pi_desc, cpu));
drivers/iommu/intel/svm.c
178
FLPT_DEFAULT_DID, __pa(mm->pgd),
drivers/iommu/io-pgtable-arm.c
398
new = paddr_to_iopte(__pa(table), data) | ARM_LPAE_PTE_TYPE_TABLE;
drivers/iommu/io-pgtable-dart.c
147
new = paddr_to_iopte(__pa(table), data) | APPLE_DART_PTE_VALID;
drivers/iommu/iommufd/selftest.c
845
__pa(mock_viommu->page),
drivers/iommu/riscv/iommu.c
170
queue->phys = __pa(queue->base);
drivers/iommu/riscv/iommu.c
628
new = phys_to_ppn(__pa(ptr)) | RISCV_IOMMU_DDTE_V;
drivers/iommu/riscv/iommu.c
705
iommu->ddt_phys = __pa(iommu->ddt_root);
drivers/mailbox/mailbox-mchp-ipc-sbi.c
285
chan_info->buf_base_tx_addr = __pa(chan_info->buf_base_tx);
drivers/mailbox/mailbox-mchp-ipc-sbi.c
293
chan_info->buf_base_rx_addr = __pa(chan_info->buf_base_rx);
drivers/mailbox/mailbox-mchp-ipc-sbi.c
311
chan_info->msg_buf_tx_addr = __pa(chan_info->msg_buf_tx);
drivers/mailbox/mailbox-mchp-ipc-sbi.c
319
chan_info->msg_buf_rx_addr = __pa(chan_info->msg_buf_rx);
drivers/mailbox/mailbox-mchp-ipc-sbi.c
399
ipc->cluster_cfg[cpuid].buf_base_addr = __pa(ipc->cluster_cfg[cpuid].buf_base);
drivers/mailbox/mailbox-mchp-ipc-sbi.c
431
ipc->buf_base_addr = __pa(ipc->buf_base);
drivers/misc/ocxl/config.c
855
rc = pnv_ocxl_set_tl_conf(dev, recv_cap, __pa(recv_rate),
drivers/mtd/devices/powernv_flash.c
69
rc = opal_flash_read(info->id, offset, __pa(buf), len, token);
drivers/mtd/devices/powernv_flash.c
72
rc = opal_flash_write(info->id, offset, __pa(buf), len, token);
drivers/mtd/mtdchar.c
1392
if (file->f_flags & O_DSYNC || map->phys >= __pa(high_memory))
drivers/net/ethernet/brocade/bna/bfa_ioc.h
59
#define bfa_alen_set(__alen, __len, __pa) \
drivers/net/ethernet/brocade/bna/bfa_ioc.h
60
__bfa_alen_set(__alen, __len, (u64)__pa)
drivers/net/ethernet/ibm/ehea/ehea_phyp.c
130
__pa(cb_addr), /* R8 */
drivers/net/ethernet/ibm/ehea/ehea_phyp.c
404
__pa(cb_addr), /* R8 */
drivers/net/ethernet/ibm/ehea/ehea_phyp.c
517
cb_logaddr = __pa(cb_addr);
drivers/net/ethernet/ibm/ehea/ehea_phyp.c
534
u64 cb_logaddr = __pa(cb_addr);
drivers/net/ethernet/ibm/ehea/ehea_phyp.c
556
u64 cb_logaddr = __pa(cb_addr);
drivers/net/ethernet/ibm/ehea/ehea_phyp.c
610
__pa(rblock), /* R6 */
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
146
rpage = __pa(vpage);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
270
rpage = __pa(vpage);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
375
rpage = __pa(vpage);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
765
index = __pa(caddr) >> SECTION_SIZE_BITS;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
797
u64 pt_abs = __pa(pt);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
805
pt[m] = __pa(pg);
drivers/net/ethernet/sun/niu.c
10008
*dma_addr = __pa(page);
drivers/net/ethernet/sun/niu.c
10038
return __pa(cpu_addr);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1730
ps3_mm_phys_to_lpar(__pa(&card->irq_status)),
drivers/net/ethernet/toshiba/ps3_gelic_wireless.c
159
ps3_mm_phys_to_lpar(__pa(cmd->buffer)) :
drivers/net/ethernet/toshiba/ps3_gelic_wireless.c
181
arg1 = ps3_mm_phys_to_lpar(__pa(cmd->buffer));
drivers/net/fjes/fjes_hw.c
270
param.req_start = __pa(hw->hw_info.req_buf);
drivers/net/fjes/fjes_hw.c
272
param.res_start = __pa(hw->hw_info.res_buf);
drivers/net/fjes/fjes_hw.c
274
param.share_start = __pa(hw->hw_info.share->ep_status);
drivers/net/fjes/fjes_main.c
471
param.req_start = __pa(hw->hw_info.req_buf);
drivers/net/fjes/fjes_main.c
473
param.res_start = __pa(hw->hw_info.res_buf);
drivers/net/fjes/fjes_main.c
474
param.share_start = __pa(hw->hw_info.share->ep_status);
drivers/of/fdt.c
1143
#define MIN_MEMBLOCK_ADDR __pa(PAGE_OFFSET)
drivers/of/fdt.c
531
memblock_reserve(__pa(initial_boot_params),
drivers/pci/controller/pci-ixp4xx.c
579
addr = __pa(PAGE_OFFSET);
drivers/pci/controller/pci-versatile.c
131
writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_0);
drivers/pci/controller/pci-versatile.c
132
writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_1);
drivers/pci/controller/pci-versatile.c
133
writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_2);
drivers/pci/controller/pci-versatile.c
94
writel(__pa(PAGE_OFFSET) >> 28, PCI_SMAP(mem));
drivers/perf/riscv_pmu_sbi.c
327
base_addr = __pa(event_info_shmem);
drivers/platform/goldfish/goldfish_pipe.c
739
(u64)(unsigned long)__pa(pipe->command_buffer);
drivers/platform/goldfish/goldfish_pipe.c
799
const unsigned long paddr = __pa(addr);
drivers/ps3/ps3-lpm.c
1122
ps3_mm_phys_to_lpar(__pa(lpm_priv->tb_cache)),
drivers/ps3/ps3-vuart.c
358
ps3_mm_phys_to_lpar(__pa(buf)), bytes, bytes_written);
drivers/ps3/ps3-vuart.c
390
ps3_mm_phys_to_lpar(__pa(buf)), bytes, bytes_read);
drivers/ps3/ps3stor_lib.c
190
dev->bounce_lpar = ps3_mm_phys_to_lpar(__pa(dev->bounce_buf));
drivers/s390/char/sclp.c
1113
if (finished_sccb != __pa(sclp_init_sccb))
drivers/s390/char/sclp.c
174
sclp_trace(prio, id, __pa(sccb), summary.b, err);
drivers/s390/char/sclp.c
507
sclp_trace(2, "RQAD", __pa(req->sccb), _RET_IP_, false);
drivers/s390/char/sclp.c
622
if (sccb == __pa(req->sccb))
drivers/s390/char/sclp.h
329
: [cmd] "d" (command), [sccb] "a" (__pa(sccb))
drivers/s390/char/sclp_sd.c
201
sclp_sd_listener_init(&listener, __pa(sccb));
drivers/s390/char/sclp_sdias.c
186
sccb->evbuf.asa = __pa(dest);
drivers/scsi/bfa/bfa_ioc.h
184
#define bfa_alen_set(__alen, __len, __pa) \
drivers/scsi/bfa/bfa_ioc.h
185
__bfa_alen_set(__alen, __len, (u64)__pa)
drivers/scsi/ps3rom.c
126
lpar = ps3_mm_phys_to_lpar(__pa(&atapi_cmnd));
drivers/tty/serial/sunhv.c
123
unsigned long ra = __pa(con_read_page);
drivers/tty/serial/sunhv.c
450
unsigned long ra = __pa(con_write_page);
drivers/tty/serial/sunhv.c
560
port->membase = (unsigned char __iomem *) __pa(port);
drivers/tty/serial/sunhv.c
65
ra = __pa(tail);
drivers/usb/early/xhci-dbc.c
102
*dma_addr = (dma_addr_t)__pa(virt);
drivers/usb/early/xhci-dbc.c
839
xdbc_write64(__pa(xdbc.evt_ring.dequeue), &xdbc.xdbc_reg->erdp);
drivers/vfio/vfio_iommu_spapr_tce.c
494
*hpa = __pa((unsigned long) page_address(page));
drivers/video/fbdev/ps3fb.c
1095
xdr_lpar = ps3_mm_phys_to_lpar(__pa(ps3fb_videomemory.address));
drivers/video/fbdev/ps3fb.c
1143
info->fix.smem_start = __pa(fb_start);
drivers/video/fbdev/ps3fb.c
1212
u64 xdr_lpar = ps3_mm_phys_to_lpar(__pa(ps3fb_videomemory.address));
drivers/video/fbdev/ssd1307fb.c
701
info->fix.smem_start = __pa(vmem);
drivers/virt/coco/sev-guest/sev-guest.c
392
ac.report_buf.pa = __pa(rep);
drivers/virt/coco/sev-guest/sev-guest.c
396
ac.manifest_buf.pa = __pa(man);
drivers/virt/coco/sev-guest/sev-guest.c
400
ac.certificates_buf.pa = __pa(certs);
drivers/virt/coco/sev-guest/sev-guest.c
403
ac.nonce.pa = __pa(desc->inblob);
drivers/watchdog/wdrtas.c
111
WDRTAS_SP_SPI, __pa(rtas_data_buf),
drivers/watchdog/wdrtas.c
163
(void *)__pa(wdrtas_logbuffer),
fs/proc/kcore.c
163
unsigned long pfn = __pa(ent->addr) >> PAGE_SHIFT;
fs/proc/kcore.c
403
phdr->p_paddr = __pa(m->addr);
fs/proc/kcore.c
556
phys = __pa(start);
fs/proc/vmcore.c
615
pfn = __pa(elfcorebuf + start) >> PAGE_SHIFT;
fs/resctrl/pseudo_lock.c
1035
physical = __pa(plr->kmem) >> PAGE_SHIFT;
include/asm-generic/io.h
1099
return __pa((unsigned long)address);
include/linux/efi.h
848
unsigned long paddr = __pa(start + i);
include/linux/hyperv.h
1759
paddr = __pa(addr);
include/linux/kexec.h
512
return phys_to_boot_phys(__pa((unsigned long)addr));
include/linux/mm.h
117
#define __pa_symbol(x) __pa(RELOC_HIDE((unsigned long)(x), 0))
include/linux/psp.h
11
#define __psp_pa(x) __pa(x)
include/xen/grant_table.h
168
map->host_addr = __pa(addr);
include/xen/grant_table.h
185
unmap->host_addr = __pa(addr);
kernel/dma/swiotlb.c
412
swiotlb_init_io_tlb_pool(mem, __pa(tlb), nslabs, false, nareas);
kernel/dma/swiotlb.c
550
memblock_free_late(__pa(mem->areas),
kernel/dma/swiotlb.c
553
memblock_free_late(__pa(mem->slots), slots_size);
kernel/liveupdate/kexec_handover.c
1236
folio = kho_restore_folio(__pa(mem));
kernel/vmcore_info.c
128
return __pa(vmcoreinfo_note);
lib/bootconfig.c
67
memblock_free_late(__pa(addr), size);
mm/cma.c
360
phys_addr_t highmem_start = __pa(high_memory - 1) + 1;
mm/cma.c
408
phys_addr_t highmem = __pa(high_memory - 1) + 1;
mm/hugetlb.c
3104
listnode = early_pfn_to_nid(PHYS_PFN(__pa(m)));
mm/hugetlb.c
3163
memblock_reserved_mark_noinit(__pa((void *)m + PAGE_SIZE),
mm/kasan/init.c
82
void *ptr = memblock_alloc_try_nid(size, size, __pa(MAX_DMA_ADDRESS),
mm/kasan/init.c
87
__func__, size, size, node, (u64)__pa(MAX_DMA_ADDRESS));
mm/kfence/core.c
723
kmemleak_ignore_phys(__pa(__kfence_pool));
mm/kfence/core.c
734
memblock_free_late(__pa(addr), KFENCE_POOL_SIZE - (addr - (unsigned long)__kfence_pool));
mm/kfence/core.c
737
memblock_free_late(__pa(kfence_metadata_init), KFENCE_METADATA_SIZE);
mm/memblock.c
2135
pg = PAGE_ALIGN(__pa(start_pg));
mm/memblock.c
2136
pgend = PAGE_ALIGN_DOWN(__pa(end_pg));
mm/memblock.c
390
addr = __pa(memblock.reserved.regions);
mm/memblock.c
400
addr = __pa(memblock.memory.regions);
mm/memblock.c
463
addr = new_array ? __pa(new_array) : 0;
mm/memblock.c
907
memblock_phys_free(__pa(ptr), size);
mm/page_ext.c
215
table_size, PAGE_SIZE, __pa(MAX_DMA_ADDRESS),
mm/percpu.c
2389
return __pa(addr);
mm/percpu.c
2950
const unsigned long goal = __pa(MAX_DMA_ADDRESS);
mm/percpu.c
2963
cpu, size, (u64)__pa(ptr));
mm/percpu.c
2970
cpu, size, node, (u64)__pa(ptr));
mm/percpu.c
3060
kmemleak_ignore_phys(__pa(ptr));
mm/percpu.c
3240
kmemleak_ignore_phys(__pa(ptr));
mm/percpu.c
3343
fc = memblock_alloc_from(unit_size, PAGE_SIZE, __pa(MAX_DMA_ADDRESS));
mm/percpu.c
3347
kmemleak_ignore_phys(__pa(fc));
mm/sparse-vmemmap.c
167
ptpfn = PHYS_PFN(__pa(p));
mm/sparse-vmemmap.c
80
__pa(MAX_DMA_ADDRESS));
mm/sparse.c
316
return __pa(pgdat);
mm/sparse.c
364
usemap_snr = pfn_to_section_nr(__pa(usage) >> PAGE_SHIFT);
mm/sparse.c
423
phys_addr_t addr = __pa(MAX_DMA_ADDRESS);
mm/sparse.c
448
phys_addr_t addr = __pa(MAX_DMA_ADDRESS);