arch/alpha/kernel/pci_iommu.c
545
__va(paddr), size, out->dma_address);
arch/alpha/kernel/pci_iommu.c
557
__va(paddr), size, out->dma_address);
arch/alpha/kernel/pci_iommu.c
583
__va(paddr), size, out->dma_address, npages);
arch/alpha/kernel/setup.c
1172
read_mem_block(__va(0), stride, size);
arch/alpha/kernel/setup.c
1176
cycles = read_mem_block(__va(0), stride, size);
arch/alpha/kernel/setup.c
1185
read_mem_block(__va(size), stride, size);
arch/alpha/kernel/setup.c
430
hwrpb = (struct hwrpb_struct*) __va(INIT_HWRPB->phys_addr);
arch/alpha/kernel/sys_nautilus.c
258
free_reserved_area(__va(alpha_mv.min_mem_address),
arch/alpha/kernel/sys_nautilus.c
259
__va(memtop), -1, NULL);
arch/arc/mm/init.c
132
initrd_start = (unsigned long)__va(phys_initrd_start);
arch/arm/common/sharpsl_param.c
28
#define param_start(x) __va(x)
arch/arm/include/asm/memory.h
330
#define pfn_to_kaddr(pfn) __va((phys_addr_t)(pfn) << PAGE_SHIFT)
arch/arm/include/asm/pgtable-3level.h
137
return __va(pud_val(pud) & PHYS_MASK & (s32)PAGE_MASK);
arch/arm/include/asm/pgtable.h
162
return __va(pmd_val(pmd) & PHYS_MASK & (s32)PAGE_MASK);
arch/arm/mach-at91/pm.c
1055
soc_pm.memcs = __va((phys_addr_t)be32_to_cpu(*reg));
arch/arm/mm/dma-mapping-nommu.c
19
dmac_map_area(__va(paddr), size, dir);
arch/arm/mm/dma-mapping-nommu.c
32
dmac_unmap_area(__va(paddr), size, dir);
arch/arm/mm/kasan_init.c
252
void *start = __va(pa_start);
arch/arm/mm/kasan_init.c
253
void *end = __va(pa_end);
arch/arm/mm/kasan_init.c
263
end = __va(arm_lowmem_limit);
arch/arm/mm/mmu.c
1253
high_memory = __va(arm_lowmem_limit - 1) + 1;
arch/arm/mm/nommu.c
149
high_memory = __va(end - 1) + 1;
arch/arm64/include/asm/memory.h
393
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/arm64/include/asm/memory.h
404
void *__addr = __va(page_to_phys(__page)); \
arch/arm64/include/asm/pgtable.h
1001
return (pud_t *)__va(p4d_page_paddr(p4d));
arch/arm64/include/asm/pgtable.h
1016
return (pud_t *)__va(p4d_page_paddr(p4d)) + pud_index(addr);
arch/arm64/include/asm/pgtable.h
1139
return (p4d_t *)__va(pgd_page_paddr(pgd)) + p4d_index(addr);
arch/arm64/include/asm/pgtable.h
851
return (unsigned long)__va(pmd_page_paddr(pmd));
arch/arm64/include/asm/pgtable.h
911
return (pmd_t *)__va(pud_page_paddr(pud));
arch/arm64/kvm/mmu.c
279
return __va(phys);
arch/arm64/mm/kasan_init.c
126
memset(__va(page_phys), KASAN_SHADOW_INIT, PAGE_SIZE);
arch/csky/abiv1/inc/abi/ckmmu.h
99
return __va(cprcr("cpcr29") & ~BIT(0));
arch/csky/abiv2/inc/abi/ckmmu.h
137
return __va(mfcr("cr<29, 15>") & ~BIT(0));
arch/csky/include/asm/page.h
82
#define pfn_to_kaddr(x) __va(PFN_PHYS(x))
arch/csky/include/asm/pgtable.h
101
return __va(ptr);
arch/csky/mm/dma-mapping.c
18
void *start = __va(page_to_phys(page));
arch/hexagon/include/asm/io.h
33
return __va(address);
arch/hexagon/include/asm/mem-layout.h
62
#define VMALLOC_START ((unsigned long) __va(high_memory + VMALLOC_OFFSET))
arch/hexagon/include/asm/page.h
123
#define page_to_virt(page) __va(page_to_phys(page))
arch/hexagon/include/asm/pgtable.h
348
return (unsigned long)__va(pmd_val(pmd) & PAGE_MASK);
arch/loongarch/include/asm/io.h
80
(unlikely(__kfence_pool == NULL)) ? __va((unsigned long)paddr) : \
arch/loongarch/include/asm/page.h
69
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/loongarch/include/asm/page.h
80
#define page_to_virt(page) __va(page_to_phys(page))
arch/loongarch/include/asm/page.h
90
(__kfence_pool == NULL) ? __va(page_to_phys(page)) : page_address(page); \
arch/loongarch/kvm/mmu.c
171
child = (kvm_pte_t *)__va(PHYSADDR(*entry));
arch/loongarch/kvm/mmu.c
193
child = (kvm_pte_t *)__va(PHYSADDR(*dir));
arch/loongarch/kvm/mmu.c
227
child = (kvm_pte_t *)__va(PHYSADDR(*dir));
arch/loongarch/mm/kasan_init.c
126
memcpy(__va(pte_phys), kasan_early_shadow_pte, sizeof(kasan_early_shadow_pte));
arch/loongarch/mm/kasan_init.c
127
pmd_populate_kernel(NULL, pmdp, (pte_t *)__va(pte_phys));
arch/loongarch/mm/kasan_init.c
139
memcpy(__va(pmd_phys), kasan_early_shadow_pmd, sizeof(kasan_early_shadow_pmd));
arch/loongarch/mm/kasan_init.c
140
pud_populate(&init_mm, pudp, (pmd_t *)__va(pmd_phys));
arch/loongarch/mm/kasan_init.c
152
memcpy(__va(pud_phys), kasan_early_shadow_pud, sizeof(kasan_early_shadow_pud));
arch/loongarch/mm/kasan_init.c
153
p4d_populate(&init_mm, p4dp, (pud_t *)__va(pud_phys));
arch/loongarch/mm/kasan_init.c
165
memcpy(__va(p4d_phys), kasan_early_shadow_p4d, sizeof(kasan_early_shadow_p4d));
arch/loongarch/mm/kasan_init.c
166
pgd_populate(&init_mm, pgdp, (p4d_t *)__va(p4d_phys));
arch/m68k/include/asm/motorola_pgtable.h
100
#define __pte_page(pte) ((unsigned long)__va(pte_val(pte) & PAGE_MASK))
arch/m68k/include/asm/motorola_pgtable.h
101
#define pmd_page_vaddr(pmd) ((unsigned long)__va(pmd_val(pmd) & _TABLE_MASK))
arch/m68k/include/asm/motorola_pgtable.h
102
#define pud_pgtable(pud) ((pmd_t *)__va(pud_val(pud) & _TABLE_MASK))
arch/m68k/include/asm/motorola_pgtable.h
110
#define pte_page(pte) virt_to_page(__va(pte_val(pte)))
arch/m68k/include/asm/motorola_pgtable.h
132
#define pud_page(pud) (mem_map + ((unsigned long)(__va(pud_val(pud)) - PAGE_OFFSET) >> PAGE_SHIFT))
arch/m68k/include/asm/page_mm.h
130
return __va(pfn << PAGE_SHIFT);
arch/m68k/include/asm/page_no.h
28
return __va(pfn << PAGE_SHIFT);
arch/m68k/include/asm/page_no.h
32
#define page_to_virt(page) __va(((((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET))
arch/m68k/include/asm/sun3_pgtable.h
19
#define PTOV(addr) __va(addr)
arch/m68k/include/asm/sun3_pgtable.h
88
(__va ((pte_val (pte) & SUN3_PAGE_PGNUM_MASK) << PAGE_SHIFT))
arch/m68k/include/asm/sun3_pgtable.h
92
return (unsigned long)__va(pmd_val(pmd) & PAGE_MASK);
arch/m68k/include/asm/virtconvert.h
28
return __va(address);
arch/m68k/kernel/traps.c
536
__va(desc), *(unsigned long *)__va(desc));
arch/m68k/kernel/traps.c
645
__va(desc), *(unsigned long *)__va(desc));
arch/m68k/mm/sun3mmu.c
69
pg_table = (pte_t *) __va ((unsigned long) pg_table);
arch/m68k/sun3/mmu_emu.c
393
pte = (pte_t *) __va ((unsigned long)(pte + offset));
arch/microblaze/include/asm/page.h
127
return __va(pfn_to_phys((pfn)));
arch/microblaze/include/asm/page.h
99
# define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
arch/mips/include/asm/page.h
207
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/mips/kernel/setup.c
169
initrd_end = (unsigned long)__va(end);
arch/mips/kernel/setup.c
170
initrd_start = (unsigned long)__va(__pa(initrd_start));
arch/mips/sgi-ip27/ip27-memory.c
361
__node_data[node] = __va(slot_freepfn << PAGE_SHIFT);
arch/nios2/include/asm/page.h
83
# define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/nios2/mm/ioremap.c
137
t_addr = __va(phys_addr);
arch/openrisc/include/asm/pgtable.h
310
return (unsigned long)__va(pte_val(pte) & PAGE_MASK);
arch/openrisc/include/asm/pgtable.h
335
return ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK));
arch/openrisc/mm/init.c
150
unsigned long *dtlb_vector = __va(0x900);
arch/openrisc/mm/init.c
151
unsigned long *itlb_vector = __va(0xa00);
arch/openrisc/mm/init.c
79
v = (u32) __va(p);
arch/parisc/include/asm/io.h
9
#define phys_to_virt(a) __va(a)
arch/parisc/include/asm/pgtable.h
298
#define pud_pgtable(pud) ((pmd_t *) __va(pud_address(pud)))
arch/parisc/include/asm/pgtable.h
368
return ((unsigned long) __va(pmd_address(pmd)));
arch/parisc/include/asm/pgtable.h
372
#define __pmd_page(pmd) ((unsigned long) __va(pmd_address(pmd)))
arch/parisc/kernel/cache.c
107
#define pfn_va(pfn) __va(PFN_PHYS(pfn))
arch/parisc/kernel/pci-dma.c
435
free_pages((unsigned long)__va(dma_handle), order);
arch/parisc/kernel/setup.c
54
strscpy(boot_command_line, (char *)__va(boot_args[1]),
arch/parisc/kernel/setup.c
74
initrd_start = (unsigned long)__va(boot_args[2]);
arch/parisc/kernel/setup.c
75
initrd_end = (unsigned long)__va(boot_args[3]);
arch/parisc/mm/init.c
593
__va(0), high_memory,
arch/parisc/mm/init.c
594
((unsigned long)high_memory - (unsigned long)__va(0)) >> 20,
arch/parisc/mm/init.c
631
map_pages((unsigned long)__va(start_paddr), start_paddr,
arch/parisc/mm/ioremap.c
34
t_addr = __va(phys_addr);
arch/powerpc/include/asm/book3s/64/pgtable.h
986
return (pud_t *)__va(p4d_val(p4d) & ~P4D_MASKED_BITS);
arch/powerpc/include/asm/book3s/64/pgtable.h
991
return (pmd_t *)__va(pud_val(pud) & ~PUD_MASKED_BITS);
arch/powerpc/include/asm/io.h
899
return (void *)__va(address);
arch/powerpc/include/asm/io.h
923
return __va(address - PCI_DRAM_OFFSET);
arch/powerpc/include/asm/page.h
227
return __va(pfn << PAGE_SHIFT);
arch/powerpc/include/asm/pgtable.h
75
return __va(pmd_val(pmd) & ~PMD_MASKED_BITS);
arch/powerpc/include/asm/sections.h
68
return start < (unsigned long)__va(real_end) &&
arch/powerpc/include/asm/sections.h
69
(unsigned long)__va(real_start) < end;
arch/powerpc/kernel/crash_dump.c
85
vaddr = __va(paddr);
arch/powerpc/kernel/fadump.c
1000
phdr->p_vaddr = (unsigned long)__va(start);
arch/powerpc/kernel/fadump.c
1112
fdh = __va(addr);
arch/powerpc/kernel/fadump.c
1470
return sprintf(buf, "%s\n", (char *)__va(fw_dump.param_area));
arch/powerpc/kernel/fadump.c
1494
params = __va(fw_dump.param_area);
arch/powerpc/kernel/fadump.c
1709
fdh = (struct fadump_crash_info_header *) __va(fw_dump.fadumphdr_addr);
arch/powerpc/kernel/fadump.c
719
fdh = __va(fw_dump.fadumphdr_addr);
arch/powerpc/kernel/setup-common.c
960
high_memory = (void *)__va(max_low_pfn * PAGE_SIZE);
arch/powerpc/kernel/setup_32.c
95
early_init_devtree(__va(dt_ptr));
arch/powerpc/kernel/setup_64.c
373
if (!dt_cpu_ftrs_init(__va(dt_ptr)))
arch/powerpc/kernel/setup_64.c
388
early_init_devtree(__va(dt_ptr));
arch/powerpc/kexec/core_64.c
135
flush_icache_range((unsigned long)__va(ranges[i].mem),
arch/powerpc/kexec/core_64.c
136
(unsigned long)__va(ranges[i].mem + ranges[i].memsz));
arch/powerpc/kexec/core_64.c
97
addr = __va(entry & PAGE_MASK);
arch/powerpc/kexec/crash.c
477
ptr = __va(mem);
arch/powerpc/kexec/crash.c
510
ptr = __va(mem);
arch/powerpc/kexec/crash.c
537
fdt = __va((void *)image->segment[fdt_index].mem);
arch/powerpc/kvm/book3s_32_mmu_host.c
379
htab = (ulong)__va(sdr1 & 0xffff0000);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
28
return __va(ppc_find_vmap_phys((unsigned long)addr));
arch/powerpc/mm/book3s64/hash_utils.c
1343
stress_hpt_struct = __va(tmp);
arch/powerpc/mm/book3s64/hash_utils.c
1386
htab_address = __va(table);
arch/powerpc/mm/book3s64/hash_utils.c
1408
base = (unsigned long)__va(base);
arch/powerpc/mm/book3s64/hash_utils.c
1434
tce_alloc_start = (unsigned long)__va(tce_alloc_start);
arch/powerpc/mm/book3s64/hash_utils.c
1435
tce_alloc_end = (unsigned long)__va(tce_alloc_end);
arch/powerpc/mm/book3s64/hash_utils.c
458
kfence_pool_start = (unsigned long) __va(kfence_pool);
arch/powerpc/mm/book3s64/hash_utils.c
470
unsigned long vaddr = (unsigned long) __va(paddr);
arch/powerpc/mm/book3s64/hash_utils.c
534
if (is_kfence_address(__va(paddr)))
arch/powerpc/mm/book3s64/radix_pgtable.c
339
vaddr = (unsigned long)__va(addr);
arch/powerpc/mm/book3s64/radix_pgtable.c
407
__kfence_pool = __va(kfence_pool);
arch/powerpc/mm/init_64.c
372
free_pages((unsigned long)(__va(addr)), page_order);
arch/powerpc/mm/kasan/init_32.c
150
ret = kasan_init_region(__va(base), top - base);
arch/powerpc/mm/mem.c
102
start = (unsigned long)__va(start);
arch/powerpc/mm/mem.c
128
high_memory = (void *)__va(max_pfn * PAGE_SIZE - 1) + 1;
arch/powerpc/mm/mem.c
84
start = (unsigned long)__va(start);
arch/powerpc/mm/nohash/8xx.c
48
return (unsigned long)__va(pa);
arch/powerpc/mm/nohash/e500.c
350
early_get_first_memblock_info(__va(dt_ptr), &size);
arch/powerpc/mm/nohash/e500.c
371
restore_to_as0(n, offset, __va(dt_ptr), 1);
arch/powerpc/mm/nohash/e500.c
376
kaslr_early_init(__va(dt_ptr), size);
arch/powerpc/platforms/44x/iss4xx.c
108
spin_table = (u32 *)__va(*spin_table_addr_prop);
arch/powerpc/platforms/44x/ppc476.c
181
spin_table = (u32 *)__va(*spin_table_addr_prop);
arch/powerpc/platforms/powernv/opal-core.c
199
memcpy(to, __va(addr), tsz);
arch/powerpc/platforms/powernv/opal-core.c
232
bufp = __va(oc_conf->cpu_state_destination_vaddr);
arch/powerpc/platforms/powernv/opal-core.c
486
opalc_metadata = __va(addr);
arch/powerpc/platforms/powernv/opal-core.c
497
opalc_cpu_metadata = __va(addr);
arch/powerpc/platforms/powernv/opal-core.c
545
oc_conf->cpu_state_destination_vaddr = (u64)__va(addr);
arch/powerpc/platforms/powernv/opal-fadump.c
193
opal_fdm = __va(fadump_conf->kernel_metadata);
arch/powerpc/platforms/powernv/opal-fadump.c
240
opal_fdm = __va(fadump_conf->kernel_metadata);
arch/powerpc/platforms/powernv/opal-fadump.c
382
(u64)__va(be64_to_cpu(opal_cpu_metadata->region[0].dest));
arch/powerpc/platforms/powernv/opal-fadump.c
431
bufp = __va(fadump_conf->cpu_state_dest_vaddr);
arch/powerpc/platforms/powernv/opal-fadump.c
529
fdh = __va(fadump_conf->fadumphdr_addr);
arch/powerpc/platforms/powernv/opal-fadump.c
695
opal_fdm_active = __va(addr);
arch/powerpc/platforms/powernv/opal-fadump.c
712
opal_cpu_metadata = __va(addr);
arch/powerpc/platforms/powernv/opal.c
1187
sg = __va(next);
arch/powerpc/platforms/powernv/opal.c
819
attr->private = __va(vals[0]);
arch/powerpc/platforms/powernv/pci-ioda-tce.c
115
tmp = __va(tce & ~(TCE_PCI_READ | TCE_PCI_WRITE));
arch/powerpc/platforms/powernv/pci-ioda-tce.c
232
pnv_pci_ioda2_table_do_free_pages(__va(hpa), size,
arch/powerpc/platforms/ps3/os-area.c
732
header = (struct os_area_header *)__va(lpar_addr);
arch/powerpc/platforms/ps3/os-area.c
733
params = (struct os_area_params *)__va(lpar_addr
arch/powerpc/platforms/ps3/os-area.c
745
db = (struct os_area_db *)__va(lpar_addr
arch/powerpc/platforms/pseries/iommu.c
494
void *uaddr = __va(start_pfn << PAGE_SHIFT);
arch/powerpc/platforms/pseries/iommu.c
606
__va(*basep), &iommu_table_pseries_ops);
arch/powerpc/platforms/pseries/ras.c
422
return __va(savep_ra);
arch/powerpc/platforms/pseries/rtas-fadump.c
397
fdh = __va(fadump_conf->fadumphdr_addr);
arch/powerpc/platforms/pseries/rtas-fadump.c
538
(char *)__va(be64_to_cpu(fdm_ptr->rgn[i].destination_address)));
arch/powerpc/platforms/pseries/rtas-fadump.c
59
fadump_conf->cpu_state_dest_vaddr = (u64)__va(addr);
arch/riscv/include/asm/page.h
186
#define pfn_to_virt(pfn) (__va(pfn_to_phys(pfn)))
arch/riscv/include/asm/page.h
197
return __va(pfn << PAGE_SHIFT);
arch/riscv/kernel/acpi.c
299
return (void __iomem *)__va(phys);
arch/riscv/mm/init.c
1269
va = (uintptr_t)__va(pa);
arch/riscv/mm/init.c
1305
__kfence_pool = __va(kfence_pool);
arch/riscv/mm/init.c
1787
unsigned long va = (unsigned long)__va(start);
arch/riscv/mm/init.c
1788
unsigned long end = (unsigned long)__va(start + size);
arch/riscv/mm/init.c
419
return (pte_t *) __va(pa);
arch/riscv/mm/init.c
506
return (pmd_t *) __va(pa);
arch/riscv/mm/init.c
570
return (pud_t *)__va(pa);
arch/riscv/mm/init.c
608
return (p4d_t *)__va(pa);
arch/riscv/mm/init.c
686
#define alloc_pgd_next(__va) (pgtable_l5_enabled ? \
arch/riscv/mm/init.c
687
pt_ops.alloc_p4d(__va) : (pgtable_l4_enabled ? \
arch/riscv/mm/init.c
688
pt_ops.alloc_pud(__va) : pt_ops.alloc_pmd(__va)))
arch/riscv/mm/init.c
692
#define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \
arch/riscv/mm/init.c
694
create_p4d_mapping(__nextp, __va, __pa, __sz, __prot) : \
arch/riscv/mm/init.c
696
create_pud_mapping((pud_t *)__nextp, __va, __pa, __sz, __prot) : \
arch/riscv/mm/init.c
697
create_pmd_mapping((pmd_t *)__nextp, __va, __pa, __sz, __prot)))
arch/riscv/mm/init.c
706
#define alloc_pgd_next(__va) pt_ops.alloc_pte(__va)
arch/riscv/mm/init.c
708
#define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \
arch/riscv/mm/init.c
709
create_pte_mapping(__nextp, __va, __pa, __sz, __prot)
arch/riscv/mm/init.c
711
#define create_p4d_mapping(__pmdp, __va, __pa, __sz, __prot) do {} while(0)
arch/riscv/mm/init.c
712
#define create_pud_mapping(__pmdp, __va, __pa, __sz, __prot) do {} while(0)
arch/riscv/mm/init.c
713
#define create_pmd_mapping(__pmdp, __va, __pa, __sz, __prot) do {} while(0)
arch/riscv/mm/kasan_init.c
102
memset(__va(phys_addr), KASAN_SHADOW_INIT, PUD_SIZE);
arch/riscv/mm/kasan_init.c
133
memset(__va(phys_addr), KASAN_SHADOW_INIT, P4D_SIZE);
arch/riscv/mm/kasan_init.c
156
memset(__va(phys_addr), KASAN_SHADOW_INIT, PGDIR_SIZE);
arch/riscv/mm/kasan_init.c
45
memset(__va(phys_addr), KASAN_SHADOW_INIT, PAGE_SIZE);
arch/riscv/mm/kasan_init.c
512
void *start = (void *)__va(p_start);
arch/riscv/mm/kasan_init.c
513
void *end = (void *)__va(p_end);
arch/riscv/mm/kasan_init.c
71
memset(__va(phys_addr), KASAN_SHADOW_INIT, PMD_SIZE);
arch/s390/include/asm/dma-types.h
45
return __va((__force unsigned long)addr);
arch/s390/include/asm/dma-types.h
80
return __va((__force unsigned long)addr);
arch/s390/include/asm/dma.h
12
#define MAX_DMA_ADDRESS __va(0x80000000)
arch/s390/include/asm/page.h
265
return __va(pfn_to_phys(pfn));
arch/s390/include/asm/pgtable.h
1357
#define p4d_deref(pud) ((unsigned long)__va(p4d_val(pud) & _REGION_ENTRY_ORIGIN))
arch/s390/include/asm/pgtable.h
1358
#define pgd_deref(pgd) ((unsigned long)__va(pgd_val(pgd) & _REGION_ENTRY_ORIGIN))
arch/s390/include/asm/pgtable.h
1367
return (unsigned long)__va(pmd_val(pmd) & origin_mask);
arch/s390/include/asm/pgtable.h
1382
return (unsigned long)__va(pud_val(pud) & origin_mask);
arch/s390/include/asm/pgtable.h
1934
((pgtable_t)__va(pmd_val(pmd) & -sizeof(pte_t)*PTRS_PER_PTE))
arch/s390/include/asm/physmem_info.h
140
range && range->end; range = range->chain ? __va(range->chain) : NULL, \
arch/s390/include/asm/physmem_info.h
152
return __va(range->chain);
arch/s390/kernel/machine_kexec.c
96
mcesa = __va(get_lowcore()->mcesad & MCESA_ORIGIN_MASK);
arch/s390/kernel/machine_kexec_file.c
197
ptr = __va(ipl_cert_list_addr);
arch/s390/kernel/machine_kexec_file.c
209
ptr = __va(ipl_cert_list_addr);
arch/s390/kernel/nmi.c
93
kfree(__va(*mcesad & MCESA_ORIGIN_MASK));
arch/s390/kernel/setup.c
673
initrd_start = (unsigned long)__va(addr);
arch/s390/kernel/setup.c
722
high_memory = __va(ident_map_size);
arch/s390/kernel/setup.c
857
ptr = __va(early_ipl_comp_list_addr);
arch/s390/kvm/gaccess.c
1108
context->exception = __cmpxchg_with_key(__va(PFN_PHYS(f->pfn) | context->offset),
arch/s390/kvm/gaccess.c
872
ptr = __va(PFN_PHYS(f->pfn) | context->offset);
arch/s390/mm/fault.c
108
table = __va(entry & _REGION_ENTRY_ORIGIN);
arch/s390/mm/fault.c
117
table = __va(entry & _SEGMENT_ENTRY_ORIGIN);
arch/s390/mm/fault.c
79
unsigned long entry, *table = __va(asce & _ASCE_ORIGIN);
arch/s390/mm/fault.c
90
table = __va(entry & _REGION_ENTRY_ORIGIN);
arch/s390/mm/fault.c
99
table = __va(entry & _REGION_ENTRY_ORIGIN);
arch/s390/mm/pgalloc.c
265
table = __va(*ste & _SEGMENT_ENTRY_ORIGIN);
arch/s390/mm/pgalloc.c
294
table = __va(*rtte & _REGION_ENTRY_ORIGIN);
arch/s390/mm/pgalloc.c
322
table = __va(*rste & _REGION_ENTRY_ORIGIN);
arch/s390/mm/pgalloc.c
350
table = __va(*rfte & _REGION_ENTRY_ORIGIN);
arch/s390/mm/pgalloc.c
369
unsigned long *table = __va(asce & _ASCE_ORIGIN);
arch/s390/mm/vmem.c
492
start = (unsigned long)__va(start);
arch/s390/mm/vmem.c
501
start = (unsigned long)__va(start);
arch/s390/mm/vmem.c
678
__set_memory_4k(__va(0), absolute_pointer(__va(0)) + ident_map_size);
arch/sh/include/asm/io.h
269
#define phys_to_virt(address) (__va(address))
arch/sh/include/asm/page.h
147
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/sh/kernel/setup.c
162
initrd_start = (unsigned long)__va(start);
arch/sh/kernel/setup.c
232
pmb_bolt_mapping((unsigned long)__va(start), start, end - start,
arch/sh/mm/init.c
298
memory_start = (unsigned long)__va(__MEMORY_START);
arch/sh/mm/numa.c
33
pmb_bolt_mapping((unsigned long)__va(start), start, end - start,
arch/sparc/include/asm/page_32.h
126
#define phys_to_virt __va
arch/sparc/include/asm/page_64.h
148
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/sparc/include/asm/page_64.h
155
#define phys_to_virt __va
arch/sparc/include/asm/pgtable_64.h
843
return ((unsigned long) __va(pfn << PAGE_SHIFT));
arch/sparc/include/asm/pgtable_64.h
853
return ((pmd_t *) __va(pfn << PAGE_SHIFT));
arch/sparc/include/asm/pgtable_64.h
862
((pud_t *) __va(p4d_val(p4d)))
arch/sparc/include/asm/pgtsrmmu.h
107
#define __nocache_va(PADDR) (__va((unsigned long)PADDR) - (unsigned long)srmmu_nocache_pool + SRMMU_NOCACHE_VADDR)
arch/sparc/include/asm/pgtsrmmu.h
108
#define __nocache_fix(VADDR) ((__typeof__(VADDR))__va(__nocache_pa(VADDR)))
arch/sparc/kernel/irq_64.c
268
bucket = (struct ino_bucket *) __va(cookie);
arch/sparc/kernel/mdesc.c
177
hp = __va(paddr);
arch/sparc/kernel/smp_64.c
467
cpu_list = __va(tb->cpu_list_pa);
arch/sparc/kernel/smp_64.c
468
mondo = __va(tb->cpu_mondo_block_pa);
arch/sparc/kernel/smp_64.c
486
cpu_list = __va(tb->cpu_list_pa);
arch/sparc/kernel/smp_64.c
487
mondo = __va(tb->cpu_mondo_block_pa);
arch/sparc/kernel/smp_64.c
655
cpu_list = __va(tb->cpu_list_pa);
arch/sparc/kernel/smp_64.c
806
mondo = __va(tb->cpu_mondo_block_pa);
arch/sparc/kernel/smp_64.c
812
cpu_list = __va(tb->cpu_list_pa);
arch/sparc/kernel/traps_64.c
2090
ent = __va(paddr);
arch/sparc/kernel/traps_64.c
2208
ent = __va(paddr);
arch/sparc/mm/init_64.c
1241
mlgroups = __va(paddr);
arch/sparc/mm/init_64.c
1282
mblocks = __va(paddr);
arch/sparc/mm/init_64.c
2541
((unsigned long) __va(kern_base)) -
arch/sparc/mm/srmmu.c
881
map_spbank((unsigned long)__va(sp_banks[i].base_addr), i);
arch/um/include/asm/pgtable-4level.h
95
#define pud_pgtable(pud) ((pmd_t *) __va(pud_val(pud) & PAGE_MASK))
arch/um/include/asm/pgtable-4level.h
98
#define p4d_pgtable(p4d) ((pud_t *) __va(p4d_val(p4d) & PAGE_MASK))
arch/um/include/asm/pgtable.h
291
#define pmd_page_vaddr(pmd) ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
arch/x86/coco/sev/core.c
433
vaddr = (unsigned long)__va(start);
arch/x86/coco/sev/vc-handle.c
48
pgd = __va(read_cr3_pa());
arch/x86/include/asm/dma.h
84
#define MAX_DMA_ADDRESS ((unsigned long)__va(MAX_DMA_PFN << PAGE_SHIFT))
arch/x86/include/asm/io.h
150
return __va(address);
arch/x86/include/asm/numachip/numachip_csr.h
41
return __va(NUMACHIP_LCSR_BASE | (1UL << 15) |
arch/x86/include/asm/numachip/numachip_csr.h
69
return (void __iomem *)__va(NUMACHIP2_LCSR_BASE |
arch/x86/include/asm/page.h
51
#ifndef __va
arch/x86/include/asm/page.h
55
#define __boot_va(x) __va(x)
arch/x86/include/asm/page.h
68
return __va(pfn << PAGE_SHIFT);
arch/x86/include/asm/pgtable.h
1024
return (unsigned long)__va(pmd_val(pmd) & pmd_pfn_mask(pmd));
arch/x86/include/asm/pgtable.h
1057
return (pmd_t *)__va(pud_val(pud) & pud_pfn_mask(pud));
arch/x86/include/asm/pgtable.h
1091
return (pud_t *)__va(p4d_val(p4d) & p4d_pfn_mask(p4d));
arch/x86/include/asm/pgtable.h
1126
return (unsigned long)__va((unsigned long)pgd_val(pgd) & PTE_PFN_MASK);
arch/x86/include/asm/realmode.h
91
real_mode_header = (struct real_mode_header *) __va(mem);
arch/x86/include/asm/unaccepted_memory.h
25
return __va(efi.unaccepted);
arch/x86/include/asm/uv/uv_hub.h
538
return __va(((unsigned long)pnode << m_val) | offset);
arch/x86/include/asm/uv/uv_hub.h
544
return __va((unsigned long)offset);
arch/x86/include/asm/uv/uv_hub.h
547
return __va(base << UV_GAM_RANGE_SHFT | offset);
arch/x86/include/asm/uv/uv_hub.h
565
return __va(UV_GLOBAL_MMR32_BASE |
arch/x86/include/asm/uv/uv_hub.h
585
return __va(UV_GLOBAL_MMR64_BASE |
arch/x86/include/asm/uv/uv_hub.h
615
return __va(UV_LOCAL_MMR_BASE | offset);
arch/x86/include/asm/xen/page.h
303
#define mfn_to_virt(m) (__va(mfn_to_pfn(m) << PAGE_SHIFT))
arch/x86/include/asm/xen/page.h
307
#define gfn_to_virt(g) (__va(gfn_to_pfn(g) << PAGE_SHIFT))
arch/x86/kernel/acpi/sleep.c
63
(struct wakeup_header *) __va(real_mode_header->wakeup_header);
arch/x86/kernel/amd_gart_64.c
780
set_memory_np((unsigned long)__va(iommu_bus_base),
arch/x86/kernel/apm_32.c
2313
(unsigned long)__va((unsigned long)apm_info.bios.cseg << 4));
arch/x86/kernel/apm_32.c
2315
(unsigned long)__va((unsigned long)apm_info.bios.cseg_16 << 4));
arch/x86/kernel/apm_32.c
2317
(unsigned long)__va((unsigned long)apm_info.bios.dseg << 4));
arch/x86/kernel/apm_32.c
423
(unsigned long)__va(0x400UL), PAGE_SIZE - 0x400 - 1);
arch/x86/kernel/check.c
129
memset(__va(start), 0, end - start);
arch/x86/kernel/check.c
149
unsigned long *addr = __va(scan_areas[i].addr);
arch/x86/kernel/cpu/common.c
2653
set_memory_4k((unsigned long)__va(0), 1);
arch/x86/kernel/ebda.c
74
bios_start = *(unsigned short *)__va(BIOS_RAM_SIZE_KB_PTR);
arch/x86/kernel/head64.c
209
command_line = __va(cmd_line_ptr);
arch/x86/kernel/head64.c
281
copy_bootdata(__va(real_mode_data));
arch/x86/kernel/head64.c
298
copy_bootdata(__va(real_mode_data));
arch/x86/kernel/machine_kexec_64.c
179
vaddr = (unsigned long)__va(control_page);
arch/x86/kernel/reboot.c
643
*((unsigned short *)__va(0x472)) = mode;
arch/x86/kernel/setup.c
449
*addr = __va(ima_kexec_buffer_phys);
arch/x86/kvm/mmu/mmu.c
2518
iterator->sptep = ((u64 *)__va(iterator->shadow_addr)) + iterator->index;
arch/x86/kvm/mmu/tdp_iter.c
73
return (tdp_ptep_t)__va(spte_to_pfn(spte) << PAGE_SHIFT);
arch/x86/kvm/svm/sev.c
5089
clflush_cache_range(__va(pfn_to_hpa(pfn)),
arch/x86/kvm/vmx/vmx.c
4067
const void *zero_page = (const void *) __va(page_to_phys(ZERO_PAGE(0)));
arch/x86/mm/fault.c
250
pmd_k = vmalloc_sync_one(__va(pgd_paddr), address);
arch/x86/mm/fault.c
296
pgd_t *base = __va(read_cr3_pa());
arch/x86/mm/fault.c
352
pgd_t *base = __va(read_cr3_pa());
arch/x86/mm/fault.c
520
pgd = __va(read_cr3_pa());
arch/x86/mm/init.c
162
adr = __va((pfn + i) << PAGE_SHIFT);
arch/x86/mm/init.c
166
return __va(pfn << PAGE_SHIFT);
arch/x86/mm/init_32.c
623
high_memory = (void *) __va(highstart_pfn * PAGE_SIZE - 1) + 1;
arch/x86/mm/init_32.c
625
high_memory = (void *) __va(max_low_pfn * PAGE_SIZE - 1) + 1;
arch/x86/mm/init_64.c
1285
start = (unsigned long)__va(start);
arch/x86/mm/init_64.c
1286
end = (unsigned long)__va(end);
arch/x86/mm/init_64.c
397
pgd = pgd_offset_k((unsigned long)__va(phys));
arch/x86/mm/init_64.c
403
p4d = p4d_offset(pgd, (unsigned long)__va(phys));
arch/x86/mm/init_64.c
409
pud = pud_offset(p4d, (unsigned long)__va(phys));
arch/x86/mm/init_64.c
612
unsigned long vaddr = (unsigned long)__va(paddr);
arch/x86/mm/init_64.c
620
vaddr = (unsigned long)__va(paddr);
arch/x86/mm/init_64.c
696
vaddr = (unsigned long)__va(paddr);
arch/x86/mm/init_64.c
697
vaddr_end = (unsigned long)__va(paddr_end);
arch/x86/mm/init_64.c
750
vaddr = (unsigned long)__va(paddr_start);
arch/x86/mm/init_64.c
751
vaddr_end = (unsigned long)__va(paddr_end);
arch/x86/mm/init_64.c
960
high_memory = (void *)__va(max_pfn * PAGE_SIZE - 1) + 1;
arch/x86/mm/ioremap.c
829
pgd_t *base = __va(read_cr3_pa());
arch/x86/mm/kaslr.c
188
vaddr = (unsigned long)__va(paddr);
arch/x86/mm/mem_encrypt_amd.c
194
__sme_early_map_unmap_mem(__va(cmdline_paddr), COMMAND_LINE_SIZE, false);
arch/x86/mm/mem_encrypt_amd.c
214
__sme_early_map_unmap_mem(__va(cmdline_paddr), COMMAND_LINE_SIZE, true);
arch/x86/mm/mem_encrypt_amd.c
344
clflush_cache_range(__va(d->pa), d->size);
arch/x86/mm/mem_encrypt_amd.c
379
early_snp_set_memory_shared((unsigned long)__va(d.pa), d.pa, 1);
arch/x86/mm/mem_encrypt_amd.c
389
early_snp_set_memory_private((unsigned long)__va(d.pa), d.pa, 1);
arch/x86/mm/mem_encrypt_amd.c
67
early_snp_set_memory_shared((unsigned long)__va(paddr), paddr, npages);
arch/x86/mm/mem_encrypt_amd.c
72
early_snp_set_memory_private((unsigned long)__va(paddr), paddr, npages);
arch/x86/mm/pat/cpa-test.c
141
addr[i] = (unsigned long)__va(pfn << PAGE_SHIFT);
arch/x86/mm/pat/cpa-test.c
58
unsigned long addr = (unsigned long)__va(i << PAGE_SHIFT);
arch/x86/mm/pat/memtype.c
824
if (ioremap_change_attr((unsigned long)__va(base), id_sz, pcm) < 0) {
arch/x86/mm/pat/set_memory.c
1915
unsigned long laddr = (unsigned long)__va(cpa->pfn << PAGE_SHIFT);
arch/x86/pci/irq.c
173
u8 * const bios_start = (u8 *)__va(0xf0000);
arch/x86/pci/irq.c
174
u8 * const bios_end = (u8 *)__va(0x100000);
arch/x86/pci/irq.c
179
rt = pirq_check_routing_table((u8 *)__va(pirq_table_addr),
arch/x86/pci/pcbios.c
314
for (check = (union bios32 *) __va(0xe0000);
arch/x86/pci/pcbios.c
315
check <= (union bios32 *) __va(0xffff0);
arch/x86/platform/efi/efi_32.c
48
va = __va(md->phys_addr);
arch/x86/platform/intel-quark/imr_selftest.c
72
imr_self_test_result(ret < 0, fmt_over, __va(base), __va(base + size));
arch/x86/platform/intel-quark/imr_selftest.c
77
imr_self_test_result(ret < 0, fmt_over, __va(base), __va(base + size));
arch/x86/platform/intel-quark/imr_selftest.c
82
imr_self_test_result(ret < 0, fmt_over, __va(base), __va(base + size));
arch/x86/platform/olpc/olpc-xo1-pm.c
75
void *pgd_addr = __va(read_cr3_pa());
arch/x86/platform/pvh/enlighten.c
47
ep = __va(pvh_start_info.memmap_paddr);
arch/x86/platform/pvh/enlighten.c
79
__va(pvh_start_info.modlist_paddr);
arch/x86/power/hibernate.c
169
pgd = (pgd_t *)__va(read_cr3_pa()) +
arch/x86/realmode/init.c
140
__va(real_mode_header->trampoline_header);
arch/x86/realmode/init.c
163
trampoline_pgd = (u64 *) __va(real_mode_header->trampoline_pgd);
arch/x86/realmode/init.c
202
(unsigned long) __va(real_mode_header->text_start);
arch/x86/virt/svm/sev.c
846
pgd = __va(read_cr3_pa());
arch/x86/virt/vmx/tdx/tdx.c
652
movdir64b(__va(phys), zero_page);
arch/x86/xen/enlighten_hvm.c
76
HYPERVISOR_shared_info = __va(PFN_PHYS(shared_info_pfn));
arch/x86/xen/enlighten_pv.c
496
va = __va(PFN_PHYS(pfn));
arch/x86/xen/enlighten_pv.c
575
virt = __va(PFN_PHYS(pfn));
arch/x86/xen/mmu_pv.c
1092
void *vaddr = __va(paddr);
arch/x86/xen/mmu_pv.c
1108
ClearPagePinned(virt_to_page(__va(pa)));
arch/x86/xen/mmu_pv.c
1249
xen_start_info->pt_base = (unsigned long)__va(__pa(xen_start_info->pt_base));
arch/x86/xen/mmu_pv.c
1403
pgd_t *user_pgd = xen_get_user_pgd(__va(cr3));
arch/x86/xen/mmu_pv.c
1542
make_lowmem_page_readonly(__va(PFN_PHYS(pfn)));
arch/x86/xen/mmu_pv.c
1552
make_lowmem_page_readonly(__va(PFN_PHYS(pfn)));
arch/x86/xen/mmu_pv.c
1560
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
arch/x86/xen/mmu_pv.c
1565
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
arch/x86/xen/mmu_pv.c
1584
unsigned long addr = (unsigned long)__va(pfn << PAGE_SHIFT);
arch/x86/xen/mmu_pv.c
1887
xen_start_info = (struct start_info *)__va(__pa(xen_start_info));
arch/x86/xen/mmu_pv.c
1990
pgd = __va(read_cr3_pa());
arch/x86/xen/mmu_pv.c
2012
make_lowmem_page_readonly(__va(pt_phys));
arch/x86/xen/mmu_pv.c
2020
make_lowmem_page_readonly(__va(pmd_phys));
arch/x86/xen/mmu_pv.c
2028
make_lowmem_page_readonly(__va(pud_phys));
arch/x86/xen/mmu_pv.c
2059
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
arch/x86/xen/setup.c
269
if (HYPERVISOR_update_va_mapping((unsigned long)__va(pfn << PAGE_SHIFT),
arch/x86/xen/setup.c
392
(unsigned long)__va(pfn << PAGE_SHIFT),
arch/xtensa/include/asm/page.h
171
#define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
arch/xtensa/kernel/setup.c
109
initrd_start = (unsigned long)__va(mi->start);
arch/xtensa/kernel/setup.c
110
initrd_end = (unsigned long)__va(mi->end);
arch/xtensa/kernel/setup.c
123
dtb_start = __va(tag->data[0]);
drivers/char/agp/alpha-agp.c
30
page = virt_to_page(__va(pa));
drivers/char/agp/uninorth-agp.c
191
flush_dcache_range((unsigned long)__va(page_to_phys(mem->pages[i])),
drivers/char/agp/uninorth-agp.c
192
(unsigned long)__va(page_to_phys(mem->pages[i]))+0x1000);
drivers/char/tpm/eventlog/of.c
93
log->bios_event_log = devm_kmemdup(&chip->dev, __va(base), size, GFP_KERNEL);
drivers/crypto/ccp/sev-dev-tio.c
41
void *va = __va(__sme_clr(sla_to_pa(sla)));
drivers/crypto/ccp/sev-dev.c
729
void *bounce_buf = __va(__sme_clr(*desc->paddr_ptr));
drivers/crypto/ccp/sev-dev.c
730
void *dst_buf = __va(__sme_clr(desc->paddr_orig));
drivers/infiniband/hw/hfi1/user_exp_rcv.c
740
phys = dma_map_single(&dd->pcidev->dev, __va(page_to_phys(pages[0])),
drivers/iommu/generic_pt/pt_common.h
342
return __va(pt_table_pa(pts));
drivers/iommu/io-pgtable-arm.c
146
#define iopte_deref(pte,d) __va(iopte_to_paddr(pte, d))
drivers/iommu/io-pgtable-dart.c
67
#define iopte_deref(pte, d) __va(iopte_to_paddr(pte, d))
drivers/iommu/riscv/iommu.c
620
ddtp = __va(ppn_to_phys(ddt));
drivers/macintosh/smu.c
507
smu->cmd_buf = __va(smu_cmdbuf_abs);
drivers/misc/sgi-gru/grufile.c
451
gru_start_vaddr = __va(gru_start_paddr);
drivers/misc/sgi-xp/xp_uv.c
48
unsigned long *dst_va = __va(uv_gpa_to_soc_phys_ram(dst_gpa));
drivers/misc/sgi-xp/xpc_partition.c
148
rp = (struct xpc_rsvd_page *)__va(xp_socket_pa(rp_pa));
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
787
return __va(ret << SECTION_SIZE_BITS);
drivers/of/fdt.c
809
initrd_start = (unsigned long)__va(start);
drivers/of/fdt.c
810
initrd_end = (unsigned long)__va(end);
drivers/of/kexec.c
150
*addr = __va(tmp_addr);
drivers/platform/x86/dell/dcdbas.c
563
for (addr = (u8 *)__va(0xf0000);
drivers/platform/x86/dell/dcdbas.c
564
addr < (u8 *)__va(0x100000 - sizeof(struct smm_eps_table));
drivers/pnp/pnpbios/bioscalls.c
458
0, data, 65536, __va(nvram_base), 65536);
drivers/pnp/pnpbios/bioscalls.c
486
(unsigned long)__va(header->fields.pm16cseg));
drivers/pnp/pnpbios/bioscalls.c
488
(unsigned long)__va(header->fields.pm16dseg));
drivers/pnp/pnpbios/bioscalls.c
64
(unsigned long)__va(0x400UL), PAGE_SIZE - 0x400 - 1);
drivers/pnp/pnpbios/core.c
436
for (check = (union pnp_bios_install_struct *)__va(0xf0000);
drivers/pnp/pnpbios/core.c
437
check < (union pnp_bios_install_struct *)__va(0xffff0);
drivers/s390/block/dcssblk.c
922
memcpy(page_addr, __va(source_addr), bvec.bv_len);
drivers/s390/block/dcssblk.c
924
memcpy(__va(source_addr), page_addr, bvec.bv_len);
drivers/s390/block/dcssblk.c
942
*kaddr = __va(dev_info->start + offset);
drivers/s390/char/monreader.c
114
return *((u8 *)__va(mon_mca_start(monmsg)) + monmsg->mca_offset + index);
drivers/s390/char/monreader.c
124
return *((u32 *)(__va(mon_mca_start(monmsg)) + monmsg->mca_offset + 4));
drivers/s390/char/monreader.c
129
return *((u32 *)(__va(mon_mca_start(monmsg)) + monmsg->mca_offset + 8));
drivers/s390/char/monreader.c
394
ret = copy_to_user(data, __va(monmsg->pos), count);
drivers/s390/char/monreader.c
407
ret = copy_to_user(data, __va(monmsg->pos), count);
drivers/s390/char/sclp.c
83
return __va(sccb_int);
drivers/s390/char/sclp_early.c
170
sclp_early_sccb = __va((unsigned long)sclp_early_sccb);
drivers/s390/char/sclp_mem.c
226
__arch_set_page_nodat((void *)__va(addr), block_size >> PAGE_SHIFT);
drivers/s390/char/sclp_mem.c
254
start = (unsigned long)kasan_mem_to_shadow(__va(addr));
drivers/video/fbdev/ssd1307fb.c
782
__free_pages(__va(info->fix.smem_start), get_order(info->fix.smem_len));
drivers/xen/mem-reservation.c
51
(unsigned long)__va(pfn << PAGE_SHIFT),
drivers/xen/mem-reservation.c
75
(unsigned long)__va(pfn << PAGE_SHIFT),
fs/proc/kcore.c
149
ent->addr = (unsigned long)__va(0);
fs/proc/kcore.c
55
return __va(phys);
include/asm-generic/io.h
1107
return __va(address);
include/asm-generic/io.h
1236
return __va(addr);
include/linux/mm.h
121
#define page_to_virt(x) __va(PFN_PHYS(page_to_pfn(x)))
include/linux/mm.h
125
#define lm_alias(x) __va(__pa_symbol(x))
include/xen/arm/page.h
84
#define gfn_to_virt(m) (__va(gfn_to_pfn(m) << XEN_PAGE_SHIFT))
init/initramfs.c
642
initrd_start = (unsigned long)__va(phys_initrd_start);
init/initramfs.c
669
unsigned long crashk_start = (unsigned long)__va(crashk_res.start);
init/initramfs.c
670
unsigned long crashk_end = (unsigned long)__va(crashk_res.end);
kernel/crash_core.c
260
phdr->p_vaddr = (unsigned long) __va(mstart);
kernel/iomem.c
36
return __va(offset);
kernel/kexec_core.c
992
virt_to_page(__va(src_page_addr));
kernel/kexec_file.c
815
zero_buf = __va(page_to_pfn(ZERO_PAGE(0)) << PAGE_SHIFT);
kernel/liveupdate/kexec_handover.c
1414
void *start = __va(kho_scratch[i].addr);
mm/kmemleak.c
1600
__va((phys_addr_t)object->pointer) :
mm/memblock.c
481
new_array = __va(addr);
mm/memory_hotplug.c
1096
ret = kasan_add_zero_shadow(__va(PFN_PHYS(pfn)), PFN_PHYS(nr_pages));
mm/memory_hotplug.c
1138
kasan_remove_zero_shadow(__va(PFN_PHYS(pfn)), PFN_PHYS(nr_pages));
mm/memremap.c
104
kasan_remove_zero_shadow(__va(range->start), range_len(range));
mm/memremap.c
215
error = kasan_add_zero_shadow(__va(range->start), range_len(range));
mm/memremap.c
252
kasan_remove_zero_shadow(__va(range->start), range_len(range));
mm/memremap.c
350
return __va(pgmap->ranges[0].start);
mm/memtest.c
48
start = __va(start_phys_aligned);
mm/mm_init.c
595
set_page_address(page, __va(pfn << PAGE_SHIFT));
mm/numa.c
31
node_data[nid] = __va(nd_pa);
mm/sparse-vmemmap.c
140
return __va(__pfn_to_phys(pfn));
security/integrity/platform_certs/load_ipl_s390.c
26
ptr = __va(ipl_cert_list_addr);
tools/include/linux/mm.h
26
return __va(address);
virt/kvm/kvm_main.c
3502
const void *zero_page = (const void *) __va(page_to_phys(ZERO_PAGE(0)));