page_to_virt
void *reloc_code = page_to_virt(kimage->control_code_page);
const void *zero_page = page_to_virt(ZERO_PAGE(0));
void *pgtable = page_to_virt(page);
const void *zero_page = page_to_virt(ZERO_PAGE(0));
#define pfn_to_virt(pfn) page_to_virt(pfn_to_page(pfn))
unsigned long paddr = (unsigned long)page_to_virt(pte_page(pte));
tbl = page_to_virt(page);
qaddr = page_to_virt(page) + (kvm_eq.qaddr & ~PAGE_MASK);
p = page_to_virt(reuse);
imsic->swfile = page_to_virt(swfile_page);
kvm->arch.pgd = page_to_virt(pgd_page);
nacl->shmem = page_to_virt(shmem_page);
reboot_code_buffer = page_to_virt(image->control_code_page);
virt = page_to_virt(page);
aift->gait = (struct zpci_gaite *)page_to_virt(page);
__set_page_unused(page_to_virt(page), 1UL << order);
__set_page_stable_dat(page_to_virt(page), 1UL << order);
__set_page_stable_nodat(page_to_virt(page), 1UL << order);
return __set_memory((unsigned long)page_to_virt(page), 1, SET_MEMORY_INV);
return __set_memory((unsigned long)page_to_virt(page), 1, SET_MEMORY_DEF);
return __set_memory((unsigned long)page_to_virt(page), nr, flags);
address = (unsigned long)page_to_virt(page + i);
trampoline_va = (u64)page_to_virt(page);
hv_crash_ptpgs[i] = page_to_virt(page);
vmx->ve_info = page_to_virt(page);
clflush_cache_range(page_to_virt(page), PAGE_SIZE);
(pmd_val(*(pmdp)) = ((unsigned long)page_to_virt(page)))
kvaddr = (unsigned long)page_to_virt(page);
scatter = page_to_virt(pg);
*desc->paddr_ptr = __psp_pa(page_to_virt(page));
adev->gart.ptr = page_to_virt(p);
vp->vp_intercept_msg_page = page_to_virt(intercept_msg_page);
vp->vp_register_page = page_to_virt(register_page);
vp->vp_ghcb_page = page_to_virt(ghcb_page);
free_pages((unsigned long)page_to_virt(shmem_page),
mpxy->shmem = page_to_virt(shmem_page);
virt = page_to_virt(pages);
char *buffer = page_to_virt(page);
dio->integrity_payload = page_to_virt((struct page *)mempool_alloc(&ic->recheck_pool, GFP_NOIO));
dio->integrity_payload = page_to_virt((struct page *)mempool_alloc(&ic->recheck_pool, GFP_NOIO));
buffer = page_to_virt(page);
return page_to_virt(rxpage->page) + rxpage->pg_off;
intf->rx_ring_cpu = page_to_virt(buffer_pg);
addr = page_to_virt(page);
data = page_to_virt(page);
buf = page_to_virt(page);
va = page_to_virt(page);
va = page_to_virt(page) + offset;
va = page_to_virt(page);
gen_pool_free_owner(p2pdma->pool, (uintptr_t)page_to_virt(page),
cpu_hw_evt->snapshot_addr = page_to_virt(snapshot_page);
session->response = (char *)page_to_virt(page);
idaws[i] = virt_to_dma64(page_to_virt(pa->pa_page[i]));
xen_grant_dma_free(dev, size, page_to_virt(vaddr), dma_handle, 0);
kioreq->ioreq = (struct ioreq *)(page_to_virt(pages[0]));
ent->addr = (unsigned long)page_to_virt(p);
#ifndef page_to_virt
return page_to_virt(page);
trace_dma_alloc_pages(dev, page_to_virt(page), *dma_handle,
trace_dma_free_pages(dev, page_to_virt(page), dma_handle, size, dir, 0);
addr = page_to_virt(page);
ret = set_memory_decrypted((unsigned long)page_to_virt(page),
ret = set_memory_encrypted((unsigned long)page_to_virt(page),
cpu_buffer->meta_page = page_to_virt(page);
to = page_to_virt(page);
copy_page(page_to_virt(walk.reuse_page),
pte_free_kernel(&init_mm, (pte_t *)page_to_virt(pmd_page(*pmd)));
pmd_free(&init_mm, (pmd_t *)page_to_virt(pud_page(*pud)));
pud_free(&init_mm, (pud_t *)page_to_virt(p4d_page(*p4d)));
p4d_free(&init_mm, (p4d_t *)page_to_virt(pgd_page(*pgd)));
__memset(page_to_virt(page), KASAN_VMALLOC_INVALID, PAGE_SIZE);
__kfence_pool = page_to_virt(pages);
kfence_metadata_init = page_to_virt(pages);
fdt = page_to_virt(fdt_page);
addr = page_to_virt(page);
addr = page_to_virt(page);
dfrag = (struct mptcp_data_frag *)(page_to_virt(pfrag->page) + offset);
va = page_to_virt(skb_frag_page(skb_frag));