arch/alpha/kernel/pci_iommu.c
889
ptes[j] = mk_iommu_pte(page_to_phys(pages[i]));
arch/arc/mm/dma.c
30
dma_cache_wback_inv(page_to_phys(page), size);
arch/arm/include/asm/pgalloc.h
144
__pmd_populate(pmdp, page_to_phys(ptep), prot);
arch/arm/kernel/patch.c
38
set_fixmap(fixmap, page_to_phys(page));
arch/arm/mm/dma-mapping.c
1202
phys_addr_t phys = page_to_phys(sg_page(s));
arch/arm/mm/dma-mapping.c
225
page_to_phys(page),
arch/arm/mm/dma-mapping.c
589
*handle = phys_to_dma(dev, page_to_phys(page));
arch/arm/mm/dma-mapping.c
977
phys_addr_t phys = page_to_phys(pages[i]);
arch/arm64/include/asm/memory.h
404
void *__addr = __va(page_to_phys(__page)); \
arch/arm64/include/asm/pgalloc.h
120
__pmd_populate(pmdp, page_to_phys(ptep),
arch/arm64/kernel/patching.c
40
phys = page_to_phys(page) + offset_in_page(addr);
arch/arm64/kvm/mmu.c
442
return page_to_phys(vmalloc_to_page(kaddr)) +
arch/arm64/mm/mmu.c
540
pa = page_to_phys(ptdesc_page(ptdesc));
arch/csky/mm/dma-mapping.c
18
void *start = __va(page_to_phys(page));
arch/csky/mm/dma-mapping.c
55
cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range);
arch/hexagon/include/asm/page.h
123
#define page_to_virt(page) __va(page_to_phys(page))
arch/loongarch/include/asm/io.h
74
page_to_phys(tlb_virt_to_page((unsigned long)kaddr)) + offset_in_page((unsigned long)kaddr);\
arch/loongarch/include/asm/page.h
80
#define page_to_virt(page) __va(page_to_phys(page))
arch/loongarch/include/asm/page.h
90
(__kfence_pool == NULL) ? __va(page_to_phys(page)) : page_address(page); \
arch/m68k/kernel/dma.c
14
cache_push(page_to_phys(page), size);
arch/m68k/mm/cache.c
117
: : "a" (page_to_phys(page)));
arch/microblaze/mm/consistent.c
20
phys_addr_t paddr = page_to_phys(page);
arch/openrisc/kernel/patching.c
30
phys = page_to_phys(page) + offset_in_page(addr);
arch/parisc/kernel/kexec.c
80
unsigned long phys = page_to_phys(image->control_code_page);
arch/parisc/kernel/patch.c
42
set_fixmap(fixmap, page_to_phys(page));
arch/powerpc/platforms/pseries/cmm.c
112
const unsigned long vpa = page_to_phys(page);
arch/powerpc/platforms/pseries/cmm.c
92
const unsigned long vpa = page_to_phys(page);
arch/riscv/kernel/patch.c
53
phys = page_to_phys(page) + offset_in_page(addr);
arch/riscv/kernel/sbi.c
600
base_addr = page_to_phys(vmalloc_to_page(bytes)) +
arch/riscv/kernel/sbi.c
629
base_addr = page_to_phys(vmalloc_to_page(bytes)) +
arch/riscv/kvm/aia_imsic.c
1122
imsic->swfile_pa = page_to_phys(swfile_page);
arch/riscv/kvm/mmu.c
577
kvm->arch.pgd_phys = page_to_phys(pgd_page);
arch/riscv/kvm/nacl.c
148
nacl->shmem_phys = page_to_phys(shmem_page);
arch/riscv/mm/dma-noncoherent.c
123
noncoherent_cache_ops.wback_inv(page_to_phys(page), size);
arch/s390/kernel/machine_kexec.c
232
data_mover = page_to_phys(image->control_code_page);
arch/s390/kvm/pci.c
259
gaddr = page_to_phys(aibv_page) + (fib->fmt0.aibv & ~PAGE_MASK);
arch/s390/kvm/vsie.c
673
*hpa = (hpa_t)page_to_phys(page) + (gpa & ~PAGE_MASK);
arch/sparc/mm/init_64.c
3139
pfrom = page_to_phys(from);
arch/sparc/mm/init_64.c
3140
pto = page_to_phys(to);
arch/sparc/mm/init_64.c
3173
pfrom = page_to_phys(from);
arch/sparc/mm/init_64.c
3174
pto = page_to_phys(to);
arch/x86/events/intel/pt.c
698
TOPA_ENTRY(&tp->topa, 1)->base = page_to_phys(p) >> TOPA_SHIFT;
arch/x86/events/intel/pt.c
793
TOPA_ENTRY(topa, -1)->base = page_to_phys(p) >> TOPA_SHIFT;
arch/x86/hyperv/hv_crash.c
556
trampoline_pa = (u32)page_to_phys(page);
arch/x86/include/asm/tdx.h
185
ret = page_to_phys(page);
arch/x86/kvm/vmx/tdx.c
2900
tdx->vp.tdvpr_pa = page_to_phys(tdx->vp.tdvpr_page);
arch/x86/kvm/vmx/vmx.c
4067
const void *zero_page = (const void *) __va(page_to_phys(ZERO_PAGE(0)));
arch/x86/kvm/vmx/vmx.c
5029
vmcs_write64(PML_ADDRESS, page_to_phys(vmx->pml_pg));
arch/x86/platform/efi/efi_64.c
239
efi_mixed_mode_stack_pa = page_to_phys(page + 1); /* stack grows down */
arch/x86/virt/vmx/tdx/tdx.c
1505
return page_to_phys(td->tdr_page);
arch/x86/virt/vmx/tdx/tdx.c
1530
.rcx = page_to_phys(tdcs_page),
arch/x86/virt/vmx/tdx/tdx.c
1544
.r8 = page_to_phys(page),
arch/x86/virt/vmx/tdx/tdx.c
1545
.r9 = page_to_phys(source),
arch/x86/virt/vmx/tdx/tdx.c
1564
.r8 = page_to_phys(page),
arch/x86/virt/vmx/tdx/tdx.c
1581
.rcx = page_to_phys(tdcx_page),
arch/x86/virt/vmx/tdx/tdx.c
1595
.r8 = page_to_phys(page),
arch/x86/virt/vmx/tdx/tdx.c
1801
.rcx = page_to_phys(page),
arch/x86/virt/vmx/tdx/tdx.c
664
tdx_quirk_reset_paddr(page_to_phys(page), PAGE_SIZE);
arch/xtensa/include/asm/highmem.h
34
return DCACHE_ALIAS(page_to_phys(page));
arch/xtensa/mm/cache.c
204
unsigned long phys = page_to_phys(pfn_to_page(pfn));
arch/xtensa/mm/cache.c
277
unsigned long phys = page_to_phys(page);
arch/xtensa/mm/cache.c
317
unsigned long phys = page_to_phys(page);
arch/xtensa/mm/cache.c
62
if (!DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) {
arch/xtensa/mm/cache.c
71
(page_to_phys(page) & DCACHE_ALIAS_MASK);
arch/xtensa/mm/cache.c
75
page_to_phys(page));
arch/xtensa/mm/cache.c
84
*paddr = page_to_phys(page);
block/bio.c
1002
phys_addr_t addr2 = page_to_phys(page) + offset + len - 1;
block/bio.c
970
phys_addr_t vec_end_addr = page_to_phys(bv->bv_page) + bv_end - 1;
block/bio.c
971
phys_addr_t page_addr = page_to_phys(page);
drivers/accel/habanalabs/common/device.c
210
page_to_phys(sg_page(sg)),
drivers/accel/habanalabs/common/device.c
253
page_to_phys(sg_page(sg)),
drivers/block/swim3.c
419
return page_to_phys(bio_page(bio)) + bio_offset(bio);
drivers/cdrom/gdrom.c
585
__raw_writel(page_to_phys(bio_page(req->bio)) + bio_offset(req->bio),
drivers/char/agp/ali-agp.c
155
page_to_phys(page)) | ALI_CACHE_FLUSH_EN ));
drivers/char/agp/ali-agp.c
183
page_to_phys(page)) | ALI_CACHE_FLUSH_EN));
drivers/char/agp/amd-k7-agp.c
317
page_to_phys(mem->pages[i]),
drivers/char/agp/amd64-agp.c
83
page_to_phys(mem->pages[i]),
drivers/char/agp/ati-agp.c
297
page_to_phys(mem->pages[i]),
drivers/char/agp/backend.c
156
bridge->scratch_page_dma = page_to_phys(page);
drivers/char/agp/efficeon-agp.c
71
unsigned long addr = page_to_phys(page);
drivers/char/agp/generic.c
1095
page_to_phys(mem->pages[i]),
drivers/char/agp/intel-gtt.c
274
new->physical = page_to_phys(new->pages[0]);
drivers/char/agp/intel-gtt.c
316
intel_private.scratch_page_dma = page_to_phys(page);
drivers/char/agp/intel-gtt.c
924
dma_addr_t addr = page_to_phys(pages[i]);
drivers/char/agp/nvidia-agp.c
229
page_to_phys(mem->pages[i]), mask_type),
drivers/char/agp/parisc-agp.c
157
paddr = page_to_phys(mem->pages[i]);
drivers/char/agp/sworks-agp.c
350
page_to_phys(mem->pages[i]), mem->type),
drivers/char/agp/uninorth-agp.c
144
page_to_phys(agp_bridge->scratch_page_page) >> 12);
drivers/char/agp/uninorth-agp.c
187
gp[i] = (page_to_phys(mem->pages[i]) >> PAGE_SHIFT) | 0x80000000UL;
drivers/char/agp/uninorth-agp.c
189
gp[i] = cpu_to_le32((page_to_phys(mem->pages[i]) & 0xFFFFF000UL) |
drivers/char/agp/uninorth-agp.c
191
flush_dcache_range((unsigned long)__va(page_to_phys(mem->pages[i])),
drivers/char/agp/uninorth-agp.c
192
(unsigned long)__va(page_to_phys(mem->pages[i]))+0x1000);
drivers/char/agp/uninorth-agp.c
431
scratch_value = (page_to_phys(agp_bridge->scratch_page_page) >> PAGE_SHIFT) | 0x80000000UL;
drivers/char/agp/uninorth-agp.c
433
scratch_value = cpu_to_le32((page_to_phys(agp_bridge->scratch_page_page) & 0xFFFFF000UL) |
drivers/crypto/ccp/sev-dev-tio.c
51
u64 pa = __sme_set(page_to_phys(pg));
drivers/crypto/nx/nx-842.h
110
return page_to_phys(vmalloc_to_page(addr)) + offset_in_page(addr);
drivers/crypto/nx/nx.c
121
sg_addr = page_to_phys(vmalloc_to_page(
drivers/crypto/nx/nx.c
91
sg_addr = page_to_phys(vmalloc_to_page(start_addr))
drivers/edac/thunderx_edac.c
339
phys = (uintptr_t)page_to_phys(lmc->mem);
drivers/firmware/efi/capsule-loader.c
193
cap_info->phys[cap_info->index] = page_to_phys(page);
drivers/firmware/efi/capsule.c
172
sglist_phys = page_to_phys(sg_pages[0]);
drivers/firmware/efi/capsule.c
267
sglist[j].data = page_to_phys(sg_pages[i + 1]);
drivers/gpu/drm/armada/armada_gem.c
103
obj->phys_addr = page_to_phys(p);
drivers/gpu/drm/drm_fbdev_dma.c
219
info->fix.smem_start = page_to_phys(virt_to_page(info->screen_buffer));
drivers/gpu/drm/etnaviv/etnaviv_dump.c
217
*bomap++ = cpu_to_le64(page_to_phys(*pages++));
drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
792
char new_bit_17 = page_to_phys(page) >> 17;
drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
833
if (page_to_phys(page) & (1 << 17))
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
88
sg_dma_address(sg) = page_to_phys(sg_page(sg));
drivers/gpu/drm/msm/msm_iommu.c
139
phys_addr_t phys = page_to_phys(iommu->prr_page);
drivers/gpu/drm/msm/msm_iommu.c
606
page_to_phys(iommu->prr_page));
drivers/gpu/drm/nouveau/nouveau_svm.c
578
args->p.phys[0] = page_to_phys(page) |
drivers/gpu/drm/nouveau/nouveau_svm.c
635
args->p.phys[0] = page_to_phys(page) |
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
382
page_to_phys(pages[n]) : engine->dmm->dummy_pa;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
875
omap_dmm->dummy_pa = page_to_phys(omap_dmm->dummy_page);
drivers/gpu/drm/panfrost/panfrost_dump.c
224
*bomap++ = page_to_phys(sg_page_iter_page(&page_iter));
drivers/infiniband/hw/hfi1/user_exp_rcv.c
740
phys = dma_map_single(&dd->pcidev->dev, __va(page_to_phys(pages[0])),
drivers/infiniband/hw/hfi1/user_exp_rcv.c
751
node->phys = page_to_phys(pages[0]);
drivers/infiniband/sw/rxe/rxe_mr.c
119
unsigned long paddr = page_to_phys(pg);
drivers/iommu/dma-iommu.c
1659
*handle = __iommu_dma_map(dev, page_to_phys(page), size, ioprot,
drivers/iommu/fsl_pamu.c
839
ppaact_phys = page_to_phys(p);
drivers/irqchip/irq-gic-v3-its.c
2343
gic_rdists->prop_table_pa = page_to_phys(page);
drivers/irqchip/irq-gic-v3-its.c
2886
table[idx] = cpu_to_le64(page_to_phys(page) | GITS_BASER_VALID);
drivers/irqchip/irq-gic-v3-its.c
3185
paddr = page_to_phys(pend_page);
drivers/irqchip/irq-gic-v3-its.c
3216
val = (page_to_phys(pend_page) |
drivers/irqchip/irq-gic-v3-its.c
3398
table[idx] = cpu_to_le64(page_to_phys(page) | GITS_BASER_VALID);
drivers/irqchip/irq-gic-v3-its.c
5470
phys_addr_t paddr = page_to_phys(pend_page);
drivers/mailbox/riscv-sbi-mpxy-mbox.c
358
page_to_phys(shmem_page), 0, 0, 0, 0, 0);
drivers/mailbox/riscv-sbi-mpxy-mbox.c
366
mpxy->shmem_phys_addr = page_to_phys(shmem_page);
drivers/misc/sgi-gru/grufault.c
190
*paddr = page_to_phys(page);
drivers/net/ethernet/intel/e1000/e1000_main.c
2919
page_to_phys(skb_frag_page(frag));
drivers/net/ethernet/sun/niu.c
10025
return page_to_phys(page) + offset;
drivers/net/fjes/fjes_hw.c
1233
(__le64)(page_to_phys(vmalloc_to_page(addr)) +
drivers/net/fjes/fjes_hw.c
533
(__le64)(page_to_phys(vmalloc_to_page(addr)) +
drivers/net/fjes/fjes_hw.c
543
(__le64)(page_to_phys(vmalloc_to_page(addr)) +
drivers/net/fjes/fjes_trace.h
114
__entry->tx = page_to_phys(vmalloc_to_page(tx)) +
drivers/net/fjes/fjes_trace.h
116
__entry->rx = page_to_phys(vmalloc_to_page(rx)) +
drivers/perf/riscv_pmu_sbi.c
686
cpu_hw_evt->snapshot_addr_phys = page_to_phys(snapshot_page);
drivers/platform/goldfish/goldfish_pipe.c
299
unsigned long xaddr = page_to_phys(pages[0]);
drivers/platform/goldfish/goldfish_pipe.c
309
xaddr = page_to_phys(pages[i]);
drivers/remoteproc/remoteproc_core.c
151
return page_to_phys(vmalloc_to_page(cpu_addr)) +
drivers/s390/crypto/vfio_ap_ops.c
457
h_nib = page_to_phys(h_page) | (nib & ~PAGE_MASK);
drivers/s390/net/qeth_core_main.c
2962
page_to_phys(pool_entry->elements[i]));
drivers/staging/media/atomisp/pci/hmm/hmm.c
431
return page_to_phys(bo->pages[idx]) + offset;
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
782
page_to_phys(bo->pages[i]), 1);
drivers/staging/media/ipu3/ipu3-dmamap.c
120
page_to_phys(pages[i]), PAGE_SIZE);
drivers/staging/media/ipu3/ipu3-mmu.c
309
phys_addr_t phys = page_to_phys(sg_page(s)) + s->offset;
drivers/tee/optee/smc_abi.c
430
optee_page = page_to_phys(*pages) +
drivers/tee/optee/smc_abi.c
448
optee_page = page_to_phys(*pages);
drivers/tee/tee_shm.c
306
dma_mem->shm.paddr = page_to_phys(page);
drivers/vdpa/vdpa_user/iova_domain.c
454
phys_addr_t pa = page_to_phys(page) + offset;
drivers/vfio/pci/mlx5/cmd.c
394
page_to_phys(page_list[i]), mapped,
drivers/virt/acrn/hsm.c
354
ret = hcall_vm_intr_monitor(vm->vmid, page_to_phys(page));
drivers/virt/acrn/ioreq.c
619
set_buffer->ioreq_buf = page_to_phys(page);
drivers/virt/acrn/mm.c
309
vm_region->service_vm_pa = page_to_phys(page);
drivers/virt/fsl_hypervisor.c
263
sg_list[0].source = page_to_phys(pages[0]) + lb_offset;
drivers/virt/fsl_hypervisor.c
267
sg_list[0].target = page_to_phys(pages[0]) + lb_offset;
drivers/virt/fsl_hypervisor.c
277
sg_list[i].source = page_to_phys(pages[i]);
drivers/virt/fsl_hypervisor.c
282
sg_list[i].target = page_to_phys(pages[i]);
drivers/virt/nitro_enclaves/ne_misc_dev.c
980
page_to_phys(ne_mem_region->pages[i]),
drivers/virt/vboxguest/vboxguest_core.c
294
req->phys_page[i] = page_to_phys(pages[i]);
drivers/virt/vboxguest/vboxguest_core.c
335
req->phys_page[i] = page_to_phys(pages[i]);
drivers/virt/vboxguest/vboxguest_utils.c
367
dst_pg_lst->pages[i] = page_to_phys(page);
drivers/virtio/virtio_mem.c
1302
const unsigned long start = page_to_phys(page);
drivers/virtio/virtio_mem.c
1360
const unsigned long addr = page_to_phys(page);
fs/btrfs/disk-io.c
197
paddrs[i] = page_to_phys(&folio->page) + offset_in_page(eb->start);
fs/btrfs/inode.c
3477
paddrs[i] = page_to_phys(folio_page(folio, pindex)) + offset_in_page(paddr);
fs/btrfs/inode.c
889
paddr = page_to_phys(bvec->bv_page) + bvec->bv_offset + bvec->bv_len - 1;
fs/btrfs/raid56.c
2753
p_paddr = page_to_phys(page);
fs/btrfs/raid56.c
2765
q_paddr = page_to_phys(page);
fs/btrfs/raid56.c
370
rbio->stripe_paddrs[i] = page_to_phys(rbio->stripe_pages[page_index]) +
fs/btrfs/scrub.c
709
return page_to_phys(folio_page(folio, 0)) + offset_in_folio(folio, offset);
include/linux/bvec.h
308
return page_to_phys(bvec->bv_page) + bvec->bv_offset;
include/linux/hyperv.h
1756
paddr = page_to_phys(vmalloc_to_page(addr)) +
include/linux/scatterlist.h
412
return page_to_phys(sg_page(sg)) + sg->offset;
kernel/dma/debug.c
1413
return page_to_phys(page) + offset_in_page(virt);
kernel/dma/debug.c
1583
entry->paddr = page_to_phys(page);
kernel/dma/debug.c
1598
.paddr = page_to_phys(page),
kernel/dma/direct.c
111
if (page && !dma_coherent_ok(dev, page_to_phys(page), size)) {
kernel/dma/direct.c
134
if (dma_coherent_ok(dev, page_to_phys(page), size) &&
kernel/dma/direct.c
142
&& !dma_coherent_ok(dev, page_to_phys(page), size)) {
kernel/dma/direct.c
181
*dma_handle = phys_to_dma_direct(dev, page_to_phys(page));
kernel/dma/direct.c
199
*dma_handle = phys_to_dma_direct(dev, page_to_phys(page));
kernel/dma/direct.c
297
*dma_handle = phys_to_dma_direct(dev, page_to_phys(page));
kernel/dma/direct.c
370
*dma_handle = phys_to_dma_direct(dev, page_to_phys(page));
kernel/dma/mapping.c
191
phys_addr_t phys = page_to_phys(page) + offset;
kernel/dma/ops_helpers.c
75
phys = page_to_phys(page);
kernel/dma/pool.c
120
ret = gen_pool_add_virt(pool, (unsigned long)addr, page_to_phys(page),
kernel/dma/swiotlb.c
1795
phys_addr_t tlb_addr = page_to_phys(page);
kernel/dma/swiotlb.c
584
paddr = page_to_phys(page);
kernel/dma/swiotlb.c
718
swiotlb_init_io_tlb_pool(pool, page_to_phys(tlb), nslabs, true, nareas);
kernel/events/core.c
8398
phys_addr = page_to_phys(p) + virt % PAGE_SIZE;
kernel/liveupdate/kexec_handover.c
1016
phys_addr_t phys = page_to_phys(vm->pages[i]);
mm/page_alloc.c
1656
accept_memory(page_to_phys(page), PAGE_SIZE << order);
mm/page_alloc.c
7628
phys_addr_t start = page_to_phys(page);
mm/page_alloc.c
7642
accept_memory(page_to_phys(page), PAGE_SIZE << MAX_PAGE_ORDER);
mm/percpu.c
2391
return page_to_phys(vmalloc_to_page(addr)) +
mm/percpu.c
2394
return page_to_phys(pcpu_addr_to_page(addr)) +
mm/vmalloc.c
674
page_to_phys(pages[i]), prot,
net/bpf/test_run.c
122
phys_to_virt(page_to_phys(netmem_to_page(netmem)));
net/bpf/test_run.c
301
head = phys_to_virt(page_to_phys(page));
sound/core/memalloc.c
296
*addr = page_to_phys(virt_to_page(p));
sound/core/memalloc.c
371
page_to_phys(vmalloc_to_page((dmab)->area + (offset)))
sound/usb/qcom/qc_audio_offload.c
1138
tr_pa = page_to_phys(pg);
sound/usb/qcom/qc_audio_offload.c
1207
er_pa = page_to_phys(pg);
sound/usb/qcom/qc_audio_offload.c
620
pa_sg = page_to_phys(sg_page(sg));
sound/virtio/virtio_pcm_msg.c
104
sg_phys(&sgs[idx]) + sgs[idx].length != page_to_phys(pg)) {
sound/virtio/virtio_pcm_msg.c
56
phys_addr_t pg_address = page_to_phys(pg);
tools/virtio/linux/dma-mapping.h
24
#define dma_map_page(d, p, o, s, dir) (page_to_phys(p) + (o))
tools/virtio/linux/dma-mapping.h
25
#define dma_map_page_attrs(d, p, o, s, dir, a) (page_to_phys(p) + (o))
tools/virtio/linux/scatterlist.h
158
return page_to_phys(sg_page(sg)) + sg->offset;
virt/kvm/kvm_main.c
3502
const void *zero_page = (const void *) __va(page_to_phys(ZERO_PAGE(0)));