XEN_PAGE_SHIFT
if (likely(set_phys_to_machine(map_ops[i].host_addr >> XEN_PAGE_SHIFT,
map_ops[i].dev_bus_addr >> XEN_PAGE_SHIFT)))
set_phys_to_machine(unmap_ops[i].host_addr >> XEN_PAGE_SHIFT,
info->intf = memremap(gfn << XEN_PAGE_SHIFT, XEN_PAGE_SIZE, MEMREMAP_WB);
static_assert(XEN_PAGE_SHIFT == EFI_PAGE_SHIFT,
return XEN_GRANT_DMA_ADDR_OFF | ((dma_addr_t)grant << XEN_PAGE_SHIFT);
return (grant_ref_t)((dma & ~XEN_GRANT_DMA_ADDR_OFF) >> XEN_PAGE_SHIFT);
int numgfns = (vma->vm_end - vma->vm_start) >> XEN_PAGE_SHIFT;
phys_addr_t paddr = (phys_addr_t)xen_pfn << XEN_PAGE_SHIFT;
phys_addr_t baddr = (phys_addr_t)bfn << XEN_PAGE_SHIFT;
phys_addr_t paddr = (xen_pfn << XEN_PAGE_SHIFT) |
phys_addr_t bus_addr = pfn_to_bfn(XEN_PFN_DOWN(p)) << XEN_PAGE_SHIFT;
memremap(xen_store_gfn << XEN_PAGE_SHIFT,
xen_store_interface = memremap(xen_store_gfn << XEN_PAGE_SHIFT,
pfn_to_gfn(virt_to_phys(v) >> XEN_PAGE_SHIFT); \
#define gfn_to_virt(m) (__va(gfn_to_pfn(m) << XEN_PAGE_SHIFT))
(pfn_to_gfn(per_cpu_ptr_to_phys(v) >> XEN_PAGE_SHIFT))
#ifndef XEN_PAGE_SHIFT
(1UL << ((order) + XEN_PAGE_SHIFT - 1))
(pfn_to_page((unsigned long)(xen_pfn) >> (PAGE_SHIFT - XEN_PAGE_SHIFT)))
((page_to_pfn(page)) << (PAGE_SHIFT - XEN_PAGE_SHIFT))
#define XEN_PFN_DOWN(x) ((x) >> XEN_PAGE_SHIFT)
#define XEN_PFN_UP(x) (((x) + XEN_PAGE_SIZE-1) >> XEN_PAGE_SHIFT)
#define XEN_PAGE_SIZE (_AC(1, UL) << XEN_PAGE_SHIFT)
.maxsize = 1 << (XEN_9PFS_RING_ORDER + XEN_PAGE_SHIFT - 2),
XEN_PAGE_SHIFT));
bytes = alloc_pages_exact(1UL << (order + XEN_PAGE_SHIFT),
free_pages_exact(bytes, 1UL << (order + XEN_PAGE_SHIFT));