VTD_PAGE_SIZE
return ALIGN_DOWN(start, VTD_PAGE_SIZE << mask);
addr = ioremap(drhd->address, VTD_PAGE_SIZE);
addr = early_ioremap(drhd->address, VTD_PAGE_SIZE);
early_iounmap(addr, VTD_PAGE_SIZE);
VTD_PAGE_SIZE);
__iommu_flush_cache(iommu, new_ce, VTD_PAGE_SIZE);
#define ROOT_SIZE VTD_PAGE_SIZE
#define CONTEXT_SIZE VTD_PAGE_SIZE
#define ROOT_ENTRY_NR (VTD_PAGE_SIZE/sizeof(struct root_entry))
if (!IS_ALIGNED(addr, VTD_PAGE_SIZE << size_order))
#define VTD_PAGE_ALIGN(addr) (((addr) + VTD_PAGE_SIZE - 1) & VTD_PAGE_MASK)
if (!IS_ALIGNED(inv_entry.addr, VTD_PAGE_SIZE) ||
clflush_cache_range(entries, VTD_PAGE_SIZE);