root_entry
void *root_entry);
tbl_wlk.rt_entry = &iommu->root_entry[bus];
struct root_entry *rt_entry;
(u64)virt_to_phys(iommu->root_entry));
struct root_entry *old_re,
struct root_entry re;
struct root_entry *old_rt;
iommu->root_entry[bus].lo = val;
iommu->root_entry[bus].hi = val;
__iommu_flush_cache(iommu, iommu->root_entry, PAGE_SIZE);
struct root_entry *root = &iommu->root_entry[bus];
if (!iommu->root_entry)
iommu_free_pages(iommu->root_entry);
iommu->root_entry = NULL;
struct root_entry *rt_entry;
if (!iommu->root_entry) {
rt_entry = &iommu->root_entry[bus];
#define ROOT_ENTRY_NR (VTD_PAGE_SIZE/sizeof(struct root_entry))
struct root_entry *root;
iommu->root_entry = root;
static phys_addr_t root_entry_lctp(struct root_entry *re)
addr = virt_to_phys(iommu->root_entry);
static phys_addr_t root_entry_uctp(struct root_entry *re)
struct root_entry *root_entry; /* virtual address */
const struct root_entry *entry = rb_entry(node, struct root_entry, node);
const struct root_entry *new_entry = rb_entry(new, struct root_entry, node);
static struct root_entry *insert_root_entry(struct rb_root *root,
struct root_entry *re)
return rb_entry_safe(node, struct root_entry, node);
static struct root_entry *lookup_root_entry(struct rb_root *root, u64 objectid)
return rb_entry_safe(node, struct root_entry, node);
struct root_entry *re;
re = rb_entry(n, struct root_entry, node);
struct root_entry *re = NULL;
re = kzalloc_obj(struct root_entry, GFP_NOFS);
struct root_entry *exist_re;
struct root_entry *re;
struct root_entry *re;
struct root_entry *re;
re = rb_entry(n, struct root_entry, node);
struct root_entry *re = NULL;
struct root_entry *tmp;
re = kmalloc_obj(struct root_entry, GFP_NOFS);