KVM_PAGES_PER_HPAGE
iterator->gfn += KVM_PAGES_PER_HPAGE(iterator->level);
gfn += KVM_PAGES_PER_HPAGE(level)) {
mask = KVM_PAGES_PER_HPAGE(fault->goal_level) - 1;
u64 page_mask = KVM_PAGES_PER_HPAGE(cur_level) -
KVM_PAGES_PER_HPAGE(cur_level - 1);
gfn_t nr_pages = KVM_PAGES_PER_HPAGE(level);
const unsigned long end = start + KVM_PAGES_PER_HPAGE(level);
for (gfn = start; gfn < end; gfn += KVM_PAGES_PER_HPAGE(level - 1)) {
gfn_t nr_pages = KVM_PAGES_PER_HPAGE(level);
gfn_t nr_pages = KVM_PAGES_PER_HPAGE(level);
return gfn & -KVM_PAGES_PER_HPAGE(level);
KVM_PAGES_PER_HPAGE(level));
__entry->pfn = fault->pfn | (fault->gfn & (KVM_PAGES_PER_HPAGE(fault->goal_level) - 1));
child_spte |= (index * KVM_PAGES_PER_HPAGE(role.level)) << PAGE_SHIFT;
iter->gfn += KVM_PAGES_PER_HPAGE(iter->level);
gfn_t end = start + KVM_PAGES_PER_HPAGE(parent->level);
gfn_t gfn = base_gfn + i * KVM_PAGES_PER_HPAGE(level);
WARN_ON_ONCE(gfn & (KVM_PAGES_PER_HPAGE(level) - 1));
pfn = pfn & ~(KVM_PAGES_PER_HPAGE(PG_LEVEL_2M) - 1);
if (slot->base_gfn & (KVM_PAGES_PER_HPAGE(level) - 1))
if ((slot->base_gfn + npages) & (KVM_PAGES_PER_HPAGE(level) - 1))
if ((slot->base_gfn ^ ugfn) & (KVM_PAGES_PER_HPAGE(level) - 1)) {