pmd_index
pmd += pmd_index(addr);
(pmd_index(MODULES_VADDR) - pmd_index(TASK_SIZE))
#define pmd_offset_phys(dir, addr) (pud_page_paddr(READ_ONCE(*(dir))) + pmd_index(addr) * sizeof(pmd_t))
pmdp += pmd_index(addr);
pmdp += pmd_index(next) - pmd_index(addr);
k = pmd_index(vaddr);
int i_min = pmd_index(start_gpa);
int i_max = pmd_index(end_gpa);
int i_min = pmd_index(start); \
int i_max = pmd_index(end); \
k = pmd_index(vaddr);
pmd = pmd_start + pmd_index(addr);
fix_bmap_spmd = fixmap_pmd[pmd_index(__fix_to_virt(FIX_BTMAP_BEGIN))];
fix_bmap_epmd = fixmap_pmd[pmd_index(__fix_to_virt(FIX_BTMAP_END))];
pmdp = pmd_base + pmd_index(addr);
uintptr_t pmd_idx = pmd_index(va);
return (pmd_t *) pud_deref(pud) + pmd_index(address);
sto = __pa(pmdp) - pmd_index(addr) * sizeof(pmd_t);
k = pmd_index(vaddr);
for (i = 0; i < pmd_index(va_text); i++)
for (; i <= pmd_index(va_end); i++)
i = pmd_index(paddr - p2v_offset);
pmd = hv_crash_ptpgs[2] + pmd_index(addr) * sizeof(pmd);
pmd_p[pmd_index(address)] = pmd;
pmd_t *pmd = pmd_page + pmd_index(addr);
int pmd_idx = pmd_index(vaddr);
pmd_idx = pmd_index(vaddr);
pmd_idx = pmd_index(vaddr);
pmd = pmd + pmd_index(vaddr);
pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
pmd = pmd_start + pmd_index(addr);
int i = pmd_index(paddr);
pmd_t *pmd = pmd_page + pmd_index(paddr);
set_pmd(pmd + pmd_index(restore_jump_address),
set_pmd(pmd + pmd_index(restore_jump_address),
pmd_t *pmd = level2_kernel_pgt + pmd_index(vaddr);
pmd = native_make_pmd(xen_read_phys_ulong(pa + pmd_index(vaddr) *
nr = last ? pmd_index(limit) + 1 : PTRS_PER_PMD;
return pud_pgtable(*pud) + pmd_index(address);
#ifndef pmd_index
#define pmd_index pmd_index
if (pmd_index(addr) != pmd_index(saddr) ||
i = addr == -1 ? 0 : pmd_index(addr) - pmd_index(*first);
for (i = pmd_index(start), addr = start; addr != end; i++, addr = next) {
ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pmd_index(vm, vaddr) * 8;
ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pmd_index(vm, gva) * 8;