set_pmd
set_pmd(pmd, __pmd((unsigned long)pte));
set_pmd(pmd, __pmd((unsigned long)page_address(pte_page)));
set_pmd(pmd, *pmd_k);
set_pmd(pmdp, __pmd(__phys_to_pmd_val(ptep) | prot));
set_pmd(pmdp, __pmd(0));
set_pmd(pmdp, new_pmd);
set_pmd(pmdp, pmd_mknoncont(pmdp_get(pmdp)));
set_pmd(pmdp, pfn_pmd(pfn, prot));
set_pmd(pmd, val);
set_pmd(dst_pmdp,
set_pmd(pmd, __pmd(__pa(pte)));
set_pmd(pmd, __pmd(__pa(page_address(pte))));
set_pmd(pmd, *pmd_k);
set_pmd(pmd, __pmd(__pa(pte)));
set_pmd(pmd, __pmd(((unsigned long)page_to_pfn(pte) << PAGE_SHIFT) |
set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
set_pmd(pmd, __pmd((unsigned long)pte));
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
set_pmd(pmdp, __pmd((unsigned long)invalid_pte_table));
set_pmd(pmd, val);
set_pmd(pmd, __pmd((unsigned long)pte));
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
set_pmd(pmd, *pmd_k);
set_pmd(pmd, __pmd((unsigned long)pte));
set_pmd(pmd, __pmd((unsigned long)pte));
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
set_pmd(pmd, *pmd_k);
set_pmd(pmd, __pmd(_KERNPG_TABLE + __pa(pte)))
set_pmd(pmd, __pmd(_KERNPG_TABLE +
set_pmd(pmd, *pmd_k);
set_pmd(pme, __pmd(_KERNPG_TABLE + __pa(pte)));
set_pmd(pmd, __pmd((PxD_FLAG_PRESENT | PxD_FLAG_VALID)
set_pmd(pmd, __pmd(0));
set_pmd(pmd, __pmd((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE));
set_pmd(pmd, __pmd((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE));
set_pmd(pmdp, __pmd(0));
set_pmd(dst_pmdp, __pmd(pmd_val(pmd) | pgprot_val(prot)));
set_pmd(kasan_early_shadow_pmd + i,
set_pmd(pmd, pfn_pmd(PFN_DOWN(__pa(p)), PAGE_TABLE));
set_pmd(pmdp, pfn_pmd(PFN_DOWN(phys_addr), PAGE_KERNEL));
set_pmd(pmdp, pfn_pmd(page_to_pfn(pte_page), PAGE_TABLE));
set_pmd(pmdp_new,
set_pmd(pmd, val);
set_pmd(pmd, new_pmd);
set_pmd(pmd, entry);
set_pmd(pmd, __pmd(_SEGMENT_ENTRY | __pa(pte)));
set_pmd(pmdp, __pmd(_SEGMENT_ENTRY_EMPTY));
set_pmd(pmdp, entry);
set_pmd(pmdp, __pmd(_SEGMENT_ENTRY_EMPTY));
set_pmd(pmdp, __pmd(pmd_addr | prot));
set_pmd(pmdp, set_pmd_bit(*pmdp, __pgprot(_SEGMENT_ENTRY_INVALID)));
set_pmd(pmdp, new);
set_pmd(pmdp, new);
set_pmd(pmd, __pmd(__pa(addr) | prot));
set_pmd(pmd, __pmd(__pa(new_page) | prot));
set_pmd(pmd, __pmd((unsigned long)pte));
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
#define pmd_clear(xp) do { set_pmd(xp, __pmd(0)); } while (0)
set_pmd(pmd, *pmd_k);
set_pmd(pmd, __pmd(_PAGE_TABLE + (unsigned long) __pa(pte)))
set_pmd(pmd, __pmd(_PAGE_TABLE + \
set_pmd(pmdp, pmd);
set_pmd(pmd, __pmd(ppd->paddr | ppd->pmd_flags));
set_pmd(pmd, __pmd(PMD_FLAGS | __pa(pte)));
set_pmd(pmd, __pmd(pmd_val(*pmd) | _PAGE_USER));
set_pmd(pmd, __pmd(_PAGE_TABLE | pa));
PVOP_VCALL2(pv_ops, mmu.set_pmd, pmdp, native_pmd_val(pmd));
set_pmd(pmdp, native_make_pmd(0));
void (*set_pmd)(pmd_t *pmdp, pmd_t pmdval);
set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE));
set_pmd(pmdp, pmd);
set_pmd(pmdp, pmd); \
set_pmd(&pmd_p[n], pmd);
set_pmd(u_pmd, *k_pmd);
set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
.mmu.set_pmd = native_set_pmd,
set_pmd(pmd, *pmd_k);
set_pmd(pmd, __pmd((addr - info->offset) | info->page_flag));
set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE));
set_pmd(pmd, __pmd(__pa(newpte)|_PAGE_TABLE));
set_pmd(pmd, pfn_pmd(pfn, init_prot));
set_pmd(pmd, pfn_pmd(pfn, prot));
set_pmd(pmd, __pmd(pte_val(entry)));
set_pmd(pmd, __pmd(phys | pgprot_val(prot)));
set_pmd(pmd, __pmd(0));
set_pmd(pmd, new_pmd);
set_pmd(pmd, _pmd);
set_pmd(pmd, _pmd);
set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
set_pmd(pmd, pmd_mkhuge(pfn_pmd(cpa->pfn,
set_pmd(pmdp, entry);
set_pmd(pmd, __pmd(0));
set_pmd(pmd, __pmd(_KERNPG_TABLE | __pa(new_pte_page)));
set_pmd(pmd, __pmd(pmd_val(*pmd) & ~_PAGE_NX));
set_pmd(pmd, pfn_pmd(pfn, PAGE_KERNEL_LARGE_EXEC));
set_pmd(pmd + pmd_index(restore_jump_address),
set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE));
set_pmd(pmd + pmd_index(restore_jump_address),
set_pmd(pmd, __pmd(0));
set_pmd(pmd, __pmd(0));
pv_ops.mmu.set_pmd = xen_set_pmd;
pv_ops.mmu.set_pmd = xen_set_pmd_hyper;
set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE));
set_pmd(pmdp,
#define pmd_clear(pmdp) do { set_pmd(pmdp, __pmd(0)); } while (0)
set_pmd(pmd, __pmd((unsigned long)kasan_early_shadow_pte));
set_pmd(pmd + i, __pmd((unsigned long)pte));
set_pmd(pmd, __pmd(((unsigned long)cur_pte) & PAGE_MASK));
#define set_pud(pudptr, pudval) set_pmd((pmd_t *)(pudptr), (pmd_t) { pudval })
goto set_pmd;
set_pmd: