pud_page
free_hotplug_page_range(pud_page(pud),
extern struct page *pud_page(pud_t pud);
extern struct page *pud_page(pud_t pud);
return pud_page(pud) == virt_to_page(lm_alias(kasan_early_shadow_pmd));
struct page *page = pud_page(*pud);
free_vmemmap_storage(pud_page(pud), PUD_SIZE, altmap);
struct page *pud_page;
pud_page = alloc_page(GFP_KERNEL);
if (!pud_page)
pudp_new = (pud_t *)page_address(pud_page);
set_p4d(p4dp, pfn_p4d(page_to_pfn(pud_page), PAGE_TABLE));
void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte);
pud_t *pud = pud_page + pud_index(addr);
static int ident_pud_init(struct x86_mapping_info *info, pud_t *pud_page,
free_pagetable(pud_page(*pud), 0);
get_page_bootmem(section_nr, pud_page(*pud), MIX_SECTION_INFO);
void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte)
pud_t *pud = pud_page + pud_index(vaddr);
phys_pud_init(pud_t *pud_page, unsigned long paddr, unsigned long paddr_end,
pud = pud_page + pud_index(vaddr);
list_add(&page_ptdesc(pud_page(*pud))->pt_list, pgtables);
#define p4d_page(p4d) (pud_page((pud_t){ p4d }))
page = pud_page(orig) + ((addr & ~PUD_MASK) >> PAGE_SHIFT);
page = pud_page(orig_pud);
page = pud_page(old_pud);
pmd_free(&init_mm, (pmd_t *)page_to_virt(pud_page(*pud)));
return pud_page(pud) == virt_to_page(lm_alias(kasan_early_shadow_pmd));
if (pud_page(val) == virt_to_page(lm_alias(kasan_early_shadow_pmd)))
return pud_page(*pud) + ((addr & ~PUD_MASK) >> PAGE_SHIFT);