hstate_vma
unsigned long stride = huge_page_size(hstate_vma(vma));
return huge_page_shift(hstate_vma(vma));
ncontig = num_contig_ptes(huge_page_size(hstate_vma(vma)), &pgsize);
ncontig = num_contig_ptes(huge_page_size(hstate_vma(vma)), &pgsize);
unsigned long psize = huge_page_size(hstate_vma(vma));
unsigned long psize = huge_page_size(hstate_vma(vma));
unsigned long sz = huge_page_size(hstate_vma(vma));
unsigned long sz = huge_page_size(hstate_vma(vma));
unsigned long sz = huge_page_size(hstate_vma(vma));
psize = huge_page_size(hstate_vma(vma));
unsigned long psize = huge_page_size(hstate_vma(vma));
struct hstate *h = hstate_vma(vma);
vma_pageshift = huge_page_shift(hstate_vma(vma));
gfn = (gpa & huge_page_mask(hstate_vma(vma))) >> PAGE_SHIFT;
stride_size = huge_page_size(hstate_vma(vma));
hugepage_size = huge_page_size(hstate_vma(vma));
? huge_page_shift(hstate_vma(vma)) \
ptep = hugetlb_walk(vma, addr, huge_page_size(hstate_vma(vma)));
ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte);
mss->shared_hugetlb += huge_page_size(hstate_vma(vma));
mss->private_hugetlb += huge_page_size(hstate_vma(vma));
ptl = huge_pte_lock(hstate_vma(vma), walk->mm, ptep);
const unsigned long psize = huge_page_size(hstate_vma(vma));
ptl = huge_pte_lock(hstate_vma(vma), vma->vm_mm, ptep);
ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
unsigned long psize = huge_page_size(hstate_vma(vma));
unsigned long psize = huge_page_size(hstate_vma(vma));
unsigned long psize = huge_page_size(hstate_vma(vma));
addr + huge_page_size(hstate_vma(vma))))
struct hstate *h = hstate_vma(walk->vma);
struct hstate *h = hstate_vma(walk->vma);
struct hstate *h = hstate_vma(vma);
ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte);
hmm_pfn_flags_order(huge_page_order(hstate_vma(vma)));
struct hstate *h = hstate_vma(vma);
struct hstate *h = hstate_vma(vma);
if (addr & ~(huge_page_mask(hstate_vma(vma))))
return huge_page_size(hstate_vma(vma));
unsigned int shift = huge_page_shift(hstate_vma(vma));
hugetlb_count_add(pages_per_huge_page(hstate_vma(vma)), vma->vm_mm);
struct hstate *h = hstate_vma(src_vma);
struct hstate *h = hstate_vma(vma);
struct hstate *h = hstate_vma(vma);
struct hstate *h = hstate_vma(vma);
struct hstate *h = hstate_vma(vma);
struct hstate *h = hstate_vma(vma);
struct hstate *h = hstate_vma(vma);
struct hstate *h = hstate_vma(vma);
huge_page_size(hstate_vma(vma)));
struct hstate *h = hstate_vma(dst_vma);
struct hstate *h = hstate_vma(vma);
unsigned int shift = huge_page_shift(hstate_vma(vma));
unsigned long sz = huge_page_size(hstate_vma(vma));
struct hstate *h = hstate_vma(vma);
if (range->start & ~huge_page_mask(hstate_vma(vma)))
range->end = ALIGN_DOWN(range->end, huge_page_size(hstate_vma(vma)));
struct hstate *h = hstate_vma(walk->vma);
!hugepage_migration_supported(hstate_vma(vma)))
*mpol = get_vma_policy(vma, addr, hstate_vma(vma)->order, &ilx);
ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
struct hstate *h = hstate_vma(vma);
spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), vma->vm_mm, ptep);
ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
tlb_change_page_size(tlb, huge_page_size(hstate_vma(vma)));
struct hstate *h __maybe_unused = hstate_vma(vrm->vma);
struct hstate *hstate = hstate_vma(vma);
struct hstate *h = hstate_vma(vma);
hsz = huge_page_size(hstate_vma(vma));
hsz = huge_page_size(hstate_vma(vma));