fs/hugetlbfs/inode.c
411
start = index * pages_per_huge_page(h);
fs/hugetlbfs/inode.c
412
end = (index + 1) * pages_per_huge_page(h);
ipc/shm.c
944
*rss_add += pages_per_huge_page(h) * mapping->nrpages;
mm/hugetlb.c
1100
resv_map->pages_per_hpage = pages_per_huge_page(h);
mm/hugetlb.c
1781
pages_per_huge_page(h), folio);
mm/hugetlb.c
1783
pages_per_huge_page(h), folio);
mm/hugetlb.c
1784
lruvec_stat_mod_folio(folio, NR_HUGETLB, -pages_per_huge_page(h));
mm/hugetlb.c
2968
idx, pages_per_huge_page(h), &h_cg);
mm/hugetlb.c
2973
ret = hugetlb_cgroup_charge_cgroup(idx, pages_per_huge_page(h), &h_cg);
mm/hugetlb.c
3004
hugetlb_cgroup_commit_charge(idx, pages_per_huge_page(h), h_cg, folio);
mm/hugetlb.c
3009
hugetlb_cgroup_commit_charge_rsvd(idx, pages_per_huge_page(h),
mm/hugetlb.c
3037
hstate_index(h), pages_per_huge_page(h), folio);
mm/hugetlb.c
3048
lruvec_stat_mod_folio(folio, NR_HUGETLB, pages_per_huge_page(h));
mm/hugetlb.c
3058
hugetlb_cgroup_uncharge_cgroup(idx, pages_per_huge_page(h), h_cg);
mm/hugetlb.c
3061
hugetlb_cgroup_uncharge_cgroup_rsvd(idx, pages_per_huge_page(h),
mm/hugetlb.c
3233
unsigned long nr_pages = pages_per_huge_page(h), i;
mm/hugetlb.c
3265
pages_per_huge_page(h));
mm/hugetlb.c
3297
pages_per_huge_page(m->hstate));
mm/hugetlb.c
3316
unsigned long npages = pages_per_huge_page(h);
mm/hugetlb.c
3390
adjust_managed_page_count(page, pages_per_huge_page(h));
mm/hugetlb.c
4004
for (i = 0; i < pages_per_huge_page(src); i += pages_per_huge_page(dst)) {
mm/hugetlb.c
4639
nr_total_pages += h->nr_huge_pages * pages_per_huge_page(h);
mm/hugetlb.c
4881
hugetlb_count_add(pages_per_huge_page(hstate_vma(vma)), vma->vm_mm);
mm/hugetlb.c
4895
unsigned long npages = pages_per_huge_page(h);
mm/hugetlb.c
510
resv->pages_per_hpage = pages_per_huge_page(h);
mm/hugetlb.c
514
VM_BUG_ON(resv->pages_per_hpage != pages_per_huge_page(h));
mm/hugetlb.c
5272
hugetlb_count_sub(pages_per_huge_page(h), mm);
mm/hugetlb.c
5897
hugetlb_count_add(pages_per_huge_page(h), mm);
mm/hugetlb.c
6389
hugetlb_count_add(pages_per_huge_page(h), dst_mm);
mm/hugetlb.c
6633
chg * pages_per_huge_page(h), &h_cg);
mm/hugetlb.c
6697
(chg - add) * pages_per_huge_page(h), h_cg);
mm/hugetlb.c
6736
chg * pages_per_huge_page(h), h_cg);
mm/hugetlb_cgroup.c
119
pages_per_huge_page(&hstates[idx]));
mm/hugetlb_cgroup.c
554
pages_per_huge_page(&hstates[idx]));
mm/hugetlb_cgroup.c
601
nr_pages = round_down(nr_pages, pages_per_huge_page(&hstates[idx]));
mm/hugetlb_vmemmap.c
673
epfn = spfn + pages_per_huge_page(h);
mm/hugetlb_vmemmap.c
806
nr_pages = pages_per_huge_page(m->hstate);
mm/hugetlb_vmemmap.c
852
nr_pages = pages_per_huge_page(h);
mm/hugetlb_vmemmap.h
38
return pages_per_huge_page(h) * sizeof(struct page);
mm/page_vma_mapped.c
211
if (!check_pte(pvmw, pages_per_huge_page(hstate)))