huge_page_size
unsigned long stride = huge_page_size(hstate_vma(vma));
unsigned long hp_size = huge_page_size(h);
ncontig = num_contig_ptes(huge_page_size(hstate_vma(vma)), &pgsize);
ncontig = num_contig_ptes(huge_page_size(hstate_vma(vma)), &pgsize);
unsigned long psize = huge_page_size(hstate_vma(vma));
unsigned long psize = huge_page_size(hstate_vma(vma));
size_t pagesize = huge_page_size(h);
unsigned long sz = huge_page_size(hstate_vma(vma));
unsigned long sz = huge_page_size(hstate_vma(vma));
unsigned long sz = huge_page_size(hstate_vma(vma));
psize = huge_page_size(hstate_vma(vma));
unsigned long psize = huge_page_size(hstate_vma(vma));
unsigned long hp_size = huge_page_size(h);
return __hugetlb_valid_size(huge_page_size(h));
stride_size = huge_page_size(hstate_vma(vma));
hugepage_size = huge_page_size(hstate_vma(vma));
huge_page_size = hdev->asic_prop.pmmu_huge.page_size;
if (addr & (huge_page_size - 1))
va_block_align = huge_page_size;
va_block_align = huge_page_size;
huge_page_size = ctx->hdev->asic_prop.pmmu_huge.page_size;
u32 pgs_in_huge_page = huge_page_size >> __ffs(page_size);
(dma_addr & (huge_page_size - 1)))
page_size = huge_page_size;
unsigned long hpage_size = huge_page_size(sbinfo->hstate);
buf->f_bsize = huge_page_size(h);
sb->s_blocksize = huge_page_size(ctx->hstate);
huge_page_size(h) / SZ_1K);
addr0 = ALIGN(addr, huge_page_size(h));
nr = huge_page_size(h);
ptep = hugetlb_walk(vma, addr, huge_page_size(hstate_vma(vma)));
end = huge_page_size(h);
loff_t hpage_size = huge_page_size(h);
loff_t hpage_size = huge_page_size(h);
mss->shared_hugetlb += huge_page_size(hstate_vma(vma));
mss->private_hugetlb += huge_page_size(hstate_vma(vma));
const unsigned long psize = huge_page_size(hstate_vma(vma));
unsigned long _sz = huge_page_size(h); \
unsigned long psize = huge_page_size(hstate_vma(vma));
unsigned long psize = huge_page_size(hstate_vma(vma));
return huge_page_size(h) / 512;
const unsigned long size = huge_page_size(h);
hugesize = ALIGN(size, huge_page_size(hs));
unsigned long psize = huge_page_size(hstate_vma(vma));
addr + huge_page_size(hstate_vma(vma))))
*priv->folio_sz = huge_page_size(h);
if (huge_page_size(h) == size)
m = memblock_alloc_exact_nid_raw(huge_page_size(h),
huge_page_size(h), 0,
m = memblock_alloc_try_nid_raw(huge_page_size(h),
huge_page_size(h), 0,
huge_page_size(h) - PAGE_SIZE);
string_get_size(huge_page_size(h), 1, STRING_UNITS_2, buf, 32);
string_get_size(huge_page_size(h), 1, STRING_UNITS_2, buf, 32);
string_get_size(huge_page_size(h), 1, STRING_UNITS_2, buf, 32);
string_get_size(huge_page_size(&default_hstate),
h->mask = ~(huge_page_size(h) - 1);
huge_page_size(h)/SZ_1K);
total += huge_page_size(h) * count;
huge_page_size(h) / SZ_1K);
huge_page_size(h) / SZ_1K);
return huge_page_size(hstate_vma(vma));
unsigned long sz = huge_page_size(h);
unsigned long sz = huge_page_size(h);
unsigned long sz = huge_page_size(h);
address + huge_page_size(h),
huge_page_size(h));
vmf->address + huge_page_size(h));
vmf->pte = hugetlb_walk(vma, vmf->address, huge_page_size(h));
huge_page_size(h));
set_huge_pte_at(mm, vmf->address, vmf->pte, new_pte, huge_page_size(h));
vmf.pte = huge_pte_alloc(mm, vma, vmf.address, huge_page_size(h));
huge_page_size(hstate_vma(vma)));
unsigned long size = huge_page_size(h);
long pages = 0, psize = huge_page_size(h);
unsigned long sz = huge_page_size(hstate_vma(vma));
unsigned long hp_size = huge_page_size(h);
if (huge_page_size(h) == PMD_SIZE)
unsigned long sz = huge_page_size(h);
mem_fmt(buf, sizeof(buf), huge_page_size(h));
m = cma_reserve_early(cma, huge_page_size(h));
m = cma_reserve_early(cma, huge_page_size(h));
psize = huge_page_size(m->hstate);
memblock_phys_free(phys, huge_page_size(h));
range->end = ALIGN_DOWN(range->end, huge_page_size(hstate_vma(vma)));
unsigned long psize = huge_page_size(h);
len = ALIGN(len, huge_page_size(hstate_file(file)));
len = ALIGN(len, huge_page_size(hs));
tlb_change_page_size(tlb, huge_page_size(hstate_vma(vma)));
vrm->old_len = ALIGN(vrm->old_len, huge_page_size(h));
vrm->new_len = ALIGN(vrm->new_len, huge_page_size(h));
unsigned long size = huge_page_size(hstate);
unsigned long boundary = (addr & huge_page_mask(h)) + huge_page_size(h);
unsigned long sz = huge_page_size(h);
hsz = huge_page_size(hstate_vma(vma));
hsz = huge_page_size(hstate_vma(vma));
addr = addr + huge_page_size;
ret = madvise(addr - base_page_size, NR_HUGE_PAGES * huge_page_size,
ret = madvise(addr, (NR_HUGE_PAGES * huge_page_size) + base_page_size,
(void)munmap(addr, NR_HUGE_PAGES * huge_page_size);
addr = mmap(NULL, NR_HUGE_PAGES * huge_page_size,
NR_HUGE_PAGES * huge_page_size - base_page_size,
((NR_HUGE_PAGES - 1) * huge_page_size) + base_page_size,
(void)munmap(addr, NR_HUGE_PAGES * huge_page_size);
addr = mmap(NULL, NR_HUGE_PAGES * huge_page_size,
if (madvise(addr, NR_HUGE_PAGES * huge_page_size, MADV_DONTNEED)) {
(void)munmap(addr, NR_HUGE_PAGES * huge_page_size);
if (fallocate(fd, 0, 0, NR_HUGE_PAGES * huge_page_size)) {
addr = mmap(NULL, NR_HUGE_PAGES * huge_page_size,
if (madvise(addr, NR_HUGE_PAGES * huge_page_size, MADV_DONTNEED)) {
if (madvise(addr, NR_HUGE_PAGES * huge_page_size, MADV_DONTNEED)) {
0, NR_HUGE_PAGES * huge_page_size)) {
(void)munmap(addr, NR_HUGE_PAGES * huge_page_size);
if (fallocate(fd, 0, 0, NR_HUGE_PAGES * huge_page_size)) {
addr = mmap(NULL, NR_HUGE_PAGES * huge_page_size,
if (madvise(addr, NR_HUGE_PAGES * huge_page_size, MADV_DONTNEED)) {
if (madvise(addr, NR_HUGE_PAGES * huge_page_size, MADV_REMOVE)) {
(void)munmap(addr, NR_HUGE_PAGES * huge_page_size);
if (fallocate(fd, 0, 0, NR_HUGE_PAGES * huge_page_size)) {
addr = mmap(NULL, NR_HUGE_PAGES * huge_page_size,
addr2 = mmap(NULL, NR_HUGE_PAGES * huge_page_size,
if (madvise(addr, NR_HUGE_PAGES * huge_page_size, MADV_DONTNEED)) {
if (madvise(addr2, NR_HUGE_PAGES * huge_page_size, MADV_DONTNEED)) {
if (madvise(addr, NR_HUGE_PAGES * huge_page_size, MADV_REMOVE)) {
(void)munmap(addr, NR_HUGE_PAGES * huge_page_size);
(void)munmap(addr2, NR_HUGE_PAGES * huge_page_size);
unsigned long huge_page_size;
*((unsigned long *)(addr + (i * huge_page_size))) = i;
force_read_pages(addr, nr_pages, huge_page_size);
huge_page_size = default_huge_page_size();
if (!huge_page_size) {
if (!huge_page_size) {
addr = mmap(NULL, (NR_HUGE_PAGES + 2) * huge_page_size,
if (munmap(addr, huge_page_size) ||
munmap(addr + (NR_HUGE_PAGES + 1) * huge_page_size,
huge_page_size)) {
static size_t huge_page_size;
madvise(huge_ptr, huge_page_size, MADV_DONTNEED);
huge_page_size = default_huge_page_size();
if (!huge_page_size)
huge_page_size / 1024);
huge_ptr = mmap(NULL, huge_page_size, PROT_READ | PROT_WRITE,
munmap(huge_ptr, huge_page_size);
long huge_page_size;
huge_page_size = read_pmd_pagesize();
if (huge_page_size <= 0)
size_t map_size = 2 * huge_page_size;
vec.iov_len = huge_page_size;
ASSERT_EQ(ret, huge_page_size);