Symbol: HPAGE_PMD_NR
arch/arc/mm/tlb.c
534
update_mmu_cache_range(NULL, vma, addr, &pte, HPAGE_PMD_NR);
drivers/base/node.c
612
pages /= HPAGE_PMD_NR;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
472
pages_per_block = HPAGE_PMD_NR;
drivers/gpu/drm/nouveau/nouveau_dmem.c
848
if (max > (unsigned long)HPAGE_PMD_NR)
drivers/gpu/drm/nouveau/nouveau_dmem.c
849
max = (unsigned long)HPAGE_PMD_NR;
include/linux/huge_mm.h
537
return READ_ONCE(huge_zero_pfn) == (pfn & ~(HPAGE_PMD_NR - 1));
include/linux/rmap.h
362
VM_WARN_ON_FOLIO(folio_nr_pages(folio) != HPAGE_PMD_NR, folio);
include/linux/rmap.h
363
VM_WARN_ON_FOLIO(nr_pages != HPAGE_PMD_NR, folio);
include/linux/rmap.h
561
__folio_dup_file_rmap(folio, page, HPAGE_PMD_NR, dst_vma, PGTABLE_LEVEL_PTE);
include/linux/rmap.h
700
return __folio_try_dup_anon_rmap(folio, page, HPAGE_PMD_NR, dst_vma,
include/linux/rmap.h
834
return __folio_try_share_anon_rmap(folio, page, HPAGE_PMD_NR,
lib/test_hmm.c
580
if (is_large && IS_ALIGNED(pfn, HPAGE_PMD_NR)
lib/test_hmm.c
581
&& (pfn + HPAGE_PMD_NR <= pfn_last)) {
lib/test_hmm.c
584
pfn += HPAGE_PMD_NR;
lib/test_hmm.c
602
mdevice->calloc += HPAGE_PMD_NR;
lib/test_hmm.c
721
nr = HPAGE_PMD_NR;
mm/filemap.c
3355
ractl._index &= ~((unsigned long)HPAGE_PMD_NR - 1);
mm/filemap.c
3356
ra->size = HPAGE_PMD_NR;
mm/filemap.c
3363
ra->async_size = HPAGE_PMD_NR;
mm/gup.c
741
*page_mask = HPAGE_PMD_NR - 1;
mm/huge_memory.c
1317
add_mm_counter(vma->vm_mm, MM_ANONPAGES, HPAGE_PMD_NR);
mm/huge_memory.c
1570
add_mm_counter(mm, mm_counter_file(fop.folio), HPAGE_PMD_NR);
mm/huge_memory.c
1841
add_mm_counter(dst_mm, MM_ANONPAGES, HPAGE_PMD_NR);
mm/huge_memory.c
1935
add_mm_counter(dst_mm, MM_ANONPAGES, HPAGE_PMD_NR);
mm/huge_memory.c
2236
task_numa_fault(last_cpupid, nid, HPAGE_PMD_NR, flags);
mm/huge_memory.c
2257
task_numa_fault(last_cpupid, nid, HPAGE_PMD_NR, flags);
mm/huge_memory.c
2396
add_mm_counter(tlb->mm, MM_ANONPAGES, -HPAGE_PMD_NR);
mm/huge_memory.c
2401
-HPAGE_PMD_NR);
mm/huge_memory.c
2641
ret = HPAGE_PMD_NR;
mm/huge_memory.c
283
return atomic_read(&huge_zero_refcount) == 1 ? HPAGE_PMD_NR : 0;
mm/huge_memory.c
294
return HPAGE_PMD_NR;
mm/huge_memory.c
2972
for (i = 0, addr = haddr; i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE) {
mm/huge_memory.c
3036
add_mm_counter(mm, mm_counter_file(folio), -HPAGE_PMD_NR);
mm/huge_memory.c
3094
folio_ref_add(folio, HPAGE_PMD_NR - 1);
mm/huge_memory.c
3098
folio_add_anon_rmap_ptes(folio, page, HPAGE_PMD_NR,
mm/huge_memory.c
3162
folio_ref_add(folio, HPAGE_PMD_NR - 1);
mm/huge_memory.c
3165
folio_add_anon_rmap_ptes(folio, page, HPAGE_PMD_NR,
mm/huge_memory.c
3188
for (i = 0, addr = haddr; i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE) {
mm/huge_memory.c
3214
for (i = 0, addr = haddr; i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE) {
mm/huge_memory.c
3253
for (i = 0; i < HPAGE_PMD_NR; i++)
mm/huge_memory.c
3256
set_ptes(mm, haddr, pte, entry, HPAGE_PMD_NR);
mm/huge_memory.c
3410
add_mm_counter(mm, MM_ANONPAGES, -HPAGE_PMD_NR);
mm/huge_memory.c
4385
if (khugepaged_max_ptes_none == HPAGE_PMD_NR - 1)
mm/huge_memory.c
4400
if (++num_filled_pages >= HPAGE_PMD_NR - khugepaged_max_ptes_none)
mm/khugepaged.c
1261
for (addr = start_addr, _pte = pte; _pte < pte + HPAGE_PMD_NR;
mm/khugepaged.c
1384
(unmapped && referenced < HPAGE_PMD_NR / 2))) {
mm/khugepaged.c
1537
i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE, pte++) {
mm/khugepaged.c
1591
for (i = 0, addr = haddr, pte = start_pte; i < HPAGE_PMD_NR;
mm/khugepaged.c
1855
pgoff_t index = 0, end = start + HPAGE_PMD_NR;
mm/khugepaged.c
1863
VM_BUG_ON(start & (HPAGE_PMD_NR - 1));
mm/khugepaged.c
2190
lruvec_stat_mod_folio(new_folio, NR_SHMEM, HPAGE_PMD_NR);
mm/khugepaged.c
2191
lruvec_stat_mod_folio(new_folio, NR_SHMEM_THPS, HPAGE_PMD_NR);
mm/khugepaged.c
2193
lruvec_stat_mod_folio(new_folio, NR_FILE_THPS, HPAGE_PMD_NR);
mm/khugepaged.c
2195
lruvec_stat_mod_folio(new_folio, NR_FILE_PAGES, HPAGE_PMD_NR);
mm/khugepaged.c
2203
folio_ref_add(new_folio, HPAGE_PMD_NR - 1);
mm/khugepaged.c
2278
trace_mm_khugepaged_collapse_file(mm, new_folio, index, addr, is_shmem, file, HPAGE_PMD_NR, result);
mm/khugepaged.c
2297
xas_for_each(&xas, folio, start + HPAGE_PMD_NR - 1) {
mm/khugepaged.c
2376
present < HPAGE_PMD_NR - khugepaged_max_ptes_none) {
mm/khugepaged.c
2489
progress += HPAGE_PMD_NR;
mm/khugepaged.c
255
if (err || max_ptes_none > HPAGE_PMD_NR - 1)
mm/khugepaged.c
280
if (err || max_ptes_swap > HPAGE_PMD_NR - 1)
mm/khugepaged.c
306
if (err || max_ptes_shared > HPAGE_PMD_NR - 1)
mm/khugepaged.c
377
khugepaged_pages_to_scan = HPAGE_PMD_NR * 8;
mm/khugepaged.c
378
khugepaged_max_ptes_none = HPAGE_PMD_NR - 1;
mm/khugepaged.c
379
khugepaged_max_ptes_swap = HPAGE_PMD_NR / 8;
mm/khugepaged.c
380
khugepaged_max_ptes_shared = HPAGE_PMD_NR / 2;
mm/khugepaged.c
542
for (_pte = pte; _pte < pte + HPAGE_PMD_NR;
mm/khugepaged.c
682
for (_pte = pte; _pte < pte + HPAGE_PMD_NR; _pte += nr_ptes,
mm/khugepaged.c
755
release_pte_pages(pte, pte + HPAGE_PMD_NR, compound_pagelist);
mm/khugepaged.c
784
for (i = 0; i < HPAGE_PMD_NR; i++) {
mm/khugepaged.c
980
unsigned long addr, end = start_addr + (HPAGE_PMD_NR * PAGE_SIZE);
mm/memory-failure.c
727
if (pfn <= hwp->pfn && hwp->pfn < pfn + HPAGE_PMD_NR) {
mm/memory.c
5456
flush_icache_pages(vma, page, HPAGE_PMD_NR);
mm/memory.c
5462
add_mm_counter(vma->vm_mm, mm_counter_file(folio), HPAGE_PMD_NR);
mm/migrate.c
1583
#define NR_MAX_BATCHED_MIGRATION HPAGE_PMD_NR
mm/migrate_device.c
875
add_mm_counter(vma->vm_mm, MM_ANONPAGES, HPAGE_PMD_NR);
mm/migrate_device.c
903
for (i = 0; i < HPAGE_PMD_NR; i++)
mm/migrate_device.c
925
for (i = 1; i < HPAGE_PMD_NR; i++)
mm/migrate_device.c
952
nr = HPAGE_PMD_NR;
mm/mprotect.c
495
if (ret == HPAGE_PMD_NR) {
mm/mprotect.c
496
pages += HPAGE_PMD_NR;
mm/page_vma_mapped.c
142
if ((pfn + HPAGE_PMD_NR - 1) < pvmw->pfn)
mm/page_vma_mapped.c
287
(pvmw->nr_pages >= HPAGE_PMD_NR)) {
mm/rmap.c
1614
__folio_add_anon_rmap(folio, page, HPAGE_PMD_NR, vma, address, flags,
mm/rmap.c
1743
__folio_add_file_rmap(folio, page, HPAGE_PMD_NR, vma, PGTABLE_LEVEL_PMD);
mm/rmap.c
1911
__folio_remove_rmap(folio, page, HPAGE_PMD_NR, vma, PGTABLE_LEVEL_PMD);
mm/shmem.c
1965
if (pages == HPAGE_PMD_NR)
mm/shmem.c
1988
if (pages == HPAGE_PMD_NR) {
mm/swap.h
12
#define SWAPFILE_CLUSTER HPAGE_PMD_NR
mm/vmscan.c
1315
if (nr_pages >= HPAGE_PMD_NR) {
mm/vmstat.c
1762
pages /= HPAGE_PMD_NR;
mm/vmstat.c
1905
v[i] /= HPAGE_PMD_NR;