Symbol: PAGES_PER_SECTION
arch/arm64/mm/mmu.c
1774
WARN_ON_ONCE(end - start > PAGES_PER_SECTION * sizeof(struct page));
arch/arm64/mm/mmu.c
1777
(end - start < PAGES_PER_SECTION * sizeof(struct page)))
arch/arm64/mm/mmu.c
2033
for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
arch/x86/mm/init_64.c
1566
if (end - start < PAGES_PER_SECTION * sizeof(struct page))
drivers/base/memory.c
227
unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
drivers/base/memory.c
283
unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
drivers/base/memory.c
471
unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
drivers/base/memory.c
568
unsigned long pages_per_block = PAGES_PER_SECTION * sections_per_block;
drivers/base/memory.c
719
const unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
drivers/virtio/virtio_mem.c
2181
for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
drivers/virtio/virtio_mem.c
2186
rc = virtio_mem_fake_offline(vm, pfn, PAGES_PER_SECTION);
drivers/virtio/virtio_mem.c
2210
for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
drivers/virtio/virtio_mem.c
2214
virtio_mem_fake_online(pfn, PAGES_PER_SECTION);
drivers/virtio/virtio_mem.c
2232
pfn += PAGES_PER_SECTION) {
drivers/virtio/virtio_mem.c
2252
pfn += PAGES_PER_SECTION) {
drivers/virtio/virtio_mem.c
2760
chunk_size = max_t(uint64_t, PFN_PHYS(PAGES_PER_SECTION),
drivers/xen/balloon.c
254
PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL);
drivers/xen/balloon.c
281
balloon_hotplug = round_up(credit, PAGES_PER_SECTION);
drivers/xen/unpopulated-alloc.c
42
unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION);
drivers/xen/unpopulated-alloc.c
57
PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL);
fs/proc/page.c
39
return round_up(max_pfn, PAGES_PER_SECTION);
include/linux/mmzone.h
1867
#define PAGE_SECTION_MASK (~(PAGES_PER_SECTION-1))
include/linux/mmzone.h
1885
#define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK)
include/linux/page_ext.h
80
return next_pfn % PAGES_PER_SECTION;
mm/bootmem_info.c
121
for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
mm/bootmem_info.c
62
mapsize = sizeof(struct page) * PAGES_PER_SECTION;
mm/bootmem_info.c
93
PAGES_PER_SECTION);
mm/compaction.c
240
return section_nr_to_pfn(start_nr) + PAGES_PER_SECTION;
mm/memblock.c
2168
start = min(start, ALIGN(prev_end, PAGES_PER_SECTION));
mm/memblock.c
2193
if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) {
mm/memblock.c
2195
free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTION));
mm/memory_hotplug.c
1001
online_pages += PAGES_PER_SECTION;
mm/memory_hotplug.c
1115
if (nr_pages >= PAGES_PER_SECTION)
mm/memory_hotplug.c
1116
online_mem_sections(pfn, ALIGN_DOWN(end_pfn, PAGES_PER_SECTION));
mm/memory_hotplug.c
1130
if (nr_pages >= PAGES_PER_SECTION)
mm/memory_hotplug.c
1131
offline_mem_sections(pfn, ALIGN_DOWN(end_pfn, PAGES_PER_SECTION));
mm/memory_hotplug.c
1167
!IS_ALIGNED(pfn + nr_pages, PAGES_PER_SECTION)))
mm/memory_hotplug.c
1929
!IS_ALIGNED(start_pfn + nr_pages, PAGES_PER_SECTION)))
mm/memory_hotplug.c
335
min_align = PAGES_PER_SECTION;
mm/memory_hotplug.c
770
if (!IS_ALIGNED(start_pfn, PAGES_PER_SECTION))
mm/memory_hotplug.c
772
if (!IS_ALIGNED(start_pfn + nr_pages, PAGES_PER_SECTION))
mm/memory_hotplug.c
994
for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
mm/mm_init.c
2154
.align = PAGES_PER_SECTION,
mm/mm_init.c
2155
.min_chunk = PAGES_PER_SECTION,
mm/mm_init.c
2216
spfn = epfn, epfn += PAGES_PER_SECTION) {
mm/mm_init.c
737
if ((nr_initialised > PAGES_PER_SECTION) &&
mm/mm_init.c
738
(pfn & (PAGES_PER_SECTION - 1)) == 0) {
mm/mm_init.c
995
end_pfn = round_up(end_pfn, PAGES_PER_SECTION);
mm/numa_memblks.c
431
if (pfn_align && pfn_align < PAGES_PER_SECTION) {
mm/numa_memblks.c
434
unsigned long sect_align_mb = PFN_PHYS(PAGES_PER_SECTION) / SZ_1M;
mm/page_alloc.c
376
pfn &= (PAGES_PER_SECTION-1);
mm/page_ext.c
298
table_size = page_ext_size * PAGES_PER_SECTION;
mm/page_ext.c
328
table_size = page_ext_size * PAGES_PER_SECTION;
mm/page_ext.c
385
for (pfn = start; !fail && pfn < end; pfn += PAGES_PER_SECTION)
mm/page_ext.c
391
end = pfn - PAGES_PER_SECTION;
mm/page_ext.c
392
for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
mm/page_ext.c
415
for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
mm/page_ext.c
420
for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
mm/page_ext.c
471
pfn = ALIGN(pfn + 1, PAGES_PER_SECTION)) {
mm/sparse.c
197
pfns = min(nr_pages, PAGES_PER_SECTION
mm/sparse.c
223
for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) {
mm/sparse.c
408
return ALIGN(sizeof(struct page) * PAGES_PER_SECTION, PMD_SIZE);
mm/sparse.c
414
return PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION);
mm/sparse.c
557
map = __populate_section_memmap(pfn, PAGES_PER_SECTION,
mm/sparse.c
567
memmap_boot_pages_add(DIV_ROUND_UP(PAGES_PER_SECTION * sizeof(struct page),
mm/sparse.c
634
for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
mm/sparse.c
652
for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
mm/sparse.c
687
unsigned long end = (unsigned long)(memmap + PAGES_PER_SECTION);
mm/sparse.c
746
PAGES_PER_SECTION), GFP_KERNEL, nid);
mm/sparse.c
761
nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page))
mm/sparse.c
901
if (nr_pages < PAGES_PER_SECTION && early_section(ms))
mm/util.c
1463
for (pfn = ALIGN(start_pfn, PAGES_PER_SECTION);
mm/util.c
1464
pfn < end_pfn; pfn += PAGES_PER_SECTION)