Symbol: zone_end_pfn
arch/arm64/kernel/hibernate.c
266
max_zone_pfn = zone_end_pfn(zone);
include/linux/mmzone.h
1124
return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone);
include/linux/mmzone.h
1277
if (start_pfn >= zone_end_pfn(zone) ||
kernel/power/snapshot.c
1256
max_zone_pfn = zone_end_pfn(zone);
kernel/power/snapshot.c
1356
max_zone_pfn = zone_end_pfn(zone);
kernel/power/snapshot.c
1418
max_zone_pfn = zone_end_pfn(zone);
kernel/power/snapshot.c
1531
max_zone_pfn = zone_end_pfn(zone);
kernel/power/snapshot.c
639
zone_end = zone_end_pfn(zone);
mm/compaction.c
1473
end_pfn = min(pageblock_end_pfn(pfn), zone_end_pfn(cc->zone));
mm/compaction.c
1649
zone_end_pfn(cc->zone)),
mm/compaction.c
1706
zone_end_pfn(zone));
mm/compaction.c
202
pageblock_start_pfn(zone_end_pfn(zone) - 1);
mm/compaction.c
2515
unsigned long end_pfn = zone_end_pfn(cc->zone);
mm/compaction.c
317
block_pfn = min(block_pfn, zone_end_pfn(zone) - 1);
mm/compaction.c
352
unsigned long free_pfn = zone_end_pfn(zone) - 1;
mm/huge_memory.c
4539
max_zone_pfn = zone_end_pfn(zone);
mm/kmemleak.c
1754
unsigned long end_pfn = zone_end_pfn(zone);
mm/memory_hotplug.c
493
zone_end_pfn(zone));
mm/memory_hotplug.c
495
zone->spanned_pages = zone_end_pfn(zone) - pfn;
mm/memory_hotplug.c
501
} else if (zone_end_pfn(zone) == end_pfn) {
mm/memory_hotplug.c
526
unsigned long end_pfn = zone_end_pfn(zone);
mm/memory_hotplug.c
706
unsigned long old_end_pfn = zone_end_pfn(zone);
mm/mm_init.c
1165
unsigned long *zone_end_pfn)
mm/mm_init.c
1172
*zone_end_pfn = min(node_end_pfn,
mm/mm_init.c
1178
*zone_end_pfn > zone_movable_pfn[nid]) {
mm/mm_init.c
1179
*zone_end_pfn = zone_movable_pfn[nid];
mm/mm_init.c
1183
*zone_start_pfn = *zone_end_pfn;
mm/mm_init.c
1224
unsigned long zone_end_pfn)
mm/mm_init.c
1229
if (zone_start_pfn == zone_end_pfn)
mm/mm_init.c
1232
nr_absent = __absent_pages_in_range(nid, zone_start_pfn, zone_end_pfn);
mm/mm_init.c
1245
zone_start_pfn, zone_end_pfn);
mm/mm_init.c
1247
zone_start_pfn, zone_end_pfn);
mm/mm_init.c
1271
unsigned long *zone_end_pfn)
mm/mm_init.c
1278
*zone_end_pfn = clamp(node_end_pfn, zone_low, zone_high);
mm/mm_init.c
1280
zone_start_pfn, zone_end_pfn);
mm/mm_init.c
1283
if (*zone_end_pfn < node_start_pfn || *zone_start_pfn > node_end_pfn)
mm/mm_init.c
1287
*zone_end_pfn = min(*zone_end_pfn, node_end_pfn);
mm/mm_init.c
1291
return *zone_end_pfn - *zone_start_pfn;
mm/mm_init.c
1345
unsigned long zone_start_pfn, zone_end_pfn;
mm/mm_init.c
1353
&zone_end_pfn);
mm/mm_init.c
1356
zone_end_pfn);
mm/mm_init.c
2147
last_pfn = SECTION_ALIGN_UP(zone_end_pfn(zone));
mm/mm_init.c
2192
if (zone_end_pfn(zone) != pgdat_end_pfn(pgdat))
mm/mm_init.c
2215
nr_pages < nr_pages_needed && spfn < zone_end_pfn(zone);
mm/mm_init.c
2270
for (; block_start_pfn < zone_end_pfn(zone);
mm/mm_init.c
2274
block_end_pfn = min(block_end_pfn, zone_end_pfn(zone));
mm/mm_init.c
873
unsigned long start_pfn, unsigned long zone_end_pfn,
mm/mm_init.c
910
if (defer_init(nid, pfn, zone_end_pfn)) {
mm/mm_init.c
947
unsigned long zone_end_pfn = zone_start_pfn + zone->spanned_pages;
mm/mm_init.c
950
start_pfn = clamp(start_pfn, zone_start_pfn, zone_end_pfn);
mm/mm_init.c
951
end_pfn = clamp(end_pfn, zone_start_pfn, zone_end_pfn);
mm/mm_init.c
957
zone_end_pfn, MEMINIT_EARLY, NULL, MIGRATE_MOVABLE,
mm/page_owner.c
432
unsigned long end_pfn = zone_end_pfn(zone);
mm/page_owner.c
775
unsigned long end_pfn = zone_end_pfn(zone);
mm/shuffle.c
84
unsigned long end_pfn = zone_end_pfn(z);
mm/vmstat.c
1635
unsigned long end_pfn = zone_end_pfn(zone);