pageblock_nr_pages
unsigned long pageblock_size = PAGE_SIZE * pageblock_nr_pages;
sb_size = PAGE_SIZE * pageblock_nr_pages;
#define CMA_MIN_ALIGNMENT_PAGES pageblock_nr_pages
#define pageblock_align(pfn) ALIGN((pfn), pageblock_nr_pages)
#define pageblock_aligned(pfn) IS_ALIGNED((pfn), pageblock_nr_pages)
#define pageblock_start_pfn(pfn) ALIGN_DOWN((pfn), pageblock_nr_pages)
#define pageblock_end_pfn(pfn) ALIGN((pfn) + 1, pageblock_nr_pages)
pfn += pageblock_nr_pages)
for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages)
pfn += pageblock_nr_pages)
block_end_pfn += pageblock_nr_pages) {
limit = pageblock_nr_pages >> 1;
highest -= pageblock_nr_pages;
block_end_pfn = min(block_start_pfn + pageblock_nr_pages,
block_start_pfn -= pageblock_nr_pages,
if (!(block_start_pfn % (COMPACT_CLUSTER_MAX * pageblock_nr_pages)))
pageblock_nr_pages);
block_start_pfn - pageblock_nr_pages;
block_end_pfn += pageblock_nr_pages) {
if (!(low_pfn % (COMPACT_CLUSTER_MAX * pageblock_nr_pages)))
for (; migrate_pfn < free_pfn; migrate_pfn += pageblock_nr_pages,
free_pfn -= pageblock_nr_pages) {
block_end_pfn += pageblock_nr_pages) {
for (i = 0; i < nr_pages; i += pageblock_nr_pages) {
recommended_min = pageblock_nr_pages * nr_zones * 2;
recommended_min += pageblock_nr_pages * nr_zones *
zone->nr_isolate_pageblock += nr_pages / pageblock_nr_pages;
zone->nr_isolate_pageblock -= nr_pages / pageblock_nr_pages;
zonesize += zone_start_pfn & (pageblock_nr_pages-1);
usemapsize = round_up(zonesize, pageblock_nr_pages);
for (i = 0; i < nr_pages; i += pageblock_nr_pages)
unsigned i = pageblock_nr_pages;
adjust_managed_page_count(page, pageblock_nr_pages);
page_zone(page)->cma_pages += pageblock_nr_pages;
adjust_managed_page_count(page, pageblock_nr_pages);
page_zone(page)->cma_pages += pageblock_nr_pages;
block_end_pfn += pageblock_nr_pages) {
VM_WARN_ON(start_pfn & (pageblock_nr_pages - 1));
if ((pageblock_nr_pages * 4) > zone_managed_pages(zone))
max_boost = max(pageblock_nr_pages, max_boost);
zone->watermark_boost = min(zone->watermark_boost + pageblock_nr_pages,
alike_pages = pageblock_nr_pages
for (; page < endpage; page += pageblock_nr_pages) {
if ((zone_managed_pages(zone) / 100) < pageblock_nr_pages)
max_managed = ALIGN((zone_managed_pages(zone) / 100), pageblock_nr_pages);
zone->nr_reserved_highatomic += pageblock_nr_pages;
pageblock_nr_pages)
size = max(pageblock_nr_pages, 1UL << order);
if (vm_total_pages < (pageblock_nr_pages * MIGRATE_TYPES))
pageblock_page += pageblock_nr_pages;
isolate_pageblock = boundary_pfn - pageblock_nr_pages;
isolate_pageblock + pageblock_nr_pages);
if (isolate_start == isolate_end - pageblock_nr_pages)
for (pfn = isolate_start + pageblock_nr_pages;
pfn < isolate_end - pageblock_nr_pages;
pfn += pageblock_nr_pages) {
page = __first_valid_page(pfn, pageblock_nr_pages);
pfn_to_page(isolate_end - pageblock_nr_pages));
pfn += pageblock_nr_pages) {
page = __first_valid_page(pfn, pageblock_nr_pages);
for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) {
page = __first_valid_page(pfn, pageblock_nr_pages);
for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) {
seq_printf(m, "Pages per block: %lu\n", pageblock_nr_pages);
#define pageblock_align(pfn) ALIGN((pfn), pageblock_nr_pages)
#define pageblock_start_pfn(pfn) ALIGN_DOWN((pfn), pageblock_nr_pages)
#define pageblock_align(pfn) ALIGN((pfn), pageblock_nr_pages)
#define pageblock_start_pfn(pfn) ALIGN_DOWN((pfn), pageblock_nr_pages)