arch/x86/mm/fault.c
305
if (!low_pfn(pgd_val(*pgd) >> PAGE_SHIFT) || !pgd_present(*pgd))
arch/x86/mm/fault.c
323
if (!low_pfn(pmd_pfn(*pmd)) || !pmd_present(*pmd) || pmd_leaf(*pmd))
drivers/iommu/iova.c
173
unsigned long high_pfn = limit_pfn, low_pfn = iovad->start_pfn;
drivers/iommu/iova.c
195
} while (curr && new_pfn <= curr_iova->pfn_hi && new_pfn >= low_pfn);
drivers/iommu/iova.c
197
if (high_pfn < size || new_pfn < low_pfn) {
drivers/iommu/iova.c
198
if (low_pfn == iovad->start_pfn && retry_pfn < limit_pfn) {
drivers/iommu/iova.c
200
low_pfn = retry_pfn + 1;
mm/compaction.c
1019
low_pfn += (1UL << freepage_order) - 1;
mm/compaction.c
1039
low_pfn += (1UL << order) - 1;
mm/compaction.c
1177
low_pfn = end_pfn;
mm/compaction.c
1188
low_pfn += folio_nr_pages(folio) - 1;
mm/compaction.c
1197
low_pfn += folio_nr_pages(folio) - 1;
mm/compaction.c
1220
++low_pfn;
mm/compaction.c
1253
if (low_pfn < next_skip_pfn) {
mm/compaction.c
1254
low_pfn = next_skip_pfn - 1;
mm/compaction.c
1270
if (unlikely(low_pfn > end_pfn))
mm/compaction.c
1271
low_pfn = end_pfn;
mm/compaction.c
1291
if (low_pfn == end_pfn && (!nr_isolated || cc->finish_pageblock)) {
mm/compaction.c
1294
update_cached_migrate(cc, low_pfn);
mm/compaction.c
1297
trace_mm_compaction_isolate_migratepages(start_pfn, low_pfn,
mm/compaction.c
1305
cc->migrate_pfn = low_pfn;
mm/compaction.c
1508
unsigned long low_pfn, min_pfn, highest = 0;
mm/compaction.c
1533
low_pfn = pageblock_start_pfn(cc->free_pfn - (distance >> 2));
mm/compaction.c
1536
if (WARN_ON_ONCE(min_pfn > low_pfn))
mm/compaction.c
1537
low_pfn = min_pfn;
mm/compaction.c
1571
if (pfn >= low_pfn) {
mm/compaction.c
1669
low_pfn = page_to_pfn(page);
mm/compaction.c
1670
fast_isolate_around(cc, low_pfn);
mm/compaction.c
1684
unsigned long low_pfn; /* lowest pfn scanner is able to scan */
mm/compaction.c
1707
low_pfn = pageblock_end_pfn(cc->migrate_pfn);
mm/compaction.c
1715
for (; block_start_pfn >= low_pfn;
mm/compaction.c
1735
block_start_pfn = max(next_pfn, low_pfn);
mm/compaction.c
2049
unsigned long low_pfn;
mm/compaction.c
2061
low_pfn = fast_find_migrateblock(cc);
mm/compaction.c
2062
block_start_pfn = pageblock_start_pfn(low_pfn);
mm/compaction.c
2071
fast_find_block = low_pfn != cc->migrate_pfn && !cc->fast_search_fail;
mm/compaction.c
2074
block_end_pfn = pageblock_end_pfn(low_pfn);
mm/compaction.c
2082
cc->migrate_pfn = low_pfn = block_end_pfn,
mm/compaction.c
2091
if (!(low_pfn % (COMPACT_CLUSTER_MAX * pageblock_nr_pages)))
mm/compaction.c
2112
if ((pageblock_aligned(low_pfn) ||
mm/compaction.c
2113
low_pfn == cc->zone->zone_start_pfn) &&
mm/compaction.c
2131
if (isolate_migratepages_block(cc, low_pfn, block_end_pfn,
mm/compaction.c
837
isolate_migratepages_block(struct compact_control *cc, unsigned long low_pfn,
mm/compaction.c
848
unsigned long start_pfn = low_pfn;
mm/compaction.c
854
cc->migrate_pfn = low_pfn;
mm/compaction.c
880
next_skip_pfn = block_end_pfn(low_pfn, cc->order);
mm/compaction.c
884
for (; low_pfn < end_pfn; low_pfn++) {
mm/compaction.c
887
if (skip_on_failure && low_pfn >= next_skip_pfn) {
mm/compaction.c
906
next_skip_pfn = block_end_pfn(low_pfn, cc->order);
mm/compaction.c
914
if (!(low_pfn % COMPACT_CLUSTER_MAX)) {
mm/compaction.c
932
page = pfn_to_page(low_pfn);
mm/compaction.c
940
if (!valid_page && (pageblock_aligned(low_pfn) ||
mm/compaction.c
941
low_pfn == cc->zone->zone_start_pfn)) {
mm/compaction.c
943
low_pfn = end_pfn;
mm/compaction.c
960
low_pfn += (1UL << order) - 1;
mm/compaction.c
982
low_pfn += (1UL << order) - 1;
mm/compaction.c
992
low_pfn += folio_nr_pages(folio) - folio_page_idx(folio, page) - 1;
mm/internal.h
1004
unsigned long low_pfn, unsigned long end_pfn);