Symbol: PAGE_OFFSET
arch/alpha/boot/bootpz.c
88
kaddr = (find_pa(vaddr) | PAGE_OFFSET);
arch/alpha/include/asm/page.h
78
#define __pa(x) ((unsigned long) (x) - PAGE_OFFSET)
arch/alpha/include/asm/page.h
79
#define __va(x) ((void *)((unsigned long) (x) + PAGE_OFFSET))
arch/alpha/include/asm/pgalloc.h
19
pmd_set(pmd, (pte_t *)(page_to_pa(pte) + PAGE_OFFSET));
arch/alpha/include/asm/pgtable.h
181
{ pmd_val(*pmdp) = _PAGE_TABLE | ((((unsigned long) ptep) - PAGE_OFFSET) << (32-PAGE_SHIFT)); }
arch/alpha/include/asm/pgtable.h
184
{ pud_val(*pudp) = _PAGE_TABLE | ((((unsigned long) pmdp) - PAGE_OFFSET) << (32-PAGE_SHIFT)); }
arch/alpha/include/asm/pgtable.h
193
return ((pmd_val(pmd) & _PFN_MASK) >> (32-PAGE_SHIFT)) + PAGE_OFFSET;
arch/alpha/include/asm/pgtable.h
202
return (pmd_t *)(PAGE_OFFSET + ((pud_val(pgd) & _PFN_MASK) >> (32-PAGE_SHIFT)));
arch/alpha/include/asm/setup.h
23
#define KERNEL_START (PAGE_OFFSET+KERNEL_START_PHYS)
arch/alpha/include/asm/setup.h
25
#define INIT_STACK (PAGE_OFFSET+KERNEL_START_PHYS+0x02000)
arch/alpha/include/asm/setup.h
26
#define EMPTY_PGT (PAGE_OFFSET+KERNEL_START_PHYS+0x04000)
arch/alpha/include/asm/setup.h
27
#define EMPTY_PGE (PAGE_OFFSET+KERNEL_START_PHYS+0x08000)
arch/alpha/include/asm/setup.h
28
#define ZERO_PGE (PAGE_OFFSET+KERNEL_START_PHYS+0x0A000)
arch/alpha/include/asm/setup.h
30
#define START_ADDR (PAGE_OFFSET+KERNEL_START_PHYS+0x10000)
arch/alpha/mm/init.c
107
if (original_pcb_ptr < PAGE_OFFSET) {
arch/alpha/mm/init.c
82
newptbr = ((unsigned long) swapper_pg_dir - PAGE_OFFSET) >> PAGE_SHIFT;
arch/arc/include/asm/highmem.h
18
#define FIXMAP_BASE (PAGE_OFFSET - FIXMAP_SIZE - PKMAP_SIZE)
arch/arc/include/asm/processor.h
87
#define VMALLOC_START (PAGE_OFFSET - (CONFIG_ARC_KVADDR_SIZE << 20))
arch/arc/kernel/setup.c
437
if (addr < PAGE_OFFSET)
arch/arc/mm/cache.c
803
if (likely(kstart > PAGE_OFFSET)) {
arch/arc/mm/highmem.c
67
BUILD_BUG_ON(PAGE_OFFSET < (VMALLOC_END + FIXMAP_SIZE + PKMAP_SIZE));
arch/arm/include/asm/dma.h
13
arm_dma_zone_size && arm_dma_zone_size < (0x100000000ULL - PAGE_OFFSET) ? \
arch/arm/include/asm/dma.h
14
(PAGE_OFFSET + arm_dma_zone_size) : 0xffffffffUL; })
arch/arm/include/asm/highmem.h
8
#define PKMAP_BASE (PAGE_OFFSET - PMD_SIZE)
arch/arm/include/asm/memory.h
127
#define MODULES_VADDR PAGE_OFFSET
arch/arm/include/asm/memory.h
280
return (phys_addr_t)x - PAGE_OFFSET + PHYS_OFFSET;
arch/arm/include/asm/memory.h
285
return x - PHYS_OFFSET + PAGE_OFFSET;
arch/arm/include/asm/memory.h
293
return (((kaddr - PAGE_OFFSET) >> PAGE_SHIFT) +
arch/arm/include/asm/memory.h
35
#define KERNEL_OFFSET (PAGE_OFFSET)
arch/arm/include/asm/memory.h
385
#define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (unsigned long)high_memory) \
arch/arm/include/asm/memory.h
60
#define MODULES_VADDR (PAGE_OFFSET - SZ_16M)
arch/arm/include/asm/memory.h
63
#define MODULES_VADDR (PAGE_OFFSET - SZ_8M)
arch/arm/include/asm/memory.h
74
#define MODULES_END (PAGE_OFFSET - PMD_SIZE)
arch/arm/include/asm/memory.h
76
#define MODULES_END (PAGE_OFFSET)
arch/arm/include/asm/pgtable-3level-hwdef.h
97
#define TTBR1_SIZE (((PAGE_OFFSET >> 30) - 1) << 16)
arch/arm/include/asm/pgtable-3level.h
52
#define USER_PTRS_PER_PGD (PAGE_OFFSET / PGDIR_SIZE)
arch/arm/kernel/atags_parse.c
201
tags = (void *)(PAGE_OFFSET + mdesc->atag_offset);
arch/arm/kernel/traps.c
126
if (sp < PAGE_OFFSET ||
arch/arm/mach-footbridge/dc21285.c
300
mem_size = (unsigned int)high_memory - PAGE_OFFSET;
arch/arm/mach-omap2/board-generic.c
111
const struct tag *tags = (const struct tag *)(PAGE_OFFSET + 0x100);
arch/arm/mach-omap2/omap-smp.c
295
if ((addr >= __pa(PAGE_OFFSET)) && (addr <= __pa(__bss_start)))
arch/arm/mach-tegra/pm.c
206
err = cpu_suspend(PHYS_OFFSET - PAGE_OFFSET, &tegra_sleep_cpu);
arch/arm/mach-tegra/pm.c
372
cpu_suspend(PHYS_OFFSET - PAGE_OFFSET, tegra_sleep_func);
arch/arm/mach-zynq/common.c
54
if (!__pa(PAGE_OFFSET))
arch/arm/mach-zynq/common.c
55
memblock_reserve(__pa(PAGE_OFFSET), 0x80000);
arch/arm/mach-zynq/platsmp.c
46
if (__pa(PAGE_OFFSET)) {
arch/arm/mach-zynq/platsmp.c
53
zero = (__force u8 __iomem *)PAGE_OFFSET;
arch/arm/mach-zynq/platsmp.c
70
if (__pa(PAGE_OFFSET))
arch/arm/mm/dump.c
27
{ PAGE_OFFSET, "Kernel Mapping" },
arch/arm/mm/init.c
244
BUILD_BUG_ON(PKMAP_BASE + LAST_PKMAP * PAGE_SIZE > PAGE_OFFSET);
arch/arm/mm/init.c
245
BUG_ON(PKMAP_BASE + LAST_PKMAP * PAGE_SIZE > PAGE_OFFSET);
arch/arm/mm/init.c
266
.start = PAGE_OFFSET,
arch/arm/mm/mmu.c
1171
vmalloc_max = VMALLOC_END - (PAGE_OFFSET + SZ_32M + VMALLOC_OFFSET);
arch/arm/mm/mmu.c
1199
PAGE_OFFSET + PHYS_OFFSET;
arch/arm/mm/mmu.c
1312
for ( ; addr < PAGE_OFFSET; addr += PMD_SIZE)
arch/arm/mm/mmu.c
986
md->virtual >= PAGE_OFFSET && md->virtual < FIXADDR_START &&
arch/arm/mm/physaddr.c
21
if (!high_memory && x >= PAGE_OFFSET)
arch/arm/mm/physaddr.c
24
if (high_memory && x >= PAGE_OFFSET && x < (unsigned long)high_memory)
arch/arm64/include/asm/memory.h
33
#define VMEMMAP_RANGE (_PAGE_END(VA_BITS_MIN) - PAGE_OFFSET)
arch/arm64/include/asm/memory.h
338
#define __is_lm_address(addr) (((u64)(addr) - PAGE_OFFSET) < (PAGE_END - PAGE_OFFSET))
arch/arm64/include/asm/memory.h
340
#define __lm_to_phys(addr) (((addr) - PAGE_OFFSET) + PHYS_OFFSET)
arch/arm64/include/asm/memory.h
358
#define __phys_to_virt(x) ((unsigned long)((x) - PHYS_OFFSET) | PAGE_OFFSET)
arch/arm64/include/asm/memory.h
412
u64 __addr = PAGE_OFFSET + (__idx * PAGE_SIZE); \
arch/arm64/include/asm/memory.h
417
u64 __idx = (__tag_reset((u64)x) - PAGE_OFFSET) / PAGE_SIZE; \
arch/arm64/include/asm/pgtable.h
27
#define VMEMMAP_UNUSED_NPAGES ((_PAGE_OFFSET(vabits_actual) - PAGE_OFFSET) >> PAGE_SHIFT)
arch/arm64/include/asm/processor.h
369
return arch_kasan_reset_tag(addr) >= PAGE_OFFSET;
arch/arm64/kernel/head.S
41
#if (PAGE_OFFSET & 0x1fffff) != 0
arch/arm64/kernel/hibernate.c
424
rc = trans_pgd_create_copy(&trans_info, &tmp_pg_dir, PAGE_OFFSET,
arch/arm64/kernel/machine_kexec.c
135
rc = trans_pgd_create_copy(&info, &trans_pgd, PAGE_OFFSET, PAGE_END);
arch/arm64/kernel/pi/map_kernel.c
27
map_range(pgd, ((u64)start + va_offset) & ~PAGE_OFFSET,
arch/arm64/kernel/pi/map_kernel.c
28
((u64)end + va_offset) & ~PAGE_OFFSET, (u64)start,
arch/arm64/kvm/mmu.c
2319
kern_hyp_va(PAGE_OFFSET),
arch/arm64/kvm/mmu.c
2322
if (hyp_idmap_start >= kern_hyp_va(PAGE_OFFSET) &&
arch/arm64/mm/mmu.c
1006
if (virt < PAGE_OFFSET) {
arch/arm64/mm/mmu.c
1931
WARN_ON((start < PAGE_OFFSET) || (end > PAGE_END));
arch/arm64/mm/mmu.c
1934
free_empty_tables(start, end, PAGE_OFFSET, PAGE_END);
arch/arm64/mm/mmu.c
979
if (virt < PAGE_OFFSET) {
arch/arm64/mm/ptdump.c
382
{ PAGE_OFFSET, "Linear Mapping start" },
arch/csky/include/asm/memory.h
13
#define VMALLOC_START (PAGE_OFFSET + LOWMEM_LIMIT + (PAGE_SIZE * 8))
arch/csky/include/asm/page.h
33
#define virt_addr_valid(kaddr) ((void *)(kaddr) >= (void *)PAGE_OFFSET && \
arch/csky/include/asm/page.h
68
#define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + va_pa_offset)
arch/csky/include/asm/page.h
69
#define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET - va_pa_offset))
arch/csky/include/asm/page.h
78
#define MAP_NR(x) PFN_DOWN((unsigned long)(x) - PAGE_OFFSET - \
arch/csky/include/asm/pgtable.h
16
#define USER_PTRS_PER_PGD (PAGE_OFFSET/PGDIR_SIZE)
arch/csky/include/asm/pgtable.h
35
(((unsigned int) addr >= PAGE_OFFSET) ? __pte(_PAGE_GLOBAL) : __pte(0)))
arch/csky/include/asm/processor.h
29
#define TASK_SIZE (PAGE_OFFSET - (PAGE_SIZE * 8))
arch/csky/mm/cachev1.c
54
unlikely(start < PAGE_OFFSET) ||
arch/csky/mm/cachev1.c
55
unlikely(start >= PAGE_OFFSET + LOWMEM_LIMIT)) {
arch/hexagon/include/asm/dma.h
12
#define MAX_DMA_ADDRESS (PAGE_OFFSET)
arch/hexagon/include/asm/mem-layout.h
41
#define TASK_SIZE (PAGE_OFFSET)
arch/hexagon/include/asm/mem-layout.h
54
#define MIN_KERNEL_SEG (PAGE_OFFSET >> PGDIR_SHIFT) /* L1 shift is 22 bits */
arch/hexagon/include/asm/page.h
83
#define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + PHYS_OFFSET)
arch/hexagon/include/asm/page.h
84
#define __va(x) ((void *)((unsigned long)(x) - PHYS_OFFSET + PAGE_OFFSET))
arch/hexagon/mm/init.c
150
segtable = segtable + (PAGE_OFFSET >> 22);
arch/hexagon/mm/init.c
183
(unsigned long) _K_init_devicetable-PAGE_OFFSET);
arch/hexagon/mm/init.c
184
*segtable = ((u32) (unsigned long) _K_init_devicetable-PAGE_OFFSET) |
arch/hexagon/mm/init.c
22
#define bootmem_startpg (PFN_UP(((unsigned long) _end) - PAGE_OFFSET + PHYS_OFFSET))
arch/loongarch/include/asm/addrspace.h
70
#ifndef PAGE_OFFSET
arch/loongarch/include/asm/dma.h
8
#define MAX_DMA_ADDRESS PAGE_OFFSET
arch/loongarch/include/asm/page.h
67
#define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET - PHYS_OFFSET))
arch/loongarch/mm/mmap.c
122
if ((vaddr < PAGE_OFFSET) || (vaddr >= vm_map_base))
arch/m68k/include/asm/dma.h
7
#define MAX_DMA_ADDRESS PAGE_OFFSET
arch/m68k/include/asm/mcf_pgalloc.h
95
memset(new_pgd, 0, PAGE_OFFSET >> PGDIR_SHIFT);
arch/m68k/include/asm/mmu_context.h
114
if (mmuar >= PAGE_OFFSET) {
arch/m68k/include/asm/mmu_context.h
140
pte = (mmuar >= PAGE_OFFSET) ? pte_offset_kernel(pmd, mmuar)
arch/m68k/include/asm/mmu_context.h
147
if (!pte_dirty(*pte) && mmuar <= PAGE_OFFSET)
arch/m68k/include/asm/mmu_context.h
164
if (pte && mmuar < PAGE_OFFSET)
arch/m68k/include/asm/motorola_pgtable.h
132
#define pud_page(pud) (mem_map + ((unsigned long)(__va(pud_val(pud)) - PAGE_OFFSET) >> PAGE_SHIFT))
arch/m68k/include/asm/page_mm.h
100
return (x - PAGE_OFFSET);
arch/m68k/include/asm/page_mm.h
111
return (void *)(x + PAGE_OFFSET);
arch/m68k/include/asm/page_mm.h
144
#define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (unsigned long)high_memory)
arch/m68k/include/asm/page_mm.h
99
if (x >= PAGE_OFFSET)
arch/m68k/include/asm/page_no.h
31
#define virt_to_page(addr) (mem_map + (((unsigned long)(addr)-PAGE_OFFSET) >> PAGE_SHIFT))
arch/m68k/include/asm/page_no.h
32
#define page_to_virt(page) __va(((((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET))
arch/m68k/include/asm/page_no.h
34
#define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET) && \
arch/m68k/include/asm/sun3_pgalloc.h
46
memset(new_pgd, 0, (PAGE_OFFSET >> PGDIR_SHIFT));
arch/m68k/kernel/setup_mm.c
242
setup_initial_init_mm((void *)PAGE_OFFSET, _etext, _edata, _end);
arch/m68k/mm/mcfmmu.c
53
pg_dir += PAGE_OFFSET >> PGDIR_SHIFT;
arch/m68k/mm/mcfmmu.c
55
address = PAGE_OFFSET;
arch/m68k/mm/motorola.c
471
m68k_memoffset = min_addr - PAGE_OFFSET;
arch/m68k/mm/sun3mmu.c
48
address = PAGE_OFFSET;
arch/m68k/mm/sun3mmu.c
60
pg_dir += PAGE_OFFSET >> PGDIR_SHIFT;
arch/m68k/sun3/config.c
144
memory_end = *(romvec->pv_sun3mem) + PAGE_OFFSET - 2*PAGE_SIZE;
arch/m68k/sun3/mmu_emu.c
192
for(seg = 0; seg < PAGE_OFFSET; seg += SUN3_PMEG_SIZE)
arch/m68k/sun3/mmu_emu.c
310
if(vaddr >= PAGE_OFFSET) {
arch/microblaze/include/asm/pgtable.h
103
#define USER_PGD_PTRS (PAGE_OFFSET >> PGDIR_SHIFT)
arch/mips/include/asm/dma.h
89
#define MAX_DMA_ADDRESS PAGE_OFFSET
arch/mips/include/asm/dma.h
91
#define MAX_DMA_ADDRESS (PAGE_OFFSET + 0x01000000)
arch/mips/include/asm/mach-generic/spaces.h
94
#ifndef PAGE_OFFSET
arch/mips/include/asm/page.h
178
return x - PAGE_OFFSET + PHYS_OFFSET;
arch/mips/include/asm/page.h
181
#define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET - PHYS_OFFSET))
arch/mips/kernel/setup.c
172
if (initrd_start < PAGE_OFFSET) {
arch/mips/mm/mmap.c
124
if ((vaddr < PAGE_OFFSET) || (vaddr >= MAP_BASE))
arch/mips/mm/physaddr.c
25
return x >= PAGE_OFFSET && (KSEGX(x) < KSEG2 ||
arch/nios2/include/asm/page.h
76
((unsigned long)(x) - PAGE_OFFSET + PHYS_OFFSET)
arch/nios2/include/asm/page.h
78
((void *)((unsigned long)(x) + PAGE_OFFSET - PHYS_OFFSET))
arch/nios2/include/asm/page.h
81
((void *)(((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET)
arch/openrisc/include/asm/page.h
21
#define KERNELBASE PAGE_OFFSET
arch/openrisc/include/asm/page.h
62
#define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET))
arch/openrisc/include/asm/page.h
63
#define __pa(x) ((unsigned long) (x) - PAGE_OFFSET)
arch/openrisc/include/asm/page.h
71
(mem_map + (((unsigned long)(addr)-PAGE_OFFSET) >> PAGE_SHIFT))
arch/openrisc/include/asm/pgtable.h
313
#define pte_pagenr(pte) ((__pte_page(pte) - PAGE_OFFSET) >> PAGE_SHIFT)
arch/openrisc/include/asm/pgtable.h
317
#define __page_address(page) (PAGE_OFFSET + (((page) - mem_map) << PAGE_SHIFT))
arch/openrisc/include/asm/pgtable.h
91
#define VMALLOC_START (PAGE_OFFSET-0x04000000UL)
arch/openrisc/include/asm/pgtable.h
92
#define VMALLOC_END (PAGE_OFFSET)
arch/openrisc/kernel/traps.c
120
if (esp < PAGE_OFFSET)
arch/openrisc/kernel/traps.c
137
if (regs->pc < PAGE_OFFSET)
arch/openrisc/mm/init.c
73
v = PAGE_OFFSET;
arch/parisc/include/asm/page.h
145
#define __pa(x) ((unsigned long)(x)-PAGE_OFFSET)
arch/parisc/include/asm/page.h
146
#define __va(x) ((void *)((unsigned long)(x)+PAGE_OFFSET))
arch/parisc/kernel/asm-offsets.c
269
DEFINE(ASM_PMD_ENTRY, ((PAGE_OFFSET & PMD_MASK) >> PMD_SHIFT));
arch/parisc/kernel/asm-offsets.c
270
DEFINE(ASM_PGD_ENTRY, PAGE_OFFSET >> PGDIR_SHIFT);
arch/powerpc/include/asm/book3s/64/hash.h
124
if (id != (PAGE_OFFSET >> 60))
arch/powerpc/include/asm/fprobe.h
9
#define FPROBE_HEADER_MSB_PATTERN (PAGE_OFFSET & ~FPROBE_HEADER_MSB_MASK)
arch/powerpc/include/asm/kasan.h
23
#define KASAN_KERN_START ALIGN_DOWN(PAGE_OFFSET - SZ_256M, SZ_256M)
arch/powerpc/include/asm/kasan.h
25
#define KASAN_KERN_START PAGE_OFFSET
arch/powerpc/include/asm/kdump.h
26
#define KDUMP_TRAMPOLINE_START (0x0100 + PAGE_OFFSET)
arch/powerpc/include/asm/kdump.h
27
#define KDUMP_TRAMPOLINE_END (0x3000 + PAGE_OFFSET)
arch/powerpc/include/asm/page.h
108
#define MEMORY_START (PHYSICAL_START + PAGE_OFFSET - KERNELBASE)
arch/powerpc/include/asm/page.h
203
VIRTUAL_WARN_ON((unsigned long)(x) >= PAGE_OFFSET); \
arch/powerpc/include/asm/page.h
204
(void *)(unsigned long)((phys_addr_t)(x) | PAGE_OFFSET); \
arch/powerpc/include/asm/page.h
209
VIRTUAL_WARN_ON((unsigned long)(x) < PAGE_OFFSET); \
arch/powerpc/include/asm/page.h
214
#define __va(x) ((void *)(unsigned long)((phys_addr_t)(x) + PAGE_OFFSET - MEMORY_START))
arch/powerpc/include/asm/page.h
215
#define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + MEMORY_START)
arch/powerpc/include/asm/page.h
234
_addr >= PAGE_OFFSET && _addr < (unsigned long)high_memory && \
arch/powerpc/include/asm/page.h
259
#define is_kernel_addr(x) ((x) >= PAGE_OFFSET)
arch/powerpc/include/asm/ppc_asm.h
570
#define tophys(rd, rs) addis rd, rs, -PAGE_OFFSET@h
arch/powerpc/include/asm/ppc_asm.h
571
#define tovirt(rd, rs) addis rd, rs, PAGE_OFFSET@h
arch/powerpc/include/asm/uaccess.h
510
const unsigned long border = PAGE_OFFSET;
arch/powerpc/kernel/btext.c
87
unsigned long vaddr = PAGE_OFFSET + 0x10000000;
arch/powerpc/kernel/head_32.h
200
addis r1, r1, emergency_ctx-PAGE_OFFSET@ha
arch/powerpc/kernel/head_32.h
202
lis r1, emergency_ctx-PAGE_OFFSET@ha
arch/powerpc/kernel/head_32.h
204
lwz r1, emergency_ctx-PAGE_OFFSET@l(r1)
arch/powerpc/kernel/traps.c
1510
bugaddr += PAGE_OFFSET;
arch/powerpc/mm/book3s32/mmu.c
151
setbat(idx, PAGE_OFFSET + base, base, size, PAGE_KERNEL_X);
arch/powerpc/mm/book3s32/mmu.c
161
unsigned long border = (unsigned long)__srwx_boundary - PAGE_OFFSET;
arch/powerpc/mm/book3s32/mmu.c
164
size = roundup_pow_of_two((unsigned long)_einittext - PAGE_OFFSET);
arch/powerpc/mm/book3s32/mmu.c
165
setibat(0, PAGE_OFFSET, 0, size, PAGE_KERNEL_X);
arch/powerpc/mm/book3s32/mmu.c
200
unsigned long base = (unsigned long)_stext - PAGE_OFFSET;
arch/powerpc/mm/book3s32/mmu.c
201
unsigned long top = ALIGN((unsigned long)_etext - PAGE_OFFSET, SZ_128K);
arch/powerpc/mm/book3s32/mmu.c
202
unsigned long border = (unsigned long)__init_begin - PAGE_OFFSET;
arch/powerpc/mm/book3s32/mmu.c
207
setibat(i++, PAGE_OFFSET + base, base, size, PAGE_KERNEL_X);
arch/powerpc/mm/book3s32/mmu.c
218
setibat(i++, PAGE_OFFSET + base, base, size, PAGE_KERNEL_X);
arch/powerpc/mm/book3s32/mmu.c
394
unsigned int hash = (unsigned int)Hash - PAGE_OFFSET;
arch/powerpc/mm/book3s64/hash_utils.c
2370
if (ea >= PAGE_OFFSET) {
arch/powerpc/mm/book3s64/radix_pgtable.c
240
for (start = PAGE_OFFSET; start < (unsigned long)_stext; start += PAGE_SIZE) {
arch/powerpc/mm/book3s64/slb.c
503
create_shadowed_slbe(PAGE_OFFSET, mmu_kernel_ssize, lflags, LINEAR_INDEX);
arch/powerpc/mm/book3s64/slb.c
513
(get_paca()->kstack & slb_esid_mask(mmu_kernel_ssize)) > PAGE_OFFSET)
arch/powerpc/mm/cacheflush.c
20
icbi((void *)PAGE_OFFSET);
arch/powerpc/mm/init_32.c
48
#if (CONFIG_LOWMEM_SIZE > (0xF0000000 - PAGE_OFFSET))
arch/powerpc/mm/mem.c
412
if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) {
arch/powerpc/mm/nohash/44x.c
179
ppc47x_pin_tlb(addr + PAGE_OFFSET, addr);
arch/powerpc/mm/nohash/44x.c
181
ppc44x_pin_tlb(addr + PAGE_OFFSET, addr);
arch/powerpc/mm/nohash/44x.c
237
ppc47x_pin_tlb(addr + PAGE_OFFSET, addr);
arch/powerpc/mm/nohash/44x.c
239
ppc44x_pin_tlb(addr + PAGE_OFFSET, addr);
arch/powerpc/mm/nohash/8xx.c
120
unsigned long v = PAGE_OFFSET + offset;
arch/powerpc/mm/nohash/8xx.c
138
flush_tlb_kernel_range(PAGE_OFFSET + v, PAGE_OFFSET + top);
arch/powerpc/mm/nohash/8xx.c
32
if (va >= PAGE_OFFSET && va < PAGE_OFFSET + block_mapped_ram)
arch/powerpc/mm/nohash/e500.c
226
unsigned long virt = PAGE_OFFSET;
arch/powerpc/mm/nohash/e500.c
240
return tlbcam_addrs[tlbcam_index - 1].limit - PAGE_OFFSET + 1;
arch/powerpc/mm/nohash/e500.c
325
virt_phys_offset = PAGE_OFFSET - memstart_addr;
arch/powerpc/mm/nohash/e500.c
368
map_mem_in_cams_addr(start, PAGE_OFFSET + offset,
arch/powerpc/mm/pgtable_32.c
103
v = PAGE_OFFSET + s;
arch/powerpc/mm/ptdump/hashpagetable.c
338
if (ea < PAGE_OFFSET)
arch/powerpc/mm/ptdump/hashpagetable.c
485
for (addr = PAGE_OFFSET; addr < PAGE_OFFSET +
arch/powerpc/mm/ptdump/hashpagetable.c
509
address_markers[0].start_address = PAGE_OFFSET;
arch/powerpc/mm/ptdump/hashpagetable.c
525
.start_address = PAGE_OFFSET,
arch/powerpc/mm/ptdump/ptdump.c
263
address_markers[i++].start_address = PAGE_OFFSET;
arch/powerpc/mm/ptdump/ptdump.c
412
ptdump_range[0].start = PAGE_OFFSET;
arch/powerpc/mm/ptdump/ptdump.c
414
ptdump_range[0].end = PAGE_OFFSET + (PGDIR_SIZE * PTRS_PER_PGD);
arch/powerpc/net/bpf_jit_comp64.c
1549
PPC_LI64(tmp2_reg, PAGE_OFFSET);
arch/powerpc/platforms/powermac/smp.c
800
unsigned int *vector = (unsigned int *)(PAGE_OFFSET+0x100);
arch/riscv/include/asm/page.h
126
((x) >= PAGE_OFFSET && (!IS_ENABLED(CONFIG_64BIT) || (x) < PAGE_OFFSET + KERN_VIRT_SIZE))
arch/riscv/include/asm/page.h
204
(unsigned long)(_addr) >= PAGE_OFFSET && pfn_valid(virt_to_pfn(_addr)); \
arch/riscv/include/asm/pgtable.h
29
#define KERNEL_LINK_ADDR PAGE_OFFSET
arch/riscv/include/asm/pgtable.h
44
#define VMALLOC_END PAGE_OFFSET
arch/riscv/include/asm/pgtable.h
45
#define VMALLOC_START (PAGE_OFFSET - VMALLOC_SIZE)
arch/riscv/include/asm/pgtable.h
52
#define BPF_JIT_REGION_START (PAGE_OFFSET - BPF_JIT_REGION_SIZE)
arch/riscv/kernel/hibernate.c
357
unsigned long start = PAGE_OFFSET;
arch/riscv/kernel/vmcore_info.c
10
vmcoreinfo_append_str("NUMBER(PAGE_OFFSET)=0x%lx\n", PAGE_OFFSET);
arch/riscv/mm/init.c
1151
0UL : PAGE_OFFSET - kernel_map.phys_addr;
arch/riscv/mm/init.c
1156
BUG_ON((PAGE_OFFSET % PGDIR_SIZE) != 0);
arch/riscv/mm/init.c
1312
if (start <= __pa(PAGE_OFFSET) &&
arch/riscv/mm/init.c
1313
__pa(PAGE_OFFSET) < end)
arch/riscv/mm/init.c
1314
start = __pa(PAGE_OFFSET);
arch/riscv/mm/init.c
1525
preallocate_pgd_pages_range(PAGE_OFFSET, PAGE_END, "direct map");
arch/riscv/mm/init.c
154
print_ml("lowmem", (unsigned long)PAGE_OFFSET,
arch/riscv/mm/init.c
1797
mhp_range.start = __pa(PAGE_OFFSET);
arch/riscv/mm/init.c
257
kernel_map.va_pa_offset = PAGE_OFFSET - phys_ram_base;
arch/riscv/mm/init.c
264
max_mapped_addr = __pa(PAGE_OFFSET) + KERN_VIRT_SIZE;
arch/riscv/mm/ptdump.c
436
address_markers[PAGE_OFFSET_NR].start_address = PAGE_OFFSET;
arch/sh/include/asm/barrier.h
31
#define ctrl_barrier() __icbi(PAGE_OFFSET)
arch/sh/include/asm/page.h
122
#define ___pa(x) ((x)-PAGE_OFFSET+__MEMORY_START)
arch/sh/include/asm/page.h
123
#define ___va(x) ((x)+PAGE_OFFSET-__MEMORY_START)
arch/sh/include/asm/page.h
125
#define ___pa(x) ((x)-PAGE_OFFSET)
arch/sh/include/asm/page.h
126
#define ___va(x) ((x)+PAGE_OFFSET)
arch/sh/include/asm/page.h
139
#define UNCAC_ADDR(addr) ((addr) - PAGE_OFFSET + uncached_start)
arch/sh/include/asm/page.h
140
#define CAC_ADDR(addr) ((addr) - uncached_start + PAGE_OFFSET)
arch/sh/kernel/traps.c
121
if (addr < PAGE_OFFSET)
arch/sparc/include/asm/page_32.h
122
#define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + phys_base)
arch/sparc/include/asm/page_32.h
123
#define __va(x) ((void *)((unsigned long) (x) - phys_base + PAGE_OFFSET))
arch/sparc/include/asm/page_32.h
131
#define virt_addr_valid(kaddr) ((((unsigned long)(kaddr)-PAGE_OFFSET)>>PAGE_SHIFT) < max_mapnr)
arch/sparc/include/asm/page_64.h
130
extern unsigned long PAGE_OFFSET;
arch/sparc/include/asm/page_64.h
145
#define __pa(x) ((unsigned long)(x) - PAGE_OFFSET)
arch/sparc/include/asm/page_64.h
146
#define __va(x) ((void *)((unsigned long) (x) + PAGE_OFFSET))
arch/sparc/include/asm/pgtable_32.h
50
#define USER_PTRS_PER_PGD PAGE_OFFSET / PGDIR_SIZE
arch/sparc/include/asm/processor_32.h
20
#define TASK_SIZE PAGE_OFFSET
arch/sparc/include/asm/processor_32.h
22
#define STACK_TOP (PAGE_OFFSET - PAGE_SIZE)
arch/sparc/kernel/sun4d_smp.c
82
while ((unsigned long)current_set[cpuid] < PAGE_OFFSET)
arch/sparc/kernel/traps_32.c
81
(((unsigned long) rw) >= PAGE_OFFSET) &&
arch/sparc/kernel/traps_64.c
1465
unsigned long vaddr = PAGE_OFFSET + paddr;
arch/sparc/kernel/traps_64.c
948
flushi(PAGE_OFFSET);
arch/sparc/mm/fault_32.c
146
if (!from_user && address >= PAGE_OFFSET)
arch/sparc/mm/init_32.c
124
initrd_start = (initrd_start - phys_base) + PAGE_OFFSET;
arch/sparc/mm/init_32.c
125
initrd_end = (initrd_end - phys_base) + PAGE_OFFSET;
arch/sparc/mm/init_64.c
1761
unsigned long vstart = PAGE_OFFSET + pstart;
arch/sparc/mm/init_64.c
1762
unsigned long vend = PAGE_OFFSET + pend;
arch/sparc/mm/init_64.c
1914
flush_tsb_kernel_range(PAGE_OFFSET + phys_start,
arch/sparc/mm/init_64.c
1915
PAGE_OFFSET + phys_end);
arch/sparc/mm/init_64.c
1920
__flush_tlb_kernel_range(PAGE_OFFSET + phys_start,
arch/sparc/mm/init_64.c
1921
PAGE_OFFSET + phys_end);
arch/sparc/mm/init_64.c
1937
unsigned long PAGE_OFFSET;
arch/sparc/mm/init_64.c
1938
EXPORT_SYMBOL(PAGE_OFFSET);
arch/sparc/mm/init_64.c
2010
PAGE_OFFSET = sparc64_va_hole_top;
arch/sparc/mm/init_64.c
2015
PAGE_OFFSET, max_phys_bits);
arch/sparc/mm/init_64.c
2223
PAGE_OFFSET;
arch/sparc/mm/init_64.c
2232
PAGE_OFFSET;
arch/sparc/mm/init_64.c
2241
PAGE_OFFSET;
arch/sparc/mm/init_64.c
2386
shift = kern_base + PAGE_OFFSET - ((unsigned long)KERNBASE);
arch/sparc/mm/init_64.c
2674
kern_linear_pte_xor[0] = _PAGE_VALID ^ PAGE_OFFSET;
arch/sparc/mm/init_64.c
2677
PAGE_OFFSET;
arch/sparc/mm/init_64.c
2721
kern_linear_pte_xor[0] = _PAGE_VALID ^ PAGE_OFFSET;
arch/sparc/mm/init_64.c
2724
PAGE_OFFSET;
arch/sparc/mm/init_64.c
512
if (kaddr >= PAGE_OFFSET)
arch/sparc/mm/init_64.c
903
initrd_start += PAGE_OFFSET;
arch/sparc/mm/init_64.c
904
initrd_end += PAGE_OFFSET;
arch/sparc/mm/srmmu.c
856
const unsigned long min_vaddr = PAGE_OFFSET;
arch/sparc/mm/srmmu.c
857
const unsigned long max_vaddr = PAGE_OFFSET + SRMMU_MAXMEM;
arch/sparc/mm/srmmu.c
877
do_large_mapping(PAGE_OFFSET, phys_base);
arch/um/include/asm/page.h
74
#define KERNELBASE PAGE_OFFSET
arch/x86/coco/sev/core.c
558
addr = PAGE_OFFSET;
arch/x86/coco/sev/core.c
559
end = PAGE_OFFSET + get_max_mapped();
arch/x86/coco/tdx/tdx.c
1037
addr = PAGE_OFFSET;
arch/x86/coco/tdx/tdx.c
1038
end = PAGE_OFFSET + get_max_mapped();
arch/x86/events/perf_event.h
1349
return ip > PAGE_OFFSET;
arch/x86/include/asm/dma.h
81
#define MAX_DMA_ADDRESS (PAGE_OFFSET + 0x1000000)
arch/x86/include/asm/io.h
236
#define __ISA_IO_base ((char __iomem *)(PAGE_OFFSET))
arch/x86/include/asm/kmsan.h
76
x = y + (__START_KERNEL_map - PAGE_OFFSET);
arch/x86/include/asm/page.h
52
#define __va(x) ((void *)((unsigned long)(x)+PAGE_OFFSET))
arch/x86/include/asm/page_32.h
9
#define __phys_addr_nodebug(x) ((x) - PAGE_OFFSET)
arch/x86/include/asm/page_64.h
28
x = y + ((x > y) ? phys_base : (__START_KERNEL_map - PAGE_OFFSET));
arch/x86/include/asm/pgtable.h
1172
#define KERNEL_PGD_BOUNDARY pgd_index(PAGE_OFFSET)
arch/x86/include/asm/pgtable_32_areas.h
51
#define MAXMEM (VMALLOC_END - PAGE_OFFSET - __VMALLOC_RESERVE)
arch/x86/kernel/cpu/microcode/core.c
256
start += PAGE_OFFSET;
arch/x86/kernel/doublefault_32.c
14
#define ptr_ok(x) ((x) > PAGE_OFFSET && (x) < PAGE_OFFSET + MAXMEM)
arch/x86/kernel/head32.c
115
*(*pl2p + ((PAGE_OFFSET >> PGDIR_SHIFT))) = pl2;
arch/x86/kernel/head32.c
145
*ptr = (unsigned long)ptep + PAGE_OFFSET;
arch/x86/kernel/head32.c
154
*ptr = (pte.pte & PTE_PFN_MASK) + PAGE_OFFSET;
arch/x86/kernel/head32.c
159
*ptr = (unsigned long)pl2p + PAGE_OFFSET;
arch/x86/kernel/head32.c
168
*ptr = (unsigned long)pl2p + PAGE_OFFSET;
arch/x86/kernel/head32.c
45
*(pl2p + ((PAGE_OFFSET >> PGDIR_SHIFT))) = (pte_t) {.pte = 0};
arch/x86/kernel/setup.c
333
initrd_start = relocated_ramdisk + PAGE_OFFSET;
arch/x86/kernel/setup.c
381
initrd_start = ramdisk_image + PAGE_OFFSET;
arch/x86/mm/dump_pagetables.c
133
[KERNEL_SPACE_NR] = { PAGE_OFFSET, "Kernel Mapping" },
arch/x86/mm/dump_pagetables.c
238
if (pcibios_enabled && st->start_address >= PAGE_OFFSET + BIOS_BEGIN &&
arch/x86/mm/dump_pagetables.c
239
addr <= PAGE_OFFSET + BIOS_END) {
arch/x86/mm/dump_pagetables.c
520
address_markers[LOW_KERNEL_NR].start_address = PAGE_OFFSET;
arch/x86/mm/init_32.c
291
pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
arch/x86/mm/init_32.c
299
pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
arch/x86/mm/init_32.c
306
unsigned int addr = pfn * PAGE_SIZE + PAGE_OFFSET;
arch/x86/mm/init_32.c
325
PAGE_OFFSET + PAGE_SIZE-1;
arch/x86/mm/init_32.c
342
pte_ofs = pte_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
arch/x86/mm/init_32.c
437
va = PAGE_OFFSET + (pfn<<PAGE_SHIFT);
arch/x86/mm/kasan_init_64.c
376
kasan_mem_to_shadow((void *)PAGE_OFFSET));
arch/x86/mm/kasan_init_64.c
391
kasan_mem_to_shadow((void *)PAGE_OFFSET + MAXMEM),
arch/x86/mm/pat/set_memory.c
1811
if (within(vaddr, PAGE_OFFSET,
arch/x86/mm/pat/set_memory.c
1812
PAGE_OFFSET + (max_pfn_mapped << PAGE_SHIFT))) {
arch/x86/mm/pat/set_memory.c
1927
if (!(within(vaddr, PAGE_OFFSET,
arch/x86/mm/pat/set_memory.c
1928
PAGE_OFFSET + (max_pfn_mapped << PAGE_SHIFT)))) {
arch/x86/mm/pat/set_memory.c
2267
decoy_addr = (pfn << PAGE_SHIFT) + (PAGE_OFFSET ^ BIT(63));
arch/x86/mm/physaddr.c
25
x = y + (__START_KERNEL_map - PAGE_OFFSET);
arch/x86/mm/physaddr.c
47
x = y + (__START_KERNEL_map - PAGE_OFFSET);
arch/x86/mm/physaddr.c
63
unsigned long phys_addr = x - PAGE_OFFSET;
arch/x86/mm/physaddr.c
65
VIRTUAL_BUG_ON(x < PAGE_OFFSET);
arch/x86/mm/physaddr.c
79
if (x < PAGE_OFFSET)
arch/x86/mm/physaddr.c
85
return pfn_valid((x - PAGE_OFFSET) >> PAGE_SHIFT);
arch/x86/mm/pti.c
183
if (address < PAGE_OFFSET) {
arch/x86/pci/pcbios.c
153
pci_indirect.address = pcibios_entry + PAGE_OFFSET;
arch/x86/pci/pcbios.c
345
bios32_indirect.address = bios32_entry + PAGE_OFFSET;
arch/x86/pci/pcbios.c
61
set_memory_x(PAGE_OFFSET + BIOS_BEGIN, (BIOS_END - BIOS_BEGIN) >> PAGE_SHIFT);
arch/x86/platform/efi/efi_64.c
116
pgd_efi = efi_pgd + pgd_index(PAGE_OFFSET);
arch/x86/platform/efi/efi_64.c
117
pgd_k = pgd_offset_k(PAGE_OFFSET);
arch/x86/platform/efi/efi_64.c
119
num_entries = pgd_index(EFI_VA_END) - pgd_index(PAGE_OFFSET);
arch/x86/platform/olpc/olpc-xo1-pm.c
31
} ofw_bios_entry = { 0xF0000 + PAGE_OFFSET, __KERNEL_CS };
arch/x86/power/hibernate_32.c
89
pgd_idx = pgd_index(PAGE_OFFSET);
arch/xtensa/include/asm/dma.h
48
#define MAX_DMA_ADDRESS (PAGE_OFFSET + XCHAL_KIO_SIZE - 1)
arch/xtensa/include/asm/page.h
144
unsigned long off = va - PAGE_OFFSET;
arch/xtensa/include/asm/page.h
165
((unsigned long) (x) - PAGE_OFFSET + PHYS_OFFSET)
arch/xtensa/include/asm/page.h
168
((void *)((unsigned long) (x) - PHYS_OFFSET + PAGE_OFFSET))
arch/xtensa/mm/init.c
101
PAGE_OFFSET, PAGE_OFFSET +
arch/xtensa/mm/tlb.c
153
if (end > start && start >= TASK_SIZE && end <= PAGE_OFFSET &&
drivers/atm/fore200e.h
595
#define FORE200E_HDL2BUF(handle) ((struct buffer*)(((u64)(handle)) | PAGE_OFFSET))
drivers/misc/lkdtm/cfi.c
69
((__force __typeof__(addr))((uintptr_t)(addr) | PAGE_OFFSET))
drivers/misc/lkdtm/powerpc.c
72
unsigned long test_address = PAGE_OFFSET, *test_ptr;
drivers/of/fdt.c
1143
#define MIN_MEMBLOCK_ADDR __pa(PAGE_OFFSET)
drivers/pci/controller/pci-ixp4xx.c
579
addr = __pa(PAGE_OFFSET);
drivers/pci/controller/pci-versatile.c
131
writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_0);
drivers/pci/controller/pci-versatile.c
132
writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_1);
drivers/pci/controller/pci-versatile.c
133
writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_2);
drivers/pci/controller/pci-versatile.c
94
writel(__pa(PAGE_OFFSET) >> 28, PCI_SMAP(mem));
fs/proc/kcore.c
45
#define kc_vaddr_to_offset(v) ((v) - PAGE_OFFSET)
fs/proc/kcore.c
48
#define kc_offset_to_vaddr(o) ((o) + PAGE_OFFSET)
include/asm-generic/dma.h
11
#define MAX_DMA_ADDRESS PAGE_OFFSET
include/linux/kvm_host.h
143
#define KVM_HVA_ERR_BAD (PAGE_OFFSET)
include/linux/kvm_host.h
144
#define KVM_HVA_ERR_RO_BAD (PAGE_OFFSET + PAGE_SIZE)
include/linux/kvm_host.h
148
return addr >= PAGE_OFFSET;
include/linux/mm.h
4848
#define PP_DMA_INDEX_BITS ((__builtin_constant_p(PAGE_OFFSET) && \
include/linux/mm.h
4849
PAGE_OFFSET >= PP_DMA_INDEX_MIN_OFFSET && \
include/linux/mm.h
4850
!(PAGE_OFFSET & (PP_DMA_INDEX_MIN_OFFSET - 1))) ? \
include/linux/mm.h
4851
MIN(32, __ffs(PAGE_OFFSET) - PP_DMA_INDEX_SHIFT) : 0)
mm/highmem.c
617
WARN_ON_ONCE(addr < PAGE_OFFSET);
mm/pgtable-generic.c
434
iommu_sva_invalidate_kva_range(PAGE_OFFSET, TLB_FLUSH_ALL);
tools/testing/selftests/powerpc/mm/bad_accesses.c
60
(fault_addr < PAGE_OFFSET || fault_addr >= kernel_virt_end));
tools/testing/selftests/powerpc/mm/bad_accesses.c
92
kernel_virt_end = PAGE_OFFSET + (7 * (512ul << 40));
tools/testing/selftests/powerpc/mm/bad_accesses.c
97
kernel_virt_end = PAGE_OFFSET + (7 * (64ul << 40));
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
570
#define tophys(rd, rs) addis rd, rs, -PAGE_OFFSET@h
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
571
#define tovirt(rd, rs) addis rd, rs, PAGE_OFFSET@h