arch/alpha/mm/init.c
200
pfn_pte(pfn, PAGE_KERNEL));
arch/alpha/mm/init.c
57
= pte_val(mk_pte(virt_to_page(ret), PAGE_KERNEL));
arch/alpha/mm/init.c
84
(newptbr << 32) | pgprot_val(PAGE_KERNEL);
arch/arc/mm/ioremap.c
35
pgprot_noncached(PAGE_KERNEL));
arch/arm/include/asm/kfence.h
21
set_pte_ext(pte + i, pfn_pte(pfn + i, PAGE_KERNEL), 0);
arch/arm/mm/copypage-v6.c
117
set_top_pte(to, mk_pte(page, PAGE_KERNEL));
arch/arm/mm/copypage-v6.c
91
set_top_pte(kfrom, mk_pte(from, PAGE_KERNEL));
arch/arm/mm/copypage-v6.c
92
set_top_pte(kto, mk_pte(to, PAGE_KERNEL));
arch/arm/mm/dma-mapping.c
1076
pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL);
arch/arm/mm/dma-mapping.c
1741
__get_dma_pgprot(attrs, PAGE_KERNEL), false,
arch/arm/mm/dma-mapping.c
202
pgprot_t prot = pgprot_dmacoherent(PAGE_KERNEL);
arch/arm/mm/dma-mapping.c
433
__dma_remap(page, size, PAGE_KERNEL);
arch/arm/mm/fault-armv.c
250
pgprot_t prot = __pgprot_modify(PAGE_KERNEL,
arch/arm/mm/flush.c
43
set_top_pte(to, pfn_pte(pfn, PAGE_KERNEL));
arch/arm/mm/flush.c
58
set_top_pte(va, pfn_pte(pfn, PAGE_KERNEL));
arch/arm/mm/kasan_init.c
296
__pgprot(pgprot_val(PAGE_KERNEL)
arch/arm/mm/kasan_init.c
67
__pgprot(pgprot_val(PAGE_KERNEL)));
arch/arm64/include/asm/io.h
278
prot = __pgprot_modify(PAGE_KERNEL, PTE_ATTRINDX_MASK,
arch/arm64/include/asm/io.h
293
__ioremap_prot((addr), (size), PAGE_KERNEL)
arch/arm64/kernel/acpi.c
287
return PAGE_KERNEL;
arch/arm64/kernel/acpi.c
373
prot = PAGE_KERNEL;
arch/arm64/kernel/efi.c
64
: pgprot_val(PAGE_KERNEL);
arch/arm64/kernel/efi.c
80
return pgprot_val(PAGE_KERNEL);
arch/arm64/kernel/pi/map_kernel.c
211
fdt, PAGE_KERNEL, IDMAP_ROOT_LEVEL,
arch/arm64/kernel/pi/map_kernel.c
45
pgprot_t data_prot = PAGE_KERNEL;
arch/arm64/kernel/pi/map_range.c
95
pgprot_t data_prot = PAGE_KERNEL;
arch/arm64/kernel/setup.c
173
void *dt_virt = fixmap_remap_fdt(dt_phys, &size, PAGE_KERNEL);
arch/arm64/kvm/nested.c
1478
prot = PAGE_KERNEL;
arch/arm64/mm/init.c
525
.pgprot = PAGE_KERNEL,
arch/arm64/mm/init.c
539
.pgprot = PAGE_KERNEL,
arch/arm64/mm/kasan_init.c
128
__set_pte(ptep, pfn_pte(__phys_to_pfn(page_phys), PAGE_KERNEL));
arch/arm64/mm/mmu.c
1078
pgprot_tagged(PAGE_KERNEL),
arch/arm64/mm/mmu.c
1164
__map_memblock(pgdp, start, end, pgprot_tagged(PAGE_KERNEL),
arch/arm64/mm/mmu.c
1179
PAGE_KERNEL, NO_CONT_MAPPINGS);
arch/arm64/mm/mmu.c
1279
KPTI_NG_TEMP_VA, PAGE_SIZE, PAGE_KERNEL,
arch/arm64/mm/mmu.c
1395
__pi_map_range(&ptep, pa, pa + sizeof(u32), pa, PAGE_KERNEL,
arch/csky/mm/init.c
76
set_pte(&kernel_pte_tables[i - PFN_DOWN(va_pa_offset)], pfn_pte(i, PAGE_KERNEL));
arch/csky/mm/tcm.c
47
set_pte(tcm_pte, pfn_pte(__phys_to_pfn(paddr), PAGE_KERNEL));
arch/csky/mm/tcm.c
66
set_pte(tcm_pte, pfn_pte(__phys_to_pfn(paddr), PAGE_KERNEL));
arch/hexagon/kernel/vdso.c
29
vdso = vmap(&vdso_page, 1, 0, PAGE_KERNEL);
arch/loongarch/include/asm/io.h
64
ioremap_prot((offset), (size), PAGE_KERNEL)
arch/loongarch/include/asm/kfence.h
33
virt_to_phys((void *)kfence_pool), PAGE_KERNEL);
arch/loongarch/kernel/setup.c
496
vmap_page_range(vaddr, vaddr + size, hw_start, pgprot_device(PAGE_KERNEL));
arch/loongarch/mm/init.c
111
entry = pfn_pmd(virt_to_pfn(p), PAGE_KERNEL);
arch/loongarch/mm/init.c
233
.pgprot = PAGE_KERNEL,
arch/loongarch/mm/kasan_init.c
183
set_pte(ptep, pfn_pte(__phys_to_pfn(page_phys), PAGE_KERNEL));
arch/loongarch/mm/pageattr.c
208
return __set_memory(addr, 1, PAGE_KERNEL, __pgprot(0));
arch/loongarch/mm/pageattr.c
230
set = PAGE_KERNEL;
arch/m68k/mm/sun3kmap.c
33
ptep = pfn_pte(phys >> PAGE_SHIFT, PAGE_KERNEL);
arch/m68k/sun3/dvma.c
32
ptep = pfn_pte(virt_to_pfn((void *)kaddr), PAGE_KERNEL);
arch/m68k/sun3x/dvma.c
129
PAGE_KERNEL));
arch/mips/loongson64/init.c
183
vmap_page_range(vaddr, vaddr + size, hw_start, pgprot_device(PAGE_KERNEL));
arch/mips/mm/init.c
133
return __kmap_pgprot(page, addr, PAGE_KERNEL);
arch/mips/mm/init.c
576
.pgprot = PAGE_KERNEL,
arch/openrisc/include/asm/io.h
35
#define _PAGE_IOREMAP (pgprot_val(PAGE_KERNEL) | _PAGE_CI)
arch/openrisc/mm/init.c
106
prot = PAGE_KERNEL;
arch/parisc/mm/init.c
403
prot = PAGE_KERNEL;
arch/parisc/mm/init.c
411
PAGE_KERNEL_EXEC : PAGE_KERNEL;
arch/parisc/mm/init.c
414
prot = PAGE_KERNEL;
arch/parisc/mm/init.c
459
PAGE_KERNEL, 0);
arch/parisc/mm/init.c
472
PAGE_KERNEL, 1);
arch/parisc/mm/init.c
498
map_pages(start, __pa(start), end - start, PAGE_KERNEL, 0);
arch/parisc/mm/init.c
632
size, PAGE_KERNEL, 0);
arch/parisc/mm/init.c
639
initrd_end - initrd_start, PAGE_KERNEL, 0);
arch/powerpc/include/asm/io.h
829
ioremap_prot((addr), (size), PAGE_KERNEL)
arch/powerpc/kernel/isa-bridge.c
50
pgprot_noncached(PAGE_KERNEL)))
arch/powerpc/kernel/isa-bridge.c
54
pgprot_noncached(PAGE_KERNEL));
arch/powerpc/kernel/pci_64.c
142
pgprot_noncached(PAGE_KERNEL))) {
arch/powerpc/kvm/e500_mmu.c
789
virt = vmap(pages, num_pages, VM_MAP, PAGE_KERNEL);
arch/powerpc/lib/code-patching.c
243
return map_kernel_page(text_poke_addr, (pfn << PAGE_SHIFT), PAGE_KERNEL);
arch/powerpc/lib/code-patching.c
300
__set_pte_at(patching_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0);
arch/powerpc/lib/code-patching.c
339
__set_pte_at(&init_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0);
arch/powerpc/lib/code-patching.c
486
__set_pte_at(patching_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0);
arch/powerpc/lib/code-patching.c
531
__set_pte_at(&init_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0);
arch/powerpc/mm/book3s64/hash_pgtable.c
119
pgprot_val(PAGE_KERNEL),
arch/powerpc/mm/book3s64/hash_pgtable.c
563
pp = htab_convert_pte_flags(pgprot_val(PAGE_KERNEL), HPTE_USE_KERNEL_KEY);
arch/powerpc/mm/book3s64/hash_utils.c
1401
prot = pgprot_val(PAGE_KERNEL);
arch/powerpc/mm/book3s64/hash_utils.c
284
unsigned long mode = htab_convert_pte_flags(pgprot_val(PAGE_KERNEL), HPTE_USE_KERNEL_KEY);
arch/powerpc/mm/book3s64/hash_utils.c
452
unsigned long prot = pgprot_val(PAGE_KERNEL);
arch/powerpc/mm/book3s64/radix_pgtable.c
1008
entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL);
arch/powerpc/mm/book3s64/radix_pgtable.c
1056
entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL);
arch/powerpc/mm/book3s64/radix_pgtable.c
403
-1, PAGE_KERNEL, PAGE_SIZE))
arch/powerpc/mm/book3s64/radix_pgtable.c
447
-1, PAGE_KERNEL, ~0UL));
arch/powerpc/mm/book3s64/radix_pgtable.c
974
ret = __map_kernel_page_nid(start, phys, PAGE_KERNEL, page_size, nid);
arch/powerpc/mm/ioremap.c
13
pgprot_t prot = pgprot_noncached(PAGE_KERNEL);
arch/powerpc/mm/ioremap.c
22
pgprot_t prot = pgprot_noncached_wc(PAGE_KERNEL);
arch/powerpc/mm/ioremap.c
31
pgprot_t prot = pgprot_cached(PAGE_KERNEL);
arch/powerpc/mm/ioremap_32.c
11
pgprot_t prot = pgprot_cached_wthru(PAGE_KERNEL);
arch/powerpc/mm/kasan/8xx.c
30
pte_t pte = pte_mkhuge(pfn_pte(PHYS_PFN(__pa(block + i * PAGE_SIZE)), PAGE_KERNEL));
arch/powerpc/mm/kasan/8xx.c
68
pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
arch/powerpc/mm/kasan/book3s_32.c
31
setbat(idx, k_nobat, phys, k_size, PAGE_KERNEL);
arch/powerpc/mm/kasan/book3s_32.c
52
pte_t pte = pfn_pte(PHYS_PFN(phys + k_cur - k_nobat), PAGE_KERNEL);
arch/powerpc/mm/kasan/init_32.c
186
kasan_populate_pte(kasan_early_shadow_pte, PAGE_KERNEL);
arch/powerpc/mm/kasan/init_32.c
49
kasan_populate_pte(new, PAGE_KERNEL);
arch/powerpc/mm/kasan/init_32.c
75
pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
arch/powerpc/mm/kasan/init_book3e_64.c
79
kasan_map_kernel_page(k_cur, __pa(va), PAGE_KERNEL);
arch/powerpc/mm/kasan/init_book3e_64.c
87
pte_t zero_pte = pfn_pte(virt_to_pfn(kasan_early_shadow_page), PAGE_KERNEL);
arch/powerpc/mm/kasan/init_book3s_64.c
35
map_kernel_page(k_cur, __pa(va), PAGE_KERNEL);
arch/powerpc/mm/kasan/init_book3s_64.c
55
pte_t zero_pte = pfn_pte(virt_to_pfn(kasan_early_shadow_page), PAGE_KERNEL);
arch/powerpc/mm/mem.c
400
pgprot_t prot = strict_module_rwx_enabled() ? PAGE_KERNEL : PAGE_KERNEL_EXEC;
arch/powerpc/mm/mem.c
447
.pgprot = PAGE_KERNEL,
arch/powerpc/mm/nohash/8xx.c
160
mmu_mapin_ram_chunk(einittext8, top, PAGE_KERNEL, true);
arch/powerpc/mm/nohash/8xx.c
180
err = mmu_mapin_ram_chunk(boundary, einittext8, PAGE_KERNEL, false);
arch/powerpc/mm/nohash/e500.c
192
pgprot_t prot = init ? PAGE_KERNEL_X : PAGE_KERNEL;
arch/powerpc/mm/pgtable_32.c
107
map_kernel_page(v, p, ktext ? PAGE_KERNEL_X : PAGE_KERNEL);
arch/powerpc/perf/vpa-dtl.c
533
buf->base = vmap(pglist, nr_pages, VM_MAP, PAGE_KERNEL);
arch/powerpc/platforms/powernv/memtrace.c
199
struct mhp_params params = { .pgprot = PAGE_KERNEL };
arch/riscv/kernel/acpi.c
304
prot = PAGE_KERNEL;
arch/riscv/kernel/acpi.c
307
prot = pgprot_writecombine(PAGE_KERNEL);
arch/riscv/kernel/efi.c
25
return PAGE_KERNEL;
arch/riscv/kernel/efi.c
40
return PAGE_KERNEL;
arch/riscv/mm/init.c
1443
pmd_set_huge(pmd, virt_to_phys(p), PAGE_KERNEL);
arch/riscv/mm/init.c
1543
.pgprot = PAGE_KERNEL,
arch/riscv/mm/init.c
1555
.pgprot = PAGE_KERNEL,
arch/riscv/mm/init.c
792
return PAGE_KERNEL;
arch/riscv/mm/init.c
806
return PAGE_KERNEL;
arch/riscv/mm/init.c
956
PMD_SIZE, PAGE_KERNEL);
arch/riscv/mm/init.c
990
pa, MAX_FDT_SIZE, PAGE_KERNEL);
arch/riscv/mm/init.c
993
pa, PMD_SIZE, PAGE_KERNEL);
arch/riscv/mm/init.c
995
pa + PMD_SIZE, PMD_SIZE, PAGE_KERNEL);
arch/riscv/mm/kasan_init.c
101
set_pud(pudp, pfn_pud(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/kasan_init.c
132
set_p4d(p4dp, pfn_p4d(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/kasan_init.c
155
set_pgd(pgdp, pfn_pgd(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/kasan_init.c
331
pfn_pte(virt_to_pfn(kasan_early_shadow_page), PAGE_KERNEL));
arch/riscv/mm/kasan_init.c
44
set_pte(ptep, pfn_pte(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/kasan_init.c
70
set_pmd(pmdp, pfn_pmd(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/pageattr.c
386
PAGE_KERNEL, __pgprot(_PAGE_EXEC));
arch/riscv/mm/pageattr.c
394
set = PAGE_KERNEL;
arch/s390/boot/vmem.c
345
entry = set_pte_bit(entry, PAGE_KERNEL);
arch/s390/include/asm/io.h
33
#define _PAGE_IOREMAP pgprot_val(PAGE_KERNEL)
arch/s390/include/asm/io.h
36
ioremap_prot((addr), (size), pgprot_writecombine(PAGE_KERNEL))
arch/s390/include/asm/pgtable.h
947
pte_flags |= pgprot_val(PAGE_KERNEL);
arch/s390/kernel/abs_lowcore.c
16
rc = __vmem_map_4k_page(addr, phys, PAGE_KERNEL, alloc);
arch/s390/mm/init.c
267
if (WARN_ON_ONCE(pgprot_val(params->pgprot) != pgprot_val(PAGE_KERNEL)))
arch/s390/mm/init.c
310
.pgprot = PAGE_KERNEL,
arch/s390/mm/pageattr.c
103
new = set_pte_bit(new, PAGE_KERNEL);
arch/s390/mm/pageattr.c
126
prot = pgprot_val(ro ? PAGE_KERNEL_RO : PAGE_KERNEL);
arch/s390/mm/vmem.c
176
prot = pgprot_val(PAGE_KERNEL);
arch/sh/include/asm/io.h
282
ioremap_prot((addr), (size), PAGE_KERNEL)
arch/sh/kernel/setup.c
233
PAGE_KERNEL);
arch/sh/mm/kmap.c
46
set_pte(kmap_coherent_pte - idx, mk_pte(page, PAGE_KERNEL));
arch/sh/mm/numa.c
34
PAGE_KERNEL);
arch/sparc/include/asm/pgtable_64.h
200
extern pgprot_t PAGE_KERNEL;
arch/sparc/mm/execmem.c
14
.pgprot = PAGE_KERNEL,
arch/sparc/mm/init_64.c
1891
PAGE_KERNEL, use_huge);
arch/sparc/mm/init_64.c
1912
(enable ? PAGE_KERNEL : __pgprot(0)), false);
arch/sparc/mm/init_64.c
2550
pgprot_t PAGE_KERNEL __read_mostly;
arch/sparc/mm/init_64.c
2551
EXPORT_SYMBOL(PAGE_KERNEL);
arch/sparc/mm/init_64.c
2657
PAGE_KERNEL = __pgprot (_PAGE_PRESENT_4U | _PAGE_VALID |
arch/sparc/mm/init_64.c
2710
PAGE_KERNEL = __pgprot (_PAGE_PRESENT_4V | _PAGE_VALID |
arch/sparc/mm/init_64.c
2714
PAGE_KERNEL_LOCKED = PAGE_KERNEL;
arch/x86/events/intel/ds.c
883
ds_update_cea(cea, buffer, bsiz, PAGE_KERNEL);
arch/x86/events/intel/ds.c
930
ds_update_cea(cea, buffer, BTS_BUFFER_SIZE, PAGE_KERNEL);
arch/x86/kernel/alternative.c
2570
pgprot = __pgprot(pgprot_val(PAGE_KERNEL) & ~_PAGE_GLOBAL);
arch/x86/kernel/amd_gart_64.c
752
PAGE_KERNEL);
arch/x86/kernel/irq_64.c
49
va = vmap(pages, IRQ_STACK_SIZE / PAGE_SIZE, VM_MAP, PAGE_KERNEL);
arch/x86/mm/cpu_entry_area.c
118
PAGE_KERNEL);
arch/x86/mm/cpu_entry_area.c
136
estacks->name## _stack, npages, PAGE_KERNEL); \
arch/x86/mm/cpu_entry_area.c
172
&per_cpu(doublefault_stack, cpu), 1, PAGE_KERNEL);
arch/x86/mm/cpu_entry_area.c
192
pgprot_t gdt_prot = PAGE_KERNEL;
arch/x86/mm/cpu_entry_area.c
193
pgprot_t tss_prot = PAGE_KERNEL;
arch/x86/mm/cpu_entry_area.c
203
PAGE_KERNEL);
arch/x86/mm/init.c
1082
pgprot = PAGE_KERNEL;
arch/x86/mm/init.c
1113
.pgprot = PAGE_KERNEL,
arch/x86/mm/init.c
1120
.pgprot = PAGE_KERNEL,
arch/x86/mm/init.c
590
init_memory_mapping(start, end, PAGE_KERNEL);
arch/x86/mm/init.c
773
init_memory_mapping(0, ISA_END_ADDRESS, PAGE_KERNEL);
arch/x86/mm/init_32.c
346
pgprot_t prot = PAGE_KERNEL;
arch/x86/mm/init_64.c
818
page_size_mask, PAGE_KERNEL,
arch/x86/mm/kasan_init_64.c
49
if (p && pmd_set_huge(pmd, __pa(p), PAGE_KERNEL))
arch/x86/mm/kasan_init_64.c
67
entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL);
arch/x86/mm/kasan_init_64.c
85
if (p && pud_set_huge(pud, __pa(p), PAGE_KERNEL))
arch/x86/mm/pti.c
471
*target_pte = pfn_pte(pa >> PAGE_SHIFT, PAGE_KERNEL);
arch/x86/xen/enlighten_pv.c
530
set_aliased_prot(ldt + i, PAGE_KERNEL);
arch/x86/xen/grant-table.c
46
mfn_pte(frames[i], PAGE_KERNEL));
arch/x86/xen/grant-table.c
68
mfn_pte(frames[i], PAGE_KERNEL));
arch/x86/xen/mmu_pv.c
1645
__set_pfn_prot(pfn, PAGE_KERNEL);
arch/x86/xen/mmu_pv.c
1740
set_page_prot_flags((void *)addr, PAGE_KERNEL, UVMF_INVLPG);
arch/x86/xen/mmu_pv.c
1745
set_page_prot_flags((void *)addr, PAGE_KERNEL, UVMF_INVLPG);
arch/x86/xen/mmu_pv.c
2007
PAGE_KERNEL);
arch/x86/xen/mmu_pv.c
2108
pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
arch/x86/xen/mmu_pv.c
2118
pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
arch/x86/xen/mmu_pv.c
2295
mfn_pte(mfn, PAGE_KERNEL), flags);
arch/x86/xen/mmu_pv.c
899
pfn_pte(pfn, PAGE_KERNEL),
arch/x86/xen/p2m.c
376
pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL));
arch/x86/xen/p2m.c
595
pfn_pte(PFN_DOWN(__pa(p2m)), PAGE_KERNEL));
arch/x86/xen/setup.c
270
mfn_pte(mfn, PAGE_KERNEL), 0)) {
arch/x86/xen/setup.c
310
set_pte_mfn(buf, mfn, PAGE_KERNEL);
arch/x86/xen/setup.c
329
set_pte_mfn(buf, mfn_save, PAGE_KERNEL);
arch/x86/xen/setup.c
466
set_pte_mfn(buf, xen_remap_mfn, PAGE_KERNEL);
arch/x86/xen/setup.c
492
set_pte_mfn(buf, mfn_save, PAGE_KERNEL);
arch/xtensa/include/asm/io.h
43
pgprot_noncached(PAGE_KERNEL));
arch/xtensa/include/asm/io.h
54
return ioremap_prot(offset, size, PAGE_KERNEL);
arch/xtensa/mm/kasan_init.c
27
PAGE_KERNEL));
arch/xtensa/mm/kasan_init.c
58
set_pte(pte + j, pfn_pte(PHYS_PFN(phys), PAGE_KERNEL));
drivers/accel/amdxdna/amdxdna_ubuf.c
113
kva = vmap(ubuf->pages, ubuf->nr_pages, VM_MAP, PAGE_KERNEL);
drivers/accel/ivpu/ivpu_mmu_context.c
63
cpu = vmap(&page, 1, VM_MAP, pgprot_writecombine(PAGE_KERNEL));
drivers/comedi/comedi_buf.c
18
#define COMEDI_PAGE_PROTECTION PAGE_KERNEL
drivers/dma-buf/heaps/cma_heap.c
217
vaddr = vmap(buffer->pages, buffer->pagecount, VM_MAP, PAGE_KERNEL);
drivers/dma-buf/heaps/system_heap.c
238
vaddr = vmap(pages, npages, VM_MAP, PAGE_KERNEL);
drivers/firewire/ohci.c
875
vaddr = vmap(pages, ARRAY_SIZE(pages), VM_MAP, PAGE_KERNEL);
drivers/firmware/efi/capsule-loader.c
120
VM_MAP, PAGE_KERNEL);
drivers/firmware/efi/earlycon.c
63
fb_prot = fb_wb ? PAGE_KERNEL : pgprot_writecombine(PAGE_KERNEL);
drivers/gpu/drm/drm_gem_shmem_helper.c
378
pgprot_t prot = PAGE_KERNEL;
drivers/gpu/drm/etnaviv/etnaviv_gem.c
356
prot = PAGE_KERNEL;
drivers/gpu/drm/etnaviv/etnaviv_gem.c
359
prot = pgprot_noncached(PAGE_KERNEL);
drivers/gpu/drm/etnaviv/etnaviv_gem.c
363
prot = pgprot_writecombine(PAGE_KERNEL);
drivers/gpu/drm/i915/gem/i915_gem_pages.c
308
pgprot = PAGE_KERNEL;
drivers/gpu/drm/i915/gt/shmem_utils.c
77
vaddr = vmap(pages, n_pages, VM_MAP_PUT_PAGES, PAGE_KERNEL);
drivers/gpu/drm/imagination/pvr_mmu.c
273
prot = PAGE_KERNEL;
drivers/gpu/drm/imagination/pvr_vm_mips.c
67
pgprot_writecombine(PAGE_KERNEL));
drivers/gpu/drm/lima/lima_sched.c
366
VM_MAP, pgprot_writecombine(PAGE_KERNEL));
drivers/gpu/drm/msm/msm_gem.c
732
VM_MAP, msm_gem_pgprot(msm_obj, PAGE_KERNEL));
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
210
pgprot_writecombine(PAGE_KERNEL));
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
137
*pmap = vmap(mem->mem, mem->pages, VM_MAP, PAGE_KERNEL);
drivers/gpu/drm/omapdrm/omap_gem.c
1082
VM_MAP, pgprot_writecombine(PAGE_KERNEL));
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
140
pgprot_writecombine(PAGE_KERNEL));
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
525
pgprot_writecombine(PAGE_KERNEL));
drivers/gpu/drm/tegra/fbdev.c
124
pgprot_writecombine(PAGE_KERNEL));
drivers/gpu/drm/tegra/gem.c
196
pgprot_writecombine(PAGE_KERNEL));
drivers/gpu/drm/ttm/ttm_bo_util.c
374
prot = ttm_io_prot(bo, mem, PAGE_KERNEL);
drivers/gpu/drm/ttm/ttm_bo_util.c
544
prot = ttm_io_prot(bo, mem, PAGE_KERNEL);
drivers/gpu/drm/ttm/ttm_tt.c
524
iter_tt->prot = ttm_prot_from_caching(tt->caching, PAGE_KERNEL);
drivers/gpu/drm/ttm/ttm_tt.c
526
iter_tt->prot = PAGE_KERNEL;
drivers/gpu/drm/v3d/v3d_bo.c
199
pgprot_writecombine(PAGE_KERNEL));
drivers/gpu/drm/vmwgfx/vmwgfx_blit.c
617
d.dst_prot = ttm_io_prot(dst, dst->resource, PAGE_KERNEL);
drivers/gpu/drm/vmwgfx/vmwgfx_blit.c
618
d.src_prot = ttm_io_prot(src, src->resource, PAGE_KERNEL);
drivers/gpu/drm/xen/xen_drm_front_gem.c
293
VM_MAP, PAGE_KERNEL);
drivers/hv/ring_buffer.c
210
confidential ? PAGE_KERNEL : pgprot_decrypted(PAGE_KERNEL));
drivers/hwtracing/coresight/coresight-tmc-etr.c
286
PAGE_KERNEL);
drivers/hwtracing/coresight/coresight-tmc-etr.c
306
PAGE_KERNEL);
drivers/hwtracing/coresight/coresight-trbe.c
763
buf->trbe_base = (unsigned long)vmap(pglist, nr_pages, VM_MAP, PAGE_KERNEL);
drivers/hwtracing/ptt/hisi_ptt.c
1057
buf->base = vmap(pagelist, nr_pages, VM_MAP, PAGE_KERNEL);
drivers/iommu/dma-iommu.c
1009
pgprot_t prot = dma_pgprot(dev, PAGE_KERNEL, attrs);
drivers/iommu/dma-iommu.c
1076
return vmap(sgt_handle(sgt)->pages, count, VM_MAP, PAGE_KERNEL);
drivers/iommu/dma-iommu.c
1614
pgprot_t prot = dma_pgprot(dev, PAGE_KERNEL, attrs);
drivers/md/dm-pcache/cache_dev.c
51
*vaddr = vmap(pages, total_pages, VM_MAP, PAGE_KERNEL);
drivers/md/dm-writecache.c
327
wc->memory_map = vmap(pages, p, VM_MAP, PAGE_KERNEL);
drivers/media/pci/intel/ipu6/ipu6-dma.c
209
info->vaddr = vmap(pages, count, VM_USERMAP, PAGE_KERNEL);
drivers/misc/vmw_vmci/vmci_queue_pair.c
761
produce_q->q_header = vmap(headers, 2, VM_MAP, PAGE_KERNEL);
drivers/net/ethernet/google/gve/gve_tx.c
42
PAGE_KERNEL);
drivers/pci/pci.c
4049
pgprot_device(PAGE_KERNEL));
drivers/perf/arm_spe_pmu.c
1030
buf->base = vmap(pglist, nr_pages, VM_MAP, PAGE_KERNEL);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
913
cached ? PAGE_KERNEL : PAGE_KERNEL_NOCACHE);
drivers/staging/media/ipu3/ipu3-dmamap.c
127
map->vaddr = vmap(pages, count, VM_USERMAP, PAGE_KERNEL);
drivers/staging/media/ipu7/ipu7-dma.c
209
info->vaddr = vmap(pages, count, VM_USERMAP, PAGE_KERNEL);
drivers/target/target_core_transport.c
2752
cmd->t_data_vmap = vmap(pages, cmd->t_data_nents, VM_MAP, PAGE_KERNEL);
drivers/virt/acrn/mm.c
254
remap_vaddr = vmap(pages, nr_pages, VM_MAP, PAGE_KERNEL);
drivers/xen/mem-reservation.c
52
mfn_pte(frames[i], PAGE_KERNEL), 0);
drivers/xen/xenbus/xenbus_client.c
702
PAGE_KERNEL);
drivers/xen/xlate_mmu.c
248
vaddr = vmap(pages, nr_pages, 0, PAGE_KERNEL);
fs/erofs/zutil.c
96
ptr = vmap(tmp_pages, nrpages, VM_MAP, PAGE_KERNEL);
fs/ntfs3/frecord.c
2252
frame_mem = vmap(pages, pages_per_frame, VM_MAP, PAGE_KERNEL);
fs/pstore/ram_core.c
428
prot = PAGE_KERNEL;
fs/pstore/ram_core.c
431
prot = pgprot_noncached(PAGE_KERNEL);
fs/pstore/ram_core.c
434
prot = pgprot_writecombine(PAGE_KERNEL);
include/asm-generic/fixmap.h
48
#define FIXMAP_PAGE_NORMAL PAGE_KERNEL
include/linux/highmem-internal.h
32
#define kmap_prot PAGE_KERNEL
include/linux/io-mapping.h
141
iomap->prot = pgprot_writecombine(PAGE_KERNEL);
include/linux/pgtable.h
2138
# define PAGE_KERNEL_RO PAGE_KERNEL
include/linux/pgtable.h
2142
# define PAGE_KERNEL_EXEC PAGE_KERNEL
io_uring/memmap.c
125
ptr = vmap(mr->pages, mr->nr_pages, VM_MAP, PAGE_KERNEL);
kernel/bpf/arena.c
137
set_pte_at(&init_mm, addr, pte, mk_pte(page, PAGE_KERNEL));
kernel/bpf/ringbuf.c
139
VM_MAP | VM_USERMAP, PAGE_KERNEL);
kernel/bpf/syscall.c
405
gfp | GFP_KERNEL | __GFP_RETRY_MAYFAIL, PAGE_KERNEL,
kernel/crash_core.c
78
safecopy = vmap(vmcoreinfo_pages, nr_pages, VM_MAP, PAGE_KERNEL);
kernel/dma/direct.c
269
pgprot_t prot = dma_pgprot(dev, PAGE_KERNEL, attrs);
kernel/dma/pool.c
105
pgprot_decrypted(pgprot_dmacoherent(PAGE_KERNEL)),
kernel/kexec_core.c
1003
vaddr = vmap(src_pages, npages, VM_MAP, PAGE_KERNEL);
kernel/liveupdate/kexec_handover.c
1139
err = vmap_pages_range(addr, addr + size, PAGE_KERNEL, pages, shift);
kernel/module/decompress.c
317
info->hdr = vmap(info->pages, info->used_pages, VM_MAP, PAGE_KERNEL);
kernel/relay.c
126
mem = vmap(buf->page_array, n_pages, VM_MAP, PAGE_KERNEL);
kernel/scs.c
47
GFP_SCS, PAGE_KERNEL, 0, node,
kernel/trace/trace.c
9697
start, pgprot_nx(PAGE_KERNEL));
lib/alloc_tag.c
442
vmap_pages_range(phys_end, phys_end + (nr << PAGE_SHIFT), PAGE_KERNEL,
lib/tests/kunit_iov_iter.c
66
buffer = vmap(pages, npages, VM_MAP | VM_MAP_PUT_PAGES, PAGE_KERNEL);
mm/execmem.c
290
p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags);
mm/execmem.c
293
p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags);
mm/execmem.c
543
r->pgprot = PAGE_KERNEL;
mm/hugetlb_vmemmap.c
221
pgprot = PAGE_KERNEL;
mm/hugetlb_vmemmap.c
259
pgprot_t pgprot = PAGE_KERNEL;
mm/hugetlb_vmemmap.c
68
pgprot_t pgprot = PAGE_KERNEL;
mm/kasan/init.c
99
PAGE_KERNEL);
mm/kasan/kasan_test_c.c
1935
v_ptr = vmap(&p_page, 1, VM_MAP, PAGE_KERNEL);
mm/kasan/shadow.c
244
PAGE_KERNEL, VM_NO_GUARD,
mm/kasan/shadow.c
313
pte = pfn_pte(page_to_pfn(page), PAGE_KERNEL);
mm/kasan/shadow.c
679
PAGE_KERNEL, VM_NO_GUARD, NUMA_NO_NODE,
mm/kmsan/kmsan_test.c
328
vbuf = vmap(pages, npages, VM_MAP, PAGE_KERNEL);
mm/kmsan/shadow.c
243
prot = PAGE_KERNEL;
mm/memory_hotplug.c
1454
pgprot_mhp(PAGE_KERNEL) };
mm/memory_hotplug.c
1500
struct mhp_params params = { .pgprot = pgprot_mhp(PAGE_KERNEL) };
mm/memremap.c
271
.pgprot = PAGE_KERNEL,
mm/percpu-vm.c
197
PAGE_KERNEL, pages, PAGE_SHIFT, GFP_KERNEL);
mm/slub.c
6781
flags, PAGE_KERNEL, allow_block ? VM_ALLOW_HUGE_VMAP:0,
mm/sparse-vmemmap.c
181
entry = pfn_pte(ptpfn, PAGE_KERNEL);
mm/vmalloc.c
3092
if (vmap_pages_range(addr, addr + size, PAGE_KERNEL,
mm/vmalloc.c
4043
if (pgprot_val(prot) == pgprot_val(PAGE_KERNEL)) {
mm/vmalloc.c
4125
gfp_mask, PAGE_KERNEL, 0, node, caller);
mm/vmalloc.c
4182
gfp_mask, PAGE_KERNEL, VM_ALLOW_HUGE_VMAP,
mm/vmalloc.c
4219
GFP_KERNEL | __GFP_ZERO, PAGE_KERNEL,
mm/vmalloc.c
4412
GFP_VMALLOC32 | __GFP_ZERO, PAGE_KERNEL,
mm/vmalloc.c
760
return vmap_pages_range(start, end, PAGE_KERNEL, pages, PAGE_SHIFT);
net/xdp/xdp_umem.c
49
umem->addrs = vmap(pages, nr_pages, VM_MAP, PAGE_KERNEL);
sound/core/memalloc.c
763
p = vmap(sgbuf->pages, sgbuf->count, VM_MAP, PAGE_KERNEL);