Symbol: init_mm
arch/alpha/kernel/smp.c
147
mmgrab(&init_mm);
arch/alpha/kernel/smp.c
148
current->active_mm = &init_mm;
arch/alpha/mm/init.c
46
init = pgd_offset(&init_mm, 0UL);
arch/arc/kernel/smp.c
162
struct mm_struct *mm = &init_mm;
arch/arc/mm/highmem.c
60
pmd_populate_kernel(&init_mm, pmd_k, pte_k);
arch/arm/include/asm/kfence.h
15
pte_t *pte = pte_alloc_one_kernel(&init_mm);
arch/arm/include/asm/kfence.h
22
pmd_populate_kernel(&init_mm, pmd, pte);
arch/arm/include/asm/mmu_context.h
144
if (mm != &init_mm)
arch/arm/include/asm/mmu_context.h
31
atomic_read(&init_mm.context.vmalloc_seq)))
arch/arm/kernel/hibernate.c
85
cpu_switch_mm(idmap_pgd, &init_mm);
arch/arm/kernel/smp.c
412
struct mm_struct *mm = &init_mm;
arch/arm/kernel/traps.c
962
atomic_inc_return_release(&init_mm.context.vmalloc_seq);
arch/arm/mm/dma-mapping.c
324
apply_to_page_range(&init_mm, start, size, __dma_update_pte, &prot);
arch/arm/mm/dump.c
448
.mm = &init_mm,
arch/arm/mm/dump.c
464
walk_pgd(&st, &init_mm, 0);
arch/arm/mm/fault.c
46
mm = &init_mm;
arch/arm/mm/fault.c
522
pgd_k = init_mm.pgd + index;
arch/arm/mm/idmap.c
114
idmap_pgd = pgd_alloc(&init_mm);
arch/arm/mm/idmap.c
137
cpu_switch_mm(idmap_pgd, &init_mm);
arch/arm/mm/idmap.c
31
pmd = pmd_alloc_one(&init_mm, addr);
arch/arm/mm/idmap.c
43
pud_populate(&init_mm, pud, pmd);
arch/arm/mm/init.c
383
set_section_perms(perms, n, true, &init_mm);
arch/arm/mm/ioremap.c
144
seq = atomic_read_acquire(&init_mm.context.vmalloc_seq);
arch/arm/mm/ioremap.c
159
} while (seq != atomic_read(&init_mm.context.vmalloc_seq));
arch/arm/mm/ioremap.c
190
atomic_inc_return_release(&init_mm.context.vmalloc_seq);
arch/arm/mm/ioremap.c
196
pte_free_kernel(&init_mm, pmd_page_vaddr(pmd));
arch/arm/mm/kasan_init.c
114
pmd_populate_kernel(&init_mm, pmdp, p);
arch/arm/mm/kasan_init.c
146
pgd_populate(&init_mm, pgdp, p);
arch/arm/mm/kasan_init.c
239
cpu_switch_mm(tmp_pgd_table, &init_mm);
arch/arm/mm/kasan_init.c
293
set_pte_at(&init_mm, KASAN_SHADOW_START + i*PAGE_SIZE,
arch/arm/mm/kasan_init.c
299
cpu_switch_mm(swapper_pg_dir, &init_mm);
arch/arm/mm/kasan_init.c
86
set_pte_at(&init_mm, addr, ptep, entry);
arch/arm/mm/mmu.c
387
pmd_populate_kernel(&init_mm, pmd, bm_pte);
arch/arm/mm/mmu.c
992
__create_mapping(&init_mm, md, early_alloc, false);
arch/arm/mm/pageattr.c
47
ret = apply_to_page_range(&init_mm, start, size, change_page_range,
arch/arm64/include/asm/mmu_context.h
100
if (mm != &init_mm && !system_uses_ttbr0_pan())
arch/arm64/include/asm/mmu_context.h
110
cpu_switch_mm(lm_alias(idmap_pg_dir), &init_mm);
arch/arm64/include/asm/mmu_context.h
210
if (mm == &init_mm)
arch/arm64/include/asm/mmu_context.h
232
update_saved_ttbr0(tsk, &init_mm);
arch/arm64/include/asm/mmu_context.h
241
if (next == &init_mm) {
arch/arm64/include/asm/pgalloc.h
111
VM_BUG_ON(mm && mm != &init_mm);
arch/arm64/include/asm/pgalloc.h
119
VM_BUG_ON(mm == &init_mm);
arch/arm64/include/asm/pgalloc.h
33
pudval |= (mm == &init_mm) ? PUD_TABLE_UXN : PUD_TABLE_PXN;
arch/arm64/include/asm/pgalloc.h
55
p4dval |= (mm == &init_mm) ? P4D_TABLE_UXN : P4D_TABLE_PXN;
arch/arm64/include/asm/pgalloc.h
84
pgdval |= (mm == &init_mm) ? PGD_TABLE_UXN : PGD_TABLE_PXN;
arch/arm64/kernel/smp.c
198
struct mm_struct *mm = &init_mm;
arch/arm64/mm/contpte.c
21
return mm != &init_mm;
arch/arm64/mm/fault.c
119
if (mm == &init_mm)
arch/arm64/mm/fault.c
137
if (mm == &init_mm) {
arch/arm64/mm/fault.c
144
mm = &init_mm;
arch/arm64/mm/fault.c
152
mm == &init_mm ? "swapper" : "user", PAGE_SIZE / SZ_1K,
arch/arm64/mm/fixmap.c
132
__pte_clear(&init_mm, addr, ptep);
arch/arm64/mm/hugetlbpage.c
209
if (mm == &init_mm)
arch/arm64/mm/mmu.c
1012
early_create_pgd_mapping(init_mm.pgd, phys, virt, size, prot, NULL,
arch/arm64/mm/mmu.c
1460
__pte_clear(&init_mm, addr, ptep);
arch/arm64/mm/mmu.c
1862
mmap_read_lock(&init_mm);
arch/arm64/mm/mmu.c
1863
mmap_read_unlock(&init_mm);
arch/arm64/mm/mmu.c
1905
mmap_read_lock(&init_mm);
arch/arm64/mm/mmu.c
1906
mmap_read_unlock(&init_mm);
arch/arm64/mm/mmu.c
1930
WARN_ON(pgdir != init_mm.pgd);
arch/arm64/mm/mmu.c
563
return __pgd_pgtable_alloc(&init_mm, gfp, pgtable_type);
arch/arm64/mm/mmu.c
984
early_create_pgd_mapping(init_mm.pgd, phys, virt, size, prot, NULL,
arch/arm64/mm/mmu.c
994
BUG_ON(mm == &init_mm);
arch/arm64/mm/ptdump.c
335
.mm = &init_mm,
arch/arm64/mm/ptdump.c
363
arm64_ptdump_walk_pgd(&st.ptdump, &init_mm);
arch/csky/include/asm/pgalloc.h
49
init = pgd_offset(&init_mm, 0UL);
arch/csky/kernel/smp.c
244
struct mm_struct *mm = &init_mm;
arch/csky/mm/fault.c
139
pgd_k = init_mm.pgd + offset;
arch/hexagon/include/asm/mmu_context.h
40
next->pgd[l1] = init_mm.pgd[l1];
arch/hexagon/kernel/smp.c
143
mmgrab(&init_mm);
arch/hexagon/kernel/smp.c
144
current->active_mm = &init_mm;
arch/hexagon/mm/init.c
75
init_mm.context.ptbase = __pa(init_mm.pgd);
arch/loongarch/include/asm/mmu_context.h
98
if (next != &init_mm)
arch/loongarch/kernel/smp.c
624
init_new_context(current, &init_mm);
arch/loongarch/kernel/traps.c
1157
mmgrab(&init_mm);
arch/loongarch/kernel/traps.c
1158
current->active_mm = &init_mm;
arch/loongarch/kernel/traps.c
1160
enter_lazy_tlb(&init_mm, current);
arch/loongarch/mm/init.c
113
set_pmd_at(&init_mm, addr, pmd, entry);
arch/loongarch/mm/init.c
153
p4d_populate(&init_mm, p4d, pud);
arch/loongarch/mm/init.c
162
pud_populate(&init_mm, pud, pmd);
arch/loongarch/mm/init.c
173
pmd_populate_kernel(&init_mm, pmd, pte);
arch/loongarch/mm/init.c
197
pte_clear(&init_mm, addr, ptep);
arch/loongarch/mm/kasan_init.c
140
pud_populate(&init_mm, pudp, (pmd_t *)__va(pmd_phys));
arch/loongarch/mm/kasan_init.c
153
p4d_populate(&init_mm, p4dp, (pud_t *)__va(pud_phys));
arch/loongarch/mm/kasan_init.c
166
pgd_populate(&init_mm, pgdp, (p4d_t *)__va(p4d_phys));
arch/loongarch/mm/pageattr.c
120
mmap_write_lock(&init_mm);
arch/loongarch/mm/pageattr.c
122
mmap_write_unlock(&init_mm);
arch/loongarch/mm/pgtable.c
30
init = pgd_offset(&init_mm, 0UL);
arch/m68k/include/asm/mmu_context.h
115
mm = &init_mm;
arch/m68k/mm/kmap.c
260
pmd_dir = pmd_alloc(&init_mm, pud_dir, virtaddr);
arch/m68k/mm/mcfmmu.c
94
mm = (!user_mode(regs) && KMAPAREA(mmuar)) ? &init_mm : current->mm;
arch/m68k/sun3x/dvma.c
101
if((pmd = pmd_alloc(&init_mm, pud, vaddr)) == NULL) {
arch/microblaze/mm/pgtable.c
155
set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT,
arch/microblaze/mm/pgtable.c
236
mm = &init_mm;
arch/mips/kernel/cpu-probe.c
2083
set_cpu_context(i, &init_mm, 0);
arch/mips/kernel/smp.c
440
init_new_context(current, &init_mm);
arch/mips/kernel/traps.c
2269
mmgrab(&init_mm);
arch/mips/kernel/traps.c
2270
current->active_mm = &init_mm;
arch/mips/kernel/traps.c
2272
enter_lazy_tlb(&init_mm, current);
arch/mips/mm/fault.c
293
pgd_k = init_mm.pgd + offset;
arch/mips/mm/pgtable.c
17
init = pgd_offset(&init_mm, 0UL);
arch/nios2/mm/fault.c
237
pgd_k = init_mm.pgd + offset;
arch/nios2/mm/ioremap.c
100
pmd = pmd_alloc(&init_mm, pud, address);
arch/nios2/mm/ioremap.c
84
dir = pgd_offset(&init_mm, address);
arch/nios2/mm/ioremap.c
94
p4d = p4d_alloc(&init_mm, dir, address);
arch/nios2/mm/ioremap.c
97
pud = pud_alloc(&init_mm, p4d, address);
arch/nios2/mm/pgtable.c
60
init = pgd_offset(&init_mm, 0UL);
arch/openrisc/kernel/dma.c
74
mmap_write_lock(&init_mm);
arch/openrisc/kernel/dma.c
77
mmap_write_unlock(&init_mm);
arch/openrisc/kernel/dma.c
88
mmap_write_lock(&init_mm);
arch/openrisc/kernel/dma.c
92
mmap_write_unlock(&init_mm);
arch/openrisc/kernel/smp.c
111
current_pgd[cpu] = init_mm.pgd;
arch/openrisc/kernel/smp.c
126
struct mm_struct *mm = &init_mm;
arch/openrisc/mm/fault.c
309
pgd_k = init_mm.pgd + offset;
arch/openrisc/mm/init.c
136
current_pgd[smp_processor_id()] = init_mm.pgd;
arch/openrisc/mm/init.c
215
set_pte_at(&init_mm, va, pte, pfn_pte(pa >> PAGE_SHIFT, prot));
arch/openrisc/mm/init.c
217
pte_clear(&init_mm, va, pte);
arch/parisc/include/asm/mmu_context.h
90
BUG_ON(next == &init_mm); /* Should never happen */
arch/parisc/include/asm/pgalloc.h
37
if (mm == &init_mm)
arch/parisc/include/asm/tlbflush.h
38
BUG_ON(mm == &init_mm); /* Should never happen */
arch/parisc/kernel/pci-dma.c
176
pte_clear(&init_mm, vaddr, pte);
arch/parisc/kernel/smp.c
294
mmgrab(&init_mm);
arch/parisc/kernel/smp.c
295
current->active_mm = &init_mm;
arch/parisc/kernel/smp.c
297
enter_lazy_tlb(&init_mm, current);
arch/parisc/mm/fixmap.c
23
set_pte_at(&init_mm, vaddr, pte, __mk_pte(phys, PAGE_KERNEL_RWX));
arch/parisc/mm/fixmap.c
35
pte_clear(&init_mm, vaddr, pte);
arch/parisc/mm/init.c
690
pmd_populate_kernel(&init_mm, pmd, pte);
arch/powerpc/include/asm/book3s/32/tlbflush.h
74
flush_range(&init_mm, start, end);
arch/powerpc/include/asm/kfence.h
51
pte_update(&init_mm, addr, kpte, _PAGE_PRESENT, 0, 0);
arch/powerpc/include/asm/kfence.h
54
pte_update(&init_mm, addr, kpte, 0, _PAGE_PRESENT, 0);
arch/powerpc/include/asm/pgalloc.h
10
if (unlikely(mm == &init_mm))
arch/powerpc/include/asm/pte-walk.h
31
pgd_t *pgdir = init_mm.pgd;
arch/powerpc/kernel/mce_power.c
39
mm = &init_mm;
arch/powerpc/kernel/setup-common.c
983
VM_WARN_ON(cpumask_test_cpu(smp_processor_id(), mm_cpumask(&init_mm)));
arch/powerpc/kernel/setup-common.c
984
cpumask_set_cpu(smp_processor_id(), mm_cpumask(&init_mm));
arch/powerpc/kernel/setup-common.c
985
inc_mm_active_cpus(&init_mm);
arch/powerpc/kernel/setup-common.c
986
mm_iommu_init(&init_mm);
arch/powerpc/kernel/smp.c
1648
mmgrab_lazy_tlb(&init_mm);
arch/powerpc/kernel/smp.c
1649
current->active_mm = &init_mm;
arch/powerpc/kernel/smp.c
1650
VM_WARN_ON(cpumask_test_cpu(smp_processor_id(), mm_cpumask(&init_mm)));
arch/powerpc/kernel/smp.c
1651
cpumask_set_cpu(cpu, mm_cpumask(&init_mm));
arch/powerpc/kernel/smp.c
1652
inc_mm_active_cpus(&init_mm);
arch/powerpc/kernel/smp.c
1802
VM_WARN_ON_ONCE(!cpumask_test_cpu(cpu, mm_cpumask(&init_mm)));
arch/powerpc/kernel/smp.c
1803
dec_mm_active_cpus(&init_mm);
arch/powerpc/kernel/smp.c
1804
cpumask_clear_cpu(cpu, mm_cpumask(&init_mm));
arch/powerpc/lib/code-patching.c
277
pte_clear(&init_mm, addr, ptep);
arch/powerpc/lib/code-patching.c
339
__set_pte_at(&init_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0);
arch/powerpc/lib/code-patching.c
346
pte_clear(&init_mm, text_poke_addr, pte);
arch/powerpc/lib/code-patching.c
531
__set_pte_at(&init_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0);
arch/powerpc/lib/code-patching.c
538
pte_clear(&init_mm, text_poke_addr, pte);
arch/powerpc/mm/book3s32/kuap.c
11
init_mm.context.sr0 |= SR_KS;
arch/powerpc/mm/book3s32/tlb.c
102
mm = (vmaddr < TASK_SIZE)? vma->vm_mm: &init_mm;
arch/powerpc/mm/book3s64/hash_pgtable.c
160
pudp = pud_alloc(&init_mm, p4dp, ea);
arch/powerpc/mm/book3s64/hash_pgtable.c
163
pmdp = pmd_alloc(&init_mm, pudp, ea);
arch/powerpc/mm/book3s64/hash_pgtable.c
169
set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot));
arch/powerpc/mm/book3s64/hash_tlb.c
220
hpte_need_flush(&init_mm, start, ptep, pte, hugepage_shift);
arch/powerpc/mm/book3s64/hash_utils.c
1533
init_mm.context.hash_context = &init_hash_mm_context;
arch/powerpc/mm/book3s64/hash_utils.c
1534
mm_ctx_set_slb_addr_limit(&init_mm.context, SLB_ADDR_LIMIT_DEFAULT);
arch/powerpc/mm/book3s64/hash_utils.c
1947
mm = &init_mm;
arch/powerpc/mm/book3s64/hash_utils.c
1973
mm = &init_mm;
arch/powerpc/mm/book3s64/pgtable.c
420
if (mm == &init_mm)
arch/powerpc/mm/book3s64/radix_pgtable.c
100
pud_populate(&init_mm, pudp, pmdp);
arch/powerpc/mm/book3s64/radix_pgtable.c
1009
set_pte_at(&init_mm, addr, ptep, entry);
arch/powerpc/mm/book3s64/radix_pgtable.c
1057
set_pte_at(&init_mm, addr, pte, entry);
arch/powerpc/mm/book3s64/radix_pgtable.c
1072
p4d_populate(&init_mm, p4dp, pud);
arch/powerpc/mm/book3s64/radix_pgtable.c
1075
return pud_alloc(&init_mm, p4dp, address);
arch/powerpc/mm/book3s64/radix_pgtable.c
1089
pud_populate(&init_mm, pudp, pmd);
arch/powerpc/mm/book3s64/radix_pgtable.c
1091
return pmd_alloc(&init_mm, pudp, address);
arch/powerpc/mm/book3s64/radix_pgtable.c
110
pmd_populate_kernel(&init_mm, pmdp, ptep);
arch/powerpc/mm/book3s64/radix_pgtable.c
1105
pmd_populate(&init_mm, pmdp, pte);
arch/powerpc/mm/book3s64/radix_pgtable.c
115
set_pte_at(&init_mm, ea, ptep, pfn_pte(pfn, flags));
arch/powerpc/mm/book3s64/radix_pgtable.c
156
pudp = pud_alloc(&init_mm, p4dp, ea);
arch/powerpc/mm/book3s64/radix_pgtable.c
1623
set_pte_at_unchecked(&init_mm, 0 /* radix unused */, ptep, new_pud);
arch/powerpc/mm/book3s64/radix_pgtable.c
163
pmdp = pmd_alloc(&init_mm, pudp, ea);
arch/powerpc/mm/book3s64/radix_pgtable.c
1653
pte_free_kernel(&init_mm, pte);
arch/powerpc/mm/book3s64/radix_pgtable.c
1657
pmd_free(&init_mm, pmd);
arch/powerpc/mm/book3s64/radix_pgtable.c
1670
set_pte_at_unchecked(&init_mm, 0 /* radix unused */, ptep, new_pmd);
arch/powerpc/mm/book3s64/radix_pgtable.c
1694
pte_free_kernel(&init_mm, pte);
arch/powerpc/mm/book3s64/radix_pgtable.c
175
set_pte_at(&init_mm, ea, ptep, pfn_pte(pfn, flags));
arch/powerpc/mm/book3s64/radix_pgtable.c
207
pudp = pud_alloc(&init_mm, p4dp, idx);
arch/powerpc/mm/book3s64/radix_pgtable.c
214
pmdp = pmd_alloc(&init_mm, pudp, idx);
arch/powerpc/mm/book3s64/radix_pgtable.c
225
radix__pte_update(&init_mm, idx, ptep, clear, 0, 0);
arch/powerpc/mm/book3s64/radix_pgtable.c
472
process_tb->prtb0 = cpu_to_be64(rts_field | __pa(init_mm.pgd) | RADIX_PGD_INDEX_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
487
init_mm.context.id = mmu_base_pid;
arch/powerpc/mm/book3s64/radix_pgtable.c
497
dw0 = rts_field | __pa(init_mm.pgd) | RADIX_PGD_INDEX_SIZE | PATB_HR;
arch/powerpc/mm/book3s64/radix_pgtable.c
653
radix__switch_mmu_context(NULL, &init_mm);
arch/powerpc/mm/book3s64/radix_pgtable.c
671
radix__switch_mmu_context(NULL, &init_mm);
arch/powerpc/mm/book3s64/radix_pgtable.c
704
pte_free_kernel(&init_mm, pte_start);
arch/powerpc/mm/book3s64/radix_pgtable.c
719
pmd_free(&init_mm, pmd_start);
arch/powerpc/mm/book3s64/radix_pgtable.c
734
pud_free(&init_mm, pud_start);
arch/powerpc/mm/book3s64/radix_pgtable.c
806
pte_clear(&init_mm, addr, pte);
arch/powerpc/mm/book3s64/radix_pgtable.c
812
pte_clear(&init_mm, addr, pte);
arch/powerpc/mm/book3s64/radix_pgtable.c
840
pte_clear(&init_mm, addr, (pte_t *)pmd);
arch/powerpc/mm/book3s64/radix_pgtable.c
846
pte_clear(&init_mm, addr, (pte_t *)pmd);
arch/powerpc/mm/book3s64/radix_pgtable.c
881
pte_clear(&init_mm, addr, (pte_t *)pud);
arch/powerpc/mm/book3s64/radix_pgtable.c
90
p4d_populate(&init_mm, p4dp, pudp);
arch/powerpc/mm/book3s64/radix_pgtable.c
903
spin_lock(&init_mm.page_table_lock);
arch/powerpc/mm/book3s64/radix_pgtable.c
920
pte_clear(&init_mm, addr, (pte_t *)pgd);
arch/powerpc/mm/book3s64/radix_pgtable.c
929
spin_unlock(&init_mm.page_table_lock);
arch/powerpc/mm/book3s64/radix_tlb.c
688
mmgrab_lazy_tlb(&init_mm);
arch/powerpc/mm/book3s64/radix_tlb.c
689
current->active_mm = &init_mm;
arch/powerpc/mm/book3s64/radix_tlb.c
690
switch_mm_irqs_off(mm, &init_mm, current);
arch/powerpc/mm/kasan/8xx.c
32
__set_pte_at(&init_mm, k_cur, ptep + i, pte, 1);
arch/powerpc/mm/kasan/8xx.c
34
pmd_populate_kernel(&init_mm, pmd, ptep);
arch/powerpc/mm/kasan/8xx.c
73
__set_pte_at(&init_mm, k_cur, pte_offset_kernel(pmd, k_cur), pte, 0);
arch/powerpc/mm/kasan/book3s_32.c
54
__set_pte_at(&init_mm, k_cur, pte_offset_kernel(pmd, k_cur), pte, 0);
arch/powerpc/mm/kasan/init_32.c
190
pmd_populate_kernel(&init_mm, pmd, kasan_early_shadow_pte);
arch/powerpc/mm/kasan/init_32.c
28
__set_pte_at(&init_mm, va, ptep, pfn_pte(PHYS_PFN(pa), prot), 1);
arch/powerpc/mm/kasan/init_32.c
50
pmd_populate_kernel(&init_mm, pmd, new);
arch/powerpc/mm/kasan/init_32.c
77
__set_pte_at(&init_mm, k_cur, pte_offset_kernel(pmd, k_cur), pte, 0);
arch/powerpc/mm/kasan/init_32.c
95
__set_pte_at(&init_mm, k_cur, ptep, pte, 0);
arch/powerpc/mm/kasan/init_book3e_64.c
101
pud_populate(&init_mm, &kasan_early_shadow_pud[i],
arch/powerpc/mm/kasan/init_book3e_64.c
105
p4d_populate(&init_mm, p4d_offset(pgd++, addr), kasan_early_shadow_pud);
arch/powerpc/mm/kasan/init_book3e_64.c
121
__set_pte_at(&init_mm, (unsigned long)kasan_early_shadow_page,
arch/powerpc/mm/kasan/init_book3e_64.c
45
p4d_populate(&init_mm, p4dp, pudp);
arch/powerpc/mm/kasan/init_book3e_64.c
51
pud_populate(&init_mm, pudp, pmdp);
arch/powerpc/mm/kasan/init_book3e_64.c
57
pmd_populate_kernel(&init_mm, pmdp, ptep);
arch/powerpc/mm/kasan/init_book3e_64.c
61
__set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot), 0);
arch/powerpc/mm/kasan/init_book3e_64.c
93
__set_pte_at(&init_mm, (unsigned long)kasan_early_shadow_page,
arch/powerpc/mm/kasan/init_book3e_64.c
97
pmd_populate_kernel(&init_mm, &kasan_early_shadow_pmd[i],
arch/powerpc/mm/kasan/init_book3s_64.c
66
__set_pte_at(&init_mm, (unsigned long)kasan_early_shadow_page,
arch/powerpc/mm/kasan/init_book3s_64.c
70
pmd_populate_kernel(&init_mm, &kasan_early_shadow_pmd[i],
arch/powerpc/mm/kasan/init_book3s_64.c
74
pud_populate(&init_mm, &kasan_early_shadow_pud[i],
arch/powerpc/mm/kasan/init_book3s_64.c
83
__set_pte_at(&init_mm, (unsigned long)kasan_early_shadow_page,
arch/powerpc/mm/mmu_context.c
51
VM_WARN_ON_ONCE(next == &init_mm);
arch/powerpc/mm/nohash/8xx.c
68
pmd_populate_kernel(&init_mm, pmdp, ptep);
arch/powerpc/mm/nohash/8xx.c
71
pmd_populate_kernel(&init_mm, pmdp + 1, ptep);
arch/powerpc/mm/nohash/8xx.c
90
set_huge_pte_at(&init_mm, va, ptep,
arch/powerpc/mm/nohash/book3e_pgtable.c
100
p4d_populate(&init_mm, p4dp, pudp);
arch/powerpc/mm/nohash/book3e_pgtable.c
105
pud_populate(&init_mm, pudp, pmdp);
arch/powerpc/mm/nohash/book3e_pgtable.c
110
pmd_populate_kernel(&init_mm, pmdp, ptep);
arch/powerpc/mm/nohash/book3e_pgtable.c
114
set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot));
arch/powerpc/mm/nohash/book3e_pgtable.c
86
pudp = pud_alloc(&init_mm, p4dp, ea);
arch/powerpc/mm/nohash/book3e_pgtable.c
89
pmdp = pmd_alloc(&init_mm, pudp, ea);
arch/powerpc/mm/nohash/mmu_context.c
375
init_mm.context.active = NR_CPUS;
arch/powerpc/mm/pageattr.c
102
return apply_to_existing_page_range(&init_mm, start, size,
arch/powerpc/mm/pageattr.c
22
return pte_update(&init_mm, addr, ptep, old & ~new, new & ~old, 0);
arch/powerpc/mm/pageattr.c
56
pte_update(&init_mm, addr, ptep, _PAGE_PRESENT, 0, 0);
arch/powerpc/mm/pageattr.c
59
pte_update(&init_mm, addr, ptep, 0, _PAGE_PRESENT, 0);
arch/powerpc/mm/pgtable.c
244
pte_clear(&init_mm, va, ptep);
arch/powerpc/mm/pgtable.c
396
if (mm == &init_mm)
arch/powerpc/mm/pgtable_32.c
46
pmd_populate_kernel(&init_mm, pmdp, ptep);
arch/powerpc/mm/pgtable_32.c
62
pmd_populate_kernel(&init_mm, pmdp, ptep);
arch/powerpc/mm/pgtable_32.c
87
set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, prot));
arch/powerpc/mm/ptdump/ptdump.c
352
ptdump_walk_pgd(&st.ptdump, &init_mm, NULL);
arch/powerpc/mm/ptdump/ptdump.c
392
ptdump_walk_pgd(&st.ptdump, &init_mm, NULL);
arch/powerpc/xmon/xmon.c
3296
mm = &init_mm;
arch/powerpc/xmon/xmon.c
3301
mm = &init_mm;
arch/powerpc/xmon/xmon.c
3314
if (mm == &init_mm)
arch/riscv/include/asm/pgalloc.h
107
init_mm.pgd + USER_PTRS_PER_PGD,
arch/riscv/kernel/smpboot.c
221
struct mm_struct *mm = &init_mm;
arch/riscv/mm/fault.c
193
pgd_k = init_mm.pgd + index;
arch/riscv/mm/fault.c
38
mm = &init_mm;
arch/riscv/mm/fault.c
42
mm == &init_mm ? (u64)__pa_symbol(mm->pgd) : virt_to_phys(mm->pgd));
arch/riscv/mm/init.c
1486
p4d = p4d_alloc(&init_mm, pgd, addr);
arch/riscv/mm/init.c
1494
pud = pud_alloc(&init_mm, p4d, addr);
arch/riscv/mm/init.c
1502
pmd = pmd_alloc(&init_mm, pud, addr);
arch/riscv/mm/init.c
1666
pte_clear(&init_mm, addr, ptep);
arch/riscv/mm/init.c
402
pte_clear(&init_mm, addr, ptep);
arch/riscv/mm/pageattr.c
278
mmap_write_lock(&init_mm);
arch/riscv/mm/pageattr.c
330
mmap_write_unlock(&init_mm);
arch/riscv/mm/pageattr.c
341
mmap_write_unlock(&init_mm);
arch/riscv/mm/pageattr.c
429
apply_to_existing_page_range(&init_mm, start, size, debug_pagealloc_set_page, &enable);
arch/riscv/mm/ptdump.c
107
.mm = &init_mm,
arch/riscv/mm/ptdump.c
399
ptdump_walk_pgd(&st.ptdump, &init_mm, NULL);
arch/s390/boot/vmem.c
142
pgd_populate(&init_mm, pgd, kasan_early_shadow_p4d);
arch/s390/boot/vmem.c
153
p4d_populate(&init_mm, p4d, kasan_early_shadow_pud);
arch/s390/boot/vmem.c
164
pud_populate(&init_mm, pud, kasan_early_shadow_pmd);
arch/s390/boot/vmem.c
175
pmd_populate(&init_mm, pmd, kasan_early_shadow_pte);
arch/s390/boot/vmem.c
376
pmd_populate(&init_mm, pmd, pte);
arch/s390/boot/vmem.c
408
pud_populate(&init_mm, pud, pmd);
arch/s390/boot/vmem.c
432
p4d_populate(&init_mm, p4d, pud);
arch/s390/boot/vmem.c
451
pgd = pgd_offset(&init_mm, addr);
arch/s390/boot/vmem.c
458
pgd_populate(&init_mm, pgd, p4d);
arch/s390/boot/vmem.c
491
init_mm_pgd = init_mm.pgd;
arch/s390/boot/vmem.c
492
init_mm.pgd = (pgd_t *)swapper_pg_dir;
arch/s390/boot/vmem.c
554
init_mm.context.asce = get_lowcore()->kernel_asce.val;
arch/s390/boot/vmem.c
555
init_mm.pgd = init_mm_pgd;
arch/s390/include/asm/mmu_context.h
72
if (next == &init_mm)
arch/s390/include/asm/tlbflush.h
71
__tlb_flush_idte(init_mm.context.asce);
arch/s390/kernel/processor.c
111
mmgrab(&init_mm);
arch/s390/kernel/processor.c
112
current->active_mm = &init_mm;
arch/s390/kernel/processor.c
114
enter_lazy_tlb(&init_mm, current);
arch/s390/kernel/smp.c
247
cpumask_set_cpu(cpu, &init_mm.context.cpu_attach_mask);
arch/s390/kernel/smp.c
248
cpumask_set_cpu(cpu, mm_cpumask(&init_mm));
arch/s390/kernel/smp.c
926
cpumask_clear_cpu(cpu, mm_cpumask(&init_mm));
arch/s390/kernel/smp.c
927
cpumask_clear_cpu(cpu, &init_mm.context.cpu_attach_mask);
arch/s390/mm/dump_pagetables.c
211
ptdump_walk_pgd(&st.ptdump, &init_mm, NULL);
arch/s390/mm/dump_pagetables.c
251
ptdump_walk_pgd(&st.ptdump, &init_mm, NULL);
arch/s390/mm/init.c
158
cpumask_set_cpu(0, &init_mm.context.cpu_attach_mask);
arch/s390/mm/init.c
159
cpumask_set_cpu(0, mm_cpumask(&init_mm));
arch/s390/mm/pgalloc.c
123
if (mm == &init_mm)
arch/s390/mm/pgalloc.c
23
if (mm == &init_mm)
arch/s390/mm/vmem.c
184
pte_clear(&init_mm, addr, pte);
arch/s390/mm/vmem.c
283
pmd_populate(&init_mm, pmd, pte);
arch/s390/mm/vmem.c
354
pud_populate(&init_mm, pud, pmd);
arch/s390/mm/vmem.c
403
p4d_populate(&init_mm, p4d, pud);
arch/s390/mm/vmem.c
460
pgd_populate(&init_mm, pgd, p4d);
arch/s390/mm/vmem.c
591
pgd_populate(&init_mm, pgd, p4d);
arch/s390/mm/vmem.c
600
p4d_populate(&init_mm, p4d, pud);
arch/s390/mm/vmem.c
609
pud_populate(&init_mm, pud, pmd);
arch/s390/mm/vmem.c
620
pmd_populate(&init_mm, pmd, pte);
arch/s390/mm/vmem.c
661
pte_clear(&init_mm, addr, ptep);
arch/s390/mm/vmem.c
77
pte = (pte_t *)page_table_alloc(&init_mm);
arch/s390/mm/vmem.c
89
page_table_free(&init_mm, table);
arch/sh/kernel/smp.c
178
struct mm_struct *mm = &init_mm;
arch/sh/kernel/smp.c
62
init_new_context(current, &init_mm);
arch/sh/mm/fault.c
129
pgd_k = init_mm.pgd + index;
arch/sh/mm/init.c
141
pud_populate(&init_mm, pud, pmd);
arch/sh/mm/init.c
154
pmd_populate_kernel(&init_mm, pmd, pte);
arch/sh/mm/kmap.c
60
pte_clear(&init_mm, vaddr, kmap_coherent_pte - idx);
arch/sh/mm/tlbflush_32.c
92
asid = cpu_asid(cpu, &init_mm);
arch/sparc/include/asm/mmu_context_64.h
86
if (unlikely(mm == &init_mm))
arch/sparc/include/asm/pgtable_64.h
892
extern struct mm_struct init_mm;
arch/sparc/include/asm/pgtable_64.h
909
if (likely(mm != &init_mm) && pte_accessible(mm, orig))
arch/sparc/kernel/leon_smp.c
95
mmgrab(&init_mm);
arch/sparc/kernel/leon_smp.c
96
current->active_mm = &init_mm;
arch/sparc/kernel/smp_64.c
130
mmgrab(&init_mm);
arch/sparc/kernel/smp_64.c
131
current->active_mm = &init_mm;
arch/sparc/kernel/sun4d_smp.c
97
mmgrab(&init_mm);
arch/sparc/kernel/sun4d_smp.c
98
current->active_mm = &init_mm;
arch/sparc/kernel/sun4m_smp.c
63
mmgrab(&init_mm);
arch/sparc/kernel/sun4m_smp.c
64
current->active_mm = &init_mm;
arch/sparc/kernel/traps_32.c
389
mmgrab(&init_mm);
arch/sparc/kernel/traps_32.c
390
current->active_mm = &init_mm;
arch/sparc/kernel/traps_64.c
2923
mmgrab(&init_mm);
arch/sparc/kernel/traps_64.c
2924
current->active_mm = &init_mm;
arch/sparc/mm/fault_32.c
279
pgd_k = init_mm.pgd + offset;
arch/sparc/mm/init_64.c
1787
pgd_populate(&init_mm, pgd, new);
arch/sparc/mm/init_64.c
1799
p4d_populate(&init_mm, p4d, new);
arch/sparc/mm/init_64.c
1815
pud_populate(&init_mm, pud, new);
arch/sparc/mm/init_64.c
1831
pmd_populate_kernel(&init_mm, pmd, new);
arch/sparc/mm/init_64.c
2396
init_mm.pgd += ((shift) / (sizeof(pgd_t)));
arch/sparc/mm/init_64.c
795
if (unlikely(!mm || mm == &init_mm))
arch/sparc/mm/srmmu.c
291
init_mm.pgd = srmmu_swapper_pg_dir;
arch/sparc/mm/srmmu.c
906
init_mm.context = (unsigned long) NO_CONTEXT;
arch/sparc/mm/tlb.c
182
if (mm == &init_mm)
arch/um/include/asm/tlbflush.h
53
um_tlb_mark_sync(&init_mm, start, end);
arch/um/include/asm/tlbflush.h
56
um_tlb_sync(&init_mm);
arch/um/kernel/smp.c
140
struct mm_struct *mm = &init_mm;
arch/um/kernel/tlb.c
174
if (mm == &init_mm) {
arch/um/kernel/trap.c
325
err = um_tlb_sync(&init_mm);
arch/x86/hyperv/hv_crash.c
235
p4d_populate(&init_mm, p4d, (pud_t *)hv_crash_ptpgs[1]);
arch/x86/hyperv/hv_crash.c
483
tramp->kernel_cr3 = __sme_pa(init_mm.pgd);
arch/x86/include/asm/pgtable_32.h
43
pte_clear(&init_mm, (vaddr), (ptep)); \
arch/x86/kernel/cpu/common.c
2493
mmgrab(&init_mm);
arch/x86/kernel/cpu/common.c
2494
cur->active_mm = &init_mm;
arch/x86/kernel/cpu/common.c
2497
enter_lazy_tlb(&init_mm, cur);
arch/x86/kernel/cpu/common.c
2505
load_mm_ldt(&init_mm);
arch/x86/kernel/espfix_64.c
115
p4d = p4d_alloc(&init_mm, pgd, ESPFIX_BASE_ADDR);
arch/x86/kernel/espfix_64.c
116
p4d_populate(&init_mm, p4d, espfix_pud_page);
arch/x86/kernel/espfix_64.c
169
paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT);
arch/x86/kernel/espfix_64.c
181
paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT);
arch/x86/kernel/tboot.c
103
MMAP_LOCK_INITIALIZER(init_mm)
arch/x86/kernel/tboot.c
104
.page_table_lock = __SPIN_LOCK_UNLOCKED(init_mm.page_table_lock),
arch/x86/kernel/tboot.c
105
.mmlist = LIST_HEAD_INIT(init_mm.mmlist),
arch/x86/mm/debug_pagetables.c
10
ptdump_walk_pgd_level_debugfs(m, &init_mm, false);
arch/x86/mm/dump_pagetables.c
501
ptdump_walk_pgd_level_core(NULL, &init_mm, pgd, true, false);
arch/x86/mm/dump_pagetables.c
510
return ptdump_walk_pgd_level_core(NULL, &init_mm, INIT_PGD, true, false);
arch/x86/mm/fault.c
1002
pgd = init_mm.pgd + pgd_index(address);
arch/x86/mm/fault.c
184
pgd_k = init_mm.pgd + index;
arch/x86/mm/init.c
1014
.loaded_mm = &init_mm,
arch/x86/mm/init_64.c
1053
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1055
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1071
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1073
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1089
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1091
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1123
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1124
pte_clear(&init_mm, addr, pte);
arch/x86/mm/init_64.c
1125
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1159
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1161
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1166
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1168
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1201
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1203
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
1329
p4d = p4d_alloc(&init_mm, pgd, addr);
arch/x86/mm/init_64.c
1347
pud = pud_alloc(&init_mm, p4d, addr);
arch/x86/mm/init_64.c
1384
if (get_gate_vma(&init_mm))
arch/x86/mm/init_64.c
271
pgd_populate(&init_mm, pgd, p4d);
arch/x86/mm/init_64.c
283
p4d_populate(&init_mm, p4d, pud);
arch/x86/mm/init_64.c
295
pud_populate(&init_mm, pud, pmd);
arch/x86/mm/init_64.c
307
pmd_populate_kernel(&init_mm, pmd, pte);
arch/x86/mm/init_64.c
549
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
554
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
580
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
584
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
592
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
593
pmd_populate_kernel_init(&init_mm, pmd, pte, init);
arch/x86/mm/init_64.c
594
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
666
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
670
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
679
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
680
pud_populate_init(&init_mm, pud, pmd, init);
arch/x86/mm/init_64.c
681
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
732
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
733
p4d_populate_init(&init_mm, p4d, pud, init);
arch/x86/mm/init_64.c
734
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
773
spin_lock(&init_mm.page_table_lock);
arch/x86/mm/init_64.c
775
pgd_populate_init(&init_mm, pgd, p4d, init);
arch/x86/mm/init_64.c
777
p4d_populate_init(&init_mm, p4d_offset(pgd, vaddr),
arch/x86/mm/init_64.c
780
spin_unlock(&init_mm.page_table_lock);
arch/x86/mm/ioremap.c
862
pmd_populate_kernel(&init_mm, pmd, bm_pte);
arch/x86/mm/ioremap.c
905
pte_clear(&init_mm, addr, pte);
arch/x86/mm/kasan_init_64.c
111
p4d_populate(&init_mm, p4d, p);
arch/x86/mm/kasan_init_64.c
131
pgd_populate(&init_mm, pgd, p);
arch/x86/mm/kasan_init_64.c
258
p4d_populate(&init_mm, p4d, p);
arch/x86/mm/kasan_init_64.c
276
pgd_populate(&init_mm, pgd, p);
arch/x86/mm/kasan_init_64.c
55
pmd_populate_kernel(&init_mm, pmd, p);
arch/x86/mm/kasan_init_64.c
68
set_pte_at(&init_mm, addr, pte, entry);
arch/x86/mm/kasan_init_64.c
91
pud_populate(&init_mm, pud, p);
arch/x86/mm/kmmio.c
153
pte_clear(&init_mm, 0, pte);
arch/x86/mm/pat/set_memory.c
1142
paravirt_alloc_pte(&init_mm, page_to_pfn(base));
arch/x86/mm/pat/set_memory.c
2378
cpa.pgd = init_mm.pgd;
arch/x86/mm/pgtable.c
182
if (mm == &init_mm)
arch/x86/mm/pgtable.c
760
pmd_free(&init_mm, pmd);
arch/x86/mm/pgtable_32.c
57
set_pte_at(&init_mm, vaddr, pte, pteval);
arch/x86/mm/pgtable_32.c
59
pte_clear(&init_mm, vaddr, pte);
arch/x86/mm/tlb.c
1083
u64 tlb_gen = atomic64_read(&init_mm.context.tlb_gen);
arch/x86/mm/tlb.c
1158
if (unlikely(loaded_mm == &init_mm))
arch/x86/mm/tlb.c
1184
switch_mm_irqs_off(NULL, &init_mm, NULL);
arch/x86/mm/tlb.c
597
if (loaded_mm == &init_mm)
arch/x86/mm/tlb.c
603
switch_mm(NULL, &init_mm, NULL);
arch/x86/mm/tlb.c
858
WARN_ON_ONCE(prev != &init_mm && !is_notrack_mm(prev) &&
arch/x86/mm/tlb.c
935
if (next != &init_mm && !cpumask_test_cpu(cpu, mm_cpumask(next)))
arch/x86/mm/tlb.c
989
if (this_cpu_read(cpu_tlbstate.loaded_mm) == &init_mm)
arch/x86/platform/efi/efi_32.c
78
ptdump_walk_pgd_level(NULL, &init_mm);
arch/x86/platform/efi/efi_64.c
81
p4d = p4d_alloc(&init_mm, pgd, EFI_VA_END);
arch/x86/platform/efi/efi_64.c
85
pud = pud_alloc(&init_mm, p4d, EFI_VA_END);
arch/x86/xen/grant-table.c
110
if (apply_to_page_range(&init_mm, (unsigned long)area->area->addr,
arch/x86/xen/grant-table.c
45
set_pte_at(&init_mm, addr, gnttab_shared_vm_area.ptes[i],
arch/x86/xen/grant-table.c
67
set_pte_at(&init_mm, addr, gnttab_status_vm_area.ptes[i],
arch/x86/xen/grant-table.c
89
set_pte_at(&init_mm, addr, ptes[i], __pte(0));
arch/x86/xen/mmu_pv.c
833
spin_lock(&init_mm.page_table_lock);
arch/x86/xen/mmu_pv.c
838
__xen_pgd_pin(&init_mm, (pgd_t *)page_address(page));
arch/x86/xen/mmu_pv.c
844
spin_unlock(&init_mm.page_table_lock);
arch/x86/xen/mmu_pv.c
865
xen_pgd_walk(&init_mm, xen_mark_pinned, FIXADDR_TOP);
arch/x86/xen/mmu_pv.c
944
spin_lock(&init_mm.page_table_lock);
arch/x86/xen/mmu_pv.c
950
__xen_pgd_unpin(&init_mm, (pgd_t *)page_address(page));
arch/x86/xen/mmu_pv.c
956
spin_unlock(&init_mm.page_table_lock);
arch/x86/xen/p2m.c
336
paravirt_alloc_pte(&init_mm, __pa(p2m_missing_pte) >> PAGE_SHIFT);
arch/x86/xen/p2m.c
338
paravirt_alloc_pte(&init_mm, __pa(p2m_identity_pte) >> PAGE_SHIFT);
arch/x86/xen/p2m.c
483
paravirt_alloc_pte(&init_mm, __pa(pte_newpg[i]) >> PAGE_SHIFT);
arch/xtensa/kernel/smp.c
122
struct mm_struct *mm = &init_mm;
arch/xtensa/mm/fault.c
45
pgd_k = init_mm.pgd + index;
drivers/xen/xenbus/xenbus_client.c
774
if (apply_to_page_range(&init_mm, (unsigned long)area->addr,
include/asm-generic/pgalloc.h
141
if (mm == &init_mm)
include/asm-generic/pgalloc.h
151
if (mm == &init_mm)
include/asm-generic/pgalloc.h
178
if (mm == &init_mm)
include/asm-generic/pgalloc.h
187
if (mm == &init_mm)
include/asm-generic/pgalloc.h
235
if (mm == &init_mm)
include/asm-generic/pgalloc.h
244
if (mm == &init_mm)
include/asm-generic/pgalloc.h
282
if (mm == &init_mm)
include/asm-generic/pgalloc.h
291
if (mm == &init_mm)
include/linux/mm.h
3543
if (mm != &init_mm && !ptlock_init(ptdesc))
include/linux/mm.h
3634
if (mm != &init_mm && !pmd_ptlock_init(ptdesc))
include/linux/mm_types.h
1415
extern struct mm_struct init_mm;
include/linux/pgalloc.h
17
pgd_populate(&init_mm, pgd, p4d); \
include/linux/pgalloc.h
24
p4d_populate(&init_mm, p4d, pud); \
include/linux/pgtable.h
156
#define pgd_offset_k(address) pgd_offset(&init_mm, (address))
init/init_task.c
115
.active_mm = &init_mm,
kernel/bpf/arena.c
137
set_pte_at(&init_mm, addr, pte, mk_pte(page, PAGE_KERNEL));
kernel/bpf/arena.c
161
pte_clear(&init_mm, addr, pte);
kernel/bpf/arena.c
172
return apply_to_page_range(&init_mm, bpf_arena_get_kern_vm_start(arena),
kernel/bpf/arena.c
288
apply_to_existing_page_range(&init_mm, bpf_arena_get_kern_vm_start(arena),
kernel/bpf/arena.c
399
ret = apply_to_page_range(&init_mm, kaddr, PAGE_SIZE, apply_range_set_cb, &data);
kernel/bpf/arena.c
612
ret = apply_to_page_range(&init_mm,
kernel/bpf/arena.c
701
apply_to_existing_page_range(&init_mm, kaddr, page_cnt << PAGE_SHIFT,
kernel/bpf/arena.c
816
apply_to_existing_page_range(&init_mm, kaddr, page_cnt << PAGE_SHIFT,
kernel/cpu.c
917
WARN_ON(mm != &init_mm);
kernel/events/core.c
8489
mm = &init_mm;
kernel/fork.c
665
current->active_mm = &init_mm;
kernel/fork.c
666
switch_mm(mm, &init_mm, current);
kernel/fork.c
720
BUG_ON(mm == &init_mm);
kernel/sched/core.c
8071
if (mm != &init_mm) {
kernel/sched/core.c
8072
mmgrab_lazy_tlb(&init_mm);
kernel/sched/core.c
8074
current->active_mm = &init_mm;
kernel/sched/core.c
8075
switch_mm_irqs_off(mm, &init_mm, current);
kernel/sched/core.c
8764
mmgrab_lazy_tlb(&init_mm);
kernel/sched/core.c
8765
enter_lazy_tlb(&init_mm, current);
mm/highmem.c
230
pte_clear(&init_mm, PKMAP_ADDR(i), &pkmap_page_table[i]);
mm/highmem.c
291
set_pte_at(&init_mm, vaddr,
mm/highmem.c
567
arch_kmap_local_set_pte(&init_mm, vaddr, kmap_pte, pteval);
mm/highmem.c
627
pte_clear(&init_mm, addr, kmap_pte);
mm/highmem.c
678
pte_clear(&init_mm, addr, kmap_pte);
mm/highmem.c
707
set_pte_at(&init_mm, addr, kmap_pte, pteval);
mm/hugetlb_vmemmap.c
109
spin_lock(&init_mm.page_table_lock);
mm/hugetlb_vmemmap.c
132
spin_unlock(&init_mm.page_table_lock);
mm/hugetlb_vmemmap.c
169
mmap_read_lock(&init_mm);
mm/hugetlb_vmemmap.c
172
mmap_read_unlock(&init_mm);
mm/hugetlb_vmemmap.c
234
set_pte_at(&init_mm, addr, pte, entry);
mm/hugetlb_vmemmap.c
276
set_pte_at(&init_mm, addr, pte, mk_pte(page, pgprot));
mm/hugetlb_vmemmap.c
60
pgtable = pte_alloc_one_kernel(&init_mm);
mm/hugetlb_vmemmap.c
64
pmd_populate_kernel(&init_mm, &__pmd, pgtable);
mm/hugetlb_vmemmap.c
72
set_pte_at(&init_mm, addr, pte, entry);
mm/hugetlb_vmemmap.c
75
spin_lock(&init_mm.page_table_lock);
mm/hugetlb_vmemmap.c
87
pmd_populate_kernel(&init_mm, pmd, pgtable);
mm/hugetlb_vmemmap.c
91
pte_free_kernel(&init_mm, pgtable);
mm/hugetlb_vmemmap.c
93
spin_unlock(&init_mm.page_table_lock);
mm/init-mm.c
32
struct mm_struct init_mm = {
mm/init-mm.c
33
.mm_mt = MTREE_INIT_EXT(mm_mt, MM_MT_FLAGS, init_mm.mmap_lock),
mm/init-mm.c
37
.write_protect_seq = SEQCNT_ZERO(init_mm.write_protect_seq),
mm/init-mm.c
38
MMAP_LOCK_INITIALIZER(init_mm)
mm/init-mm.c
39
.page_table_lock = __SPIN_LOCK_UNLOCKED(init_mm.page_table_lock),
mm/init-mm.c
40
.arg_lock = __SPIN_LOCK_UNLOCKED(init_mm.arg_lock),
mm/init-mm.c
41
.mmlist = LIST_HEAD_INIT(init_mm.mmlist),
mm/init-mm.c
43
.vma_writer_wait = __RCUWAIT_INITIALIZER(init_mm.vma_writer_wait),
mm/init-mm.c
44
.mm_lock_seq = SEQCNT_ZERO(init_mm.mm_lock_seq),
mm/init-mm.c
48
.mm_cid.lock = __RAW_SPIN_LOCK_UNLOCKED(init_mm.mm_cid.lock),
mm/init-mm.c
51
INIT_MM_CONTEXT(init_mm)
mm/init-mm.c
57
init_mm.start_code = (unsigned long)start_code;
mm/init-mm.c
58
init_mm.end_code = (unsigned long)end_code;
mm/init-mm.c
59
init_mm.end_data = (unsigned long)end_data;
mm/init-mm.c
60
init_mm.brk = (unsigned long)brk;
mm/kasan/init.c
103
set_pte_at(&init_mm, addr, pte, zero_pte);
mm/kasan/init.c
119
pmd_populate_kernel(&init_mm, pmd,
mm/kasan/init.c
128
p = pte_alloc_one_kernel(&init_mm);
mm/kasan/init.c
136
pmd_populate_kernel(&init_mm, pmd, p);
mm/kasan/init.c
155
pud_populate(&init_mm, pud,
mm/kasan/init.c
158
pmd_populate_kernel(&init_mm, pmd,
mm/kasan/init.c
167
p = pmd_alloc(&init_mm, pud, addr);
mm/kasan/init.c
173
pud_populate(&init_mm, pud, p);
mm/kasan/init.c
197
pud_populate(&init_mm, pud,
mm/kasan/init.c
200
pmd_populate_kernel(&init_mm, pmd,
mm/kasan/init.c
209
p = pud_alloc(&init_mm, p4d, addr);
mm/kasan/init.c
260
pud_populate(&init_mm, pud,
mm/kasan/init.c
263
pmd_populate_kernel(&init_mm, pmd,
mm/kasan/init.c
271
if (!p4d_alloc(&init_mm, pgd, addr))
mm/kasan/init.c
295
pte_free_kernel(&init_mm, (pte_t *)page_to_virt(pmd_page(*pmd)));
mm/kasan/init.c
310
pmd_free(&init_mm, (pmd_t *)page_to_virt(pud_page(*pud)));
mm/kasan/init.c
325
pud_free(&init_mm, (pud_t *)page_to_virt(p4d_page(*p4d)));
mm/kasan/init.c
340
p4d_free(&init_mm, (p4d_t *)page_to_virt(pgd_page(*pgd)));
mm/kasan/init.c
362
pte_clear(&init_mm, addr, pte);
mm/kasan/shadow.c
315
spin_lock(&init_mm.page_table_lock);
mm/kasan/shadow.c
317
set_pte_at(&init_mm, addr, ptep, pte);
mm/kasan/shadow.c
320
spin_unlock(&init_mm.page_table_lock);
mm/kasan/shadow.c
381
ret = apply_to_page_range(&init_mm, start, nr_pages * PAGE_SIZE,
mm/kasan/shadow.c
476
spin_lock(&init_mm.page_table_lock);
mm/kasan/shadow.c
480
pte_clear(&init_mm, addr, ptep);
mm/kasan/shadow.c
481
spin_unlock(&init_mm.page_table_lock);
mm/kasan/shadow.c
602
apply_to_existing_page_range(&init_mm,
mm/memory.c
3225
mapped_pte = pte = (mm == &init_mm) ?
mm/memory.c
3231
mapped_pte = pte = (mm == &init_mm) ?
mm/memory.c
3253
if (mm != &init_mm)
mm/memory.c
478
pte_t *new = pte_alloc_one_kernel(&init_mm);
mm/memory.c
482
spin_lock(&init_mm.page_table_lock);
mm/memory.c
485
pmd_populate_kernel(&init_mm, pmd, new);
mm/memory.c
488
spin_unlock(&init_mm.page_table_lock);
mm/memory.c
490
pte_free_kernel(&init_mm, new);
mm/page_table_check.c
151
if (&init_mm == mm)
mm/page_table_check.c
163
if (&init_mm == mm)
mm/page_table_check.c
175
if (&init_mm == mm)
mm/page_table_check.c
207
if (&init_mm == mm)
mm/page_table_check.c
237
if (&init_mm == mm)
mm/page_table_check.c
255
if (&init_mm == mm)
mm/page_table_check.c
269
if (&init_mm == mm)
mm/pagewalk.c
637
mmap_assert_locked(&init_mm);
mm/pagewalk.c
654
.mm = &init_mm,
mm/pagewalk.c
696
if (mm == &init_mm)
mm/pagewalk.c
76
if (walk->mm == &init_mm || addr >= TASK_SIZE)
mm/pagewalk.c
82
if (walk->mm != &init_mm && addr < TASK_SIZE)
mm/percpu.c
3166
pud_populate(&init_mm, pud, pmd);
mm/percpu.c
3174
pmd_populate_kernel(&init_mm, pmd, new);
mm/rmap.c
2302
list_add(&mm->mmlist, &init_mm.mmlist);
mm/sparse-vmemmap.c
182
set_pte_at(&init_mm, addr, pte, entry);
mm/sparse-vmemmap.c
206
pmd_populate_kernel(&init_mm, pmd, p);
mm/sparse-vmemmap.c
219
pud_populate(&init_mm, pud, p);
mm/sparse-vmemmap.c
339
pte_clear(&init_mm, maddr, pte);
mm/sparse-vmemmap.c
349
pte_clear(&init_mm, maddr, pte);
mm/sparse-vmemmap.c
377
ptep_set_wrprotect(&init_mm, maddr, pte);
mm/swapfile.c
2418
prev_mm = &init_mm;
mm/swapfile.c
2422
p = &init_mm.mmlist;
mm/swapfile.c
2425
(p = p->next) != &init_mm.mmlist) {
mm/swapfile.c
2516
list_for_each_safe(p, next, &init_mm.mmlist)
mm/vmalloc.c
128
set_huge_pte_at(&init_mm, addr, pte, entry, size);
mm/vmalloc.c
133
set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot));
mm/vmalloc.c
175
pmd = pmd_alloc_track(&init_mm, pud, addr, mask);
mm/vmalloc.c
227
pud = pud_alloc_track(&init_mm, p4d, addr, mask);
mm/vmalloc.c
279
p4d = p4d_alloc_track(&init_mm, pgd, addr, mask);
mm/vmalloc.c
3591
set_pte_at(&init_mm, addr, pte, ptent);
mm/vmalloc.c
3615
if (apply_to_page_range(&init_mm, (unsigned long)area->addr,
mm/vmalloc.c
384
ptent = huge_ptep_get_and_clear(&init_mm, addr, pte, size);
mm/vmalloc.c
389
ptent = ptep_get_and_clear(&init_mm, addr, pte);
mm/vmalloc.c
559
set_pte_at(&init_mm, addr, pte, mk_pte(page, prot));
mm/vmalloc.c
576
pmd = pmd_alloc_track(&init_mm, pud, addr, mask);
mm/vmalloc.c
594
pud = pud_alloc_track(&init_mm, p4d, addr, mask);
mm/vmalloc.c
612
p4d = p4d_alloc_track(&init_mm, pgd, addr, mask);