Symbol: set_pmd
arch/arc/include/asm/pgalloc.h
46
set_pmd(pmd, __pmd((unsigned long)pte));
arch/arc/include/asm/pgalloc.h
51
set_pmd(pmd, __pmd((unsigned long)page_address(pte_page)));
arch/arc/mm/fault.c
64
set_pmd(pmd, *pmd_k);
arch/arm64/include/asm/pgalloc.h
101
set_pmd(pmdp, __pmd(__phys_to_pmd_val(ptep) | prot));
arch/arm64/include/asm/pgtable.h
841
set_pmd(pmdp, __pmd(0));
arch/arm64/mm/mmu.c
1818
set_pmd(pmdp, new_pmd);
arch/arm64/mm/mmu.c
628
set_pmd(pmdp, pmd_mknoncont(pmdp_get(pmdp)));
arch/arm64/mm/mmu.c
655
set_pmd(pmdp, pfn_pmd(pfn, prot));
arch/arm64/mm/pageattr.c
59
set_pmd(pmd, val);
arch/arm64/mm/trans_pgd.c
112
set_pmd(dst_pmdp,
arch/csky/include/asm/pgalloc.h
16
set_pmd(pmd, __pmd(__pa(pte)));
arch/csky/include/asm/pgalloc.h
22
set_pmd(pmd, __pmd(__pa(page_address(pte))));
arch/csky/mm/fault.c
160
set_pmd(pmd, *pmd_k);
arch/csky/mm/init.c
115
set_pmd(pmd, __pmd(__pa(pte)));
arch/hexagon/include/asm/pgalloc.h
51
set_pmd(pmd, __pmd(((unsigned long)page_to_pfn(pte) << PAGE_SHIFT) |
arch/hexagon/include/asm/pgalloc.h
77
set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
arch/hexagon/include/asm/pgalloc.h
85
set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
arch/loongarch/include/asm/pgalloc.h
19
set_pmd(pmd, __pmd((unsigned long)pte));
arch/loongarch/include/asm/pgalloc.h
24
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
arch/loongarch/include/asm/pgtable.h
262
set_pmd(pmdp, __pmd((unsigned long)invalid_pte_table));
arch/loongarch/mm/pageattr.c
74
set_pmd(pmd, val);
arch/mips/include/asm/pgalloc.h
23
set_pmd(pmd, __pmd((unsigned long)pte));
arch/mips/include/asm/pgalloc.h
29
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
arch/mips/mm/fault.c
313
set_pmd(pmd, *pmd_k);
arch/mips/mm/init.c
257
set_pmd(pmd, __pmd((unsigned long)pte));
arch/nios2/include/asm/pgalloc.h
20
set_pmd(pmd, __pmd((unsigned long)pte));
arch/nios2/include/asm/pgalloc.h
26
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
arch/nios2/mm/fault.c
255
set_pmd(pmd, *pmd_k);
arch/openrisc/include/asm/pgalloc.h
29
set_pmd(pmd, __pmd(_KERNPG_TABLE + __pa(pte)))
arch/openrisc/include/asm/pgalloc.h
34
set_pmd(pmd, __pmd(_KERNPG_TABLE +
arch/openrisc/mm/fault.c
339
set_pmd(pmd, *pmd_k);
arch/openrisc/mm/init.c
99
set_pmd(pme, __pmd(_KERNPG_TABLE + __pa(pte)));
arch/parisc/include/asm/pgalloc.h
53
set_pmd(pmd, __pmd((PxD_FLAG_PRESENT | PxD_FLAG_VALID)
arch/parisc/include/asm/pgtable.h
292
set_pmd(pmd, __pmd(0));
arch/riscv/include/asm/pgalloc.h
23
set_pmd(pmd, __pmd((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE));
arch/riscv/include/asm/pgalloc.h
31
set_pmd(pmd, __pmd((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE));
arch/riscv/include/asm/pgtable.h
252
set_pmd(pmdp, __pmd(0));
arch/riscv/kernel/hibernate.c
223
set_pmd(dst_pmdp, __pmd(pmd_val(pmd) | pgprot_val(prot)));
arch/riscv/mm/kasan_init.c
334
set_pmd(kasan_early_shadow_pmd + i,
arch/riscv/mm/kasan_init.c
36
set_pmd(pmd, pfn_pmd(PFN_DOWN(__pa(p)), PAGE_TABLE));
arch/riscv/mm/kasan_init.c
70
set_pmd(pmdp, pfn_pmd(PFN_DOWN(phys_addr), PAGE_KERNEL));
arch/riscv/mm/pageattr.c
128
set_pmd(pmdp, pfn_pmd(page_to_pfn(pte_page), PAGE_TABLE));
arch/riscv/mm/pageattr.c
164
set_pmd(pmdp_new,
arch/riscv/mm/pageattr.c
62
set_pmd(pmd, val);
arch/riscv/mm/pgtable.c
114
set_pmd(pmd, new_pmd);
arch/s390/boot/vmem.c
371
set_pmd(pmd, entry);
arch/s390/include/asm/pgalloc.h
154
set_pmd(pmd, __pmd(_SEGMENT_ENTRY | __pa(pte)));
arch/s390/include/asm/pgtable.h
1004
set_pmd(pmdp, __pmd(_SEGMENT_ENTRY_EMPTY));
arch/s390/include/asm/pgtable.h
1706
set_pmd(pmdp, entry);
arch/s390/include/asm/pgtable.h
1730
set_pmd(pmdp, __pmd(_SEGMENT_ENTRY_EMPTY));
arch/s390/mm/pageattr.c
218
set_pmd(pmdp, __pmd(pmd_addr | prot));
arch/s390/mm/pgtable.c
230
set_pmd(pmdp, set_pmd_bit(*pmdp, __pgprot(_SEGMENT_ENTRY_INVALID)));
arch/s390/mm/pgtable.c
246
set_pmd(pmdp, new);
arch/s390/mm/pgtable.c
259
set_pmd(pmdp, new);
arch/s390/mm/vmem.c
257
set_pmd(pmd, __pmd(__pa(addr) | prot));
arch/s390/mm/vmem.c
272
set_pmd(pmd, __pmd(__pa(new_page) | prot));
arch/sh/include/asm/pgalloc.h
26
set_pmd(pmd, __pmd((unsigned long)pte));
arch/sh/include/asm/pgalloc.h
32
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
arch/sh/include/asm/pgtable_32.h
329
#define pmd_clear(xp) do { set_pmd(xp, __pmd(0)); } while (0)
arch/sh/mm/fault.c
153
set_pmd(pmd, *pmd_k);
arch/um/include/asm/pgalloc.h
16
set_pmd(pmd, __pmd(_PAGE_TABLE + (unsigned long) __pa(pte)))
arch/um/include/asm/pgalloc.h
19
set_pmd(pmd, __pmd(_PAGE_TABLE + \
arch/x86/boot/compressed/ident_map_64.c
225
set_pmd(pmdp, pmd);
arch/x86/boot/startup/sme.c
160
set_pmd(pmd, __pmd(ppd->paddr | ppd->pmd_flags));
arch/x86/boot/startup/sme.c
178
set_pmd(pmd, __pmd(PMD_FLAGS | __pa(pte)));
arch/x86/entry/vsyscall/vsyscall_64.c
358
set_pmd(pmd, __pmd(pmd_val(*pmd) | _PAGE_USER));
arch/x86/hyperv/hv_crash.c
526
set_pmd(pmd, __pmd(_PAGE_TABLE | pa));
arch/x86/include/asm/paravirt.h
384
PVOP_VCALL2(pv_ops, mmu.set_pmd, pmdp, native_pmd_val(pmd));
arch/x86/include/asm/paravirt.h
481
set_pmd(pmdp, native_make_pmd(0));
arch/x86/include/asm/paravirt_types.h
148
void (*set_pmd)(pmd_t *pmdp, pmd_t pmdval);
arch/x86/include/asm/pgalloc.h
66
set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
arch/x86/include/asm/pgalloc.h
82
set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE));
arch/x86/include/asm/pgtable.h
1218
set_pmd(pmdp, pmd);
arch/x86/include/asm/pgtable.h
1724
set_pmd(pmdp, pmd); \
arch/x86/kernel/espfix_64.c
183
set_pmd(&pmd_p[n], pmd);
arch/x86/kernel/ldt.c
244
set_pmd(u_pmd, *k_pmd);
arch/x86/kernel/machine_kexec_32.c
95
set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
arch/x86/kernel/machine_kexec_64.c
211
set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
arch/x86/kernel/paravirt.c
180
.mmu.set_pmd = native_set_pmd,
arch/x86/mm/fault.c
208
set_pmd(pmd, *pmd_k);
arch/x86/mm/ident_map.c
90
set_pmd(pmd, __pmd((addr - info->offset) | info->page_flag));
arch/x86/mm/init_32.c
100
set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE));
arch/x86/mm/init_32.c
181
set_pmd(pmd, __pmd(__pa(newpte)|_PAGE_TABLE));
arch/x86/mm/init_32.c
333
set_pmd(pmd, pfn_pmd(pfn, init_prot));
arch/x86/mm/init_32.c
335
set_pmd(pmd, pfn_pmd(pfn, prot));
arch/x86/mm/init_64.c
1525
set_pmd(pmd, __pmd(pte_val(entry)));
arch/x86/mm/init_64.c
417
set_pmd(pmd, __pmd(phys | pgprot_val(prot)));
arch/x86/mm/init_64.c
463
set_pmd(pmd, __pmd(0));
arch/x86/mm/kmmio.c
144
set_pmd(pmd, new_pmd);
arch/x86/mm/pat/set_memory.c
1288
set_pmd(pmd, _pmd);
arch/x86/mm/pat/set_memory.c
1305
set_pmd(pmd, _pmd);
arch/x86/mm/pat/set_memory.c
1546
set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
arch/x86/mm/pat/set_memory.c
1630
set_pmd(pmd, pmd_mkhuge(pfn_pmd(cpa->pfn,
arch/x86/mm/pgtable.c
413
set_pmd(pmdp, entry);
arch/x86/mm/pti.c
265
set_pmd(pmd, __pmd(0));
arch/x86/mm/pti.c
277
set_pmd(pmd, __pmd(_KERNPG_TABLE | __pa(new_pte_page)));
arch/x86/power/hibernate.c
183
set_pmd(pmd, __pmd(pmd_val(*pmd) & ~_PAGE_NX));
arch/x86/power/hibernate_32.c
110
set_pmd(pmd, pfn_pmd(pfn, PAGE_KERNEL_LARGE_EXEC));
arch/x86/power/hibernate_32.c
158
set_pmd(pmd + pmd_index(restore_jump_address),
arch/x86/power/hibernate_32.c
66
set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE));
arch/x86/power/hibernate_64.c
68
set_pmd(pmd + pmd_index(restore_jump_address),
arch/x86/xen/mmu_pv.c
1080
set_pmd(pmd, __pmd(0));
arch/x86/xen/mmu_pv.c
1131
set_pmd(pmd, __pmd(0));
arch/x86/xen/mmu_pv.c
2164
pv_ops.mmu.set_pmd = xen_set_pmd;
arch/x86/xen/mmu_pv.c
2209
pv_ops.mmu.set_pmd = xen_set_pmd_hyper;
arch/x86/xen/p2m.c
396
set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE));
arch/x86/xen/p2m.c
494
set_pmd(pmdp,
arch/xtensa/include/asm/pgtable.h
248
#define pmd_clear(pmdp) do { set_pmd(pmdp, __pmd(0)); } while (0)
arch/xtensa/mm/kasan_init.c
31
set_pmd(pmd, __pmd((unsigned long)kasan_early_shadow_pte));
arch/xtensa/mm/kasan_init.c
63
set_pmd(pmd + i, __pmd((unsigned long)pte));
arch/xtensa/mm/mmu.c
47
set_pmd(pmd, __pmd(((unsigned long)cur_pte) & PAGE_MASK));
include/asm-generic/pgtable-nopmd.h
44
#define set_pud(pudptr, pudval) set_pmd((pmd_t *)(pudptr), (pmd_t) { pudval })
mm/huge_memory.c
1876
goto set_pmd;
mm/huge_memory.c
1943
set_pmd: