Symbol: PMD_MASK
arch/arc/include/asm/pgtable-levels.h
143
#define pmd_pfn(pmd) ((pmd_val(pmd) & PMD_MASK) >> PAGE_SHIFT)
arch/arm/include/asm/kfence.h
38
if (split_pmd_page(pmd, addr & PMD_MASK))
arch/arm/include/asm/pgtable-3level.h
210
#define pmd_pfn(pmd) (((pmd_val(pmd) & PMD_MASK) & PHYS_MASK) >> PAGE_SHIFT)
arch/arm/include/asm/tlb.h
42
addr = (addr & PMD_MASK) + SZ_1M;
arch/arm/mach-sa1100/assabet.c
603
unsigned long phys = __PREG(GPLR) & PMD_MASK;
arch/arm/mm/idmap.c
50
*pmd = __pmd((addr & PMD_MASK) | prot);
arch/arm/mm/idmap.c
60
addr = (addr & PMD_MASK) | prot;
arch/arm/mm/init.c
445
#define MODULES_VADDR (((unsigned long)_exiprom + ~PMD_MASK) & PMD_MASK)
arch/arm/mm/ioremap.c
335
} else if (!((paddr | size | addr) & ~PMD_MASK)) {
arch/arm/mm/mmu.c
1095
if ((addr & ~PMD_MASK) == SECTION_SIZE) {
arch/arm/mm/mmu.c
1098
pmd_empty_section_gap(addr & PMD_MASK);
arch/arm/mm/mmu.c
1107
if ((addr & ~PMD_MASK) == SECTION_SIZE) {
arch/arm/mm/mmu.c
1114
next = (addr + PMD_SIZE - 1) & PMD_MASK;
arch/arm/mm/mmu.c
1310
addr = ((unsigned long)_exiprom + PMD_SIZE - 1) & PMD_MASK;
arch/arm/mm/mmu.c
1382
for (addr = VMALLOC_START; addr < (FIXADDR_TOP & PMD_MASK); addr += PMD_SIZE)
arch/arm64/include/asm/pgtable.h
633
#define pmd_pfn(pmd) ((__pmd_to_phys(pmd) & PMD_MASK) >> PAGE_SHIFT)
arch/arm64/kvm/mmu.c
1422
*ipap &= PMD_MASK;
arch/arm64/mm/mmu.c
1597
if (!pgtable_range_aligned(start, end, floor, ceiling, PMD_MASK))
arch/arm64/mm/mmu.c
1817
VM_BUG_ON(phys & ~PMD_MASK);
arch/arm64/mm/mmu.c
265
if (((addr | next | phys) & ~PMD_MASK) == 0 &&
arch/csky/mm/init.c
130
vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK;
arch/loongarch/include/asm/pgtable.h
116
#define vmemmap ((struct page *)((VMALLOC_END + PMD_SIZE) & PMD_MASK))
arch/m68k/sun3x/dvma.c
120
if((end2 & PMD_MASK) > (vaddr & PMD_MASK))
arch/m68k/sun3x/dvma.c
121
end3 = (vaddr + (PMD_SIZE-1)) & PMD_MASK;
arch/microblaze/include/asm/highmem.h
46
& PMD_MASK)
arch/mips/mm/pgtable-32.c
63
fixrange_init(vaddr & PMD_MASK, vaddr + FIXADDR_SIZE, pgd_base);
arch/mips/mm/pgtable-32.c
70
fixrange_init(vaddr & PMD_MASK, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base);
arch/mips/mm/pgtable-64.c
117
vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK;
arch/nios2/mm/ioremap.c
31
address &= ~PMD_MASK;
arch/nios2/mm/ioremap.c
70
address = (address + PMD_SIZE) & PMD_MASK;
arch/parisc/kernel/asm-offsets.c
269
DEFINE(ASM_PMD_ENTRY, ((PAGE_OFFSET & PMD_MASK) >> PMD_SHIFT));
arch/parisc/kernel/pci-dma.c
120
vaddr = (vaddr + PMD_SIZE) & PMD_MASK;
arch/parisc/kernel/pci-dma.c
168
vaddr &= ~PMD_MASK;
arch/parisc/kernel/pci-dma.c
210
vaddr = (vaddr + PMD_SIZE) & PMD_MASK;
arch/parisc/kernel/pci-dma.c
83
vaddr &= ~PMD_MASK;
arch/powerpc/include/asm/highmem.h
53
#define PKMAP_BASE ((FIXADDR_START - PAGE_SIZE*(LAST_PKMAP + 1)) & PMD_MASK)
arch/powerpc/kvm/book3s_64_mmu_radix.c
1386
gpa = (gpa & PMD_MASK) + PMD_SIZE;
arch/powerpc/kvm/book3s_64_mmu_radix.c
713
unsigned long lgpa = gpa & PMD_MASK;
arch/powerpc/mm/book3s64/radix_tlb.c
1081
hstart = (start + PMD_SIZE - 1) & PMD_MASK;
arch/powerpc/mm/book3s64/radix_tlb.c
1082
hend = end & PMD_MASK;
arch/powerpc/mm/nohash/tlb_64e.c
59
unsigned long start = address & PMD_MASK;
arch/riscv/include/asm/pgtable.h
787
return ((__pmd_to_phys(pmd) & PMD_MASK) >> PAGE_SHIFT);
arch/riscv/kvm/mmu.c
438
*gpa &= PMD_MASK;
arch/riscv/mm/init.c
235
vmlinux_end = (vmlinux_end + PMD_SIZE - 1) & PMD_MASK;
arch/riscv/mm/init.c
246
phys_ram_base = memblock_start_of_DRAM() & PMD_MASK;
arch/riscv/mm/init.c
861
uintptr_t set_satp_mode_pmd = ((unsigned long)set_satp_mode) & PMD_MASK;
arch/riscv/mm/pageattr.c
108
vaddr <= (vaddr & PMD_MASK) && end >= next)
arch/s390/kernel/vdso.c
95
end = (start + len + PMD_SIZE - 1) & PMD_MASK;
arch/s390/mm/pageattr.c
158
new = __pmd(pmd_val(new) & PMD_MASK);
arch/s390/mm/pageattr.c
179
need_split |= !!(addr & ~PMD_MASK);
arch/s390/mm/vmem.c
293
try_free_pte_table(pmd, addr & PMD_MASK);
arch/sh/mm/init.c
320
vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK;
arch/sh/mm/init.c
321
end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK;
arch/sparc/include/asm/pgtable_32.h
17
#define PMD_ALIGN(__addr) (((__addr) + ~PMD_MASK) & PMD_MASK)
arch/sparc/include/asm/viking.h
241
vaddr &= ~PMD_MASK;
arch/sparc/mm/init_64.c
1751
if (guard && !(vstart & ~PMD_MASK) && (vend - vstart) >= PMD_SIZE)
arch/sparc/mm/init_64.c
1835
this_end = (vstart + PMD_SIZE) & PMD_MASK;
arch/sparc/mm/init_64.c
2585
vstart = vstart & PMD_MASK;
arch/sparc/mm/srmmu.c
703
start = (start + PMD_SIZE) & PMD_MASK;
arch/sparc/mm/srmmu.c
738
start = (start + PMD_SIZE) & PMD_MASK;
arch/sparc/mm/srmmu.c
792
if (!(start & ~(PMD_MASK))) {
arch/x86/boot/compressed/ident_map_64.c
205
address = __address & PMD_MASK;
arch/x86/boot/compressed/ident_map_64.c
369
address &= PMD_MASK;
arch/x86/boot/startup/map_kernel.c
115
if (load_delta & ~PMD_MASK)
arch/x86/boot/startup/sme.c
225
ppd->vaddr_end = vaddr_end & PMD_MASK;
arch/x86/include/asm/page_types.h
17
#define PHYSICAL_PMD_PAGE_MASK (((signed long)PMD_MASK) & __PHYSICAL_MASK)
arch/x86/include/asm/pgtable_32_areas.h
31
((FIXADDR_TOT_START - PAGE_SIZE*(CPU_ENTRY_AREA_PAGES+1)) & PMD_MASK)
arch/x86/include/asm/pgtable_32_areas.h
34
((CPU_ENTRY_AREA_BASE - PAGE_SIZE) & PMD_MASK)
arch/x86/include/asm/pgtable_32_areas.h
39
((LDT_BASE_ADDR - PAGE_SIZE) & PMD_MASK)
arch/x86/kernel/head64.c
154
pmd = (physaddr & PMD_MASK) + early_pmd_flags;
arch/x86/kernel/head64.c
231
BUILD_BUG_ON((__START_KERNEL_map & ~PMD_MASK) != 0);
arch/x86/kernel/head64.c
232
BUILD_BUG_ON((MODULES_VADDR & ~PMD_MASK) != 0);
arch/x86/mm/cpu_entry_area.c
252
BUG_ON(CPU_ENTRY_AREA_BASE & ~PMD_MASK);
arch/x86/mm/fault.c
269
for (addr = start & PMD_MASK;
arch/x86/mm/ident_map.c
83
addr &= PMD_MASK;
arch/x86/mm/init.c
842
if (((text_poke_mm_addr + PAGE_SIZE) & ~PMD_MASK) == 0)
arch/x86/mm/init_32.c
323
pfn &= PMD_MASK >> PAGE_SHIFT;
arch/x86/mm/init_32.c
487
vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK;
arch/x86/mm/init_32.c
488
end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK;
arch/x86/mm/init_64.c
395
BUG_ON((phys & ~PMD_MASK) || (size & ~PMD_MASK));
arch/x86/mm/init_64.c
536
paddr_next = (paddr & PMD_MASK) + PMD_SIZE;
arch/x86/mm/init_64.c
539
!e820__mapped_any(paddr & PMD_MASK, paddr_next,
arch/x86/mm/init_64.c
541
!e820__mapped_any(paddr & PMD_MASK, paddr_next,
arch/x86/mm/mem_encrypt_amd.c
166
pmd = map ? (paddr & PMD_MASK) + pmd_flags : 0;
arch/x86/mm/pat/set_memory.c
1153
lpaddr = address & PMD_MASK;
arch/x86/mm/pat/set_memory.c
1257
addr &= PMD_MASK;
arch/x86/mm/pat/set_memory.c
1263
if (PFN_PHYS(pfn) & ~PMD_MASK)
arch/x86/mm/pat/set_memory.c
1378
addr &= PMD_MASK;
arch/x86/mm/pat/set_memory.c
1461
unsigned long next_page = (start + PMD_SIZE) & PMD_MASK;
arch/x86/mm/pat/set_memory.c
1592
unsigned long next_page = (start + PMD_SIZE) & PMD_MASK;
arch/x86/mm/pat/set_memory.c
418
start = addr & PMD_MASK;
arch/x86/mm/pat/set_memory.c
867
offset = virt_addr & ~PMD_MASK;
arch/x86/mm/pti.c
351
WARN_ON_ONCE(addr & ~PMD_MASK);
arch/x86/platform/efi/efi_64.c
332
efi_va &= PMD_MASK;
arch/x86/platform/efi/efi_64.c
338
efi_va = (efi_va & PMD_MASK) + pa_offset;
arch/x86/power/hibernate_32.c
159
__pmd((jump_address_phys & PMD_MASK) | pgprot_val(PAGE_KERNEL_LARGE_EXEC)));
arch/x86/power/hibernate_64.c
69
__pmd((jump_address_phys & PMD_MASK) | pgprot_val(pmd_text_prot)));
arch/x86/xen/mmu_pv.c
1190
vaddr &= PMD_MASK;
arch/x86/xen/mmu_pv.c
1938
return pa + (vaddr & ~PMD_MASK);
arch/xtensa/include/asm/fixmap.h
32
#define FIXADDR_START ((FIXADDR_END - FIXADDR_SIZE) & PMD_MASK)
arch/xtensa/include/asm/highmem.h
22
(LAST_PKMAP + 1) * PAGE_SIZE) & PMD_MASK)
drivers/dax/device.c
146
unsigned long pmd_addr = vmf->address & PMD_MASK;
drivers/gpu/drm/drm_gem_shmem_helper.c
561
bool aligned = (vmf->address & ~PMD_MASK) == (paddr & ~PMD_MASK);
drivers/gpu/drm/drm_gem_shmem_helper.c
565
pfn &= PMD_MASK >> PAGE_SHIFT;
fs/dax.c
1942
unsigned long pmd_addr = vmf->address & PMD_MASK;
fs/proc/task_mmu.c
1859
#define PAGEMAP_WALK_MASK (PMD_MASK)
fs/proc/task_mmu.c
2005
unsigned int idx = (addr & ~PMD_MASK) >> PAGE_SHIFT;
include/linux/pgtable.h
1422
({ unsigned long __boundary = ((addr) + PMD_SIZE) & PMD_MASK; \
mm/debug_vm_pgtable.c
1137
args->fixed_pmd_pfn = __phys_to_pfn(phys & PMD_MASK);
mm/filemap.c
3771
(addr0 & PMD_MASK) == ((addr0 + folio_size(folio) - 1) & PMD_MASK)) {
mm/gup.c
2939
page = pmd_page(orig) + ((addr & ~PMD_MASK) >> PAGE_SHIFT);
mm/hmm.c
214
pfn = pmd_pfn(pmd) + ((addr & ~PMD_MASK) >> PAGE_SHIFT);
mm/huge_memory.c
1610
unsigned long addr = vmf->address & PMD_MASK;
mm/huge_memory.c
1637
unsigned long addr = vmf->address & PMD_MASK;
mm/hugetlb_vmemmap.c
136
return vmemmap_split_pmd(pmd, head, addr & PMD_MASK, vmemmap_walk);
mm/memory.c
344
addr &= PMD_MASK;
mm/memory.c
351
ceiling &= PMD_MASK;
mm/memory.c
4643
vmf->address & PMD_MASK, &ptl);
mm/memory.c
4903
if (unlikely(folio_start < max(address & PMD_MASK, vma->vm_start)))
mm/memory.c
5157
pte = pte_offset_map(vmf->pmd, vmf->address & PMD_MASK);
mm/memory.c
6853
pmd_pfn(pmd), PMD_MASK, pmd_write(pmd),
mm/mremap.c
552
mask = PMD_MASK;
mm/mremap.c
814
try_realign_addr(pmc, PMD_MASK);
mm/swap_state.c
806
*start = max3(left, vma->vm_start, faddr & PMD_MASK);
mm/swap_state.c
807
*end = min3(right, vma->vm_end, (faddr & PMD_MASK) + PMD_SIZE);
mm/vmalloc.c
842
return pmd_page(*pmd) + ((addr & ~PMD_MASK) >> PAGE_SHIFT);
mm/vmscan.c
3511
pte = pte_offset_map_rw_nolock(args->mm, pmd, start & PMD_MASK, &pmdval, &ptl);
mm/vmscan.c
3563
if (i < PTRS_PER_PTE && get_next_vma(PMD_MASK, PAGE_SIZE, args, &start, &end))
mm/vmscan.c
3614
addr = i ? (*first & PMD_MASK) + i * PMD_SIZE : *first;
mm/vmscan.c
4237
start = max(addr & PMD_MASK, vma->vm_start);
mm/vmscan.c
4238
end = min(addr | ~PMD_MASK, vma->vm_end - 1) + 1;