arch/arm/mm/fault-armv.c
127
const unsigned long pmd_start_addr = ALIGN_DOWN(addr, PMD_SIZE);
arch/arm/mm/init.c
136
ALIGN_DOWN(addr, pageblock_size),
arch/arm64/include/asm/kvm_pkvm.h
85
start = ALIGN_DOWN(start, PAGE_SIZE);
arch/arm64/kernel/ftrace.c
371
unsigned long literal = ALIGN_DOWN(rec->ip - 12, 8);
arch/arm64/kernel/io.c
22
const u##bits *_end_from = _from + ALIGN_DOWN(_count, 8); \
arch/arm64/kernel/module-plts.c
55
p = ALIGN_DOWN((u64)a, SZ_4K);
arch/arm64/kernel/module-plts.c
56
q = ALIGN_DOWN((u64)b, SZ_4K);
arch/arm64/kernel/rsi.c
116
base = ALIGN_DOWN(base, RSI_GRANULE_SIZE);
arch/arm64/kernel/static_call.c
15
literal = ALIGN_DOWN((u64)tramp + 4, SZ_4K) +
arch/arm64/kvm/hyp/nvhe/mem_protect.c
519
cur.start = ALIGN_DOWN(addr, granule);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
851
u64 cur, start = ALIGN_DOWN((u64)from, PAGE_SIZE);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
887
u64 cur, start = ALIGN_DOWN((u64)from, PAGE_SIZE);
arch/arm64/kvm/hyp/nvhe/mm.c
156
start = ALIGN_DOWN((u64)hyp_phys_to_page(start), PAGE_SIZE);
arch/arm64/kvm/hyp/nvhe/mm.c
414
start = ALIGN_DOWN(start, PAGE_SIZE);
arch/arm64/kvm/hyp/pgtable.c
1096
.phys = ALIGN_DOWN(phys, PAGE_SIZE),
arch/arm64/kvm/hyp/pgtable.c
201
data->addr = ALIGN_DOWN(data->addr, kvm_granule_size(level));
arch/arm64/kvm/hyp/pgtable.c
272
.start = ALIGN_DOWN(addr, PAGE_SIZE),
arch/arm64/kvm/hyp/pgtable.c
273
.addr = ALIGN_DOWN(addr, PAGE_SIZE),
arch/arm64/kvm/hyp/pgtable.c
316
ret = kvm_pgtable_walk(pgt, ALIGN_DOWN(addr, PAGE_SIZE),
arch/arm64/kvm/hyp/pgtable.c
456
.phys = ALIGN_DOWN(phys, PAGE_SIZE),
arch/arm64/kvm/hyp/pgtable.c
852
u64 addr = ALIGN_DOWN(ctx->addr, size);
arch/arm64/kvm/mmu.c
1449
ALIGN_DOWN(hva, PUD_SIZE) >= vma->vm_start &&
arch/arm64/kvm/mmu.c
1455
ALIGN_DOWN(hva, PMD_SIZE) >= vma->vm_start &&
arch/arm64/kvm/mmu.c
1762
fault_ipa = ALIGN_DOWN(fault_ipa, vma_pagesize);
arch/arm64/kvm/mmu.c
1763
ipa = ALIGN_DOWN(ipa, vma_pagesize);
arch/arm64/kvm/mmu.c
2305
hyp_idmap_start = ALIGN_DOWN(hyp_idmap_start, PAGE_SIZE);
arch/arm64/kvm/mmu.c
43
phys_addr_t boundary = ALIGN_DOWN(addr + size, size);
arch/arm64/kvm/mmu.c
553
start = ALIGN_DOWN(__pa(from), PAGE_SIZE);
arch/arm64/kvm/mmu.c
573
start = ALIGN_DOWN(__pa(from), PAGE_SIZE);
arch/arm64/kvm/mmu.c
720
base = ALIGN_DOWN(io_map_base - size, size);
arch/arm64/lib/insn.c
1220
offset = (addr - ALIGN_DOWN(pc, SZ_4K)) >> 12;
arch/arm64/mm/contpte.c
272
cont_start = ALIGN_DOWN(addr, CONT_PTE_SIZE);
arch/arm64/mm/contpte.c
278
pfn = ALIGN_DOWN(pte_pfn(pte), CONT_PTES);
arch/arm64/mm/contpte.c
42
*start = ALIGN_DOWN(*start, CONT_PTE_SIZE);
arch/arm64/mm/contpte.c
676
start_addr = addr = ALIGN_DOWN(addr, CONT_PTE_SIZE);
arch/arm64/mm/contpte.c
78
start_addr = addr = ALIGN_DOWN(addr, CONT_PTE_SIZE);
arch/arm64/mm/contpte.c
79
pte = pfn_pte(ALIGN_DOWN(pte_pfn(pte), CONT_PTES), pte_pgprot(pte));
arch/arm64/mm/kasan_init.c
391
shadow_start = ALIGN_DOWN(shadow_start, PAGE_SIZE);
arch/arm64/mm/mmu.c
680
if (ALIGN_DOWN(addr, PGDIR_SIZE) == addr)
arch/arm64/mm/mmu.c
691
if (ALIGN_DOWN(addr, P4D_SIZE) == addr)
arch/arm64/mm/mmu.c
703
if (ALIGN_DOWN(addr, PUD_SIZE) == addr)
arch/arm64/mm/mmu.c
720
if (ALIGN_DOWN(addr, CONT_PMD_SIZE) == addr)
arch/arm64/mm/mmu.c
733
if (ALIGN_DOWN(addr, PMD_SIZE) == addr)
arch/arm64/mm/mmu.c
745
if (ALIGN_DOWN(addr, CONT_PTE_SIZE) == addr)
arch/loongarch/kernel/relocate.c
104
size = ALIGN_DOWN(size - diff, sizeof(hash));
arch/loongarch/kvm/mmu.c
638
return (hva >= ALIGN(start, PMD_SIZE)) && (hva < ALIGN_DOWN(end, PMD_SIZE));
arch/mips/kernel/relocate.c
204
size = ALIGN_DOWN(size - diff, sizeof(hash));
arch/mips/mm/init.c
286
cfg->upper = ALIGN_DOWN(PFN_PHYS(start_pfn + nr_pages), maar_align) - 1;
arch/openrisc/kernel/setup.c
81
unsigned long aligned_start = ALIGN_DOWN(initrd_start, PAGE_SIZE);
arch/powerpc/include/asm/book3s/32/pgtable.h
193
#define VMALLOC_END ALIGN_DOWN(ioremap_bot, PAGE_SIZE << KASAN_SHADOW_SCALE_SHIFT)
arch/powerpc/include/asm/fixmap.h
85
ALIGN_DOWN(FIXADDR_START, PGDIR_SIZE))
arch/powerpc/include/asm/kasan.h
23
#define KASAN_KERN_START ALIGN_DOWN(PAGE_OFFSET - SZ_256M, SZ_256M)
arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h
26
ptep = pte_offset_kernel((pmd_t *)ptep, ALIGN_DOWN(addr, SZ_8M));
arch/powerpc/include/asm/nohash/32/pgtable.h
111
#define VMALLOC_END ALIGN_DOWN(ioremap_bot, PAGE_SIZE << KASAN_SHADOW_SCALE_SHIFT)
arch/powerpc/include/asm/nohash/32/pte-8xx.h
172
return (pmd_t *)ptep == pmd_off(mm, ALIGN_DOWN(addr, SZ_8M));
arch/powerpc/kernel/fadump.c
107
end = ALIGN_DOWN(end, CMA_MIN_ALIGNMENT_BYTES);
arch/powerpc/kernel/hw_breakpoint.c
162
start_addr = ALIGN_DOWN(hw->address, HW_BREAKPOINT_SIZE);
arch/powerpc/kernel/hw_breakpoint.c
170
(ALIGN_DOWN(start_addr, SZ_512) != ALIGN_DOWN(end_addr - 1, SZ_512)))
arch/powerpc/kernel/hw_breakpoint_constraints.c
25
hw_start_addr = ALIGN_DOWN(info->address, HW_BREAKPOINT_SIZE);
arch/powerpc/kernel/hw_breakpoint_constraints.c
44
hw_start_addr = ALIGN_DOWN(info->address, align_size);
arch/powerpc/kernel/pci_64.c
157
phys_page = ALIGN_DOWN(hose->io_base_phys, PAGE_SIZE);
arch/powerpc/kernel/process.c
835
unsigned long start_addr = ALIGN_DOWN(brk->address, HW_BREAKPOINT_SIZE);
arch/powerpc/kernel/prom.c
103
return (start + size) > ALIGN_DOWN(initrd_start, PAGE_SIZE) &&
arch/powerpc/kernel/prom.c
690
memblock_reserve(ALIGN_DOWN(__pa(initrd_start), PAGE_SIZE),
arch/powerpc/kernel/prom.c
692
ALIGN_DOWN(initrd_start, PAGE_SIZE));
arch/powerpc/kernel/prom.c
870
memory_limit = ALIGN_DOWN(memory_limit ?: memblock_phys_mem_size(), SZ_16M);
arch/powerpc/kernel/prom_init.c
1552
addr = ALIGN_DOWN(alloc_top_high - size, align);
arch/powerpc/kernel/prom_init.c
1570
base = ALIGN_DOWN(alloc_top - size, align);
arch/powerpc/kernel/prom_init.c
1572
base = ALIGN_DOWN(base - 0x100000, align)) {
arch/powerpc/kernel/prom_init.c
1638
base = ALIGN_DOWN(base, PAGE_SIZE);
arch/powerpc/kernel/prom_init.c
821
prom_memory_limit = ALIGN_DOWN(prom_memory_limit, SZ_16M);
arch/powerpc/kernel/ptrace/ptrace-noadv.c
217
brk.address = ALIGN_DOWN(bp_info->addr, HW_BREAKPOINT_SIZE);
arch/powerpc/mm/book3s32/mmu.c
189
if (addr < ALIGN_DOWN(MODULES_VADDR, SZ_256M))
arch/powerpc/mm/book3s64/hash_pgtable.c
511
start = ALIGN_DOWN(start, step);
arch/powerpc/mm/book3s64/hash_tlb.c
197
start = ALIGN_DOWN(start, PAGE_SIZE);
arch/powerpc/mm/book3s64/hash_tlb.c
232
addr = ALIGN_DOWN(addr, PMD_SIZE);
arch/powerpc/mm/book3s64/hash_utils.c
640
vend = ALIGN_DOWN(vend, step);
arch/powerpc/mm/book3s64/hash_utils.c
722
vaddr = ALIGN_DOWN(vstart, step);
arch/powerpc/mm/book3s64/radix_pgtable.c
1141
start = ALIGN_DOWN(start, PAGE_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
1143
start = ALIGN_DOWN(start, PMD_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
1356
unsigned long pfn_offset = addr_pfn - ALIGN_DOWN(addr_pfn, nr_pages);
arch/powerpc/mm/book3s64/radix_pgtable.c
198
start = ALIGN_DOWN(start, PAGE_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
315
end = ALIGN_DOWN(end, PAGE_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
741
unsigned long start = ALIGN_DOWN(addr, PMD_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
748
unsigned long start = ALIGN_DOWN(addr, PAGE_SIZE);
arch/powerpc/mm/init_64.c
210
start = ALIGN_DOWN(start, page_size);
arch/powerpc/mm/init_64.c
328
start = ALIGN_DOWN(start, page_size);
arch/powerpc/mm/kasan/8xx.c
53
kasan_init_shadow_8M(k_start, ALIGN_DOWN(k_end, SZ_8M), block);
arch/powerpc/mm/kasan/8xx.c
54
k_cur = ALIGN_DOWN(k_end, SZ_8M);
arch/powerpc/mm/kasan/8xx.c
70
if (k_cur < ALIGN_DOWN(k_end, SZ_512K))
arch/powerpc/mm/kasan/init_book3e_64.c
74
k_start = ALIGN_DOWN((unsigned long)kasan_mem_to_shadow(start), PAGE_SIZE);
arch/powerpc/mm/kasan/init_book3s_64.c
30
k_start = ALIGN_DOWN((unsigned long)kasan_mem_to_shadow(start), PAGE_SIZE);
arch/powerpc/mm/mem.c
390
for (va = ALIGN_DOWN(MODULES_VADDR, PGDIR_SIZE); va < MODULES_END; va += PGDIR_SIZE)
arch/powerpc/mm/nohash/8xx.c
130
for (; p < ALIGN_DOWN(top, SZ_8M) && p < top && !err; p += SZ_8M, v += SZ_8M)
arch/powerpc/mm/nohash/8xx.c
132
for (; p < ALIGN_DOWN(top, SZ_512K) && p < top && !err; p += SZ_512K, v += SZ_512K)
arch/powerpc/mm/nohash/8xx.c
134
for (; p < ALIGN_DOWN(top, SZ_16K) && p < top && !err; p += SZ_16K, v += SZ_16K)
arch/powerpc/mm/pageattr.c
77
unsigned long start = ALIGN_DOWN(addr, PAGE_SIZE);
arch/powerpc/mm/pgtable_32.c
40
unsigned long addr = ALIGN_DOWN(FIXADDR_START, PGDIR_SIZE);
arch/powerpc/platforms/powernv/opal-fadump.c
677
fadump_conf->max_copy_size = ALIGN_DOWN(U32_MAX, PAGE_SIZE);
arch/powerpc/platforms/powernv/pci-ioda.c
267
start = ALIGN_DOWN(r->start - base, sgsz);
arch/powerpc/platforms/ps3/mm.c
269
r->size = ALIGN_DOWN(size, 1 << PAGE_SHIFT_16M);
arch/powerpc/platforms/ps3/mm.c
401
unsigned long aligned_bus = ALIGN_DOWN(bus_addr, 1 << r->page_size);
arch/powerpc/platforms/ps3/mm.c
430
unsigned long aligned_lpar = ALIGN_DOWN(lpar_addr, 1 << r->page_size);
arch/powerpc/platforms/ps3/mm.c
782
unsigned long aligned_phys = ALIGN_DOWN(phys_addr, 1 << r->page_size);
arch/powerpc/platforms/ps3/mm.c
837
unsigned long aligned_phys = ALIGN_DOWN(phys_addr, 1 << r->page_size);
arch/powerpc/platforms/ps3/mm.c
896
unsigned long aligned_bus = ALIGN_DOWN(bus_addr,
arch/powerpc/platforms/ps3/mm.c
933
unsigned long aligned_bus = ALIGN_DOWN(bus_addr,
arch/powerpc/platforms/pseries/htmdump.c
94
page = ALIGN_DOWN(*ppos, PAGE_SIZE);
arch/powerpc/platforms/pseries/papr_scm.c
1301
u64 aligned_addr = ALIGN_DOWN(phys_addr, L1_CACHE_BYTES);
arch/powerpc/platforms/pseries/rtas-fadump.c
608
fadump_conf->max_copy_size = ALIGN_DOWN(U64_MAX, PAGE_SIZE);
arch/riscv/kernel/ftrace.c
98
unsigned long literal = ALIGN_DOWN(rec->ip - 12, 8);
arch/riscv/kvm/mmu.c
363
return (hva >= ALIGN(uaddr_start, PMD_SIZE)) && (hva < ALIGN_DOWN(uaddr_end, PMD_SIZE));
arch/s390/kvm/dat.c
1314
for (cur = ALIGN_DOWN(gfn, _PAGE_ENTRIES); cur < gfn + count; cur += _PAGE_ENTRIES) {
arch/s390/kvm/dat.c
546
for (cur = ALIGN_DOWN(start, cur_size); cur < end; idx++, cur = next) {
arch/s390/kvm/dat.h
578
return (struct crst_table *)ALIGN_DOWN((unsigned long)crstep, _CRST_TABLE_SIZE);
arch/s390/kvm/dat.h
583
return (struct page_table *)ALIGN_DOWN((unsigned long)ptep, _PAGE_TABLE_SIZE);
arch/s390/kvm/gmap.c
1071
kvm_s390_vsie_gmap_notifier(sg, ALIGN_DOWN(gaddr, align), ALIGN(gaddr + 1, align));
arch/s390/kvm/gmap.h
212
gfn = ALIGN_DOWN(gfn, align);
arch/s390/kvm/kvm-s390.c
3272
.memory_size = ALIGN_DOWN(TASK_SIZE, _SEGMENT_SIZE),
arch/s390/mm/vmem.c
138
unsigned long page = ALIGN_DOWN(start, PMD_SIZE);
arch/s390/mm/vmem.c
160
unsigned long page = ALIGN_DOWN(start, PMD_SIZE);
arch/sh/drivers/pci/pcie-sh7786.c
369
memstart = ALIGN_DOWN(memstart, memsize);
arch/x86/boot/compressed/kaslr.c
495
pud_end = ALIGN_DOWN(region->start + region->size, PUD_SIZE);
arch/x86/kernel/cpu/sgx/encl.c
658
align = ALIGN_DOWN(addr + i, sizeof(unsigned long));
arch/x86/kernel/traps.c
1084
sp = ALIGN_DOWN(sp, 8) - sizeof(*regs_ret);
arch/x86/kvm/svm/sev.c
4973
kvm_pfn_t pfn_aligned = ALIGN_DOWN(pfn, PTRS_PER_PMD);
arch/x86/kvm/svm/sev.c
5013
pfn_aligned = ALIGN_DOWN(pfn, PTRS_PER_PMD);
arch/x86/kvm/svm/sev.c
5014
gfn_aligned = ALIGN_DOWN(gfn, PTRS_PER_PMD);
arch/x86/mm/init.c
647
real_end = max(map_start, ALIGN_DOWN(map_end, PMD_SIZE));
arch/x86/mm/init_64.c
870
unsigned long start = ALIGN_DOWN(addr, PMD_SIZE);
arch/x86/mm/init_64.c
920
const unsigned long page = ALIGN_DOWN(start, PMD_SIZE);
arch/x86/virt/svm/sev.c
197
pa = ALIGN_DOWN(pa, PMD_SIZE);
arch/x86/virt/svm/sev.c
818
pfn_i = ALIGN_DOWN(pfn, PTRS_PER_PMD);
arch/x86/virt/vmx/tdx/tdx.c
358
#define TDMR_ALIGN_DOWN(_addr) ALIGN_DOWN((_addr), TDMR_ALIGNMENT)
block/blk-merge.c
408
bytes = ALIGN_DOWN(bytes, bio_split_alignment(bio, lim));
block/fops.c
449
iomap->offset = ALIGN_DOWN(offset, bdev_logical_block_size(bdev));
drivers/accel/habanalabs/common/memory.c
1548
dma_max_seg_size = ALIGN_DOWN(dma_get_max_seg_size(dev), PAGE_SIZE);
drivers/accel/ivpu/ivpu_mmu_context.c
362
u64 start = ALIGN_DOWN(vpu_addr, IVPU_MMU_CONT_PAGES_SIZE);
drivers/acpi/nfit/mce.c
67
ALIGN_DOWN(mce->addr, align), align);
drivers/block/null_blk/main.c
1363
first_bad = ALIGN_DOWN(first_bad, block_sectors);
drivers/block/null_blk/zoned.c
108
min(ALIGN_DOWN(dev->zone_append_max_sectors,
drivers/cache/hisi_soc_hha.c
75
addr = ALIGN_DOWN(addr, HISI_HHA_MAINT_ALIGN);
drivers/cache/sifive_ccache.c
155
for (line = ALIGN_DOWN(start, SIFIVE_CCACHE_LINE_SIZE); line < end;
drivers/crypto/ccp/sev-dev.c
388
paddr = __sme_clr(ALIGN_DOWN(paddr, PAGE_SIZE));
drivers/crypto/exynos-rng.c
104
slen = ALIGN_DOWN(slen, 4);
drivers/crypto/stm32/stm32-cryp.c
1058
align_size = ALIGN_DOWN(cryp->payload_in, cryp->hw_blocksize);
drivers/crypto/stm32/stm32-cryp.c
1563
align_size = ALIGN_DOWN(cryp->payload_in, cryp->hw_blocksize);
drivers/crypto/stm32/stm32-cryp.c
1615
align_size = ALIGN_DOWN(cryp->header_in, cryp->hw_blocksize);
drivers/crypto/stm32/stm32-cryp.c
1634
align_size = ALIGN_DOWN(cryp->payload_in, cryp->hw_blocksize);
drivers/crypto/stm32/stm32-cryp.c
995
align_size = ALIGN_DOWN(cryp->header_in, cryp->hw_blocksize);
drivers/cxl/core/atl.c
116
hpa_range.start = ALIGN_DOWN(hpa_range.start, SZ_256M);
drivers/dax/device.c
96
ALIGN_DOWN(vmf->address, fault_size));
drivers/dax/kmem.c
38
r->end = ALIGN_DOWN(range->end + 1, memory_block_size_bytes()) - 1;
drivers/dma/stm32/stm32-dma.c
130
ALIGN_DOWN(STM32_DMA_MAX_DATA_ITEMS, 16)
drivers/dma/stm32/stm32-dma3.c
232
#define STM32_DMA3_MAX_BLOCK_SIZE ALIGN_DOWN(CBR1_BNDT, 64)
drivers/dpll/zl3073x/flash.c
349
max_block_size = ALIGN_DOWN(ZL_FLASH_MAX_BLOCK_SIZE, sector_size);
drivers/firmware/efi/efi.c
781
p = early_memremap(ALIGN_DOWN(prsv, PAGE_SIZE),
drivers/firmware/efi/libstub/alignedmem.c
38
alloc_addr = ALIGN_DOWN(max + 1, align) - 1;
drivers/firmware/efi/libstub/mem.c
95
alloc_addr = ALIGN_DOWN(max + 1, EFI_ALLOC_ALIGN) - 1;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
809
return ALIGN_DOWN(tmp, PAGE_SIZE);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1696
available = ALIGN_DOWN(available, PAGE_SIZE);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1698
available = ALIGN_DOWN(vram_available, VRAM_AVAILABLITY_ALIGN);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1508
uint64_t aligned_pos = ALIGN_DOWN(pos, 4);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
996
start = max(ALIGN_DOWN(addr, size), prange->start);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1148
unsigned long last_align_down = ALIGN_DOWN(prange->last, 512);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1172
unsigned long last_align_down = ALIGN_DOWN(prange->last, 512);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2146
l = min(last, ALIGN_DOWN(start + max_pages, max_pages) - 1);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2817
(unsigned long)ALIGN_DOWN(addr, 1UL << p->svms.default_granularity));
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3199
start = max_t(unsigned long, ALIGN_DOWN(addr, size), prange->start);
drivers/gpu/drm/ast/ast_cursor.c
62
return ALIGN_DOWN(ast->vram_size - size, SZ_8);
drivers/gpu/drm/drm_gpusvm.c
531
notifier->itree.start = ALIGN_DOWN(fault_addr, gpusvm->notifier_size);
drivers/gpu/drm/drm_gpusvm.c
640
range->itree.start = ALIGN_DOWN(fault_addr, chunk_size);
drivers/gpu/drm/drm_gpusvm.c
899
start = ALIGN_DOWN(fault_addr, gpusvm->chunk_sizes[i]);
drivers/gpu/drm/drm_mipi_dbi.c
1352
max_chunk = ALIGN_DOWN(max_chunk, 2);
drivers/gpu/drm/drm_pagemap.c
1064
start = ALIGN_DOWN(fault_addr, size);
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
904
cmdbuf_paddr = ALIGN_DOWN(etnaviv_cmdbuf_get_pa(&gpu->buffer), SZ_128M);
drivers/gpu/drm/gud/gud_pipe.c
114
rect->x1 = ALIGN_DOWN(rect->x1, block_width);
drivers/gpu/drm/gud/gud_pipe.c
67
rect->x1 = ALIGN_DOWN(rect->x1, block_width);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
970
dsm_size = ALIGN_DOWN(lmem_size - dsm_base, SZ_1M);
drivers/gpu/drm/i915/gvt/kvmgt.c
785
ALIGN_DOWN(off, PAGE_SIZE),
drivers/gpu/drm/mediatek/mtk_disp_ovl_adaptor.c
169
align_width = ALIGN_DOWN(pending->width, 2);
drivers/gpu/drm/mediatek/mtk_ethdr.c
163
unsigned int align_width = ALIGN_DOWN(pending->width, 2);
drivers/gpu/drm/msm/msm_iommu.c
315
pt_count = ((ALIGN(iova + len, 1ull << 39) - ALIGN_DOWN(iova, 1ull << 39)) >> 39) +
drivers/gpu/drm/msm/msm_iommu.c
316
((ALIGN(iova + len, 1ull << 30) - ALIGN_DOWN(iova, 1ull << 30)) >> 30) +
drivers/gpu/drm/msm/msm_iommu.c
317
((ALIGN(iova + len, 1ull << 21) - ALIGN_DOWN(iova, 1ull << 21)) >> 21);
drivers/gpu/drm/nouveau/nouveau_dmem.c
221
args.start = ALIGN_DOWN(vmf->address, (PAGE_SIZE << order));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
249
meta->gspFwWprEnd = ALIGN_DOWN(gsp->fb.bios.vga_workspace.addr, 0x20000);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
323
gsp->fb.wpr2.frts.addr = ALIGN_DOWN(gsp->fb.bios.addr, 0x20000) - gsp->fb.wpr2.frts.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
326
gsp->fb.wpr2.boot.addr = ALIGN_DOWN(gsp->fb.wpr2.frts.addr - gsp->fb.wpr2.boot.size, 0x1000);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
329
gsp->fb.wpr2.elf.addr = ALIGN_DOWN(gsp->fb.wpr2.boot.addr - gsp->fb.wpr2.elf.size, 0x10000);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
333
gsp->fb.wpr2.heap.addr = ALIGN_DOWN(gsp->fb.wpr2.elf.addr - gsp->fb.wpr2.heap.size, 0x100000);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
334
gsp->fb.wpr2.heap.size = ALIGN_DOWN(gsp->fb.wpr2.elf.addr - gsp->fb.wpr2.heap.addr, 0x100000);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
336
gsp->fb.wpr2.addr = ALIGN_DOWN(gsp->fb.wpr2.heap.addr - sizeof(GspFwWprMeta), 0x100000);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1806
tail = ALIGN_DOWN(tail, vmm->func->page_block);
drivers/gpu/drm/panthor/panthor_mmu.c
1256
pt_count = ((ALIGN(va + size, 1ull << 39) - ALIGN_DOWN(va, 1ull << 39)) >> 39) +
drivers/gpu/drm/panthor/panthor_mmu.c
1257
((ALIGN(va + size, 1ull << 30) - ALIGN_DOWN(va, 1ull << 30)) >> 30) +
drivers/gpu/drm/panthor/panthor_mmu.c
1258
((ALIGN(va + size, 1ull << 21) - ALIGN_DOWN(va, 1ull << 21)) >> 21);
drivers/gpu/drm/panthor/panthor_mmu.c
2124
aligned_unmap_start = ALIGN_DOWN(*unmap_start, SZ_2M);
drivers/gpu/drm/xe/xe_eu_stall.c
582
aligned_count = ALIGN_DOWN(count, stream->data_record_size);
drivers/gpu/drm/xe/xe_ggtt.c
1161
hole_end = ALIGN_DOWN(hole_end, alignment);
drivers/gpu/drm/xe/xe_ggtt.c
983
hole_end = ALIGN_DOWN(hole_end, alignment);
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1147
return ALIGN_DOWN(GUC_ID_MAX, SZ_1K);
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1850
fair = ALIGN_DOWN(fair, alignment);
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
742
fair = ALIGN_DOWN(fair, alignment);
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
751
u64 shareable = ALIGN_DOWN(GUC_GGTT_TOP, SZ_512M);
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
755
return ALIGN_DOWN(shareable, alignment);
drivers/gpu/drm/xe/xe_guc_tlb_inval.c
154
start = ALIGN_DOWN(start, align);
drivers/gpu/drm/xe/xe_guc_tlb_inval.c
159
start = ALIGN_DOWN(orig_start, length);
drivers/gpu/drm/xe/xe_guc_tlb_inval.c
168
start = ALIGN_DOWN(orig_start, length);
drivers/gpu/drm/xe/xe_migrate.c
2039
chunk = ALIGN_DOWN(chunk, PAGE_SIZE / XE_PAGE_SIZE);
drivers/gpu/drm/xe/xe_pt.c
1923
start = ALIGN_DOWN(start, 0x1ull << xe_pt_shift(level));
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
121
return ALIGN_DOWN(stolen_size, SZ_1M);
drivers/hv/hv_balloon.c
996
rg_start = ALIGN_DOWN(pg_start, ha_pages_in_chunk);
drivers/hv/mshv_regions.c
483
page_offset = ALIGN_DOWN(gfn - region->start_gfn,
drivers/infiniband/core/umem_dmabuf.c
39
start = ALIGN_DOWN(umem_dmabuf->umem.address, PAGE_SIZE);
drivers/infiniband/core/umem_odp.c
354
range.start = ALIGN_DOWN(user_virt, 1UL << page_shift);
drivers/infiniband/core/umem_odp.c
72
start = ALIGN_DOWN(umem_odp->umem.address, page_size);
drivers/infiniband/hw/hfi1/pin_system.c
207
u64 start = ALIGN_DOWN(req_start, PAGE_SIZE);
drivers/infiniband/hw/hfi1/pin_system.c
310
page_offset = start - ALIGN_DOWN(start, PAGE_SIZE);
drivers/iommu/generic_pt/kunit_iommu_pt.h
361
start = ALIGN_DOWN(start, priv->smallest_pgsz);
drivers/iommu/intel/cache.c
290
return ALIGN_DOWN(start, VTD_PAGE_SIZE << mask);
drivers/iommu/intel/iommu.h
1100
addr = ALIGN_DOWN(addr, align);
drivers/iommu/iommufd/io_pagetable.c
497
start_byte = start - ALIGN_DOWN(start, PAGE_SIZE);
drivers/iommu/iommufd/pages.c
1404
(void __user *)ALIGN_DOWN((uintptr_t)uptr, PAGE_SIZE);
drivers/iommu/iommufd/selftest.c
1593
ALIGN_DOWN(iova, PAGE_SIZE)) /
drivers/iommu/iommufd/selftest.c
1619
uptr - (iova - ALIGN_DOWN(iova, PAGE_SIZE)), pages,
drivers/iommu/virtio-iommu.c
409
limit = ALIGN_DOWN(limit + 1, granule) - 1;
drivers/iommu/virtio-iommu.c
412
u64 resv_start = ALIGN_DOWN(resv->start, granule);
drivers/mailbox/mailbox-mpfs.c
139
u8 byte_off = ALIGN_DOWN(msg->cmd_data_size, 4);
drivers/media/i2c/ov5693.c
765
return ALIGN_DOWN(OV5693_PIXEL_RATE / OV5693_FIXED_PPL / tgt_fps, 2);
drivers/media/platform/arm/mali-c55/mali-c55-tpg.c
75
def_vblank = ALIGN_DOWN(def_vblank, 2);
drivers/media/platform/nxp/imx8-isi/imx8-isi-video.c
1226
fsize->stepwise.max_width = ALIGN_DOWN(max_width, h_align);
drivers/media/platform/nxp/imx8-isi/imx8-isi-video.c
1227
fsize->stepwise.max_height = ALIGN_DOWN(MXC_ISI_MAX_HEIGHT, v_align);
drivers/mfd/lpc_ich.c
1326
res->start = ALIGN_DOWN(spi_base, SPIBASE_BYT_SZ);
drivers/misc/lkdtm/kstack_erase.c
72
untracked_high = ALIGN_DOWN(untracked_high, sizeof(unsigned long));
drivers/mtd/devices/mtd_intel_dg.c
248
to4 = ALIGN_DOWN(to, sizeof(u32));
drivers/mtd/devices/mtd_intel_dg.c
278
len8 = ALIGN_DOWN(len_s, sizeof(u64));
drivers/mtd/devices/mtd_intel_dg.c
321
from4 = ALIGN_DOWN(from, sizeof(u32));
drivers/mtd/devices/mtd_intel_dg.c
352
len8 = ALIGN_DOWN(len_s, sizeof(u64));
drivers/mtd/mtdpstore.c
114
off = ALIGN_DOWN(off, mtd->erasesize);
drivers/mtd/mtdpstore.c
156
off = ALIGN_DOWN(off, mtd->erasesize);
drivers/mtd/mtdpstore.c
172
off = ALIGN_DOWN(off, mtd->erasesize);
drivers/mtd/mtdpstore.c
189
off = ALIGN_DOWN(off, cxt->mtd->erasesize);
drivers/mtd/mtdpstore.c
33
off = ALIGN_DOWN(off, mtd->erasesize);
drivers/mtd/mtdpstore.c
55
off = ALIGN_DOWN(off, mtd->erasesize);
drivers/mtd/mtdpstore.c
87
off = ALIGN_DOWN(off, mtd->erasesize);
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1043
return ALIGN_DOWN(raw_len, this->bch_geometry.eccn_chunk_size);
drivers/mtd/nand/raw/loongson-nand-controller.c
329
col0 = ALIGN_DOWN(op->addrs[0], chip->buf_align);
drivers/net/ethernet/broadcom/asp2/bcmasp.c
288
ALIGN_DOWN(offset, 4));
drivers/net/ethernet/broadcom/asp2/bcmasp.c
291
ALIGN_DOWN(offset, 4));
drivers/net/ethernet/broadcom/asp2/bcmasp.c
303
match_val, ALIGN_DOWN(offset, 4));
drivers/net/ethernet/broadcom/asp2/bcmasp.c
305
mask_val, ALIGN_DOWN(offset, 4));
drivers/net/ethernet/broadcom/bnge/bnge_hwrm.c
496
offset = ALIGN_DOWN(offset, BNGE_HWRM_DMA_ALIGN);
drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c
791
offset = ALIGN_DOWN(offset, BNXT_HWRM_DMA_ALIGN);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3512
priv->rx_buf_size = ALIGN_DOWN(DPAA2_ETH_RX_BUF_SIZE, rx_buf_align);
drivers/net/ethernet/intel/ice/ice_xsk.c
776
batched = ALIGN_DOWN(nb_pkts, PKTS_PER_BATCH);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3745
idpf_rx_buf_hw_update(bufq, ALIGN_DOWN(bufq_nta,
drivers/net/ethernet/intel/idpf/idpf_txrx.c
664
idpf_rx_buf_hw_update(bufq, ALIGN_DOWN(bufq->next_to_alloc,
drivers/net/ethernet/intel/idpf/idpf_txrx.h
230
ALIGN_DOWN(IDPF_TX_MAX_DESC_DATA, IDPF_TX_MAX_READ_REQ_SIZE)
drivers/net/ethernet/intel/idpf/idpf_txrx.h
39
#define IDPF_MAX_TXQ_DESC ALIGN_DOWN(IDPF_MAX_DESCS, IDPF_REQ_DESC_MULTIPLE)
drivers/net/ethernet/intel/idpf/idpf_txrx.h
40
#define IDPF_MAX_RXQ_DESC ALIGN_DOWN(IDPF_MAX_DESCS, IDPF_REQ_RXQ_DESC_MULTIPLE)
drivers/net/ethernet/intel/idpf/xsk.c
379
count = ALIGN_DOWN(bufq->pending - 1, rx_thresh);
drivers/net/ethernet/intel/libeth/rx.c
27
len = min3(len, ALIGN_DOWN(max_len ? : U32_MAX, LIBETH_RX_BUF_STRIDE),
drivers/net/ethernet/intel/libeth/rx.c
53
len = ALIGN_DOWN(len, LIBETH_RX_BUF_STRIDE) ? : LIBETH_RX_BUF_STRIDE;
drivers/net/ethernet/intel/libeth/rx.c
54
len = min3(len, ALIGN_DOWN(max_len ? : U32_MAX, LIBETH_RX_BUF_STRIDE),
drivers/net/ethernet/mellanox/mlx5/core/en.h
110
(ALIGN_DOWN(MLX5_SEND_WQE_MAX_SIZE - sizeof(struct mlx5e_umr_wqe), \
drivers/net/ethernet/mellanox/mlx5/core/en.h
116
(ALIGN_DOWN(U16_MAX, 4) * 2) /* Fits into u16 and aligned by WQEBB. */
drivers/net/ethernet/mellanox/mlx5/core/en.h
217
wqebbs = ALIGN_DOWN(wqebbs, 2);
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
116
max_pages_per_wqe = ALIGN_DOWN(max_wqe_size - sizeof(struct mlx5e_umr_wqe),
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h
53
ALIGN_DOWN(MLX5E_KSM_MAX_ENTRIES_PER_WQE(wqe_size), MLX5_UMR_KSM_NUM_ENTRIES_ALIGNMENT)
drivers/net/ethernet/mscc/ocelot_fdma.c
117
dcb->datap = ALIGN_DOWN(dma_addr, 4);
drivers/net/ethernet/mscc/ocelot_fdma.c
118
dcb->datal = ALIGN_DOWN(size, 4);
drivers/net/ethernet/mscc/ocelot_fdma.h
59
#define OCELOT_FDMA_RXB_SIZE ALIGN_DOWN(OCELOT_FDMA_RX_SIZE - OCELOT_FDMA_SKBFRAG_OVR, 4)
drivers/net/mdio/mdio-mux-bcm-iproc.c
230
res->start = ALIGN_DOWN(res->start, SZ_4K);
drivers/net/wireguard/noise.c
508
now.tv_nsec = ALIGN_DOWN(now.tv_nsec,
drivers/net/wireless/realtek/rtw89/efuse_be.c
233
min = ALIGN_DOWN(page_offset, 2);
drivers/net/wireless/realtek/rtw89/efuse_be.c
524
dump_addr = ALIGN_DOWN(EF_FV_OFSET_BE_V1, 4);
drivers/net/wireless/realtek/rtw89/pci.h
1489
#define TXADDR_INFO_LENTHG_V1_MAX ALIGN_DOWN(BIT(11) - 1, 4)
drivers/nvdimm/dimm_devs.c
744
end = ALIGN_DOWN(res->end + 1, align) - 1;
drivers/nvdimm/dimm_devs.c
784
start = ALIGN_DOWN(res->start, align);
drivers/nvdimm/dimm_devs.c
802
return ALIGN_DOWN(nd_mapping->size - busy, align);
drivers/nvdimm/namespace_devs.c
404
valid->end = ALIGN_DOWN(valid->end + 1, align) - 1;
drivers/nvdimm/pfn_devs.c
389
nsoff = ALIGN_DOWN((nd_region->ndr_start
drivers/nvdimm/pfn_devs.c
774
end_trunc = start + size - ALIGN_DOWN(start + size, align);
drivers/nvmem/stm32-bsec-optee-ta.c
162
start = ALIGN_DOWN(offset, 4);
drivers/pci/controller/pci-aardvark.c
1177
reg = ALIGN_DOWN(PCIE_ECAM_OFFSET(bus->number, devfn, where), 4);
drivers/pci/controller/pci-aardvark.c
1258
reg = ALIGN_DOWN(PCIE_ECAM_OFFSET(bus->number, devfn, where), 4);
drivers/pci/controller/pcie-iproc.c
469
val = ALIGN_DOWN(PCIE_ECAM_OFFSET(busno, devfn, where), 4) |
drivers/pci/controller/pcie-iproc.c
962
axi_addr = ALIGN_DOWN(axi_addr, window_size);
drivers/pci/controller/pcie-iproc.c
963
pci_addr = ALIGN_DOWN(pci_addr, window_size);
drivers/pci/controller/pcie-xilinx.c
210
phys_addr_t pa = ALIGN_DOWN(virt_to_phys(pcie), SZ_4K);
drivers/pci/controller/pcie-xilinx.c
481
phys_addr_t pa = ALIGN_DOWN(virt_to_phys(pcie), SZ_4K);
drivers/pci/controller/plda/pcie-microchip-host.c
628
val = ALIGN_DOWN(lower_32_bits(pcie_addr), SZ_4K);
drivers/pci/controller/plda/pcie-plda-host.c
506
val = ALIGN_DOWN(lower_32_bits(axi_addr), SZ_4K);
drivers/pci/endpoint/functions/pci-epf-ntb.c
524
msg_addr = ALIGN_DOWN(msix_tbl[i].msg_addr, align);
drivers/pci/endpoint/pci-epf-core.c
408
epf_bar[bar].phys_addr = ALIGN_DOWN(bar_addr, aligned_mem_size);
drivers/pci/pci.h
149
__pos = ALIGN_DOWN(__pos, 4); \
drivers/pci/setup-bus.c
1900
((align) ? ALIGN_DOWN((addr), (align)) : (addr))
drivers/remoteproc/mtk_scp_ipi.c
93
ptr = (void __iomem *)ALIGN_DOWN((unsigned long)dst, 4);
drivers/rpmsg/qcom_glink_rpm.c
181
aligned_dlen = ALIGN_DOWN(dlen, 4);
drivers/rpmsg/qcom_glink_rpm.c
189
pad = ALIGN(tlen, 8) - ALIGN_DOWN(tlen, 4);
drivers/s390/block/dasd.c
337
max_bytes = ALIGN_DOWN(max_bytes, PAGE_SIZE);
drivers/scsi/sd.c
3275
end = (void *)buffer + ALIGN_DOWN(data.header_length + data.length,
drivers/spi/spi-atcspi200.c
265
op->data.nbytes = ALIGN_DOWN(op->data.nbytes, 4);
drivers/spi/spi-fsl-qspi.c
592
for (i = 0; i < ALIGN_DOWN(op->data.nbytes, 4); i += 4) {
drivers/spi/spi-fsl-qspi.c
618
for (i = 0; i < ALIGN_DOWN(op->data.nbytes, 4); i += 4) {
drivers/spi/spi-fsl-qspi.c
741
op->data.nbytes = ALIGN_DOWN(op->data.nbytes, 8);
drivers/spi/spi-nxp-fspi.c
1084
op->data.nbytes = ALIGN_DOWN(op->data.nbytes, 8);
drivers/spi/spi-nxp-fspi.c
905
for (i = 0; i < ALIGN_DOWN(op->data.nbytes, 8); i += 8) {
drivers/spi/spi-nxp-fspi.c
947
for (i = 0; i < ALIGN_DOWN(len, 8); i += 8) {
drivers/spi/spi-nxp-xspi.c
821
if (i == ALIGN_DOWN(op->data.nbytes, 4)) {
drivers/usb/cdns3/cdns3-gadget.c
1276
if (ALIGN_DOWN(trb->buffer, SZ_4K) !=
drivers/usb/cdns3/cdns3-gadget.c
1277
ALIGN_DOWN(trb->buffer + length, SZ_4K))
drivers/vfio/pci/nvgrace-gpu/main.c
312
addr = ALIGN_DOWN(vmf->address, PAGE_SIZE << order);
drivers/vfio/vfio_main.c
1683
device->iommufd_access, ALIGN_DOWN(iova, PAGE_SIZE),
drivers/vfio/vfio_main.c
1716
ALIGN_DOWN(iova, PAGE_SIZE),
drivers/virtio/virtio_mem.c
2715
addr = ALIGN_DOWN(addr, vm->device_block_size);
drivers/virtio/virtio_mem.c
2773
cur_end = ALIGN_DOWN(cur_start + chunk_size, chunk_size);
drivers/watchdog/ziirave_wdt.c
273
(addr - ALIGN_DOWN(addr, ZIIRAVE_FIRM_PAGE_SIZE));
fs/btrfs/bio.c
741
return ALIGN_DOWN(sector_offset << SECTOR_SHIFT, fs_info->sectorsize);
fs/btrfs/inode.c
10380
isize = ALIGN_DOWN(inode->i_size, fs_info->sectorsize);
fs/btrfs/inode.c
9734
start = ALIGN_DOWN(iocb->ki_pos, fs_info->sectorsize);
fs/btrfs/ioctl.c
4367
start = ALIGN_DOWN(kiocb.ki_pos, fs_info->sectorsize);
fs/btrfs/ioctl.c
4777
start = ALIGN_DOWN(pos, fs_info->sectorsize);
fs/btrfs/reflink.c
722
const u64 wb_start = ALIGN_DOWN(inode->i_size, bs);
fs/btrfs/reflink.c
826
wb_len = ALIGN(inode_in->vfs_inode.i_size, bs) - ALIGN_DOWN(pos_in, bs);
fs/btrfs/reflink.c
851
ret = btrfs_wait_ordered_range(inode_in, ALIGN_DOWN(pos_in, bs), wb_len);
fs/btrfs/reflink.c
854
ret = btrfs_wait_ordered_range(inode_out, ALIGN_DOWN(pos_out, bs), wb_len);
fs/btrfs/relocation.c
4191
u64 new_length = ALIGN_DOWN(dest_length, fs_info->nodesize);
fs/btrfs/relocation.c
5017
u64 new_length = ALIGN_DOWN(length, fs_info->nodesize);
fs/btrfs/send.c
5981
slen = ALIGN_DOWN(src_end - clone_root->offset,
fs/btrfs/space-info.c
507
avail = ALIGN_DOWN(avail, fs_info->zone_size);
fs/btrfs/zoned.c
131
u64 bytenr = ALIGN_DOWN(zone_end, BTRFS_SUPER_INFO_SIZE) -
fs/btrfs/zoned.c
763
fs_info->max_zone_append_size = ALIGN_DOWN(
fs/btrfs/zoned.c
864
wp = ALIGN_DOWN(zone_end << SECTOR_SHIFT,
fs/exfat/balloc.c
261
ent_idx = ALIGN_DOWN(CLUSTER_TO_BITMAP_ENT(clu), BITS_PER_LONG);
fs/f2fs/f2fs.h
3536
return ALIGN_DOWN(addrs, F2FS_I(inode)->i_cluster_size);
fs/f2fs/file.c
431
ALIGN_DOWN(dn->ofs_in_node, F2FS_I(inode)->i_cluster_size));
fs/fuse/file.c
3127
ALIGN_DOWN(pos_out, PAGE_SIZE),
fs/iomap/buffered-io.c
1317
start_byte = ALIGN_DOWN(start_byte, PAGE_SIZE) +
fs/iomap/ioend.c
463
sector_offset = ALIGN_DOWN(sector_offset << SECTOR_SHIFT,
fs/iomap/swapfile.c
148
.len = ALIGN_DOWN(i_size_read(inode), PAGE_SIZE),
fs/iomap/swapfile.c
46
next_ppage = ALIGN_DOWN(iomap->addr + iomap->length, PAGE_SIZE) >>
fs/pstore/ram.c
535
zone_sz = ALIGN_DOWN(zone_sz, 2);
fs/remap_range.c
81
count = ALIGN_DOWN(count, bs);
fs/xfs/xfs_zone_gc.c
771
split_sectors = ALIGN_DOWN(split_sectors << SECTOR_SHIFT,
fs/zonefs/file.c
42
iomap->offset = ALIGN_DOWN(offset, sb->s_blocksize);
fs/zonefs/file.c
92
iomap->offset = ALIGN_DOWN(offset, sb->s_blocksize);
include/linux/huge_mm.h
233
haddr = ALIGN_DOWN(addr, hpage_size);
include/linux/iova.h
70
return ALIGN_DOWN(size, iovad->granule);
include/linux/mm.h
230
#define PAGE_ALIGN_DOWN(addr) ALIGN_DOWN(addr, PAGE_SIZE)
include/linux/pageblock-flags.h
78
#define pageblock_start_pfn(pfn) ALIGN_DOWN((pfn), pageblock_nr_pages)
include/net/ip_vs.h
427
ALIGN_DOWN(2 * 1000 * 10 / IPVS_EST_LOAD_DIVISOR / IPVS_EST_NTICKS, 8)
include/net/libeth/rx.h
31
ALIGN_DOWN(SKB_MAX_ORDER(hr, LIBETH_RX_PAGE_ORDER), \
include/net/libeth/xdp.h
544
batched = ALIGN_DOWN(this, LIBETH_XDP_TX_BATCH);
include/net/xdp_sock_drv.h
72
return ALIGN_DOWN(frame_size, 128);
include/rdma/ib_umem.h
70
ALIGN_DOWN(umem->iova, pgsz))) /
include/vdso/align.h
12
#define PTR_ALIGN_DOWN(p, a) ((typeof(p))ALIGN_DOWN((unsigned long)(p), (a)))
include/video/udlfb.h
101
#define DL_ALIGN_DOWN(x, a) ALIGN_DOWN(x, a)
init/initramfs.c
656
unsigned long aligned_start = ALIGN_DOWN(start, PAGE_SIZE);
kernel/kexec_file.c
484
temp_start = ALIGN_DOWN(temp_start, kbuf->buf_align);
kernel/trace/trace_events_user.c
189
*addr = ALIGN_DOWN(*addr, sizeof(long));
lib/alloc_tag.c
418
unsigned long phys_end = ALIGN_DOWN(module_tags.start_addr, PAGE_SIZE) +
lib/scatterlist.c
471
max_segment = ALIGN_DOWN(max_segment, PAGE_SIZE);
lib/test_hmm.c
1651
args.start = ALIGN_DOWN(vmf->address, (PAGE_SIZE << order));
mm/cma.c
587
end = ALIGN_DOWN(end, align);
mm/cma.c
592
size = ALIGN_DOWN(size, (PAGE_SIZE << order_per_bit));
mm/damon/core.c
1749
sz_to_skip = ALIGN_DOWN(quota->charge_addr_from -
mm/damon/core.c
1797
start = ALIGN_DOWN(filter->addr_range.start, min_region_sz);
mm/damon/core.c
1798
end = ALIGN_DOWN(filter->addr_range.end, min_region_sz);
mm/damon/core.c
1968
sz = ALIGN_DOWN(quota->esz - quota->charged_sz,
mm/damon/core.c
250
ALIGN_DOWN(range->start,
mm/damon/core.c
2525
sz_sub = ALIGN_DOWN(damon_rand(1, 10) *
mm/damon/core.c
258
first->ar.start = ALIGN_DOWN(range->start,
mm/damon/paddr.c
94
if (ALIGN_DOWN(last_addr, last_folio_sz) ==
mm/damon/paddr.c
95
ALIGN_DOWN(sampling_addr, last_folio_sz)) {
mm/damon/vaddr.c
553
if (same_target && (ALIGN_DOWN(last_addr, last_folio_sz) ==
mm/damon/vaddr.c
554
ALIGN_DOWN(r->sampling_addr, last_folio_sz))) {
mm/damon/vaddr.c
81
sz_piece = ALIGN_DOWN(sz_orig / nr_pieces, DAMON_MIN_REGION_SZ);
mm/huge_memory.c
3311
range_in_vma(vma, ALIGN_DOWN(address, HPAGE_PMD_SIZE),
mm/hugetlb.c
6851
v_end = ALIGN_DOWN(vma->vm_end, PUD_SIZE);
mm/hugetlb.c
6863
*start = ALIGN_DOWN(*start, PUD_SIZE);
mm/hugetlb.c
7324
ALIGN_DOWN(vma->vm_end, PUD_SIZE),
mm/internal.h
142
.val = ALIGN_DOWN(entry.val, folio_nr_pages(folio)),
mm/kasan/shadow.c
576
region_end = ALIGN_DOWN(end, KASAN_MEMORY_PER_SHADOW_PAGE);
mm/kasan/shadow.c
584
free_region_end = ALIGN_DOWN(free_region_end, KASAN_MEMORY_PER_SHADOW_PAGE);
mm/kfence/core.c
249
return !KFENCE_WARN_ON(!kfence_protect_page(ALIGN_DOWN(addr, PAGE_SIZE), true));
mm/kfence/core.c
254
return !KFENCE_WARN_ON(!kfence_protect_page(ALIGN_DOWN(addr, PAGE_SIZE), false));
mm/kfence/core.c
274
if (KFENCE_WARN_ON(ALIGN_DOWN(meta->addr, PAGE_SIZE) != pageaddr))
mm/kfence/core.c
356
const unsigned long pageaddr = ALIGN_DOWN(meta->addr, PAGE_SIZE);
mm/kfence/core.c
366
addr = ALIGN_DOWN(meta->addr + meta->size, sizeof(u64));
mm/kfence/core.c
374
const unsigned long pageaddr = ALIGN_DOWN(meta->addr, PAGE_SIZE);
mm/kfence/core.c
476
meta->addr = ALIGN_DOWN(meta->addr, cache->align);
mm/kfence/core.c
537
kcsan_begin_scoped_access((void *)ALIGN_DOWN((unsigned long)addr, PAGE_SIZE), PAGE_SIZE,
mm/kfence/core.c
546
memzero_explicit((void *)ALIGN_DOWN(meta->unprotected_page, PAGE_SIZE), PAGE_SIZE);
mm/kmsan/core.c
341
if (ALIGN_DOWN(cur_addr + size - 1, PAGE_SIZE) ==
mm/kmsan/core.c
342
ALIGN_DOWN(cur_addr, PAGE_SIZE))
mm/kmsan/core.c
95
(u32 *)ALIGN_DOWN((u64)shadow_dst, KMSAN_ORIGIN_SIZE);
mm/kmsan/init.c
39
nstart = ALIGN_DOWN(nstart, PAGE_SIZE);
mm/kmsan/shadow.c
131
addr = ALIGN_DOWN(addr, KMSAN_ORIGIN_SIZE);
mm/madvise.c
897
range->end = ALIGN_DOWN(range->end, huge_page_size(hstate_vma(vma)));
mm/memory.c
4555
addr = ALIGN_DOWN(vmf->address, nr_pages * PAGE_SIZE);
mm/memory.c
4653
addr = ALIGN_DOWN(vmf->address, PAGE_SIZE << order);
mm/memory.c
4664
addr = ALIGN_DOWN(vmf->address, PAGE_SIZE << order);
mm/memory.c
5168
addr = ALIGN_DOWN(vmf->address, PAGE_SIZE << order);
mm/memory.c
5182
addr = ALIGN_DOWN(vmf->address, PAGE_SIZE << order);
mm/memory.c
5273
addr = ALIGN_DOWN(vmf->address, nr_pages * PAGE_SIZE);
mm/memory.c
5743
from_pte = max(ALIGN_DOWN(pte_off, nr_pages),
mm/memory.c
6016
unsigned long pt_start = ALIGN_DOWN(addr, PMD_SIZE);
mm/memory.c
7337
const unsigned long base_addr = ALIGN_DOWN(addr_hint, folio_size(folio));
mm/memory.c
7373
unsigned long addr = ALIGN_DOWN(addr_hint, folio_size(dst));
mm/memory_hotplug.c
1116
online_mem_sections(pfn, ALIGN_DOWN(end_pfn, PAGES_PER_SECTION));
mm/memory_hotplug.c
1131
offline_mem_sections(pfn, ALIGN_DOWN(end_pfn, PAGES_PER_SECTION));
mm/memory_hotplug.c
992
pfn = ALIGN_DOWN(pfn, group->d.unit_pages);
mm/numa_emulation.c
270
size = ALIGN_DOWN(size, FAKE_NODE_MIN_SIZE);
mm/page_isolation.c
348
start_pfn = max(ALIGN_DOWN(isolate_pageblock, MAX_ORDER_NR_PAGES),
mm/shuffle.c
113
ALIGN_DOWN(get_random_long() % z->spanned_pages,
mm/slub.c
7735
s->offset = ALIGN_DOWN(s->object_size / 2, sizeof(void *));
mm/swapfile.c
919
unsigned long start = ALIGN_DOWN(offset, SWAPFILE_CLUSTER);
mm/vmalloc.c
381
addr = ALIGN_DOWN(addr, size);
net/ipv4/tcp_output.c
922
align_size = ALIGN_DOWN(size, 4);
net/xdp/xsk_buff_pool.c
223
frame_size = ALIGN_DOWN(frame_size, 128);
sound/core/memalloc.c
392
start = ALIGN_DOWN(ofs, PAGE_SIZE);
sound/core/memalloc.c
650
start = ALIGN_DOWN(ofs, PAGE_SIZE);
sound/soc/sof/debug.c
45
pos = ALIGN_DOWN(pos, 4);
tools/include/linux/mm.h
16
#define PAGE_ALIGN_DOWN(addr) ALIGN_DOWN(addr, PAGE_SIZE)
tools/testing/memblock/linux/mmzone.h
27
#define pageblock_start_pfn(pfn) ALIGN_DOWN((pfn), pageblock_nr_pages)
tools/testing/selftests/mm/pkey-helpers.h
178
((typeof(p))ALIGN_DOWN((unsigned long)(p), ptr_align_to))
tools/testing/vma/linux/mmzone.h
27
#define pageblock_start_pfn(pfn) ALIGN_DOWN((pfn), pageblock_nr_pages)
virt/kvm/guest_memfd.c
105
index = ALIGN_DOWN(index, folio_nr_pages(folio));