rounddown2
bit = rounddown2(bit, BITS_PER_MAP);
if (rounddown2((long)src, sizeof(struct in_addr))) {
if (rounddown2((long)src, sizeof(struct in6_addr) / 2)) {
#define aarch64_trunc_page(x, size) rounddown2((kvaddr_t)(x), size)
return (rounddown2(dump_avail_n(kd, i), page_size) +
return (rounddown2(x, page_size));
offset = (uintptr_t)pagepool_start - rounddown2(
(void) lseek(dirp->dd_fd, base + rounddown2(loc, DIRBLKSIZ), SEEK_SET);
addr = rounddown2(addr, cl_size);
addr = rounddown2(addr, M(2));
sz = MIN(sz, rounddown2(mem_avail * 45 / 100, SEGALIGN));
rounddown2((commit_limit - committed_as) * 95 / 100, SEGALIGN));
loaded_segments[nkexec_segments].mem = (void *)rounddown2(dest,SEGALIGN);
alignsector = rounddown2(lastsector * DEV_BSIZE, DEV_GELIBOOT_BSIZE);
alnstart = rounddown2(reqstart, (int)gdesc->gdev->md.md_sectorsize);
fp->f_seekp = rounddown2(fp->f_seekp, bsize) + bsize;
eblock = rounddown2((uint64_t)si->mr[i].start +
subldr = rounddown2((uintptr_t)_start, KERN_ALIGN);
if ((amd_feature & AMDID_PAGE1GB) != 0 && rounddown2(spa + len,
if (error != 0 && rounddown2(spa + len, NBPDR) >= roundup2(spa,
sva = rounddown2(sva, PAGE_SIZE);
sva = rounddown2(sva, NBPML5);
sva = rounddown2(sva, NBPML4);
sva = rounddown2(sva, NBPDP);
sva = rounddown2(sva, NBPDP);
sva = rounddown2(sva, NBPDR);
sva = rounddown2(sva, NBPDR);
kernphys - rounddown2(kernphys, NBPDP), NBPDP);
for (i = 0, pax = rounddown2(kernphys, NBPDP);
j = rounddown2(kernphys, NBPDP) >> PDPSHIFT;
destp = rounddown2(destp, sizeof(uint32_t));
destp = rounddown2(destp, sizeof(uint32_t));
destp = rounddown2(destp, sizeof(uint32_t));
addr = rounddown2(addr, 4);
phys = rounddown2(vtophys(rman_get_virtual(sc->sc_its_res)) +
pa = rounddown2(end_pa - L2_SIZE, L2_SIZE);
KASSERT(rounddown2(sva, L2_SIZE) + L2_SIZE == roundup2(eva, L2_SIZE),
rounddown2(pa, L2_SIZE)) >> L2_SHIFT;
pa = rounddown2(pa, L2_SIZE);
(roundup2(va + size, L2_SIZE) - rounddown2(va, L2_SIZE)) >> L2_SHIFT;
va_trunc = rounddown2(va, L2_SIZE) + offset;
vmm_base = rounddown2(vmm_base, L2_SIZE);
destp = rounddown2(destp, sizeof(uint32_t));
destp = rounddown2(destp, sizeof(uint32_t));
destp = rounddown2(destp, sizeof(uint32_t));
destp = rounddown2(destp, sizeof(uint32_t));
destp = rounddown2(destp, sizeof(void *));
destp = rounddown2(destp, sizeof(void *));
destp = rounddown2(destp, sizeof(void *));
#define ALIGN_DOWN(x, y) rounddown2(x, y)
todo = rounddown2(resid, CHACHA20_NATIVE_BLOCK_LEN);
todo = rounddown2(resid, CHACHA20_NATIVE_BLOCK_LEN);
todo = rounddown2(resid, CHACHA20_NATIVE_BLOCK_LEN);
todo = rounddown2(resid, CHACHA20_NATIVE_BLOCK_LEN);
if (rounddown2(buttons, 1 << sc->hw.buttons)
sc->pg_offset = rounddown2(apbase & (64 * 1024 * 1024 - 1),
mask = (0xfULL << 32) | rounddown2(0xfffff000, size) | 0x800;
start = rounddown2(addr, 64);
offset = rounddown2(offset, AES_BLOCK_LEN);
tlsp->ghash_offset = rounddown2(offset + plen,
*offset = rounddown2(*offset, AES_BLOCK_LEN);
rx_len = rounddown2(rx_len, 512);
tx_len = rounddown2(tx_len, 512);
end = rounddown2(start + MAX_DDP_BUFFER_SIZE, CACHE_LINE_SIZE);
#define ALIGN_DOWN(x, y) rounddown2((x), (y))
#define round_down(x, y) rounddown2((x), (y))
sc->mem[win].cardaddr = rounddown2(cardaddr, EXCA_MEM_PAGESIZE);
uint32_t sector_base = rounddown2(offset, sc->erasesize);
pArray->VDeviceCapacity = rounddown2(capacity, pArray->u.array.bStripeWitch) *
"%s: available size is %u\n", __func__, rounddown2(ret, 8));
return (rounddown2(ret, 8));
sndbuf = rounddown2(sndbuf, PAGE_SIZE);
rcvbuf = rounddown2(rcvbuf, PAGE_SIZE);
IWM_WRITE(sc, wreg, rounddown2(hw, 8));
rounddown2(val, mpsmin));
pn = rounddown2(pe - 1, SYM_CONF_DMA_BOUNDARY);
CSR_WRITE_4(sc, TI_WINBASE, rounddown2(segptr, TI_WINLEN));
CSR_WRITE_4(sc, TI_WINBASE, rounddown2(segptr, TI_WINLEN));
CSR_WRITE_4(sc, TI_WINBASE, rounddown2(segptr, TI_WINLEN));
CSR_WRITE_4(sc, TI_WINBASE, rounddown2(segptr, TI_WINLEN));
CSR_WRITE_4(sc, TI_WINBASE, rounddown2(segptr, TI_WINLEN));
pg->physaddr = rounddown2(segs->ds_addr, USB_PAGE_SIZE);
pg->physaddr = rounddown2(segs->ds_addr + off, USB_PAGE_SIZE);
dp->i_diroff = rounddown2(i_offset, DIRBLKSIZ);
rounddown2(pp->mediasize - table->d_bbase * pp->sectorsize,
v = rounddown2(pp->sectorsize + offsetof(struct disklabel64, d_magic),
sva = rounddown2(sva, NBPDR);
#define NEW_MAX_LD(num) rounddown2(num + LD_PER_PAGE, LD_PER_PAGE)
#define aligned(a, t) (rounddown2((u_long)(a), sizeof(t)) == (u_long)(a))
stack_top -= rounddown2(stack_off & PAGE_MASK, sizeof(void *));
destp = rounddown2(destp, sizeof(void *));
destp = rounddown2(destp, sizeof(void *));
destp = rounddown2(destp, sizeof(void *));
destp = rounddown2(destp, sizeof(void *));
destp = rounddown2(destp, sizeof(void *));
attach_va = rounddown2((vm_offset_t)shmaddr, SHMLBA);
*first = rounddown2(pidx, pagesizes[psind] / PAGE_SIZE);
*before = pindex - rounddown2(pindex, pagesizes[psind] /
pgsin_b = OFF_TO_IDX(lb - rounddown2(lb, bo_bs));
off1 = rounddown2(off1, KASAN_SHADOW_SCALE);
off2 = rounddown2(off2, KASAN_SHADOW_SCALE);
(m)->m_data += rounddown2(MCLBYTES - (len), sizeof(long)); \
todo = rounddown2(MIN(resid, MIN(inlen, outlen)),
todo = rounddown2(MIN(resid, MIN(inlen, outlen)),
todo = rounddown2(MIN(resid, MIN(inlen, outlen)), blksz);
inlen = rounddown2(crp->crp_aad_length, blksz);
inlen = rounddown2(MIN(inlen, resid), blksz);
todo = rounddown2(MIN(resid, MIN(inlen, outlen)),
todo = rounddown2(MIN(resid, inlen), blksz);
todo = rounddown2(MIN(resid, MIN(inlen, outlen)),
todo = rounddown2(MIN(resid, MIN(inlen, outlen)),
todo = rounddown2(MIN(resid, MIN(inlen, outlen)),
hash &= rounddown2(0xfffff, VSID_NBPW);
hash &= rounddown2(VSID_HASHMASK, VSID_NBPW);
ppa = rounddown2(pa, gran);
ppa = rounddown2(pa, gran);
while (rounddown2(va, size) + size < va + sz)
va = rounddown2(va, size);
pa = rounddown2(pa, size);
for (line = rounddown2(paddr, SIFIVE_CCACHE_LINE_SIZE);
#define trunc_page(x) rounddown2(x, PAGE_SIZE)
firstino = rounddown2(inodedep->id_ino, INOPB(fs));
blkoff = rounddown2(offset, DIRBLKSIZ); /* offset of start of block */
rounddown2(I_OFFSET(dp), DIRBLKSIZ));
rounddown2(I_OFFSET(dp), DIRBLKSIZ));
dp->i_diroff = rounddown2(i_offset, DIRBLKSIZ);
end = rounddown2(vaddr, size);
while (rounddown2(vaddr, pagesizes[psind]) < fs->entry->start ||
m_super += rounddown2(m - m_super,
vaddr = rounddown2(vaddr, pagesizes[psind]);
if ((frag = rounddown2(base, DEV_BSIZE)) != base &&
if ((frag = rounddown2(endoff, DEV_BSIZE)) != endoff &&
if ((frag = rounddown2(base, DEV_BSIZE)) != base &&
if ((frag = rounddown2(endoff, DEV_BSIZE)) != endoff &&
index = rounddown2(m - rv->pages, npages);
return (rv->pages + rounddown2(m - rv->pages,
index = rounddown2(index, 128) / 64;
rv->pages[rounddown2(index, VM_SUBLEVEL_0_NPAGES)].psind = 0;
rv->pages[rounddown2(index, VM_SUBLEVEL_0_NPAGES)].psind = 1;
end = rounddown2(end, DEV_BSIZE);
rounddown2(mtrr_physmask, mrd->mr_len);
#ifndef rounddown2
address = rounddown2(end - length, alignment);
table_offset = rounddown2(pi->pi_msix.table_offset, 4096);
table_offset = rounddown2(pi->pi_msix.table_offset, 4096);
fsopts->size += rounddown2(file_len, ffs_opts->bsize);
rounddown2(file_len, ffs_opts->bsize), ffs_opts->fsize);
vdevsize = rounddown2(fsopts->maxsize, 1 << zfs->ashift);
vdevsize1 = rounddown2(asize, mssize) + VDEV_LABEL_SPACE;
nbits = rounddown2(zfs->asize, zfs->mssize) >> zfs->ashift;