btodb
dp->e2di_nblock = htole32(btodb(sblock.e2fs_bsize));
idesc->id_entryno *= btodb(sblock.e2fs_bsize);
DIP_SET(dp, di_blocks, DIP(dp, di_blocks) + btodb(sblock.fs_bsize));
DIP_SET(dp, di_blocks, DIP(dp, di_blocks) - btodb(sblock.fs_bsize));
DIP_SET(dp, di_blocks, btodb(sblock.fs_fsize));
idesc->id_entryno *= btodb(sblock.fs_fsize);
node.dp1.di_blocks = btodb(fragroundup(&sblock, node.dp1.di_size));
node.dp1.di_blocks = btodb(fragroundup(&sblock, node.dp1.di_size));
node.dp2.di_blocks = btodb(fragroundup(&sblock, node.dp2.di_size));
node.dp2.di_blocks = btodb(fragroundup(&sblock, node.dp2.di_size));
blkno += btodb(n); /* XXX? */
blkno += btodb(n); /* XXX? */
blkno += btodb(PAGE_SIZE);
blkno += btodb(n); /* XXX? */
btodb(LIF_VOLSTART)));
blkno += btodb(n);
btodb(foff), count, p, &bsize)))
blkno = dumplo + btodb(maddr) + 1;
for (j = npg; j--; maddr += NBPG, blkno += btodb(NBPG)) {
if (cpu_dumpsize() > btodb(sizeof buf)) {
*blknop += btodb(sizeof buf);
return (btodb(roundup(size, dbtob(1))));
dumplo = nblks - btodb(ptoa(physmem));
if (dumplo < btodb(PAGE_SIZE))
dumplo = btodb(PAGE_SIZE);
blkno += btodb(PAGE_SIZE);
blkno = dumplo + btodb(maddr) + 1;
j--, maddr += PAGE_SIZE, blkno+= btodb(PAGE_SIZE))
dumplo = nblks - btodb(ptoa(physmem));
if (dumplo < btodb(PAGE_SIZE))
dumplo = btodb(PAGE_SIZE);
if (dumplo < btodb(1))
dumplo = btodb(1);
blkno += btodb(NBPG);
blkno += btodb(n);
return btodb(sz + DEV_BSIZE - 1);
(iso_blknum + sess) * btodb(iso_bsize),
bp->b_blkno = iso_blknum * btodb(ISO_DEFAULT_BLOCK_SIZE);
bp->b_blkno = sector * btodb(bsize);
bp->b_blkno = sector * btodb(bsize);
if ((error = bread(devvp, sector * btodb(bsize), bsize, &bp)) != 0)
if ((error = bread(devvp, sector * btodb(bsize), bsize,
bp->b_blkno = btodb(uio->uio_offset);
bscale = btodb(bsize);
bn = btodb(uio->uio_offset) & ~(bscale - 1);
bscale = btodb(bsize);
bn = btodb(uio->uio_offset) & ~(bscale - 1);
*blkctr += btodb(PAGE_SIZE);
const uint64_t blks = btodb(size);
blkctr += btodb(PAGE_SIZE);
if (offset + blks > btodb(HIBERNATE_CHUNK_TABLE_SIZE)) {
blkctr += btodb(used);
chunktable_size = btodb(HIBERNATE_CHUNK_TABLE_SIZE);
i += MAXPHYS, blkctr += btodb(MAXPHYS)) {
blkctr += btodb(read_size);
btodb(HIBERNATE_CHUNK_TABLE_SIZE);
btodb(HIBERNATE_CHUNK_TABLE_SIZE));
if ((err = hibernate_write(hib, btodb(i),
inc = btodb(size);
btodb(f->f_offset), bcount, dest, &resid);
btodb(f->f_offset), bcount, dest, &resid);
*ap->a_bnp = ap->a_bn * btodb(vp->v_mount->mnt_stat.f_iosize);
#ifndef btodb
ip->i_e2fs_nblock += btodb(fs->e2fs_bsize);
ip->i_e2fs_nblock -= btodb(deallocated);
nblocks = btodb(fs->e2fs_bsize);
blocksreleased += btodb(fs->e2fs_bsize);
nblocks = btodb(fs->e2fs_bsize);
if ((error = ufs_quota_alloc_blocks(ip, btodb(size), cred)) != 0)
DIP_ADD(ip, blocks, btodb(size));
(void) ufs_quota_free_blocks(ip, btodb(size), cred);
if ((error = ufs_quota_alloc_blocks(ip, btodb(nsize - osize), cred))
quota_updated = btodb(nsize - osize);
DIP_ADD(ip, blocks, btodb(nsize - osize));
DIP_ADD(ip, blocks, btodb(nsize - osize));
(void)ufs_quota_free_blocks(ip, btodb(deallocated), cred);
ip->i_ffs1_blocks -= btodb(deallocated);
(void) ufs_quota_free_blocks(ip, btodb(deallocated), cred);
ip->i_ffs2_blocks -= btodb(deallocated);
nblocks = btodb(fs->fs_bsize);
blocksreleased += btodb(bsize);
blocksreleased += btodb(oldspace - newspace);
nblocks = btodb(fs->fs_bsize);
bn = btodb((u_int64_t)pageno << PAGE_SHIFT); /* convert to diskblock */
nbp->vb_buf.b_blkno = nbn + btodb(off);
startblk = btodb((u_int64_t)startslot << PAGE_SHIFT);
block += btodb(PAGE_SIZE);
block += btodb(PAGE_SIZE);
btodb((u_int64_t)sdp->swd_npginuse <<
nblocks = (int)btodb(va.va_size);
dqblk.dqb_curblocks = btodb((u_quad_t)
dqblk.dqb_bsoftlimit = btodb((u_quad_t)
dqblk.dqb_bhardlimit = btodb((u_quad_t)