BTOBB
.numblks = BTOBB(mp->m_sb.sb_blocksize),
.numblks = BTOBB(mp->m_sb.sb_blocksize),
.numblks = BTOBB(mp->m_sb.sb_blocksize),
.numblks = BTOBB(mp->m_sb.sb_blocksize),
.numblks = BTOBB(mp->m_sb.sb_blocksize),
.numblks = BTOBB(mp->m_sb.sb_blocksize),
bno += BTOBB(blksize);
bno += BTOBB(blksize);
bno += BTOBB(blksize);
bno += BTOBB(blksize);
BTOBB(1 << (xfs_has_logv2(log->l_mp) ? \
BTOBB(XLOG_MAX_ICLOGS << (xfs_has_logv2(log->l_mp) ? \
lsunit = BTOBB(mp->m_sb.sb_logsunit);
min_logblks = roundup_64(BTOBB(max_logres), lsunit) +
min_logblks = BTOBB(max_logres) + 2 * BBSIZE;
error = xfs_buf_read(mp->m_ddev_targp, d, BTOBB(byte_cnt), 0,
BTOBB(byte_cnt), 0, &bp);
start = BTOBB(range.start);
for (i = 0; i < BTOBB(iclog->ic_offset + roundoff); i++) {
if (bno + BTOBB(count) > log->l_logBBsize) {
XFS_STATS_ADD(log->l_mp, xs_log_blocks, BTOBB(count));
if (bno + BTOBB(count) > log->l_logBBsize)
log->l_curr_block += BTOBB(eventual_size)+BTOBB(log->l_iclog_hsize);
uint32_t sunit_bb = BTOBB(log->l_iclog_roundoff);
if (blocks < BTOBB(iclog->ic_offset) +
BTOBB(log->l_iclog_hsize)) {
if (blocks < BTOBB(iclog->ic_offset) + 1) {
rhead_blk + hblks + BTOBB(be32_to_cpu(rhead->h_len)));
for (i = 0; i < BTOBB(be32_to_cpu(rhead->h_len)); i++) {
dbp = xlog_alloc_buffer(log, BTOBB(h_size));
bblks = (int)BTOBB(be32_to_cpu(rhead->h_len));
bblks = (int)BTOBB(be32_to_cpu(rhead->h_len));
BTOBB(be32_to_cpu(head->h_len)) + xhdrs)
if (tail_distance > BTOBB(XLOG_MAX_ICLOGS * hsize))
BTOBB(sector_size), &bp, buf_ops);
*daddr = BTOBB(offset);
*bblen = BTOBB(len);
#define BLKDEV_BB BTOBB(BLKDEV_IOSIZE)