Symbol: BBTOB
fs/xfs/libxfs/xfs_attr_remote.c
147
len = BBTOB(bp->b_length);
fs/xfs/libxfs/xfs_attr_remote.c
212
len = BBTOB(bp->b_length);
fs/xfs/libxfs/xfs_attr_remote.c
302
unsigned int len = BBTOB(bp->b_length);
fs/xfs/libxfs/xfs_attr_remote.c
351
unsigned int len = BBTOB(bp->b_length);
fs/xfs/libxfs/xfs_dir2_data.c
306
bp->b_addr, BBTOB(bp->b_length), __FILE__, __LINE__,
fs/xfs/libxfs/xfs_dir2_leaf.c
131
bp->b_addr, BBTOB(bp->b_length), __FILE__, __LINE__,
fs/xfs/libxfs/xfs_dir2_node.c
91
bp->b_addr, BBTOB(bp->b_length), __FILE__, __LINE__,
fs/xfs/libxfs/xfs_dquot_buf.c
28
return BBTOB(nbblks) / sizeof(struct xfs_dqblk);
fs/xfs/libxfs/xfs_ialloc.c
373
xfs_buf_zero(fbuf, 0, BBTOB(fbuf->b_length));
fs/xfs/libxfs/xfs_rtgroup.c
625
if (memchr_inv(rsb + 1, 0, BBTOB(bp->b_length) - sizeof(*rsb)))
fs/xfs/libxfs/xfs_sb.c
1438
xfs_buf_zero(bp, 0, BBTOB(bp->b_length));
fs/xfs/libxfs/xfs_sb.c
1739
BBTOB(mp->m_dalign), BBTOB(mp->m_swidth));
fs/xfs/libxfs/xfs_symlink_remote.c
193
ASSERT(BBTOB(bp->b_length) >=
fs/xfs/scrub/agheader.c
421
if (memchr_inv((char *)sb + sblen, 0, BBTOB(bp->b_length) - sblen))
fs/xfs/scrub/agheader_repair.c
1697
xfs_trans_log_buf(sc->tp, agi_bp, 0, BBTOB(agi_bp->b_length) - 1);
fs/xfs/scrub/agheader_repair.c
208
memset(agf, 0, BBTOB(agf_bp->b_length));
fs/xfs/scrub/agheader_repair.c
325
xfs_trans_log_buf(sc->tp, agf_bp, 0, BBTOB(agf_bp->b_length) - 1);
fs/xfs/scrub/agheader_repair.c
63
xfs_buf_zero(bp, 0, BBTOB(bp->b_length));
fs/xfs/scrub/agheader_repair.c
688
memset(agfl, 0xFF, BBTOB(agfl_bp->b_length));
fs/xfs/scrub/agheader_repair.c
707
xfs_trans_log_buf(sc->tp, agfl_bp, 0, BBTOB(agfl_bp->b_length) - 1);
fs/xfs/scrub/agheader_repair.c
80
xfs_trans_log_buf(sc->tp, bp, 0, BBTOB(bp->b_length) - 1);
fs/xfs/scrub/agheader_repair.c
901
memset(agi, 0, BBTOB(agi_bp->b_length));
fs/xfs/scrub/dir_repair.c
481
end = min_t(unsigned int, BBTOB(bp->b_length),
fs/xfs/scrub/ialloc.c
440
if (imap.im_boffset >= BBTOB(cluster_bp->b_length)) {
fs/xfs/scrub/ialloc_repair.c
159
if (offset >= BBTOB(cluster_bp->b_length))
fs/xfs/scrub/quota_repair.c
384
xfs_trans_log_buf(sc->tp, bp, 0, BBTOB(bp->b_length) - 1);
fs/xfs/scrub/reap.c
1300
chunks = DIV_ROUND_UP(BBTOB(bp->b_maps[i].bm_len),
fs/xfs/xfs_buf.c
106
unsigned int size = BBTOB(bp->b_length);
fs/xfs/xfs_buf.c
1172
roundup(BBTOB(bp->b_length), PAGE_SIZE));
fs/xfs/xfs_buf.c
1333
unsigned int len = BBTOB(bp->b_length);
fs/xfs/xfs_buf.c
179
size_t size = BBTOB(bp->b_length);
fs/xfs/xfs_buf.c
391
ASSERT(!(BBTOB(map->bm_len) < btp->bt_meta_sectorsize));
fs/xfs/xfs_buf.c
392
ASSERT(!(BBTOB(map->bm_bn) & (xfs_off_t)btp->bt_meta_sectormask));
fs/xfs/xfs_buf.h
358
return xfs_verify_cksum(bp->b_addr, BBTOB(bp->b_length),
fs/xfs/xfs_buf.h
365
xfs_update_cksum(bp->b_addr, BBTOB(bp->b_length),
fs/xfs/xfs_buf_item.c
1021
end = start + BBTOB(bp->b_maps[i].bm_len) - 1;
fs/xfs/xfs_buf_item.c
1025
start += BBTOB(bp->b_maps[i].bm_len);
fs/xfs/xfs_buf_item.c
1042
start += BBTOB(bp->b_maps[i].bm_len);
fs/xfs/xfs_buf_item.c
252
offset += BBTOB(bp->b_maps[i].bm_len);
fs/xfs/xfs_buf_item.c
398
offset += BBTOB(bp->b_maps[i].bm_len);
fs/xfs/xfs_buf_item.c
822
bp->b_addr, BBTOB(bp->b_length), fa);
fs/xfs/xfs_buf_item.c
889
chunks = DIV_ROUND_UP(BBTOB(bp->b_maps[i].bm_len),
fs/xfs/xfs_buf_item.c
899
BBTOB(bp->b_maps[i].bm_len));
fs/xfs/xfs_buf_item_recover.c
1137
(BBTOB(bp->b_length) != M_IGEO(log->l_mp)->inode_cluster_size)) {
fs/xfs/xfs_buf_item_recover.c
492
ASSERT(BBTOB(bp->b_length) >=
fs/xfs/xfs_buf_item_recover.c
632
inodes_per_buf = BBTOB(bp->b_length) >> mp->m_sb.sb_inodelog;
fs/xfs/xfs_buf_item_recover.c
674
ASSERT((reg_buf_offset + reg_buf_bytes) <= BBTOB(bp->b_length));
fs/xfs/xfs_buf_mem.c
137
loff_t pos = BBTOB(xfs_buf_daddr(bp));
fs/xfs/xfs_buf_mem.c
144
if (BBTOB(bp->b_length) != XMBUF_BLOCKSIZE)
fs/xfs/xfs_buf_mem.c
195
pos = BBTOB(xfs_buf_daddr(bp));
fs/xfs/xfs_buf_mem.c
196
shmem_truncate_range(inode, pos, pos + BBTOB(bp->b_length) - 1);
fs/xfs/xfs_dquot.c
267
memset(d, 0, BBTOB(q->qi_dqchunklen));
fs/xfs/xfs_dquot.c
303
xfs_trans_log_buf(tp, bp, 0, BBTOB(q->qi_dqchunklen) - 1);
fs/xfs/xfs_fsmap.c
40
dest->fmr_physical = BBTOB(src->fmr_physical);
fs/xfs/xfs_fsmap.c
42
dest->fmr_offset = BBTOB(src->fmr_offset);
fs/xfs/xfs_fsmap.c
43
dest->fmr_length = BBTOB(src->fmr_length);
fs/xfs/xfs_iomap.c
126
iomap->addr = BBTOB(daddr);
fs/xfs/xfs_log.c
1345
log->l_logsize = BBTOB(num_bblks);
fs/xfs/xfs_log.c
1650
unsigned int split_offset = BBTOB(log->l_logBBsize - bno);
fs/xfs/xfs_log_priv.h
393
min_t(int, (log)->l_logsize >> 3, BBTOB(XLOG_TOTAL_REC_SHIFT(log)) << 4)
fs/xfs/xfs_log_priv.h
624
return BBTOB(hi_block - lo_block);
fs/xfs/xfs_log_priv.h
626
return (uint64_t)log->l_logsize - BBTOB(lo_block - hi_block);
fs/xfs/xfs_log_recover.c
104
return kvzalloc(BBTOB(nbblks), GFP_KERNEL | __GFP_RETRY_MAYFAIL);
fs/xfs/xfs_log_recover.c
116
return BBTOB(blk_no & ((xfs_daddr_t)log->l_sectBBsize - 1));
fs/xfs/xfs_log_recover.c
141
BBTOB(nbblks), data, op);
fs/xfs/xfs_log_recover.c
1562
buffer + BBTOB(ealign - start_block));
fs/xfs/xfs_log_recover.c
3144
offset + BBTOB(split_hblks));
fs/xfs/xfs_log_recover.c
3205
offset + BBTOB(split_bblks));
fs/xfs/xfs_mount.c
314
if ((BBTOB(mp->m_dalign) & mp->m_blockmask) ||
fs/xfs/xfs_mount.c
315
(BBTOB(mp->m_swidth) & mp->m_blockmask)) {
fs/xfs/xfs_notify_failure.c
173
u64 dev_len = BBTOB(btp->bt_nr_sectors);
fs/xfs/xfs_trans_buf.c
556
ASSERT(first <= last && last < BBTOB(bp->b_length));
fs/xfs/xfs_verify_media.c
186
ASSERT(BBTOB(bbcount) >= btp->bt_logical_sectorsize);
fs/xfs/xfs_verify_media.c
188
return clamp(iosize, btp->bt_logical_sectorsize, BBTOB(bbcount));
fs/xfs/xfs_verify_media.c
300
if (!IS_ALIGNED(BBTOB(me->me_start_daddr | me->me_end_daddr),