blk_rq_bytes
io_req->io_desc[0].length = blk_rq_bytes(req);
blkg_rwstat_add(&bfqg->stats.bytes, rq->cmd_flags, blk_rq_bytes(rq));
return bio_crypt_ctx_mergeable(req->crypt_ctx, blk_rq_bytes(req),
return bio_crypt_ctx_mergeable(req->crypt_ctx, blk_rq_bytes(req),
req->__data_len += blk_rq_bytes(next);
if (blk_rq_bytes(req) < blk_rq_cur_bytes(req)) {
if (blk_rq_bytes(req) & (bdev_logical_block_size(bio->bi_bdev) - 1))
if (blk_update_request(rq, error, blk_rq_bytes(rq)))
blk_rq_bytes(last) >= BLK_PLUG_FLUSH_SIZE)) {
if (q->disk && should_fail_request(q->disk->part0, blk_rq_bytes(rq)))
rq->__data_len = blk_rq_bytes(rq_src);
rq->bio, rq->biotail, blk_rq_bytes(rq));
int total_bytes = blk_rq_bytes(req);
if (blk_crypto_rq_has_keyslot(req) && nr_bytes >= blk_rq_bytes(req))
buf->payload_len = blk_rq_bytes(req);
req_blocks = blk_rq_bytes(rq) / scmd->device->sector_size;
blk_rq_bytes(fd_request)));
if (remaining > blk_rq_bytes(current_req) && CT(raw_cmd->cmd[COMMAND]) == FD_WRITE) {
ret = file->f_op->fallocate(file, mode, pos, blk_rq_bytes(rq));
if (cmd->ret < 0 || cmd->ret == blk_rq_bytes(rq) ||
iov_iter_bvec(&iter, rw, bvec, nr_bvec, blk_rq_bytes(rq));
blk_rq_bytes(req), (req->timeout / HZ) * cmd->retries);
request.len = htonl(blk_rq_bytes(req));
(unsigned long long)blk_rq_pos(req) << 9, blk_rq_bytes(req));
if (atomic_long_sub_return(blk_rq_bytes(rq), &nullb->cur_bytes) < 0) {
u64 len = blk_rq_bytes(rq);
u64 length = blk_rq_bytes(rq);
msg.bi_size = cpu_to_le32(blk_rq_bytes(rq));
const unsigned int rq_bytes = blk_rq_bytes(req);
const unsigned int rq_bytes = blk_rq_bytes(req);
if (unlikely(mapped_bytes != blk_rq_bytes(req))) {
data_len = blk_rq_bytes(req);
iov_iter_bvec(&iter, rw, cmd->bvec, nr_bvec, blk_rq_bytes(rq));
nr_bvec, blk_rq_bytes(rq));
if (cmd->ret >= 0 && cmd->ret != blk_rq_bytes(rq)) {
if (cmd->ret >= 0 && cmd->ret != blk_rq_bytes(rq)) {
zone_no, cmd->ret, blk_rq_bytes(rq));
size_t nr_bytes = blk_rq_bytes(rq);
blk_rq_bytes(req), &len);
blk_rq_bytes(req), &len);
count = blk_rq_bytes(msb->block_req);
t_len = blk_rq_bytes(msb->block_req);
} else if (!blk_rq_bytes(req)) {
int to_read = blk_rq_bytes(req);
cpu_to_le16((blk_rq_bytes(req) >> ns->head->lba_shift) - 1);
blk_rq_bytes(req) >> ns->head->lba_shift,
cpu_to_le16((blk_rq_bytes(req) >> ns->head->lba_shift) - 1);
if (blk_rq_bytes(req) > queue_atomic_write_unit_max_bytes(q))
u64 end = start + blk_rq_bytes(req) - 1;
if (blk_rq_bytes(req) > boundary_bytes)
blk_rq_bytes(rq) >> SECTOR_SHIFT,
blk_rq_bytes(req));
data_size = blk_rq_bytes(req);
aidaw = scm_aidaw_fetch(scmrq, blk_rq_bytes(req));
if (likely(blk_rq_bytes(req) > 0 || blk_stat == BLK_STS_OK)) {
if (scsi_end_request(req, blk_stat, blk_rq_bytes(req)))
if (blk_rq_bytes(rq) & rq->q->limits.dma_pad_mask) {
(rq->q->limits.dma_pad_mask & ~blk_rq_bytes(rq)) + 1;
BUG_ON(blk_rq_bytes(req));
cmd->transfersize = blk_rq_bytes(req);
if (blk_rq_bytes(req))
return blk_rq_bytes(rq);
BUG_ON(blk_rq_bytes(rq) && !bytes);
} else if (blk_rq_bytes(req) == 0 && sense_current) {
good_bytes = blk_rq_bytes(req);
scsi_set_resid(SCpnt, blk_rq_bytes(req));
transfer_len = blk_rq_bytes(rq);
return blk_rq_bytes(rq) >> SECTOR_SHIFT;
return blk_rq_bytes(rq);
return blk_rq_bytes(scsi_cmd_to_rq(scmd)) >> shift;
__entry->bytes = blk_rq_bytes(rq);
imu->len = blk_rq_bytes(rq);
blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_INSERT,
blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_ISSUE,
blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_BACKMERGE,
blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_REQUEUE,
blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_ZONE_APPEND,
__blk_add_trace(bt, blk_rq_pos(rq), blk_rq_bytes(rq),
__blk_add_trace(bt, blk_rq_trace_sector(rq), blk_rq_bytes(rq), 0,