REQ_RAHEAD
if (bio->bi_opf & REQ_RAHEAD)
if (bio->bi_opf & REQ_RAHEAD)
if (bio->bi_opf & REQ_RAHEAD)
else if (bio->bi_opf & REQ_RAHEAD)
!(req->master_bio->bi_opf & REQ_RAHEAD) &&
if (bio->bi_opf & REQ_RAHEAD)
if (bio->bi_opf & REQ_RAHEAD) {
if ((bio->bi_opf & (REQ_RAHEAD|REQ_BACKGROUND))) {
if (bio->bi_opf & REQ_RAHEAD)
if (bio->bi_opf & REQ_RAHEAD)
if (bio->bi_opf & REQ_RAHEAD)
static blk_opf_t PASSTHROUGH_FLAGS = (REQ_PRIO | REQ_META | REQ_SYNC | REQ_RAHEAD);
unlikely(!(bio->bi_opf & REQ_RAHEAD)) &&
(bio->bi_opf & REQ_RAHEAD))) {
if (bio->bi_opf & REQ_RAHEAD) {
else if (bio->bi_opf & REQ_RAHEAD)
return !(bio->bi_opf & (REQ_RAHEAD | REQ_NOWAIT));
if (bi->bi_opf & REQ_RAHEAD)
if (req->cmd_flags & REQ_RAHEAD)
if (req->cmd_flags & (REQ_FAILFAST_DEV | REQ_RAHEAD))
else if (!(bio->bi_opf & REQ_RAHEAD))
.opf = REQ_OP_READ | REQ_RAHEAD,
bh_readahead(bh, REQ_RAHEAD);
bio->bi_opf |= REQ_RAHEAD;
(ignore_locked ? REQ_RAHEAD : 0),
bio->bi_opf |= REQ_RAHEAD;
ext4_read_bh_nowait(bh, REQ_RAHEAD, NULL, false);
.op_flags = sync ? (REQ_META | REQ_PRIO) : REQ_RAHEAD,
return rac ? REQ_RAHEAD : 0;
REQ_RAHEAD, true, NULL);
err = read_node_folio(afolio, REQ_RAHEAD);
submit_bh(REQ_OP_READ | REQ_RAHEAD | REQ_META |
submit_bh(REQ_OP_READ | REQ_RAHEAD | REQ_META |
bh_readahead(bh, REQ_RAHEAD | REQ_META | REQ_PRIO);
bio->bi_opf |= REQ_RAHEAD;
opf |= REQ_RAHEAD;
if (opf & REQ_RAHEAD) {
REQ_OP_READ | REQ_RAHEAD,
if (opf & REQ_RAHEAD) {
REQ_OP_READ | REQ_RAHEAD, &bh);
bio->bi_opf |= REQ_RAHEAD;
bh_readahead_batch(num, bha, REQ_RAHEAD);
op |= REQ_RAHEAD;
#define F2FS_OP_FLAGS (REQ_RAHEAD | REQ_SYNC | REQ_META | REQ_PRIO | \
{ (__force u32)REQ_RAHEAD, "R" }, \
if (opf & REQ_RAHEAD)