Symbol: req_op
arch/um/drivers/ubd_kern.c
1178
if (req_op(req->req) == REQ_OP_READ) {
arch/um/drivers/ubd_kern.c
1198
enum req_op op = req_op(req);
arch/um/drivers/ubd_kern.c
1261
enum req_op op = req_op(req);
arch/um/drivers/ubd_kern.c
1298
switch (req_op(req)) {
arch/um/drivers/ubd_kern.c
1416
if (req_op(req->req) == REQ_OP_FLUSH) {
arch/um/drivers/ubd_kern.c
1437
switch (req_op(req->req)) {
arch/um/drivers/ubd_kern.c
451
if (req_op(io_req->req) == REQ_OP_DISCARD)
arch/um/drivers/ubd_kern.c
453
else if (req_op(io_req->req) == REQ_OP_WRITE_ZEROES)
block/bio.c
1527
size_t len, enum req_op op)
block/blk-core.c
1040
unsigned long bdev_start_io_acct(struct block_device *bdev, enum req_op op,
block/blk-core.c
1064
void bdev_end_io_acct(struct block_device *bdev, enum req_op op,
block/blk-core.c
124
inline const char *blk_op_str(enum req_op op)
block/blk-iocost.c
2586
switch (req_op(rq)) {
block/blk-iocost.c
2817
switch (req_op(rq)) {
block/blk-map.c
368
enum req_op op = req_op(rq);
block/blk-merge.c
537
req_op(rq) == REQ_OP_DISCARD ||
block/blk-merge.c
538
req_op(rq) == REQ_OP_SECURE_ERASE)
block/blk-merge.c
554
if (req_op(req) == REQ_OP_DISCARD)
block/blk-merge.c
725
part_stat_inc(req->part, merges[op_stat_group(req_op(req))]);
block/blk-merge.c
727
in_flight[op_is_write(req_op(req))]);
block/blk-merge.c
783
if (req_op(req) != req_op(next))
block/blk-merge.c
903
if (req_op(rq) != bio_op(bio))
block/blk-merge.c
939
part_stat_inc(req->part, merges[op_stat_group(req_op(req))]);
block/blk-mq-debugfs.c
266
const enum req_op op = req_op(rq);
block/blk-mq.c
1079
const int sgrp = op_stat_group(req_op(req));
block/blk-mq.c
1086
in_flight[op_is_write(req_op(req))]);
block/blk-mq.c
1146
part_stat_local_inc(req->part, in_flight[op_is_write(req_op(req))]);
block/blk-mq.c
1388
if (blk_integrity_rq(rq) && req_op(rq) == REQ_OP_WRITE)
block/blk-mq.c
2641
} else if (req_op(rq) == REQ_OP_FLUSH) {
block/blk-mq.c
859
const int sgrp = op_stat_group(req_op(req));
block/blk-mq.c
874
blk_rq_pos(req), (__force u32)req_op(req),
block/blk-mq.c
875
blk_op_str(req_op(req)),
block/blk-mq.c
896
if (blk_integrity_rq(req) && req_op(req) == REQ_OP_READ)
block/blk-mq.c
966
if (blk_integrity_rq(req) && req_op(req) == REQ_OP_READ &&
block/blk-wbt.c
703
const enum req_op op = req_op(rq);
block/blk-zoned.c
292
int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op,
block/blk-zoned.c
440
enum req_op op;
block/blk.h
166
if (req_op(rq) == REQ_OP_FLUSH)
block/blk.h
169
if (req_op(rq) == REQ_OP_WRITE_ZEROES)
block/blk.h
172
if (req_op(rq) == REQ_OP_ZONE_APPEND)
block/blk.h
193
if (req_op(req) == REQ_OP_DISCARD &&
block/blk.h
201
if (req_op(rq) == REQ_OP_DISCARD)
block/blk.h
209
enum req_op op = req_op(rq);
block/blk.h
495
return req_op(rq) == REQ_OP_ZONE_APPEND ||
drivers/block/drbd/drbd_actlog.c
127
sector_t sector, enum req_op op)
drivers/block/drbd/drbd_actlog.c
178
sector_t sector, enum req_op op)
drivers/block/drbd/drbd_bitmap.c
993
enum req_op op = ctx->flags & BM_AIO_READ ? REQ_OP_READ : REQ_OP_WRITE;
drivers/block/drbd/drbd_int.h
1420
struct drbd_backing_dev *bdev, sector_t sector, enum req_op op);
drivers/block/drbd/drbd_receiver.c
2212
static enum req_op wire_flags_to_bio_op(u32 dpf)
drivers/block/loop.c
1863
switch (req_op(rq)) {
drivers/block/loop.c
1899
const bool write = op_is_write(req_op(rq));
drivers/block/loop.c
290
req_op(rq) != REQ_OP_READ) {
drivers/block/loop.c
326
if (req_op(rq) == REQ_OP_WRITE)
drivers/block/loop.c
418
switch (req_op(rq)) {
drivers/block/nbd.c
419
switch (req_op(req)) {
drivers/block/null_blk/main.c
1301
op_is_write(req_op(rq)), pos,
drivers/block/null_blk/main.c
1372
blk_status_t null_handle_memory_backed(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/main.c
1389
if (!dev->memory_backed && req_op(rq) == REQ_OP_READ) {
drivers/block/null_blk/main.c
1423
blk_status_t null_process_cmd(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/main.c
1443
sector_t nr_sectors, enum req_op op)
drivers/block/null_blk/main.c
1608
cmd->error = null_process_cmd(cmd, req_op(req), blk_rq_pos(req),
drivers/block/null_blk/main.c
1703
null_handle_cmd(cmd, sector, nr_sectors, req_op(rq));
drivers/block/null_blk/null_blk.h
134
blk_status_t null_process_cmd(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/null_blk.h
138
blk_status_t null_handle_memory_backed(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/null_blk.h
148
blk_status_t null_process_zoned_cmd(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/null_blk.h
167
enum req_op op, sector_t sector, sector_t nr_sectors)
drivers/block/null_blk/trace.h
44
__field_struct(enum req_op, op)
drivers/block/null_blk/trace.h
49
__entry->op = req_op(blk_mq_rq_from_pdu(cmd));
drivers/block/null_blk/zoned.c
646
static blk_status_t null_zone_mgmt(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/zoned.c
708
blk_status_t null_process_zoned_cmd(struct nullb_cmd *cmd, enum req_op op,
drivers/block/ps3disk.c
172
switch (req_op(req)) {
drivers/block/ps3disk.c
235
if (req_op(req) == REQ_OP_FLUSH) {
drivers/block/rbd.c
4790
switch (req_op(bd->rq)) {
drivers/block/rbd.c
4804
rbd_warn(rbd_dev, "unknown req_op %d", req_op(bd->rq));
drivers/block/rnbd/rnbd-clt.c
1014
if ((req_op(rq) != REQ_OP_DISCARD) && (req_op(rq) != REQ_OP_WRITE_ZEROES))
drivers/block/rnbd/rnbd-proto.h
278
switch (req_op(rq)) {
drivers/block/rnbd/rnbd-proto.h
302
(__force u32)req_op(rq),
drivers/block/ublk_drv.c
1346
return ublk_rq_has_data(req) && req_op(req) == REQ_OP_WRITE;
drivers/block/ublk_drv.c
1352
(req_op(req) == REQ_OP_READ || req_op(req) == REQ_OP_DRV_IN);
drivers/block/ublk_drv.c
1437
switch (req_op(req)) {
drivers/block/ublk_drv.c
1490
if (!io->res && req_op(req) == REQ_OP_READ)
drivers/block/ublk_drv.c
1504
if (req_op(req) != REQ_OP_READ && req_op(req) != REQ_OP_WRITE &&
drivers/block/ublk_drv.c
1505
req_op(req) != REQ_OP_DRV_IN)
drivers/block/ublk_drv.c
3215
req_op(req) == REQ_OP_READ))
drivers/block/ublk_drv.c
3217
} else if (req_op(req) != REQ_OP_ZONE_APPEND && buf_addr) {
drivers/block/ublk_drv.c
3353
if (req_op(req) == REQ_OP_ZONE_APPEND)
drivers/block/ublk_drv.c
3686
if (req_op(req) == REQ_OP_ZONE_APPEND)
drivers/block/ublk_drv.c
3920
if ((req_op(req) == REQ_OP_READ || req_op(req) == REQ_OP_DRV_IN) &&
drivers/block/ublk_drv.c
3925
if ((req_op(req) == REQ_OP_WRITE ||
drivers/block/ublk_drv.c
3926
req_op(req) == REQ_OP_ZONE_APPEND) &&
drivers/block/ublk_drv.c
622
switch (req_op(req)) {
drivers/block/virtio_blk.c
247
if (!IS_ENABLED(CONFIG_BLK_DEV_ZONED) && op_is_zone_mgmt(req_op(req)))
drivers/block/virtio_blk.c
253
switch (req_op(req)) {
drivers/block/virtio_blk.c
343
if (req_op(req) == REQ_OP_ZONE_APPEND)
drivers/block/xen-blkback/blkback.c
1299
enum req_op operation;
drivers/block/xen-blkback/blkback.c
450
enum req_op operation)
drivers/block/xen-blkfront.c
2085
if (req_op(shadow[j].request) == REQ_OP_FLUSH ||
drivers/block/xen-blkfront.c
2086
req_op(shadow[j].request) == REQ_OP_DISCARD ||
drivers/block/xen-blkfront.c
2087
req_op(shadow[j].request) == REQ_OP_SECURE_ERASE ||
drivers/block/xen-blkfront.c
566
if (req_op(req) == REQ_OP_SECURE_ERASE && info->feature_secdiscard)
drivers/block/xen-blkfront.c
771
BUG_ON(req_op(req) == REQ_OP_FLUSH || req->cmd_flags & REQ_FUA);
drivers/block/xen-blkfront.c
783
if (req_op(req) == REQ_OP_FLUSH ||
drivers/block/xen-blkfront.c
784
(req_op(req) == REQ_OP_WRITE && (req->cmd_flags & REQ_FUA))) {
drivers/block/xen-blkfront.c
876
if (unlikely(req_op(req) == REQ_OP_DISCARD ||
drivers/block/xen-blkfront.c
877
req_op(req) == REQ_OP_SECURE_ERASE))
drivers/block/xen-blkfront.c
915
if (unlikely(req_op(qd->rq) == REQ_OP_FLUSH && !info->feature_flush))
drivers/block/zloop.c
388
bool is_append = req_op(rq) == REQ_OP_ZONE_APPEND;
drivers/block/zloop.c
389
bool is_write = req_op(rq) == REQ_OP_WRITE || is_append;
drivers/block/zloop.c
569
switch (req_op(rq)) {
drivers/block/zloop.c
599
pr_err("Unsupported operation %d\n", req_op(rq));
drivers/block/zloop.c
625
switch (req_op(rq)) {
drivers/block/zloop.c
644
req_op(rq) == REQ_OP_WRITE ? "" : "append ",
drivers/block/zloop.c
662
if (req_op(rq) == REQ_OP_ZONE_APPEND)
drivers/block/zloop.c
719
if (zlo->ordered_zone_append && req_op(rq) == REQ_OP_ZONE_APPEND) {
drivers/cdrom/gdrom.c
639
switch (req_op(bd->rq)) {
drivers/crypto/hisilicon/sec2/sec.h
181
const struct sec_req_op *req_op;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1490
ret = ctx->req_op->buf_map(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1494
ctx->req_op->do_transfer(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1496
ret = ctx->req_op->bd_fill(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1503
ctx->req_op->buf_unmap(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1509
ctx->req_op->buf_unmap(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1986
ret = ctx->req_op->bd_send(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2072
ctx->req_op = &sec_skcipher_req_ops;
drivers/crypto/hisilicon/sec2/sec_crypto.c
2075
ctx->req_op = &sec_skcipher_req_ops_v3;
drivers/crypto/hisilicon/sec2/sec_crypto.c
2105
ctx->req_op = &sec_aead_req_ops;
drivers/crypto/hisilicon/sec2/sec_crypto.c
2108
ctx->req_op = &sec_aead_req_ops_v3;
drivers/crypto/hisilicon/sec2/sec_crypto.c
2355
return ctx->req_op->process(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
244
ctx->req_op->buf_unmap(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
2583
return ctx->req_op->process(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
307
ctx->req_op->buf_unmap(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
308
ctx->req_op->callback(ctx, req, err);
drivers/crypto/hisilicon/sec2/sec_crypto.c
332
ctx->req_op->buf_unmap(ctx, req);
drivers/crypto/hisilicon/sec2/sec_crypto.c
333
ctx->req_op->callback(ctx, req, err);
drivers/hwmon/drivetemp.c
165
enum req_op op;
drivers/md/dm-bufio.c
1297
static void use_dmio(struct dm_buffer *b, enum req_op op, sector_t sector,
drivers/md/dm-bufio.c
1337
static void use_bio(struct dm_buffer *b, enum req_op op, sector_t sector,
drivers/md/dm-bufio.c
1377
static void submit_io(struct dm_buffer *b, enum req_op op, unsigned short ioprio,
drivers/md/dm-ebs-target.c
122
static int __ebs_rw_bio(struct ebs_c *ec, enum req_op op, struct bio *bio)
drivers/md/dm-ebs-target.c
65
static int __ebs_rw_bvec(struct ebs_c *ec, enum req_op op, struct bio_vec *bv,
drivers/md/dm-integrity.c
317
enum req_op op;
drivers/md/dm-integrity.c
544
const enum req_op op = opf & REQ_OP_MASK;
drivers/md/dm-io.c
319
const enum req_op op = opf & REQ_OP_MASK;
drivers/md/dm-kcopyd.c
361
enum req_op op;
drivers/md/dm-log.c
299
static int rw_header(struct log_c *lc, enum req_op op)
drivers/md/dm-rq.c
225
if (req_op(clone) == REQ_OP_DISCARD &&
drivers/md/dm-rq.c
228
else if (req_op(clone) == REQ_OP_WRITE_ZEROES &&
drivers/md/dm-zoned-metadata.c
2042
unsigned int chunk, enum req_op op)
drivers/md/dm-zoned-metadata.c
735
static int dmz_rdwr_block(struct dmz_dev *dev, enum req_op op,
drivers/md/dm-zoned.h
249
unsigned int chunk, enum req_op op);
drivers/md/dm.c
1328
enum req_op op = bio_op(bio);
drivers/md/raid10.c
2586
int sectors, struct page *page, enum req_op op)
drivers/md/raid5.c
1159
enum req_op op;
drivers/mmc/core/block.c
2425
switch (req_op(req)) {
drivers/mmc/core/block.c
2449
switch (req_op(req)) {
drivers/mmc/core/queue.c
46
switch (req_op(req)) {
drivers/mmc/core/queue.c
67
if (req_op(req) == REQ_OP_READ || req_op(req) == REQ_OP_WRITE)
drivers/mtd/mtd_blkdevs.c
57
switch (req_op(req)) {
drivers/mtd/ubi/block.c
305
switch (req_op(bd->rq)) {
drivers/nvdimm/btt.c
1419
enum req_op op, sector_t sector)
drivers/nvme/host/core.c
1089
switch (req_op(req)) {
drivers/nvme/host/core.c
427
req_op(req) == REQ_OP_ZONE_APPEND) {
drivers/nvme/host/multipath.c
196
nvme_req(rq)->start_time = bdev_start_io_acct(disk->part0, req_op(rq),
drivers/nvme/host/multipath.c
210
bdev_end_io_acct(ns->head->disk->part0, req_op(rq),
drivers/nvme/host/nvme.h
900
static inline enum req_op nvme_req_op(struct nvme_command *cmd)
drivers/nvme/host/pci.c
1949
nvmeq->qid, blk_op_str(req_op(req)), req_op(req),
drivers/nvme/host/zns.c
244
if (req_op(req) == REQ_OP_ZONE_RESET_ALL)
drivers/nvme/target/passthru.c
274
ARRAY_SIZE(req->inline_bvec), req_op(rq));
drivers/nvme/target/passthru.c
276
bio = bio_alloc(NULL, bio_max_segs(req->sg_cnt), req_op(rq),
drivers/nvme/target/zns.c
321
static inline enum req_op zsa_req_op(u8 zsa)
drivers/nvme/target/zns.c
476
enum req_op op = zsa_req_op(req->cmd->zms.zsa);
drivers/s390/block/dasd_fba.c
555
if (req_op(req) == REQ_OP_DISCARD || req_op(req) == REQ_OP_WRITE_ZEROES)
drivers/scsi/ch.c
188
void *buffer, unsigned int buflength, enum req_op op)
drivers/scsi/scsi_lib.c
1114
!op_is_write(req_op(rq)) &&
drivers/scsi/scsi_lib.c
2159
enum req_op op = data_direction == DMA_TO_DEVICE ? REQ_OP_DRV_OUT :
drivers/scsi/scsi_transport_spi.c
108
enum req_op op, void *buffer, unsigned int bufflen,
drivers/scsi/sd.c
1474
switch (req_op(rq)) {
drivers/scsi/sd.c
2322
switch (req_op(req)) {
drivers/scsi/sd_zbc.c
369
if (op_is_zone_mgmt(req_op(rq)) &&
drivers/scsi/sr.c
396
switch (req_op(rq)) {
fs/btrfs/raid56.c
1239
unsigned int sector_nr, enum req_op op)
fs/buffer.c
2781
const enum req_op op = opf & REQ_OP_MASK;
fs/direct-io.c
169
const enum req_op dio_op = dio->opf & REQ_OP_MASK;
fs/direct-io.c
246
const enum req_op dio_op = dio->opf & REQ_OP_MASK;
fs/direct-io.c
336
const enum req_op dio_op = dio->opf & REQ_OP_MASK;
fs/direct-io.c
427
const enum req_op dio_op = dio->opf & REQ_OP_MASK;
fs/direct-io.c
502
const enum req_op dio_op = dio->opf & REQ_OP_MASK;
fs/direct-io.c
606
const enum req_op dio_op = dio->opf & REQ_OP_MASK;
fs/direct-io.c
789
const enum req_op dio_op = dio->opf & REQ_OP_MASK;
fs/direct-io.c
906
const enum req_op dio_op = dio->opf & REQ_OP_MASK;
fs/f2fs/f2fs.h
1342
enum req_op op; /* contains REQ_OP_ */
fs/nfs/blocklayout/blocklayout.c
124
do_add_page_to_bio(struct bio *bio, int npg, enum req_op op, sector_t isect,
fs/xfs/xfs_bio_io.c
18
enum req_op op)
fs/xfs/xfs_log_recover.c
125
enum req_op op)
fs/xfs/xfs_platform.h
225
char *data, enum req_op op);
fs/zonefs/file.c
203
enum req_op op;
fs/zonefs/super.c
100
struct zonefs_zone *z, enum req_op op)
fs/zonefs/super.c
128
int zonefs_inode_zone_mgmt(struct inode *inode, enum req_op op)
fs/zonefs/trace.h
24
enum req_op op),
fs/zonefs/trace.h
29
__field(enum req_op, op)
fs/zonefs/zonefs.h
260
int zonefs_inode_zone_mgmt(struct inode *inode, enum req_op op);
include/linux/bio.h
467
size_t len, enum req_op op);
include/linux/blk-mq.h
229
static inline enum req_op req_op(const struct request *req)
include/linux/blk-mq.h
246
#define rq_data_dir(rq) (op_is_write(req_op(rq)) ? WRITE : READ)
include/linux/blk-mq.h
249
(op_is_write(req_op(rq)) ? DMA_TO_DEVICE : DMA_FROM_DEVICE)
include/linux/blk_types.h
454
static inline enum req_op bio_op(const struct bio *bio)
include/linux/blk_types.h
492
static inline bool op_is_zone_mgmt(enum req_op op)
include/linux/blk_types.h
506
static inline int op_stat_group(enum req_op op)
include/linux/blkdev.h
1030
extern const char *blk_op_str(enum req_op op);
include/linux/blkdev.h
1704
unsigned long bdev_start_io_acct(struct block_device *bdev, enum req_op op,
include/linux/blkdev.h
1706
void bdev_end_io_acct(struct block_device *bdev, enum req_op op,
include/linux/blkdev.h
447
int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op,
include/linux/blkdev.h
872
enum req_op op = bio_op(bio);
include/trace/events/f2fs.h
1184
__field(enum req_op, op)
include/trace/events/f2fs.h
1240
__field(enum req_op, op)
include/trace/events/nilfs2.h
195
enum req_op mode),
include/trace/events/nilfs2.h
207
__field_struct(enum req_op, mode)
kernel/trace/blktrace.c
324
const enum req_op op = opf & REQ_OP_MASK;