HCTX_TYPE_READ
[HCTX_TYPE_READ] = "read",
CTX_RQ_SEQ_OPS(read, HCTX_TYPE_READ);
(type != HCTX_TYPE_READ || rq->mq_hctx->type != HCTX_TYPE_DEFAULT))
type = HCTX_TYPE_READ;
case HCTX_TYPE_READ:
set->map[HCTX_TYPE_READ].nr_queues = num_online_cpus();
set->map[HCTX_TYPE_READ].queue_offset = 0;
blk_mq_map_queues(&set->map[HCTX_TYPE_READ]);
set->map[HCTX_TYPE_POLL].queue_offset = set->map[HCTX_TYPE_READ].queue_offset +
set->map[HCTX_TYPE_READ].nr_queues;
set->map[HCTX_TYPE_READ].nr_queues,
set->map[HCTX_TYPE_READ].nr_queues);
vblk->io_queues[HCTX_TYPE_READ] = 0;
vblk->io_queues[HCTX_TYPE_READ],
io_queues[HCTX_TYPE_READ] = opts->nr_io_queues;
nr_io_queues -= io_queues[HCTX_TYPE_READ];
if (opts->nr_write_queues && io_queues[HCTX_TYPE_READ]) {
set->map[HCTX_TYPE_READ].nr_queues =
io_queues[HCTX_TYPE_READ];
set->map[HCTX_TYPE_READ].queue_offset =
set->map[HCTX_TYPE_READ].nr_queues =
set->map[HCTX_TYPE_READ].queue_offset = 0;
blk_mq_map_queues(&set->map[HCTX_TYPE_READ]);
io_queues[HCTX_TYPE_READ];
io_queues[HCTX_TYPE_READ],
dev->io_queues[HCTX_TYPE_READ];
dev->io_queues[HCTX_TYPE_READ] = nr_read_queues;
affd->set_size[HCTX_TYPE_READ] = nr_read_queues;
dev->io_queues[HCTX_TYPE_READ] = 0;
dev->io_queues[HCTX_TYPE_READ],
if (dev->io_queues[HCTX_TYPE_READ])
queue->ctrl->io_queues[HCTX_TYPE_READ];
ctrl->io_queues[HCTX_TYPE_READ];
ctrl->io_queues[HCTX_TYPE_READ] +
mq_map = set->map[HCTX_TYPE_READ].mq_map;
shost->tag_set.map[HCTX_TYPE_READ].nr_queues = 0;
vscsi->io_queues[HCTX_TYPE_READ] = 0;
vscsi->io_queues[HCTX_TYPE_READ],
hba->nr_queues[HCTX_TYPE_READ] = read_queues;
rem -= hba->nr_queues[HCTX_TYPE_READ];
hba->nr_queues[HCTX_TYPE_READ] = 0;
hba->nr_queues[HCTX_TYPE_READ], hba->nr_queues[HCTX_TYPE_POLL],