arch/um/drivers/ubd_kern.c
159
struct blk_mq_tag_set tag_set;
arch/um/drivers/ubd_kern.c
784
blk_mq_free_tag_set(&ubd_dev->tag_set);
arch/um/drivers/ubd_kern.c
865
ubd_dev->tag_set.ops = &ubd_mq_ops;
arch/um/drivers/ubd_kern.c
866
ubd_dev->tag_set.queue_depth = 64;
arch/um/drivers/ubd_kern.c
867
ubd_dev->tag_set.numa_node = NUMA_NO_NODE;
arch/um/drivers/ubd_kern.c
868
ubd_dev->tag_set.driver_data = ubd_dev;
arch/um/drivers/ubd_kern.c
869
ubd_dev->tag_set.nr_hw_queues = 1;
arch/um/drivers/ubd_kern.c
871
err = blk_mq_alloc_tag_set(&ubd_dev->tag_set);
arch/um/drivers/ubd_kern.c
875
disk = blk_mq_alloc_disk(&ubd_dev->tag_set, &lim, ubd_dev);
arch/um/drivers/ubd_kern.c
907
blk_mq_free_tag_set(&ubd_dev->tag_set);
block/blk-mq-debugfs.c
359
blk_mq_tagset_busy_iter(hctx->queue->tag_set, hctx_show_busy_rq,
block/blk-mq-sched.c
553
struct blk_mq_tag_set *set = q->tag_set;
block/blk-mq-sched.c
617
unsigned int flags = q->tag_set->flags;
block/blk-mq-sched.c
675
if (blk_mq_is_shared_tags(q->tag_set->flags)) {
block/blk-mq-sched.c
676
blk_mq_free_rqs(q->tag_set, q->sched_shared_tags,
block/blk-mq-sched.c
681
blk_mq_free_rqs(q->tag_set,
block/blk-mq-sysfs.c
233
mutex_lock(&q->tag_set->tag_list_lock);
block/blk-mq-sysfs.c
239
mutex_unlock(&q->tag_set->tag_list_lock);
block/blk-mq-sysfs.c
247
mutex_unlock(&q->tag_set->tag_list_lock);
block/blk-mq-sysfs.c
260
mutex_lock(&q->tag_set->tag_list_lock);
block/blk-mq-sysfs.c
263
mutex_unlock(&q->tag_set->tag_list_lock);
block/blk-mq-tag.c
271
struct blk_mq_tag_set *set = q->tag_set;
block/blk-mq-tag.c
509
srcu_idx = srcu_read_lock(&q->tag_set->tags_srcu);
block/blk-mq-tag.c
510
if (blk_mq_is_shared_tags(q->tag_set->flags)) {
block/blk-mq-tag.c
511
struct blk_mq_tags *tags = q->tag_set->shared_tags;
block/blk-mq-tag.c
539
srcu_read_unlock(&q->tag_set->tags_srcu, srcu_idx);
block/blk-mq-tag.c
629
nr - q->tag_set->reserved_tags);
block/blk-mq.c
1770
blk_mq_wait_quiesce_done(q->tag_set);
block/blk-mq.c
304
blk_mq_wait_quiesce_done(q->tag_set);
block/blk-mq.c
3695
srcu_idx = srcu_read_lock(&hctx->queue->tag_set->tags_srcu);
block/blk-mq.c
3697
srcu_read_unlock(&hctx->queue->tag_set->tags_srcu, srcu_idx);
block/blk-mq.c
4090
struct blk_mq_tag_set *set = q->tag_set;
block/blk-mq.c
4177
struct blk_mq_tag_set *set = q->tag_set;
block/blk-mq.c
4325
struct blk_mq_tag_set *set = q->tag_set;
block/blk-mq.c
4640
q->tag_set = set;
block/blk-mq.c
4684
struct blk_mq_tag_set *set = q->tag_set;
block/blk-mq.c
4997
struct blk_mq_tag_set *set = q->tag_set;
block/blk-mq.c
5080
lockdep_assert_held_write(&q->tag_set->update_nr_hwq_lock);
block/blk-mq.h
441
if ((q)->tag_set->flags & BLK_MQ_F_BLOCKING) { \
block/blk-mq.h
442
struct blk_mq_tag_set *__tag_set = (q)->tag_set; \
block/blk-mq.h
462
q->tag_set->map[HCTX_TYPE_POLL].nr_queues;
block/blk-mq.h
87
return queue_hctx((q), (q->tag_set->map[type].mq_map[cpu]));
block/blk-sysfs.c
68
struct blk_mq_tag_set *set = q->tag_set;
block/bsg-lib.c
22
struct blk_mq_tag_set tag_set;
block/bsg-lib.c
279
container_of(q->tag_set, struct bsg_set, tag_set);
block/bsg-lib.c
324
container_of(q->tag_set, struct bsg_set, tag_set);
block/bsg-lib.c
329
blk_mq_free_tag_set(&bset->tag_set);
block/bsg-lib.c
338
container_of(rq->q->tag_set, struct bsg_set, tag_set);
block/bsg-lib.c
378
set = &bset->tag_set;
block/elevator.c
591
q->nr_requests = q->tag_set->queue_depth;
block/elevator.c
592
q->async_depth = q->tag_set->queue_depth;
block/elevator.c
622
blk_mq_free_sched_res(&ctx->res, ctx->type, q->tag_set);
block/elevator.c
639
blk_mq_free_sched_res(&res, ctx->old->type, q->tag_set);
block/elevator.c
656
struct blk_mq_tag_set *set = q->tag_set;
block/elevator.c
703
struct blk_mq_tag_set *set = q->tag_set;
block/elevator.c
740
if (q->tag_set->flags & BLK_MQ_F_NO_SCHED_BY_DEFAULT)
block/elevator.c
753
blk_mq_is_shared_tags(q->tag_set->flags))) {
block/elevator.c
793
struct blk_mq_tag_set *set = q->tag_set;
block/genhd.c
594
set = disk->queue->tag_set;
block/genhd.c
782
struct blk_mq_tag_set *set = q->tag_set;
block/genhd.c
817
set = disk->queue->tag_set;
drivers/block/amiflop.c
1786
disk = blk_mq_alloc_disk(&unit[drive].tag_set, &lim, NULL);
drivers/block/amiflop.c
1816
memset(&unit[drive].tag_set, 0, sizeof(unit[drive].tag_set));
drivers/block/amiflop.c
1817
unit[drive].tag_set.ops = &amiflop_mq_ops;
drivers/block/amiflop.c
1818
unit[drive].tag_set.nr_hw_queues = 1;
drivers/block/amiflop.c
1819
unit[drive].tag_set.nr_maps = 1;
drivers/block/amiflop.c
1820
unit[drive].tag_set.queue_depth = 2;
drivers/block/amiflop.c
1821
unit[drive].tag_set.numa_node = NUMA_NO_NODE;
drivers/block/amiflop.c
1822
if (blk_mq_alloc_tag_set(&unit[drive].tag_set))
drivers/block/amiflop.c
1832
blk_mq_free_tag_set(&unit[drive].tag_set);
drivers/block/amiflop.c
205
struct blk_mq_tag_set tag_set;
drivers/block/aoe/aoe.h
175
struct blk_mq_tag_set tag_set;
drivers/block/aoe/aoeblk.c
365
set = &d->tag_set;
drivers/block/aoe/aoedev.c
293
blk_mq_free_tag_set(&d->tag_set);
drivers/block/ataflop.c
2001
disk = blk_mq_alloc_disk(&unit[drive].tag_set, &lim, NULL);
drivers/block/ataflop.c
2055
blk_mq_free_tag_set(&unit[i].tag_set);
drivers/block/ataflop.c
2073
blk_mq_free_tag_set(&fs->tag_set);
drivers/block/ataflop.c
2086
memset(&unit[i].tag_set, 0, sizeof(unit[i].tag_set));
drivers/block/ataflop.c
2087
unit[i].tag_set.ops = &ataflop_mq_ops;
drivers/block/ataflop.c
2088
unit[i].tag_set.nr_hw_queues = 1;
drivers/block/ataflop.c
2089
unit[i].tag_set.nr_maps = 1;
drivers/block/ataflop.c
2090
unit[i].tag_set.queue_depth = 2;
drivers/block/ataflop.c
2091
unit[i].tag_set.numa_node = NUMA_NO_NODE;
drivers/block/ataflop.c
2092
ret = blk_mq_alloc_tag_set(&unit[i].tag_set);
drivers/block/ataflop.c
2098
blk_mq_free_tag_set(&unit[i].tag_set);
drivers/block/ataflop.c
305
struct blk_mq_tag_set tag_set;
drivers/block/loop.c
2037
lo->tag_set.ops = &loop_mq_ops;
drivers/block/loop.c
2038
lo->tag_set.nr_hw_queues = 1;
drivers/block/loop.c
2039
lo->tag_set.queue_depth = hw_queue_depth;
drivers/block/loop.c
2040
lo->tag_set.numa_node = NUMA_NO_NODE;
drivers/block/loop.c
2041
lo->tag_set.cmd_size = sizeof(struct loop_cmd);
drivers/block/loop.c
2042
lo->tag_set.flags = BLK_MQ_F_STACKING | BLK_MQ_F_NO_SCHED_BY_DEFAULT;
drivers/block/loop.c
2043
lo->tag_set.driver_data = lo;
drivers/block/loop.c
2045
err = blk_mq_alloc_tag_set(&lo->tag_set);
drivers/block/loop.c
2049
disk = lo->lo_disk = blk_mq_alloc_disk(&lo->tag_set, &lim, lo);
drivers/block/loop.c
2106
blk_mq_free_tag_set(&lo->tag_set);
drivers/block/loop.c
2121
blk_mq_free_tag_set(&lo->tag_set);
drivers/block/loop.c
73
struct blk_mq_tag_set tag_set;
drivers/block/nbd.c
1053
blk_mq_tagset_busy_iter(&nbd->tag_set, nbd_clear_req, NULL);
drivers/block/nbd.c
117
struct blk_mq_tag_set tag_set;
drivers/block/nbd.c
1343
if (nbd->tag_set.timeout)
drivers/block/nbd.c
1344
sock->sk->sk_sndtimeo = nbd->tag_set.timeout;
drivers/block/nbd.c
1463
nbd->tag_set.timeout = 0;
drivers/block/nbd.c
1489
blk_mq_update_nr_hw_queues(&nbd->tag_set, num_connections);
drivers/block/nbd.c
1529
if (nbd->tag_set.timeout)
drivers/block/nbd.c
1531
nbd->tag_set.timeout;
drivers/block/nbd.c
1585
nbd->tag_set.timeout = timeout * HZ;
drivers/block/nbd.c
1838
debugfs_create_u32("timeout", 0444, dir, &nbd->tag_set.timeout);
drivers/block/nbd.c
1923
nbd->tag_set.ops = &nbd_mq_ops;
drivers/block/nbd.c
1924
nbd->tag_set.nr_hw_queues = 1;
drivers/block/nbd.c
1925
nbd->tag_set.queue_depth = 128;
drivers/block/nbd.c
1926
nbd->tag_set.numa_node = NUMA_NO_NODE;
drivers/block/nbd.c
1927
nbd->tag_set.cmd_size = sizeof(struct nbd_cmd);
drivers/block/nbd.c
1928
nbd->tag_set.flags = BLK_MQ_F_BLOCKING;
drivers/block/nbd.c
1929
nbd->tag_set.driver_data = nbd;
drivers/block/nbd.c
1933
err = blk_mq_alloc_tag_set(&nbd->tag_set);
drivers/block/nbd.c
1954
disk = blk_mq_alloc_disk(&nbd->tag_set, &lim, NULL);
drivers/block/nbd.c
2004
blk_mq_free_tag_set(&nbd->tag_set);
drivers/block/nbd.c
269
blk_mq_free_tag_set(&nbd->tag_set);
drivers/block/nbd.c
480
(config->num_connections == 1 && nbd->tag_set.timeout)) {
drivers/block/nbd.c
513
if (!nbd->tag_set.timeout) {
drivers/block/nbd.c
886
if (hwq < nbd->tag_set.nr_hw_queues)
drivers/block/nbd.c
887
req = blk_mq_tag_to_rq(nbd->tag_set.tags[hwq],
drivers/block/null_blk/main.c
1780
if (nullb->tag_set == &nullb->__tag_set)
drivers/block/null_blk/main.c
1781
blk_mq_free_tag_set(nullb->tag_set);
drivers/block/null_blk/main.c
1845
if (tag_set.ops)
drivers/block/null_blk/main.c
1848
tag_set.nr_hw_queues = g_submit_queues;
drivers/block/null_blk/main.c
1849
tag_set.queue_depth = g_hw_queue_depth;
drivers/block/null_blk/main.c
1850
tag_set.numa_node = g_home_node;
drivers/block/null_blk/main.c
1852
tag_set.flags |= BLK_MQ_F_NO_SCHED_BY_DEFAULT;
drivers/block/null_blk/main.c
1854
tag_set.flags |= BLK_MQ_F_TAG_HCTX_SHARED;
drivers/block/null_blk/main.c
1856
tag_set.flags |= BLK_MQ_F_BLOCKING;
drivers/block/null_blk/main.c
1858
error = null_init_tag_set(&tag_set, g_poll_queues);
drivers/block/null_blk/main.c
1860
tag_set.ops = NULL;
drivers/block/null_blk/main.c
1867
nullb->tag_set = &tag_set;
drivers/block/null_blk/main.c
1871
nullb->tag_set = &nullb->__tag_set;
drivers/block/null_blk/main.c
1872
nullb->tag_set->driver_data = nullb;
drivers/block/null_blk/main.c
1873
nullb->tag_set->nr_hw_queues = nullb->dev->submit_queues;
drivers/block/null_blk/main.c
1874
nullb->tag_set->queue_depth = nullb->dev->hw_queue_depth;
drivers/block/null_blk/main.c
1875
nullb->tag_set->numa_node = nullb->dev->home_node;
drivers/block/null_blk/main.c
1877
nullb->tag_set->flags |= BLK_MQ_F_NO_SCHED_BY_DEFAULT;
drivers/block/null_blk/main.c
1879
nullb->tag_set->flags |= BLK_MQ_F_TAG_HCTX_SHARED;
drivers/block/null_blk/main.c
1881
nullb->tag_set->flags |= BLK_MQ_F_BLOCKING;
drivers/block/null_blk/main.c
1882
return null_init_tag_set(nullb->tag_set, nullb->dev->poll_queues);
drivers/block/null_blk/main.c
2008
nullb->disk = blk_mq_alloc_disk(nullb->tag_set, &lim, nullb);
drivers/block/null_blk/main.c
2069
if (nullb->tag_set == &nullb->__tag_set)
drivers/block/null_blk/main.c
2070
blk_mq_free_tag_set(nullb->tag_set);
drivers/block/null_blk/main.c
2212
if (tag_set.ops)
drivers/block/null_blk/main.c
2213
blk_mq_free_tag_set(&tag_set);
drivers/block/null_blk/main.c
407
set = dev->nullb->tag_set;
drivers/block/null_blk/main.c
72
static struct blk_mq_tag_set tag_set;
drivers/block/null_blk/null_blk.h
121
struct blk_mq_tag_set *tag_set;
drivers/block/ps3disk.c
32
struct blk_mq_tag_set tag_set;
drivers/block/ps3disk.c
441
error = blk_mq_alloc_sq_tag_set(&priv->tag_set, &ps3disk_mq_ops, 1, 0);
drivers/block/ps3disk.c
445
gendisk = blk_mq_alloc_disk(&priv->tag_set, &lim, dev);
drivers/block/ps3disk.c
478
blk_mq_free_tag_set(&priv->tag_set);
drivers/block/ps3disk.c
504
blk_mq_free_tag_set(&priv->tag_set);
drivers/block/rbd.c
440
struct blk_mq_tag_set tag_set;
drivers/block/rbd.c
4827
blk_mq_free_tag_set(&rbd_dev->tag_set);
drivers/block/rbd.c
4972
memset(&rbd_dev->tag_set, 0, sizeof(rbd_dev->tag_set));
drivers/block/rbd.c
4973
rbd_dev->tag_set.ops = &rbd_mq_ops;
drivers/block/rbd.c
4974
rbd_dev->tag_set.queue_depth = rbd_dev->opts->queue_depth;
drivers/block/rbd.c
4975
rbd_dev->tag_set.numa_node = NUMA_NO_NODE;
drivers/block/rbd.c
4976
rbd_dev->tag_set.nr_hw_queues = num_present_cpus();
drivers/block/rbd.c
4977
rbd_dev->tag_set.cmd_size = sizeof(struct rbd_img_request);
drivers/block/rbd.c
4979
err = blk_mq_alloc_tag_set(&rbd_dev->tag_set);
drivers/block/rbd.c
4992
disk = blk_mq_alloc_disk(&rbd_dev->tag_set, &lim, rbd_dev);
drivers/block/rbd.c
5012
blk_mq_free_tag_set(&rbd_dev->tag_set);
drivers/block/rnbd/rnbd-clt.c
1208
struct blk_mq_tag_set *tag_set = &sess->tag_set;
drivers/block/rnbd/rnbd-clt.c
1210
memset(tag_set, 0, sizeof(*tag_set));
drivers/block/rnbd/rnbd-clt.c
1211
tag_set->ops = &rnbd_mq_ops;
drivers/block/rnbd/rnbd-clt.c
1212
tag_set->queue_depth = sess->queue_depth;
drivers/block/rnbd/rnbd-clt.c
1213
tag_set->numa_node = NUMA_NO_NODE;
drivers/block/rnbd/rnbd-clt.c
1214
tag_set->flags = BLK_MQ_F_TAG_QUEUE_SHARED;
drivers/block/rnbd/rnbd-clt.c
1215
tag_set->cmd_size = sizeof(struct rnbd_iu) + RNBD_RDMA_SGL_SIZE;
drivers/block/rnbd/rnbd-clt.c
1218
tag_set->nr_maps = sess->nr_poll_queues ? HCTX_MAX_TYPES : 2;
drivers/block/rnbd/rnbd-clt.c
1223
tag_set->nr_hw_queues = num_online_cpus() + sess->nr_poll_queues;
drivers/block/rnbd/rnbd-clt.c
1224
tag_set->driver_data = sess;
drivers/block/rnbd/rnbd-clt.c
1226
return blk_mq_alloc_tag_set(tag_set);
drivers/block/rnbd/rnbd-clt.c
1395
dev->gd = blk_mq_alloc_disk(&dev->sess->tag_set, &lim, dev);
drivers/block/rnbd/rnbd-clt.c
739
if (sess->tag_set.tags)
drivers/block/rnbd/rnbd-clt.c
740
blk_mq_free_tag_set(&sess->tag_set);
drivers/block/rnbd/rnbd-clt.h
89
struct blk_mq_tag_set tag_set;
drivers/block/sunvdc.c
1078
blk_mq_free_tag_set(&port->tag_set);
drivers/block/sunvdc.c
830
err = blk_mq_alloc_sq_tag_set(&port->tag_set, &vdc_mq_ops,
drivers/block/sunvdc.c
835
g = blk_mq_alloc_disk(&port->tag_set, &lim, port);
drivers/block/sunvdc.c
86
struct blk_mq_tag_set tag_set;
drivers/block/sunvdc.c
891
blk_mq_free_tag_set(&port->tag_set);
drivers/block/swim.c
191
struct blk_mq_tag_set tag_set;
drivers/block/swim.c
785
blk_mq_free_tag_set(&fs->tag_set);
drivers/block/swim.c
820
err = blk_mq_alloc_sq_tag_set(&swd->unit[drive].tag_set,
drivers/block/swim.c
826
blk_mq_alloc_disk(&swd->unit[drive].tag_set, &lim,
drivers/block/swim.c
829
blk_mq_free_tag_set(&swd->unit[drive].tag_set);
drivers/block/swim3.c
1212
rc = blk_mq_alloc_sq_tag_set(&fs->tag_set, &swim3_mq_ops, 2, 0);
drivers/block/swim3.c
1216
disk = blk_mq_alloc_disk(&fs->tag_set, &lim, fs);
drivers/block/swim3.c
1245
blk_mq_free_tag_set(&fs->tag_set);
drivers/block/swim3.c
206
struct blk_mq_tag_set tag_set;
drivers/block/ublk_drv.c
1746
struct request *req = blk_mq_tag_to_rq(ub->tag_set.tags[ubq->q_id], tag);
drivers/block/ublk_drv.c
2606
ub->tag_set.tags[ubq->q_id], tag);
drivers/block/ublk_drv.c
2683
req = blk_mq_tag_to_rq(ub->tag_set.tags[ubq->q_id], tag);
drivers/block/ublk_drv.c
2847
blk_mq_tagset_busy_iter(&ub->tag_set,
drivers/block/ublk_drv.c
297
struct blk_mq_tag_set tag_set;
drivers/block/ublk_drv.c
3396
req = blk_mq_tag_to_rq(ub->tag_set.tags[q_id], tag);
drivers/block/ublk_drv.c
4085
if (ub->tag_set.map[HCTX_TYPE_DEFAULT].mq_map[cpu] == q_id)
drivers/block/ublk_drv.c
4203
blk_mq_free_tag_set(&ub->tag_set);
drivers/block/ublk_drv.c
4255
ub->tag_set.ops = &ublk_batch_mq_ops;
drivers/block/ublk_drv.c
4257
ub->tag_set.ops = &ublk_mq_ops;
drivers/block/ublk_drv.c
4258
ub->tag_set.nr_hw_queues = ub->dev_info.nr_hw_queues;
drivers/block/ublk_drv.c
4259
ub->tag_set.queue_depth = ub->dev_info.queue_depth;
drivers/block/ublk_drv.c
4260
ub->tag_set.numa_node = NUMA_NO_NODE;
drivers/block/ublk_drv.c
4261
ub->tag_set.driver_data = ub;
drivers/block/ublk_drv.c
4262
return blk_mq_alloc_tag_set(&ub->tag_set);
drivers/block/ublk_drv.c
4403
disk = blk_mq_alloc_disk(&ub->tag_set, &lim, NULL);
drivers/block/ublk_drv.c
4492
if (ub->tag_set.map[HCTX_TYPE_DEFAULT].mq_map[i] == queue)
drivers/block/ublk_drv.c
4700
blk_mq_free_tag_set(&ub->tag_set);
drivers/block/ublk_drv.c
5050
blk_mq_tagset_busy_iter(&ubq->dev->tag_set, ublk_count_busy_req, &data);
drivers/block/virtio_blk.c
1484
memset(&vblk->tag_set, 0, sizeof(vblk->tag_set));
drivers/block/virtio_blk.c
1485
vblk->tag_set.ops = &virtio_mq_ops;
drivers/block/virtio_blk.c
1486
vblk->tag_set.queue_depth = queue_depth;
drivers/block/virtio_blk.c
1487
vblk->tag_set.numa_node = NUMA_NO_NODE;
drivers/block/virtio_blk.c
1488
vblk->tag_set.cmd_size =
drivers/block/virtio_blk.c
1491
vblk->tag_set.driver_data = vblk;
drivers/block/virtio_blk.c
1492
vblk->tag_set.nr_hw_queues = vblk->num_vqs;
drivers/block/virtio_blk.c
1493
vblk->tag_set.nr_maps = 1;
drivers/block/virtio_blk.c
1495
vblk->tag_set.nr_maps = 3;
drivers/block/virtio_blk.c
1497
err = blk_mq_alloc_tag_set(&vblk->tag_set);
drivers/block/virtio_blk.c
1508
vblk->disk = blk_mq_alloc_disk(&vblk->tag_set, &lim, vblk);
drivers/block/virtio_blk.c
1550
blk_mq_free_tag_set(&vblk->tag_set);
drivers/block/virtio_blk.c
1570
blk_mq_free_tag_set(&vblk->tag_set);
drivers/block/virtio_blk.c
71
struct blk_mq_tag_set tag_set;
drivers/block/xen-blkfront.c
1120
memset(&info->tag_set, 0, sizeof(info->tag_set));
drivers/block/xen-blkfront.c
1121
info->tag_set.ops = &blkfront_mq_ops;
drivers/block/xen-blkfront.c
1122
info->tag_set.nr_hw_queues = info->nr_rings;
drivers/block/xen-blkfront.c
1130
info->tag_set.queue_depth = BLK_RING_SIZE(info) / 2;
drivers/block/xen-blkfront.c
1132
info->tag_set.queue_depth = BLK_RING_SIZE(info);
drivers/block/xen-blkfront.c
1133
info->tag_set.numa_node = NUMA_NO_NODE;
drivers/block/xen-blkfront.c
1134
info->tag_set.cmd_size = sizeof(struct blkif_req);
drivers/block/xen-blkfront.c
1135
info->tag_set.driver_data = info;
drivers/block/xen-blkfront.c
1137
err = blk_mq_alloc_tag_set(&info->tag_set);
drivers/block/xen-blkfront.c
1142
gd = blk_mq_alloc_disk(&info->tag_set, &lim, info);
drivers/block/xen-blkfront.c
1177
blk_mq_free_tag_set(&info->tag_set);
drivers/block/xen-blkfront.c
2111
blk_mq_update_nr_hw_queues(&info->tag_set, info->nr_rings);
drivers/block/xen-blkfront.c
226
struct blk_mq_tag_set tag_set;
drivers/block/xen-blkfront.c
2398
blk_mq_free_tag_set(&info->tag_set);
drivers/block/xen-blkfront.c
2483
blk_mq_free_tag_set(&info->tag_set);
drivers/block/z2ram.c
309
static struct blk_mq_tag_set tag_set;
drivers/block/z2ram.c
320
disk = blk_mq_alloc_disk(&tag_set, NULL, NULL);
drivers/block/z2ram.c
351
tag_set.ops = &z2_mq_ops;
drivers/block/z2ram.c
352
tag_set.nr_hw_queues = 1;
drivers/block/z2ram.c
353
tag_set.nr_maps = 1;
drivers/block/z2ram.c
354
tag_set.queue_depth = 16;
drivers/block/z2ram.c
355
tag_set.numa_node = NUMA_NO_NODE;
drivers/block/z2ram.c
356
ret = blk_mq_alloc_tag_set(&tag_set);
drivers/block/z2ram.c
369
blk_mq_free_tag_set(&tag_set);
drivers/block/z2ram.c
385
blk_mq_free_tag_set(&tag_set);
drivers/block/zloop.c
1092
zlo->tag_set.ops = &zloop_mq_ops;
drivers/block/zloop.c
1093
zlo->tag_set.nr_hw_queues = opts->nr_queues;
drivers/block/zloop.c
1094
zlo->tag_set.queue_depth = opts->queue_depth;
drivers/block/zloop.c
1095
zlo->tag_set.numa_node = NUMA_NO_NODE;
drivers/block/zloop.c
1096
zlo->tag_set.cmd_size = sizeof(struct zloop_cmd);
drivers/block/zloop.c
1097
zlo->tag_set.driver_data = zlo;
drivers/block/zloop.c
1099
ret = blk_mq_alloc_tag_set(&zlo->tag_set);
drivers/block/zloop.c
1105
zlo->disk = blk_mq_alloc_disk(&zlo->tag_set, &lim, zlo);
drivers/block/zloop.c
1145
blk_mq_free_tag_set(&zlo->tag_set);
drivers/block/zloop.c
115
struct blk_mq_tag_set tag_set;
drivers/block/zloop.c
809
blk_mq_free_tag_set(&zlo->tag_set);
drivers/cdrom/gdrom.c
104
struct blk_mq_tag_set tag_set;
drivers/cdrom/gdrom.c
779
err = blk_mq_alloc_sq_tag_set(&gd.tag_set, &gdrom_mq_ops, 1,
drivers/cdrom/gdrom.c
784
gd.disk = blk_mq_alloc_disk(&gd.tag_set, &lim, NULL);
drivers/cdrom/gdrom.c
825
blk_mq_free_tag_set(&gd.tag_set);
drivers/cdrom/gdrom.c
837
blk_mq_free_tag_set(&gd.tag_set);
drivers/md/dm-core.h
120
struct blk_mq_tag_set *tag_set;
drivers/md/dm-rq.c
550
md->tag_set = kzalloc_node(sizeof(struct blk_mq_tag_set), GFP_KERNEL, md->numa_node_id);
drivers/md/dm-rq.c
551
if (!md->tag_set)
drivers/md/dm-rq.c
554
md->tag_set->ops = &dm_mq_ops;
drivers/md/dm-rq.c
555
md->tag_set->queue_depth = dm_get_blk_mq_queue_depth();
drivers/md/dm-rq.c
556
md->tag_set->numa_node = md->numa_node_id;
drivers/md/dm-rq.c
557
md->tag_set->flags = BLK_MQ_F_STACKING;
drivers/md/dm-rq.c
558
md->tag_set->nr_hw_queues = dm_get_blk_mq_nr_hw_queues();
drivers/md/dm-rq.c
559
md->tag_set->driver_data = md;
drivers/md/dm-rq.c
561
md->tag_set->cmd_size = sizeof(struct dm_rq_target_io);
drivers/md/dm-rq.c
565
md->tag_set->cmd_size += immutable_tgt->per_io_data_size;
drivers/md/dm-rq.c
569
err = blk_mq_alloc_tag_set(md->tag_set);
drivers/md/dm-rq.c
573
err = blk_mq_init_allocated_queue(md->tag_set, md->queue);
drivers/md/dm-rq.c
579
blk_mq_free_tag_set(md->tag_set);
drivers/md/dm-rq.c
581
kfree(md->tag_set);
drivers/md/dm-rq.c
582
md->tag_set = NULL;
drivers/md/dm-rq.c
589
if (md->tag_set) {
drivers/md/dm-rq.c
590
blk_mq_free_tag_set(md->tag_set);
drivers/md/dm-rq.c
591
kfree(md->tag_set);
drivers/md/dm-rq.c
592
md->tag_set = NULL;
drivers/memstick/core/ms_block.c
2096
rc = blk_mq_alloc_sq_tag_set(&msb->tag_set, &msb_mq_ops, 2, 0);
drivers/memstick/core/ms_block.c
2100
msb->disk = blk_mq_alloc_disk(&msb->tag_set, &lim, card);
drivers/memstick/core/ms_block.c
2140
blk_mq_free_tag_set(&msb->tag_set);
drivers/memstick/core/ms_block.c
2196
blk_mq_free_tag_set(&msb->tag_set);
drivers/memstick/core/ms_block.h
150
struct blk_mq_tag_set tag_set;
drivers/memstick/core/mspro_block.c
1138
rc = blk_mq_alloc_sq_tag_set(&msb->tag_set, &mspro_mq_ops, 2, 0);
drivers/memstick/core/mspro_block.c
1142
msb->disk = blk_mq_alloc_disk(&msb->tag_set, &lim, card);
drivers/memstick/core/mspro_block.c
1175
blk_mq_free_tag_set(&msb->tag_set);
drivers/memstick/core/mspro_block.c
1258
blk_mq_free_tag_set(&msb->tag_set);
drivers/memstick/core/mspro_block.c
140
struct blk_mq_tag_set tag_set;
drivers/mmc/core/block.c
2649
blk_mq_free_tag_set(&md->queue.tag_set);
drivers/mmc/core/queue.c
384
disk = blk_mq_alloc_disk(&mq->tag_set, &lim, mq);
drivers/mmc/core/queue.c
432
memset(&mq->tag_set, 0, sizeof(mq->tag_set));
drivers/mmc/core/queue.c
433
mq->tag_set.ops = &mmc_mq_ops;
drivers/mmc/core/queue.c
439
mq->tag_set.queue_depth =
drivers/mmc/core/queue.c
442
mq->tag_set.queue_depth = MMC_QUEUE_DEPTH;
drivers/mmc/core/queue.c
443
mq->tag_set.numa_node = NUMA_NO_NODE;
drivers/mmc/core/queue.c
444
mq->tag_set.flags = BLK_MQ_F_BLOCKING;
drivers/mmc/core/queue.c
445
mq->tag_set.nr_hw_queues = 1;
drivers/mmc/core/queue.c
446
mq->tag_set.cmd_size = sizeof(struct mmc_queue_req);
drivers/mmc/core/queue.c
447
mq->tag_set.driver_data = mq;
drivers/mmc/core/queue.c
461
ret = blk_mq_alloc_tag_set(&mq->tag_set);
drivers/mmc/core/queue.c
468
blk_mq_free_tag_set(&mq->tag_set);
drivers/mmc/core/queue.c
507
blk_mq_free_tag_set(&mq->tag_set);
drivers/mmc/core/queue.h
77
struct blk_mq_tag_set tag_set;
drivers/mtd/mtd_blkdevs.c
327
new->tag_set = kzalloc_obj(*new->tag_set);
drivers/mtd/mtd_blkdevs.c
328
if (!new->tag_set)
drivers/mtd/mtd_blkdevs.c
33
blk_mq_free_tag_set(dev->tag_set);
drivers/mtd/mtd_blkdevs.c
331
ret = blk_mq_alloc_sq_tag_set(new->tag_set, &mtd_mq_ops, 2,
drivers/mtd/mtd_blkdevs.c
34
kfree(dev->tag_set);
drivers/mtd/mtd_blkdevs.c
343
gd = blk_mq_alloc_disk(new->tag_set, &lim, new);
drivers/mtd/mtd_blkdevs.c
396
blk_mq_free_tag_set(new->tag_set);
drivers/mtd/mtd_blkdevs.c
398
kfree(new->tag_set);
drivers/mtd/ubi/block.c
383
dev->tag_set.ops = &ubiblock_mq_ops;
drivers/mtd/ubi/block.c
384
dev->tag_set.queue_depth = 64;
drivers/mtd/ubi/block.c
385
dev->tag_set.numa_node = NUMA_NO_NODE;
drivers/mtd/ubi/block.c
386
dev->tag_set.flags = BLK_MQ_F_BLOCKING;
drivers/mtd/ubi/block.c
387
dev->tag_set.cmd_size = sizeof(struct ubiblock_pdu);
drivers/mtd/ubi/block.c
388
dev->tag_set.driver_data = dev;
drivers/mtd/ubi/block.c
389
dev->tag_set.nr_hw_queues = 1;
drivers/mtd/ubi/block.c
391
ret = blk_mq_alloc_tag_set(&dev->tag_set);
drivers/mtd/ubi/block.c
400
gd = blk_mq_alloc_disk(&dev->tag_set, &lim, dev);
drivers/mtd/ubi/block.c
442
blk_mq_free_tag_set(&dev->tag_set);
drivers/mtd/ubi/block.c
460
blk_mq_free_tag_set(&dev->tag_set);
drivers/mtd/ubi/block.c
85
struct blk_mq_tag_set tag_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1070
curr_fc->tag_set = &hws_definer_l3_type_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1218
curr_fc->tag_set = &hws_definer_outer_second_vlan_type_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1226
curr_fc->tag_set = &hws_definer_inner_second_vlan_type_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1260
curr_fc->tag_set = HWS_IS_FLD_SET(match_param,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
133
(fc)->tag_set = &hws_definer_generic_set; \
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1449
curr_fc->tag_set = &hws_definer_icmp_dw1_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1470
curr_fc->tag_set = &hws_definer_icmpv6_dw1_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1495
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1510
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1526
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1542
curr_fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1557
curr_fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1572
curr_fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1678
if (fc[i].tag_set)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1700
if (!fc[i].tag_set)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1720
if (!fc[i].tag_set)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
2045
fc->tag_set(fc, match_param, tag);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
2058
fc->tag_set(fc, match_param, tag);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
557
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
562
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
567
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
572
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
577
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
582
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
587
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
592
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
612
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
617
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
622
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
627
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
632
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
637
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
642
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
647
fc->tag_set = &hws_definer_generic_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
781
curr_fc->tag_set = &hws_definer_outer_vlan_type_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
889
curr_fc->tag_set = &hws_definer_l3_type_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
958
curr_fc->tag_set = &hws_definer_inner_vlan_type_set;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.h
272
void (*tag_set)(struct mlx5hws_definer_fc *fc,
drivers/nvme/host/core.c
5288
blk_mq_wait_quiesce_done(ctrl->admin_q->tag_set);
drivers/nvme/host/fc.c
167
struct blk_mq_tag_set tag_set;
drivers/nvme/host/fc.c
2116
int queue_idx = (set == &ctrl->tag_set) ? hctx_idx + 1 : 0;
drivers/nvme/host/fc.c
2467
blk_mq_tagset_busy_iter(&ctrl->tag_set,
drivers/nvme/host/fc.c
2469
blk_mq_tagset_wait_completed_request(&ctrl->tag_set);
drivers/nvme/host/fc.c
2881
ret = nvme_alloc_io_tag_set(&ctrl->ctrl, &ctrl->tag_set,
drivers/nvme/host/fc.c
2945
blk_mq_update_nr_hw_queues(&ctrl->tag_set, nr_io_queues);
drivers/nvme/host/rdma.c
108
struct blk_mq_tag_set tag_set;
drivers/nvme/host/rdma.c
1654
return queue->ctrl->tag_set.tags[queue_idx - 1];
drivers/nvme/host/rdma.c
299
int queue_idx = (set == &ctrl->tag_set) ? hctx_idx + 1 : 0;
drivers/nvme/host/rdma.c
763
return nvme_alloc_io_tag_set(ctrl, &to_rdma_ctrl(ctrl)->tag_set,
drivers/nvme/host/rdma.c
883
nr_queues = min(ctrl->tag_set.nr_hw_queues + 1, ctrl->ctrl.queue_count);
drivers/nvme/host/rdma.c
911
ctrl->tag_set.nr_hw_queues + 1);
drivers/nvme/host/tcp.c
1640
struct blk_mq_tag_set *set = &ctrl->tag_set;
drivers/nvme/host/tcp.c
188
struct blk_mq_tag_set tag_set;
drivers/nvme/host/tcp.c
2176
ret = nvme_alloc_io_tag_set(ctrl, &to_tcp_ctrl(ctrl)->tag_set,
drivers/nvme/host/tcp.c
261
return queue->ctrl->tag_set.tags[queue_idx - 1];
drivers/nvme/host/tcp.c
556
int queue_idx = (set == &ctrl->tag_set) ? hctx_idx + 1 : 0;
drivers/nvme/target/loop.c
213
(set == &ctrl->tag_set) ? hctx_idx + 1 : 0);
drivers/nvme/target/loop.c
35
struct blk_mq_tag_set tag_set;
drivers/nvme/target/loop.c
482
blk_mq_update_nr_hw_queues(&ctrl->tag_set,
drivers/nvme/target/loop.c
525
ret = nvme_alloc_io_tag_set(&ctrl->ctrl, &ctrl->tag_set,
drivers/nvme/target/loop.c
90
return queue->ctrl->tag_set.tags[queue_idx - 1];
drivers/s390/block/dasd_genhd.c
100
blk_mq_free_tag_set(&block->tag_set);
drivers/s390/block/dasd_genhd.c
152
blk_mq_free_tag_set(&block->tag_set);
drivers/s390/block/dasd_genhd.c
89
block->tag_set.ops = &dasd_mq_ops;
drivers/s390/block/dasd_genhd.c
90
block->tag_set.cmd_size = sizeof(struct dasd_ccw_req);
drivers/s390/block/dasd_genhd.c
91
block->tag_set.nr_hw_queues = nr_hw_queues;
drivers/s390/block/dasd_genhd.c
92
block->tag_set.queue_depth = queue_depth;
drivers/s390/block/dasd_genhd.c
93
block->tag_set.numa_node = NUMA_NO_NODE;
drivers/s390/block/dasd_genhd.c
94
rc = blk_mq_alloc_tag_set(&block->tag_set);
drivers/s390/block/dasd_genhd.c
98
gdp = blk_mq_alloc_disk(&block->tag_set, &lim, block);
drivers/s390/block/dasd_int.h
621
struct blk_mq_tag_set tag_set;
drivers/s390/block/scm_blk.c
458
bdev->tag_set.ops = &scm_mq_ops;
drivers/s390/block/scm_blk.c
459
bdev->tag_set.cmd_size = sizeof(blk_status_t);
drivers/s390/block/scm_blk.c
460
bdev->tag_set.nr_hw_queues = nr_requests;
drivers/s390/block/scm_blk.c
461
bdev->tag_set.queue_depth = nr_requests_per_io * nr_requests;
drivers/s390/block/scm_blk.c
462
bdev->tag_set.numa_node = NUMA_NO_NODE;
drivers/s390/block/scm_blk.c
464
ret = blk_mq_alloc_tag_set(&bdev->tag_set);
drivers/s390/block/scm_blk.c
468
bdev->gendisk = blk_mq_alloc_disk(&bdev->tag_set, &lim, scmdev);
drivers/s390/block/scm_blk.c
500
blk_mq_free_tag_set(&bdev->tag_set);
drivers/s390/block/scm_blk.c
510
blk_mq_free_tag_set(&bdev->tag_set);
drivers/s390/block/scm_blk.h
20
struct blk_mq_tag_set tag_set;
drivers/scsi/fnic/fnic_main.c
682
struct blk_mq_queue_map *qmap = &host->tag_set.map[HCTX_TYPE_DEFAULT];
drivers/scsi/hisi_sas/hisi_sas_main.c
579
struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT];
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3557
struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT];
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
3372
qmap = &shost->tag_set.map[i];
drivers/scsi/hosts.c
629
blk_mq_tagset_busy_iter(&shost->tag_set,
drivers/scsi/hosts.c
730
blk_mq_tagset_busy_iter(&shost->tag_set, complete_all_cmds_iter,
drivers/scsi/hosts.c
766
blk_mq_tagset_busy_iter(&shost->tag_set, __scsi_host_busy_iter_fn,
drivers/scsi/libsas/sas_discover.c
387
blk_mq_tagset_busy_iter(&shost->tag_set, sas_abort_cmd, dev);
drivers/scsi/megaraid/megaraid_sas_base.c
3194
map = &shost->tag_set.map[HCTX_TYPE_DEFAULT];
drivers/scsi/megaraid/megaraid_sas_base.c
3202
shost->tag_set.map[HCTX_TYPE_READ].nr_queues = 0;
drivers/scsi/megaraid/megaraid_sas_base.c
3205
map = &shost->tag_set.map[HCTX_TYPE_POLL];
drivers/scsi/mpi3mr/mpi3mr_os.c
4106
blk_mq_tagset_busy_iter(&mrioc->shost->tag_set,
drivers/scsi/mpi3mr/mpi3mr_os.c
4114
blk_mq_tagset_busy_iter(&mrioc->shost->tag_set,
drivers/scsi/mpi3mr/mpi3mr_os.c
4191
map = &shost->tag_set.map[i];
drivers/scsi/mpi3mr/mpi3mr_os.c
4253
blk_mq_tagset_busy_iter(&shost->tag_set,
drivers/scsi/mpi3mr/mpi3mr_os.c
628
blk_mq_tagset_busy_iter(&shost->tag_set,
drivers/scsi/mpi3mr/mpi3mr_os.c
661
blk_mq_tagset_busy_iter(&shost->tag_set,
drivers/scsi/mpt3sas/mpt3sas_scsih.c
13152
map = &shost->tag_set.map[i];
drivers/scsi/pm8001/pm8001_init.c
105
struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT];
drivers/scsi/qla2xxx/qla_os.c
8079
struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT];
drivers/scsi/scsi_debug.c
6825
blk_mq_tagset_busy_iter(&shost->tag_set, sdebug_stop_cmnd, NULL);
drivers/scsi/scsi_debug.c
6900
blk_mq_tagset_busy_iter(&shost->tag_set,
drivers/scsi/scsi_debug.c
7605
blk_mq_tagset_busy_iter(&host->tag_set, sdebug_submit_queue_iter,
drivers/scsi/scsi_debug.c
9067
struct blk_mq_queue_map *map = &shost->tag_set.map[i];
drivers/scsi/scsi_debug.c
9151
blk_mq_tagset_busy_iter(&shost->tag_set, sdebug_blk_mq_poll_iter,
drivers/scsi/scsi_lib.c
2012
struct Scsi_Host *shost = container_of(set, struct Scsi_Host, tag_set);
drivers/scsi/scsi_lib.c
2106
struct blk_mq_tag_set *tag_set = &shost->tag_set;
drivers/scsi/scsi_lib.c
2115
memset(tag_set, 0, sizeof(*tag_set));
drivers/scsi/scsi_lib.c
2117
tag_set->ops = &scsi_mq_ops;
drivers/scsi/scsi_lib.c
2119
tag_set->ops = &scsi_mq_ops_no_commit;
drivers/scsi/scsi_lib.c
2120
tag_set->nr_hw_queues = shost->nr_hw_queues ? : 1;
drivers/scsi/scsi_lib.c
2121
tag_set->nr_maps = shost->nr_maps ? : 1;
drivers/scsi/scsi_lib.c
2122
tag_set->queue_depth = shost->can_queue + shost->nr_reserved_cmds;
drivers/scsi/scsi_lib.c
2123
tag_set->reserved_tags = shost->nr_reserved_cmds;
drivers/scsi/scsi_lib.c
2124
tag_set->cmd_size = cmd_size;
drivers/scsi/scsi_lib.c
2125
tag_set->numa_node = dev_to_node(shost->dma_dev);
drivers/scsi/scsi_lib.c
2127
tag_set->flags |= BLK_MQ_F_TAG_RR;
drivers/scsi/scsi_lib.c
2129
tag_set->flags |= BLK_MQ_F_BLOCKING;
drivers/scsi/scsi_lib.c
2130
tag_set->driver_data = shost;
drivers/scsi/scsi_lib.c
2132
tag_set->flags |= BLK_MQ_F_TAG_HCTX_SHARED;
drivers/scsi/scsi_lib.c
2134
return blk_mq_alloc_tag_set(tag_set);
drivers/scsi/scsi_lib.c
2142
blk_mq_free_tag_set(&shost->tag_set);
drivers/scsi/scsi_lib.c
3106
blk_mq_wait_quiesce_done(&shost->tag_set);
drivers/scsi/scsi_lib.c
3166
blk_mq_wait_quiesce_done(&shost->tag_set);
drivers/scsi/scsi_scan.c
339
q = blk_mq_alloc_queue(&sdev->host->tag_set, &lim, sdev);
drivers/scsi/scsi_sysfs.c
400
struct blk_mq_tag_set *tag_set = &shost->tag_set;
drivers/scsi/scsi_sysfs.c
402
return snprintf(buf, 20, "%d\n", tag_set->nr_hw_queues);
drivers/scsi/smartpqi/smartpqi_init.c
6600
blk_mq_map_hw_queues(&shost->tag_set.map[HCTX_TYPE_DEFAULT],
drivers/scsi/smartpqi/smartpqi_init.c
6603
blk_mq_map_queues(&shost->tag_set.map[HCTX_TYPE_DEFAULT]);
drivers/scsi/virtio_scsi.c
738
struct blk_mq_queue_map *map = &shost->tag_set.map[i];
drivers/target/loopback/tcm_loop.c
330
blk_mq_tagset_busy_iter(&sh->tag_set, tcm_loop_flush_work_iter, NULL);
drivers/ufs/core/ufshcd-priv.h
377
struct blk_mq_tags *tags = hba->host->tag_set.shared_tags;
drivers/ufs/core/ufshcd.c
1438
blk_mq_quiesce_tagset(&hba->host->tag_set);
drivers/ufs/core/ufshcd.c
1447
blk_mq_unquiesce_tagset(&hba->host->tag_set);
drivers/ufs/core/ufshcd.c
1463
blk_mq_unquiesce_tagset(&hba->host->tag_set);
drivers/ufs/core/ufshcd.c
3008
struct blk_mq_queue_map *map = &shost->tag_set.map[i];
drivers/ufs/core/ufshcd.c
5788
blk_mq_tagset_busy_iter(&hba->host->tag_set,
drivers/ufs/core/ufshcd.c
6556
blk_mq_quiesce_tagset(&hba->host->tag_set);
drivers/ufs/core/ufshcd.c
6562
blk_mq_unquiesce_tagset(&hba->host->tag_set);
drivers/ufs/core/ufshcd.c
663
blk_mq_tagset_busy_iter(&hba->host->tag_set, ufshcd_print_tr_iter, &pr_prdt);
drivers/ufs/core/ufshcd.c
6678
blk_mq_tagset_busy_iter(&hba->host->tag_set, ufshcd_abort_one, &ret);
drivers/ufs/core/ufshcd.c
7683
blk_mq_tagset_busy_iter(&hba->host->tag_set,
drivers/ufs/host/ufs-mediatek.c
847
struct blk_mq_tag_set *tag_set = &hba->host->tag_set;
drivers/ufs/host/ufs-mediatek.c
848
struct blk_mq_queue_map *map = &tag_set->map[HCTX_TYPE_DEFAULT];
include/linux/blkdev.h
640
struct blk_mq_tag_set *tag_set;
include/linux/mtd/blktrans.h
34
struct blk_mq_tag_set *tag_set;
include/scsi/scsi_host.h
589
struct blk_mq_tag_set tag_set;
include/scsi/scsi_tcq.h
32
if (hwq < shost->tag_set.nr_hw_queues) {
include/scsi/scsi_tcq.h
33
req = blk_mq_tag_to_rq(shost->tag_set.tags[hwq],
lib/radix-tree.c
440
tag_set(node, tag, 0);
lib/radix-tree.c
945
tag_set(node, tag, offset);
lib/radix-tree.c
984
tag_set(parent, tag, offset);