Symbol: bdev_get_queue
block/bio-integrity.c
164
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
block/bio-integrity.c
307
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
block/bio.c
1762
trace_block_bio_complete(bdev_get_queue(bio->bi_bdev), bio);
block/blk-cgroup.c
2132
blkg_get(bdev_get_queue(bio->bi_bdev)->root_blkg);
block/blk-cgroup.c
2133
bio->bi_blkg = bdev_get_queue(bio->bi_bdev)->root_blkg;
block/blk-cgroup.c
981
spin_unlock_irq(&bdev_get_queue(ctx->bdev)->queue_lock);
block/blk-core.c
681
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
block/blk-core.c
699
if (q == bdev_get_queue(bio->bi_bdev))
block/blk-core.c
783
struct request_queue *q = bdev_get_queue(bdev);
block/blk-core.c
953
q = bdev_get_queue(bdev);
block/blk-crypto-profile.c
487
bdev_get_queue(bdev)->crypto_profile;
block/blk-crypto.c
356
return __blk_crypto_cfg_supported(bdev_get_queue(bdev)->crypto_profile,
block/blk-crypto.c
421
struct request_queue *q = bdev_get_queue(bdev);
block/blk-crypto.c
567
bdev_get_queue(bdev)->crypto_profile;
block/blk-iocost.c
3425
q = bdev_get_queue(ctx.bdev);
block/blk-mq.c
109
blk_mq_queue_tag_busy_iter(bdev_get_queue(part), blk_mq_check_in_driver,
block/blk-mq.c
3143
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
block/blk-rq-qos.h
146
q = bdev_get_queue(bio->bi_bdev);
block/blk-settings.c
1042
struct request_queue *q = bdev_get_queue(bdev);
block/blk-settings.c
1055
struct request_queue *q = bdev_get_queue(bdev);
block/blk-throttle.c
1732
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
block/blk.h
86
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
block/genhd.c
1115
struct request_queue *q = bdev_get_queue(bdev);
block/ioctl.c
704
queue_max_sectors(bdev_get_queue(bdev)));
drivers/block/drbd/drbd_main.c
912
struct request_queue *q = bdev_get_queue(bdev);
drivers/block/rnbd/rnbd-srv.c
563
cpu_to_le32(queue_max_hw_sectors(bdev_get_queue(bdev)));
drivers/md/bcache/super.c
1024
q = bdev_get_queue(dc->bdev);
drivers/md/bcache/super.c
1406
struct request_queue *q = bdev_get_queue(dc->bdev);
drivers/md/dm-io.c
316
struct request_queue *q = bdev_get_queue(where->bdev);
drivers/md/dm-mpath.c
1631
struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev);
drivers/md/dm-mpath.c
2219
struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev);
drivers/md/dm-mpath.c
538
q = bdev_get_queue(bdev);
drivers/md/dm-mpath.c
881
struct request_queue *q = bdev_get_queue(bdev);
drivers/md/dm-mpath.c
951
q = bdev_get_queue(p->path.dev->bdev);
drivers/md/dm-table.c
1221
bdev_get_queue(bdev)->crypto_profile;
drivers/md/dm-table.c
1353
bdev_get_queue(dev->bdev)->crypto_profile;
drivers/md/dm-table.c
1897
struct request_queue *q = bdev_get_queue(dev->bdev);
drivers/md/dm-table.c
415
struct request_queue *q = bdev_get_queue(bdev);
drivers/md/dm-table.c
890
struct request_queue *q = bdev_get_queue(bdev);
drivers/md/dm-zoned-target.c
589
if (blk_queue_dying(bdev_get_queue(dmz_dev->bdev))) {
drivers/target/target_core_device.c
847
struct request_queue *q = bdev_get_queue(bdev);
drivers/target/target_core_iblock.c
133
q = bdev_get_queue(bd);
fs/xfs/xfs_zone_gc.c
755
&bdev_get_queue(chunk->bio.bi_bdev)->limits;
include/linux/blkdev.h
1312
return &bdev_get_queue(bdev)->limits;
include/linux/blkdev.h
1362
return queue_emulates_zone_append(bdev_get_queue(bdev));
include/linux/blkdev.h
1373
return queue_max_segments(bdev_get_queue(bdev));
include/linux/blkdev.h
1390
return queue_logical_block_size(bdev_get_queue(bdev));
include/linux/blkdev.h
1400
return queue_physical_block_size(bdev_get_queue(bdev));
include/linux/blkdev.h
1410
return queue_io_min(bdev_get_queue(bdev));
include/linux/blkdev.h
1420
return queue_io_opt(bdev_get_queue(bdev));
include/linux/blkdev.h
1432
return queue_zone_write_granularity(bdev_get_queue(bdev));
include/linux/blkdev.h
1467
return blk_queue_rot(bdev_get_queue(bdev));
include/linux/blkdev.h
1482
struct request_queue *q = bdev_get_queue(bdev);
include/linux/blkdev.h
1498
return blk_queue_write_cache(bdev_get_queue(bdev));
include/linux/blkdev.h
1513
return blk_queue_is_zoned(bdev_get_queue(bdev));
include/linux/blkdev.h
1523
struct request_queue *q = bdev_get_queue(bdev);
include/linux/blkdev.h
1595
return queue_dma_alignment(bdev_get_queue(bdev));
include/linux/blkdev.h
1862
return queue_atomic_write_unit_min_bytes(bdev_get_queue(bdev));
include/linux/blkdev.h
1870
return queue_atomic_write_unit_max_bytes(bdev_get_queue(bdev));
kernel/trace/blktrace.c
2042
struct request_queue *q = bdev_get_queue(bdev);
kernel/trace/blktrace.c
2076
struct request_queue *q = bdev_get_queue(bdev);
kernel/trace/blktrace.c
926
struct request_queue *q = bdev_get_queue(bdev);