bdev_get_queue
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
trace_block_bio_complete(bdev_get_queue(bio->bi_bdev), bio);
blkg_get(bdev_get_queue(bio->bi_bdev)->root_blkg);
bio->bi_blkg = bdev_get_queue(bio->bi_bdev)->root_blkg;
spin_unlock_irq(&bdev_get_queue(ctx->bdev)->queue_lock);
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
if (q == bdev_get_queue(bio->bi_bdev))
struct request_queue *q = bdev_get_queue(bdev);
q = bdev_get_queue(bdev);
bdev_get_queue(bdev)->crypto_profile;
return __blk_crypto_cfg_supported(bdev_get_queue(bdev)->crypto_profile,
struct request_queue *q = bdev_get_queue(bdev);
bdev_get_queue(bdev)->crypto_profile;
q = bdev_get_queue(ctx.bdev);
blk_mq_queue_tag_busy_iter(bdev_get_queue(part), blk_mq_check_in_driver,
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
q = bdev_get_queue(bio->bi_bdev);
struct request_queue *q = bdev_get_queue(bdev);
struct request_queue *q = bdev_get_queue(bdev);
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
struct request_queue *q = bdev_get_queue(bio->bi_bdev);
struct request_queue *q = bdev_get_queue(bdev);
queue_max_sectors(bdev_get_queue(bdev)));
struct request_queue *q = bdev_get_queue(bdev);
cpu_to_le32(queue_max_hw_sectors(bdev_get_queue(bdev)));
q = bdev_get_queue(dc->bdev);
struct request_queue *q = bdev_get_queue(dc->bdev);
struct request_queue *q = bdev_get_queue(where->bdev);
struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev);
struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev);
q = bdev_get_queue(bdev);
struct request_queue *q = bdev_get_queue(bdev);
q = bdev_get_queue(p->path.dev->bdev);
bdev_get_queue(bdev)->crypto_profile;
bdev_get_queue(dev->bdev)->crypto_profile;
struct request_queue *q = bdev_get_queue(dev->bdev);
struct request_queue *q = bdev_get_queue(bdev);
struct request_queue *q = bdev_get_queue(bdev);
if (blk_queue_dying(bdev_get_queue(dmz_dev->bdev))) {
struct request_queue *q = bdev_get_queue(bdev);
q = bdev_get_queue(bd);
&bdev_get_queue(chunk->bio.bi_bdev)->limits;
return &bdev_get_queue(bdev)->limits;
return queue_emulates_zone_append(bdev_get_queue(bdev));
return queue_max_segments(bdev_get_queue(bdev));
return queue_logical_block_size(bdev_get_queue(bdev));
return queue_physical_block_size(bdev_get_queue(bdev));
return queue_io_min(bdev_get_queue(bdev));
return queue_io_opt(bdev_get_queue(bdev));
return queue_zone_write_granularity(bdev_get_queue(bdev));
return blk_queue_rot(bdev_get_queue(bdev));
struct request_queue *q = bdev_get_queue(bdev);
return blk_queue_write_cache(bdev_get_queue(bdev));
return blk_queue_is_zoned(bdev_get_queue(bdev));
struct request_queue *q = bdev_get_queue(bdev);
return queue_dma_alignment(bdev_get_queue(bdev));
return queue_atomic_write_unit_min_bytes(bdev_get_queue(bdev));
return queue_atomic_write_unit_max_bytes(bdev_get_queue(bdev));
struct request_queue *q = bdev_get_queue(bdev);
struct request_queue *q = bdev_get_queue(bdev);
struct request_queue *q = bdev_get_queue(bdev);