/linux/include/linux/ ! |
H A D | blkdev.h | 1020 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue() function 1290 return &bdev_get_queue(bdev)->limits; in bdev_limits() 1340 return queue_emulates_zone_append(bdev_get_queue(bdev)); in bdev_emulates_zone_append() 1351 return queue_max_segments(bdev_get_queue(bdev)); in bdev_max_segments() 1368 return queue_logical_block_size(bdev_get_queue(bdev)); in bdev_logical_block_size() 1378 return queue_physical_block_size(bdev_get_queue(bdev)); in bdev_physical_block_size() 1388 return queue_io_min(bdev_get_queue(bdev)); in bdev_io_min() 1398 return queue_io_opt(bdev_get_queue(bdev)); in bdev_io_opt() 1410 return queue_zone_write_granularity(bdev_get_queue(bdev)); in bdev_zone_write_granularity() 1445 return blk_queue_nonrot(bdev_get_queue(bdev)); in bdev_nonrot() [all …]
|
/linux/block/ ! |
H A D | blk-crypto.c | 388 return __blk_crypto_cfg_supported(bdev_get_queue(bdev)->crypto_profile, in blk_crypto_config_supported_natively() 453 struct request_queue *q = bdev_get_queue(bdev); in blk_crypto_evict_key() 599 bdev_get_queue(bdev)->crypto_profile; in blk_crypto_ioctl()
|
H A D | blk-core.c | 681 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in __submit_bio_noacct() 699 if (q == bdev_get_queue(bio->bi_bdev)) in __submit_bio_noacct() 780 struct request_queue *q = bdev_get_queue(bdev); in submit_bio_noacct() 943 q = bdev_get_queue(bdev); in bio_poll()
|
H A D | blk-rq-qos.h | 144 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in rq_qos_done_bio()
|
H A D | blk-settings.c | 1011 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset() 1024 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment()
|
H A D | blk-crypto-profile.c | 489 bdev_get_queue(bdev)->crypto_profile; in blk_crypto_derive_sw_secret()
|
H A D | blk-cgroup.c | 997 spin_unlock_irq(&bdev_get_queue(ctx->bdev)->queue_lock); in blkg_conf_exit() 2150 blkg_get(bdev_get_queue(bio->bi_bdev)->root_blkg); in bio_associate_blkg_from_css() 2151 bio->bi_blkg = bdev_get_queue(bio->bi_bdev)->root_blkg; in bio_associate_blkg_from_css()
|
H A D | blk.h | 85 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_queue_enter()
|
H A D | ioctl.c | 610 queue_max_sectors(bdev_get_queue(bdev))); in blkdev_common_ioctl()
|
H A D | genhd.c | 1115 struct request_queue *q = bdev_get_queue(bdev); in part_inflight_show()
|
H A D | bio.c | 1621 trace_block_bio_complete(bdev_get_queue(bio->bi_bdev), bio); in bio_endio()
|
H A D | blk-throttle.c | 1753 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in __blk_throtl_bio()
|
H A D | blk-mq.c | 108 blk_mq_queue_tag_busy_iter(bdev_get_queue(part), blk_mq_check_in_driver, in blk_mq_in_driver_rw() 3113 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in blk_mq_submit_bio()
|
H A D | blk-iocost.c | 3427 q = bdev_get_queue(ctx.bdev); in ioc_cost_model_write()
|
/linux/drivers/md/ ! |
H A D | dm-table.c | 425 struct request_queue *q = bdev_get_queue(bdev); in dm_set_device_limits() 906 struct request_queue *q = bdev_get_queue(bdev); in device_is_not_rq_stackable() 1237 bdev_get_queue(bdev)->crypto_profile; in dm_wrappedkey_op_callback() 1375 bdev_get_queue(dev->bdev)->crypto_profile; in device_intersect_crypto_capabilities() 1919 struct request_queue *q = bdev_get_queue(dev->bdev); in device_not_write_zeroes_capable()
|
H A D | dm-mpath.c | 544 q = bdev_get_queue(bdev); in multipath_clone_and_map() 887 struct request_queue *q = bdev_get_queue(bdev); in setup_scsi_dh() 963 q = bdev_get_queue(p->path.dev->bdev); in parse_path() 1630 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in activate_or_offline_path() 2221 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in pgpath_busy()
|
H A D | dm-io.c | 316 struct request_queue *q = bdev_get_queue(where->bdev); in do_region()
|
H A D | dm-zoned-target.c | 589 if (blk_queue_dying(bdev_get_queue(dmz_dev->bdev))) { in dmz_bdev_is_dying()
|
/linux/kernel/trace/ ! |
H A D | blktrace.c | 705 struct request_queue *q = bdev_get_queue(bdev); in blk_trace_ioctl() 1728 struct request_queue *q = bdev_get_queue(bdev); in sysfs_blk_trace_attr_show() 1762 struct request_queue *q = bdev_get_queue(bdev); in sysfs_blk_trace_attr_store()
|
/linux/drivers/block/rnbd/ ! |
H A D | rnbd-srv.c | 548 cpu_to_le32(queue_max_hw_sectors(bdev_get_queue(bdev))); in rnbd_srv_fill_msg_open_rsp()
|
/linux/drivers/target/ ! |
H A D | target_core_iblock.c | 134 q = bdev_get_queue(bd); in iblock_configure_device()
|
/linux/fs/xfs/ ! |
H A D | xfs_zone_gc.c | 754 &bdev_get_queue(chunk->bio.bi_bdev)->limits; in xfs_zone_gc_split_write()
|
/linux/drivers/md/bcache/ ! |
H A D | super.c | 1016 q = bdev_get_queue(dc->bdev); in cached_dev_status_update() 1398 struct request_queue *q = bdev_get_queue(dc->bdev); in cached_dev_init()
|
/linux/drivers/block/drbd/ ! |
H A D | drbd_main.c | 918 struct request_queue *q = bdev_get_queue(bdev); in drbd_send_sizes()
|