Searched refs:bdev_get_queue (Results 1 – 25 of 26) sorted by relevance
12
1042 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue() function1312 return &bdev_get_queue(bdev)->limits; in bdev_limits()1362 return queue_emulates_zone_append(bdev_get_queue(bdev)); in bdev_emulates_zone_append()1373 return queue_max_segments(bdev_get_queue(bdev)); in bdev_max_segments()1390 return queue_logical_block_size(bdev_get_queue(bdev)); in bdev_logical_block_size()1400 return queue_physical_block_size(bdev_get_queue(bdev)); in bdev_physical_block_size()1410 return queue_io_min(bdev_get_queue(bdev)); in bdev_io_min()1420 return queue_io_opt(bdev_get_queue(bdev)); in bdev_io_opt()1432 return queue_zone_write_granularity(bdev_get_queue(bdev)); in bdev_zone_write_granularity()1467 return blk_queue_rot(bdev_get_queue(bdev)); in bdev_rot()[all …]
356 return __blk_crypto_cfg_supported(bdev_get_queue(bdev)->crypto_profile, in blk_crypto_config_supported_natively()421 struct request_queue *q = bdev_get_queue(bdev); in blk_crypto_evict_key()567 bdev_get_queue(bdev)->crypto_profile; in blk_crypto_ioctl()
681 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in __submit_bio_noacct()699 if (q == bdev_get_queue(bio->bi_bdev)) in __submit_bio_noacct()783 struct request_queue *q = bdev_get_queue(bdev); in submit_bio_noacct()953 q = bdev_get_queue(bdev); in bio_poll()
164 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_integrity_add_page()307 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_integrity_map_user()
146 q = bdev_get_queue(bio->bi_bdev); in rq_qos_done_bio()
1042 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset()1055 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment()
487 bdev_get_queue(bdev)->crypto_profile; in blk_crypto_derive_sw_secret()
981 spin_unlock_irq(&bdev_get_queue(ctx->bdev)->queue_lock); in blkg_conf_exit()2132 blkg_get(bdev_get_queue(bio->bi_bdev)->root_blkg); in bio_associate_blkg_from_css()2133 bio->bi_blkg = bdev_get_queue(bio->bi_bdev)->root_blkg; in bio_associate_blkg_from_css()
86 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_queue_enter()
704 queue_max_sectors(bdev_get_queue(bdev))); in blkdev_common_ioctl()
1115 struct request_queue *q = bdev_get_queue(bdev); in part_inflight_show()
1732 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in __blk_throtl_bio()
1762 trace_block_bio_complete(bdev_get_queue(bio->bi_bdev), bio); in bio_endio()
109 blk_mq_queue_tag_busy_iter(bdev_get_queue(part), blk_mq_check_in_driver, in blk_mq_in_driver_rw()3143 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in blk_mq_submit_bio()
3425 q = bdev_get_queue(ctx.bdev); in ioc_cost_model_write()
415 struct request_queue *q = bdev_get_queue(bdev); in dm_set_device_limits()890 struct request_queue *q = bdev_get_queue(bdev); in device_is_not_rq_stackable()1221 bdev_get_queue(bdev)->crypto_profile; in dm_wrappedkey_op_callback()1353 bdev_get_queue(dev->bdev)->crypto_profile; in device_intersect_crypto_capabilities()1897 struct request_queue *q = bdev_get_queue(dev->bdev); in device_not_write_zeroes_capable()
538 q = bdev_get_queue(bdev); in multipath_clone_and_map()881 struct request_queue *q = bdev_get_queue(bdev); in setup_scsi_dh()951 q = bdev_get_queue(p->path.dev->bdev); in parse_path()1631 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in activate_or_offline_path()2219 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in pgpath_busy()
316 struct request_queue *q = bdev_get_queue(where->bdev); in do_region()
589 if (blk_queue_dying(bdev_get_queue(dmz_dev->bdev))) { in dmz_bdev_is_dying()
926 struct request_queue *q = bdev_get_queue(bdev); in blk_trace_ioctl()2042 struct request_queue *q = bdev_get_queue(bdev); in sysfs_blk_trace_attr_show()2076 struct request_queue *q = bdev_get_queue(bdev); in sysfs_blk_trace_attr_store()
563 cpu_to_le32(queue_max_hw_sectors(bdev_get_queue(bdev))); in rnbd_srv_fill_msg_open_rsp()
133 q = bdev_get_queue(bd); in iblock_configure_device()
847 struct request_queue *q = bdev_get_queue(bdev); in target_configure_write_atomic_from_bdev()
755 &bdev_get_queue(chunk->bio.bi_bdev)->limits; in xfs_zone_gc_split_write()
1024 q = bdev_get_queue(dc->bdev); in cached_dev_status_update()1406 struct request_queue *q = bdev_get_queue(dc->bdev); in cached_dev_init()