Searched refs:elv_tbl (Results 1 – 3 of 3) sorted by relevance
| /linux/block/ |
| H A D | blk-mq-sched.c | 445 void blk_mq_free_sched_res_batch(struct xarray *elv_tbl, in blk_mq_free_sched_res_batch() argument 462 ctx = xa_load(elv_tbl, q->id); in blk_mq_free_sched_res_batch() 472 void blk_mq_free_sched_ctx_batch(struct xarray *elv_tbl) in blk_mq_free_sched_ctx_batch() argument 477 xa_for_each(elv_tbl, i, ctx) { in blk_mq_free_sched_ctx_batch() 478 xa_erase(elv_tbl, i); in blk_mq_free_sched_ctx_batch() 483 int blk_mq_alloc_sched_ctx_batch(struct xarray *elv_tbl, in blk_mq_alloc_sched_ctx_batch() argument 496 if (xa_insert(elv_tbl, q->id, ctx, GFP_KERNEL)) { in blk_mq_alloc_sched_ctx_batch() 569 int blk_mq_alloc_sched_res_batch(struct xarray *elv_tbl, in blk_mq_alloc_sched_res_batch() argument 587 ctx = xa_load(elv_tbl, q->id); in blk_mq_alloc_sched_res_batch() 604 ctx = xa_load(elv_tbl, q->id); in blk_mq_alloc_sched_res_batch()
|
| H A D | blk-mq-sched.h | 32 int blk_mq_alloc_sched_res_batch(struct xarray *elv_tbl, 34 int blk_mq_alloc_sched_ctx_batch(struct xarray *elv_tbl, 36 void blk_mq_free_sched_ctx_batch(struct xarray *elv_tbl);
|
| H A D | blk-mq.c | 5057 struct xarray *elv_tbl) in blk_mq_elv_switch_back() argument 5059 struct elv_change_ctx *ctx = xa_load(elv_tbl, q->id); in blk_mq_elv_switch_back() 5076 struct xarray *elv_tbl) in blk_mq_elv_switch_none() argument 5090 ctx = xa_load(elv_tbl, q->id); in blk_mq_elv_switch_none() 5122 struct xarray elv_tbl; in __blk_mq_update_nr_hw_queues() local 5137 xa_init(&elv_tbl); in __blk_mq_update_nr_hw_queues() 5138 if (blk_mq_alloc_sched_ctx_batch(&elv_tbl, set) < 0) in __blk_mq_update_nr_hw_queues() 5141 if (blk_mq_alloc_sched_res_batch(&elv_tbl, set, nr_hw_queues) < 0) in __blk_mq_update_nr_hw_queues() 5155 if (blk_mq_elv_switch_none(q, &elv_tbl)) in __blk_mq_update_nr_hw_queues() 5195 blk_mq_elv_switch_back(q, &elv_tbl); in __blk_mq_update_nr_hw_queues() [all …]
|