Home
last modified time | relevance | path

Searched refs:mq_hctx (Results 1 – 14 of 14) sorted by relevance

/linux/block/
H A Dblk-mq-sched.c53 return rqa->mq_hctx > rqb->mq_hctx; in sched_rq_cmp()
59 list_first_entry(rq_list, struct request, queuelist)->mq_hctx; in blk_mq_dispatch_hctx_list()
64 if (rq->mq_hctx != hctx) { in blk_mq_dispatch_hctx_list()
139 if (rq->mq_hctx != hctx) in __blk_mq_do_dispatch_sched()
262 } while (blk_mq_dispatch_rq_list(rq->mq_hctx, &rq_list, false)); in blk_mq_do_dispatch_ctx()
H A Dblk-mq.c420 rq->mq_hctx = hctx; in blk_mq_rq_ctx_init()
647 if (blk_mq_get_hctx_type(opf) != rq->mq_hctx->type) in blk_mq_alloc_cached_request()
803 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in __blk_mq_free_request()
808 rq->mq_hctx = NULL; in __blk_mq_free_request()
1233 if (nr_tags == TAG_COMP_BATCH || cur_hctx != rq->mq_hctx) { in blk_mq_end_request_batch()
1237 cur_hctx = rq->mq_hctx; in blk_mq_end_request_batch()
1328 if ((rq->mq_hctx->nr_ctx == 1 && in blk_mq_complete_request_remote()
1386 rq->mq_hctx->tags->rqs[rq->tag] = rq; in blk_mq_start_request()
1392 WRITE_ONCE(rq->bio->bi_cookie, rq->mq_hctx->queue_num); in blk_mq_start_request()
1448 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in blk_execute_rq_nowait()
[all …]
H A Dblk-flush.c311 flush_rq->mq_hctx = first_rq->mq_hctx; in blk_kick_flush()
343 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in mq_flush_data_end_io()
H A Dblk-mq-tag.c291 if (rq->q == q && (!hctx || rq->mq_hctx == hctx)) in bt_iter()
646 return (rq->mq_hctx->queue_num << BLK_MQ_UNIQUE_TAG_BITS) | in blk_mq_unique_tag()
H A Dblk-mq.h367 __blk_mq_put_driver_tag(rq->mq_hctx, rq); in blk_mq_put_driver_tag()
H A Dblk-mq-debugfs.c344 if (rq->mq_hctx == params->hctx) in hctx_show_busy_rq()
/linux/drivers/nvme/host/
H A Dpci.c757 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_pci_metadata_use_sgls()
769 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_pci_use_sgls()
835 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_free_descriptors()
860 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_free_prps()
872 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_free_sgls()
891 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_unmap_metadata()
931 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_unmap_data()
973 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_pci_prp_save_mapping()
1003 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_pci_setup_data_prp()
1138 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_pci_setup_data_sgl()
[all …]
H A Dnvme.h276 return req->mq_hctx->queue_num + 1; in nvme_req_qid()
H A Dcore.c740 if (req->mq_hctx->type == HCTX_TYPE_POLL) in nvme_init_request()
/linux/drivers/block/
H A Dvirtio_blk.c338 struct virtio_blk *vblk = req->mq_hctx->queue->queuedata; in virtblk_request_done()
467 struct virtio_blk *vblk = req->mq_hctx->queue->queuedata; in virtblk_prep_rq_batch()
470 return virtblk_prep_rq(req->mq_hctx, vblk, req, vbr) == BLK_STS_OK; in virtblk_prep_rq_batch()
509 struct virtio_blk_vq *this_vq = get_virtio_blk_vq(req->mq_hctx); in virtio_queue_rqs()
H A Dublk_drv.c2017 ublk_dispatch_req(rq->mq_hctx->driver_data, rq); in ublk_cmd_list_tw_cb()
2034 struct ublk_queue *ubq = rq->mq_hctx->driver_data; in ublk_timeout()
2180 struct ublk_queue *this_q = req->mq_hctx->driver_data; in ublk_queue_rqs()
2233 struct ublk_queue *this_q = req->mq_hctx->driver_data; in ublk_batch_queue_rqs()
3065 struct ublk_queue *ubq = rq->mq_hctx->driver_data; in ublk_io_release()
5038 if (!blk_mq_request_started(rq) && rq->mq_hctx->driver_data == idle->ubq) in ublk_count_busy_req()
/linux/drivers/ufs/core/
H A Dufs-mcq.c118 struct blk_mq_hw_ctx *hctx = READ_ONCE(req->mq_hctx); in ufshcd_mcq_req_to_hwq()
/linux/include/linux/
H A Dblk-mq.h108 struct blk_mq_hw_ctx *mq_hctx; member
/linux/drivers/block/null_blk/
H A Dmain.c1621 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in null_timeout_rq()
1717 ret = null_queue_rq(rq->mq_hctx, &bd); in null_queue_rqs()