Home
last modified time | relevance | path

Searched refs:nr_io_queues (Results 1 – 10 of 10) sorted by relevance

/linux/drivers/nvme/host/
H A Dfabrics.c729 opts->nr_io_queues = num_online_cpus(); in nvmf_parse_options()
831 opts->nr_io_queues = min_t(unsigned int, in nvmf_parse_options()
1076 opts->nr_io_queues = 0; in nvmf_parse_options()
1123 void nvmf_set_io_queues(struct nvmf_ctrl_options *opts, u32 nr_io_queues, in nvmf_set_io_queues() argument
1126 if (opts->nr_write_queues && opts->nr_io_queues < nr_io_queues) { in nvmf_set_io_queues()
1132 io_queues[HCTX_TYPE_READ] = opts->nr_io_queues; in nvmf_set_io_queues()
1133 nr_io_queues -= io_queues[HCTX_TYPE_READ]; in nvmf_set_io_queues()
1135 min(opts->nr_write_queues, nr_io_queues); in nvmf_set_io_queues()
1136 nr_io_queues -= io_queues[HCTX_TYPE_DEFAULT]; in nvmf_set_io_queues()
1144 min(opts->nr_io_queues, nr_io_queues); in nvmf_set_io_queues()
[all …]
H A Dfabrics.h124 unsigned int nr_io_queues; member
215 return min(opts->nr_io_queues, num_online_cpus()) + in nvmf_nr_io_queues()
239 void nvmf_set_io_queues(struct nvmf_ctrl_options *opts, u32 nr_io_queues,
H A Dfc.c2863 unsigned int nr_io_queues; in nvme_fc_create_io_queues() local
2866 nr_io_queues = min3(opts->nr_io_queues, num_online_cpus(), in nvme_fc_create_io_queues()
2868 ret = nvme_set_queue_count(&ctrl->ctrl, &nr_io_queues); in nvme_fc_create_io_queues()
2875 ctrl->ctrl.queue_count = nr_io_queues + 1; in nvme_fc_create_io_queues()
2876 if (!nr_io_queues) in nvme_fc_create_io_queues()
2917 unsigned int nr_io_queues; in nvme_fc_recreate_io_queues() local
2920 nr_io_queues = min3(opts->nr_io_queues, num_online_cpus(), in nvme_fc_recreate_io_queues()
2922 ret = nvme_set_queue_count(&ctrl->ctrl, &nr_io_queues); in nvme_fc_recreate_io_queues()
2929 if (!nr_io_queues && prior_ioq_cnt) { in nvme_fc_recreate_io_queues()
2936 ctrl->ctrl.queue_count = nr_io_queues + 1; in nvme_fc_recreate_io_queues()
[all …]
H A Dpci.c2052 static int nvme_cmb_qdepth(struct nvme_dev *dev, int nr_io_queues, in nvme_cmb_qdepth() argument
2059 if (q_size_aligned * nr_io_queues > dev->cmb_size) { in nvme_cmb_qdepth()
2060 u64 mem_per_q = div_u64(dev->cmb_size, nr_io_queues); in nvme_cmb_qdepth()
2274 static unsigned long db_bar_size(struct nvme_dev *dev, unsigned nr_io_queues) in db_bar_size() argument
2276 return NVME_REG_DBS + ((nr_io_queues + 1) * 8 * dev->db_stride); in db_bar_size()
2858 static int nvme_setup_irqs(struct nvme_dev *dev, unsigned int nr_io_queues) in nvme_setup_irqs() argument
2873 poll_queues = min(dev->nr_poll_queues, nr_io_queues - 1); in nvme_setup_irqs()
2890 irq_queues += (nr_io_queues - poll_queues); in nvme_setup_irqs()
2913 unsigned int nr_io_queues; in nvme_setup_io_queues() local
2941 nr_io_queues = min(nvme_max_io_queues(dev), in nvme_setup_io_queues()
[all …]
H A Drdma.c719 unsigned int nr_io_queues; in nvme_rdma_alloc_io_queues() local
722 nr_io_queues = nvmf_nr_io_queues(opts); in nvme_rdma_alloc_io_queues()
723 ret = nvme_set_queue_count(&ctrl->ctrl, &nr_io_queues); in nvme_rdma_alloc_io_queues()
727 if (nr_io_queues == 0) { in nvme_rdma_alloc_io_queues()
733 ctrl->ctrl.queue_count = nr_io_queues + 1; in nvme_rdma_alloc_io_queues()
735 "creating %d I/O queues.\n", nr_io_queues); in nvme_rdma_alloc_io_queues()
737 nvmf_set_io_queues(opts, nr_io_queues, ctrl->io_queues); in nvme_rdma_alloc_io_queues()
2287 ctrl->ctrl.queue_count = opts->nr_io_queues + opts->nr_write_queues + in nvme_rdma_alloc_ctrl()
H A Dapple.c1024 unsigned int nr_io_queues = 1; in apple_nvme_reset_work() local
1192 nr_io_queues = 1; in apple_nvme_reset_work()
1193 ret = nvme_set_queue_count(&anv->ctrl, &nr_io_queues); in apple_nvme_reset_work()
1196 if (nr_io_queues != 1) { in apple_nvme_reset_work()
1201 anv->ctrl.queue_count = nr_io_queues + 1; in apple_nvme_reset_work()
H A Dtcp.c2144 unsigned int nr_io_queues; in nvme_tcp_alloc_io_queues() local
2147 nr_io_queues = nvmf_nr_io_queues(ctrl->opts); in nvme_tcp_alloc_io_queues()
2148 ret = nvme_set_queue_count(ctrl, &nr_io_queues); in nvme_tcp_alloc_io_queues()
2152 if (nr_io_queues == 0) { in nvme_tcp_alloc_io_queues()
2158 ctrl->queue_count = nr_io_queues + 1; in nvme_tcp_alloc_io_queues()
2160 "creating %d I/O queues.\n", nr_io_queues); in nvme_tcp_alloc_io_queues()
2162 nvmf_set_io_queues(ctrl->opts, nr_io_queues, in nvme_tcp_alloc_io_queues()
2901 ctrl->ctrl.queue_count = opts->nr_io_queues + opts->nr_write_queues + in nvme_tcp_alloc_ctrl()
H A Dauth.c56 return ctrl->opts->nr_io_queues + ctrl->opts->nr_write_queues + in ctrl_max_dhchaps()
H A Dcore.c1724 int status, nr_io_queues; in nvme_set_queue_count() local
1746 nr_io_queues = min(result & 0xffff, result >> 16) + 1; in nvme_set_queue_count()
1747 *count = min(*count, nr_io_queues); in nvme_set_queue_count()
/linux/drivers/nvme/target/
H A Dloop.c320 unsigned int nr_io_queues; in nvme_loop_init_io_queues() local
323 nr_io_queues = min(opts->nr_io_queues, num_online_cpus()); in nvme_loop_init_io_queues()
324 ret = nvme_set_queue_count(&ctrl->ctrl, &nr_io_queues); in nvme_loop_init_io_queues()
325 if (ret || !nr_io_queues) in nvme_loop_init_io_queues()
328 dev_info(ctrl->ctrl.device, "creating %d I/O queues.\n", nr_io_queues); in nvme_loop_init_io_queues()
330 for (i = 1; i <= nr_io_queues; i++) { in nvme_loop_init_io_queues()
595 ctrl->queues = kzalloc_objs(*ctrl->queues, opts->nr_io_queues + 1); in nvme_loop_create_ctrl()
612 if (opts->nr_io_queues) { in nvme_loop_create_ctrl()