Home
last modified time | relevance | path

Searched refs:num_queues (Results 1 – 25 of 59) sorted by relevance

123

/src/sys/dev/gve/
H A Dgve_main.c149 priv->tx_cfg.num_queues; in gve_tx_timeout_service_callback()
278 if (new_queue_cnt < priv->rx_cfg.num_queues) { in gve_adjust_rx_queues()
283 gve_free_rx_rings(priv, new_queue_cnt, priv->rx_cfg.num_queues); in gve_adjust_rx_queues()
285 err = gve_alloc_rx_rings(priv, priv->rx_cfg.num_queues, new_queue_cnt); in gve_adjust_rx_queues()
294 priv->rx_cfg.num_queues = new_queue_cnt; in gve_adjust_rx_queues()
312 if (new_queue_cnt < priv->tx_cfg.num_queues) { in gve_adjust_tx_queues()
317 gve_free_tx_rings(priv, new_queue_cnt, priv->tx_cfg.num_queues); in gve_adjust_tx_queues()
319 err = gve_alloc_tx_rings(priv, priv->tx_cfg.num_queues, new_queue_cnt); in gve_adjust_tx_queues()
328 priv->tx_cfg.num_queues = new_queue_cnt; in gve_adjust_tx_queues()
348 gve_free_rx_rings(priv, 0, priv->rx_cfg.num_queues); in gve_adjust_ring_sizes()
[all …]
H A Dgve_adminq.c243 gve_adminq_destroy_rx_queues(struct gve_priv *priv, uint32_t num_queues) in gve_adminq_destroy_rx_queues() argument
248 for (i = 0; i < num_queues; i++) { in gve_adminq_destroy_rx_queues()
259 device_printf(priv->dev, "Destroyed %d rx queues\n", num_queues); in gve_adminq_destroy_rx_queues()
264 gve_adminq_destroy_tx_queues(struct gve_priv *priv, uint32_t num_queues) in gve_adminq_destroy_tx_queues() argument
269 for (i = 0; i < num_queues; i++) { in gve_adminq_destroy_tx_queues()
280 device_printf(priv->dev, "Destroyed %d tx queues\n", num_queues); in gve_adminq_destroy_tx_queues()
332 gve_adminq_create_rx_queues(struct gve_priv *priv, uint32_t num_queues) in gve_adminq_create_rx_queues() argument
337 for (i = 0; i < num_queues; i++) { in gve_adminq_create_rx_queues()
347 device_printf(priv->dev, "Created %d rx queues\n", num_queues); in gve_adminq_create_rx_queues()
388 gve_adminq_create_tx_queues(struct gve_priv *priv, uint32_t num_queues) in gve_adminq_create_tx_queues() argument
[all …]
H A Dgve_sysctl.c206 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_setup_queue_stat_sysctl()
209 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_setup_queue_stat_sysctl()
321 val = priv->tx_cfg.num_queues; in gve_sysctl_num_tx_queues()
330 if (val != priv->tx_cfg.num_queues) { in gve_sysctl_num_tx_queues()
346 val = priv->rx_cfg.num_queues; in gve_sysctl_num_rx_queues()
356 if (val != priv->rx_cfg.num_queues) { in gve_sysctl_num_rx_queues()
492 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_accum_stats()
499 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_accum_stats()
H A Dgve_qpl.c172 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_register_qpls()
186 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_register_qpls()
214 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_unregister_qpls()
225 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_unregister_qpls()
H A Dgve_adminq.h440 int gve_adminq_create_rx_queues(struct gve_priv *priv, uint32_t num_queues);
441 int gve_adminq_create_tx_queues(struct gve_priv *priv, uint32_t num_queues);
442 int gve_adminq_destroy_tx_queues(struct gve_priv *priv, uint32_t num_queues);
443 int gve_adminq_destroy_rx_queues(struct gve_priv *priv, uint32_t num_queues);
H A Dgve_tx.c280 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_create_tx_rings()
287 err = gve_adminq_create_tx_queues(priv, priv->tx_cfg.num_queues); in gve_create_tx_rings()
294 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_create_tx_rings()
337 for (i = 0; i < priv->tx_cfg.num_queues; i++) in gve_destroy_tx_rings()
341 err = gve_adminq_destroy_tx_queues(priv, priv->tx_cfg.num_queues); in gve_destroy_tx_rings()
902 i = mbuf->m_pkthdr.flowid % priv->tx_cfg.num_queues; in gve_xmit_ifp()
904 i = curcpu % priv->tx_cfg.num_queues; in gve_xmit_ifp()
948 for (i = 0; i < priv->tx_cfg.num_queues; ++i) { in gve_qflush()
H A Dgve_utils.c411 for (idx = 0; idx < priv->tx_cfg.num_queues; idx++) { in gve_unmask_all_queue_irqs()
420 for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) { in gve_unmask_all_queue_irqs()
433 for (int idx = 0; idx < priv->tx_cfg.num_queues; idx++) { in gve_mask_all_queue_irqs()
437 for (int idx = 0; idx < priv->rx_cfg.num_queues; idx++) { in gve_mask_all_queue_irqs()
H A Dgve_rx.c313 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_create_rx_rings()
316 err = gve_adminq_create_rx_queues(priv, priv->rx_cfg.num_queues); in gve_create_rx_rings()
323 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_create_rx_rings()
363 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_destroy_rx_rings()
367 err = gve_adminq_destroy_rx_queues(priv, priv->rx_cfg.num_queues); in gve_destroy_rx_rings()
/src/sys/dev/ixl/
H A Dixl_pf_qmgr.c40 ixl_pf_qmgr_init(struct ixl_pf_qmgr *qmgr, u16 num_queues) in ixl_pf_qmgr_init() argument
42 if (num_queues < 1) in ixl_pf_qmgr_init()
45 qmgr->num_queues = num_queues; in ixl_pf_qmgr_init()
46 qmgr->qinfo = malloc(num_queues * sizeof(struct ixl_pf_qmgr_qinfo), in ixl_pf_qmgr_init()
117 for (i = 0; i < qmgr->num_queues; i++) { in ixl_pf_qmgr_alloc_scattered()
157 return (qmgr->num_queues); in ixl_pf_qmgr_get_num_queues()
169 for (int i = 0; i < qmgr->num_queues; i++) { in ixl_pf_qmgr_get_num_free()
182 if (start > qmgr->num_queues - 1) in ixl_pf_qmgr_get_first_free()
185 for (i = start; i < qmgr->num_queues; i++) { in ixl_pf_qmgr_get_first_free()
305 for (i = 0; i < qmgr->num_queues; i++) { in ixl_pf_qmgr_find_free_contiguous_block()
H A Dixl_pf_qmgr.h63 u16 num_queues; member
85 int ixl_pf_qmgr_init(struct ixl_pf_qmgr *qmgr, u16 num_queues);
H A Dixl_pf_iov.c75 static int ixl_vf_reserve_queues(struct ixl_pf *pf, struct ixl_vf *vf, int num_queues);
1740 ixl_vf_reserve_queues(struct ixl_pf *pf, struct ixl_vf *vf, int num_queues) in ixl_vf_reserve_queues() argument
1746 if (num_queues < 1 || num_queues > 16) in ixl_vf_reserve_queues()
1748 num_queues, vf->vf_num); in ixl_vf_reserve_queues()
1749 if (num_queues < 1) { in ixl_vf_reserve_queues()
1751 num_queues = 1; in ixl_vf_reserve_queues()
1752 } else if (num_queues > IAVF_MAX_QUEUES) { in ixl_vf_reserve_queues()
1754 num_queues = IAVF_MAX_QUEUES; in ixl_vf_reserve_queues()
1756 error = ixl_pf_qmgr_alloc_scattered(&pf->qmgr, num_queues, &vf->qtag); in ixl_vf_reserve_queues()
1759 num_queues, vf->vf_num); in ixl_vf_reserve_queues()
/src/sys/dev/ntb/if_ntb/
H A Dif_ntb.c79 SYSCTL_UINT(_hw_if_ntb, OID_AUTO, num_queues, CTLFLAG_RWTUN,
97 int num_queues; member
142 sc->num_queues = min(g_if_ntb_num_queues, in ntb_net_attach()
144 sc->queues = malloc(sc->num_queues * sizeof(struct ntb_net_queue), in ntb_net_attach()
147 for (i = 0; i < sc->num_queues; i++) { in ntb_net_attach()
164 sc->num_queues = i; in ntb_net_attach()
165 device_printf(dev, "%d queue(s)\n", sc->num_queues); in ntb_net_attach()
185 for (i = 0; i < sc->num_queues; i++) in ntb_net_attach()
197 for (i = 0; i < sc->num_queues; i++) in ntb_net_detach()
202 for (i = 0; i < sc->num_queues; i++) { in ntb_net_detach()
[all …]
/src/sys/contrib/dev/iwlwifi/fw/
H A Dinit.c108 int i, num_queues, size, ret; in iwl_configure_rxq() local
126 num_queues = fwrt->trans->info.num_rxqs - 1; in iwl_configure_rxq()
128 size = struct_size(cmd, data, num_queues); in iwl_configure_rxq()
134 cmd->num_queues = num_queues; in iwl_configure_rxq()
136 for (i = 0; i < num_queues; i++) { in iwl_configure_rxq()
/src/sys/dev/xen/netfront/
H A Dnetfront.c220 u_int num_queues; member
478 for (i = 0; i < np->num_queues; i++) { in netfront_suspend()
483 for (i = 0; i < np->num_queues; i++) { in netfront_suspend()
503 for (i = 0; i < info->num_queues; i++) { in netfront_resume()
508 for (i = 0; i < info->num_queues; i++) { in netfront_resume()
582 unsigned long num_queues, max_queues = 0; in talk_to_backend() local
595 num_queues = xn_num_queues; in talk_to_backend()
596 if (num_queues > max_queues) in talk_to_backend()
597 num_queues = max_queues; in talk_to_backend()
599 err = setup_device(dev, info, num_queues); in talk_to_backend()
[all …]
/src/tools/tools/nvmf/nvmfd/
H A Dio.c197 uint32_t num_queues; in handle_set_features() local
205 num_queues = le32toh(cmd->cdw11) & 0xffff; in handle_set_features()
208 if (num_queues == 65535) in handle_set_features()
212 if (le32toh(cmd->cdw11) >> 16 != num_queues) in handle_set_features()
216 num_queues++; in handle_set_features()
220 ioc->num_io_queues = num_queues; in handle_set_features()
221 ioc->io_qpairs = calloc(num_queues, sizeof(*ioc->io_qpairs)); in handle_set_features()
222 ioc->io_sockets = calloc(num_queues, sizeof(*ioc->io_sockets)); in handle_set_features()
/src/lib/libnvmf/
H A Dnvmf_host.c794 struct nvmf_qpair *admin_qp, u_int num_queues, in prepare_queues_for_handoff() argument
803 if (num_queues == 0) in prepare_queues_for_handoff()
813 for (i = 0; i < num_queues; i++) { in prepare_queues_for_handoff()
822 nvlist_add_number(nvl_rparams, "num_io_queues", num_queues); in prepare_queues_for_handoff()
863 for (i = 0; i < num_queues; i++) { in prepare_queues_for_handoff()
881 const char *hostnqn, struct nvmf_qpair *admin_qp, u_int num_queues, in nvmf_handoff_host() argument
896 num_queues, io_queues, cdata, reconnect_delay, in nvmf_handoff_host()
908 for (i = 0; i < num_queues; i++) in nvmf_handoff_host()
992 const char *hostnqn, struct nvmf_qpair *admin_qp, u_int num_queues, in nvmf_reconnect_host() argument
1001 num_queues, io_queues, cdata, reconnect_delay, in nvmf_reconnect_host()
[all …]
H A Dlibnvmf.h351 const char *hostnqn, struct nvmf_qpair *admin_qp, u_int num_queues,
380 const char *hostnqn, struct nvmf_qpair *admin_qp, u_int num_queues,
/src/sys/dev/mana/
H A Dmana_sysctl.c70 for (i = 0; i < apc->num_queues; i++) { in mana_sysctl_rx_stat_agg_u64()
79 for (i = 0; i < apc->num_queues; i++) { in mana_sysctl_rx_stat_agg_u64()
135 for (i = 0; i < apc->num_queues; i++) { in mana_sysctl_tx_stat_agg_u64()
144 for (i = 0; i < apc->num_queues; i++) { in mana_sysctl_tx_stat_agg_u64()
276 for (i = 0; i < apc->num_queues; i++) { in mana_sysctl_add_queues()
H A Dmana_en.c532 % apc->num_queues; in mana_xmit()
834 apc->num_queues; in mana_start_xmit()
836 txq_id = m->m_pkthdr.flowid % apc->num_queues; in mana_start_xmit()
920 apc->rxqs = mallocarray(apc->num_queues, sizeof(struct mana_rxq *), in mana_init_port_context()
1394 for (rxq_idx = 0; rxq_idx < apc->num_queues; rxq_idx++) { in mana_fence_rqs()
2058 for (i = 0; i < apc->num_queues; i++) { in mana_destroy_txq()
2086 apc->tx_qp = mallocarray(apc->num_queues, sizeof(struct mana_tx_qp), in mana_create_txq()
2107 for (i = 0; i < apc->num_queues; i++) { in mana_create_txq()
2565 for (i = 0; i < apc->num_queues; i++) { in mana_add_rx_queues()
2586 for (rxq_idx = 0; rxq_idx < apc->num_queues; rxq_idx++) { in mana_destroy_vport()
[all …]
/src/sys/net/
H A Dif_epair.c128 int num_queues; member
231 bucket %= sc->num_queues; in epair_select_queue()
601 sc->num_queues = epair_tasks.tasks; in epair_alloc_sc()
602 sc->queues = mallocarray(sc->num_queues, sizeof(struct epair_queue), in epair_alloc_sc()
604 for (int i = 0; i < sc->num_queues; i++) { in epair_alloc_sc()
701 for (int i = 0; i < sc->num_queues; i++) { in epair_free_sc()
845 for (int i = 0; i < sc->num_queues; i++) { in epair_drain_rings()
/src/sys/dev/nvme/
H A Dnvme_ctrlr_cmd.c216 uint32_t num_queues, nvme_cb_fn_t cb_fn, void *cb_arg) in nvme_ctrlr_cmd_set_num_queues() argument
220 cdw11 = ((num_queues - 1) << 16) | (num_queues - 1); in nvme_ctrlr_cmd_set_num_queues()
/src/sys/dev/nvmf/controller/
H A Dnvmft_controller.c751 uint32_t num_queues; in handle_set_features() local
754 num_queues = le32toh(cmd->cdw11) & 0xffff; in handle_set_features()
757 if (num_queues == 65535) in handle_set_features()
761 if (le32toh(cmd->cdw11) >> 16 != num_queues) in handle_set_features()
765 num_queues++; in handle_set_features()
767 io_qpairs = mallocarray(num_queues, sizeof(*io_qpairs), in handle_set_features()
780 ctrlr->num_io_queues = num_queues; in handle_set_features()
/src/sys/dev/virtio/pci/
H A Dvirtio_pci_modern_var.h78 uint16_t num_queues; /* read-only */ member
/src/sys/dev/virtio/block/
H A Dvirtio_blk.h91 uint16_t num_queues; member
/src/sys/dev/sfxge/common/
H A Def10_rx.c208 __in uint32_t num_queues, in efx_mcdi_rss_context_alloc() argument
218 if (num_queues > EFX_MAXRSS) { in efx_mcdi_rss_context_alloc()
251 MCDI_IN_SET_DWORD(req, RSS_CONTEXT_ALLOC_IN_NUM_QUEUES, num_queues); in efx_mcdi_rss_context_alloc()
583 __in uint32_t num_queues, in ef10_rx_scale_context_alloc() argument
588 rc = efx_mcdi_rss_context_alloc(enp, type, num_queues, rss_contextp); in ef10_rx_scale_context_alloc()

123