| /src/sys/dev/gve/ |
| H A D | gve_main.c | 149 priv->tx_cfg.num_queues; in gve_tx_timeout_service_callback() 278 if (new_queue_cnt < priv->rx_cfg.num_queues) { in gve_adjust_rx_queues() 283 gve_free_rx_rings(priv, new_queue_cnt, priv->rx_cfg.num_queues); in gve_adjust_rx_queues() 285 err = gve_alloc_rx_rings(priv, priv->rx_cfg.num_queues, new_queue_cnt); in gve_adjust_rx_queues() 294 priv->rx_cfg.num_queues = new_queue_cnt; in gve_adjust_rx_queues() 312 if (new_queue_cnt < priv->tx_cfg.num_queues) { in gve_adjust_tx_queues() 317 gve_free_tx_rings(priv, new_queue_cnt, priv->tx_cfg.num_queues); in gve_adjust_tx_queues() 319 err = gve_alloc_tx_rings(priv, priv->tx_cfg.num_queues, new_queue_cnt); in gve_adjust_tx_queues() 328 priv->tx_cfg.num_queues = new_queue_cnt; in gve_adjust_tx_queues() 348 gve_free_rx_rings(priv, 0, priv->rx_cfg.num_queues); in gve_adjust_ring_sizes() [all …]
|
| H A D | gve_adminq.c | 243 gve_adminq_destroy_rx_queues(struct gve_priv *priv, uint32_t num_queues) in gve_adminq_destroy_rx_queues() argument 248 for (i = 0; i < num_queues; i++) { in gve_adminq_destroy_rx_queues() 259 device_printf(priv->dev, "Destroyed %d rx queues\n", num_queues); in gve_adminq_destroy_rx_queues() 264 gve_adminq_destroy_tx_queues(struct gve_priv *priv, uint32_t num_queues) in gve_adminq_destroy_tx_queues() argument 269 for (i = 0; i < num_queues; i++) { in gve_adminq_destroy_tx_queues() 280 device_printf(priv->dev, "Destroyed %d tx queues\n", num_queues); in gve_adminq_destroy_tx_queues() 332 gve_adminq_create_rx_queues(struct gve_priv *priv, uint32_t num_queues) in gve_adminq_create_rx_queues() argument 337 for (i = 0; i < num_queues; i++) { in gve_adminq_create_rx_queues() 347 device_printf(priv->dev, "Created %d rx queues\n", num_queues); in gve_adminq_create_rx_queues() 388 gve_adminq_create_tx_queues(struct gve_priv *priv, uint32_t num_queues) in gve_adminq_create_tx_queues() argument [all …]
|
| H A D | gve_sysctl.c | 206 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_setup_queue_stat_sysctl() 209 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_setup_queue_stat_sysctl() 321 val = priv->tx_cfg.num_queues; in gve_sysctl_num_tx_queues() 330 if (val != priv->tx_cfg.num_queues) { in gve_sysctl_num_tx_queues() 346 val = priv->rx_cfg.num_queues; in gve_sysctl_num_rx_queues() 356 if (val != priv->rx_cfg.num_queues) { in gve_sysctl_num_rx_queues() 492 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_accum_stats() 499 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_accum_stats()
|
| H A D | gve_qpl.c | 172 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_register_qpls() 186 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_register_qpls() 214 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_unregister_qpls() 225 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_unregister_qpls()
|
| H A D | gve_adminq.h | 440 int gve_adminq_create_rx_queues(struct gve_priv *priv, uint32_t num_queues); 441 int gve_adminq_create_tx_queues(struct gve_priv *priv, uint32_t num_queues); 442 int gve_adminq_destroy_tx_queues(struct gve_priv *priv, uint32_t num_queues); 443 int gve_adminq_destroy_rx_queues(struct gve_priv *priv, uint32_t num_queues);
|
| H A D | gve_tx.c | 280 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_create_tx_rings() 287 err = gve_adminq_create_tx_queues(priv, priv->tx_cfg.num_queues); in gve_create_tx_rings() 294 for (i = 0; i < priv->tx_cfg.num_queues; i++) { in gve_create_tx_rings() 337 for (i = 0; i < priv->tx_cfg.num_queues; i++) in gve_destroy_tx_rings() 341 err = gve_adminq_destroy_tx_queues(priv, priv->tx_cfg.num_queues); in gve_destroy_tx_rings() 902 i = mbuf->m_pkthdr.flowid % priv->tx_cfg.num_queues; in gve_xmit_ifp() 904 i = curcpu % priv->tx_cfg.num_queues; in gve_xmit_ifp() 948 for (i = 0; i < priv->tx_cfg.num_queues; ++i) { in gve_qflush()
|
| H A D | gve_utils.c | 411 for (idx = 0; idx < priv->tx_cfg.num_queues; idx++) { in gve_unmask_all_queue_irqs() 420 for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) { in gve_unmask_all_queue_irqs() 433 for (int idx = 0; idx < priv->tx_cfg.num_queues; idx++) { in gve_mask_all_queue_irqs() 437 for (int idx = 0; idx < priv->rx_cfg.num_queues; idx++) { in gve_mask_all_queue_irqs()
|
| H A D | gve_rx.c | 313 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_create_rx_rings() 316 err = gve_adminq_create_rx_queues(priv, priv->rx_cfg.num_queues); in gve_create_rx_rings() 323 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_create_rx_rings() 363 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_destroy_rx_rings() 367 err = gve_adminq_destroy_rx_queues(priv, priv->rx_cfg.num_queues); in gve_destroy_rx_rings()
|
| /src/sys/dev/ixl/ |
| H A D | ixl_pf_qmgr.c | 40 ixl_pf_qmgr_init(struct ixl_pf_qmgr *qmgr, u16 num_queues) in ixl_pf_qmgr_init() argument 42 if (num_queues < 1) in ixl_pf_qmgr_init() 45 qmgr->num_queues = num_queues; in ixl_pf_qmgr_init() 46 qmgr->qinfo = malloc(num_queues * sizeof(struct ixl_pf_qmgr_qinfo), in ixl_pf_qmgr_init() 117 for (i = 0; i < qmgr->num_queues; i++) { in ixl_pf_qmgr_alloc_scattered() 157 return (qmgr->num_queues); in ixl_pf_qmgr_get_num_queues() 169 for (int i = 0; i < qmgr->num_queues; i++) { in ixl_pf_qmgr_get_num_free() 182 if (start > qmgr->num_queues - 1) in ixl_pf_qmgr_get_first_free() 185 for (i = start; i < qmgr->num_queues; i++) { in ixl_pf_qmgr_get_first_free() 305 for (i = 0; i < qmgr->num_queues; i++) { in ixl_pf_qmgr_find_free_contiguous_block()
|
| H A D | ixl_pf_qmgr.h | 63 u16 num_queues; member 85 int ixl_pf_qmgr_init(struct ixl_pf_qmgr *qmgr, u16 num_queues);
|
| H A D | ixl_pf_iov.c | 75 static int ixl_vf_reserve_queues(struct ixl_pf *pf, struct ixl_vf *vf, int num_queues); 1740 ixl_vf_reserve_queues(struct ixl_pf *pf, struct ixl_vf *vf, int num_queues) in ixl_vf_reserve_queues() argument 1746 if (num_queues < 1 || num_queues > 16) in ixl_vf_reserve_queues() 1748 num_queues, vf->vf_num); in ixl_vf_reserve_queues() 1749 if (num_queues < 1) { in ixl_vf_reserve_queues() 1751 num_queues = 1; in ixl_vf_reserve_queues() 1752 } else if (num_queues > IAVF_MAX_QUEUES) { in ixl_vf_reserve_queues() 1754 num_queues = IAVF_MAX_QUEUES; in ixl_vf_reserve_queues() 1756 error = ixl_pf_qmgr_alloc_scattered(&pf->qmgr, num_queues, &vf->qtag); in ixl_vf_reserve_queues() 1759 num_queues, vf->vf_num); in ixl_vf_reserve_queues()
|
| /src/sys/dev/ntb/if_ntb/ |
| H A D | if_ntb.c | 79 SYSCTL_UINT(_hw_if_ntb, OID_AUTO, num_queues, CTLFLAG_RWTUN, 97 int num_queues; member 142 sc->num_queues = min(g_if_ntb_num_queues, in ntb_net_attach() 144 sc->queues = malloc(sc->num_queues * sizeof(struct ntb_net_queue), in ntb_net_attach() 147 for (i = 0; i < sc->num_queues; i++) { in ntb_net_attach() 164 sc->num_queues = i; in ntb_net_attach() 165 device_printf(dev, "%d queue(s)\n", sc->num_queues); in ntb_net_attach() 185 for (i = 0; i < sc->num_queues; i++) in ntb_net_attach() 197 for (i = 0; i < sc->num_queues; i++) in ntb_net_detach() 202 for (i = 0; i < sc->num_queues; i++) { in ntb_net_detach() [all …]
|
| /src/sys/contrib/dev/iwlwifi/fw/ |
| H A D | init.c | 108 int i, num_queues, size, ret; in iwl_configure_rxq() local 126 num_queues = fwrt->trans->info.num_rxqs - 1; in iwl_configure_rxq() 128 size = struct_size(cmd, data, num_queues); in iwl_configure_rxq() 134 cmd->num_queues = num_queues; in iwl_configure_rxq() 136 for (i = 0; i < num_queues; i++) { in iwl_configure_rxq()
|
| /src/sys/dev/xen/netfront/ |
| H A D | netfront.c | 220 u_int num_queues; member 478 for (i = 0; i < np->num_queues; i++) { in netfront_suspend() 483 for (i = 0; i < np->num_queues; i++) { in netfront_suspend() 503 for (i = 0; i < info->num_queues; i++) { in netfront_resume() 508 for (i = 0; i < info->num_queues; i++) { in netfront_resume() 582 unsigned long num_queues, max_queues = 0; in talk_to_backend() local 595 num_queues = xn_num_queues; in talk_to_backend() 596 if (num_queues > max_queues) in talk_to_backend() 597 num_queues = max_queues; in talk_to_backend() 599 err = setup_device(dev, info, num_queues); in talk_to_backend() [all …]
|
| /src/tools/tools/nvmf/nvmfd/ |
| H A D | io.c | 197 uint32_t num_queues; in handle_set_features() local 205 num_queues = le32toh(cmd->cdw11) & 0xffff; in handle_set_features() 208 if (num_queues == 65535) in handle_set_features() 212 if (le32toh(cmd->cdw11) >> 16 != num_queues) in handle_set_features() 216 num_queues++; in handle_set_features() 220 ioc->num_io_queues = num_queues; in handle_set_features() 221 ioc->io_qpairs = calloc(num_queues, sizeof(*ioc->io_qpairs)); in handle_set_features() 222 ioc->io_sockets = calloc(num_queues, sizeof(*ioc->io_sockets)); in handle_set_features()
|
| /src/lib/libnvmf/ |
| H A D | nvmf_host.c | 794 struct nvmf_qpair *admin_qp, u_int num_queues, in prepare_queues_for_handoff() argument 803 if (num_queues == 0) in prepare_queues_for_handoff() 813 for (i = 0; i < num_queues; i++) { in prepare_queues_for_handoff() 822 nvlist_add_number(nvl_rparams, "num_io_queues", num_queues); in prepare_queues_for_handoff() 863 for (i = 0; i < num_queues; i++) { in prepare_queues_for_handoff() 881 const char *hostnqn, struct nvmf_qpair *admin_qp, u_int num_queues, in nvmf_handoff_host() argument 896 num_queues, io_queues, cdata, reconnect_delay, in nvmf_handoff_host() 908 for (i = 0; i < num_queues; i++) in nvmf_handoff_host() 992 const char *hostnqn, struct nvmf_qpair *admin_qp, u_int num_queues, in nvmf_reconnect_host() argument 1001 num_queues, io_queues, cdata, reconnect_delay, in nvmf_reconnect_host() [all …]
|
| H A D | libnvmf.h | 351 const char *hostnqn, struct nvmf_qpair *admin_qp, u_int num_queues, 380 const char *hostnqn, struct nvmf_qpair *admin_qp, u_int num_queues,
|
| /src/sys/dev/mana/ |
| H A D | mana_sysctl.c | 70 for (i = 0; i < apc->num_queues; i++) { in mana_sysctl_rx_stat_agg_u64() 79 for (i = 0; i < apc->num_queues; i++) { in mana_sysctl_rx_stat_agg_u64() 135 for (i = 0; i < apc->num_queues; i++) { in mana_sysctl_tx_stat_agg_u64() 144 for (i = 0; i < apc->num_queues; i++) { in mana_sysctl_tx_stat_agg_u64() 276 for (i = 0; i < apc->num_queues; i++) { in mana_sysctl_add_queues()
|
| H A D | mana_en.c | 532 % apc->num_queues; in mana_xmit() 834 apc->num_queues; in mana_start_xmit() 836 txq_id = m->m_pkthdr.flowid % apc->num_queues; in mana_start_xmit() 920 apc->rxqs = mallocarray(apc->num_queues, sizeof(struct mana_rxq *), in mana_init_port_context() 1394 for (rxq_idx = 0; rxq_idx < apc->num_queues; rxq_idx++) { in mana_fence_rqs() 2058 for (i = 0; i < apc->num_queues; i++) { in mana_destroy_txq() 2086 apc->tx_qp = mallocarray(apc->num_queues, sizeof(struct mana_tx_qp), in mana_create_txq() 2107 for (i = 0; i < apc->num_queues; i++) { in mana_create_txq() 2565 for (i = 0; i < apc->num_queues; i++) { in mana_add_rx_queues() 2586 for (rxq_idx = 0; rxq_idx < apc->num_queues; rxq_idx++) { in mana_destroy_vport() [all …]
|
| /src/sys/net/ |
| H A D | if_epair.c | 128 int num_queues; member 231 bucket %= sc->num_queues; in epair_select_queue() 601 sc->num_queues = epair_tasks.tasks; in epair_alloc_sc() 602 sc->queues = mallocarray(sc->num_queues, sizeof(struct epair_queue), in epair_alloc_sc() 604 for (int i = 0; i < sc->num_queues; i++) { in epair_alloc_sc() 701 for (int i = 0; i < sc->num_queues; i++) { in epair_free_sc() 845 for (int i = 0; i < sc->num_queues; i++) { in epair_drain_rings()
|
| /src/sys/dev/nvme/ |
| H A D | nvme_ctrlr_cmd.c | 216 uint32_t num_queues, nvme_cb_fn_t cb_fn, void *cb_arg) in nvme_ctrlr_cmd_set_num_queues() argument 220 cdw11 = ((num_queues - 1) << 16) | (num_queues - 1); in nvme_ctrlr_cmd_set_num_queues()
|
| /src/sys/dev/nvmf/controller/ |
| H A D | nvmft_controller.c | 751 uint32_t num_queues; in handle_set_features() local 754 num_queues = le32toh(cmd->cdw11) & 0xffff; in handle_set_features() 757 if (num_queues == 65535) in handle_set_features() 761 if (le32toh(cmd->cdw11) >> 16 != num_queues) in handle_set_features() 765 num_queues++; in handle_set_features() 767 io_qpairs = mallocarray(num_queues, sizeof(*io_qpairs), in handle_set_features() 780 ctrlr->num_io_queues = num_queues; in handle_set_features()
|
| /src/sys/dev/virtio/pci/ |
| H A D | virtio_pci_modern_var.h | 78 uint16_t num_queues; /* read-only */ member
|
| /src/sys/dev/virtio/block/ |
| H A D | virtio_blk.h | 91 uint16_t num_queues; member
|
| /src/sys/dev/sfxge/common/ |
| H A D | ef10_rx.c | 208 __in uint32_t num_queues, in efx_mcdi_rss_context_alloc() argument 218 if (num_queues > EFX_MAXRSS) { in efx_mcdi_rss_context_alloc() 251 MCDI_IN_SET_DWORD(req, RSS_CONTEXT_ALLOC_IN_NUM_QUEUES, num_queues); in efx_mcdi_rss_context_alloc() 583 __in uint32_t num_queues, in ef10_rx_scale_context_alloc() argument 588 rc = efx_mcdi_rss_context_alloc(enp, type, num_queues, rss_contextp); in ef10_rx_scale_context_alloc()
|