| /src/sys/dev/sfxge/ |
| H A D | sfxge_tx.c | 143 static void sfxge_tx_qdpl_service(struct sfxge_txq *txq); 144 static void sfxge_tx_qlist_post(struct sfxge_txq *txq); 145 static void sfxge_tx_qunblock(struct sfxge_txq *txq); 146 static int sfxge_tx_queue_tso(struct sfxge_txq *txq, struct mbuf *mbuf, 151 sfxge_next_stmp(struct sfxge_txq *txq, struct sfxge_tx_mapping **pstmp) in sfxge_next_stmp() argument 155 &txq->stmp[txq->ptr_mask])) in sfxge_next_stmp() 156 *pstmp = &txq->stmp[0]; in sfxge_next_stmp() 162 sfxge_tx_maybe_toggle_cksum_offload(struct sfxge_txq *txq, struct mbuf *mbuf, in sfxge_tx_maybe_toggle_cksum_offload() argument 182 if (new_hw_cksum_flags == txq->hw_cksum_flags) in sfxge_tx_maybe_toggle_cksum_offload() 185 desc = &txq->pend_desc[txq->n_pend_desc]; in sfxge_tx_maybe_toggle_cksum_offload() [all …]
|
| H A D | sfxge_ev.c | 54 struct sfxge_txq *txq; in sfxge_ev_qcomplete() local 62 if ((txq = evq->txq) != NULL) { in sfxge_ev_qcomplete() 63 evq->txq = NULL; in sfxge_ev_qcomplete() 64 evq->txqs = &(evq->txq); in sfxge_ev_qcomplete() 69 next = txq->next; in sfxge_ev_qcomplete() 70 txq->next = NULL; in sfxge_ev_qcomplete() 72 KASSERT(txq->evq_index == index, in sfxge_ev_qcomplete() 75 if (txq->pending != txq->completed) in sfxge_ev_qcomplete() 76 sfxge_tx_qcomplete(txq, evq); in sfxge_ev_qcomplete() 78 txq = next; in sfxge_ev_qcomplete() [all …]
|
| /src/sys/contrib/dev/iwlwifi/pcie/gen1_2/ |
| H A D | tx.c | 86 struct iwl_txq *txq) in iwl_pcie_txq_inc_wr_ptr() argument 89 int txq_id = txq->id; in iwl_pcie_txq_inc_wr_ptr() 91 lockdep_assert_held(&txq->lock); in iwl_pcie_txq_inc_wr_ptr() 114 txq->need_update = true; in iwl_pcie_txq_inc_wr_ptr() 123 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr); in iwl_pcie_txq_inc_wr_ptr() 124 if (!txq->block) in iwl_pcie_txq_inc_wr_ptr() 126 txq->write_ptr | (txq_id << 8)); in iwl_pcie_txq_inc_wr_ptr() 135 struct iwl_txq *txq = trans_pcie->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() local 140 spin_lock_bh(&txq->lock); in iwl_pcie_txq_check_wrptrs() 141 if (txq->need_update) { in iwl_pcie_txq_check_wrptrs() [all …]
|
| H A D | tx-gen2.c | 299 struct iwl_txq *txq, in iwl_txq_gen2_build_tx_amsdu() argument 306 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_txq_gen2_build_tx_amsdu() 307 struct iwl_tfh_tfd *tfd = iwl_txq_get_tfd(trans, txq, idx); in iwl_txq_gen2_build_tx_amsdu() 312 tb_phys = iwl_txq_get_first_tb_dma(txq, idx); in iwl_txq_gen2_build_tx_amsdu() 348 memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE); in iwl_txq_gen2_build_tx_amsdu() 387 struct iwl_txq *txq, in iwl_txq_gen2_build_tx() argument 395 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_txq_gen2_build_tx() 396 struct iwl_tfh_tfd *tfd = iwl_txq_get_tfd(trans, txq, idx); in iwl_txq_gen2_build_tx() 402 tb_phys = iwl_txq_get_first_tb_dma(txq, idx); in iwl_txq_gen2_build_tx() 405 memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE); in iwl_txq_gen2_build_tx() [all …]
|
| /src/sys/dev/ath/ |
| H A D | if_ath_tx_edma.c | 143 ath_tx_alq_edma_push(struct ath_softc *sc, int txq, int nframes, in ath_tx_alq_edma_push() argument 148 aq.txq = htobe32(txq); in ath_tx_alq_edma_push() 171 ath_tx_edma_push_staging_list(struct ath_softc *sc, struct ath_txq *txq, in ath_tx_edma_push_staging_list() argument 179 ATH_TXQ_LOCK_ASSERT(txq); in ath_tx_edma_push_staging_list() 184 txq->axq_qnum, in ath_tx_edma_push_staging_list() 185 txq->axq_fifo_depth, in ath_tx_edma_push_staging_list() 186 !! (TAILQ_EMPTY(&txq->axq_q))); in ath_tx_edma_push_staging_list() 191 if (txq->axq_fifo_depth >= HAL_TXFIFO_DEPTH) in ath_tx_edma_push_staging_list() 194 if (TAILQ_EMPTY(&txq->axq_q)) in ath_tx_edma_push_staging_list() 206 bf = ATH_TXQ_FIRST(txq); in ath_tx_edma_push_staging_list() [all …]
|
| H A D | if_ath.c | 188 static int ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq, 3990 ath_txq_init(struct ath_softc *sc, struct ath_txq *txq, int qnum) in ath_txq_init() argument 3992 txq->axq_qnum = qnum; in ath_txq_init() 3993 txq->axq_ac = 0; in ath_txq_init() 3994 txq->axq_depth = 0; in ath_txq_init() 3995 txq->axq_aggr_depth = 0; in ath_txq_init() 3996 txq->axq_intrcnt = 0; in ath_txq_init() 3997 txq->axq_link = NULL; in ath_txq_init() 3998 txq->axq_softc = sc; in ath_txq_init() 3999 TAILQ_INIT(&txq->axq_q); in ath_txq_init() [all …]
|
| H A D | if_ath_tx.c | 697 ath_tx_handoff_mcast(struct ath_softc *sc, struct ath_txq *txq, in ath_tx_handoff_mcast() argument 713 txq->axq_qnum); in ath_tx_handoff_mcast() 716 ATH_TXQ_LOCK(txq); in ath_tx_handoff_mcast() 717 if (ATH_TXQ_LAST(txq, axq_q_s) != NULL) { in ath_tx_handoff_mcast() 718 struct ath_buf *bf_last = ATH_TXQ_LAST(txq, axq_q_s); in ath_tx_handoff_mcast() 732 ATH_TXQ_INSERT_TAIL(txq, bf, bf_list); in ath_tx_handoff_mcast() 733 ATH_TXQ_UNLOCK(txq); in ath_tx_handoff_mcast() 740 ath_tx_handoff_hw(struct ath_softc *sc, struct ath_txq *txq, in ath_tx_handoff_hw() argument 757 KASSERT(txq->axq_qnum != ATH_TXQ_SWQ, in ath_tx_handoff_hw() 777 ATH_TXQ_LOCK(txq); in ath_tx_handoff_hw() [all …]
|
| /src/sys/dev/xen/netfront/ |
| H A D | netfront.c | 206 struct netfront_txq *txq; member 222 struct netfront_txq *txq; member 357 SLIST_INSERT_HEAD(&ref->txq->tags, ref, next); in tag_free() 480 XN_TX_LOCK(&np->txq[i]); in netfront_suspend() 485 XN_TX_UNLOCK(&np->txq[i]); in netfront_suspend() 505 XN_TX_LOCK(&info->txq[i]); in netfront_resume() 510 XN_TX_UNLOCK(&info->txq[i]); in netfront_resume() 522 struct netfront_txq *txq, in write_queue_xenstore_keys() argument 531 KASSERT(rxq->id == txq->id, ("Mismatch between RX and TX queue ids")); in write_queue_xenstore_keys() 533 KASSERT(rxq->xen_intr_handle == txq->xen_intr_handle, in write_queue_xenstore_keys() [all …]
|
| /src/sys/net/ |
| H A D | iflib.c | 469 #define TXQ_AVAIL(txq) ((txq->ift_size - txq->ift_pad) -\ argument 470 get_inuse(txq->ift_size, txq->ift_cidx, txq->ift_pidx, txq->ift_gen)) 526 #define CALLOUT_LOCK(txq) mtx_lock(&txq->ift_mtx) argument 527 #define CALLOUT_UNLOCK(txq) mtx_unlock(&txq->ift_mtx) argument 677 static int iflib_tx_credits_update(if_ctx_t ctx, iflib_txq_t txq); 682 static void iflib_txq_check_drain(iflib_txq_t txq, int budget); 695 static void iflib_ifmp_purge(iflib_txq_t txq); 703 static __inline int iflib_completed_tx_reclaim(iflib_txq_t txq, 986 iflib_txq_t txq = &ctx->ifc_txqs[kring->ring_id]; in iflib_netmap_txsync() local 988 bus_dmamap_sync(txq->ift_ifdi->idi_tag, txq->ift_ifdi->idi_map, in iflib_netmap_txsync() [all …]
|
| /src/sys/dev/cxgb/ |
| H A D | cxgb_sge.c | 232 #define TXQ_RING_EMPTY(qs) drbr_empty((qs)->port->ifp, (qs)->txq[TXQ_ETH].txq_mr) 234 drbr_needs_enqueue((qs)->port->ifp, (qs)->txq[TXQ_ETH].txq_mr) 235 #define TXQ_RING_FLUSH(qs) drbr_flush((qs)->port->ifp, (qs)->txq[TXQ_ETH].txq_mr) 237 drbr_dequeue_cond((qs)->port->ifp, (qs)->txq[TXQ_ETH].txq_mr, func, arg) 239 drbr_dequeue((qs)->port->ifp, (qs)->txq[TXQ_ETH].txq_mr) 257 struct sge_txq *txq; in check_pkt_coalesce() local 262 txq = &qs->txq[TXQ_ETH]; in check_pkt_coalesce() 276 if (*fill != 0 && (txq->in_use <= cxgb_tx_coalesce_enable_stop) && in check_pkt_coalesce() 279 else if (*fill == 0 && (txq->in_use >= cxgb_tx_coalesce_enable_start)) in check_pkt_coalesce() 372 struct sge_txq *q = &qs->txq[queue]; in reclaim_completed_tx() [all …]
|
| /src/sys/dev/ice/ |
| H A D | ice_iflib_txrx.c | 50 static int _ice_ift_txd_encap(struct ice_tx_queue *txq, if_pkt_info_t pi); 51 static int _ice_ift_txd_credits_update(struct ice_softc *sc, struct ice_tx_queue *txq, bool clear); 135 _ice_ift_txd_encap(struct ice_tx_queue *txq, if_pkt_info_t pi) in _ice_ift_txd_encap() argument 152 i = ice_tso_setup(txq, pi); in _ice_ift_txd_encap() 154 ice_tx_setup_offload(txq, pi, &cmd, &off); in _ice_ift_txd_encap() 159 mask = txq->desc_count - 1; in _ice_ift_txd_encap() 163 txd = &txq->tx_base[i]; in _ice_ift_txd_encap() 174 txq->stats.tx_bytes += seglen; in _ice_ift_txd_encap() 185 txq->tx_rsq[txq->tx_rs_pidx] = pidx_last; in _ice_ift_txd_encap() 186 txq->tx_rs_pidx = (txq->tx_rs_pidx+1) & mask; in _ice_ift_txd_encap() [all …]
|
| H A D | ice_common_txrx.h | 145 ice_tso_setup(struct ice_tx_queue *txq, if_pkt_info_t pi) in ice_tso_setup() argument 153 txd = (struct ice_tx_ctx_desc *)&txq->tx_base[idx]; in ice_tso_setup() 160 txq->stats.mss_too_small++; in ice_tso_setup() 172 txq->stats.tso++; in ice_tso_setup() 174 return ((idx + 1) & (txq->desc_count-1)); in ice_tso_setup() 192 ice_tx_setup_offload(struct ice_tx_queue *txq, if_pkt_info_t pi, u32 *cmd, u32 *off) in ice_tx_setup_offload() argument 201 txq->stats.cso[ICE_CSO_STAT_TX_IP4]++; in ice_tx_setup_offload() 214 txq->stats.cso[ICE_CSO_STAT_TX_IP6]++; in ice_tx_setup_offload() 218 txq->stats.cso[ICE_CSO_STAT_TX_L3_ERR]++; in ice_tx_setup_offload() 234 txq->stats.cso[ICE_CSO_STAT_TX_TCP]++; in ice_tx_setup_offload() [all …]
|
| /src/sys/dev/virtio/network/ |
| H A D | if_vtnet.c | 867 struct vtnet_txq *txq; in vtnet_init_txq() local 869 txq = &sc->vtnet_txqs[id]; in vtnet_init_txq() 871 snprintf(txq->vtntx_name, sizeof(txq->vtntx_name), "%s-tx%d", in vtnet_init_txq() 873 mtx_init(&txq->vtntx_mtx, txq->vtntx_name, NULL, MTX_DEF); in vtnet_init_txq() 875 txq->vtntx_sc = sc; in vtnet_init_txq() 876 txq->vtntx_id = id; in vtnet_init_txq() 878 txq->vtntx_sg = sglist_alloc(sc->vtnet_tx_nsegs, M_NOWAIT); in vtnet_init_txq() 879 if (txq->vtntx_sg == NULL) in vtnet_init_txq() 883 txq->vtntx_br = buf_ring_alloc(VTNET_DEFAULT_BUFRING_SIZE, M_DEVBUF, in vtnet_init_txq() 884 M_NOWAIT, &txq->vtntx_mtx); in vtnet_init_txq() [all …]
|
| /src/sys/dev/cxgbe/crypto/ |
| H A D | t6_kern_tls.c | 72 struct sge_txq *txq; member 104 const struct ktls_session *tls, struct sge_txq *txq); 276 write_set_tcb_field_ulp(struct tlspcb *tlsp, void *dst, struct sge_txq *txq, in write_set_tcb_field_ulp() argument 288 V_ULP_TXPKT_FID(txq->eq.cntxt_id) | V_ULP_TXPKT_RO(1)); in write_set_tcb_field_ulp() 311 ktls_set_tcb_fields(struct tlspcb *tlsp, struct tcpcb *tp, struct sge_txq *txq) in ktls_set_tcb_fields() argument 340 write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_T_FLAGS, in ktls_set_tcb_fields() 346 write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_SND_UNA_RAW, in ktls_set_tcb_fields() 352 write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_SND_MAX_RAW, in ktls_set_tcb_fields() 357 write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_TIMESTAMP_OFFSET, in ktls_set_tcb_fields() 366 error = mp_ring_enqueue(txq->r, items, 1, 1); in ktls_set_tcb_fields() [all …]
|
| H A D | t7_kern_tls.c | 75 struct sge_txq *txq; member 111 const struct ktls_session *tls, struct sge_txq *txq); 145 tlsp->txq = NULL; in alloc_tlspcb() 160 struct sge_txq *txq; in t7_tls_tag_alloc() local 263 tlsp->txq = txq = &sc->sge.txq[qidx]; in t7_tls_tag_alloc() 266 error = ktls_setup_keys(tlsp, tls, txq); in t7_tls_tag_alloc() 337 TXQ_LOCK(txq); in t7_tls_tag_alloc() 339 txq->kern_tls_gcm++; in t7_tls_tag_alloc() 341 txq->kern_tls_cbc++; in t7_tls_tag_alloc() 342 TXQ_UNLOCK(txq); in t7_tls_tag_alloc() [all …]
|
| /src/sys/dev/mana/ |
| H A D | mana_en.c | 487 mana_xmit(struct mana_txq *txq) in mana_xmit() argument 491 if_t ndev = txq->ndev; in mana_xmit() 506 gdma_sq = txq->gdma_sq; in mana_xmit() 507 cq = &apc->tx_qp[txq->idx].tx_cq; in mana_xmit() 508 tx_stats = &txq->stats; in mana_xmit() 512 next_to_use = txq->next_to_use; in mana_xmit() 514 while ((mbuf = drbr_peek(ndev, txq->txq_br)) != NULL) { in mana_xmit() 517 drbr_putback(ndev, txq->txq_br, mbuf); in mana_xmit() 529 stops > wakeups && txq->alt_txq_idx == txq->idx) { in mana_xmit() 530 txq->alt_txq_idx = in mana_xmit() [all …]
|
| /src/sys/contrib/dev/iwlwifi/mld/ |
| H A D | tx.c | 48 iwl_mld_get_queue_size(struct iwl_mld *mld, struct ieee80211_txq *txq) in iwl_mld_get_queue_size() argument 50 struct ieee80211_sta *sta = txq->sta; in iwl_mld_get_queue_size() 57 for_each_sta_active_link(txq->vif, sta, link_sta, link_id) { in iwl_mld_get_queue_size() 70 static int iwl_mld_allocate_txq(struct iwl_mld *mld, struct ieee80211_txq *txq) in iwl_mld_allocate_txq() argument 72 u8 tid = txq->tid == IEEE80211_NUM_TIDS ? IWL_MGMT_TID : txq->tid; in iwl_mld_allocate_txq() 73 u32 fw_sta_mask = iwl_mld_fw_sta_id_mask(mld, txq->sta); in iwl_mld_allocate_txq() 77 unsigned int watchdog_timeout = txq->vif->type == NL80211_IFTYPE_AP ? in iwl_mld_allocate_txq() 88 size = iwl_mld_get_queue_size(mld, txq); in iwl_mld_allocate_txq() 100 static int iwl_mld_add_txq(struct iwl_mld *mld, struct ieee80211_txq *txq) in iwl_mld_add_txq() argument 102 struct iwl_mld_txq *mld_txq = iwl_mld_txq_from_mac80211(txq); in iwl_mld_add_txq() [all …]
|
| H A D | tx.h | 46 iwl_mld_txq_from_mac80211(struct ieee80211_txq *txq) in iwl_mld_txq_from_mac80211() argument 48 return (void *)txq->drv_priv; in iwl_mld_txq_from_mac80211() 52 void iwl_mld_remove_txq(struct iwl_mld *mld, struct ieee80211_txq *txq); 56 void iwl_mld_tx_from_txq(struct iwl_mld *mld, struct ieee80211_txq *txq); 60 int iwl_mld_ensure_queue(struct iwl_mld *mld, struct ieee80211_txq *txq); 75 struct ieee80211_txq *txq);
|
| /src/sys/dev/malo/ |
| H A D | if_malo.c | 548 malo_txdma_setup(struct malo_softc *sc, struct malo_txq *txq) in malo_txdma_setup() argument 554 error = malo_desc_setup(sc, "tx", &txq->dma, in malo_txdma_setup() 568 txq->dma.dd_bufptr = bf; in malo_txdma_setup() 570 STAILQ_INIT(&txq->free); in malo_txdma_setup() 571 txq->nfree = 0; in malo_txdma_setup() 572 ds = txq->dma.dd_desc; in malo_txdma_setup() 575 bf->bf_daddr = DS2PHYS(&txq->dma, ds); in malo_txdma_setup() 584 STAILQ_INSERT_TAIL(&txq->free, bf, bf_list); in malo_txdma_setup() 585 txq->nfree++; in malo_txdma_setup() 626 malo_txdma_cleanup(struct malo_softc *sc, struct malo_txq *txq) in malo_txdma_cleanup() argument [all …]
|
| /src/sys/contrib/dev/mediatek/mt76/ |
| H A D | tx.c | 9 mt76_txq_get_qid(struct ieee80211_txq *txq) in mt76_txq_get_qid() argument 11 if (!txq->sta) in mt76_txq_get_qid() 14 return txq->ac; in mt76_txq_get_qid() 21 struct ieee80211_txq *txq; in mt76_tx_check_agg_ssn() local 30 txq = sta->txq[tid]; in mt76_tx_check_agg_ssn() 31 mtxq = (struct mt76_txq *)txq->drv_priv; in mt76_tx_check_agg_ssn() 375 struct ieee80211_txq *txq = mtxq_to_txq(mtxq); in mt76_txq_dequeue() local 379 skb = ieee80211_tx_dequeue(phy->hw, txq); in mt76_txq_dequeue() 419 struct ieee80211_txq *txq = sta->txq[i]; in mt76_release_buffered_frames() local 420 struct mt76_txq *mtxq = (struct mt76_txq *)txq->drv_priv; in mt76_release_buffered_frames() [all …]
|
| /src/sys/contrib/dev/athk/ath10k/ |
| H A D | txrx.c | 50 struct ieee80211_txq *txq; in ath10k_txrx_tx_unref() local 76 txq = skb_cb->txq; in ath10k_txrx_tx_unref() 78 if (txq) { in ath10k_txrx_tx_unref() 79 artxq = (void *)txq->drv_priv; in ath10k_txrx_tx_unref() 89 if (txq && txq->sta && skb_cb->airtime_est) in ath10k_txrx_tx_unref() 90 ieee80211_sta_register_airtime(txq->sta, txq->tid, in ath10k_txrx_tx_unref() 138 if (txq) in ath10k_txrx_tx_unref() 139 status.sta = txq->sta; in ath10k_txrx_tx_unref()
|
| /src/sys/dev/iavf/ |
| H A D | iavf_vc_iflib.c | 88 vqpi->txq.vsi_id = vqci->vsi_id; in iavf_configure_queues() 89 vqpi->txq.queue_id = i; in iavf_configure_queues() 90 vqpi->txq.ring_len = scctx->isc_ntxd[0]; in iavf_configure_queues() 91 vqpi->txq.dma_ring_addr = txr->tx_paddr; in iavf_configure_queues() 94 vqpi->txq.headwb_enabled = 0; in iavf_configure_queues() 95 vqpi->txq.dma_headwb_addr = 0; in iavf_configure_queues() 97 vqpi->txq.headwb_enabled = 1; in iavf_configure_queues() 98 vqpi->txq.dma_headwb_addr = txr->tx_paddr + in iavf_configure_queues()
|
| /src/sys/dev/ral/ |
| H A D | rt2661.c | 240 error = rt2661_alloc_tx_ring(sc, &sc->txq[ac], in rt2661_attach() 324 rt2661_free_tx_ring(sc, &sc->txq[ac]); in rt2661_attach() 342 rt2661_free_tx_ring(sc, &sc->txq[0]); in rt2661_detach() 343 rt2661_free_tx_ring(sc, &sc->txq[1]); in rt2661_detach() 344 rt2661_free_tx_ring(sc, &sc->txq[2]); in rt2661_detach() 345 rt2661_free_tx_ring(sc, &sc->txq[3]); in rt2661_detach() 852 struct rt2661_tx_ring *txq; in rt2661_tx_intr() local 868 txq = (qid <= 3) ? &sc->txq[qid] : &sc->mgtq; in rt2661_tx_intr() 871 data = &txq->data[txq->stat]; in rt2661_tx_intr() 911 DPRINTFN(sc, 15, "tx done q=%d idx=%u\n", qid, txq->stat); in rt2661_tx_intr() [all …]
|
| /src/sys/dev/mwl/ |
| H A D | if_mwl.c | 974 struct mwl_txq *txq = &sc->sc_txq[i]; in mwl_setupdma() local 975 sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr; in mwl_setupdma() 1264 mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq) in mwl_gettxbuf() argument 1271 MWL_TXQ_LOCK(txq); in mwl_gettxbuf() 1272 bf = STAILQ_FIRST(&txq->free); in mwl_gettxbuf() 1274 STAILQ_REMOVE_HEAD(&txq->free, bf_list); in mwl_gettxbuf() 1275 txq->nfree--; in mwl_gettxbuf() 1277 MWL_TXQ_UNLOCK(txq); in mwl_gettxbuf() 1280 "%s: out of xmit buffers on q %d\n", __func__, txq->qnum); in mwl_gettxbuf() 1291 mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf) in mwl_puttxbuf_head() argument [all …]
|
| /src/sys/dev/cxgbe/ |
| H A D | t4_sge.c | 1154 struct sge_txq *txq; in t4_setup_vi_queues() local 1235 for_each_txq(vi, i, txq) { in t4_setup_vi_queues() 1236 rc = alloc_txq(vi, txq, i); in t4_setup_vi_queues() 1262 struct sge_txq *txq; in t4_teardown_vi_queues() local 1291 for_each_txq(vi, i, txq) { in t4_teardown_vi_queues() 1292 free_txq(vi, txq); in t4_teardown_vi_queues() 3143 set_txupdate_flags(struct sge_txq *txq, u_int avail, in set_txupdate_flags() argument 3146 struct sge_eq *eq = &txq->eq; in set_txupdate_flags() 3147 struct txpkts *txp = &txq->txp; in set_txupdate_flags() 3164 record_eth_tx_time(struct sge_txq *txq) in record_eth_tx_time() argument [all …]
|