Home
last modified time | relevance | path

Searched refs:txq (Results 1 – 25 of 149) sorted by relevance

123456

/src/sys/dev/sfxge/
H A Dsfxge_tx.c143 static void sfxge_tx_qdpl_service(struct sfxge_txq *txq);
144 static void sfxge_tx_qlist_post(struct sfxge_txq *txq);
145 static void sfxge_tx_qunblock(struct sfxge_txq *txq);
146 static int sfxge_tx_queue_tso(struct sfxge_txq *txq, struct mbuf *mbuf,
151 sfxge_next_stmp(struct sfxge_txq *txq, struct sfxge_tx_mapping **pstmp) in sfxge_next_stmp() argument
155 &txq->stmp[txq->ptr_mask])) in sfxge_next_stmp()
156 *pstmp = &txq->stmp[0]; in sfxge_next_stmp()
162 sfxge_tx_maybe_toggle_cksum_offload(struct sfxge_txq *txq, struct mbuf *mbuf, in sfxge_tx_maybe_toggle_cksum_offload() argument
182 if (new_hw_cksum_flags == txq->hw_cksum_flags) in sfxge_tx_maybe_toggle_cksum_offload()
185 desc = &txq->pend_desc[txq->n_pend_desc]; in sfxge_tx_maybe_toggle_cksum_offload()
[all …]
H A Dsfxge_ev.c54 struct sfxge_txq *txq; in sfxge_ev_qcomplete() local
62 if ((txq = evq->txq) != NULL) { in sfxge_ev_qcomplete()
63 evq->txq = NULL; in sfxge_ev_qcomplete()
64 evq->txqs = &(evq->txq); in sfxge_ev_qcomplete()
69 next = txq->next; in sfxge_ev_qcomplete()
70 txq->next = NULL; in sfxge_ev_qcomplete()
72 KASSERT(txq->evq_index == index, in sfxge_ev_qcomplete()
75 if (txq->pending != txq->completed) in sfxge_ev_qcomplete()
76 sfxge_tx_qcomplete(txq, evq); in sfxge_ev_qcomplete()
78 txq = next; in sfxge_ev_qcomplete()
[all …]
/src/sys/contrib/dev/iwlwifi/pcie/gen1_2/
H A Dtx.c86 struct iwl_txq *txq) in iwl_pcie_txq_inc_wr_ptr() argument
89 int txq_id = txq->id; in iwl_pcie_txq_inc_wr_ptr()
91 lockdep_assert_held(&txq->lock); in iwl_pcie_txq_inc_wr_ptr()
114 txq->need_update = true; in iwl_pcie_txq_inc_wr_ptr()
123 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr); in iwl_pcie_txq_inc_wr_ptr()
124 if (!txq->block) in iwl_pcie_txq_inc_wr_ptr()
126 txq->write_ptr | (txq_id << 8)); in iwl_pcie_txq_inc_wr_ptr()
135 struct iwl_txq *txq = trans_pcie->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() local
140 spin_lock_bh(&txq->lock); in iwl_pcie_txq_check_wrptrs()
141 if (txq->need_update) { in iwl_pcie_txq_check_wrptrs()
[all …]
H A Dtx-gen2.c299 struct iwl_txq *txq, in iwl_txq_gen2_build_tx_amsdu() argument
306 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_txq_gen2_build_tx_amsdu()
307 struct iwl_tfh_tfd *tfd = iwl_txq_get_tfd(trans, txq, idx); in iwl_txq_gen2_build_tx_amsdu()
312 tb_phys = iwl_txq_get_first_tb_dma(txq, idx); in iwl_txq_gen2_build_tx_amsdu()
348 memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE); in iwl_txq_gen2_build_tx_amsdu()
387 struct iwl_txq *txq, in iwl_txq_gen2_build_tx() argument
395 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_txq_gen2_build_tx()
396 struct iwl_tfh_tfd *tfd = iwl_txq_get_tfd(trans, txq, idx); in iwl_txq_gen2_build_tx()
402 tb_phys = iwl_txq_get_first_tb_dma(txq, idx); in iwl_txq_gen2_build_tx()
405 memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE); in iwl_txq_gen2_build_tx()
[all …]
/src/sys/dev/ath/
H A Dif_ath_tx_edma.c143 ath_tx_alq_edma_push(struct ath_softc *sc, int txq, int nframes, in ath_tx_alq_edma_push() argument
148 aq.txq = htobe32(txq); in ath_tx_alq_edma_push()
171 ath_tx_edma_push_staging_list(struct ath_softc *sc, struct ath_txq *txq, in ath_tx_edma_push_staging_list() argument
179 ATH_TXQ_LOCK_ASSERT(txq); in ath_tx_edma_push_staging_list()
184 txq->axq_qnum, in ath_tx_edma_push_staging_list()
185 txq->axq_fifo_depth, in ath_tx_edma_push_staging_list()
186 !! (TAILQ_EMPTY(&txq->axq_q))); in ath_tx_edma_push_staging_list()
191 if (txq->axq_fifo_depth >= HAL_TXFIFO_DEPTH) in ath_tx_edma_push_staging_list()
194 if (TAILQ_EMPTY(&txq->axq_q)) in ath_tx_edma_push_staging_list()
206 bf = ATH_TXQ_FIRST(txq); in ath_tx_edma_push_staging_list()
[all …]
H A Dif_ath.c188 static int ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq,
3990 ath_txq_init(struct ath_softc *sc, struct ath_txq *txq, int qnum) in ath_txq_init() argument
3992 txq->axq_qnum = qnum; in ath_txq_init()
3993 txq->axq_ac = 0; in ath_txq_init()
3994 txq->axq_depth = 0; in ath_txq_init()
3995 txq->axq_aggr_depth = 0; in ath_txq_init()
3996 txq->axq_intrcnt = 0; in ath_txq_init()
3997 txq->axq_link = NULL; in ath_txq_init()
3998 txq->axq_softc = sc; in ath_txq_init()
3999 TAILQ_INIT(&txq->axq_q); in ath_txq_init()
[all …]
H A Dif_ath_tx.c697 ath_tx_handoff_mcast(struct ath_softc *sc, struct ath_txq *txq, in ath_tx_handoff_mcast() argument
713 txq->axq_qnum); in ath_tx_handoff_mcast()
716 ATH_TXQ_LOCK(txq); in ath_tx_handoff_mcast()
717 if (ATH_TXQ_LAST(txq, axq_q_s) != NULL) { in ath_tx_handoff_mcast()
718 struct ath_buf *bf_last = ATH_TXQ_LAST(txq, axq_q_s); in ath_tx_handoff_mcast()
732 ATH_TXQ_INSERT_TAIL(txq, bf, bf_list); in ath_tx_handoff_mcast()
733 ATH_TXQ_UNLOCK(txq); in ath_tx_handoff_mcast()
740 ath_tx_handoff_hw(struct ath_softc *sc, struct ath_txq *txq, in ath_tx_handoff_hw() argument
757 KASSERT(txq->axq_qnum != ATH_TXQ_SWQ, in ath_tx_handoff_hw()
777 ATH_TXQ_LOCK(txq); in ath_tx_handoff_hw()
[all …]
/src/sys/dev/xen/netfront/
H A Dnetfront.c206 struct netfront_txq *txq; member
222 struct netfront_txq *txq; member
357 SLIST_INSERT_HEAD(&ref->txq->tags, ref, next); in tag_free()
480 XN_TX_LOCK(&np->txq[i]); in netfront_suspend()
485 XN_TX_UNLOCK(&np->txq[i]); in netfront_suspend()
505 XN_TX_LOCK(&info->txq[i]); in netfront_resume()
510 XN_TX_UNLOCK(&info->txq[i]); in netfront_resume()
522 struct netfront_txq *txq, in write_queue_xenstore_keys() argument
531 KASSERT(rxq->id == txq->id, ("Mismatch between RX and TX queue ids")); in write_queue_xenstore_keys()
533 KASSERT(rxq->xen_intr_handle == txq->xen_intr_handle, in write_queue_xenstore_keys()
[all …]
/src/sys/net/
H A Diflib.c469 #define TXQ_AVAIL(txq) ((txq->ift_size - txq->ift_pad) -\ argument
470 get_inuse(txq->ift_size, txq->ift_cidx, txq->ift_pidx, txq->ift_gen))
526 #define CALLOUT_LOCK(txq) mtx_lock(&txq->ift_mtx) argument
527 #define CALLOUT_UNLOCK(txq) mtx_unlock(&txq->ift_mtx) argument
677 static int iflib_tx_credits_update(if_ctx_t ctx, iflib_txq_t txq);
682 static void iflib_txq_check_drain(iflib_txq_t txq, int budget);
695 static void iflib_ifmp_purge(iflib_txq_t txq);
703 static __inline int iflib_completed_tx_reclaim(iflib_txq_t txq,
986 iflib_txq_t txq = &ctx->ifc_txqs[kring->ring_id]; in iflib_netmap_txsync() local
988 bus_dmamap_sync(txq->ift_ifdi->idi_tag, txq->ift_ifdi->idi_map, in iflib_netmap_txsync()
[all …]
/src/sys/dev/cxgb/
H A Dcxgb_sge.c232 #define TXQ_RING_EMPTY(qs) drbr_empty((qs)->port->ifp, (qs)->txq[TXQ_ETH].txq_mr)
234 drbr_needs_enqueue((qs)->port->ifp, (qs)->txq[TXQ_ETH].txq_mr)
235 #define TXQ_RING_FLUSH(qs) drbr_flush((qs)->port->ifp, (qs)->txq[TXQ_ETH].txq_mr)
237 drbr_dequeue_cond((qs)->port->ifp, (qs)->txq[TXQ_ETH].txq_mr, func, arg)
239 drbr_dequeue((qs)->port->ifp, (qs)->txq[TXQ_ETH].txq_mr)
257 struct sge_txq *txq; in check_pkt_coalesce() local
262 txq = &qs->txq[TXQ_ETH]; in check_pkt_coalesce()
276 if (*fill != 0 && (txq->in_use <= cxgb_tx_coalesce_enable_stop) && in check_pkt_coalesce()
279 else if (*fill == 0 && (txq->in_use >= cxgb_tx_coalesce_enable_start)) in check_pkt_coalesce()
372 struct sge_txq *q = &qs->txq[queue]; in reclaim_completed_tx()
[all …]
/src/sys/dev/ice/
H A Dice_iflib_txrx.c50 static int _ice_ift_txd_encap(struct ice_tx_queue *txq, if_pkt_info_t pi);
51 static int _ice_ift_txd_credits_update(struct ice_softc *sc, struct ice_tx_queue *txq, bool clear);
135 _ice_ift_txd_encap(struct ice_tx_queue *txq, if_pkt_info_t pi) in _ice_ift_txd_encap() argument
152 i = ice_tso_setup(txq, pi); in _ice_ift_txd_encap()
154 ice_tx_setup_offload(txq, pi, &cmd, &off); in _ice_ift_txd_encap()
159 mask = txq->desc_count - 1; in _ice_ift_txd_encap()
163 txd = &txq->tx_base[i]; in _ice_ift_txd_encap()
174 txq->stats.tx_bytes += seglen; in _ice_ift_txd_encap()
185 txq->tx_rsq[txq->tx_rs_pidx] = pidx_last; in _ice_ift_txd_encap()
186 txq->tx_rs_pidx = (txq->tx_rs_pidx+1) & mask; in _ice_ift_txd_encap()
[all …]
H A Dice_common_txrx.h145 ice_tso_setup(struct ice_tx_queue *txq, if_pkt_info_t pi) in ice_tso_setup() argument
153 txd = (struct ice_tx_ctx_desc *)&txq->tx_base[idx]; in ice_tso_setup()
160 txq->stats.mss_too_small++; in ice_tso_setup()
172 txq->stats.tso++; in ice_tso_setup()
174 return ((idx + 1) & (txq->desc_count-1)); in ice_tso_setup()
192 ice_tx_setup_offload(struct ice_tx_queue *txq, if_pkt_info_t pi, u32 *cmd, u32 *off) in ice_tx_setup_offload() argument
201 txq->stats.cso[ICE_CSO_STAT_TX_IP4]++; in ice_tx_setup_offload()
214 txq->stats.cso[ICE_CSO_STAT_TX_IP6]++; in ice_tx_setup_offload()
218 txq->stats.cso[ICE_CSO_STAT_TX_L3_ERR]++; in ice_tx_setup_offload()
234 txq->stats.cso[ICE_CSO_STAT_TX_TCP]++; in ice_tx_setup_offload()
[all …]
/src/sys/dev/virtio/network/
H A Dif_vtnet.c867 struct vtnet_txq *txq; in vtnet_init_txq() local
869 txq = &sc->vtnet_txqs[id]; in vtnet_init_txq()
871 snprintf(txq->vtntx_name, sizeof(txq->vtntx_name), "%s-tx%d", in vtnet_init_txq()
873 mtx_init(&txq->vtntx_mtx, txq->vtntx_name, NULL, MTX_DEF); in vtnet_init_txq()
875 txq->vtntx_sc = sc; in vtnet_init_txq()
876 txq->vtntx_id = id; in vtnet_init_txq()
878 txq->vtntx_sg = sglist_alloc(sc->vtnet_tx_nsegs, M_NOWAIT); in vtnet_init_txq()
879 if (txq->vtntx_sg == NULL) in vtnet_init_txq()
883 txq->vtntx_br = buf_ring_alloc(VTNET_DEFAULT_BUFRING_SIZE, M_DEVBUF, in vtnet_init_txq()
884 M_NOWAIT, &txq->vtntx_mtx); in vtnet_init_txq()
[all …]
/src/sys/dev/cxgbe/crypto/
H A Dt6_kern_tls.c72 struct sge_txq *txq; member
104 const struct ktls_session *tls, struct sge_txq *txq);
276 write_set_tcb_field_ulp(struct tlspcb *tlsp, void *dst, struct sge_txq *txq, in write_set_tcb_field_ulp() argument
288 V_ULP_TXPKT_FID(txq->eq.cntxt_id) | V_ULP_TXPKT_RO(1)); in write_set_tcb_field_ulp()
311 ktls_set_tcb_fields(struct tlspcb *tlsp, struct tcpcb *tp, struct sge_txq *txq) in ktls_set_tcb_fields() argument
340 write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_T_FLAGS, in ktls_set_tcb_fields()
346 write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_SND_UNA_RAW, in ktls_set_tcb_fields()
352 write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_SND_MAX_RAW, in ktls_set_tcb_fields()
357 write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_TIMESTAMP_OFFSET, in ktls_set_tcb_fields()
366 error = mp_ring_enqueue(txq->r, items, 1, 1); in ktls_set_tcb_fields()
[all …]
H A Dt7_kern_tls.c75 struct sge_txq *txq; member
111 const struct ktls_session *tls, struct sge_txq *txq);
145 tlsp->txq = NULL; in alloc_tlspcb()
160 struct sge_txq *txq; in t7_tls_tag_alloc() local
263 tlsp->txq = txq = &sc->sge.txq[qidx]; in t7_tls_tag_alloc()
266 error = ktls_setup_keys(tlsp, tls, txq); in t7_tls_tag_alloc()
337 TXQ_LOCK(txq); in t7_tls_tag_alloc()
339 txq->kern_tls_gcm++; in t7_tls_tag_alloc()
341 txq->kern_tls_cbc++; in t7_tls_tag_alloc()
342 TXQ_UNLOCK(txq); in t7_tls_tag_alloc()
[all …]
/src/sys/dev/mana/
H A Dmana_en.c487 mana_xmit(struct mana_txq *txq) in mana_xmit() argument
491 if_t ndev = txq->ndev; in mana_xmit()
506 gdma_sq = txq->gdma_sq; in mana_xmit()
507 cq = &apc->tx_qp[txq->idx].tx_cq; in mana_xmit()
508 tx_stats = &txq->stats; in mana_xmit()
512 next_to_use = txq->next_to_use; in mana_xmit()
514 while ((mbuf = drbr_peek(ndev, txq->txq_br)) != NULL) { in mana_xmit()
517 drbr_putback(ndev, txq->txq_br, mbuf); in mana_xmit()
529 stops > wakeups && txq->alt_txq_idx == txq->idx) { in mana_xmit()
530 txq->alt_txq_idx = in mana_xmit()
[all …]
/src/sys/contrib/dev/iwlwifi/mld/
H A Dtx.c48 iwl_mld_get_queue_size(struct iwl_mld *mld, struct ieee80211_txq *txq) in iwl_mld_get_queue_size() argument
50 struct ieee80211_sta *sta = txq->sta; in iwl_mld_get_queue_size()
57 for_each_sta_active_link(txq->vif, sta, link_sta, link_id) { in iwl_mld_get_queue_size()
70 static int iwl_mld_allocate_txq(struct iwl_mld *mld, struct ieee80211_txq *txq) in iwl_mld_allocate_txq() argument
72 u8 tid = txq->tid == IEEE80211_NUM_TIDS ? IWL_MGMT_TID : txq->tid; in iwl_mld_allocate_txq()
73 u32 fw_sta_mask = iwl_mld_fw_sta_id_mask(mld, txq->sta); in iwl_mld_allocate_txq()
77 unsigned int watchdog_timeout = txq->vif->type == NL80211_IFTYPE_AP ? in iwl_mld_allocate_txq()
88 size = iwl_mld_get_queue_size(mld, txq); in iwl_mld_allocate_txq()
100 static int iwl_mld_add_txq(struct iwl_mld *mld, struct ieee80211_txq *txq) in iwl_mld_add_txq() argument
102 struct iwl_mld_txq *mld_txq = iwl_mld_txq_from_mac80211(txq); in iwl_mld_add_txq()
[all …]
H A Dtx.h46 iwl_mld_txq_from_mac80211(struct ieee80211_txq *txq) in iwl_mld_txq_from_mac80211() argument
48 return (void *)txq->drv_priv; in iwl_mld_txq_from_mac80211()
52 void iwl_mld_remove_txq(struct iwl_mld *mld, struct ieee80211_txq *txq);
56 void iwl_mld_tx_from_txq(struct iwl_mld *mld, struct ieee80211_txq *txq);
60 int iwl_mld_ensure_queue(struct iwl_mld *mld, struct ieee80211_txq *txq);
75 struct ieee80211_txq *txq);
/src/sys/dev/malo/
H A Dif_malo.c548 malo_txdma_setup(struct malo_softc *sc, struct malo_txq *txq) in malo_txdma_setup() argument
554 error = malo_desc_setup(sc, "tx", &txq->dma, in malo_txdma_setup()
568 txq->dma.dd_bufptr = bf; in malo_txdma_setup()
570 STAILQ_INIT(&txq->free); in malo_txdma_setup()
571 txq->nfree = 0; in malo_txdma_setup()
572 ds = txq->dma.dd_desc; in malo_txdma_setup()
575 bf->bf_daddr = DS2PHYS(&txq->dma, ds); in malo_txdma_setup()
584 STAILQ_INSERT_TAIL(&txq->free, bf, bf_list); in malo_txdma_setup()
585 txq->nfree++; in malo_txdma_setup()
626 malo_txdma_cleanup(struct malo_softc *sc, struct malo_txq *txq) in malo_txdma_cleanup() argument
[all …]
/src/sys/contrib/dev/mediatek/mt76/
H A Dtx.c9 mt76_txq_get_qid(struct ieee80211_txq *txq) in mt76_txq_get_qid() argument
11 if (!txq->sta) in mt76_txq_get_qid()
14 return txq->ac; in mt76_txq_get_qid()
21 struct ieee80211_txq *txq; in mt76_tx_check_agg_ssn() local
30 txq = sta->txq[tid]; in mt76_tx_check_agg_ssn()
31 mtxq = (struct mt76_txq *)txq->drv_priv; in mt76_tx_check_agg_ssn()
375 struct ieee80211_txq *txq = mtxq_to_txq(mtxq); in mt76_txq_dequeue() local
379 skb = ieee80211_tx_dequeue(phy->hw, txq); in mt76_txq_dequeue()
419 struct ieee80211_txq *txq = sta->txq[i]; in mt76_release_buffered_frames() local
420 struct mt76_txq *mtxq = (struct mt76_txq *)txq->drv_priv; in mt76_release_buffered_frames()
[all …]
/src/sys/contrib/dev/athk/ath10k/
H A Dtxrx.c50 struct ieee80211_txq *txq; in ath10k_txrx_tx_unref() local
76 txq = skb_cb->txq; in ath10k_txrx_tx_unref()
78 if (txq) { in ath10k_txrx_tx_unref()
79 artxq = (void *)txq->drv_priv; in ath10k_txrx_tx_unref()
89 if (txq && txq->sta && skb_cb->airtime_est) in ath10k_txrx_tx_unref()
90 ieee80211_sta_register_airtime(txq->sta, txq->tid, in ath10k_txrx_tx_unref()
138 if (txq) in ath10k_txrx_tx_unref()
139 status.sta = txq->sta; in ath10k_txrx_tx_unref()
/src/sys/dev/iavf/
H A Diavf_vc_iflib.c88 vqpi->txq.vsi_id = vqci->vsi_id; in iavf_configure_queues()
89 vqpi->txq.queue_id = i; in iavf_configure_queues()
90 vqpi->txq.ring_len = scctx->isc_ntxd[0]; in iavf_configure_queues()
91 vqpi->txq.dma_ring_addr = txr->tx_paddr; in iavf_configure_queues()
94 vqpi->txq.headwb_enabled = 0; in iavf_configure_queues()
95 vqpi->txq.dma_headwb_addr = 0; in iavf_configure_queues()
97 vqpi->txq.headwb_enabled = 1; in iavf_configure_queues()
98 vqpi->txq.dma_headwb_addr = txr->tx_paddr + in iavf_configure_queues()
/src/sys/dev/ral/
H A Drt2661.c240 error = rt2661_alloc_tx_ring(sc, &sc->txq[ac], in rt2661_attach()
324 rt2661_free_tx_ring(sc, &sc->txq[ac]); in rt2661_attach()
342 rt2661_free_tx_ring(sc, &sc->txq[0]); in rt2661_detach()
343 rt2661_free_tx_ring(sc, &sc->txq[1]); in rt2661_detach()
344 rt2661_free_tx_ring(sc, &sc->txq[2]); in rt2661_detach()
345 rt2661_free_tx_ring(sc, &sc->txq[3]); in rt2661_detach()
852 struct rt2661_tx_ring *txq; in rt2661_tx_intr() local
868 txq = (qid <= 3) ? &sc->txq[qid] : &sc->mgtq; in rt2661_tx_intr()
871 data = &txq->data[txq->stat]; in rt2661_tx_intr()
911 DPRINTFN(sc, 15, "tx done q=%d idx=%u\n", qid, txq->stat); in rt2661_tx_intr()
[all …]
/src/sys/dev/mwl/
H A Dif_mwl.c974 struct mwl_txq *txq = &sc->sc_txq[i]; in mwl_setupdma() local
975 sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr; in mwl_setupdma()
1264 mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq) in mwl_gettxbuf() argument
1271 MWL_TXQ_LOCK(txq); in mwl_gettxbuf()
1272 bf = STAILQ_FIRST(&txq->free); in mwl_gettxbuf()
1274 STAILQ_REMOVE_HEAD(&txq->free, bf_list); in mwl_gettxbuf()
1275 txq->nfree--; in mwl_gettxbuf()
1277 MWL_TXQ_UNLOCK(txq); in mwl_gettxbuf()
1280 "%s: out of xmit buffers on q %d\n", __func__, txq->qnum); in mwl_gettxbuf()
1291 mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf) in mwl_puttxbuf_head() argument
[all …]
/src/sys/dev/cxgbe/
H A Dt4_sge.c1154 struct sge_txq *txq; in t4_setup_vi_queues() local
1235 for_each_txq(vi, i, txq) { in t4_setup_vi_queues()
1236 rc = alloc_txq(vi, txq, i); in t4_setup_vi_queues()
1262 struct sge_txq *txq; in t4_teardown_vi_queues() local
1291 for_each_txq(vi, i, txq) { in t4_teardown_vi_queues()
1292 free_txq(vi, txq); in t4_teardown_vi_queues()
3143 set_txupdate_flags(struct sge_txq *txq, u_int avail, in set_txupdate_flags() argument
3146 struct sge_eq *eq = &txq->eq; in set_txupdate_flags()
3147 struct txpkts *txp = &txq->txp; in set_txupdate_flags()
3164 record_eth_tx_time(struct sge_txq *txq) in record_eth_tx_time() argument
[all …]

123456