Home
last modified time | relevance | path

Searched refs:tx_q (Results 1 – 25 of 33) sorted by relevance

12

/linux/drivers/net/ethernet/stmicro/stmmac/
H A Dring_mode.c17 static int jumbo_frm(struct stmmac_tx_queue *tx_q, struct sk_buff *skb, in jumbo_frm() argument
21 struct stmmac_priv *priv = tx_q->priv_data; in jumbo_frm()
22 unsigned int entry = tx_q->cur_tx; in jumbo_frm()
27 desc = (struct dma_desc *)(tx_q->dma_etx + entry); in jumbo_frm()
29 desc = tx_q->dma_tx + entry; in jumbo_frm()
46 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
47 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm()
48 tx_q->tx_skbuff_dma[entry].is_jumbo = true; in jumbo_frm()
53 tx_q->tx_skbuff[entry] = NULL; in jumbo_frm()
57 desc = (struct dma_desc *)(tx_q in jumbo_frm()
119 clean_desc3(struct stmmac_tx_queue * tx_q,struct dma_desc * p) clean_desc3() argument
[all...]
H A Dchain_mode.c17 static int jumbo_frm(struct stmmac_tx_queue *tx_q, struct sk_buff *skb, in jumbo_frm() argument
21 struct stmmac_priv *priv = tx_q->priv_data; in jumbo_frm()
22 unsigned int entry = tx_q->cur_tx; in jumbo_frm()
27 desc = tx_q->dma_tx + entry; in jumbo_frm()
41 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
42 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm()
48 tx_q->tx_skbuff[entry] = NULL; in jumbo_frm()
50 desc = tx_q->dma_tx + entry; in jumbo_frm()
59 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
60 tx_q in jumbo_frm()
143 clean_desc3(struct stmmac_tx_queue * tx_q,struct dma_desc * p) clean_desc3() argument
[all...]
H A Dstmmac_main.c394 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_tx_avail() local
397 if (tx_q->dirty_tx > tx_q->cur_tx) in stmmac_tx_avail()
398 avail = tx_q->dirty_tx - tx_q->cur_tx - 1; in stmmac_tx_avail()
400 avail = priv->dma_conf.dma_tx_size - tx_q->cur_tx + tx_q->dirty_tx - 1; in stmmac_tx_avail()
430 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_eee_tx_busy() local
432 if (tx_q->dirty_tx != tx_q in stmmac_eee_tx_busy()
1381 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; stmmac_display_tx_rings() local
1478 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; stmmac_clear_tx_descriptors() local
1603 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; stmmac_free_tx_buffer() local
1874 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; __init_dma_tx_desc_rings() local
1972 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; dma_free_tx_skbufs() local
2060 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; __free_dma_tx_desc_resources() local
2226 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; __alloc_dma_tx_desc_resources() local
2585 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; stmmac_xdp_xmit_zc() local
2730 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; stmmac_tx_clean() local
2921 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; stmmac_tx_err() local
2989 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; stmmac_napi_check() local
3122 struct stmmac_tx_queue *tx_q; stmmac_init_dma_engine() local
3183 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; stmmac_tx_timer_arm() local
3214 struct stmmac_tx_queue *tx_q = container_of(t, struct stmmac_tx_queue, txtimer); stmmac_tx_timer() local
3249 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; stmmac_init_coalesce() local
3597 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; stmmac_hw_setup() local
3619 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; stmmac_hw_setup() local
3993 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[chan]; stmmac_setup_dma_desc() local
4179 stmmac_vlan_insert(struct stmmac_priv * priv,struct sk_buff * skb,struct stmmac_tx_queue * tx_q) stmmac_vlan_insert() argument
4223 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; stmmac_tso_allocator() local
4258 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; stmmac_flush_tx_descriptors() local
4321 struct stmmac_tx_queue *tx_q; stmmac_tso_xmit() local
4595 struct stmmac_tx_queue *tx_q; stmmac_xmit() local
4985 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; stmmac_xdp_xmit_xdpf() local
6168 struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)data; stmmac_msi_intr_tx() local
6391 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; stmmac_rings_status_show() local
6886 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; stmmac_enable_tx_queue() local
6964 struct stmmac_tx_queue *tx_q; stmmac_xdp_open() local
7069 struct stmmac_tx_queue *tx_q; stmmac_xsk_wakeup() local
7896 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; stmmac_reset_tx_queue() local
[all...]
/linux/drivers/net/ethernet/qualcomm/emac/
H A Demac-mac.c305 writel(upper_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config()
308 writel(lower_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config()
311 writel(adpt->tx_q.tpd.count & TPD_RING_SIZE_BMSK, in emac_mac_dma_rings_config()
587 struct emac_tx_queue *tx_q = &adpt->tx_q; in emac_tx_q_descs_free() local
592 if (!tx_q->tpd.tpbuff) in emac_tx_q_descs_free()
595 for (i = 0; i < tx_q->tpd.count; i++) { in emac_tx_q_descs_free()
596 struct emac_buffer *tpbuf = GET_TPD_BUFFER(tx_q, i); in emac_tx_q_descs_free()
610 size = sizeof(struct emac_buffer) * tx_q->tpd.count; in emac_tx_q_descs_free()
611 memset(tx_q in emac_tx_q_descs_free()
662 struct emac_tx_queue *tx_q = &adpt->tx_q; emac_tx_q_bufs_free() local
675 emac_tx_q_desc_alloc(struct emac_adapter * adpt,struct emac_tx_queue * tx_q) emac_tx_q_desc_alloc() argument
1026 emac_tx_tpd_create(struct emac_adapter * adpt,struct emac_tx_queue * tx_q,struct emac_tpd * tpd) emac_tx_tpd_create() argument
1044 emac_tx_tpd_mark_last(struct emac_adapter * adpt,struct emac_tx_queue * tx_q) emac_tx_tpd_mark_last() argument
1168 emac_tpd_num_free_descs(struct emac_tx_queue * tx_q) emac_tpd_num_free_descs() argument
1179 emac_mac_tx_process(struct emac_adapter * adpt,struct emac_tx_queue * tx_q) emac_mac_tx_process() argument
1246 emac_tso_csum(struct emac_adapter * adpt,struct emac_tx_queue * tx_q,struct sk_buff * skb,struct emac_tpd * tpd) emac_tso_csum() argument
1331 emac_tx_fill_tpd(struct emac_adapter * adpt,struct emac_tx_queue * tx_q,struct sk_buff * skb,struct emac_tpd * tpd) emac_tx_fill_tpd() argument
1438 emac_mac_tx_buf_send(struct emac_adapter * adpt,struct emac_tx_queue * tx_q,struct sk_buff * skb) emac_mac_tx_buf_send() argument
[all...]
H A Demac-mac.h231 struct emac_tx_queue *tx_q,
233 void emac_mac_tx_process(struct emac_adapter *adpt, struct emac_tx_queue *tx_q);
H A Demac.h341 struct emac_tx_queue tx_q; member
H A Demac.c122 return emac_mac_tx_buf_send(adpt, &adpt->tx_q, skb); in emac_start_xmit()
160 emac_mac_tx_process(adpt, &adpt->tx_q); in emac_isr()
/linux/drivers/net/ethernet/intel/idpf/
H A Didpf_txrx.c192 * @tx_q: queue for which the buffers are allocated
196 static int idpf_tx_buf_alloc_all(struct idpf_tx_queue *tx_q) in idpf_tx_buf_alloc_all() argument
205 buf_size = sizeof(struct idpf_tx_buf) * tx_q->desc_count; in idpf_tx_buf_alloc_all()
206 tx_q->tx_buf = kzalloc(buf_size, GFP_KERNEL); in idpf_tx_buf_alloc_all()
207 if (!tx_q->tx_buf) in idpf_tx_buf_alloc_all()
210 if (!idpf_queue_has(FLOW_SCH_EN, tx_q)) in idpf_tx_buf_alloc_all()
213 buf_stack = &tx_q->stash->buf_stack; in idpf_tx_buf_alloc_all()
218 buf_stack->bufs = kcalloc(tx_q->desc_count, sizeof(*buf_stack->bufs), in idpf_tx_buf_alloc_all()
223 buf_stack->size = tx_q->desc_count; in idpf_tx_buf_alloc_all()
224 buf_stack->top = tx_q in idpf_tx_buf_alloc_all()
244 idpf_tx_desc_alloc(const struct idpf_vport * vport,struct idpf_tx_queue * tx_q) idpf_tx_desc_alloc() argument
1644 idpf_tx_handle_sw_marker(struct idpf_tx_queue * tx_q) idpf_tx_handle_sw_marker() argument
1815 idpf_tx_splitq_clean(struct idpf_tx_queue * tx_q,u16 end,int napi_budget,struct libeth_sq_napi_stats * cleaned,bool descs_only) idpf_tx_splitq_clean() argument
2039 struct idpf_tx_queue *tx_q; idpf_tx_clean_complq() local
2116 struct idpf_tx_queue *tx_q = complq->txq_grp->txqs[i]; idpf_tx_clean_complq() local
2190 idpf_txq_has_room(struct idpf_tx_queue * tx_q,u32 size) idpf_txq_has_room() argument
2207 idpf_tx_maybe_stop_splitq(struct idpf_tx_queue * tx_q,unsigned int descs_needed) idpf_tx_maybe_stop_splitq() argument
2232 idpf_tx_buf_hw_update(struct idpf_tx_queue * tx_q,u32 val,bool xmit_more) idpf_tx_buf_hw_update() argument
2377 idpf_tx_splitq_map(struct idpf_tx_queue * tx_q,struct idpf_tx_splitq_params * params,struct idpf_tx_buf * first) idpf_tx_splitq_map() argument
2763 idpf_tx_drop_skb(struct idpf_tx_queue * tx_q,struct sk_buff * skb) idpf_tx_drop_skb() argument
2785 idpf_tx_tstamp(struct idpf_tx_queue * tx_q,struct sk_buff * skb,struct idpf_tx_offload_params * off) idpf_tx_tstamp() argument
2832 idpf_tx_tstamp(struct idpf_tx_queue * tx_q,struct sk_buff * skb,struct idpf_tx_offload_params * off) idpf_tx_tstamp() argument
2851 idpf_tx_splitq_frame(struct sk_buff * skb,struct idpf_tx_queue * tx_q) idpf_tx_splitq_frame() argument
2953 struct idpf_tx_queue *tx_q; idpf_tx_start() local
[all...]
H A Didpf_ptp.c630 * @tx_q: Transmit queue on which the Tx timestamp is requested
640 int idpf_ptp_request_ts(struct idpf_tx_queue *tx_q, struct sk_buff *skb, in idpf_ptp_request_ts() argument
647 spin_lock(&tx_q->cached_tstamp_caps->latches_lock); in idpf_ptp_request_ts()
649 head = &tx_q->cached_tstamp_caps->latches_free; in idpf_ptp_request_ts()
651 spin_unlock(&tx_q->cached_tstamp_caps->latches_lock); in idpf_ptp_request_ts()
664 &tx_q->cached_tstamp_caps->latches_in_use); in idpf_ptp_request_ts()
665 spin_unlock(&tx_q->cached_tstamp_caps->latches_lock); in idpf_ptp_request_ts()
H A Didpf_ptp.h286 int idpf_ptp_request_ts(struct idpf_tx_queue *tx_q, struct sk_buff *skb,
371 static inline int idpf_ptp_request_ts(struct idpf_tx_queue *tx_q, in idpf_ptp_request_ts() argument
H A Didpf_txrx.h1037 void idpf_tx_buf_hw_update(struct idpf_tx_queue *tx_q, u32 val,
1040 netdev_tx_t idpf_tx_drop_skb(struct idpf_tx_queue *tx_q, struct sk_buff *skb);
1047 struct idpf_tx_queue *tx_q);
/linux/drivers/net/wwan/t7xx/
H A Dt7xx_hif_dpmaif.c211 struct dpmaif_tx_queue *tx_q; in t7xx_dpmaif_rxtx_sw_allocs() local
236 tx_q = &dpmaif_ctrl->txq[tx_idx]; in t7xx_dpmaif_rxtx_sw_allocs()
237 tx_q->index = tx_idx; in t7xx_dpmaif_rxtx_sw_allocs()
238 tx_q->dpmaif_ctrl = dpmaif_ctrl; in t7xx_dpmaif_rxtx_sw_allocs()
239 ret = t7xx_dpmaif_txq_init(tx_q); in t7xx_dpmaif_rxtx_sw_allocs()
261 tx_q = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_rxtx_sw_allocs()
262 t7xx_dpmaif_txq_free(tx_q); in t7xx_dpmaif_rxtx_sw_allocs()
282 struct dpmaif_tx_queue *tx_q; in t7xx_dpmaif_sw_release() local
289 tx_q = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_sw_release()
290 t7xx_dpmaif_txq_free(tx_q); in t7xx_dpmaif_sw_release()
[all...]
/linux/drivers/net/ethernet/intel/ice/
H A Dice_dcb_lib.h43 * @tx_q: hung_queue/tx_queue
45 * finds if queue 'tx_q' falls between the two offsets of any given TC
47 static inline bool ice_find_q_in_range(u16 low, u16 high, unsigned int tx_q) in ice_find_q_in_range() argument
49 return (tx_q >= low) && (tx_q < high); in ice_find_q_in_range()
/linux/net/nfc/nci/
H A Duart.c44 skb = skb_dequeue(&nu->tx_q); in nci_uart_dequeue()
56 return skb_queue_empty(&nu->tx_q); in nci_uart_queue_empty()
122 skb_queue_head_init(&nu->tx_q); in nci_uart_set_driver()
186 skb_queue_purge(&nu->tx_q); in nci_uart_tty_close()
363 skb_queue_tail(&nu->tx_q, skb); in nci_uart_send()
H A Ddata.c146 spin_lock_irqsave(&ndev->tx_q.lock, flags); in nci_queue_tx_data_frags()
149 __skb_queue_tail(&ndev->tx_q, skb_frag); in nci_queue_tx_data_frags()
151 spin_unlock_irqrestore(&ndev->tx_q.lock, flags); in nci_queue_tx_data_frags()
185 skb_queue_tail(&ndev->tx_q, skb); in nci_send_data()
H A Dcore.c543 skb_queue_purge(&ndev->tx_q); in nci_open_device()
576 skb_queue_purge(&ndev->tx_q); in nci_close_device()
1269 skb_queue_head_init(&ndev->tx_q); in nci_register_device()
1506 skb = skb_dequeue(&ndev->tx_q); in nci_tx_work()
H A Dntf.c78 if (!skb_queue_empty(&ndev->tx_q)) in nci_core_conn_credits_ntf_packet()
739 skb_queue_purge(&ndev->tx_q); in nci_rf_deactivate_ntf_packet()
/linux/drivers/net/wireless/mediatek/mt7601u/
H A Ddma.c312 struct mt7601u_tx_queue *q = &dev->tx_q[ep]; in mt7601u_dma_submit_tx()
477 if (!dev->tx_q) in mt7601u_free_tx()
481 mt7601u_free_tx_queue(&dev->tx_q[i]); in mt7601u_free_tx()
505 dev->tx_q = devm_kcalloc(dev->dev, __MT_EP_OUT_MAX, in mt7601u_alloc_tx()
506 sizeof(*dev->tx_q), GFP_KERNEL); in mt7601u_alloc_tx()
507 if (!dev->tx_q) in mt7601u_alloc_tx()
511 if (mt7601u_alloc_tx_queue(dev, &dev->tx_q[i])) in mt7601u_alloc_tx()
H A Dmt7601u.h140 * @tx_lock: protects @tx_q and changes of MT7601U_STATE_*_STATS
201 struct mt7601u_tx_queue *tx_q; member
/linux/drivers/net/fddi/skfp/h/
H A Dhwmtm.h242 #define HWM_GET_TX_USED(smc,queue) (int) (smc)->hw.fp.tx_q[queue].tx_used
261 (smc)->hw.fp.tx_q[queue].tx_curr_put
H A Dfplustm.h191 struct s_smt_tx_queue tx_q[USED_QUEUES] ; member
/linux/drivers/net/wireless/mediatek/mt76/
H A Dmt76_connac_mac.c92 if (wcid && pm->tx_q[i].wcid != wcid) in mt76_connac_free_pending_tx_skbs()
95 dev_kfree_skb(pm->tx_q[i].skb); in mt76_connac_free_pending_tx_skbs()
96 pm->tx_q[i].skb = NULL; in mt76_connac_free_pending_tx_skbs()
111 if (!pm->tx_q[qid].skb) { in mt76_connac_pm_queue_skb()
113 pm->tx_q[qid].wcid = wcid; in mt76_connac_pm_queue_skb()
114 pm->tx_q[qid].skb = skb; in mt76_connac_pm_queue_skb()
130 struct mt76_wcid *wcid = pm->tx_q[i].wcid; in mt76_connac_pm_dequeue_skbs()
133 if (!pm->tx_q[i].skb) in mt76_connac_pm_dequeue_skbs()
140 mt76_tx(phy, sta, wcid, pm->tx_q[i].skb); in mt76_connac_pm_dequeue_skbs()
141 pm->tx_q[ in mt76_connac_pm_dequeue_skbs()
[all...]
/linux/net/bluetooth/
H A Dl2cap_core.c449 skb_queue_head_init(&chan->tx_q); in l2cap_chan_create()
572 skb_queue_head_init(&chan->tx_q); in l2cap_le_flowctl_init()
683 skb_queue_purge(&chan->tx_q); in l2cap_chan_del()
698 skb_queue_purge(&chan->tx_q); in l2cap_chan_del()
1932 skb_queue_splice_tail_init(skbs, &chan->tx_q); in l2cap_streaming_send()
1934 while (!skb_queue_empty(&chan->tx_q)) { in l2cap_streaming_send()
1936 skb = skb_dequeue(&chan->tx_q); in l2cap_streaming_send()
2012 if (skb_queue_is_last(&chan->tx_q, skb)) in l2cap_ertm_send()
2015 chan->tx_send_head = skb_queue_next(&chan->tx_q, skb); in l2cap_ertm_send()
2022 chan->unacked_frames, skb_queue_len(&chan->tx_q)); in l2cap_ertm_send()
[all...]
/linux/include/net/nfc/
H A Dnci_core.h227 struct sk_buff_head tx_q; member
458 struct sk_buff_head tx_q; member
/linux/drivers/net/ethernet/chelsio/cxgb3/
H A Dadapter.h208 struct netdev_queue *tx_q; /* associated netdev TX queue */ member

12