Lines Matching defs:txr
357 struct bnxt_tx_ring_info *txr;
367 txr = &bp->tx_ring[bp->tx_ring_map[i]];
368 prod = txr->tx_prod;
370 free_size = bnxt_tx_avail(bp, txr);
380 txbd = &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)];
384 tx_buf = &txr->tx_buf_ring[prod];
401 struct tx_push_buffer *tx_push_buf = txr->tx_push;
404 void __iomem *db = txr->tx_db.doorbell;
447 txbd->tx_bd_haddr = txr->data_mapping;
449 txbd = &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)];
454 txr->tx_prod = prod;
499 &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)];
542 txbd = &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)];
551 tx_buf = &txr->tx_buf_ring[prod];
571 txr->tx_prod = prod;
574 bnxt_db_write(bp, &txr->tx_db, prod);
578 if (unlikely(bnxt_tx_avail(bp, txr) <= MAX_SKB_FRAGS + 1)) {
580 bnxt_db_write(bp, &txr->tx_db, prod);
590 if (bnxt_tx_avail(bp, txr) > bp->tx_wake_thresh)
599 prod = txr->tx_prod;
600 tx_buf = &txr->tx_buf_ring[prod];
609 tx_buf = &txr->tx_buf_ring[prod];
621 struct bnxt_tx_ring_info *txr = bnapi->tx_ring;
622 struct netdev_queue *txq = netdev_get_tx_queue(bp->dev, txr->txq_index);
623 u16 cons = txr->tx_cons;
633 tx_buf = &txr->tx_buf_ring[cons];
649 tx_buf = &txr->tx_buf_ring[cons];
665 txr->tx_cons = cons;
675 (bnxt_tx_avail(bp, txr) > bp->tx_wake_thresh)) {
678 bnxt_tx_avail(bp, txr) > bp->tx_wake_thresh &&
679 txr->dev_state != BNXT_DEV_STATE_CLOSING)
2273 struct bnxt_tx_ring_info *txr = bnapi->tx_ring;
2274 u16 prod = txr->tx_prod;
2279 bnxt_db_write_relaxed(bp, &txr->tx_db, prod);
2527 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i];
2531 struct bnxt_sw_tx_bd *tx_buf = &txr->tx_buf_ring[j];
2573 tx_buf = &txr->tx_buf_ring[ring_idx];
2934 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i];
2937 if (txr->tx_push) {
2939 txr->tx_push, txr->tx_push_mapping);
2940 txr->tx_push = NULL;
2943 ring = &txr->tx_ring_struct;
2970 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i];
2974 ring = &txr->tx_ring_struct;
2980 ring->grp_idx = txr->bnapi->index;
2987 txr->tx_push = dma_alloc_coherent(&pdev->dev,
2989 &txr->tx_push_mapping,
2992 if (!txr->tx_push)
2995 mapping = txr->tx_push_mapping +
2997 txr->data_mapping = cpu_to_le64(mapping);
3133 struct bnxt_tx_ring_info *txr;
3171 txr = bnapi->tx_ring;
3172 if (!txr)
3175 ring = &txr->tx_ring_struct;
3179 rmem->pg_arr = (void **)txr->tx_desc_ring;
3180 rmem->dma_arr = txr->tx_desc_mapping;
3182 rmem->vmem = (void **)&txr->tx_buf_ring;
3346 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i];
3347 struct bnxt_ring_struct *ring = &txr->tx_ring_struct;
4017 struct bnxt_tx_ring_info *txr;
4025 txr = bnapi->tx_ring;
4026 if (txr) {
4027 txr->tx_prod = 0;
4028 txr->tx_cons = 0;
4188 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i];
4191 txr->tx_ring_struct.ring_mem.flags =
4193 txr->bnapi = bp->bnapi[j];
4194 bp->bnapi[j]->tx_ring = txr;
4197 txr->txq_index = i - bp->tx_nr_rings_xdp;
4988 static u16 bnxt_cp_ring_for_tx(struct bnxt *bp, struct bnxt_tx_ring_info *txr)
4991 struct bnxt_napi *bnapi = txr->bnapi;
4997 return bnxt_cp_ring_from_grp(bp, &txr->tx_ring_struct);
5494 struct bnxt_tx_ring_info *txr;
5496 txr = container_of(ring, struct bnxt_tx_ring_info,
5501 req.cmpl_ring_id = cpu_to_le16(bnxt_cp_ring_for_tx(bp, txr));
5684 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i];
5689 struct bnxt_napi *bnapi = txr->bnapi;
5705 ring = &txr->tx_ring_struct;
5710 bnxt_set_db(bp, &txr->tx_db, type, map_idx, ring->fw_ring_id);
5807 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i];
5808 struct bnxt_ring_struct *ring = &txr->tx_ring_struct;
5811 u32 cmpl_ring_id = bnxt_cp_ring_for_tx(bp, txr);
8848 struct bnxt_tx_ring_info *txr;
8852 txr = &bp->tx_ring[i];
8853 txr->dev_state = BNXT_DEV_STATE_CLOSING;
8864 struct bnxt_tx_ring_info *txr;
8867 txr = &bp->tx_ring[i];
8868 txr->dev_state = 0;
10535 struct bnxt_tx_ring_info *txr = bnapi->tx_ring;
10538 if (!txr)
10542 i, txr->tx_ring_struct.fw_ring_id, txr->tx_prod,
10543 txr->tx_cons);