Searched refs:nb_pkts (Results 1 – 7 of 7) sorted by relevance
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | test_xsk.c | 346 if (pkt_stream->current_pkt_nb >= pkt_stream->nb_pkts) in pkt_stream_get_next_tx_pkt() 354 while (pkt_stream->current_pkt_nb < pkt_stream->nb_pkts) { in pkt_stream_get_next_rx_pkt() 385 static struct pkt_stream *__pkt_stream_alloc(u32 nb_pkts) in __pkt_stream_alloc() argument 393 pkt_stream->pkts = calloc(nb_pkts, sizeof(*pkt_stream->pkts)); in __pkt_stream_alloc() 399 pkt_stream->nb_pkts = nb_pkts; in __pkt_stream_alloc() 422 while (next_frag++ < pkt_stream->nb_pkts) { in pkt_nb_frags() 456 static struct pkt_stream *__pkt_stream_generate(u32 nb_pkts, u32 pkt_len, u32 nb_start, u32 nb_off) in __pkt_stream_generate() argument 461 pkt_stream = __pkt_stream_alloc(nb_pkts); in __pkt_stream_generate() 465 pkt_stream->nb_pkts = nb_pkts; in __pkt_stream_generate() 467 for (i = 0; i < nb_pkts; i++) { in __pkt_stream_generate() [all …]
|
| H A D | test_xsk.h | 172 u32 nb_pkts; member 186 struct pkt_stream *pkt_stream_generate(u32 nb_pkts, u32 pkt_len);
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_xsk.c | 550 static void i40e_fill_tx_hw_ring(struct i40e_ring *xdp_ring, struct xdp_desc *descs, u32 nb_pkts, in i40e_fill_tx_hw_ring() argument 555 batched = nb_pkts & ~(PKTS_PER_BATCH - 1); in i40e_fill_tx_hw_ring() 556 leftover = nb_pkts & (PKTS_PER_BATCH - 1); in i40e_fill_tx_hw_ring() 582 u32 nb_pkts, nb_processed = 0; in i40e_xmit_zc() local 585 nb_pkts = xsk_tx_peek_release_desc_batch(xdp_ring->xsk_pool, budget); in i40e_xmit_zc() 586 if (!nb_pkts) in i40e_xmit_zc() 589 if (xdp_ring->next_to_use + nb_pkts >= xdp_ring->count) { in i40e_xmit_zc() 595 i40e_fill_tx_hw_ring(xdp_ring, &descs[nb_processed], nb_pkts - nb_processed, in i40e_xmit_zc() 602 i40e_update_tx_stats(xdp_ring, nb_pkts, total_bytes); in i40e_xmit_zc() 604 return nb_pkts < budget; in i40e_xmit_zc()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_xsk.c | 771 struct xdp_desc *descs, u32 nb_pkts, in ice_fill_tx_hw_ring() argument 776 batched = ALIGN_DOWN(nb_pkts, PKTS_PER_BATCH); in ice_fill_tx_hw_ring() 777 leftover = nb_pkts & (PKTS_PER_BATCH - 1); in ice_fill_tx_hw_ring() 794 u32 nb_pkts, nb_processed = 0; in ice_xmit_zc() local 807 nb_pkts = xsk_tx_peek_release_desc_batch(xsk_pool, budget); in ice_xmit_zc() 808 if (!nb_pkts) in ice_xmit_zc() 811 if (xdp_ring->next_to_use + nb_pkts >= xdp_ring->count) { in ice_xmit_zc() 819 nb_pkts - nb_processed, &total_bytes); in ice_xmit_zc() 823 ice_update_tx_ring_stats(xdp_ring, nb_pkts, total_bytes); in ice_xmit_zc() 828 return nb_pkts < budget; in ice_xmit_zc()
|
| /linux/net/xdp/ |
| H A D | xsk.c | 487 u32 nb_pkts = 0; in xsk_tx_peek_release_fallback() local 489 while (nb_pkts < max_entries && xsk_tx_peek_desc(pool, &descs[nb_pkts])) in xsk_tx_peek_release_fallback() 490 nb_pkts++; in xsk_tx_peek_release_fallback() 493 return nb_pkts; in xsk_tx_peek_release_fallback() 496 u32 xsk_tx_peek_release_desc_batch(struct xsk_buff_pool *pool, u32 nb_pkts) in xsk_tx_peek_release_desc_batch() argument 504 return xsk_tx_peek_release_fallback(pool, nb_pkts); in xsk_tx_peek_release_desc_batch() 509 nb_pkts = 0; in xsk_tx_peek_release_desc_batch() 513 nb_pkts = xskq_cons_nb_entries(xs->tx, nb_pkts); in xsk_tx_peek_release_desc_batch() 521 nb_pkts = xskq_prod_nb_free(pool->cq, nb_pkts); in xsk_tx_peek_release_desc_batch() 522 if (!nb_pkts) in xsk_tx_peek_release_desc_batch() [all …]
|
| /linux/drivers/net/ethernet/intel/igb/ |
| H A D | igb_xsk.c | 468 u32 cmd_type, olinfo_status, nb_pkts, i = 0; in igb_xmit_zc() local 481 nb_pkts = xsk_tx_peek_release_desc_batch(xsk_pool, budget); in igb_xmit_zc() 482 if (!nb_pkts) in igb_xmit_zc() 485 for (; i < nb_pkts; i++) { in igb_xmit_zc() 524 return nb_pkts < budget; in igb_xmit_zc()
|
| /linux/drivers/net/ |
| H A D | virtio_net.c | 1522 u32 nb_pkts, i; in virtnet_xsk_xmit_batch() local 1527 nb_pkts = xsk_tx_peek_release_desc_batch(pool, budget); in virtnet_xsk_xmit_batch() 1528 if (!nb_pkts) in virtnet_xsk_xmit_batch() 1531 for (i = 0; i < nb_pkts; i++) { in virtnet_xsk_xmit_batch() 1534 xsk_tx_completed(sq->xsk_pool, nb_pkts - i); in virtnet_xsk_xmit_batch()
|