Lines Matching +full:queue +full:- +full:rx

1 /* SPDX-License-Identifier: GPL-2.0-only */
19 /* Mailbox Queue */
27 /* Number of descriptors in a queue should be a multiple of 32. RX queue
66 * given RX completion queue has descriptors. This includes _ALL_ buffer
68 * you have a total of 1024 buffers so your RX queue _must_ have at least that
69 * many descriptors. This macro divides a given number of RX descriptors by
70 * number of buffer queues to calculate how many descriptors each buffer queue
71 * can have without overrunning the RX queue.
83 #define IDPF_RX_BUFQ_WORKING_SET(rxq) ((rxq)->desc_count - 1)
87 if (unlikely(++(ntc) == (rxq)->desc_count)) { \
95 if (unlikely(++(idx) == (q)->desc_count)) \
117 ((((txq)->next_to_clean > (txq)->next_to_use) ? 0 : (txq)->desc_count) + \
118 (txq)->next_to_clean - (txq)->next_to_use - 1)
120 #define IDPF_TX_BUF_RSV_UNUSED(txq) ((txq)->stash->buf_stack.top)
122 (txq)->desc_count >> 2)
124 #define IDPF_TX_COMPLQ_OVERFLOW_THRESH(txcq) ((txcq)->desc_count >> 1)
126 * completions that are expected to arrive on the TX completion queue.
129 (((txq)->num_completions_pending >= (txq)->complq->num_completions ? \
131 (txq)->num_completions_pending - (txq)->complq->num_completions)
136 ((++(txq)->compl_tag_cur_gen) >= (txq)->compl_tag_gen_max ? \
137 0 : (txq)->compl_tag_cur_gen)
147 struct idpf_flex_tx_desc q; /* queue based scheduling */
154 * struct idpf_buf_lifo - LIFO for managing OOO completions
166 * struct idpf_tx_offload_params - Offload parameters for a given packet
168 * @hdr_offsets: Offset parameter for single queue model
169 * @cd_tunneling: Type of tunneling enabled for single queue model
227 /* The size limit for a transmit buffer in a descriptor is (16K - 1).
232 #define IDPF_TX_MAX_DESC_DATA (SZ_16K - 1)
243 DIV_ROUND_DOWN_ULL((IDPF_CTLQ_MAX_BUF_LEN - IDPF_RX_PTYPE_HDR_SZ), \
246 #define IDPF_GET_PTYPE_SIZE(p) struct_size((p), proto_id, (p)->proto_id_count)
279 * 0->1 or 1->0 on each ring wrap. SW maintains its own
288 * @__IDPF_Q_SW_MARKER: Used to indicate TX queue marker completions
291 * @__IDPF_Q_HSPLIT_EN: enable header split on Rx (splitq)
306 #define idpf_queue_set(f, q) __set_bit(__IDPF_Q_##f, (q)->flags)
307 #define idpf_queue_clear(f, q) __clear_bit(__IDPF_Q_##f, (q)->flags)
308 #define idpf_queue_change(f, q) __change_bit(__IDPF_Q_##f, (q)->flags)
309 #define idpf_queue_has(f, q) test_bit(__IDPF_Q_##f, (q)->flags)
312 __test_and_clear_bit(__IDPF_Q_##f, (q)->flags)
314 __assign_bit(__IDPF_Q_##f, (q)->flags, v)
340 * @rx_itr: RX ITR register
364 * @num_rxq: Number of RX queues
368 * @rx: Array of RX queues to service
380 * @rx_dim: Data for RX net_dim algorithm
381 * @rx_itr_value: RX interrupt throttling rate
383 * @rx_itr_idx: RX ITR index
394 struct idpf_rx_queue **rx; member
465 * struct idpf_txq_stash - Tx buffer stash for Flow-based scheduling mode
476 * struct idpf_rx_queue - software structure representing a receive queue
477 * @rx: universal receive descriptor array
481 * @napi: NAPI instance corresponding to this queue (splitq)
484 * @netdev: &net_device corresponding to this queue
485 * @tail: Tail offset. Used for both queue models single and split.
487 * @idx: For RX queue, it is used to index to total RX queue across groups and
490 * @rxdids: Supported RX descriptor ids
491 * @rx_ptype_lkup: LUT of Rx ptypes
494 * @next_to_alloc: RX buffer to allocate at
499 * @q_id: Queue id
503 * @rx_buffer_low_watermark: RX buffer low watermark
506 * @rx_max_pkt_size: RX max packet size
511 union virtchnl2_rx_desc *rx; member
567 * struct idpf_tx_queue - software structure representing a transmit queue
576 * @tail: Tail offset. Used for both queue models single and split
578 * @idx: For TX queue, it is used as index to map between TX queue group and
589 * --------------------------------
590 * | GEN=0-1023 |IDX = 0-63|
591 * --------------------------------
597 * --------------------------------
598 * |GEN | IDX = 0-8159 |
599 * --------------------------------
602 * @netdev: &net_device corresponding to this queue
606 * the TX completion queue, it can be for any TXQ associated
607 * with that completion queue. This means we can clean up to
608 * N TXQs during a single call to clean the completion queue.
610 * that single call to clean the completion queue. By doing so,
613 * @clean_budget: singleq only, queue cleaning budget
615 * @tx_max_bufs: Max buffers that can be transmitted with scatter-gather
616 * @stash: Tx buffer stash for Flow-based scheduling mode
622 * @q_id: Queue id
686 * struct idpf_buf_queue - software structure representing a buffer queue
697 * @next_to_alloc: RX buffer to allocate at
700 * @q_id: Queue id
704 * @rx_buffer_low_watermark: RX buffer low watermark
745 * struct idpf_compl_queue - software structure representing a completion queue
750 * @clean_budget: queue cleaning budget
751 * @netdev: &net_device corresponding to this queue
755 * @num_completions: Only relevant for TX completion queue. It tracks the
759 * @q_id: Queue id
824 * @rxq: RX queue
838 * @bufq: Buffer queue
860 * @singleq: Struct with single queue related members
861 * @singleq.num_rxq: Number of RX queues associated
862 * @singleq.rxqs: Array of RX queue pointers
863 * @splitq: Struct with split queue related members
864 * @splitq.num_rxq_sets: Number of RX queue sets
865 * @splitq.rxq_sets: Array of RX queue sets
866 * @splitq.bufq_sets: Buffer queue set pointer
892 * @txqs: Array of TX queue pointers
894 * @complq: Associated completion queue pointer, split queue only
896 * completion queue, acculumated for all TX queues
897 * associated with that completion queue.
922 cpu = cpumask_first(&q_vector->napi.config->affinity_mask); in idpf_q_vector_to_mem()
928 * idpf_size_to_txd_count - Get number of descriptors needed for large Tx frag
941 * idpf_tx_singleq_build_ctob - populate command tag offset and size
966 * idpf_tx_splitq_build_desc - determine which type of data descriptor to build
976 if (params->dtype == IDPF_TX_DESC_DTYPE_FLEX_L2TAG1_L2TAG2) in idpf_tx_splitq_build_desc()
983 * idpf_vport_intr_set_wb_on_itr - enable descriptor writeback on disabled interrupts
984 * @q_vector: pointer to queue vector struct
990 if (q_vector->wb_on_itr) in idpf_vport_intr_set_wb_on_itr()
993 q_vector->wb_on_itr = true; in idpf_vport_intr_set_wb_on_itr()
994 reg = &q_vector->intr_reg; in idpf_vport_intr_set_wb_on_itr()
996 writel(reg->dyn_ctl_wb_on_itr_m | reg->dyn_ctl_intena_msk_m | in idpf_vport_intr_set_wb_on_itr()
997 (IDPF_NO_ITR_UPDATE_IDX << reg->dyn_ctl_itridx_s), in idpf_vport_intr_set_wb_on_itr()
998 reg->dyn_ctl); in idpf_vport_intr_set_wb_on_itr()
1043 return !netif_subqueue_maybe_stop(tx_q->netdev, tx_q->idx, in idpf_tx_maybe_stop_common()