Lines Matching refs:rx_buf

44 				     struct efx_rx_buffer *rx_buf,
48 unsigned max_len = rx_buf->len - efx->type->rx_buffer_padding;
56 rx_buf->flags |= EFX_RX_PKT_DISCARD;
68 struct efx_rx_buffer *rx_buf,
84 EFX_WARN_ON_ONCE_PARANOID(rx_buf->len < hdr_len);
92 if (rx_buf->len > hdr_len) {
93 rx_buf->page_offset += hdr_len;
94 rx_buf->len -= hdr_len;
98 rx_buf->page, rx_buf->page_offset,
99 rx_buf->len, efx->rx_buffer_truesize);
100 rx_buf->page = NULL;
105 rx_buf = efx_rx_buf_next(&channel->rx_queue, rx_buf);
108 __free_pages(rx_buf->page, efx->rx_buffer_order);
109 rx_buf->page = NULL;
126 struct efx_rx_buffer *rx_buf;
130 rx_buf = efx_rx_buffer(rx_queue, index);
131 rx_buf->flags |= flags;
136 efx_rx_packet__check_len(rx_queue, rx_buf, len);
144 WARN_ON(!(len == 0 && rx_buf->flags & EFX_RX_PKT_DISCARD));
145 rx_buf->flags |= EFX_RX_PKT_DISCARD;
152 (rx_buf->flags & EFX_RX_PKT_CSUMMED) ? " [SUMMED]" : "",
153 (rx_buf->flags & EFX_RX_PKT_DISCARD) ? " [DISCARD]" : "");
158 if (unlikely(rx_buf->flags & EFX_RX_PKT_DISCARD)) {
160 efx_siena_discard_rx_packet(channel, rx_buf, n_frags);
165 rx_buf->len = len;
170 efx_sync_rx_buffer(efx, rx_buf, rx_buf->len);
175 prefetch(efx_rx_buf_va(rx_buf));
177 rx_buf->page_offset += efx->rx_prefix_size;
178 rx_buf->len -= efx->rx_prefix_size;
187 rx_buf = efx_rx_buf_next(rx_queue, rx_buf);
190 efx_sync_rx_buffer(efx, rx_buf, efx->rx_dma_len);
192 rx_buf->len = len - (n_frags - 1) * efx->rx_dma_len;
193 efx_sync_rx_buffer(efx, rx_buf, rx_buf->len);
197 rx_buf = efx_rx_buffer(rx_queue, index);
198 efx_siena_recycle_rx_pages(channel, rx_buf, n_frags);
209 struct efx_rx_buffer *rx_buf,
213 u16 hdr_len = min_t(u16, rx_buf->len, EFX_SKB_HEADERS);
215 skb = efx_rx_mk_skb(channel, rx_buf, n_frags, eh, hdr_len);
220 efx_siena_free_rx_buffers(rx_queue, rx_buf, n_frags);
227 if (likely(rx_buf->flags & EFX_RX_PKT_CSUMMED)) {
229 skb->csum_level = !!(rx_buf->flags & EFX_RX_PKT_CSUM_LEVEL);
252 struct efx_rx_buffer *rx_buf, u8 **ehp)
271 efx_siena_free_rx_buffers(rx_queue, rx_buf,
281 dma_sync_single_for_cpu(&efx->pci_dev->dev, rx_buf->dma_addr,
282 rx_buf->len, DMA_FROM_DEVICE);
292 rx_buf->len, false);
303 rx_buf->page_offset += offset;
304 rx_buf->len -= offset;
315 efx_siena_free_rx_buffers(rx_queue, rx_buf, 1);
329 efx_siena_free_rx_buffers(rx_queue, rx_buf, 1);
342 efx_siena_free_rx_buffers(rx_queue, rx_buf, 1);
351 efx_siena_free_rx_buffers(rx_queue, rx_buf, 1);
363 struct efx_rx_buffer *rx_buf =
365 u8 *eh = efx_rx_buf_va(rx_buf);
370 if (rx_buf->flags & EFX_RX_PKT_PREFIX_LEN)
371 rx_buf->len = le16_to_cpup((__le16 *)
375 * loopback layer, and free the rx_buf here
380 efx_siena_loopback_rx_packet(efx, eh, rx_buf->len);
382 efx_siena_free_rx_buffers(rx_queue, rx_buf,
387 if (!efx_do_xdp(efx, channel, rx_buf, &eh))
391 rx_buf->flags &= ~EFX_RX_PKT_CSUMMED;
393 if ((rx_buf->flags & EFX_RX_PKT_TCP) && !channel->type->receive_skb)
394 efx_siena_rx_packet_gro(channel, rx_buf,
397 efx_rx_deliver(channel, eh, rx_buf, channel->rx_pkt_n_frags);