Lines Matching full:xdp
76 static inline dma_addr_t xsk_buff_xdp_get_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_dma() argument
78 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_xdp_get_dma()
83 static inline dma_addr_t xsk_buff_xdp_get_frame_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_frame_dma() argument
85 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_xdp_get_frame_dma()
101 static inline u32 xsk_buff_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) in xsk_buff_alloc_batch() argument
103 return xp_alloc_batch(pool, xdp, max); in xsk_buff_alloc_batch()
111 static inline void xsk_buff_free(struct xdp_buff *xdp) in xsk_buff_free() argument
113 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_free()
117 if (likely(!xdp_buff_has_frags(xdp))) in xsk_buff_free()
125 xdp_get_shared_info_from_buff(xdp)->nr_frags = 0; in xsk_buff_free()
131 struct xdp_buff *xdp) in xsk_buff_add_frag() argument
133 const void *data = xdp->data; in xsk_buff_add_frag()
137 offset_in_page(data), xdp->data_end - data, in xsk_buff_add_frag()
138 xdp->frame_sz, false)) in xsk_buff_add_frag()
141 frag = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_add_frag()
149 struct xdp_buff_xsk *xskb = container_of(first, struct xdp_buff_xsk, xdp); in xsk_buff_get_frag()
157 ret = &frag->xdp; in xsk_buff_get_frag()
165 struct xdp_buff_xsk *xskb = container_of(tail, struct xdp_buff_xsk, xdp); in xsk_buff_del_tail()
172 struct xdp_buff_xsk *xskb = container_of(first, struct xdp_buff_xsk, xdp); in xsk_buff_get_tail()
177 return &frag->xdp; in xsk_buff_get_tail()
180 static inline void xsk_buff_set_size(struct xdp_buff *xdp, u32 size) in xsk_buff_set_size() argument
182 xdp->data = xdp->data_hard_start + XDP_PACKET_HEADROOM; in xsk_buff_set_size()
183 xdp->data_meta = xdp->data; in xsk_buff_set_size()
184 xdp->data_end = xdp->data + size; in xsk_buff_set_size()
185 xdp->flags = 0; in xsk_buff_set_size()
249 static inline void xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp) in xsk_buff_dma_sync_for_cpu() argument
251 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_dma_sync_for_cpu()
347 static inline dma_addr_t xsk_buff_xdp_get_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_dma() argument
352 static inline dma_addr_t xsk_buff_xdp_get_frame_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_frame_dma() argument
367 static inline u32 xsk_buff_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) in xsk_buff_alloc_batch() argument
377 static inline void xsk_buff_free(struct xdp_buff *xdp) in xsk_buff_free() argument
382 struct xdp_buff *xdp) in xsk_buff_add_frag() argument
401 static inline void xsk_buff_set_size(struct xdp_buff *xdp, u32 size) in xsk_buff_set_size() argument
439 static inline void xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp) in xsk_buff_dma_sync_for_cpu() argument