Lines Matching +full:xo +full:- +full:1
1 // SPDX-License-Identifier: GPL-2.0-only
34 struct xfrm_offload *xo; in esp4_gro_receive() local
46 xo = xfrm_offload(skb); in esp4_gro_receive()
47 if (!xo || !(xo->flags & CRYPTO_DONE)) { in esp4_gro_receive()
53 if (sp->len == XFRM_MAX_DEPTH) in esp4_gro_receive()
56 x = xfrm_state_lookup(dev_net(skb->dev), skb->mark, in esp4_gro_receive()
57 (xfrm_address_t *)&ip_hdr(skb)->daddr, in esp4_gro_receive()
62 skb->mark = xfrm_smark_get(skb->mark, x); in esp4_gro_receive()
64 sp->xvec[sp->len++] = x; in esp4_gro_receive()
65 sp->olen++; in esp4_gro_receive()
67 xo = xfrm_offload(skb); in esp4_gro_receive()
68 if (!xo) in esp4_gro_receive()
72 xo->flags |= XFRM_GRO; in esp4_gro_receive()
74 if (NAPI_GRO_CB(skb)->proto == IPPROTO_UDP) in esp4_gro_receive()
77 XFRM_TUNNEL_SKB_CB(skb)->tunnel.ip4 = NULL; in esp4_gro_receive()
78 XFRM_SPI_SKB_CB(skb)->family = AF_INET; in esp4_gro_receive()
79 XFRM_SPI_SKB_CB(skb)->daddroff = offsetof(struct iphdr, daddr); in esp4_gro_receive()
80 XFRM_SPI_SKB_CB(skb)->seq = seq; in esp4_gro_receive()
86 return ERR_PTR(-EINPROGRESS); in esp4_gro_receive()
91 NAPI_GRO_CB(skb)->same_flow = 0; in esp4_gro_receive()
92 NAPI_GRO_CB(skb)->flush = 1; in esp4_gro_receive()
101 struct xfrm_offload *xo = xfrm_offload(skb); in esp4_gso_encap() local
102 int proto = iph->protocol; in esp4_gso_encap()
104 skb_push(skb, -skb_network_offset(skb)); in esp4_gso_encap()
108 esph->spi = x->id.spi; in esp4_gso_encap()
109 esph->seq_no = htonl(XFRM_SKB_CB(skb)->seq.output.low); in esp4_gso_encap()
111 xo->proto = proto; in esp4_gso_encap()
118 __be16 type = x->inner_mode.family == AF_INET6 ? htons(ETH_P_IPV6) in xfrm4_tunnel_gso_segment()
129 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_transport_gso_segment()
130 struct xfrm_offload *xo = xfrm_offload(skb); in xfrm4_transport_gso_segment() local
132 skb->transport_header += x->props.header_len; in xfrm4_transport_gso_segment()
133 ops = rcu_dereference(inet_offloads[xo->proto]); in xfrm4_transport_gso_segment()
134 if (likely(ops && ops->callbacks.gso_segment)) in xfrm4_transport_gso_segment()
135 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_transport_gso_segment()
144 struct xfrm_offload *xo = xfrm_offload(skb); in xfrm4_beet_gso_segment() local
145 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_beet_gso_segment()
147 u8 proto = xo->proto; in xfrm4_beet_gso_segment()
149 skb->transport_header += x->props.header_len; in xfrm4_beet_gso_segment()
151 if (x->sel.family != AF_INET6) { in xfrm4_beet_gso_segment()
154 (struct ip_beet_phdr *)skb->data; in xfrm4_beet_gso_segment()
156 skb->transport_header += ph->hdrlen * 8; in xfrm4_beet_gso_segment()
157 proto = ph->nexthdr; in xfrm4_beet_gso_segment()
159 skb->transport_header -= IPV4_BEET_PHMAXLEN; in xfrm4_beet_gso_segment()
164 skb->transport_header += in xfrm4_beet_gso_segment()
167 skb_shinfo(skb)->gso_type |= SKB_GSO_TCPV4; in xfrm4_beet_gso_segment()
171 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP4; in xfrm4_beet_gso_segment()
175 if (likely(ops && ops->callbacks.gso_segment)) in xfrm4_beet_gso_segment()
176 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_beet_gso_segment()
185 switch (x->outer_mode.encap) { in xfrm4_outer_mode_gso_segment()
194 return ERR_PTR(-EOPNOTSUPP); in xfrm4_outer_mode_gso_segment()
204 struct xfrm_offload *xo = xfrm_offload(skb); in esp4_gso_segment() local
207 if (!xo) in esp4_gso_segment()
208 return ERR_PTR(-EINVAL); in esp4_gso_segment()
210 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_ESP)) in esp4_gso_segment()
211 return ERR_PTR(-EINVAL); in esp4_gso_segment()
214 x = sp->xvec[sp->len - 1]; in esp4_gso_segment()
215 aead = x->data; in esp4_gso_segment()
218 if (esph->spi != x->id.spi) in esp4_gso_segment()
219 return ERR_PTR(-EINVAL); in esp4_gso_segment()
222 return ERR_PTR(-EINVAL); in esp4_gso_segment()
226 skb->encap_hdr_csum = 1; in esp4_gso_segment()
228 if ((!(skb->dev->gso_partial_features & NETIF_F_HW_ESP) && in esp4_gso_segment()
229 !(features & NETIF_F_HW_ESP)) || x->xso.dev != skb->dev) in esp4_gso_segment()
233 !(skb->dev->gso_partial_features & NETIF_F_HW_ESP_TX_CSUM)) in esp4_gso_segment()
237 xo->flags |= XFRM_GSO_SEGMENT; in esp4_gso_segment()
244 struct crypto_aead *aead = x->data; in esp_input_tail()
245 struct xfrm_offload *xo = xfrm_offload(skb); in esp_input_tail() local
248 return -EINVAL; in esp_input_tail()
250 if (!(xo->flags & CRYPTO_DONE)) in esp_input_tail()
251 skb->ip_summed = CHECKSUM_NONE; in esp_input_tail()
261 struct xfrm_offload *xo; in esp_xmit() local
270 xo = xfrm_offload(skb); in esp_xmit()
272 if (!xo) in esp_xmit()
273 return -EINVAL; in esp_xmit()
276 !(skb->dev->gso_partial_features & NETIF_F_HW_ESP)) || in esp_xmit()
277 x->xso.dev != skb->dev) { in esp_xmit()
278 xo->flags |= CRYPTO_FALLBACK; in esp_xmit()
282 esp.proto = xo->proto; in esp_xmit()
286 aead = x->data; in esp_xmit()
293 esp.clen = ALIGN(skb->len + 2 + esp.tfclen, blksize); in esp_xmit()
294 esp.plen = esp.clen - skb->len - esp.tfclen; in esp_xmit()
306 seq = xo->seq.low; in esp_xmit()
309 esph->spi = x->id.spi; in esp_xmit()
311 skb_push(skb, -skb_network_offset(skb)); in esp_xmit()
313 if (xo->flags & XFRM_GSO_SEGMENT) { in esp_xmit()
314 esph->seq_no = htonl(seq); in esp_xmit()
317 xo->seq.low++; in esp_xmit()
319 xo->seq.low += skb_shinfo(skb)->gso_segs; in esp_xmit()
322 if (xo->seq.low < seq) in esp_xmit()
323 xo->seq.hi++; in esp_xmit()
325 esp.seqno = cpu_to_be64(seq + ((u64)xo->seq.hi << 32)); in esp_xmit()
327 ip_hdr(skb)->tot_len = htons(skb->len); in esp_xmit()
332 return -ENOMEM; in esp_xmit()
334 xo = xfrm_offload(skb); in esp_xmit()
335 if (!xo) in esp_xmit()
336 return -EINVAL; in esp_xmit()
338 xo->flags |= XFRM_XMIT; in esp_xmit()
348 if (skb_needs_linearize(skb, skb->dev->features) && in esp_xmit()
350 return -ENOMEM; in esp_xmit()
373 return -EAGAIN; in esp4_offload_init()