Lines Matching +full:xo +full:- +full:1

1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * xfrm_device.c - IPsec device offloading code.
26 struct xfrm_offload *xo = xfrm_offload(skb); in __xfrm_transport_prep() local
29 if (xo->flags & XFRM_GSO_SEGMENT) in __xfrm_transport_prep()
30 skb->transport_header -= x->props.header_len; in __xfrm_transport_prep()
32 pskb_pull(skb, skb_transport_offset(skb) + x->props.header_len); in __xfrm_transport_prep()
39 struct xfrm_offload *xo = xfrm_offload(skb); in __xfrm_mode_tunnel_prep() local
41 if (xo->flags & XFRM_GSO_SEGMENT) in __xfrm_mode_tunnel_prep()
42 skb->transport_header = skb->network_header + hsize; in __xfrm_mode_tunnel_prep()
45 pskb_pull(skb, skb->mac_len + x->props.header_len); in __xfrm_mode_tunnel_prep()
51 struct xfrm_offload *xo = xfrm_offload(skb); in __xfrm_mode_beet_prep() local
54 if (xo->flags & XFRM_GSO_SEGMENT) in __xfrm_mode_beet_prep()
55 skb->transport_header = skb->network_header + hsize; in __xfrm_mode_beet_prep()
58 if (x->sel.family != AF_INET6) { in __xfrm_mode_beet_prep()
60 if (x->outer_mode.family == AF_INET6) in __xfrm_mode_beet_prep()
61 phlen += sizeof(struct ipv6hdr) - sizeof(struct iphdr); in __xfrm_mode_beet_prep()
64 pskb_pull(skb, skb->mac_len + hsize + (x->props.header_len - phlen)); in __xfrm_mode_beet_prep()
70 switch (x->outer_mode.encap) { in xfrm_outer_mode_prep()
72 if (x->outer_mode.family == AF_INET) in xfrm_outer_mode_prep()
75 if (x->outer_mode.family == AF_INET6) in xfrm_outer_mode_prep()
80 if (x->outer_mode.family == AF_INET) in xfrm_outer_mode_prep()
83 if (x->outer_mode.family == AF_INET6) in xfrm_outer_mode_prep()
88 if (x->outer_mode.family == AF_INET) in xfrm_outer_mode_prep()
91 if (x->outer_mode.family == AF_INET6) in xfrm_outer_mode_prep()
103 struct xfrm_offload *xo = xfrm_offload(skb); in xmit_xfrm_check_overflow() local
104 __u32 seq = xo->seq.low; in xmit_xfrm_check_overflow()
106 seq += skb_shinfo(skb)->gso_segs; in xmit_xfrm_check_overflow()
107 if (unlikely(seq < xo->seq.low)) in xmit_xfrm_check_overflow()
121 struct xfrm_offload *xo = xfrm_offload(skb); in validate_xmit_xfrm() local
122 struct net_device *dev = skb->dev; in validate_xmit_xfrm()
125 if (!xo || (xo->flags & XFRM_XMIT)) in validate_xmit_xfrm()
132 x = sp->xvec[sp->len - 1]; in validate_xmit_xfrm()
133 if (xo->flags & XFRM_GRO || x->xso.dir == XFRM_DEV_OFFLOAD_IN) in validate_xmit_xfrm()
140 if (x->xso.type == XFRM_DEV_OFFLOAD_PACKET && x->xso.dev != dev) { in validate_xmit_xfrm()
147 if ((x->xso.dev != dev) && (x->xso.real_dev == dev)) in validate_xmit_xfrm()
152 err = !skb_queue_empty(&sd->xfrm_backlog); in validate_xmit_xfrm()
160 if (skb_is_gso(skb) && (unlikely(x->xso.dev != dev) || in validate_xmit_xfrm()
178 if (!skb->next) { in validate_xmit_xfrm()
179 esp_features |= skb->dev->gso_partial_features; in validate_xmit_xfrm()
182 xo->flags |= XFRM_DEV_RESUME; in validate_xmit_xfrm()
184 err = x->type_offload->xmit(x, skb, esp_features); in validate_xmit_xfrm()
186 if (err == -EINPROGRESS) in validate_xmit_xfrm()
194 skb_push(skb, skb->data - skb_mac_header(skb)); in validate_xmit_xfrm()
200 esp_features |= skb->dev->gso_partial_features; in validate_xmit_xfrm()
203 xo = xfrm_offload(skb2); in validate_xmit_xfrm()
204 xo->flags |= XFRM_DEV_RESUME; in validate_xmit_xfrm()
208 err = x->type_offload->xmit(x, skb2, esp_features); in validate_xmit_xfrm()
210 skb2->next = nskb; in validate_xmit_xfrm()
211 } else if (err != -EINPROGRESS) { in validate_xmit_xfrm()
213 skb2->next = nskb; in validate_xmit_xfrm()
220 pskb->next = nskb; in validate_xmit_xfrm()
225 skb_push(skb2, skb2->data - skb_mac_header(skb2)); in validate_xmit_xfrm()
240 struct xfrm_dev_offload *xso = &x->xso; in xfrm_dev_state_add()
245 if (!x->type_offload) { in xfrm_dev_state_add()
247 return -EINVAL; in xfrm_dev_state_add()
250 if (xuo->flags & in xfrm_dev_state_add()
253 return -EINVAL; in xfrm_dev_state_add()
256 is_packet_offload = xuo->flags & XFRM_OFFLOAD_PACKET; in xfrm_dev_state_add()
259 if ((!is_packet_offload && x->encap) || x->tfcpad) { in xfrm_dev_state_add()
261 return -EINVAL; in xfrm_dev_state_add()
264 dev = dev_get_by_index(net, xuo->ifindex); in xfrm_dev_state_add()
266 if (!(xuo->flags & XFRM_OFFLOAD_INBOUND)) { in xfrm_dev_state_add()
267 saddr = &x->props.saddr; in xfrm_dev_state_add()
268 daddr = &x->id.daddr; in xfrm_dev_state_add()
270 saddr = &x->id.daddr; in xfrm_dev_state_add()
271 daddr = &x->props.saddr; in xfrm_dev_state_add()
275 x->props.family, in xfrm_dev_state_add()
278 return (is_packet_offload) ? -EINVAL : 0; in xfrm_dev_state_add()
280 dev = dst->dev; in xfrm_dev_state_add()
286 if (!dev->xfrmdev_ops || !dev->xfrmdev_ops->xdo_dev_state_add) { in xfrm_dev_state_add()
287 xso->dev = NULL; in xfrm_dev_state_add()
289 return (is_packet_offload) ? -EINVAL : 0; in xfrm_dev_state_add()
292 if (!is_packet_offload && x->props.flags & XFRM_STATE_ESN && in xfrm_dev_state_add()
293 !dev->xfrmdev_ops->xdo_dev_state_advance_esn) { in xfrm_dev_state_add()
295 xso->dev = NULL; in xfrm_dev_state_add()
297 return -EINVAL; in xfrm_dev_state_add()
300 xso->dev = dev; in xfrm_dev_state_add()
301 netdev_tracker_alloc(dev, &xso->dev_tracker, GFP_ATOMIC); in xfrm_dev_state_add()
302 xso->real_dev = dev; in xfrm_dev_state_add()
304 if (xuo->flags & XFRM_OFFLOAD_INBOUND) in xfrm_dev_state_add()
305 xso->dir = XFRM_DEV_OFFLOAD_IN; in xfrm_dev_state_add()
307 xso->dir = XFRM_DEV_OFFLOAD_OUT; in xfrm_dev_state_add()
310 xso->type = XFRM_DEV_OFFLOAD_PACKET; in xfrm_dev_state_add()
312 xso->type = XFRM_DEV_OFFLOAD_CRYPTO; in xfrm_dev_state_add()
314 err = dev->xfrmdev_ops->xdo_dev_state_add(x, extack); in xfrm_dev_state_add()
316 xso->dev = NULL; in xfrm_dev_state_add()
317 xso->dir = 0; in xfrm_dev_state_add()
318 xso->real_dev = NULL; in xfrm_dev_state_add()
319 netdev_put(dev, &xso->dev_tracker); in xfrm_dev_state_add()
320 xso->type = XFRM_DEV_OFFLOAD_UNSPECIFIED; in xfrm_dev_state_add()
327 * authors to do not return -EOPNOTSUPP in packet offload mode. in xfrm_dev_state_add()
329 WARN_ON(err == -EOPNOTSUPP && is_packet_offload); in xfrm_dev_state_add()
330 if (err != -EOPNOTSUPP || is_packet_offload) { in xfrm_dev_state_add()
344 struct xfrm_dev_offload *xdo = &xp->xdo; in xfrm_dev_policy_add()
348 if (!xuo->flags || xuo->flags & ~XFRM_OFFLOAD_PACKET) { in xfrm_dev_policy_add()
353 return -EINVAL; in xfrm_dev_policy_add()
356 dev = dev_get_by_index(net, xuo->ifindex); in xfrm_dev_policy_add()
358 return -EINVAL; in xfrm_dev_policy_add()
360 if (!dev->xfrmdev_ops || !dev->xfrmdev_ops->xdo_dev_policy_add) { in xfrm_dev_policy_add()
361 xdo->dev = NULL; in xfrm_dev_policy_add()
364 return -EINVAL; in xfrm_dev_policy_add()
367 xdo->dev = dev; in xfrm_dev_policy_add()
368 netdev_tracker_alloc(dev, &xdo->dev_tracker, GFP_ATOMIC); in xfrm_dev_policy_add()
369 xdo->real_dev = dev; in xfrm_dev_policy_add()
370 xdo->type = XFRM_DEV_OFFLOAD_PACKET; in xfrm_dev_policy_add()
373 xdo->dir = XFRM_DEV_OFFLOAD_IN; in xfrm_dev_policy_add()
376 xdo->dir = XFRM_DEV_OFFLOAD_OUT; in xfrm_dev_policy_add()
379 xdo->dir = XFRM_DEV_OFFLOAD_FWD; in xfrm_dev_policy_add()
382 xdo->dev = NULL; in xfrm_dev_policy_add()
383 netdev_put(dev, &xdo->dev_tracker); in xfrm_dev_policy_add()
385 return -EINVAL; in xfrm_dev_policy_add()
388 err = dev->xfrmdev_ops->xdo_dev_policy_add(xp, extack); in xfrm_dev_policy_add()
390 xdo->dev = NULL; in xfrm_dev_policy_add()
391 xdo->real_dev = NULL; in xfrm_dev_policy_add()
392 xdo->type = XFRM_DEV_OFFLOAD_UNSPECIFIED; in xfrm_dev_policy_add()
393 xdo->dir = 0; in xfrm_dev_policy_add()
394 netdev_put(dev, &xdo->dev_tracker); in xfrm_dev_policy_add()
408 struct net_device *dev = x->xso.dev; in xfrm_dev_offload_ok()
410 if (!x->type_offload) in xfrm_dev_offload_ok()
413 if (x->xso.type == XFRM_DEV_OFFLOAD_PACKET || in xfrm_dev_offload_ok()
414 ((!dev || (dev == xfrm_dst_path(dst)->dev)) && in xfrm_dev_offload_ok()
415 !xdst->child->xfrm)) { in xfrm_dev_offload_ok()
416 mtu = xfrm_state_mtu(x, xdst->child_mtu_cached); in xfrm_dev_offload_ok()
417 if (skb->len <= mtu) in xfrm_dev_offload_ok()
427 if (dev && dev->xfrmdev_ops && dev->xfrmdev_ops->xdo_dev_offload_ok) in xfrm_dev_offload_ok()
428 return x->xso.dev->xfrmdev_ops->xdo_dev_offload_ok(skb, x); in xfrm_dev_offload_ok()
436 struct net_device *dev = skb->dev; in xfrm_dev_resume()
453 skb_queue_tail(&sd->xfrm_backlog, skb); in xfrm_dev_resume()
463 struct sk_buff_head *xfrm_backlog = &sd->xfrm_backlog; in xfrm_dev_backlog()
472 spin_lock(&xfrm_backlog->lock); in xfrm_dev_backlog()
474 spin_unlock(&xfrm_backlog->lock); in xfrm_dev_backlog()
487 if ((dev->features & NETIF_F_HW_ESP_TX_CSUM) && in xfrm_api_check()
488 !(dev->features & NETIF_F_HW_ESP)) in xfrm_api_check()
491 if ((dev->features & NETIF_F_HW_ESP) && in xfrm_api_check()
492 (!(dev->xfrmdev_ops && in xfrm_api_check()
493 dev->xfrmdev_ops->xdo_dev_state_add && in xfrm_api_check()
494 dev->xfrmdev_ops->xdo_dev_state_delete))) in xfrm_api_check()
497 if (dev->features & (NETIF_F_HW_ESP | NETIF_F_HW_ESP_TX_CSUM)) in xfrm_api_check()
506 if (dev->features & NETIF_F_HW_ESP) { in xfrm_dev_down()