Home
last modified time | relevance | path

Searched refs:ack_seq (Results 1 – 25 of 37) sorted by relevance

12

/linux/net/netfilter/
H A Dnf_conntrack_seqadj.c198 if (after(ntohl(tcph->ack_seq) - other_way->offset_before, in nf_ct_seq_adjust()
204 newack = htonl(ntohl(tcph->ack_seq) - ackoff); in nf_ct_seq_adjust()
205 inet_proto_csum_replace4(&tcph->check, skb, tcph->ack_seq, newack, in nf_ct_seq_adjust()
208 ntohl(tcph->seq), ntohl(newseq), ntohl(tcph->ack_seq), in nf_ct_seq_adjust()
210 tcph->ack_seq = newack; in nf_ct_seq_adjust()
H A Dnf_log_syslog.c211 ntohl(th->seq), ntohl(th->ack_seq)); in nf_log_dump_tcp_header()
/linux/net/netfilter/ipvs/
H A Dip_vs_app.c312 * Fixes th->ack_seq based on ip_vs_seq info.
317 __u32 ack_seq = ntohl(th->ack_seq); in vs_fix_ack_seq() local
320 * Adjust ack_seq with delta-offset for in vs_fix_ack_seq()
325 /* since ack_seq is the number of octet that is expected in vs_fix_ack_seq()
327 if(after(ack_seq, vseq->init_seq+vseq->delta)) { in vs_fix_ack_seq()
328 th->ack_seq = htonl(ack_seq - vseq->delta); in vs_fix_ack_seq()
330 "(%d) from ack_seq\n", __func__, vseq->delta); in vs_fix_ack_seq()
333 th->ack_seq in vs_fix_ack_seq()
[all...]
/linux/drivers/net/slip/
H A Dslhc.c384 if((deltaA = ntohl(th->ack_seq) - ntohl(oth->ack_seq)) != 0L){ in slhc_compress()
554 thp->ack_seq = htonl( ntohl(thp->ack_seq) + i); in slhc_uncompress()
583 thp->ack_seq = htonl( ntohl(thp->ack_seq) + x); in slhc_uncompress()
/linux/net/mptcp/
H A Doptions.c563 u64 ack_seq; in mptcp_established_options_dss() local
594 ack_seq = READ_ONCE(msk->ack_seq); in mptcp_established_options_dss()
597 opts->ext_copy.data_ack = ack_seq; in mptcp_established_options_dss()
601 opts->ext_copy.data_ack32 = (uint32_t)ack_seq; in mptcp_established_options_dss()
1097 mptcp_expand_seq(READ_ONCE(msk->ack_seq), data_fin_seq, use_64bit)); in mptcp_update_rcv_data_fin()
1204 /* we can't wait for recvmsg() to update the ack_seq, otherwise in mptcp_incoming_options()
1262 u64 ack_seq, rcv_wnd_old, rcv_wnd_new; in mptcp_set_rwin() local
1270 ack_seq = READ_ONCE(msk->ack_seq); in mptcp_set_rwin()
[all...]
H A Dsubflow.c495 WRITE_ONCE(msk->ack_seq, subflow->iasn); in subflow_set_remote_key()
1190 map_seq = mptcp_expand_seq(READ_ONCE(msk->ack_seq), mpext->data_seq, mpext->dsn64); in get_mapping_status()
1302 mptcp_update_rcv_data_fin(msk, READ_ONCE(msk->ack_seq), true)) in subflow_sched_work_if_closed()
1357 u64 ack_seq; in subflow_check_data_avail() local
1376 old_ack = READ_ONCE(msk->ack_seq); in subflow_check_data_avail()
1377 ack_seq = mptcp_subflow_get_mapped_dsn(subflow); in subflow_check_data_avail()
1378 pr_debug("msk ack_seq=%llx subflow ack_seq=%llx\n", old_ack, in subflow_check_data_avail()
1379 ack_seq); in subflow_check_data_avail()
1380 if (unlikely(before64(ack_seq, old_ac in subflow_check_data_avail()
[all...]
H A Dprotocol.c328 if (MPTCP_SKB_CB(skb)->map_seq == msk->ack_seq) { in __mptcp_move_skb()
331 WRITE_ONCE(msk->ack_seq, msk->ack_seq + copy_len); in __mptcp_move_skb()
339 } else if (after64(MPTCP_SKB_CB(skb)->map_seq, msk->ack_seq)) { in __mptcp_move_skb()
416 if (READ_ONCE(msk->ack_seq) == rcv_data_fin_seq) { in mptcp_pending_data_fin()
552 WRITE_ONCE(msk->ack_seq, msk->ack_seq + 1); in mptcp_check_data_fin()
673 if (after64(MPTCP_SKB_CB(skb)->map_seq, msk->ack_seq)) in __mptcp_ofo_queue()
680 msk->ack_seq))) { in __mptcp_ofo_queue()
689 int delta = msk->ack_seq in __mptcp_ofo_queue()
[all...]
/linux/net/ipv4/netfilter/
H A Dnf_reject_ipv4.c206 tcph->seq = oth->ack_seq; in nf_reject_ip_tcphdr_put()
208 tcph->ack_seq = htonl(ntohl(oth->seq) + oth->syn + oth->fin + in nf_reject_ip_tcphdr_put()
/linux/net/ipv4/
H A Dtcp_input.c123 void (*cad)(struct sock *sk, u32 ack_seq)) in clean_acked_data_enable() argument
1304 if (before(start_seq_0, TCP_SKB_CB(ack_skb)->ack_seq)) { in tcp_check_dsack()
1872 if (before(TCP_SKB_CB(ack_skb)->ack_seq, prior_snd_una - tp->max_window)) in tcp_sacktag_write_queue()
1898 if ((TCP_SKB_CB(ack_skb)->ack_seq != tp->snd_una) && in tcp_sacktag_write_queue()
3499 const u32 ack, const u32 ack_seq, in tcp_may_update_window() argument
3503 after(ack_seq, tp->snd_wl1) || in tcp_may_update_window()
3504 (ack_seq == tp->snd_wl1 && (nwin > tp->snd_wnd || !nwin)); in tcp_may_update_window()
3569 u32 ack_seq) in tcp_ack_update_window() argument
3578 if (tcp_may_update_window(tp, ack, ack_seq, nwin)) { in tcp_ack_update_window()
3580 tcp_update_wl(tp, ack_seq); in tcp_ack_update_window()
3813 u32 ack_seq = TCP_SKB_CB(skb)->seq; tcp_ack() local
[all...]
H A Dsyncookies.c185 __u32 cookie = ntohl(th->ack_seq) - 1; in __cookie_v4_check()
286 treq->snt_isn = ntohl(th->ack_seq) - 1; in cookie_tcp_reqsk_init()
H A Dtcp_offload.c336 flush |= (__force int)(th->ack_seq ^ th2->ack_seq); in tcp_gro_receive()
H A Dtcp_ipv4.c781 rep.th.seq = th->ack_seq; in tcp_v4_send_reset()
784 rep.th.ack_seq = htonl(ntohl(th->seq) + th->syn + th->fin + in tcp_v4_send_reset()
972 rep.th.ack_seq = htonl(ack); in tcp_v4_send_ack()
2089 if (likely(!before(TCP_SKB_CB(skb)->ack_seq, TCP_SKB_CB(tail)->ack_seq))) {
2090 TCP_SKB_CB(tail)->ack_seq = TCP_SKB_CB(skb)->ack_seq;
2183 TCP_SKB_CB(skb)->ack_seq = ntohl(th->ack_seq);
H A Dtcp_minisocks.c763 A: sends SYN|ACK, seq=7, ack_seq=8 in tcp_check_req()
764 B: sends SYN|ACK, seq=7, ack_seq=8 in tcp_check_req()
799 (TCP_SKB_CB(skb)->ack_seq != in tcp_check_req()
/linux/net/rds/
H A Dib_cm.c78 __be64 ack_seq = 0; in rds_ib_cm_connect_complete() local
96 ack_seq = get_unaligned(&dp->ricp_v6.dp_ack_seq); in rds_ib_cm_connect_complete()
103 ack_seq = get_unaligned(&dp->ricp_v4.dp_ack_seq); in rds_ib_cm_connect_complete()
154 if (ack_seq) in rds_ib_cm_connect_complete()
155 rds_send_drop_acked(conn, be64_to_cpu(ack_seq), in rds_ib_cm_connect_complete()
/linux/net/ipv6/netfilter/
H A Dnf_reject_ipv6.c235 tcph->seq = oth->ack_seq; in nf_reject_ip6_tcphdr_put()
237 tcph->ack_seq = htonl(ntohl(oth->seq) + oth->syn + oth->fin + in nf_reject_ip6_tcphdr_put()
/linux/tools/testing/selftests/bpf/progs/
H A Dtest_btf_skc_cls_ingress.c61 recv_cookie = bpf_ntohl(th->ack_seq) - 1; in test_syncookie_helper()
H A Dtest_tcp_custom_syncookie.c418 ctx->tcp->ack_seq = bpf_htonl(bpf_ntohl(ctx->tcp->seq) + 1); in tcp_handle_syn()
466 u32 cookie = bpf_ntohl(ctx->tcp->ack_seq) - 1; in tcp_validate_cookie()
H A Dxdp_synproxy_kern.c525 tcp_header->ack_seq = bpf_htonl(bpf_ntohl(tcp_header->seq) + 1); in tcp_gen_synack()
/linux/include/linux/ceph/
H A Dmsgr.h202 __le64 ack_seq; member
H A Dmessenger.h532 void ceph_con_discard_sent(struct ceph_connection *con, u64 ack_seq);
/linux/drivers/infiniband/hw/irdma/
H A Dcm.c355 tcph->ack_seq = htonl(cm_node->tcp_cntxt.loc_ack_num); in irdma_form_ah_cm_frame()
358 tcph->ack_seq = 0; in irdma_form_ah_cm_frame()
543 tcph->ack_seq = htonl(cm_node->tcp_cntxt.loc_ack_num); in irdma_form_uda_cm_frame()
546 tcph->ack_seq = 0; in irdma_form_uda_cm_frame()
2623 if (ntohl(tcph->ack_seq) != cm_node->tcp_cntxt.loc_seq_num) { in irdma_check_syn()
2639 u32 ack_seq; in irdma_check_seq() local
2646 ack_seq = ntohl(tcph->ack_seq); in irdma_check_seq()
2648 if (ack_seq != loc_seq_num || in irdma_check_seq()
2761 cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->ack_seq); in irdma_handle_synack_pkt()
[all...]
/linux/tools/include/uapi/linux/
H A Dtcp.h29 __be32 ack_seq; member
/linux/include/uapi/linux/
H A Dtcp.h29 __be32 ack_seq; member
/linux/drivers/net/ethernet/chelsio/cxgb/
H A Dcpl5_cmd.h393 u32 ack_seq; member
/linux/net/sched/
H A Dsch_cake.c1046 u32 ack_seq_a = ntohl(tcph_a->ack_seq); in cake_tcph_sack_compare()
1294 after(ntohl(tcph_check->ack_seq), ntohl(tcph->ack_seq))) in cake_ack_filter()
1306 (ntohl(tcph_check->ack_seq) == ntohl(tcph->ack_seq) && in cake_ack_filter()

12