/linux/net/netfilter/ |
H A D | nf_conntrack_seqadj.c | 198 if (after(ntohl(tcph->ack_seq) - other_way->offset_before, in nf_ct_seq_adjust() 204 newack = htonl(ntohl(tcph->ack_seq) - ackoff); in nf_ct_seq_adjust() 205 inet_proto_csum_replace4(&tcph->check, skb, tcph->ack_seq, newack, in nf_ct_seq_adjust() 208 ntohl(tcph->seq), ntohl(newseq), ntohl(tcph->ack_seq), in nf_ct_seq_adjust() 210 tcph->ack_seq = newack; in nf_ct_seq_adjust()
|
H A D | nf_log_syslog.c | 211 ntohl(th->seq), ntohl(th->ack_seq)); in nf_log_dump_tcp_header()
|
/linux/net/netfilter/ipvs/ |
H A D | ip_vs_app.c | 312 * Fixes th->ack_seq based on ip_vs_seq info. 317 __u32 ack_seq = ntohl(th->ack_seq); in vs_fix_ack_seq() local 320 * Adjust ack_seq with delta-offset for in vs_fix_ack_seq() 325 /* since ack_seq is the number of octet that is expected in vs_fix_ack_seq() 327 if(after(ack_seq, vseq->init_seq+vseq->delta)) { in vs_fix_ack_seq() 328 th->ack_seq = htonl(ack_seq - vseq->delta); in vs_fix_ack_seq() 330 "(%d) from ack_seq\n", __func__, vseq->delta); in vs_fix_ack_seq() 333 th->ack_seq in vs_fix_ack_seq() [all...] |
/linux/drivers/net/slip/ |
H A D | slhc.c | 384 if((deltaA = ntohl(th->ack_seq) - ntohl(oth->ack_seq)) != 0L){ in slhc_compress() 554 thp->ack_seq = htonl( ntohl(thp->ack_seq) + i); in slhc_uncompress() 583 thp->ack_seq = htonl( ntohl(thp->ack_seq) + x); in slhc_uncompress()
|
/linux/net/mptcp/ |
H A D | options.c | 563 u64 ack_seq; in mptcp_established_options_dss() local 594 ack_seq = READ_ONCE(msk->ack_seq); in mptcp_established_options_dss() 597 opts->ext_copy.data_ack = ack_seq; in mptcp_established_options_dss() 601 opts->ext_copy.data_ack32 = (uint32_t)ack_seq; in mptcp_established_options_dss() 1097 mptcp_expand_seq(READ_ONCE(msk->ack_seq), data_fin_seq, use_64bit)); in mptcp_update_rcv_data_fin() 1204 /* we can't wait for recvmsg() to update the ack_seq, otherwise in mptcp_incoming_options() 1262 u64 ack_seq, rcv_wnd_old, rcv_wnd_new; in mptcp_set_rwin() local 1270 ack_seq = READ_ONCE(msk->ack_seq); in mptcp_set_rwin() [all...] |
H A D | subflow.c | 495 WRITE_ONCE(msk->ack_seq, subflow->iasn); in subflow_set_remote_key() 1190 map_seq = mptcp_expand_seq(READ_ONCE(msk->ack_seq), mpext->data_seq, mpext->dsn64); in get_mapping_status() 1302 mptcp_update_rcv_data_fin(msk, READ_ONCE(msk->ack_seq), true)) in subflow_sched_work_if_closed() 1357 u64 ack_seq; in subflow_check_data_avail() local 1376 old_ack = READ_ONCE(msk->ack_seq); in subflow_check_data_avail() 1377 ack_seq = mptcp_subflow_get_mapped_dsn(subflow); in subflow_check_data_avail() 1378 pr_debug("msk ack_seq=%llx subflow ack_seq=%llx\n", old_ack, in subflow_check_data_avail() 1379 ack_seq); in subflow_check_data_avail() 1380 if (unlikely(before64(ack_seq, old_ac in subflow_check_data_avail() [all...] |
H A D | protocol.c | 328 if (MPTCP_SKB_CB(skb)->map_seq == msk->ack_seq) { in __mptcp_move_skb() 331 WRITE_ONCE(msk->ack_seq, msk->ack_seq + copy_len); in __mptcp_move_skb() 339 } else if (after64(MPTCP_SKB_CB(skb)->map_seq, msk->ack_seq)) { in __mptcp_move_skb() 416 if (READ_ONCE(msk->ack_seq) == rcv_data_fin_seq) { in mptcp_pending_data_fin() 552 WRITE_ONCE(msk->ack_seq, msk->ack_seq + 1); in mptcp_check_data_fin() 673 if (after64(MPTCP_SKB_CB(skb)->map_seq, msk->ack_seq)) in __mptcp_ofo_queue() 680 msk->ack_seq))) { in __mptcp_ofo_queue() 689 int delta = msk->ack_seq in __mptcp_ofo_queue() [all...] |
/linux/net/ipv4/netfilter/ |
H A D | nf_reject_ipv4.c | 206 tcph->seq = oth->ack_seq; in nf_reject_ip_tcphdr_put() 208 tcph->ack_seq = htonl(ntohl(oth->seq) + oth->syn + oth->fin + in nf_reject_ip_tcphdr_put()
|
/linux/net/ipv4/ |
H A D | tcp_input.c | 123 void (*cad)(struct sock *sk, u32 ack_seq)) in clean_acked_data_enable() argument 1304 if (before(start_seq_0, TCP_SKB_CB(ack_skb)->ack_seq)) { in tcp_check_dsack() 1872 if (before(TCP_SKB_CB(ack_skb)->ack_seq, prior_snd_una - tp->max_window)) in tcp_sacktag_write_queue() 1898 if ((TCP_SKB_CB(ack_skb)->ack_seq != tp->snd_una) && in tcp_sacktag_write_queue() 3499 const u32 ack, const u32 ack_seq, in tcp_may_update_window() argument 3503 after(ack_seq, tp->snd_wl1) || in tcp_may_update_window() 3504 (ack_seq == tp->snd_wl1 && (nwin > tp->snd_wnd || !nwin)); in tcp_may_update_window() 3569 u32 ack_seq) in tcp_ack_update_window() argument 3578 if (tcp_may_update_window(tp, ack, ack_seq, nwin)) { in tcp_ack_update_window() 3580 tcp_update_wl(tp, ack_seq); in tcp_ack_update_window() 3813 u32 ack_seq = TCP_SKB_CB(skb)->seq; tcp_ack() local [all...] |
H A D | syncookies.c | 185 __u32 cookie = ntohl(th->ack_seq) - 1; in __cookie_v4_check() 286 treq->snt_isn = ntohl(th->ack_seq) - 1; in cookie_tcp_reqsk_init()
|
H A D | tcp_offload.c | 336 flush |= (__force int)(th->ack_seq ^ th2->ack_seq); in tcp_gro_receive()
|
H A D | tcp_ipv4.c | 781 rep.th.seq = th->ack_seq; in tcp_v4_send_reset() 784 rep.th.ack_seq = htonl(ntohl(th->seq) + th->syn + th->fin + in tcp_v4_send_reset() 972 rep.th.ack_seq = htonl(ack); in tcp_v4_send_ack() 2089 if (likely(!before(TCP_SKB_CB(skb)->ack_seq, TCP_SKB_CB(tail)->ack_seq))) { 2090 TCP_SKB_CB(tail)->ack_seq = TCP_SKB_CB(skb)->ack_seq; 2183 TCP_SKB_CB(skb)->ack_seq = ntohl(th->ack_seq);
|
H A D | tcp_minisocks.c | 763 A: sends SYN|ACK, seq=7, ack_seq=8 in tcp_check_req() 764 B: sends SYN|ACK, seq=7, ack_seq=8 in tcp_check_req() 799 (TCP_SKB_CB(skb)->ack_seq != in tcp_check_req()
|
/linux/net/rds/ |
H A D | ib_cm.c | 78 __be64 ack_seq = 0; in rds_ib_cm_connect_complete() local 96 ack_seq = get_unaligned(&dp->ricp_v6.dp_ack_seq); in rds_ib_cm_connect_complete() 103 ack_seq = get_unaligned(&dp->ricp_v4.dp_ack_seq); in rds_ib_cm_connect_complete() 154 if (ack_seq) in rds_ib_cm_connect_complete() 155 rds_send_drop_acked(conn, be64_to_cpu(ack_seq), in rds_ib_cm_connect_complete()
|
/linux/net/ipv6/netfilter/ |
H A D | nf_reject_ipv6.c | 235 tcph->seq = oth->ack_seq; in nf_reject_ip6_tcphdr_put() 237 tcph->ack_seq = htonl(ntohl(oth->seq) + oth->syn + oth->fin + in nf_reject_ip6_tcphdr_put()
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | test_btf_skc_cls_ingress.c | 61 recv_cookie = bpf_ntohl(th->ack_seq) - 1; in test_syncookie_helper()
|
H A D | test_tcp_custom_syncookie.c | 418 ctx->tcp->ack_seq = bpf_htonl(bpf_ntohl(ctx->tcp->seq) + 1); in tcp_handle_syn() 466 u32 cookie = bpf_ntohl(ctx->tcp->ack_seq) - 1; in tcp_validate_cookie()
|
H A D | xdp_synproxy_kern.c | 525 tcp_header->ack_seq = bpf_htonl(bpf_ntohl(tcp_header->seq) + 1); in tcp_gen_synack()
|
/linux/include/linux/ceph/ |
H A D | msgr.h | 202 __le64 ack_seq; member
|
H A D | messenger.h | 532 void ceph_con_discard_sent(struct ceph_connection *con, u64 ack_seq);
|
/linux/drivers/infiniband/hw/irdma/ |
H A D | cm.c | 355 tcph->ack_seq = htonl(cm_node->tcp_cntxt.loc_ack_num); in irdma_form_ah_cm_frame() 358 tcph->ack_seq = 0; in irdma_form_ah_cm_frame() 543 tcph->ack_seq = htonl(cm_node->tcp_cntxt.loc_ack_num); in irdma_form_uda_cm_frame() 546 tcph->ack_seq = 0; in irdma_form_uda_cm_frame() 2623 if (ntohl(tcph->ack_seq) != cm_node->tcp_cntxt.loc_seq_num) { in irdma_check_syn() 2639 u32 ack_seq; in irdma_check_seq() local 2646 ack_seq = ntohl(tcph->ack_seq); in irdma_check_seq() 2648 if (ack_seq != loc_seq_num || in irdma_check_seq() 2761 cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->ack_seq); in irdma_handle_synack_pkt() [all...] |
/linux/tools/include/uapi/linux/ |
H A D | tcp.h | 29 __be32 ack_seq; member
|
/linux/include/uapi/linux/ |
H A D | tcp.h | 29 __be32 ack_seq; member
|
/linux/drivers/net/ethernet/chelsio/cxgb/ |
H A D | cpl5_cmd.h | 393 u32 ack_seq; member
|
/linux/net/sched/ |
H A D | sch_cake.c | 1046 u32 ack_seq_a = ntohl(tcph_a->ack_seq); in cake_tcph_sack_compare() 1294 after(ntohl(tcph_check->ack_seq), ntohl(tcph->ack_seq))) in cake_ack_filter() 1306 (ntohl(tcph_check->ack_seq) == ntohl(tcph->ack_seq) && in cake_ack_filter()
|