Lines Matching refs:icsk
229 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_measure_rcv_mss() local
230 const unsigned int lss = icsk->icsk_ack.last_seg_size; in tcp_measure_rcv_mss()
233 icsk->icsk_ack.last_seg_size = 0; in tcp_measure_rcv_mss()
239 if (len >= icsk->icsk_ack.rcv_mss) { in tcp_measure_rcv_mss()
244 if (unlikely(len != icsk->icsk_ack.rcv_mss)) { in tcp_measure_rcv_mss()
261 icsk->icsk_ack.rcv_mss = min_t(unsigned int, len, in tcp_measure_rcv_mss()
264 DO_ONCE_LITE_IF(len > icsk->icsk_ack.rcv_mss + MAX_TCP_OPTION_SPACE, in tcp_measure_rcv_mss()
278 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED; in tcp_measure_rcv_mss()
299 icsk->icsk_ack.last_seg_size = len; in tcp_measure_rcv_mss()
301 icsk->icsk_ack.rcv_mss = len; in tcp_measure_rcv_mss()
305 if (icsk->icsk_ack.pending & ICSK_ACK_PUSHED) in tcp_measure_rcv_mss()
306 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED2; in tcp_measure_rcv_mss()
307 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED; in tcp_measure_rcv_mss()
313 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_incr_quickack() local
314 unsigned int quickacks = tcp_sk(sk)->rcv_wnd / (2 * icsk->icsk_ack.rcv_mss); in tcp_incr_quickack()
319 if (quickacks > icsk->icsk_ack.quick) in tcp_incr_quickack()
320 icsk->icsk_ack.quick = quickacks; in tcp_incr_quickack()
325 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_quickack_mode() local
329 icsk->icsk_ack.ato = TCP_ATO_MIN; in tcp_enter_quickack_mode()
338 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_quickack_mode() local
340 return icsk->icsk_ack.dst_quick_ack || in tcp_in_quickack_mode()
341 (icsk->icsk_ack.quick && !inet_csk_in_pingpong_mode(sk)); in tcp_in_quickack_mode()
781 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_window() local
785 icsk->icsk_ack.quick = 0; in tcp_clamp_window()
999 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_save_lrcv_flowlabel() local
1002 icsk->icsk_ack.lrcv_flowlabel = ntohl(ip6_flowlabel(ipv6_hdr(skb))); in tcp_save_lrcv_flowlabel()
1019 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_recv() local
1030 if (!icsk->icsk_ack.ato) { in tcp_event_data_recv()
1035 icsk->icsk_ack.ato = TCP_ATO_MIN; in tcp_event_data_recv()
1037 int m = now - icsk->icsk_ack.lrcvtime; in tcp_event_data_recv()
1041 icsk->icsk_ack.ato = (icsk->icsk_ack.ato >> 1) + TCP_ATO_MIN / 2; in tcp_event_data_recv()
1042 } else if (m < icsk->icsk_ack.ato) { in tcp_event_data_recv()
1043 icsk->icsk_ack.ato = (icsk->icsk_ack.ato >> 1) + m; in tcp_event_data_recv()
1044 if (icsk->icsk_ack.ato > icsk->icsk_rto) in tcp_event_data_recv()
1045 icsk->icsk_ack.ato = icsk->icsk_rto; in tcp_event_data_recv()
1046 } else if (m > icsk->icsk_rto) { in tcp_event_data_recv()
1053 icsk->icsk_ack.lrcvtime = now; in tcp_event_data_recv()
2555 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_loss() local
2558 bool new_recovery = icsk->icsk_ca_state < TCP_CA_Recovery; in tcp_enter_loss()
2564 if (icsk->icsk_ca_state <= TCP_CA_Disorder || in tcp_enter_loss()
2566 (icsk->icsk_ca_state == TCP_CA_Loss && !icsk->icsk_retransmits)) { in tcp_enter_loss()
2569 tp->snd_ssthresh = icsk->icsk_ca_ops->ssthresh(sk); in tcp_enter_loss()
2581 if (icsk->icsk_ca_state <= TCP_CA_Disorder && in tcp_enter_loss()
2596 (new_recovery || icsk->icsk_retransmits) && in tcp_enter_loss()
2857 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_undo_cwnd_reduction() local
2859 tcp_snd_cwnd_set(tp, icsk->icsk_ca_ops->undo_cwnd(sk)); in tcp_undo_cwnd_reduction()
3075 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_probe_failed() local
3077 icsk->icsk_mtup.search_high = icsk->icsk_mtup.probe_size - 1; in tcp_mtup_probe_failed()
3078 icsk->icsk_mtup.probe_size = 0; in tcp_mtup_probe_failed()
3085 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_probe_success() local
3091 do_div(val, icsk->icsk_mtup.probe_size); in tcp_mtup_probe_success()
3099 icsk->icsk_mtup.search_low = icsk->icsk_mtup.probe_size; in tcp_mtup_probe_success()
3100 icsk->icsk_mtup.probe_size = 0; in tcp_mtup_probe_success()
3101 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_mtup_probe_success()
3117 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_non_congestion_loss_retransmit() local
3120 if (icsk->icsk_ca_state != TCP_CA_Loss) { in tcp_non_congestion_loss_retransmit()
3331 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastretrans_alert() local
3353 if (icsk->icsk_ca_state == TCP_CA_Open) { in tcp_fastretrans_alert()
3357 switch (icsk->icsk_ca_state) { in tcp_fastretrans_alert()
3378 switch (icsk->icsk_ca_state) { in tcp_fastretrans_alert()
3390 if (icsk->icsk_ca_state != TCP_CA_Recovery) { in tcp_fastretrans_alert()
3401 if (icsk->icsk_ca_state != TCP_CA_Loss) in tcp_fastretrans_alert()
3404 if (!(icsk->icsk_ca_state == TCP_CA_Open || in tcp_fastretrans_alert()
3416 if (icsk->icsk_ca_state <= TCP_CA_Disorder) in tcp_fastretrans_alert()
3426 if (icsk->icsk_ca_state < TCP_CA_CWR && in tcp_fastretrans_alert()
3427 icsk->icsk_mtup.probe_size && in tcp_fastretrans_alert()
3515 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cong_avoid() local
3517 icsk->icsk_ca_ops->cong_avoid(sk, ack, acked); in tcp_cong_avoid()
3526 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rearm_rto() local
3540 if (icsk->icsk_pending == ICSK_TIME_REO_TIMEOUT || in tcp_rearm_rto()
3541 icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) { in tcp_rearm_rto()
3606 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clean_rtx_queue() local
3737 if (unlikely(icsk->icsk_mtup.probe_size && in tcp_clean_rtx_queue()
3772 if (icsk->icsk_ca_ops->pkts_acked) { in tcp_clean_rtx_queue()
3778 icsk->icsk_ca_ops->pkts_acked(sk, &sample); in tcp_clean_rtx_queue()
3786 icsk = inet_csk(sk); in tcp_clean_rtx_queue()
3789 tp->lost_out, icsk->icsk_ca_state); in tcp_clean_rtx_queue()
3794 tp->sacked_out, icsk->icsk_ca_state); in tcp_clean_rtx_queue()
3799 tp->retrans_out, icsk->icsk_ca_state); in tcp_clean_rtx_queue()
3809 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ack_probe() local
3817 icsk->icsk_backoff = 0; in tcp_ack_probe()
3818 icsk->icsk_probes_tstamp = 0; in tcp_ack_probe()
3861 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cong_control() local
3863 if (icsk->icsk_ca_ops->cong_control) { in tcp_cong_control()
3864 icsk->icsk_ca_ops->cong_control(sk, ack, flag, rs); in tcp_cong_control()
4144 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_ack_event() local
4146 if (icsk->icsk_ca_ops->in_ack_event) { in tcp_in_ack_event()
4157 icsk->icsk_ca_ops->in_ack_event(sk, ack_ev_flags); in tcp_in_ack_event()
4248 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ack() local
4297 WRITE_ONCE(icsk->icsk_retransmits, 0); in tcp_ack()
4360 WRITE_ONCE(icsk->icsk_probes_out, 0); in tcp_ack()
6727 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_transfer() local
6731 icsk->icsk_af_ops->rebuild_header(sk); in tcp_init_transfer()
6748 if (!icsk->icsk_ca_initialized) in tcp_init_transfer()
6756 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_finish_connect() local
6760 icsk->icsk_ack.lrcvtime = tcp_jiffies32; in tcp_finish_connect()
6763 icsk->icsk_af_ops->sk_rx_dst_set(sk, skb); in tcp_finish_connect()
6875 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rcv_synsent_state_process() local
6898 if (icsk->icsk_retransmits == 0) in tcp_rcv_synsent_state_process()
6982 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_rcv_synsent_state_process()
7006 READ_ONCE(icsk->icsk_accept_queue.rskq_defer_accept) || in tcp_rcv_synsent_state_process()
7083 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_rcv_synsent_state_process()
7173 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rcv_state_process() local
7202 icsk->icsk_af_ops->conn_request(sk, skb); in tcp_rcv_state_process()