| /linux/net/ipv4/ |
| H A D | tcp_timer.c | 30 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_rto_to_user_timeout() local 35 user_timeout = READ_ONCE(icsk->icsk_user_timeout); in tcp_clamp_rto_to_user_timeout() 37 return icsk->icsk_rto; in tcp_clamp_rto_to_user_timeout() 47 return min_t(u32, icsk->icsk_rto, msecs_to_jiffies(remaining)); in tcp_clamp_rto_to_user_timeout() 52 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_probe0_to_user_timeout() local 56 user_timeout = READ_ONCE(icsk->icsk_user_timeout); in tcp_clamp_probe0_to_user_timeout() 57 if (!user_timeout || !icsk->icsk_probes_tstamp) in tcp_clamp_probe0_to_user_timeout() 60 elapsed = tcp_jiffies32 - icsk->icsk_probes_tstamp; in tcp_clamp_probe0_to_user_timeout() 165 static void tcp_mtu_probing(struct inet_connection_sock *icsk, struct sock *sk) in tcp_mtu_probing() argument 174 if (!icsk->icsk_mtup.enabled) { in tcp_mtu_probing() [all …]
|
| H A D | tcp_cong.c | 40 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_ca_state() local 44 if (icsk->icsk_ca_ops->set_state) in tcp_set_ca_state() 45 icsk->icsk_ca_ops->set_state(sk, ca_state); in tcp_set_ca_state() 46 icsk->icsk_ca_state = ca_state; in tcp_set_ca_state() 219 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_assign_congestion_control() local 226 icsk->icsk_ca_ops = ca; in tcp_assign_congestion_control() 229 memset(icsk->icsk_ca_priv, 0, sizeof(icsk->icsk_ca_priv)); in tcp_assign_congestion_control() 238 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_congestion_control() local 241 if (icsk->icsk_ca_ops->init) in tcp_init_congestion_control() 242 icsk->icsk_ca_ops->init(sk); in tcp_init_congestion_control() [all …]
|
| H A D | tcp_ulp.c | 106 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_ulp() local 108 if (icsk->icsk_ulp_ops->update) in tcp_update_ulp() 109 icsk->icsk_ulp_ops->update(sk, proto, write_space); in tcp_update_ulp() 114 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_ulp() local 120 if (!icsk->icsk_ulp_ops) in tcp_cleanup_ulp() 123 if (icsk->icsk_ulp_ops->release) in tcp_cleanup_ulp() 124 icsk->icsk_ulp_ops->release(sk); in tcp_cleanup_ulp() 125 module_put(icsk->icsk_ulp_ops->owner); in tcp_cleanup_ulp() 127 icsk->icsk_ulp_ops = NULL; in tcp_cleanup_ulp() 132 struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_set_ulp() local [all …]
|
| H A D | inet_connection_sock.c | 604 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_wait_for_connect() local 626 if (reqsk_queue_empty(&icsk->icsk_accept_queue)) in inet_csk_wait_for_connect() 631 if (!reqsk_queue_empty(&icsk->icsk_accept_queue)) in inet_csk_wait_for_connect() 652 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_accept() local 653 struct request_sock_queue *queue = &icsk->icsk_accept_queue; in inet_csk_accept() 725 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_init_xmit_timers() local 728 timer_setup(&icsk->icsk_delack_timer, delack_handler, 0); in inet_csk_init_xmit_timers() 729 timer_setup(&icsk->icsk_keepalive_timer, keepalive_handler, 0); in inet_csk_init_xmit_timers() 730 icsk->icsk_pending = icsk->icsk_ack.pending = 0; in inet_csk_init_xmit_timers() 735 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers() local [all …]
|
| H A D | tcp_output.c | 91 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_new_data_sent() local 104 if (!prior_packets || icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) in tcp_event_new_data_sent() 188 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_sent() local 199 if ((u32)(now - icsk->icsk_ack.lrcvtime) < icsk->icsk_ack.ato) in tcp_event_data_sent() 1515 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_transmit_skb() local 1662 INDIRECT_CALL_INET(icsk->icsk_af_ops->send_check, in __tcp_transmit_skb() 1693 err = INDIRECT_CALL_INET(icsk->icsk_af_ops->queue_xmit, in __tcp_transmit_skb() 1976 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_mtu_to_mss() local 1982 mss_now = pmtu - icsk->icsk_af_ops->net_header_len - sizeof(struct tcphdr); in __tcp_mtu_to_mss() 1989 mss_now -= icsk->icsk_ext_hdr_len; in __tcp_mtu_to_mss() [all …]
|
| H A D | tcp_input.c | 229 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_measure_rcv_mss() local 230 const unsigned int lss = icsk->icsk_ack.last_seg_size; in tcp_measure_rcv_mss() 233 icsk->icsk_ack.last_seg_size = 0; in tcp_measure_rcv_mss() 239 if (len >= icsk->icsk_ack.rcv_mss) { in tcp_measure_rcv_mss() 244 if (unlikely(len != icsk->icsk_ack.rcv_mss)) { in tcp_measure_rcv_mss() 261 icsk->icsk_ack.rcv_mss = min_t(unsigned int, len, in tcp_measure_rcv_mss() 264 DO_ONCE_LITE_IF(len > icsk->icsk_ack.rcv_mss + MAX_TCP_OPTION_SPACE, in tcp_measure_rcv_mss() 278 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED; in tcp_measure_rcv_mss() 299 icsk->icsk_ack.last_seg_size = len; in tcp_measure_rcv_mss() 301 icsk->icsk_ack.rcv_mss = len; in tcp_measure_rcv_mss() [all …]
|
| H A D | inet_timewait_sock.c | 109 const struct inet_connection_sock *icsk = inet_csk(sk); in inet_twsk_hashdance_schedule() local 125 tw->tw_tb = icsk->icsk_bind_hash; in inet_twsk_hashdance_schedule() 126 WARN_ON(!icsk->icsk_bind_hash); in inet_twsk_hashdance_schedule() 128 tw->tw_tb2 = icsk->icsk_bind2_hash; in inet_twsk_hashdance_schedule() 129 WARN_ON(!icsk->icsk_bind2_hash); in inet_twsk_hashdance_schedule()
|
| H A D | tcp_htcp.c | 84 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_rtt() local 92 if (icsk->icsk_ca_state == TCP_CA_Open) { in measure_rtt() 104 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_achieved_throughput() local 109 if (icsk->icsk_ca_state == TCP_CA_Open) in measure_achieved_throughput() 119 if (!((1 << icsk->icsk_ca_state) & (TCPF_CA_Open | TCPF_CA_Disorder))) { in measure_achieved_throughput()
|
| H A D | tcp.c | 431 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_sock() local 441 icsk->icsk_rto = TCP_TIMEOUT_INIT; in tcp_init_sock() 444 icsk->icsk_rto_max = msecs_to_jiffies(rto_max_ms); in tcp_init_sock() 447 icsk->icsk_rto_min = usecs_to_jiffies(rto_min_us); in tcp_init_sock() 448 icsk->icsk_delack_max = TCP_DELACK_MAX; in tcp_init_sock() 479 icsk->icsk_sync_mss = tcp_sync_mss; in tcp_init_sock() 1572 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_cleanup_rbuf() local 1575 tp->rcv_nxt - tp->rcv_wup > icsk->icsk_ack.rcv_mss || in __tcp_cleanup_rbuf() 1583 ((icsk->icsk_ack.pending & ICSK_ACK_PUSHED2) || in __tcp_cleanup_rbuf() 1584 ((icsk->icsk_ack.pending & ICSK_ACK_PUSHED) && in __tcp_cleanup_rbuf() [all …]
|
| H A D | tcp_minisocks.c | 329 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_time_wait() local 338 const int rto = (icsk->icsk_rto << 2) - (icsk->icsk_rto >> 1); in tcp_time_wait() 501 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_openreq_child() local 511 icsk->icsk_ca_dst_locked = tcp_ca_dst_locked(dst); in tcp_ca_openreq_child() 512 icsk->icsk_ca_ops = ca; in tcp_ca_openreq_child() 520 (!icsk->icsk_ca_setsockopt || in tcp_ca_openreq_child() 521 !bpf_try_module_get(icsk->icsk_ca_ops, icsk->icsk_ca_ops->owner))) in tcp_ca_openreq_child()
|
| H A D | inet_diag.c | 209 int inet_sk_diag_fill(struct sock *sk, struct inet_connection_sock *icsk, in inet_sk_diag_fill() argument 278 if (!icsk) { in inet_sk_diag_fill() 283 icsk_pending = smp_load_acquire(&icsk->icsk_pending); in inet_sk_diag_fill() 288 r->idiag_retrans = READ_ONCE(icsk->icsk_retransmits); in inet_sk_diag_fill() 293 r->idiag_retrans = READ_ONCE(icsk->icsk_probes_out); in inet_sk_diag_fill() 296 } else if (timer_pending(&icsk->icsk_keepalive_timer)) { in inet_sk_diag_fill() 298 r->idiag_retrans = READ_ONCE(icsk->icsk_probes_out); in inet_sk_diag_fill() 300 jiffies_delta_to_msecs(icsk->icsk_keepalive_timer.expires - jiffies); in inet_sk_diag_fill() 317 ca_ops = READ_ONCE(icsk->icsk_ca_ops); in inet_sk_diag_fill() 337 ca_ops = READ_ONCE(icsk->icsk_ca_ops); in inet_sk_diag_fill()
|
| H A D | tcp_nv.c | 241 const struct inet_connection_sock *icsk = inet_csk(sk); in tcpnv_acked() local 255 if (icsk->icsk_ca_state != TCP_CA_Open && in tcpnv_acked() 256 icsk->icsk_ca_state != TCP_CA_Disorder) in tcpnv_acked()
|
| H A D | tcp_bic.c | 192 const struct inet_connection_sock *icsk = inet_csk(sk); in bictcp_acked() local 194 if (icsk->icsk_ca_state == TCP_CA_Open) { in bictcp_acked()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | bpf_iter_tcp4.c | 64 inet_csk_in_pingpong_mode(const struct inet_connection_sock *icsk) in inet_csk_in_pingpong_mode() argument 66 return icsk->icsk_ack.pingpong >= TCP_PINGPONG_THRESH; in inet_csk_in_pingpong_mode() 77 const struct inet_connection_sock *icsk; in dump_tcp_sock() local 88 icsk = &tp->inet_conn; in dump_tcp_sock() 89 inet = &icsk->icsk_inet; in dump_tcp_sock() 91 fastopenq = &icsk->icsk_accept_queue.fastopenq; in dump_tcp_sock() 98 if (icsk->icsk_pending == ICSK_TIME_RETRANS || in dump_tcp_sock() 99 icsk->icsk_pending == ICSK_TIME_REO_TIMEOUT || in dump_tcp_sock() 100 icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) { in dump_tcp_sock() 103 } else if (icsk->icsk_pending == ICSK_TIME_PROBE0) { in dump_tcp_sock() [all …]
|
| H A D | bpf_iter_tcp6.c | 64 inet_csk_in_pingpong_mode(const struct inet_connection_sock *icsk) in inet_csk_in_pingpong_mode() argument 66 return icsk->icsk_ack.pingpong >= TCP_PINGPONG_THRESH; in inet_csk_in_pingpong_mode() 77 const struct inet_connection_sock *icsk; in dump_tcp6_sock() local 88 icsk = &tp->tcp.inet_conn; in dump_tcp6_sock() 89 inet = &icsk->icsk_inet; in dump_tcp6_sock() 91 fastopenq = &icsk->icsk_accept_queue.fastopenq; in dump_tcp6_sock() 98 if (icsk->icsk_pending == ICSK_TIME_RETRANS || in dump_tcp6_sock() 99 icsk->icsk_pending == ICSK_TIME_REO_TIMEOUT || in dump_tcp6_sock() 100 icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) { in dump_tcp6_sock() 103 } else if (icsk->icsk_pending == ICSK_TIME_PROBE0) { in dump_tcp6_sock() [all …]
|
| H A D | mptcp_subflow.c | 88 struct inet_connection_sock *icsk; in _check_getsockopt_subflow_cc() local 93 icsk = bpf_core_cast(ssk, struct inet_connection_sock); in _check_getsockopt_subflow_cc() 96 __builtin_memcmp(icsk->icsk_ca_ops->name, cc, TCP_CA_NAME_MAX)) { in _check_getsockopt_subflow_cc()
|
| H A D | sock_destroy_prog.c | 77 const struct inet_connection_sock *icsk; in iter_tcp6_server() local 92 icsk = &tcp_sk->tcp.inet_conn; in iter_tcp6_server() 93 inet = &icsk->icsk_inet; in iter_tcp6_server()
|
| /linux/include/net/ |
| H A D | inet_connection_sock.h | 200 icsk_delack_timeout(const struct inet_connection_sock *icsk) in icsk_delack_timeout() argument 202 return READ_ONCE(icsk->icsk_delack_timer.expires); in icsk_delack_timeout() 207 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timer() local 210 smp_store_release(&icsk->icsk_pending, 0); in inet_csk_clear_xmit_timer() 215 smp_store_release(&icsk->icsk_ack.pending, 0); in inet_csk_clear_xmit_timer() 216 icsk->icsk_ack.retry = 0; in inet_csk_clear_xmit_timer() 218 sk_stop_timer(sk, &icsk->icsk_delack_timer); in inet_csk_clear_xmit_timer() 232 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_reset_xmit_timer() local 243 smp_store_release(&icsk->icsk_pending, what); in inet_csk_reset_xmit_timer() 246 smp_store_release(&icsk->icsk_ack.pending, in inet_csk_reset_xmit_timer() [all …]
|
| H A D | espintcp.h | 35 const struct inet_connection_sock *icsk = inet_csk(sk); in espintcp_getctx() local 38 return (__force void *)icsk->icsk_ulp_data; in espintcp_getctx()
|
| H A D | tcp.h | 401 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_dec_quickack_mode() local 403 if (icsk->icsk_ack.quick) { in tcp_dec_quickack_mode() 407 if (pkts >= icsk->icsk_ack.quick) { in tcp_dec_quickack_mode() 408 icsk->icsk_ack.quick = 0; in tcp_dec_quickack_mode() 410 icsk->icsk_ack.ato = TCP_ATO_MIN; in tcp_dec_quickack_mode() 412 icsk->icsk_ack.quick -= pkts; in tcp_dec_quickack_mode() 1374 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_ecn() local 1376 return icsk->icsk_ca_ops->flags & TCP_CONG_NEEDS_ECN; in tcp_ca_needs_ecn() 1381 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_accecn() local 1383 return icsk->icsk_ca_ops->flags & TCP_CONG_NEEDS_ACCECN; in tcp_ca_needs_accecn() [all …]
|
| /linux/net/mptcp/ |
| H A D | token_test.c | 34 struct inet_connection_sock *icsk; in build_icsk() local 36 icsk = kunit_kzalloc(test, sizeof(struct inet_connection_sock), in build_icsk() 38 KUNIT_EXPECT_NOT_ERR_OR_NULL(test, icsk); in build_icsk() 39 return icsk; in build_icsk() 73 struct inet_connection_sock *icsk = build_icsk(test); in mptcp_token_test_msk_basic() local 79 rcu_assign_pointer(icsk->icsk_ulp_data, ctx); in mptcp_token_test_msk_basic() 84 mptcp_token_new_connect((struct sock *)icsk)); in mptcp_token_test_msk_basic()
|
| H A D | subflow.c | 770 struct inet_connection_sock *icsk = inet_csk(sk); in subflow_ulp_fallback() local 773 icsk->icsk_ulp_ops = NULL; in subflow_ulp_fallback() 774 rcu_assign_pointer(icsk->icsk_ulp_data, NULL); in subflow_ulp_fallback() 1566 struct inet_connection_sock *icsk = inet_csk(sk); in mptcpv6_handle_mapped() local 1572 subflow, sk->sk_family, icsk->icsk_af_ops, target, mapped); in mptcpv6_handle_mapped() 1574 if (likely(icsk->icsk_af_ops == target)) in mptcpv6_handle_mapped() 1577 subflow->icsk_af_ops = icsk->icsk_af_ops; in mptcpv6_handle_mapped() 1578 icsk->icsk_af_ops = target; in mptcpv6_handle_mapped() 1844 struct inet_connection_sock *icsk = inet_csk(sk); in subflow_create_ctx() local 1851 rcu_assign_pointer(icsk->icsk_ulp_data, ctx); in subflow_create_ctx() [all …]
|
| /linux/net/psp/ |
| H A D | psp_sock.c | 181 struct inet_connection_sock *icsk; in psp_sock_assoc_set_tx() local 238 icsk = inet_csk(sk); in psp_sock_assoc_set_tx() 239 icsk->icsk_ext_hdr_len += psp_sk_overhead(sk); in psp_sock_assoc_set_tx() 240 icsk->icsk_sync_mss(sk, icsk->icsk_pmtu_cookie); in psp_sock_assoc_set_tx()
|
| /linux/net/tls/ |
| H A D | tls_toe.c | 48 struct inet_connection_sock *icsk = inet_csk(sk); in tls_toe_sk_destruct() local 53 rcu_assign_pointer(icsk->icsk_ulp_data, NULL); in tls_toe_sk_destruct()
|
| /linux/net/ipv6/ |
| H A D | tcp_ipv6.c | 137 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_v6_connect() local 220 u32 exthdrlen = icsk->icsk_ext_hdr_len; in tcp_v6_connect() 231 WRITE_ONCE(icsk->icsk_af_ops, &ipv6_mapped); in tcp_v6_connect() 242 icsk->icsk_ext_hdr_len = exthdrlen; in tcp_v6_connect() 244 WRITE_ONCE(icsk->icsk_af_ops, &ipv6_specific); in tcp_v6_connect() 302 icsk->icsk_ext_hdr_len = psp_sk_overhead(sk); in tcp_v6_connect() 304 icsk->icsk_ext_hdr_len += opt->opt_flen + in tcp_v6_connect() 2080 struct inet_connection_sock *icsk = inet_csk(sk); local 2084 icsk->icsk_af_ops = &ipv6_specific; 2136 const struct inet_connection_sock *icsk = inet_csk(sp); local [all …]
|