| /linux/include/net/ |
| H A D | inet_connection_sock.h | 152 #define inet_csk(ptr) container_of_const(ptr, struct inet_connection_sock, icsk_inet.sk) macro 156 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca() 181 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED; in inet_csk_schedule_ack() 186 return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED; in inet_csk_ack_scheduled() 191 memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack)); in inet_csk_delack_init() 207 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timer() 232 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_reset_xmit_timer() 283 reqsk_queue_added(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_added() 288 return reqsk_queue_len(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_len() 308 return !reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue) ? in inet_csk_listen_poll() [all …]
|
| H A D | tcp.h | 401 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_dec_quickack_mode() 855 return READ_ONCE(inet_csk(sk)->icsk_rto_max); in tcp_rto_max() 860 inet_csk(sk)->icsk_rto = min(inet_csk(sk)->icsk_rto, tcp_rto_max(sk)); in tcp_bound_rto() 882 u32 rto_min = READ_ONCE(inet_csk(sk)->icsk_rto_min); in tcp_rto_min() 1374 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_ecn() 1381 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_accecn() 1388 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_ect_1_negotiation() 1395 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_no_fallback_rfc3168() 1402 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_event() 1484 (1 << inet_csk(sk)->icsk_ca_state); in tcp_in_cwnd_reduction() [all …]
|
| H A D | espintcp.h | 35 const struct inet_connection_sock *icsk = inet_csk(sk); in espintcp_getctx()
|
| /linux/net/ipv4/ |
| H A D | tcp_recovery.c | 13 if (inet_csk(sk)->icsk_ca_state >= TCP_CA_Recovery) in tcp_rack_reo_wnd() 109 timeout, inet_csk(sk)->icsk_rto); in tcp_rack_mark_lost() 126 if (inet_csk(sk)->icsk_ca_state != TCP_CA_Recovery) { in tcp_rack_reo_timeout() 128 if (!inet_csk(sk)->icsk_ca_ops->cong_control) in tcp_rack_reo_timeout() 133 if (inet_csk(sk)->icsk_pending != ICSK_TIME_RETRANS) in tcp_rack_reo_timeout() 144 const u8 state = inet_csk(sk)->icsk_ca_state; in tcp_newreno_mark_lost()
|
| H A D | tcp_timer.c | 30 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_rto_to_user_timeout() 52 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_probe0_to_user_timeout() 221 if (!inet_csk(sk)->icsk_retransmits) in retransmits_timed_out() 244 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timeout() 310 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_delack_timer_handler() 390 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_probe_timer() 441 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_rto_stats() 458 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastopen_synack_timer() 497 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rtx_probe0_timed_out() 538 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_retransmit_timer() [all …]
|
| H A D | inet_connection_sock.c | 578 if (!inet_csk(sk)->icsk_bind_hash) in inet_csk_get_port() 580 WARN_ON(inet_csk(sk)->icsk_bind_hash != tb); in inet_csk_get_port() 581 WARN_ON(inet_csk(sk)->icsk_bind2_hash != tb2); in inet_csk_get_port() 604 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_wait_for_connect() 652 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_accept() 725 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_init_xmit_timers() 735 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers() 747 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers_sync() 1008 reqsk_queue_removed(&inet_csk(sk)->icsk_accept_queue, req); in __inet_csk_reqsk_queue_drop() 1055 reqsk_queue_migrated(&inet_csk(nsk)->icsk_accept_queue, req); in reqsk_timer_handler() [all …]
|
| H A D | tcp_ulp.c | 106 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_ulp() 114 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_ulp() 132 struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_set_ulp()
|
| H A D | tcp_cong.c | 40 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_ca_state() 219 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_assign_congestion_control() 238 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_congestion_control() 253 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_reinit_congestion_control() 272 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_congestion_control() 415 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_congestion_control()
|
| H A D | tcp_fastopen.c | 57 fastopenq = &inet_csk(lsk)->icsk_accept_queue.fastopenq; in reqsk_fastopen_remove() 130 inet_csk(sk)->icsk_accept_queue.fastopenq.ctx, 1); in tcp_fastopen_destroy_cipher() 169 q = &inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_reset_cipher() 331 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; in tcp_fastopen_create_child() 335 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_fastopen_create_child() 401 fastopenq = &inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_queue_check() 676 u32 timeouts = inet_csk(sk)->icsk_retransmits; in tcp_fastopen_active_detect_blackhole()
|
| H A D | tcp_dctcp.h | 29 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update() 33 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
|
| H A D | inet_hashtables.c | 204 inet_csk(sk)->icsk_bind_hash = tb; in inet_bind_hash() 205 inet_csk(sk)->icsk_bind2_hash = tb2; in inet_bind_hash() 225 tb = inet_csk(sk)->icsk_bind_hash; in __inet_put_port() 226 inet_csk(sk)->icsk_bind_hash = NULL; in __inet_put_port() 231 if (inet_csk(sk)->icsk_bind2_hash) { in __inet_put_port() 232 struct inet_bind2_bucket *tb2 = inet_csk(sk)->icsk_bind2_hash; in __inet_put_port() 235 inet_csk(sk)->icsk_bind2_hash = NULL; in __inet_put_port() 270 tb = inet_csk(sk)->icsk_bind_hash; in __inet_inherit_port() 271 tb2 = inet_csk(sk)->icsk_bind2_hash; in __inet_inherit_port() 761 struct inet_bind_bucket *tb = inet_csk(sk)->icsk_bind_hash; in inet_reuseport_add_sock() [all …]
|
| H A D | tcp_input.c | 229 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_measure_rcv_mss() 313 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_incr_quickack() 325 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_quickack_mode() 338 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_quickack_mode() 538 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in __tcp_accecn_process() 555 ((1 << inet_csk(sk)->icsk_ca_state) & in __tcp_accecn_process() 608 const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops; in tcp_sndbuf_expand() 673 return 2 * inet_csk(sk)->icsk_ack.rcv_mss; in __tcp_grow_window() 727 inet_csk(sk)->icsk_ack.quick |= 1; in tcp_grow_window() 781 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_window() [all …]
|
| H A D | tcp_output.c | 91 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_new_data_sent() 177 while ((delta -= inet_csk(sk)->icsk_rto) > 0 && cwnd > restart_cwnd) in tcp_cwnd_restart() 188 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_sent() 292 if (unlikely(inet_csk(sk)->icsk_ack.pending & ICSK_ACK_NOMEM)) { in tcp_select_window() 1045 inet_csk(sk)->icsk_retransmits < 2 && in tcp_syn_options() 1355 inet_csk(sk)->icsk_af_ops->mtu_reduced(sk); in tcp_release_cb() 1515 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_transmit_skb() 1976 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_mtu_to_mss() 2010 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mss_to_mtu() 2023 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_init() [all …]
|
| H A D | tcp_dctcp.c | 114 inet_csk(sk)->icsk_ca_ops = &dctcp_reno; in dctcp_init() 186 new_state != inet_csk(sk)->icsk_ca_state) in dctcp_state() 227 if (inet_csk(sk)->icsk_ca_ops != &dctcp_reno) { in dctcp_get_info()
|
| H A D | tcp_minisocks.c | 329 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_time_wait() 501 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_openreq_child() 564 newicsk = inet_csk(newsk); in tcp_create_openreq_child() 914 if (req->num_timeout < READ_ONCE(inet_csk(sk)->icsk_accept_queue.rskq_defer_accept) && in tcp_check_req() 927 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_check_req() 937 reqsk_queue_removed(&inet_csk(req->rsk_listener)->icsk_accept_queue, req); in tcp_check_req()
|
| H A D | tcp_diag.c | 118 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_diag_get_aux() 148 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_diag_get_aux_size() 284 return inet_sk_diag_fill(sk, inet_csk(sk), skb, cb, r, nlmsg_flags, in sk_diag_fill() 369 if (inet_sk_diag_fill(sk, inet_csk(sk), skb, in tcp_diag_dump()
|
| H A D | tcp_htcp.c | 84 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_rtt() 104 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_achieved_throughput()
|
| H A D | tcp.c | 431 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_sock() 1572 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_cleanup_rbuf() 3052 if (inet_csk(sk)->icsk_bind_hash && in tcp_set_state() 3397 inet_csk(sk)->icsk_backoff = 0; in tcp_write_queue_purge() 3403 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_disconnect() 3735 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_PUSHED; in __tcp_sock_set_quickack() 3755 WRITE_ONCE(inet_csk(sk)->icsk_syn_retries, val); in tcp_sock_set_syncnt() 3768 WRITE_ONCE(inet_csk(sk)->icsk_user_timeout, val); in tcp_sock_set_user_timeout() 3880 struct inet_connection_sock *icsk = inet_csk(sk); in do_tcp_setsockopt() 3981 WRITE_ONCE(inet_csk(sk)->icsk_rto_max, msecs_to_jiffies(val)); in do_tcp_setsockopt() [all …]
|
| /linux/net/mptcp/ |
| H A D | mptcp_diag.c | 23 return inet_sk_diag_fill(sk, inet_csk(sk), skb, cb, req, NLM_F_MULTI, in sk_diag_dump() 52 err = inet_sk_diag_fill(sk, inet_csk(sk), rep, cb, req, 0, in mptcp_diag_dump_one() 104 if (!ctx || strcmp(inet_csk(sk)->icsk_ulp_ops->name, "mptcp")) in mptcp_diag_dump_listeners()
|
| H A D | protocol.h | 598 const struct inet_connection_sock *icsk = inet_csk(sk); in mptcp_subflow_ctx() 872 inet_csk(sk)->icsk_af_ops = ctx->icsk_af_ops; in mptcp_subflow_tcp_fallback() 893 if (!inet_csk(sk)->icsk_mtup.probe_timestamp) in mptcp_stop_tout_timer() 896 sk_stop_timer(sk, &inet_csk(sk)->mptcp_tout_timer); in mptcp_stop_tout_timer() 897 inet_csk(sk)->icsk_mtup.probe_timestamp = 0; in mptcp_stop_tout_timer() 903 inet_csk(sk)->icsk_mtup.probe_timestamp = tout ? : 1; in mptcp_set_close_tout()
|
| H A D | subflow.c | 770 struct inet_connection_sock *icsk = inet_csk(sk); in subflow_ulp_fallback() 788 if (inet_csk(ssk)->icsk_ulp_ops) { in mptcp_subflow_drop_ctx() 1520 if (reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue)) in subflow_data_ready() 1537 (tcp_sk(sk)->rcv_nxt - tcp_sk(sk)->rcv_wup) > inet_csk(sk)->icsk_ack.rcv_mss) in subflow_data_ready() 1538 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in subflow_data_ready() 1566 struct inet_connection_sock *icsk = inet_csk(sk); in mptcpv6_handle_mapped() 1844 struct inet_connection_sock *icsk = inet_csk(sk); in subflow_create_ctx() 1895 struct request_sock_queue *queue = &inet_csk(listener_ssk)->icsk_accept_queue; in mptcp_subflow_queue_clean() 1966 struct inet_connection_sock *icsk = inet_csk(sk); in subflow_ulp_init()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | bpf_dctcp.c | 161 new_state != BPF_CORE_READ_BITFIELD(inet_csk(sk), icsk_ca_state)) in BPF_PROG() 193 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update() 197 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
|
| H A D | bpf_tracing_net.h | 163 static inline struct inet_connection_sock *inet_csk(const struct sock *sk) in inet_csk() function 170 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca()
|
| H A D | bpf_cc_cubic.c | 123 (1 << inet_csk(sk)->icsk_ca_state)) { in BPF_PROG() 130 inet_csk(sk)->icsk_ca_state == TCP_CA_CWR) { in BPF_PROG()
|
| /linux/net/tls/ |
| H A D | tls_toe.c | 48 struct inet_connection_sock *icsk = inet_csk(sk); in tls_toe_sk_destruct()
|