Home
last modified time | relevance | path

Searched refs:inet_csk (Results 1 – 25 of 56) sorted by relevance

123

/linux/include/net/
H A Dinet_connection_sock.h152 #define inet_csk(ptr) container_of_const(ptr, struct inet_connection_sock, icsk_inet.sk) macro
156 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca()
181 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED; in inet_csk_schedule_ack()
186 return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED; in inet_csk_ack_scheduled()
191 memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack)); in inet_csk_delack_init()
207 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timer()
232 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_reset_xmit_timer()
283 reqsk_queue_added(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_added()
288 return reqsk_queue_len(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_len()
308 return !reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue) ? in inet_csk_listen_poll()
[all …]
H A Dtcp.h401 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_dec_quickack_mode()
855 return READ_ONCE(inet_csk(sk)->icsk_rto_max); in tcp_rto_max()
860 inet_csk(sk)->icsk_rto = min(inet_csk(sk)->icsk_rto, tcp_rto_max(sk)); in tcp_bound_rto()
882 u32 rto_min = READ_ONCE(inet_csk(sk)->icsk_rto_min); in tcp_rto_min()
1374 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_ecn()
1381 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_accecn()
1388 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_ect_1_negotiation()
1395 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_no_fallback_rfc3168()
1402 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_event()
1484 (1 << inet_csk(sk)->icsk_ca_state); in tcp_in_cwnd_reduction()
[all …]
H A Despintcp.h35 const struct inet_connection_sock *icsk = inet_csk(sk); in espintcp_getctx()
/linux/net/ipv4/
H A Dtcp_recovery.c13 if (inet_csk(sk)->icsk_ca_state >= TCP_CA_Recovery) in tcp_rack_reo_wnd()
109 timeout, inet_csk(sk)->icsk_rto); in tcp_rack_mark_lost()
126 if (inet_csk(sk)->icsk_ca_state != TCP_CA_Recovery) { in tcp_rack_reo_timeout()
128 if (!inet_csk(sk)->icsk_ca_ops->cong_control) in tcp_rack_reo_timeout()
133 if (inet_csk(sk)->icsk_pending != ICSK_TIME_RETRANS) in tcp_rack_reo_timeout()
144 const u8 state = inet_csk(sk)->icsk_ca_state; in tcp_newreno_mark_lost()
H A Dtcp_timer.c30 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_rto_to_user_timeout()
52 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_probe0_to_user_timeout()
221 if (!inet_csk(sk)->icsk_retransmits) in retransmits_timed_out()
244 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timeout()
310 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_delack_timer_handler()
390 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_probe_timer()
441 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_rto_stats()
458 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastopen_synack_timer()
497 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rtx_probe0_timed_out()
538 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_retransmit_timer()
[all …]
H A Dinet_connection_sock.c578 if (!inet_csk(sk)->icsk_bind_hash) in inet_csk_get_port()
580 WARN_ON(inet_csk(sk)->icsk_bind_hash != tb); in inet_csk_get_port()
581 WARN_ON(inet_csk(sk)->icsk_bind2_hash != tb2); in inet_csk_get_port()
604 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_wait_for_connect()
652 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_accept()
725 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_init_xmit_timers()
735 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers()
747 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers_sync()
1008 reqsk_queue_removed(&inet_csk(sk)->icsk_accept_queue, req); in __inet_csk_reqsk_queue_drop()
1055 reqsk_queue_migrated(&inet_csk(nsk)->icsk_accept_queue, req); in reqsk_timer_handler()
[all …]
H A Dtcp_ulp.c106 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_ulp()
114 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_ulp()
132 struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_set_ulp()
H A Dtcp_cong.c40 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_ca_state()
219 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_assign_congestion_control()
238 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_congestion_control()
253 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_reinit_congestion_control()
272 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_congestion_control()
415 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_congestion_control()
H A Dtcp_fastopen.c57 fastopenq = &inet_csk(lsk)->icsk_accept_queue.fastopenq; in reqsk_fastopen_remove()
130 inet_csk(sk)->icsk_accept_queue.fastopenq.ctx, 1); in tcp_fastopen_destroy_cipher()
169 q = &inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_reset_cipher()
331 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; in tcp_fastopen_create_child()
335 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_fastopen_create_child()
401 fastopenq = &inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_queue_check()
676 u32 timeouts = inet_csk(sk)->icsk_retransmits; in tcp_fastopen_active_detect_blackhole()
H A Dtcp_dctcp.h29 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update()
33 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
H A Dinet_hashtables.c204 inet_csk(sk)->icsk_bind_hash = tb; in inet_bind_hash()
205 inet_csk(sk)->icsk_bind2_hash = tb2; in inet_bind_hash()
225 tb = inet_csk(sk)->icsk_bind_hash; in __inet_put_port()
226 inet_csk(sk)->icsk_bind_hash = NULL; in __inet_put_port()
231 if (inet_csk(sk)->icsk_bind2_hash) { in __inet_put_port()
232 struct inet_bind2_bucket *tb2 = inet_csk(sk)->icsk_bind2_hash; in __inet_put_port()
235 inet_csk(sk)->icsk_bind2_hash = NULL; in __inet_put_port()
270 tb = inet_csk(sk)->icsk_bind_hash; in __inet_inherit_port()
271 tb2 = inet_csk(sk)->icsk_bind2_hash; in __inet_inherit_port()
761 struct inet_bind_bucket *tb = inet_csk(sk)->icsk_bind_hash; in inet_reuseport_add_sock()
[all …]
H A Dtcp_input.c229 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_measure_rcv_mss()
313 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_incr_quickack()
325 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_quickack_mode()
338 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_quickack_mode()
538 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in __tcp_accecn_process()
555 ((1 << inet_csk(sk)->icsk_ca_state) & in __tcp_accecn_process()
608 const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops; in tcp_sndbuf_expand()
673 return 2 * inet_csk(sk)->icsk_ack.rcv_mss; in __tcp_grow_window()
727 inet_csk(sk)->icsk_ack.quick |= 1; in tcp_grow_window()
781 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_window()
[all …]
H A Dtcp_output.c91 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_new_data_sent()
177 while ((delta -= inet_csk(sk)->icsk_rto) > 0 && cwnd > restart_cwnd) in tcp_cwnd_restart()
188 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_sent()
292 if (unlikely(inet_csk(sk)->icsk_ack.pending & ICSK_ACK_NOMEM)) { in tcp_select_window()
1045 inet_csk(sk)->icsk_retransmits < 2 && in tcp_syn_options()
1355 inet_csk(sk)->icsk_af_ops->mtu_reduced(sk); in tcp_release_cb()
1515 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_transmit_skb()
1976 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_mtu_to_mss()
2010 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mss_to_mtu()
2023 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_init()
[all …]
H A Dtcp_dctcp.c114 inet_csk(sk)->icsk_ca_ops = &dctcp_reno; in dctcp_init()
186 new_state != inet_csk(sk)->icsk_ca_state) in dctcp_state()
227 if (inet_csk(sk)->icsk_ca_ops != &dctcp_reno) { in dctcp_get_info()
H A Dtcp_minisocks.c329 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_time_wait()
501 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_openreq_child()
564 newicsk = inet_csk(newsk); in tcp_create_openreq_child()
914 if (req->num_timeout < READ_ONCE(inet_csk(sk)->icsk_accept_queue.rskq_defer_accept) && in tcp_check_req()
927 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_check_req()
937 reqsk_queue_removed(&inet_csk(req->rsk_listener)->icsk_accept_queue, req); in tcp_check_req()
H A Dtcp_diag.c118 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_diag_get_aux()
148 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_diag_get_aux_size()
284 return inet_sk_diag_fill(sk, inet_csk(sk), skb, cb, r, nlmsg_flags, in sk_diag_fill()
369 if (inet_sk_diag_fill(sk, inet_csk(sk), skb, in tcp_diag_dump()
H A Dtcp_htcp.c84 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_rtt()
104 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_achieved_throughput()
H A Dtcp.c431 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_sock()
1572 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_cleanup_rbuf()
3052 if (inet_csk(sk)->icsk_bind_hash && in tcp_set_state()
3397 inet_csk(sk)->icsk_backoff = 0; in tcp_write_queue_purge()
3403 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_disconnect()
3735 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_PUSHED; in __tcp_sock_set_quickack()
3755 WRITE_ONCE(inet_csk(sk)->icsk_syn_retries, val); in tcp_sock_set_syncnt()
3768 WRITE_ONCE(inet_csk(sk)->icsk_user_timeout, val); in tcp_sock_set_user_timeout()
3880 struct inet_connection_sock *icsk = inet_csk(sk); in do_tcp_setsockopt()
3981 WRITE_ONCE(inet_csk(sk)->icsk_rto_max, msecs_to_jiffies(val)); in do_tcp_setsockopt()
[all …]
/linux/net/mptcp/
H A Dmptcp_diag.c23 return inet_sk_diag_fill(sk, inet_csk(sk), skb, cb, req, NLM_F_MULTI, in sk_diag_dump()
52 err = inet_sk_diag_fill(sk, inet_csk(sk), rep, cb, req, 0, in mptcp_diag_dump_one()
104 if (!ctx || strcmp(inet_csk(sk)->icsk_ulp_ops->name, "mptcp")) in mptcp_diag_dump_listeners()
H A Dprotocol.h598 const struct inet_connection_sock *icsk = inet_csk(sk); in mptcp_subflow_ctx()
872 inet_csk(sk)->icsk_af_ops = ctx->icsk_af_ops; in mptcp_subflow_tcp_fallback()
893 if (!inet_csk(sk)->icsk_mtup.probe_timestamp) in mptcp_stop_tout_timer()
896 sk_stop_timer(sk, &inet_csk(sk)->mptcp_tout_timer); in mptcp_stop_tout_timer()
897 inet_csk(sk)->icsk_mtup.probe_timestamp = 0; in mptcp_stop_tout_timer()
903 inet_csk(sk)->icsk_mtup.probe_timestamp = tout ? : 1; in mptcp_set_close_tout()
H A Dsubflow.c770 struct inet_connection_sock *icsk = inet_csk(sk); in subflow_ulp_fallback()
788 if (inet_csk(ssk)->icsk_ulp_ops) { in mptcp_subflow_drop_ctx()
1520 if (reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue)) in subflow_data_ready()
1537 (tcp_sk(sk)->rcv_nxt - tcp_sk(sk)->rcv_wup) > inet_csk(sk)->icsk_ack.rcv_mss) in subflow_data_ready()
1538 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in subflow_data_ready()
1566 struct inet_connection_sock *icsk = inet_csk(sk); in mptcpv6_handle_mapped()
1844 struct inet_connection_sock *icsk = inet_csk(sk); in subflow_create_ctx()
1895 struct request_sock_queue *queue = &inet_csk(listener_ssk)->icsk_accept_queue; in mptcp_subflow_queue_clean()
1966 struct inet_connection_sock *icsk = inet_csk(sk); in subflow_ulp_init()
/linux/tools/testing/selftests/bpf/progs/
H A Dbpf_dctcp.c161 new_state != BPF_CORE_READ_BITFIELD(inet_csk(sk), icsk_ca_state)) in BPF_PROG()
193 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update()
197 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
H A Dbpf_tracing_net.h163 static inline struct inet_connection_sock *inet_csk(const struct sock *sk) in inet_csk() function
170 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca()
H A Dbpf_cc_cubic.c123 (1 << inet_csk(sk)->icsk_ca_state)) { in BPF_PROG()
130 inet_csk(sk)->icsk_ca_state == TCP_CA_CWR) { in BPF_PROG()
/linux/net/tls/
H A Dtls_toe.c48 struct inet_connection_sock *icsk = inet_csk(sk); in tls_toe_sk_destruct()

123