Searched refs:sk_pacing_rate (Results 1 – 16 of 16) sorted by relevance
| /linux/tools/testing/selftests/bpf/progs/ ! |
| H A D | tcp_ca_write_sk_pacing.c | 40 sk->sk_pacing_rate = min(rate, sk->sk_max_pacing_rate); in BPF_PROG()
|
| H A D | bpf_cc_cubic.c | 43 /* set sk_pacing_rate to 200 % of current rate (mss * cwnd / srtt) */ in tcp_update_pacing_rate() 47 * In Slow Start [1], set sk_pacing_rate to 200 % the current rate. in tcp_update_pacing_rate() 64 sk->sk_pacing_rate = min(rate, sk->sk_max_pacing_rate); in tcp_update_pacing_rate()
|
| H A D | bpf_cubic.c | 437 rate = sk->sk_pacing_rate; in hystart_ack_delay()
|
| /linux/net/ipv4/ ! |
| H A D | bpf_tcp_ca.c | 75 case offsetof(struct sock, sk_pacing_rate): in bpf_tcp_ca_btf_struct_access() 76 end = offsetofend(struct sock, sk_pacing_rate); in bpf_tcp_ca_btf_struct_access()
|
| H A D | tcp_bbr.c | 282 WRITE_ONCE(sk->sk_pacing_rate, in bbr_init_pacing_rate_from_rtt() 295 if (bbr_full_bw_reached(sk) || rate > READ_ONCE(sk->sk_pacing_rate)) in bbr_set_pacing_rate() 296 WRITE_ONCE(sk->sk_pacing_rate, rate); in bbr_set_pacing_rate() 302 return READ_ONCE(sk->sk_pacing_rate) < (bbr_min_tso_rate >> 3) ? 1 : 2; in bbr_min_tso_segs() 314 READ_ONCE(sk->sk_pacing_rate) >> READ_ONCE(sk->sk_pacing_shift), in bbr_tso_segs_goal()
|
| H A D | tcp_cubic.c | 376 rate = READ_ONCE(sk->sk_pacing_rate); in hystart_ack_delay()
|
| H A D | tcp_output.c | 1445 unsigned long rate = READ_ONCE(sk->sk_pacing_rate); in tcp_update_skb_after_send() 2259 bytes = READ_ONCE(sk->sk_pacing_rate) >> READ_ONCE(sk->sk_pacing_shift); in tcp_tso_segs() 2860 READ_ONCE(sk->sk_pacing_rate) >> READ_ONCE(sk->sk_pacing_shift)); in tcp_small_queue_check() 2867 u64 extra_bytes = (u64)READ_ONCE(sk->sk_pacing_rate) * in tcp_small_queue_check()
|
| H A D | tcp.c | 3620 * tp->rtt_min, icsk_rto and sk->sk_pacing_rate. in tcp_repair_options_est() 4229 rate = READ_ONCE(sk->sk_pacing_rate); in tcp_get_info_chrono_stats() 4435 rate = READ_ONCE(sk->sk_pacing_rate); in tcp_skb_ttl_or_hop_limit()
|
| H A D | tcp_input.c | 1143 /* set sk_pacing_rate to 200 % of current rate (mss * cwnd / srtt) */ in tcp_update_pacing_rate() 1147 * In Slow Start [1], set sk_pacing_rate to 200 % the current rate. in tcp_update_pacing_rate() 1164 /* WRITE_ONCE() is needed because sch_fq fetches sk_pacing_rate in tcp_update_pacing_rate() 1168 WRITE_ONCE(sk->sk_pacing_rate, in tcp_update_pacing_rate()
|
| /linux/include/trace/events/ ! |
| H A D | mptcp.h | 53 __entry->pace = READ_ONCE(ssk->sk_pacing_rate);
|
| /linux/net/core/ ! |
| H A D | sock.c | 1270 pacing_rate = READ_ONCE(sk->sk_pacing_rate); in sk_setsockopt() 1272 WRITE_ONCE(sk->sk_pacing_rate, ulval); in sk_setsockopt() 3755 sk->sk_pacing_rate = ~0UL; in sock_init_data_uid() 4545 CACHELINE_ASSERT_GROUP_MEMBER(struct sock, sock_write_tx, sk_pacing_rate); in sock_struct_check()
|
| /linux/net/sched/ ! |
| H A D | sch_fq.c | 17 * Transport (eg TCP) can set in sk->sk_pacing_rate a rate, enqueue a 364 * they do not contain sk_pacing_rate in fq_classify() 745 rate = min(READ_ONCE(skb->sk->sk_pacing_rate), rate); in fq_dequeue()
|
| H A D | Kconfig | 305 set by TCP stack into sk->sk_pacing_rate (for locally generated
|
| /linux/net/mptcp/ ! |
| H A D | protocol.c | 1521 subflow->avg_pacing_rate = READ_ONCE(ssk->sk_pacing_rate); in mptcp_subflow_get_send() 1561 READ_ONCE(ssk->sk_pacing_rate) * burst, in mptcp_update_post_push()
|
| /linux/include/net/ ! |
| H A D | sock.h | 271 * @sk_pacing_rate: Pacing rate (if supported by transport/packet scheduler) 498 unsigned long sk_pacing_rate; /* bytes per second */ 493 unsigned long sk_pacing_rate; /* bytes per second */ global() member
|
| /linux/Documentation/networking/ ! |
| H A D | ip-sysctl.rst | 1117 tso_packet_size = sk->sk_pacing_rate / 1024; 1132 sk->sk_pacing_rate is set by TCP stack using a ratio applied 1141 sk->sk_pacing_rate is set by TCP stack using a ratio applied
|