/linux/include/linux/ |
H A D | atmdev.h | 247 refcount_add(skb->truesize, &sk_atm(vcc)->sk_wmem_alloc); in atm_account_tx() 255 &sk_atm(vcc)->sk_wmem_alloc)); in atm_return_tx() 272 return (size + refcount_read(&sk_atm(vcc)->sk_wmem_alloc)) < in atm_may_send()
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | bpf_iter_netlink.c | 46 s->sk_wmem_alloc.refs.counter - 1, in dump_netlink()
|
H A D | bpf_iter_udp4.c | 62 inet->sk.sk_wmem_alloc.refs.counter - 1, in dump_udp4()
|
H A D | bpf_iter_udp6.c | 70 inet->sk.sk_wmem_alloc.refs.counter - 1, in dump_udp6()
|
/linux/net/atm/ |
H A D | common.c | 84 if (refcount_read(&sk->sk_wmem_alloc)) in vcc_sock_destruct() 86 __func__, refcount_read(&sk->sk_wmem_alloc)); in vcc_sock_destruct() 105 refcount_read(&sk->sk_wmem_alloc)) <= sk->sk_sndbuf; in vcc_writable() 160 refcount_set(&sk->sk_wmem_alloc, 1); in vcc_create()
|
H A D | signaling.c | 71 WARN_ON(refcount_sub_and_test(skb->truesize, &sk_atm(vcc)->sk_wmem_alloc)); in sigd_send()
|
/linux/net/core/ |
H A D | sock.c | 2316 refcount_set(&sk->sk_wmem_alloc, 1); in sk_alloc() 2342 refcount_read(&sk->sk_wmem_alloc) == 0); in __sk_destruct() 2418 * We subtract one from sk_wmem_alloc and can know if in sk_free() 2422 if (refcount_dec_and_test(&sk->sk_wmem_alloc)) in sk_free() 2497 /* sk_wmem_alloc set to one (see sk_free() and sock_wfree()) */ in sk_clone_lock() 2498 refcount_set(&newsk->sk_wmem_alloc, 1); in sk_clone_lock() 2653 free = refcount_sub_and_test(len, &sk->sk_wmem_alloc); in sock_wfree() 2662 * Keep a reference on sk_wmem_alloc, this will be released in sock_wfree() 2665 WARN_ON(refcount_sub_and_test(len - 1, &sk->sk_wmem_alloc)); in sock_wfree() 2670 * if sk_wmem_alloc reache in sock_wfree() [all...] |
H A D | datagram.c | 754 refcount_add(truesize, &skb->sk->sk_wmem_alloc); in __zerocopy_sg_from_iter()
|
/linux/net/rds/ |
H A D | tcp_send.c | 202 if ((refcount_read(&sk->sk_wmem_alloc) << 1) <= sk->sk_sndbuf && in rds_tcp_write_space()
|
/linux/net/ipv4/ |
H A D | tcp_offload.c | 246 refcount_add(delta, &skb->sk->sk_wmem_alloc); in tcp_gso_segment() 248 WARN_ON_ONCE(refcount_sub_and_test(-delta, &skb->sk->sk_wmem_alloc)); in tcp_gso_segment()
|
H A D | inet_fragment.c | 543 * Update sk_wmem_alloc. in inet_frag_reasm_prepare() 547 refcount_add(ts_delta, &sk->sk_wmem_alloc); in inet_frag_reasm_prepare() 624 refcount_add(sum_truesize - head_truesize, &sk->sk_wmem_alloc); in inet_frag_reasm_finish()
|
H A D | udp_offload.c | 629 refcount_add(delta, &sk->sk_wmem_alloc); in __udp_gso_segment() 631 WARN_ON_ONCE(refcount_sub_and_test(-delta, &sk->sk_wmem_alloc)); in __udp_gso_segment()
|
H A D | tcp_output.c | 1067 * The invariant being skb->truesize subtracted from sk->sk_wmem_alloc 1209 /* Keep one reference on sk_wmem_alloc. in tcp_wfree() 1212 WARN_ON(refcount_sub_and_test(skb->truesize - 1, &sk->sk_wmem_alloc)); in tcp_wfree() 1221 if (refcount_read(&sk->sk_wmem_alloc) >= SKB_TRUESIZE(1) && this_cpu_ksoftirqd() == current) in tcp_wfree() 1387 refcount_add(skb->truesize, &sk->sk_wmem_alloc); in __tcp_transmit_skb() 2626 /* TSQ is based on skb truesize sum (sk_wmem_alloc), so we in tcp_small_queue_check() 2634 if (refcount_read(&sk->sk_wmem_alloc) > limit) { in tcp_small_queue_check() 2649 if (refcount_read(&sk->sk_wmem_alloc) > limit) in tcp_small_queue_check() 3755 * sk->sk_wmem_alloc in an atomic, we can promote to rw. in tcp_make_synack()
|
H A D | esp4.c | 468 refcount_add(tailen, &sk->sk_wmem_alloc); in esp_output_head()
|
/linux/net/kcm/ |
H A D | kcmproc.c | 152 refcount_read(&psock->sk->sk_wmem_alloc)); in kcm_format_psock()
|
/linux/include/trace/events/ |
H A D | sock.h | 120 __entry->wmem_alloc = refcount_read(&sk->sk_wmem_alloc);
|
/linux/net/sctp/ |
H A D | proc.c | 284 refcount_read(&sk->sk_wmem_alloc), in sctp_assocs_seq_show()
|
H A D | output.c | 431 refcount_add(skb->truesize, &head->sk->sk_wmem_alloc); in sctp_packet_gso_append()
|
/linux/net/rxrpc/ |
H A D | af_rxrpc.c | 53 return refcount_read(&sk->sk_wmem_alloc) < (size_t) sk->sk_sndbuf; in rxrpc_writable() 916 WARN_ON(refcount_read(&sk->sk_wmem_alloc)); in rxrpc_sock_destructor()
|
/linux/include/net/ |
H A D | sock.h | 253 * @sk_wmem_alloc: transmit queue bytes committed 465 refcount_t sk_wmem_alloc; member 2290 * Return: sk_wmem_alloc minus initial offset of one 2294 return refcount_read(&sk->sk_wmem_alloc) - 1; in sk_wmem_alloc_get() 2486 amt = sk->sk_sndbuf - refcount_read(&sk->sk_wmem_alloc); in sock_wspace() 2584 return refcount_read(&sk->sk_wmem_alloc) < (READ_ONCE(sk->sk_sndbuf) >> 1); in sock_writeable()
|
/linux/net/ipv6/ |
H A D | ip6_output.c | 1669 if (refcount_read(&sk->sk_wmem_alloc) + wmem_alloc_delta <= in __ip6_append_data() 1813 refcount_add(wmem_alloc_delta, &sk->sk_wmem_alloc); in __ip6_append_data() 1822 refcount_add(wmem_alloc_delta, &sk->sk_wmem_alloc); in __ip6_append_data()
|
/linux/net/phonet/ |
H A D | socket.c | 348 refcount_read(&sk->sk_wmem_alloc) < sk->sk_sndbuf && in pn_socket_poll()
|
/linux/net/xdp/ |
H A D | xsk.c | 655 refcount_add(ts, &xs->sk.sk_wmem_alloc); in xsk_build_skb_zerocopy() 718 refcount_add(PAGE_SIZE, &xs->sk.sk_wmem_alloc); in xsk_build_skb()
|
/linux/net/caif/ |
H A D | caif_socket.c | 1005 caif_assert(!refcount_read(&sk->sk_wmem_alloc)); in caif_sock_destructor()
|
/linux/net/unix/ |
H A D | af_unix.c | 594 (refcount_read(&sk->sk_wmem_alloc) << 2) <= READ_ONCE(sk->sk_sndbuf); in unix_writable() 641 DEBUG_NET_WARN_ON_ONCE(refcount_read(&sk->sk_wmem_alloc)); in unix_sock_destructor() 2480 refcount_add(size, &sk->sk_wmem_alloc); in unix_stream_sendmsg()
|