| /linux/net/x25/ |
| H A D | x25_out.c | 100 skb_queue_tail(&sk->sk_write_queue, skbn); in x25_output() 106 skb_queue_tail(&sk->sk_write_queue, skb); in x25_output() 159 if (!skb_peek(&sk->sk_write_queue)) in x25_kick() 177 skb = skb_dequeue(&sk->sk_write_queue); in x25_kick() 181 skb_queue_head(&sk->sk_write_queue, skb); in x25_kick() 200 (skb = skb_dequeue(&sk->sk_write_queue)) != NULL); in x25_kick()
|
| H A D | x25_subr.c | 38 skb_queue_purge(&sk->sk_write_queue); in x25_clear_queues() 80 skb_queue_head(&sk->sk_write_queue, skb); in x25_requeue_frames() 82 skb_append(skb_prev, skb, &sk->sk_write_queue); in x25_requeue_frames()
|
| /linux/net/nfc/ |
| H A D | rawsock.c | 41 spin_lock_bh(&sk->sk_write_queue.lock); in rawsock_write_queue_purge() 42 __skb_queue_purge(&sk->sk_write_queue); in rawsock_write_queue_purge() 44 spin_unlock_bh(&sk->sk_write_queue.lock); in rawsock_write_queue_purge() 170 spin_lock_bh(&sk->sk_write_queue.lock); in rawsock_data_exchange_complete() 171 if (!skb_queue_empty(&sk->sk_write_queue)) in rawsock_data_exchange_complete() 175 spin_unlock_bh(&sk->sk_write_queue.lock); in rawsock_data_exchange_complete() 203 skb = skb_dequeue(&sk->sk_write_queue); in rawsock_tx_work() 241 spin_lock_bh(&sk->sk_write_queue.lock); in rawsock_sendmsg() 242 __skb_queue_tail(&sk->sk_write_queue, skb); in rawsock_sendmsg() 247 spin_unlock_bh(&sk->sk_write_queue.lock); in rawsock_sendmsg()
|
| /linux/net/rose/ |
| H A D | rose_out.c | 57 if (!skb_peek(&sk->sk_write_queue)) in rose_kick() 73 skb = skb_dequeue(&sk->sk_write_queue); in rose_kick() 77 skb_queue_head(&sk->sk_write_queue, skb); in rose_kick() 96 (skb = skb_dequeue(&sk->sk_write_queue)) != NULL); in rose_kick()
|
| H A D | rose_subr.c | 34 skb_queue_purge(&sk->sk_write_queue); in rose_clear_queues() 71 skb_queue_head(&sk->sk_write_queue, skb); in rose_requeue_frames() 73 skb_append(skb_prev, skb, &sk->sk_write_queue); in rose_requeue_frames()
|
| H A D | af_rose.c | 1262 skb_queue_tail(&sk->sk_write_queue, skbn); /* Throw it on the queue */ in rose_sendmsg() 1268 skb_queue_tail(&sk->sk_write_queue, skb); /* Throw it on the queue */ in rose_sendmsg() 1271 skb_queue_tail(&sk->sk_write_queue, skb); /* Shove it onto the queue */ in rose_sendmsg()
|
| /linux/net/netrom/ |
| H A D | nr_out.c | 66 skb_queue_tail(&sk->sk_write_queue, skbn); /* Throw it on the queue */ in nr_output() 71 skb_queue_tail(&sk->sk_write_queue, skb); /* Throw it on the queue */ in nr_output() 136 if (!skb_peek(&sk->sk_write_queue)) in nr_kick() 155 skb = skb_dequeue(&sk->sk_write_queue); in nr_kick() 159 skb_queue_head(&sk->sk_write_queue, skb); in nr_kick() 178 (skb = skb_dequeue(&sk->sk_write_queue)) != NULL); in nr_kick()
|
| H A D | nr_subr.c | 35 skb_queue_purge(&sk->sk_write_queue); in nr_clear_queues() 74 skb_queue_head(&sk->sk_write_queue, skb); in nr_requeue_frames() 76 skb_append(skb_prev, skb, &sk->sk_write_queue); in nr_requeue_frames()
|
| /linux/net/bluetooth/bnep/ |
| H A D | netdev.c | 104 skb_queue_tail(&sk->sk_write_queue, skb); in bnep_net_set_mc_list() 192 skb_queue_tail(&sk->sk_write_queue, skb); in bnep_net_xmit() 195 if (skb_queue_len(&sk->sk_write_queue) >= BNEP_TX_QUEUE_LEN) { in bnep_net_xmit()
|
| H A D | core.c | 508 while ((skb = skb_dequeue(&sk->sk_write_queue))) in bnep_session()
|
| /linux/net/llc/ |
| H A D | llc_conn.c | 181 skb_queue_tail(&sk->sk_write_queue, skb); in llc_conn_send_pdu() 232 skb_queue_tail(&sk->sk_write_queue, skb); in llc_conn_resend_i_pdu_as_cmd() 273 skb_queue_tail(&sk->sk_write_queue, skb); in llc_conn_resend_i_pdu_as_rsp() 332 while ((skb = skb_dequeue(&sk->sk_write_queue)) != NULL) { in llc_conn_send_pdus() 983 skb_queue_len(&sk->sk_write_queue)); in llc_sk_free() 986 skb_queue_purge(&sk->sk_write_queue); in llc_sk_free() 1015 skb_queue_purge(&sk->sk_write_queue); in llc_sk_reset()
|
| /linux/net/kcm/ |
| H A D | kcmsock.c | 599 if (skb_queue_empty(&sk->sk_write_queue)) in kcm_write_msgs() 602 kcm_tx_msg(skb_peek(&sk->sk_write_queue))->started_tx = false; in kcm_write_msgs() 606 while ((head = skb_peek(&sk->sk_write_queue))) { in kcm_write_msgs() 694 skb_dequeue(&sk->sk_write_queue); in kcm_write_msgs() 701 WARN_ON(!skb_queue_empty(&sk->sk_write_queue)); in kcm_write_msgs() 896 bool not_busy = skb_queue_empty(&sk->sk_write_queue); in kcm_sendmsg() 900 __skb_queue_tail(&sk->sk_write_queue, head); in kcm_sendmsg() 969 if (unlikely(skb_queue_len(&sk->sk_write_queue) == 0 && err == -EAGAIN)) in kcm_sendmsg() 982 if (skb_queue_empty_lockless(&sk->sk_write_queue)) in kcm_splice_eof() 1703 __skb_queue_purge(&sk->sk_write_queue); in kcm_release()
|
| H A D | kcmproc.c | 121 kcm->sk.sk_write_queue.qlen, in kcm_format_sock() 151 psock->sk->sk_write_queue.qlen, in kcm_format_psock()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | nested_acquire.c | 17 ptr = bpf_kfunc_nested_acquire_nonzero_offset_test(&sk->sk_write_queue); in BPF_PROG()
|
| /linux/net/core/ |
| H A D | stream.c | 212 WARN_ON_ONCE(!skb_queue_empty(&sk->sk_write_queue)); in sk_stream_kill_queues()
|
| /linux/include/net/ |
| H A D | tcp.h | 2193 return skb_peek_tail(&sk->sk_write_queue); in tcp_write_queue_tail() 2197 skb_queue_walk_from_safe(&(sk)->sk_write_queue, skb, tmp) 2201 return skb_peek(&sk->sk_write_queue); in tcp_send_head() 2207 return skb_queue_is_last(&sk->sk_write_queue, skb); in tcp_skb_is_last() 2236 __skb_queue_tail(&sk->sk_write_queue, skb); in tcp_add_write_queue_tail() 2239 if (sk->sk_write_queue.next == skb) in tcp_add_write_queue_tail() 2248 __skb_queue_before(&sk->sk_write_queue, skb, new); in tcp_insert_write_queue_before() 2254 __skb_unlink(skb, &sk->sk_write_queue); in tcp_unlink_write_queue()
|
| H A D | udp.h | 243 skb_queue_walk(&sk->sk_write_queue, skb) { in udp_csum_outgoing()
|
| /linux/net/tipc/ |
| H A D | socket.c | 564 __skb_queue_purge(&sk->sk_write_queue); in __tipc_shutdown() 1272 struct sk_buff_head *txq = &tsk->sk.sk_write_queue; in tipc_sk_push_backlog() 1504 if (unlikely(syn && !tipc_msg_skb_clone(&pkts, &sk->sk_write_queue))) { in __tipc_sendmsg() 1557 struct sk_buff_head *txq = &sk->sk_write_queue; in __tipc_sendstream() 1677 __skb_queue_purge(&sk->sk_write_queue); in tipc_sk_finish_conn() 2233 if (skb_queue_empty(&sk->sk_write_queue)) in tipc_sk_filter_connect() 2254 if (!skb_queue_empty(&sk->sk_write_queue)) in tipc_sk_filter_connect() 2862 tipc_msg_skb_clone(&sk->sk_write_queue, list); in tipc_sk_retry_connect() 3658 skb_queue_len(&sk->sk_write_queue)) || in tipc_sk_fill_sock_diag() 3991 i += tipc_list_dump(&sk->sk_write_queue, false, buf + i); in tipc_sk_dump()
|
| /linux/net/ipv6/ |
| H A D | raw.c | 526 skb = skb_peek(&sk->sk_write_queue); in rawv6_push_pending_frames() 542 if (skb_queue_len(&sk->sk_write_queue) == 1) { in rawv6_push_pending_frames() 551 skb_queue_walk(&sk->sk_write_queue, skb) { in rawv6_push_pending_frames()
|
| H A D | icmp.c | 282 skb = skb_peek(&sk->sk_write_queue); in icmpv6_push_pending_frames() 290 if (skb_queue_len(&sk->sk_write_queue) == 1) { in icmpv6_push_pending_frames() 300 skb_queue_walk(&sk->sk_write_queue, skb) { in icmpv6_push_pending_frames()
|
| H A D | ip6_output.c | 1839 if (skb_queue_empty(&sk->sk_write_queue)) { in ip6_append_data() 1857 return __ip6_append_data(sk, &sk->sk_write_queue, &inet->cork, in ip6_append_data() 2027 __ip6_flush_pending_frames(sk, &sk->sk_write_queue, in ip6_flush_pending_frames()
|
| /linux/net/ipv4/ |
| H A D | ip_output.c | 1370 if (skb_queue_empty(&sk->sk_write_queue)) { in ip_append_data() 1378 return __ip_append_data(sk, fl4, &sk->sk_write_queue, &inet->cork.base, in ip_append_data() 1548 __ip_flush_pending_frames(sk, &sk->sk_write_queue, &inet_sk(sk)->cork.base); in ip_flush_pending_frames() 1663 nskb = skb_peek(&sk->sk_write_queue); in ip_send_unicast_reply()
|
| H A D | icmp.c | 394 } else if ((skb = skb_peek(&sk->sk_write_queue)) != NULL) { in icmp_push_reply() 402 skb_queue_walk(&sk->sk_write_queue, skb1) { in icmp_push_reply()
|
| /linux/net/l2tp/ |
| H A D | l2tp_ip6.c | 502 skb = skb_peek(&sk->sk_write_queue); in l2tp_ip6_push_pending_frames() 654 ulen = len + (skb_queue_empty(&sk->sk_write_queue) ? transhdrlen : 0); in l2tp_ip6_sendmsg()
|
| /linux/drivers/net/ |
| H A D | tun.c | 230 struct sk_buff_head *queue = &tfile->sk.sk_write_queue; in tun_napi_receive() 567 skb_queue_purge(&tfile->sk.sk_write_queue); in tun_queue_purge() 1477 struct sk_buff_head *queue = &tfile->sk.sk_write_queue; in tun_rx_batched() 1932 struct sk_buff_head *queue = &tfile->sk.sk_write_queue; in tun_get_user() 2480 queue = &tfile->sk.sk_write_queue; in tun_xdp_one()
|