Home
last modified time | relevance | path

Searched refs:sk_receive_queue (Results 1 – 25 of 90) sorted by relevance

1234

/linux/net/unix/
H A Daf_unix.c184 a = container_of(_a, struct sock, sk_receive_queue.lock.dep_map); in unix_recvq_lock_cmp_fn()
185 b = container_of(_b, struct sock, sk_receive_queue.lock.dep_map); in unix_recvq_lock_cmp_fn()
290 return skb_queue_len_lockless(&sk->sk_receive_queue) > sk->sk_max_ack_backlog; in unix_recvq_full_lockless()
618 if (!skb_queue_empty(&sk->sk_receive_queue)) { in unix_dgram_disconnected()
619 skb_queue_purge_reason(&sk->sk_receive_queue, in unix_dgram_disconnected()
639 skb_queue_purge_reason(&sk->sk_receive_queue, SKB_DROP_REASON_SOCKET_CLOSE); in unix_sock_destructor()
699 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in unix_release_sock()
703 skb = skb_peek_next(skb, &sk->sk_receive_queue); in unix_release_sock()
721 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in unix_release_sock()
881 spin_lock(&sk->sk_receive_queue.lock); in unix_count_nr_fds()
[all …]
H A Ddiag.c69 spin_lock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons()
72 sk->sk_receive_queue.qlen * sizeof(u32)); in sk_diag_dump_icons()
78 skb_queue_walk(&sk->sk_receive_queue, skb) in sk_diag_dump_icons()
81 spin_unlock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons()
87 spin_unlock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons()
96 rql.udiag_rqueue = skb_queue_len_lockless(&sk->sk_receive_queue); in sk_diag_show_rqlen()
H A Dunix_bpf.c11 ({ !skb_queue_empty(&__sk->sk_receive_queue) || \
69 if (!skb_queue_empty(&sk->sk_receive_queue) && in unix_bpf_recvmsg()
/linux/net/mptcp/
H A Dfastopen.c28 skb = skb_peek(&ssk->sk_receive_queue); in mptcp_fastopen_subflow_synack_set_params()
33 __skb_unlink(skb, &ssk->sk_receive_queue); in mptcp_fastopen_subflow_synack_set_params()
56 __skb_queue_tail(&sk->sk_receive_queue, skb); in mptcp_fastopen_subflow_synack_set_params()
H A Dsubflow.c1072 if (skb_queue_is_last(&ssk->sk_receive_queue, skb)) { in validate_data_csum()
1078 while ((skb = skb_peek(&ssk->sk_receive_queue))) in validate_data_csum()
1119 skb = skb_peek(&ssk->sk_receive_queue); in get_mapping_status()
1299 if (!skb_queue_empty(&ssk->sk_receive_queue)) in subflow_sched_work_if_closed()
1359 if (!skb_peek(&ssk->sk_receive_queue)) in subflow_check_data_avail()
1370 trace_subflow_check_data_avail(status, skb_peek(&ssk->sk_receive_queue)); in subflow_check_data_avail()
1378 skb = skb_peek(&ssk->sk_receive_queue); in subflow_check_data_avail()
1431 while ((skb = skb_peek(&ssk->sk_receive_queue))) in subflow_check_data_avail()
1439 skb = skb_peek(&ssk->sk_receive_queue); in subflow_check_data_avail()
/linux/net/phonet/
H A Ddatagram.c37 spin_lock_bh(&sk->sk_receive_queue.lock); in pn_ioctl()
38 skb = skb_peek(&sk->sk_receive_queue); in pn_ioctl()
40 spin_unlock_bh(&sk->sk_receive_queue.lock); in pn_ioctl()
61 skb_queue_purge(&sk->sk_receive_queue); in pn_destruct()
H A Dpep.c405 queue = &sk->sk_receive_queue; in pipe_do_rcv()
464 skb_queue_purge(&sk->sk_receive_queue); in pipe_destruct()
577 skb_queue_tail(&sk->sk_receive_queue, skb); in pipe_handler_do_rcv()
684 skb_queue_head(&sk->sk_receive_queue, skb); in pep_do_rcv()
941 q = &sk->sk_receive_queue; in pep_first_packet_length()
1257 struct sk_buff *skb = skb_dequeue(&sk->sk_receive_queue); in pep_read()
/linux/net/bluetooth/
H A Daf_bluetooth.c393 if (!skb_queue_empty(&sk->sk_receive_queue)) in bt_sock_data_wait()
436 skb = skb_dequeue(&sk->sk_receive_queue); in bt_sock_stream_recvmsg()
462 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg()
501 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg()
508 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg()
556 if (!skb_queue_empty_lockless(&sk->sk_receive_queue)) in bt_sock_poll()
680 spin_lock(&sk->sk_receive_queue.lock); in bt_sock_ioctl()
681 skb = skb_peek(&sk->sk_receive_queue); in bt_sock_ioctl()
683 spin_unlock(&sk->sk_receive_queue.lock); in bt_sock_ioctl()
/linux/net/atm/
H A Dsignaling.c63 skb_queue_tail(&sk_atm(sigd)->sk_receive_queue, skb); in sigd_put_skb()
159 skb_queue_tail(&sk->sk_receive_queue, skb); in sigd_send()
256 if (skb_peek(&sk_atm(vcc)->sk_receive_queue)) in sigd_close()
258 skb_queue_purge(&sk_atm(vcc)->sk_receive_queue); in sigd_close()
H A Dioctl.c82 spin_lock_irq(&sk->sk_receive_queue.lock); in do_vcc_ioctl()
83 skb = skb_peek(&sk->sk_receive_queue); in do_vcc_ioctl()
85 spin_unlock_irq(&sk->sk_receive_queue.lock); in do_vcc_ioctl()
H A Draw.c28 skb_queue_tail(&sk->sk_receive_queue, skb); in atm_push_raw()
/linux/net/sctp/
H A Dulpqueue.c135 &sk->sk_receive_queue); in sctp_clear_pd()
152 __skb_queue_tail(&sk->sk_receive_queue, in sctp_clear_pd()
213 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event()
227 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event()
236 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event()
251 if (queue == &sk->sk_receive_queue && !sp->data_ready_signalled) { in sctp_ulpq_tail_event()
1079 if (skb_queue_empty(&asoc->base.sk->sk_receive_queue)) { in sctp_ulpq_renege()
1119 __skb_queue_tail(&sk->sk_receive_queue, sctp_event2skb(ev)); in sctp_ulpq_abort_pd()
/linux/net/caif/
H A Dcaif_socket.c124 struct sk_buff_head *list = &sk->sk_receive_queue; in caif_queue_rcv_skb()
313 if (!skb_queue_empty(&sk->sk_receive_queue) || in caif_stream_data_wait()
377 skb = skb_dequeue(&sk->sk_receive_queue); in caif_stream_recvmsg()
422 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg()
436 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg()
446 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg()
906 spin_lock_bh(&sk->sk_receive_queue.lock); in caif_release()
908 spin_unlock_bh(&sk->sk_receive_queue.lock); in caif_release()
949 if (!skb_queue_empty_lockless(&sk->sk_receive_queue) || in caif_poll()
/linux/net/kcm/
H A Dkcmproc.c119 kcm->sk.sk_receive_queue.qlen, in kcm_format_sock()
149 psock->sk->sk_receive_queue.qlen, in kcm_format_psock()
167 if (psock->sk->sk_receive_queue.qlen) { in kcm_format_psock()
/linux/net/tipc/
H A Dsocket.c279 kfree_skb(__skb_dequeue(&sk->sk_receive_queue)); in tsk_advance_rx_queue()
310 while ((skb = __skb_dequeue(&sk->sk_receive_queue))) in tsk_rej_rx_queue()
567 skb = skb_peek(&sk->sk_receive_queue); in __tipc_shutdown()
569 __skb_unlink(skb, &sk->sk_receive_queue); in __tipc_shutdown()
585 skb = __skb_dequeue(&sk->sk_receive_queue); in __tipc_shutdown()
587 __skb_queue_purge(&sk->sk_receive_queue); in __tipc_shutdown()
603 __skb_queue_purge(&sk->sk_receive_queue); in __tipc_shutdown()
811 if (!skb_queue_empty_lockless(&sk->sk_receive_queue)) in tipc_poll()
819 if (skb_queue_empty_lockless(&sk->sk_receive_queue)) in tipc_poll()
1839 if (timeo && skb_queue_empty(&sk->sk_receive_queue)) { in tipc_wait_for_rcvmsg()
[all …]
/linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/
H A Dchtls_cm.h174 __skb_unlink(skb, &sk->sk_receive_queue); in chtls_free_skb()
181 __skb_unlink(skb, &sk->sk_receive_queue); in chtls_kfree_skb()
H A Dchtls_io.c1371 skb = skb_peek(&sk->sk_receive_queue); in chtls_pt_recvmsg()
1432 __skb_unlink(skb, &sk->sk_receive_queue); in chtls_pt_recvmsg()
1512 next_skb = skb_peek(&sk->sk_receive_queue); in chtls_pt_recvmsg()
1557 skb_queue_walk(&sk->sk_receive_queue, skb) { in peekmsg()
1679 skb_queue_empty_lockless(&sk->sk_receive_queue) && in chtls_recvmsg()
1709 skb = skb_peek(&sk->sk_receive_queue); in chtls_recvmsg()
1825 !skb_peek(&sk->sk_receive_queue)) in chtls_recvmsg()
/linux/drivers/xen/
H A Dpvcalls-back.c117 spin_lock_irqsave(&map->sock->sk->sk_receive_queue.lock, flags); in pvcalls_conn_back_read()
118 if (skb_queue_empty(&map->sock->sk->sk_receive_queue)) { in pvcalls_conn_back_read()
120 spin_unlock_irqrestore(&map->sock->sk->sk_receive_queue.lock, in pvcalls_conn_back_read()
124 spin_unlock_irqrestore(&map->sock->sk->sk_receive_queue.lock, flags); in pvcalls_conn_back_read()
149 spin_lock_irqsave(&map->sock->sk->sk_receive_queue.lock, flags); in pvcalls_conn_back_read()
150 if (ret > 0 && !skb_queue_empty(&map->sock->sk->sk_receive_queue)) in pvcalls_conn_back_read()
152 spin_unlock_irqrestore(&map->sock->sk->sk_receive_queue.lock, flags); in pvcalls_conn_back_read()
/linux/net/nfc/
H A Dllcp_sock.c572 if (!skb_queue_empty_lockless(&sk->sk_receive_queue)) in llcp_sock_poll()
841 skb_queue_empty(&sk->sk_receive_queue)) { in llcp_sock_recvmsg()
868 skb_queue_head(&sk->sk_receive_queue, skb); in llcp_sock_recvmsg()
899 skb_queue_head(&sk->sk_receive_queue, skb); in llcp_sock_recvmsg()
964 skb_queue_purge(&sk->sk_receive_queue); in llcp_sock_destruct()
/linux/net/rds/
H A Dtcp_recv.c258 skb_queue_len(&tc->t_sock->sk->sk_receive_queue)); in rds_tcp_data_recv()
281 if (skb_queue_empty_lockless(&sock->sk->sk_receive_queue) && in rds_tcp_read_sock()
/linux/net/ipv4/
H A Dtcp_bpf.c197 !skb_queue_empty_lockless(&sk->sk_receive_queue), &wait); in tcp_msg_wait_data()
254 if (unlikely(!skb_queue_empty(&sk->sk_receive_queue))) { in tcp_bpf_recvmsg_parser()
260 if (unlikely(!skb_queue_empty(&sk->sk_receive_queue))) { in tcp_bpf_recvmsg_parser()
369 if (!skb_queue_empty(&sk->sk_receive_queue) && in tcp_bpf_recvmsg()
/linux/net/packet/
H A Daf_packet.c689 spin_lock(&po->sk.sk_receive_queue.lock); in prb_retire_rx_blk_timer_expired()
732 spin_unlock(&po->sk.sk_receive_queue.lock); in prb_retire_rx_blk_timer_expired()
2203 spin_lock(&sk->sk_receive_queue.lock); in packet_rcv()
2207 __skb_queue_tail(&sk->sk_receive_queue, skb); in packet_rcv()
2208 spin_unlock(&sk->sk_receive_queue.lock); in packet_rcv()
2348 spin_lock(&sk->sk_receive_queue.lock); in tpacket_rcv()
2386 __skb_queue_tail(&sk->sk_receive_queue, copy_skb); in tpacket_rcv()
2388 spin_unlock(&sk->sk_receive_queue.lock); in tpacket_rcv()
2481 spin_lock(&sk->sk_receive_queue.lock); in tpacket_rcv()
2484 spin_unlock(&sk->sk_receive_queue.lock); in tpacket_rcv()
[all …]
/linux/net/l2tp/
H A Dl2tp_ip.c594 spin_lock_bh(&sk->sk_receive_queue.lock); in l2tp_ioctl()
595 skb = skb_peek(&sk->sk_receive_queue); in l2tp_ioctl()
597 spin_unlock_bh(&sk->sk_receive_queue.lock); in l2tp_ioctl()
/linux/net/llc/
H A Daf_llc.c714 if (skb_queue_empty(&sk->sk_receive_queue)) { in llc_ui_accept()
721 skb = skb_dequeue(&sk->sk_receive_queue); in llc_ui_accept()
807 skb = skb_peek(&sk->sk_receive_queue); in llc_ui_recvmsg()
895 skb_unlink(skb, &sk->sk_receive_queue); in llc_ui_recvmsg()
913 skb_unlink(skb, &sk->sk_receive_queue); in llc_ui_recvmsg()
/linux/net/rose/
H A Daf_rose.c370 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in rose_destroy_socket()
967 skb = skb_dequeue(&sk->sk_receive_queue); in rose_accept()
1106 skb_queue_head(&sk->sk_receive_queue, skb); in rose_rx_call_request()
1363 spin_lock_irq(&sk->sk_receive_queue.lock); in rose_ioctl()
1364 if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) in rose_ioctl()
1366 spin_unlock_irq(&sk->sk_receive_queue.lock); in rose_ioctl()

1234