Lines Matching +full:send +full:- +full:empty
1 // SPDX-License-Identifier: GPL-2.0-only
5 * Copyright (c) 2003-2016, Intel Corporation.
12 * ishtp_cl_alloc_rx_ring() - Allocate RX ring buffers
17 * Return: 0 on success else -ENOMEM
21 size_t len = cl->device->fw_client->props.max_msg_length; in ishtp_cl_alloc_rx_ring()
27 for (j = 0; j < cl->rx_ring_size; ++j) { in ishtp_cl_alloc_rx_ring()
30 ret = -ENOMEM; in ishtp_cl_alloc_rx_ring()
36 spin_lock_irqsave(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring()
37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring()
38 spin_unlock_irqrestore(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring()
44 dev_err(&cl->device->dev, "error in allocating Rx buffers\n"); in ishtp_cl_alloc_rx_ring()
50 * ishtp_cl_alloc_tx_ring() - Allocate TX ring buffers
55 * Return: 0 on success else -ENOMEM
59 size_t len = cl->device->fw_client->props.max_msg_length; in ishtp_cl_alloc_tx_ring()
63 cl->tx_ring_free_size = 0; in ishtp_cl_alloc_tx_ring()
66 for (j = 0; j < cl->tx_ring_size; ++j) { in ishtp_cl_alloc_tx_ring()
73 tx_buf->send_buf.data = kmalloc(len, GFP_KERNEL); in ishtp_cl_alloc_tx_ring()
74 if (!tx_buf->send_buf.data) { in ishtp_cl_alloc_tx_ring()
79 spin_lock_irqsave(&cl->tx_free_list_spinlock, flags); in ishtp_cl_alloc_tx_ring()
80 list_add_tail(&tx_buf->list, &cl->tx_free_list.list); in ishtp_cl_alloc_tx_ring()
81 ++cl->tx_ring_free_size; in ishtp_cl_alloc_tx_ring()
82 spin_unlock_irqrestore(&cl->tx_free_list_spinlock, flags); in ishtp_cl_alloc_tx_ring()
86 dev_err(&cl->device->dev, "error in allocating Tx pool\n"); in ishtp_cl_alloc_tx_ring()
88 return -ENOMEM; in ishtp_cl_alloc_tx_ring()
92 * ishtp_cl_free_rx_ring() - Free RX ring buffers
102 /* release allocated memory - pass over free_rb_list */ in ishtp_cl_free_rx_ring()
103 spin_lock_irqsave(&cl->free_list_spinlock, flags); in ishtp_cl_free_rx_ring()
104 while (!list_empty(&cl->free_rb_list.list)) { in ishtp_cl_free_rx_ring()
105 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, in ishtp_cl_free_rx_ring()
107 list_del(&rb->list); in ishtp_cl_free_rx_ring()
108 kfree(rb->buffer.data); in ishtp_cl_free_rx_ring()
111 spin_unlock_irqrestore(&cl->free_list_spinlock, flags); in ishtp_cl_free_rx_ring()
112 /* release allocated memory - pass over in_process_list */ in ishtp_cl_free_rx_ring()
113 spin_lock_irqsave(&cl->in_process_spinlock, flags); in ishtp_cl_free_rx_ring()
114 while (!list_empty(&cl->in_process_list.list)) { in ishtp_cl_free_rx_ring()
115 rb = list_entry(cl->in_process_list.list.next, in ishtp_cl_free_rx_ring()
117 list_del(&rb->list); in ishtp_cl_free_rx_ring()
118 kfree(rb->buffer.data); in ishtp_cl_free_rx_ring()
121 spin_unlock_irqrestore(&cl->in_process_spinlock, flags); in ishtp_cl_free_rx_ring()
125 * ishtp_cl_free_tx_ring() - Free TX ring buffers
135 spin_lock_irqsave(&cl->tx_free_list_spinlock, flags); in ishtp_cl_free_tx_ring()
136 /* release allocated memory - pass over tx_free_list */ in ishtp_cl_free_tx_ring()
137 while (!list_empty(&cl->tx_free_list.list)) { in ishtp_cl_free_tx_ring()
138 tx_buf = list_entry(cl->tx_free_list.list.next, in ishtp_cl_free_tx_ring()
140 list_del(&tx_buf->list); in ishtp_cl_free_tx_ring()
141 --cl->tx_ring_free_size; in ishtp_cl_free_tx_ring()
142 kfree(tx_buf->send_buf.data); in ishtp_cl_free_tx_ring()
145 spin_unlock_irqrestore(&cl->tx_free_list_spinlock, flags); in ishtp_cl_free_tx_ring()
147 spin_lock_irqsave(&cl->tx_list_spinlock, flags); in ishtp_cl_free_tx_ring()
148 /* release allocated memory - pass over tx_list */ in ishtp_cl_free_tx_ring()
149 while (!list_empty(&cl->tx_list.list)) { in ishtp_cl_free_tx_ring()
150 tx_buf = list_entry(cl->tx_list.list.next, in ishtp_cl_free_tx_ring()
152 list_del(&tx_buf->list); in ishtp_cl_free_tx_ring()
153 kfree(tx_buf->send_buf.data); in ishtp_cl_free_tx_ring()
156 spin_unlock_irqrestore(&cl->tx_list_spinlock, flags); in ishtp_cl_free_tx_ring()
160 * ishtp_io_rb_free() - Free IO request block
170 kfree(rb->buffer.data); in ishtp_io_rb_free()
175 * ishtp_io_rb_init() - Allocate and init IO request block
190 INIT_LIST_HEAD(&rb->list); in ishtp_io_rb_init()
191 rb->cl = cl; in ishtp_io_rb_init()
192 rb->buf_idx = 0; in ishtp_io_rb_init()
197 * ishtp_io_rb_alloc_buf() - Allocate and init response buffer
203 * Return: 0 on success else -ENOMEM
208 return -EINVAL; in ishtp_io_rb_alloc_buf()
213 rb->buffer.data = kmalloc(length, GFP_KERNEL); in ishtp_io_rb_alloc_buf()
214 if (!rb->buffer.data) in ishtp_io_rb_alloc_buf()
215 return -ENOMEM; in ishtp_io_rb_alloc_buf()
217 rb->buffer.size = length; in ishtp_io_rb_alloc_buf()
222 * ishtp_cl_io_rb_recycle() - Recycle IO request blocks
225 * Re-append rb to its client's free list and send flow control if needed
227 * Return: 0 on success else -EFAULT
235 if (!rb || !rb->cl) in ishtp_cl_io_rb_recycle()
236 return -EFAULT; in ishtp_cl_io_rb_recycle()
238 cl = rb->cl; in ishtp_cl_io_rb_recycle()
239 spin_lock_irqsave(&cl->free_list_spinlock, flags); in ishtp_cl_io_rb_recycle()
240 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_io_rb_recycle()
241 spin_unlock_irqrestore(&cl->free_list_spinlock, flags); in ishtp_cl_io_rb_recycle()
244 * If we returned the first buffer to empty 'free' list, in ishtp_cl_io_rb_recycle()
245 * send flow control in ishtp_cl_io_rb_recycle()
247 if (!cl->out_flow_ctrl_creds) in ishtp_cl_io_rb_recycle()
255 * ishtp_cl_tx_empty() -test whether client device tx buffer is empty
258 * Look client device tx buffer list, and check whether this list is empty
260 * Return: true if client tx buffer list is empty else false
267 spin_lock_irqsave(&cl->tx_list_spinlock, tx_flags); in ishtp_cl_tx_empty()
268 tx_list_empty = list_empty(&cl->tx_list.list); in ishtp_cl_tx_empty()
269 spin_unlock_irqrestore(&cl->tx_list_spinlock, tx_flags); in ishtp_cl_tx_empty()
276 * ishtp_cl_rx_get_rb() -Get a rb from client device rx buffer list
279 * Check client device in-processing buffer list and get a rb from it.
281 * Return: rb pointer if buffer list isn't empty else NULL
288 spin_lock_irqsave(&cl->in_process_spinlock, rx_flags); in ishtp_cl_rx_get_rb()
289 rb = list_first_entry_or_null(&cl->in_process_list.list, in ishtp_cl_rx_get_rb()
292 list_del_init(&rb->list); in ishtp_cl_rx_get_rb()
293 spin_unlock_irqrestore(&cl->in_process_spinlock, rx_flags); in ishtp_cl_rx_get_rb()