Searched refs:buf_states (Results 1 – 3 of 3) sorted by relevance
25 buf_state = &rx->dqo.buf_states[buffer_id]; in gve_alloc_buf_state()43 s16 buffer_id = buf_state - rx->dqo.buf_states; in gve_buf_state_is_allocated()51 s16 buffer_id = buf_state - rx->dqo.buf_states; in gve_free_buf_state()67 buf_state = &rx->dqo.buf_states[buffer_id]; in gve_dequeue_buf_state()83 s16 buffer_id = buf_state - rx->dqo.buf_states; in gve_enqueue_buf_state()93 rx->dqo.buf_states[tail].next = buffer_id; in gve_enqueue_buf_state()312 desc->buf_id = cpu_to_le16(buf_state - rx->dqo.buf_states); in gve_alloc_buffer()334 desc->buf_id = cpu_to_le16(buf_state - rx->dqo.buf_states); in gve_alloc_buffer()
55 if (rx->dqo.buf_states) { in gve_rx_init_ring_state_dqo()57 rx->dqo.buf_states[i].next = i + 1; in gve_rx_init_ring_state_dqo()58 rx->dqo.buf_states[rx->dqo.num_buf_states - 1].next = -1; in gve_rx_init_ring_state_dqo()96 if (rx->dqo.buf_states) { in gve_rx_reset_ring_dqo()98 struct gve_rx_buf_state_dqo *bs = &rx->dqo.buf_states[i]; in gve_rx_reset_ring_dqo()147 struct gve_rx_buf_state_dqo *bs = &rx->dqo.buf_states[i]; in gve_rx_free_ring_dqo()180 kvfree(rx->dqo.buf_states); in gve_rx_free_ring_dqo()181 rx->dqo.buf_states = NULL; in gve_rx_free_ring_dqo()250 rx->dqo.buf_states = kvcalloc_node(rx->dqo.num_buf_states, in gve_rx_alloc_ring_dqo()251 sizeof(rx->dqo.buf_states[0]), in gve_rx_alloc_ring_dqo()[all …]
268 struct gve_rx_buf_state_dqo *buf_states; member