Home
last modified time | relevance | path

Searched refs:queue_id (Results 1 – 25 of 244) sorted by relevance

12345678910

/linux/drivers/net/ethernet/engleder/
H A Dtsnep_xdp.c22 struct xsk_buff_pool *pool, u16 queue_id) in tsnep_xdp_enable_pool() argument
27 if (queue_id >= adapter->num_rx_queues || in tsnep_xdp_enable_pool()
28 queue_id >= adapter->num_tx_queues) in tsnep_xdp_enable_pool()
31 queue = &adapter->queue[queue_id]; in tsnep_xdp_enable_pool()
32 if (queue->rx->queue_index != queue_id || in tsnep_xdp_enable_pool()
33 queue->tx->queue_index != queue_id) { in tsnep_xdp_enable_pool()
58 static int tsnep_xdp_disable_pool(struct tsnep_adapter *adapter, u16 queue_id) in tsnep_xdp_disable_pool() argument
63 if (queue_id >= adapter->num_rx_queues || in tsnep_xdp_disable_pool()
64 queue_id >= adapter->num_tx_queues) in tsnep_xdp_disable_pool()
67 pool = xsk_get_pool_from_qid(adapter->netdev, queue_id); in tsnep_xdp_disable_pool()
[all …]
/linux/arch/mips/include/asm/octeon/
H A Dcvmx-cmd-queue.h170 cvmx_cmd_queue_result_t cvmx_cmd_queue_initialize(cvmx_cmd_queue_id_t queue_id,
183 cvmx_cmd_queue_result_t cvmx_cmd_queue_shutdown(cvmx_cmd_queue_id_t queue_id);
193 int cvmx_cmd_queue_length(cvmx_cmd_queue_id_t queue_id);
205 void *cvmx_cmd_queue_buffer(cvmx_cmd_queue_id_t queue_id);
214 static inline int __cvmx_cmd_queue_get_index(cvmx_cmd_queue_id_t queue_id) in __cvmx_cmd_queue_get_index() argument
223 int unit = queue_id >> 16; in __cvmx_cmd_queue_get_index()
224 int q = (queue_id >> 4) & 0xf; in __cvmx_cmd_queue_get_index()
225 int core = queue_id & 0xf; in __cvmx_cmd_queue_get_index()
236 static inline void __cvmx_cmd_queue_lock(cvmx_cmd_queue_id_t queue_id, in __cvmx_cmd_queue_lock() argument
278 … "=" GCC_OFF_SMALL_ASM()(__cvmx_cmd_queue_state_ptr->ticket[__cvmx_cmd_queue_get_index(queue_id)]), in __cvmx_cmd_queue_lock()
[all …]
/linux/drivers/net/ethernet/intel/igc/
H A Digc_xdp.c57 struct xsk_buff_pool *pool, u16 queue_id) in igc_xdp_enable_pool() argument
67 if (queue_id >= adapter->num_rx_queues || in igc_xdp_enable_pool()
68 queue_id >= adapter->num_tx_queues) in igc_xdp_enable_pool()
89 rx_ring = adapter->rx_ring[queue_id]; in igc_xdp_enable_pool()
90 tx_ring = adapter->tx_ring[queue_id]; in igc_xdp_enable_pool()
108 err = igc_xsk_wakeup(ndev, queue_id, XDP_WAKEUP_RX); in igc_xdp_enable_pool()
118 static int igc_xdp_disable_pool(struct igc_adapter *adapter, u16 queue_id) in igc_xdp_disable_pool() argument
125 if (queue_id >= adapter->num_rx_queues || in igc_xdp_disable_pool()
126 queue_id >= adapter->num_tx_queues) in igc_xdp_disable_pool()
129 pool = xsk_get_pool_from_qid(adapter->netdev, queue_id); in igc_xdp_disable_pool()
[all …]
/linux/arch/mips/cavium-octeon/executive/
H A Dcvmx-cmd-queue.c108 cvmx_cmd_queue_result_t cvmx_cmd_queue_initialize(cvmx_cmd_queue_id_t queue_id, in cvmx_cmd_queue_initialize() argument
117 qstate = __cvmx_cmd_queue_get_state(queue_id); in cvmx_cmd_queue_initialize()
188 ticket[__cvmx_cmd_queue_get_index(queue_id)] = 0; in cvmx_cmd_queue_initialize()
203 cvmx_cmd_queue_result_t cvmx_cmd_queue_shutdown(cvmx_cmd_queue_id_t queue_id) in cvmx_cmd_queue_shutdown() argument
205 __cvmx_cmd_queue_state_t *qptr = __cvmx_cmd_queue_get_state(queue_id); in cvmx_cmd_queue_shutdown()
212 if (cvmx_cmd_queue_length(queue_id) > 0) { in cvmx_cmd_queue_shutdown()
218 __cvmx_cmd_queue_lock(queue_id, qptr); in cvmx_cmd_queue_shutdown()
238 int cvmx_cmd_queue_length(cvmx_cmd_queue_id_t queue_id) in cvmx_cmd_queue_length() argument
241 if (__cvmx_cmd_queue_get_state(queue_id) == NULL) in cvmx_cmd_queue_length()
249 switch ((cvmx_cmd_queue_id_t) (queue_id & 0xff0000)) { in cvmx_cmd_queue_length()
[all …]
/linux/drivers/scsi/lpfc/
H A Dlpfc_debugfs.h416 q->queue_id, q->type, q->subtype, in lpfc_debug_dump_q()
456 qtypestr, wqidx, wq->queue_id); in lpfc_debug_dump_wq()
459 qtypestr, wq->queue_id); in lpfc_debug_dump_wq()
504 if (cq->assoc_qid == eq->queue_id) in lpfc_debug_dump_cq()
516 qtypestr, wqidx, wq->queue_id, wqidx, cq->queue_id, in lpfc_debug_dump_cq()
517 eqidx, eq->queue_id); in lpfc_debug_dump_cq()
521 qtypestr, wq->queue_id, cq->queue_id, in lpfc_debug_dump_cq()
522 eqidx, eq->queue_id); in lpfc_debug_dump_cq()
542 pr_err("EQ[Idx:%d|Qid:%d]\n", qidx, qp->queue_id); in lpfc_debug_dump_hba_eq()
557 phba->sli4_hba.dat_rq->queue_id); in lpfc_debug_dump_dat_rq()
[all …]
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_amdkfd_gfx_v10_3.c58 uint32_t queue_id) in acquire_queue() argument
63 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue()
67 uint32_t pipe_id, uint32_t queue_id) in get_queue_mask() argument
70 queue_id; in get_queue_mask()
131 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument
161 + queue_id * (mmSDMA0_RLC1_RB_CNTL - mmSDMA0_RLC0_RB_CNTL); in get_sdma_rlc_reg_offset()
164 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset()
180 uint32_t pipe_id, uint32_t queue_id, in hqd_load_v10_3() argument
190 pr_debug("Load hqd of pipe %d queue %d\n", pipe_id, queue_id); in hqd_load_v10_3()
191 acquire_queue(adev, pipe_id, queue_id); in hqd_load_v10_3()
[all …]
H A Damdgpu_amdkfd_gfx_v11.c56 uint32_t queue_id) in acquire_queue() argument
61 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue()
65 uint32_t pipe_id, uint32_t queue_id) in get_queue_mask() argument
68 queue_id; in get_queue_mask()
127 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument
146 + queue_id * (regSDMA0_QUEUE1_RB_CNTL - regSDMA0_QUEUE0_RB_CNTL); in get_sdma_rlc_reg_offset()
149 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset()
165 uint32_t queue_id, uint32_t __user *wptr, in hqd_load_v11() argument
175 pr_debug("Load hqd of pipe %d queue %d\n", pipe_id, queue_id); in hqd_load_v11()
176 acquire_queue(adev, pipe_id, queue_id); in hqd_load_v11()
[all …]
H A Damdgpu_amdkfd_gfx_v9.h32 uint32_t queue_id, uint32_t __user *wptr,
36 uint32_t pipe_id, uint32_t queue_id,
39 uint32_t pipe_id, uint32_t queue_id,
43 uint32_t queue_id, uint32_t inst);
47 uint32_t queue_id, uint32_t inst);
62 uint32_t queue_id, uint32_t inst);
64 uint32_t pipe_id, uint32_t queue_id);
108 uint32_t queue_id,
112 uint32_t queue_id,
H A Damdgpu_amdkfd_gfx_v10.c58 uint32_t queue_id) in acquire_queue() argument
63 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue()
67 uint32_t pipe_id, uint32_t queue_id) in get_queue_mask() argument
70 queue_id; in get_queue_mask()
162 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument
178 + queue_id * (mmSDMA0_RLC1_RB_CNTL - mmSDMA0_RLC0_RB_CNTL); in get_sdma_rlc_reg_offset()
181 queue_id, retval); in get_sdma_rlc_reg_offset()
209 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_load() argument
219 pr_debug("Load hqd of pipe %d queue %d\n", pipe_id, queue_id); in kgd_hqd_load()
220 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load()
[all …]
H A Damdgpu_amdkfd_gfx_v9.c64 uint32_t queue_id, uint32_t inst) in kgd_gfx_v9_acquire_queue() argument
69 kgd_gfx_v9_lock_srbm(adev, mec, pipe, queue_id, 0, inst); in kgd_gfx_v9_acquire_queue()
73 uint32_t pipe_id, uint32_t queue_id) in kgd_gfx_v9_get_queue_mask() argument
76 queue_id; in kgd_gfx_v9_get_queue_mask()
182 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument
204 + queue_id * (mmSDMA0_RLC1_RB_CNTL - mmSDMA0_RLC0_RB_CNTL); in get_sdma_rlc_reg_offset()
207 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset()
223 uint32_t pipe_id, uint32_t queue_id, in kgd_gfx_v9_hqd_load() argument
234 kgd_gfx_v9_acquire_queue(adev, pipe_id, queue_id, inst); in kgd_gfx_v9_hqd_load()
286 (uint32_t)kgd_gfx_v9_get_queue_mask(adev, pipe_id, queue_id)); in kgd_gfx_v9_hqd_load()
[all …]
H A Damdgpu_amdkfd_gfx_v8.c58 uint32_t queue_id) in acquire_queue() argument
63 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue()
155 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_load() argument
166 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load()
176 mec, pipe, queue_id); in kgd_hqd_load()
179 ((mec << 5) | (pipe << 3) | queue_id | 0x80)); in kgd_hqd_load()
216 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load()
229 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_dump() argument
245 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_dump()
318 uint32_t engine_id, uint32_t queue_id, in kgd_hqd_sdma_dump() argument
[all …]
H A Damdgpu_amdkfd_gfx_v7.c64 uint32_t queue_id) in acquire_queue() argument
69 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue()
160 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_load() argument
171 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load()
192 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load()
205 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_dump() argument
221 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_dump()
295 uint32_t engine_id, uint32_t queue_id, in kgd_hqd_sdma_dump() argument
299 queue_id * KFD_CIK_SDMA_QUEUE_OFFSET; in kgd_hqd_sdma_dump()
322 uint32_t queue_id, uint32_t inst) in kgd_hqd_is_occupied() argument
[all …]
H A Damdgpu_amdkfd_gfx_v12.c44 uint32_t queue_id) in acquire_queue() argument
49 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue()
78 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument
97 + queue_id * (regSDMA0_QUEUE1_RB_CNTL - regSDMA0_QUEUE0_RB_CNTL); in get_sdma_rlc_reg_offset()
100 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset()
106 uint32_t pipe_id, uint32_t queue_id, in hqd_dump_v12() argument
122 acquire_queue(adev, pipe_id, queue_id); in hqd_dump_v12()
137 uint32_t engine_id, uint32_t queue_id, in hqd_sdma_dump_v12() argument
141 engine_id, queue_id); in hqd_sdma_dump_v12()
H A Damdgpu_userq.c909 args->out.queue_id = qid; in amdgpu_userq_create()
1024 queue = __xa_erase(&fpriv->userq_mgr.userq_xa, args->in.queue_id); in amdgpu_userq_ioctl()
1045 unsigned long queue_id; in amdgpu_userq_restore_all() local
1049 xa_for_each(&uq_mgr->userq_xa, queue_id, queue) { in amdgpu_userq_restore_all()
1050 queue = amdgpu_userq_get(uq_mgr, queue_id); in amdgpu_userq_restore_all()
1294 unsigned long queue_id; in amdgpu_userq_evict_all() local
1299 xa_for_each(&uq_mgr->userq_xa, queue_id, queue) { in amdgpu_userq_evict_all()
1300 queue = amdgpu_userq_get(uq_mgr, queue_id); in amdgpu_userq_evict_all()
1335 unsigned long queue_id; in amdgpu_userq_wait_for_signal() local
1338 xa_for_each(&uq_mgr->userq_xa, queue_id, queue) { in amdgpu_userq_wait_for_signal()
[all …]
/linux/drivers/net/wireless/st/cw1200/
H A Dqueue.c32 queue->queue_id); in __cw1200_queue_lock()
33 ieee80211_stop_queue(stats->priv->hw, queue->queue_id); in __cw1200_queue_lock()
43 queue->queue_id); in __cw1200_queue_unlock()
44 ieee80211_wake_queue(stats->priv->hw, queue->queue_id); in __cw1200_queue_unlock()
49 u8 *queue_id, u8 *item_generation, in cw1200_queue_parse_id() argument
54 *queue_id = (packet_id >> 16) & 0xFF; in cw1200_queue_parse_id()
58 static inline u32 cw1200_queue_mk_packet_id(u8 queue_generation, u8 queue_id, in cw1200_queue_mk_packet_id() argument
63 ((u32)queue_id << 16) | in cw1200_queue_mk_packet_id()
165 u8 queue_id, in cw1200_queue_init() argument
174 queue->queue_id = queue_id; in cw1200_queue_init()
[all …]
/linux/drivers/mailbox/
H A Dti-msgmgr.c47 u8 queue_id; member
108 u8 queue_id; member
473 "rx_%03d", d->is_sproxy ? qinst->proxy_id : qinst->queue_id); in ti_msgmgr_queue_rx_irq_req()
486 qinst->queue_id, qinst->proxy_id, in ti_msgmgr_queue_rx_irq_req()
614 if (req_qid == qinst->queue_id && req_pid == qinst->proxy_id) in ti_msgmgr_of_xlate()
648 qinst->queue_id = qd->queue_id; in ti_msgmgr_queue_setup()
650 if (qinst->queue_id > d->queue_count) { in ti_msgmgr_queue_setup()
652 idx, qinst->queue_id, d->queue_count); in ti_msgmgr_queue_setup()
674 Q_DATA_OFFSET(qinst->proxy_id, qinst->queue_id, in ti_msgmgr_queue_setup()
677 Q_DATA_OFFSET(qinst->proxy_id, qinst->queue_id, in ti_msgmgr_queue_setup()
[all …]
/linux/drivers/staging/media/atomisp/pci/runtime/bufq/src/
H A Dbufq.c186 int queue_id; in unmap_buffer_type_to_queue_id() local
193 queue_id = buffer_type_to_queue_id_map[thread_id][buf_type]; in unmap_buffer_type_to_queue_id()
195 queue_availability[thread_id][queue_id] = true; in unmap_buffer_type_to_queue_id()
319 int queue_id, in ia_css_bufq_enqueue_buffer() argument
325 IA_CSS_ENTER_PRIVATE("queue_id=%d", queue_id); in ia_css_bufq_enqueue_buffer()
327 (queue_id == SH_CSS_INVALID_QUEUE_ID)) in ia_css_bufq_enqueue_buffer()
332 queue_id, in ia_css_bufq_enqueue_buffer()
346 int queue_id, in ia_css_bufq_dequeue_buffer() argument
352 IA_CSS_ENTER_PRIVATE("queue_id=%d", queue_id); in ia_css_bufq_dequeue_buffer()
354 (queue_id <= SH_CSS_INVALID_QUEUE_ID) || in ia_css_bufq_dequeue_buffer()
[all …]
/linux/drivers/gpu/drm/amd/include/
H A Dkgd_kfd_interface.h234 uint32_t queue_id, uint32_t __user *wptr,
239 uint32_t pipe_id, uint32_t queue_id,
246 uint32_t pipe_id, uint32_t queue_id,
250 uint32_t engine_id, uint32_t queue_id,
255 uint32_t queue_id, uint32_t inst);
260 uint32_t queue_id, uint32_t inst);
329 uint32_t pipe_id, uint32_t queue_id,
332 uint32_t pipe_id, uint32_t queue_id,
/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_mqd_manager.c211 uint32_t pipe_id, uint32_t queue_id, in kfd_hiq_load_mqd_kiq() argument
215 queue_id, p->doorbell_off, 0); in kfd_hiq_load_mqd_kiq()
220 uint32_t pipe_id, uint32_t queue_id) in kfd_destroy_mqd_cp() argument
223 pipe_id, queue_id, 0); in kfd_destroy_mqd_cp()
239 uint32_t queue_id) in kfd_is_occupied_cp() argument
242 pipe_id, queue_id, 0); in kfd_is_occupied_cp()
246 uint32_t pipe_id, uint32_t queue_id, in kfd_load_mqd_sdma() argument
261 uint32_t queue_id) in kfd_destroy_mqd_sdma() argument
268 uint32_t queue_id) in kfd_is_occupied_sdma() argument
/linux/drivers/net/wireless/realtek/rtlwifi/rtl8192cu/
H A Dphy.c381 u8 i, queue_id; in _rtl92cu_phy_set_rf_power_state() local
416 for (queue_id = 0, i = 0; in _rtl92cu_phy_set_rf_power_state()
417 queue_id < RTL_PCI_MAX_TX_QUEUE_COUNT;) { in _rtl92cu_phy_set_rf_power_state()
418 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92cu_phy_set_rf_power_state()
420 queue_id == BEACON_QUEUE) { in _rtl92cu_phy_set_rf_power_state()
421 queue_id++; in _rtl92cu_phy_set_rf_power_state()
427 queue_id, in _rtl92cu_phy_set_rf_power_state()
436 queue_id, in _rtl92cu_phy_set_rf_power_state()
459 for (queue_id = 0, i = 0; in _rtl92cu_phy_set_rf_power_state()
460 queue_id < RTL_PCI_MAX_TX_QUEUE_COUNT;) { in _rtl92cu_phy_set_rf_power_state()
[all …]
/linux/drivers/net/ethernet/netronome/nfp/
H A Dnfp_net_xsk.c111 struct xsk_buff_pool *pool, u16 queue_id) in nfp_net_xsk_setup_pool() argument
143 prev_pool = dp->xsk_pools[queue_id]; in nfp_net_xsk_setup_pool()
144 dp->xsk_pools[queue_id] = pool; in nfp_net_xsk_setup_pool()
162 int nfp_net_xsk_wakeup(struct net_device *netdev, u32 queue_id, u32 flags) in nfp_net_xsk_wakeup() argument
171 napi_schedule(&nn->r_vecs[queue_id].napi); in nfp_net_xsk_wakeup()
/linux/drivers/media/platform/raspberrypi/rp1-cfe/
H A Dcfe-trace.h17 TP_PROTO(u32 node_id, u32 buf_idx, u32 queue_id),
18 TP_ARGS(node_id, buf_idx, queue_id),
22 __field(u32, queue_id)
27 __entry->queue_id = queue_id;
30 __entry->buf_idx, __entry->queue_id)
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dbwc.c13 hws_bwc_get_burst_th(struct mlx5hws_context *ctx, u16 queue_id) in hws_bwc_get_burst_th() argument
15 return min(ctx->send_queue[queue_id].num_entries / 2, in hws_bwc_get_burst_th()
90 rule_attr.queue_id = mlx5hws_bwc_get_queue_id(ctx, i); in hws_bwc_matcher_move_all_simple()
110 hws_bwc_get_burst_th(ctx, rule_attr.queue_id); in hws_bwc_matcher_move_all_simple()
112 rule_attr.queue_id, in hws_bwc_matcher_move_all_simple()
132 queue = &ctx->send_queue[rule_attr.queue_id]; in hws_bwc_matcher_move_all_simple()
135 rule_attr.queue_id, in hws_bwc_matcher_move_all_simple()
407 u16 queue_id, in mlx5hws_bwc_queue_poll() argument
414 u16 burst_th = hws_bwc_get_burst_th(ctx, queue_id); in mlx5hws_bwc_queue_poll()
425 queue_full = mlx5hws_send_engine_full(&ctx->send_queue[queue_id]); in mlx5hws_bwc_queue_poll()
[all …]
/linux/drivers/media/platform/qcom/iris/
H A Diris_hfi_queue.c199 static void iris_hfi_queue_set_header(struct iris_core *core, u32 queue_id, in iris_hfi_queue_set_header() argument
205 iface_q->qhdr->queue_type = queue_id; in iris_hfi_queue_set_header()
221 if (queue_id == IFACEQ_DBGQ_ID) in iris_hfi_queue_set_header()
226 iris_hfi_queue_init(struct iris_core *core, u32 queue_id, struct iris_iface_q_info *iface_q) in iris_hfi_queue_init() argument
229 u32 offset = sizeof(*q_tbl_hdr) + (queue_id * IFACEQ_QUEUE_SIZE); in iris_hfi_queue_init()
234 iface_q->qhdr = &q_tbl_hdr->q_hdr[queue_id]; in iris_hfi_queue_init()
236 iris_hfi_queue_set_header(core, queue_id, iface_q); in iris_hfi_queue_init()
/linux/net/xdp/
H A Dxsk_buff_pool.c148 bpf.xsk.queue_id = pool->queue_id; in xp_disable_drv_zc()
158 struct net_device *netdev, u16 queue_id, u16 flags) in xp_assign_dev() argument
172 if (xsk_get_pool_from_qid(netdev, queue_id)) in xp_assign_dev()
176 pool->queue_id = queue_id; in xp_assign_dev()
177 err = xsk_reg_pool_at_qid(netdev, pool, queue_id); in xp_assign_dev()
215 bpf.xsk.queue_id = queue_id; in xp_assign_dev()
237 xsk_clear_pool_at_qid(netdev, queue_id); in xp_assign_dev()
244 struct net_device *dev, u16 queue_id) in xp_assign_dev_shared() argument
253 return xp_assign_dev(pool, dev, queue_id, flags); in xp_assign_dev_shared()
265 xsk_clear_pool_at_qid(pool->netdev, pool->queue_id); in xp_clear_dev()

12345678910