/linux/drivers/net/ethernet/engleder/ |
H A D | tsnep_xdp.c | 22 struct xsk_buff_pool *pool, u16 queue_id) in tsnep_xdp_enable_pool() argument 27 if (queue_id >= adapter->num_rx_queues || in tsnep_xdp_enable_pool() 28 queue_id >= adapter->num_tx_queues) in tsnep_xdp_enable_pool() 31 queue = &adapter->queue[queue_id]; in tsnep_xdp_enable_pool() 32 if (queue->rx->queue_index != queue_id || in tsnep_xdp_enable_pool() 33 queue->tx->queue_index != queue_id) { in tsnep_xdp_enable_pool() 58 static int tsnep_xdp_disable_pool(struct tsnep_adapter *adapter, u16 queue_id) in tsnep_xdp_disable_pool() argument 63 if (queue_id >= adapter->num_rx_queues || in tsnep_xdp_disable_pool() 64 queue_id >= adapter->num_tx_queues) in tsnep_xdp_disable_pool() 67 pool = xsk_get_pool_from_qid(adapter->netdev, queue_id); in tsnep_xdp_disable_pool() 81 tsnep_xdp_setup_pool(struct tsnep_adapter * adapter,struct xsk_buff_pool * pool,u16 queue_id) tsnep_xdp_setup_pool() argument [all...] |
/linux/arch/mips/include/asm/octeon/ |
H A D | cvmx-cmd-queue.h | 163 * @queue_id: Hardware command queue to initialize. 170 cvmx_cmd_queue_result_t cvmx_cmd_queue_initialize(cvmx_cmd_queue_id_t queue_id, 179 * @queue_id: Queue to shutdown 183 cvmx_cmd_queue_result_t cvmx_cmd_queue_shutdown(cvmx_cmd_queue_id_t queue_id); 189 * @queue_id: Hardware command queue to query 193 int cvmx_cmd_queue_length(cvmx_cmd_queue_id_t queue_id); 201 * @queue_id: Command queue to query 205 void *cvmx_cmd_queue_buffer(cvmx_cmd_queue_id_t queue_id); 210 * @queue_id: Queue ID to get an index for 214 static inline int __cvmx_cmd_queue_get_index(cvmx_cmd_queue_id_t queue_id) in __cvmx_cmd_queue_get_index() argument 236 __cvmx_cmd_queue_lock(cvmx_cmd_queue_id_t queue_id,__cvmx_cmd_queue_state_t * qptr) __cvmx_cmd_queue_lock() argument 303 __cvmx_cmd_queue_get_state(cvmx_cmd_queue_id_t queue_id) __cvmx_cmd_queue_get_state() argument 326 cvmx_cmd_queue_write(cvmx_cmd_queue_id_t queue_id,int use_locking,int cmd_count,uint64_t * cmds) cvmx_cmd_queue_write() argument 424 cvmx_cmd_queue_write2(cvmx_cmd_queue_id_t queue_id,int use_locking,uint64_t cmd1,uint64_t cmd2) cvmx_cmd_queue_write2() argument 524 cvmx_cmd_queue_write3(cvmx_cmd_queue_id_t queue_id,int use_locking,uint64_t cmd1,uint64_t cmd2,uint64_t cmd3) cvmx_cmd_queue_write3() argument [all...] |
/linux/drivers/net/ethernet/intel/igc/ |
H A D | igc_xdp.c | 57 struct xsk_buff_pool *pool, u16 queue_id) in igc_xdp_enable_pool() argument 67 if (queue_id >= adapter->num_rx_queues || in igc_xdp_enable_pool() 68 queue_id >= adapter->num_tx_queues) in igc_xdp_enable_pool() 89 rx_ring = adapter->rx_ring[queue_id]; in igc_xdp_enable_pool() 90 tx_ring = adapter->tx_ring[queue_id]; in igc_xdp_enable_pool() 108 err = igc_xsk_wakeup(ndev, queue_id, XDP_WAKEUP_RX); in igc_xdp_enable_pool() 118 static int igc_xdp_disable_pool(struct igc_adapter *adapter, u16 queue_id) in igc_xdp_disable_pool() argument 125 if (queue_id >= adapter->num_rx_queues || in igc_xdp_disable_pool() 126 queue_id >= adapter->num_tx_queues) in igc_xdp_disable_pool() 129 pool = xsk_get_pool_from_qid(adapter->netdev, queue_id); in igc_xdp_disable_pool() 160 igc_xdp_setup_pool(struct igc_adapter * adapter,struct xsk_buff_pool * pool,u16 queue_id) igc_xdp_setup_pool() argument [all...] |
/linux/drivers/scsi/lpfc/ |
H A D | lpfc_debugfs.h | 413 q->queue_id, q->type, q->subtype, in lpfc_debug_dump_q() 453 qtypestr, wqidx, wq->queue_id); in lpfc_debug_dump_wq() 456 qtypestr, wq->queue_id); in lpfc_debug_dump_wq() 501 if (cq->assoc_qid == eq->queue_id) in lpfc_debug_dump_cq() 513 qtypestr, wqidx, wq->queue_id, wqidx, cq->queue_id, in lpfc_debug_dump_cq() 514 eqidx, eq->queue_id); in lpfc_debug_dump_cq() 518 qtypestr, wq->queue_id, cq->queue_id, in lpfc_debug_dump_cq() 519 eqidx, eq->queue_id); in lpfc_debug_dump_cq() [all...] |
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_amdkfd_gfx_v10_3.c | 58 uint32_t queue_id) in acquire_queue() argument 63 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 67 uint32_t pipe_id, uint32_t queue_id) in get_queue_mask() argument 70 queue_id; in get_queue_mask() 131 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 161 + queue_id * (mmSDMA0_RLC1_RB_CNTL - mmSDMA0_RLC0_RB_CNTL); in get_sdma_rlc_reg_offset() 164 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset() 180 uint32_t pipe_id, uint32_t queue_id, in hqd_load_v10_3() argument 190 pr_debug("Load hqd of pipe %d queue %d\n", pipe_id, queue_id); in hqd_load_v10_3() 191 acquire_queue(adev, pipe_id, queue_id); in hqd_load_v10_3() 277 hiq_mqd_load_v10_3(struct amdgpu_device * adev,void * mqd,uint32_t pipe_id,uint32_t queue_id,uint32_t doorbell_off,uint32_t inst) hiq_mqd_load_v10_3() argument 329 hqd_dump_v10_3(struct amdgpu_device * adev,uint32_t pipe_id,uint32_t queue_id,uint32_t (** dump)[2],uint32_t * n_regs,uint32_t inst) hqd_dump_v10_3() argument 429 hqd_sdma_dump_v10_3(struct amdgpu_device * adev,uint32_t engine_id,uint32_t queue_id,uint32_t (** dump)[2],uint32_t * n_regs) hqd_sdma_dump_v10_3() argument 461 hqd_is_occupied_v10_3(struct amdgpu_device * adev,uint64_t queue_address,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) hqd_is_occupied_v10_3() argument 503 hqd_destroy_v10_3(struct amdgpu_device * adev,void * mqd,enum kfd_preempt_type reset_type,unsigned int utimeout,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) hqd_destroy_v10_3() argument [all...] |
H A D | amdgpu_amdkfd_gfx_v11.c | 56 uint32_t queue_id) in acquire_queue() argument 61 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 65 uint32_t pipe_id, uint32_t queue_id) in get_queue_mask() argument 68 queue_id; in get_queue_mask() 127 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 146 + queue_id * (regSDMA0_QUEUE1_RB_CNTL - regSDMA0_QUEUE0_RB_CNTL); in get_sdma_rlc_reg_offset() 149 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset() 165 uint32_t queue_id, uint32_t __user *wptr, in hqd_load_v11() argument 175 pr_debug("Load hqd of pipe %d queue %d\n", pipe_id, queue_id); in hqd_load_v11() 176 acquire_queue(adev, pipe_id, queue_id); in hqd_load_v11() 262 hiq_mqd_load_v11(struct amdgpu_device * adev,void * mqd,uint32_t pipe_id,uint32_t queue_id,uint32_t doorbell_off,uint32_t inst) hiq_mqd_load_v11() argument 314 hqd_dump_v11(struct amdgpu_device * adev,uint32_t pipe_id,uint32_t queue_id,uint32_t (** dump)[2],uint32_t * n_regs,uint32_t inst) hqd_dump_v11() argument 414 hqd_sdma_dump_v11(struct amdgpu_device * adev,uint32_t engine_id,uint32_t queue_id,uint32_t (** dump)[2],uint32_t * n_regs) hqd_sdma_dump_v11() argument 450 hqd_is_occupied_v11(struct amdgpu_device * adev,uint64_t queue_address,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) hqd_is_occupied_v11() argument 491 hqd_destroy_v11(struct amdgpu_device * adev,void * mqd,enum kfd_preempt_type reset_type,unsigned int utimeout,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) hqd_destroy_v11() argument 790 kgd_gfx_v11_hqd_get_pq_addr(struct amdgpu_device * adev,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) kgd_gfx_v11_hqd_get_pq_addr() argument 797 kgd_gfx_v11_hqd_reset(struct amdgpu_device * adev,uint32_t pipe_id,uint32_t queue_id,uint32_t inst,unsigned int utimeout) kgd_gfx_v11_hqd_reset() argument [all...] |
H A D | amdgpu_amdkfd_gfx_v9.h | 32 uint32_t queue_id, uint32_t __user *wptr, 36 uint32_t pipe_id, uint32_t queue_id, 39 uint32_t pipe_id, uint32_t queue_id, 43 uint32_t queue_id, uint32_t inst); 47 uint32_t queue_id, uint32_t inst); 62 uint32_t queue_id, uint32_t inst); 64 uint32_t pipe_id, uint32_t queue_id); 108 uint32_t queue_id, 112 uint32_t queue_id,
|
H A D | amdgpu_userq.c | 104 int queue_id) in amdgpu_userq_cleanup() argument 111 idr_remove(&uq_mgr->userq_idr, queue_id); in amdgpu_userq_cleanup() 119 int queue_id; in amdgpu_userq_active() local 124 idr_for_each_entry(&uq_mgr->userq_idr, queue, queue_id) in amdgpu_userq_active() 297 amdgpu_userq_destroy(struct drm_file *filp, int queue_id) in amdgpu_userq_destroy() argument 308 queue = amdgpu_userq_find(uq_mgr, queue_id); in amdgpu_userq_destroy() 326 amdgpu_userq_cleanup(uq_mgr, queue, queue_id); in amdgpu_userq_destroy() 536 args->out.queue_id = qid; in amdgpu_userq_create() 574 r = amdgpu_userq_destroy(filp, args->in.queue_id); in amdgpu_userq_ioctl() 591 int queue_id; in amdgpu_userq_restore_all() local 748 int queue_id; amdgpu_userq_evict_all() local 767 int queue_id, ret; amdgpu_userq_wait_for_signal() local 839 uint32_t queue_id; amdgpu_userq_mgr_fini() local 868 int queue_id; amdgpu_userq_suspend() local 894 int queue_id; amdgpu_userq_resume() local 920 int queue_id; amdgpu_userq_stop_sched_for_enforce_isolation() local 955 int queue_id; amdgpu_userq_start_sched_for_enforce_isolation() local [all...] |
H A D | amdgpu_amdkfd_gfx_v10.c | 58 uint32_t queue_id) in acquire_queue() argument 63 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 67 uint32_t pipe_id, uint32_t queue_id) in get_queue_mask() argument 70 queue_id; in get_queue_mask() 162 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 178 + queue_id * (mmSDMA0_RLC1_RB_CNTL - mmSDMA0_RLC0_RB_CNTL); in get_sdma_rlc_reg_offset() 181 queue_id, retval); in get_sdma_rlc_reg_offset() 209 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_load() argument 219 pr_debug("Load hqd of pipe %d queue %d\n", pipe_id, queue_id); in kgd_hqd_load() 220 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load() 291 kgd_hiq_mqd_load(struct amdgpu_device * adev,void * mqd,uint32_t pipe_id,uint32_t queue_id,uint32_t doorbell_off,uint32_t inst) kgd_hiq_mqd_load() argument 343 kgd_hqd_dump(struct amdgpu_device * adev,uint32_t pipe_id,uint32_t queue_id,uint32_t (** dump)[2],uint32_t * n_regs,uint32_t inst) kgd_hqd_dump() argument 443 kgd_hqd_sdma_dump(struct amdgpu_device * adev,uint32_t engine_id,uint32_t queue_id,uint32_t (** dump)[2],uint32_t * n_regs) kgd_hqd_sdma_dump() argument 475 kgd_hqd_is_occupied(struct amdgpu_device * adev,uint64_t queue_address,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) kgd_hqd_is_occupied() argument 516 kgd_hqd_destroy(struct amdgpu_device * adev,void * mqd,enum kfd_preempt_type reset_type,unsigned int utimeout,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) kgd_hqd_destroy() argument 1074 kgd_gfx_v10_hqd_get_pq_addr(struct amdgpu_device * adev,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) kgd_gfx_v10_hqd_get_pq_addr() argument 1081 kgd_gfx_v10_hqd_reset(struct amdgpu_device * adev,uint32_t pipe_id,uint32_t queue_id,uint32_t inst,unsigned int utimeout) kgd_gfx_v10_hqd_reset() argument [all...] |
H A D | amdgpu_amdkfd_gfx_v8.c | 58 uint32_t queue_id) in acquire_queue() argument 63 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 155 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_load() argument 166 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load() 176 mec, pipe, queue_id); in kgd_hqd_load() 179 ((mec << 5) | (pipe << 3) | queue_id | 0x80)); in kgd_hqd_load() 216 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load() 229 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_dump() argument 245 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_dump() 318 uint32_t engine_id, uint32_t queue_id, in kgd_hqd_sdma_dump() argument 354 kgd_hqd_is_occupied(struct amdgpu_device * adev,uint64_t queue_address,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) kgd_hqd_is_occupied() argument 394 kgd_hqd_destroy(struct amdgpu_device * adev,void * mqd,enum kfd_preempt_type reset_type,unsigned int utimeout,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) kgd_hqd_destroy() argument [all...] |
H A D | amdgpu_amdkfd_gfx_v9.c | 64 uint32_t queue_id, uint32_t inst) in kgd_gfx_v9_acquire_queue() argument 69 kgd_gfx_v9_lock_srbm(adev, mec, pipe, queue_id, 0, inst); in kgd_gfx_v9_acquire_queue() 73 uint32_t pipe_id, uint32_t queue_id) in kgd_gfx_v9_get_queue_mask() argument 76 queue_id; in kgd_gfx_v9_get_queue_mask() 182 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 204 + queue_id * (mmSDMA0_RLC1_RB_CNTL - mmSDMA0_RLC0_RB_CNTL); in get_sdma_rlc_reg_offset() 207 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset() 223 uint32_t pipe_id, uint32_t queue_id, in kgd_gfx_v9_hqd_load() argument 234 kgd_gfx_v9_acquire_queue(adev, pipe_id, queue_id, inst); in kgd_gfx_v9_hqd_load() 286 (uint32_t)kgd_gfx_v9_get_queue_mask(adev, pipe_id, queue_id)); in kgd_gfx_v9_hqd_load() 302 kgd_gfx_v9_hiq_mqd_load(struct amdgpu_device * adev,void * mqd,uint32_t pipe_id,uint32_t queue_id,uint32_t doorbell_off,uint32_t inst) kgd_gfx_v9_hiq_mqd_load() argument 354 kgd_gfx_v9_hqd_dump(struct amdgpu_device * adev,uint32_t pipe_id,uint32_t queue_id,uint32_t (** dump)[2],uint32_t * n_regs,uint32_t inst) kgd_gfx_v9_hqd_dump() argument 454 kgd_hqd_sdma_dump(struct amdgpu_device * adev,uint32_t engine_id,uint32_t queue_id,uint32_t (** dump)[2],uint32_t * n_regs) kgd_hqd_sdma_dump() argument 486 kgd_gfx_v9_hqd_is_occupied(struct amdgpu_device * adev,uint64_t queue_address,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) kgd_gfx_v9_hqd_is_occupied() argument 527 kgd_gfx_v9_hqd_destroy(struct amdgpu_device * adev,void * mqd,enum kfd_preempt_type reset_type,unsigned int utimeout,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) kgd_gfx_v9_hqd_destroy() argument 1128 kgd_gfx_v9_hqd_get_pq_addr(struct amdgpu_device * adev,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) kgd_gfx_v9_hqd_get_pq_addr() argument 1176 kgd_gfx_v9_hqd_reset(struct amdgpu_device * adev,uint32_t pipe_id,uint32_t queue_id,uint32_t inst,unsigned int utimeout) kgd_gfx_v9_hqd_reset() argument [all...] |
H A D | amdgpu_amdkfd_gfx_v7.c | 64 uint32_t queue_id) in acquire_queue() argument 69 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 160 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_load() argument 171 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load() 192 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load() 205 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_dump() argument 221 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_dump() 295 uint32_t engine_id, uint32_t queue_id, in kgd_hqd_sdma_dump() argument 299 queue_id * KFD_CIK_SDMA_QUEUE_OFFSET; in kgd_hqd_sdma_dump() 322 uint32_t queue_id, uint32_ in kgd_hqd_is_occupied() argument 362 kgd_hqd_destroy(struct amdgpu_device * adev,void * mqd,enum kfd_preempt_type reset_type,unsigned int utimeout,uint32_t pipe_id,uint32_t queue_id,uint32_t inst) kgd_hqd_destroy() argument [all...] |
H A D | amdgpu_amdkfd_gfx_v12.c | 44 uint32_t queue_id) in acquire_queue() argument 49 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 78 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 97 + queue_id * (regSDMA0_QUEUE1_RB_CNTL - regSDMA0_QUEUE0_RB_CNTL); in get_sdma_rlc_reg_offset() 100 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset() 106 uint32_t pipe_id, uint32_t queue_id, in hqd_dump_v12() argument 122 acquire_queue(adev, pipe_id, queue_id); in hqd_dump_v12() 137 uint32_t engine_id, uint32_t queue_id, in hqd_sdma_dump_v12() argument 141 engine_id, queue_id); in hqd_sdma_dump_v12()
|
/linux/drivers/net/wireless/st/cw1200/ |
H A D | queue.c | 32 queue->queue_id); in __cw1200_queue_lock() 33 ieee80211_stop_queue(stats->priv->hw, queue->queue_id); in __cw1200_queue_lock() 43 queue->queue_id); in __cw1200_queue_unlock() 44 ieee80211_wake_queue(stats->priv->hw, queue->queue_id); in __cw1200_queue_unlock() 49 u8 *queue_id, u8 *item_generation, in cw1200_queue_parse_id() argument 54 *queue_id = (packet_id >> 16) & 0xFF; in cw1200_queue_parse_id() 58 static inline u32 cw1200_queue_mk_packet_id(u8 queue_generation, u8 queue_id, in cw1200_queue_mk_packet_id() argument 63 ((u32)queue_id << 16) | in cw1200_queue_mk_packet_id() 166 u8 queue_id, in cw1200_queue_init() argument 175 queue->queue_id in cw1200_queue_init() 375 u8 queue_generation, queue_id, item_generation, item_id; cw1200_queue_requeue() local 417 u8 queue_generation, queue_id, item_generation, item_id; cw1200_queue_remove() local 470 u8 queue_generation, queue_id, item_generation, item_id; cw1200_queue_get_skb() local [all...] |
/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_mqd_manager.h | 79 uint32_t pipe_id, uint32_t queue_id, 90 uint32_t queue_id); 97 uint32_t queue_id); 169 uint32_t pipe_id, uint32_t queue_id, 174 uint32_t pipe_id, uint32_t queue_id); 181 uint32_t queue_id); 184 uint32_t pipe_id, uint32_t queue_id, 189 uint32_t pipe_id, uint32_t queue_id); 193 uint32_t queue_id);
|
H A D | kfd_mqd_manager.c | 209 uint32_t pipe_id, uint32_t queue_id, in kfd_hiq_load_mqd_kiq() argument 213 queue_id, p->doorbell_off, 0); in kfd_hiq_load_mqd_kiq() 218 uint32_t pipe_id, uint32_t queue_id) in kfd_destroy_mqd_cp() argument 221 pipe_id, queue_id, 0); in kfd_destroy_mqd_cp() 237 uint32_t queue_id) in kfd_is_occupied_cp() argument 240 pipe_id, queue_id, 0); in kfd_is_occupied_cp() 244 uint32_t pipe_id, uint32_t queue_id, in kfd_load_mqd_sdma() argument 259 uint32_t queue_id) in kfd_destroy_mqd_sdma() argument 266 uint32_t queue_id) in kfd_is_occupied_sdma() argument
|
/linux/drivers/mailbox/ |
H A D | ti-msgmgr.c | 42 * @queue_id: Queue Number for this path 47 u8 queue_id; member 94 * @queue_id: Queue Identifier as mapped on SoC 108 u8 queue_id; member 473 "rx_%03d", d->is_sproxy ? qinst->proxy_id : qinst->queue_id); in ti_msgmgr_queue_rx_irq_req() 486 qinst->queue_id, qinst->proxy_id, in ti_msgmgr_queue_rx_irq_req() 614 if (req_qid == qinst->queue_id && req_pid == qinst->proxy_id) in ti_msgmgr_of_xlate() 648 qinst->queue_id = qd->queue_id; in ti_msgmgr_queue_setup() 650 if (qinst->queue_id > in ti_msgmgr_queue_setup() [all...] |
/linux/drivers/staging/media/atomisp/pci/runtime/bufq/src/ |
H A D | bufq.c | 186 int queue_id; in unmap_buffer_type_to_queue_id() local 193 queue_id = buffer_type_to_queue_id_map[thread_id][buf_type]; in unmap_buffer_type_to_queue_id() 195 queue_availability[thread_id][queue_id] = true; in unmap_buffer_type_to_queue_id() 319 int queue_id, in ia_css_bufq_enqueue_buffer() argument 325 IA_CSS_ENTER_PRIVATE("queue_id=%d", queue_id); in ia_css_bufq_enqueue_buffer() 327 (queue_id == SH_CSS_INVALID_QUEUE_ID)) in ia_css_bufq_enqueue_buffer() 332 queue_id, in ia_css_bufq_enqueue_buffer() 346 int queue_id, in ia_css_bufq_dequeue_buffer() argument 352 IA_CSS_ENTER_PRIVATE("queue_id in ia_css_bufq_dequeue_buffer() [all...] |
/linux/drivers/gpu/drm/amd/include/ |
H A D | kgd_kfd_interface.h | 234 uint32_t queue_id, uint32_t __user *wptr, 239 uint32_t pipe_id, uint32_t queue_id, 246 uint32_t pipe_id, uint32_t queue_id, 250 uint32_t engine_id, uint32_t queue_id, 255 uint32_t queue_id, uint32_t inst); 260 uint32_t queue_id, uint32_t inst); 329 uint32_t pipe_id, uint32_t queue_id, 332 uint32_t pipe_id, uint32_t queue_id,
|
/linux/drivers/net/ethernet/netronome/nfp/ |
H A D | nfp_net_xsk.c | 111 struct xsk_buff_pool *pool, u16 queue_id) in nfp_net_xsk_setup_pool() argument 143 prev_pool = dp->xsk_pools[queue_id]; in nfp_net_xsk_setup_pool() 144 dp->xsk_pools[queue_id] = pool; in nfp_net_xsk_setup_pool() 162 int nfp_net_xsk_wakeup(struct net_device *netdev, u32 queue_id, u32 flags) in nfp_net_xsk_wakeup() argument 166 /* queue_id comes from a zero-copy socket, installed with XDP_SETUP_XSK_POOL, in nfp_net_xsk_wakeup() 171 napi_schedule(&nn->r_vecs[queue_id].napi); in nfp_net_xsk_wakeup()
|
/linux/drivers/media/platform/raspberrypi/rp1-cfe/ |
H A D | cfe-trace.h | 17 TP_PROTO(u32 node_id, u32 buf_idx, u32 queue_id), 18 TP_ARGS(node_id, buf_idx, queue_id), 22 __field(u32, queue_id) 27 __entry->queue_id = queue_id; 30 __entry->buf_idx, __entry->queue_id)
|
/linux/drivers/media/platform/qcom/iris/ |
H A D | iris_hfi_queue.c | 200 static void iris_hfi_queue_set_header(struct iris_core *core, u32 queue_id, in iris_hfi_queue_set_header() argument 206 iface_q->qhdr->queue_type = queue_id; in iris_hfi_queue_set_header() 222 if (queue_id == IFACEQ_DBGQ_ID) in iris_hfi_queue_set_header() 227 iris_hfi_queue_init(struct iris_core *core, u32 queue_id, struct iris_iface_q_info *iface_q) in iris_hfi_queue_init() argument 230 u32 offset = sizeof(*q_tbl_hdr) + (queue_id * IFACEQ_QUEUE_SIZE); in iris_hfi_queue_init() 235 iface_q->qhdr = &q_tbl_hdr->q_hdr[queue_id]; in iris_hfi_queue_init() 237 iris_hfi_queue_set_header(core, queue_id, iface_q); in iris_hfi_queue_init()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
H A D | bwc.c | 13 hws_bwc_get_burst_th(struct mlx5hws_context *ctx, u16 queue_id) in hws_bwc_get_burst_th() argument 15 return min(ctx->send_queue[queue_id].num_entries / 2, in hws_bwc_get_burst_th() 95 rule_attr.queue_id = mlx5hws_bwc_get_queue_id(ctx, i); in hws_bwc_matcher_move_all_simple() 111 rule_attr.queue_id, in hws_bwc_matcher_move_all_simple() 123 queue = &ctx->send_queue[rule_attr.queue_id]; in hws_bwc_matcher_move_all_simple() 126 rule_attr.queue_id, in hws_bwc_matcher_move_all_simple() 378 u16 queue_id, in mlx5hws_bwc_queue_poll() argument 385 u16 burst_th = hws_bwc_get_burst_th(ctx, queue_id); in mlx5hws_bwc_queue_poll() 396 queue_full = mlx5hws_send_engine_full(&ctx->send_queue[queue_id]); in mlx5hws_bwc_queue_poll() 398 ret = mlx5hws_send_queue_poll(ctx, queue_id, com in mlx5hws_bwc_queue_poll() [all...] |
/linux/net/xdp/ |
H A D | xsk_buff_pool.c | 152 bpf.xsk.queue_id = pool->queue_id; in xp_disable_drv_zc() 166 struct net_device *netdev, u16 queue_id, u16 flags) in xp_assign_dev() argument 180 if (xsk_get_pool_from_qid(netdev, queue_id)) in xp_assign_dev() 184 pool->queue_id = queue_id; in xp_assign_dev() 185 err = xsk_reg_pool_at_qid(netdev, pool, queue_id); in xp_assign_dev() 223 bpf.xsk.queue_id = queue_id; in xp_assign_dev() 245 xsk_clear_pool_at_qid(netdev, queue_id); in xp_assign_dev() 252 xp_assign_dev_shared(struct xsk_buff_pool * pool,struct xdp_sock * umem_xs,struct net_device * dev,u16 queue_id) xp_assign_dev_shared() argument [all...] |
/linux/tools/testing/selftests/bpf/ |
H A D | xsk.c | 71 __u32 queue_id; member 459 __u32 queue_id) in xsk_get_ctx() argument 467 if (ctx->ifindex == ifindex && ctx->queue_id == queue_id) { in xsk_get_ctx() 504 __u32 queue_id, in xsk_create_ctx() argument 530 ctx->queue_id = queue_id; in xsk_create_ctx() 540 __u32 queue_id, struct xsk_umem *umem, in xsk_socket__create_shared() argument 580 ctx = xsk_get_ctx(umem, ifindex, queue_id); in xsk_socket__create_shared() 587 ctx = xsk_create_ctx(xsk, umem, ifindex, queue_id, fil in xsk_socket__create_shared() 709 xsk_socket__create(struct xsk_socket ** xsk_ptr,int ifindex,__u32 queue_id,struct xsk_umem * umem,struct xsk_ring_cons * rx,struct xsk_ring_prod * tx,const struct xsk_socket_config * usr_config) xsk_socket__create() argument [all...] |