/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_ring.c | 40 * Most engines on the GPU are fed via ring buffers. Ring 46 * pointers are equal, the ring is idle. When the host 47 * writes commands to the ring buffer, it increments the 55 * @type: ring type for which to return the limit. 73 * amdgpu_ring_alloc - allocate space on the ring buffer 75 * @ring: amdgpu_ring structure holding ring information 76 * @ndw: number of dwords to allocate in the ring buffer 78 * Allocate @ndw dwords in the ring buffer (all asics). 81 int amdgpu_ring_alloc(struct amdgpu_ring *ring, unsigne argument 112 amdgpu_ring_alloc_reemit(struct amdgpu_ring * ring,unsigned int ndw) amdgpu_ring_alloc_reemit() argument 132 amdgpu_ring_insert_nop(struct amdgpu_ring * ring,uint32_t count) amdgpu_ring_insert_nop() argument 160 amdgpu_ring_generic_pad_ib(struct amdgpu_ring * ring,struct amdgpu_ib * ib) amdgpu_ring_generic_pad_ib() argument 175 amdgpu_ring_commit(struct amdgpu_ring * ring) amdgpu_ring_commit() argument 204 amdgpu_ring_undo(struct amdgpu_ring * ring) amdgpu_ring_undo() argument 212 amdgpu_ring_get_gpu_addr(ring,offset) global() argument 215 amdgpu_ring_get_cpu_addr(ring,offset) global() argument 232 amdgpu_ring_init(struct amdgpu_device * adev,struct amdgpu_ring * ring,unsigned int max_dw,struct amdgpu_irq_src * irq_src,unsigned int irq_type,unsigned int hw_prio,atomic_t * sched_score) amdgpu_ring_init() argument 400 amdgpu_ring_fini(struct amdgpu_ring * ring) amdgpu_ring_fini() argument 440 amdgpu_ring_emit_reg_write_reg_wait_helper(struct amdgpu_ring * ring,uint32_t reg0,uint32_t reg1,uint32_t ref,uint32_t mask) amdgpu_ring_emit_reg_write_reg_wait_helper() argument 457 amdgpu_ring_soft_recovery(struct amdgpu_ring * ring,unsigned int vmid,struct dma_fence * fence) amdgpu_ring_soft_recovery() argument 504 struct amdgpu_ring *ring = file_inode(f)->i_private; amdgpu_debugfs_ring_read() local 585 struct amdgpu_ring *ring = file_inode(f)->i_private; amdgpu_debugfs_virt_ring_read() local 611 struct amdgpu_ring *ring = file_inode(f)->i_private; amdgpu_debugfs_mqd_read() local 633 struct amdgpu_ring *ring = data; amdgpu_debugfs_ring_error() local 645 amdgpu_debugfs_ring_init(struct amdgpu_device * adev,struct amdgpu_ring * ring) amdgpu_debugfs_ring_init() argument 685 amdgpu_ring_test_helper(struct amdgpu_ring * ring) amdgpu_ring_test_helper() argument 703 amdgpu_ring_to_mqd_prop(struct amdgpu_ring * ring,struct amdgpu_mqd_prop * prop) amdgpu_ring_to_mqd_prop() argument 736 amdgpu_ring_init_mqd(struct amdgpu_ring * ring) amdgpu_ring_init_mqd() argument 754 amdgpu_ring_ib_begin(struct amdgpu_ring * ring) amdgpu_ring_ib_begin() argument 760 amdgpu_ring_ib_end(struct amdgpu_ring * ring) amdgpu_ring_ib_end() argument 766 amdgpu_ring_ib_on_emit_cntl(struct amdgpu_ring * ring) amdgpu_ring_ib_on_emit_cntl() argument 772 amdgpu_ring_ib_on_emit_ce(struct amdgpu_ring * ring) amdgpu_ring_ib_on_emit_ce() argument 778 amdgpu_ring_ib_on_emit_de(struct amdgpu_ring * ring) amdgpu_ring_ib_on_emit_de() argument 784 amdgpu_ring_sched_ready(struct amdgpu_ring * ring) amdgpu_ring_sched_ready() argument 795 amdgpu_ring_reset_helper_begin(struct amdgpu_ring * ring,struct amdgpu_fence * guilty_fence) amdgpu_ring_reset_helper_begin() argument 804 amdgpu_ring_reset_helper_end(struct amdgpu_ring * ring,struct amdgpu_fence * guilty_fence) amdgpu_ring_reset_helper_end() argument 830 amdgpu_ring_is_reset_type_supported(struct amdgpu_ring * ring,u32 reset_type) amdgpu_ring_is_reset_type_supported() argument [all...] |
H A D | jpeg_v1_0.c | 37 static void jpeg_v1_0_ring_begin_use(struct amdgpu_ring *ring); 42 static void jpeg_v1_0_decode_ring_patch_wreg(struct amdgpu_ring *ring, uint32_t *ptr, uint32_t reg_offset, uint32_t val) in jpeg_v1_0_decode_ring_patch_wreg() argument 44 struct amdgpu_device *adev = ring->adev; in jpeg_v1_0_decode_ring_patch_wreg() 45 ring->ring[(*ptr)++] = PACKETJ(SOC15_REG_OFFSET(JPEG, 0, mmUVD_JRBC_EXTERNAL_REG_BASE), 0, 0, PACKETJ_TYPE0); in jpeg_v1_0_decode_ring_patch_wreg() 48 ring->ring[(*ptr)++] = 0; in jpeg_v1_0_decode_ring_patch_wreg() 49 ring->ring[(*ptr)++] = PACKETJ((reg_offset >> 2), 0, 0, PACKETJ_TYPE0); in jpeg_v1_0_decode_ring_patch_wreg() 51 ring in jpeg_v1_0_decode_ring_patch_wreg() 57 jpeg_v1_0_decode_ring_set_patch_ring(struct amdgpu_ring * ring,uint32_t ptr) jpeg_v1_0_decode_ring_set_patch_ring() argument 140 jpeg_v1_0_decode_ring_get_rptr(struct amdgpu_ring * ring) jpeg_v1_0_decode_ring_get_rptr() argument 154 jpeg_v1_0_decode_ring_get_wptr(struct amdgpu_ring * ring) jpeg_v1_0_decode_ring_get_wptr() argument 168 jpeg_v1_0_decode_ring_set_wptr(struct amdgpu_ring * ring) jpeg_v1_0_decode_ring_set_wptr() argument 182 jpeg_v1_0_decode_ring_insert_start(struct amdgpu_ring * ring) jpeg_v1_0_decode_ring_insert_start() argument 201 jpeg_v1_0_decode_ring_insert_end(struct amdgpu_ring * ring) jpeg_v1_0_decode_ring_insert_end() argument 223 jpeg_v1_0_decode_ring_emit_fence(struct amdgpu_ring * ring,u64 addr,u64 seq,unsigned flags) jpeg_v1_0_decode_ring_emit_fence() argument 297 jpeg_v1_0_decode_ring_emit_ib(struct amdgpu_ring * ring,struct amdgpu_job * job,struct amdgpu_ib * ib,uint32_t flags) jpeg_v1_0_decode_ring_emit_ib() argument 353 jpeg_v1_0_decode_ring_emit_reg_wait(struct amdgpu_ring * ring,uint32_t reg,uint32_t val,uint32_t mask) jpeg_v1_0_decode_ring_emit_reg_wait() argument 383 jpeg_v1_0_decode_ring_emit_vm_flush(struct amdgpu_ring * ring,unsigned vmid,uint64_t pd_addr) jpeg_v1_0_decode_ring_emit_vm_flush() argument 398 jpeg_v1_0_decode_ring_emit_wreg(struct amdgpu_ring * ring,uint32_t reg,uint32_t val) jpeg_v1_0_decode_ring_emit_wreg() argument 419 jpeg_v1_0_decode_ring_nop(struct amdgpu_ring * ring,uint32_t count) jpeg_v1_0_decode_ring_nop() argument 487 struct amdgpu_ring *ring; jpeg_v1_0_sw_init() local 533 struct amdgpu_ring *ring = adev->jpeg.inst->ring_dec; jpeg_v1_0_start() local 604 jpeg_v1_0_ring_begin_use(struct amdgpu_ring * ring) jpeg_v1_0_ring_begin_use() argument [all...] |
H A D | amdgpu_fence.c | 63 * @ring: ring the fence is associated with 68 static void amdgpu_fence_write(struct amdgpu_ring *ring, u32 seq) in amdgpu_fence_write() argument 70 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_fence_write() 79 * @ring: ring the fence is associated with 84 static u32 amdgpu_fence_read(struct amdgpu_ring *ring) in amdgpu_fence_read() argument 86 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_fence_read() 98 * amdgpu_fence_emit - emit a fence on the requested ring 100 * @ring 108 amdgpu_fence_emit(struct amdgpu_ring * ring,struct dma_fence ** f,struct amdgpu_fence * af,unsigned int flags) amdgpu_fence_emit() argument 187 amdgpu_fence_emit_polling(struct amdgpu_ring * ring,uint32_t * s,uint32_t timeout) amdgpu_fence_emit_polling() argument 218 amdgpu_fence_schedule_fallback(struct amdgpu_ring * ring) amdgpu_fence_schedule_fallback() argument 235 amdgpu_fence_process(struct amdgpu_ring * ring) amdgpu_fence_process() argument 296 struct amdgpu_ring *ring = timer_container_of(ring, t, amdgpu_fence_fallback() local 313 amdgpu_fence_wait_empty(struct amdgpu_ring * ring) amdgpu_fence_wait_empty() argument 346 amdgpu_fence_wait_polling(struct amdgpu_ring * ring,uint32_t wait_seq,signed long timeout) amdgpu_fence_wait_polling() argument 366 amdgpu_fence_count_emitted(struct amdgpu_ring * ring) amdgpu_fence_count_emitted() argument 386 amdgpu_fence_last_unsignaled_time_us(struct amdgpu_ring * ring) amdgpu_fence_last_unsignaled_time_us() argument 417 amdgpu_fence_update_start_timestamp(struct amdgpu_ring * ring,uint32_t seq,ktime_t timestamp) amdgpu_fence_update_start_timestamp() argument 443 amdgpu_fence_driver_start_ring(struct amdgpu_ring * ring,struct amdgpu_irq_src * irq_src,unsigned int irq_type) amdgpu_fence_driver_start_ring() argument 479 amdgpu_fence_driver_init_ring(struct amdgpu_ring * ring) amdgpu_fence_driver_init_ring() argument 536 amdgpu_fence_need_ring_interrupt_restore(struct amdgpu_ring * ring) amdgpu_fence_need_ring_interrupt_restore() argument 574 struct amdgpu_ring *ring = adev->rings[i]; amdgpu_fence_driver_hw_fini() local 604 struct amdgpu_ring *ring = adev->rings[i]; amdgpu_fence_driver_isr_toggle() local 621 struct amdgpu_ring *ring = adev->rings[i]; amdgpu_fence_driver_sw_fini() local 660 struct amdgpu_ring *ring = adev->rings[i]; amdgpu_fence_driver_hw_init() local 679 amdgpu_fence_driver_clear_job_fences(struct amdgpu_ring * ring) amdgpu_fence_driver_clear_job_fences() argument 710 amdgpu_fence_driver_set_error(struct amdgpu_ring * ring,int error) amdgpu_fence_driver_set_error() argument 733 amdgpu_fence_driver_force_completion(struct amdgpu_ring * ring) amdgpu_fence_driver_force_completion() argument 776 amdgpu_ring_backup_unprocessed_command(struct amdgpu_ring * ring,u64 start_wptr,u32 end_wptr) amdgpu_ring_backup_unprocessed_command() argument 788 amdgpu_ring_backup_unprocessed_commands(struct amdgpu_ring * ring,struct amdgpu_fence * guilty_fence) amdgpu_ring_backup_unprocessed_commands() argument 954 struct amdgpu_ring *ring = adev->rings[i]; amdgpu_debugfs_fence_info_show() local [all...] |
H A D | amdgpu_ib.c | 45 * command ring and the hw will fetch the commands from the IB 48 * put in IBs for execution by the requested ring. 103 * amdgpu_ib_schedule - schedule an IB (Indirect Buffer) on the ring 105 * @ring: ring index the IB is associated with 111 * Schedule an IB on the associated ring (all asics). 114 * On SI, there are two parallel engines fed from the primary ring, 121 * a CONST_IB), it will be put on the ring prior to the DE IB. Prior 124 int amdgpu_ib_schedule(struct amdgpu_ring *ring, unsigned int num_ibs, in amdgpu_ib_schedule() argument 128 struct amdgpu_device *adev = ring in amdgpu_ib_schedule() 419 struct amdgpu_ring *ring = adev->rings[i]; amdgpu_ib_ring_tests() local [all...] |
H A D | amdgpu_ring_mux.c | 44 struct amdgpu_ring *ring) in amdgpu_ring_mux_sw_entry() argument 46 return ring->entry_index < mux->ring_entry_size ? in amdgpu_ring_mux_sw_entry() 47 &mux->ring_entry[ring->entry_index] : NULL; in amdgpu_ring_mux_sw_entry() 50 /* copy packages on sw ring range[begin, end) */ 52 struct amdgpu_ring *ring, in amdgpu_ring_mux_copy_pkt_from_sw_ring() argument 58 start = s_start & ring->buf_mask; in amdgpu_ring_mux_copy_pkt_from_sw_ring() 59 end = s_end & ring->buf_mask; in amdgpu_ring_mux_copy_pkt_from_sw_ring() 62 DRM_ERROR("no more data copied from sw ring\n"); in amdgpu_ring_mux_copy_pkt_from_sw_ring() 66 amdgpu_ring_alloc(real_ring, (ring->ring_size >> 2) + end - start); in amdgpu_ring_mux_copy_pkt_from_sw_ring() 67 amdgpu_ring_write_multiple(real_ring, (void *)&ring in amdgpu_ring_mux_copy_pkt_from_sw_ring() 150 amdgpu_ring_mux_init(struct amdgpu_ring_mux * mux,struct amdgpu_ring * ring,unsigned int entry_size) amdgpu_ring_mux_init() argument 195 amdgpu_ring_mux_add_sw_ring(struct amdgpu_ring_mux * mux,struct amdgpu_ring * ring) amdgpu_ring_mux_add_sw_ring() argument 213 amdgpu_ring_mux_set_wptr(struct amdgpu_ring_mux * mux,struct amdgpu_ring * ring,u64 wptr) amdgpu_ring_mux_set_wptr() argument 253 amdgpu_ring_mux_get_wptr(struct amdgpu_ring_mux * mux,struct amdgpu_ring * ring) amdgpu_ring_mux_get_wptr() argument 282 amdgpu_ring_mux_get_rptr(struct amdgpu_ring_mux * mux,struct amdgpu_ring * ring) amdgpu_ring_mux_get_rptr() argument 316 amdgpu_sw_ring_get_rptr_gfx(struct amdgpu_ring * ring) amdgpu_sw_ring_get_rptr_gfx() argument 325 amdgpu_sw_ring_get_wptr_gfx(struct amdgpu_ring * ring) amdgpu_sw_ring_get_wptr_gfx() argument 334 amdgpu_sw_ring_set_wptr_gfx(struct amdgpu_ring * ring) amdgpu_sw_ring_set_wptr_gfx() argument 344 amdgpu_sw_ring_insert_nop(struct amdgpu_ring * ring,uint32_t count) amdgpu_sw_ring_insert_nop() argument 364 struct amdgpu_ring *ring; amdgpu_mcbp_scan() local 393 amdgpu_sw_ring_ib_begin(struct amdgpu_ring * ring) amdgpu_sw_ring_ib_begin() argument 408 amdgpu_sw_ring_ib_end(struct amdgpu_ring * ring) amdgpu_sw_ring_ib_end() argument 419 amdgpu_sw_ring_ib_mark_offset(struct amdgpu_ring * ring,enum amdgpu_ring_mux_offset_type type) amdgpu_sw_ring_ib_mark_offset() argument 433 amdgpu_ring_mux_start_ib(struct amdgpu_ring_mux * mux,struct amdgpu_ring * ring) amdgpu_ring_mux_start_ib() argument 462 scan_and_remove_signaled_chunk(struct amdgpu_ring_mux * mux,struct amdgpu_ring * ring) scan_and_remove_signaled_chunk() argument 485 amdgpu_ring_mux_ib_mark_offset(struct amdgpu_ring_mux * mux,struct amdgpu_ring * ring,u64 offset,enum amdgpu_ring_mux_offset_type type) amdgpu_ring_mux_ib_mark_offset() argument 519 amdgpu_ring_mux_end_ib(struct amdgpu_ring_mux * mux,struct amdgpu_ring * ring) amdgpu_ring_mux_end_ib() argument 545 struct amdgpu_ring *ring = NULL; amdgpu_mcbp_handle_trailing_fence_irq() local [all...] |
/linux/drivers/gpu/drm/radeon/ |
H A D | radeon_ring.c | 39 * Most engines on the GPU are fed via ring buffers. Ring 45 * pointers are equal, the ring is idle. When the host 46 * writes commands to the ring buffer, it increments the 50 static void radeon_debugfs_ring_init(struct radeon_device *rdev, struct radeon_ring *ring); 53 * radeon_ring_supports_scratch_reg - check if the ring supports 57 * @ring: radeon_ring structure holding ring information 59 * Check if a specific ring supports writing to scratch registers (all asics). 60 * Returns true if the ring supports writing to scratch regs, false if not. 63 struct radeon_ring *ring) in radeon_ring_supports_scratch_reg() argument 83 radeon_ring_free_size(struct radeon_device * rdev,struct radeon_ring * ring) radeon_ring_free_size() argument 109 radeon_ring_alloc(struct radeon_device * rdev,struct radeon_ring * ring,unsigned ndw) radeon_ring_alloc() argument 145 radeon_ring_lock(struct radeon_device * rdev,struct radeon_ring * ring,unsigned ndw) radeon_ring_lock() argument 169 radeon_ring_commit(struct radeon_device * rdev,struct radeon_ring * ring,bool hdp_flush) radeon_ring_commit() argument 200 radeon_ring_unlock_commit(struct radeon_device * rdev,struct radeon_ring * ring,bool hdp_flush) radeon_ring_unlock_commit() argument 214 radeon_ring_undo(struct radeon_ring * ring) radeon_ring_undo() argument 227 radeon_ring_unlock_undo(struct radeon_device * rdev,struct radeon_ring * ring) radeon_ring_unlock_undo() argument 242 radeon_ring_lockup_update(struct radeon_device * rdev,struct radeon_ring * ring) radeon_ring_lockup_update() argument 254 radeon_ring_test_lockup(struct radeon_device * rdev,struct radeon_ring * ring) radeon_ring_test_lockup() argument 285 radeon_ring_backup(struct radeon_device * rdev,struct radeon_ring * ring,uint32_t ** data) radeon_ring_backup() argument 349 radeon_ring_restore(struct radeon_device * rdev,struct radeon_ring * ring,unsigned size,uint32_t * data) radeon_ring_restore() argument 383 radeon_ring_init(struct radeon_device * rdev,struct radeon_ring * ring,unsigned ring_size,unsigned rptr_offs,u32 nop) radeon_ring_init() argument 439 radeon_ring_fini(struct radeon_device * rdev,struct radeon_ring * ring) radeon_ring_fini() argument 469 struct radeon_ring *ring = m->private; radeon_debugfs_ring_info_show() local 549 radeon_debugfs_ring_init(struct radeon_device * rdev,struct radeon_ring * ring) radeon_debugfs_ring_init() argument [all...] |
H A D | evergreen_dma.c | 31 * evergreen_dma_fence_ring_emit - emit a fence on the DMA ring 36 * Add a DMA fence packet to the ring to write 43 struct radeon_ring *ring = &rdev->ring[fence->ring]; in evergreen_dma_fence_ring_emit() local 44 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in evergreen_dma_fence_ring_emit() 46 radeon_ring_write(ring, DMA_PACKET(DMA_PACKET_FENCE, 0, 0)); in evergreen_dma_fence_ring_emit() 47 radeon_ring_write(ring, addr & 0xfffffffc); in evergreen_dma_fence_ring_emit() 48 radeon_ring_write(ring, (upper_32_bits(addr) & 0xff)); in evergreen_dma_fence_ring_emit() 49 radeon_ring_write(ring, fenc in evergreen_dma_fence_ring_emit() 69 struct radeon_ring *ring = &rdev->ring[ib->ring]; evergreen_dma_ring_ib_execute() local 115 struct radeon_ring *ring = &rdev->ring[ring_index]; evergreen_copy_dma() local 170 evergreen_dma_is_lockup(struct radeon_device * rdev,struct radeon_ring * ring) evergreen_dma_is_lockup() argument [all...] |
H A D | r600_dma.c | 34 * to the 3D engine (ring buffer, IBs, etc.), but the 46 * @ring: radeon ring pointer 51 struct radeon_ring *ring) in r600_dma_get_rptr() argument 56 rptr = rdev->wb.wb[ring->rptr_offs/4]; in r600_dma_get_rptr() 67 * @ring: radeon ring pointer 72 struct radeon_ring *ring) in r600_dma_get_wptr() argument 81 * @ring: radeon ring pointe 86 r600_dma_set_wptr(struct radeon_device * rdev,struct radeon_ring * ring) r600_dma_set_wptr() argument 121 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; r600_dma_resume() local 207 r600_dma_is_lockup(struct radeon_device * rdev,struct radeon_ring * ring) r600_dma_is_lockup() argument 230 r600_dma_ring_test(struct radeon_device * rdev,struct radeon_ring * ring) r600_dma_ring_test() argument 289 struct radeon_ring *ring = &rdev->ring[fence->ring]; r600_dma_fence_ring_emit() local 313 r600_dma_semaphore_ring_emit(struct radeon_device * rdev,struct radeon_ring * ring,struct radeon_semaphore * semaphore,bool emit_wait) r600_dma_semaphore_ring_emit() argument 336 r600_dma_ib_test(struct radeon_device * rdev,struct radeon_ring * ring) r600_dma_ib_test() argument 406 struct radeon_ring *ring = &rdev->ring[ib->ring]; r600_dma_ring_ib_execute() local 451 struct radeon_ring *ring = &rdev->ring[ring_index]; r600_copy_dma() local [all...] |
H A D | uvd_v1_0.c | 35 * @ring: radeon_ring pointer 40 struct radeon_ring *ring) in uvd_v1_0_get_rptr() argument 49 * @ring: radeon_ring pointer 54 struct radeon_ring *ring) in uvd_v1_0_get_wptr() argument 63 * @ring: radeon_ring pointer 68 struct radeon_ring *ring) in uvd_v1_0_set_wptr() argument 70 WREG32(UVD_RBC_RB_WPTR, ring->wptr); in uvd_v1_0_set_wptr() 79 * Write a fence and a trap command to the ring. 84 struct radeon_ring *ring = &rdev->ring[fenc in uvd_v1_0_fence_emit() local 159 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; uvd_v1_0_init() local 250 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; uvd_v1_0_fini() local 265 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; uvd_v1_0_start() local 421 uvd_v1_0_ring_test(struct radeon_device * rdev,struct radeon_ring * ring) uvd_v1_0_ring_test() argument 466 uvd_v1_0_semaphore_emit(struct radeon_device * rdev,struct radeon_ring * ring,struct radeon_semaphore * semaphore,bool emit_wait) uvd_v1_0_semaphore_emit() argument 484 struct radeon_ring *ring = &rdev->ring[ib->ring]; uvd_v1_0_ib_execute() local 500 uvd_v1_0_ib_test(struct radeon_device * rdev,struct radeon_ring * ring) uvd_v1_0_ib_test() argument [all...] |
H A D | cik_sdma.c | 42 * and each one supports 1 ring buffer used for gfx 46 * (ring buffer, IBs, etc.), but sDMA has it's own 58 * @ring: radeon ring pointer 63 struct radeon_ring *ring) in cik_sdma_get_rptr() argument 68 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cik_sdma_get_rptr() 70 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in cik_sdma_get_rptr() 85 * @ring: radeon ring pointer 90 struct radeon_ring *ring) in cik_sdma_get_wptr() argument 111 cik_sdma_set_wptr(struct radeon_device * rdev,struct radeon_ring * ring) cik_sdma_set_wptr() argument 135 struct radeon_ring *ring = &rdev->ring[ib->ring]; cik_sdma_ring_ib_execute() local 171 struct radeon_ring *ring = &rdev->ring[ridx]; cik_sdma_hdp_flush_ring_emit() local 202 struct radeon_ring *ring = &rdev->ring[fence->ring]; cik_sdma_fence_ring_emit() local 228 cik_sdma_semaphore_ring_emit(struct radeon_device * rdev,struct radeon_ring * ring,struct radeon_semaphore * semaphore,bool emit_wait) cik_sdma_semaphore_ring_emit() argument 366 struct radeon_ring *ring; cik_sdma_gfx_resume() local 586 struct radeon_ring *ring = &rdev->ring[ring_index]; cik_copy_dma() local 645 cik_sdma_ring_test(struct radeon_device * rdev,struct radeon_ring * ring) cik_sdma_ring_test() argument 701 cik_sdma_ib_test(struct radeon_device * rdev,struct radeon_ring * ring) cik_sdma_ib_test() argument 774 cik_sdma_is_lockup(struct radeon_device * rdev,struct radeon_ring * ring) cik_sdma_is_lockup() argument 944 cik_dma_vm_flush(struct radeon_device * rdev,struct radeon_ring * ring,unsigned vm_id,uint64_t pd_addr) cik_dma_vm_flush() argument [all...] |
/linux/drivers/net/wireless/broadcom/b43legacy/ |
H A D | dma.c | 32 struct b43legacy_dmadesc32 *op32_idx2desc(struct b43legacy_dmaring *ring, in op32_idx2desc() argument 38 *meta = &(ring->meta[slot]); in op32_idx2desc() 39 desc = ring->descbase; in op32_idx2desc() 45 static void op32_fill_descriptor(struct b43legacy_dmaring *ring, in op32_fill_descriptor() argument 50 struct b43legacy_dmadesc32 *descbase = ring->descbase; in op32_fill_descriptor() 57 B43legacy_WARN_ON(!(slot >= 0 && slot < ring->nr_slots)); in op32_fill_descriptor() 62 addr |= ring->dev->dma.translation; in op32_fill_descriptor() 63 ctl = (bufsize - ring->frameoffset) in op32_fill_descriptor() 65 if (slot == ring->nr_slots - 1) in op32_fill_descriptor() 80 static void op32_poke_tx(struct b43legacy_dmaring *ring, in argument 86 op32_tx_suspend(struct b43legacy_dmaring * ring) op32_tx_suspend() argument 93 op32_tx_resume(struct b43legacy_dmaring * ring) op32_tx_resume() argument 100 op32_get_current_rxslot(struct b43legacy_dmaring * ring) op32_get_current_rxslot() argument 110 op32_set_current_rxslot(struct b43legacy_dmaring * ring,int slot) op32_set_current_rxslot() argument 117 free_slots(struct b43legacy_dmaring * ring) free_slots() argument 122 next_slot(struct b43legacy_dmaring * ring,int slot) next_slot() argument 131 update_max_used_slots(struct b43legacy_dmaring * ring,int current_used_slots) update_max_used_slots() argument 146 update_max_used_slots(struct b43legacy_dmaring * ring,int current_used_slots) update_max_used_slots() argument 153 request_slot(struct b43legacy_dmaring * ring) request_slot() argument 175 struct b43legacy_dmaring *ring; priority_to_txring() local 226 map_descbuffer(struct b43legacy_dmaring * ring,unsigned char * buf,size_t len,int tx) map_descbuffer() argument 246 unmap_descbuffer(struct b43legacy_dmaring * ring,dma_addr_t addr,size_t len,int tx) unmap_descbuffer() argument 262 sync_descbuffer_for_cpu(struct b43legacy_dmaring * ring,dma_addr_t addr,size_t len) sync_descbuffer_for_cpu() argument 273 sync_descbuffer_for_device(struct b43legacy_dmaring * ring,dma_addr_t addr,size_t len) sync_descbuffer_for_device() argument 284 free_descriptor_buffer(struct b43legacy_dmaring * ring,struct b43legacy_dmadesc_meta * meta,int irq_context) free_descriptor_buffer() argument 297 alloc_ringmemory(struct b43legacy_dmaring * ring) alloc_ringmemory() argument 309 free_ringmemory(struct b43legacy_dmaring * ring) free_ringmemory() argument 390 b43legacy_dma_mapping_error(struct b43legacy_dmaring * ring,dma_addr_t addr,size_t buffersize,bool dma_to_device) b43legacy_dma_mapping_error() argument 419 setup_rx_descbuffer(struct b43legacy_dmaring * ring,struct b43legacy_dmadesc32 * desc,struct b43legacy_dmadesc_meta * meta,gfp_t gfp_flags) setup_rx_descbuffer() argument 469 alloc_initial_descbuffers(struct b43legacy_dmaring * ring) alloc_initial_descbuffers() argument 506 dmacontroller_setup(struct b43legacy_dmaring * ring) dmacontroller_setup() argument 548 dmacontroller_cleanup(struct b43legacy_dmaring * ring) dmacontroller_cleanup() argument 561 free_all_descbuffers(struct b43legacy_dmaring * ring) free_all_descbuffers() argument 608 struct b43legacy_dmaring *ring; b43legacy_setup_dmaring() local 707 b43legacy_destroy_dmaring(struct b43legacy_dmaring * ring) b43legacy_destroy_dmaring() argument 758 struct b43legacy_dmaring *ring; b43legacy_dma_init() local 852 generate_cookie(struct b43legacy_dmaring * ring,int slot) generate_cookie() argument 895 struct b43legacy_dmaring *ring = NULL; parse_cookie() local 925 dma_tx_fragment(struct b43legacy_dmaring * ring,struct sk_buff ** in_skb) dma_tx_fragment() argument 1026 should_inject_overflow(struct b43legacy_dmaring * ring) should_inject_overflow() argument 1051 struct b43legacy_dmaring *ring; b43legacy_dma_tx() local 1104 struct b43legacy_dmaring *ring; b43legacy_dma_handle_txstatus() local 1220 dma_rx(struct b43legacy_dmaring * ring,int * slot) dma_rx() argument 1318 b43legacy_dma_rx(struct b43legacy_dmaring * ring) b43legacy_dma_rx() argument 1338 b43legacy_dma_tx_suspend_ring(struct b43legacy_dmaring * ring) b43legacy_dma_tx_suspend_ring() argument 1344 b43legacy_dma_tx_resume_ring(struct b43legacy_dmaring * ring) b43legacy_dma_tx_resume_ring() argument [all...] |
/linux/drivers/bus/mhi/ep/ |
H A D | ring.c | 10 size_t mhi_ep_ring_addr2offset(struct mhi_ep_ring *ring, u64 ptr) in mhi_ep_ring_addr2offset() argument 12 return (ptr - ring->rbase) / sizeof(struct mhi_ring_element); in mhi_ep_ring_addr2offset() 15 static u32 mhi_ep_ring_num_elems(struct mhi_ep_ring *ring) in mhi_ep_ring_num_elems() argument 19 memcpy_fromio(&rlen, (void __iomem *) &ring->ring_ctx->generic.rlen, sizeof(u64)); in mhi_ep_ring_num_elems() 24 void mhi_ep_ring_inc_index(struct mhi_ep_ring *ring) in mhi_ep_ring_inc_index() argument 26 ring->rd_offset = (ring->rd_offset + 1) % ring->ring_size; in mhi_ep_ring_inc_index() 29 static int __mhi_ep_cache_ring(struct mhi_ep_ring *ring, size_t end) in __mhi_ep_cache_ring() argument 31 struct mhi_ep_cntrl *mhi_cntrl = ring in __mhi_ep_cache_ring() 79 mhi_ep_cache_ring(struct mhi_ep_ring * ring,u64 wr_ptr) mhi_ep_cache_ring() argument 96 mhi_ep_update_wr_offset(struct mhi_ep_ring * ring) mhi_ep_update_wr_offset() argument 106 mhi_ep_ring_add_element(struct mhi_ep_ring * ring,struct mhi_ring_element * el) mhi_ep_ring_add_element() argument 153 mhi_ep_ring_init(struct mhi_ep_ring * ring,enum mhi_ep_ring_type type,u32 id) mhi_ep_ring_init() argument 171 struct mhi_ep_ring *ring = container_of(work, struct mhi_ep_ring, intmodt_work.work); mhi_ep_raise_irq() local 178 mhi_ep_ring_start(struct mhi_ep_cntrl * mhi_cntrl,struct mhi_ep_ring * ring,union mhi_ep_ring_ctx * ctx) mhi_ep_ring_start() argument 225 mhi_ep_ring_reset(struct mhi_ep_cntrl * mhi_cntrl,struct mhi_ep_ring * ring) mhi_ep_ring_reset() argument [all...] |
/linux/drivers/net/wireless/broadcom/b43/ |
H A D | dma.c | 72 struct b43_dmadesc_generic *op32_idx2desc(struct b43_dmaring *ring, in op32_idx2desc() argument 78 *meta = &(ring->meta[slot]); in op32_idx2desc() 79 desc = ring->descbase; in op32_idx2desc() 85 static void op32_fill_descriptor(struct b43_dmaring *ring, in op32_fill_descriptor() argument 90 struct b43_dmadesc32 *descbase = ring->descbase; in op32_fill_descriptor() 97 B43_WARN_ON(!(slot >= 0 && slot < ring->nr_slots)); in op32_fill_descriptor() 99 addr = b43_dma_address(&ring->dev->dma, dmaaddr, B43_DMA_ADDR_LOW); in op32_fill_descriptor() 100 addrext = b43_dma_address(&ring->dev->dma, dmaaddr, B43_DMA_ADDR_EXT); in op32_fill_descriptor() 103 if (slot == ring->nr_slots - 1) in op32_fill_descriptor() 118 static void op32_poke_tx(struct b43_dmaring *ring, in argument 124 op32_tx_suspend(struct b43_dmaring * ring) op32_tx_suspend() argument 130 op32_tx_resume(struct b43_dmaring * ring) op32_tx_resume() argument 136 op32_get_current_rxslot(struct b43_dmaring * ring) op32_get_current_rxslot() argument 146 op32_set_current_rxslot(struct b43_dmaring * ring,int slot) op32_set_current_rxslot() argument 164 op64_idx2desc(struct b43_dmaring * ring,int slot,struct b43_dmadesc_meta ** meta) op64_idx2desc() argument 177 op64_fill_descriptor(struct b43_dmaring * ring,struct b43_dmadesc_generic * desc,dma_addr_t dmaaddr,u16 bufsize,int start,int end,int irq) op64_fill_descriptor() argument 213 op64_poke_tx(struct b43_dmaring * ring,int slot) op64_poke_tx() argument 219 op64_tx_suspend(struct b43_dmaring * ring) op64_tx_suspend() argument 225 op64_tx_resume(struct b43_dmaring * ring) op64_tx_resume() argument 231 op64_get_current_rxslot(struct b43_dmaring * ring) op64_get_current_rxslot() argument 241 op64_set_current_rxslot(struct b43_dmaring * ring,int slot) op64_set_current_rxslot() argument 257 free_slots(struct b43_dmaring * ring) free_slots() argument 262 next_slot(struct b43_dmaring * ring,int slot) next_slot() argument 270 prev_slot(struct b43_dmaring * ring,int slot) prev_slot() argument 279 update_max_used_slots(struct b43_dmaring * ring,int current_used_slots) update_max_used_slots() argument 294 update_max_used_slots(struct b43_dmaring * ring,int current_used_slots) update_max_used_slots() argument 300 request_slot(struct b43_dmaring * ring) request_slot() argument 347 map_descbuffer(struct b43_dmaring * ring,unsigned char * buf,size_t len,int tx) map_descbuffer() argument 364 unmap_descbuffer(struct b43_dmaring * ring,dma_addr_t addr,size_t len,int tx) unmap_descbuffer() argument 377 sync_descbuffer_for_cpu(struct b43_dmaring * ring,dma_addr_t addr,size_t len) sync_descbuffer_for_cpu() argument 386 sync_descbuffer_for_device(struct b43_dmaring * ring,dma_addr_t addr,size_t len) sync_descbuffer_for_device() argument 395 free_descriptor_buffer(struct b43_dmaring * ring,struct b43_dmadesc_meta * meta) free_descriptor_buffer() argument 407 alloc_ringmemory(struct b43_dmaring * ring) alloc_ringmemory() argument 430 free_ringmemory(struct b43_dmaring * ring) free_ringmemory() argument 538 b43_dma_mapping_error(struct b43_dmaring * ring,dma_addr_t addr,size_t buffersize,bool dma_to_device) b43_dma_mapping_error() argument 570 b43_rx_buffer_is_poisoned(struct b43_dmaring * ring,struct sk_buff * skb) b43_rx_buffer_is_poisoned() argument 577 b43_poison_rx_buffer(struct b43_dmaring * ring,struct sk_buff * skb) b43_poison_rx_buffer() argument 592 setup_rx_descbuffer(struct b43_dmaring * ring,struct b43_dmadesc_generic * desc,struct b43_dmadesc_meta * meta,gfp_t gfp_flags) setup_rx_descbuffer() argument 636 alloc_initial_descbuffers(struct b43_dmaring * ring) alloc_initial_descbuffers() argument 672 dmacontroller_setup(struct b43_dmaring * ring) dmacontroller_setup() argument 753 dmacontroller_cleanup(struct b43_dmaring * ring) dmacontroller_cleanup() argument 774 free_all_descbuffers(struct b43_dmaring * ring) free_all_descbuffers() argument 837 struct b43_dmaring *ring; b43_setup_dmaring() local 966 b43_destroy_dmaring(struct b43_dmaring * ring,const char * ringname) b43_destroy_dmaring() argument 1012 destroy_ring(dma,ring) global() argument 1136 generate_cookie(struct b43_dmaring * ring,int slot) generate_cookie() argument 1160 struct b43_dmaring *ring = NULL; parse_cookie() local 1189 dma_tx_fragment(struct b43_dmaring * ring,struct sk_buff * skb) dma_tx_fragment() argument 1288 should_inject_overflow(struct b43_dmaring * ring) should_inject_overflow() argument 1313 struct b43_dmaring *ring; select_ring_by_priority() local 1342 struct b43_dmaring *ring; b43_dma_tx() local 1418 struct b43_dmaring *ring; b43_dma_handle_txstatus() local 1582 dma_rx(struct b43_dmaring * ring,int * slot) dma_rx() argument 1668 b43_dma_handle_rx_overflow(struct b43_dmaring * ring) b43_dma_handle_rx_overflow() argument 1687 b43_dma_rx(struct b43_dmaring * ring) b43_dma_rx() argument 1707 b43_dma_tx_suspend_ring(struct b43_dmaring * ring) b43_dma_tx_suspend_ring() argument 1713 b43_dma_tx_resume_ring(struct b43_dmaring * ring) b43_dma_tx_resume_ring() argument [all...] |
/linux/drivers/thunderbolt/ |
H A D | nhi.c | 28 #define RING_TYPE(ring) ((ring)->is_tx ? "TX ring" : "RX ring") argument 53 static int ring_interrupt_index(const struct tb_ring *ring) in ring_interrupt_index() argument 55 int bit = ring->hop; in ring_interrupt_index() 56 if (!ring->is_tx) in ring_interrupt_index() 57 bit += ring->nhi->hop_count; in ring_interrupt_index() 61 static void nhi_mask_interrupt(struct tb_nhi *nhi, int mask, int ring) in nhi_mask_interrupt() argument 66 val = ioread32(nhi->iobase + REG_RING_INTERRUPT_BASE + ring); in nhi_mask_interrupt() 73 nhi_clear_interrupt(struct tb_nhi * nhi,int ring) nhi_clear_interrupt() argument 86 ring_interrupt_active(struct tb_ring * ring,bool active) ring_interrupt_active() argument 176 ring_desc_base(struct tb_ring * ring) ring_desc_base() argument 184 ring_options_base(struct tb_ring * ring) ring_options_base() argument 192 ring_iowrite_cons(struct tb_ring * ring,u16 cons) ring_iowrite_cons() argument 202 ring_iowrite_prod(struct tb_ring * ring,u16 prod) ring_iowrite_prod() argument 208 ring_iowrite32desc(struct tb_ring * ring,u32 value,u32 offset) ring_iowrite32desc() argument 213 ring_iowrite64desc(struct tb_ring * ring,u64 value,u32 offset) ring_iowrite64desc() argument 219 ring_iowrite32options(struct tb_ring * ring,u32 value,u32 offset) ring_iowrite32options() argument 224 ring_full(struct tb_ring * ring) ring_full() argument 229 ring_empty(struct tb_ring * ring) ring_empty() argument 239 ring_write_descriptors(struct tb_ring * ring) ring_write_descriptors() argument 275 struct tb_ring *ring = container_of(work, typeof(*ring), work); ring_work() local 323 __tb_ring_enqueue(struct tb_ring * ring,struct ring_frame * frame) __tb_ring_enqueue() argument 349 tb_ring_poll(struct tb_ring * ring) tb_ring_poll() argument 381 __ring_interrupt_mask(struct tb_ring * ring,bool mask) __ring_interrupt_mask() argument 397 __ring_interrupt(struct tb_ring * ring) __ring_interrupt() argument 417 tb_ring_poll_complete(struct tb_ring * ring) tb_ring_poll_complete() argument 430 ring_clear_msix(const struct tb_ring * ring) ring_clear_msix() argument 447 struct tb_ring *ring = data; ring_msix() local 459 ring_request_msix(struct tb_ring * ring,bool no_suspend) ring_request_msix() argument 493 ring_release_msix(struct tb_ring * ring) ring_release_msix() argument 504 nhi_alloc_hop(struct tb_nhi * nhi,struct tb_ring * ring) nhi_alloc_hop() argument 582 struct tb_ring *ring = NULL; tb_ring_alloc() local 680 tb_ring_start(struct tb_ring * ring) tb_ring_start() argument 764 tb_ring_stop(struct tb_ring * ring) tb_ring_stop() argument 809 tb_ring_free(struct tb_ring * ring) tb_ring_free() argument 912 struct tb_ring *ring; nhi_interrupt_work() local [all...] |
/linux/drivers/net/ethernet/apm/xgene/ |
H A D | xgene_enet_ring2.c | 12 static void xgene_enet_ring_init(struct xgene_enet_desc_ring *ring) in xgene_enet_ring_init() argument 14 u32 *ring_cfg = ring->state; in xgene_enet_ring_init() 15 u64 addr = ring->dma; in xgene_enet_ring_init() 17 if (xgene_enet_ring_owner(ring->id) == RING_OWNER_CPU) { in xgene_enet_ring_init() 18 ring_cfg[0] |= SET_VAL(X2_INTLINE, ring->id & RING_BUFNUM_MASK); in xgene_enet_ring_init() 27 ring_cfg[3] |= SET_VAL(RINGSIZE, ring->cfgsize) in xgene_enet_ring_init() 34 static void xgene_enet_ring_set_type(struct xgene_enet_desc_ring *ring) in xgene_enet_ring_set_type() argument 36 u32 *ring_cfg = ring->state; in xgene_enet_ring_set_type() 40 is_bufpool = xgene_enet_is_bufpool(ring->id); in xgene_enet_ring_set_type() 47 static void xgene_enet_ring_set_recombbuf(struct xgene_enet_desc_ring *ring) in xgene_enet_ring_set_recombbuf() argument 55 xgene_enet_ring_wr32(struct xgene_enet_desc_ring * ring,u32 offset,u32 data) xgene_enet_ring_wr32() argument 63 xgene_enet_write_ring_state(struct xgene_enet_desc_ring * ring) xgene_enet_write_ring_state() argument 75 xgene_enet_clr_ring_state(struct xgene_enet_desc_ring * ring) xgene_enet_clr_ring_state() argument 81 xgene_enet_set_ring_state(struct xgene_enet_desc_ring * ring) xgene_enet_set_ring_state() argument 95 xgene_enet_set_ring_id(struct xgene_enet_desc_ring * ring) xgene_enet_set_ring_id() argument 118 xgene_enet_clr_desc_ring_id(struct xgene_enet_desc_ring * ring) xgene_enet_clr_desc_ring_id() argument 128 xgene_enet_setup_ring(struct xgene_enet_desc_ring * ring) xgene_enet_setup_ring() argument 152 xgene_enet_clear_ring(struct xgene_enet_desc_ring * ring) xgene_enet_clear_ring() argument 158 xgene_enet_wr_cmd(struct xgene_enet_desc_ring * ring,int count) xgene_enet_wr_cmd() argument 171 xgene_enet_ring_len(struct xgene_enet_desc_ring * ring) xgene_enet_ring_len() argument 182 xgene_enet_setup_coalescing(struct xgene_enet_desc_ring * ring) xgene_enet_setup_coalescing() argument [all...] |
/linux/net/rds/ |
H A D | ib_ring.c | 66 void rds_ib_ring_init(struct rds_ib_work_ring *ring, u32 nr) in rds_ib_ring_init() argument 68 memset(ring, 0, sizeof(*ring)); in rds_ib_ring_init() 69 ring->w_nr = nr; in rds_ib_ring_init() 70 rdsdebug("ring %p nr %u\n", ring, ring->w_nr); in rds_ib_ring_init() 73 static inline u32 __rds_ib_ring_used(struct rds_ib_work_ring *ring) in __rds_ib_ring_used() argument 78 diff = ring->w_alloc_ctr - (u32) atomic_read(&ring in __rds_ib_ring_used() 84 rds_ib_ring_resize(struct rds_ib_work_ring * ring,u32 nr) rds_ib_ring_resize() argument 92 __rds_ib_ring_empty(struct rds_ib_work_ring * ring) __rds_ib_ring_empty() argument 97 rds_ib_ring_alloc(struct rds_ib_work_ring * ring,u32 val,u32 * pos) rds_ib_ring_alloc() argument 117 rds_ib_ring_free(struct rds_ib_work_ring * ring,u32 val) rds_ib_ring_free() argument 127 rds_ib_ring_unalloc(struct rds_ib_work_ring * ring,u32 val) rds_ib_ring_unalloc() argument 133 rds_ib_ring_empty(struct rds_ib_work_ring * ring) rds_ib_ring_empty() argument 138 rds_ib_ring_low(struct rds_ib_work_ring * ring) rds_ib_ring_low() argument 147 rds_ib_ring_oldest(struct rds_ib_work_ring * ring) rds_ib_ring_oldest() argument 156 rds_ib_ring_completed(struct rds_ib_work_ring * ring,u32 wr_id,u32 oldest) rds_ib_ring_completed() argument [all...] |
/linux/drivers/net/ethernet/marvell/octeon_ep_vf/ |
H A D | octep_vf_regs_cnxk.h | 34 #define CNXK_VF_SDP_R_ERR_TYPE(ring) \ argument 35 (CNXK_VF_SDP_R_ERR_TYPE_START + ((ring) * CNXK_VF_RING_OFFSET)) 37 #define CNXK_VF_SDP_R_IN_CONTROL(ring) \ argument 38 (CNXK_VF_SDP_R_IN_CONTROL_START + ((ring) * CNXK_VF_RING_OFFSET)) 40 #define CNXK_VF_SDP_R_IN_ENABLE(ring) \ argument 41 (CNXK_VF_SDP_R_IN_ENABLE_START + ((ring) * CNXK_VF_RING_OFFSET)) 43 #define CNXK_VF_SDP_R_IN_INSTR_BADDR(ring) \ argument 44 (CNXK_VF_SDP_R_IN_INSTR_BADDR_START + ((ring) * CNXK_VF_RING_OFFSET)) 46 #define CNXK_VF_SDP_R_IN_INSTR_RSIZE(ring) \ argument 47 (CNXK_VF_SDP_R_IN_INSTR_RSIZE_START + ((ring) * CNXK_VF_RING_OFFSE 49 CNXK_VF_SDP_R_IN_INSTR_DBELL(ring) global() argument 52 CNXK_VF_SDP_R_IN_CNTS(ring) global() argument 55 CNXK_VF_SDP_R_IN_INT_LEVELS(ring) global() argument 58 CNXK_VF_SDP_R_IN_PKT_CNT(ring) global() argument 61 CNXK_VF_SDP_R_IN_BYTE_CNT(ring) global() argument 97 CNXK_VF_SDP_R_OUT_CONTROL(ring) global() argument 100 CNXK_VF_SDP_R_OUT_ENABLE(ring) global() argument 103 CNXK_VF_SDP_R_OUT_SLIST_BADDR(ring) global() argument 106 CNXK_VF_SDP_R_OUT_SLIST_RSIZE(ring) global() argument 109 CNXK_VF_SDP_R_OUT_SLIST_DBELL(ring) global() argument 112 CNXK_VF_SDP_R_OUT_WMARK(ring) global() argument 115 CNXK_VF_SDP_R_OUT_CNTS(ring) global() argument 118 CNXK_VF_SDP_R_OUT_INT_LEVELS(ring) global() argument 121 CNXK_VF_SDP_R_OUT_PKT_CNT(ring) global() argument 124 CNXK_VF_SDP_R_OUT_BYTE_CNT(ring) global() argument 154 CNXK_VF_SDP_R_MBOX_PF_VF_DATA(ring) global() argument 157 CNXK_VF_SDP_R_MBOX_PF_VF_INT(ring) global() argument 160 CNXK_VF_SDP_R_MBOX_VF_PF_DATA(ring) global() argument [all...] |
H A D | octep_vf_regs_cn9k.h | 33 #define CN93_VF_SDP_R_IN_CONTROL(ring) \ argument 34 (CN93_VF_SDP_R_IN_CONTROL_START + ((ring) * CN93_VF_RING_OFFSET)) 36 #define CN93_VF_SDP_R_IN_ENABLE(ring) \ argument 37 (CN93_VF_SDP_R_IN_ENABLE_START + ((ring) * CN93_VF_RING_OFFSET)) 39 #define CN93_VF_SDP_R_IN_INSTR_BADDR(ring) \ argument 40 (CN93_VF_SDP_R_IN_INSTR_BADDR_START + ((ring) * CN93_VF_RING_OFFSET)) 42 #define CN93_VF_SDP_R_IN_INSTR_RSIZE(ring) \ argument 43 (CN93_VF_SDP_R_IN_INSTR_RSIZE_START + ((ring) * CN93_VF_RING_OFFSET)) 45 #define CN93_VF_SDP_R_IN_INSTR_DBELL(ring) \ argument 46 (CN93_VF_SDP_R_IN_INSTR_DBELL_START + ((ring) * CN93_VF_RING_OFFSE 48 CN93_VF_SDP_R_IN_CNTS(ring) global() argument 51 CN93_VF_SDP_R_IN_INT_LEVELS(ring) global() argument 54 CN93_VF_SDP_R_IN_PKT_CNT(ring) global() argument 57 CN93_VF_SDP_R_IN_BYTE_CNT(ring) global() argument 92 CN93_VF_SDP_R_OUT_CONTROL(ring) global() argument 95 CN93_VF_SDP_R_OUT_ENABLE(ring) global() argument 98 CN93_VF_SDP_R_OUT_SLIST_BADDR(ring) global() argument 101 CN93_VF_SDP_R_OUT_SLIST_RSIZE(ring) global() argument 104 CN93_VF_SDP_R_OUT_SLIST_DBELL(ring) global() argument 107 CN93_VF_SDP_R_OUT_CNTS(ring) global() argument 110 CN93_VF_SDP_R_OUT_INT_LEVELS(ring) global() argument 113 CN93_VF_SDP_R_OUT_PKT_CNT(ring) global() argument 116 CN93_VF_SDP_R_OUT_BYTE_CNT(ring) global() argument 146 CN93_VF_SDP_R_MBOX_PF_VF_DATA(ring) global() argument 149 CN93_VF_SDP_R_MBOX_PF_VF_INT(ring) global() argument 152 CN93_VF_SDP_R_MBOX_VF_PF_DATA(ring) global() argument [all...] |
/linux/drivers/net/ethernet/marvell/octeon_ep/ |
H A D | octep_regs_cnxk_pf.h | 73 #define CNXK_SDP_R_IN_CONTROL(ring) \ argument 74 (CNXK_SDP_R_IN_CONTROL_START + ((ring) * CNXK_RING_OFFSET)) 76 #define CNXK_SDP_R_IN_ENABLE(ring) \ argument 77 (CNXK_SDP_R_IN_ENABLE_START + ((ring) * CNXK_RING_OFFSET)) 79 #define CNXK_SDP_R_IN_INSTR_BADDR(ring) \ argument 80 (CNXK_SDP_R_IN_INSTR_BADDR_START + ((ring) * CNXK_RING_OFFSET)) 82 #define CNXK_SDP_R_IN_INSTR_RSIZE(ring) \ argument 83 (CNXK_SDP_R_IN_INSTR_RSIZE_START + ((ring) * CNXK_RING_OFFSET)) 85 #define CNXK_SDP_R_IN_INSTR_DBELL(ring) \ argument 86 (CNXK_SDP_R_IN_INSTR_DBELL_START + ((ring) * CNXK_RING_OFFSE 88 CNXK_SDP_R_IN_CNTS(ring) global() argument 91 CNXK_SDP_R_IN_INT_LEVELS(ring) global() argument 94 CNXK_SDP_R_IN_PKT_CNT(ring) global() argument 97 CNXK_SDP_R_IN_BYTE_CNT(ring) global() argument 131 CNXK_SDP_R_OUT_CONTROL(ring) global() argument 134 CNXK_SDP_R_OUT_ENABLE(ring) global() argument 137 CNXK_SDP_R_OUT_SLIST_BADDR(ring) global() argument 140 CNXK_SDP_R_OUT_SLIST_RSIZE(ring) global() argument 143 CNXK_SDP_R_OUT_SLIST_DBELL(ring) global() argument 146 CNXK_SDP_R_OUT_WMARK(ring) global() argument 149 CNXK_SDP_R_OUT_CNTS(ring) global() argument 152 CNXK_SDP_R_OUT_INT_LEVELS(ring) global() argument 155 CNXK_SDP_R_OUT_PKT_CNT(ring) global() argument 158 CNXK_SDP_R_OUT_BYTE_CNT(ring) global() argument 190 CNXK_SDP_R_IN_INT_MDRT_CTL0(ring) global() argument 193 CNXK_SDP_R_IN_INT_MDRT_CTL1(ring) global() argument 196 CNXK_SDP_R_IN_INT_MDRT_DBG(ring) global() argument 199 CNXK_SDP_R_OUT_INT_MDRT_CTL0(ring) global() argument 202 CNXK_SDP_R_OUT_INT_MDRT_CTL1(ring) global() argument 205 CNXK_SDP_R_OUT_INT_MDRT_DBG(ring) global() argument 208 CNXK_SDP_R_MBOX_ISM(ring) global() argument 211 CNXK_SDP_R_OUT_CNTS_ISM(ring) global() argument 214 CNXK_SDP_R_IN_CNTS_ISM(ring) global() argument 232 CNXK_SDP_R_MBOX_PF_VF_DATA(ring) global() argument 235 CNXK_SDP_R_MBOX_PF_VF_INT(ring) global() argument 238 CNXK_SDP_R_MBOX_VF_PF_DATA(ring) global() argument 241 CNXK_SDP_MBOX_VF_PF_DATA(ring) global() argument 244 CNXK_SDP_MBOX_PF_VF_DATA(ring) global() argument 250 CNXK_SDP_R_ERR_TYPE(ring) global() argument 257 CNXK_SDP_R_MBOX_ISM(ring) global() argument 260 CNXK_SDP_R_OUT_CNTS_ISM(ring) global() argument 263 CNXK_SDP_R_IN_CNTS_ISM(ring) global() argument 379 CNXK_SDP_EPVF_RING(ring) global() argument 381 CNXK_SDP_IN_RING_TB_MAP(ring) global() argument 383 CNXK_SDP_IN_RATE_LIMIT(ring) global() argument [all...] |
H A D | octep_regs_cn9k_pf.h | 72 #define CN93_SDP_R_IN_CONTROL(ring) \ argument 73 (CN93_SDP_R_IN_CONTROL_START + ((ring) * CN93_RING_OFFSET)) 75 #define CN93_SDP_R_IN_ENABLE(ring) \ argument 76 (CN93_SDP_R_IN_ENABLE_START + ((ring) * CN93_RING_OFFSET)) 78 #define CN93_SDP_R_IN_INSTR_BADDR(ring) \ argument 79 (CN93_SDP_R_IN_INSTR_BADDR_START + ((ring) * CN93_RING_OFFSET)) 81 #define CN93_SDP_R_IN_INSTR_RSIZE(ring) \ argument 82 (CN93_SDP_R_IN_INSTR_RSIZE_START + ((ring) * CN93_RING_OFFSET)) 84 #define CN93_SDP_R_IN_INSTR_DBELL(ring) \ argument 85 (CN93_SDP_R_IN_INSTR_DBELL_START + ((ring) * CN93_RING_OFFSE 87 CN93_SDP_R_IN_CNTS(ring) global() argument 90 CN93_SDP_R_IN_INT_LEVELS(ring) global() argument 93 CN93_SDP_R_IN_PKT_CNT(ring) global() argument 96 CN93_SDP_R_IN_BYTE_CNT(ring) global() argument 129 CN93_SDP_R_OUT_CONTROL(ring) global() argument 132 CN93_SDP_R_OUT_ENABLE(ring) global() argument 135 CN93_SDP_R_OUT_SLIST_BADDR(ring) global() argument 138 CN93_SDP_R_OUT_SLIST_RSIZE(ring) global() argument 141 CN93_SDP_R_OUT_SLIST_DBELL(ring) global() argument 144 CN93_SDP_R_OUT_CNTS(ring) global() argument 147 CN93_SDP_R_OUT_INT_LEVELS(ring) global() argument 150 CN93_SDP_R_OUT_PKT_CNT(ring) global() argument 153 CN93_SDP_R_OUT_BYTE_CNT(ring) global() argument 181 CN93_SDP_R_IN_INT_MDRT_CTL0(ring) global() argument 184 CN93_SDP_R_IN_INT_MDRT_CTL1(ring) global() argument 187 CN93_SDP_R_IN_INT_MDRT_DBG(ring) global() argument 190 CN93_SDP_R_OUT_INT_MDRT_CTL0(ring) global() argument 193 CN93_SDP_R_OUT_INT_MDRT_CTL1(ring) global() argument 196 CN93_SDP_R_OUT_INT_MDRT_DBG(ring) global() argument 214 CN93_SDP_R_MBOX_PF_VF_DATA(ring) global() argument 217 CN93_SDP_R_MBOX_PF_VF_INT(ring) global() argument 220 CN93_SDP_R_MBOX_VF_PF_DATA(ring) global() argument 223 CN93_SDP_MBOX_VF_PF_DATA(ring) global() argument 226 CN93_SDP_MBOX_PF_VF_DATA(ring) global() argument 232 CN93_SDP_R_ERR_TYPE(ring) global() argument 239 CN93_SDP_R_MBOX_ISM(ring) global() argument 242 CN93_SDP_R_OUT_CNTS_ISM(ring) global() argument 245 CN93_SDP_R_IN_CNTS_ISM(ring) global() argument 361 CN93_SDP_EPVF_RING(ring) global() argument 363 CN93_SDP_IN_RING_TB_MAP(ring) global() argument 365 CN93_SDP_IN_RATE_LIMIT(ring) global() argument [all...] |
/linux/drivers/gpu/drm/i915/gt/ |
H A D | intel_ring.c | 19 unsigned int intel_ring_update_space(struct intel_ring *ring) in intel_ring_update_space() argument 23 space = __intel_ring_space(ring->head, ring->emit, ring->size); in intel_ring_update_space() 25 ring->space = space; in intel_ring_update_space() 29 void __intel_ring_pin(struct intel_ring *ring) in __intel_ring_pin() argument 31 GEM_BUG_ON(!atomic_read(&ring->pin_count)); in __intel_ring_pin() 32 atomic_inc(&ring->pin_count); in __intel_ring_pin() 35 int intel_ring_pin(struct intel_ring *ring, struct i915_gem_ww_ctx *ww) in intel_ring_pin() argument 37 struct i915_vma *vma = ring in intel_ring_pin() 85 intel_ring_reset(struct intel_ring * ring,u32 tail) intel_ring_reset() argument 94 intel_ring_unpin(struct intel_ring * ring) intel_ring_unpin() argument 149 struct intel_ring *ring; intel_engine_create_ring() local 186 struct intel_ring *ring = container_of(ref, typeof(*ring), ref); intel_ring_free() local 193 wait_for_space(struct intel_ring * ring,struct intel_timeline * tl,unsigned int bytes) wait_for_space() argument 232 struct intel_ring *ring = rq->ring; intel_ring_begin() local [all...] |
H A D | selftest_ring.c | 8 struct intel_ring *ring; in mock_ring() local 10 ring = kzalloc(sizeof(*ring) + sz, GFP_KERNEL); in mock_ring() 11 if (!ring) in mock_ring() 14 kref_init(&ring->ref); in mock_ring() 15 ring->size = sz; in mock_ring() 16 ring->wrap = BITS_PER_TYPE(ring->size) - ilog2(sz); in mock_ring() 17 ring->effective_size = sz; in mock_ring() 18 ring in mock_ring() 26 mock_ring_free(struct intel_ring * ring) mock_ring_free() argument 31 check_ring_direction(struct intel_ring * ring,u32 next,u32 prev,int expected) check_ring_direction() argument 52 check_ring_step(struct intel_ring * ring,u32 x,u32 step) check_ring_step() argument 65 check_ring_offset(struct intel_ring * ring,u32 x,u32 step) check_ring_offset() argument 78 struct intel_ring *ring; igt_ring_direction() local [all...] |
/linux/drivers/crypto/inside-secure/ |
H A D | safexcel_ring.c | 21 /* Actual command descriptor ring */ in safexcel_init_ring_descriptors() 32 /* Command descriptor shadow ring for storing additional token data */ in safexcel_init_ring_descriptors() 78 struct safexcel_desc_ring *ring, in safexcel_ring_next_cwptr() argument 82 void *ptr = ring->write; in safexcel_ring_next_cwptr() 85 *atoken = ring->shwrite; in safexcel_ring_next_cwptr() 87 if ((ring->write == ring->read - ring->offset) || in safexcel_ring_next_cwptr() 88 (ring->read == ring in safexcel_ring_next_cwptr() 103 safexcel_ring_next_rwptr(struct safexcel_crypto_priv * priv,struct safexcel_desc_ring * ring,struct result_data_desc ** rtoken) safexcel_ring_next_rwptr() argument 124 safexcel_ring_next_rptr(struct safexcel_crypto_priv * priv,struct safexcel_desc_ring * ring) safexcel_ring_next_rptr() argument 140 safexcel_ring_curr_rptr(struct safexcel_crypto_priv * priv,int ring) safexcel_ring_curr_rptr() argument 148 safexcel_ring_first_rdr_index(struct safexcel_crypto_priv * priv,int ring) safexcel_ring_first_rdr_index() argument 156 safexcel_ring_rdr_rdesc_index(struct safexcel_crypto_priv * priv,int ring,struct safexcel_result_desc * rdesc) safexcel_ring_rdr_rdesc_index() argument 165 safexcel_ring_rollback_wptr(struct safexcel_crypto_priv * priv,struct safexcel_desc_ring * ring) safexcel_ring_rollback_wptr() argument [all...] |
/linux/tools/testing/selftests/net/ |
H A D | psock_tpacket.c | 66 struct ring { struct 71 void (*walk)(int sock, struct ring *ring); argument 220 static void walk_v1_v2_rx(int sock, struct ring *ring) in walk_v1_v2_rx() argument 227 bug_on(ring->type != PACKET_RX_RING); in walk_v1_v2_rx() 239 while (__v1_v2_rx_kernel_ready(ring->rd[frame_num].iov_base, in walk_v1_v2_rx() 240 ring->version)) { in walk_v1_v2_rx() 241 ppd.raw = ring->rd[frame_num].iov_base; in walk_v1_v2_rx() 243 switch (ring in walk_v1_v2_rx() 72 versionring global() argument 73 __anon368def29010aring global() argument 354 get_next_frame(struct ring * ring,int n) get_next_frame() argument 369 walk_tx(int sock,struct ring * ring) walk_tx() argument 500 walk_v1_v2(int sock,struct ring * ring) walk_v1_v2() argument 581 walk_v3_rx(int sock,struct ring * ring) walk_v3_rx() argument 622 walk_v3(int sock,struct ring * ring) walk_v3() argument 630 __v1_v2_fill(struct ring * ring,unsigned int blocks) __v1_v2_fill() argument 646 __v3_fill(struct ring * ring,unsigned int blocks,int type) __v3_fill() argument 667 setup_ring(int sock,struct ring * ring,int version,int type) setup_ring() argument 708 mmap_ring(int sock,struct ring * ring) mmap_ring() argument 726 bind_ring(int sock,struct ring * ring) bind_ring() argument 746 walk_ring(int sock,struct ring * ring) walk_ring() argument 751 unmap_ring(int sock,struct ring * ring) unmap_ring() argument 805 struct ring ring; test_tpacket() local [all...] |
/linux/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_transport.c | 40 static int adf_reserve_ring(struct adf_etr_bank_data *bank, u32 ring) in adf_reserve_ring() argument 43 if (bank->ring_mask & (1 << ring)) { in adf_reserve_ring() 47 bank->ring_mask |= (1 << ring); in adf_reserve_ring() 52 static void adf_unreserve_ring(struct adf_etr_bank_data *bank, u32 ring) in adf_unreserve_ring() argument 55 bank->ring_mask &= ~(1 << ring); in adf_unreserve_ring() 59 static void adf_enable_ring_irq(struct adf_etr_bank_data *bank, u32 ring) in adf_enable_ring_irq() argument 64 bank->irq_mask |= (1 << ring); in adf_enable_ring_irq() 72 static void adf_disable_ring_irq(struct adf_etr_bank_data *bank, u32 ring) in adf_disable_ring_irq() argument 77 bank->irq_mask &= ~(1 << ring); in adf_disable_ring_irq() 83 bool adf_ring_nearly_full(struct adf_etr_ring_data *ring) in adf_ring_nearly_full() argument 88 adf_send_message(struct adf_etr_ring_data * ring,u32 * msg) adf_send_message() argument 112 adf_handle_response(struct adf_etr_ring_data * ring) adf_handle_response() argument 136 adf_configure_tx_ring(struct adf_etr_ring_data * ring) adf_configure_tx_ring() argument 147 adf_configure_rx_ring(struct adf_etr_ring_data * ring) adf_configure_rx_ring() argument 160 adf_init_ring(struct adf_etr_ring_data * ring) adf_init_ring() argument 203 adf_cleanup_ring(struct adf_etr_ring_data * ring) adf_cleanup_ring() argument 226 struct adf_etr_ring_data *ring; adf_create_ring() local 304 adf_remove_ring(struct adf_etr_ring_data * ring) adf_remove_ring() argument 392 struct adf_etr_ring_data *ring; adf_init_bank() local 530 struct adf_etr_ring_data *ring = &bank->rings[i]; cleanup_bank() local [all...] |