| /linux/drivers/firmware/xilinx/ |
| H A D | zynqmp-crypto.c | 28 ret = zynqmp_pm_invoke_fn(PM_SECURE_AES, ret_payload, 2, upper_32_bits(address), in zynqmp_pm_aes_engine() 55 u32 upper_addr = upper_32_bits(address); in zynqmp_pm_sha_hash() 110 upper_32_bits(keyaddr)); in versal_pm_aes_key_write() 141 upper_32_bits(hw_req)); in versal_pm_aes_op_init() 158 upper_32_bits(aad_addr), in versal_pm_aes_update_aad() 175 upper_32_bits(in_params), in versal_pm_aes_enc_update() 177 upper_32_bits(in_addr)); in versal_pm_aes_enc_update() 191 upper_32_bits(gcm_addr)); in versal_pm_aes_enc_final() 207 upper_32_bits(in_params), in versal_pm_aes_dec_update() 209 upper_32_bits(in_addr)); in versal_pm_aes_dec_update() [all …]
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | si_dma.c | 82 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in si_dma_vm_copy_pages() 83 ib->ptr[ib->length_dw++] = upper_32_bits(src) & 0xff; in si_dma_vm_copy_pages() 121 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in si_dma_vm_write_pages() 133 ib->ptr[ib->length_dw++] = upper_32_bits(value); in si_dma_vm_write_pages() 173 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in si_dma_vm_set_pages() 177 ib->ptr[ib->length_dw++] = upper_32_bits(value); in si_dma_vm_set_pages() 265 radeon_ring_write(ring, upper_32_bits(dst_offset) & 0xff); in si_copy_dma() 266 radeon_ring_write(ring, upper_32_bits(src_offset) & 0xff); in si_copy_dma()
|
| H A D | ni_dma.c | 134 radeon_ring_write(ring, upper_32_bits(ring->next_rptr_gpu_addr) & 0xff); in cayman_dma_ring_ib_execute() 145 radeon_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in cayman_dma_ring_ib_execute() 222 upper_32_bits(rdev->wb.gpu_addr + wb_offset) & 0xFF); in cayman_dma_resume() 330 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in cayman_dma_vm_copy_pages() 331 ib->ptr[ib->length_dw++] = upper_32_bits(src) & 0xff; in cayman_dma_vm_copy_pages() 370 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in cayman_dma_vm_write_pages() 382 ib->ptr[ib->length_dw++] = upper_32_bits(value); in cayman_dma_vm_write_pages() 422 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in cayman_dma_vm_set_pages() 426 ib->ptr[ib->length_dw++] = upper_32_bits(value); in cayman_dma_vm_set_pages()
|
| H A D | r600_dma.c | 143 upper_32_bits(rdev->wb.gpu_addr + R600_WB_DMA_RPTR_OFFSET) & 0xFF); in r600_dma_resume() 255 radeon_ring_write(ring, upper_32_bits(gpu_addr) & 0xff); in r600_dma_ring_test() 295 radeon_ring_write(ring, (upper_32_bits(addr) & 0xff)); in r600_dma_fence_ring_emit() 322 radeon_ring_write(ring, upper_32_bits(addr) & 0xff); in r600_dma_semaphore_ring_emit() 360 ib.ptr[2] = upper_32_bits(gpu_addr) & 0xff; in r600_dma_ib_test() 415 radeon_ring_write(ring, upper_32_bits(ring->next_rptr_gpu_addr) & 0xff); in r600_dma_ring_ib_execute() 426 radeon_ring_write(ring, (ib->length_dw << 16) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in r600_dma_ring_ib_execute() 478 radeon_ring_write(ring, (((upper_32_bits(dst_offset) & 0xff) << 16) | in r600_copy_dma() 479 (upper_32_bits(src_offset) & 0xff))); in r600_copy_dma()
|
| H A D | evergreen_dma.c | 48 radeon_ring_write(ring, (upper_32_bits(addr) & 0xff)); in evergreen_dma_fence_ring_emit() 78 radeon_ring_write(ring, upper_32_bits(ring->next_rptr_gpu_addr) & 0xff); in evergreen_dma_ring_ib_execute() 89 radeon_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in evergreen_dma_ring_ib_execute() 142 radeon_ring_write(ring, upper_32_bits(dst_offset) & 0xff); in evergreen_copy_dma() 143 radeon_ring_write(ring, upper_32_bits(src_offset) & 0xff); in evergreen_copy_dma()
|
| H A D | cik_sdma.c | 145 radeon_ring_write(ring, upper_32_bits(ring->next_rptr_gpu_addr)); in cik_sdma_ring_ib_execute() 155 radeon_ring_write(ring, upper_32_bits(ib->gpu_addr)); in cik_sdma_ring_ib_execute() 208 radeon_ring_write(ring, upper_32_bits(addr)); in cik_sdma_fence_ring_emit() 237 radeon_ring_write(ring, upper_32_bits(addr)); in cik_sdma_semaphore_ring_emit() 400 upper_32_bits(rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF); in cik_sdma_gfx_resume() 614 radeon_ring_write(ring, upper_32_bits(src_offset)); in cik_copy_dma() 616 radeon_ring_write(ring, upper_32_bits(dst_offset)); in cik_copy_dma() 670 radeon_ring_write(ring, upper_32_bits(gpu_addr)); in cik_sdma_ring_test() 728 ib.ptr[2] = upper_32_bits(gpu_addr); in cik_sdma_ib_test() 817 ib->ptr[ib->length_dw++] = upper_32_bits(src); in cik_sdma_vm_copy_pages() [all …]
|
| /linux/drivers/pci/controller/mobiveil/ |
| H A D | pcie-mobiveil.c | 154 mobiveil_csr_writel(pcie, upper_32_bits(size64), in program_ib_windows() 159 mobiveil_csr_writel(pcie, upper_32_bits(cpu_addr), in program_ib_windows() 164 mobiveil_csr_writel(pcie, upper_32_bits(pci_addr), in program_ib_windows() 195 mobiveil_csr_writel(pcie, upper_32_bits(size64), in program_ob_windows() 205 mobiveil_csr_writel(pcie, upper_32_bits(cpu_addr), in program_ob_windows() 210 mobiveil_csr_writel(pcie, upper_32_bits(pci_addr), in program_ob_windows()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | sdma_v6_0.c | 152 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v6_0_ring_init_cond_exec() 218 upper_32_bits(ring->wptr << 2)); in sdma_v6_0_ring_set_wptr() 232 upper_32_bits(ring->wptr << 2)); in sdma_v6_0_ring_set_wptr() 238 upper_32_bits(ring->wptr << 2)); in sdma_v6_0_ring_set_wptr() 287 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v6_0_ring_emit_ib() 290 amdgpu_ring_write(ring, upper_32_bits(csa_mc_addr)); in sdma_v6_0_ring_emit_ib() 366 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v6_0_ring_emit_fence() 377 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v6_0_ring_emit_fence() 378 amdgpu_ring_write(ring, upper_32_bits(seq)); in sdma_v6_0_ring_emit_fence() 512 …WREG32_SOC15_IP(GC, sdma_v6_0_get_reg_offset(adev, i, regSDMA0_QUEUE0_RB_RPTR_HI), upper_32_bits(r… in sdma_v6_0_gfx_resume_instance() [all …]
|
| H A D | sdma_v5_2.c | 150 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v5_2_ring_init_cond_exec() 223 upper_32_bits(ring->wptr << 2)); in sdma_v5_2_ring_set_wptr() 238 upper_32_bits(ring->wptr << 2)); in sdma_v5_2_ring_set_wptr() 247 upper_32_bits(ring->wptr << 2)); in sdma_v5_2_ring_set_wptr() 251 upper_32_bits(ring->wptr << 2)); in sdma_v5_2_ring_set_wptr() 300 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v5_2_ring_emit_ib() 303 amdgpu_ring_write(ring, upper_32_bits(csa_mc_addr)); in sdma_v5_2_ring_emit_ib() 383 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v5_2_ring_emit_fence() 394 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v5_2_ring_emit_fence() 395 amdgpu_ring_write(ring, upper_32_bits(seq)); in sdma_v5_2_ring_emit_fence() [all …]
|
| H A D | sdma_v2_4.c | 261 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v2_4_ring_emit_ib() 314 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v2_4_ring_emit_fence() 322 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v2_4_ring_emit_fence() 323 amdgpu_ring_write(ring, upper_32_bits(seq)); in sdma_v2_4_ring_emit_fence() 445 upper_32_bits(ring->rptr_gpu_addr) & 0xFFFFFFFF); in sdma_v2_4_gfx_resume() 555 amdgpu_ring_write(ring, upper_32_bits(gpu_addr)); in sdma_v2_4_ring_test_ring() 610 ib.ptr[2] = upper_32_bits(gpu_addr); in sdma_v2_4_ring_test_ib() 664 ib->ptr[ib->length_dw++] = upper_32_bits(src); in sdma_v2_4_vm_copy_pte() 666 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v2_4_vm_copy_pte() 689 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v2_4_vm_write_pte() [all …]
|
| H A D | sdma_v5_0.c | 310 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v5_0_ring_init_cond_exec() 383 upper_32_bits(ring->wptr << 2)); in sdma_v5_0_ring_set_wptr() 397 upper_32_bits(ring->wptr << 2)); in sdma_v5_0_ring_set_wptr() 403 upper_32_bits(ring->wptr << 2)); in sdma_v5_0_ring_set_wptr() 452 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v5_0_ring_emit_ib() 455 amdgpu_ring_write(ring, upper_32_bits(csa_mc_addr)); in sdma_v5_0_ring_emit_ib() 533 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v5_0_ring_emit_fence() 544 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v5_0_ring_emit_fence() 545 amdgpu_ring_write(ring, upper_32_bits(seq)); in sdma_v5_0_ring_emit_fence() 718 …WREG32_SOC15_IP(GC, sdma_v5_0_get_reg_offset(adev, i, mmSDMA0_GFX_RB_RPTR_HI), upper_32_bits(ring-… in sdma_v5_0_gfx_resume_instance() [all …]
|
| H A D | cik_sdma.c | 235 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xffffffff); in cik_sdma_ring_emit_ib() 285 amdgpu_ring_write(ring, upper_32_bits(addr)); in cik_sdma_ring_emit_fence() 293 amdgpu_ring_write(ring, upper_32_bits(addr)); in cik_sdma_ring_emit_fence() 294 amdgpu_ring_write(ring, upper_32_bits(seq)); in cik_sdma_ring_emit_fence() 470 upper_32_bits(ring->rptr_gpu_addr) & 0xFFFFFFFF); in cik_sdma_gfx_resume() 623 amdgpu_ring_write(ring, upper_32_bits(gpu_addr)); in cik_sdma_ring_test_ring() 678 ib.ptr[2] = upper_32_bits(gpu_addr); in cik_sdma_ring_test_ib() 728 ib->ptr[ib->length_dw++] = upper_32_bits(src); in cik_sdma_vm_copy_pte() 730 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in cik_sdma_vm_copy_pte() 753 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in cik_sdma_vm_write_pte() [all …]
|
| H A D | sdma_v3_0.c | 437 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v3_0_ring_emit_ib() 490 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v3_0_ring_emit_fence() 498 amdgpu_ring_write(ring, upper_32_bits(addr)); in sdma_v3_0_ring_emit_fence() 499 amdgpu_ring_write(ring, upper_32_bits(seq)); in sdma_v3_0_ring_emit_fence() 686 upper_32_bits(ring->rptr_gpu_addr) & 0xFFFFFFFF); in sdma_v3_0_gfx_resume() 712 upper_32_bits(wptr_gpu_addr)); in sdma_v3_0_gfx_resume() 829 amdgpu_ring_write(ring, upper_32_bits(gpu_addr)); in sdma_v3_0_ring_test_ring() 884 ib.ptr[2] = upper_32_bits(gpu_addr); in sdma_v3_0_ring_test_ib() 937 ib->ptr[ib->length_dw++] = upper_32_bits(src); in sdma_v3_0_vm_copy_pte() 939 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v3_0_vm_copy_pte() [all …]
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_mqd_manager_vi.c | 118 m->cp_mqd_base_addr_hi = upper_32_bits(addr); in init_mqd() 132 m->compute_tba_hi = upper_32_bits(q->tba_addr >> 8); in init_mqd() 134 m->compute_tma_hi = upper_32_bits(q->tma_addr >> 8); in init_mqd() 145 upper_32_bits(q->ctx_save_restore_area_address); in init_mqd() 186 m->cp_hqd_pq_base_hi = upper_32_bits((uint64_t)q->queue_address >> 8); in __update_mqd() 189 m->cp_hqd_pq_rptr_report_addr_hi = upper_32_bits((uint64_t)q->read_ptr); in __update_mqd() 191 m->cp_hqd_pq_wptr_poll_addr_hi = upper_32_bits((uint64_t)q->write_ptr); in __update_mqd() 218 upper_32_bits(q->eop_ring_buffer_address >> 8); in __update_mqd() 372 m->sdmax_rlcx_rb_base_hi = upper_32_bits(q->queue_address >> 8); in update_mqd_sdma() 374 m->sdmax_rlcx_rb_rptr_addr_hi = upper_32_bits((uint64_t)q->read_ptr); in update_mqd_sdma()
|
| H A D | kfd_mqd_manager_v12.c | 128 m->cp_mqd_base_addr_hi = upper_32_bits(addr); in init_mqd() 153 upper_32_bits(q->ctx_save_restore_area_address); in init_mqd() 194 m->cp_hqd_pq_base_hi = upper_32_bits((uint64_t)q->queue_address >> 8); in update_mqd() 197 m->cp_hqd_pq_rptr_report_addr_hi = upper_32_bits((uint64_t)q->read_ptr); in update_mqd() 199 m->cp_hqd_pq_wptr_poll_addr_hi = upper_32_bits((uint64_t)q->write_ptr); in update_mqd() 221 upper_32_bits(q->eop_ring_buffer_address >> 8); in update_mqd() 336 m->sdmax_rlcx_rb_base_hi = upper_32_bits(q->queue_address >> 8); in update_mqd_sdma() 338 m->sdmax_rlcx_rb_rptr_addr_hi = upper_32_bits((uint64_t)q->read_ptr); in update_mqd_sdma() 340 m->sdmax_rlcx_rb_wptr_poll_addr_hi = upper_32_bits((uint64_t)q->write_ptr); in update_mqd_sdma()
|
| H A D | kfd_mqd_manager_v12_1.c | 184 m->cp_mqd_base_addr_hi = upper_32_bits(addr); in init_mqd() 209 upper_32_bits(q->ctx_save_restore_area_address); in init_mqd() 251 m->cp_hqd_pq_base_hi = upper_32_bits((uint64_t)q->queue_address >> 8); in update_mqd() 264 upper_32_bits((q->queue_address + q->queue_size) >> 8); in update_mqd() 275 m->cp_hqd_pq_rptr_report_addr_hi = upper_32_bits((uint64_t)q->read_ptr); in update_mqd() 277 m->cp_hqd_pq_wptr_poll_addr_hi = upper_32_bits((uint64_t)q->write_ptr); in update_mqd() 299 upper_32_bits(q->eop_ring_buffer_address >> 8); in update_mqd() 411 m->sdmax_rlcx_rb_base_hi = upper_32_bits(q->queue_address >> 8); in update_mqd_sdma() 413 m->sdmax_rlcx_rb_rptr_addr_hi = upper_32_bits((uint64_t)q->read_ptr); in update_mqd_sdma() 415 m->sdmax_rlcx_rb_wptr_poll_addr_hi = upper_32_bits((uint64_t)q->write_ptr); in update_mqd_sdma() [all …]
|
| H A D | kfd_mqd_manager_v11.c | 149 m->cp_mqd_base_addr_hi = upper_32_bits(addr); in init_mqd() 178 upper_32_bits(q->ctx_save_restore_area_address); in init_mqd() 219 m->cp_hqd_pq_base_hi = upper_32_bits((uint64_t)q->queue_address >> 8); in update_mqd() 222 m->cp_hqd_pq_rptr_report_addr_hi = upper_32_bits((uint64_t)q->read_ptr); in update_mqd() 224 m->cp_hqd_pq_wptr_poll_addr_hi = upper_32_bits((uint64_t)q->write_ptr); in update_mqd() 246 upper_32_bits(q->eop_ring_buffer_address >> 8); in update_mqd() 423 m->sdmax_rlcx_rb_base_hi = upper_32_bits(q->queue_address >> 8); in update_mqd_sdma() 425 m->sdmax_rlcx_rb_rptr_addr_hi = upper_32_bits((uint64_t)q->read_ptr); in update_mqd_sdma() 427 m->sdmax_rlcx_rb_wptr_poll_addr_hi = upper_32_bits((uint64_t)q->write_ptr); in update_mqd_sdma()
|
| H A D | kfd_mqd_manager_v10.c | 116 m->cp_mqd_base_addr_hi = upper_32_bits(addr); in init_mqd() 138 upper_32_bits(q->ctx_save_restore_area_address); in init_mqd() 180 m->cp_hqd_pq_base_hi = upper_32_bits((uint64_t)q->queue_address >> 8); in update_mqd() 183 m->cp_hqd_pq_rptr_report_addr_hi = upper_32_bits((uint64_t)q->read_ptr); in update_mqd() 185 m->cp_hqd_pq_wptr_poll_addr_hi = upper_32_bits((uint64_t)q->write_ptr); in update_mqd() 207 upper_32_bits(q->eop_ring_buffer_address >> 8); in update_mqd() 378 m->sdmax_rlcx_rb_base_hi = upper_32_bits(q->queue_address >> 8); in update_mqd_sdma() 380 m->sdmax_rlcx_rb_rptr_addr_hi = upper_32_bits((uint64_t)q->read_ptr); in update_mqd_sdma()
|
| /linux/drivers/iio/test/ |
| H A D | iio-test-format.c | 212 values[1] = upper_32_bits(value); in iio_test_iio_format_value_integer_64() 218 values[1] = upper_32_bits(value); in iio_test_iio_format_value_integer_64() 224 values[1] = upper_32_bits(value); in iio_test_iio_format_value_integer_64() 230 values[1] = upper_32_bits(value); in iio_test_iio_format_value_integer_64() 236 values[1] = upper_32_bits(value); in iio_test_iio_format_value_integer_64() 242 values[1] = upper_32_bits(value); in iio_test_iio_format_value_integer_64() 248 values[1] = upper_32_bits(value); in iio_test_iio_format_value_integer_64()
|
| /linux/drivers/pci/controller/ |
| H A D | pci-xgene.c | 296 val = (val32 & 0x0000ffff) | (upper_32_bits(mask) << 16); in xgene_pcie_set_ib_mask() 300 val = (val32 & 0xffff0000) | (upper_32_bits(mask) >> 16); in xgene_pcie_set_ib_mask() 386 xgene_pcie_writel(port, offset + 0x04, upper_32_bits(cpu_addr)); in xgene_pcie_setup_ob_reg() 388 xgene_pcie_writel(port, offset + 0x0c, upper_32_bits(mask)); in xgene_pcie_setup_ob_reg() 390 xgene_pcie_writel(port, offset + 0x14, upper_32_bits(pci_addr)); in xgene_pcie_setup_ob_reg() 398 xgene_pcie_writel(port, CFGBARH, upper_32_bits(addr)); in xgene_pcie_setup_cfg_reg() 448 upper_32_bits(pim) | EN_COHERENCY); in xgene_pcie_setup_pims() 450 xgene_pcie_writel(port, pim_reg + 0x14, upper_32_bits(size)); in xgene_pcie_setup_pims() 507 writel(upper_32_bits(cpu_addr), bar_addr + 0x4); in xgene_pcie_setup_ib_reg() 517 xgene_pcie_writel(port, IBAR3L + 0x4, upper_32_bits(cpu_addr)); in xgene_pcie_setup_ib_reg() [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
| H A D | gm20b.c | 74 hdr.code_dma_base1 = upper_32_bits((addr + adjust) >> 8); in gm20b_pmu_acr_bld_patch() 77 hdr.data_dma_base1 = upper_32_bits((addr + adjust) >> 8); in gm20b_pmu_acr_bld_patch() 80 hdr.overlay_dma_base1 = upper_32_bits((addr + adjust) << 8); in gm20b_pmu_acr_bld_patch() 104 .code_dma_base1 = upper_32_bits(code), in gm20b_pmu_acr_bld_write() 105 .data_dma_base1 = upper_32_bits(data), in gm20b_pmu_acr_bld_write() 106 .overlay_dma_base1 = upper_32_bits(code), in gm20b_pmu_acr_bld_write()
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
| H A D | gv100.c | 48 nvkm_wo32(chan->inst, 0x00c, upper_32_bits(userd)); in gv100_chan_ramfc_write() 52 nvkm_wo32(chan->inst, 0x04c, upper_32_bits(offset) | (limit2 << 16)); in gv100_chan_ramfc_write() 102 nvkm_wo32(chan->inst, 0x214, upper_32_bits(addr)); in gv100_ectx_bind() 122 nvkm_wo32(chan->inst, 0x224, upper_32_bits(bar2)); in gv100_ectx_ce_bind() 188 nvkm_wo32(memory, offset + 0x4, upper_32_bits(user)); in gv100_runl_insert_chan() 190 nvkm_wo32(memory, offset + 0xc, upper_32_bits(inst)); in gv100_runl_insert_chan()
|
| /linux/drivers/net/ethernet/apm/xgene-v2/ |
| H A D | ring.c | 28 dma_h = upper_32_bits(next_dma); in xge_setup_desc() 40 xge_wr_csr(pdata, DMATXDESCH, upper_32_bits(dma_addr)); in xge_update_tx_desc_addr() 52 xge_wr_csr(pdata, DMARXDESCH, upper_32_bits(dma_addr)); in xge_update_rx_desc_addr()
|
| /linux/drivers/media/pci/pt3/ |
| H A D | pt3_dma.c | 54 iowrite32(upper_32_bits(adap->desc_buf[0].b_addr), in pt3_start_dma() 185 d->next_h = upper_32_bits(desc_addr); in pt3_alloc_dmabuf() 191 d->addr_h = upper_32_bits(data_addr); in pt3_alloc_dmabuf() 196 d->next_h = upper_32_bits(desc_addr); in pt3_alloc_dmabuf() 205 d->next_h = upper_32_bits(desc_addr); in pt3_alloc_dmabuf()
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
| H A D | gm20b.c | 42 hdr.code_dma_base1 = upper_32_bits((addr + adjust) >> 8); in gm20b_gr_acr_bld_patch() 45 hdr.data_dma_base1 = upper_32_bits((addr + adjust) >> 8); in gm20b_gr_acr_bld_patch() 66 .code_dma_base1 = upper_32_bits(code), in gm20b_gr_acr_bld_write() 67 .data_dma_base1 = upper_32_bits(data), in gm20b_gr_acr_bld_write()
|