Home
last modified time | relevance | path

Searched refs:dma_addr (Results 1 – 25 of 795) sorted by relevance

12345678910>>...32

/linux/drivers/net/ethernet/cisco/enic/
H A Denic_res.h34 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_ex() argument
45 (u64)dma_addr | VNIC_PADDR_TARGET, in enic_queue_wq_desc_ex()
55 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop, desc_skip_cnt, in enic_queue_wq_desc_ex()
60 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_cont() argument
63 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_cont()
69 dma_addr_t dma_addr, unsigned int len, int vlan_tag_insert, in enic_queue_wq_desc() argument
72 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc()
79 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_csum() argument
83 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_csum()
91 void *os_buf, dma_addr_t dma_addr, unsigne in enic_queue_wq_desc_csum_l4() argument
102 enic_queue_wq_desc_tso(struct vnic_wq * wq,void * os_buf,dma_addr_t dma_addr,unsigned int len,unsigned int mss,unsigned int hdr_len,int vlan_tag_insert,unsigned int vlan_tag,int eop,int loopback) enic_queue_wq_desc_tso() argument
114 enic_queue_rq_desc(struct vnic_rq * rq,void * os_buf,unsigned int os_buf_index,dma_addr_t dma_addr,unsigned int len) enic_queue_rq_desc() argument
[all...]
/linux/include/trace/events/
H A Ddma.h37 TP_PROTO(struct device *dev, phys_addr_t phys_addr, dma_addr_t dma_addr,
39 TP_ARGS(dev, phys_addr, dma_addr, size, dir, attrs),
44 __field(u64, dma_addr)
53 __entry->dma_addr = dma_addr;
59 TP_printk("%s dir=%s dma_addr=%llx size=%zu phys_addr=%llx attrs=%s",
62 __entry->dma_addr,
70 TP_PROTO(struct device *dev, phys_addr_t phys_addr, dma_addr_t dma_addr, \
72 TP_ARGS(dev, phys_addr, dma_addr, size, dir, attrs))
98 TP_printk("%s dir=%s dma_addr
[all...]
/linux/sound/soc/bcm/
H A Dbcm63xx-pcm-whistler.c20 dma_addr_t dma_addr; member
26 dma_addr_t dma_addr; member
153 dma_desc->dma_addr = runtime->dma_addr; in bcm63xx_pcm_prepare()
168 regmap_write(regmap_i2s, regaddr_descaddr, dma_desc->dma_addr); in bcm63xx_pcm_prepare()
181 prtd->dma_addr_next = substream->runtime->dma_addr; in bcm63xx_pcm_pointer()
184 prtd->dma_addr_next - substream->runtime->dma_addr); in bcm63xx_pcm_pointer()
274 dma_desc->dma_addr += in i2s_dma_isr()
278 if (dma_desc->dma_addr - runtime->dma_addr > in i2s_dma_isr()
[all...]
/linux/drivers/net/ethernet/apm/xgene-v2/
H A Dring.c24 next_dma = ring->dma_addr + (offset * XGENE_ENET_DESC_SIZE); in xge_setup_desc()
37 dma_addr_t dma_addr = ring->dma_addr; in xge_update_tx_desc_addr() local
39 xge_wr_csr(pdata, DMATXDESCL, dma_addr); in xge_update_tx_desc_addr()
40 xge_wr_csr(pdata, DMATXDESCH, upper_32_bits(dma_addr)); in xge_update_tx_desc_addr()
49 dma_addr_t dma_addr = ring->dma_addr; in xge_update_rx_desc_addr() local
51 xge_wr_csr(pdata, DMARXDESCL, dma_addr); in xge_update_rx_desc_addr()
52 xge_wr_csr(pdata, DMARXDESCH, upper_32_bits(dma_addr)); in xge_update_rx_desc_addr()
/linux/drivers/gpu/drm/
H A Ddrm_pagemap.c205 * @dma_addr: Array to store DMA addresses corresponding to mapped pages
212 * corresponding page, and stores the DMA address in the provided @dma_addr
218 dma_addr_t *dma_addr, in drm_pagemap_migrate_map_pages() argument
234 dma_addr[i] = dma_map_page(dev, page, 0, PAGE_SIZE, dir); in drm_pagemap_migrate_map_pages()
235 if (dma_mapping_error(dev, dma_addr[i])) in drm_pagemap_migrate_map_pages()
245 * @dma_addr: Array of DMA addresses corresponding to mapped pages
250 * Memory (SVM). It iterates over each DMA address provided in @dma_addr, checks
254 dma_addr_t *dma_addr, in drm_pagemap_migrate_unmap_pages() argument
261 if (!dma_addr[i] || dma_mapping_error(dev, dma_addr[ in drm_pagemap_migrate_unmap_pages()
317 dma_addr_t *dma_addr; drm_pagemap_migrate_to_devmem() local
512 dma_addr_t *dma_addr; drm_pagemap_evict_to_ram() local
615 dma_addr_t *dma_addr; __drm_pagemap_migrate_to_ram() local
[all...]
/linux/arch/mips/sgi-ip32/
H A Dip32-dma.c23 dma_addr_t dma_addr = paddr & RAM_OFFSET_MASK; in phys_to_dma() local
26 dma_addr += CRIME_HI_MEM_BASE; in phys_to_dma()
27 return dma_addr; in phys_to_dma()
30 phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dma_addr) in dma_to_phys() argument
32 phys_addr_t paddr = dma_addr & RAM_OFFSET_MASK; in dma_to_phys()
34 if (dma_addr >= 256*1024*1024) in dma_to_phys()
/linux/drivers/crypto/marvell/octeontx2/
H A Dotx2_cpt_reqmgr.h107 dma_addr_t dma_addr; member
197 if (req->out[i].dma_addr) in otx2_cpt_info_destroy()
199 req->out[i].dma_addr, in otx2_cpt_info_destroy()
205 if (req->in[i].dma_addr) in otx2_cpt_info_destroy()
207 req->in[i].dma_addr, in otx2_cpt_info_destroy()
231 list[i].dma_addr = dma_map_single(&pdev->dev, list[i].vptr, in setup_sgio_components()
234 if (unlikely(dma_mapping_error(&pdev->dev, list[i].dma_addr))) { in setup_sgio_components()
246 sg_ptr->ptr0 = cpu_to_be64(list[i * SG_COMPS_MAX + 0].dma_addr); in setup_sgio_components()
247 sg_ptr->ptr1 = cpu_to_be64(list[i * SG_COMPS_MAX + 1].dma_addr); in setup_sgio_components()
248 sg_ptr->ptr2 = cpu_to_be64(list[i * SG_COMPS_MAX + 2].dma_addr); in setup_sgio_components()
[all...]
/linux/drivers/net/ethernet/freescale/dpaa2/
H A Ddpaa2-eth-trace.h17 #define TR_BUF_FMT "[%s] vaddr=%p size=%zu dma_addr=%pad map_size=%zu bpid=%d"
116 dma_addr_t dma_addr,
122 TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid),
132 __field(dma_addr_t, dma_addr)
144 __entry->dma_addr = dma_addr;
157 &__entry->dma_addr,
167 dma_addr_t dma_addr,
171 TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid)
179 dma_addr_t dma_addr,
[all...]
/linux/drivers/gpu/drm/xe/
H A Dxe_res_cursor.h58 /** @dma_addr: Current element in a struct drm_pagemap_device_addr array */
59 const struct drm_pagemap_device_addr *dma_addr; member
64 * This may be different to @dma_addr.addr since elements in
95 cur->dma_addr = NULL; in xe_res_first()
173 const struct drm_pagemap_device_addr *addr = cur->dma_addr; in __xe_res_dma_next()
192 cur->dma_addr = addr; in __xe_res_dma_next()
216 cur->dma_addr = NULL; in xe_res_first_sg()
223 * xe_res_first_dma - initialize a xe_res_cursor with dma_addr array
225 * @dma_addr: struct drm_pagemap_device_addr array to walk
232 static inline void xe_res_first_dma(const struct drm_pagemap_device_addr *dma_addr, in xe_res_first_dma() argument
[all...]
/linux/drivers/media/platform/mediatek/vcodec/decoder/vdec/
H A Dvdec_h264_req_multi_if.c138 u64 dma_addr; member
494 inst->vsi_ext->bs.dma_addr = (u64)bs->dma_addr; in vdec_h264_slice_setup_lat_buffer_ext()
499 inst->vsi_ext->mv_buf_dma[i].dma_addr = mem->dma_addr; in vdec_h264_slice_setup_lat_buffer_ext()
502 inst->vsi_ext->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr; in vdec_h264_slice_setup_lat_buffer_ext()
505 inst->vsi_ext->row_info.dma_addr = 0; in vdec_h264_slice_setup_lat_buffer_ext()
508 inst->vsi_ext->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr; in vdec_h264_slice_setup_lat_buffer_ext()
[all...]
H A Dvdec_hevc_req_multi_if.c274 dma_addr_t dma_addr; member
740 inst->vsi->bs.dma_addr = (u64)bs->dma_addr; in vdec_hevc_slice_setup_lat_buffer()
761 inst->vsi->mv_buf_dma[i].dma_addr = mem->dma_addr; in vdec_hevc_slice_setup_lat_buffer()
765 inst->vsi->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr; in vdec_hevc_slice_setup_lat_buffer()
768 inst->vsi->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr; in vdec_hevc_slice_setup_lat_buffer()
771 inst->vsi->slice_bc.dma_addr in vdec_hevc_slice_setup_lat_buffer()
[all...]
/linux/drivers/net/ethernet/sfc/
H A Dtx_tso.c42 * @dma_addr: DMA address of current position
65 dma_addr_t dma_addr; member
96 * @dma_addr: DMA address of fragment
103 dma_addr_t dma_addr, unsigned int len, in efx_tx_queue_insert() argument
119 buffer->dma_addr = dma_addr; in efx_tx_queue_insert()
122 dma_addr, len); in efx_tx_queue_insert()
130 dma_addr += dma_len; in efx_tx_queue_insert()
175 dma_addr_t dma_addr; in tso_start() local
198 dma_addr in tso_start()
[all...]
H A Dmcdi_functions.c80 dma_addr_t dma_addr; in efx_mcdi_ev_init() local
115 dma_addr = channel->eventq.dma_addr; in efx_mcdi_ev_init()
117 MCDI_SET_ARRAY_QWORD(inbuf, INIT_EVQ_IN_DMA_ADDR, i, dma_addr); in efx_mcdi_ev_init()
118 dma_addr += EFX_BUF_SIZE; in efx_mcdi_ev_init()
172 dma_addr_t dma_addr; in efx_mcdi_tx_init() local
185 dma_addr = tx_queue->txd.dma_addr; in efx_mcdi_tx_init()
188 tx_queue->queue, entries, (u64)dma_addr); in efx_mcdi_tx_init()
191 MCDI_SET_ARRAY_QWORD(inbuf, INIT_TXQ_IN_DMA_ADDR, i, dma_addr); in efx_mcdi_tx_init()
285 dma_addr_t dma_addr; efx_mcdi_rx_init() local
[all...]
/linux/kernel/dma/
H A Ddebug.h14 int direction, dma_addr_t dma_addr,
28 dma_addr_t dma_addr, void *virt,
36 dma_addr_t dma_addr,
39 extern void debug_dma_unmap_resource(struct device *dev, dma_addr_t dma_addr,
60 int direction, dma_addr_t dma_addr, in debug_dma_map_page() argument
83 dma_addr_t dma_addr, void *virt, in debug_dma_alloc_coherent() argument
95 dma_addr_t dma_addr, in debug_dma_map_resource() argument
101 dma_addr_t dma_addr, size_t size, in debug_dma_unmap_resource() argument
H A Ddirect.c35 dma_addr_t dma_addr) in dma_direct_to_page() argument
37 return pfn_to_page(PHYS_PFN(dma_to_phys(dev, dma_addr))); in dma_direct_to_page()
72 dma_addr_t dma_addr = phys_to_dma_direct(dev, phys); in dma_coherent_ok() local
74 if (dma_addr == DMA_MAPPING_ERROR) in dma_coherent_ok()
76 return dma_addr + size - 1 <= in dma_coherent_ok()
316 void *cpu_addr, dma_addr_t dma_addr, unsigned long attrs) in dma_direct_free() argument
330 arch_dma_free(dev, size, cpu_addr, dma_addr, attrs); in dma_direct_free()
355 __dma_direct_free_pages(dev, dma_direct_to_page(dev, dma_addr), size); in dma_direct_free()
382 struct page *page, dma_addr_t dma_addr, in dma_direct_free_pages() argument
508 dma_addr_t dma_addr in dma_direct_map_resource() local
522 dma_direct_get_sgtable(struct device * dev,struct sg_table * sgt,void * cpu_addr,dma_addr_t dma_addr,size_t size,unsigned long attrs) dma_direct_get_sgtable() argument
541 dma_direct_mmap(struct device * dev,struct vm_area_struct * vma,void * cpu_addr,dma_addr_t dma_addr,size_t size,unsigned long attrs) dma_direct_mmap() argument
644 dma_direct_need_sync(struct device * dev,dma_addr_t dma_addr) dma_direct_need_sync() argument
[all...]
H A Ddirect.h14 void *cpu_addr, dma_addr_t dma_addr, size_t size,
18 void *cpu_addr, dma_addr_t dma_addr, size_t size,
20 bool dma_direct_need_sync(struct device *dev, dma_addr_t dma_addr);
88 dma_addr_t dma_addr = phys_to_dma(dev, phys); in dma_direct_map_page() local
96 if (unlikely(!dma_capable(dev, dma_addr, size, true)) || in dma_direct_map_page()
105 &dma_addr, size, *dev->dma_mask, dev->bus_dma_limit); in dma_direct_map_page()
111 return dma_addr; in dma_direct_map_page()
/linux/drivers/net/ethernet/qualcomm/emac/
H A Demac-mac.c305 writel(upper_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config()
308 writel(lower_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config()
315 writel(upper_32_bits(adpt->rx_q.rfd.dma_addr), in emac_mac_dma_rings_config()
318 writel(lower_32_bits(adpt->rx_q.rfd.dma_addr), in emac_mac_dma_rings_config()
320 writel(lower_32_bits(adpt->rx_q.rrd.dma_addr), in emac_mac_dma_rings_config()
598 if (tpbuf->dma_addr) { in emac_tx_q_descs_free()
600 tpbuf->dma_addr, tpbuf->length, in emac_tx_q_descs_free()
602 tpbuf->dma_addr = 0; in emac_tx_q_descs_free()
635 if (rfbuf->dma_addr) { in emac_rx_q_free_descs()
636 dma_unmap_single(dev, rfbuf->dma_addr, rfbu in emac_rx_q_free_descs()
[all...]
/linux/drivers/net/ethernet/netronome/nfp/nfd3/
H A Ddp.c266 dma_addr_t dma_addr; in nfp_nfd3_tx() local
302 dma_addr = dma_map_single(dp->dev, skb->data, skb_headlen(skb), in nfp_nfd3_tx()
304 if (dma_mapping_error(dp->dev, dma_addr)) in nfp_nfd3_tx()
312 txbuf->dma_addr = dma_addr; in nfp_nfd3_tx()
321 nfp_desc_set_dma_addr_40b(txd, dma_addr); in nfp_nfd3_tx()
350 dma_addr = skb_frag_dma_map(dp->dev, frag, 0, in nfp_nfd3_tx()
352 if (dma_mapping_error(dp->dev, dma_addr)) in nfp_nfd3_tx()
357 tx_ring->txbufs[wr_idx].dma_addr = dma_addr; in nfp_nfd3_tx()
552 nfp_nfd3_napi_alloc_one(struct nfp_net_dp * dp,dma_addr_t * dma_addr) nfp_nfd3_napi_alloc_one() argument
589 nfp_nfd3_rx_give_one(const struct nfp_net_dp * dp,struct nfp_net_rx_ring * rx_ring,void * frag,dma_addr_t dma_addr) nfp_nfd3_rx_give_one() argument
1215 dma_addr_t dma_addr; nfp_nfd3_ctrl_tx_one() local
[all...]
/linux/drivers/media/pci/solo6x10/
H A Dsolo6x10-p2m.c32 dma_addr_t dma_addr; in solo_p2m_dma() local
40 dma_addr = dma_map_single(&solo_dev->pdev->dev, sys_addr, size, in solo_p2m_dma()
42 if (dma_mapping_error(&solo_dev->pdev->dev, dma_addr)) in solo_p2m_dma()
45 ret = solo_p2m_dma_t(solo_dev, wr, dma_addr, ext_addr, size, in solo_p2m_dma()
48 dma_unmap_single(&solo_dev->pdev->dev, dma_addr, size, in solo_p2m_dma()
93 desc[1].dma_addr); in solo_p2m_dma_desc()
125 dma_addr_t dma_addr, u32 ext_addr, u32 size, in solo_p2m_fill_desc() argument
128 WARN_ON_ONCE(dma_addr & 0x03); in solo_p2m_fill_desc()
141 desc->dma_addr = dma_addr; in solo_p2m_fill_desc()
146 solo_p2m_dma_t(struct solo_dev * solo_dev,int wr,dma_addr_t dma_addr,u32 ext_addr,u32 size,int repeat,u32 ext_size) solo_p2m_dma_t() argument
[all...]
/linux/arch/mips/loongson2ef/lemote-2f/
H A Ddma.c9 phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dma_addr) in dma_to_phys() argument
11 if (dma_addr > 0x8fffffff) in dma_to_phys()
12 return dma_addr; in dma_to_phys()
13 return dma_addr & 0x0fffffff; in dma_to_phys()
/linux/drivers/xen/
H A Dswiotlb-xen.c72 dma_addr_t dma_addr) in xen_dma_to_phys() argument
74 return xen_bus_to_phys(dev, dma_to_phys(dev, dma_addr)); in xen_dma_to_phys()
100 dma_addr_t dma_addr) in xen_swiotlb_find_pool() argument
102 unsigned long bfn = XEN_PFN_DOWN(dma_to_phys(dev, dma_addr)); in xen_swiotlb_find_pool()
259 * Unmap a single streaming mode DMA translation. The dma_addr and size must
289 xen_swiotlb_sync_single_for_cpu(struct device *dev, dma_addr_t dma_addr, in xen_swiotlb_sync_single_for_cpu() argument
292 phys_addr_t paddr = xen_dma_to_phys(dev, dma_addr); in xen_swiotlb_sync_single_for_cpu()
296 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr)))) in xen_swiotlb_sync_single_for_cpu()
299 xen_dma_sync_for_cpu(dev, dma_addr, size, dir); in xen_swiotlb_sync_single_for_cpu()
302 pool = xen_swiotlb_find_pool(dev, dma_addr); in xen_swiotlb_sync_single_for_cpu()
308 xen_swiotlb_sync_single_for_device(struct device * dev,dma_addr_t dma_addr,size_t size,enum dma_data_direction dir) xen_swiotlb_sync_single_for_device() argument
[all...]
/linux/drivers/media/platform/mediatek/vcodec/encoder/
H A Dvenc_vpu_if.c270 if ((frm_buf->fb_addr[0].dma_addr % 16 == 0) && in vpu_enc_encode_32bits()
271 (frm_buf->fb_addr[1].dma_addr % 16 == 0) && in vpu_enc_encode_32bits()
272 (frm_buf->fb_addr[2].dma_addr % 16 == 0)) { in vpu_enc_encode_32bits()
273 out.base.input_addr[0] = frm_buf->fb_addr[0].dma_addr; in vpu_enc_encode_32bits()
274 out.base.input_addr[1] = frm_buf->fb_addr[1].dma_addr; in vpu_enc_encode_32bits()
275 out.base.input_addr[2] = frm_buf->fb_addr[2].dma_addr; in vpu_enc_encode_32bits()
277 mtk_venc_err(vpu->ctx, "dma_addr not align to 16"); in vpu_enc_encode_32bits()
282 out.base.bs_addr = bs_buf->dma_addr; in vpu_enc_encode_32bits()
316 if ((frm_buf->fb_addr[0].dma_addr % 16 == 0) && in vpu_enc_encode_34bits()
317 (frm_buf->fb_addr[1].dma_addr in vpu_enc_encode_34bits()
[all...]
/linux/drivers/crypto/caam/
H A Dkey_gen.c50 dma_addr_t dma_addr; in gen_split_key() local
75 dma_addr = dma_map_single(jrdev, key_out, local_max, DMA_BIDIRECTIONAL); in gen_split_key()
76 if (dma_mapping_error(jrdev, dma_addr)) { in gen_split_key()
82 append_key(desc, dma_addr, keylen, CLASS_2 | KEY_DEST_CLASS_REG); in gen_split_key()
100 append_fifo_store(desc, dma_addr, adata->keylen, in gen_split_key()
121 dma_unmap_single(jrdev, dma_addr, local_max, DMA_BIDIRECTIONAL); in gen_split_key()
/linux/drivers/net/ethernet/sfc/falcon/
H A Dtx.c38 buffer->dma_addr = page_buf->dma_addr + offset; in ef4_tx_get_copy_buffer()
50 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in ef4_dequeue_buffer()
174 dma_addr_t dma_addr, in ef4_tx_map_chunk() argument
184 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in ef4_tx_map_chunk()
187 buffer->dma_addr = dma_addr; in ef4_tx_map_chunk()
190 dma_addr += dma_len; in ef4_tx_map_chunk()
204 dma_addr_t dma_addr, unmap_addr; in ef4_tx_map_data() local
213 dma_addr in ef4_tx_map_data()
[all...]
/linux/drivers/crypto/intel/qat/qat_common/
H A Dadf_rl_admin.c12 prep_admin_req_msg(struct rl_sla *sla, dma_addr_t dma_addr, in prep_admin_req_msg() argument
17 req->init_cfg_ptr = dma_addr; in prep_admin_req_msg()
76 dma_addr_t dma_addr; in adf_rl_send_admin_add_update_msg() local
80 &dma_addr, GFP_KERNEL); in adf_rl_send_admin_add_update_msg()
85 prep_admin_req_msg(sla, dma_addr, fw_params, &req, is_update); in adf_rl_send_admin_add_update_msg()
89 dma_addr); in adf_rl_send_admin_add_update_msg()

12345678910>>...32