| /linux/drivers/crypto/caam/ |
| H A D | caampkc.c | 49 static void rsa_io_unmap(struct device *dev, struct rsa_edesc *edesc, in rsa_io_unmap() argument 54 dma_unmap_sg(dev, req->dst, edesc->dst_nents, DMA_FROM_DEVICE); in rsa_io_unmap() 55 dma_unmap_sg(dev, req_ctx->fixup_src, edesc->src_nents, DMA_TO_DEVICE); in rsa_io_unmap() 57 if (edesc->sec4_sg_bytes) in rsa_io_unmap() 58 dma_unmap_single(dev, edesc->sec4_sg_dma, edesc->sec4_sg_bytes, in rsa_io_unmap() 62 static void rsa_pub_unmap(struct device *dev, struct rsa_edesc *edesc, in rsa_pub_unmap() argument 68 struct rsa_pub_pdb *pdb = &edesc->pdb.pub; in rsa_pub_unmap() 74 static void rsa_priv_f1_unmap(struct device *dev, struct rsa_edesc *edesc, in rsa_priv_f1_unmap() argument 80 struct rsa_priv_f1_pdb *pdb = &edesc->pdb.priv_f1; in rsa_priv_f1_unmap() 86 static void rsa_priv_f2_unmap(struct device *dev, struct rsa_edesc *edesc, in rsa_priv_f2_unmap() argument [all …]
|
| H A D | caamhash.c | 124 struct ahash_edesc *edesc; member 553 struct ahash_edesc *edesc, in ahash_unmap() argument 558 if (edesc->src_nents) in ahash_unmap() 559 dma_unmap_sg(dev, req->src, edesc->src_nents, DMA_TO_DEVICE); in ahash_unmap() 561 if (edesc->sec4_sg_bytes) in ahash_unmap() 562 dma_unmap_single(dev, edesc->sec4_sg_dma, in ahash_unmap() 563 edesc->sec4_sg_bytes, DMA_TO_DEVICE); in ahash_unmap() 573 struct ahash_edesc *edesc, in ahash_unmap_ctx() argument 582 ahash_unmap(dev, edesc, req, dst_len); in ahash_unmap_ctx() 590 struct ahash_edesc *edesc; in ahash_done_cpy() local [all …]
|
| H A D | caamalg.c | 138 struct skcipher_edesc *edesc; member 143 struct aead_edesc *edesc; member 1026 struct aead_edesc *edesc, in aead_unmap() argument 1030 edesc->src_nents, edesc->dst_nents, 0, 0, in aead_unmap() 1031 edesc->sec4_sg_dma, edesc->sec4_sg_bytes); in aead_unmap() 1034 static void skcipher_unmap(struct device *dev, struct skcipher_edesc *edesc, in skcipher_unmap() argument 1041 edesc->src_nents, edesc->dst_nents, in skcipher_unmap() 1042 edesc->iv_dma, ivsize, in skcipher_unmap() 1043 edesc->sec4_sg_dma, edesc->sec4_sg_bytes); in skcipher_unmap() 1052 struct aead_edesc *edesc; in aead_crypt_done() local [all …]
|
| H A D | caamalg_qi2.c | 367 struct aead_edesc *edesc; in aead_edesc_alloc() local 376 edesc = qi_cache_zalloc(flags); in aead_edesc_alloc() 377 if (unlikely(!edesc)) { in aead_edesc_alloc() 390 qi_cache_free(edesc); in aead_edesc_alloc() 398 qi_cache_free(edesc); in aead_edesc_alloc() 407 qi_cache_free(edesc); in aead_edesc_alloc() 421 qi_cache_free(edesc); in aead_edesc_alloc() 435 qi_cache_free(edesc); in aead_edesc_alloc() 443 qi_cache_free(edesc); in aead_edesc_alloc() 473 sg_table = &edesc->sgt[0]; in aead_edesc_alloc() [all …]
|
| H A D | caamalg_qi.c | 894 struct aead_edesc *edesc, in aead_unmap() argument 900 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in aead_unmap() 901 edesc->iv_dma, ivsize, DMA_TO_DEVICE, edesc->qm_sg_dma, in aead_unmap() 902 edesc->qm_sg_bytes); in aead_unmap() 903 dma_unmap_single(dev, edesc->assoclen_dma, 4, DMA_TO_DEVICE); in aead_unmap() 906 static void skcipher_unmap(struct device *dev, struct skcipher_edesc *edesc, in skcipher_unmap() argument 912 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in skcipher_unmap() 913 edesc->iv_dma, ivsize, DMA_BIDIRECTIONAL, edesc->qm_sg_dma, in skcipher_unmap() 914 edesc->qm_sg_bytes); in skcipher_unmap() 920 struct aead_edesc *edesc; in aead_done() local [all …]
|
| H A D | caampkc.h | 113 struct rsa_edesc *edesc; member
|
| H A D | caamalg_qi2.h | 189 void *edesc; member
|
| /linux/drivers/dma/ti/ |
| H A D | edma.c | 218 struct edma_desc *edesc; member 722 struct edma_desc *edesc; in edma_execute() local 726 if (!echan->edesc) { in edma_execute() 732 echan->edesc = to_edma_desc(&vdesc->tx); in edma_execute() 735 edesc = echan->edesc; in edma_execute() 738 left = edesc->pset_nr - edesc->processed; in edma_execute() 740 edesc->sg_len = 0; in edma_execute() 744 j = i + edesc->processed; in edma_execute() 745 edma_write_slot(ecc, echan->slot[i], &edesc->pset[j].param); in edma_execute() 746 edesc->sg_len += edesc->pset[j].len; in edma_execute() [all …]
|
| /linux/drivers/crypto/ |
| H A D | talitos.c | 326 struct talitos_edesc *edesc; in get_request_hdr() local 334 edesc = container_of(request->desc, struct talitos_edesc, desc); in get_request_hdr() 336 return ((struct talitos_desc *)(edesc->buf + edesc->dma_len))->hdr1; in get_request_hdr() 491 struct talitos_edesc *edesc; local 493 edesc = container_of(priv->chan[ch].fifo[iter].desc, 496 (edesc->buf + edesc->dma_len))->hdr; 960 struct talitos_edesc *edesc, in talitos_sg_unmap() argument 967 unsigned int src_nents = edesc->src_nents ? : 1; in talitos_sg_unmap() 968 unsigned int dst_nents = edesc->dst_nents ? : 1; in talitos_sg_unmap() 971 dma_sync_single_for_device(dev, edesc->dma_link_tbl + offset, in talitos_sg_unmap() [all …]
|
| /linux/drivers/dma/ |
| H A D | fsl-edma-common.c | 51 if (!fsl_chan->edesc) { in fsl_edma_tx_chan_handler() 57 if (!fsl_chan->edesc->iscyclic) { in fsl_edma_tx_chan_handler() 58 list_del(&fsl_chan->edesc->vdesc.node); in fsl_edma_tx_chan_handler() 59 vchan_cookie_complete(&fsl_chan->edesc->vdesc); in fsl_edma_tx_chan_handler() 60 fsl_chan->edesc = NULL; in fsl_edma_tx_chan_handler() 63 vchan_cyclic_callback(&fsl_chan->edesc->vdesc); in fsl_edma_tx_chan_handler() 66 if (!fsl_chan->edesc) in fsl_edma_tx_chan_handler() 244 fsl_chan->edesc = NULL; in fsl_edma_terminate_all() 262 if (fsl_chan->edesc) { in fsl_edma_pause() 276 if (fsl_chan->edesc) { in fsl_edma_resume() [all …]
|
| H A D | fsl-edma-common.h | 168 struct fsl_edma_desc *edesc; member
|
| H A D | fsl-edma-main.c | 70 if (!fsl_chan->edesc) in fsl_edma3_err_check()
|
| /linux/drivers/net/ethernet/stmicro/stmmac/ |
| H A D | stmmac.h | 110 struct dma_edesc *edesc; member
|