| /linux/drivers/crypto/qce/ |
| H A D | skcipher.c | 51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_done() 75 int dst_nents, src_nents, ret; in qce_skcipher_async_req_handle() local 87 rctx->dst_nents = sg_nents_for_len(req->dst, req->cryptlen); in qce_skcipher_async_req_handle() 89 rctx->dst_nents = rctx->src_nents; in qce_skcipher_async_req_handle() 94 if (rctx->dst_nents < 0) { in qce_skcipher_async_req_handle() 96 return -rctx->dst_nents; in qce_skcipher_async_req_handle() 99 rctx->dst_nents += 1; in qce_skcipher_async_req_handle() 104 ret = sg_alloc_table(&rctx->dst_tbl, rctx->dst_nents, gfp); in qce_skcipher_async_req_handle() 126 dst_nents = dma_map_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_async_req_handle() 127 if (!dst_nents) { in qce_skcipher_async_req_handle() [all …]
|
| H A D | aead.c | 51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_aead_done() 125 rctx->dst_nents = sg_nents_for_len(req->dst, totallen); in qce_aead_prepare_dst_buf() 126 if (rctx->dst_nents < 0) { in qce_aead_prepare_dst_buf() 131 rctx->dst_nents += 2; in qce_aead_prepare_dst_buf() 133 rctx->dst_nents += 1; in qce_aead_prepare_dst_buf() 137 ret = sg_alloc_table(&rctx->dst_tbl, rctx->dst_nents, gfp); in qce_aead_prepare_dst_buf() 175 rctx->dst_nents = sg_nents_for_len(rctx->dst_sg, totallen) + 1; in qce_aead_prepare_dst_buf() 289 rctx->dst_nents = rctx->src_nents + 1; in qce_aead_ccm_prepare_buf_assoclen() 291 rctx->dst_nents = rctx->src_nents; in qce_aead_ccm_prepare_buf_assoclen() 323 rctx->src_nents = rctx->dst_nents - 1; in qce_aead_prepare_buf() [all …]
|
| H A D | cipher.h | 39 int dst_nents; member
|
| H A D | aead.h | 31 int dst_nents; member
|
| /linux/drivers/crypto/inside-secure/eip93/ |
| H A D | eip93-common.c | 212 int src_nents, dst_nents; in check_valid_request() local 231 dst_nents = sg_nents_for_len(dst, totlen_dst); in check_valid_request() 232 if (dst_nents < 0) in check_valid_request() 233 return dst_nents; in check_valid_request() 236 src_nents = max(src_nents, dst_nents); in check_valid_request() 237 dst_nents = src_nents; in check_valid_request() 245 if (unlikely(totlen_dst && !dst_nents)) in check_valid_request() 250 if (dst_nents == 1 && src_nents == 1) { in check_valid_request() 285 dst_nents = sg_nents_for_len(rctx->sg_dst, totlen_dst); in check_valid_request() 286 if (dst_nents < 0) in check_valid_request() [all …]
|
| H A D | eip93-cipher.h | 42 int dst_nents; member
|
| /linux/drivers/crypto/caam/ |
| H A D | caamalg_qi.c | 804 int dst_nents; member 826 int dst_nents; member 874 int dst_nents, dma_addr_t iv_dma, int ivsize, in caam_unmap() argument 881 if (dst_nents) in caam_unmap() 882 dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE); in caam_unmap() 900 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in aead_unmap() 912 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in skcipher_unmap() 951 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local 1004 dst_nents = sg_nents_for_len(req->dst, dst_len); in aead_edesc_alloc() 1005 if (unlikely(dst_nents < 0)) { in aead_edesc_alloc() [all …]
|
| H A D | caamalg_qi2.h | 112 int dst_nents; member 132 int dst_nents; member
|
| H A D | caamalg.c | 966 int dst_nents; member 992 int dst_nents; member 1005 int dst_nents, in caam_unmap() argument 1012 if (dst_nents) in caam_unmap() 1013 dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE); in caam_unmap() 1030 edesc->src_nents, edesc->dst_nents, 0, 0, in aead_unmap() 1041 edesc->src_nents, edesc->dst_nents, in skcipher_unmap() 1121 edesc->dst_nents > 1 ? 100 : req->cryptlen, 1); in skcipher_crypt_done() 1395 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local 1412 dst_nents = sg_nents_for_len(req->dst, dst_len); in aead_edesc_alloc() [all …]
|
| H A D | caamalg_qi2.c | 153 int dst_nents, dma_addr_t iv_dma, int ivsize, in caam_unmap() argument 160 if (dst_nents) in caam_unmap() 161 dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE); in caam_unmap() 365 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local 394 dst_nents = sg_nents_for_len(req->dst, dst_len); in aead_edesc_alloc() 395 if (unlikely(dst_nents < 0)) { in aead_edesc_alloc() 399 return ERR_PTR(dst_nents); in aead_edesc_alloc() 414 if (dst_nents) { in aead_edesc_alloc() 415 mapped_dst_nents = dma_map_sg(dev, req->dst, dst_nents, in aead_edesc_alloc() 479 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in aead_edesc_alloc() [all …]
|
| H A D | caampkc.c | 54 dma_unmap_sg(dev, req->dst, edesc->dst_nents, DMA_FROM_DEVICE); in rsa_io_unmap() 262 int src_nents, dst_nents; in rsa_edesc_alloc() local 292 dst_nents = sg_nents_for_len(req->dst, req->dst_len); in rsa_edesc_alloc() 300 mapped_dst_nents = dma_map_sg(dev, req->dst, dst_nents, in rsa_edesc_alloc() 340 edesc->dst_nents = dst_nents; in rsa_edesc_alloc() 368 dma_unmap_sg(dev, req->dst, dst_nents, DMA_FROM_DEVICE); in rsa_edesc_alloc()
|
| H A D | caampkc.h | 133 int dst_nents; member
|
| /linux/drivers/crypto/intel/keembay/ |
| H A D | keembay-ocs-aes-core.c | 86 int dst_nents; member 387 dma_unmap_sg(dev, req->dst, rctx->dst_nents, rctx->in_place ? in kmb_ocs_sk_dma_cleanup() 419 sg_swap_blocks(req->dst, rctx->dst_nents, in kmb_ocs_sk_prepare_inplace() 425 rctx->dst_nents, DMA_BIDIRECTIONAL); in kmb_ocs_sk_prepare_inplace() 475 rctx->dst_nents, DMA_FROM_DEVICE); in kmb_ocs_sk_prepare_notinplace() 523 rctx->dst_nents = sg_nents_for_len(req->dst, req->cryptlen); in kmb_ocs_sk_run() 524 if (rctx->dst_nents < 0) in kmb_ocs_sk_run() 560 sg_swap_blocks(req->dst, rctx->dst_nents, in kmb_ocs_sk_run() 694 dma_unmap_sg(dev, req->dst, rctx->dst_nents, rctx->in_place ? in kmb_ocs_aead_dma_cleanup() 783 rctx->dst_nents = sg_nents_for_len(req->dst, dst_size); in kmb_ocs_aead_dma_prepare() [all …]
|
| /linux/drivers/crypto/marvell/cesa/ |
| H A D | cipher.c | 64 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_cleanup() 134 len = mv_cesa_sg_copy_from_sram(engine, req->dst, creq->dst_nents, in mv_cesa_skcipher_std_process() 334 ret = dma_map_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_req_init() 398 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_req_init() 440 creq->dst_nents = sg_nents_for_len(req->dst, req->cryptlen); in mv_cesa_skcipher_req_init() 441 if (creq->dst_nents < 0) { in mv_cesa_skcipher_req_init() 443 return creq->dst_nents; in mv_cesa_skcipher_req_init()
|
| /linux/drivers/crypto/ti/ |
| H A D | dthev2-aes.c | 274 int dst_nents; in dthe_aes_run() local 318 dst_nents = src_nents; in dthe_aes_run() 321 dst_nents = sg_nents_for_len(dst, len); in dthe_aes_run() 322 dst_mapped_nents = dma_map_sg(rx_dev, dst, dst_nents, dst_dir); in dthe_aes_run() 391 dma_unmap_sg(rx_dev, dst, dst_nents, dst_dir); in dthe_aes_run()
|
| /linux/drivers/crypto/ |
| H A D | talitos.c | 968 unsigned int dst_nents = edesc->dst_nents ? : 1; in talitos_sg_unmap() local 970 if (is_sec1 && dst && dst_nents > 1) { in talitos_sg_unmap() 973 sg_pcopy_from_buffer(dst, dst_nents, edesc->buf + offset, len, in talitos_sg_unmap() 980 if (dst && (dst_nents == 1 || !is_sec1)) in talitos_sg_unmap() 981 dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE); in talitos_sg_unmap() 1012 unsigned int dst_nents = edesc->dst_nents ? : 1; in ipsec_esp_unmap() local 1014 sg_pcopy_to_buffer(areq->dst, dst_nents, ctx->iv, ivsize, in ipsec_esp_unmap() 1265 sg_count = edesc->dst_nents ? : 1; in ipsec_esp() 1332 int src_nents, dst_nents, alloc_len, dma_len, src_len, dst_len; in talitos_edesc_alloc() local 1353 dst_nents = dst ? src_nents : 0; in talitos_edesc_alloc() [all …]
|
| H A D | sa2ul.c | 1088 int sg_nents, src_nents, dst_nents; in sa_run() local 1199 dst_nents = src_nents; in sa_run() 1202 dst_nents = sg_nents_for_len(req->dst, req->size); in sa_run() 1205 if (dst_nents == 1 && split_size <= req->dst->length) { in sa_run() 1207 dst_nents = 1; in sa_run() 1213 mapped_sg->sgt.orig_nents = dst_nents; in sa_run() 1223 mapped_sg->sgt.orig_nents = dst_nents; in sa_run() 1233 0, 1, &split_size, &dst, &dst_nents, in sa_run() 1236 dst_nents = mapped_sg->sgt.nents; in sa_run() 1244 rxd->tx_in = dmaengine_prep_slave_sg(dma_rx, dst, dst_nents, in sa_run()
|
| H A D | talitos.h | 62 int dst_nents; member
|
| /linux/drivers/crypto/aspeed/ |
| H A D | aspeed-hace-crypto.c | 144 dma_unmap_sg(dev, req->dst, rctx->dst_nents, DMA_FROM_DEVICE); in aspeed_sk_transfer_sg() 164 nbytes = sg_copy_from_buffer(out_sg, rctx->dst_nents, in aspeed_sk_transfer() 174 "nb_out_sg", rctx->dst_nents, in aspeed_sk_transfer() 259 rctx->dst_nents, DMA_FROM_DEVICE); in aspeed_sk_start_sg() 348 dma_unmap_sg(hace_dev->dev, req->dst, rctx->dst_nents, in aspeed_sk_start_sg() 380 rctx->dst_nents = sg_nents(req->dst); in aspeed_hace_skcipher_trigger()
|
| H A D | aspeed-hace.h | 229 int dst_nents; member
|
| /linux/drivers/crypto/virtio/ |
| H A D | virtio_crypto_skcipher_algs.c | 331 int src_nents, dst_nents; in __virtio_crypto_skcipher_do_req() local 347 dst_nents = sg_nents(req->dst); in __virtio_crypto_skcipher_do_req() 350 src_nents, dst_nents); in __virtio_crypto_skcipher_do_req() 353 sg_total = src_nents + dst_nents + 3; in __virtio_crypto_skcipher_do_req()
|
| /linux/drivers/crypto/ccp/ |
| H A D | ccp-dmaengine.c | 354 unsigned int dst_nents, in ccp_create_desc() argument 375 if (!dst_nents || !src_nents) in ccp_create_desc() 406 dst_nents--; in ccp_create_desc() 407 if (!dst_nents) in ccp_create_desc()
|
| /linux/drivers/crypto/starfive/ |
| H A D | jh7110-aes.c | 506 unsigned int len, src_nents, dst_nents; in starfive_aes_map_sg() local 516 dst_nents = src_nents; in starfive_aes_map_sg() 532 dst_nents = dma_map_sg(cryp->dev, dtsg, 1, DMA_FROM_DEVICE); in starfive_aes_map_sg() 533 if (dst_nents == 0) in starfive_aes_map_sg()
|
| /linux/drivers/crypto/bcm/ |
| H A D | cipher.h | 295 int dst_nents; /* Number of dst entries with data */ member
|
| H A D | cipher.c | 161 rctx->dst_nents, chunksize); in spu_skcipher_rx_sg_create() 347 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize); in handle_skcipher_req() 432 rx_frag_num += rctx->dst_nents; in handle_skcipher_req() 1101 rctx->dst_nents, resp_len); in spu_aead_rx_sg_create() 1339 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize); in handle_aead_req() 1467 rx_frag_num += rctx->dst_nents; in handle_aead_req() 1486 rx_frag_num -= rctx->dst_nents; in handle_aead_req() 1732 rctx->dst_nents = 0; in skcipher_enqueue() 1920 rctx->dst_nents = 0; in ahash_enqueue() 2641 rctx->dst_nents = 0; in aead_enqueue()
|