Home
last modified time | relevance | path

Searched refs:out_sg (Results 1 – 25 of 28) sorted by relevance

12

/linux/lib/
H A Dsg_split.c17 struct scatterlist *out_sg; member
80 struct scatterlist *in_sg, *out_sg; in sg_split_phys() local
85 out_sg = split->out_sg; in sg_split_phys()
86 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys()
87 *out_sg = *in_sg; in sg_split_phys()
89 out_sg->offset += split->skip_sg0; in sg_split_phys()
90 out_sg->length -= split->skip_sg0; in sg_split_phys()
92 sg_dma_address(out_sg) = 0; in sg_split_phys()
93 sg_dma_len(out_sg) = 0; in sg_split_phys()
96 out_sg[-1].length = split->length_last_sg; in sg_split_phys()
[all …]
/linux/drivers/crypto/nx/
H A Dnx-aes-xcbc.c62 struct nx_sg *in_sg, *out_sg; in nx_xcbc_empty() local
86 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *) keys, &len, in nx_xcbc_empty()
93 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_xcbc_empty()
114 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, in nx_xcbc_empty()
121 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_xcbc_empty()
172 struct nx_sg *out_sg; in nx_xcbc_update() local
194 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, in nx_xcbc_update()
202 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_xcbc_update()
250 struct nx_sg *in_sg, *out_sg; in nx_xcbc_finup() local
284 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, in nx_xcbc_finup()
[all …]
H A Dnx-sha512.c70 struct nx_sg *out_sg; in nx_sha512_update() local
88 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, in nx_sha512_update()
90 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha512_update()
150 struct nx_sg *in_sg, *out_sg; in nx_sha512_finup() local
187 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, in nx_sha512_finup()
191 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha512_finup()
H A Dnx-sha256.c69 struct nx_sg *out_sg; in nx_sha256_update() local
87 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, in nx_sha256_update()
89 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha256_update()
147 struct nx_sg *in_sg, *out_sg; in nx_sha256_finup() local
179 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, max_sg_len); in nx_sha256_finup()
187 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha256_finup()
H A Dnx.c246 struct nx_sg *nx_outsg = nx_ctx->out_sg; in nx_build_sg_lists()
271 nx_ctx->op.outlen = trim_sg_list(nx_ctx->out_sg, nx_outsg, delta, nbytes); in nx_build_sg_lists()
291 nx_ctx->op.out = __pa(nx_ctx->out_sg); in nx_ctx_init()
299 nx_ctx->op_aead.out = __pa(nx_ctx->out_sg); in nx_ctx_init()
654 nx_ctx->out_sg = (struct nx_sg *)((u8 *)nx_ctx->in_sg + NX_PAGE_SIZE); in nx_crypto_ctx_init()
658 (struct nx_csbcpb *)((u8 *)nx_ctx->out_sg + in nx_crypto_ctx_init()
730 nx_ctx->out_sg = NULL; in nx_crypto_ctx_exit()
H A Dnx-aes-gcm.c246 struct nx_sg *in_sg, *out_sg; in gcm_empty() local
273 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *) out, &len, in gcm_empty()
280 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in gcm_empty()
H A Dnx.h131 struct nx_sg *out_sg; /* aligned pointer into kmem to an sg list */ member
H A Dnx-aes-ccm.c162 struct nx_sg *nx_outsg = nx_ctx->out_sg; in generate_pat()
249 nx_ctx->op.outlen = (nx_ctx->out_sg - nx_outsg) * in generate_pat()
/linux/drivers/crypto/
H A Domap-des.c146 struct scatterlist *out_sg; member
370 struct scatterlist *in_sg, struct scatterlist *out_sg, in omap_des_crypt_dma() argument
427 tx_out = dmaengine_prep_slave_sg(dd->dma_lch_out, out_sg, out_sg_len, in omap_des_crypt_dma()
466 err = dma_map_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_des_crypt_dma_start()
474 err = omap_des_crypt_dma(tfm, dd->in_sg, dd->out_sg, dd->in_sg_len, in omap_des_crypt_dma_start()
478 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_des_crypt_dma_start()
531 dd->out_sg = req->dst; in omap_des_prepare_req()
544 ret = omap_crypto_align_sg(&dd->out_sg, dd->total, DES_BLOCK_SIZE, in omap_des_prepare_req()
554 dd->out_sg_len = sg_nents_for_len(dd->out_sg, dd->total); in omap_des_prepare_req()
592 dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, in omap_des_done_task()
[all …]
H A Domap-aes-gcm.c37 dd->out_sg = NULL; in omap_aes_gcm_finish_req()
56 dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_gcm_done_task()
59 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, DMA_FROM_DEVICE); in omap_aes_gcm_done_task()
62 omap_crypto_cleanup(dd->out_sg, dd->orig_out, in omap_aes_gcm_done_task()
152 dd->out_sg = req->dst; in omap_aes_gcm_copy_buffers()
155 dd->out_sg = scatterwalk_ffwd(sg_arr, req->dst, req->assoclen); in omap_aes_gcm_copy_buffers()
158 if (req->src == req->dst || dd->out_sg == sg_arr) in omap_aes_gcm_copy_buffers()
162 ret = omap_crypto_align_sg(&dd->out_sg, cryptlen, in omap_aes_gcm_copy_buffers()
171 dd->out_sg_len = sg_nents_for_len(dd->out_sg, clen); in omap_aes_gcm_copy_buffers()
H A Datmel-tdes.c108 struct scatterlist *out_sg; member
294 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); in atmel_tdes_crypt_pdc_stop()
301 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, in atmel_tdes_crypt_pdc_stop()
462 out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) && in atmel_tdes_crypt_start()
463 IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size); in atmel_tdes_crypt_start()
466 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) in atmel_tdes_crypt_start()
473 count = min_t(size_t, count, sg_dma_len(dd->out_sg)); in atmel_tdes_crypt_start()
481 err = dma_map_sg(dd->dev, dd->out_sg, 1, in atmel_tdes_crypt_start()
491 addr_out = sg_dma_address(dd->out_sg); in atmel_tdes_crypt_start()
515 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); in atmel_tdes_crypt_start()
[all …]
H A Domap-aes.c268 struct scatterlist *out_sg, in omap_aes_crypt_dma() argument
326 tx_out = dmaengine_prep_slave_sg(dd->dma_lch_out, out_sg, in omap_aes_crypt_dma()
376 err = dma_map_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_crypt_dma_start()
385 err = omap_aes_crypt_dma(dd, dd->in_sg, dd->out_sg, dd->in_sg_len, in omap_aes_crypt_dma_start()
390 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_crypt_dma_start()
441 dd->out_sg = req->dst; in omap_aes_prepare_req()
454 ret = omap_crypto_align_sg(&dd->out_sg, dd->total, AES_BLOCK_SIZE, in omap_aes_prepare_req()
464 dd->out_sg_len = sg_nents_for_len(dd->out_sg, dd->total); in omap_aes_prepare_req()
506 dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_done_task()
509 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_done_task()
[all …]
H A Dsahara.c214 struct scatterlist *out_sg; member
463 dev->nb_out_sg = sg_nents_for_len(dev->out_sg, dev->total); in sahara_hw_descriptor_create()
481 ret = dma_map_sg(dev->device, dev->out_sg, dev->nb_out_sg, in sahara_hw_descriptor_create()
506 sg = dev->out_sg; in sahara_hw_descriptor_create()
572 dev->out_sg = req->dst; in sahara_aes_process()
603 dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, in sahara_aes_process()
/linux/drivers/crypto/stm32/
H A Dstm32-cryp.c216 struct scatterlist *out_sg; member
924 dma_sync_sg_for_device(cryp->dev, cryp->out_sg, cryp->out_sg_len, DMA_FROM_DEVICE); in stm32_cryp_dma_callback()
926 if (cryp->in_sg != cryp->out_sg) in stm32_cryp_dma_callback()
929 dma_unmap_sg(cryp->dev, cryp->out_sg, cryp->out_sg_len, DMA_FROM_DEVICE); in stm32_cryp_dma_callback()
938 kfree(cryp->out_sg); in stm32_cryp_dma_callback()
942 if (cryp->out_sg != cryp->req->dst) in stm32_cryp_dma_callback()
943 kfree(cryp->out_sg); in stm32_cryp_dma_callback()
1019 if (cryp->in_sg != cryp->out_sg) { in stm32_cryp_dma_start()
1027 ret = dma_map_sg(cryp->dev, cryp->out_sg, cryp->out_sg_len, DMA_FROM_DEVICE); in stm32_cryp_dma_start()
1046 tx_out = dmaengine_prep_slave_sg(cryp->dma_lch_out, cryp->out_sg, cryp->out_sg_len, in stm32_cryp_dma_start()
[all …]
/linux/drivers/crypto/gemini/
H A Dsl3516-ce-cipher.c31 struct scatterlist *out_sg; in sl3516_ce_need_fallback() local
88 out_sg = areq->dst; in sl3516_ce_need_fallback()
89 while (in_sg && out_sg) { in sl3516_ce_need_fallback()
90 if (in_sg->length != out_sg->length) { in sl3516_ce_need_fallback()
95 out_sg = sg_next(out_sg); in sl3516_ce_need_fallback()
97 if (in_sg || out_sg) in sl3516_ce_need_fallback()
/linux/net/vmw_vsock/
H A Dvirtio_transport.c104 int ret, in_sg = 0, out_sg = 0; in virtio_transport_send_skb() local
108 sg_init_one(sgs[out_sg], virtio_vsock_hdr(skb), in virtio_transport_send_skb()
110 out_sg++; in virtio_transport_send_skb()
114 sg_init_one(sgs[out_sg], skb->data, skb->len); in virtio_transport_send_skb()
115 out_sg++; in virtio_transport_send_skb()
139 sg_init_one(sgs[out_sg], in virtio_transport_send_skb()
142 out_sg++; in virtio_transport_send_skb()
146 ret = virtqueue_add_sgs(vq, sgs, out_sg, in_sg, skb, gfp); in virtio_transport_send_skb()
/linux/drivers/crypto/allwinner/sun8i-ss/
H A Dsun8i-ss-cipher.c29 struct scatterlist *out_sg = areq->dst; in sun8i_ss_need_fallback() local
77 out_sg = areq->dst; in sun8i_ss_need_fallback()
78 while (in_sg && out_sg) { in sun8i_ss_need_fallback()
79 if (in_sg->length != out_sg->length) in sun8i_ss_need_fallback()
82 out_sg = sg_next(out_sg); in sun8i_ss_need_fallback()
84 if (in_sg || out_sg) in sun8i_ss_need_fallback()
/linux/drivers/crypto/allwinner/sun4i-ss/
H A Dsun4i-ss-cipher.c183 struct scatterlist *out_sg = areq->dst; in sun4i_ss_cipher_poll() local
230 while (out_sg && no_chunk == 1) { in sun4i_ss_cipher_poll()
231 if ((out_sg->length | out_sg->offset) & 3u) in sun4i_ss_cipher_poll()
233 out_sg = sg_next(out_sg); in sun4i_ss_cipher_poll()
/linux/crypto/
H A Drsa-pkcs1pad.c29 struct scatterlist in_sg[2], out_sg[1]; member
256 pkcs1pad_sg_set_buf(req_ctx->out_sg, req_ctx->out_buf, in pkcs1pad_decrypt()
265 req_ctx->out_sg, req->src_len, in pkcs1pad_decrypt()
/linux/drivers/crypto/starfive/
H A Djh7110-rsa.c278 sg_copy_buffer(rctx->out_sg, sg_nents(rctx->out_sg), in starfive_rsa_enc_core()
310 rctx->out_sg = req->dst; in starfive_rsa_enc()
341 rctx->out_sg = req->dst; in starfive_rsa_dec()
H A Djh7110-aes.c313 scatterwalk_map_and_copy(cryp->tag_out, rctx->out_sg, in starfive_aes_read_authtag()
582 rctx->out_sg = req->dst; in starfive_aes_do_one_req()
595 ret = starfive_aes_map_sg(cryp, rctx->in_sg, rctx->out_sg); in starfive_aes_do_one_req()
651 rctx->out_sg = rctx->in_sg; in starfive_aes_aead_do_one_req()
653 rctx->out_sg = scatterwalk_ffwd(_dst, req->dst, cryp->assoclen); in starfive_aes_aead_do_one_req()
711 ret = starfive_aes_map_sg(cryp, rctx->in_sg, rctx->out_sg); in starfive_aes_aead_do_one_req()
H A Djh7110-cryp.h218 struct scatterlist *out_sg; member
/linux/arch/um/drivers/
H A Dvirtio_pcidev.c95 struct scatterlist out_sg, extra_sg, in_sg; in virtio_pcidev_send_cmd() local
97 [0] = &out_sg, in virtio_pcidev_send_cmd()
150 sg_init_one(&out_sg, cmd, cmd_size); in virtio_pcidev_send_cmd()
/linux/drivers/crypto/aspeed/
H A Daspeed-hace-crypto.c155 struct scatterlist *out_sg; in aspeed_sk_transfer() local
161 out_sg = req->dst; in aspeed_sk_transfer()
164 nbytes = sg_copy_from_buffer(out_sg, rctx->dst_nents, in aspeed_sk_transfer()
/linux/drivers/virtio/
H A Dvirtio_rtc_driver.c416 struct scatterlist out_sg[1]; in viortc_msg_xfer() local
424 sgs[0] = out_sg; in viortc_msg_xfer()
427 sg_init_one(out_sg, msg->req, msg->req_size); in viortc_msg_xfer()

12