| /linux/crypto/ |
| H A D | authencesn.c | 92 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_genicv_tail() local 98 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_genicv_tail() 101 scatterwalk_map_and_copy(hash, dst, assoclen + cryptlen, authsize, 1); in crypto_authenc_esn_genicv_tail() 124 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_genicv() local 134 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 1); in crypto_authenc_esn_genicv() 140 ahash_request_set_crypt(ahreq, dst, hash, assoclen + cryptlen); in crypto_authenc_esn_genicv() 168 unsigned int cryptlen = req->cryptlen; in crypto_authenc_esn_encrypt() local 188 skcipher_request_set_crypt(skreq, src, dst, cryptlen, req->iv); in crypto_authenc_esn_encrypt() 208 unsigned int cryptlen = req->cryptlen - authsize; in crypto_authenc_esn_decrypt_tail() local 221 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_decrypt_tail() [all …]
|
| H A D | chacha20poly1305.c | 46 unsigned int cryptlen; member 87 req->assoclen + rctx->cryptlen, in poly_verify_tag() 107 if (rctx->cryptlen == 0) in chacha_decrypt() 121 rctx->cryptlen, creq->iv); in chacha_decrypt() 141 __le64 cryptlen; in poly_hash() member 152 if (rctx->cryptlen == req->cryptlen) /* encrypting */ in poly_hash() 172 total = rctx->cryptlen; in poly_hash() 181 padlen = -rctx->cryptlen % POLY1305_BLOCK_SIZE; in poly_hash() 185 tail.cryptlen = cpu_to_le64(rctx->cryptlen); in poly_hash() 190 if (rctx->cryptlen != req->cryptlen) in poly_hash() [all …]
|
| H A D | aegis128-core.c | 344 u64 assoclen, u64 cryptlen) in crypto_aegis128_final() argument 347 u64 cryptbits = cryptlen * 8; in crypto_aegis128_final() 392 unsigned int cryptlen = req->cryptlen; in crypto_aegis128_encrypt_generic() local 401 crypto_aegis128_final(&state, &tag, req->assoclen, cryptlen); in crypto_aegis128_encrypt_generic() 403 scatterwalk_map_and_copy(tag.bytes, req->dst, req->assoclen + cryptlen, in crypto_aegis128_encrypt_generic() 414 unsigned int cryptlen = req->cryptlen - authsize; in crypto_aegis128_decrypt_generic() local 419 scatterwalk_map_and_copy(tag.bytes, req->src, req->assoclen + cryptlen, in crypto_aegis128_decrypt_generic() 427 crypto_aegis128_final(&state, &tag, req->assoclen, cryptlen); in crypto_aegis128_decrypt_generic() 453 unsigned int cryptlen = req->cryptlen; in crypto_aegis128_encrypt_simd() local 465 crypto_aegis128_final_simd(&state, &tag, req->assoclen, cryptlen, 0); in crypto_aegis128_encrypt_simd() [all …]
|
| H A D | ccm.c | 125 unsigned int cryptlen) in format_input() argument 143 return set_msg_len(info + 16 - l, cryptlen, l); in format_input() 166 unsigned int cryptlen) in crypto_ccm_auth() argument 179 err = format_input(odata, req, cryptlen); in crypto_ccm_auth() 215 cryptlen += ilen; in crypto_ccm_auth() 218 ahash_request_set_crypt(ahreq, plain, odata, cryptlen); in crypto_ccm_auth() 233 req->assoclen + req->cryptlen, in crypto_ccm_encrypt_done() 289 unsigned int cryptlen = req->cryptlen; in crypto_ccm_encrypt() local 298 err = crypto_ccm_auth(req, sg_next(pctx->src), cryptlen); in crypto_ccm_encrypt() 309 skcipher_request_set_crypt(skreq, pctx->src, dst, cryptlen + 16, iv); in crypto_ccm_encrypt() [all …]
|
| H A D | gcm.c | 56 unsigned int cryptlen; member 174 unsigned int cryptlen) in crypto_gcm_init_crypt() argument 186 cryptlen + sizeof(pctx->auth_tag), in crypto_gcm_init_crypt() 227 lengths.b = cpu_to_be64(gctx->cryptlen * 8); in gcm_hash_len() 287 remain = gcm_remain(gctx->cryptlen); in gcm_hash_crypt_continue() 316 if (gctx->cryptlen) in gcm_hash_assoc_remain_continue() 318 gctx->src, gctx->cryptlen, flags) ?: in gcm_hash_assoc_remain_continue() 413 req->assoclen + req->cryptlen, in gcm_enc_copy_hash() 424 gctx->cryptlen = req->cryptlen; in gcm_encrypt_continue() 452 crypto_gcm_init_crypt(req, req->cryptlen); in crypto_gcm_encrypt() [all …]
|
| H A D | echainiv.c | 36 if (req->cryptlen < ivsize) in echainiv_encrypt() 45 req->assoclen + req->cryptlen); in echainiv_encrypt() 50 req->cryptlen, info); in echainiv_encrypt() 82 if (req->cryptlen < ivsize) in echainiv_decrypt() 92 req->cryptlen - ivsize, req->iv); in echainiv_decrypt()
|
| H A D | xts.c | 87 const bool cts = (req->cryptlen % XTS_BLOCK_SIZE); in xts_xor_tweak() 164 int offset = req->cryptlen & ~(XTS_BLOCK_SIZE - 1); in xts_cts_final() 167 int tail = req->cryptlen % XTS_BLOCK_SIZE; in xts_cts_final() 209 if (!err && unlikely(req->cryptlen % XTS_BLOCK_SIZE)) { in xts_encrypt_done() 229 if (!err && unlikely(req->cryptlen % XTS_BLOCK_SIZE)) { in xts_decrypt_done() 247 if (req->cryptlen < XTS_BLOCK_SIZE) in xts_init_crypt() 253 req->cryptlen & ~(XTS_BLOCK_SIZE - 1), NULL); in xts_init_crypt() 272 if (err || likely((req->cryptlen % XTS_BLOCK_SIZE) == 0)) in xts_encrypt() 289 if (err || likely((req->cryptlen % XTS_BLOCK_SIZE) == 0)) in xts_decrypt()
|
| H A D | seqiv.c | 59 if (req->cryptlen < ivsize) in seqiv_aead_encrypt() 70 req->assoclen + req->cryptlen); in seqiv_aead_encrypt() 87 req->cryptlen - ivsize, info); in seqiv_aead_encrypt() 108 if (req->cryptlen < ivsize + crypto_aead_authsize(geniv)) in seqiv_aead_decrypt() 118 req->cryptlen - ivsize, req->iv); in seqiv_aead_decrypt()
|
| /linux/arch/arm/crypto/ |
| H A D | aes-ce-glue.c | 269 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2; in cts_cbc_encrypt() 280 if (req->cryptlen <= AES_BLOCK_SIZE) { in cts_cbc_encrypt() 281 if (req->cryptlen < AES_BLOCK_SIZE) in cts_cbc_encrypt() 296 if (req->cryptlen == AES_BLOCK_SIZE) in cts_cbc_encrypt() 299 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt() 302 subreq.cryptlen); in cts_cbc_encrypt() 307 req->cryptlen - cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_encrypt() 327 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2; in cts_cbc_decrypt() 338 if (req->cryptlen <= AES_BLOCK_SIZE) { in cts_cbc_decrypt() 339 if (req->cryptlen < AES_BLOCK_SIZE) in cts_cbc_decrypt() [all …]
|
| /linux/drivers/crypto/allwinner/sun8i-ce/ |
| H A D | sun8i-ce-cipher.c | 34 if (sg_nents_for_len(areq->src, areq->cryptlen) > MAX_SG || in sun8i_ce_cipher_need_fallback() 35 sg_nents_for_len(areq->dst, areq->cryptlen) > MAX_SG) { in sun8i_ce_cipher_need_fallback() 42 if (areq->cryptlen < crypto_skcipher_ivsize(tfm)) { in sun8i_ce_cipher_need_fallback() 49 if (areq->cryptlen == 0) { in sun8i_ce_cipher_need_fallback() 56 if (areq->cryptlen % 16) { in sun8i_ce_cipher_need_fallback() 63 len = areq->cryptlen; in sun8i_ce_cipher_need_fallback() 83 len = areq->cryptlen; in sun8i_ce_cipher_need_fallback() 126 areq->cryptlen, areq->iv); in sun8i_ce_cipher_fallback() 150 int ns = sg_nents_for_len(areq->src, areq->cryptlen); in sun8i_ce_cipher_prepare() 151 int nd = sg_nents_for_len(areq->dst, areq->cryptlen); in sun8i_ce_cipher_prepare() [all …]
|
| /linux/arch/x86/crypto/ |
| H A D | aegis128-aesni-glue.c | 61 unsigned int cryptlen); 179 unsigned int cryptlen, bool enc) in crypto_aegis128_aesni_crypt() argument 200 aegis128_aesni_final(&state, tag_xor, req->assoclen, cryptlen); in crypto_aegis128_aesni_crypt() 210 unsigned int cryptlen = req->cryptlen; in crypto_aegis128_aesni_encrypt() local 213 err = crypto_aegis128_aesni_crypt(req, &tag, cryptlen, true); in crypto_aegis128_aesni_encrypt() 218 req->assoclen + cryptlen, authsize, 1); in crypto_aegis128_aesni_encrypt() 229 unsigned int cryptlen = req->cryptlen - authsize; in crypto_aegis128_aesni_decrypt() local 233 req->assoclen + cryptlen, authsize, 0); in crypto_aegis128_aesni_decrypt() 235 err = crypto_aegis128_aesni_crypt(req, &tag, cryptlen, false); in crypto_aegis128_aesni_decrypt()
|
| H A D | aesni-intel_glue.c | 213 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2; in cts_cbc_encrypt() 224 if (req->cryptlen <= AES_BLOCK_SIZE) { in cts_cbc_encrypt() 225 if (req->cryptlen < AES_BLOCK_SIZE) in cts_cbc_encrypt() 239 if (req->cryptlen == AES_BLOCK_SIZE) in cts_cbc_encrypt() 242 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt() 245 subreq.cryptlen); in cts_cbc_encrypt() 250 req->cryptlen - cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_encrypt() 269 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2; in cts_cbc_decrypt() 280 if (req->cryptlen <= AES_BLOCK_SIZE) { in cts_cbc_decrypt() 281 if (req->cryptlen < AES_BLOCK_SIZE) in cts_cbc_decrypt() [all …]
|
| /linux/drivers/crypto/qce/ |
| H A D | aead.c | 71 totallen = req->cryptlen + req->assoclen; in qce_aead_done() 80 totallen = req->cryptlen + req->assoclen - ctx->authsize; in qce_aead_done() 124 totallen = rctx->cryptlen + assoclen; in qce_aead_prepare_dst_buf() 152 sg = qce_sgtable_add(&rctx->dst_tbl, msg_sg, rctx->cryptlen); in qce_aead_prepare_dst_buf() 157 totallen = rctx->cryptlen + rctx->assoclen; in qce_aead_prepare_dst_buf() 192 unsigned int adata_header_len, cryptlen, totallen; in qce_aead_ccm_prepare_buf_assoclen() local 198 cryptlen = rctx->cryptlen + ctx->authsize; in qce_aead_ccm_prepare_buf_assoclen() 200 cryptlen = rctx->cryptlen; in qce_aead_ccm_prepare_buf_assoclen() 201 totallen = cryptlen + req->assoclen; in qce_aead_ccm_prepare_buf_assoclen() 257 sg = qce_sgtable_add(&rctx->src_tbl, msg_sg, cryptlen); in qce_aead_ccm_prepare_buf_assoclen() [all …]
|
| H A D | skcipher.c | 79 rctx->cryptlen = req->cryptlen; in qce_skcipher_async_req_handle() 85 rctx->src_nents = sg_nents_for_len(req->src, req->cryptlen); in qce_skcipher_async_req_handle() 87 rctx->dst_nents = sg_nents_for_len(req->dst, req->cryptlen); in qce_skcipher_async_req_handle() 110 sg = qce_sgtable_add(&rctx->dst_tbl, req->dst, req->cryptlen); in qce_skcipher_async_req_handle() 271 if (!req->cryptlen) in qce_skcipher_crypt() 279 if (!IS_ALIGNED(req->cryptlen, blocksize)) in qce_skcipher_crypt() 292 (IS_XTS(rctx->flags) && ((req->cryptlen <= aes_sw_max_len) || in qce_skcipher_crypt() 293 (req->cryptlen > QCE_SECTOR_SIZE && in qce_skcipher_crypt() 294 req->cryptlen % QCE_SECTOR_SIZE))))) { in qce_skcipher_crypt() 301 req->dst, req->cryptlen, req->iv); in qce_skcipher_crypt()
|
| /linux/drivers/crypto/allwinner/sun8i-ss/ |
| H A D | sun8i-ss-cipher.c | 33 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sun8i_ss_need_fallback() 38 if (sg_nents_for_len(areq->src, areq->cryptlen) > 8 || in sun8i_ss_need_fallback() 39 sg_nents_for_len(areq->dst, areq->cryptlen) > 8) { in sun8i_ss_need_fallback() 44 len = areq->cryptlen; in sun8i_ss_need_fallback() 59 len = areq->cryptlen; in sun8i_ss_need_fallback() 112 areq->cryptlen, areq->iv); in sun8i_ss_cipher_fallback() 128 unsigned int len = areq->cryptlen; in sun8i_ss_setup_ivs() 137 offset = areq->cryptlen - ivsize; in sun8i_ss_setup_ivs() 197 int nsgs = sg_nents_for_len(areq->src, areq->cryptlen); in sun8i_ss_cipher() 198 int nsgd = sg_nents_for_len(areq->dst, areq->cryptlen); in sun8i_ss_cipher() [all …]
|
| /linux/drivers/crypto/ |
| H A D | omap-aes-gcm.c | 93 int alen, clen, cryptlen, assoclen, ret; in omap_aes_gcm_copy_buffers() local 101 cryptlen = req->cryptlen; in omap_aes_gcm_copy_buffers() 107 cryptlen -= authlen; in omap_aes_gcm_copy_buffers() 110 clen = ALIGN(cryptlen, AES_BLOCK_SIZE); in omap_aes_gcm_copy_buffers() 112 nsg = !!(assoclen && cryptlen); in omap_aes_gcm_copy_buffers() 130 if (cryptlen) { in omap_aes_gcm_copy_buffers() 136 ret = omap_crypto_align_sg(&tmp, cryptlen, in omap_aes_gcm_copy_buffers() 148 dd->total = cryptlen; in omap_aes_gcm_copy_buffers() 161 if (cryptlen) { in omap_aes_gcm_copy_buffers() 162 ret = omap_crypto_align_sg(&dd->out_sg, cryptlen, in omap_aes_gcm_copy_buffers() [all …]
|
| /linux/drivers/crypto/allwinner/sun4i-ss/ |
| H A D | sun4i-ss-cipher.c | 30 unsigned int ileft = areq->cryptlen; in sun4i_ss_opti_poll() 31 unsigned int oleft = areq->cryptlen; in sun4i_ss_opti_poll() 41 if (!areq->cryptlen) in sun4i_ss_opti_poll() 51 areq->cryptlen - ivsize, ivsize, 0); in sun4i_ss_opti_poll() 57 algt->stat_bytes += areq->cryptlen; in sun4i_ss_opti_poll() 74 ileft = areq->cryptlen / 4; in sun4i_ss_opti_poll() 75 oleft = areq->cryptlen / 4; in sun4i_ss_opti_poll() 137 scatterwalk_map_and_copy(areq->iv, areq->dst, areq->cryptlen - ivsize, in sun4i_ss_opti_poll() 166 areq->cryptlen, areq->iv); in sun4i_ss_cipher_poll_fallback() 196 unsigned int ileft = areq->cryptlen; in sun4i_ss_cipher_poll() [all …]
|
| /linux/drivers/crypto/cavium/nitrox/ |
| H A D | nitrox_aead.c | 165 creq->gph.param0 = cpu_to_be16(rctx->cryptlen); in nitrox_set_creq() 166 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen); in nitrox_set_creq() 227 rctx->cryptlen = areq->cryptlen; in nitrox_aes_gcm_enc() 229 rctx->srclen = areq->assoclen + areq->cryptlen; in nitrox_aes_gcm_enc() 261 rctx->cryptlen = areq->cryptlen - aead->authsize; in nitrox_aes_gcm_dec() 263 rctx->srclen = areq->cryptlen + areq->assoclen; in nitrox_aes_gcm_dec() 449 aead_rctx->cryptlen = areq->cryptlen; in nitrox_rfc4106_enc() 451 aead_rctx->srclen = aead_rctx->assoclen + aead_rctx->cryptlen; in nitrox_rfc4106_enc() 481 aead_rctx->cryptlen = areq->cryptlen - aead->authsize; in nitrox_rfc4106_dec() 484 areq->cryptlen - GCM_RFC4106_IV_SIZE + areq->assoclen; in nitrox_rfc4106_dec()
|
| /linux/drivers/crypto/tegra/ |
| H A D | tegra-se-aes.c | 60 unsigned int cryptlen; member 109 offset = req->cryptlen - ctx->ivsize; in tegra_cbc_iv_copyback() 124 num = req->cryptlen / ctx->ivsize; in tegra_aes_update_iv() 125 if (req->cryptlen % ctx->ivsize) in tegra_aes_update_iv() 273 rctx->len = req->cryptlen; in tegra_aes_do_one_req() 291 scatterwalk_map_and_copy(rctx->datbuf.buf, req->src, 0, req->cryptlen, 0); in tegra_aes_do_one_req() 322 scatterwalk_map_and_copy(rctx->datbuf.buf, req->dst, 0, req->cryptlen, 1); in tegra_aes_do_one_req() 489 if (!IS_ALIGNED(req->cryptlen, crypto_skcipher_blocksize(tfm))) { in tegra_aes_crypt() 490 dev_dbg(ctx->se->dev, "invalid length (%d)", req->cryptlen); in tegra_aes_crypt() 493 } else if (req->cryptlen < XTS_BLOCK_SIZE) { in tegra_aes_crypt() [all …]
|
| /linux/arch/arm64/crypto/ |
| H A D | aes-glue.c | 276 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2; in cts_cbc_encrypt() 286 if (req->cryptlen <= AES_BLOCK_SIZE) { in cts_cbc_encrypt() 287 if (req->cryptlen < AES_BLOCK_SIZE) in cts_cbc_encrypt() 302 if (req->cryptlen == AES_BLOCK_SIZE) in cts_cbc_encrypt() 305 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt() 308 subreq.cryptlen); in cts_cbc_encrypt() 313 req->cryptlen - cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_encrypt() 332 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2; in cts_cbc_decrypt() 342 if (req->cryptlen <= AES_BLOCK_SIZE) { in cts_cbc_decrypt() 343 if (req->cryptlen < AES_BLOCK_SIZE) in cts_cbc_decrypt() [all …]
|
| /linux/drivers/crypto/gemini/ |
| H A D | sl3516-ce-cipher.c | 34 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sl3516_ce_need_fallback() 119 areq->cryptlen, areq->iv); in sl3516_ce_cipher_fallback() 147 areq->cryptlen, in sl3516_ce_cipher() 179 len = areq->cryptlen; in sl3516_ce_cipher() 189 areq->cryptlen, i, rctx->t_src[i].len, sg->offset, todo); in sl3516_ce_cipher() 196 dev_err(ce->dev, "remaining len %d/%u nr_sgs=%d\n", len, areq->cryptlen, nr_sgs); in sl3516_ce_cipher() 201 len = areq->cryptlen; in sl3516_ce_cipher() 211 areq->cryptlen, i, rctx->t_dst[i].len, sg->offset, todo); in sl3516_ce_cipher() 234 ecb->cipher.algorithm_len = areq->cryptlen; in sl3516_ce_cipher()
|
| /linux/drivers/crypto/intel/keembay/ |
| H A D | keembay-ocs-aes-core.c | 261 if (req->cryptlen % AES_BLOCK_SIZE != 0) in kmb_ocs_sk_validate_input() 268 if (req->cryptlen % AES_BLOCK_SIZE != 0) in kmb_ocs_sk_validate_input() 289 if (req->cryptlen < AES_BLOCK_SIZE) in kmb_ocs_sk_validate_input() 326 req->cryptlen, req->iv); in kmb_ocs_sk_common() 343 if (!req->cryptlen && mode != OCS_MODE_CTS) in kmb_ocs_sk_common() 415 req->cryptlen - iv_size, iv_size, 0); in kmb_ocs_sk_prepare_inplace() 420 req->cryptlen - AES_BLOCK_SIZE, in kmb_ocs_sk_prepare_inplace() 421 req->cryptlen - (2 * AES_BLOCK_SIZE)); in kmb_ocs_sk_prepare_inplace() 434 req->cryptlen, 0); in kmb_ocs_sk_prepare_inplace() 454 rctx->src_nents = sg_nents_for_len(req->src, req->cryptlen); in kmb_ocs_sk_prepare_notinplace() [all …]
|
| /linux/drivers/crypto/aspeed/ |
| H A D | aspeed-hace-crypto.c | 38 areq->cryptlen, areq->iv); in aspeed_crypto_do_fallback() 52 if (areq->cryptlen == 0) in aspeed_crypto_need_fallback() 56 !IS_ALIGNED(areq->cryptlen, DES_BLOCK_SIZE)) in aspeed_crypto_need_fallback() 60 !IS_ALIGNED(areq->cryptlen, AES_BLOCK_SIZE)) in aspeed_crypto_need_fallback() 165 crypto_engine->cipher_addr, req->cryptlen); in aspeed_sk_transfer() 168 "nbytes", nbytes, "cryptlen", req->cryptlen); in aspeed_sk_transfer() 173 "nbytes", nbytes, "req->cryptlen", req->cryptlen, in aspeed_sk_transfer() 193 crypto_engine->cipher_addr, req->cryptlen); in aspeed_sk_start() 196 "nbytes", nbytes, "req->cryptlen", req->cryptlen, in aspeed_sk_start() 202 "nbytes", nbytes, "cryptlen", req->cryptlen); in aspeed_sk_start() [all …]
|
| /linux/drivers/crypto/virtio/ |
| H A D | virtio_crypto_skcipher_algs.c | 341 src_nents = sg_nents_for_len(req->src, req->cryptlen); in __virtio_crypto_skcipher_do_req() 383 cpu_to_le32(req->cryptlen); in __virtio_crypto_skcipher_do_req() 392 dst_len = min_t(unsigned int, req->cryptlen, dst_len); in __virtio_crypto_skcipher_do_req() 394 req->cryptlen, dst_len); in __virtio_crypto_skcipher_do_req() 396 if (unlikely(req->cryptlen + dst_len + ivsize + in __virtio_crypto_skcipher_do_req() 425 req->cryptlen - AES_BLOCK_SIZE, in __virtio_crypto_skcipher_do_req() 475 if (!req->cryptlen) in virtio_crypto_skcipher_encrypt() 477 if (req->cryptlen % AES_BLOCK_SIZE) in virtio_crypto_skcipher_encrypt() 498 if (!req->cryptlen) in virtio_crypto_skcipher_decrypt() 500 if (req->cryptlen % AES_BLOCK_SIZE) in virtio_crypto_skcipher_decrypt() [all …]
|
| /linux/drivers/crypto/stm32/ |
| H A D | stm32-cryp.c | 573 return is_encrypt(cryp) ? cryp->areq->cryptlen : in stm32_cryp_get_input_text_len() 574 cryp->areq->cryptlen - cryp->authsize; in stm32_cryp_get_input_text_len() 1236 if (req->cryptlen % AES_BLOCK_SIZE) in stm32_cryp_aes_ecb_encrypt() 1239 if (req->cryptlen == 0) in stm32_cryp_aes_ecb_encrypt() 1247 if (req->cryptlen % AES_BLOCK_SIZE) in stm32_cryp_aes_ecb_decrypt() 1250 if (req->cryptlen == 0) in stm32_cryp_aes_ecb_decrypt() 1258 if (req->cryptlen % AES_BLOCK_SIZE) in stm32_cryp_aes_cbc_encrypt() 1261 if (req->cryptlen == 0) in stm32_cryp_aes_cbc_encrypt() 1269 if (req->cryptlen % AES_BLOCK_SIZE) in stm32_cryp_aes_cbc_decrypt() 1272 if (req->cryptlen == 0) in stm32_cryp_aes_cbc_decrypt() [all …]
|