| /linux/drivers/net/ovpn/ |
| H A D | crypto_aead.c | 36 crypto_aead_authsize(ks->encrypt); /* Auth Tag */ in ovpn_aead_encap_overhead() 42 const unsigned int tag_size = crypto_aead_authsize(ks->encrypt); in ovpn_aead_encrypt() 133 req = aead_request_alloc(ks->encrypt, GFP_ATOMIC); in ovpn_aead_encrypt() 140 aead_request_set_tfm(req, ks->encrypt); in ovpn_aead_encrypt() 302 crypto_free_aead(ks->encrypt); in ovpn_aead_crypto_key_slot_destroy() 326 if (kc->encrypt.nonce_tail_size != OVPN_NONCE_TAIL_SIZE || in ovpn_aead_crypto_key_slot_new() 335 ks->encrypt = NULL; in ovpn_aead_crypto_key_slot_new() 340 ks->encrypt = ovpn_aead_init("encrypt", alg_name, in ovpn_aead_crypto_key_slot_new() 341 kc->encrypt.cipher_key, in ovpn_aead_crypto_key_slot_new() 342 kc->encrypt.cipher_key_size); in ovpn_aead_crypto_key_slot_new() [all …]
|
| /linux/drivers/crypto/ |
| H A D | padlock-aes.c | 66 struct cword encrypt; member 134 ctx->cword.encrypt.rounds = 10 + (key_len - 16) / 4; in aes_set_key() 135 ctx->cword.decrypt.rounds = ctx->cword.encrypt.rounds; in aes_set_key() 136 ctx->cword.encrypt.ksize = (key_len - 16) / 8; in aes_set_key() 137 ctx->cword.decrypt.ksize = ctx->cword.encrypt.ksize; in aes_set_key() 144 ctx->cword.encrypt.keygen = 1; in aes_set_key() 155 if (&ctx->cword.encrypt == per_cpu(paes_last_cword, cpu) || in aes_set_key() 310 padlock_reset_key(&ctx->cword.encrypt); in padlock_aes_encrypt() 311 ecb_crypt(in, out, ctx->E, &ctx->cword.encrypt, 1); in padlock_aes_encrypt() 312 padlock_store_cword(&ctx->cword.encrypt); in padlock_aes_encrypt() [all …]
|
| /linux/drivers/crypto/caam/ |
| H A D | caamalg.c | 1140 bool all_contig, bool encrypt) in init_aead_job() argument 1152 sh_desc = encrypt ? ctx->sh_desc_enc : ctx->sh_desc_dec; in init_aead_job() 1153 ptr = encrypt ? ctx->sh_desc_enc_dma : ctx->sh_desc_dec_dma; in init_aead_job() 1189 if (encrypt) in init_aead_job() 1201 bool all_contig, bool encrypt) in init_gcm_job() argument 1210 init_aead_job(req, edesc, all_contig, encrypt); in init_gcm_job() 1215 if (encrypt && generic_gcm && !(req->assoclen + req->cryptlen)) in init_gcm_job() 1231 bool encrypt) in init_chachapoly_job() argument 1239 init_aead_job(req, edesc, all_contig, encrypt); in init_chachapoly_job() 1265 bool all_contig, bool encrypt) in init_authenc_job() argument [all …]
|
| H A D | caamalg_qi.c | 942 bool encrypt) in aead_edesc_alloc() argument 962 drv_ctx = get_drv_ctx(ctx, encrypt ? ENCRYPT : DECRYPT); in aead_edesc_alloc() 975 (encrypt ? authsize : 0); in aead_edesc_alloc() 994 dst_len = src_len + (encrypt ? authsize : (-authsize)); in aead_edesc_alloc() 1040 if ((alg->caam.rfc3686 && encrypt) || !alg->caam.geniv) in aead_edesc_alloc() 1136 (encrypt ? ctx->authsize : (-ctx->authsize)); in aead_edesc_alloc() 1161 static inline int aead_crypt(struct aead_request *req, bool encrypt) in aead_crypt() argument 1172 edesc = aead_edesc_alloc(req, encrypt); in aead_crypt() 1255 bool encrypt) in skcipher_edesc_alloc() argument 1272 drv_ctx = get_drv_ctx(ctx, encrypt ? ENCRYPT : DECRYPT); in skcipher_edesc_alloc() [all …]
|
| H A D | caamalg_qi2.c | 353 bool encrypt) in aead_edesc_alloc() argument 384 dst_len = src_len + (encrypt ? authsize : (-authsize)); in aead_edesc_alloc() 429 (encrypt ? authsize : 0); in aead_edesc_alloc() 448 if ((alg->caam.rfc3686 && encrypt) || !alg->caam.geniv) in aead_edesc_alloc() 550 (encrypt ? ctx->authsize : (-ctx->authsize)); in aead_edesc_alloc() 1684 .encrypt = skcipher_encrypt, 1700 .encrypt = skcipher_encrypt, 1716 .encrypt = skcipher_encrypt, 1732 .encrypt = skcipher_encrypt, 1750 .encrypt = skcipher_encrypt, [all …]
|
| /linux/arch/sparc/crypto/ |
| H A D | des_glue.c | 94 static int __ecb_crypt(struct skcipher_request *req, bool encrypt) in __ecb_crypt() argument 106 if (encrypt) in __ecb_crypt() 135 static int __cbc_crypt(struct skcipher_request *req, bool encrypt) in __cbc_crypt() argument 147 if (encrypt) in __cbc_crypt() 152 if (encrypt) in __cbc_crypt() 243 static int __ecb3_crypt(struct skcipher_request *req, bool encrypt) in __ecb3_crypt() argument 256 if (encrypt) in __ecb3_crypt() 289 static int __cbc3_crypt(struct skcipher_request *req, bool encrypt) in __cbc3_crypt() argument 302 if (encrypt) in __cbc3_crypt() 308 if (encrypt) in __cbc3_crypt() [all …]
|
| /linux/drivers/crypto/intel/ixp4xx/ |
| H A D | ixp4xx_crypto.c | 153 bool encrypt; member 163 int encrypt; member 179 struct ix_sa_dir encrypt; member 359 if (req_ctx->encrypt) { in finish_scattered_hmac() 399 if (req_ctx->encrypt) { in one_packet() 623 ret = init_sa_dir(&ctx->encrypt); in init_tfm() 628 free_sa_dir(&ctx->encrypt); in init_tfm() 665 free_sa_dir(&ctx->encrypt); in exit_tfm() 739 static int setup_auth(struct crypto_tfm *tfm, int encrypt, unsigned int authsize, in setup_auth() argument 750 dir = encrypt ? &ctx->encrypt : &ctx->decrypt; in setup_auth() [all …]
|
| /linux/drivers/crypto/inside-secure/eip93/ |
| H A D | eip93-aead.c | 266 .encrypt = eip93_aead_encrypt, 294 .encrypt = eip93_aead_encrypt, 322 .encrypt = eip93_aead_encrypt, 350 .encrypt = eip93_aead_encrypt, 379 .encrypt = eip93_aead_encrypt, 408 .encrypt = eip93_aead_encrypt, 437 .encrypt = eip93_aead_encrypt, 466 .encrypt = eip93_aead_encrypt, 494 .encrypt = eip93_aead_encrypt, 522 .encrypt = eip93_aead_encrypt, [all …]
|
| H A D | eip93-cipher.c | 208 .encrypt = eip93_skcipher_encrypt, 235 .encrypt = eip93_skcipher_encrypt, 262 .encrypt = eip93_skcipher_encrypt, 289 .encrypt = eip93_skcipher_encrypt, 316 .encrypt = eip93_skcipher_encrypt, 342 .encrypt = eip93_skcipher_encrypt, 368 .encrypt = eip93_skcipher_encrypt, 394 .encrypt = eip93_skcipher_encrypt,
|
| /linux/drivers/crypto/tegra/ |
| H A D | tegra-se-aes.c | 38 bool encrypt; member 62 bool encrypt; member 111 if (rctx->encrypt) in tegra_cbc_iv_copyback() 132 static int tegra234_aes_crypto_cfg(u32 alg, bool encrypt) in tegra234_aes_crypto_cfg() argument 141 if (encrypt) in tegra234_aes_crypto_cfg() 146 if (encrypt) in tegra234_aes_crypto_cfg() 151 if (encrypt) in tegra234_aes_crypto_cfg() 168 static int tegra234_aes_cfg(u32 alg, bool encrypt) in tegra234_aes_cfg() argument 175 if (encrypt) in tegra234_aes_cfg() 181 if (encrypt) in tegra234_aes_cfg() [all …]
|
| /linux/drivers/crypto/virtio/ |
| H A D | virtio_crypto_skcipher_algs.c | 32 bool encrypt; member 113 int encrypt) in virtio_crypto_alg_skcipher_init_session() argument 117 int op = encrypt ? VIRTIO_CRYPTO_OP_ENCRYPT : VIRTIO_CRYPTO_OP_DECRYPT; in virtio_crypto_alg_skcipher_init_session() 178 if (encrypt) in virtio_crypto_alg_skcipher_init_session() 192 int encrypt) in virtio_crypto_alg_skcipher_close_session() argument 217 if (encrypt) in virtio_crypto_alg_skcipher_close_session() 369 if (vc_sym_req->encrypt) { in __virtio_crypto_skcipher_do_req() 423 if (!vc_sym_req->encrypt) in __virtio_crypto_skcipher_do_req() 482 vc_sym_req->encrypt = true; in virtio_crypto_skcipher_encrypt() 505 vc_sym_req->encrypt = false; in virtio_crypto_skcipher_decrypt() [all …]
|
| /linux/arch/arm64/crypto/ |
| H A D | sm4-ce-glue.c | 150 struct sm4_ctx *ctx, bool encrypt) in sm4_cbc_crypt() argument 168 if (encrypt) in sm4_cbc_crypt() 199 static int sm4_cbc_cts_crypt(struct skcipher_request *req, bool encrypt) in sm4_cbc_cts_crypt() argument 215 return sm4_cbc_crypt(req, ctx, encrypt); in sm4_cbc_cts_crypt() 228 err = sm4_cbc_crypt(&subreq, ctx, encrypt); in sm4_cbc_cts_crypt() 248 if (encrypt) in sm4_cbc_cts_crypt() 310 static int sm4_xts_crypt(struct skcipher_request *req, bool encrypt) in sm4_xts_crypt() argument 354 if (encrypt) in sm4_xts_crypt() 386 if (encrypt) in sm4_xts_crypt() 422 .encrypt = sm4_ecb_encrypt, [all …]
|
| /linux/net/ceph/ |
| H A D | crypto.c | 264 static int ceph_aes_crypt(const struct ceph_crypto_key *key, bool encrypt, in ceph_aes_crypt() argument 272 int crypt_len = encrypt ? in_len + pad_byte : in_len; in ceph_aes_crypt() 276 if (encrypt) in ceph_aes_crypt() 293 if (encrypt) in ceph_aes_crypt() 300 encrypt ? "en" : "de", ret); in ceph_aes_crypt() 308 if (encrypt) { in ceph_aes_crypt() 389 int ceph_crypt(const struct ceph_crypto_key *key, int usage_slot, bool encrypt, in ceph_crypt() argument 397 return ceph_aes_crypt(key, encrypt, buf, buf_len, in_len, in ceph_crypt() 400 return encrypt ? in ceph_crypt()
|
| /linux/drivers/crypto/ccree/ |
| H A D | cc_cipher.c | 1010 .encrypt = cc_cipher_encrypt, 1028 .encrypt = cc_cipher_encrypt, 1046 .encrypt = cc_cipher_encrypt, 1064 .encrypt = cc_cipher_encrypt, 1082 .encrypt = cc_cipher_encrypt, 1100 .encrypt = cc_cipher_encrypt, 1122 .encrypt = cc_cipher_encrypt, 1139 .encrypt = cc_cipher_encrypt, 1156 .encrypt = cc_cipher_encrypt, 1173 .encrypt = cc_cipher_encrypt, [all …]
|
| /linux/drivers/net/wireless/intel/ipw2x00/ |
| H A D | libipw_tx.c | 248 int encrypt, host_encrypt, host_encrypt_msdu; in libipw_xmit() local 284 encrypt = !(ether_type == htons(ETH_P_PAE) && ieee->ieee802_1x) && in libipw_xmit() 285 ieee->sec.encrypt; in libipw_xmit() 287 host_encrypt = ieee->host_encrypt && encrypt && crypt; in libipw_xmit() 288 host_encrypt_msdu = ieee->host_encrypt_msdu && encrypt && crypt; in libipw_xmit() 290 if (!encrypt && ieee->ieee802_1x && in libipw_xmit() 418 txb->encrypted = encrypt; in libipw_xmit()
|
| /linux/drivers/crypto/inside-secure/ |
| H A D | safexcel_cipher.c | 1296 .encrypt = safexcel_encrypt, 1333 .encrypt = safexcel_encrypt, 1407 .encrypt = safexcel_encrypt, 1469 .encrypt = safexcel_encrypt, 1508 .encrypt = safexcel_encrypt, 1568 .encrypt = safexcel_encrypt, 1607 .encrypt = safexcel_encrypt, 1680 .encrypt = safexcel_aead_encrypt, 1716 .encrypt = safexcel_aead_encrypt, 1752 .encrypt = safexcel_aead_encrypt, [all …]
|
| /linux/lib/crypto/ |
| H A D | chacha20poly1305.c | 215 int encrypt) in chacha20poly1305_crypt_sg_inplace() argument 264 if (!encrypt) in chacha20poly1305_crypt_sg_inplace() 294 if (encrypt) in chacha20poly1305_crypt_sg_inplace() 307 if (encrypt) { in chacha20poly1305_crypt_sg_inplace() 323 sg_copy_buffer(src, sg_nents(src), b.mac[encrypt], in chacha20poly1305_crypt_sg_inplace() 324 sizeof(b.mac[1]), src_len, !encrypt); in chacha20poly1305_crypt_sg_inplace() 325 ret = encrypt || in chacha20poly1305_crypt_sg_inplace()
|
| /linux/drivers/hv/ |
| H A D | hv.c | 145 static int hv_free_page(void **page, bool encrypt, const char *note) in hv_free_page() argument 152 if (encrypt) in hv_free_page() 161 note, ret, encrypt); in hv_free_page() 246 const bool encrypt = !vmbus_is_confidential(); in hv_synic_free() local 254 encrypt, "post msg"); in hv_synic_free() 257 encrypt, "hypervisor SynIC event"); in hv_synic_free() 259 encrypt, "hypervisor SynIC msg"); in hv_synic_free()
|
| /linux/drivers/crypto/hisilicon/sec2/ |
| H A D | sec_crypto.c | 121 bool encrypt); 123 struct skcipher_request *sreq, bool encrypt); 250 req->c_req.encrypt); in sec_alg_send_backlog_soft() 253 req->c_req.encrypt); in sec_alg_send_backlog_soft() 991 if (!c_req->encrypt && ctx->alg_type == SEC_AEAD) { in GEN_SEC_SETKEY_FUNC() 1182 if (!c_req->encrypt && ctx->alg_type == SEC_AEAD) { in sec_cipher_map_inner() 1250 if (!c_req->encrypt && is_aead) { in sec_cipher_map() 1550 if (c_req->encrypt) in sec_skcipher_bd_fill() 1602 if (c_req->encrypt) in sec_skcipher_bd_fill_v3() 1653 sgl = req->c_req.encrypt ? sk_req->dst : sk_req->src; in sec_update_iv() [all …]
|
| /linux/arch/x86/crypto/ |
| H A D | sm4_aesni_avx2_glue.c | 61 .encrypt = sm4_avx_ecb_encrypt, 77 .encrypt = sm4_cbc_encrypt, 94 .encrypt = ctr_crypt,
|
| /linux/drivers/fpga/ |
| H A D | socfpga-a10.c | 170 bool encrypt, bool compress) in socfpga_a10_fpga_get_cd_ratio() argument 188 if (!compress && !encrypt) in socfpga_a10_fpga_get_cd_ratio() 209 int encrypt, compress; in socfpga_a10_fpga_set_cdratio() local 211 encrypt = socfpga_a10_fpga_encrypted((u32 *)buf, count / 4); in socfpga_a10_fpga_set_cdratio() 212 if (encrypt < 0) in socfpga_a10_fpga_set_cdratio() 219 cd_ratio = socfpga_a10_fpga_get_cd_ratio(cfg_width, encrypt, compress); in socfpga_a10_fpga_set_cdratio()
|
| /linux/drivers/crypto/ccp/ |
| H A D | ccp-crypto-aes.c | 67 static int ccp_aes_crypt(struct skcipher_request *req, bool encrypt) in ccp_aes_crypt() argument 99 (encrypt) ? CCP_AES_ACTION_ENCRYPT : CCP_AES_ACTION_DECRYPT; in ccp_aes_crypt() 159 static int ccp_aes_rfc3686_crypt(struct skcipher_request *req, bool encrypt) in ccp_aes_rfc3686_crypt() argument 180 return ccp_aes_crypt(req, encrypt); in ccp_aes_rfc3686_crypt() 207 .encrypt = ccp_aes_encrypt, 225 .encrypt = ccp_aes_rfc3686_encrypt,
|
| /linux/arch/arm/crypto/ |
| H A D | aes-neonbs-glue.c | 255 static int __xts_crypt(struct skcipher_request *req, bool encrypt, in __xts_crypt() argument 289 int reorder_last_tweak = !encrypt && tail > 0; in __xts_crypt() 316 if (encrypt) in __xts_crypt() 350 .encrypt = ecb_encrypt, 365 .encrypt = cbc_encrypt, 381 .encrypt = ctr_encrypt, 396 .encrypt = xts_encrypt,
|
| /linux/drivers/crypto/cavium/nitrox/ |
| H A D | nitrox_skcipher.c | 399 .encrypt = nitrox_aes_encrypt, 418 .encrypt = nitrox_aes_encrypt, 437 .encrypt = nitrox_aes_encrypt, 458 .encrypt = nitrox_aes_encrypt, 475 .encrypt = nitrox_aes_encrypt, 494 .encrypt = nitrox_3des_encrypt, 513 .encrypt = nitrox_3des_encrypt,
|
| /linux/drivers/crypto/hisilicon/sec/ |
| H A D | sec_algs.c | 630 *sec_alg_alloc_and_fill_el(struct sec_bd_info *template, int encrypt, in sec_alg_alloc_and_fill_el() argument 648 if (encrypt) in sec_alg_alloc_and_fill_el() 708 bool encrypt) in sec_alg_skcipher_crypto() argument 774 encrypt ? 1 : 0, in sec_alg_skcipher_crypto() 948 .encrypt = sec_alg_skcipher_encrypt, 968 .encrypt = sec_alg_skcipher_encrypt, 988 .encrypt = sec_alg_skcipher_encrypt, 1008 .encrypt = sec_alg_skcipher_encrypt, 1029 .encrypt = sec_alg_skcipher_encrypt, 1049 .encrypt = sec_alg_skcipher_encrypt, [all …]
|