| /linux/drivers/crypto/nx/ |
| H A D | nx-aes-gcm.c | 28 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in gcm_aes_nx_set_key() local 35 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_128); in gcm_aes_nx_set_key() 40 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_192); in gcm_aes_nx_set_key() 45 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_256); in gcm_aes_nx_set_key() 53 csbcpb->cpb.hdr.mode = NX_MODE_AES_GCM; in gcm_aes_nx_set_key() 54 memcpy(csbcpb->cpb.aes_gcm.key, in_key, key_len); in gcm_aes_nx_set_key() 171 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in gmac() local 178 csbcpb->cpb.hdr.mode = NX_MODE_AES_GMAC; in gmac() 180 NX_CPB_FDM(csbcpb) &= ~NX_FDM_CONTINUATION; in gmac() 189 memcpy(csbcpb->cpb.aes_gcm.iv_or_cnt, iv, AES_BLOCK_SIZE); in gmac() [all …]
|
| H A D | nx-aes-xcbc.c | 32 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in nx_xcbc_set_key() local 42 memcpy(csbcpb->cpb.aes_xcbc.key, in_key, key_len); in nx_xcbc_set_key() 61 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in nx_xcbc_empty() local 69 csbcpb->cpb.hdr.mode = NX_MODE_AES_ECB; in nx_xcbc_empty() 70 memcpy(key, csbcpb->cpb.aes_xcbc.key, AES_BLOCK_SIZE); in nx_xcbc_empty() 71 memcpy(csbcpb->cpb.aes_ecb.key, key, AES_BLOCK_SIZE); in nx_xcbc_empty() 72 NX_CPB_FDM(csbcpb) |= NX_FDM_ENDE_ENCRYPT; in nx_xcbc_empty() 106 memcpy(csbcpb->cpb.aes_ecb.key, keys[0], AES_BLOCK_SIZE); in nx_xcbc_empty() 130 csbcpb->cpb.hdr.mode = NX_MODE_AES_XCBC_MAC; in nx_xcbc_empty() 131 memcpy(csbcpb->cpb.aes_xcbc.key, key, AES_BLOCK_SIZE); in nx_xcbc_empty() [all …]
|
| H A D | nx-aes-ccm.c | 28 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in ccm_aes_nx_set_key() local 35 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_128); in ccm_aes_nx_set_key() 43 csbcpb->cpb.hdr.mode = NX_MODE_AES_CCM; in ccm_aes_nx_set_key() 44 memcpy(csbcpb->cpb.aes_ccm.key, in_key, key_len); in ccm_aes_nx_set_key() 186 b0 = nx_ctx->csbcpb->cpb.aes_ccm.in_pat_or_b0; in generate_pat() 191 b0 = nx_ctx->csbcpb->cpb.aes_ccm.in_pat_or_b0; in generate_pat() 252 NX_CPB_FDM(nx_ctx->csbcpb) |= NX_FDM_ENDE_ENCRYPT; in generate_pat() 253 NX_CPB_FDM(nx_ctx->csbcpb) |= NX_FDM_INTERMEDIATE; in generate_pat() 255 result = nx_ctx->csbcpb->cpb.aes_ccm.out_pat_or_mac; in generate_pat() 329 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in ccm_nx_decrypt() local [all …]
|
| H A D | nx-aes-cbc.c | 28 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in cbc_aes_nx_set_key() local 34 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_128); in cbc_aes_nx_set_key() 38 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_192); in cbc_aes_nx_set_key() 42 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_256); in cbc_aes_nx_set_key() 49 csbcpb->cpb.hdr.mode = NX_MODE_AES_CBC; in cbc_aes_nx_set_key() 50 memcpy(csbcpb->cpb.aes_cbc.key, in_key, key_len); in cbc_aes_nx_set_key() 60 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in cbc_aes_nx_crypt() local 68 NX_CPB_FDM(csbcpb) |= NX_FDM_ENDE_ENCRYPT; in cbc_aes_nx_crypt() 70 NX_CPB_FDM(csbcpb) &= ~NX_FDM_ENDE_ENCRYPT; in cbc_aes_nx_crypt() 77 csbcpb->cpb.aes_cbc.iv); in cbc_aes_nx_crypt() [all …]
|
| H A D | nx-aes-ecb.c | 28 struct nx_csbcpb *csbcpb = (struct nx_csbcpb *)nx_ctx->csbcpb; in ecb_aes_nx_set_key() local 34 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_128); in ecb_aes_nx_set_key() 38 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_192); in ecb_aes_nx_set_key() 42 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_256); in ecb_aes_nx_set_key() 49 csbcpb->cpb.hdr.mode = NX_MODE_AES_ECB; in ecb_aes_nx_set_key() 50 memcpy(csbcpb->cpb.aes_ecb.key, in_key, key_len); in ecb_aes_nx_set_key() 60 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in ecb_aes_nx_crypt() local 68 NX_CPB_FDM(csbcpb) |= NX_FDM_ENDE_ENCRYPT; in ecb_aes_nx_crypt() 70 NX_CPB_FDM(csbcpb) &= ~NX_FDM_ENDE_ENCRYPT; in ecb_aes_nx_crypt() 91 atomic64_add(be32_to_cpu(csbcpb->csb.processed_byte_count), in ecb_aes_nx_crypt()
|
| H A D | nx-aes-ctr.c | 29 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in ctr_aes_nx_set_key() local 35 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_128); in ctr_aes_nx_set_key() 39 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_192); in ctr_aes_nx_set_key() 43 NX_CPB_SET_KEY_SIZE(csbcpb, NX_KS_AES_256); in ctr_aes_nx_set_key() 50 csbcpb->cpb.hdr.mode = NX_MODE_AES_CTR; in ctr_aes_nx_set_key() 51 memcpy(csbcpb->cpb.aes_ctr.key, in_key, key_len); in ctr_aes_nx_set_key() 78 struct nx_csbcpb *csbcpb = nx_ctx->csbcpb; in ctr_aes_nx_crypt() local 90 csbcpb->cpb.aes_ctr.iv); in ctr_aes_nx_crypt() 104 memcpy(iv, csbcpb->cpb.aes_cbc.cv, AES_BLOCK_SIZE); in ctr_aes_nx_crypt() 107 atomic64_add(be32_to_cpu(csbcpb->csb.processed_byte_count), in ctr_aes_nx_crypt()
|
| H A D | nx-sha512.c | 40 NX_CPB_SET_DIGEST_SIZE(nx_ctx->csbcpb, NX_DS_SHA512); in nx_crypto_ctx_sha512_init() 68 struct nx_csbcpb *csbcpb = (struct nx_csbcpb *)nx_ctx->csbcpb; in nx_sha512_update() local 78 memcpy(csbcpb->cpb.sha512.message_digest, sctx->state, SHA512_DIGEST_SIZE); in nx_sha512_update() 79 NX_CPB_FDM(csbcpb) |= NX_FDM_INTERMEDIATE; in nx_sha512_update() 80 NX_CPB_FDM(csbcpb) |= NX_FDM_CONTINUATION; in nx_sha512_update() 115 memcpy(csbcpb->cpb.sha512.input_partial_digest, in nx_sha512_update() 116 csbcpb->cpb.sha512.message_digest, in nx_sha512_update() 138 memcpy(sctx->state, csbcpb->cpb.sha512.message_digest, SHA512_DIGEST_SIZE); in nx_sha512_update() 149 struct nx_csbcpb *csbcpb = (struct nx_csbcpb *)nx_ctx->csbcpb; in nx_sha512_finup() local 167 memcpy(csbcpb->cpb.sha512.input_partial_digest, sctx->state, SHA512_DIGEST_SIZE); in nx_sha512_finup() [all …]
|
| H A D | nx-sha256.c | 40 NX_CPB_SET_DIGEST_SIZE(nx_ctx->csbcpb, NX_DS_SHA256); in nx_crypto_ctx_sha256_init() 67 struct nx_csbcpb *csbcpb = (struct nx_csbcpb *)nx_ctx->csbcpb; in nx_sha256_update() local 77 memcpy(csbcpb->cpb.sha256.message_digest, sctx->state, SHA256_DIGEST_SIZE); in nx_sha256_update() 78 NX_CPB_FDM(csbcpb) |= NX_FDM_INTERMEDIATE; in nx_sha256_update() 79 NX_CPB_FDM(csbcpb) |= NX_FDM_CONTINUATION; in nx_sha256_update() 114 memcpy(csbcpb->cpb.sha256.input_partial_digest, in nx_sha256_update() 115 csbcpb->cpb.sha256.message_digest, in nx_sha256_update() 135 memcpy(sctx->state, csbcpb->cpb.sha256.message_digest, SHA256_DIGEST_SIZE); in nx_sha256_update() 146 struct nx_csbcpb *csbcpb = (struct nx_csbcpb *)nx_ctx->csbcpb; in nx_sha256_finup() local 163 memcpy(csbcpb->cpb.sha256.input_partial_digest, sctx->state, SHA256_DIGEST_SIZE); in nx_sha256_finup() [all …]
|
| H A D | nx-common-pseries.c | 96 struct nx_csbcpb csbcpb; member 313 struct nx_csbcpb *csbcpb; in nx842_pseries_compress() local 346 csbcpb = &workmem->csbcpb; in nx842_pseries_compress() 347 memset(csbcpb, 0, sizeof(*csbcpb)); in nx842_pseries_compress() 348 op.csbcpb = nx842_get_pa(csbcpb); in nx842_pseries_compress() 390 ret = nx842_validate_result(dev, &csbcpb->csb); in nx842_pseries_compress() 394 *outlen = be32_to_cpu(csbcpb->csb.processed_byte_count); in nx842_pseries_compress() 443 struct nx_csbcpb *csbcpb; in nx842_pseries_decompress() local 478 csbcpb = &workmem->csbcpb; in nx842_pseries_decompress() 479 memset(csbcpb, 0, sizeof(*csbcpb)); in nx842_pseries_decompress() [all …]
|
| H A D | nx.c | 286 nx_ctx->csbcpb->csb.valid |= NX_CSB_VALID_BIT; in nx_ctx_init() 289 nx_ctx->op.csbcpb = __pa(nx_ctx->csbcpb); in nx_ctx_init() 297 nx_ctx->op_aead.csbcpb = __pa(nx_ctx->csbcpb_aead); in nx_ctx_init() 651 nx_ctx->csbcpb = (struct nx_csbcpb *)(round_up((u64)nx_ctx->kmem, in nx_crypto_ctx_init() 653 nx_ctx->in_sg = (struct nx_sg *)((u8 *)nx_ctx->csbcpb + NX_PAGE_SIZE); in nx_crypto_ctx_init() 727 nx_ctx->csbcpb = NULL; in nx_crypto_ctx_exit()
|
| H A D | nx.h | 125 struct nx_csbcpb *csbcpb; /* aligned page given to phyp @ hcall time */ member
|
| /linux/arch/powerpc/include/asm/ |
| H A D | vio.h | 74 u64 csbcpb; member
|
| /linux/arch/powerpc/platforms/pseries/ |
| H A D | vio.c | 1112 op->outlen, op->csbcpb); in vio_h_cop_sync()
|