| /linux/arch/s390/crypto/ |
| H A D | ghash_s390.c | 21 u8 key[GHASH_BLOCK_SIZE]; 25 u8 icv[GHASH_BLOCK_SIZE]; 26 u8 key[GHASH_BLOCK_SIZE]; 35 memcpy(dctx->key, ctx->key, GHASH_BLOCK_SIZE); in ghash_init() 45 if (keylen != GHASH_BLOCK_SIZE) in ghash_setkey() 48 memcpy(ctx->key, key, GHASH_BLOCK_SIZE); in ghash_setkey() 59 n = srclen & ~(GHASH_BLOCK_SIZE - 1); in ghash_update() 68 u8 buf[GHASH_BLOCK_SIZE] = {}; in ghash_flush() 71 cpacf_kimd(CPACF_KIMD_GHASH, dctx, buf, GHASH_BLOCK_SIZE); in ghash_flush() 82 memcpy(dst, dctx->icv, GHASH_BLOCK_SIZE); in ghash_finup() [all …]
|
| H A D | aes_s390.c | 838 u8 j0[GHASH_BLOCK_SIZE];/* initial counter value */ in gcm_aes_crypt() 933 .ivsize = GHASH_BLOCK_SIZE - sizeof(u32),
|
| /linux/crypto/ |
| H A D | ghash-generic.c | 61 if (keylen != GHASH_BLOCK_SIZE) in ghash_setkey() 67 BUILD_BUG_ON(sizeof(k) != GHASH_BLOCK_SIZE); in ghash_setkey() 68 memcpy(&k, key, GHASH_BLOCK_SIZE); /* avoid violating alignment rules */ in ghash_setkey() 70 memzero_explicit(&k, GHASH_BLOCK_SIZE); in ghash_setkey() 86 crypto_xor(dst, src, GHASH_BLOCK_SIZE); in ghash_update() 88 src += GHASH_BLOCK_SIZE; in ghash_update() 89 srclen -= GHASH_BLOCK_SIZE; in ghash_update() 90 } while (srclen >= GHASH_BLOCK_SIZE); in ghash_update() 115 memcpy(dst, buf, GHASH_BLOCK_SIZE); in ghash_finup() 139 .cra_blocksize = GHASH_BLOCK_SIZE,
|
| /linux/arch/riscv/crypto/ |
| H A D | ghash-riscv64-glue.c | 41 if (keylen != GHASH_BLOCK_SIZE) in riscv64_ghash_setkey() 44 memcpy(&tctx->key, key, GHASH_BLOCK_SIZE); in riscv64_ghash_setkey() 71 GHASH_BLOCK_SIZE); in riscv64_ghash_blocks() 73 src += GHASH_BLOCK_SIZE; in riscv64_ghash_blocks() 74 srclen -= GHASH_BLOCK_SIZE; in riscv64_ghash_blocks() 86 round_down(srclen, GHASH_BLOCK_SIZE)); in riscv64_ghash_update() 87 return srclen - round_down(srclen, GHASH_BLOCK_SIZE); in riscv64_ghash_update() 97 u8 buf[GHASH_BLOCK_SIZE] = {}; in riscv64_ghash_finup() 100 riscv64_ghash_blocks(tctx, dctx, buf, GHASH_BLOCK_SIZE); in riscv64_ghash_finup() 116 .cra_blocksize = GHASH_BLOCK_SIZE,
|
| /linux/arch/powerpc/crypto/ |
| H A D | ghash.c | 56 if (keylen != GHASH_BLOCK_SIZE) in p8_ghash_setkey() 67 memcpy(&ctx->key, key, GHASH_BLOCK_SIZE); in p8_ghash_setkey() 80 gcm_ghash_p8(dctx->shash, ctx->htable, src, GHASH_BLOCK_SIZE); in __ghash_block() 85 crypto_xor((u8 *)dctx->shash, src, GHASH_BLOCK_SIZE); in __ghash_block() 94 int remain = srclen - round_down(srclen, GHASH_BLOCK_SIZE); in __ghash_blocks() 108 crypto_xor((u8 *)dctx->shash, src, GHASH_BLOCK_SIZE); in __ghash_blocks() 110 srclen -= GHASH_BLOCK_SIZE; in __ghash_blocks() 111 src += GHASH_BLOCK_SIZE; in __ghash_blocks() 134 u8 buf[GHASH_BLOCK_SIZE] = {}; in p8_ghash_finup() 156 .cra_blocksize = GHASH_BLOCK_SIZE,
|
| /linux/arch/arm64/crypto/ |
| H A D | ghash-ce-glue.c | 83 #define MAX_BLOCKS (SZ_64K / GHASH_BLOCK_SIZE) 92 blocks = len / GHASH_BLOCK_SIZE; in ghash_update() 93 len -= blocks * GHASH_BLOCK_SIZE; in ghash_update() 101 src += chunk * GHASH_BLOCK_SIZE; in ghash_update() 133 u8 buf[GHASH_BLOCK_SIZE] = {}; in ghash_finup() 159 if (keylen != GHASH_BLOCK_SIZE) in ghash_setkey() 163 memcpy(&key->k, inkey, GHASH_BLOCK_SIZE); in ghash_setkey() 174 .base.cra_blocksize = GHASH_BLOCK_SIZE, 193 u8 key[GHASH_BLOCK_SIZE]; in gcm_aes_setkey() 204 memcpy(&ctx->ghash_key.k, key, GHASH_BLOCK_SIZE); in gcm_aes_setkey() [all …]
|
| H A D | sm4-ce-gcm-glue.c | 34 #define GHASH_BLOCK_SIZE 16 macro 75 u8 __aligned(8) buffer[GHASH_BLOCK_SIZE]; in gcm_calculate_auth_mac() 90 if (n + buflen < GHASH_BLOCK_SIZE) { in gcm_calculate_auth_mac() 97 unsigned int l = GHASH_BLOCK_SIZE - buflen; in gcm_calculate_auth_mac() 107 nblocks = n / GHASH_BLOCK_SIZE; in gcm_calculate_auth_mac() 111 p += nblocks * GHASH_BLOCK_SIZE; in gcm_calculate_auth_mac() 114 buflen = n % GHASH_BLOCK_SIZE; in gcm_calculate_auth_mac() 125 memset(&buffer[buflen], 0, GHASH_BLOCK_SIZE - buflen); in gcm_calculate_auth_mac()
|
| /linux/arch/arm/crypto/ |
| H A D | ghash-ce-glue.c | 87 blocks = len / GHASH_BLOCK_SIZE; in ghash_update() 89 return len - blocks * GHASH_BLOCK_SIZE; in ghash_update() 119 u8 buf[GHASH_BLOCK_SIZE] = {}; in ghash_finup() 144 if (keylen != GHASH_BLOCK_SIZE) in ghash_setkey() 148 memcpy(&key->k, inkey, GHASH_BLOCK_SIZE); in ghash_setkey() 181 .base.cra_blocksize = GHASH_BLOCK_SIZE, 250 int buf_added = min(count, GHASH_BLOCK_SIZE - *buf_count); in gcm_update_mac() 259 if (count >= GHASH_BLOCK_SIZE || *buf_count == GHASH_BLOCK_SIZE) { in gcm_update_mac() 260 int blocks = count / GHASH_BLOCK_SIZE; in gcm_update_mac() 265 src += blocks * GHASH_BLOCK_SIZE; in gcm_update_mac() [all …]
|
| /linux/include/crypto/ |
| H A D | ghash.h | 11 #define GHASH_BLOCK_SIZE 16 macro 21 u8 buffer[GHASH_BLOCK_SIZE];
|
| /linux/arch/x86/crypto/ |
| H A D | ghash-clmulni-intel_glue.c | 46 if (keylen != GHASH_BLOCK_SIZE) in ghash_setkey() 116 memcpy(dst, buf, GHASH_BLOCK_SIZE); in ghash_finup() 133 .cra_blocksize = GHASH_BLOCK_SIZE,
|
| /linux/lib/crypto/ |
| H A D | aesgcm.c | 67 crypto_xor((u8 *)ghash, src, min(len, GHASH_BLOCK_SIZE)); in aesgcm_ghash() 70 src += GHASH_BLOCK_SIZE; in aesgcm_ghash() 71 len -= GHASH_BLOCK_SIZE; in aesgcm_ghash()
|
| /linux/drivers/crypto/chelsio/ |
| H A D | chcr_crypto.h | 39 #define GHASH_BLOCK_SIZE 16 macro
|
| /linux/drivers/crypto/inside-secure/ |
| H A D | safexcel_cipher.c | 2559 ctx->state_sz = GHASH_BLOCK_SIZE; in safexcel_aead_gcm_cra_init()
|