Home
last modified time | relevance | path

Searched refs:GHASH_BLOCK_SIZE (Results 1 – 13 of 13) sorted by relevance

/linux/arch/s390/crypto/
H A Dghash_s390.c21 u8 key[GHASH_BLOCK_SIZE];
25 u8 icv[GHASH_BLOCK_SIZE];
26 u8 key[GHASH_BLOCK_SIZE];
35 memcpy(dctx->key, ctx->key, GHASH_BLOCK_SIZE); in ghash_init()
45 if (keylen != GHASH_BLOCK_SIZE) in ghash_setkey()
48 memcpy(ctx->key, key, GHASH_BLOCK_SIZE); in ghash_setkey()
59 n = srclen & ~(GHASH_BLOCK_SIZE - 1); in ghash_update()
68 u8 buf[GHASH_BLOCK_SIZE] = {}; in ghash_flush()
71 cpacf_kimd(CPACF_KIMD_GHASH, dctx, buf, GHASH_BLOCK_SIZE); in ghash_flush()
82 memcpy(dst, dctx->icv, GHASH_BLOCK_SIZE); in ghash_finup()
[all …]
H A Daes_s390.c838 u8 j0[GHASH_BLOCK_SIZE];/* initial counter value */ in gcm_aes_crypt()
933 .ivsize = GHASH_BLOCK_SIZE - sizeof(u32),
/linux/crypto/
H A Dghash-generic.c61 if (keylen != GHASH_BLOCK_SIZE) in ghash_setkey()
67 BUILD_BUG_ON(sizeof(k) != GHASH_BLOCK_SIZE); in ghash_setkey()
68 memcpy(&k, key, GHASH_BLOCK_SIZE); /* avoid violating alignment rules */ in ghash_setkey()
70 memzero_explicit(&k, GHASH_BLOCK_SIZE); in ghash_setkey()
86 crypto_xor(dst, src, GHASH_BLOCK_SIZE); in ghash_update()
88 src += GHASH_BLOCK_SIZE; in ghash_update()
89 srclen -= GHASH_BLOCK_SIZE; in ghash_update()
90 } while (srclen >= GHASH_BLOCK_SIZE); in ghash_update()
115 memcpy(dst, buf, GHASH_BLOCK_SIZE); in ghash_finup()
139 .cra_blocksize = GHASH_BLOCK_SIZE,
/linux/arch/riscv/crypto/
H A Dghash-riscv64-glue.c41 if (keylen != GHASH_BLOCK_SIZE) in riscv64_ghash_setkey()
44 memcpy(&tctx->key, key, GHASH_BLOCK_SIZE); in riscv64_ghash_setkey()
71 GHASH_BLOCK_SIZE); in riscv64_ghash_blocks()
73 src += GHASH_BLOCK_SIZE; in riscv64_ghash_blocks()
74 srclen -= GHASH_BLOCK_SIZE; in riscv64_ghash_blocks()
86 round_down(srclen, GHASH_BLOCK_SIZE)); in riscv64_ghash_update()
87 return srclen - round_down(srclen, GHASH_BLOCK_SIZE); in riscv64_ghash_update()
97 u8 buf[GHASH_BLOCK_SIZE] = {}; in riscv64_ghash_finup()
100 riscv64_ghash_blocks(tctx, dctx, buf, GHASH_BLOCK_SIZE); in riscv64_ghash_finup()
116 .cra_blocksize = GHASH_BLOCK_SIZE,
/linux/arch/powerpc/crypto/
H A Dghash.c56 if (keylen != GHASH_BLOCK_SIZE) in p8_ghash_setkey()
67 memcpy(&ctx->key, key, GHASH_BLOCK_SIZE); in p8_ghash_setkey()
80 gcm_ghash_p8(dctx->shash, ctx->htable, src, GHASH_BLOCK_SIZE); in __ghash_block()
85 crypto_xor((u8 *)dctx->shash, src, GHASH_BLOCK_SIZE); in __ghash_block()
94 int remain = srclen - round_down(srclen, GHASH_BLOCK_SIZE); in __ghash_blocks()
108 crypto_xor((u8 *)dctx->shash, src, GHASH_BLOCK_SIZE); in __ghash_blocks()
110 srclen -= GHASH_BLOCK_SIZE; in __ghash_blocks()
111 src += GHASH_BLOCK_SIZE; in __ghash_blocks()
134 u8 buf[GHASH_BLOCK_SIZE] = {}; in p8_ghash_finup()
156 .cra_blocksize = GHASH_BLOCK_SIZE,
/linux/arch/arm64/crypto/
H A Dghash-ce-glue.c83 #define MAX_BLOCKS (SZ_64K / GHASH_BLOCK_SIZE)
92 blocks = len / GHASH_BLOCK_SIZE; in ghash_update()
93 len -= blocks * GHASH_BLOCK_SIZE; in ghash_update()
101 src += chunk * GHASH_BLOCK_SIZE; in ghash_update()
133 u8 buf[GHASH_BLOCK_SIZE] = {}; in ghash_finup()
159 if (keylen != GHASH_BLOCK_SIZE) in ghash_setkey()
163 memcpy(&key->k, inkey, GHASH_BLOCK_SIZE); in ghash_setkey()
174 .base.cra_blocksize = GHASH_BLOCK_SIZE,
193 u8 key[GHASH_BLOCK_SIZE]; in gcm_aes_setkey()
204 memcpy(&ctx->ghash_key.k, key, GHASH_BLOCK_SIZE); in gcm_aes_setkey()
[all …]
H A Dsm4-ce-gcm-glue.c34 #define GHASH_BLOCK_SIZE 16 macro
75 u8 __aligned(8) buffer[GHASH_BLOCK_SIZE]; in gcm_calculate_auth_mac()
90 if (n + buflen < GHASH_BLOCK_SIZE) { in gcm_calculate_auth_mac()
97 unsigned int l = GHASH_BLOCK_SIZE - buflen; in gcm_calculate_auth_mac()
107 nblocks = n / GHASH_BLOCK_SIZE; in gcm_calculate_auth_mac()
111 p += nblocks * GHASH_BLOCK_SIZE; in gcm_calculate_auth_mac()
114 buflen = n % GHASH_BLOCK_SIZE; in gcm_calculate_auth_mac()
125 memset(&buffer[buflen], 0, GHASH_BLOCK_SIZE - buflen); in gcm_calculate_auth_mac()
/linux/arch/arm/crypto/
H A Dghash-ce-glue.c87 blocks = len / GHASH_BLOCK_SIZE; in ghash_update()
89 return len - blocks * GHASH_BLOCK_SIZE; in ghash_update()
119 u8 buf[GHASH_BLOCK_SIZE] = {}; in ghash_finup()
144 if (keylen != GHASH_BLOCK_SIZE) in ghash_setkey()
148 memcpy(&key->k, inkey, GHASH_BLOCK_SIZE); in ghash_setkey()
181 .base.cra_blocksize = GHASH_BLOCK_SIZE,
250 int buf_added = min(count, GHASH_BLOCK_SIZE - *buf_count); in gcm_update_mac()
259 if (count >= GHASH_BLOCK_SIZE || *buf_count == GHASH_BLOCK_SIZE) { in gcm_update_mac()
260 int blocks = count / GHASH_BLOCK_SIZE; in gcm_update_mac()
265 src += blocks * GHASH_BLOCK_SIZE; in gcm_update_mac()
[all …]
/linux/include/crypto/
H A Dghash.h11 #define GHASH_BLOCK_SIZE 16 macro
21 u8 buffer[GHASH_BLOCK_SIZE];
/linux/arch/x86/crypto/
H A Dghash-clmulni-intel_glue.c46 if (keylen != GHASH_BLOCK_SIZE) in ghash_setkey()
116 memcpy(dst, buf, GHASH_BLOCK_SIZE); in ghash_finup()
133 .cra_blocksize = GHASH_BLOCK_SIZE,
/linux/lib/crypto/
H A Daesgcm.c67 crypto_xor((u8 *)ghash, src, min(len, GHASH_BLOCK_SIZE)); in aesgcm_ghash()
70 src += GHASH_BLOCK_SIZE; in aesgcm_ghash()
71 len -= GHASH_BLOCK_SIZE; in aesgcm_ghash()
/linux/drivers/crypto/chelsio/
H A Dchcr_crypto.h39 #define GHASH_BLOCK_SIZE 16 macro
/linux/drivers/crypto/inside-secure/
H A Dsafexcel_cipher.c2559 ctx->state_sz = GHASH_BLOCK_SIZE; in safexcel_aead_gcm_cra_init()