Home
last modified time | relevance | path

Searched refs:xmm2 (Results 1 – 25 of 37) sorted by relevance

12

/linux/lib/crypto/x86/
H A Dblake2s-core.S84 movdqa .Liv(%rip),%xmm2 // v[8..11] = iv[0..3]
106 paddd %xmm3,%xmm2
107 pxor %xmm2,%xmm1
127 paddd %xmm3,%xmm2
128 pxor %xmm2,%xmm1
135 pshufd $0x39,%xmm2,%xmm2
151 paddd %xmm3,%xmm2
152 pxor %xmm2,%xmm1
172 paddd %xmm3,%xmm2
173 pxor %xmm2,%xmm1
[all …]
H A Dchacha-ssse3-x86_64.S48 paddd %xmm3,%xmm2
49 pxor %xmm2,%xmm1
61 paddd %xmm3,%xmm2
62 pxor %xmm2,%xmm1
71 pshufd $0x4e,%xmm2,%xmm2
81 paddd %xmm3,%xmm2
82 pxor %xmm2,%xmm1
94 paddd %xmm3,%xmm2
95 pxor %xmm2,%xmm1
104 pshufd $0x4e,%xmm2,%xmm2
[all …]
H A Daes-aesni.S59 movdqa \b, %xmm2
60 pshufb MASK, %xmm2
61 aesenclast RCON, %xmm2
67 pxor %xmm2, \a
144 pshufd $0xff, %xmm0, %xmm2 // Get four copies of last dword
145 aesenclast %xmm5, %xmm2 // Just does SubBytes
147 pxor %xmm2, %xmm1
H A Dpolyval-pclmul-avx.S92 vpclmulqdq $0x01, (16*\i)(KEY_POWERS), %xmm0, %xmm2
96 vpxor %xmm2, MI, MI
109 vpclmulqdq $0x10, %xmm0, %xmm1, %xmm2
112 vpxor %xmm2, MI, MI
H A Dsha512-ssse3-asm.S171 movdqa W_t(idx), %xmm2 # XMM2 = W[t-2]
174 movdqa %xmm2, %xmm0 # XMM0 = W[t-2]
191 pxor %xmm2, %xmm0 # XMM0 = (W[t-2] >> 42) ^ W[t-2]
203 pxor %xmm2, %xmm0 # XMM0 = (((W[t-2]>>42)^W[t-2])>>13)^W[t-2]
215 movdqa %xmm2, %xmm1 # XMM1 = W[t-2]
228 pxor %xmm2, %xmm1 # XMM1 = (W[t-2] << 42)^W[t-2]
H A Dsha1-ni-asm.S64 #define E1 %xmm2
H A Dsha512-avx-asm.S190 vpsrlq $6, %xmm4, %xmm2 # XMM2 = W[t-2]>>6
203 vpxor %xmm3, %xmm2, %xmm2 # XMM2 = W[t-2]>>6 ^ W[t-2]<<3
223 vpxor %xmm2, %xmm0, %xmm0 # XMM0 = s1(W[t-2])
H A Dnh-sse2.S14 #define PASS2_SUMS %xmm2
H A Dsha256-ni-asm.S66 #define STATE1 %xmm2
204 #define STATE1_A %xmm2
H A Dchacha-avx512vl-x86_64.S133 vextracti128 $1,%ymm7,%xmm2
155 vmovdqa %xmm2,%xmm7
345 vextracti128 $1,%ymm10,%xmm2
367 vmovdqa %xmm2,%xmm10
H A Dchacha-avx2-x86_64.S160 vextracti128 $1,%ymm7,%xmm2
182 vmovdqa %xmm2,%xmm7
411 vextracti128 $1,%ymm10,%xmm2
433 vmovdqa %xmm2,%xmm10
/linux/arch/x86/crypto/
H A Daria-aesni-avx-asm_64.S889 inpack16_post(%xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
893 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
895 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7,
899 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
901 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7,
905 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
907 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7,
911 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
913 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7,
917 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
[all …]
H A Daesni-intel_asm.S33 #define KEY %xmm2
89 movaps %xmm2, %xmm5
90 movaps %xmm2, %xmm6
93 pxor %xmm3, %xmm2
94 pxor %xmm5, %xmm2
99 shufps $0b01001110, %xmm2, %xmm1
113 movaps %xmm2, %xmm5
116 pxor %xmm3, %xmm2
117 pxor %xmm5, %xmm2
126 shufps $0b00010000, %xmm2, %xmm4
[all …]
H A Dcamellia-aesni-avx-asm_64.S193 roundsm16(%xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
201 roundsm16(%xmm4, %xmm5, %xmm6, %xmm7, %xmm0, %xmm1, %xmm2, %xmm3,
729 inpack16_post(%xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
733 enc_rounds16(%xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
737 fls16(%rax, %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
745 enc_rounds16(%xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
749 fls16(%rax, %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
757 enc_rounds16(%xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
776 outunpack16(%xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
787 fls16(%rax, %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
[all …]
H A Daes-gcm-aesni-x86_64.S502 .set H_CUR, %xmm2
583 .set BSWAP_MASK, %xmm2
709 .set TMP2, %xmm2
1060 _ghash_mul_step \i, H_POW1, H_POW1_X64, GHASH_ACC, GFPOLY, %xmm1, %xmm2
1063 _ghash_mul_step 9, H_POW1, H_POW1_X64, GHASH_ACC, GFPOLY, %xmm1, %xmm2
H A Daes-gcm-vaes-avx2.S233 .set TMP2_XMM, %xmm2
512 .set TMP2_XMM, %xmm2
721 .set TMP0_XMM, %xmm2
1094 %xmm1, %xmm2, %xmm3
1098 %xmm1, %xmm2, %xmm3
H A Daes-gcm-vaes-avx512.S514 .set GHASHDATA2_XMM, %xmm2
1007 _horizontal_xor HI, HI_XMM, GHASH_ACC_XMM, %xmm0, %xmm1, %xmm2
1114 %xmm1, %xmm2, %xmm3
1118 %xmm1, %xmm2, %xmm3
H A Dghash-clmulni-intel_asm.S26 #define T1 %xmm2
H A Daes-xts-avx-x86_64.S815 vpshufb (%r9, LEN64, 1), %xmm0, %xmm2
816 vmovdqu %xmm2, (DST, LEN64, 1)
H A Dcast6-avx-x86_64-asm_64.S39 #define RC1 %xmm2
H A Dtwofish-avx-x86_64-asm_64.S39 #define RC1 %xmm2
H A Dsm4-aesni-avx2-asm_64.S47 #define MASK_4BITx %xmm2
H A Dcast5-avx-x86_64-asm_64.S39 #define RL2 %xmm2
/linux/lib/crc/x86/
H A Dcrc-pclmul-template.S407 _load_data 16, "-16(BUF,LEN)", BSWAP_MASK_XMM, %xmm2
435 pblendvb %xmm2, %xmm1 // uses %xmm0 as implicit operand
440 vpblendvb %xmm3, %xmm2, %xmm1, %xmm1
532 _cond_vex punpcklqdq, %xmm1, %xmm2, %xmm2
535 _cond_vex pxor, %xmm2, %xmm0, %xmm0
538 vpternlogq $0x96, %xmm2, %xmm1, %xmm0
/linux/arch/x86/entry/vdso/vdso64/
H A Dvgetrandom-chacha.S34 .set state1, %xmm2

12