/linux/arch/riscv/crypto/ |
H A D | aes-riscv64-zvkned.S | 54 #define INP a1 macro 60 vle32.v v16, (INP) 93 vle32.v v16, (INP) 96 add INP, INP, t1 128 vle32.v v17, (INP) // Load plaintext block 132 addi INP, INP, 16 146 vle32.v v20, (INP) // Load ciphertext blocks 155 add INP, IN [all...] |
H A D | aes-riscv64-zvkned-zvkb.S | 53 #define INP a1 macro 113 vle8.v v20, (INP) 118 add INP, INP, t0
|
H A D | aes-riscv64-zvkned-zvbb-zvkg.S | 54 #define INP a1 macro 178 vle32.v TMP0, (INP) 186 add INP, INP, t0 245 vle32.v TMP0, (INP) 253 addi t0, INP, 16
|
/linux/lib/crypto/s390/ |
H A D | chacha-s390.S | 44 #define INP %r3 macro 293 VLM XT0,XT3,0,INP,0 302 la INP,0x40(INP) 319 VLM XT0,XT3,0,INP,0 328 la INP,0x40(INP) 346 VLM XT0,XT3,0,INP,0 355 la INP,0x40(INP) 428 #define INP global() macro [all...] |
/linux/lib/crypto/x86/ |
H A D | sha256-avx2-asm.S | 93 INP = %rsi # 2nd arg define 98 y3 = %esi # clobbers INP 538 lea -64(INP, NUM_BLKS), NUM_BLKS # pointer to last block 541 cmp NUM_BLKS, INP 562 VMOVDQ 0*32(INP),XTMP0 563 VMOVDQ 1*32(INP),XTMP1 564 VMOVDQ 2*32(INP),XTMP2 565 VMOVDQ 3*32(INP),XTMP3 580 add $64, INP 581 mov INP, _IN [all...] |
H A D | sha256-avx-asm.S | 98 INP = %rsi # 2nd arg define 101 SRND = %rsi # clobbers INP 360 add INP, NUM_BLKS # pointer to end of data 380 COPY_XMM_AND_BSWAP X0, 0*16(INP), BYTE_FLIP_MASK 381 COPY_XMM_AND_BSWAP X1, 1*16(INP), BYTE_FLIP_MASK 382 COPY_XMM_AND_BSWAP X2, 2*16(INP), BYTE_FLIP_MASK 383 COPY_XMM_AND_BSWAP X3, 3*16(INP), BYTE_FLIP_MASK 385 mov INP, _INP(%rsp) 443 mov _INP(%rsp), INP 444 add $64, INP [all...] |
H A D | sha256-ssse3-asm.S | 91 INP = %rsi # 2nd arg define 94 SRND = %rsi # clobbers INP 367 add INP, NUM_BLKS 388 COPY_XMM_AND_BSWAP X0, 0*16(INP), BYTE_FLIP_MASK 389 COPY_XMM_AND_BSWAP X1, 1*16(INP), BYTE_FLIP_MASK 390 COPY_XMM_AND_BSWAP X2, 2*16(INP), BYTE_FLIP_MASK 391 COPY_XMM_AND_BSWAP X3, 3*16(INP), BYTE_FLIP_MASK 393 mov INP, _INP(%rsp) 454 mov _INP(%rsp), INP 455 add $64, INP [all...] |
H A D | sha512-avx2-asm.S | 75 INP = %rsi define 585 add INP, NUM_BLKS # pointer to end of data 607 COPY_YMM_AND_BSWAP Y_0, (INP), BYTE_FLIP_MASK 608 COPY_YMM_AND_BSWAP Y_1, 1*32(INP), BYTE_FLIP_MASK 609 COPY_YMM_AND_BSWAP Y_2, 2*32(INP), BYTE_FLIP_MASK 610 COPY_YMM_AND_BSWAP Y_3, 3*32(INP), BYTE_FLIP_MASK 612 mov INP, frame_INP(%rsp) 665 mov frame_INP(%rsp), INP 666 add $128, INP 667 cmp frame_INPEND(%rsp), INP [all...] |
/linux/lib/crypto/riscv/ |
H A D | chacha-riscv64-zvkb.S | 50 #define INP a1 macro 206 vlsseg8e32.v v16, (INP), STRIDE 221 addi TMP, INP, 32 279 add INP, INP, TMP
|