Lines Matching defs:x0
57 * x0..x7: byte-sliced AB state
61 * x0..x7: new byte-sliced CD state
63 #define roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \
76 vpshufb t4, x0, x0; \
87 filter_8bit(x0, t5, t6, t7, t4); \
89 vextracti128 $1, x0, t0##_x; \
105 vaesenclast t4##_x, x0##_x, x0##_x; \
107 vinserti128 $1, t0##_x, x0, x0; \
137 filter_8bit(x0, t0, t1, t7, t6); \
161 vpxor x5, x0, x0; \
174 vpxor x0, x6, x6; \
181 vpxor x7, x0, x0; \
187 vpxor x0, x5, x5; \
200 vpxor t7, x0, x0; \
201 vpxor 4 * 32(mem_cd), x0, x0; \
241 * x0..x7: byte-sliced AB state preloaded
245 #define two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
250 vmovdqu x0, 4 * 32(mem_cd); \
262 store_ab(x0, x1, x2, x3, x4, x5, x6, x7, mem_ab);
264 #define dummy_store(x0, x1, x2, x3, x4, x5, x6, x7, mem_ab) /* do nothing */
266 #define store_ab_state(x0, x1, x2, x3, x4, x5, x6, x7, mem_ab) \
272 vmovdqu x0, 0 * 32(mem_ab); \
277 #define enc_rounds32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
279 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
281 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
283 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
286 #define dec_rounds32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
288 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
290 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
292 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
329 * x0..x7: new byte-sliced CD state
452 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \
453 vpunpckhdq x1, x0, t2; \
454 vpunpckldq x1, x0, x0; \
459 vpunpckhqdq t1, x0, x1; \
460 vpunpcklqdq t1, x0, x0; \
515 #define inpack32_pre(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
517 vpbroadcastq key, x0; \
518 vpshufb .Lpack_bswap(%rip), x0, x0; \
520 vpxor 0 * 32(rio), x0, y7; \
521 vpxor 1 * 32(rio), x0, y6; \
522 vpxor 2 * 32(rio), x0, y5; \
523 vpxor 3 * 32(rio), x0, y4; \
524 vpxor 4 * 32(rio), x0, y3; \
525 vpxor 5 * 32(rio), x0, y2; \
526 vpxor 6 * 32(rio), x0, y1; \
527 vpxor 7 * 32(rio), x0, y0; \
528 vpxor 8 * 32(rio), x0, x7; \
529 vpxor 9 * 32(rio), x0, x6; \
530 vpxor 10 * 32(rio), x0, x5; \
531 vpxor 11 * 32(rio), x0, x4; \
532 vpxor 12 * 32(rio), x0, x3; \
533 vpxor 13 * 32(rio), x0, x2; \
534 vpxor 14 * 32(rio), x0, x1; \
535 vpxor 15 * 32(rio), x0, x0;
538 #define inpack32_post(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
540 byteslice_16x16b_fast(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, \
543 vmovdqu x0, 0 * 32(mem_ab); \
561 #define outunpack32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, \
563 byteslice_16x16b_fast(y0, y4, x0, x4, y1, y5, x1, x5, y2, y6, x2, x6, \
566 vmovdqu x0, stack_tmp0; \
568 vpbroadcastq key, x0; \
569 vpshufb .Lpack_bswap(%rip), x0, x0; \
571 vpxor x0, y7, y7; \
572 vpxor x0, y6, y6; \
573 vpxor x0, y5, y5; \
574 vpxor x0, y4, y4; \
575 vpxor x0, y3, y3; \
576 vpxor x0, y2, y2; \
577 vpxor x0, y1, y1; \
578 vpxor x0, y0, y0; \
579 vpxor x0, x7, x7; \
580 vpxor x0, x6, x6; \
581 vpxor x0, x5, x5; \
582 vpxor x0, x4, x4; \
583 vpxor x0, x3, x3; \
584 vpxor x0, x2, x2; \
585 vpxor x0, x1, x1; \
586 vpxor stack_tmp0, x0, x0;
588 #define write_output(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \
590 vmovdqu x0, 0 * 32(rio); \