/arch/x86/crypto/ |
D | sm4-aesni-avx-asm_64.S | 45 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \ macro 187 transpose_4x4(RA0, RA1, RA2, RA3, RTMP0, RTMP1); 230 transpose_4x4(RA0, RA1, RA2, RA3, RTMP0, RTMP1); 275 transpose_4x4(RA0, RA1, RA2, RA3, RTMP0, RTMP1); 276 transpose_4x4(RB0, RB1, RB2, RB3, RTMP0, RTMP1); 347 transpose_4x4(RA0, RA1, RA2, RA3, RTMP0, RTMP1); 348 transpose_4x4(RB0, RB1, RB2, RB3, RTMP0, RTMP1);
|
D | sm4-aesni-avx2-asm_64.S | 61 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \ macro 181 transpose_4x4(RA0, RA1, RA2, RA3, RTMP0, RTMP1); 182 transpose_4x4(RB0, RB1, RB2, RB3, RTMP0, RTMP1); 259 transpose_4x4(RA0, RA1, RA2, RA3, RTMP0, RTMP1); 260 transpose_4x4(RB0, RB1, RB2, RB3, RTMP0, RTMP1);
|
D | camellia-aesni-avx-asm_64.S | 419 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \ macro 436 transpose_4x4(a0, a1, a2, a3, d2, d3); \ 437 transpose_4x4(b0, b1, b2, b3, d2, d3); \ 443 transpose_4x4(c0, c1, c2, c3, a0, a1); \ 444 transpose_4x4(d0, d1, d2, d3, a0, a1); \ 468 transpose_4x4(a0, b0, c0, d0, d2, d3); \ 469 transpose_4x4(a1, b1, c1, d1, d2, d3); \ 475 transpose_4x4(a2, b2, c2, d2, b0, b1); \ 476 transpose_4x4(a3, b3, c3, d3, b0, b1); \
|
D | camellia-aesni-avx2-asm_64.S | 453 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \ macro 470 transpose_4x4(a0, a1, a2, a3, d2, d3); \ 471 transpose_4x4(b0, b1, b2, b3, d2, d3); \ 477 transpose_4x4(c0, c1, c2, c3, a0, a1); \ 478 transpose_4x4(d0, d1, d2, d3, a0, a1); \ 502 transpose_4x4(a0, b0, c0, d0, d2, d3); \ 503 transpose_4x4(a1, b1, c1, d1, d2, d3); \ 509 transpose_4x4(a2, b2, c2, d2, b0, b1); \ 510 transpose_4x4(a3, b3, c3, d3, b0, b1); \
|
D | cast6-avx-x86_64-asm_64.S | 186 #define transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ macro 203 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) 206 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \
|
D | twofish-avx-x86_64-asm_64.S | 204 #define transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ macro 221 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) 224 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \
|
D | serpent-sse2-i586-asm_32.S | 453 #define transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ macro 474 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) 477 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ 485 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \
|
D | serpent-sse2-x86_64-asm_64.S | 575 #define transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ macro 596 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) 599 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ 607 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \
|
D | serpent-avx2-asm_64.S | 536 #define transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ macro 548 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) 551 transpose_4x4(x0, x1, x2, x3, t0, t1, t2)
|
D | serpent-avx-x86_64-asm_64.S | 536 #define transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ macro 548 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) 551 transpose_4x4(x0, x1, x2, x3, t0, t1, t2)
|