| /Linux-v5.4/arch/x86/crypto/ |
| D | twofish-avx-x86_64-asm_64.S | 47 #define RA2 %xmm4 macro 254 inpack_blocks(RA2, RB2, RC2, RD2, RK1, RX0, RY0, RK2); 273 outunpack_blocks(RC2, RD2, RA2, RB2, RK1, RX0, RY0, RK2); 295 inpack_blocks(RC2, RD2, RA2, RB2, RK1, RX0, RY0, RK2); 296 rotate_1l(RA2); 313 outunpack_blocks(RA2, RB2, RC2, RD2, RK1, RX0, RY0, RK2); 328 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 332 store_8way(%r11, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2); 348 load_8way(%rdx, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2); 352 store_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); [all …]
|
| D | cast6-avx-x86_64-asm_64.S | 42 #define RA2 %xmm4 macro 268 inpack_blocks(RA2, RB2, RC2, RD2, RTMP, RX, RKRF, RKM); 292 outunpack_blocks(RA2, RB2, RC2, RD2, RTMP, RX, RKRF, RKM); 316 inpack_blocks(RA2, RB2, RC2, RD2, RTMP, RX, RKRF, RKM); 339 outunpack_blocks(RA2, RB2, RC2, RD2, RTMP, RX, RKRF, RKM); 356 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 360 store_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 379 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 383 store_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 404 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); [all …]
|
| D | serpent-avx-x86_64-asm_64.S | 41 #define RA2 %xmm6 macro 569 read_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 606 write_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 623 read_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 673 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 677 store_8way(%rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 691 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 709 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 728 load_ctr_8way(%rcx, .Lbswap128_mask, RA1, RB1, RC1, RD1, RA2, RB2, RC2, 733 store_ctr_8way(%rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); [all …]
|
| D | serpent-avx2-asm_64.S | 41 #define RA2 %ymm3 macro 575 read_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 612 write_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 629 read_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 681 load_16way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 685 store_16way(%rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 703 load_16way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 725 load_16way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 749 load_ctr_16way(%rcx, .Lbswap128_mask, RA1, RB1, RC1, RD1, RA2, RB2, RC2, 755 store_ctr_16way(%rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); [all …]
|
| D | serpent-sse2-x86_64-asm_64.S | 28 #define RA2 %xmm5 macro 634 read_blocks(%rax, RA2, RB2, RC2, RD2, RK0, RK1, RK2); 676 write_blocks(%rax, RA2, RB2, RC2, RD2, RK0, RK1, RK2); 682 xor_blocks(%rax, RA2, RB2, RC2, RD2, RK0, RK1, RK2); 698 read_blocks(%rax, RA2, RB2, RC2, RD2, RK0, RK1, RK2);
|