| /linux/arch/x86/crypto/ |
| H A D | sm3-avx-asm_64.S | 124 #define XTMP0 %xmm6 macro 241 vmovdqu 0*16(RDATA), XTMP0; /* XTMP0: w3, w2, w1, w0 */ \ 245 vpshufb BSWAP_REG, XTMP0, XTMP0; \ 249 vpxor XTMP0, XTMP1, XTMP4; \ 253 vmovdqa XTMP0, IW_W1_ADDR(0, 0); \ 263 vpshufd $0b00000000, XTMP0, W0; /* W0: xx, w0, xx, xx */ \ 264 vpshufd $0b11111001, XTMP0, W1; /* W1: xx, w3, w2, w1 */ \ 273 vpshufd $0b10111111, w0, XTMP0; \ 274 vpalignr $12, XTMP0, w1, XTMP0; /* XTMP0: xx, w2, w1, w0 */ \ 280 vpxor w3, XTMP0, XTMP0; [all …]
|
| /linux/arch/arm64/crypto/ |
| H A D | sm3-neon-core.S | 76 #define XTMP0 v6 macro 197 rev32 XTMP0.16b, W0.16b; 205 eor XTMP4.16b, XTMP1.16b, XTMP0.16b; 209 st1 {XTMP0.16b}, [addr0], #16; 216 ext W0.16b, XTMP0.16b, XTMP0.16b, #8; /* W0: xx, w0, xx, xx */ 223 ext W1.16b, XTMP0.16b, XTMP0.16b, #4; /* W1: xx, w3, w2, w1 */ 247 ext XTMP0.16b, w0.16b, w0.16b, #12; /* XTMP0: w0, xx, xx, xx */ 251 ext XTMP0.16b, XTMP0.16b, w1.16b, #12; /* XTMP0: xx, w2, w1, w0 */ 257 eor XTMP0.16b, XTMP0.16b, w3.16b; 271 eor XTMP0.16b, XTMP0.16b, XTMP2.16b; [all …]
|
| /linux/lib/crypto/x86/ |
| H A D | sha256-avx-asm.S | 85 XTMP0 = %xmm0 define 160 vpalignr $4, X2, X3, XTMP0 # XTMP0 = W[-7] 167 vpaddd X0, XTMP0, XTMP0 # XTMP0 = W[-7] + W[-16] 227 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 261 vpaddd XTMP4, XTMP0, XTMP0 # XTMP0 = {..., ..., W[1], W[0]} 265 vpshufd $0b01010000, XTMP0, XTMP2 # XTMP2 = W[-2] {DDCC} 299 vpaddd XTMP0, XTMP5, X0 # X0 = {W[3], W[2], W[1], W[0]}
|
| H A D | sha256-avx2-asm.S | 78 XTMP0 = %ymm0 define 162 vpalignr $4, X2, X3, XTMP0 # XTMP0 = W[-7] 168 vpaddd X0, XTMP0, XTMP0 # XTMP0 = W[-7] + W[-16]# y1 = (e >> 6)# S1 244 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 285 vpaddd XTMP4, XTMP0, XTMP0 # XTMP0 = {..., ..., W[1], W[0]} 291 vpshufd $0b01010000, XTMP0, XTMP2 # XTMP2 = W[-2] {DDCC} 341 vpaddd XTMP0, XTMP5, X0 # X0 = {W[3], W[2], W[1], W[0]} 562 VMOVDQ 0*32(INP),XTMP0 568 vpshufb BYTE_FLIP_MASK, XTMP0, XTMP0 574 vperm2i128 $0x20, XTMP2, XTMP0, X0 [all …]
|
| H A D | sha256-ssse3-asm.S | 79 XTMP0 = %xmm0 define 150 movdqa X3, XTMP0 154 palignr $4, X2, XTMP0 # XTMP0 = W[-7] 162 paddd X0, XTMP0 # XTMP0 = W[-7] + W[-16] 227 paddd XTMP1, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 264 paddd XTMP4, XTMP0 # XTMP0 = {..., ..., W[1], W[0]} 268 pshufd $0b01010000, XTMP0, XTMP2 # XTMP2 = W[-2] {BBAA} 305 paddd XTMP0, X0 # X0 = {W[3], W[2], W[1], W[0]}
|