/linux/arch/x86/crypto/ |
H A D | sm3-avx-asm_64.S | 124 #define XTMP0 %xmm6 macro 241 vmovdqu 0*16(RDATA), XTMP0; /* XTMP0: w3, w2, w1, w0 */ \ 245 vpshufb BSWAP_REG, XTMP0, XTMP0; \ 249 vpxor XTMP0, XTMP1, XTMP4; \ 253 vmovdqa XTMP0, IW_W1_ADDR(0, 0); \ 263 vpshufd $0b00000000, XTMP0, W0; /* W0: xx, w0, xx, xx */ \ 264 vpshufd $0b11111001, XTMP0, W1; /* W1: xx, w3, w2, w1 */ \ 273 vpshufd $0b10111111, w0, XTMP0; \ 274 vpalignr $12, XTMP0, w1, XTMP0; /* XTMP0: xx, w2, w1, w0 */ \ 280 vpxor w3, XTMP0, XTMP0; [all …]
|
H A D | sha256-avx-asm.S | 86 XTMP0 = %xmm0 define 161 vpalignr $4, X2, X3, XTMP0 # XTMP0 = W[-7] 168 vpaddd X0, XTMP0, XTMP0 # XTMP0 = W[-7] + W[-16] 228 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 262 vpaddd XTMP4, XTMP0, XTMP0 # XTMP0 = {..., ..., W[1], W[0]} 266 vpshufd $0b01010000, XTMP0, XTMP2 # XTMP2 = W[-2] {DDCC} 300 vpaddd XTMP0, XTMP5, X0 # X0 = {W[3], W[2], W[1], W[0]}
|
H A D | sha256-avx2-asm.S | 79 XTMP0 = %ymm0 define 163 vpalignr $4, X2, X3, XTMP0 # XTMP0 = W[-7] 169 vpaddd X0, XTMP0, XTMP0 # XTMP0 = W[-7] + W[-16]# y1 = (e >> 6)# S1 245 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 286 vpaddd XTMP4, XTMP0, XTMP0 # XTMP0 = {..., ..., W[1], W[0]} 292 vpshufd $0b01010000, XTMP0, XTMP2 # XTMP2 = W[-2] {DDCC} 342 vpaddd XTMP0, XTMP5, X0 # X0 = {W[3], W[2], W[1], W[0]} 566 VMOVDQ 0*32(INP),XTMP0 572 vpshufb BYTE_FLIP_MASK, XTMP0, XTMP0 578 vperm2i128 $0x20, XTMP2, XTMP0, X0 [all …]
|
H A D | sha256-ssse3-asm.S | 80 XTMP0 = %xmm0 define 151 movdqa X3, XTMP0 155 palignr $4, X2, XTMP0 # XTMP0 = W[-7] 163 paddd X0, XTMP0 # XTMP0 = W[-7] + W[-16] 228 paddd XTMP1, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 265 paddd XTMP4, XTMP0 # XTMP0 = {..., ..., W[1], W[0]} 269 pshufd $0b01010000, XTMP0, XTMP2 # XTMP2 = W[-2] {BBAA} 306 paddd XTMP0, X0 # X0 = {W[3], W[2], W[1], W[0]}
|
/linux/arch/arm64/crypto/ |
H A D | sm3-neon-core.S | 76 #define XTMP0 v6 macro 197 rev32 XTMP0.16b, W0.16b; 205 eor XTMP4.16b, XTMP1.16b, XTMP0.16b; 209 st1 {XTMP0.16b}, [addr0], #16; 216 ext W0.16b, XTMP0.16b, XTMP0.16b, #8; /* W0: xx, w0, xx, xx */ 223 ext W1.16b, XTMP0.16b, XTMP0.16b, #4; /* W1: xx, w3, w2, w1 */ 247 ext XTMP0.16b, w0.16b, w0.16b, #12; /* XTMP0: w0, xx, xx, xx */ 251 ext XTMP0.16b, XTMP0.16b, w1.16b, #12; /* XTMP0: xx, w2, w1, w0 */ 257 eor XTMP0.16b, XTMP0.16b, w3.16b; 271 eor XTMP0.16b, XTMP0.16b, XTMP2.16b; [all …]
|