/linux/arch/x86/crypto/ |
H A D | ghash-clmulni-intel_asm.S | 28 #define T3 %xmm4 macro 44 * T3 49 pshufd $0b01001110, SHASH, T3 51 pxor SHASH, T3 55 pclmulqdq $0x00, T3, T2 # T2 = (a1 + a0) * (b1 + b0) 59 movaps T2, T3 60 pslldq $8, T3 62 pxor T3, DATA 67 movaps DATA, T3 68 psllq $1, T3 [all …]
|
H A D | nh-avx2-x86_64.S | 29 #define T3 %ymm11 macro 43 vpaddd \k0, T3, T0 44 vpaddd \k1, T3, T1 45 vpaddd \k2, T3, T2 46 vpaddd \k3, T3, T3 55 vpshufd $0x10, T3, T7 56 vpshufd $0x32, T3, T3 60 vpmuludq T7, T3, T3 64 vpaddq T3, PASS3_SUMS, PASS3_SUMS 86 vmovdqu (MESSAGE), T3 [all …]
|
H A D | nh-sse2-x86_64.S | 24 #define T3 %xmm11 macro 44 movdqa T1, T3 48 paddd \k3, T3 57 pshufd $0x10, T3, T7 58 pshufd $0x32, T3, T3 62 pmuludq T7, T3 66 paddq T3, PASS3_SUMS
|
H A D | poly1305-x86_64-cryptogams.pl | 419 my ($H0,$H1,$H2,$H3,$H4, $T0,$T1,$T2,$T3,$T4, $D0,$D1,$D2,$D3,$D4, $MASK) = 891 vpsrldq \$6,$T1,$T3 894 vpunpcklqdq $T3,$T2,$T3 # 2:3 899 vpsrlq \$4,$T3,$T2 901 vpsrlq \$30,$T3,$T3 903 vpand $MASK,$T3,$T3 # 3 984 vpmuludq $T3,$D4,$D3 # d3 = h3*r0 990 vpmuludq $T3,$H2,$H1 # h3*r1 1012 vpmuludq $T3,$H4,$H4 # h3*s2 1023 vpmuludq $T3,$H3,$H1 # h3*s3 [all …]
|
H A D | camellia-aesni-avx2-asm_64.S | 62 #define roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument 72 vbroadcasti128 .Lpre_tf_hi_s4(%rip), t3; \ 90 filter_8bit(x3, t2, t3, t7, t4); \ 91 filter_8bit(x6, t2, t3, t7, t4); \ 92 vextracti128 $1, x3, t3##_x; \ 111 vaesenclast t4##_x, t3##_x, t3##_x; \ 112 vinserti128 $1, t3##_x, x3, x3; \ 116 vextracti128 $1, x1, t3##_x; \ 127 vaesenclast t4##_x, t3##_x, t3##_x; \ 128 vinserti128 $1, t3##_x, x1, x1; \ [all …]
|
H A D | camellia-aesni-avx-asm_64.S | 50 #define roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument 72 vmovdqa .Lpre_tf_hi_s4(%rip), t3; \ 82 filter_8bit(x3, t2, t3, t7, t6); \ 83 filter_8bit(x6, t2, t3, t7, t6); \ 99 vmovdqa .Lpost_tf_hi_s3(%rip), t3; \ 108 filter_8bit(x2, t2, t3, t7, t6); \ 109 filter_8bit(x5, t2, t3, t7, t6); \ 121 vpsrldq $3, t0, t3; \ 126 vpshufb t6, t3, t3; \ 158 vpxor t3, x4, x4; \ [all …]
|
/linux/arch/riscv/lib/ |
H A D | memmove.S | 35 * Forward Copy: t3 - Index counter of dest 52 mv t3, a0 67 andi t5, t3, -SZREG 70 * If dest(Register t3) rounded down to the nearest naturally 78 beq t5, t3, 1f 100 sub a5, a1, t3 /* Find the difference between src and dest */ 135 addi t3, t3, (2 * SZREG) 139 REG_S t2, ((0 * SZREG) - (2 * SZREG))(t3) 141 beq t3, a2, 2f 148 REG_S t2, ((1 * SZREG) - (2 * SZREG))(t3) [all …]
|
H A D | strlen.S | 58 * t0, t1, t2, t3 67 li t3, SZREG 68 sub t3, t3, t2 98 bgtu t3, a0, 2f 102 li t3, -1 113 beq t1, t3, 1b
|
/linux/arch/alpha/lib/ |
H A D | ev67-strrchr.S | 42 sll t5, 8, t3 # U : 00000000ch000000 46 or t5, t3, t3 # E : 00000000chch0000 51 or t2, t3, t2 # E : 0000chchchch0000 60 cmpbge zero, t2, t3 # E : bits set iff byte == c 62 andnot t3, t4, t3 # E : clear garbage from char test 68 cmovne t3, v0, t6 # E : save previous comparisons match 72 cmovne t3, t3, t8 # E : Latency=2, extra map slot 78 cmpbge zero, t2, t3 # E : bits set iff byte == c 89 and t3, t4, t3 # E : mask out char matches after null 90 cmovne t3, t3, t8 # E : save it, if match found Latency=2, extra map slot [all …]
|
H A D | strrchr.S | 38 cmpbge zero, t2, t3 # e0 : bits set iff byte == c 40 andnot t3, t4, t3 # e0 : clear garbage from char test 46 cmovne t3, v0, t6 # .. e1 : save previous comparisons match 47 cmovne t3, t3, t8 # e0 : 51 cmpbge zero, t2, t3 # e0 : bits set iff byte == c 61 and t3, t4, t3 # e0 : mask out char matches after null 62 cmovne t3, t3, t8 # .. e1 : save it, if match found 63 cmovne t3, v0, t6 # e0 :
|
H A D | strchr.S | 36 cmpbge zero, t1, t3 # e0 : bits set iff byte == c 37 or t2, t3, t0 # e1 : bits set iff char match or zero match 46 cmpbge zero, t1, t3 # .. e1 : bits set iff byte == c 47 or t2, t3, t0 # e0 : 53 and t0, t3, t1 # e0 : bit set iff byte was the char 57 and t0, 0xcc, t3 # .. e1 : 60 cmovne t3, 2, t3 # e0 : 62 addq t2, t3, t2 # e0 :
|
H A D | ev67-strchr.S | 33 and a1, 0xff, t3 # E : 00000000000000ch 37 insbl t3, 6, a3 # U : 00ch000000000000 38 or t5, t3, a1 # E : 000000000000chch 55 cmpbge zero, t1, t3 # E : bits set iff byte == c 56 or t2, t3, t0 # E : bits set iff char match or zero match 74 cmpbge zero, t1, t3 # E : bits set iff byte == c 75 or t2, t3, t0 # E : 76 cttz t3, a2 # U0 : speculative (in case we get a match) 81 and t0, t3, t1 # E : bit set iff byte was the char
|
/linux/arch/arm/crypto/ |
H A D | aes-neonbs-core.S | 169 y0, y1, y2, y3, t0, t1, t2, t3 argument 172 mul_gf4 \x0, \x1, \y0, \y1, \t2, \t3 175 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x2, \x3, \y2, \y3, \t2 182 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x6, \x7, \y2, \y3, \t2 185 mul_gf4 \x4, \x5, \y0, \y1, \t2, \t3 193 t0, t1, t2, t3, s0, s1, s2, s3 194 veor \t3, \x4, \x6 199 veor \s3, \t3, \t0 201 vand \s2, \t3, \s0 202 vorr \t3, \t3, \s0 [all …]
|
H A D | sha512-armv4.pl | 76 $t3="r12"; 104 ldr $t3,[sp,#$Hoff+4] @ h.hi 119 adc $Thi,$Thi,$t3 @ T += h 120 ldr $t3,[sp,#$Goff+4] @ g.hi 124 eor $t1,$t1,$t3 132 eor $t1,$t1,$t3 @ Ch(e,f,g) 133 ldr $t3,[$Ktbl,#$hi] @ K[i].hi 141 adc $Thi,$Thi,$t3 @ T += K[i] 147 ldr $t3,[sp,#$Coff+0] @ c.lo 174 and $Alo,$Alo,$t3 [all …]
|
H A D | nh-neon-core.S | 45 T3 .req q11 52 vld1.8 {T3}, [MESSAGE]! 58 vadd.u32 T0, T3, \k0 59 vadd.u32 T1, T3, \k1 60 vadd.u32 T2, T3, \k2 61 vadd.u32 T3, T3, \k3
|
H A D | sha256-armv4.pl | 53 $T1="r3"; $t3="r3"; 130 and $t3,$t3,$t2 @ (b^c)&=(a^b) 132 eor $t3,$t3,$b @ Maj(a,b,c) 134 @ add $h,$h,$t3 @ h+=Maj(a,b,c) 136 ($t2,$t3)=($t3,$t2); 239 eor $t3,$B,$C @ magic 249 ldreq $t3,[sp,#16*4] @ pull ctx 253 ldr $t0,[$t3,#0] 254 ldr $t1,[$t3,#4] 255 ldr $t2,[$t3,#8] [all …]
|
/linux/arch/mips/lib/ |
H A D | csum_partial.S | 28 #undef t3 32 #define t3 $11 macro 182 CSUM_BIGCHUNK1(src, 0x00, sum, t0, t1, t3, t4) 193 CSUM_BIGCHUNK(src, 0x00, sum, t0, t1, t3, t4) 194 CSUM_BIGCHUNK(src, 0x20, sum, t0, t1, t3, t4) 195 CSUM_BIGCHUNK(src, 0x40, sum, t0, t1, t3, t4) 196 CSUM_BIGCHUNK(src, 0x60, sum, t0, t1, t3, t4) 208 CSUM_BIGCHUNK(src, 0x00, sum, t0, t1, t3, t4) 209 CSUM_BIGCHUNK(src, 0x20, sum, t0, t1, t3, t4) 217 CSUM_BIGCHUNK(src, 0x00, sum, t0, t1, t3, t4) [all …]
|
/linux/arch/loongarch/kernel/ |
H A D | mcount.S | 50 la.pcrel t3, ftrace_graph_return 51 PTR_L t3, t3, 0 52 bne t1, t3, ftrace_graph_caller 54 la.pcrel t3, ftrace_graph_entry 55 PTR_L t3, t3, 0 56 bne t1, t3, ftrace_graph_caller
|
/linux/arch/mips/kernel/ |
H A D | octeon_switch.S | 124 dmfc2 t3, 0x0082 131 sd t3, OCTEON_CP2_3DES_KEY+16(a0) 132 dmfc2 t3, 0x0103 139 sd t3, OCTEON_CP2_AES_IV+8(a0) 140 dmfc2 t3, 0x0107 147 sd t3, OCTEON_CP2_AES_KEY+24(a0) 160 dmfc2 t3, 0x0242 167 sd t3, OCTEON_CP2_HSH_DATW+16(a0) 168 dmfc2 t3, 0x0246 175 sd t3, OCTEON_CP2_HSH_DATW+48(a0) [all …]
|
/linux/arch/arm64/crypto/ |
H A D | nh-neon-core.S | 29 T3 .req v11 38 ld1 {T3.16b}, [MESSAGE], #16 44 add T0.4s, T3.4s, \k0\().4s 45 add T1.4s, T3.4s, \k1\().4s 46 add T2.4s, T3.4s, \k2\().4s 47 add T3.4s, T3.4s, \k3\().4s 53 mov T7.d[0], T3.d[1] 57 umlal PASS3_SUMS.2d, T3.2s, T7.2s
|
H A D | aes-neonbs-core.S | 115 y0, y1, y2, y3, t0, t1, t2, t3 argument 118 mul_gf4 \x0, \x1, \y0, \y1, \t2, \t3 121 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x2, \x3, \y2, \y3, \t2 128 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x6, \x7, \y2, \y3, \t2 131 mul_gf4 \x4, \x5, \y0, \y1, \t2, \t3 139 t0, t1, t2, t3, s0, s1, s2, s3 140 eor \t3, \x4, \x6 145 eor \s3, \t3, \t0 147 and \s2, \t3, \s0 148 orr \t3, \t3, \s0 [all …]
|
H A D | sha512-armv8.pl | 104 ($t0,$t1,$t2,$t3)=map("$reg_t$_",(16,17,19,28)); 153 and $t3,$t3,$t2 // (b^c)&=(a^b) 155 eor $t3,$t3,$b // Maj(a,b,c) 157 add $h,$h,$t3 // h+=Maj(a,b,c) 158 ldr $t3,[$Ktbl],#$SZ // *K++, $t2 in next round 177 and $t3,$t3,$t2 // (b^c)&=(a^b) 181 eor $t3,$t3,$b // Maj(a,b,c) 186 add $h,$h,$t3 // h+=Maj(a,b,c) 187 ldr $t3,[$Ktbl],#$SZ // *K++, $t2 in next round 192 ($t2,$t3)=($t3,$t2); [all …]
|
/linux/include/crypto/ |
H A D | aria.h | 344 u32 *t3) in aria_sbox_layer1_with_pre_diff() argument 358 *t3 = s1[get_u8(*t3, 0)] ^ in aria_sbox_layer1_with_pre_diff() 359 s2[get_u8(*t3, 1)] ^ in aria_sbox_layer1_with_pre_diff() 360 x1[get_u8(*t3, 2)] ^ in aria_sbox_layer1_with_pre_diff() 361 x2[get_u8(*t3, 3)]; in aria_sbox_layer1_with_pre_diff() 366 u32 *t3) in aria_sbox_layer2_with_pre_diff() argument 380 *t3 = x1[get_u8(*t3, 0)] ^ in aria_sbox_layer2_with_pre_diff() 381 x2[get_u8(*t3, 1)] ^ in aria_sbox_layer2_with_pre_diff() 382 s1[get_u8(*t3, 2)] ^ in aria_sbox_layer2_with_pre_diff() 383 s2[get_u8(*t3, 3)]; in aria_sbox_layer2_with_pre_diff() [all …]
|
/linux/arch/mips/cavium-octeon/ |
H A D | octeon-memcpy.S | 108 #undef t3 112 #define t3 $11 macro 190 EXC( LOAD t3, UNIT(3)(src), l_exc_copy) 195 EXC( STORE t3, UNIT(3)(dst), s_exc_p13u) 199 EXC( LOAD t3, UNIT(7)(src), l_exc_copy) 204 EXC( STORE t3, UNIT(7)(dst), s_exc_p9u) 209 EXC( LOAD t3, UNIT(-5)(src), l_exc_copy_rewind16) 213 EXC( STORE t3, UNIT(-5)(dst), s_exc_p5u) 217 EXC( LOAD t3, UNIT(-1)(src), l_exc_copy_rewind16) 221 EXC( STORE t3, UNIT(-1)(dst), s_exc_p1u) [all …]
|
/linux/arch/alpha/include/asm/ |
H A D | word-at-a-time.h | 42 unsigned long t1, t2, t3; in find_zero() 48 t3 = bits & 0xaa; in find_zero() 51 if (t3) t3 = 1; in find_zero() 52 return t1 + t2 + t3; in find_zero()
|