| /linux/arch/x86/crypto/ |
| H A D | serpent-avx2-asm_64.S | 51 #define S0_1(x0, x1, x2, x3, x4) \ argument 52 vpor x0, x3, tp; \ 53 vpxor x3, x0, x0; \ 54 vpxor x2, x3, x4; \ 56 vpxor x1, tp, x3; \ 60 #define S0_2(x0, x1, x2, x3, x4) \ argument 61 vpxor x3, x0, x0; \ 65 vpxor x2, x3, x3; \ 70 #define S1_1(x0, x1, x2, x3, x4) \ argument 72 vpxor x3, x0, x0; \ [all …]
|
| H A D | serpent-sse2-x86_64-asm_64.S | 41 #define S0_1(x0, x1, x2, x3, x4) \ argument 42 movdqa x3, x4; \ 43 por x0, x3; \ 47 pxor x1, x3; \ 51 #define S0_2(x0, x1, x2, x3, x4) \ argument 52 pxor x3, x0; \ 56 pxor x2, x3; \ 61 #define S1_1(x0, x1, x2, x3, x4) \ argument 64 pxor x3, x0; \ 65 pxor RNOT, x3; \ [all …]
|
| H A D | serpent-sse2-i586-asm_32.S | 42 #define K(x0, x1, x2, x3, x4, i) \ argument 50 pxor x4, x3; 52 #define LK(x0, x1, x2, x3, x4, i) \ argument 69 pxor x2, x3; \ 70 pxor x4, x3; \ 71 movdqa x3, x4; \ 72 pslld $7, x3; \ 74 por x4, x3; \ 78 pxor x3, x0; \ 79 pxor x3, x2; \ [all …]
|
| H A D | aria-aesni-avx-asm_64.S | 53 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \ argument 57 vpunpckldq x3, x2, t1; \ 58 vpunpckhdq x3, x2, x2; \ 63 vpunpckhqdq x2, t2, x3; \ 171 #define inpack16_pre(x0, x1, x2, x3, \ argument 179 vmovdqu (3 * 16)(rio), x3; \ 194 #define inpack16_post(x0, x1, x2, x3, \ argument 199 byteslice_16x16b(x0, x1, x2, x3, \ 208 vmovdqu x3, 3 * 16(mem_ab); \ 222 #define write_output(x0, x1, x2, x3, \ argument [all …]
|
| H A D | aria-aesni-avx2-asm_64.S | 69 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \ argument 73 vpunpckldq x3, x2, t1; \ 74 vpunpckhdq x3, x2, x2; \ 79 vpunpckhqdq x2, t2, x3; \ 187 #define inpack16_pre(x0, x1, x2, x3, \ argument 195 vmovdqu (3 * 32)(rio), x3; \ 210 #define inpack16_post(x0, x1, x2, x3, \ argument 215 byteslice_16x16b(x0, x1, x2, x3, \ 224 vmovdqu x3, 3 * 32(mem_ab); \ 238 #define write_output(x0, x1, x2, x3, \ argument [all …]
|
| H A D | glue_helper-asm-avx.S | 8 #define load_8way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 12 vmovdqu (3*16)(src), x3; \ 18 #define store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 22 vmovdqu x3, (3*16)(dst); \ 28 #define store_cbc_8way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 31 vpxor (2*16)(src), x3, x3; \ 36 store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7);
|
| H A D | glue_helper-asm-avx2.S | 8 #define load_16way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 12 vmovdqu (3*32)(src), x3; \ 18 #define store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 22 vmovdqu x3, (3*32)(dst); \ 28 #define store_cbc_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7, t0) \ argument 34 vpxor (2*32+16)(src), x3, x3; \ 39 store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7);
|
| H A D | aria-gfni-avx512-asm_64.S | 53 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \ argument 57 vpunpckldq x3, x2, t1; \ 58 vpunpckhdq x3, x2, x2; \ 63 vpunpckhqdq x2, t2, x3; \ 171 #define inpack16_pre(x0, x1, x2, x3, \ argument 179 vmovdqu64 (3 * 64)(rio), x3; \ 194 #define inpack16_post(x0, x1, x2, x3, \ argument 199 byteslice_16x16b(x0, x1, x2, x3, \ 208 vmovdqu64 x3, 3 * 64(mem_ab); \ 222 #define write_output(x0, x1, x2, x3, \ argument [all …]
|
| H A D | cast6-avx-x86_64-asm_64.S | 190 #define transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ argument 193 vpunpckldq x3, x2, t1; \ 194 vpunpckhdq x3, x2, x3; \ 198 vpunpcklqdq x3, t2, x2; \ 199 vpunpckhqdq x3, t2, x3; 201 #define inpack_blocks(x0, x1, x2, x3, t0, t1, t2, rmask) \ argument 205 vpshufb rmask, x3, x3; \ 207 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) 209 #define outunpack_blocks(x0, x1, x2, x3, t0, t1, t2, rmask) \ argument 210 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ [all …]
|
| /linux/arch/arm64/lib/ |
| H A D | tishift.S | 12 mov x3, #64 13 sub x3, x3, x2 14 cmp x3, #0 17 lsr x3, x0, x3 19 orr x1, x1, x3 34 mov x3, #64 35 sub x3, x3, x2 36 cmp x3, #0 39 lsl x3, x1, x3 41 orr x0, x0, x3 [all …]
|
| H A D | mte.S | 72 mov x3, x1 74 1: ldgm x4, [x3] 77 add x3, x3, x5 93 mov x3, x1 104 2: sub x0, x1, x3 // update the number of tags set 118 mov x3, x0 130 2: sub x0, x0, x3 // update the number of tags copied
|
| /linux/arch/arm64/mm/ |
| H A D | cache.S | 34 mov x3, x1 35 dcache_by_line_op cvau, ish, x2, x3, x4, x5, \fixup 41 invalidate_icache_by_line x0, x1, x2, x3, \fixup 72 uaccess_ttbr0_enable x2, x3, x4 98 invalidate_icache_by_line x0, x1, x2, x3 112 dcache_by_line_op civac, sy, x0, x1, x2, x3 131 dcache_by_line_op cvau, ish, x0, x1, x2, x3 146 dcache_line_size x2, x3 147 sub x3, x2, #1 148 tst x1, x3 // end cache line aligned? [all …]
|
| /linux/lib/crc/arm64/ |
| H A D | crc32-core.S | 66 ldp x3, x4, [x1] 70 \order x3, x4, x5, x6 73 crc32\c\()x w8, w0, x3 74 csel x3, x3, x4, eq 77 lsr x4, x3, #32 79 csel x3, x3, x4, eq 95 32: ldp x3, x4, [x1], #32 98 \order x3, x4, x5, x6 99 crc32\c\()x w0, w0, x3 108 ldr x3, [x1], #8 [all …]
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_msg_arm64.h | 56 register u64 x3 asm("x3") = VMWARE_HYPERVISOR_PORT; in vmware_hypercall1() 65 : "r" (x1), "r" (x2), "r" (x3), "r" (x7) in vmware_hypercall1() 79 register u64 x3 asm("x3") = in3 | VMWARE_HYPERVISOR_PORT; in vmware_hypercall5() 90 : "r" (x1), "r" (x3), "r" (x4), "r" (x5), "r" (x7) in vmware_hypercall5() 105 register u64 x3 asm("x3") = in3 | VMWARE_HYPERVISOR_PORT; in vmware_hypercall6() 115 : "+r" (x0), "+r" (x2), "+r" (x3), "=r" (x4), "=r" (x5) in vmware_hypercall6() 120 *out3 = x3; in vmware_hypercall6() 135 register u64 x3 asm("x3") = in3 | VMWARE_HYPERVISOR_PORT; in vmware_hypercall7() 145 : "+r" (x0), "+r" (x1), "+r" (x2), "+r" (x3) in vmware_hypercall7() 151 *out3 = x3; in vmware_hypercall7() [all …]
|
| /linux/arch/powerpc/boot/dts/fsl/ |
| H A D | mpc8572ds.dtsi | 169 reg = <0x3>; 251 0x8800 0x0 0x0 0x2 &mpic 0x3 0x1 0 0 252 0x8800 0x0 0x0 0x3 &mpic 0x4 0x1 0 0 257 0x8900 0x0 0x0 0x2 &mpic 0x3 0x1 0 0 258 0x8900 0x0 0x0 0x3 &mpic 0x4 0x1 0 0 263 0x8a00 0x0 0x0 0x2 &mpic 0x3 0x1 0 0 264 0x8a00 0x0 0x0 0x3 &mpic 0x4 0x1 0 0 269 0x8b00 0x0 0x0 0x2 &mpic 0x3 0x1 0 0 270 0x8b00 0x0 0x0 0x3 &mpic 0x4 0x1 0 0 275 0x8c00 0x0 0x0 0x2 &mpic 0x3 0x1 0 0 [all …]
|
| H A D | mpc8569mds.dts | 35 0x3 0x0 0x0 0xfc000000 0x00008000 144 0x2 0x1e 0x3 0x0 0x2 0x0 /* QE_MUX_MDIO */ 146 0x0 0x0 0x1 0x0 0x3 0x0 /* ENET1_TXD0_SER1_TXD0 */ 147 0x0 0x1 0x1 0x0 0x3 0x0 /* ENET1_TXD1_SER1_TXD1 */ 149 0x0 0x3 0x1 0x0 0x2 0x0 /* ENET1_TXD3_SER1_TXD3 */ 150 0x0 0x6 0x2 0x0 0x3 0x0 /* ENET1_RXD0_SER1_RXD0 */ 155 0x0 0xc 0x2 0x0 0x3 0x0 /* ENET1_RX_DV_SER1_CTS_B */ 164 0x2 0x1e 0x3 0x0 0x2 0x0 /* QE_MUX_MDIO */ 165 0x2 0x10 0x2 0x0 0x3 0x0 /* CLK17 */ 175 0x0 0x1a 0x2 0x0 0x3 0x0 /* ENET2_RX_DV_SER2_CTS_B */ [all …]
|
| /linux/arch/arm64/kvm/hyp/ |
| H A D | entry.S | 75 ldp x2, x3, [x29, #CPU_XREG_OFFSET(2)] 132 stp x2, x3, [x1, #CPU_XREG_OFFSET(2)] 135 ldp x2, x3, [sp], #16 // x0, x1 138 stp x2, x3, [x1, #CPU_XREG_OFFSET(0)] 153 adr_this_cpu x2, kvm_hyp_ctxt, x3 160 ptrauth_switch_to_hyp x1, x2, x3, x4, x5 163 mte_switch_to_hyp x1, x2, x3 166 restore_sp_el0 x2, x3 171 set_loaded_vcpu xzr, x2, x3 197 mrs x3, esr_el2 [all …]
|
| /linux/lib/crypto/x86/ |
| H A D | chacha-avx512vl-x86_64.S | 56 # x0 += x1, x3 = rotl32(x3 ^ x0, 16) 61 # x2 += x3, x1 = rotl32(x1 ^ x2, 12) 66 # x0 += x1, x3 = rotl32(x3 ^ x0, 8) 71 # x2 += x3, x1 = rotl32(x1 ^ x2, 7) 80 # x3 = shuffle32(x3, MASK(2, 1, 0, 3)) 83 # x0 += x1, x3 = rotl32(x3 ^ x0, 16) 88 # x2 += x3, x1 = rotl32(x1 ^ x2, 12) 93 # x0 += x1, x3 = rotl32(x3 ^ x0, 8) 98 # x2 += x3, x1 = rotl32(x1 ^ x2, 7) 107 # x3 = shuffle32(x3, MASK(0, 3, 2, 1)) [all …]
|
| /linux/arch/sparc/include/asm/ |
| H A D | sfp-machine_32.h | 108 #define __FP_FRAC_ADD_4(r3,r2,r1,r0,x3,x2,x1,x0,y3,y2,y1,y0) \ argument 120 : "%rJ" ((USItype)(x3)), \ 133 #define __FP_FRAC_SUB_4(r3,r2,r1,r0,x3,x2,x1,x0,y3,y2,y1,y0) \ argument 145 : "%rJ" ((USItype)(x3)), \ 160 #define __FP_FRAC_DEC_4(x3,x2,x1,x0,y3,y2,y1,y0) __FP_FRAC_SUB_4(x3,x2,x1,x0,x3,x2,x1,x0,y3,y2,y1,y… argument 162 #define __FP_FRAC_ADDI_4(x3,x2,x1,x0,i) \ argument 167 : "=&r" (x3), \ 172 "0" ((USItype)(x3)), \
|
| /linux/lib/crypto/powerpc/ |
| H A D | curve25519.h | 112 fe51 x1, x2, z2, x3, z3; in curve25519_fe51() local 124 x3[0] = x1[0]; in curve25519_fe51() 125 x3[1] = x1[1]; in curve25519_fe51() 126 x3[2] = x1[2]; in curve25519_fe51() 127 x3[3] = x1[3]; in curve25519_fe51() 128 x3[4] = x1[4]; in curve25519_fe51() 143 x25519_cswap(x2, x3, swap); in curve25519_fe51() 149 fsub(d, x3, z3); // D = x_3 - z_3 in curve25519_fe51() 150 fadd(c, x3, z3); // C = x_3 + z_3 in curve25519_fe51() 164 fsqr(x3, dacb_p); // x3 = (DA + CB)^2 in curve25519_fe51()
|
| /linux/arch/arm/crypto/ |
| H A D | aes-neonbs-core.S | 151 .macro mul_gf4_n_gf4, x0, x1, y0, y1, t0, x2, x3, y2, y3, t1 157 veor \x2, \x2, \x3 159 vand \x3, \x3, \y2 163 veor \x2, \x2, \x3 165 veor \x3, \x3, \t1 168 .macro mul_gf16_2, x0, x1, x2, x3, x4, x5, x6, x7, \ 171 veor \t1, \x1, \x3 175 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x2, \x3, \y2, \y3, \t2 179 veor \x3, \x3, \t1 192 .macro inv_gf256, x0, x1, x2, x3, x4, x5, x6, x7, \ [all …]
|
| /linux/arch/arm64/crypto/ |
| H A D | aes-neonbs-core.S | 97 .macro mul_gf4_n_gf4, x0, x1, y0, y1, t0, x2, x3, y2, y3, t1 103 eor \x2, \x2, \x3 105 and \x3, \x3, \y2 109 eor \x2, \x2, \x3 111 eor \x3, \x3, \t1 114 .macro mul_gf16_2, x0, x1, x2, x3, x4, x5, x6, x7, \ 117 eor \t1, \x1, \x3 121 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x2, \x3, \y2, \y3, \t2 125 eor \x3, \x3, \t1 138 .macro inv_gf256, x0, x1, x2, x3, x4, x5, x6, x7, \ [all …]
|
| /linux/arch/arm64/boot/dts/qcom/ |
| H A D | pmi8998.dtsi | 51 reg = <0x3 SPMI_USID>; 59 interrupts = <0x3 0xdc 0x2 IRQ_TYPE_EDGE_RISING>, 60 <0x3 0xdc 0x0 IRQ_TYPE_LEVEL_HIGH>; 65 interrupts = <0x3 0xde 0x1 IRQ_TYPE_EDGE_RISING>, 66 <0x3 0xde 0x0 IRQ_TYPE_LEVEL_LOW>; 90 interrupts = <0x3 0xd8 0x1 IRQ_TYPE_EDGE_RISING>, 91 <0x3 0xd8 0x2 IRQ_TYPE_EDGE_RISING>;
|
| /linux/arch/arm64/kernel/ |
| H A D | hibernate-asm.S | 61 mov x19, x3 67 copy_page x0, x1, x2, x3, x4, x5, x6, x7, x8, x9 71 raw_dcache_line_size x2, x3 72 sub x3, x2, #1 73 bic x4, x10, x3
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| H A D | host.S | 21 stp x2, x3, [x0, #CPU_XREG_OFFSET(2)] 24 ldp x2, x3, [sp], #16 // x0, x1 27 stp x2, x3, [x0, #CPU_XREG_OFFSET(0)] 79 ldp x2, x3, [x29, #CPU_XREG_OFFSET(2)] 135 mov x4, x3 136 mov x3, x2 137 hyp_pa x3, x6 220 mrs x3, par_el1 271 ldp x2, x3, [x18, #CPU_XREG_OFFSET(2)] 283 stp x2, x3, [x18, #CPU_XREG_OFFSET(2)]
|