Home
last modified time | relevance | path

Searched refs:x3 (Results 1 – 25 of 518) sorted by relevance

12345678910>>...21

/linux/arch/x86/crypto/
H A Dserpent-avx2-asm_64.S51 #define S0_1(x0, x1, x2, x3, x4) \ argument
52 vpor x0, x3, tp; \
53 vpxor x3, x0, x0; \
54 vpxor x2, x3, x4; \
56 vpxor x1, tp, x3; \
60 #define S0_2(x0, x1, x2, x3, x4) \ argument
61 vpxor x3, x0, x0; \
65 vpxor x2, x3, x3; \
70 #define S1_1(x0, x1, x2, x3, x4) \ argument
72 vpxor x3, x0, x0; \
[all …]
H A Dserpent-sse2-x86_64-asm_64.S41 #define S0_1(x0, x1, x2, x3, x4) \ argument
42 movdqa x3, x4; \
43 por x0, x3; \
47 pxor x1, x3; \
51 #define S0_2(x0, x1, x2, x3, x4) \ argument
52 pxor x3, x0; \
56 pxor x2, x3; \
61 #define S1_1(x0, x1, x2, x3, x4) \ argument
64 pxor x3, x0; \
65 pxor RNOT, x3; \
[all …]
H A Dserpent-sse2-i586-asm_32.S42 #define K(x0, x1, x2, x3, x4, i) \ argument
50 pxor x4, x3;
52 #define LK(x0, x1, x2, x3, x4, i) \ argument
69 pxor x2, x3; \
70 pxor x4, x3; \
71 movdqa x3, x4; \
72 pslld $7, x3; \
74 por x4, x3; \
78 pxor x3, x0; \
79 pxor x3, x2; \
[all …]
H A Daria-aesni-avx-asm_64.S53 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \ argument
57 vpunpckldq x3, x2, t1; \
58 vpunpckhdq x3, x2, x2; \
63 vpunpckhqdq x2, t2, x3; \
171 #define inpack16_pre(x0, x1, x2, x3, \ argument
179 vmovdqu (3 * 16)(rio), x3; \
194 #define inpack16_post(x0, x1, x2, x3, \ argument
199 byteslice_16x16b(x0, x1, x2, x3, \
208 vmovdqu x3, 3 * 16(mem_ab); \
222 #define write_output(x0, x1, x2, x3, \ argument
[all …]
H A Daria-aesni-avx2-asm_64.S69 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \ argument
73 vpunpckldq x3, x2, t1; \
74 vpunpckhdq x3, x2, x2; \
79 vpunpckhqdq x2, t2, x3; \
187 #define inpack16_pre(x0, x1, x2, x3, \ argument
195 vmovdqu (3 * 32)(rio), x3; \
210 #define inpack16_post(x0, x1, x2, x3, \ argument
215 byteslice_16x16b(x0, x1, x2, x3, \
224 vmovdqu x3, 3 * 32(mem_ab); \
238 #define write_output(x0, x1, x2, x3, \ argument
[all …]
H A Dglue_helper-asm-avx.S8 #define load_8way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ argument
12 vmovdqu (3*16)(src), x3; \
18 #define store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument
22 vmovdqu x3, (3*16)(dst); \
28 #define store_cbc_8way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument
31 vpxor (2*16)(src), x3, x3; \
36 store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7);
H A Dglue_helper-asm-avx2.S8 #define load_16way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ argument
12 vmovdqu (3*32)(src), x3; \
18 #define store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument
22 vmovdqu x3, (3*32)(dst); \
28 #define store_cbc_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7, t0) \ argument
34 vpxor (2*32+16)(src), x3, x3; \
39 store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7);
H A Daria-gfni-avx512-asm_64.S53 #define transpose_4x4(x0, x1, x2, x3, t1, t2) \ argument
57 vpunpckldq x3, x2, t1; \
58 vpunpckhdq x3, x2, x2; \
63 vpunpckhqdq x2, t2, x3; \
171 #define inpack16_pre(x0, x1, x2, x3, \ argument
179 vmovdqu64 (3 * 64)(rio), x3; \
194 #define inpack16_post(x0, x1, x2, x3, \ argument
199 byteslice_16x16b(x0, x1, x2, x3, \
208 vmovdqu64 x3, 3 * 64(mem_ab); \
222 #define write_output(x0, x1, x2, x3, \ argument
[all …]
H A Dcast6-avx-x86_64-asm_64.S190 #define transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \ argument
193 vpunpckldq x3, x2, t1; \
194 vpunpckhdq x3, x2, x3; \
198 vpunpcklqdq x3, t2, x2; \
199 vpunpckhqdq x3, t2, x3;
201 #define inpack_blocks(x0, x1, x2, x3, t0, t1, t2, rmask) \ argument
205 vpshufb rmask, x3, x3; \
207 transpose_4x4(x0, x1, x2, x3, t0, t1, t2)
209 #define outunpack_blocks(x0, x1, x2, x3, t0, t1, t2, rmask) \ argument
210 transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \
[all …]
/linux/arch/arm64/lib/
H A Dtishift.S12 mov x3, #64
13 sub x3, x3, x2
14 cmp x3, #0
17 lsr x3, x0, x3
19 orr x1, x1, x3
34 mov x3, #64
35 sub x3, x3, x2
36 cmp x3, #0
39 lsl x3, x1, x3
41 orr x0, x0, x3
[all …]
H A Dmte.S72 mov x3, x1
74 1: ldgm x4, [x3]
77 add x3, x3, x5
93 mov x3, x1
104 2: sub x0, x1, x3 // update the number of tags set
118 mov x3, x0
130 2: sub x0, x0, x3 // update the number of tags copied
/linux/arch/arm64/mm/
H A Dcache.S34 mov x3, x1
35 dcache_by_line_op cvau, ish, x2, x3, x4, x5, \fixup
41 invalidate_icache_by_line x0, x1, x2, x3, \fixup
72 uaccess_ttbr0_enable x2, x3, x4
98 invalidate_icache_by_line x0, x1, x2, x3
112 dcache_by_line_op civac, sy, x0, x1, x2, x3
131 dcache_by_line_op cvau, ish, x0, x1, x2, x3
146 dcache_line_size x2, x3
147 sub x3, x2, #1
148 tst x1, x3 // end cache line aligned?
[all …]
/linux/lib/crc/arm64/
H A Dcrc32-core.S66 ldp x3, x4, [x1]
70 \order x3, x4, x5, x6
73 crc32\c\()x w8, w0, x3
74 csel x3, x3, x4, eq
77 lsr x4, x3, #32
79 csel x3, x3, x4, eq
95 32: ldp x3, x4, [x1], #32
98 \order x3, x4, x5, x6
99 crc32\c\()x w0, w0, x3
108 ldr x3, [x1], #8
[all …]
/linux/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_msg_arm64.h56 register u64 x3 asm("x3") = VMWARE_HYPERVISOR_PORT; in vmware_hypercall1()
65 : "r" (x1), "r" (x2), "r" (x3), "r" (x7) in vmware_hypercall1()
79 register u64 x3 asm("x3") = in3 | VMWARE_HYPERVISOR_PORT; in vmware_hypercall5()
90 : "r" (x1), "r" (x3), "r" (x4), "r" (x5), "r" (x7) in vmware_hypercall5()
105 register u64 x3 asm("x3") = in3 | VMWARE_HYPERVISOR_PORT; in vmware_hypercall6()
115 : "+r" (x0), "+r" (x2), "+r" (x3), "=r" (x4), "=r" (x5) in vmware_hypercall6()
120 *out3 = x3; in vmware_hypercall6()
135 register u64 x3 asm("x3") = in3 | VMWARE_HYPERVISOR_PORT; in vmware_hypercall7()
145 : "+r" (x0), "+r" (x1), "+r" (x2), "+r" (x3) in vmware_hypercall7()
151 *out3 = x3; in vmware_hypercall7()
[all …]
/linux/arch/powerpc/boot/dts/fsl/
H A Dmpc8572ds.dtsi169 reg = <0x3>;
251 0x8800 0x0 0x0 0x2 &mpic 0x3 0x1 0 0
252 0x8800 0x0 0x0 0x3 &mpic 0x4 0x1 0 0
257 0x8900 0x0 0x0 0x2 &mpic 0x3 0x1 0 0
258 0x8900 0x0 0x0 0x3 &mpic 0x4 0x1 0 0
263 0x8a00 0x0 0x0 0x2 &mpic 0x3 0x1 0 0
264 0x8a00 0x0 0x0 0x3 &mpic 0x4 0x1 0 0
269 0x8b00 0x0 0x0 0x2 &mpic 0x3 0x1 0 0
270 0x8b00 0x0 0x0 0x3 &mpic 0x4 0x1 0 0
275 0x8c00 0x0 0x0 0x2 &mpic 0x3 0x1 0 0
[all …]
H A Dmpc8569mds.dts35 0x3 0x0 0x0 0xfc000000 0x00008000
144 0x2 0x1e 0x3 0x0 0x2 0x0 /* QE_MUX_MDIO */
146 0x0 0x0 0x1 0x0 0x3 0x0 /* ENET1_TXD0_SER1_TXD0 */
147 0x0 0x1 0x1 0x0 0x3 0x0 /* ENET1_TXD1_SER1_TXD1 */
149 0x0 0x3 0x1 0x0 0x2 0x0 /* ENET1_TXD3_SER1_TXD3 */
150 0x0 0x6 0x2 0x0 0x3 0x0 /* ENET1_RXD0_SER1_RXD0 */
155 0x0 0xc 0x2 0x0 0x3 0x0 /* ENET1_RX_DV_SER1_CTS_B */
164 0x2 0x1e 0x3 0x0 0x2 0x0 /* QE_MUX_MDIO */
165 0x2 0x10 0x2 0x0 0x3 0x0 /* CLK17 */
175 0x0 0x1a 0x2 0x0 0x3 0x0 /* ENET2_RX_DV_SER2_CTS_B */
[all …]
/linux/arch/arm64/kvm/hyp/
H A Dentry.S75 ldp x2, x3, [x29, #CPU_XREG_OFFSET(2)]
132 stp x2, x3, [x1, #CPU_XREG_OFFSET(2)]
135 ldp x2, x3, [sp], #16 // x0, x1
138 stp x2, x3, [x1, #CPU_XREG_OFFSET(0)]
153 adr_this_cpu x2, kvm_hyp_ctxt, x3
160 ptrauth_switch_to_hyp x1, x2, x3, x4, x5
163 mte_switch_to_hyp x1, x2, x3
166 restore_sp_el0 x2, x3
171 set_loaded_vcpu xzr, x2, x3
197 mrs x3, esr_el2
[all …]
/linux/lib/crypto/x86/
H A Dchacha-avx512vl-x86_64.S56 # x0 += x1, x3 = rotl32(x3 ^ x0, 16)
61 # x2 += x3, x1 = rotl32(x1 ^ x2, 12)
66 # x0 += x1, x3 = rotl32(x3 ^ x0, 8)
71 # x2 += x3, x1 = rotl32(x1 ^ x2, 7)
80 # x3 = shuffle32(x3, MASK(2, 1, 0, 3))
83 # x0 += x1, x3 = rotl32(x3 ^ x0, 16)
88 # x2 += x3, x1 = rotl32(x1 ^ x2, 12)
93 # x0 += x1, x3 = rotl32(x3 ^ x0, 8)
98 # x2 += x3, x1 = rotl32(x1 ^ x2, 7)
107 # x3 = shuffle32(x3, MASK(0, 3, 2, 1))
[all …]
/linux/arch/sparc/include/asm/
H A Dsfp-machine_32.h108 #define __FP_FRAC_ADD_4(r3,r2,r1,r0,x3,x2,x1,x0,y3,y2,y1,y0) \ argument
120 : "%rJ" ((USItype)(x3)), \
133 #define __FP_FRAC_SUB_4(r3,r2,r1,r0,x3,x2,x1,x0,y3,y2,y1,y0) \ argument
145 : "%rJ" ((USItype)(x3)), \
160 #define __FP_FRAC_DEC_4(x3,x2,x1,x0,y3,y2,y1,y0) __FP_FRAC_SUB_4(x3,x2,x1,x0,x3,x2,x1,x0,y3,y2,y1,y… argument
162 #define __FP_FRAC_ADDI_4(x3,x2,x1,x0,i) \ argument
167 : "=&r" (x3), \
172 "0" ((USItype)(x3)), \
/linux/lib/crypto/powerpc/
H A Dcurve25519.h112 fe51 x1, x2, z2, x3, z3; in curve25519_fe51() local
124 x3[0] = x1[0]; in curve25519_fe51()
125 x3[1] = x1[1]; in curve25519_fe51()
126 x3[2] = x1[2]; in curve25519_fe51()
127 x3[3] = x1[3]; in curve25519_fe51()
128 x3[4] = x1[4]; in curve25519_fe51()
143 x25519_cswap(x2, x3, swap); in curve25519_fe51()
149 fsub(d, x3, z3); // D = x_3 - z_3 in curve25519_fe51()
150 fadd(c, x3, z3); // C = x_3 + z_3 in curve25519_fe51()
164 fsqr(x3, dacb_p); // x3 = (DA + CB)^2 in curve25519_fe51()
/linux/arch/arm/crypto/
H A Daes-neonbs-core.S151 .macro mul_gf4_n_gf4, x0, x1, y0, y1, t0, x2, x3, y2, y3, t1
157 veor \x2, \x2, \x3
159 vand \x3, \x3, \y2
163 veor \x2, \x2, \x3
165 veor \x3, \x3, \t1
168 .macro mul_gf16_2, x0, x1, x2, x3, x4, x5, x6, x7, \
171 veor \t1, \x1, \x3
175 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x2, \x3, \y2, \y3, \t2
179 veor \x3, \x3, \t1
192 .macro inv_gf256, x0, x1, x2, x3, x4, x5, x6, x7, \
[all …]
/linux/arch/arm64/crypto/
H A Daes-neonbs-core.S97 .macro mul_gf4_n_gf4, x0, x1, y0, y1, t0, x2, x3, y2, y3, t1
103 eor \x2, \x2, \x3
105 and \x3, \x3, \y2
109 eor \x2, \x2, \x3
111 eor \x3, \x3, \t1
114 .macro mul_gf16_2, x0, x1, x2, x3, x4, x5, x6, x7, \
117 eor \t1, \x1, \x3
121 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x2, \x3, \y2, \y3, \t2
125 eor \x3, \x3, \t1
138 .macro inv_gf256, x0, x1, x2, x3, x4, x5, x6, x7, \
[all …]
/linux/arch/arm64/boot/dts/qcom/
H A Dpmi8998.dtsi51 reg = <0x3 SPMI_USID>;
59 interrupts = <0x3 0xdc 0x2 IRQ_TYPE_EDGE_RISING>,
60 <0x3 0xdc 0x0 IRQ_TYPE_LEVEL_HIGH>;
65 interrupts = <0x3 0xde 0x1 IRQ_TYPE_EDGE_RISING>,
66 <0x3 0xde 0x0 IRQ_TYPE_LEVEL_LOW>;
90 interrupts = <0x3 0xd8 0x1 IRQ_TYPE_EDGE_RISING>,
91 <0x3 0xd8 0x2 IRQ_TYPE_EDGE_RISING>;
/linux/arch/arm64/kernel/
H A Dhibernate-asm.S61 mov x19, x3
67 copy_page x0, x1, x2, x3, x4, x5, x6, x7, x8, x9
71 raw_dcache_line_size x2, x3
72 sub x3, x2, #1
73 bic x4, x10, x3
/linux/arch/arm64/kvm/hyp/nvhe/
H A Dhost.S21 stp x2, x3, [x0, #CPU_XREG_OFFSET(2)]
24 ldp x2, x3, [sp], #16 // x0, x1
27 stp x2, x3, [x0, #CPU_XREG_OFFSET(0)]
79 ldp x2, x3, [x29, #CPU_XREG_OFFSET(2)]
135 mov x4, x3
136 mov x3, x2
137 hyp_pa x3, x6
220 mrs x3, par_el1
271 ldp x2, x3, [x18, #CPU_XREG_OFFSET(2)]
283 stp x2, x3, [x18, #CPU_XREG_OFFSET(2)]

12345678910>>...21