Lines Matching full:asm
86 asm volatile("movdqu %0, %%xmm0" :: "m" ((ctx)->sse[0])); \
87 asm volatile("movdqu %0, %%xmm1" :: "m" ((ctx)->sse[1])); \
88 asm volatile("movdqu %0, %%xmm2" :: "m" ((ctx)->sse[2])); \
89 asm volatile("movdqu %0, %%xmm3" :: "m" ((ctx)->sse[3])); \
94 asm volatile("movdqu %%xmm0, %0" : "=m" ((ctx)->sse[0])); \
95 asm volatile("movdqu %%xmm1, %0" : "=m" ((ctx)->sse[1])); \
96 asm volatile("movdqu %%xmm2, %0" : "=m" ((ctx)->sse[2])); \
97 asm volatile("movdqu %%xmm3, %0" : "=m" ((ctx)->sse[3])); \
108 asm volatile("pxor %xmm4, %xmm4"); in fletcher_4_sse2_native()
111 asm volatile("movdqu %0, %%xmm5" :: "m"(*ip)); in fletcher_4_sse2_native()
112 asm volatile("movdqa %xmm5, %xmm6"); in fletcher_4_sse2_native()
113 asm volatile("punpckldq %xmm4, %xmm5"); in fletcher_4_sse2_native()
114 asm volatile("punpckhdq %xmm4, %xmm6"); in fletcher_4_sse2_native()
115 asm volatile("paddq %xmm5, %xmm0"); in fletcher_4_sse2_native()
116 asm volatile("paddq %xmm0, %xmm1"); in fletcher_4_sse2_native()
117 asm volatile("paddq %xmm1, %xmm2"); in fletcher_4_sse2_native()
118 asm volatile("paddq %xmm2, %xmm3"); in fletcher_4_sse2_native()
119 asm volatile("paddq %xmm6, %xmm0"); in fletcher_4_sse2_native()
120 asm volatile("paddq %xmm0, %xmm1"); in fletcher_4_sse2_native()
121 asm volatile("paddq %xmm1, %xmm2"); in fletcher_4_sse2_native()
122 asm volatile("paddq %xmm2, %xmm3"); in fletcher_4_sse2_native()
139 asm volatile("movd %0, %%xmm5" :: "r"(scratch1)); in fletcher_4_sse2_byteswap()
140 asm volatile("movd %0, %%xmm6" :: "r"(scratch2)); in fletcher_4_sse2_byteswap()
141 asm volatile("punpcklqdq %xmm6, %xmm5"); in fletcher_4_sse2_byteswap()
142 asm volatile("paddq %xmm5, %xmm0"); in fletcher_4_sse2_byteswap()
143 asm volatile("paddq %xmm0, %xmm1"); in fletcher_4_sse2_byteswap()
144 asm volatile("paddq %xmm1, %xmm2"); in fletcher_4_sse2_byteswap()
145 asm volatile("paddq %xmm2, %xmm3"); in fletcher_4_sse2_byteswap()
183 asm volatile("movdqu %0, %%xmm7"::"m" (mask)); in fletcher_4_ssse3_byteswap()
184 asm volatile("pxor %xmm4, %xmm4"); in fletcher_4_ssse3_byteswap()
187 asm volatile("movdqu %0, %%xmm5"::"m" (*ip)); in fletcher_4_ssse3_byteswap()
188 asm volatile("pshufb %xmm7, %xmm5"); in fletcher_4_ssse3_byteswap()
189 asm volatile("movdqa %xmm5, %xmm6"); in fletcher_4_ssse3_byteswap()
190 asm volatile("punpckldq %xmm4, %xmm5"); in fletcher_4_ssse3_byteswap()
191 asm volatile("punpckhdq %xmm4, %xmm6"); in fletcher_4_ssse3_byteswap()
192 asm volatile("paddq %xmm5, %xmm0"); in fletcher_4_ssse3_byteswap()
193 asm volatile("paddq %xmm0, %xmm1"); in fletcher_4_ssse3_byteswap()
194 asm volatile("paddq %xmm1, %xmm2"); in fletcher_4_ssse3_byteswap()
195 asm volatile("paddq %xmm2, %xmm3"); in fletcher_4_ssse3_byteswap()
196 asm volatile("paddq %xmm6, %xmm0"); in fletcher_4_ssse3_byteswap()
197 asm volatile("paddq %xmm0, %xmm1"); in fletcher_4_ssse3_byteswap()
198 asm volatile("paddq %xmm1, %xmm2"); in fletcher_4_ssse3_byteswap()
199 asm volatile("paddq %xmm2, %xmm3"); in fletcher_4_ssse3_byteswap()