Lines Matching refs:__asm
90 __asm("vmovdqu %0, %%ymm0" :: "m" ((ctx)->avx[0])); \
91 __asm("vmovdqu %0, %%ymm1" :: "m" ((ctx)->avx[1])); \
92 __asm("vmovdqu %0, %%ymm2" :: "m" ((ctx)->avx[2])); \
93 __asm("vmovdqu %0, %%ymm3" :: "m" ((ctx)->avx[3])); \
98 __asm("vmovdqu %%ymm0, %0" : "=m" ((ctx)->avx[0])); \
99 __asm("vmovdqu %%ymm1, %0" : "=m" ((ctx)->avx[1])); \
100 __asm("vmovdqu %%ymm2, %0" : "=m" ((ctx)->avx[2])); \
101 __asm("vmovdqu %%ymm3, %0" : "=m" ((ctx)->avx[3])); \
113 __asm("vpmovzxdq %0, %%ymm4"::"m" (*ip)); in fletcher_4_avx2_native()
114 __asm("vpaddq %ymm4, %ymm0, %ymm0"); in fletcher_4_avx2_native()
115 __asm("vpaddq %ymm0, %ymm1, %ymm1"); in fletcher_4_avx2_native()
116 __asm("vpaddq %ymm1, %ymm2, %ymm2"); in fletcher_4_avx2_native()
117 __asm("vpaddq %ymm2, %ymm3, %ymm3"); in fletcher_4_avx2_native()
121 __asm("vzeroupper"); in fletcher_4_avx2_native()
138 __asm("vmovdqu %0, %%ymm5" :: "m" (mask)); in fletcher_4_avx2_byteswap()
141 __asm("vpmovzxdq %0, %%ymm4"::"m" (*ip)); in fletcher_4_avx2_byteswap()
142 __asm("vpshufb %ymm5, %ymm4, %ymm4"); in fletcher_4_avx2_byteswap()
144 __asm("vpaddq %ymm4, %ymm0, %ymm0"); in fletcher_4_avx2_byteswap()
145 __asm("vpaddq %ymm0, %ymm1, %ymm1"); in fletcher_4_avx2_byteswap()
146 __asm("vpaddq %ymm1, %ymm2, %ymm2"); in fletcher_4_avx2_byteswap()
147 __asm("vpaddq %ymm2, %ymm3, %ymm3"); in fletcher_4_avx2_byteswap()
151 __asm("vzeroupper"); in fletcher_4_avx2_byteswap()