Lines Matching refs:__v4df

17 typedef double __v4df __attribute__ ((__vector_size__ (32)));  typedef
76 return (__m256d)((__v4df)__a+(__v4df)__b); in _mm256_add_pd()
112 return (__m256d)((__v4df)__a-(__v4df)__b); in _mm256_sub_pd()
149 return (__m256d)__builtin_ia32_addsubpd256((__v4df)__a, (__v4df)__b); in _mm256_addsub_pd()
186 return (__m256d)((__v4df)__a/(__v4df)__b); in _mm256_div_pd()
225 return (__m256d)__builtin_ia32_maxpd256((__v4df)__a, (__v4df)__b); in _mm256_max_pd()
267 return (__m256d)__builtin_ia32_minpd256((__v4df)__a, (__v4df)__b); in _mm256_min_pd()
306 return (__m256d)((__v4df)__a * (__v4df)__b); in _mm256_mul_pd()
341 return (__m256d)__builtin_ia32_sqrtpd256((__v4df)__a); in _mm256_sqrt_pd()
425 ((__m256d)__builtin_ia32_roundpd256((__v4df)(__m256d)(V), (M)))
700 return (__m256d)__builtin_ia32_haddpd256((__v4df)__a, (__v4df)__b); in _mm256_hadd_pd()
746 return (__m256d)__builtin_ia32_hsubpd256((__v4df)__a, (__v4df)__b); in _mm256_hsub_pd()
838 return (__m256d)__builtin_ia32_vpermilvarpd256((__v4df)__a, (__v4di)__c); in _mm256_permutevar_pd()
1055 ((__m256d)__builtin_ia32_vpermilpd256((__v4df)(__m256d)(A), (int)(C)))
1243 ((__m256d)__builtin_ia32_vperm2f128_pd256((__v4df)(__m256d)(V1), \
1244 (__v4df)(__m256d)(V2), (int)(M)))
1353 ((__m256d)__builtin_ia32_blendpd256((__v4df)(__m256d)(V1), \
1354 (__v4df)(__m256d)(V2), (int)(M)))
1409 (__v4df)__a, (__v4df)__b, (__v4df)__c); in _mm256_blendv_pd()
1582 ((__m256d)__builtin_ia32_shufpd256((__v4df)(__m256d)(a), \
1583 (__v4df)(__m256d)(b), (int)(mask)))
1786 ((__m256d)__builtin_ia32_cmppd256((__v4df)(__m256d)(a), \
1787 (__v4df)(__m256d)(b), (c)))
2179 return (__m256d)__builtin_convertvector((__v4si)__a, __v4df); in _mm256_cvtepi32_pd()
2210 return (__m128)__builtin_ia32_cvtpd2ps256((__v4df) __a); in _mm256_cvtpd_ps()
2245 return (__m256d)__builtin_convertvector((__v4sf)__a, __v4df); in _mm256_cvtps_pd()
2266 return (__m128i)__builtin_ia32_cvttpd2dq256((__v4df) __a); in _mm256_cvttpd_epi32()
2286 return (__m128i)__builtin_ia32_cvtpd2dq256((__v4df) __a); in _mm256_cvtpd_epi32()
2428 return __builtin_shufflevector((__v4df)__a, (__v4df)__a, 0, 0, 2, 2); in _mm256_movedup_pd()
2451 return __builtin_shufflevector((__v4df)__a, (__v4df)__b, 1, 5, 1+2, 5+2); in _mm256_unpackhi_pd()
2473 return __builtin_shufflevector((__v4df)__a, (__v4df)__b, 0, 4, 0+2, 4+2); in _mm256_unpacklo_pd()
2733 return __builtin_ia32_vtestzpd256((__v4df)__a, (__v4df)__b); in _mm256_testz_pd()
2762 return __builtin_ia32_vtestcpd256((__v4df)__a, (__v4df)__b); in _mm256_testc_pd()
2792 return __builtin_ia32_vtestnzcpd256((__v4df)__a, (__v4df)__b); in _mm256_testnzc_pd()
2978 return __builtin_ia32_movmskpd256((__v4df)__a); in _mm256_movemask_pd()
3064 return __extension__ (__m256d)(__v4df){ __d, __d, __d, __d }; in _mm256_broadcast_sd()
3420 return (__m256d)__builtin_ia32_maskloadpd256((const __v4df *)__p, in _mm256_maskload_pd()
3542 __builtin_ia32_maskstorepd256((__v4df *)__p, (__v4di)__m, (__v4df)__a); in _mm256_maskstore_pd()
3606 typedef __v4df __v4df_aligned __attribute__((aligned(32))); in _mm256_stream_pd()
4475 return __builtin_shufflevector((__v4df)__a, (__v4df)__a, 0, 1); in _mm256_castpd256_pd128()
4709 ((__m256d)__builtin_ia32_vinsertf128_pd256((__v4df)(__m256d)(V1), \
4801 ((__m128d)__builtin_ia32_vextractf128_pd256((__v4df)(__m256d)(V), (int)(M)))