/freebsd/contrib/llvm-project/clang/lib/Headers/ |
H A D | xmmintrin.h | 20 typedef float __v4sf __attribute__((__vector_size__(16))); typedef 80 return (__m128)((__v4sf)__a + (__v4sf)__b); in _mm_add_ps() 123 return (__m128)((__v4sf)__a - (__v4sf)__b); in _mm_sub_ps() 165 return (__m128)((__v4sf)__a * (__v4sf)__b); in _mm_mul_ps() 206 return (__m128)((__v4sf)__a / (__v4sf)__b); in _mm_div_ps() 224 return (__m128)__builtin_ia32_sqrtss((__v4sf)__a); in _mm_sqrt_ss() 241 return __builtin_ia32_sqrtps((__v4sf)__a); in _mm_sqrt_ps() 259 return (__m128)__builtin_ia32_rcpss((__v4sf)__a); in _mm_rcp_ss() 276 return (__m128)__builtin_ia32_rcpps((__v4sf)__a); in _mm_rcp_ps() 295 return __builtin_ia32_rsqrtss((__v4sf)__a); in _mm_rsqrt_ss() [all …]
|
H A D | fma4intrin.h | 26 return (__m128)__builtin_ia32_vfmaddps((__v4sf)__A, (__v4sf)__B, (__v4sf)__C); in _mm_macc_ps() 38 return (__m128)__builtin_ia32_vfmaddss((__v4sf)__A, (__v4sf)__B, (__v4sf)__C); in _mm_macc_ss() 50 return (__m128)__builtin_ia32_vfmaddps((__v4sf)__A, (__v4sf)__B, -(__v4sf)__C); in _mm_msub_ps() 62 return (__m128)__builtin_ia32_vfmaddss((__v4sf)__A, (__v4sf)__B, -(__v4sf)__C); in _mm_msub_ss() 74 return (__m128)__builtin_ia32_vfmaddps(-(__v4sf)__A, (__v4sf)__B, (__v4sf)__C); in _mm_nmacc_ps() 86 return (__m128)__builtin_ia32_vfmaddss(-(__v4sf)__A, (__v4sf)__B, (__v4sf)__C); in _mm_nmacc_ss() 98 return (__m128)__builtin_ia32_vfmaddps(-(__v4sf)__A, (__v4sf)__B, -(__v4sf)__C); in _mm_nmsub_ps() 110 return (__m128)__builtin_ia32_vfmaddss(-(__v4sf)__A, (__v4sf)__B, -(__v4sf)__C); in _mm_nmsub_ss() 122 return (__m128)__builtin_ia32_vfmaddsubps((__v4sf)__A, (__v4sf)__B, (__v4sf)__C); in _mm_maddsub_ps() 134 return (__m128)__builtin_ia32_vfmaddsubps((__v4sf)__A, (__v4sf)__B, -(__v4sf)__C); in _mm_msubadd_ps()
|
H A D | fmaintrin.h | 38 return (__m128)__builtin_ia32_vfmaddps((__v4sf)__A, (__v4sf)__B, (__v4sf)__C); in _mm_fmadd_ps() 87 return (__m128)__builtin_ia32_vfmaddss3((__v4sf)__A, (__v4sf)__B, (__v4sf)__C); in _mm_fmadd_ss() 136 return (__m128)__builtin_ia32_vfmaddps((__v4sf)__A, (__v4sf)__B, -(__v4sf)__C); in _mm_fmsub_ps() 185 return (__m128)__builtin_ia32_vfmaddss3((__v4sf)__A, (__v4sf)__B, -(__v4sf)__C); in _mm_fmsub_ss() 234 return (__m128)__builtin_ia32_vfmaddps(-(__v4sf)__A, (__v4sf)__B, (__v4sf)__C); in _mm_fnmadd_ps() 283 return (__m128)__builtin_ia32_vfmaddss3((__v4sf)__A, -(__v4sf)__B, (__v4sf)__C); in _mm_fnmadd_ss() 332 return (__m128)__builtin_ia32_vfmaddps(-(__v4sf)__A, (__v4sf)__B, -(__v4sf)__C); in _mm_fnmsub_ps() 381 return (__m128)__builtin_ia32_vfmaddss3((__v4sf)__A, -(__v4sf)__B, -(__v4sf)__C); in _mm_fnmsub_ss() 437 return (__m128)__builtin_ia32_vfmaddsubps((__v4sf)__A, (__v4sf)__B, (__v4sf)__C); in _mm_fmaddsub_ps() 489 return (__m128)__builtin_ia32_vfmaddsubps((__v4sf)__A, (__v4sf)__B, -(__v4sf)__C); in _mm_fmsubadd_ps()
|
H A D | avx512fp16intrin.h | 1523 ((__m128)__builtin_ia32_vcvtsh2ss_round_mask((__v4sf)(A), (__v8hf)(B), \ 1524 (__v4sf)_mm_undefined_ps(), \ 1529 (__v4sf)(A), (__v8hf)(B), (__v4sf)(W), (__mmask8)(U), (int)(R))) 1532 ((__m128)__builtin_ia32_vcvtsh2ss_round_mask((__v4sf)(A), (__v8hf)(B), \ 1533 (__v4sf)_mm_setzero_ps(), \ 1539 (__v4sf)__A, (__v8hf)__B, (__v4sf)_mm_undefined_ps(), (__mmask8)-1, in _mm_cvtsh_ss() 1547 return (__m128)__builtin_ia32_vcvtsh2ss_round_mask((__v4sf)__A, (__v8hf)__B, in _mm_mask_cvtsh_ss() 1548 (__v4sf)__W, (__mmask8)__U, in _mm_mask_cvtsh_ss() 1556 (__v4sf)__A, (__v8hf)__B, (__v4sf)_mm_setzero_ps(), (__mmask8)__U, in _mm_maskz_cvtsh_ss() 1561 ((__m128h)__builtin_ia32_vcvtss2sh_round_mask((__v8hf)(A), (__v4sf)(B), \ [all …]
|
H A D | smmintrin.h | 240 ((__m128)__builtin_ia32_roundps((__v4sf)(__m128)(X), (M))) 281 ((__m128)__builtin_ia32_roundss((__v4sf)(__m128)(X), (__v4sf)(__m128)(Y), \ 413 ((__m128)__builtin_ia32_blendps((__v4sf)(__m128)(V1), (__v4sf)(__m128)(V2), \ 466 return (__m128)__builtin_ia32_blendvps((__v4sf)__V1, (__v4sf)__V2, in _mm_blendv_ps() 467 (__v4sf)__M); in _mm_blendv_ps() 597 ((__m128)__builtin_ia32_dpps((__v4sf)(__m128)(X), (__v4sf)(__m128)(Y), (M))) 864 int, __builtin_ia32_vec_ext_v4sf((__v4sf)(__m128)(X), (int)(N))) 870 (D) = __builtin_ia32_vec_ext_v4sf((__v4sf)(__m128)(X), (int)(N)); \
|
H A D | f16cintrin.h | 42 __v4sf __r = __builtin_ia32_vcvtph2ps(__v); in _cvtsh_ss() 69 (unsigned short)(((__v8hi)__builtin_ia32_vcvtps2ph((__v4sf){a, 0, 0, 0}, \ 96 ((__m128i)__builtin_ia32_vcvtps2ph((__v4sf)(__m128)(a), (imm)))
|
H A D | avxintrin.h | 893 return (__m128)__builtin_ia32_vpermilvarps((__v4sf)__a, (__v4si)__c); in _mm_permutevar_ps() 1111 ((__m128)__builtin_ia32_vpermilps((__v4sf)(__m128)(A), (int)(C))) 2245 return (__m256d)__builtin_convertvector((__v4sf)__a, __v4df); in _mm256_cvtps_pd() 2645 return __builtin_ia32_vtestzps((__v4sf)__a, (__v4sf)__b); in _mm_testz_ps() 2674 return __builtin_ia32_vtestcps((__v4sf)__a, (__v4sf)__b); in _mm_testc_ps() 2704 return __builtin_ia32_vtestnzcps((__v4sf)__a, (__v4sf)__b); in _mm_testnzc_ps() 3125 return (__m256)__builtin_shufflevector((__v4sf)__b, (__v4sf)__b, in _mm256_broadcast_ps() 3445 return (__m128)__builtin_ia32_maskloadps((const __v4sf *)__p, (__v4si)__m); in _mm_maskload_ps() 3566 __builtin_ia32_maskstoreps((__v4sf *)__p, (__v4si)__m, (__v4sf)__a); in _mm_maskstore_ps() 4551 return __builtin_shufflevector((__v4sf)__a, in _mm256_castps128_ps256() [all …]
|
H A D | avxneconvertintrin.h | 449 return (__m128bh)__builtin_ia32_vcvtneps2bf16128((__v4sf)__A); in _mm_cvtneps_avx_pbh()
|
H A D | xopintrin.h | 723 ((__m128)__builtin_ia32_vpermil2ps((__v4sf)(__m128)(X), (__v4sf)(__m128)(Y), \ 734 return (__m128)__builtin_ia32_vfrczss((__v4sf)__A); in _mm_frcz_ss() 746 return (__m128)__builtin_ia32_vfrczps((__v4sf)__A); in _mm_frcz_ps()
|
H A D | emmintrin.h | 1296 __builtin_shufflevector((__v4sf)__a, (__v4sf)__a, 0, 1), __v2df); in _mm_cvtps_pd() 1381 return (__m128)__builtin_ia32_cvtsd2ss((__v4sf)__a, (__v2df)__b); in _mm_cvtsd_ss() 3314 return (__m128) __builtin_convertvector((__v4si)__a, __v4sf); in _mm_cvtepi32_ps() 3332 return (__m128i)__builtin_ia32_cvtps2dq((__v4sf)__a); in _mm_cvtps_epi32() 3351 return (__m128i)__builtin_ia32_cvttps2dq((__v4sf)__a); in _mm_cvttps_epi32()
|
/freebsd/contrib/llvm-project/clang/lib/Headers/ppc_wrappers/ |
H A D | xmmintrin.h | 73 typedef vector float __v4sf; typedef 94 return ((__m128)vec_ld(0, (__v4sf *)__P)); in _mm_load_ps() 108 __v4sf __tmp; in _mm_loadr_ps() 114 __tmp = vec_ld(0, (__v4sf *)__P); in _mm_loadr_ps() 123 return __extension__(__m128)(__v4sf){__F, __F, __F, __F}; in _mm_set1_ps() 136 return __extension__(__m128)(__v4sf){__W, __X, __Y, __Z}; in _mm_set_ps() 143 return __extension__(__m128)(__v4sf){__Z, __Y, __X, __W}; in _mm_setr_ps() 150 vec_st((__v4sf)__A, 0, (__v4sf *)__P); in _mm_store_ps() 164 __v4sf __tmp; in _mm_storer_ps() 178 __v4sf __va = vec_splat((__v4sf)__A, 0); in _mm_store1_ps() [all …]
|
H A D | pmmintrin.h | 51 const __v4sf __even_n0 = {-0.0, 0.0, -0.0, 0.0}; in _mm_addsub_ps() 52 __v4sf __even_neg_Y = vec_xor(__Y, __even_n0); in _mm_addsub_ps() 73 return (__m128)vec_add(vec_perm((__v4sf)__X, (__v4sf)__Y, __xform2), in _mm_hadd_ps() 74 vec_perm((__v4sf)__X, (__v4sf)__Y, __xform1)); in _mm_hadd_ps() 86 return (__m128)vec_sub(vec_perm((__v4sf)__X, (__v4sf)__Y, __xform2), in _mm_hsub_ps() 87 vec_perm((__v4sf)__X, (__v4sf)__Y, __xform1)); in _mm_hsub_ps()
|
H A D | emmintrin.h | 881 __v4sf __result; in _mm_cvtpd_ps() 893 __result = (__v4sf)vec_vpkudum((__vector long long)__temp, in _mm_cvtpd_ps() 899 __result = (__v4sf)vec_perm((__v16qu)__temp, (__v16qu)__vzero, __pkperm); in _mm_cvtpd_ps() 967 __v4sf __rounded; in _mm_cvtps_epi32() 970 __rounded = vec_rint((__v4sf)__A); in _mm_cvtps_epi32() 980 __result = vec_cts((__v4sf)__A, 0); in _mm_cvttps_epi32() 989 return (__m128d)vec_doubleh((__v4sf)__A); in _mm_cvtps_pd() 993 __v4sf __a = (__v4sf)__A; in _mm_cvtps_pd() 994 __v4sf __temp; in _mm_cvtps_pd() 1067 __v4sf __result = (__v4sf)__A; in _mm_cvtsd_ss() [all …]
|