1 /*===---- pmmintrin.h - Implementation of SSE3 intrinsics on PowerPC -------=== 2 * 3 * Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 * See https://llvm.org/LICENSE.txt for license information. 5 * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 * 7 *===-----------------------------------------------------------------------=== 8 */ 9 10 /* Implemented from the specification included in the Intel C++ Compiler 11 User Guide and Reference, version 9.0. */ 12 13 #ifndef NO_WARN_X86_INTRINSICS 14 /* This header is distributed to simplify porting x86_64 code that 15 makes explicit use of Intel intrinsics to powerpc64le. 16 It is the user's responsibility to determine if the results are 17 acceptable and make additional changes as necessary. 18 Note that much code that uses Intel intrinsics can be rewritten in 19 standard C or GNU C extensions, which are more portable and better 20 optimized across multiple targets. 21 22 In the specific case of X86 SSE3 intrinsics, the PowerPC VMX/VSX ISA 23 is a good match for most SIMD operations. However the Horizontal 24 add/sub requires the data pairs be permuted into a separate 25 registers with vertical even/odd alignment for the operation. 26 And the addsub operation requires the sign of only the even numbered 27 elements be flipped (xored with -0.0). 28 For larger blocks of code using these intrinsic implementations, 29 the compiler be should be able to schedule instructions to avoid 30 additional latency. 31 32 In the specific case of the monitor and mwait instructions there are 33 no direct equivalent in the PowerISA at this time. So those 34 intrinsics are not implemented. */ 35 #error \ 36 "Please read comment above. Use -DNO_WARN_X86_INTRINSICS to disable this warning." 37 #endif 38 39 #ifndef PMMINTRIN_H_ 40 #define PMMINTRIN_H_ 41 42 #if defined(__ppc64__) && \ 43 (defined(__linux__) || defined(__FreeBSD__) || defined(_AIX)) 44 45 /* We need definitions from the SSE2 and SSE header files*/ 46 #include <emmintrin.h> 47 48 extern __inline __m128 49 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 50 _mm_addsub_ps(__m128 __X, __m128 __Y) { 51 const __v4sf __even_n0 = {-0.0, 0.0, -0.0, 0.0}; 52 __v4sf __even_neg_Y = vec_xor(__Y, __even_n0); 53 return (__m128)vec_add(__X, __even_neg_Y); 54 } 55 56 extern __inline __m128d 57 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 58 _mm_addsub_pd(__m128d __X, __m128d __Y) { 59 const __v2df __even_n0 = {-0.0, 0.0}; 60 __v2df __even_neg_Y = vec_xor(__Y, __even_n0); 61 return (__m128d)vec_add(__X, __even_neg_Y); 62 } 63 64 extern __inline __m128 65 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 66 _mm_hadd_ps(__m128 __X, __m128 __Y) { 67 __vector unsigned char __xform2 = {0x00, 0x01, 0x02, 0x03, 0x08, 0x09, 68 0x0A, 0x0B, 0x10, 0x11, 0x12, 0x13, 69 0x18, 0x19, 0x1A, 0x1B}; 70 __vector unsigned char __xform1 = {0x04, 0x05, 0x06, 0x07, 0x0C, 0x0D, 71 0x0E, 0x0F, 0x14, 0x15, 0x16, 0x17, 72 0x1C, 0x1D, 0x1E, 0x1F}; 73 return (__m128)vec_add(vec_perm((__v4sf)__X, (__v4sf)__Y, __xform2), 74 vec_perm((__v4sf)__X, (__v4sf)__Y, __xform1)); 75 } 76 77 extern __inline __m128 78 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 79 _mm_hsub_ps(__m128 __X, __m128 __Y) { 80 __vector unsigned char __xform2 = {0x00, 0x01, 0x02, 0x03, 0x08, 0x09, 81 0x0A, 0x0B, 0x10, 0x11, 0x12, 0x13, 82 0x18, 0x19, 0x1A, 0x1B}; 83 __vector unsigned char __xform1 = {0x04, 0x05, 0x06, 0x07, 0x0C, 0x0D, 84 0x0E, 0x0F, 0x14, 0x15, 0x16, 0x17, 85 0x1C, 0x1D, 0x1E, 0x1F}; 86 return (__m128)vec_sub(vec_perm((__v4sf)__X, (__v4sf)__Y, __xform2), 87 vec_perm((__v4sf)__X, (__v4sf)__Y, __xform1)); 88 } 89 90 extern __inline __m128d 91 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 92 _mm_hadd_pd(__m128d __X, __m128d __Y) { 93 return (__m128d)vec_add(vec_mergeh((__v2df)__X, (__v2df)__Y), 94 vec_mergel((__v2df)__X, (__v2df)__Y)); 95 } 96 97 extern __inline __m128d 98 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 99 _mm_hsub_pd(__m128d __X, __m128d __Y) { 100 return (__m128d)vec_sub(vec_mergeh((__v2df)__X, (__v2df)__Y), 101 vec_mergel((__v2df)__X, (__v2df)__Y)); 102 } 103 104 #ifdef _ARCH_PWR8 105 extern __inline __m128 106 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 107 _mm_movehdup_ps(__m128 __X) { 108 return (__m128)vec_mergeo((__v4su)__X, (__v4su)__X); 109 } 110 #endif 111 112 #ifdef _ARCH_PWR8 113 extern __inline __m128 114 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 115 _mm_moveldup_ps(__m128 __X) { 116 return (__m128)vec_mergee((__v4su)__X, (__v4su)__X); 117 } 118 #endif 119 120 extern __inline __m128d 121 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 122 _mm_loaddup_pd(double const *__P) { 123 return (__m128d)vec_splats(*__P); 124 } 125 126 extern __inline __m128d 127 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 128 _mm_movedup_pd(__m128d __X) { 129 return _mm_shuffle_pd(__X, __X, _MM_SHUFFLE2(0, 0)); 130 } 131 132 extern __inline __m128i 133 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 134 _mm_lddqu_si128(__m128i const *__P) { 135 return (__m128i)(vec_vsx_ld(0, (signed int const *)__P)); 136 } 137 138 /* POWER8 / POWER9 have no equivalent for _mm_monitor nor _mm_wait. */ 139 140 #else 141 #include_next <pmmintrin.h> 142 #endif /* defined(__ppc64__) && 143 * (defined(__linux__) || defined(__FreeBSD__) || defined(_AIX)) */ 144 145 #endif /* PMMINTRIN_H_ */ 146