|
66 | 66 | (__v2df)_mm_setzero_pd(), (__mmask8)(U))) |
67 | 67 |
|
68 | 68 | #define _mm256_minmax_pd(A, B, C) \ |
69 | | - ((__m256d)__builtin_ia32_vminmaxpd256_round_mask( \ |
| 69 | + ((__m256d)__builtin_ia32_vminmaxpd256_mask( \ |
70 | 70 | (__v4df)(__m256d)(A), (__v4df)(__m256d)(B), (int)(C), \ |
71 | | - (__v4df)_mm256_setzero_pd(), (__mmask8)-1, _MM_FROUND_NO_EXC)) |
| 71 | + (__v4df)_mm256_setzero_pd(), (__mmask8)-1)) |
72 | 72 |
|
73 | 73 | #define _mm256_mask_minmax_pd(W, U, A, B, C) \ |
74 | | - ((__m256d)__builtin_ia32_vminmaxpd256_round_mask( \ |
| 74 | + ((__m256d)__builtin_ia32_vminmaxpd256_mask( \ |
75 | 75 | (__v4df)(__m256d)(A), (__v4df)(__m256d)(B), (int)(C), \ |
76 | | - (__v4df)(__m256d)(W), (__mmask8)(U), _MM_FROUND_NO_EXC)) |
| 76 | + (__v4df)(__m256d)(W), (__mmask8)(U))) |
77 | 77 |
|
78 | 78 | #define _mm256_maskz_minmax_pd(U, A, B, C) \ |
79 | | - ((__m256d)__builtin_ia32_vminmaxpd256_round_mask( \ |
| 79 | + ((__m256d)__builtin_ia32_vminmaxpd256_mask( \ |
80 | 80 | (__v4df)(__m256d)(A), (__v4df)(__m256d)(B), (int)(C), \ |
81 | | - (__v4df)_mm256_setzero_pd(), (__mmask8)(U), _MM_FROUND_NO_EXC)) |
82 | | - |
83 | | -#define _mm256_minmax_round_pd(A, B, C, R) \ |
84 | | - ((__m256d)__builtin_ia32_vminmaxpd256_round_mask( \ |
85 | | - (__v4df)(__m256d)(A), (__v4df)(__m256d)(B), (int)(C), \ |
86 | | - (__v4df)_mm256_undefined_pd(), (__mmask8)-1, (int)(R))) |
87 | | - |
88 | | -#define _mm256_mask_minmax_round_pd(W, U, A, B, C, R) \ |
89 | | - ((__m256d)__builtin_ia32_vminmaxpd256_round_mask( \ |
90 | | - (__v4df)(__m256d)(A), (__v4df)(__m256d)(B), (int)(C), \ |
91 | | - (__v4df)(__m256d)(W), (__mmask8)(U), (int)(R))) |
92 | | - |
93 | | -#define _mm256_maskz_minmax_round_pd(U, A, B, C, R) \ |
94 | | - ((__m256d)__builtin_ia32_vminmaxpd256_round_mask( \ |
95 | | - (__v4df)(__m256d)(A), (__v4df)(__m256d)(B), (int)(C), \ |
96 | | - (__v4df)_mm256_setzero_pd(), (__mmask8)(U), (int)(R))) |
| 81 | + (__v4df)_mm256_setzero_pd(), (__mmask8)(U))) |
97 | 82 |
|
98 | 83 | #define _mm_minmax_ph(A, B, C) \ |
99 | 84 | ((__m128h)__builtin_ia32_vminmaxph128_mask( \ |
|
111 | 96 | (__v8hf)_mm_setzero_ph(), (__mmask8)(U))) |
112 | 97 |
|
113 | 98 | #define _mm256_minmax_ph(A, B, C) \ |
114 | | - ((__m256h)__builtin_ia32_vminmaxph256_round_mask( \ |
| 99 | + ((__m256h)__builtin_ia32_vminmaxph256_mask( \ |
115 | 100 | (__v16hf)(__m256h)(A), (__v16hf)(__m256h)(B), (int)(C), \ |
116 | | - (__v16hf)_mm256_setzero_ph(), (__mmask16)-1, _MM_FROUND_NO_EXC)) |
| 101 | + (__v16hf)_mm256_setzero_ph(), (__mmask16)-1)) |
117 | 102 |
|
118 | 103 | #define _mm256_mask_minmax_ph(W, U, A, B, C) \ |
119 | | - ((__m256h)__builtin_ia32_vminmaxph256_round_mask( \ |
| 104 | + ((__m256h)__builtin_ia32_vminmaxph256_mask( \ |
120 | 105 | (__v16hf)(__m256h)(A), (__v16hf)(__m256h)(B), (int)(C), \ |
121 | | - (__v16hf)(__m256h)(W), (__mmask16)(U), _MM_FROUND_NO_EXC)) |
| 106 | + (__v16hf)(__m256h)(W), (__mmask16)(U))) |
122 | 107 |
|
123 | 108 | #define _mm256_maskz_minmax_ph(U, A, B, C) \ |
124 | | - ((__m256h)__builtin_ia32_vminmaxph256_round_mask( \ |
125 | | - (__v16hf)(__m256h)(A), (__v16hf)(__m256h)(B), (int)(C), \ |
126 | | - (__v16hf)_mm256_setzero_ph(), (__mmask16)(U), _MM_FROUND_NO_EXC)) |
127 | | - |
128 | | -#define _mm256_minmax_round_ph(A, B, C, R) \ |
129 | | - ((__m256h)__builtin_ia32_vminmaxph256_round_mask( \ |
| 109 | + ((__m256h)__builtin_ia32_vminmaxph256_mask( \ |
130 | 110 | (__v16hf)(__m256h)(A), (__v16hf)(__m256h)(B), (int)(C), \ |
131 | | - (__v16hf)_mm256_undefined_ph(), (__mmask16)-1, (int)(R))) |
132 | | - |
133 | | -#define _mm256_mask_minmax_round_ph(W, U, A, B, C, R) \ |
134 | | - ((__m256h)__builtin_ia32_vminmaxph256_round_mask( \ |
135 | | - (__v16hf)(__m256h)(A), (__v16hf)(__m256h)(B), (C), \ |
136 | | - (__v16hf)(__m256h)(W), (__mmask16)(U), (int)(R))) |
137 | | - |
138 | | -#define _mm256_maskz_minmax_round_ph(U, A, B, C, R) \ |
139 | | - ((__m256h)__builtin_ia32_vminmaxph256_round_mask( \ |
140 | | - (__v16hf)(__m256h)(A), (__v16hf)(__m256h)(B), (int)(C), \ |
141 | | - (__v16hf)_mm256_setzero_ph(), (__mmask16)(U), (int)(R))) |
| 111 | + (__v16hf)_mm256_setzero_ph(), (__mmask16)(U))) |
142 | 112 |
|
143 | 113 | #define _mm_minmax_ps(A, B, C) \ |
144 | 114 | ((__m128)__builtin_ia32_vminmaxps128_mask( \ |
|
156 | 126 | (__v4sf)_mm_setzero_ps(), (__mmask8)(U))) |
157 | 127 |
|
158 | 128 | #define _mm256_minmax_ps(A, B, C) \ |
159 | | - ((__m256)__builtin_ia32_vminmaxps256_round_mask( \ |
| 129 | + ((__m256)__builtin_ia32_vminmaxps256_mask( \ |
160 | 130 | (__v8sf)(__m256)(A), (__v8sf)(__m256)(B), (int)(C), \ |
161 | | - (__v8sf)_mm256_setzero_ps(), (__mmask8)-1, _MM_FROUND_NO_EXC)) |
| 131 | + (__v8sf)_mm256_setzero_ps(), (__mmask8)-1)) |
162 | 132 |
|
163 | 133 | #define _mm256_mask_minmax_ps(W, U, A, B, C) \ |
164 | | - ((__m256)__builtin_ia32_vminmaxps256_round_mask( \ |
| 134 | + ((__m256)__builtin_ia32_vminmaxps256_mask( \ |
165 | 135 | (__v8sf)(__m256)(A), (__v8sf)(__m256)(B), (int)(C), (__v8sf)(__m256)(W), \ |
166 | | - (__mmask8)(U), _MM_FROUND_NO_EXC)) |
| 136 | + (__mmask8)(U))) |
167 | 137 |
|
168 | 138 | #define _mm256_maskz_minmax_ps(U, A, B, C) \ |
169 | | - ((__m256)__builtin_ia32_vminmaxps256_round_mask( \ |
170 | | - (__v8sf)(__m256)(A), (__v8sf)(__m256)(B), (int)(C), \ |
171 | | - (__v8sf)_mm256_setzero_ps(), (__mmask8)(U), _MM_FROUND_NO_EXC)) |
172 | | - |
173 | | -#define _mm256_minmax_round_ps(A, B, C, R) \ |
174 | | - ((__m256)__builtin_ia32_vminmaxps256_round_mask( \ |
175 | | - (__v8sf)(__m256)(A), (__v8sf)(__m256)(B), (int)(C), \ |
176 | | - (__v8sf)_mm256_undefined_ps(), (__mmask8)-1, (int)(R))) |
177 | | - |
178 | | -#define _mm256_mask_minmax_round_ps(W, U, A, B, C, R) \ |
179 | | - ((__m256)__builtin_ia32_vminmaxps256_round_mask( \ |
180 | | - (__v8sf)(__m256)(A), (__v8sf)(__m256)(B), (int)(C), (__v8sf)(__m256)(W), \ |
181 | | - (__mmask8)(U), (int)(R))) |
182 | | - |
183 | | -#define _mm256_maskz_minmax_round_ps(U, A, B, C, R) \ |
184 | | - ((__m256)__builtin_ia32_vminmaxps256_round_mask( \ |
| 139 | + ((__m256)__builtin_ia32_vminmaxps256_mask( \ |
185 | 140 | (__v8sf)(__m256)(A), (__v8sf)(__m256)(B), (int)(C), \ |
186 | | - (__v8sf)_mm256_setzero_ps(), (__mmask8)(U), (int)(R))) |
| 141 | + (__v8sf)_mm256_setzero_ps(), (__mmask8)(U))) |
187 | 142 |
|
188 | 143 | #define _mm_minmax_sd(A, B, C) \ |
189 | 144 | ((__m128d)__builtin_ia32_vminmaxsd_round_mask( \ |
|
0 commit comments