|
1 |
| -// RUN: %clang_cc1 -ffreestanding %s -O0 -triple=x86_64-apple-darwin -target-cpu skylake-avx512 -emit-llvm -o - -Wall -Werror | FileCheck %s |
| 1 | +// RUN: %clang_cc1 -x c -ffreestanding %s -O0 -triple=x86_64-apple-darwin -target-cpu skylake-avx512 -emit-llvm -o - -Wall -Werror | FileCheck %s |
| 2 | +// RUN: %clang_cc1 -x c -ffreestanding %s -O0 -triple=i386-apple-darwin -target-cpu skylake-avx512 -emit-llvm -o - -Wall -Werror | FileCheck %s |
| 3 | +// RUN: %clang_cc1 -x c++ -ffreestanding %s -O0 -triple=x86_64-apple-darwin -target-cpu skylake-avx512 -emit-llvm -o - -Wall -Werror | FileCheck %s |
| 4 | +// RUN: %clang_cc1 -x c++ -ffreestanding %s -O0 -triple=i386-apple-darwin -target-cpu skylake-avx512 -emit-llvm -o - -Wall -Werror | FileCheck %s |
2 | 5 |
|
3 | 6 | #include <immintrin.h>
|
4 | 7 | #include "builtin_test_helpers.h"
|
5 | 8 |
|
6 | 9 | long long test_mm512_reduce_add_epi64(__m512i __W){
|
7 |
| -// CHECK-LABEL: @test_mm512_reduce_add_epi64( |
8 |
| -// CHECK: call i64 @llvm.vector.reduce.add.v8i64(<8 x i64> %{{.*}}) |
| 10 | +// CHECK-LABEL: test_mm512_reduce_add_epi64 |
| 11 | +// CHECK: call {{.*}}i64 @llvm.vector.reduce.add.v8i64(<8 x i64> %{{.*}}) |
9 | 12 | return _mm512_reduce_add_epi64(__W);
|
10 | 13 | }
|
11 | 14 | TEST_CONSTEXPR(_mm512_reduce_add_epi64((__m512i)(__v8di){-4, -3, -2, -1, 0, 1, 2, 3}) == -4);
|
12 | 15 |
|
13 | 16 | long long test_mm512_reduce_mul_epi64(__m512i __W){
|
14 |
| -// CHECK-LABEL: @test_mm512_reduce_mul_epi64( |
15 |
| -// CHECK: call i64 @llvm.vector.reduce.mul.v8i64(<8 x i64> %{{.*}}) |
| 17 | +// CHECK-LABEL: test_mm512_reduce_mul_epi64 |
| 18 | +// CHECK: call {{.*}}i64 @llvm.vector.reduce.mul.v8i64(<8 x i64> %{{.*}}) |
16 | 19 | return _mm512_reduce_mul_epi64(__W);
|
17 | 20 | }
|
18 | 21 | TEST_CONSTEXPR(_mm512_reduce_mul_epi64((__m512i)(__v8di){1, 2, 3, 4, 5, 6, 7, 8}) == 40320);
|
19 | 22 |
|
20 | 23 | long long test_mm512_reduce_or_epi64(__m512i __W){
|
21 |
| -// CHECK-LABEL: @test_mm512_reduce_or_epi64( |
22 |
| -// CHECK: call i64 @llvm.vector.reduce.or.v8i64(<8 x i64> %{{.*}}) |
| 24 | +// CHECK-LABEL: test_mm512_reduce_or_epi64 |
| 25 | +// CHECK: call {{.*}}i64 @llvm.vector.reduce.or.v8i64(<8 x i64> %{{.*}}) |
23 | 26 | return _mm512_reduce_or_epi64(__W);
|
24 | 27 | }
|
25 | 28 | TEST_CONSTEXPR(_mm512_reduce_or_epi64((__m512i)(__v8di){0x100, 0x200, 0x400, 0x800, 0, 0, 0, 0}) == 0xF00);
|
26 | 29 |
|
27 | 30 | long long test_mm512_reduce_and_epi64(__m512i __W){
|
28 |
| -// CHECK-LABEL: @test_mm512_reduce_and_epi64( |
29 |
| -// CHECK: call i64 @llvm.vector.reduce.and.v8i64(<8 x i64> %{{.*}}) |
| 31 | +// CHECK-LABEL: test_mm512_reduce_and_epi64 |
| 32 | +// CHECK: call {{.*}}i64 @llvm.vector.reduce.and.v8i64(<8 x i64> %{{.*}}) |
30 | 33 | return _mm512_reduce_and_epi64(__W);
|
31 | 34 | }
|
32 | 35 | TEST_CONSTEXPR(_mm512_reduce_and_epi64((__m512i)(__v8di){0xFFFF, 0xFF00, 0x00FF, 0xFFFF, 0xFFFF, 0xFFFF, 0xFF00, 0x00FF}) == 0x0000);
|
33 | 36 |
|
34 | 37 | long long test_mm512_mask_reduce_add_epi64(__mmask8 __M, __m512i __W){
|
35 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_add_epi64( |
| 38 | +// CHECK-LABEL: test_mm512_mask_reduce_add_epi64 |
36 | 39 | // CHECK: select <8 x i1> %{{.*}}, <8 x i64> %{{.*}}, <8 x i64> %{{.*}}
|
37 |
| -// CHECK: call i64 @llvm.vector.reduce.add.v8i64(<8 x i64> %{{.*}}) |
| 40 | +// CHECK: call {{.*}}i64 @llvm.vector.reduce.add.v8i64(<8 x i64> %{{.*}}) |
38 | 41 | return _mm512_mask_reduce_add_epi64(__M, __W);
|
39 | 42 | }
|
40 | 43 |
|
41 | 44 | long long test_mm512_mask_reduce_mul_epi64(__mmask8 __M, __m512i __W){
|
42 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_mul_epi64( |
| 45 | +// CHECK-LABEL: test_mm512_mask_reduce_mul_epi64 |
43 | 46 | // CHECK: select <8 x i1> %{{.*}}, <8 x i64> %{{.*}}, <8 x i64> %{{.*}}
|
44 |
| -// CHECK: call i64 @llvm.vector.reduce.mul.v8i64(<8 x i64> %{{.*}}) |
| 47 | +// CHECK: call {{.*}}i64 @llvm.vector.reduce.mul.v8i64(<8 x i64> %{{.*}}) |
45 | 48 | return _mm512_mask_reduce_mul_epi64(__M, __W);
|
46 | 49 | }
|
47 | 50 |
|
48 | 51 | long long test_mm512_mask_reduce_and_epi64(__mmask8 __M, __m512i __W){
|
49 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_and_epi64( |
| 52 | +// CHECK-LABEL: test_mm512_mask_reduce_and_epi64 |
50 | 53 | // CHECK: select <8 x i1> %{{.*}}, <8 x i64> %{{.*}}, <8 x i64> %{{.*}}
|
51 |
| -// CHECK: call i64 @llvm.vector.reduce.and.v8i64(<8 x i64> %{{.*}}) |
| 54 | +// CHECK: call {{.*}}i64 @llvm.vector.reduce.and.v8i64(<8 x i64> %{{.*}}) |
52 | 55 | return _mm512_mask_reduce_and_epi64(__M, __W);
|
53 | 56 | }
|
54 | 57 |
|
55 | 58 | long long test_mm512_mask_reduce_or_epi64(__mmask8 __M, __m512i __W){
|
56 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_or_epi64( |
| 59 | +// CHECK-LABEL: test_mm512_mask_reduce_or_epi64 |
57 | 60 | // CHECK: select <8 x i1> %{{.*}}, <8 x i64> %{{.*}}, <8 x i64> %{{.*}}
|
58 |
| -// CHECK: call i64 @llvm.vector.reduce.or.v8i64(<8 x i64> %{{.*}}) |
| 61 | +// CHECK: call {{.*}}i64 @llvm.vector.reduce.or.v8i64(<8 x i64> %{{.*}}) |
59 | 62 | return _mm512_mask_reduce_or_epi64(__M, __W);
|
60 | 63 | }
|
61 | 64 |
|
62 | 65 | int test_mm512_reduce_add_epi32(__m512i __W){
|
63 |
| -// CHECK-LABEL: @test_mm512_reduce_add_epi32( |
64 |
| -// CHECK: call i32 @llvm.vector.reduce.add.v16i32(<16 x i32> %{{.*}}) |
| 66 | +// CHECK-LABEL: test_mm512_reduce_add_epi32 |
| 67 | +// CHECK: call {{.*}}i32 @llvm.vector.reduce.add.v16i32(<16 x i32> %{{.*}}) |
65 | 68 | return _mm512_reduce_add_epi32(__W);
|
66 | 69 | }
|
67 | 70 | TEST_CONSTEXPR(_mm512_reduce_add_epi32((__m512i)(__v16si){-8, -7, -6, -5, -4, -3, -2, -1, 0, 1, 2, 3, 4, 5, 6, 7}) == -8);
|
68 | 71 |
|
69 | 72 | int test_mm512_reduce_mul_epi32(__m512i __W){
|
70 |
| -// CHECK-LABEL: @test_mm512_reduce_mul_epi32( |
71 |
| -// CHECK: call i32 @llvm.vector.reduce.mul.v16i32(<16 x i32> %{{.*}}) |
| 73 | +// CHECK-LABEL: test_mm512_reduce_mul_epi32 |
| 74 | +// CHECK: call {{.*}}i32 @llvm.vector.reduce.mul.v16i32(<16 x i32> %{{.*}}) |
72 | 75 | return _mm512_reduce_mul_epi32(__W);
|
73 | 76 | }
|
74 | 77 | TEST_CONSTEXPR(_mm512_reduce_mul_epi32((__m512i)(__v16si){1, 1, 1, 1, 1, 1, 1, 1, 2, 2, 3, 1, 1, -3, 1, 1}) == -36);
|
75 | 78 |
|
76 | 79 | int test_mm512_reduce_or_epi32(__m512i __W){
|
77 |
| -// CHECK: call i32 @llvm.vector.reduce.or.v16i32(<16 x i32> %{{.*}}) |
| 80 | +// CHECK: call {{.*}}i32 @llvm.vector.reduce.or.v16i32(<16 x i32> %{{.*}}) |
78 | 81 | return _mm512_reduce_or_epi32(__W);
|
79 | 82 | }
|
80 | 83 | TEST_CONSTEXPR(_mm512_reduce_or_epi32((__m512i)(__v16si){0x01, 0x02, 0x04, 0x08, 0x10, 0x20, 0x40, 0x80, 0, 0, 0, 0, 0, 0, 0, 0}) == 0xFF);
|
81 | 84 |
|
82 | 85 | int test_mm512_reduce_and_epi32(__m512i __W){
|
83 |
| -// CHECK-LABEL: @test_mm512_reduce_and_epi32( |
84 |
| -// CHECK: call i32 @llvm.vector.reduce.and.v16i32(<16 x i32> %{{.*}}) |
| 86 | +// CHECK-LABEL: test_mm512_reduce_and_epi32 |
| 87 | +// CHECK: call {{.*}}i32 @llvm.vector.reduce.and.v16i32(<16 x i32> %{{.*}}) |
85 | 88 | return _mm512_reduce_and_epi32(__W);
|
86 | 89 | }
|
87 | 90 | TEST_CONSTEXPR(_mm512_reduce_and_epi32((__m512i)(__v16si){0xFF, 0xF0, 0x0F, 0xFF, 0xFF, 0xFF, 0xF0, 0x0F, 0xFF, 0xFF, 0xFF, 0xFF, 0xF0, 0xF0, 0x0F, 0x0F}) == 0x00);
|
88 | 91 |
|
89 | 92 | int test_mm512_mask_reduce_add_epi32(__mmask16 __M, __m512i __W){
|
90 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_add_epi32( |
| 93 | +// CHECK-LABEL: test_mm512_mask_reduce_add_epi32 |
91 | 94 | // CHECK: select <16 x i1> %{{.*}}, <16 x i32> %{{.*}}, <16 x i32> %{{.*}}
|
92 |
| -// CHECK: call i32 @llvm.vector.reduce.add.v16i32(<16 x i32> %{{.*}}) |
| 95 | +// CHECK: call {{.*}}i32 @llvm.vector.reduce.add.v16i32(<16 x i32> %{{.*}}) |
93 | 96 | return _mm512_mask_reduce_add_epi32(__M, __W);
|
94 | 97 | }
|
95 | 98 |
|
96 | 99 | int test_mm512_mask_reduce_mul_epi32(__mmask16 __M, __m512i __W){
|
97 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_mul_epi32( |
| 100 | +// CHECK-LABEL: test_mm512_mask_reduce_mul_epi32 |
98 | 101 | // CHECK: select <16 x i1> %{{.*}}, <16 x i32> %{{.*}}, <16 x i32> %{{.*}}
|
99 |
| -// CHECK: call i32 @llvm.vector.reduce.mul.v16i32(<16 x i32> %{{.*}}) |
| 102 | +// CHECK: call {{.*}}i32 @llvm.vector.reduce.mul.v16i32(<16 x i32> %{{.*}}) |
100 | 103 | return _mm512_mask_reduce_mul_epi32(__M, __W);
|
101 | 104 | }
|
102 | 105 |
|
103 | 106 | int test_mm512_mask_reduce_and_epi32(__mmask16 __M, __m512i __W){
|
104 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_and_epi32( |
| 107 | +// CHECK-LABEL: test_mm512_mask_reduce_and_epi32 |
105 | 108 | // CHECK: select <16 x i1> %{{.*}}, <16 x i32> %{{.*}}, <16 x i32> %{{.*}}
|
106 |
| -// CHECK: call i32 @llvm.vector.reduce.and.v16i32(<16 x i32> %{{.*}}) |
| 109 | +// CHECK: call {{.*}}i32 @llvm.vector.reduce.and.v16i32(<16 x i32> %{{.*}}) |
107 | 110 | return _mm512_mask_reduce_and_epi32(__M, __W);
|
108 | 111 | }
|
109 | 112 |
|
110 | 113 | int test_mm512_mask_reduce_or_epi32(__mmask16 __M, __m512i __W){
|
111 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_or_epi32( |
| 114 | +// CHECK-LABEL: test_mm512_mask_reduce_or_epi32 |
112 | 115 | // CHECK: select <16 x i1> %{{.*}}, <16 x i32> %{{.*}}, <16 x i32> %{{.*}}
|
113 |
| -// CHECK: call i32 @llvm.vector.reduce.or.v16i32(<16 x i32> %{{.*}}) |
| 116 | +// CHECK: call {{.*}}i32 @llvm.vector.reduce.or.v16i32(<16 x i32> %{{.*}}) |
114 | 117 | return _mm512_mask_reduce_or_epi32(__M, __W);
|
115 | 118 | }
|
116 | 119 |
|
117 | 120 | double test_mm512_reduce_add_pd(__m512d __W, double ExtraAddOp){
|
118 |
| -// CHECK-LABEL: @test_mm512_reduce_add_pd( |
| 121 | +// CHECK-LABEL: test_mm512_reduce_add_pd |
119 | 122 | // CHECK-NOT: reassoc
|
120 |
| -// CHECK: call reassoc double @llvm.vector.reduce.fadd.v8f64(double -0.000000e+00, <8 x double> %{{.*}}) |
| 123 | +// CHECK: call reassoc {{.*}}double @llvm.vector.reduce.fadd.v8f64(double -0.000000e+00, <8 x double> %{{.*}}) |
121 | 124 | // CHECK-NOT: reassoc
|
122 | 125 | return _mm512_reduce_add_pd(__W) + ExtraAddOp;
|
123 | 126 | }
|
124 | 127 |
|
125 | 128 | double test_mm512_reduce_mul_pd(__m512d __W, double ExtraMulOp){
|
126 |
| -// CHECK-LABEL: @test_mm512_reduce_mul_pd( |
| 129 | +// CHECK-LABEL: test_mm512_reduce_mul_pd |
127 | 130 | // CHECK-NOT: reassoc
|
128 |
| -// CHECK: call reassoc double @llvm.vector.reduce.fmul.v8f64(double 1.000000e+00, <8 x double> %{{.*}}) |
| 131 | +// CHECK: call reassoc {{.*}}double @llvm.vector.reduce.fmul.v8f64(double 1.000000e+00, <8 x double> %{{.*}}) |
129 | 132 | // CHECK-NOT: reassoc
|
130 | 133 | return _mm512_reduce_mul_pd(__W) * ExtraMulOp;
|
131 | 134 | }
|
132 | 135 |
|
133 | 136 | float test_mm512_reduce_add_ps(__m512 __W){
|
134 |
| -// CHECK-LABEL: @test_mm512_reduce_add_ps( |
135 |
| -// CHECK: call reassoc float @llvm.vector.reduce.fadd.v16f32(float -0.000000e+00, <16 x float> %{{.*}}) |
| 137 | +// CHECK-LABEL: test_mm512_reduce_add_ps |
| 138 | +// CHECK: call reassoc {{.*}}float @llvm.vector.reduce.fadd.v16f32(float -0.000000e+00, <16 x float> %{{.*}}) |
136 | 139 | return _mm512_reduce_add_ps(__W);
|
137 | 140 | }
|
138 | 141 |
|
139 | 142 | float test_mm512_reduce_mul_ps(__m512 __W){
|
140 |
| -// CHECK-LABEL: @test_mm512_reduce_mul_ps( |
141 |
| -// CHECK: call reassoc float @llvm.vector.reduce.fmul.v16f32(float 1.000000e+00, <16 x float> %{{.*}}) |
| 143 | +// CHECK-LABEL: test_mm512_reduce_mul_ps |
| 144 | +// CHECK: call reassoc {{.*}}float @llvm.vector.reduce.fmul.v16f32(float 1.000000e+00, <16 x float> %{{.*}}) |
142 | 145 | return _mm512_reduce_mul_ps(__W);
|
143 | 146 | }
|
144 | 147 |
|
145 | 148 | double test_mm512_mask_reduce_add_pd(__mmask8 __M, __m512d __W){
|
146 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_add_pd( |
| 149 | +// CHECK-LABEL: test_mm512_mask_reduce_add_pd |
147 | 150 | // CHECK: select <8 x i1> %{{.*}}, <8 x double> %{{.*}}, <8 x double> %{{.*}}
|
148 |
| -// CHECK: call reassoc double @llvm.vector.reduce.fadd.v8f64(double -0.000000e+00, <8 x double> %{{.*}}) |
| 151 | +// CHECK: call reassoc {{.*}}double @llvm.vector.reduce.fadd.v8f64(double -0.000000e+00, <8 x double> %{{.*}}) |
149 | 152 | return _mm512_mask_reduce_add_pd(__M, __W);
|
150 | 153 | }
|
151 | 154 |
|
152 | 155 | double test_mm512_mask_reduce_mul_pd(__mmask8 __M, __m512d __W){
|
153 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_mul_pd( |
| 156 | +// CHECK-LABEL: test_mm512_mask_reduce_mul_pd |
154 | 157 | // CHECK: select <8 x i1> %{{.*}}, <8 x double> %{{.*}}, <8 x double> %{{.*}}
|
155 |
| -// CHECK: call reassoc double @llvm.vector.reduce.fmul.v8f64(double 1.000000e+00, <8 x double> %{{.*}}) |
| 158 | +// CHECK: call reassoc {{.*}}double @llvm.vector.reduce.fmul.v8f64(double 1.000000e+00, <8 x double> %{{.*}}) |
156 | 159 | return _mm512_mask_reduce_mul_pd(__M, __W);
|
157 | 160 | }
|
158 | 161 |
|
159 | 162 | float test_mm512_mask_reduce_add_ps(__mmask16 __M, __m512 __W){
|
160 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_add_ps( |
| 163 | +// CHECK-LABEL: test_mm512_mask_reduce_add_ps |
161 | 164 | // CHECK: select <16 x i1> %{{.*}}, <16 x float> {{.*}}, <16 x float> {{.*}}
|
162 |
| -// CHECK: call reassoc float @llvm.vector.reduce.fadd.v16f32(float -0.000000e+00, <16 x float> %{{.*}}) |
| 165 | +// CHECK: call reassoc {{.*}}float @llvm.vector.reduce.fadd.v16f32(float -0.000000e+00, <16 x float> %{{.*}}) |
163 | 166 | return _mm512_mask_reduce_add_ps(__M, __W);
|
164 | 167 | }
|
165 | 168 |
|
166 | 169 | float test_mm512_mask_reduce_mul_ps(__mmask16 __M, __m512 __W){
|
167 |
| -// CHECK-LABEL: @test_mm512_mask_reduce_mul_ps( |
| 170 | +// CHECK-LABEL: test_mm512_mask_reduce_mul_ps |
168 | 171 | // CHECK: select <16 x i1> %{{.*}}, <16 x float> {{.*}}, <16 x float> %{{.*}}
|
169 |
| -// CHECK: call reassoc float @llvm.vector.reduce.fmul.v16f32(float 1.000000e+00, <16 x float> %{{.*}}) |
| 172 | +// CHECK: call reassoc {{.*}}float @llvm.vector.reduce.fmul.v16f32(float 1.000000e+00, <16 x float> %{{.*}}) |
170 | 173 | return _mm512_mask_reduce_mul_ps(__M, __W);
|
171 | 174 | }
|
0 commit comments