@@ -105,16 +105,14 @@ TEST(TensorCopy, Tensor) {
105
105
}
106
106
107
107
TEST (TensorFromVector, Tensor) {
108
- using namespace paddle ::framework;
109
- using namespace paddle ::platform;
110
108
{
111
109
std::vector<int > src_vec = {1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 };
112
- Tensor cpu_tensor;
110
+ paddle::framework:: Tensor cpu_tensor;
113
111
114
112
// Copy to CPU Tensor
115
- cpu_tensor.Resize (make_ddim ({3 , 3 }));
113
+ cpu_tensor.Resize (paddle::framework:: make_ddim ({3 , 3 }));
116
114
auto cpu_place = new paddle::platform::CPUPlace ();
117
- TensorFromVector<int >(src_vec, &cpu_tensor);
115
+ paddle::framework:: TensorFromVector<int >(src_vec, &cpu_tensor);
118
116
119
117
// Compare Tensors
120
118
const int * cpu_ptr = cpu_tensor.data <int >();
@@ -125,8 +123,8 @@ TEST(TensorFromVector, Tensor) {
125
123
}
126
124
127
125
src_vec.erase (src_vec.begin (), src_vec.begin () + 5 );
128
- cpu_tensor.Resize (make_ddim ({2 , 2 }));
129
- TensorFromVector<int >(src_vec, &cpu_tensor);
126
+ cpu_tensor.Resize (paddle::framework:: make_ddim ({2 , 2 }));
127
+ paddle::framework:: TensorFromVector<int >(src_vec, &cpu_tensor);
130
128
cpu_ptr = cpu_tensor.data <int >();
131
129
src_ptr = src_vec.data ();
132
130
ASSERT_NE (src_ptr, cpu_ptr);
@@ -140,23 +138,23 @@ TEST(TensorFromVector, Tensor) {
140
138
#ifdef PADDLE_WITH_CUDA
141
139
{
142
140
std::vector<int > src_vec = {1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 };
143
- Tensor cpu_tensor;
144
- Tensor gpu_tensor;
145
- Tensor dst_tensor;
141
+ paddle::framework:: Tensor cpu_tensor;
142
+ paddle::framework:: Tensor gpu_tensor;
143
+ paddle::framework:: Tensor dst_tensor;
146
144
147
145
// Copy to CPU Tensor
148
146
cpu_tensor.Resize (make_ddim ({3 , 3 }));
149
147
auto cpu_place = new paddle::platform::CPUPlace ();
150
- CPUDeviceContext cpu_ctx (*cpu_place);
151
- TensorFromVector<int >(src_vec, cpu_ctx, &cpu_tensor);
148
+ paddle::platform:: CPUDeviceContext cpu_ctx (*cpu_place);
149
+ paddle::framework:: TensorFromVector<int >(src_vec, cpu_ctx, &cpu_tensor);
152
150
153
151
// Copy to GPUTensor
154
- gpu_tensor.Resize (make_ddim ({3 , 3 }));
152
+ gpu_tensor.Resize (paddle::framework:: make_ddim ({3 , 3 }));
155
153
auto gpu_place = new paddle::platform::CUDAPlace ();
156
- CUDADeviceContext gpu_ctx (*gpu_place);
157
- TensorFromVector<int >(src_vec, gpu_ctx, &gpu_tensor);
154
+ paddle::platform:: CUDADeviceContext gpu_ctx (*gpu_place);
155
+ paddle::framework:: TensorFromVector<int >(src_vec, gpu_ctx, &gpu_tensor);
158
156
// Copy from GPU to CPU tensor for comparison
159
- TensorCopy (gpu_tensor, *cpu_place, gpu_ctx, &dst_tensor);
157
+ paddle::framework:: TensorCopy (gpu_tensor, *cpu_place, gpu_ctx, &dst_tensor);
160
158
161
159
// Sync before Compare Tensors
162
160
gpu_ctx.Wait ();
@@ -172,11 +170,11 @@ TEST(TensorFromVector, Tensor) {
172
170
173
171
src_vec.erase (src_vec.begin (), src_vec.begin () + 5 );
174
172
175
- cpu_tensor.Resize (make_ddim ({2 , 2 }));
176
- TensorFromVector<int >(src_vec, cpu_ctx, &cpu_tensor);
177
- gpu_tensor.Resize (make_ddim ({2 , 2 }));
178
- TensorFromVector<int >(src_vec, gpu_ctx, &gpu_tensor);
179
- TensorCopy (gpu_tensor, *cpu_place, gpu_ctx, &dst_tensor);
173
+ cpu_tensor.Resize (paddle::framework:: make_ddim ({2 , 2 }));
174
+ paddle::framework:: TensorFromVector<int >(src_vec, cpu_ctx, &cpu_tensor);
175
+ gpu_tensor.Resize (paddle::framework:: make_ddim ({2 , 2 }));
176
+ paddle::framework:: TensorFromVector<int >(src_vec, gpu_ctx, &gpu_tensor);
177
+ paddle::framework:: TensorCopy (gpu_tensor, *cpu_place, gpu_ctx, &dst_tensor);
180
178
181
179
// Sync before Compare Tensors
182
180
gpu_ctx.Wait ();
@@ -197,18 +195,16 @@ TEST(TensorFromVector, Tensor) {
197
195
}
198
196
199
197
TEST (TensorToVector, Tensor) {
200
- using namespace paddle ::framework;
201
- using namespace paddle ::platform;
202
198
{
203
- Tensor src;
204
- int * src_ptr = src.mutable_data <int >({3 , 3 }, CPUPlace ());
199
+ paddle::framework:: Tensor src;
200
+ int * src_ptr = src.mutable_data <int >({3 , 3 }, paddle::platform:: CPUPlace ());
205
201
for (int i = 0 ; i < 3 * 3 ; ++i) {
206
202
src_ptr[i] = i;
207
203
}
208
204
209
- CPUPlace place;
205
+ paddle::platform:: CPUPlace place;
210
206
std::vector<int > dst;
211
- TensorToVector<int >(src, &dst);
207
+ paddle::framework:: TensorToVector<int >(src, &dst);
212
208
213
209
for (int i = 0 ; i < 3 * 3 ; ++i) {
214
210
EXPECT_EQ (src_ptr[i], dst[i]);
@@ -217,13 +213,13 @@ TEST(TensorToVector, Tensor) {
217
213
#ifdef PADDLE_WITH_CUDA
218
214
{
219
215
std::vector<int > src_vec = {1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 };
220
- Tensor gpu_tensor;
221
- CUDAPlace place;
222
- CUDADeviceContext gpu_ctx (place);
223
- TensorFromVector<int >(src_vec, gpu_ctx, &gpu_tensor);
216
+ paddle::framework:: Tensor gpu_tensor;
217
+ paddle::platform:: CUDAPlace place;
218
+ paddle::platform:: CUDADeviceContext gpu_ctx (place);
219
+ paddle::framework:: TensorFromVector<int >(src_vec, gpu_ctx, &gpu_tensor);
224
220
225
221
std::vector<int > dst;
226
- TensorToVector<int >(gpu_tensor, gpu_ctx, &dst);
222
+ paddle::framework:: TensorToVector<int >(gpu_tensor, gpu_ctx, &dst);
227
223
228
224
for (int i = 0 ; i < 3 * 3 ; ++i) {
229
225
EXPECT_EQ (src_vec[i], dst[i]);
@@ -233,54 +229,54 @@ TEST(TensorToVector, Tensor) {
233
229
}
234
230
235
231
TEST (TensorContainsNAN, CPU) {
236
- using namespace paddle ::framework;
237
- using namespace paddle ::platform;
238
232
{
239
- Tensor src;
240
- float * buf = src.mutable_data <float >({3 }, CPUPlace ());
233
+ paddle::framework:: Tensor src;
234
+ float * buf = src.mutable_data <float >({3 }, paddle::platform:: CPUPlace ());
241
235
buf[0 ] = 0.0 ;
242
236
buf[1 ] = NAN;
243
237
buf[2 ] = 0.0 ;
244
- ASSERT_TRUE (TensorContainsNAN (src));
238
+ ASSERT_TRUE (paddle::framework:: TensorContainsNAN (src));
245
239
buf[1 ] = 0.0 ;
246
- ASSERT_FALSE (TensorContainsNAN (src));
240
+ ASSERT_FALSE (paddle::framework:: TensorContainsNAN (src));
247
241
}
248
242
249
243
{
250
- Tensor src;
251
- float16* buf = src.mutable_data <float16>({3 }, CPUPlace ());
244
+ paddle::framework::Tensor src;
245
+ paddle::platform::float16* buf =
246
+ src.mutable_data <paddle::platform::float16>(
247
+ {3 }, paddle::platform::CPUPlace ());
252
248
buf[0 ] = 0.0 ;
253
249
buf[1 ].x = 0x7fff ;
254
250
buf[2 ] = 0.0 ;
255
- ASSERT_TRUE (TensorContainsNAN (src));
251
+ ASSERT_TRUE (paddle::framework:: TensorContainsNAN (src));
256
252
buf[1 ] = 0.0 ;
257
- ASSERT_FALSE (TensorContainsNAN (src));
253
+ ASSERT_FALSE (paddle::framework:: TensorContainsNAN (src));
258
254
}
259
255
}
260
256
261
257
TEST (TensorContainsInf, CPU) {
262
- using namespace paddle ::framework;
263
- using namespace paddle ::platform;
264
258
{
265
- Tensor src;
266
- double * buf = src.mutable_data <double >({3 }, CPUPlace ());
259
+ paddle::framework:: Tensor src;
260
+ double * buf = src.mutable_data <double >({3 }, paddle::platform:: CPUPlace ());
267
261
buf[0 ] = 1.0 ;
268
262
buf[1 ] = INFINITY;
269
263
buf[2 ] = 0.0 ;
270
- ASSERT_TRUE (TensorContainsInf (src));
264
+ ASSERT_TRUE (paddle::framework:: TensorContainsInf (src));
271
265
buf[1 ] = 1.0 ;
272
- ASSERT_FALSE (TensorContainsInf (src));
266
+ ASSERT_FALSE (paddle::framework:: TensorContainsInf (src));
273
267
}
274
268
275
269
{
276
- Tensor src;
277
- float16* buf = src.mutable_data <float16>({3 }, CPUPlace ());
270
+ paddle::framework::Tensor src;
271
+ paddle::platform::float16* buf =
272
+ src.mutable_data <paddle::platform::float16>(
273
+ {3 }, paddle::platform::CPUPlace ());
278
274
buf[0 ] = 1.0 ;
279
275
buf[1 ].x = 0x7c00 ;
280
276
buf[2 ] = 0.0 ;
281
- ASSERT_TRUE (TensorContainsInf (src));
277
+ ASSERT_TRUE (paddle::framework:: TensorContainsInf (src));
282
278
buf[1 ] = 1.0 ;
283
- ASSERT_FALSE (TensorContainsInf (src));
279
+ ASSERT_FALSE (paddle::framework:: TensorContainsInf (src));
284
280
}
285
281
}
286
282
0 commit comments