@@ -105,7 +105,7 @@ void TestInference(const std::string& dirname,
105
105
state = paddle::platform::ProfilerState::kCPU ;
106
106
} else {
107
107
#ifdef PADDLE_WITH_CUDA
108
- state = paddle::platform::ProfilerState::kCUDA ;
108
+ state = paddle::platform::ProfilerState::kAll ;
109
109
// The default device_id of paddle::platform::CUDAPlace is 0.
110
110
// Users can get the device_id using:
111
111
// int device_id = place.GetDeviceId();
@@ -143,7 +143,7 @@ void TestInference(const std::string& dirname,
143
143
}
144
144
// Disable the profiler and print the timing information
145
145
paddle::platform::DisableProfiler (paddle::platform::EventSortingKey::kDefault ,
146
- " load_program_profiler.txt " );
146
+ " load_program_profiler" );
147
147
paddle::platform::ResetProfiler ();
148
148
149
149
// 3. Get the feed_target_names and fetch_target_names
@@ -208,7 +208,7 @@ void TestInference(const std::string& dirname,
208
208
// Disable the profiler and print the timing information
209
209
paddle::platform::DisableProfiler (
210
210
paddle::platform::EventSortingKey::kDefault ,
211
- " run_inference_profiler.txt " );
211
+ " run_inference_profiler" );
212
212
paddle::platform::ResetProfiler ();
213
213
}
214
214
0 commit comments