We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 683c635 commit 0e40429Copy full SHA for 0e40429
paddle/fluid/inference/api/api_tensorrt_subgraph_engine.cc
@@ -32,8 +32,8 @@ class TensorRTSubgraphPredictor : public NativePaddlePredictor {
32
: NativePaddlePredictor(config), config_(config) {}
33
34
bool Init(const std::shared_ptr<framework::Scope>& parent_scope) {
35
- FLAGS_inference_analysis_enable_tensorrt_subgraph_engine = true;
36
VLOG(3) << "Predictor::init()";
+ FLAGS_inference_analysis_enable_tensorrt_subgraph_engine = true;
37
FLAGS_tensorrt_max_batch_size = config_.max_batch_size;
38
FLAGS_tensorrt_workspace_size = config_.workspace_size;
39
if (config_.use_gpu) {
0 commit comments