File tree Expand file tree Collapse file tree 14 files changed +16
-17
lines changed Expand file tree Collapse file tree 14 files changed +16
-17
lines changed Original file line number Diff line number Diff line change @@ -112,8 +112,8 @@ void DataFlowGraph::Build(const framework::proto::ProgramDesc &prog) {
112
112
out_alias->SetPbMsg (out->pb_msg ());
113
113
var2id[out_alias->name ()] =
114
114
out_alias->id (); // update variable's alias Node
115
- LOG (INFO ) << " loop found in graph, create SSA alias node ["
116
- << out_alias->repr () << " ] for [" << out->repr () << " ]" ;
115
+ VLOG ( 40 ) << " loop found in graph, create SSA alias node ["
116
+ << out_alias->repr () << " ] for [" << out->repr () << " ]" ;
117
117
out = out_alias;
118
118
}
119
119
out->inlinks .push_back (o);
Original file line number Diff line number Diff line change @@ -27,7 +27,7 @@ class ActivationOpConverter : public OpConverter {
27
27
// Here the two nullptr looks strange, that's because the
28
28
// framework::OpDesc's constructor is strange.
29
29
framework::OpDesc op_desc (op, nullptr );
30
- LOG (INFO )
30
+ VLOG ( 3 )
31
31
<< " convert a fluid Activation op to tensorrt activation layer whose "
32
32
" type is "
33
33
<< op_type_;
Original file line number Diff line number Diff line change @@ -23,7 +23,7 @@ class BatchNormOpConverter : public OpConverter {
23
23
public:
24
24
void operator ()(const framework::proto::OpDesc& op,
25
25
const framework::Scope& scope, bool test_mode) override {
26
- LOG (INFO ) << " convert a fluid batch norm op to tensorrt batch_norm" ;
26
+ VLOG ( 3 ) << " convert a fluid batch norm op to tensorrt batch_norm" ;
27
27
28
28
framework::OpDesc op_desc (op, nullptr );
29
29
PADDLE_ENFORCE_EQ (op_desc.Input (" X" ).size (), 1 );
Original file line number Diff line number Diff line change @@ -25,7 +25,7 @@ class ConcatOpConverter : public OpConverter {
25
25
public:
26
26
void operator ()(const framework::proto::OpDesc& op,
27
27
const framework::Scope& scope, bool test_mode) override {
28
- VLOG (40 ) << " convert a fluid mul op to tensorrt mul layer without bias" ;
28
+ VLOG (3 ) << " convert a fluid mul op to tensorrt mul layer without bias" ;
29
29
30
30
framework::OpDesc op_desc (op, nullptr );
31
31
// Declare inputs
Original file line number Diff line number Diff line change @@ -37,8 +37,7 @@ class Conv2dOpConverter : public OpConverter {
37
37
public:
38
38
void operator ()(const framework::proto::OpDesc& op,
39
39
const framework::Scope& scope, bool test_mode) override {
40
- LOG (INFO)
41
- << " convert a fluid conv2d op to tensorrt conv layer without bias" ;
40
+ VLOG (3 ) << " convert a fluid conv2d op to tensorrt conv layer without bias" ;
42
41
43
42
framework::OpDesc op_desc (op, nullptr );
44
43
PADDLE_ENFORCE_EQ (op_desc.Input (" Input" ).size (), 1 );
Original file line number Diff line number Diff line change @@ -25,7 +25,7 @@ class DropoutOpConverter : public OpConverter {
25
25
public:
26
26
void operator ()(const framework::proto::OpDesc& op,
27
27
const framework::Scope& scope, bool test_mode) override {
28
- VLOG (40 ) << " convert a fluid dropout op to tensorrt dropout layer" ;
28
+ VLOG (3 ) << " convert a fluid dropout op to tensorrt dropout layer" ;
29
29
framework::OpDesc op_desc (op, nullptr );
30
30
// Declare inputs
31
31
auto * input1 = engine_->GetITensor (op_desc.Input (" X" )[0 ]);
Original file line number Diff line number Diff line change @@ -26,7 +26,7 @@ class ElementwiseWeightOpConverter : public OpConverter {
26
26
// Here the two nullptr looks strange, that's because the
27
27
// framework::OpDesc's constructor is strange.
28
28
framework::OpDesc op_desc (op, nullptr );
29
- LOG (INFO ) << " convert a fluid elementwise op to tensorrt IScaleLayer" ;
29
+ VLOG ( 3 ) << " convert a fluid elementwise op to tensorrt IScaleLayer" ;
30
30
31
31
PADDLE_ENFORCE_EQ (op_desc.Input (" X" ).size (), 1 );
32
32
PADDLE_ENFORCE_EQ (op_desc.Input (" Y" ).size (), 1 ); // Y is a weight
@@ -108,7 +108,7 @@ class ElementwiseTensorOpConverter : public OpConverter {
108
108
// Here the two nullptr looks strange, that's because the
109
109
// framework::OpDesc's constructor is strange.
110
110
framework::OpDesc op_desc (op, nullptr );
111
- LOG (INFO ) << " convert a fluid elementwise op to tensorrt IScaleLayer" ;
111
+ VLOG ( 3 ) << " convert a fluid elementwise op to tensorrt IScaleLayer" ;
112
112
113
113
PADDLE_ENFORCE_EQ (op_desc.Input (" X" ).size (), 1 );
114
114
PADDLE_ENFORCE_EQ (op_desc.Input (" Y" ).size (), 1 ); // Y is a weight
Original file line number Diff line number Diff line change @@ -52,7 +52,7 @@ class FcOpConverter : public OpConverter {
52
52
public:
53
53
void operator ()(const framework::proto::OpDesc& op,
54
54
const framework::Scope& scope, bool test_mode) override {
55
- VLOG (40 ) << " convert a fluid fc op to tensorrt fc layer without bias" ;
55
+ VLOG (3 ) << " convert a fluid fc op to tensorrt fc layer without bias" ;
56
56
57
57
framework::OpDesc op_desc (op, nullptr );
58
58
PADDLE_ENFORCE_EQ (op_desc.Input (" X" ).size (), 1 );
Original file line number Diff line number Diff line change @@ -25,7 +25,7 @@ class MulOpConverter : public OpConverter {
25
25
public:
26
26
void operator ()(const framework::proto::OpDesc& op,
27
27
const framework::Scope& scope, bool test_mode) override {
28
- VLOG (40 ) << " convert a fluid mul op to tensorrt mul layer without bias" ;
28
+ VLOG (3 ) << " convert a fluid mul op to tensorrt mul layer without bias" ;
29
29
30
30
framework::OpDesc op_desc (op, nullptr );
31
31
// Declare inputs
Original file line number Diff line number Diff line change @@ -25,7 +25,7 @@ class PadOpConverter : public OpConverter {
25
25
public:
26
26
void operator ()(const framework::proto::OpDesc& op,
27
27
const framework::Scope& scope, bool test_mode) override {
28
- VLOG (40 ) << " convert a fluid transpose op to tensorrt tranpose layer" ;
28
+ VLOG (3 ) << " convert a fluid transpose op to tensorrt tranpose layer" ;
29
29
30
30
framework::OpDesc op_desc (op, nullptr );
31
31
// Declare inputs
You can’t perform that action at this time.
0 commit comments