Skip to content

Commit 0d2668a

Browse files
singlaiiitWei Wei
authored andcommitted
[fx2trt] Add more profiling logging (#78)
Summary: Pull Request resolved: pytorch/fx2trt#78 This is to analyze more on the lowering latency. Reviewed By: yyetim, khabinov Differential Revision: D36352458 fbshipit-source-id: d3f40d02b89b815fd7a0c2ad90ab2ab0a1b24207
1 parent b5262ae commit 0d2668a

File tree

1 file changed

+14
-0
lines changed

1 file changed

+14
-0
lines changed

fx/fx2trt.py

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,6 @@
1+
import logging
12
import warnings
3+
from datetime import datetime
24
from typing import Any, Callable, Dict, List, NamedTuple, Optional, Sequence
35

46
import numpy
@@ -15,6 +17,10 @@
1517
from .input_tensor_spec import InputTensorSpec
1618
from .utils import get_dynamic_dims, LowerPrecision, torch_dtype_to_trt
1719

20+
21+
_LOGGER: logging.Logger = logging.getLogger(__name__)
22+
23+
1824
TRT_INTERPRETER_CALL_PRE_OBSERVER: Observer[
1925
Callable[[torch.fx.GraphModule], None]
2026
] = Observer("TRT_INTERPRETER_CALL_PRE_OBSERVER")
@@ -179,7 +185,12 @@ def run(
179185
warnings.warn("Current platform doesn't support fast native fp16!")
180186

181187
self.input_specs_iter = 0
188+
run_module_start_time = datetime.now()
182189
super().run()
190+
_LOGGER.info(
191+
f"Run Module elapsed time: {datetime.now() - run_module_start_time}"
192+
)
193+
build_engine_start_time = datetime.now()
183194

184195
self.builder.max_batch_size = max_batch_size
185196
builder_config = self.builder.create_builder_config()
@@ -227,6 +238,9 @@ def run(
227238
if builder_config.get_timing_cache()
228239
else bytearray()
229240
)
241+
_LOGGER.info(
242+
f"Build TRT engine elapsed time: {datetime.now() - build_engine_start_time}"
243+
)
230244

231245
return TRTInterpreterResult(
232246
engine, self._input_names, self._output_names, serialized_cache

0 commit comments

Comments
 (0)