Skip to content

Commit e92a813

Browse files
committed
chore: Fix fp16 functionality for fx2trt backend
Signed-off-by: Dheeraj Peri <[email protected]>
1 parent d108f87 commit e92a813

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

tools/perf/perf_run.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -120,13 +120,15 @@ def run_fx2trt(model, input_tensors, params, precision, batch_size):
120120
precision = LowerPrecision.FP32
121121
elif precision == "fp16":
122122
precision = LowerPrecision.FP16
123+
model.half()
124+
input_tensors = [tensor.half() for tensor in input_tensors]
123125
# Run lowering eager mode benchmark
124126
model = lower_to_trt(
125127
model,
126128
input_tensors,
127129
max_batch_size=batch_size,
128130
lower_precision=precision,
129-
verbose_log=True,
131+
verbose_log=False,
130132
)
131133

132134
iters = params.get('iterations', 20)

0 commit comments

Comments
 (0)