diff --git a/py/torch_tensorrt/dynamo/conversion/custom_ops_converters.py b/py/torch_tensorrt/dynamo/conversion/custom_ops_converters.py index 17850fabce..6fedd712b1 100644 --- a/py/torch_tensorrt/dynamo/conversion/custom_ops_converters.py +++ b/py/torch_tensorrt/dynamo/conversion/custom_ops_converters.py @@ -54,7 +54,6 @@ def fused_nccl_reduce_scatter( [args[0]], ) - breakpoint() else: _LOGGER.debug( "Did not load torch.distributed converters since TensorRT-LLM is not available"