@@ -71,13 +71,11 @@ def compile(
7171 engine_capability : EngineCapability = ENGINE_CAPABILITY ,
7272 refit : bool = REFIT ,
7373 debug : bool = DEBUG ,
74- capability : EngineCapability = EngineCapability .default ,
7574 num_avg_timing_iters : int = NUM_AVG_TIMING_ITERS ,
7675 workspace_size : int = WORKSPACE_SIZE ,
7776 dla_sram_size : int = DLA_SRAM_SIZE ,
7877 dla_local_dram_size : int = DLA_LOCAL_DRAM_SIZE ,
7978 dla_global_dram_size : int = DLA_GLOBAL_DRAM_SIZE ,
80- calibrator : object = None ,
8179 truncate_long_and_double : bool = TRUNCATE_LONG_AND_DOUBLE ,
8280 require_full_compilation : bool = REQUIRE_FULL_COMPILATION ,
8381 min_block_size : int = MIN_BLOCK_SIZE ,
@@ -156,6 +154,12 @@ def compile(
156154 if debug :
157155 set_log_level (logger .parent , logging .DEBUG )
158156
157+ if torch_executed_modules is not None and torch_executed_modules :
158+ logger .warning (
159+ f"Detected torch_executed_modules was non-empty: { torch_executed_modules } "
160+ "\n This feature is unimplemented in Torch-TRT Dynamo currently."
161+ )
162+
159163 if not isinstance (inputs , collections .abc .Sequence ):
160164 inputs = [inputs ]
161165
@@ -214,6 +218,7 @@ def compile(
214218 "use_python_runtime" : use_python_runtime ,
215219 "truncate_long_and_double" : truncate_long_and_double ,
216220 "use_fast_partitioner" : use_fast_partitioner ,
221+ "num_avg_timing_iters" : num_avg_timing_iters ,
217222 "enable_experimental_decompositions" : enable_experimental_decompositions ,
218223 "require_full_compilation" : require_full_compilation ,
219224 "disable_tf32" : disable_tf32 ,
0 commit comments