We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent c17231e commit 8425f78Copy full SHA for 8425f78
vllm/platforms/cuda.py
@@ -182,9 +182,6 @@ def check_and_update_config(cls, vllm_config: "VllmConfig") -> None:
182
compilation_config.use_cudagraph = False
183
if model_config is not None:
184
model_config.enforce_eager = True
185
- # TODO (varun): Turning this ON gives incorrect results for the
186
- # Deepseek-V2-lite model.
187
- vllm_config.compilation_config.use_inductor = False
188
189
@classmethod
190
def get_current_memory_usage(cls,
0 commit comments