Skip to content

Commit 6a1d7ad

Browse files
Fix not cleared globals in runtime config (#1983)
Related change in extension: HabanaAI/vllm-hpu-extension#370 The PR fixes case when we run vllm multiple times in the same process, e.g. running some pytest scenarios. Signed-off-by: Artur Fierka <[email protected]>
1 parent 552bb0f commit 6a1d7ad

File tree

2 files changed

+4
-2
lines changed

2 files changed

+4
-2
lines changed

requirements/hpu.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ ray<2.49.0
77
triton==3.1.0
88
setuptools>=77.0.3
99
setuptools-scm>=8
10-
vllm-hpu-extension @ git+https://github.com/HabanaAI/vllm-hpu-extension.git@048015b
10+
vllm-hpu-extension @ git+https://github.com/HabanaAI/vllm-hpu-extension.git@7717587
1111

1212
# Dependencies for HPU vllm docker image
1313
datasets

vllm/worker/hpu_model_runner.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,8 @@
3636
HabanaMemoryProfiler,
3737
HabanaProfilerCounterHelper,
3838
format_bytes)
39-
from vllm_hpu_extension.runtime import finalize_config, get_config
39+
from vllm_hpu_extension.runtime import (clear_config, finalize_config,
40+
get_config)
4041

4142
import vllm.envs as envs
4243
from vllm.attention import AttentionMetadata, get_attn_backend
@@ -4114,6 +4115,7 @@ def _make_decode_output(
41144115
return SamplerOutput(sampler_outputs)
41154116

41164117
def __del__(self):
4118+
clear_config()
41174119
self.shutdown_inc()
41184120

41194121
def _patch_prev_output(self):

0 commit comments

Comments
 (0)