Skip to content

Commit 63a6dcf

Browse files
HappyAmazonianredhairdragonLokiiiiii
authored
merge lora fix & autoawq build issue (#2963)
Co-authored-by: Shen Teng <shen_teng@engineering.ucla.edu> Co-authored-by: Loki <lokravi@amazon.com>
1 parent bfe6b3e commit 63a6dcf

File tree

2 files changed

+4
-1
lines changed

2 files changed

+4
-1
lines changed

engines/python/setup/djl_python/lmi_vllm/vllm_async_service.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -273,6 +273,8 @@ async def add_lora(self, lora_name: str, lora_alias: str, lora_path: str):
273273
logging.info(f"Adding LoRA {lora_name} from {lora_path}")
274274
lora_id = self.lora_id_counter.inc(1)
275275
lora_request = create_lora_request(lora_name, lora_id, lora_path, None)
276+
# Register the LoRA request with the model registry so vLLM can find it
277+
self.model_registry.lora_requests[lora_name] = lora_request
276278
self.lora_requests[lora_request.lora_name] = lora_request
277279
result = await self.vllm_engine.add_lora(lora_request)
278280
logging.info(f"LoRA {lora_name} added to engine: {result}")
@@ -285,6 +287,7 @@ async def remove_lora(self, lora_name: str, lora_alias: str):
285287
lora_request = get_lora_request(lora_name, self.lora_requests)
286288
result = await self.vllm_engine.remove_lora(lora_request.lora_int_id)
287289
del self.lora_requests[lora_name]
290+
del self.model_registry.lora_requests[lora_name]
288291
return result
289292

290293
async def pin_lora(self, lora_name: str, lora_alias: str):

serving/docker/lmi-container-requirements.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ scipy==1.16.0
2323
onnx==1.19.0
2424
sentence_transformers
2525
onnxruntime-gpu==1.20.0
26-
autoawq
26+
autoawq==0.2.9
2727
tokenizers
2828
pydantic==2.11.7
2929
optimum==1.23.2

0 commit comments

Comments
 (0)