Skip to content

Commit 05719cc

Browse files
authored
fix vlm deploy lora & agent (#1371)
1 parent 465f7e6 commit 05719cc

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

swift/llm/deploy.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -116,13 +116,13 @@ async def _prepare_request(request: Union[ChatCompletionRequest, CompletionReque
116116
f'the model `{model_or_engine.model_type}` is in text generation format. '
117117
'Please use the `completions` API.')
118118
messages = request.messages
119-
# For agent, check if response is endwith observations and join tool observation
120-
messages_join_observation(messages)
121119
images = request.images
122120
if _args.is_multimodal:
123121
compat_openai(messages, images, template.template_type)
124122
messages = decode_base64(messages=messages)['messages']
125123
images = decode_base64(images=images)['images']
124+
# For agent, check if response is endwith observations and join tool observation
125+
messages_join_observation(messages)
126126
example = messages_to_history(messages)
127127
if len(images) > 0:
128128
example['images'] = images

swift/llm/infer.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -210,7 +210,7 @@ def prepare_model_template(args: InferArguments,
210210
f'args.max_model_len: {args.max_model_len}, model.max_model_len: {model.max_model_len}')
211211
# Preparing LoRA
212212
if is_adapter(args.sft_type) and args.ckpt_dir is not None:
213-
if is_quant_model(args.model_type, model):
213+
if args.lora_request_list is not None and (is_quant_model(args.model_type, model) or args.is_multimodal):
214214
# gptq awq does not support lora switching
215215
args.lora_request_list = None
216216
logger.warning('The current model does not support LoRA switching. '

0 commit comments

Comments
 (0)