Skip to content

Commit ad594c7

Browse files
authored
Merge pull request #547 from foundation-model-stack/v2.8.2-rc1
chore(release): merge set of changes for v2.8.2
2 parents 4fa54e1 + 4bd099a commit ad594c7

File tree

2 files changed

+15
-21
lines changed

2 files changed

+15
-21
lines changed

README.md

Lines changed: 9 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -909,6 +909,8 @@ For information on supported dataset formats and how to tune a vision-language m
909909

910910
### Supported vision model
911911

912+
Note that vision models are supported starting with `fms-hf-tuning` v2.8.1 or later.
913+
912914
- Legend:
913915

914916
✅ Ready and available
@@ -921,12 +923,13 @@ For information on supported dataset formats and how to tune a vision-language m
921923

922924
Model Name & Size | Model Architecture | LoRA Tuning | Full Finetuning |
923925
-------------------- | ---------------- | --------------- | --------------- |
924-
Llama 3.2-11B Vision | MllamaForConditionalGeneration | ✅* | ✅* |
925-
Llava 1.5-7B | LlavaForConditionalGeneration | ✅* | ✅* |
926-
Granite 3.1-2B Vision | LlavaNextForConditionalGeneration | ✅* | ✅* |
927-
Llava Mistral 1.6-7B | LlavaNextForConditionalGeneration | ✅* | ✅* |
928-
929-
(*) - Supported with `fms-hf-tuning` v2.8.0 or later.
926+
Llama 3.2-11B Vision | MllamaForConditionalGeneration | ✅ | ✅ |
927+
Llama 3.2-90B Vision | MllamaForConditionalGeneration | ✔️ | ✔️ |
928+
Granite 3.2-2B Vision | LlavaNextForConditionalGeneration | ✅ | ✅ |
929+
Llava Mistral 1.6-7B | LlavaNextForConditionalGeneration | ✅ | ✅ |
930+
Llava 1.6-34B | LlavaNextForConditionalGeneration | ✔️ | ✔️ |
931+
Llava 1.5-7B | LlavaForConditionalGeneration | ✅ | ✅ |
932+
Llava 1.5-13B | LlavaForConditionalGeneration | ✔️ | ✔️ |
930933

931934
**Note**: vLLM currently does not support inference with LoRA-tuned vision models. To use a tuned LoRA adapter of vision model, please merge it with the base model before running vLLM inference.
932935

tuning/sft_trainer.py

Lines changed: 6 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -167,26 +167,17 @@ def train(
167167
"`--padding_free` argument was called with `packing=True`, "
168168
"Trainer should not perform packing when using `--padding_free`"
169169
)
170-
170+
if fast_moe_config is not None and fast_moe_config.fast_moe is None:
171+
fast_moe_config = None
171172
if fast_moe_config is not None:
172-
# Checking for unsupported modules with Scatter MoE for LoRA
173-
# Only raise an error for `all-linear`
174-
restricted_modules = ["all-linear"]
173+
# If LoRA with ScatterMoE detected, raise warning
174+
accepted_layers = ["all-linear"]
175175
if (
176176
peft_config is not None
177177
and hasattr(peft_config, "target_modules")
178-
and any(
179-
module in (peft_config.target_modules or [])
180-
for module in restricted_modules
181-
)
178+
and fast_moe_config.fast_moe is not None
179+
and peft_config.target_modules != accepted_layers
182180
):
183-
raise ValueError(
184-
"`--fast_moe` with LoRA does not currently support `all-linear`, as "
185-
"target modules at this time. Please explicitly specify target "
186-
"modules when using `--fast_moe` with LoRA."
187-
)
188-
# If other common non-linear modules, raise warning
189-
if peft_config is not None and hasattr(peft_config, "target_modules"):
190181
logger.warning(
191182
"You are running lora with the ScatterMoE plugin, please note that "
192183
"passing target modules that are part of the moe module can cause unexpected "

0 commit comments

Comments
 (0)