Skip to content

Commit 80bb1e8

Browse files
authored
Officially support SmolLM3 using the Transformers backend (#22665)
Signed-off-by: Harry Mellor <[email protected]>
1 parent d030b01 commit 80bb1e8

File tree

3 files changed

+10
-0
lines changed

3 files changed

+10
-0
lines changed

docs/models/supported_models.md

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -409,6 +409,12 @@ th {
409409
| `MiniMaxText01ForCausalLM` | MiniMax-Text | `MiniMaxAI/MiniMax-Text-01`, etc. | | | ✅︎ |
410410
| `Zamba2ForCausalLM` | Zamba2 | `Zyphra/Zamba2-7B-instruct`, `Zyphra/Zamba2-2.7B-instruct`, `Zyphra/Zamba2-1.2B-instruct`, etc. | | | ✅︎ |
411411

412+
Some models are supported only via the [Transformers backend](#transformers). The purpose of the table below is to acknowledge models which we officially support in this way. The logs will say that the Transformers backend is being used, and you will see no warning that this is fallback behaviour. This means that, if you have issues with any of the models listed below, please [make an issue](https://github.com/vllm-project/vllm/issues/new/choose) and we'll do our best to fix it!
413+
414+
| Architecture | Models | Example HF Models | [LoRA](../features/lora.md) | [PP](../serving/parallelism_scaling.md) | [V1](gh-issue:8779) |
415+
|--------------|--------|-------------------|----------------------|---------------------------|---------------------|
416+
| `SmolLM3ForCausalLM` | SmolLM3 | `HuggingFaceTB/SmolLM3-3B` | ✅︎ | ✅︎ | ✅︎ |
417+
412418
!!! note
413419
Currently, the ROCm version of vLLM supports Mistral and Mixtral only for context lengths up to 4096.
414420

tests/models/registry.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -291,6 +291,7 @@ def check_available_online(
291291
"Qwen3ForCausalLM": _HfExamplesInfo("Qwen/Qwen3-8B"),
292292
"Qwen3MoeForCausalLM": _HfExamplesInfo("Qwen/Qwen3-30B-A3B"),
293293
"RWForCausalLM": _HfExamplesInfo("tiiuae/falcon-40b"),
294+
"SmolLM3ForCausalLM": _HfExamplesInfo("HuggingFaceTB/SmolLM3-3B"),
294295
"StableLMEpochForCausalLM": _HfExamplesInfo("stabilityai/stablelm-zephyr-3b"), # noqa: E501
295296
"StableLmForCausalLM": _HfExamplesInfo("stabilityai/stablelm-3b-4e1t"),
296297
"Starcoder2ForCausalLM": _HfExamplesInfo("bigcode/starcoder2-3b"),

vllm/model_executor/models/registry.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -271,6 +271,9 @@
271271
}
272272

273273
_TRANSFORMERS_SUPPORTED_MODELS = {
274+
# Text generation models
275+
"SmolLM3ForCausalLM": ("transformers", "TransformersForCausalLM"),
276+
# Multimodal models
274277
"Emu3ForConditionalGeneration": ("transformers", "TransformersForMultimodalLM"), # noqa: E501
275278
}
276279

0 commit comments

Comments
 (0)