You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: smolvla.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,7 +16,7 @@ authors:
16
16
# SmolVLA: Efficient Vision-Language-Action Model trained on Lerobot Community Data
17
17
## 🧭TL;DR
18
18
Today, we introduce [SmolVLA](https://huggingface.co/lerobot/smolvla_base), a compact (450M), open-source Vision-Language-Action model for robotics that runs on consumer hardware.
19
-
- Pretrained only on open-source community-shared datasets under the [lerobot](https://huggingface.co/datasets?other=lerobot&sort=trending) tag.
19
+
- Pretrained only on compatibly licensed, open-source community-shared datasets under the [lerobot](https://huggingface.co/datasets?other=lerobot&sort=trending) tag.
20
20
- SmolVLA-450M outperforms much larger VLAs and strong baselines such as [ACT](https://huggingface.co/papers/2401.02117) on simulation (LIBERO, Meta-World) and real-world tasks ([SO100, SO101](https://github.com/TheRobotStudio/SO-ARM100)).
21
21
- Supports *asynchronous inference* for **30% faster response** and **2× task throughput**.
0 commit comments