You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
**\[ English | [中文](README_zh.md) | [日本語](README_ja.md)\]**
19
17
20
-
**LLMC** is an off-the-shell tool designed for compressing LLM, leveraging state-of-the-art compression algorithms to enhance efficiency and reduce model size without compromising performance.
21
-
22
-
**English doc** is [here](https://llmc-en.readthedocs.io/en/latest/).
23
-
24
-
**Chinese doc** is [here](https://llmc-zhcn.readthedocs.io/en/latest/).
25
-
26
-
**Docker hub** is [here](https://hub.docker.com/r/llmcompression/llmc).
You can download the Docker image that can run llmc with the following command. Users in mainland China are recommended to use Alibaba Cloud Docker.
18
+
</div>
31
19
32
-
docker hub
20
+
**LLMC** is an off-the-shell tool designed for compressing LLM, leveraging state-of-the-art compression algorithms to enhance efficiency and reduce model size without compromising performance. You can download the Docker image that can run llmc with the following command. Users in mainland China are recommended to use Alibaba Cloud Docker.
-**May 12, 2025:** 🔥 We now fully support quantization for the **`Wan2.1`** series of video generation models and provide export of truly quantized **INT8/FP8** weights, compatible with the [lightx2v](https://github.com/ModelTC/lightx2v) inference framework. For details, please refer to the [lightx2v documentation](https://llmc-en.readthedocs.io/en/latest/backend/lightx2v.html).
0 commit comments