Skip to content

Commit 9fafeb4

Browse files
authored
docs: update README with CUDA 13.0, Python 3.11-3.13, new GPU archs (#175)
1 parent 014c716 commit 9fafeb4

File tree

1 file changed

+5
-10
lines changed

1 file changed

+5
-10
lines changed

README.md

Lines changed: 5 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -34,33 +34,28 @@ Features:
3434
- [x] Export to ONNX model, inference by OnnxRuntime
3535

3636
*Latest News* 🔥
37+
- [2026/03] CUDA 13.0 support, PyTorch 2.10, Python 3.11-3.13
38+
- [2026/03] Support H100/H200 (sm_90), B200/B300 (sm_100), RTX 5090 (sm_120)
3739
- [2024/03] ONNX Models export API
3840
- [2024/01] Support [HQQ](https://github.com/mobiusml/hqq) algorithm
3941
- [2023/12] The first PyPi package released
4042

4143
## Installation
42-
Easy to install qllm from PyPi [cu124]
44+
Easy to install qllm from PyPi
4345

4446
`pip install qllm`
4547

4648

47-
Install from release package, CUDA-124 is supported.
48-
[py310,py311,py312] https://github.com/wejoncy/QLLM/releases
49+
Install from release package, CUDA 13.0 is supported.
50+
[py311, py312, py313] https://github.com/wejoncy/QLLM/releases
4951

5052
Build from Source
5153

5254
**Please set ENV EXCLUDE_EXTENTION_FOR_FAST_BUILD=1 for fast build**
5355

54-
If you are using CUDA-124
5556
```
5657
pip install git+https://github.com/wejoncy/QLLM.git --no-build-isolation
5758
```
58-
OR CUDA-118/121
59-
```
60-
git clone https://github.com/wejoncy/QLLM.git
61-
cd QLLM
62-
python setup.py install
63-
```
6459

6560
# How to use it
6661

0 commit comments

Comments
 (0)