Skip to content

Commit a0c2aae

Browse files
Merge pull request #85 from chichun-charlie-liu/main
fix: minor fix from last PR regarding py3.12 upgrades
2 parents 33fe3d0 + 94c8738 commit a0c2aae

File tree

2 files changed

+8
-7
lines changed

2 files changed

+8
-7
lines changed

README.md

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -36,9 +36,7 @@ FMS Model Optimizer is a framework for developing reduced precision neural netwo
3636
### Requirements
3737

3838
1. **🐧 Linux system with Nvidia GPU (V100/A100/H100)**
39-
2. Python 3.9 to Python 3.11
40-
41-
📋 Python 3.12 is currently not supported due to PyTorch Dynamo constraint
39+
2. Python 3.10 to Python 3.12
4240
3. CUDA >=12
4341

4442
*Optional packages based on optimization functionality required:*
@@ -47,9 +45,12 @@ FMS Model Optimizer is a framework for developing reduced precision neural netwo
4745
- [auto_gptq](https://pypi.org/project/auto-gptq/) or build from [source](https://github.com/AutoGPTQ/AutoGPTQ)
4846
- If you want to experiment with **INT8** deployment in [QAT](./examples/QAT_INT8/) and [PTQ](./examples/PTQ_INT8/) examples:
4947
- Nvidia GPU with compute capability > 8.0 (A100 family or higher)
50-
- [Ninja](https://ninja-build.org/)
51-
- Clone the [CUTLASS](https://github.com/NVIDIA/cutlass) repository
52-
- `PyTorch 2.3.1` (as newer version will cause issue for the custom CUDA kernel used in these examples)
48+
- Option 1:
49+
- [Ninja](https://ninja-build.org/)
50+
- Clone the [CUTLASS](https://github.com/NVIDIA/cutlass) repository
51+
- `PyTorch 2.3.1` (as newer version will cause issue for the custom CUDA kernel used in these examples)
52+
- Option 2:
53+
- use triton kernel included. But this kernel is currently not faster than FP16.
5354
- **FP8** is a reduced precision format like **INT8**:
5455
- Nvidia A100 family or higher
5556
- [llm-compressor](https://github.com/vllm-project/llm-compressor)

tox.ini

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -91,6 +91,6 @@ commands =
9191

9292
[gh]
9393
python =
94-
3.11 = 3.12-{unitcov}
94+
3.12 = 3.12-{unitcov}
9595
3.11 = 3.11-{unitcov}
9696
3.10 = 3.10-{unitcov}

0 commit comments

Comments
 (0)