Skip to content

Commit ed79ecd

Browse files
committed
Small updates to the docs
Signed-off-by: Martin Hickey <[email protected]>
1 parent e53751d commit ed79ecd

File tree

3 files changed

+8
-5
lines changed

3 files changed

+8
-5
lines changed

.spellcheck-en-custom.txt

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -50,6 +50,7 @@ lossy
5050
LSTM
5151
matmul
5252
matmuls
53+
matplotlib
5354
maxperCh
5455
maxpertoken
5556
Miniforge

README.md

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010

1111
## Introduction
1212

13-
FMS Model Optimizer is a framework for developing reduced precision neural network models. Quantization techniques, such as [quantization-aware-training (QAT)](https://arxiv.org/abs/2407.11062), [post-training quantization (PTQ)](https://arxiv.org/abs/2102.05426), and several other optimization techniques on popular deep learning workloads are supported.
13+
FMS Model Optimizer is a framework for developing reduced precision neural network models. [Quantization](https://www.ibm.com/think/topics/quantization) techniques, such as [quantization-aware-training (QAT)](https://arxiv.org/abs/2407.11062), [post-training quantization (PTQ)](https://arxiv.org/abs/2102.05426), and several other optimization techniques on popular deep learning workloads are supported.
1414

1515
## Highlights
1616

@@ -37,10 +37,11 @@ FMS Model Optimizer is a framework for developing reduced precision neural netwo
3737

3838
1. **🐧 Linux system with Nvidia GPU (V100/A100/H100)**
3939
2. Python 3.9 to Python 3.11
40+
4041
📋 Python 3.12 is currently not supported due to PyTorch Dynamo constraint
4142
3. CUDA >=12
4243

43-
*Optional packages based on optimization functionalities required:*
44+
*Optional packages based on optimization functionality required:*
4445

4546
- **GPTQ** is a popular compression method for LLMs:
4647
- [auto_gptq](https://pypi.org/project/auto-gptq/) or build from [source](https://github.com/AutoGPTQ/AutoGPTQ)
@@ -53,15 +54,16 @@ FMS Model Optimizer is a framework for developing reduced precision neural netwo
5354
- Nvidia H100 family or higher
5455
- [llm-compressor](https://github.com/vllm-project/llm-compressor)
5556
- To enable compute graph plotting function (mostly for troubleshooting purpose):
57+
- [matplotlib](https://matplotlib.org/)
5658
- [graphviz](https://graphviz.org/)
5759
- [pygraphviz](https://pygraphviz.github.io/)
5860

5961
> [!NOTE]
60-
> PyTorch version should be < 2.4 if you would like to experiment deployment with external INT8 kernel.
62+
> PyTorch version should be < 2.5 if you would like to experiment deployment with external INT8 kernel.
6163
6264
### Installation
6365

64-
We recommend using a Python virtual environment with Python 3.10+. Here is how to setup a virtual environment using [Python venv](https://docs.python.org/3/library/venv.html):
66+
We recommend using a Python virtual environment with Python 3.9+. Here is how to setup a virtual environment using [Python venv](https://docs.python.org/3/library/venv.html):
6567

6668
```
6769
python3 -m venv fms_mo_venv

examples/FP8_QUANT/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ This is an example of mature FP8, which under the hood leverages some functional
1010
## Requirements
1111

1212
- [FMS Model Optimizer requirements](../../README.md#requirements)
13-
- Nvidia A100 family or higher
13+
- Nvidia H100 family or higher
1414
- The [llm-compressor](https://github.com/vllm-project/llm-compressor) library can be installed using pip:
1515

1616
```bash

0 commit comments

Comments
 (0)