Skip to content

Commit 7daae3b

Browse files
committed
chore: Fixed MX vocab errors
Signed-off-by: Brandon Groth <[email protected]>
1 parent d05401f commit 7daae3b

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

examples/MX/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# `microscaling` Examples Using a Toy Model and Direct Quantization (DQ)
22
Here, we provide two simple examples of using MX format in `fms-mo`.
3-
"MX format", such as `MXFP8`, is a different format compared to typical IEEE formats, e.g. PyTorch FP8s (`e4m3` or `e5m2`, see our other [FP8 example](../FP8_QUANT/README.md).) Mainly all the `mx` format are group-based where each member of the group is using the specified format, e.g. FP8 for MXFP8 while each group has a shared (usualy 8-bit) "scale". Group size could be as small as 32 or 16, depending on hardware design.
3+
"MX format", such as `MXFP8`, is a different format compared to typical IEEE formats, e.g. PyTorch FP8s (`e4m3` or `e5m2`, see our other [FP8 example](../FP8_QUANT/README.md).) Mainly all the `mx` format are group-based where each member of the group is using the specified format, e.g. FP8 for MXFP8 while each group has a shared (usually 8-bit) "scale". Group size could be as small as 32 or 16, depending on hardware design.
44
> [!NOTE]
55
It is important to keep in mind that `mx` is not natively supported by Hopper GPUs yet (some will be supported by Blackwell), which means the quantization configurations and corresponding behavior are simulated, i.e. no real "speed up" should be expected.
66

@@ -23,7 +23,7 @@ Expected output includes:
2323

2424
```
2525

26-
The second example is the same as in the [DQ](../DQ_SQ/README.md) folder, except using [microscaling](https://arxiv.org/abs/2310.10537) format. We demonstrate the effect of MXINT8, MXFP8, MXFP6, MXFP4 for weights, activations, and/or KV-cache.
26+
The second example is the same as in the [DQ](../DQ_SQ/README.md) folder, except using [microxcaling](https://arxiv.org/abs/2310.10537) format. We demonstrate the effect of MXINT8, MXFP8, MXFP6, MXFP4 for weights, activations, and/or KV-cache.
2727

2828
**1. Prepare Data** for calibration process by converting into its tokenized form. An example of tokenization using `LLAMA-3-8B`'s tokenizer is below.
2929

patches/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ To make a git diff patch file, first make your desired changes to the repository
1515
```
1616
git diff > <package>.patch
1717
```
18-
Packages may include files that differ by whitespaces even if you didn't change them.
18+
Packages may include files that differ by white spaces even if you didn't change them.
1919
To address this, add `--ignore-all-spaces` to the `git diff` command.
2020

2121
To test the patch file, copy the `<package>.patch` file to `fms-model-optimizer/patches`.

0 commit comments

Comments
 (0)