You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/source/en/quantization/torchao.md
+33Lines changed: 33 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -88,6 +88,39 @@ Some quantization methods are aliases (for example, `int8wo` is the commonly use
88
88
89
89
Refer to the official torchao documentation for a better understanding of the available quantization methods and the exhaustive list of configuration options available.
90
90
91
+
## Serializing and Deserializing quantized models
92
+
93
+
To serialize a quantized model in a given dtype, first load the model with the desired quantization dtype and then save it using the [`~ModelMixin.save_pretrained`] method.
94
+
95
+
```python
96
+
import torch
97
+
from diffusers import FluxTransformer2DModel, TorchAoConfig
0 commit comments