Version 2.21.0
-
Bug fixes and Improvements
-
ONNX
- Fix IndexError when Conv or Linear layers are reused in the model (65c4b3b)
- Add optional argument
export_int32_biasto aimet-onnx export (3b8e0f0) - Unpin PyTorch version in aimet-onnx (d99b6c4)
- Align NaN handling with ORT CPU Execution Provider (e4c49eb)
- Fix quantization axis handling for transposed MatMul operations (6ca06d6)
-
PyTorch
- Fix quantization logic to enable input quantizers for layers following ignored layers (
80fb4fe_)
- Fix quantization logic to enable input quantizers for layers following ignored layers (
-