Skip to content

Commit 74caa6b

Browse files
AlexanderDokuchaevNikita Savelyevdaniil-lyakhovl-batljaljushkin
authored
Post release 2.18.0 actions (#3654)
### Changes Bump OV version to 2025.3 Update docs Cherry-pick from release branch: - #3637 - #3634 - #3633 - #3629 ### Reason Changes from release branch ### Related tickets 172462 ### Tests https://github.com/openvinotoolkit/nncf/actions/runs/17545330049 https://github.com/openvinotoolkit/nncf/actions/runs/17545486898 --------- Co-authored-by: Nikita Savelyev <nikita.savelyev@intel.com> Co-authored-by: Daniil Lyakhov <daniil.lyakhov@intel.com> Co-authored-by: Liubov Talamanova <liubov.talamanova@intel.com> Co-authored-by: Lyalyushkin Nikolay <nikolay.lyalyushkin@intel.com> Co-authored-by: Andrey Churkin <andrey.churkin@intel.com> Co-authored-by: andreyanufr <andrey.anufriev@intel.com> Co-authored-by: Alexander Suslov <alexander.suslov@intel.com>
1 parent eeae24e commit 74caa6b

File tree

29 files changed

+84
-35
lines changed

29 files changed

+84
-35
lines changed

ReleaseNotes.md

Lines changed: 39 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,44 @@
11
# Release Notes
22

3+
## New in Release 2.18.0
4+
5+
Post-training Quantization:
6+
7+
- Features:
8+
- (OpenVINO) Introduced new compression data types CB4_F8E4M3 and CODEBOOK. CB4_F8E4M3 is a fixed codebook with 16 fp8 values based on NF4 data type values. CODEBOOK is an arbitrary user-selectable codebook that can be used to experiment with different data types. Both data types are used for weight compression. The AWQ and scale estimation algorithms are supported for these data types.
9+
- (OpenVINO) Added support for compressing FP8 (f8e4m3 and f8e5m2) weights to 4-bit data types, which is particularly beneficial for models like DeepSeek-R1.
10+
- Added `group_size_fallback_mode` parameter for advanced weight compression. It controls how nodes that do not support the default group size are handled. By default (`IGNORE`), such nodes are skipped. With `ERROR`, an exception is raised if the channel size is not divisible by the group size, while `ADJUST` attempts to modify the group size so it becomes valid.
11+
- (TorchFX) Added support for external quantizers in the `quantize_pt2e` API, including [XNNPACKQuantizer](https://docs.pytorch.org/executorch/stable/backends-xnnpack.html#quantization) and [CoreMLQuantizer](https://docs.pytorch.org/executorch/stable/backends-coreml.html#quantization). Users now can quantize their models in [ExecuTorch](https://github.com/pytorch/executorch) for the XNNPACK and CoreML backends via the nncf `quantize_pt2e` employing smooth quant, bias correction algorithms and a wide range of statistic collectors.
12+
- (ONNX) Added support for data-aware weight compression in the ONNX backend, including the AWQ and Scale Estimation algorithms. Provided an [example](https://github.com/openvinotoolkit/nncf/tree/develop/examples/llm_compression/onnx/tiny_llama_scale_estimation) demonstrating the data-aware weight compression pipeline using the `TinyLlama/TinyLlama-1.1B-Chat-v1.0` model in ONNX format.
13+
- Improvements:
14+
- Support of weight compression for models with the Rotary Positional Embedding block.
15+
- Support of weight compression for models with stateful self-attention blocks.
16+
- Tutorials:
17+
- [Post-Training Optimization of Qwen-Agent](https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/llm-agent-mcp/llm-agent-mcp.ipynb)
18+
- [Post-Training Optimization of FLUX.1 Kontext Model](https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/flux.1-kontext/flux.1-kontext.ipynb)
19+
- [Post-Training Optimization of Qwen3 Embedding Model](https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/qwen3-embedding/qwen3-embedding.ipynb)
20+
- [Post-Training Optimization of GLM-4.1V-9B-Thinking Model](https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/glm4.1-v-thinking/glm4.1-v-thinking.ipynb)
21+
22+
Compression-aware training:
23+
24+
- Features:
25+
- (PyTorch) Enhanced initialization for "QAT with absorbable LoRA" using advanced compression methods (AWQ + Scale Estimation). This improvement replaces the previous basic data-free compression approach, enabling QAT to start with a more accurate model baseline and achieve [superior final accuracy](https://github.com/openvinotoolkit/nncf/pull/3577).
26+
- Improvements:
27+
- (PyTorch) Streamlined "QAT with absorbable LoRA" by removing checkpoint selection based on validation set. This change significantly reduces overall tuning time and maximum allocated memory. While [the results on Wikitext](/examples/llm_compression/torch/distillation_qat_with_lora/README.md#results-on-wikitext) are slightly worse, it provides a more efficient and faster tuning pipeline (e.g. reduced from 32 minutes to 25 minutes for SmoLM-1.7B).
28+
- Tutorials:
29+
- (TorchFX) Added [example](examples/llm_compression/torch_fx/tiny_llama/README.md) for compression of TinnyLama-1.1B.
30+
- Updated [example](examples/llm_compression/onnx/tiny_llama/main.py) to meet NPU implementation.
31+
- Implemented fast evaluation and improved output in [example](examples/llm_compression/torch/downstream_qat_with_nls/README.md).
32+
33+
Deprecations/Removals:
34+
35+
- Removed examples that used `create_compressed_model` API.
36+
37+
Requirements:
38+
39+
- Updated PyTorch (2.8.0) and Torchvision (0.23.0) versions.
40+
- Set require `setuptools>=77` to build package.
41+
342
## New in Release 2.17.0
443

544
Post-training Quantization:

constraints.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# Openvino
2-
openvino==2025.2.0
2+
openvino==2025.3.0
33

44
# Pytorch
55
torch==2.8.0

docs/Installation.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,8 @@ as well as the supported versions of Python:
4949

5050
| NNCF | OpenVINO | PyTorch | ONNX | TensorFlow | Python |
5151
|-----------|------------|----------|----------|------------|--------|
52-
| `develop` | `2025.2.0` | `2.8.0` | `1.17.0` | `2.15.1` | `3.10` |
52+
| `develop` | `2025.3.0` | `2.8.0` | `1.17.0` | `2.15.1` | `3.10` |
53+
| `2.18.0` | `2025.3.0` | `2.8.0` | `1.17.0` | `2.15.1` | `3.10` |
5354
| `2.17.0` | `2025.2.0` | `2.7.1` | `1.17.0` | `2.15.1` | `3.10` |
5455
| `2.16.0` | `2025.1.0` | `2.6.0` | `1.17.0` | `2.15.1` | `3.10` |
5556
| `2.15.0` | `2025.0.0` | `2.5.1` | `1.17.0` | `2.15.1` | `3.10` |
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
transformers
2-
openvino==2025.2.0
2+
openvino==2025.3.0
33
optimum-intel[openvino]
44
git+https://github.com/onnx/onnx.git@c25eebcf51b781dbfcc75a9c8bdf5dd1781367fe # onnx-1.19.0.dev
55
onnxruntime==1.21.1

examples/llm_compression/onnx/tiny_llama_scale_estimation/requirements.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
torch==2.8.0
22
transformers>=4.48.0
3-
openvino==2025.2.0
3+
openvino==2025.3.0
44
optimum-intel[openvino]
55
onnx==1.17.0
66
onnxruntime==1.21.1
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
openvino==2025.2.0
1+
openvino==2025.3.0
22
optimum-intel[openvino]>=1.22.0
33
transformers>=4.48.0
44
onnx==1.17.0
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
datasets
2-
openvino==2025.2.0
2+
openvino==2025.3.0
33
optimum-intel[openvino]>=1.22.0
44
transformers>=4.48.0
55
onnx==1.17.0
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
transformers>=4.48.0
22
datasets==2.14.7
3-
openvino==2025.2.0
3+
openvino==2025.3.0
44
optimum-intel[openvino]>=1.22.0
55
onnx==1.17.0
Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
datasets
2-
whowhatbench @ git+https://github.com/openvinotoolkit/openvino.genai@2025.2.0.0#subdirectory=tools/who_what_benchmark
2+
whowhatbench @ git+https://github.com/openvinotoolkit/openvino.genai@2025.3.0.0#subdirectory=tools/who_what_benchmark
33
numpy>=1.23.5,<2
4-
openvino==2025.2.0
4+
openvino==2025.3.0
55
optimum-intel>=1.22.0
66
transformers>=4.48.0
77
onnx==1.17.0
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
torch==2.8.0
22
datasets==3.0.1
33
numpy>=1.23.5,<2
4-
openvino==2025.2.0
4+
openvino==2025.3.0
55
optimum-intel>=1.22.0
66
transformers>=4.48.0
77
onnx==1.17.0

0 commit comments

Comments
 (0)