Skip to content

Commit aa698a2

Browse files
ai-edge-botniuchl
authored andcommitted
Remove alpha reference and update project status
LiteRT-PiperOrigin-RevId: 846879237
1 parent eaf7d63 commit aa698a2

File tree

2 files changed

+80
-25
lines changed

2 files changed

+80
-25
lines changed

README.md

Lines changed: 3 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -16,22 +16,12 @@ platforms, via efficient conversion, runtime, and optimization
1616
LiteRT continues the legacy of TensorFlow Lite as the trusted, high-performance
1717
runtime for on-device AI.
1818

19-
LiteRT V1 supports the TensorFlow Lite APIs, and is the recommended
20-
solution for existing apps using those older APIs.
21-
22-
LiteRT V2 (aka Next as announced at Google IO '25), introduces a new set of
23-
APIs, featuring advanced GPU/NPU acceleration, delivering superior performance,
24-
and making on-device ML inference easier than ever.
25-
26-
### 🚀 Status: General Availability (for V1) / Alpha (for V2)
27-
28-
- LiteRT V1 supports the TensorFlow Lite API, and its release status is
29-
[general availability](https://support.google.com/a/answer/11202276?hl=en).
30-
- LiteRT V2 is an alpha release and under active development.
19+
LiteRT features advanced GPU/NPU acceleration, delivers superior ML & GenAI
20+
performance, making on-device ML inference easier than ever.
3121

3222
### 🌟 What's New
3323

34-
- **🆕 New LiteRT v2 API**: Streamline development with automated accelerator
24+
- **🆕 New LiteRT Compiled Model API**: Streamline development with automated accelerator
3525
selection, true async execution, and efficient I/O buffer handling.
3626

3727
- Automated accelerator selection vs explicit delegate creation
@@ -157,17 +147,6 @@ deployment. Our product strategies are:
157147
- **Platform Support**: Enhancing support for core platforms and exploring new
158148
ones.
159149

160-
## 🗺 What's Next:
161-
162-
**V2 Beta by Dec 2025:**
163-
164-
- Achieve feature parity with TensorFlow Lite
165-
- Expand GPU backend support
166-
- Proactively increase ML and GenAI model coverage
167-
- Broader LiteRT Runtime/Converter upgrades from TensorFlow Lite
168-
169-
**V2 General Availability by Google IO, May 2026**
170-
171150
## 🙌 Contributing
172151

173152
We welcome contributions to LiteRT. Please see the
@@ -204,7 +183,6 @@ Check out these other projects from Google:
204183
- **[XNNPACK](https://github.com/google/XNNPACK)**: A highly optimized library
205184
of neural network inference operators for ARM, x86, and WebAssembly
206185
architectures that provides high-performance CPU acceleration for LiteRT.
207-
- **V2 GPU Delegate** - Coming soon
208186
- **[MediaPipe](https://github.com/google-ai-edge/mediapipe)**: A framework for
209187
building cross-platform, customizable ML solutions for live and streaming
210188
media.

RELEASE.md

Lines changed: 77 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,80 @@
1+
# Release 2.1.0
2+
3+
**Release 2.1.0 is the LiteRT** beta release.
4+
5+
LiteRT APIs are stable and have achieved feature parity. This milestone marks a significant step forward, introducing full feature parity with TensorFlow Lite, stable LiteRT APIs, and critical performance enhancements for GPU and NPU acceleration. With this release, we are officially recommending that developers begin their transition to LiteRT.
6+
7+
## Major Features and Improvements
8+
9+
### LiteRT Runtime
10+
11+
* Custom op is supported through [custom op dispatcher](https://github.com/google-ai-edge/LiteRT/blob/main/g3doc/apis/Custom_Op_Dispatcher.md).
12+
* CMake Build is supported in addition to Bazel
13+
* Released LiteRT C++ SDK using prebuilt libLiteRt.so file
14+
* Added Profiler API in CompiledModel
15+
* Added ErrorReporter API in CompiledModel
16+
* Added ResizeInputTensor API in CompiledModel
17+
18+
### LiteRT NPU
19+
20+
* Introduced LiteRT Accelerator Test Suite for coverage and regression testing
21+
* Introduced LiteRT graph transformation APIs for compiler plugins
22+
* Qualcomm
23+
* Added support for Qualcomm Snapdragon Gen5
24+
* Added support for NPU JIT mode
25+
* LiteRT Op coverage improvements
26+
* MediaTek
27+
* Added support for NPU JIT mode
28+
* LiteRT Op coverage improvements
29+
30+
### LiteRT GPU
31+
32+
* Increased GPU coverage with WebGPU/Dawn and OpenCL including Android, Linux, MacOS, Windows, iOS, IoT devices
33+
* Added asynchronous execution to Metal, WebGPU backends
34+
* Improved performance and memory footprint
35+
* Added an option to control GPU inference priority
36+
* Better error handling (without crashing) on Delegation errors
37+
38+
### LLM Support
39+
40+
* Provided Desktop GPU backends prebuilt for Linux (x64, arm64), MacOS (arm64), Windows (x64)
41+
* Improved memory utilization when executing on GPUs
42+
* Published new LLMs on [https://huggingface.co/litert-community](https://huggingface.co/litert-community)
43+
* litert-community/FastVLM-0.5B
44+
* litert-community/Qwen3-0.6B
45+
* litert-community/embeddinggemma-300m with new NPU precompiled models
46+
* litert-community/gemma-3-270m-it with new NPU precompiled model
47+
* Published Function Gemma on [https://huggingface.co/google](https://huggingface.co/google)
48+
* google/functiongemma-270m-it
49+
50+
### LiteRT on Android
51+
52+
* Added Interpreter API (CPU only) in the Maven v2.1.0+ packages
53+
* Added [Instruction](https://ai.google.dev/edge/litert/next/android_cpp_sdk) to use pre-built CompiledModell C++ API from the Maven package.
54+
55+
## Bug Fixes and Other Changes
56+
57+
Fixes Android min SDK version and it’s 23 now.
58+
59+
LiteRT NPU: Fixes partition algorithm when the full model cannot be offloaded to NPU.
60+
61+
## Breaking Changes
62+
63+
* Removed direct C headers usage. Users no longer need to include C headers.
64+
* TensorBuffer::CreateManaged() requires Environment always.
65+
* All TensorBuffer creation requires Environment except HostMemory types.
66+
* LiteRT C++ constructors are hidden. All LiteRT C++ objects should be created by Create() methods.
67+
* Moved internal only C++ APIs(such as litert\_logging.h) to litert/cc/internal
68+
* Removed Tensor, Subgraph, Signature access from litert::Model. Instead users can access SimpleTensor, SimpleSignature.
69+
* The CompiledModel::Create() API no longer needs litert::Model. They can be created from filename, model buffers directly.
70+
* Users can access SimpleTensor and SimpleSignature from CompiledModel.
71+
* Annotation, Metrics APIs are removed from CompiledModel.
72+
* Removed individual OpaqueOptions creation. These OpaqueOptions objects are obtained by Options directly.
73+
* Options::GetCpuOptions()
74+
* Options::GetGpuOptions()
75+
* Options::GetRuntimeOptions()
76+
*
77+
178
# Release 2.0.2a1
279

380
## LiteRT

0 commit comments

Comments
 (0)