Skip to content

Commit 8ea459f

Browse files
authored
bump version to v0.12.3 (#4493)
1 parent 2539582 commit 8ea459f

3 files changed

Lines changed: 5 additions & 3 deletions

File tree

README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,6 +26,7 @@ ______________________________________________________________________
2626
<details open>
2727
<summary><b>2026</b></summary>
2828

29+
- \[2026/04\] The LMDeploy project on PyPI has reached its storage quota, so pre-built wheels for new releases cannot be uploaded for the time being. You can download packages from the [GitHub Releases](https://github.com/InternLM/lmdeploy/releases) page or install from source instead. We will update this notice when wheel uploads to PyPI resume. Affected versions: >=0.12.2
2930
- \[2026/02\] Support [Qwen3.5](https://huggingface.co/collections/Qwen/qwen35)
3031
- \[2026/02\] Support [vllm-project/llm-compressor](https://github.com/vllm-project/llm-compressor) 4bit symmetric/asymmetric quantization. Refer [here](./docs/en/quantization/llm_compressor.md) for detailed guide
3132

@@ -228,7 +229,7 @@ Since v0.3.0, the default prebuilt package is compiled on **CUDA 12**. Starting
228229
If you are using a GeForce RTX 50 series graphics card, please install the LMDeploy prebuilt package compiled with **CUDA 12.8** as follows:
229230

230231
```shell
231-
export LMDEPLOY_VERSION=0.12.2
232+
export LMDEPLOY_VERSION=0.12.3
232233
export PYTHON_VERSION=312
233234
pip install https://github.com/InternLM/lmdeploy/releases/download/v${LMDEPLOY_VERSION}/lmdeploy-${LMDEPLOY_VERSION}+cu128-cp${PYTHON_VERSION}-cp${PYTHON_VERSION}-manylinux2014_x86_64.whl --extra-index-url https://download.pytorch.org/whl/cu128
234235
```

README_zh-CN.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,6 +26,7 @@ ______________________________________________________________________
2626
<details open>
2727
<summary><b>2026</b></summary>
2828

29+
- \[2026/04\] 由于 LMDeploy 在 PyPI 上的项目存储配额已满,新版本目前无法上传预编译安装包(wheels)。用户可以通过 [GitHub Releases](https://github.com/InternLM/lmdeploy/releases) 页面下载安装包,或者通过源码安装等方式使用最新版本;预编译包恢复上传后我们会另行通知。受影响版本:>=0.12.2
2930
- \[2026/02\] 支持 [Qwen3.5](https://huggingface.co/collections/Qwen/qwen35)
3031
- \[2026/02\] 支持 [vllm-project/llm-compressor](https://github.com/vllm-project/llm-compressor) 4bit 对称和非对称量化。 具体操作指南详见[此处](./docs/zh_cn/quantization/llm_compressor.md)
3132

@@ -230,7 +231,7 @@ pip install lmdeploy
230231
若使用 GeForce RTX 50 系列显卡,请按照如下方式安装基于 **CUDA 12.8** 编译的 LMDeploy 预编译包。
231232

232233
```shell
233-
export LMDEPLOY_VERSION=0.12.2
234+
export LMDEPLOY_VERSION=0.12.3
234235
export PYTHON_VERSION=312
235236
pip install https://github.com/InternLM/lmdeploy/releases/download/v${LMDEPLOY_VERSION}/lmdeploy-${LMDEPLOY_VERSION}+cu128-cp${PYTHON_VERSION}-cp${PYTHON_VERSION}-manylinux2014_x86_64.whl --extra-index-url https://download.pytorch.org/whl/cu128
236237
```

lmdeploy/version.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# Copyright (c) OpenMMLab. All rights reserved.
22

3-
__version__ = '0.12.2'
3+
__version__ = '0.12.3'
44
short_version = __version__
55

66

0 commit comments

Comments
 (0)