Skip to content

Commit a0ca5b4

Browse files
authored
release: update xllm release version to v0.6.1. (#297)
1 parent 8c3960e commit a0ca5b4

File tree

7 files changed

+44
-33
lines changed

7 files changed

+44
-33
lines changed

README.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -112,22 +112,22 @@ Supported models list:
112112
First, download the image we provide:
113113
```bash
114114
# A2 x86
115-
docker pull xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts
115+
docker pull xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86
116116
# A2 arm
117-
docker pull xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts-aarch64
117+
docker pull xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-arm
118118
# A3 arm
119-
docker pull xllm/xllm-ai:xllm-0.6.0-dev-hc-rc2-py3.11-oe24.03-lts-aarch64
119+
docker pull xllm/xllm-ai:xllm-0.6.1-dev-hc-rc2-arm
120120
# or
121121
# A2 x86
122-
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts
122+
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86
123123
# A2 arm
124-
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts-aarch64
124+
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-arm
125125
# A3 arm
126-
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hc-rc2-py3.11-oe24.03-lts-aarch64
126+
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hc-rc2-arm
127127
```
128128
Then create the corresponding container:
129129
```bash
130-
sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts
130+
sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ -v /etc/hccn.conf:/etc/hccn.conf xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86
131131
```
132132

133133
Install official repo and submodules:

README_zh.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -112,22 +112,22 @@ xLLM 提供了强大的智能计算能力,通过硬件系统的算力优化与
112112
首先下载我们提供的镜像:
113113
```bash
114114
# A2 x86
115-
docker pull xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts
115+
docker pull xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86
116116
# A2 arm
117-
docker pull xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts-aarch64
117+
docker pull xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-arm
118118
# A3 arm
119-
docker pull xllm/xllm-ai:xllm-0.6.0-dev-hc-rc2-py3.11-oe24.03-lts-aarch64
119+
docker pull xllm/xllm-ai:xllm-0.6.1-dev-hc-rc2-arm
120120
# 或者
121121
# A2 x86
122-
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts
122+
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86
123123
# A2 arm
124-
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts-aarch64
124+
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-arm
125125
# A3 arm
126-
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hc-rc2-py3.11-oe24.03-lts-aarch64
126+
docker pull quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hc-rc2-arm
127127
```
128128
然后创建对应的容器
129129
```bash
130-
sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts
130+
sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ -v /etc/hccn.conf:/etc/hccn.conf xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86
131131
```
132132

133133
下载官方仓库与模块依赖:

RELEASE.md

Lines changed: 11 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,14 @@
1+
# Release xllm 0.6.1
2+
3+
## **Major Features and Improvements**
4+
5+
### Bugfix
6+
7+
- Skip cancelled requests when processing stream output.
8+
- Resolve segmentation fault during qwen3 quantized inference.
9+
- Fix the alignment of monitoring metrics format for Prometheus.
10+
- Clear outdated tensors to save memory when loading model weights.
11+
112
# Release xllm 0.6.0
213

314
## **Major Features and Improvements**

cibuild/build.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ function error() {
66
exit 1
77
}
88

9-
IMAGE="quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts"
9+
IMAGE="quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86"
1010

1111
RUN_OPTS=(
1212
--rm

docs/en/getting_started/compile.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -3,28 +3,28 @@
33
## Container Environment Preparation
44
First, download the image we provide:
55
```bash
6-
docker pull xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts
6+
docker pull xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86
77
```
88
Then create the corresponding container:
99
```bash
10-
sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /mnt/cfs/9n-das-admin/llm_models:/mnt/cfs/9n-das-admin/llm_models -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts
10+
sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /mnt/cfs/9n-das-admin/llm_models:/mnt/cfs/9n-das-admin/llm_models -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ -v /etc/hccn.conf:/etc/hccn.conf xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86
1111
```
1212

1313
### Docker images
1414

1515
| Device | Arch | Images |
1616
|:---------:|:-----------:|:-------------:|
17-
| A2 | x86 | xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts |
18-
| A2 | arm | xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts-aarch64 |
19-
| A3 | arm | xllm/xllm-ai:xllm-0.6.0-dev-hc-rc2-py3.11-oe24.03-lts-aarch64 |
17+
| A2 | x86 | xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86 |
18+
| A2 | arm | xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-arm |
19+
| A3 | arm | xllm/xllm-ai:xllm-0.6.1-dev-hc-rc2-arm |
2020

2121
If you can't download it, you can use the following source instead:
2222

2323
| Device | Arch | Images |
2424
|:---------:|:-----------:|:-------------:|
25-
| A2 | x86 | quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts |
26-
| A2 | arm | quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts-aarch64 |
27-
| A3 | arm | quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hc-rc2-py3.11-oe24.03-lts-aarch64 |
25+
| A2 | x86 | quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86 |
26+
| A2 | arm | quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-arm |
27+
| A3 | arm | quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hc-rc2-arm |
2828

2929
## Installation
3030
After entering the container, download and compile using our [official repository](https://github.com/jd-opensource/xllm):

docs/zh/getting_started/compile.md

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -4,28 +4,28 @@
44
## 容器环境准备
55
首先下载我们提供的镜像:
66
```bash
7-
docker pull xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts
7+
docker pull xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86
88
```
99
然后创建对应的容器
1010
```bash
11-
sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /mnt/cfs/9n-das-admin/llm_models:/mnt/cfs/9n-das-admin/llm_models -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts
11+
sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /mnt/cfs/9n-das-admin/llm_models:/mnt/cfs/9n-das-admin/llm_models -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ -v /etc/hccn.conf:/etc/hccn.conf xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86
1212
```
1313

1414
### 镜像列表
1515

1616
| Device | Arch | Images |
1717
|:---------:|:-----------:|:-------------:|
18-
| A2 | x86 | xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts |
19-
| A2 | arm | xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts-aarch64 |
20-
| A3 | arm | xllm/xllm-ai:xllm-0.6.0-dev-hc-rc2-py3.11-oe24.03-lts-aarch64 |
18+
| A2 | x86 | xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86 |
19+
| A2 | arm | xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-arm |
20+
| A3 | arm | xllm/xllm-ai:xllm-0.6.1-dev-hc-rc2-arm |
2121

22-
如果下载不下来可以换用下面的源
22+
若无法下载可换用如下镜像源
2323

2424
| Device | Arch | Images |
2525
|:---------:|:-----------:|:-------------:|
26-
| A2 | x86 | quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts |
27-
| A2 | arm | quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hb-rc2-py3.11-oe24.03-lts-aarch64 |
28-
| A3 | arm | quay.io/jd_xllm/xllm-ai:xllm-0.6.0-dev-hc-rc2-py3.11-oe24.03-lts-aarch64 |
26+
| A2 | x86 | quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-x86 |
27+
| A2 | arm | quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hb-rc2-arm |
28+
| A3 | arm | quay.io/jd_xllm/xllm-ai:xllm-0.6.1-dev-hc-rc2-arm |
2929

3030

3131
## 安装

version.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
0.6.0
1+
0.6.1

0 commit comments

Comments
 (0)