Skip to content

Commit 2f46579

Browse files
sooraj-satheeshgemini-code-assist[bot]bigPYJ1151
authored
[doc] Update x86 CPU-inference installation doc to reflect optionality of AVX512f (#22707)
Signed-off-by: Sooraj S <[email protected]> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> Co-authored-by: Li, Jiang <[email protected]>
1 parent 3a7e3bb commit 2f46579

File tree

1 file changed

+3
-2
lines changed

1 file changed

+3
-2
lines changed

docs/getting_started/installation/cpu/x86.inc.md

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ vLLM supports basic model inferencing and serving on x86 CPU platform, with data
66
# --8<-- [start:requirements]
77

88
- OS: Linux
9-
- CPU flags: `avx512f`, `avx512_bf16` (Optional), `avx512_vnni` (Optional)
9+
- CPU flags: `avx512f` (Recommended), `avx512_bf16` (Optional), `avx512_vnni` (Optional)
1010

1111
!!! tip
1212
Use `lscpu` to check the CPU flags.
@@ -28,7 +28,7 @@ vLLM supports basic model inferencing and serving on x86 CPU platform, with data
2828
[https://gallery.ecr.aws/q9t5s3a7/vllm-cpu-release-repo](https://gallery.ecr.aws/q9t5s3a7/vllm-cpu-release-repo)
2929

3030
!!! warning
31-
If deploying the pre-built images on machines only contain `avx512f`, `Illegal instruction` error may be raised. It is recommended to build images for these machines with `--build-arg VLLM_CPU_AVX512BF16=false` and `--build-arg VLLM_CPU_AVX512VNNI=false`.
31+
If deploying the pre-built images on machines without `avx512f`, `avx512_bf16`, or `avx512_vnni` support, an `Illegal instruction` error may be raised. It is recommended to build images for these machines with the appropriate build arguments (e.g., `--build-arg VLLM_CPU_DISABLE_AVX512=true`, `--build-arg VLLM_CPU_AVX512BF16=false`, or `--build-arg VLLM_CPU_AVX512VNNI=false`) to disable unsupported features. Please note that without `avx512f`, AVX2 will be used and this version is not recommended because it only has basic feature support.
3232

3333
# --8<-- [end:pre-built-images]
3434
# --8<-- [start:build-image-from-source]
@@ -37,6 +37,7 @@ vLLM supports basic model inferencing and serving on x86 CPU platform, with data
3737
docker build -f docker/Dockerfile.cpu \
3838
--build-arg VLLM_CPU_AVX512BF16=false (default)|true \
3939
--build-arg VLLM_CPU_AVX512VNNI=false (default)|true \
40+
--build-arg VLLM_CPU_DISABLE_AVX512=false (default)|true \
4041
--tag vllm-cpu-env \
4142
--target vllm-openai .
4243

0 commit comments

Comments
 (0)