We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4098b72 commit 32d6692Copy full SHA for 32d6692
docs/source/index.md
@@ -43,7 +43,7 @@ vLLM is flexible and easy to use with:
43
- Tensor parallelism and pipeline parallelism support for distributed inference
44
- Streaming outputs
45
- OpenAI-compatible API server
46
-- Support NVIDIA GPUs, AMD CPUs and GPUs, Intel CPUs, Gaudi® accelerators and GPUs, PowerPC CPUs, TPU, and AWS Trainium and Inferentia Accelerators.
+- Support NVIDIA GPUs, AMD CPUs and GPUs, Intel CPUs, Gaudi® accelerators and GPUs, IBM Power CPUs, TPU, and AWS Trainium and Inferentia Accelerators.
47
- Prefix caching support
48
- Multi-lora support
49
0 commit comments