|
85 | 85 | <div data-md-component="skip"> |
86 | 86 |
|
87 | 87 |
|
88 | | - <a href="#key-features" class="md-skip"> |
| 88 | + <a href="#tested-inference-engines-gpus-and-models" class="md-skip"> |
89 | 89 | Skip to content |
90 | 90 | </a> |
91 | 91 |
|
|
448 | 448 | <ul class="md-nav__list" data-md-component="toc" data-md-scrollfix> |
449 | 449 |
|
450 | 450 | <li class="md-nav__item"> |
451 | | - <a href="#key-features" class="md-nav__link"> |
| 451 | + <a href="#tested-inference-engines-gpus-and-models" class="md-nav__link"> |
452 | 452 | <span class="md-ellipsis"> |
453 | | - Key Features |
| 453 | + Tested Inference Engines, GPUs, and Models |
454 | 454 | </span> |
455 | 455 | </a> |
456 | 456 |
|
457 | 457 | </li> |
458 | 458 |
|
459 | 459 | <li class="md-nav__item"> |
460 | | - <a href="#supported-accelerators" class="md-nav__link"> |
| 460 | + <a href="#architecture" class="md-nav__link"> |
461 | 461 | <span class="md-ellipsis"> |
462 | | - Supported Accelerators |
463 | | - </span> |
464 | | - </a> |
465 | | - |
466 | | -</li> |
467 | | - |
468 | | - <li class="md-nav__item"> |
469 | | - <a href="#supported-models" class="md-nav__link"> |
470 | | - <span class="md-ellipsis"> |
471 | | - Supported Models |
472 | | - </span> |
473 | | - </a> |
474 | | - |
475 | | -</li> |
476 | | - |
477 | | - <li class="md-nav__item"> |
478 | | - <a href="#openai-compatible-apis" class="md-nav__link"> |
479 | | - <span class="md-ellipsis"> |
480 | | - OpenAI-Compatible APIs |
| 462 | + Architecture |
481 | 463 | </span> |
482 | 464 | </a> |
483 | 465 |
|
|
3562 | 3544 | <ul class="md-nav__list" data-md-component="toc" data-md-scrollfix> |
3563 | 3545 |
|
3564 | 3546 | <li class="md-nav__item"> |
3565 | | - <a href="#key-features" class="md-nav__link"> |
3566 | | - <span class="md-ellipsis"> |
3567 | | - Key Features |
3568 | | - </span> |
3569 | | - </a> |
3570 | | - |
3571 | | -</li> |
3572 | | - |
3573 | | - <li class="md-nav__item"> |
3574 | | - <a href="#supported-accelerators" class="md-nav__link"> |
3575 | | - <span class="md-ellipsis"> |
3576 | | - Supported Accelerators |
3577 | | - </span> |
3578 | | - </a> |
3579 | | - |
3580 | | -</li> |
3581 | | - |
3582 | | - <li class="md-nav__item"> |
3583 | | - <a href="#supported-models" class="md-nav__link"> |
| 3547 | + <a href="#tested-inference-engines-gpus-and-models" class="md-nav__link"> |
3584 | 3548 | <span class="md-ellipsis"> |
3585 | | - Supported Models |
| 3549 | + Tested Inference Engines, GPUs, and Models |
3586 | 3550 | </span> |
3587 | 3551 | </a> |
3588 | 3552 |
|
3589 | 3553 | </li> |
3590 | 3554 |
|
3591 | 3555 | <li class="md-nav__item"> |
3592 | | - <a href="#openai-compatible-apis" class="md-nav__link"> |
| 3556 | + <a href="#architecture" class="md-nav__link"> |
3593 | 3557 | <span class="md-ellipsis"> |
3594 | | - OpenAI-Compatible APIs |
| 3558 | + Architecture |
3595 | 3559 | </span> |
3596 | 3560 | </a> |
3597 | 3561 |
|
@@ -3655,54 +3619,44 @@ <h1>Overview</h1> |
3655 | 3619 | <a class="github-button" href="https://github.com/gpustack/gpustack/fork" data-show-count="true" data-icon="octicon-repo-forked" data-size="large" aria-label="Fork">Fork</a> |
3656 | 3620 | </p> |
3657 | 3621 |
|
3658 | | -<p>GPUStack is an open-source GPU cluster manager for running AI models.</p> |
3659 | | -<h3 id="key-features">Key Features</h3> |
| 3622 | +<p>GPUStack is an open-source GPU cluster manager designed for efficient AI model deployment. It lets you run models efficiently on your own GPU hardware by choosing the best inference engines, scheduling GPU resources, analyzing model architectures, and automatically configuring deployment parameters.</p> |
| 3623 | +<p>The following figure shows how GPUStack delivers improved inference throughput over the unoptimized vLLM baseline:</p> |
| 3624 | +<p><a class="glightbox" href="../assets/a100-throughput-comparison.png" data-type="image" data-width="auto" data-height="auto" data-desc-position="bottom"><img alt="a100-throughput-comparison" src="../assets/a100-throughput-comparison.png" /></a></p> |
| 3625 | +<p>For detailed benchmarking methods and results, visit our <a href="https://docs.gpustack.ai/latest/performance-lab/overview/">Inference Performance Lab</a>.</p> |
| 3626 | +<h2 id="tested-inference-engines-gpus-and-models">Tested Inference Engines, GPUs, and Models</h2> |
| 3627 | +<p>GPUStack uses a plug-in architecture that makes it easy to add new AI models, inference engines, and GPU hardware. We work closely with partners and the open-source community to test and optimize emerging models across different inference engines and GPUs. Below is the current list of supported inference engines, GPUs, and models, which will continue to expand over time.</p> |
| 3628 | +<p><strong>Tested Inference Engines:</strong></p> |
3660 | 3629 | <ul> |
3661 | | -<li><strong>High Performance:</strong> Optimized for high-throughput and low-latency inference.</li> |
3662 | | -<li><strong>GPU Cluster Management:</strong> Efficiently manage multiple GPU clusters across different providers, including Docker-based, Kubernetes, and cloud platforms such as DigitalOcean.</li> |
3663 | | -<li><strong>Broad GPU Compatibility:</strong> Seamless support for GPUs from various vendors.</li> |
3664 | | -<li><strong>Extensive Model Support:</strong> Supports a wide range of models, including LLMs, VLMs, image models, audio models, embedding models, and rerank models.</li> |
3665 | | -<li><strong>Flexible Inference Backends:</strong> Built-in support for fast inference engines such as vLLM and SGLang, with the ability to integrate custom backends.</li> |
3666 | | -<li><strong>Multi-Version Backend Support:</strong> Run multiple versions of inference backends concurrently to meet diverse runtime requirements.</li> |
3667 | | -<li><strong>Distributed Inference:</strong> Supports single-node and multi-node, multi-GPU inference, including heterogeneous GPUs across vendors and environments.</li> |
3668 | | -<li><strong>Scalable GPU Architecture:</strong> Easily scale by adding more GPUs, nodes, or clusters to your infrastructure.</li> |
3669 | | -<li><strong>Robust Model Stability:</strong> Ensures high availability through automatic failure recovery, multi-instance redundancy, and intelligent load balancing.</li> |
3670 | | -<li><strong>Intelligent Deployment Evaluation:</strong> Automatically assesses model resource requirements, backend and architecture compatibility, OS compatibility, and other deployment factors.</li> |
3671 | | -<li><strong>Automated Scheduling:</strong> Dynamically allocates models based on available resources.</li> |
3672 | | -<li><strong>OpenAI-Compatible APIs:</strong> Fully compatible with OpenAI API specifications for seamless integration.</li> |
3673 | | -<li><strong>User & API Key Management:</strong> Simplified management of users and API keys.</li> |
3674 | | -<li><strong>Real-Time GPU Monitoring:</strong> Monitor GPU performance and utilization in real time.</li> |
3675 | | -<li><strong>Token and Rate Metrics:</strong> Track token usage and API request rates.</li> |
| 3630 | +<li>vLLM</li> |
| 3631 | +<li>SGLang</li> |
| 3632 | +<li>TensorRT-LLM</li> |
| 3633 | +<li>MindIE</li> |
3676 | 3634 | </ul> |
3677 | | -<h2 id="supported-accelerators">Supported Accelerators</h2> |
3678 | | -<p>GPUStack supports a variety of General-Purpose Accelerators, including:</p> |
3679 | | -<ul class="task-list"> |
3680 | | -<li class="task-list-item"><label class="task-list-control"><input type="checkbox" disabled checked/><span class="task-list-indicator"></span></label> NVIDIA GPU</li> |
3681 | | -<li class="task-list-item"><label class="task-list-control"><input type="checkbox" disabled checked/><span class="task-list-indicator"></span></label> AMD GPU</li> |
3682 | | -<li class="task-list-item"><label class="task-list-control"><input type="checkbox" disabled checked/><span class="task-list-indicator"></span></label> Ascend NPU</li> |
3683 | | -<li class="task-list-item"><label class="task-list-control"><input type="checkbox" disabled checked/><span class="task-list-indicator"></span></label> Hygon DCU (Experimental)</li> |
3684 | | -<li class="task-list-item"><label class="task-list-control"><input type="checkbox" disabled checked/><span class="task-list-indicator"></span></label> MThreads GPU (Experimental)</li> |
3685 | | -<li class="task-list-item"><label class="task-list-control"><input type="checkbox" disabled checked/><span class="task-list-indicator"></span></label> Iluvatar GPU (Experimental)</li> |
3686 | | -<li class="task-list-item"><label class="task-list-control"><input type="checkbox" disabled checked/><span class="task-list-indicator"></span></label> MetaX GPU (Experimental)</li> |
3687 | | -<li class="task-list-item"><label class="task-list-control"><input type="checkbox" disabled checked/><span class="task-list-indicator"></span></label> Cambricon MLU (Experimental)</li> |
| 3635 | +<p><strong>Tested GPUs:</strong></p> |
| 3636 | +<ul> |
| 3637 | +<li>NVIDIA A100</li> |
| 3638 | +<li>NVIDIA H100/H200</li> |
| 3639 | +<li>Ascend 910B</li> |
| 3640 | +</ul> |
| 3641 | +<p><strong>Tuned Models:</strong></p> |
| 3642 | +<ul> |
| 3643 | +<li>Qwen3</li> |
| 3644 | +<li>gpt-oss</li> |
| 3645 | +<li>GLM-4.5-Air</li> |
| 3646 | +<li>GLM-4.5/4.6</li> |
| 3647 | +<li>DeepSeek-R1</li> |
| 3648 | +</ul> |
| 3649 | +<h2 id="architecture">Architecture</h2> |
| 3650 | +<p>GPUStack enables development teams, IT organizations, and service providers to deliver Model-as-a-Service at scale. It supports industry-standard APIs for LLM, voice, image, and video models. The platform includes built-in user authentication and access control, real-time monitoring of GPU performance and utilization, and detailed metering of token usage and API request rates.</p> |
| 3651 | +<p>The figure below illustrates how a single GPUStack server can manage multiple GPU clusters across both on-premises and cloud environments. The GPUStack scheduler allocates GPUs to maximize resource utilization and selects the appropriate inference engines for optimal performance. Administrators also gain full visibility into system health and metrics through integrated Grafana and Prometheus dashboards.</p> |
| 3652 | +<p><a class="glightbox" href="../assets/gpustack-v2-architecture.png" data-type="image" data-width="auto" data-height="auto" data-desc-position="bottom"><img alt="gpustack-v2-architecture" src="../assets/gpustack-v2-architecture.png" /></a></p> |
| 3653 | +<p>GPUStack provides a powerful framework for deploying AI models. Its core features include:</p> |
| 3654 | +<ul> |
| 3655 | +<li><strong>Multi-Cluster GPU Management.</strong> Manages GPU clusters across multiple environments. This includes on-premises servers, Kubernetes clusters, and cloud providers.</li> |
| 3656 | +<li><strong>Pluggable Inference Engines.</strong> Automatically configures high-performance inference engines such as vLLM, SGLang, and TensorRT-LLM. You can also add custom inference engines as needed.</li> |
| 3657 | +<li><strong>Performance-Optimized Configurations.</strong> Offers pre-tuned modes for low latency or high throughput. GPUStack supports extended KV cache systems like LMCache and HiCache to reduce TTFT. It also includes built-in support for speculative decoding methods such as EAGLE3, MTP, and N-grams.</li> |
| 3658 | +<li><strong>Enterprise-Grade Operations.</strong> Offers support for automated failure recovery, load balancing, monitoring, authentication, and access control.</li> |
3688 | 3659 | </ul> |
3689 | | -<h2 id="supported-models">Supported Models</h2> |
3690 | | -<p>GPUStack uses <a href="https://github.com/vllm-project/vllm">vLLM</a>, <a href="https://github.com/sgl-project/sglang">SGLang</a>, <a href="https://www.hiascend.com/en/software/mindie">MindIE</a> and <a href="https://github.com/gpustack/vox-box">vox-box</a> as built-in inference backends, and it also supports any custom backend that can run in a container and expose a serving API. This allows GPUStack to work with a wide range of models.</p> |
3691 | | -<p>Models can come from the following sources:</p> |
3692 | | -<ol> |
3693 | | -<li> |
3694 | | -<p><a href="https://huggingface.co/">Hugging Face</a></p> |
3695 | | -</li> |
3696 | | -<li> |
3697 | | -<p><a href="https://modelscope.cn/">ModelScope</a></p> |
3698 | | -</li> |
3699 | | -<li> |
3700 | | -<p>Local File Path</p> |
3701 | | -</li> |
3702 | | -</ol> |
3703 | | -<p>For information on which models are supported by each built-in inference backend, please refer to the supported models section in the <a href="../user-guide/built-in-inference-backends/">Built-in Inference Backends</a> documentation.</p> |
3704 | | -<h2 id="openai-compatible-apis">OpenAI-Compatible APIs</h2> |
3705 | | -<p>GPUStack serves OpenAI compatible APIs. For details, please refer to <a href="../user-guide/openai-compatible-apis/">OpenAI Compatible APIs</a></p> |
3706 | 3660 |
|
3707 | 3661 |
|
3708 | 3662 |
|
|
0 commit comments