You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
vLLM CPU installation automatically builds from source with the following process:
152
+
153
+
### Independent Installation (Alternative)
154
+
155
+
If heimdall installation fails, you can install independently:
156
+
157
+
> 🔗 **Reference**: For the latest installation methods, check the [Official vLLM CPU Installation Documentation](https://docs.vllm.ai/en/stable/getting_started/installation/cpu.html).
133
158
134
-
### GPU Mode
135
-
```bash
136
-
pip install vllm
137
-
```
138
159
139
-
After independent installation, you can run benchmarks by executing the scripts directly.
140
160
141
161
## NUMA Configuration
142
162
@@ -198,7 +218,7 @@ Used datasets:
198
218
-**Format**: 4-bit quantized GGUF
199
219
200
220
### vLLM
201
-
-**CPU**: Meta-Llama-3-8B
221
+
-**CPU**: Meta-Llama-3-8B (v0.9.1)
202
222
-**GPU**: Meta-Llama-3-70B
203
223
-**Source**: Automatic download from Hugging Face Hub
204
224
@@ -212,7 +232,7 @@ Used datasets:
212
232
```
213
233
214
234
2.**Memory Shortage**
215
-
- Ensure sufficient RAM for model size
235
+
- Ensure sufficient RAM for model size (32GB+ for 8B models)
216
236
- Sufficient VRAM required for vLLM GPU mode
217
237
218
238
3.**Check NUMA Configuration**
@@ -223,12 +243,35 @@ Used datasets:
223
243
4.**Permission Issues**
224
244
- Permission settings required for perf commands
225
245
226
-
### vLLM Special Configuration
246
+
5.**uv Command Not Found**
247
+
```bash
248
+
# uv should be available as part of heimdall setup
0 commit comments