You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The compilation depends on [vcpkg](https://github.com/microsoft/vcpkg). The Docker image already includes VCPKG_ROOT preconfigured. If you want to manually set it up, you can:
156
-
```bash
157
-
git clone https://gitcode.com/xLLM-AI/vcpkg.git
158
-
cd vcpkg && git checkout ffc42e97c866ce9692f5c441394832b86548422c
159
-
export VCPKG_ROOT=/your/path/to/vcpkg
160
-
```
161
-
162
-
#### Compilation
163
-
When compiling, generate executable files `build/xllm/core/server/xllm` under `build/`:
164
-
```bash
165
-
python setup.py build
166
-
```
167
-
Or, compile directly using the following command to generate the whl package under `dist/`:
168
-
```bash
169
-
python setup.py bdist_wheel
170
-
```
171
-
172
-
#### Launch
173
-
Run the following command to start xLLM engine:
174
-
```bash
175
-
./build/xllm/core/server/xllm \ # launch xllm server
176
-
--model=/path/to/your/llm \ # model path(to replace with your own path)
177
-
--port=9977 \ # set service port to 9977
178
-
--max_memory_utilization 0.90 # set the maximal utilization of device memory
179
-
```
84
+
Please refer to [Quick Start](docs/en/getting_started/quick_start.md) for more details. Besides, please check the model support status at [Model Support List](docs/en/supported_models.md).
180
85
181
86
---
182
87
183
-
## 5. Contributing
88
+
## 4. Contributing
184
89
There are several ways you can contribute to xLLM:
185
90
186
91
1. Reporting Issues (Bugs & Errors)
@@ -200,14 +105,14 @@ If you have problems about development, please check our document: **[Document](
200
105
201
106
---
202
107
203
-
## 6. Community & Support
108
+
## 5. Community & Support
204
109
If you encounter any issues along the way, you are welcomed to submit reproducible steps and log snippets in the project's Issues area, or contact the xLLM Core team directly via your internal Slack. In addition, we have established official WeChat groups. You can access the following QR code to join. Welcome to contact us!
This project was made possible thanks to the following open-source projects:
213
118
-[ScaleLLM](https://github.com/vectorch-ai/ScaleLLM) - xLLM draws inspiration from ScaleLLM's graph construction method and references its runtime execution.
@@ -217,6 +122,7 @@ This project was made possible thanks to the following open-source projects:
217
122
-[safetensors](https://github.com/huggingface/safetensors) - xLLM relies on the C binding safetensors capability.
218
123
-[Partial JSON Parser](https://github.com/promplate/partial-json-parser) - Implement xLLM's C++ JSON parser with insights from Python and Go implementations.
219
124
-[concurrentqueue](https://github.com/cameron314/concurrentqueue) - A fast multi-producer, multi-consumer lock-free concurrent queue for C++11.
0 commit comments