You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Built from: <https://github.com/EZForever/llama.cpp-static>
6
6
7
7
## Usage
8
8
9
-
Please refer to [llama.cpp docker guide](https://github.com/ggerganov/llama.cpp/blob/master/README.md#docker) and [server README](https://github.com/ggerganov/llama.cpp/tree/master/examples/server).
9
+
Please refer to [llama.cpp docker guide](https://github.com/ggml-org/llama.cpp/blob/master/docs/docker.md) and [server README](https://github.com/ggml-org/llama.cpp/blob/master/tools/server/README.md).
10
10
11
11
## Tag format
12
12
@@ -25,7 +25,7 @@ Server images are tagged in the format of `server-<ssl>-<avx>`.
25
25
-`avx`: (Only) AVX instruction set is enabled. Might be useful if you are using some old CPUs that don't support AVX2.
26
26
-`avx2`: AVX2 instruction set is enabled. This build should support most modern/recent CPUs with reasonable performance.
27
27
-`avx512`: AVX512 base instruction set is enabled. Currently only some high-end or server-grade CPUs support this instruction set, so check your hardware specs before using this build.
28
-
-`oneapi`: Experimental build with the Intel oneAPI compiler, inspired by ggerganov/llama.cpp#5067. Offers a ~30% speed increase (~20tok/s vs ~15tok/s) in prompt processing on my machine compared to `avx2` builds. Not updated daily.
28
+
-`oneapi`: Experimental build with the Intel oneAPI compiler, inspired by ggml-org/llama.cpp#5067. Offers a ~30% speed increase (~20tok/s vs ~15tok/s) in prompt processing on my machine compared to `avx2` builds. Not updated daily.
29
29
30
-
RPC server images are tagged in the format of `rpc-server-<ssl>-<avx>`. Refer to [rpc README](https://github.com/ggerganov/llama.cpp/tree/master/examples/rpc) for detailed information.
30
+
RPC server images are tagged in the format of `rpc-server-<ssl>-<avx>`. Refer to [rpc README](https://github.com/ggml-org/llama.cpp/blob/master/tools/rpc/README.md) for detailed information.
0 commit comments