Skip to content

Commit 0ec5b62

Browse files
committed
more build.md updates
1 parent bd8c5a8 commit 0ec5b62

File tree

1 file changed

+15
-11
lines changed

1 file changed

+15
-11
lines changed

docs/build.md

Lines changed: 15 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -63,11 +63,11 @@ When built with Metal support, you can explicitly disable GPU inference with the
6363

6464
Building the program with BLAS support may lead to some performance improvements in prompt processing using batch sizes higher than 32 (the default is 512). Using BLAS doesn't affect the generation performance. There are currently several different BLAS implementations available for build and use:
6565
66-
### Accelerate Framework:
66+
### Accelerate Framework
6767
6868
This is only available on Mac PCs and it's enabled by default. You can just build using the normal instructions.
6969

70-
### OpenBLAS:
70+
### OpenBLAS
7171

7272
This provides BLAS acceleration using only the CPU. Make sure to have OpenBLAS installed on your machine.
7373

@@ -82,15 +82,7 @@ This provides BLAS acceleration using only the CPU. Make sure to have OpenBLAS i
8282

8383
Check [BLIS.md](./backend/BLIS.md) for more information.
8484

85-
## SYCL
86-
87-
SYCL is a higher-level programming model to improve programming productivity on various hardware accelerators.
88-
89-
llama.cpp based on SYCL is used to **support Intel GPU** (Data Center Max series, Flex series, Arc series, Built-in GPU and iGPU).
90-
91-
For detailed info, please refer to [llama.cpp for SYCL](./backend/SYCL.md).
92-
93-
## Intel oneMKL
85+
### Intel oneMKL
9486

9587
Building through oneAPI compilers will make avx_vnni instruction set available for intel processors that do not support avx512 and avx512_vnni. Please note that this build config **does not support Intel GPU**. For Intel GPU support, please refer to [llama.cpp for SYCL](./backend/SYCL.md).
9688

@@ -107,6 +99,18 @@ Building through oneAPI compilers will make avx_vnni instruction set available f
10799

108100
Check [Optimizing and Running LLaMA2 on Intel® CPU](https://www.intel.com/content/www/us/en/content-details/791610/optimizing-and-running-llama2-on-intel-cpu.html) for more information.
109101

102+
### Other BLAS libraries
103+
104+
Any other BLAS library can be used by setting the `GGML_BLAS_VENDOR` option. See the [CMake documentation](https://cmake.org/cmake/help/latest/module/FindBLAS.html#blas-lapack-vendors) for a list of supported vendors.
105+
106+
## SYCL
107+
108+
SYCL is a higher-level programming model to improve programming productivity on various hardware accelerators.
109+
110+
llama.cpp based on SYCL is used to **support Intel GPU** (Data Center Max series, Flex series, Arc series, Built-in GPU and iGPU).
111+
112+
For detailed info, please refer to [llama.cpp for SYCL](./backend/SYCL.md).
113+
110114
## CUDA
111115

112116
This provides GPU acceleration using an NVIDIA GPU. Make sure to have the CUDA toolkit installed. You can download it from your Linux distro's package manager (e.g. `apt install nvidia-cuda-toolkit`) or from here: [CUDA Toolkit](https://developer.nvidia.com/cuda-downloads).

0 commit comments

Comments
 (0)