File tree Expand file tree Collapse file tree 1 file changed +2
-7
lines changed Expand file tree Collapse file tree 1 file changed +2
-7
lines changed Original file line number Diff line number Diff line change @@ -10,13 +10,8 @@ Inference of Meta's [LLaMA](https://arxiv.org/abs/2302.13971) model (and others)
10
10
11
11
### Hot topics
12
12
13
- - Remove LLAMA_MAX_DEVICES and LLAMA_SUPPORTS_GPU_OFFLOAD: https://github.com/ggerganov/llama.cpp/pull/5240
14
- - Incoming backends: https://github.com/ggerganov/llama.cpp/discussions/5138
15
- - [ SYCL backend] ( README-sycl.md ) is ready (1/28/2024), support Linux/Windows in Intel GPUs (iGPU, Arc/Flex/Max series)
16
- - New SOTA quantized models, including pure 2-bits: https://huggingface.co/ikawrakow
17
- - Collecting Apple Silicon performance stats:
18
- - M-series: https://github.com/ggerganov/llama.cpp/discussions/4167
19
- - A-series: https://github.com/ggerganov/llama.cpp/discussions/4508
13
+ - Support for Gemma models: https://github.com/ggerganov/llama.cpp/pull/5631
14
+ - Non-linear quantization IQ4_NL: https://github.com/ggerganov/llama.cpp/pull/5590
20
15
- Looking for contributions to improve and maintain the ` server ` example: https://github.com/ggerganov/llama.cpp/issues/4216
21
16
22
17
----
You can’t perform that action at this time.
0 commit comments