Skip to content

Commit c14f72d

Browse files
authored
readme : update hot topics
1 parent cc6cac0 commit c14f72d

File tree

1 file changed

+2
-7
lines changed

1 file changed

+2
-7
lines changed

README.md

Lines changed: 2 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -10,13 +10,8 @@ Inference of Meta's [LLaMA](https://arxiv.org/abs/2302.13971) model (and others)
1010

1111
### Hot topics
1212

13-
- Remove LLAMA_MAX_DEVICES and LLAMA_SUPPORTS_GPU_OFFLOAD: https://github.com/ggerganov/llama.cpp/pull/5240
14-
- Incoming backends: https://github.com/ggerganov/llama.cpp/discussions/5138
15-
- [SYCL backend](README-sycl.md) is ready (1/28/2024), support Linux/Windows in Intel GPUs (iGPU, Arc/Flex/Max series)
16-
- New SOTA quantized models, including pure 2-bits: https://huggingface.co/ikawrakow
17-
- Collecting Apple Silicon performance stats:
18-
- M-series: https://github.com/ggerganov/llama.cpp/discussions/4167
19-
- A-series: https://github.com/ggerganov/llama.cpp/discussions/4508
13+
- Support for Gemma models: https://github.com/ggerganov/llama.cpp/pull/5631
14+
- Non-linear quantization IQ4_NL: https://github.com/ggerganov/llama.cpp/pull/5590
2015
- Looking for contributions to improve and maintain the `server` example: https://github.com/ggerganov/llama.cpp/issues/4216
2116

2217
----

0 commit comments

Comments
 (0)