v3.0.0-beta.13
Pre-release
Pre-release
3.0.0-beta.13 (2024-03-03)
Bug Fixes
- adapt to
llama.cpp
breaking change (#175) (5a70576) - return user-defined llama tokens (#175) (5a70576)
Features
- gguf parser (#168) (bcaab4f)
- use the best compute layer available by default (#175) (5a70576)
- more guardrails to prevent loading an incompatible prebuilt binary (#175) (5a70576)
inspect
command (#175) (5a70576)GemmaChatWrapper
(#175) (5a70576)TemplateChatWrapper
(#175) (5a70576)
Shipped with llama.cpp
release b2329
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)