v3.0.0-beta.37
Pre-release
Pre-release
3.0.0-beta.37 (2024-07-05)
Features
- flash attention (#264) (c2e322c)
- manually exclude specific GPU types from
gpu: "auto"
(#264) (c2e322c)
Shipped with llama.cpp
release b3322
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)