Skip to content

Releases: withcatai/node-llama-cpp

v3.8.0

17 May 22:11
f2cb873
Compare
Choose a tag to compare

3.8.0 (2025-05-17)

Features

Bug Fixes

  • adapt to breaking llama.cpp changes (#460) (f2cb873)
  • capture multi-token segment separators (#460) (f2cb873)
  • race condition when reading extremely long gguf metadata (#460) (f2cb873)
  • adapt memory estimation to newly added model architectures (#460) (f2cb873)
  • skip binary testing on certain problematic conditions (#460) (f2cb873)
  • improve GPU backend loading error description (#460) (f2cb873)

Shipped with llama.cpp release b5414

To use the latest llama.cpp release available, run npx -n node-llama-cpp source download --release latest. (learn more)

v3.7.0

28 Mar 01:07
c070e81
Compare
Choose a tag to compare

3.7.0 (2025-03-28)

Features

  • extract function calling syntax from a Jinja template (#444) (c070e81)
  • Full support for Qwen and QwQ via QwenChatWrapper (#444) (c070e81)
  • export a llama instance getter on a model instance (#444) (c070e81)

Bug Fixes

  • better handling for function calling with empty parameters (#444) (c070e81)
  • reranking edge case crash (#444) (c070e81)
  • limit the context size by default in the node-typescript template (#444) (c070e81)
  • adapt to breaking llama.cpp changes (#444) (c070e81)
  • bump min nodejs version to 20 due to dependencies' requirements (#444) (c070e81)
  • defineChatSessionFunction type (#444) (c070e81)

Shipped with llama.cpp release b4980

To use the latest llama.cpp release available, run npx -n node-llama-cpp source download --release latest. (learn more)

v3.6.0

21 Feb 19:00
599a161
Compare
Choose a tag to compare

DeepSeek R1 is here!

Read about the release in the blog post


3.6.0 (2025-02-21)

Features

Bug Fixes


Shipped with llama.cpp release b4753

To use the latest llama.cpp release available, run npx -n node-llama-cpp source download --release latest. (learn more)

v3.5.0

31 Jan 01:09
63a1066
Compare
Choose a tag to compare

3.5.0 (2025-01-31)

Features

Bug Fixes

  • add missing Jinja features for DeepSeek (#425) (6e4bf3d)

Shipped with llama.cpp release b4600

To use the latest llama.cpp release available, run npx -n node-llama-cpp source download --release latest. (learn more)

v3.4.3

30 Jan 22:52
6e4bf3d
Compare
Choose a tag to compare

3.4.3 (2025-01-30)

Bug Fixes


Shipped with llama.cpp release b4599

To use the latest llama.cpp release available, run npx -n node-llama-cpp source download --release latest. (learn more)

v3.4.2

27 Jan 19:17
314d7e8
Compare
Choose a tag to compare

3.4.2 (2025-01-27)

Bug Fixes


Shipped with llama.cpp release b4567

To use the latest llama.cpp release available, run npx -n node-llama-cpp source download --release latest. (learn more)

v3.4.1

23 Jan 19:30
86e1bee
Compare
Choose a tag to compare

3.4.1 (2025-01-23)

Bug Fixes


Shipped with llama.cpp release b4529

To use the latest llama.cpp release available, run npx -n node-llama-cpp source download --release latest. (learn more)

v3.4.0

08 Jan 00:33
d1b4416
Compare
Choose a tag to compare

3.4.0 (2025-01-08)

Features

Bug Fixes

  • check for Rosetta usage on macOS x64 when using the inspect gpu command (#405) (632a7bf)
  • detect running under Rosetta on Apple Silicone and show an error message instead of crashing (#405) (632a7bf)
  • switch from "nextTick" to "nextCycle" for the default batch dispatcher (#405) (632a7bf)
  • remove deprecated CLS token (#405) (632a7bf)
  • pipe error logs in inspect gpu command (#405) (632a7bf)

Shipped with llama.cpp release b4435

To use the latest llama.cpp release available, run npx -n node-llama-cpp source download --release latest. (learn more)

v3.3.2

27 Dec 22:18
e2c5c3f
Compare
Choose a tag to compare

3.3.2 (2024-12-27)

Bug Fixes


Shipped with llama.cpp release b4291

To use the latest llama.cpp release available, run npx -n node-llama-cpp source download --release latest. (learn more)

v3.3.1

09 Dec 00:33
6a54163
Compare
Choose a tag to compare

3.3.1 (2024-12-09)

Bug Fixes

  • align embedding input with WPM vocabulary type models (#393) (28c7984)

Shipped with llama.cpp release b4291

To use the latest llama.cpp release available, run npx -n node-llama-cpp source download --release latest. (learn more)