Skip to content

Releases: JamePeng/llama-cpp-python

v0.3.14-cu128-AVX2-win-20250806

06 Aug 19:00

Choose a tag to compare

  1. Sync llama.cpp API 20250806
  2. Update vendor/llama.cpp to 0d8831543cdc368fb248bae6f1b4aa5516684edc
  3. Add strftime_now function into Jinja2ChatFormatter Class for gpt-oss and gemma3 chat_template need

v0.3.14-cu128-AVX2-linux-20250806

06 Aug 15:02

Choose a tag to compare

  1. Sync llama.cpp API 20250806
  2. Update vendor/llama.cpp to 0d8831543cdc368fb248bae6f1b4aa5516684edc
  3. Add strftime_now function into Jinja2ChatFormatter Class for gpt-oss and gemma3 chat_template need

v0.3.14-cu126-AVX2-win-20250806

06 Aug 18:01

Choose a tag to compare

  1. Sync llama.cpp API 20250806
  2. Update vendor/llama.cpp to 0d8831543cdc368fb248bae6f1b4aa5516684edc
  3. Add strftime_now function into Jinja2ChatFormatter Class for gpt-oss and gemma3 chat_template need

v0.3.14-cu126-AVX2-linux-20250806

06 Aug 14:51

Choose a tag to compare

  1. Sync llama.cpp API 20250806
  2. Update vendor/llama.cpp to 0d8831543cdc368fb248bae6f1b4aa5516684edc
  3. Add strftime_now function into Jinja2ChatFormatter Class for gpt-oss and gemma3 chat_template need

v0.3.14-cu124-AVX2-win-20250806

06 Aug 17:54

Choose a tag to compare

  1. Sync llama.cpp API 20250806
  2. Update vendor/llama.cpp to 0d8831543cdc368fb248bae6f1b4aa5516684edc
  3. Add strftime_now function into Jinja2ChatFormatter Class for gpt-oss and gemma3 chat_template need

v0.3.14-cu124-AVX2-linux-20250806

06 Aug 14:51

Choose a tag to compare

  1. Sync llama.cpp API 20250806
  2. Update vendor/llama.cpp to 0d8831543cdc368fb248bae6f1b4aa5516684edc
  3. Add strftime_now function into Jinja2ChatFormatter Class for gpt-oss and gemma3 chat_template need

v0.3.14-cu128-AVX2-win-20250801

01 Aug 17:24

Choose a tag to compare

  1. Compile with cuda12.8.1 for Blackwell architecture (sm_100 and sm_120) computing cards
  2. Sync llama.cpp API 20250801
  3. Remove sm_70 in cuda12.8.1 action flow

v0.3.14-cu128-AVX2-linux-20250801

01 Aug 14:34

Choose a tag to compare

  1. Compile with cuda12.8.1 for Blackwell architecture (sm_100 and sm_120) computing cards
  2. Sync llama.cpp API 20250801
  3. Remove sm_70 in cuda12.8.1 action flow

v0.3.14-cu126-AVX2-win-20250801

01 Aug 16:26

Choose a tag to compare

Sync llama.cpp API 20250801

v0.3.14-cu126-AVX2-linux-20250801

01 Aug 13:24

Choose a tag to compare

Sync llama.cpp API 20250801