Skip to content

Conversation

@robbiemu
Copy link
Contributor

@robbiemu robbiemu commented Oct 18, 2024

SentencePiece by default uses BPE, which by default also uses a tokenizer.json. This does not have to be customized with tokens you cannot read from tokenizer.model, but as the BSC-LT/salamanda-7b and related models show, it can be. modified the LlamaModel class to augment the vocabulatry from the json file if it is present.

related issue: #9899

ggerganov and others added 7 commits October 18, 2024 07:32
add intel amx isa detection

add vnni kernel for gemv cases

add vnni and amx kernel support for block_q8_0

code cleanup

fix packing B issue

enable openmp

fine tune amx kernel

switch to aten parallel pattern

add error message for nested parallelism

code cleanup

add f16 support in ggml-amx

add amx kernels for QK_K quant formats: Q4_K, Q5_K, Q6_K and IQ4_XS

update CMakeList

update README

fix some compilation warning

fix compiler warning when amx is not enabled

minor change

ggml-ci

move ggml_amx_init from ggml.c to ggml-amx/mmq.cpp

ggml-ci

update CMakeLists with -mamx-tile, -mamx-int8 and -mamx-bf16

ggml-ci

add amx as an ggml-backend

update header file, the old path for immintrin.h has changed to ggml-cpu-impl.h

minor change

update CMakeLists.txt

minor change

apply weight prepacking in set_tensor method in ggml-backend

fix compile error

ggml-ci

minor change

ggml-ci

update CMakeLists.txt

ggml-ci

add march dependency

minor change

ggml-ci

change ggml_backend_buffer_is_host to return false for amx backend

ggml-ci

fix supports_op

use device reg for AMX backend

ggml-ci

minor change

ggml-ci

minor change

fix rebase

set .buffer_from_host_ptr to be false for AMX backend
…rg#9705)

* implemented missing SYCL event APIs

* sycl : Added device and backend reg interfaces

* Restructured ggml-sycl.cpp
* rpc : refactor backend

Use structs for RPC request/response messages

* rpc : refactor server
…l-org#9745)

* refactor llama_batch_get_one

* adapt all examples

* fix simple.cpp

* fix llama_bench

* fix

* fix context shifting

* free batch before return

* use common_batch_add, reuse llama_batch in loop

* null terminated seq_id list

* fix save-load-state example

* fix perplexity

* correct token pos in llama_batch_allocr
@github-actions github-actions bot added build Compilation issues android Issues specific to Android examples server ggml changes relating to the ggml tensor library for machine learning SYCL https://en.wikipedia.org/wiki/SYCL - GPU programming language labels Oct 19, 2024
icppWorld and others added 4 commits October 20, 2024 19:01
llama_cpp_canister allows you to run llama.cpp as a Smart Contract on the Internet Computer. The smart contract runs as WebAssembly in a so-called 'canister'.
Update the binding list by adding LM-Kit.NET (C# & VB.NET)
@robbiemu
Copy link
Contributor Author

robbiemu commented Oct 20, 2024

I have no idea why I am seeing so many files changed on this page right now...

git status
On branch llamacasuallm-sp-bpe
Your branch is up to date with 'origin/llamacasuallm-sp-bpe'.

nothing to commit, working tree clean

git --no-pager diff --name-only upstream/master HEAD
.gitignore
convert_hf_to_gguf.py

git diff origin/llamacasuallm-sp-bpe --name-only

git remote get-url upstream
https://github.com/ggerganov/llama.cpp.git

I've been syncing and rebasing to keep my PR fresh to master.. I think I must have force pushed instead of merging first the last time, I will recreate this

@robbiemu robbiemu closed this Oct 20, 2024
@robbiemu robbiemu deleted the llamacasuallm-sp-bpe branch October 20, 2024 22:57
@robbiemu robbiemu restored the llamacasuallm-sp-bpe branch October 20, 2024 22:57
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

android Issues specific to Android build Compilation issues examples ggml changes relating to the ggml tensor library for machine learning python python script changes server SYCL https://en.wikipedia.org/wiki/SYCL - GPU programming language

Projects

None yet

Development

Successfully merging this pull request may close these issues.

8 participants