Skip to content

Conversation

makaveli10
Copy link

Convert 1bitLLM/bitnet_b1_58-large to TQ2_0:

python convert_hf_to_gguf.py 1bitLLM/bitnet_b1_58-large --remote --outtype f32
./build_vulkan/bin/llama-quantize 1bitLLM-bitnet_b1_58-large-f32.gguf TQ2_0

Run inference on vulkan with:

./build_vulkan/bin/llama-cli -m ggml-model-TQ2_0.gguf -if -p "Hello" -ngl 999

makaveli10 and others added 23 commits July 30, 2025 16:58
This fixes the vkDeviceLostError on Mali
@makaveli10 makaveli10 changed the title Integrate TQ2_0 into vulkan Draft: Integrate TQ2_0 into vulkan Sep 29, 2025
@zoq zoq mentioned this pull request Oct 9, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants