Skip to content

Conversation

@ngxson
Copy link
Collaborator

@ngxson ngxson commented Mar 21, 2025

Fix #12483

Some fine-tuned models have it, we should include it in the converted model to make sure it runs correctly.

While we can remove this as suggested in the issue, I think it's not worth the risk. We are not sure if the fine tuning code produces different token_embd and lm_head tensor or not.

@ngxson ngxson requested a review from ggerganov March 21, 2025 21:12
@github-actions github-actions bot added the python python script changes label Mar 21, 2025
@ngxson ngxson merged commit fbdfefe into ggml-org:master Mar 22, 2025
49 of 50 checks passed
Ivy233 pushed a commit to Ivy233/llama.cpp that referenced this pull request Mar 23, 2025
…-org#12506)

* llama : gemma3 : use output tensor if it exists in model weight

* also add to the llm_tensor_names
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

python python script changes

Projects

None yet

Development

Successfully merging this pull request may close these issues.

convert_hf_to_gguf.py: Can not map tensor 'lm_head.weight' on Gemma-3-12b-it

2 participants