Skip to content

Conversation

@YASSERRMD
Copy link
Owner

Summary

  • support optional qkv bias for CodeGen models when converting HF checkpoints
  • expose attention.use_qkv_bias metadata and writer API
  • create zero bias when config expects bias but weights omit it
  • add unit test for CodeGen bias handling

Testing

  • flake8 gguf-py/gguf/constants.py gguf-py/gguf/gguf_writer.py convert_hf_to_gguf.py gguf-py/tests/test_codegen_bias.py
  • pytest gguf-py/tests/test_codegen_bias.py -k bias -q (skipped: torch not installed)

https://chatgpt.com/codex/tasks/task_e_687e677a4cb0832886d3a41f49a663b0

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants