Skip to content

Conversation

@taronaeo
Copy link
Collaborator

This PR fixes the erroneous byteswap for v.head.ffn_up.bias tensor for Big-Endian systems. Please refer to #12863 for more information.

I have verified this change on both Little-Endian and Big-Endian systems and it is working as intended. For reference these are the test machines:

  1. Little-Endian: MacBook Air M3
  2. Big-Endian: IBM z15 Mainframe

@github-actions github-actions bot added examples python python script changes labels Apr 12, 2025
@ngxson
Copy link
Collaborator

ngxson commented Apr 12, 2025

If I understand correctly from your original issue, the problem one appears on the last tensor of the file, but not a particular tensor name. So I think this fix is not valid. Imagine in another model, the last tensor is another one, now you will end up with a gguf with 2 broken tensors.

@ngxson
Copy link
Collaborator

ngxson commented Apr 12, 2025

I'm closing this because it's obviously not a valid fix. Even with the same model, you may end up with different tensor order, which may still result in a broken gguf.

@ngxson ngxson closed this Apr 12, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

examples python python script changes

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants