Replies: 1 comment
-
You mean, instead of having 256 weights from the same row in a block of 256, we could have used 32 x 8 from 8 different consecutive rows? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi there! I much admire your work in this project.
One thing I’ve been wondering… I believe weights are already repacked to make MatMul more efficient for the ffn... now I don’t understand the code well enough… are we (or could we possibly) also interleaving weight of w1,w2,w3 into panels? And then quantize based on this panels structures instead of individual blocked weight matrixes?
Maybe this doesn’t make my sense at all.. but I’ve been thinking about it for a while now, and it seems to me this could also open other possibilities like selecting variable Bitrate for each panel. Or sorting the panels by importance (derived from imatrix), and only calculating the most important ones (like top 50%).
I apologize if some of this seems stupid, it probably is 🙈…
Beta Was this translation helpful? Give feedback.
All reactions