We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent a0a4dea commit b8fbb6cCopy full SHA for b8fbb6c
examples/models/llava/export_llava.py
@@ -216,6 +216,7 @@ def quant_embedding(model):
216
print("qval_copy", qval_copy)
217
print("qval", qval)
218
print("MATCHING", (qval_copy == qval).to(torch.float32).mean())
219
+ print("MAX DIFF", (qval_copy.to(torch.int32) - qval.to(torch.int32)).abs().max())
220
221
print("scale_copy", scale_copy)
222
print("scale", scale)
@@ -246,7 +247,7 @@ def quant_embedding(model):
246
247
248
print("token_embedding_ep_copy", token_embedding_ep_copy)
249
print("token_embedding_ep", token_embedding_ep)
- return token_embedding_ep
250
+ return token_embedding_ep_copy
251
252
253
def export_all(llava_model: LlavaModel):
0 commit comments