We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 46375c2 commit 73c08f5Copy full SHA for 73c08f5
examples/models/llava/export_llava.py
@@ -210,6 +210,7 @@ def quant_embedding(model):
210
print("qval_copy", qval_copy)
211
print("qval", qval)
212
print("MATCHING", (qval_copy == qval).to(torch.float32).mean())
213
+ print("MAX DIFF", (qval_copy.to(torch.int32) - qval.to(torch.int32)).abs().max())
214
215
print("scale_copy", scale_copy)
216
print("scale", scale)
@@ -240,7 +241,7 @@ def quant_embedding(model):
240
241
242
print("token_embedding_ep_copy", token_embedding_ep_copy)
243
print("token_embedding_ep", token_embedding_ep)
- return token_embedding_ep
244
+ return token_embedding_ep_copy
245
246
247
def export_all(llava_model: LlavaModel):
0 commit comments