We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent febcb22 commit 54c6f2dCopy full SHA for 54c6f2d
examples/models/llava/export_llava.py
@@ -186,7 +186,7 @@ def quant_embedding(model):
186
packed=False,
187
).quantized_model()
188
189
- quantized_token_embed = quant_embedding(llava.model_.language_model.model)
+ quantized_token_embed = quant_embedding(llava.model_.model.language_model)
190
token_dim_1 = Dim("token_dim_1", min=2, max=llava.text_model_args.max_seq_len)
191
dynamic_shapes = [{1: token_dim_1}]
192
with torch.no_grad():
0 commit comments