GPT-2-LM-Head ONNX file input #7370
Unanswered
cayleyhamilton
asked this question in
Other Q&A
Replies: 1 comment
-
I suspect the shape of the model's input is incorrect. It should indeed be (batch_size, sequence_length). cc @neginraoof - can you please check? Also, this issue is better filed here https://github.com/onnx/models. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
https://github.com/onnx/models/tree/master/text/machine_comprehension/gpt-2
The ONNX file's input is int64[input1_dyanamic_axes_1, input1_dynamic_axes_2, input1_dynamic_axes_3]
The referring text says 'long tensor of dynamic shape (batch_size, sequence_length). This makes intuitive sense to me.
A tensor with a token-id for each batch element and word piece.
How do I interpret the 3 axes in the ONNX input?
How is the variable sequence length within a batch represented?
Thanks for your help, in advance.
Beta Was this translation helpful? Give feedback.
All reactions