-
Beta Was this translation helpful? Give feedback.
Answered by
snakers4
Feb 18, 2026
Replies: 1 comment 3 replies
-
|
Hi, Looks like you are using the torch jit-model. It is inevitable that is has some warm-up during first inference. You can compare the jit-model with the onnx model. |
Beta Was this translation helpful? Give feedback.
3 replies
Answer selected by
snakers4
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment








Hi,
Looks like you are using the torch jit-model. It is inevitable that is has some warm-up during first inference.
If the inference takes a second, it is a rather lengthy file (i.e. 1 minute). You can warm-up the model on a shorter file.
You can compare the jit-model with the onnx model.
Also if the session is ended properly (see
.reset_state), you do not need to recreate the model each time you run new inference.