Skip to content
Discussion options

You must be logged in to vote

Hi,

Looks like you are using the torch jit-model. It is inevitable that is has some warm-up during first inference.
If the inference takes a second, it is a rather lengthy file (i.e. 1 minute). You can warm-up the model on a shorter file.

You can compare the jit-model with the onnx model.
Also if the session is ended properly (see .reset_state), you do not need to recreate the model each time you run new inference.

Replies: 1 comment 3 replies

Comment options

You must be logged in to vote
3 replies
@hvassard
Comment options

@snakers4
Comment options

@hvassard
Comment options

Answer selected by snakers4
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
help wanted Extra attention is needed
2 participants
Converted from issue

This discussion was converted from issue #756 on February 18, 2026 19:06.