what is the right way to run inference with GPU(cuda) accelaration? #8918
Unanswered
hunterchenghx
asked this question in
Other Q&A
Replies: 1 comment 1 reply
-
|
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I use a NGC pytorch docker container with onnxruntime-gpu installed.
My container environment is like: ubuntu18.04, python3.6.10, onnx1.7.0, onnxruntime1.8.1, cuda11.0.3, cudnn8.0.4
my code is like this
But when I did the inference, only CPU is working. My nvidia-smi didn't show GPU running.
Could anyone help me? Is my code wrong or sth. happened with my onnxruntime installation?
Beta Was this translation helpful? Give feedback.
All reactions