Running inference using CUDAExecutionProvider without cuDNN #18083
-
Hi all, In Python, using onnxruntime-gpu, is it possible to use the CUDAExecutionProvider without having the cuDNN library available? I try to not include cuDNN in our project, as it significantly increases loading times. I already tried setting different SessionOptions and options to the CUDAExecutionProvider, e.g., disabling optimizations performed by cuDNN, but that seems to be not enough. Does anyone have an idea or knows if that's possible at all? Maybe something similar to PyTorch, where there is a compilation flag to disable cuDNN? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
No, the CUDAExecutionProvider uses cuDNN for some of the kernels it supports. So there's no way to build the CUDA provider without it: |
Beta Was this translation helpful? Give feedback.
No, the CUDAExecutionProvider uses cuDNN for some of the kernels it supports. So there's no way to build the CUDA provider without it:
https://onnxruntime.ai/docs/build/eps.html#cuda