-
Notifications
You must be signed in to change notification settings - Fork 22
Open
autowarefoundation/tvm_vendor
#14Description
This is a followup issue from Enable OpenCL Backend for TVM
We may want to bring up CUDA backend for TVM for two reason:
- Future benchmarking tasks. I was trying to compare the performance of TensorRT and TVM deployment. According to the paper, Ansor: Generating High-Performance Tensor Programs for Deep Learning, the evaluation indicates that optimized TVM code beats TensorRT for some models. It would make more sense to benchmark with CUDA backend.
- General performance reason. According to a performance comparison survey. opencl is 13% ~ 67% slower than cuda.
Reference
And the reason it was not done in the previous issue, Enable OpenCL Backend for TVM is that I was not able to compile Lidar CenterPoint models due to some errors and I did not have time to look into the details. However, to proceed development, I believe it is necessary to enable CUDA.
Regarding Comments from Ambroise, it is true that CUDA libraries need to be handled beforehand. I think we can try to detect the existence of CUDA components and patch tvm_vendor accordingly.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels