Skip to content

πŸ› [Bug] TensorRT-RTX: need to remove timing cacheΒ #3817

@lanluo-nvidia

Description

@lanluo-nvidia

Bug Description

Timing cache is used during build time for storing autotuning results in TRT Enterprise. As TRT-RTX does not use autotuning, [@hongyu Miao (HONGYUM)] removed the timing cache API from TRT-RTX.

Runtime cache is used in inference time to store and prevent repeated JIT compilation of kernels/graphs. This is a separate API. Thanks!

To Reproduce

Steps to reproduce the behavior:

Expected behavior

Environment

Build information about Torch-TensorRT can be found by turning on debug messages

  • Torch-TensorRT Version (e.g. 1.0.0):
  • PyTorch Version (e.g. 1.0):
  • CPU Architecture:
  • OS (e.g., Linux):
  • How you installed PyTorch (conda, pip, libtorch, source):
  • Build command you used (if compiling from source):
  • Are you using local sources or building from archives:
  • Python version:
  • CUDA version:
  • GPU models and configuration:
  • Any other relevant information:

Additional context

Metadata

Metadata

Assignees

Labels

bugSomething isn't working

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions