Skip to content
Discussion options

You must be logged in to vote

Hi @AakashKT

There's a feature called KernelHistory which allows us to get the actual runtimes of the CUDA/Optix kernels. I've given a brief explanation on how to use it in this comment.

If the kernel runtimes are similiar for the two mi.render calls then the only difference is the time it takes to trace through the Python integrator to build the JIT graph.(Note: there should typically be two kernels per mi.render call)
If the kernel runtimes differ widely, it most likely has to do with the nature of your integrators. The recommended steps to profile this any deeper would be to remove parts of you integrators to understand the runtime costs that they entail.

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@AakashKT
Comment options

Answer selected by AakashKT
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants