-
Notifications
You must be signed in to change notification settings - Fork 26
Update benchmark_step.jl for CUDA benchmarking with useful kernel names
#4055
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
|
|
||
| - group: "Reproducibility infrastructure" | ||
| steps: | ||
|
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
These changes were made by a YAML auto-formatter in VS Code. Is there a style guide I might be breaking here?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm not sure... this is something I have been wondering as well. I considered following this example, which is used in Buildkite's docs.
|
@dennisYatunin @imreddyTeja thoughts on the timeouts here? Should I increase the limit or disable kernel renaming? |
What is the advantage of using kernel renaming in climaatmos-ci for buildkite steps that don't profile? Is the idea to run the profiler at the end of each simulation? Comparing the buildkite for this PR to the main's last buildkite run shows a ~40% slowdown, but everything after the first step doesn't seem to be affected. I'm not sure if that cost is worth it at the moment. |
I believe that was Dennis's vision, and then we'd process and summarize all of the profiling results together in a later step. |
This uses CliMA/ClimaCore.jl#2376 to provide more useful CUDA kernel names in benchmarks.
TODO
benchmark.jl, notbenchmark_step.jl?