Skip to content

Commit 34ce279

Browse files
author
Max Andriychuk
committed
Fix spelling
1 parent bf2ea1d commit 34ce279

File tree

2 files changed

+5
-4
lines changed

2 files changed

+5
-4
lines changed

.github/actions/spelling/allow/terms.txt

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,7 @@ ICHEP
1717
IIT
1818
JIT'd
1919
Jacobians
20+
JMU
2021
Jurgaityt
2122
LHC
2223
LLMs

_posts/2025-14-07-activty-analysis-cuda.md renamed to _posts/2025-14-07-activity-analysis-cuda.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -14,25 +14,25 @@ tags: gsoc c++ clang root auto-differentiation
1414
Hi! I’m Maksym Andriichuk, a third-year student of JMU Wuerzburg studying Mathematics. I am exited to be a part of Clad team fo this year's Google Summer of Code.
1515

1616
### Project description
17-
My project focuses on removing atomic operations when differentiating CUDA kernels. When accessing gpu global memory inside of a gradinet of a kernel data races inevitably occur and atomic operation are used instead, due to how reverse mode differentiation works in Clad. However, in some cases we can guarantee that no data race occur which enables us to drop atomic operations and drastically speeds the execution time of the gradient.
17+
My project focuses on removing atomic operations when differentiating CUDA kernels. When accessing gpu global memory inside of a gradient of a kernel data races inevitably occur and atomic operation are used instead, due to how reverse mode differentiation works in Clad. However, in some cases we can guarantee that no data race occur which enables us to drop atomic operations and drastically speeds the execution time of the gradient.
1818

1919
### Project goals
2020
The main goals of this project are:
2121

2222
- Implement a mechanism to check whether data races occur in various scenarios.
2323

24-
- Compare Clad with other tools on benchmarks uncluding RSBench and LULESH.
24+
- Compare Clad with other tools on benchmarks including RSBench and LULESH.
2525

2626
### Implementation strategy
2727
- Solve minor CUDA-related issues to get familiar with the codebase.
2828

29-
- Implement series of visitors to distinguish between different types of scenarious where atomic operations could be dropped
29+
- Implement series of visitors to distinguish between different types of scenarios where atomic operations could be dropped
3030

3131
- Use the existing benchmarks to compare the speedup from the implemented analysis.
3232

3333
## Conclusion
3434

35-
By integrating an analysis for (CUDA) GPU kernels we aim to speedup the execution of the gradient by removing atomic operation where posiible. To declare success, we would compare Clad to the other AD tools using different benchmarks. I am exited to be a part of the Clad team this summer and can not wait to share my progress.
35+
By integrating an analysis for (CUDA) GPU kernels we aim to speedup the execution of the gradient by removing atomic operation where possible. To declare success, we would compare Clad to the other AD tools using different benchmarks. I am exited to be a part of the Clad team this summer and can not wait to share my progress.
3636

3737
### Related Links
3838

0 commit comments

Comments
 (0)