[kernel] Recompilation optimization triggered by triton function para…#7481
[kernel] Recompilation optimization triggered by triton function para…#7481HarpsealCC wants to merge 1 commit intovllm-project:mainfrom
Conversation
Summary of ChangesHello, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request implements an optimization strategy for Triton kernels by modifying how certain parameters are handled. By removing Highlights
🧠 New Feature in Public Preview: You can now enable Memory to help Gemini Code Assist learn from your team's feedback. This makes future code reviews more consistent and personalized to your project's style. Click here to enable Memory in your admin console. Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for GitHub and other Google products, sign up here. Footnotes
|
There was a problem hiding this comment.
Code Review
This pull request optimizes Triton kernels to reduce recompilation by changing several compile-time constant parameters (tl.constexpr) to runtime arguments. This is a valid approach to improve performance by avoiding repeated kernel compilations.
While reviewing the changes, I found one area of concern in vllm_ascend/ops/triton/fused_gdn_gating.py where a dynamic calculation for BLK_BATCHES has been replaced with a hardcoded value. This could potentially lead to runtime errors or suboptimal performance under certain conditions. I have added a specific comment with a suggestion to revert this change.
Additionally, as per the repository's contribution guidelines, I am providing a suggested PR title and summary to better reflect the changes and adhere to the project's formatting standards.
Suggested PR Title:
[kernel][Misc] Optimize Triton kernels to reduce recompilationSuggested PR Summary:
### What this PR does / why we need it?
This PR optimizes Triton kernels (`chunk_gated_delta_rule_fwd_kernel_h_blockdim64` and `fused_gdn_gating_kernel`) by converting several parameters from compile-time constants (`tl.constexpr`) to runtime arguments. This change avoids frequent kernel recompilations when these parameters vary, which can reduce model loading times and improve first-token latency.
### Does this PR introduce _any_ user-facing change?
No, this is a backend performance optimization and does not introduce any user-facing changes.
### How was this patch tested?
CI passed with existing tests.9e1a610 to
5e27da4
Compare
|
👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:
If CI fails, you can run linting and testing checks locally according Contributing and Testing. |
8f87c59 to
a3fc72b
Compare
…meter optimization Signed-off-by: l30072083 <liuchengzhuo1@h-partners.com>
…meter optimization,
What this PR does / why we need it?
Some parameters of Triton operators are unnecessarily modified with the "constexpr" modifier. When these parameters change, recompilation is triggered, which significantly affects the model performance. Therefore, these parameters need to be rectified.
Does this PR introduce any user-facing change?
No
How was this patch tested?
Signed-off-by: HarpSealCC 844291270@qq.com