Skip to content

Conversation

@depaulmillz
Copy link
Contributor

This switches from using cuDriverGetVersion to using cudaDriverGetVersion to resolve the issue reported in Dao-AILab/flash-attention#2116.

@HydraQYH
Copy link
Contributor

HydraQYH commented Jan 5, 2026

Great. I think this is a better solution for Dao-AILab/flash-attention#2116. We are also updating flash attention in sgl-kernel(sgl-project/sglang#16032), and we might need this fix as well. I hope this PR can be merged as soon as possible.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants