File tree Expand file tree Collapse file tree 1 file changed +4
-3
lines changed
Expand file tree Collapse file tree 1 file changed +4
-3
lines changed Original file line number Diff line number Diff line change @@ -14,9 +14,10 @@ Efficient Triton implementations for [Native Sparse Attention: Hardware-Aligned
1414
1515## News
1616
17- - [ 2025/02/25] Introduced an online top‑k selection kernel that avoids materializing the attention matrix during selection.
18- - [ 2025/02/24] Added support for a fused Triton kernel combining selected attention with sliding attention.
19- - [ 2025/02/21] Enabled handling of a variable number of selected blocks for queries across different positions and batches.
17+ - [ 2025-03-15] We now add full support for NSA training with ` flame ` .
18+ - [ 2025-02-25] Introduce an online top‑k selection kernel that avoids materializing the attention matrix during selection.
19+ - [ 2025-02-24] Add support for a fused Triton kernel combining selected attention with sliding attention.
20+ - [ 2025-02-21] Enable handling of a variable number of selected blocks for queries across different positions and batches.
2021
2122### Setup
2223
You can’t perform that action at this time.
0 commit comments