Skip to content

Commit 08392c8

Browse files
LoserCheemsCopilot
andauthored
Update flash_dmattn/flash_dmattn_interface.py
Co-authored-by: Copilot <[email protected]>
1 parent 3b1b402 commit 08392c8

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

flash_dmattn/flash_dmattn_interface.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -648,7 +648,7 @@ def flash_dmattn_func(
648648
shape ({batch_size|1}, {nheads|nheads_k|1}, {seqlen_q|1}, {seqlen_k|1}) to apply to the attention scores.
649649
If None, no mask is applied.
650650
attn_bias: torch.Tensor, optional. The attention bias float tensor of
651-
shape (batch_size, {nheads|nheads_k|1}, {seqlen_q|1}, {seqlen_k|1}) to add to the attention scores.
651+
shape ({batch_size|1}, {nheads|nheads_k|1}, {seqlen_q|1}, {seqlen_k|1}) to add to the attention scores.
652652
If None, no bias is applied.
653653
softmax_scale: float. The scaling of QK^T before applying softmax.
654654
Default to 1 / sqrt(headdim).

0 commit comments

Comments
 (0)