Skip to content

Commit 3ebe17e

Browse files
authored
Merge pull request #699 from kvcache-ai/Atream-patch-1
Update DeepseekR1_V3_tutorial.md
2 parents 9650893 + 369f4d9 commit 3ebe17e

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

doc/en/DeepseekR1_V3_tutorial.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -160,7 +160,7 @@ is speed up which is inspiring. So our showcase makes use of this finding*
160160
### V0.2.2 longer context & FP8 kernel
161161
#### longer context
162162
To use this feature, [install flashinfer](https://github.com/flashinfer-ai/flashinfer) first.
163-
163+
Note: The latest MLA kernel in FlashInfer still has a few minor issues. They are continuously fixing them on the main branch. If you are using FlashInfer, please install it from the main source code.
164164

165165
If you want to use long context(longer than 20K) for prefill, enable the matrix absorption MLA during the prefill phase, which will significantly reduce the size of the kv cache. Modify yaml file like this:
166166
```

0 commit comments

Comments
 (0)