Commit a91eb31
Add support to export XNNPACK based static_llama
Summary:
Add support to export XNNPACK based static_llama
- static_llama is the QNN backend hybrid/prefill+decode model with KV cache as the inference input
- https://www.internalfb.com/code/fbsource/fbcode/executorch/examples/qualcomm/oss_scripts/llama2/model/static_llama.py
Differential Revision: D678671901 parent 68c0208 commit a91eb31
1 file changed
+2
-1
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
79 | 79 | | |
80 | 80 | | |
81 | 81 | | |
82 | | - | |
| 82 | + | |
83 | 83 | | |
84 | 84 | | |
85 | 85 | | |
| |||
649 | 649 | | |
650 | 650 | | |
651 | 651 | | |
| 652 | + | |
652 | 653 | | |
653 | 654 | | |
654 | 655 | | |
| |||
0 commit comments