Commit 68298e1
Add support to export XNNPACK based static_llama (#7535)
Summary:
Add support to export XNNPACK based static_llama
- static_llama is the QNN backend hybrid/prefill+decode model with KV cache as the inference input
- https://www.internalfb.com/code/fbsource/fbcode/executorch/examples/qualcomm/oss_scripts/llama2/model/static_llama.py
Reviewed By: tarun292
Differential Revision: D678671901 parent 68c0208 commit 68298e1
1 file changed
+8
-1
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
79 | 79 | | |
80 | 80 | | |
81 | 81 | | |
82 | | - | |
| 82 | + | |
| 83 | + | |
| 84 | + | |
| 85 | + | |
| 86 | + | |
| 87 | + | |
| 88 | + | |
| 89 | + | |
83 | 90 | | |
84 | 91 | | |
85 | 92 | | |
| |||
0 commit comments