You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Qualcomm AI Engine Direct - GA Static Phi-4-mini (#13179)
### Summary
- Support Phi-4-mini-instruct for static llama path
- add P-ROPE for phi-4-mini
- add EOS tok for Phi-4-mini
### Test plan
```
python examples/qualcomm/oss_scripts/llama/llama.py -b build-android -s $DEVICE -m SM8750 --prompt "I would like to learn python, could you teach me with a simple example?" --temperature 0 --model_mode hybrid --prefill_ar_len 32 --max_seq_len 128 --ptq 16a8w --decoder_model phi_4_mini --num_sharding 4
```
cc: @haowhsu-quic, @shewu-quic, @winskuo-quic, @cccclai
# TODO: Encountered the following error during runtime, so switched behavior for now.
1181
-
# Error: libc++abi: terminating due to uncaught exception of type std::runtime_error: Unsupported Normalizer type: NFC.
1182
-
data.pop("normalizer")
1186
+
# Error: libc++abi: terminating due to uncaught exception of type std::runtime_error: invert=true is not supported for Split PreTokenizer. Only invert=false is supported.
0 commit comments