@@ -442,7 +442,7 @@ jobs:
442442
443443 # Compile only as weight sharing is not applicable on x86
444444 test-static-llama-size-qnn-linux :
445- name : test-static-llama-runner -qnn-linux
445+ name : test-static-llama-size -qnn-linux
446446 uses : pytorch/test-infra/.github/workflows/linux_job_v2.yml@main
447447 permissions :
448448 id-token : write
@@ -460,22 +460,27 @@ jobs:
460460 CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
461461 conda activate "${CONDA_ENV}"
462462
463+ BUILD_TOOL="cmake"
464+
463465 PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh
464466 PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh
465467
466468 # Setup executorch
467469 PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
468470
471+ # Setup install_requirements for llama
472+ PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh
473+
469474 # Retrieve 110M Stories Llama Artifacts
475+ PYTHON_EXECUTABLE=python . .ci/scripts/utils.sh
470476 PYTHON_EXECUTABLE=python download_stories_model_artifacts
471- $PYTHON_EXECUTABLE -m extension.llm.tokenizer.tokenizer -t tokenizer.model -o tokenizer.bin
472477
473478 # Test static llama stories110m pte size
474- PYTHON_EXECUTABLE= python backends/qualcomm/tests/test_qnn_delegate.py -k TestExampleScript.test_stories_single_llama --model SM8650 --build_folder build-android/ --executorch_root . --artifact_dir . --compile_only"
479+ PYTHONPATH="${PWD}" python backends/qualcomm/tests/test_qnn_delegate.py -k TestExampleScript.test_stories_single_llama --model SM8650 --build_folder build-android/ --executorch_root . --artifact_dir . --compile_only"
475480
476481 # Checks accuracy with weight sharing disabled since x86 does not support weight sharing.
477482 test-static-llama-accuracy-qnn-linux :
478- name : test-static-llama-runner -qnn-linux
483+ name : test-static-llama-accuracy -qnn-linux
479484 uses : pytorch/test-infra/.github/workflows/linux_job.yml@main
480485 strategy :
481486 fail-fast : false
@@ -490,18 +495,23 @@ jobs:
490495 CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
491496 conda activate "${CONDA_ENV}"
492497
498+ BUILD_TOOL="cmake"
499+
493500 PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh
494501 PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh
495502
496503 # Setup executorch
497504 PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
498505
506+ # Setup install_requirements for llama
507+ PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh
508+
499509 # Retrieve 110M Stories Llama Artifacts
510+ PYTHON_EXECUTABLE=python . .ci/scripts/utils.sh
500511 PYTHON_EXECUTABLE=python download_stories_model_artifacts
501- $PYTHON_EXECUTABLE -m extension.llm.tokenizer.tokenizer -t tokenizer.model -o tokenizer.bin
502512
503513 # Test static llama stories110m accuracy
504- PYTHON_EXECUTABLE= python backends/qualcomm/tests/test_qnn_delegate.py -k TestExampleScript.test_stories_single_llama --model SM8650 --build_folder build-x86_64/ --executorch_root . --artifact_dir . --enable_x86_64"
514+ PYTHONPATH="${PWD}" python backends/qualcomm/tests/test_qnn_delegate.py -k TestExampleScript.test_stories_single_llama --model SM8650 --build_folder build-x86_64/ --executorch_root . --artifact_dir . --enable_x86_64"
505515
506516 test-qnn-models-linux :
507517 name : test-qnn-models-linux
0 commit comments