@@ -131,7 +131,7 @@ jobs:
131131
132132 test-arm-backend-delegation :
133133 name : test-arm-backend-delegation
134- uses : pytorch/test-infra/.github/workflows/linux_job .yml@main
134+ uses : pytorch/test-infra/.github/workflows/linux_job_v2 .yml@main
135135 with :
136136 runner : linux.2xlarge
137137 docker-image : executorch-ubuntu-22.04-arm-sdk
@@ -157,7 +157,7 @@ jobs:
157157
158158 test-arm-reference-delegation :
159159 name : test-arm-reference-delegation
160- uses : pytorch/test-infra/.github/workflows/linux_job .yml@main
160+ uses : pytorch/test-infra/.github/workflows/linux_job_v2 .yml@main
161161 with :
162162 runner : linux.2xlarge
163163 docker-image : executorch-ubuntu-22.04-arm-sdk
@@ -290,7 +290,7 @@ jobs:
290290 # ${CONDA_RUN} python -m unittest examples.models.llava.test.test_llava
291291
292292 # # run e2e (export, tokenizer and runner)
293- # PYTHON_EXECUTABLE=python ${CONDA_RUN} bash .ci/scripts/test_llava.sh Release
293+ # PYTHON_EXECUTABLE=python ${CONDA_RUN} bash .ci/scripts/test_llava.sh
294294
295295 test-qnn-model :
296296 name : test-qnn-model
@@ -351,6 +351,8 @@ jobs:
351351 done
352352
353353 test-huggingface-transformers :
354+ # NB: Don't run this on fork PRs because they won't have access to the secret and would fail anyway
355+ if : ${{ !github.event.pull_request.head.repo.fork }}
354356 name : test-huggingface-transformers
355357 uses : pytorch/test-infra/.github/workflows/linux_job.yml@main
356358 secrets : inherit
@@ -441,3 +443,39 @@ jobs:
441443
442444 cmake-out/examples/models/llama/llama_main --model_path=${ET_MODEL_NAME}.pte --tokenizer_path=${TOKENIZER_BIN_FILE} --prompt="My name is"
443445 echo "::endgroup::"
446+
447+
448+ test-llama-runner-qnn-linux :
449+ name : test-llama-runner-qnn-linux
450+ uses : pytorch/test-infra/.github/workflows/linux_job.yml@main
451+ strategy :
452+ matrix :
453+ dtype : [fp32]
454+ pt2e_quantize : [qnn_16a16w, qnn_8a8w]
455+ mode : [qnn]
456+ fail-fast : false
457+ with :
458+ runner : linux.2xlarge
459+ docker-image : executorch-ubuntu-22.04-qnn-sdk
460+ submodules : ' true'
461+ ref : ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
462+ timeout : 900
463+ script : |
464+ # The generic Linux job chooses to use base env, not the one setup by the image
465+ CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
466+ conda activate "${CONDA_ENV}"
467+
468+ BUILD_TOOL="cmake"
469+ DTYPE=${{ matrix.dtype }}
470+ MODE=${{ matrix.mode }}
471+ PT2E_QUANTIZE=${{ matrix.pt2e_quantize }}
472+
473+ PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh
474+ PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh
475+
476+ # Setup executorch
477+ PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
478+ # Install requirements for export_llama
479+ PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh
480+ # Test llama2
481+ PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}"
0 commit comments