Skip to content

Commit a635bdb

Browse files
committed
fix pull.yml
1 parent c8a1b36 commit a635bdb

File tree

1 file changed

+15
-39
lines changed

1 file changed

+15
-39
lines changed

.github/workflows/pull.yml

Lines changed: 15 additions & 39 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,6 @@ concurrency:
1212
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.sha }}-${{ github.event_name == 'workflow_dispatch' }}-${{ github.event_name == 'schedule' }}
1313
cancel-in-progress: true
1414

15-
1615
jobs:
1716
test-qnn-wheel-packages-linux:
1817
name: test-qnn-wheel-packages-linux
@@ -592,7 +591,6 @@ jobs:
592591
# Test static llama weight sharing and accuracy
593592
PYTHON_EXECUTABLE=python bash .ci/scripts/test_qnn_static_llama.sh
594593
595-
596594
test-static-llama-qnn-eval-linux:
597595
name: test-static-llama-qnn-eval-linux
598596
uses: pytorch/test-infra/.github/workflows/linux_job_v2.yml@main
@@ -625,25 +623,19 @@ jobs:
625623
# The generic Linux job chooses to use base env, not the one setup by the image
626624
CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
627625
conda activate "${CONDA_ENV}"
628-
629626
BUILD_TOOL="cmake"
630-
631627
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh
632628
PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh
633-
634629
# Setup executorch
635630
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}"
636-
637631
# Setup install_requirements for llama
638632
PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh
639-
640633
# Test static llama weight sharing and accuracy
641634
echo ">>> Running config: ${{ matrix.config.name }}"
642635
PYTHON_EXECUTABLE=python bash .ci/scripts/test_qnn_static_llama_eval.sh \
643636
--flags "${{ matrix.config.flags }}" \
644637
--threshold "${{ matrix.config.threshold }}"
645638
646-
647639
test-qnn-models-linux:
648640
name: test-qnn-models-linux
649641
uses: pytorch/test-infra/.github/workflows/linux_job_v2.yml@main
@@ -802,25 +794,8 @@ jobs:
802794
# install a recent version of torchtune.
803795
PYTHON_EXECUTABLE=python python -m pip install torchtune==0.7.0.dev20250730 --extra-index-url https://download.pytorch.org/whl/nightly/cpu
804796
805-
test-openvino-linux:
806-
name: test-openvino-linux
807-
uses: pytorch/test-infra/.github/workflows/linux_job_v2.yml@main
808-
permissions:
809-
id-token: write
810-
contents: read
811-
strategy:
812-
fail-fast: false
813-
if: false # TODO Re-enable after fixing timeouts (#14314)
814-
with:
815-
runner: linux.2xlarge
816-
docker-image: ci-image:executorch-ubuntu-22.04-gcc9
817-
submodules: 'recursive'
818-
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
819-
timeout: 90
820-
script: |
821-
# The generic Linux job chooses to use base env, not the one setup by the image
822-
CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
823-
conda activate "${CONDA_ENV}"
797+
# run llama runner in eager mode
798+
PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama_lora.sh
824799
825800
test-mediatek-models-linux:
826801
name: test-mediatek-models-linux
@@ -857,6 +832,7 @@ jobs:
857832
contents: read
858833
strategy:
859834
fail-fast: false
835+
if: false # TODO Re-enable after fixing timeouts (#14314)
860836
with:
861837
runner: linux.2xlarge
862838
docker-image: ci-image:executorch-ubuntu-22.04-gcc9
@@ -988,25 +964,25 @@ jobs:
988964
script: |
989965
set -ex
990966
991-
# # The generic Linux job chooses to use base env, not the one setup by the image
992-
# CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
993-
# conda activate "${CONDA_ENV}"
967+
# The generic Linux job chooses to use base env, not the one setup by the image
968+
CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
969+
conda activate "${CONDA_ENV}"
994970
995-
# # Setup python
996-
# PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "cmake"
971+
# Setup python
972+
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "cmake"
997973
998974
# Setup Samsung SDK (AI Lite Core) and install enn backend
999975
export SAMSUNG_AI_LITECORE_KEY=$SECRET_SAMSUNG_AI_LITECORE_KEY
1000976
source .ci/scripts/setup-samsung-linux-deps.sh
1001977
1002-
# # Test models serially
1003-
# models="mv2 ic3 resnet18 resnet50 mv3 ic4 dl3 edsr vit w2l"
1004-
# for model in $models; do
1005-
# python -m executorch.examples.samsung.aot_compiler --model_name=$model -c E9955
1006-
# done
978+
# Test models serially
979+
models="mv2 ic3 resnet18 resnet50 mv3 ic4 dl3 edsr vit w2l"
980+
for model in $models; do
981+
python -m executorch.examples.samsung.aot_compiler --model_name=$model -c E9955
982+
done
1007983
1008-
# # Test ops
1009-
# python -m unittest discover -s backends/samsung/test/ops -p "test_*.py"
984+
# Test ops
985+
python -m unittest discover -s backends/samsung/test/ops -p "test_*.py"
1010986
1011987
1012988
test-vulkan-models-linux:

0 commit comments

Comments
 (0)