| 
38 | 38 |         # Build and test ExecuTorch with the add model on portable backend.  | 
39 | 39 |         PYTHON_EXECUTABLE=python bash .ci/scripts/test_model.sh "add" "${BUILD_TOOL}" "portable"  | 
40 | 40 | 
  | 
41 |  | -  test-models-linux:  | 
42 |  | -    name: test-models-linux  | 
 | 41 | +  test-models-linux-basic:  | 
 | 42 | +    name: test-models-linux-basic  | 
43 | 43 |     uses: pytorch/test-infra/.github/workflows/linux_job_v2.yml@main  | 
44 | 44 |     permissions:  | 
45 | 45 |       id-token: write  | 
 | 
70 | 70 |         # Build and test ExecuTorch  | 
71 | 71 |         PYTHON_EXECUTABLE=python bash .ci/scripts/test_model.sh "${MODEL_NAME}" "${BUILD_TOOL}" "${BACKEND}"  | 
72 | 72 | 
  | 
 | 73 | +  test-models-linux:  | 
 | 74 | +    name: test-models-linux  | 
 | 75 | +    uses: pytorch/test-infra/.github/workflows/linux_job_v2.yml@main  | 
 | 76 | +    permissions:  | 
 | 77 | +      id-token: write  | 
 | 78 | +      contents: read  | 
 | 79 | +    strategy:  | 
 | 80 | +      matrix:  | 
 | 81 | +        model: [linear, add, add_mul, ic3, mv2, resnet18, resnet50, mobilebert, emformer_transcribe]  | 
 | 82 | +        backend: [portable, xnnpack-quantization-delegation]  | 
 | 83 | +        runner: [linux.2xlarge]  | 
 | 84 | +        include:  | 
 | 85 | +          - model: ic4  | 
 | 86 | +            backend: portable  | 
 | 87 | +            runner: linux.4xlarge.memory  | 
 | 88 | +          - model: ic4  | 
 | 89 | +            backend: xnnpack-quantization-delegation  | 
 | 90 | +            runner: linux.4xlarge.memory  | 
 | 91 | +          - model: emformer_join  | 
 | 92 | +            backend: portable  | 
 | 93 | +            runner: linux.4xlarge.memory  | 
 | 94 | +          - model: emformer_join  | 
 | 95 | +            backend: xnnpack-quantization-delegation  | 
 | 96 | +            runner: linux.4xlarge.memory  | 
 | 97 | +          - model: phi-4-mini  | 
 | 98 | +            backend: portable  | 
 | 99 | +            runner: linux.4xlarge.memory  | 
 | 100 | +          - model: llama3_2_vision_encoder  | 
 | 101 | +            backend: portable  | 
 | 102 | +            runner: linux.4xlarge.memory  | 
 | 103 | +          - model: w2l  | 
 | 104 | +            backend: portable  | 
 | 105 | +            runner: linux.4xlarge.memory  | 
 | 106 | +      fail-fast: false  | 
 | 107 | +    with:  | 
 | 108 | +      runner: ${{ matrix.runner }}  | 
 | 109 | +      docker-image: executorch-ubuntu-22.04-clang12  | 
 | 110 | +      submodules: 'true'  | 
 | 111 | +      ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}  | 
 | 112 | +      timeout: 90  | 
 | 113 | +      script: |  | 
 | 114 | +        # The generic Linux job chooses to use base env, not the one setup by the image  | 
 | 115 | +        CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")  | 
 | 116 | +        conda activate "${CONDA_ENV}"  | 
 | 117 | +
  | 
 | 118 | +        MODEL_NAME=${{ matrix.model }}  | 
 | 119 | +        BUILD_TOOL=cmake  | 
 | 120 | +        BACKEND=${{ matrix.backend }}  | 
 | 121 | +
  | 
 | 122 | +        PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}"  | 
 | 123 | +        # Build and test ExecuTorch  | 
 | 124 | +        PYTHON_EXECUTABLE=python bash .ci/scripts/test_model.sh "${MODEL_NAME}" "${BUILD_TOOL}" "${BACKEND}"  | 
 | 125 | +
  | 
73 | 126 |   test-llama-runner-linux:  | 
74 | 127 |     name: test-llama-runner-linux  | 
75 | 128 |     uses: pytorch/test-infra/.github/workflows/linux_job_v2.yml@main  | 
 | 
0 commit comments