@@ -223,7 +223,6 @@ jobs:
223223    strategy :
224224      matrix :
225225        dtype : [fp32] 
226-         build-tool : [buck2, cmake] 
227226        mode : [portable, xnnpack+kv+custom, mps, coreml] 
228227      fail-fast : false 
229228    with :
@@ -235,25 +234,12 @@ jobs:
235234      script : | 
236235
237236        DTYPE=${{ matrix.dtype }} 
238-         BUILD_TOOL=${{ matrix.build-tool }} 
239237        MODE=${{ matrix.mode }} 
240238
241-         if [[ "${BUILD_TOOL}" == "buck2" ]]; then 
242-           # TODO: Will add more modes that don't support buck2 
243-           if [[ "${MODE}" == "mps" ]]; then 
244-             echo "mps doesn't support buck2." 
245-             exit 0 
246-           fi 
247-           if [[ "${MODE}" == "coreml" ]]; then 
248-             echo "coreml doesn't support buck2." 
249-             exit 0 
250-           fi 
251-         fi 
252- 
253239        bash .ci/scripts/setup-conda.sh 
254240
255241        # Setup executorch 
256-         PYTHON_EXECUTABLE=python ${CONDA_RUN} bash .ci/scripts/setup-macos.sh "${BUILD_TOOL}"  
242+         PYTHON_EXECUTABLE=python ${CONDA_RUN} bash .ci/scripts/setup-macos.sh cmake  
257243
258244        if [[ "${MODE}" == "mps" ]]; then 
259245          # Install mps delegate 
@@ -268,7 +254,7 @@ jobs:
268254        # Install requirements for export_llama 
269255        PYTHON_EXECUTABLE=python ${CONDA_RUN} bash examples/models/llama2/install_requirements.sh 
270256        # Test llama2 
271-         PYTHON_EXECUTABLE=python ${CONDA_RUN} bash .ci/scripts/test_llama.sh stories110M "${BUILD_TOOL}"  "${DTYPE}" "${MODE}" 
257+         PYTHON_EXECUTABLE=python ${CONDA_RUN} bash .ci/scripts/test_llama.sh stories110M cmake  "${DTYPE}" "${MODE}" 
272258
273259   #  # TODO(jackzhxng): Runner consistently runs out of memory before test finishes. Try to find a more powerful runner.
274260  #  test-llava-runner-macos:
0 commit comments