@@ -223,7 +223,6 @@ jobs:
223223 strategy :
224224 matrix :
225225 dtype : [fp32]
226- build-tool : [buck2, cmake]
227226 mode : [portable, xnnpack+kv+custom, mps, coreml]
228227 fail-fast : false
229228 with :
@@ -235,25 +234,12 @@ jobs:
235234 script : |
236235
237236 DTYPE=${{ matrix.dtype }}
238- BUILD_TOOL=${{ matrix.build-tool }}
239237 MODE=${{ matrix.mode }}
240238
241- if [[ "${BUILD_TOOL}" == "buck2" ]]; then
242- # TODO: Will add more modes that don't support buck2
243- if [[ "${MODE}" == "mps" ]]; then
244- echo "mps doesn't support buck2."
245- exit 0
246- fi
247- if [[ "${MODE}" == "coreml" ]]; then
248- echo "coreml doesn't support buck2."
249- exit 0
250- fi
251- fi
252-
253239 bash .ci/scripts/setup-conda.sh
254240
255241 # Setup executorch
256- PYTHON_EXECUTABLE=python ${CONDA_RUN} bash .ci/scripts/setup-macos.sh "${BUILD_TOOL}"
242+ PYTHON_EXECUTABLE=python ${CONDA_RUN} bash .ci/scripts/setup-macos.sh cmake
257243
258244 if [[ "${MODE}" == "mps" ]]; then
259245 # Install mps delegate
@@ -268,7 +254,7 @@ jobs:
268254 # Install requirements for export_llama
269255 PYTHON_EXECUTABLE=python ${CONDA_RUN} bash examples/models/llama2/install_requirements.sh
270256 # Test llama2
271- PYTHON_EXECUTABLE=python ${CONDA_RUN} bash .ci/scripts/test_llama.sh stories110M "${BUILD_TOOL}" "${DTYPE}" "${MODE}"
257+ PYTHON_EXECUTABLE=python ${CONDA_RUN} bash .ci/scripts/test_llama.sh stories110M cmake "${DTYPE}" "${MODE}"
272258
273259 # # TODO(jackzhxng): Runner consistently runs out of memory before test finishes. Try to find a more powerful runner.
274260 # test-llava-runner-macos:
0 commit comments