@@ -29,27 +29,22 @@ cmake -DPYTHON_EXECUTABLE=python \
2929 -DEXECUTORCH_ENABLE_LOGGING=1 \
3030 -DCMAKE_BUILD_TYPE=Release \
3131 -DEXECUTORCH_BUILD_EXTENSION_DATA_LOADER=ON \
32- -DEXECUTORCH_BUILD_EXTENSION_LLM=ON \
33- -DEXECUTORCH_BUILD_EXTENSION_LLM_RUNNER=ON \
34- -DEXECUTORCH_BUILD_EXTENSION_MODULE=ON \
3532 -DEXECUTORCH_BUILD_EXTENSION_FLAT_TENSOR=ON \
33+ -DEXECUTORCH_BUILD_EXTENSION_MODULE=ON \
3634 -DEXECUTORCH_BUILD_EXTENSION_TENSOR=ON \
3735 -DEXECUTORCH_BUILD_XNNPACK=OFF \
3836 -DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON \
3937 -DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
38+ -DEXECUTORCH_BUILD_KERNELS_TORCHAO=ON \
39+ -DEXECUTORCH_BUILD_EXTENSION_LLM_RUNNER=ON \
40+ -DEXECUTORCH_BUILD_EXTENSION_LLM=ON \
4041 -DEXECUTORCH_BUILD_KERNELS_LLM=ON \
4142 -Bcmake-out .
42- cmake --build cmake-out -j16 --target install --config Release
43+ cmake --build cmake-out -j16 --config Release --target install
4344
4445# Install llama runner with torchao
4546cmake -DPYTHON_EXECUTABLE=python \
46- -DBUILD_TESTING=OFF \
4747 -DCMAKE_BUILD_TYPE=Release \
48- -DEXECUTORCH_BUILD_KERNELS_LLM=ON \
49- -DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
50- -DEXECUTORCH_BUILD_XNNPACK=OFF \
51- -DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON \
52- -DEXECUTORCH_BUILD_TORCHAO=ON \
5348 -Bcmake-out/examples/models/llama \
5449 examples/models/llama
5550cmake --build cmake-out/examples/models/llama -j16 --config Release
0 commit comments