Skip to content

Commit bc00cae

Browse files
committed
up
1 parent 61db1b6 commit bc00cae

File tree

4 files changed

+9
-15
lines changed

4 files changed

+9
-15
lines changed

.Package.swift/kernels_torchao/dummy.swift

Whitespace-only changes.

.Package.swift/kernels_torchao_debug/dummy.swift

Whitespace-only changes.

.ci/scripts/test_llama_torchao_lowbit.sh

Lines changed: 5 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -29,27 +29,22 @@ cmake -DPYTHON_EXECUTABLE=python \
2929
-DEXECUTORCH_ENABLE_LOGGING=1 \
3030
-DCMAKE_BUILD_TYPE=Release \
3131
-DEXECUTORCH_BUILD_EXTENSION_DATA_LOADER=ON \
32-
-DEXECUTORCH_BUILD_EXTENSION_LLM=ON \
33-
-DEXECUTORCH_BUILD_EXTENSION_LLM_RUNNER=ON \
34-
-DEXECUTORCH_BUILD_EXTENSION_MODULE=ON \
3532
-DEXECUTORCH_BUILD_EXTENSION_FLAT_TENSOR=ON \
33+
-DEXECUTORCH_BUILD_EXTENSION_MODULE=ON \
3634
-DEXECUTORCH_BUILD_EXTENSION_TENSOR=ON \
3735
-DEXECUTORCH_BUILD_XNNPACK=OFF \
3836
-DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON \
3937
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
38+
-DEXECUTORCH_BUILD_KERNELS_TORCHAO=ON \
39+
-DEXECUTORCH_BUILD_EXTENSION_LLM_RUNNER=ON \
40+
-DEXECUTORCH_BUILD_EXTENSION_LLM=ON \
4041
-DEXECUTORCH_BUILD_KERNELS_LLM=ON \
4142
-Bcmake-out .
42-
cmake --build cmake-out -j16 --target install --config Release
43+
cmake --build cmake-out -j16 --config Release --target install
4344

4445
# Install llama runner with torchao
4546
cmake -DPYTHON_EXECUTABLE=python \
46-
-DBUILD_TESTING=OFF \
4747
-DCMAKE_BUILD_TYPE=Release \
48-
-DEXECUTORCH_BUILD_KERNELS_LLM=ON \
49-
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
50-
-DEXECUTORCH_BUILD_XNNPACK=OFF \
51-
-DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON \
52-
-DEXECUTORCH_BUILD_TORCHAO=ON \
5348
-Bcmake-out/examples/models/llama \
5449
examples/models/llama
5550
cmake --build cmake-out/examples/models/llama -j16 --config Release

examples/models/llama/README.md

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -383,7 +383,6 @@ Once the model is exported, we need to build ExecuTorch and the runner with the
383383
The first step is to install ExecuTorch (the same as step 3.1 above):
384384
385385
```
386-
cmake -DPYTHON_EXECUTABLE=python \
387386
cmake -DPYTHON_EXECUTABLE=python \
388387
-DCMAKE_INSTALL_PREFIX=cmake-out \
389388
-DEXECUTORCH_ENABLE_LOGGING=1 \
@@ -395,12 +394,12 @@ cmake -DPYTHON_EXECUTABLE=python \
395394
-DEXECUTORCH_BUILD_XNNPACK=OFF \
396395
-DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON \
397396
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
398-
-DEXECUTORCH_BUILD_KERNELS_TORCHAO=ON \
399-
-DEXECUTORCH_BUILD_EXTENSION_LLM_RUNNER=ON \
400-
-DEXECUTORCH_BUILD_EXTENSION_LLM=ON \
397+
-DEXECUTORCH_BUILD_KERNELS_TORCHAO=ON \
398+
-DEXECUTORCH_BUILD_EXTENSION_LLM_RUNNER=ON \
399+
-DEXECUTORCH_BUILD_EXTENSION_LLM=ON \
401400
-DEXECUTORCH_BUILD_KERNELS_LLM=ON \
402401
-Bcmake-out .
403-
cmake --build cmake-out -j16 --target install --config Release
402+
cmake --build cmake-out -j16 --config Release --target install
404403
```
405404
406405
Next install the llama runner with torchao kernels enabled (similar to step 3.2 above):

0 commit comments

Comments
 (0)