Skip to content

Commit 1071d8e

Browse files
authored
Merge branch 'main' into gh/swolchok/140/head
2 parents 6977c03 + 57a09f4 commit 1071d8e

File tree

43 files changed

+356
-191
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

43 files changed

+356
-191
lines changed

.ci/scripts/utils.sh

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -17,17 +17,17 @@ retry () {
1717
}
1818

1919
clean_executorch_install_folders() {
20-
./install_requirements.sh --clean
20+
./install_executorch.sh --clean
2121
}
2222

2323
install_executorch() {
2424
which pip
2525
# Install executorch, this assumes that Executorch is checked out in the
2626
# current directory.
2727
if [[ "${1:-}" == "use-pt-pinned-commit" ]]; then
28-
./install_requirements.sh --pybind xnnpack --use-pt-pinned-commit
28+
./install_executorch.sh --pybind xnnpack --use-pt-pinned-commit
2929
else
30-
./install_requirements.sh --pybind xnnpack
30+
./install_executorch.sh --pybind xnnpack
3131
fi
3232
# Just print out the list of packages for debugging
3333
pip list

.github/workflows/apple.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ on:
99
paths:
1010
- .ci/scripts/setup-ios.sh
1111
- .github/workflows/apple.yml
12-
- install_requirements.sh
12+
- install_executorch.sh
1313
- backends/apple/**
1414
- build/build_apple_frameworks.sh
1515
- build/build_apple_llm_demo.sh

.github/workflows/pull.yml

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -200,7 +200,7 @@ jobs:
200200
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "cmake"
201201
202202
# install pybind
203-
bash install_requirements.sh --pybind xnnpack
203+
bash install_executorch.sh --pybind xnnpack
204204
205205
# install Llava requirements
206206
bash examples/models/llama/install_requirements.sh
@@ -436,7 +436,7 @@ jobs:
436436
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "cmake"
437437
438438
# install pybind
439-
bash install_requirements.sh --pybind xnnpack
439+
bash install_executorch.sh --pybind xnnpack
440440
441441
# install phi-3-mini requirements
442442
bash examples/models/phi-3-mini/install_requirements.sh
@@ -463,7 +463,7 @@ jobs:
463463
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "cmake"
464464
465465
# install pybind
466-
bash install_requirements.sh --pybind xnnpack
466+
bash install_executorch.sh --pybind xnnpack
467467
468468
# install llama requirements
469469
bash examples/models/llama/install_requirements.sh
@@ -490,7 +490,7 @@ jobs:
490490
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "cmake"
491491
492492
# install pybind
493-
bash install_requirements.sh --pybind xnnpack
493+
bash install_executorch.sh --pybind xnnpack
494494
495495
# install llama requirements
496496
bash examples/models/llama/install_requirements.sh
@@ -517,7 +517,7 @@ jobs:
517517
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "cmake"
518518
519519
# install pybind
520-
bash install_requirements.sh --pybind xnnpack
520+
bash install_executorch.sh --pybind xnnpack
521521
522522
# install llama requirements
523523
bash examples/models/llama/install_requirements.sh

backends/apple/mps/setup.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -97,7 +97,7 @@ I 00:00:00.122615 executorch:mps_executor_runner.mm:501] Model verified successf
9797
### [Optional] Run the generated model directly using pybind
9898
1. Make sure `pybind` MPS support was installed:
9999
```bash
100-
./install_requirements.sh --pybind mps
100+
./install_executorch.sh --pybind mps
101101
```
102102
2. Run the `mps_example` script to trace the model and run it directly from python:
103103
```bash

backends/cadence/build_cadence_fusionG3.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ unset XTENSA_CORE
1212
export XTENSA_CORE=FCV_FG3GP
1313
git submodule sync
1414
git submodule update --init
15-
./install_requirements.sh
15+
./install_executorch.sh
1616

1717
rm -rf cmake-out
1818

backends/cadence/build_cadence_hifi4.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ unset XTENSA_CORE
1212
export XTENSA_CORE=nxp_rt600_RI23_11_newlib
1313
git submodule sync
1414
git submodule update --init
15-
./install_requirements.sh
15+
./install_executorch.sh
1616

1717
rm -rf cmake-out
1818

backends/cadence/reference/operators/quantized_conv_out.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -119,7 +119,7 @@ __attribute__((noinline)) void conv2d_nchw_core_generic(
119119
if (((_h + d0 * _wh - p0) >= 0) &&
120120
((_h + d0 * _wh - p0) < h) &&
121121
((_w + d1 * _ww - p1) >= 0) &&
122-
((_w + d1 * _ww - p1 < w))) {
122+
((_w + d1 * _ww - p1) < w)) {
123123
int ioff =
124124
(_h + d0 * _wh - p0) * w + (_w + d1 * _ww - p1);
125125
int woff = _wh * ww + _ww;

backends/vulkan/docs/android_demo.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -81,7 +81,7 @@ First, build and install ExecuTorch libraries, then build the LLaMA runner
8181
binary using the Android NDK toolchain.
8282

8383
```shell
84-
./install_requirements.sh --clean
84+
./install_executorch.sh --clean
8585
(mkdir cmake-android-out && \
8686
cmake . -DCMAKE_INSTALL_PREFIX=cmake-android-out \
8787
-DCMAKE_TOOLCHAIN_FILE=$ANDROID_NDK/build/cmake/android.toolchain.cmake \

backends/vulkan/runtime/graph/ops/glsl/conv2d_pw.glsl

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,9 @@ ${layout_declare_ubo(8, "float", "out_min", "float", "out_max")}
3333
layout(local_size_x_id = 0, local_size_y_id = 1, local_size_z_id = 2) in;
3434

3535
// shared memory to hold calculated positions, this would reduce register usage thus improving performance.
36-
shared ivec2 pos_shared[gl_WorkGroupSize.x * gl_WorkGroupSize.y * gl_WorkGroupSize.z * TILE_SIZE * TILE_SIZE];
36+
// 64 is the number of threads in the local wg
37+
$num_shared = 64 * TILE_SIZE * TILE_SIZE
38+
shared ivec2 pos_shared[${num_shared}];
3739

3840
/*
3941
* Computes a 2D pointwise convolution of an NxN output tile. Calculating an

backends/xnnpack/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -98,7 +98,7 @@ After exporting the XNNPACK Delegated model, we can now try running it with exam
9898
cd executorch
9999

100100
# Get a clean cmake-out directory
101-
./install_requirements.sh --clean
101+
./install_executorch.sh --clean
102102
mkdir cmake-out
103103

104104
# Configure cmake

0 commit comments

Comments
 (0)