Skip to content

Commit 5816b94

Browse files
committed
Merge remote-tracking branch 'origin/main' into toupstream/cat_annotator
2 parents c4d5892 + c045c35 commit 5816b94

File tree

191 files changed

+7137
-1318
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

191 files changed

+7137
-1318
lines changed
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
export-D64151426
1+
bd5482c7c3e1197e10c46ff739027f917d9c1fcc

.github/workflows/_android.yml

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -66,6 +66,16 @@ jobs:
6666
# avoid permission issue
6767
sudo chown -R "${USER}" /opt/android
6868
69+
- name: Download Artifacts
70+
shell: bash
71+
run: |
72+
set -eux
73+
curl -O https://gha-artifacts.s3.amazonaws.com/${{ github.repository }}/${{ github.run_id }}/artifacts/llm_demo/app-debug.apk
74+
curl -O https://gha-artifacts.s3.amazonaws.com/${{ github.repository }}/${{ github.run_id }}/artifacts/llm_demo/app-debug-androidTest.apk
75+
curl -O https://gha-artifacts.s3.amazonaws.com/${{ github.repository }}/${{ github.run_id }}/artifacts/fp32-xnnpack-custom/model.zip
76+
unzip model.zip
77+
mv *.pte model.pte
78+
6979
- name: Gradle cache
7080
uses: gradle/actions/setup-gradle@v3
7181

.github/workflows/ghstack_land.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,7 @@ on:
1111
- 'gh/kimishpatel/[0-9]+/base'
1212
- 'gh/kirklandsign/[0-9]+/base'
1313
- 'gh/larryliu0820/[0-9]+/base'
14+
- 'gh/lucylq/[0-9]+/base'
1415
- 'gh/manuelcandales/[0-9]+/base'
1516
- 'gh/mcr229/[0-9]+/base'
1617
- 'gh/swolchok/[0-9]+/base'

.github/workflows/pull.yml

Lines changed: 17 additions & 46 deletions
Original file line numberDiff line numberDiff line change
@@ -35,9 +35,6 @@ jobs:
3535
name: test-setup-linux-gcc
3636
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
3737
strategy:
38-
matrix:
39-
include:
40-
- build-tool: cmake
4138
fail-fast: false
4239
with:
4340
runner: linux.2xlarge
@@ -50,7 +47,7 @@ jobs:
5047
CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
5148
conda activate "${CONDA_ENV}"
5249
53-
BUILD_TOOL=${{ matrix.build-tool }}
50+
BUILD_TOOL="cmake"
5451
5552
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
5653
# Build and test ExecuTorch with the add model on portable backend.
@@ -89,20 +86,11 @@ jobs:
8986
strategy:
9087
matrix:
9188
dtype: [fp32]
92-
build-tool: [buck2, cmake]
9389
mode: [portable, xnnpack+custom, xnnpack+custom+qe]
9490
include:
9591
- dtype: bf16
96-
build-tool: cmake
97-
mode: portable
98-
- dtype: bf16
99-
build-tool: buck2
10092
mode: portable
10193
- dtype: bf16
102-
build-tool: cmake
103-
mode: custom
104-
- dtype: bf16
105-
build-tool: buck2
10694
mode: custom
10795
fail-fast: false
10896
with:
@@ -111,29 +99,30 @@ jobs:
11199
submodules: 'true'
112100
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
113101
timeout: 900
102+
upload-artifact: android-models
103+
upload-artifact-to-s3: true
114104
script: |
115105
# The generic Linux job chooses to use base env, not the one setup by the image
116106
CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
117107
conda activate "${CONDA_ENV}"
118108
119109
DTYPE=${{ matrix.dtype }}
120-
BUILD_TOOL=${{ matrix.build-tool }}
110+
BUILD_TOOL="cmake"
121111
MODE=${{ matrix.mode }}
112+
ARTIFACTS_DIR_NAME="artifacts-to-be-uploaded/${DTYPE}-${MODE}"
113+
ARTIFACTS_DIR_NAME="${ARTIFACTS_DIR_NAME/+/-}"
122114
123115
# Setup executorch
124-
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh buck2
116+
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
125117
# Install requirements for export_llama
126118
PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh
127119
# Test llama2
128-
PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh stories110M "${BUILD_TOOL}" "${DTYPE}" "${MODE}"
120+
PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh stories110M "${BUILD_TOOL}" "${DTYPE}" "${MODE}" "${ARTIFACTS_DIR_NAME}"
129121
130122
test-llama-runner-linux-android:
131123
name: test-llama-runner-linux-android
132124
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
133125
strategy:
134-
matrix:
135-
include:
136-
- build-tool: cmake
137126
fail-fast: false
138127
with:
139128
runner: linux.2xlarge
@@ -146,18 +135,14 @@ jobs:
146135
CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
147136
conda activate "${CONDA_ENV}"
148137
149-
BUILD_TOOL=${{ matrix.build-tool }}
138+
BUILD_TOOL="cmake"
150139
PYTHON_EXECUTABLE=python \
151140
bash .ci/scripts/build_llama_android.sh "${BUILD_TOOL}"
152141
153142
test-custom-ops-linux:
154143
name: test-custom-ops-linux
155144
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
156145
strategy:
157-
matrix:
158-
include:
159-
- build-tool: buck2
160-
- build-tool: cmake
161146
fail-fast: false
162147
with:
163148
runner: linux.2xlarge
@@ -170,7 +155,7 @@ jobs:
170155
CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
171156
conda activate "${CONDA_ENV}"
172157
173-
BUILD_TOOL=${{ matrix.build-tool }}
158+
BUILD_TOOL="cmake"
174159
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
175160
# Test custom ops
176161
PYTHON_EXECUTABLE=python bash examples/portable/custom_ops/test_custom_ops.sh "${BUILD_TOOL}"
@@ -179,10 +164,6 @@ jobs:
179164
name: test-selective-build-linux
180165
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
181166
strategy:
182-
matrix:
183-
include:
184-
- build-tool: buck2
185-
- build-tool: cmake
186167
fail-fast: false
187168
with:
188169
runner: linux.2xlarge
@@ -195,7 +176,7 @@ jobs:
195176
CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
196177
conda activate "${CONDA_ENV}"
197178
198-
BUILD_TOOL=${{ matrix.build-tool }}
179+
BUILD_TOOL="cmake"
199180
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
200181
# Test selective build
201182
PYTHON_EXECUTABLE=python bash examples/selective_build/test_selective_build.sh "${BUILD_TOOL}"
@@ -235,9 +216,6 @@ jobs:
235216
name: test-quantized-aot-lib-linux
236217
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
237218
strategy:
238-
matrix:
239-
include:
240-
- build-tool: cmake
241219
fail-fast: false
242220
with:
243221
runner: linux.2xlarge
@@ -250,17 +228,14 @@ jobs:
250228
CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
251229
conda activate "${CONDA_ENV}"
252230
253-
BUILD_TOOL=${{ matrix.build-tool }}
231+
BUILD_TOOL="cmake"
254232
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
255233
PYTHON_EXECUTABLE=python bash examples/xnnpack/quantization/test_quantize.sh "${BUILD_TOOL}" mv2
256234
257235
test-pybind-build-linux:
258236
name: test-pybind-build-linux
259237
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
260238
strategy:
261-
matrix:
262-
include:
263-
- build-tool: cmake
264239
fail-fast: false
265240
with:
266241
runner: linux.2xlarge
@@ -274,7 +249,7 @@ jobs:
274249
conda activate "${CONDA_ENV}"
275250
276251
# build module for executorch.extension.pybindings.portable_lib
277-
BUILD_TOOL=${{ matrix.build-tool }}
252+
BUILD_TOOL="cmake"
278253
PYTHON_EXECUTABLE=python \
279254
EXECUTORCH_BUILD_XNNPACK=ON \
280255
EXECUTORCH_BUILD_PYBIND=ON \
@@ -349,6 +324,7 @@ jobs:
349324
350325
android:
351326
uses: ./.github/workflows/_android.yml
327+
needs: test-llama-runner-linux
352328

353329
unittest:
354330
uses: ./.github/workflows/_unittest.yml
@@ -357,10 +333,6 @@ jobs:
357333

358334
unittest-arm:
359335
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
360-
strategy:
361-
matrix:
362-
include:
363-
- build-tool: buck2
364336
with:
365337
runner: linux.2xlarge
366338
docker-image: executorch-ubuntu-22.04-arm-sdk
@@ -374,7 +346,7 @@ jobs:
374346
CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]")
375347
conda activate "${CONDA_ENV}"
376348
377-
BUILD_TOOL=${{ matrix.build-tool }}
349+
BUILD_TOOL="cmake"
378350
379351
# Setup MacOS dependencies as there is no Docker support on MacOS atm
380352
PYTHON_EXECUTABLE=python \
@@ -396,7 +368,6 @@ jobs:
396368
strategy:
397369
matrix:
398370
dtype: [fp32]
399-
build-tool: [cmake]
400371
mode: [qnn]
401372
fail-fast: false
402373
with:
@@ -411,14 +382,14 @@ jobs:
411382
conda activate "${CONDA_ENV}"
412383
413384
DTYPE=${{ matrix.dtype }}
414-
BUILD_TOOL=${{ matrix.build-tool }}
385+
BUILD_TOOL="cmake"
415386
MODE=${{ matrix.mode }}
416387
417388
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh
418389
PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh
419390
420391
# Setup executorch
421-
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh buck2
392+
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
422393
# Install requirements for export_llama
423394
PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh
424395
# Test llama2

.github/workflows/update-viablestrict.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ jobs:
2020
with:
2121
repository: pytorch/executorch
2222
stable-branch: viable/strict
23-
requires: '[\"pull\", \"lint\", \"trunk\", \"Build documentation\", \"^Android$\", \"^Apple$\"]'
23+
requires: '[\"pull\", \"lint\", \"trunk\", \"Build documentation\", \"^Apple$\"]'
2424
secret-bot-token: ${{ secrets.UPDATEBOT_TOKEN }}
2525
clickhouse-url: ${{ secrets.CLICKHOUSE_URL }}
2626
clickhouse-username: ${{ secrets.CLICKHOUSE_VIABLESTRICT_USERNAME }}

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -92,7 +92,7 @@ tools.
9292
├── runtime # Core C++ runtime.
9393
| ├── backend # Backend delegate runtime APIs.
9494
| ├── core # Core structures used across all levels of the runtime.
95-
| ├── executor # Model loading, initalization, and execution.
95+
| ├── executor # Model loading, initialization, and execution.
9696
| ├── kernel # Kernel registration and management.
9797
| ├── platform # Layer between architecture specific code and portable C++.
9898
├── schema # ExecuTorch PTE file format flatbuffer

backends/apple/coreml/compiler/coreml_preprocess.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -425,12 +425,15 @@ def preprocess(
425425
CoreMLBackend.op_linear_quantizer_config_from_compile_specs(compile_specs)
426426
)
427427

428+
# Load the model if MODEL_TYPE is 'COMPILED_MODEL'. This step is necessary because
429+
# get_compiled_model_path() requires a loaded model.
430+
skip_model_load = model_type != CoreMLBackend.MODEL_TYPE.COMPILED_MODEL
428431
mlmodel = ct.convert(
429432
model=edge_program,
430433
source="pytorch",
431434
convert_to="mlprogram",
432435
pass_pipeline=ct.PassPipeline.DEFAULT,
433-
skip_model_load=True,
436+
skip_model_load=skip_model_load,
434437
compute_precision=model_compute_precision,
435438
minimum_deployment_target=minimum_deployment_target,
436439
compute_units=compute_units,

backends/apple/coreml/runtime/delegate/backend_delegate.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ class BackendDelegate {
2828
// Max models cache size in bytes.
2929
size_t max_models_cache_size = 10 * size_t(1024) * size_t(1024) * size_t(1024);
3030
// If set to `true`, delegate pre-warms the most recently used asset.
31-
bool should_prewarm_asset = true;
31+
bool should_prewarm_asset = false;
3232
// If set to `true`, delegate pre-warms the model in `init`.
3333
bool should_prewarm_model = true;
3434
};

backends/apple/coreml/runtime/delegate/com.apple.executorchcoreml_config.plist

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
<plist version="1.0">
44
<dict>
55
<key>shouldPrewarmAsset</key>
6-
<true/>
6+
<false/>
77
<key>shouldPrewarmModel</key>
88
<true/>
99
<key>maxAssetsSizeInBytes</key>

backends/apple/coreml/runtime/test/CoreMLBackendDelegateTests.mm

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -209,6 +209,12 @@ - (void)testStateProgramExecute {
209209
}
210210
#endif
211211

212+
- (void)testAddMulCompiledProgramExecute {
213+
NSURL *modelURL = [[self class] bundledResourceWithName:@"add_mul_compiled_coreml_all" extension:@"pte"];
214+
XCTAssertNotNil(modelURL);
215+
[self executeModelAtURL:modelURL nLoads:1 nExecutions:2];
216+
}
217+
212218
- (void)executeMultipleModelsConcurrently:(NSArray<NSURL *> *)modelURLs
213219
nLoads:(NSUInteger)nLoads
214220
nExecutions:(NSUInteger)nExecutions

0 commit comments

Comments
 (0)