-
Couldn't load subscription status.
- Fork 700
MTK update rule #5776
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Closed
MTK update rule #5776
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Summary: Fix the build errors I introduced in #5318 Pull Request resolved: #5513 Test Plan: Built the model using the instructions at https://github.com/pytorch/executorch/blob/main/examples/models/phi-3-mini/README.md Reviewed By: helunwencser Differential Revision: D63133231 Pulled By: dbort fbshipit-source-id: 81d5402e76fd86c919a13a1b1b83687feca72ab7
Summary: Pull Request resolved: #5514 #5475 moved the images to static folder, #5493 updated some of it but missed out on the delegate specific docs. This diff, fixes that by replacing the image URLs that are broken Reviewed By: cmodi-meta, svekars, kirklandsign Differential Revision: D63135522 fbshipit-source-id: 10071b0428c203eef38500852e64f71e13ec1315
Summary: Took me a bit to figure out how to make cmake work; I had it installed in homebrew but it wasn't working. Add instructions for installing cmake globally in a way that Xcode can find it. Pull Request resolved: #5533 Reviewed By: shoumikhin Differential Revision: D63153558 Pulled By: dbort fbshipit-source-id: b7df7c0b9723c6e60cafea67a39dd326276a6148
Summary: Pull Request resolved: #5505 Reviewed By: SS-JIA Differential Revision: D63000056 fbshipit-source-id: 959127e874b30c7ebc069499d99e8c5881b3b272
Summary: Pull Request resolved: #5444 TIL about the `target` clang/GCC function attribute, which allows building a particular function under an `-march` flag instead of a whole file. ghstack-source-id: 243858419 Reviewed By: malfet Differential Revision: D62905047 fbshipit-source-id: a89c8169fea315aa653bbca819a672357c3dff77
Summary: Pull Request resolved: #5534 Reviewed By: dltn Differential Revision: D63199241 fbshipit-source-id: 9b097c3cc6c492fc13d9183c8679de9d00c80d21
Summary: Pull Request resolved: #5535 . Reviewed By: dltn Differential Revision: D63201286 fbshipit-source-id: 1767a1c0cf876f7a3b6b4534a83c912c3de0eabf
Summary: Pull Request resolved: #5539 Reviewed By: dltn Differential Revision: D63222601 fbshipit-source-id: 9a0bd6e10f7d2d2c769b7617e0b0e605af616eb0
Summary: Removed [!TIP] format and replaced with a subheader Pull Request resolved: #5517 Reviewed By: kirklandsign Differential Revision: D63145938 Pulled By: cmodi-meta fbshipit-source-id: 5503901957b8aeffe23cdc756bd4c73a007dd35e
Summary: This logs the metrics from the size command when building with run.sh Pull Request resolved: #5342 Reviewed By: manuelcandales Differential Revision: D62874679 Pulled By: digantdesai fbshipit-source-id: f69bfa12c48101e540e684a590f78b546903cb42
Summary: Adding "px" unit for PyTorch site (i.e. https://pytorch.org/executorch/main/llm/llama-demo-android.html) will have same image widths as readme in github Pull Request resolved: #5540 Reviewed By: Riandy, kirklandsign Differential Revision: D63226892 Pulled By: cmodi-meta fbshipit-source-id: 5cfa30ee0ab156c1e004405cdc7dd99a0f61d2c2
Summary: The code under examples/... is a proxy for user code, and users should never declare code under the `torch::` or `executorch::` namespaces. Move this code under the `example::` namespace to make it more clear that users should use their own namespaces when writing code like this. Pull Request resolved: #5478 Test Plan: - Built using the instructions at https://github.com/pytorch/executorch/blob/main/examples/mediatek/README.md Reviewed By: JacobSzwejbka, cccclai Differential Revision: D62992974 Pulled By: dbort fbshipit-source-id: b01f1b33d2853a0555ae19d79769a5bb6d0ba853
Summary: examples/ code should use the new `executorch::` namespaces. Pull Request resolved: #5516 Test Plan: Built the app using the instructions at https://github.com/pytorch/executorch/blob/main/examples/demo-apps/apple_ios/LLaMA/README.md Reviewed By: larryliu0820 Differential Revision: D63138639 Pulled By: dbort fbshipit-source-id: fffb6d35d425dd733eead1b24ee8b9f2831e65c0
Summary: Pull Request resolved: #5546 The last prompt sent would be included in `getConversationHistory()` + adding it prior to sending it with the generate(). It looks like this got move during the rebasing. To fix this we now call `getConversationHistory()` prior to adding the rawPrompt to a Message. In regards to model response, I noticed that it did not really change the quality of the response. (tested with Llama 3.1) Reviewed By: Riandy Differential Revision: D62761977 fbshipit-source-id: 2f975983965fe837147f1ffb8b5dcfa8f2061895
Summary: Example code should use the new `executorch::` namespace wherever possible, and should not define code under the `torch::` namespace. Pull Request resolved: #5512 Test Plan: - Built llava changes with `bash .ci/scripts/test_llava.sh` Reviewed By: JacobSzwejbka, larryliu0820 Differential Revision: D63133181 Pulled By: dbort fbshipit-source-id: 5796b85eef053f3b3e4ba0e27a3a26ae48747b5a
Summary: Use the names in the new `executorch::` namespace. Pull Request resolved: #5495 Test Plan: ``` ./examples/devtools/build_example_runner.sh ``` Reviewed By: larryliu0820 Differential Revision: D63047148 Pulled By: dbort fbshipit-source-id: e0e3af1c130aaf409ecc142c28d75f0a44d88fa3
Summary: Pull Request resolved: #5548 Converting the input to and from float32 is faster than not using the op. h/t to torchchat, which does this already (though it had a bug, which I sent a patch for). Reviewed By: kimishpatel Differential Revision: D63158951 fbshipit-source-id: 58c90d141ee403536c03a3b731f8547790fc9440
Summary: This PR adds a CI job for phi-3-mini Pull Request resolved: #5532 Test Plan: The CI Job is green: https://github.com/pytorch/executorch/actions/runs/10967809307/job/30458161933?pr=5532 Reviewed By: iseeyuan Differential Revision: D63157703 Pulled By: helunwencser fbshipit-source-id: fc7f54e166062443f396e7a304712f7b60e5db90
Summary: Preview in GitHub for consistency: https://github.com/pytorch/executorch/blob/fix-images/examples/demo-apps/android/LlamaDemo/README.md Pull Request resolved: #5550 Test Plan: Doc preview: https://docs-preview.pytorch.org/pytorch/executorch/5550/llm/llama-demo-android.html Rendered GitHub preview: https://github.com/pytorch/executorch/blob/fix-images/examples/demo-apps/android/LlamaDemo/README.md?rgh-link-date=2024-09-23T20%3A12%3A32Z Reviewed By: cmodi-meta, kirklandsign Differential Revision: D63281004 Pulled By: svekars fbshipit-source-id: cc5710624cb9bdab9056558c94f127b3bc12b96c
…5515) Summary: Pull Request resolved: #5515 Storing QMat2 in a texture gives way to two main problems: - Indexing is a mess and additional computation is required to take into account the fact that we are reading ivec4's and only using half of the values - There is no texel fetching in int8. The texel is read in int32 and needs to be casted Keeping QMat2 in a buffer performs better because, although reading from buffers is slower, removing the extra computation compensates for this. {F1863459327} This diff also moves the scales_and_zeros tensor to Channels Packed in texture implementations because it just makes more sense, I had done some terrible indexing shennanigans before. ghstack-source-id: 244258611 exported-using-ghexport Reviewed By: yipjustin Differential Revision: D62504978 fbshipit-source-id: df2fdf87f75140be0a316576c8ffad67feefd6d7
Summary: Pull Request resolved: #5499 Seems to block bfloat16 stories110M as exported by torchchat (and we should have op coverage for bfloat16 anyway). ghstack-source-id: 243857968 Reviewed By: larryliu0820 Differential Revision: D63054001 fbshipit-source-id: 530b479872643f878912592c7b260d71e6e05804
Summary: Pull Request resolved: #5500 ghstack-source-id: 243857969 Reviewed By: digantdesai, larryliu0820 Differential Revision: D63057744 fbshipit-source-id: 9e1fb6f6479adb1575c5aed61b9da3c774586ba3
Summary: Pull Request resolved: #5519 Discovered these were missing while trying to use the following diff. ghstack-source-id: 243867517 exported-using-ghexport Reviewed By: digantdesai Differential Revision: D63147276 fbshipit-source-id: bf75fb0fe452e2e68a34271ca1250cdb90657e5a
Summary: Pull Request resolved: #5553 Adding support to load Llama Guard model and run prompt classification task Reviewed By: cmodi-meta, kirklandsign Differential Revision: D63148252 fbshipit-source-id: 482559e694da05bdec75b9a2dbd76163c686e47d
Summary: Pull Request resolved: #5560 ## Context Refactor operator test code generation scripts, such that components can be re-used to generate operator benchmarks. In broad strokes, the refactors implemented by this diff are as follows: * Improve granularity of Python modules * Replace `test` with `correctness_test`, to make it clear that we are generating correctness tests. **Note that I haven't changed the top level target name `compute_graph_op_tests_bin` since I believe it would be too verbose. ghstack-source-id: 244283559 exported-using-ghexport Reviewed By: nathanaelsee Differential Revision: D63286131 fbshipit-source-id: 1177ea381e6381045f1c97491dd7ec006690f574
Summary: Pull Request resolved: #5561 ## Context Use the automatic test generation infrastructure to generate operator benchmarks. The overall concept is the same as the test generation; we just structure the generated code in the style of the google benchmark library instead of GTEST. ghstack-source-id: 244287193 Reviewed By: derekxu, nathanaelsee Differential Revision: D63286132 fbshipit-source-id: 25c379accf6664dfca8232db81772b638b41c758
Summary: Add separate tests for Ethos-U85 to all backend operator tests. Updated ethos-u-vela version to support more operators. Signed-off-by: Per Åstrand <[[email protected]](mailto:[email protected])> Signed-off-by: Tom Allsop <[[email protected]](mailto:[email protected])> Pull Request resolved: #5346 Reviewed By: manuelcandales Differential Revision: D62875027 Pulled By: digantdesai fbshipit-source-id: 3bf238d81957258ee93ae235d575beff8a575191
Summary: Pull Request resolved: #5565 Swap to using method meta so we can be finer grained about this check Reviewed By: dbort Differential Revision: D62983475 fbshipit-source-id: c4599c5ecad0409cd8b2670464c4e9e8809b49ad
Summary: Update the output of executor runner on the tutorial `add.pte` Pull Request resolved: #5452 Reviewed By: mergennachin Differential Revision: D63934482 Pulled By: dvorjackz fbshipit-source-id: 4933508d21f6965e0fcddd0caf5bf0ec1acc69ee
Summary: Pull Request resolved: #5909 Reviewed By: mergennachin Differential Revision: D63930714 Pulled By: dvorjackz fbshipit-source-id: 46134ac5750315606dffbf246da55e712dd192e8
Summary: The script didn't handle invalid JSON well when the benchmark results is empty https://github.com/pytorch/executorch/actions/runs/11197448597/job/31128464106. It's better to upload whatever it finds instead of crashing in this case. Looking a bit close at the benchmark jobs, there are some cases where the `benchmark_results.json` file is empty. I'm not sure why yet, but it can be fixed in another PR: * https://github.com/pytorch/executorch/actions/runs/11197448597/job/31127998979#step:15:286 (stories110M, qnn) * https://github.com/pytorch/executorch/actions/runs/11197448597/job/31127999221#step:15:946 (vit, xnnpack) The test spec already waits for 3 minutes to see if the file is there before giving up Pull Request resolved: #5916 Reviewed By: guangy10 Differential Revision: D63950793 Pulled By: huydhn fbshipit-source-id: 160d1465395c025e028b0e2cb69b5837a8f0208a
Summary: Pull Request resolved: #5882 Lots of things are redundant and a few need to move to utils. Subsequent changes will split the export function and separate the run part. Main changes: - call `fuse_pt2` after `convert_pt2` instead of `quantize_pt2`, and avoid calling `convert_pt2` twice - move `print_ops_info` into `export_to_cadence` - remove the need to call `export_to_edge` in `export_model` - move the serialization utils to `utils.py` Reviewed By: zonglinpeng Differential Revision: D63795843 fbshipit-source-id: 7eb482b0daccf64d3f1ca73ffb5e5148584a6678
Summary: - e2e script for https://github.com/apple/ml-fastvit (fastvit_s18) - add pass to handle mismatched tensor shape for broadcast ops when doing layout transform - add ParamObserver for params with lots of outliers - refactor & breakage fix Pull Request resolved: #5543 Reviewed By: kimishpatel Differential Revision: D63965451 Pulled By: cccclai fbshipit-source-id: 40cd85f60a8a539e6600cac1bfe16cdac4bb0465
Summary: Pull Request resolved: #5903 As title, they were tested in more recent version. 2.13 likely breaks Reviewed By: mergennachin Differential Revision: D63922556 fbshipit-source-id: 926fe6fd172d65f02d470f689016c687f1504647
Summary: Pull Request resolved: #5935 . Reviewed By: kirklandsign Differential Revision: D63988442 fbshipit-source-id: a0517166adc9ef3a1b2ee26d6003293b96ff7314
Summary: Pull Request resolved: #5868 the schema file copy is an actual AOT step. It's a workaround but it has been here a while, probably easier to have it as part of the build script instead of manually doing it everytime Reviewed By: kirklandsign Differential Revision: D63878765 fbshipit-source-id: 7fc4b848b72e2859364e21cbea3dd1b433d6cd8e
Summary: Pull Request resolved: #5861 We should use this option during exporting 1B/3B models as bf16 because KVCache is always fp32. Otherwise, we see regressed performance for 1B/3B in bf16 format. ghstack-source-id: 246391007 Reviewed By: mergennachin Differential Revision: D63871048 fbshipit-source-id: 6b3ff80dbc689a04c70e2fcc5c98698bb74f899b
Summary: Fix ExecuTorch capitalization. Pull Request resolved: #5945 Reviewed By: kirklandsign Differential Revision: D63997153 Pulled By: shoumikhin fbshipit-source-id: 5a3bd82d295c709ff9217b6325582e5cc2fab09c
Summary: Pull Request resolved: #5879 Port from LiteInterpreter's [`flip.glsl`](https://github.com/pytorch/pytorch/blob/main/aten/src/ATen/native/vulkan/glsl/flip.glsl). ``` - func: flip(Tensor self, int[] dims) -> Tensor ``` Will use this to verify AHB image is interpreted correctly. ghstack-source-id: 246657861 Reviewed By: SS-JIA Differential Revision: D63843843 fbshipit-source-id: e48ad206e7e563538142fa587b3acaf9c2919972
Summary: Pull Request resolved: #5799 ## Context As title, this diff adds an implementation for a fused SDPA + KV-Cache update operator which will be used in LLaMA models. Currently the SDPA portion of the operator is implemented via it's consituent operators, but a future optimization opportunity would be to implement a single flash attention shader. ## Reference Implementation For future reference, a reference implementation of the SDPA + KV cache update mechanism is shown below. This reference implementation was originally used to check intermediate outputs but in the end I decided to compare against the `sdpa_with_kv_cache` operator in `extension/llm` for simplicity. ``` at::Tensor convert_boolean_attn_mask( const at::Tensor& attn_mask, caffe2::TypeMeta dtype) { // Convert boolean mask to additive mask; need to invert mask to indicate what // to mask *out*. if (attn_mask.dtype() == at::kBool) { return at::where( attn_mask.logical_not(), -std::numeric_limits<double>::infinity(), at::scalar_tensor( 0.0, at::TensorOptions().dtype(dtype).device(attn_mask.device()))); } // Otherwise, attn_mask represents an additive attention tensor return attn_mask; } at::Tensor construct_attention_mask( const at::Tensor& q, const at::Tensor& k_cache, const int start_pos) { const int max_seq_len = k_cache.size(1); const int seq_len = q.size(1); at::Tensor attn_mask_base = at::ones({max_seq_len, start_pos + seq_len}, q.options().dtype(at::kBool)) .tril(); at::Tensor attn_mask_sliced = at::slice(attn_mask_base, 0, start_pos, start_pos + seq_len); attn_mask_sliced = convert_boolean_attn_mask(attn_mask_sliced, q.dtype()); return attn_mask_sliced; } std::vector<at::Tensor> sdpa_reference_impl( const at::Tensor& q_projected, const at::Tensor& k_projected, const at::Tensor& v_projected, at::Tensor& key_cache, at::Tensor& value_cache, const int64_t start_pos, const int64_t seq_len, const c10::optional<at::Tensor> __attn_mask_ignored, const double dropout_p, const bool is_causal, const c10::optional<double> scale) { at::Tensor attn_mask = construct_attention_mask(q_projected, key_cache, start_pos); at::Tensor key_cache_updated = at::slice_scatter( key_cache, k_projected, 1, start_pos, start_pos + k_projected.size(1)); at::Tensor value_cache_updated = at::slice_scatter( value_cache, v_projected, 1, start_pos, start_pos + v_projected.size(1)); at::Tensor key_cache_sliced = at::slice(key_cache_updated, 1, 0, start_pos + q_projected.size(1)); at::Tensor value_cache_sliced = at::slice(value_cache_updated, 1, 0, start_pos + q_projected.size(1)); at::Tensor q_transposed = q_projected.transpose(1, 2); at::Tensor k_transposed = key_cache_sliced.transpose(1, 2); at::Tensor v_transposed = value_cache_sliced.transpose(1, 2); // Skip doing repeat_interleave; assume that num_attention_heads == // num_kv_heads float scale_factor = 1.0 / sqrt(q_transposed.size(-1)); at::Tensor k_transposed_2 = k_transposed.transpose(-2, -1); at::Tensor attn_weight_prescale = at::matmul(q_transposed, k_transposed_2); at::Tensor attn_weight = attn_weight_prescale * scale_factor + attn_mask; at::Tensor attn_weight_softmax = at::softmax(attn_weight, -1); at::Tensor out = at::matmul(attn_weight_softmax, v_transposed); return { out.transpose(1, 2), key_cache_sliced, value_cache_sliced, q_transposed, k_transposed, v_transposed, k_transposed_2, attn_weight_prescale, attn_weight, attn_weight_softmax, out, }; } ``` ghstack-source-id: 246640547 Reviewed By: kimishpatel Differential Revision: D63724114 fbshipit-source-id: c85afc2f8eade8e0ac6e348eabbe608e5a0efce6
Summary: Changing `xnnpack.passes` to `xnnpack._passes` to indicate that these passes are not covered under the API stability guarantee. Pull Request resolved: #5917 Reviewed By: Olivia-liu, helunwencser Differential Revision: D63925008 fbshipit-source-id: 3d9f13c0a3bd61c66d07cebd62047a3e24f8af1d
Summary: Fix XNNPack tutorial to use `source_fn_stack` instead of `source_fn`. Pull Request resolved: #5948 Reviewed By: dvorjackz Differential Revision: D63950962 fbshipit-source-id: 5b4ced1c7edee4f5d60e9bffb8ab7a4a82788fcb
Summary: Add missing `export_for_training` import in bundledio tutorial. Pull Request resolved: #5949 Reviewed By: Olivia-liu, dvorjackz Differential Revision: D63951201 fbshipit-source-id: 42648d6d3e09edfd77714cf4a474d64dcd6c5e07
Summary: Bumps [numpy](https://github.com/numpy/numpy) from 1.21.3 to 1.22.0. <details> <summary>Release notes</summary> <p><em>Sourced from <a href="https://github.com/numpy/numpy/releases">numpy's releases</a>.</em></p> <blockquote> <h2>v1.22.0</h2> <h1>NumPy 1.22.0 Release Notes</h1> <p>NumPy 1.22.0 is a big release featuring the work of 153 contributors spread over 609 pull requests. There have been many improvements, highlights are:</p> <ul> <li>Annotations of the main namespace are essentially complete. Upstream is a moving target, so there will likely be further improvements, but the major work is done. This is probably the most user visible enhancement in this release.</li> <li>A preliminary version of the proposed Array-API is provided. This is a step in creating a standard collection of functions that can be used across application such as CuPy and JAX.</li> <li>NumPy now has a DLPack backend. DLPack provides a common interchange format for array (tensor) data.</li> <li>New methods for <code>quantile</code>, <code>percentile</code>, and related functions. The new methods provide a complete set of the methods commonly found in the literature.</li> <li>A new configurable allocator for use by downstream projects.</li> </ul> <p>These are in addition to the ongoing work to provide SIMD support for commonly used functions, improvements to F2PY, and better documentation.</p> <p>The Python versions supported in this release are 3.8-3.10, Python 3.7 has been dropped. Note that 32 bit wheels are only provided for Python 3.8 and 3.9 on Windows, all other wheels are 64 bits on account of Ubuntu, Fedora, and other Linux distributions dropping 32 bit support. All 64 bit wheels are also linked with 64 bit integer OpenBLAS, which should fix the occasional problems encountered by folks using truly huge arrays.</p> <h2>Expired deprecations</h2> <h3>Deprecated numeric style dtype strings have been removed</h3> <p>Using the strings <code>"Bytes0"</code>, <code>"Datetime64"</code>, <code>"Str0"</code>, <code>"Uint32"</code>, and <code>"Uint64"</code> as a dtype will now raise a <code>TypeError</code>.</p> <p>(<a href="https://redirect.github.com/numpy/numpy/pull/19539">gh-19539</a>)</p> <h3>Expired deprecations for <code>loads</code>, <code>ndfromtxt</code>, and <code>mafromtxt</code> in npyio</h3> <p><code>numpy.loads</code> was deprecated in v1.15, with the recommendation that users use <code>pickle.loads</code> instead. <code>ndfromtxt</code> and <code>mafromtxt</code> were both deprecated in v1.17 - users should use <code>numpy.genfromtxt</code> instead with the appropriate value for the <code>usemask</code> parameter.</p> <p>(<a href="https://redirect.github.com/numpy/numpy/pull/19615">gh-19615</a>)</p> </blockquote> <p>... (truncated)</p> </details> <details> <summary>Commits</summary> <ul> <li><a href="https://github.com/numpy/numpy/commit/4adc87dff15a247e417d50f10cc4def8e1c17a03"><code>4adc87d</code></a> Merge pull request <a href="https://redirect.github.com/numpy/numpy/issues/20685">#20685</a> from charris/prepare-for-1.22.0-release</li> <li><a href="https://github.com/numpy/numpy/commit/fd66547557f57c430d41be2fc0764f74a62e8ccf"><code>fd66547</code></a> REL: Prepare for the NumPy 1.22.0 release.</li> <li><a href="https://github.com/numpy/numpy/commit/125304b035effcd82e366e601b102e7347eaa9ba"><code>125304b</code></a> wip</li> <li><a href="https://github.com/numpy/numpy/commit/c283859128b1a4b57014581570a23ed7950a24ea"><code>c283859</code></a> Merge pull request <a href="https://redirect.github.com/numpy/numpy/issues/20682">#20682</a> from charris/backport-20416</li> <li><a href="https://github.com/numpy/numpy/commit/5399c03d4a069fe81a1616be0184c9749d7271ee"><code>5399c03</code></a> Merge pull request <a href="https://redirect.github.com/numpy/numpy/issues/20681">#20681</a> from charris/backport-20954</li> <li><a href="https://github.com/numpy/numpy/commit/f9c45f8ebf31340b1a5a0371bfca25afcfc4794e"><code>f9c45f8</code></a> Merge pull request <a href="https://redirect.github.com/numpy/numpy/issues/20680">#20680</a> from charris/backport-20663</li> <li><a href="https://github.com/numpy/numpy/commit/794b36f7e1bf2a8c42774ab0db86a74bd32f674b"><code>794b36f</code></a> Update armccompiler.py</li> <li><a href="https://github.com/numpy/numpy/commit/d93b14e3d7abaa1d837825e51671f817788e120f"><code>d93b14e</code></a> Update test_public_api.py</li> <li><a href="https://github.com/numpy/numpy/commit/7662c0789cc6a70d5ad4d950ee2e95f3afef7df6"><code>7662c07</code></a> Update <strong>init</strong>.py</li> <li><a href="https://github.com/numpy/numpy/commit/311ab52488a7d096ac3bc4c2de0fdae17ecd13ef"><code>311ab52</code></a> Update armccompiler.py</li> <li>Additional commits viewable in <a href="https://github.com/numpy/numpy/compare/v1.21.3...v1.22.0">compare view</a></li> </ul> </details> <br /> [](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) You can trigger a rebase of this PR by commenting `dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) --- <details> <summary>Dependabot commands and options</summary> <br /> You can trigger Dependabot actions by commenting on this PR: - `dependabot rebase` will rebase this PR - `dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `dependabot merge` will merge this PR after your CI passes on it - `dependabot squash and merge` will squash and merge this PR after your CI passes on it - `dependabot cancel merge` will cancel a previously requested merge and block automerging - `dependabot reopen` will reopen this PR if it is closed - `dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `dependabot show <dependency name> ignore conditions` will show all of the ignore conditions of the specified dependency - `dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself) You can disable automated security fix PRs for this repo from the [Security Alerts page](https://github.com/pytorch/executorch/network/alerts). </details> > **Note** > Automatic rebases have been disabled on this pull request as it has been open for over 30 days. Pull Request resolved: #4514 Reviewed By: tarun292 Differential Revision: D64004841 Pulled By: larryliu0820 fbshipit-source-id: 30509aaa9b7b74ab7377e66430447cdd545c03e5
Summary: Pull Request resolved: #5952 Reviewed By: kirklandsign Differential Revision: D64005568 fbshipit-source-id: 7cd8ab9fe33d5745064aca7720d34ce1d9f4f06b
Summary: Pull Request resolved: #5953 Improve the logic used to generate a debug handle for delegate nodes. The existing logic is causing conflicts with other nodes debug handles. Reviewed By: Olivia-liu Differential Revision: D63779522 fbshipit-source-id: ff1583ee21bf05748c91a1c8b8e489996f80d7d9
Summary: Pull Request resolved: #5904 Move the Arm backend out of the `torch::` namespace, and update to avoid using the `torch::` or `exec_aten::` namespaces. Also, move the VelaBinStream code into a namespace to avoid declaring symbols in the global namespace. Reviewed By: mergennachin Differential Revision: D63923290 fbshipit-source-id: a98e4c1ede8072a9fd96fb7fc2e2fcc4e82bc28a
Summary: * Rename `_portable_lib.cpython-3.<distribution info>.so` to `_portable_lib.so` so it can be found by CMake `find_library()`. This can be achieved by setting `SETUPTOOLS_EXT_SUFFIX`. * Since `executorch-config.cmake` is also being used to find installed libraries such as `executorch.a`, `xnnpack_backend.a`, add a condition to tell if `executorch-config.cmake` is being used in cmake-out or site-packages. Pull Request resolved: #5961 Reviewed By: metascroy Differential Revision: D64014291 Pulled By: larryliu0820 fbshipit-source-id: 2757f2883d3f836e9efd45676f792c12f742e63d
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
CLA Signed
This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
No description provided.