Skip to content

Commit c609f11

Browse files
Add Llama 3.2 and subsections to Example dir README (#5701)
Add Llama 3.2 and subsections to Example dir README (#5661) Summary: Pull Request resolved: #5661 Reviewed By: mergennachin Differential Revision: D63416533 Pulled By: cmodi-meta fbshipit-source-id: 25259965bf697cfa36cb7d78ac22e214b659968b (cherry picked from commit ba0958a) Co-authored-by: cmodi-meta <[email protected]>
1 parent cd1fa99 commit c609f11

File tree

1 file changed

+19
-13
lines changed

1 file changed

+19
-13
lines changed

examples/README.md

Lines changed: 19 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -31,39 +31,45 @@ examples
3131

3232
A user's journey may commence by exploring the demos located in the [`portable/`](./portable) directory. Here, you will gain insights into the fundamental end-to-end workflow to generate a binary file from a ML model in [portable mode](../docs/source/concepts.md##portable-mode-lean-mode) and run it on the ExecuTorch runtime.
3333

34-
## Demo of Llama 2 and Llama 3
34+
## Demos Apps
3535

36-
[This page](./models/llama2/README.md) demonstrates how to run Llama 2 7B and Llama 3 8B models on mobile via ExecuTorch. We use XNNPACK to accelerate the performance and 4-bit groupwise PTQ quantization to fit the model on Android and iOS mobile phones.
36+
Explore mobile apps with ExecuTorch models integrated and deployable on [Android](./demo-apps/android) and [iOS]((./demo-apps/apple_ios)). This provides end-to-end instructions on how to export Llama models, load on device, build the app, and run it on device.
3737

38-
## Demo of Selective Build
38+
For specific details related to models and backend, you can explore the various subsections.
3939

40-
To understand how to deploy the ExecuTorch runtime with optimization for binary size, explore the demos available in the [`selective_build/`](./selective_build) directory. These demos are specifically designed to illustrate the [Selective Build](../docs/source/kernel-library-selective_build.md), offering insights into reducing the binary size while maintaining efficiency.
40+
### Llama Models
4141

42-
## Demo of ExecuTorch Developer Tools
42+
[This page](./models/llama2/README.md) demonstrates how to run Llama 3.2 (1B, 3B), Llama 3.1 (8B), Llama 3 (8B), and Llama 2 7B models on mobile via ExecuTorch. We use XNNPACK, QNNPACK, MediaTek, and MPS to accelerate the performance and 4-bit groupwise PTQ quantization to fit the model on Android and iOS mobile phones.
4343

44-
You will find demos of [ExecuTorch Developer Tools](./devtools/) in the [`devtools/`](./devtools/) directory. The examples focuses on exporting and executing BundledProgram for ExecuTorch model verification and ETDump for collecting profiling and debug data.
44+
### Llava1.5 7B
4545

46-
## Demo Apps
46+
[This page](./models/llava/README.md) demonstrates how to run [Llava 1.5 7B](https://github.com/haotian-liu/LLaVA) model on mobile via ExecuTorch. We use XNNPACK to accelerate the performance and 4-bit groupwise PTQ quantization to fit the model on Android and iOS mobile phones.
4747

48-
Explore mobile apps with ExecuTorch models integrated and deployable on Android and iOS in the [`demo-apps/android/`](./demo-apps/android) and [`demo-apps/apple_ios/`](./demo-apps/apple_ios) directories, respectively.
48+
### Selective Build
49+
50+
To understand how to deploy the ExecuTorch runtime with optimization for binary size, explore the demos available in the [`selective_build/`](./selective_build) directory. These demos are specifically designed to illustrate the [Selective Build](../docs/source/kernel-library-selective_build.md), offering insights into reducing the binary size while maintaining efficiency.
51+
52+
### Developer Tools
53+
54+
You will find demos of [ExecuTorch Developer Tools](./devtools/) in the [`devtools/`](./devtools/) directory. The examples focuses on exporting and executing BundledProgram for ExecuTorch model verification and ETDump for collecting profiling and debug data.
4955

50-
## Demo of XNNPACK delegation
56+
### XNNPACK delegation
5157

5258
The demos in the [`xnnpack/`](./xnnpack) directory provide valuable insights into the process of lowering and executing an ExecuTorch model with built-in performance enhancements. These demos specifically showcase the workflow involving [XNNPACK backend](https://github.com/pytorch/executorch/tree/main/backends/xnnpack) delegation and quantization.
5359

54-
## Demo of ExecuTorch Apple Backend
60+
### Apple Backend
5561

5662
You will find demos of [ExecuTorch Core ML Backend](./apple/coreml/) in the [`apple/coreml/`](./apple/coreml) directory and [MPS Backend](./apple/mps/) in the [`apple/mps/`](./apple/mps) directory.
5763

58-
## Demo of ExecuTorch on ARM Cortex-M55 + Ethos-U55
64+
### ARM Cortex-M55 + Ethos-U55 Backend
5965

6066
The [`arm/`](./arm) directory contains scripts to help you run a PyTorch model on a ARM Corstone-300 platform via ExecuTorch.
6167

62-
## Demo of ExecuTorch QNN Backend
68+
### QNN Backend
6369

6470
You will find demos of [ExecuTorch QNN Backend](./qualcomm) in the [`qualcomm/`](./qualcomm) directory.
6571

66-
## Demo of ExecuTorch on Cadence HiFi4 DSP
72+
### Cadence HiFi4 DSP
6773

6874
The [`Cadence/`](./cadence) directory hosts a demo that showcases the process of exporting and executing a model on Xtensa Hifi4 DSP. You can utilize [this tutorial](../docs/source/build-run-xtensa.md) to guide you in configuring the demo and running it.
6975

0 commit comments

Comments
 (0)