Skip to content

Commit ba0958a

Browse files
cmodi-metafacebook-github-bot
authored andcommitted
Add Llama 3.2 and subsections to Example dir README (#5661)
Summary: Pull Request resolved: #5661 Reviewed By: mergennachin Differential Revision: D63416533 Pulled By: cmodi-meta fbshipit-source-id: 25259965bf697cfa36cb7d78ac22e214b659968b
1 parent 984986e commit ba0958a

File tree

1 file changed

+16
-14
lines changed

1 file changed

+16
-14
lines changed

examples/README.md

Lines changed: 16 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -31,43 +31,45 @@ examples
3131

3232
A user's journey may commence by exploring the demos located in the [`portable/`](./portable) directory. Here, you will gain insights into the fundamental end-to-end workflow to generate a binary file from a ML model in [portable mode](../docs/source/concepts.md##portable-mode-lean-mode) and run it on the ExecuTorch runtime.
3333

34-
## Demo of Llama 2 and Llama 3
34+
## Demos Apps
3535

36-
[This page](./models/llama2/README.md) demonstrates how to run Llama 2 7B and Llama 3 8B models on mobile via ExecuTorch. We use XNNPACK to accelerate the performance and 4-bit groupwise PTQ quantization to fit the model on Android and iOS mobile phones.
36+
Explore mobile apps with ExecuTorch models integrated and deployable on [Android](./demo-apps/android) and [iOS]((./demo-apps/apple_ios)). This provides end-to-end instructions on how to export Llama models, load on device, build the app, and run it on device.
3737

38-
## Demo of Llava1.5 7B
38+
For specific details related to models and backend, you can explore the various subsections.
39+
40+
### Llama Models
41+
42+
[This page](./models/llama2/README.md) demonstrates how to run Llama 3.2 (1B, 3B), Llama 3.1 (8B), Llama 3 (8B), and Llama 2 7B models on mobile via ExecuTorch. We use XNNPACK, QNNPACK, MediaTek, and MPS to accelerate the performance and 4-bit groupwise PTQ quantization to fit the model on Android and iOS mobile phones.
43+
44+
### Llava1.5 7B
3945

4046
[This page](./models/llava/README.md) demonstrates how to run [Llava 1.5 7B](https://github.com/haotian-liu/LLaVA) model on mobile via ExecuTorch. We use XNNPACK to accelerate the performance and 4-bit groupwise PTQ quantization to fit the model on Android and iOS mobile phones.
4147

42-
## Demo of Selective Build
48+
### Selective Build
4349

4450
To understand how to deploy the ExecuTorch runtime with optimization for binary size, explore the demos available in the [`selective_build/`](./selective_build) directory. These demos are specifically designed to illustrate the [Selective Build](../docs/source/kernel-library-selective_build.md), offering insights into reducing the binary size while maintaining efficiency.
4551

46-
## Demo of ExecuTorch Developer Tools
52+
### Developer Tools
4753

4854
You will find demos of [ExecuTorch Developer Tools](./devtools/) in the [`devtools/`](./devtools/) directory. The examples focuses on exporting and executing BundledProgram for ExecuTorch model verification and ETDump for collecting profiling and debug data.
4955

50-
## Demo Apps
51-
52-
Explore mobile apps with ExecuTorch models integrated and deployable on Android and iOS in the [`demo-apps/android/`](./demo-apps/android) and [`demo-apps/apple_ios/`](./demo-apps/apple_ios) directories, respectively.
53-
54-
## Demo of XNNPACK delegation
56+
### XNNPACK delegation
5557

5658
The demos in the [`xnnpack/`](./xnnpack) directory provide valuable insights into the process of lowering and executing an ExecuTorch model with built-in performance enhancements. These demos specifically showcase the workflow involving [XNNPACK backend](https://github.com/pytorch/executorch/tree/main/backends/xnnpack) delegation and quantization.
5759

58-
## Demo of ExecuTorch Apple Backend
60+
### Apple Backend
5961

6062
You will find demos of [ExecuTorch Core ML Backend](./apple/coreml/) in the [`apple/coreml/`](./apple/coreml) directory and [MPS Backend](./apple/mps/) in the [`apple/mps/`](./apple/mps) directory.
6163

62-
## Demo of ExecuTorch on ARM Cortex-M55 + Ethos-U55
64+
### ARM Cortex-M55 + Ethos-U55 Backend
6365

6466
The [`arm/`](./arm) directory contains scripts to help you run a PyTorch model on a ARM Corstone-300 platform via ExecuTorch.
6567

66-
## Demo of ExecuTorch QNN Backend
68+
### QNN Backend
6769

6870
You will find demos of [ExecuTorch QNN Backend](./qualcomm) in the [`qualcomm/`](./qualcomm) directory.
6971

70-
## Demo of ExecuTorch on Cadence HiFi4 DSP
72+
### Cadence HiFi4 DSP
7173

7274
The [`Cadence/`](./cadence) directory hosts a demo that showcases the process of exporting and executing a model on Xtensa Hifi4 DSP. You can utilize [this tutorial](../docs/source/build-run-xtensa.md) to guide you in configuring the demo and running it.
7375

0 commit comments

Comments
 (0)