Skip to content

Commit b93b69a

Browse files
authored
New Projects
New Projects - Edge AI and Mobile
2 parents e062481 + 21f31ae commit b93b69a

11 files changed

+1292
-0
lines changed
Lines changed: 76 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,76 @@
1+
---
2+
title: "SME2 on vivo X300: Mobile Edge AI Projects for multi-modal inference, built on Arm Lumex"
3+
description: "Leverage the latest SME2 (Scalable Matrix Extension 2) available on the newest vivo X300 smartphones (built on Arm Lumex CSS) for advanced image/video, audio and text processing edge AI. Explore how SME2, via KleidiAI, enables larger matrix workloads, higher throughput, and novel applications on device without cloud connectivity required."
4+
subjects:
5+
- "ML"
6+
- "Performance and Architecture"
7+
- "Libraries"
8+
requires-team:
9+
- "No"
10+
platform:
11+
- "Mobile, Graphics, and Gaming"
12+
- "AI"
13+
- "IoT"
14+
sw-hw:
15+
- "Software"
16+
support-level:
17+
- "Self-Service"
18+
- "Arm Ambassador Support"
19+
publication-date: 2025-11-27
20+
license:
21+
status:
22+
- "Published"
23+
---
24+
25+
## Description
26+
27+
### Why is this important?
28+
29+
SME2 (Scalable Matrix Extension 2) is the latest CPU extension on Arm Lumex CSS. Designed to accelerate matrix-oriented compute workloads directly on device, SME2 improves AI/ML performance. This is by accelerating models that rely on operations like matrix multiplication, common in transformers, convolutional neural networks (CNNs), and large language models (LLMs). Via KleidiAI, SME2 is seamlessly integrated into frameworks such as ExecuTorch, LiteRT, ONNX Runtime so it is automatically leveraged for applications depending on whether SME2 is present on the host device.
30+
31+
[SME2](https://www.arm.com/technologies/sme2)
32+
33+
The vivo X300 is built on Arm Lumex. SME2 now enables AI compute that previously was too heavy or inaccessible on mobile. Developers can now utilise these advancements to deliver advanced applications on-device, reducing latency, increasing data privacy, and unlocking novel use-cases.
34+
35+
[vivo X300, built on Arm Lumex](https://www.arm.com/company/success-library/vivo-x300-smartphones)
36+
37+
### Project Summary
38+
39+
Select a **mobile edge AI application** that benefits from large matrix operations, multi-modal fusion, or transformer-based processing enabled by SME2. Build and optimize a proof-of-concept application on a vivo X300 phone or other device supporting SME2.
40+
41+
Example project areas:
42+
- Real-time video semantic segmentation (e.g., background removal + AR compositing)
43+
- Live object detection + natural-language description (text summary of what the camera sees)
44+
- Multi-sensor fusion (camera + IMU + microphone) for gesture + voice recognition
45+
- On-device lightweight LLM or encoder-only transformer processing for mobile assistants
46+
47+
Identify a model architecture that maps to wide matrix operations (e.g., ViT, MLP-Mixer, multi-branch CNN with large FC layers). Utilise a mobile-friendly framework (e.g., ExecuTorch, LiteRT, ONNX Runtime, MediaPipe) to leverage SME2 optimizations. Optimize quantization, memory layout, and verify that the large matrix multiplications get scheduled efficiently on the SME2-enabled CPU. Build a mobile app (Android) that executes the model and utilises it for a compelling use-case.
48+
49+
Utilise the resources and learning paths below and create an exciting and challenging application. Optionally, you could also compare performance vs a reference phone without SME2.
50+
51+
---
52+
53+
## Resources from Arm and our partners
54+
55+
- Arm Developer: [Launchpad - Mobile AI](https://developer.arm.com/mobile-graphics-and-gaming/ai-mobile)
56+
- Learning Path: [Mobile AI/ML Performance Profiling](https://learn.arm.com/learning-paths/mobile-graphics-and-gaming/profiling-ml-on-arm/)
57+
- Learning Path: [Build an Android chat app with Llama, KleidiAI, ExecuTorch, and XNNPACK](https://learn.arm.com/learning-paths/mobile-graphics-and-gaming/build-llama3-chat-android-app-using-executorch-and-xnnpack/)
58+
- Learning Path: [Vision LLM Inference on Android with KleidiAI](https://learn.arm.com/learning-paths/mobile-graphics-and-gaming/vision-llm-inference-on-android-with-kleidiai-and-mnn/)
59+
- Learning Path: [Build a Hands-Free Selfie Android Application with MediaPipe](https://learn.arm.com/learning-paths/mobile-graphics-and-gaming/build-android-selfie-app-using-mediapipe-multimodality/)
60+
- Repository: [AI on Arm course](https://github.com/arm-university/AI-on-Arm)
61+
- Arm / Cambridge University edX course: [AI at the Edge on Arm (Mobile)](https://www.edx.org/learn/computer-science/arm-education-ai-at-the-edge-on-arm)
62+
63+
---
64+
65+
## Support Level
66+
67+
This project is designed to be self-serve but comes with opportunity of some community support from Arm Ambassadors, who are part of the Arm Developer program. If you are not already part of our program, [click here to join](https://www.arm.com/resources/developer-program?#register).
68+
69+
## Benefits
70+
71+
Standout project contributions will result in digital badges for CV building, recognised by Arm Talent Acquisition. We are currently discussing with national agencies the potential for funding streams for Arm Developer Labs projects, which would flow to you, not us.
72+
73+
74+
To receive the benefits, you must show us your project through our [online form](https://forms.office.com/e/VZnJQLeRhD). Please do not include any confidential information in your contribution. Additionally if you are affiliated with an academic institution, please ensure you have the right to share your material.
75+
76+
---
Lines changed: 80 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,80 @@
1+
---
2+
title: "Edge AI with NPU: always-on-AI with ExecuTorch on Cortex-M55 + Ethos-U85 → Cortex-A"
3+
description: "The vision of Edge AI compute is to embed low-power intelligent sensing, perception, and decision systems everywhere. A low-power always-on-AI island continuously monitors sensory inputs to detect triggers. When a trigger is detected, it wakes up a more capable processor to carry out high-value inference, interaction, or control tasks."
4+
subjects:
5+
- "ML"
6+
- "Performance and Architecture"
7+
- "Embedded Linux"
8+
- "RTOS Fundamentals"
9+
requires-team:
10+
- "No"
11+
platform:
12+
- "IoT"
13+
- "Embedded and Microcontrollers"
14+
- "AI"
15+
sw-hw:
16+
- "Software"
17+
- "Hardware"
18+
support-level:
19+
- "Self-Service"
20+
- "Arm Ambassador Support"
21+
publication-date: 2025-11-27
22+
license:
23+
status:
24+
- "Published"
25+
donation:
26+
---
27+
28+
![educate_on_arm](../../images/Educate_on_Arm_banner.png)
29+
30+
31+
## Description
32+
33+
**Why is this important?**
34+
35+
The vision of Edge AI compute is to embed intelligent low-power sensing, perception, and decision systems everywhere (in homes, wearables, infrastructure) so devices can react to subtle cues, adapt to context, and wake higher-power systems only when needed. Rather than sending everything to the cloud or running full-scale models continuously, this Edge AI system operates as a layered hierarchy:
36+
- A low-power always-on-AI model continuously monitors sensory inputs (audio, motion, video) to detect triggers or anomalies.
37+
- When a trigger is detected, it wakes up a more capable processor (e.g. Cortex-A running a rich OS such as Linux) to carry out further tasks. This could be high-value inference, interaction, or control tasks. It could also involve connecting to other IoT devices or to a Neoverse cloud instance.
38+
39+
This architecture is key to bridging the gap between battery-constrained devices and rich AI services, making systems smarter, more efficient, and responsive without draining resources.
40+
41+
**Project Summary**
42+
43+
Using equipment such as the Alif Ensemble development kit (e.g. E6/E8, which includes Cortex-A, Cortex-M55, and Ethos-U85 cores - or E4 (M55+U85) + Raspberry Pi for Cortex-A), and the ExecuTorch framework, build an Edge AI prototype that implements:
44+
45+
1. A “wake-up” path: deploy a TOSA-compliant optimized model on the Cortex-M55 + Ethos-U85 pair to continuously monitor sensory signals (audio, motion, video) for wake-word, anomalies, or triggers.
46+
2. A subsequent workload path: when a trigger is detected, activate a Cortex-A core to perform more complex tasks, e.g. use an LLM optimised for CPU inference, connect to and manage other IoT devices, or connect to a Neoverse cloud instance for heavier inference.
47+
3. Evaluation and documentation: measure accuracy, latency, power consumption, robustness, and compare trade-offs between modalities (audio, video, motion). Demonstrate an end-to-end use case of your choice (e.g. smart assistant, anomaly alert system, gesture control, environment monitoring).
48+
49+
*Note that the Cortex-A32 included on the Alif DevKits will not be suitable for LLM inference. If using the onboard core for the project, target cloud/IoT connectivity. For LLM inference, consider connecting a Raspberry Pi 5 or similar.*
50+
51+
Example: Use a microphone input to detect “Hey Arm”. After wake-up, launch an optimised LLM on Raspberry Pi Cortex-A to answer questions or control local devices.
52+
53+
You are free to mix and match sensors, modalities, and tasks — as long as the core architecture (wake-on M55/U85, main task on A) is preserved.
54+
55+
Many of these DevKits come with additional Ethos-U55 NPUs onboard - feel free to be creative and distribute different tasks across the different NPUs - what use-cases and applications can you achieve?
56+
57+
## What will you use?
58+
You should either be familiar with, or willing to learn about, the following:
59+
- Programming: Python, C++, Embedded C
60+
- ExecuTorch, plus knowledge of model quantization, pruning, conversion. Use of the Vela compiler and TOSA.
61+
- Edge/Embedded development: bare-metal or RTOS (e.g. Zephyr), and embedded Linux (e.g. Yocto) or Raspberry Pi OS
62+
63+
64+
## Resources from Arm and our partners
65+
- Arm Developer: [Edge AI](https://developer.arm.com/edge-ai)
66+
- Learning Path: [Navigating Machine Learning with Ethos-U processors](https://learn.arm.com/learning-paths/microcontrollers/nav-mlek/)
67+
- Repository: [AI on Arm course](https://github.com/arm-university/AI-on-Arm)
68+
- Example Board: [Alif Ensemble DevKit E8](https://www.keil.arm.com/boards/alif-semiconductor-devkit-e8-gen-1-2558a7b/features/)
69+
- PyTorch Blog: [ExecuTorch support for Ethos-U85](https://pytorch.org/blog/pt-executorch-ethos-u85/)
70+
71+
## Support Level
72+
73+
This project is designed to be self-serve but comes with opportunity of some community support from Arm Ambassadors, who are part of the Arm Developer program. If you are not already part of our program, [click here to join](https://www.arm.com/resources/developer-program?#register).
74+
75+
## Benefits
76+
77+
Standout project contributions will result in digital badges for CV building, recognised by Arm Talent Acquisition. We are currently discussing with national agencies the potential for funding streams for Arm Developer Labs projects, which would flow to you, not us.
78+
79+
80+
To receive the benefits, you must show us your project through our [online form](https://forms.office.com/e/VZnJQLeRhD). Please do not include any confidential information in your contribution. Additionally if you are affiliated with an academic institution, please ensure you have the right to share your material.
Lines changed: 117 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,117 @@
1+
---
2+
title: "Ethos-U85 NPU Applications with TOSA Model Explorer: Exploring Next-Gen Edge AI Inference"
3+
description: "Push the limits of Edge AI by deploying the heaviest inference applications possible on Ethos-U85. Students will explore transformer-based and TOSA-optimized workloads that demonstrate performance levels on the next-gen of Ethos NPUs."
4+
subjects:
5+
- "ML"
6+
- "Performance and Architecture"
7+
requires-team:
8+
- "No"
9+
platform:
10+
- "IoT"
11+
- "Embedded and Microcontrollers"
12+
- "AI"
13+
sw-hw:
14+
- "Software"
15+
- "Hardware"
16+
support-level:
17+
- "Self-Service"
18+
- "Arm Ambassador Support"
19+
publication-date: 2025-11-27
20+
license:
21+
status:
22+
- "Published"
23+
donation:
24+
---
25+
26+
![educate_on_arm](../../images/Educate_on_Arm_banner.png)
27+
28+
## Description
29+
30+
**Why is this important?**
31+
32+
The Arm Ethos-U85 NPU represents a major leap in bringing *heavy inference* to constrained embedded systems. With its full transformer operator support, expanded MAC throughput, and native TOSA compatibility, the Ethos-U85 enables developers to deploy models and workloads that were previously too intensive for MCU-class devices.
33+
34+
This project challenges you to explore the boundaries of what’s possible on Ethos-U85. The goal is to demonstrate inference performance and model complexity that is now achievable due to the architectural improvements and transformer acceleration capabilities of the Ethos-U85.
35+
36+
[Ethos-U85 Launch](https://newsroom.arm.com/blog/ethos-u85)
37+
38+
**Project Summary**
39+
40+
Using hardware such as the Alif Ensemble E4/E6/E8 DevKits (all include Ethos-U85) or a comparable platform or Arm Fixed Virtual Platform Corstone-320, your task is to design and benchmark an advanced edge inference application that exploits the Ethos-U85’s compute and transformer capabilities.
41+
42+
Your project should include:
43+
44+
1. Model Deployment and Optimization
45+
Select a computationally intensive model — ideally transformer-based or multi-branch convolutional — and deploy it on the Ethos-U85 using:
46+
- The TOSA Model Explorer extension to inspect and adapt unsupported or experimental models for TOSA compliance.
47+
- The Vela compiler for optimization.
48+
49+
These tools can be used to:
50+
- Convert and visualize model graphs in TOSA format.
51+
- Identify unsupported operators.
52+
- Modify or substitute layers for compatibility using the Flatbuffers schema before re-exporting.
53+
- Run Vela for optimized compilation targeting Ethos-U85.
54+
55+
2. Application Demonstration
56+
Implement a working example that highlights the Ethos-U85’s strengths in real-world inference. Possible categories include:
57+
- Transformers on Edge: lightweight BERT, ViT, or audio transformers (e.g. speech or sound event classification).
58+
- High-resolution Vision: semantic segmentation, object detection on large input sizes, or multi-head perception networks.
59+
- Multi-modal Fusion: combining audio, image, or sensor streams for contextual understanding.
60+
61+
3. Analysis and Benchmarking
62+
Report quantitative results on:
63+
- Inference latency, throughput (FPS or tokens/s), and memory footprint.
64+
- Power efficiency under load (optional).
65+
- Comparative performance versus Ethos-U55/U65 (use available benchmarks for reference or utilise the other Ethos-U NPUs provided in the Alif DevKits).
66+
- The effect of TOSA optimization — demonstrate measurable improvements from graph conversion and operator fusion.
67+
68+
---
69+
70+
## What kind of projects should you target?
71+
72+
To clearly demonstrate the leap from Ethos-U55/U65 to U85, choose projects that meet at least one of the following criteria:
73+
74+
- Transformer-heavy architectures: e.g. attention blocks, transformer encoders, ViTs, or hybrid CNN+transformer models.
75+
- *Example:* an audio event detection transformer that must process longer sequences or higher-resolution spectrograms.
76+
- High-resolution or multi-branch networks: models with high input dimensionality or multiple processing paths that saturate NPU throughput.
77+
- *Example:* 512×512 semantic segmentation or multi-object detection.
78+
- Dense post-processing or large fully connected layers: cases where U55/U65 memory limits or MAC bandwidth previously restricted performance.
79+
- *Example:* large MLP heads or transformer token mixers.
80+
- Multi-modal pipelines: combining multiple sensor inputs (e.g. image + IMU + audio) where the NPU must maintain concurrency or shared intermediate representations.
81+
82+
The Ethos-U85 is ideal for projects where model performance is constrained by attention layers, large activations, or operator types that previously required fallback to the CPU. Use the Ethos-U85 to eliminate those fallbacks and achieve full-NPU execution of advanced topologies.
83+
84+
---
85+
86+
## What will you use?
87+
You should be familiar with, or willing to learn about:
88+
- Programming: Python, C/C++
89+
- ExecuTorch or TensorFlow Lite (Micro/LiteRT)
90+
- Techniques for optimising AI models for the edge (quantization, pruning, etc.)
91+
- Optimization Tools:
92+
- TOSA Model Explorer
93+
- .tflite to .tosa converter (if using Tensorflow rather than ExecuTorch)
94+
- Vela compiler for Ethos-U
95+
- Bare-metal or RTOS (e.g., Zephyr)
96+
97+
---
98+
99+
## Resources from Arm and our partners
100+
- Arm Developer: [Edge AI](https://developer.arm.com/edge-ai)
101+
- Learning Path: [Navigating Machine Learning with Ethos-U processors](https://learn.arm.com/learning-paths/microcontrollers/nav-mlek/)
102+
- Repository: [AI on Arm course](https://github.com/arm-university/AI-on-Arm)
103+
- Example Board: [Alif Ensemble DevKit E8](https://www.keil.arm.com/boards/alif-semiconductor-devkit-e8-gen-1-2558a7b/features/)
104+
- Documentation: [TOSA Specification](https://www.mlplatform.org/tosa/), [TOSA Model Explorer](https://github.com/arm/tosa-adapter-model-explorer), and [TOSA Reference Model](https://gitlab.arm.com/tosa/tosa-reference-model)
105+
- PyTorch Blog: [ExecuTorch support for Ethos-U85](https://pytorch.org/blog/pt-executorch-ethos-u85/)
106+
---
107+
108+
## Support Level
109+
110+
This project is designed to be self-serve but comes with opportunity of some community support from Arm Ambassadors, who are part of the Arm Developer program. If you are not already part of our program, [click here to join](https://www.arm.com/resources/developer-program?#register).
111+
112+
## Benefits
113+
114+
Standout project contributions will result in digital badges for CV building, recognised by Arm Talent Acquisition. We are currently discussing with national agencies the potential for funding streams for Arm Developer Labs projects, which would flow to you, not us.
115+
116+
117+
To receive the benefits, you must show us your project through our [online form](https://forms.office.com/e/VZnJQLeRhD). Please do not include any confidential information in your contribution. Additionally if you are affiliated with an academic institution, please ensure you have the right to share your material.

0 commit comments

Comments
 (0)