Skip to content

Commit dd8d5be

Browse files
cmodi-metafacebook-github-bot
authored andcommitted
Cleanup xnnpack_README.md (pytorch#5662)
Summary: Pull Request resolved: pytorch#5662 Reviewed By: kirklandsign Differential Revision: D63416131 Pulled By: cmodi-meta fbshipit-source-id: f7b37a7ee78fc556072d47ae9ab884e23826d1a2
1 parent f3fa9fa commit dd8d5be

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

examples/demo-apps/apple_ios/LLaMA/docs/delegates/xnnpack_README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# Building Llama iOS Demo for XNNPack Backend
22

3-
**[UPDATE - 09/25]** We have added support for running [Llama 3.2 models](#for-llama-3.2-1b-and-3b-models) on the XNNPack backend. We currently support inference on their original data type (BFloat16).
3+
**[UPDATE - 09/25]** We have added support for running [Llama 3.2 models](#for-llama-32-1b-and-3b-models) on the XNNPack backend. We currently support inference on their original data type (BFloat16).
44

55
This tutorial covers the end to end workflow for building an iOS demo app using XNNPack backend on device.
66
More specifically, it covers:
@@ -49,7 +49,7 @@ sh examples/models/llama2/install_requirements.sh
4949

5050
### For Llama 3.2 1B and 3B models
5151
We have supported BFloat16 as a data type on the XNNPack backend for Llama 3.2 1B/3B models.
52-
* You can download original model weights for Llama through Meta official [website](https://llama.meta.com/), or via Huggingface (Link to specific 3.2 1B repo)
52+
* You can download original model weights for Llama through Meta official [website](https://llama.meta.com/).
5353
* For chat use-cases, download the instruct models instead of pretrained.
5454
* Run “examples/models/llama2/install_requirements.sh” to install dependencies.
5555
* The 1B model in BFloat16 format can run on mobile devices with 8GB RAM (iPhone 15 Pro and later). The 3B model will require 12GB+ RAM and hence will not fit on 8GB RAM phones.

0 commit comments

Comments
 (0)