Skip to content
This repository was archived by the owner on Sep 10, 2025. It is now read-only.

Commit 4dda7f3

Browse files
committed
Update
[ghstack-poisoned]
2 parents 3d7723d + eebb8b7 commit 4dda7f3

File tree

2 files changed

+6
-12
lines changed

2 files changed

+6
-12
lines changed

README.md

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -182,7 +182,7 @@ python3 torchchat.py generate llama3.1 --prompt "write me a story about a boy an
182182
[skip default]: end
183183

184184
### Server
185-
This mode exposes a REST API for interacting with a model.
185+
This mode exposes a REST API for interacting with a model.
186186
The server follows the [OpenAI API specification](https://platform.openai.com/docs/api-reference/chat) for chat completions.
187187

188188
To test out the REST API, **you'll need 2 terminals**: one to host the server, and one to send the request.
@@ -440,7 +440,7 @@ The following assumes you've completed the steps for [Setting up ExecuTorch](#se
440440

441441
1. Download the AAR file, which contains the Java library and corresponding JNI library, to build and run the app.
442442

443-
- [executorch-llama-tiktoken-rc3-0719.aar](https://ossci-android.s3.amazonaws.com/executorch/main/executorch-llama-tiktoken-rc3-0719.aar) (SHASUM: c3e5d2a97708f033c2b1839a89f12f737e3bbbef)
443+
- [executorch-240919.aar](https://ossci-android.s3.amazonaws.com/executorch/main/executorch-240919.aar) (SHASUM: c8a5d38ead03bfa28ee8469f6355840ad0d182ba)
444444

445445
2. Rename the downloaded AAR file to `executorch.aar` and move the file to `android/torchchat/app/libs/`. You may need to create directory `android/torchchat/app/libs/` if it does not exist.
446446

@@ -460,9 +460,8 @@ The following assumes you've completed the steps for [Setting up ExecuTorch](#se
460460

461461
6. Follow the app's UI guidelines to pick the model and tokenizer files from the local filesystem. Then issue a prompt.
462462
463-
**Note:** The AAR file listed in Step 1 has the tiktoken tokenizer, which is used for Llama 3. To tweak or use a custom tokenizer and runtime, modify the ExecuTorch code
464-
and use [this script](https://github.com/pytorch/executorch/blob/main/build/build_android_llm_demo.sh) to build the AAR library. For convenience, we also provide an AAR
465-
for sentencepiece tokenizer (e.g. Llama 2): [executorch-llama-bpe-rc3-0719.aar](https://ossci-android.s3.amazonaws.com/executorch/main/executorch-llama-bpe-rc3-0719.aar) (SHASUM: d5fe81d9a4700c36b50ae322e6bf34882134edb0)
463+
**Note:** The AAR file listed in Step 1 has the tiktoken and sentensepiece tokenizer. To tweak or use a custom tokenizer and runtime, modify the ExecuTorch code
464+
and use [this script](https://github.com/pytorch/executorch/blob/main/build/build_android_llm_demo.sh) to build the AAR library.
466465
467466
<p align="center">
468467
<img src="https://pytorch.org/executorch/main/_static/img/android_llama_app.png" width="600" alt="Android app running a LlaMA model">

torchchat/utils/scripts/android_example.sh

Lines changed: 2 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -30,13 +30,8 @@ else
3030
exit -1
3131
fi
3232

33-
if [ "${USE_TIKTOKEN:-OFF}" == "ON" ]; then
34-
LLAMA_AAR_URL="https://ossci-android.s3.amazonaws.com/executorch/main/executorch-llama-tiktoken-rc3-0719.aar"
35-
LLAMA_AAR_SHASUM="c3e5d2a97708f033c2b1839a89f12f737e3bbbef"
36-
else
37-
LLAMA_AAR_URL="https://ossci-android.s3.amazonaws.com/executorch/main/executorch-llama-bpe-rc3-0719.aar"
38-
LLAMA_AAR_SHASUM="d5fe81d9a4700c36b50ae322e6bf34882134edb0"
39-
fi
33+
LLAMA_AAR_URL="https://ossci-android.s3.amazonaws.com/executorch/main/executorch-240919.aar"
34+
LLAMA_AAR_SHASUM="c8a5d38ead03bfa28ee8469f6355840ad0d182ba"
4035

4136
mkdir -p ${TORCHCHAT_ROOT}/build/android
4237

0 commit comments

Comments
 (0)