Skip to content

Commit 8eac09b

Browse files
haozha111copybara-github
authored andcommitted
Update generative example Readme file.
PiperOrigin-RevId: 727032806
1 parent 46e4c41 commit 8eac09b

File tree

1 file changed

+9
-1
lines changed

1 file changed

+9
-1
lines changed

ai_edge_torch/generative/examples/README.md

Lines changed: 9 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -61,9 +61,17 @@ Alibaba's [Qwen 2.5 VL](https://huggingface.co/Qwen/Qwen2.5-VL-3B-Instruct)
6161
model based on Qwen is also provided as an example.
6262

6363
## AMD-Llama-135m
64-
6564
[AMD-Llama-135m](https://huggingface.co/amd/AMD-Llama-135m) is a 135M parameter model based on the Llama2 architecture and uses the same tokenizer as Llama2. It was trained on AMD Instinct MI250 accelerators.
6665

66+
## Moonshine
67+
[Moonshine](https://huggingface.co/UsefulSensors/moonshine) is a sequence-to-sequence ASR model trained by
68+
Useful Sensors, capable of transcribing English speech audio into English text. The model has two variants:
69+
tiny(27M) and base(61M).
70+
71+
## T5
72+
[T5](https://huggingface.co/docs/transformers/en/model_doc/t5) is an encoder-decoder model pre-trained on a multi-task mixture of unsupervised and supervised tasks and for which each task is converted into a text-to-text format. t5-small
73+
has 60M parameters.
74+
6775
## Overall workflow
6876
To support a new LLM with the Edge Generative API, we need to go through the process of model (re)authoring, checkpoint mapping/loading, model quantization (via PT2E), model conversion to flatbuffer schema, model quality evaluation, benchmarking and on-device inference pipeline authoring.
6977

0 commit comments

Comments
 (0)