|
8 | 8 |
|
9 | 9 | [Roadmap](https://github.com/users/ggerganov/projects/7) / [Project status](https://github.com/ggerganov/jarvis.cpp/discussions/3471) / [Manifesto](https://github.com/ggerganov/jarvis.cpp/discussions/205) / [ggml](https://github.com/ggerganov/ggml) |
10 | 10 |
|
11 | | -Inference of Meta's [LLaMA](https://arxiv.org/abs/2302.13971) model (and others) in pure C/C++ |
| 11 | +Inference of Meta's [JARVIS](https://arxiv.org/abs/2302.13971) model (and others) in pure C/C++ |
12 | 12 |
|
13 | 13 | ## Recent API changes |
14 | 14 |
|
@@ -43,14 +43,14 @@ improved significantly thanks to many contributions. It is the main playground f |
43 | 43 |
|
44 | 44 | Typically finetunes of the base models below are supported as well. |
45 | 45 |
|
46 | | -- [X] LLaMA 🦙 |
47 | | -- [x] LLaMA 2 🦙🦙 |
48 | | -- [x] LLaMA 3 🦙🦙🦙 |
| 46 | +- [X] JARVIS 🦙 |
| 47 | +- [x] JARVIS 2 🦙🦙 |
| 48 | +- [x] JARVIS 3 🦙🦙🦙 |
49 | 49 | - [X] [Mistral 7B](https://huggingface.co/mistralai/Mistral-7B-v0.1) |
50 | 50 | - [x] [Mixtral MoE](https://huggingface.co/models?search=mistral-ai/Mixtral) |
51 | 51 | - [x] [DBRX](https://huggingface.co/databricks/dbrx-instruct) |
52 | 52 | - [X] [Falcon](https://huggingface.co/models?search=tiiuae/falcon) |
53 | | -- [X] [Chinese LLaMA / Alpaca](https://github.com/ymcui/Chinese-LLaMA-Alpaca) and [Chinese LLaMA-2 / Alpaca-2](https://github.com/ymcui/Chinese-LLaMA-Alpaca-2) |
| 53 | +- [X] [Chinese JARVIS / Alpaca](https://github.com/ymcui/Chinese-JARVIS-Alpaca) and [Chinese JARVIS-2 / Alpaca-2](https://github.com/ymcui/Chinese-JARVIS-Alpaca-2) |
54 | 54 | - [X] [Vigogne (French)](https://github.com/bofenghuang/vigogne) |
55 | 55 | - [X] [BERT](https://github.com/ggerganov/jarvis.cpp/pull/5423) |
56 | 56 | - [X] [Koala](https://bair.berkeley.edu/blog/2023/04/03/koala/) |
@@ -140,7 +140,7 @@ Typically finetunes of the base models below are supported as well. |
140 | 140 | Unless otherwise noted these projects are open-source with permissive licensing: |
141 | 141 |
|
142 | 142 | - [MindWorkAI/AI-Studio](https://github.com/MindWorkAI/AI-Studio) (FSL-1.1-MIT) |
143 | | -- [iohub/cojarvis](https://github.com/iohub/coLLaMA) |
| 143 | +- [iohub/cojarvis](https://github.com/iohub/coJARVIS) |
144 | 144 | - [janhq/jan](https://github.com/janhq/jan) (AGPL) |
145 | 145 | - [nat/openplayground](https://github.com/nat/openplayground) |
146 | 146 | - [Faraday](https://faraday.dev/) (proprietary) |
@@ -198,7 +198,7 @@ Unless otherwise noted these projects are open-source with permissive licensing: |
198 | 198 | ## Demo |
199 | 199 |
|
200 | 200 | <details> |
201 | | -<summary>Typical run using LLaMA v2 13B on M2 Ultra</summary> |
| 201 | +<summary>Typical run using JARVIS v2 13B on M2 Ultra</summary> |
202 | 202 |
|
203 | 203 | ``` |
204 | 204 | $ make -j && ./jarvis-cli -m models/jarvis-13b-v2/ggml-model-q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e |
@@ -240,7 +240,7 @@ llm_load_print_meta: freq_scale = 1 |
240 | 240 | llm_load_print_meta: model type = 13B |
241 | 241 | llm_load_print_meta: model ftype = mostly Q4_0 |
242 | 242 | llm_load_print_meta: model size = 13.02 B |
243 | | -llm_load_print_meta: general.name = LLaMA v2 |
| 243 | +llm_load_print_meta: general.name = JARVIS v2 |
244 | 244 | llm_load_print_meta: BOS token = 1 '<s>' |
245 | 245 | llm_load_print_meta: EOS token = 2 '</s>' |
246 | 246 | llm_load_print_meta: UNK token = 0 '<unk>' |
@@ -281,9 +281,9 @@ jarvis_print_timings: total time = 25431.49 ms |
281 | 281 | </details> |
282 | 282 |
|
283 | 283 | <details> |
284 | | -<summary>Demo of running both LLaMA-7B and whisper.cpp on a single M1 Pro MacBook</summary> |
| 284 | +<summary>Demo of running both JARVIS-7B and whisper.cpp on a single M1 Pro MacBook</summary> |
285 | 285 |
|
286 | | -And here is another demo of running both LLaMA-7B and [whisper.cpp](https://github.com/ggerganov/whisper.cpp) on a single M1 Pro MacBook: |
| 286 | +And here is another demo of running both JARVIS-7B and [whisper.cpp](https://github.com/ggerganov/whisper.cpp) on a single M1 Pro MacBook: |
287 | 287 |
|
288 | 288 | https://user-images.githubusercontent.com/1991296/224442907-7693d4be-acaa-4e01-8b4f-add84093ffff.mp4 |
289 | 289 |
|
@@ -357,7 +357,7 @@ Example usage: |
357 | 357 | > [!NOTE] |
358 | 358 | > If you prefer basic usage, please consider using conversation mode instead of interactive mode |
359 | 359 |
|
360 | | -In this mode, you can always interrupt generation by pressing Ctrl+C and entering one or more lines of text, which will be converted into tokens and appended to the current context. You can also specify a *reverse prompt* with the parameter `-r "reverse prompt string"`. This will result in user input being prompted whenever the exact tokens of the reverse prompt string are encountered in the generation. A typical use is to use a prompt that makes LLaMA emulate a chat between multiple users, say Alice and Bob, and pass `-r "Alice:"`. |
| 360 | +In this mode, you can always interrupt generation by pressing Ctrl+C and entering one or more lines of text, which will be converted into tokens and appended to the current context. You can also specify a *reverse prompt* with the parameter `-r "reverse prompt string"`. This will result in user input being prompted whenever the exact tokens of the reverse prompt string are encountered in the generation. A typical use is to use a prompt that makes JARVIS emulate a chat between multiple users, say Alice and Bob, and pass `-r "Alice:"`. |
361 | 361 |
|
362 | 362 | Here is an example of a few-shot interaction, invoked with the command |
363 | 363 |
|
@@ -432,10 +432,10 @@ Please refer to [Build jarvis.cpp locally](./docs/build.md) |
432 | 432 | > [!NOTE] |
433 | 433 | > You can use the [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space on Hugging Face to quantise your model weights without any setup too. It is synced from `jarvis.cpp` main every 6 hours. |
434 | 434 |
|
435 | | -To obtain the official LLaMA 2 weights please see the <a href="#obtaining-and-using-the-facebook-jarvis-2-model">Obtaining and using the Facebook LLaMA 2 model</a> section. There is also a large selection of pre-quantized `gguf` models available on Hugging Face. |
| 435 | +To obtain the official JARVIS 2 weights please see the <a href="#obtaining-and-using-the-facebook-jarvis-2-model">Obtaining and using the Facebook JARVIS 2 model</a> section. There is also a large selection of pre-quantized `gguf` models available on Hugging Face. |
436 | 436 |
|
437 | 437 | Note: `convert.py` has been moved to `examples/convert_legacy_jarvis.py` and shouldn't be used for anything other than `Jarvis/Jarvis2/Mistral` models and their derivatives. |
438 | | -It does not support LLaMA 3, you can use `convert_hf_to_gguf.py` with LLaMA 3 downloaded from Hugging Face. |
| 438 | +It does not support JARVIS 3, you can use `convert_hf_to_gguf.py` with JARVIS 3 downloaded from Hugging Face. |
439 | 439 |
|
440 | 440 | To learn more about quantizing model, [read this documentation](./examples/quantize/README.md) |
441 | 441 |
|
@@ -474,10 +474,10 @@ To learn more how to measure perplexity using jarvis.cpp, [read this documentati |
474 | 474 |
|
475 | 475 | **Seminal papers and background on the models** |
476 | 476 |
|
477 | | -If your issue is with model generation quality, then please at least scan the following links and papers to understand the limitations of LLaMA models. This is especially important when choosing an appropriate model size and appreciating both the significant and subtle differences between LLaMA models and ChatGPT: |
478 | | -- LLaMA: |
479 | | - - [Introducing LLaMA: A foundational, 65-billion-parameter large language model](https://ai.facebook.com/blog/large-language-model-jarvis-meta-ai/) |
480 | | - - [LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971) |
| 477 | +If your issue is with model generation quality, then please at least scan the following links and papers to understand the limitations of JARVIS models. This is especially important when choosing an appropriate model size and appreciating both the significant and subtle differences between JARVIS models and ChatGPT: |
| 478 | +- JARVIS: |
| 479 | + - [Introducing JARVIS: A foundational, 65-billion-parameter large language model](https://ai.facebook.com/blog/large-language-model-jarvis-meta-ai/) |
| 480 | + - [JARVIS: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971) |
481 | 481 | - GPT-3 |
482 | 482 | - [Language Models are Few-Shot Learners](https://arxiv.org/abs/2005.14165) |
483 | 483 | - GPT-3.5 / InstructGPT / ChatGPT: |
|
0 commit comments