Skip to content

Commit 31a90b3

Browse files
committed
test
1 parent 24c46d9 commit 31a90b3

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

43 files changed

+100
-568
lines changed

.devops/jarvis-cpp-cuda.srpm.spec

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
Name: jarvis.cpp-cuda
1616
Version: %( date "+%%Y%%m%%d" )
1717
Release: 1%{?dist}
18-
Summary: CPU Inference of LLaMA model in pure C/C++ (no CUDA/OpenCL)
18+
Summary: CPU Inference of JARVIS model in pure C/C++ (no CUDA/OpenCL)
1919
License: MIT
2020
Source0: https://github.com/ggerganov/jarvis.cpp/archive/refs/heads/master.tar.gz
2121
BuildRequires: coreutils make gcc-c++ git cuda-toolkit

.devops/jarvis-cpp.srpm.spec

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@
1616
Name: jarvis.cpp
1717
Version: %( date "+%%Y%%m%%d" )
1818
Release: 1%{?dist}
19-
Summary: CPU Inference of LLaMA model in pure C/C++ (no CUDA/OpenCL)
19+
Summary: CPU Inference of JARVIS model in pure C/C++ (no CUDA/OpenCL)
2020
License: MIT
2121
Source0: https://github.com/ggerganov/jarvis.cpp/archive/refs/heads/master.tar.gz
2222
BuildRequires: coreutils make gcc-c++ git libstdc++-devel

.devops/nix/package.nix

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -218,7 +218,7 @@ effectiveStdenv.mkDerivation (finalAttrs: {
218218
# overridden by importing Nixpkgs with `allowBroken = true`.
219219
broken = (useMetalKit && !effectiveStdenv.isDarwin);
220220

221-
description = "Inference of LLaMA model in pure C/C++${descriptionSuffix}";
221+
description = "Inference of JARVIS model in pure C/C++${descriptionSuffix}";
222222
homepage = "https://github.com/ggerganov/jarvis.cpp/";
223223
license = lib.licenses.mit;
224224

Binary file not shown.

LLMCLI.java

Lines changed: 0 additions & 74 deletions
This file was deleted.

README.md

Lines changed: 17 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@
88

99
[Roadmap](https://github.com/users/ggerganov/projects/7) / [Project status](https://github.com/ggerganov/jarvis.cpp/discussions/3471) / [Manifesto](https://github.com/ggerganov/jarvis.cpp/discussions/205) / [ggml](https://github.com/ggerganov/ggml)
1010

11-
Inference of Meta's [LLaMA](https://arxiv.org/abs/2302.13971) model (and others) in pure C/C++
11+
Inference of Meta's [JARVIS](https://arxiv.org/abs/2302.13971) model (and others) in pure C/C++
1212

1313
## Recent API changes
1414

@@ -43,14 +43,14 @@ improved significantly thanks to many contributions. It is the main playground f
4343

4444
Typically finetunes of the base models below are supported as well.
4545

46-
- [X] LLaMA 🦙
47-
- [x] LLaMA 2 🦙🦙
48-
- [x] LLaMA 3 🦙🦙🦙
46+
- [X] JARVIS 🦙
47+
- [x] JARVIS 2 🦙🦙
48+
- [x] JARVIS 3 🦙🦙🦙
4949
- [X] [Mistral 7B](https://huggingface.co/mistralai/Mistral-7B-v0.1)
5050
- [x] [Mixtral MoE](https://huggingface.co/models?search=mistral-ai/Mixtral)
5151
- [x] [DBRX](https://huggingface.co/databricks/dbrx-instruct)
5252
- [X] [Falcon](https://huggingface.co/models?search=tiiuae/falcon)
53-
- [X] [Chinese LLaMA / Alpaca](https://github.com/ymcui/Chinese-LLaMA-Alpaca) and [Chinese LLaMA-2 / Alpaca-2](https://github.com/ymcui/Chinese-LLaMA-Alpaca-2)
53+
- [X] [Chinese JARVIS / Alpaca](https://github.com/ymcui/Chinese-JARVIS-Alpaca) and [Chinese JARVIS-2 / Alpaca-2](https://github.com/ymcui/Chinese-JARVIS-Alpaca-2)
5454
- [X] [Vigogne (French)](https://github.com/bofenghuang/vigogne)
5555
- [X] [BERT](https://github.com/ggerganov/jarvis.cpp/pull/5423)
5656
- [X] [Koala](https://bair.berkeley.edu/blog/2023/04/03/koala/)
@@ -140,7 +140,7 @@ Typically finetunes of the base models below are supported as well.
140140
Unless otherwise noted these projects are open-source with permissive licensing:
141141

142142
- [MindWorkAI/AI-Studio](https://github.com/MindWorkAI/AI-Studio) (FSL-1.1-MIT)
143-
- [iohub/cojarvis](https://github.com/iohub/coLLaMA)
143+
- [iohub/cojarvis](https://github.com/iohub/coJARVIS)
144144
- [janhq/jan](https://github.com/janhq/jan) (AGPL)
145145
- [nat/openplayground](https://github.com/nat/openplayground)
146146
- [Faraday](https://faraday.dev/) (proprietary)
@@ -198,7 +198,7 @@ Unless otherwise noted these projects are open-source with permissive licensing:
198198
## Demo
199199

200200
<details>
201-
<summary>Typical run using LLaMA v2 13B on M2 Ultra</summary>
201+
<summary>Typical run using JARVIS v2 13B on M2 Ultra</summary>
202202

203203
```
204204
$ make -j && ./jarvis-cli -m models/jarvis-13b-v2/ggml-model-q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e
@@ -240,7 +240,7 @@ llm_load_print_meta: freq_scale = 1
240240
llm_load_print_meta: model type = 13B
241241
llm_load_print_meta: model ftype = mostly Q4_0
242242
llm_load_print_meta: model size = 13.02 B
243-
llm_load_print_meta: general.name = LLaMA v2
243+
llm_load_print_meta: general.name = JARVIS v2
244244
llm_load_print_meta: BOS token = 1 '<s>'
245245
llm_load_print_meta: EOS token = 2 '</s>'
246246
llm_load_print_meta: UNK token = 0 '<unk>'
@@ -281,9 +281,9 @@ jarvis_print_timings: total time = 25431.49 ms
281281
</details>
282282

283283
<details>
284-
<summary>Demo of running both LLaMA-7B and whisper.cpp on a single M1 Pro MacBook</summary>
284+
<summary>Demo of running both JARVIS-7B and whisper.cpp on a single M1 Pro MacBook</summary>
285285

286-
And here is another demo of running both LLaMA-7B and [whisper.cpp](https://github.com/ggerganov/whisper.cpp) on a single M1 Pro MacBook:
286+
And here is another demo of running both JARVIS-7B and [whisper.cpp](https://github.com/ggerganov/whisper.cpp) on a single M1 Pro MacBook:
287287

288288
https://user-images.githubusercontent.com/1991296/224442907-7693d4be-acaa-4e01-8b4f-add84093ffff.mp4
289289

@@ -357,7 +357,7 @@ Example usage:
357357
> [!NOTE]
358358
> If you prefer basic usage, please consider using conversation mode instead of interactive mode
359359
360-
In this mode, you can always interrupt generation by pressing Ctrl+C and entering one or more lines of text, which will be converted into tokens and appended to the current context. You can also specify a *reverse prompt* with the parameter `-r "reverse prompt string"`. This will result in user input being prompted whenever the exact tokens of the reverse prompt string are encountered in the generation. A typical use is to use a prompt that makes LLaMA emulate a chat between multiple users, say Alice and Bob, and pass `-r "Alice:"`.
360+
In this mode, you can always interrupt generation by pressing Ctrl+C and entering one or more lines of text, which will be converted into tokens and appended to the current context. You can also specify a *reverse prompt* with the parameter `-r "reverse prompt string"`. This will result in user input being prompted whenever the exact tokens of the reverse prompt string are encountered in the generation. A typical use is to use a prompt that makes JARVIS emulate a chat between multiple users, say Alice and Bob, and pass `-r "Alice:"`.
361361

362362
Here is an example of a few-shot interaction, invoked with the command
363363

@@ -432,10 +432,10 @@ Please refer to [Build jarvis.cpp locally](./docs/build.md)
432432
> [!NOTE]
433433
> You can use the [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space on Hugging Face to quantise your model weights without any setup too. It is synced from `jarvis.cpp` main every 6 hours.
434434
435-
To obtain the official LLaMA 2 weights please see the <a href="#obtaining-and-using-the-facebook-jarvis-2-model">Obtaining and using the Facebook LLaMA 2 model</a> section. There is also a large selection of pre-quantized `gguf` models available on Hugging Face.
435+
To obtain the official JARVIS 2 weights please see the <a href="#obtaining-and-using-the-facebook-jarvis-2-model">Obtaining and using the Facebook JARVIS 2 model</a> section. There is also a large selection of pre-quantized `gguf` models available on Hugging Face.
436436

437437
Note: `convert.py` has been moved to `examples/convert_legacy_jarvis.py` and shouldn't be used for anything other than `Jarvis/Jarvis2/Mistral` models and their derivatives.
438-
It does not support LLaMA 3, you can use `convert_hf_to_gguf.py` with LLaMA 3 downloaded from Hugging Face.
438+
It does not support JARVIS 3, you can use `convert_hf_to_gguf.py` with JARVIS 3 downloaded from Hugging Face.
439439

440440
To learn more about quantizing model, [read this documentation](./examples/quantize/README.md)
441441

@@ -474,10 +474,10 @@ To learn more how to measure perplexity using jarvis.cpp, [read this documentati
474474

475475
**Seminal papers and background on the models**
476476

477-
If your issue is with model generation quality, then please at least scan the following links and papers to understand the limitations of LLaMA models. This is especially important when choosing an appropriate model size and appreciating both the significant and subtle differences between LLaMA models and ChatGPT:
478-
- LLaMA:
479-
- [Introducing LLaMA: A foundational, 65-billion-parameter large language model](https://ai.facebook.com/blog/large-language-model-jarvis-meta-ai/)
480-
- [LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)
477+
If your issue is with model generation quality, then please at least scan the following links and papers to understand the limitations of JARVIS models. This is especially important when choosing an appropriate model size and appreciating both the significant and subtle differences between JARVIS models and ChatGPT:
478+
- JARVIS:
479+
- [Introducing JARVIS: A foundational, 65-billion-parameter large language model](https://ai.facebook.com/blog/large-language-model-jarvis-meta-ai/)
480+
- [JARVIS: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)
481481
- GPT-3
482482
- [Language Models are Few-Shot Learners](https://arxiv.org/abs/2005.14165)
483483
- GPT-3.5 / InstructGPT / ChatGPT:

SECURITY.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ For maximum security when handling untrusted inputs, you may need to employ the
2626

2727
* Sandboxing: Isolate the environment where the inference happens.
2828
* Pre-analysis: Check how the model performs by default when exposed to prompt injection (e.g. using [fuzzing for prompt injection](https://github.com/FonduAI/awesome-prompt-injection?tab=readme-ov-file#tools)). This will give you leads on how hard you will have to work on the next topics.
29-
* Updates: Keep both LLaMA C++ and your libraries updated with the latest security patches.
29+
* Updates: Keep both JARVIS C++ and your libraries updated with the latest security patches.
3030
* Input Sanitation: Before feeding data to the model, sanitize inputs rigorously. This involves techniques such as:
3131
* Validation: Enforce strict rules on allowed characters and data types.
3232
* Filtering: Remove potentially malicious scripts or code fragments.
@@ -57,7 +57,7 @@ If you intend to run multiple models in parallel with shared memory, it is your
5757

5858
## Reporting a vulnerability
5959

60-
Beware that none of the topics under [Using jarvis.cpp securely](#using-jarviscpp-securely) are considered vulnerabilities of LLaMA C++.
60+
Beware that none of the topics under [Using jarvis.cpp securely](#using-jarviscpp-securely) are considered vulnerabilities of JARVIS C++.
6161

6262
<!-- normal version -->
6363
However, If you have discovered a security vulnerability in this project, please report it privately. **Do not disclose it as a public issue.** This gives us time to work with you to fix the issue before public exposure, reducing the chance that the exploit will be used before a patch is released.

ci/run.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -390,7 +390,7 @@ function gg_run_open_jarvis_7b_v2 {
390390
function gg_sum_open_jarvis_7b_v2 {
391391
gg_printf '### %s\n\n' "${ci}"
392392

393-
gg_printf 'OpenLLaMA 7B-v2:\n'
393+
gg_printf 'OpenJARVIS 7B-v2:\n'
394394
gg_printf '- status: %s\n' "$(cat $OUT/${ci}.exit)"
395395
gg_printf '- perplexity:\n%s\n' "$(cat $OUT/${ci}-ppl.log)"
396396
gg_printf '- imatrix:\n```\n%s\n```\n' "$(cat $OUT/${ci}-imatrix-sum.log)"

cmake/jarvis.pc.in

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ libdir=${exec_prefix}/lib
44
includedir=${prefix}/include
55

66
Name: jarvis
7-
Description: Port of Facebook's LLaMA model in C/C++
7+
Description: Port of Facebook's JARVIS model in C/C++
88
Version: @PROJECT_VERSION@
99
Libs: -L${libdir} -ljarvis
1010
Cflags: -I${includedir}

convert_hf_to_gguf.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1515,7 +1515,7 @@ def prepare_tensors(self):
15151515
raise ValueError(f"Unprocessed norms: {norms}")
15161516

15171517

1518-
@Model.register("LLaMAForCausalLM", "JarvisForCausalLM", "MistralForCausalLM", "MixtralForCausalLM")
1518+
@Model.register("JARVISForCausalLM", "JarvisForCausalLM", "MistralForCausalLM", "MixtralForCausalLM")
15191519
class JarvisModel(Model):
15201520
model_arch = gguf.MODEL_ARCH.JARVIS
15211521

0 commit comments

Comments
 (0)