Skip to content

Commit 585abd1

Browse files
committed
refactor: update to the new GitHub org used for llama.cpp
1 parent eba30d4 commit 585abd1

File tree

13 files changed

+26
-26
lines changed

13 files changed

+26
-26
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -94,7 +94,7 @@ console.log("AI: " + a2);
9494
To contribute to `node-llama-cpp` read the [contribution guide](https://node-llama-cpp.withcat.ai/guide/contributing).
9595

9696
## Acknowledgements
97-
* llama.cpp: [ggerganov/llama.cpp](https://github.com/ggerganov/llama.cpp)
97+
* llama.cpp: [ggml-org/llama.cpp](https://github.com/ggml-org/llama.cpp)
9898

9999

100100
<br />

docs/blog/v3.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ image:
1515
---
1616
[`node-llama-cpp`](https://node-llama-cpp.withcat.ai) 3.0 is finally here.
1717

18-
With [`node-llama-cpp`](https://node-llama-cpp.withcat.ai), you can run large language models locally on your machine using the power of [`llama.cpp`](https://github.com/ggerganov/llama.cpp) with a simple and easy-to-use API.
18+
With [`node-llama-cpp`](https://node-llama-cpp.withcat.ai), you can run large language models locally on your machine using the power of [`llama.cpp`](https://github.com/ggml-org/llama.cpp) with a simple and easy-to-use API.
1919

2020
It includes everything you need, from downloading models, to running them in the most optimized way for your hardware, and integrating them in your projects.
2121

@@ -43,7 +43,7 @@ While `llama.cpp` is an amazing project, it's also highly technical and can be c
4343
`node-llama-cpp` bridge that gap, making `llama.cpp` accessible to everyone, regardless of their experience level.
4444

4545
### Performance
46-
[`node-llama-cpp`](https://node-llama-cpp.withcat.ai) is built on top of [`llama.cpp`](https://github.com/ggerganov/llama.cpp), a highly optimized C++ library for running large language models.
46+
[`node-llama-cpp`](https://node-llama-cpp.withcat.ai) is built on top of [`llama.cpp`](https://github.com/ggml-org/llama.cpp), a highly optimized C++ library for running large language models.
4747

4848
`llama.cpp` supports many compute backends, including Metal, CUDA, and Vulkan. It also uses [Accelerate](https://developer.apple.com/accelerate/) on Mac.
4949

@@ -116,7 +116,7 @@ npx -y node-llama-cpp chat
116116
Check out the [getting started guide](../guide/index.md) to learn how to use `node-llama-cpp`.
117117

118118
## Thank You
119-
`node-llama-cpp` is only possible thanks to the amazing work done on [`llama.cpp`](https://github.com/ggerganov/llama.cpp) by [Georgi Gerganov](https://github.com/ggerganov), [Slaren](https://github.com/slaren) and all the contributors from the community.
119+
`node-llama-cpp` is only possible thanks to the amazing work done on [`llama.cpp`](https://github.com/ggml-org/llama.cpp) by [Georgi Gerganov](https://github.com/ggerganov), [Slaren](https://github.com/slaren) and all the contributors from the community.
120120

121121
## What's next?
122122
Version 3.0 is a major milestone, but there's plenty more planned for the future.

docs/guide/Vulkan.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -135,7 +135,7 @@ watch -d "npx --no node-llama-cpp inspect gpu"
135135
```
136136

137137
## Vulkan Caveats
138-
[At the moment](https://github.com/ggerganov/llama.cpp/issues/7575),
138+
[At the moment](https://github.com/ggml-org/llama.cpp/issues/7575),
139139
Vulkan doesn't work well when using multiple contexts at the same time,
140140
so it's recommended to use a single context with Vulkan,
141141
and to manually dispose a context (using [`.dispose()`](../api/classes/LlamaContext.md#dispose)) before creating a new one.

docs/guide/building-from-source.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -172,7 +172,7 @@ or pass the code snippet that is printed after the build finishes.
172172
Every new release of `node-llama-cpp` ships with the latest release of `llama.cpp` that was available at the time of the release,
173173
so relying on the latest version of `node-llama-cpp` should be enough for most use cases.
174174

175-
However, you may want to download a newer release of `llama.cpp` ([`llama.cpp` releases](https://github.com/ggerganov/llama.cpp/releases))
175+
However, you may want to download a newer release of `llama.cpp` ([`llama.cpp` releases](https://github.com/ggml-org/llama.cpp/releases))
176176
and build it from source to get the latest features and bug fixes before a new version of `node-llama-cpp` is released.
177177

178178
A new release may contain breaking changes, so it won't necessarily work properly or even compile at all, so do this with caution.
@@ -182,7 +182,7 @@ You can do this by specifying the `--release` option with the release tag you wa
182182
npx --no node-llama-cpp source download --release "b1350"
183183
```
184184

185-
> You can find the release tag on the [`llama.cpp` releases page](https://github.com/ggerganov/llama.cpp/releases):
185+
> You can find the release tag on the [`llama.cpp` releases page](https://github.com/ggml-org/llama.cpp/releases):
186186
187187
You can also opt to download the latest release available:
188188
```shell

docs/guide/choosing-a-model.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -142,7 +142,7 @@ Here are a few concepts to be aware of when choosing a model:
142142
If you plan to feed the model with a lot of data at once, you'll need a model that supports a large context size.
143143
The larger the context size is, the more data the model can process at once.
144144

145-
You can only create a context with a size that is smaller or equal to the context size the model was trained on (although there are techniques around that, like [RoPE](https://github.com/ggerganov/llama.cpp/discussions/1965)).
145+
You can only create a context with a size that is smaller or equal to the context size the model was trained on (although there are techniques around that, like [RoPE](https://github.com/ggml-org/llama.cpp/discussions/1965)).
146146
The larger the context size is, the more memory the model will require to run.
147147
If you plan to feed the model with a lot of data at once, you may want to choose a smaller model that uses less memory, so you can create a larger context.
148148

docs/guide/cmakeOptions.data.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ const loader = {
1616
const clonedRepoReleaseInfo = await getClonedLlamaCppRepoReleaseInfo();
1717
const release = clonedRepoReleaseInfo?.tag ?? await getBinariesGithubRelease();
1818

19-
const githubFileUrl = `https://github.com/ggerganov/llama.cpp/blob/${encodeURIComponent(release)}/ggml/CMakeLists.txt`;
19+
const githubFileUrl = `https://github.com/ggml-org/llama.cpp/blob/${encodeURIComponent(release)}/ggml/CMakeLists.txt`;
2020

2121
return {
2222
cmakeOptionsFileUrl: githubFileUrl,

docs/guide/grammar.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ so it's recommended to use it together with `maxTokens` set to the context size
2626
## Using a Builtin Grammar {#builtin-grammar}
2727
The [`llama.getGrammarFor("<format>")`](../api/classes/Llama.md#getgrammarfor) method reads a GBNF grammar file that's originally provided by `llama.cpp` and is included inside of `node-llama-cpp`.
2828

29-
You can see the full list of supported grammar files [here](https://github.com/ggerganov/llama.cpp/tree/master/grammars).
29+
You can see the full list of supported grammar files [here](https://github.com/ggml-org/llama.cpp/tree/master/grammars).
3030

3131
```typescript
3232
import {fileURLToPath} from "url";
@@ -174,7 +174,7 @@ so there's no need to explain the schema in the prompt.
174174
:::
175175

176176
## Creating Your Own Grammar {#custom-grammar}
177-
To create your own grammar, read the [GBNF guide](https://github.com/ggerganov/llama.cpp/blob/f5fe98d11bdf9e7797bcfb05c0c3601ffc4b9d26/grammars/README.md) to create a GBNF grammar file.
177+
To create your own grammar, read the [GBNF guide](https://github.com/ggml-org/llama.cpp/blob/f5fe98d11bdf9e7797bcfb05c0c3601ffc4b9d26/grammars/README.md) to create a GBNF grammar file.
178178

179179
To use your custom grammar file, load it via the [`llama.createGrammar(...)`](../api/classes/Llama.md#creategrammar) method:
180180
```typescript

docs/guide/tips-and-tricks.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -67,8 +67,8 @@ or provide additional information regarding flash attention when used.
6767
6868
OpenMP can help improve inference performance on Linux and Windows, but requires additional installation and setup.
6969

70-
The performance improvement can be [up to 8% faster](https://github.com/ggerganov/llama.cpp/pull/7606) inference times (on specific conditions).
71-
Setting the `OMP_PROC_BIND` environment variable to `TRUE` on systems that support many threads (assume 36 as the minimum) can improve performance [by up to 23%](https://github.com/ggerganov/llama.cpp/pull/7606).
70+
The performance improvement can be [up to 8% faster](https://github.com/ggml-org/llama.cpp/pull/7606) inference times (on specific conditions).
71+
Setting the `OMP_PROC_BIND` environment variable to `TRUE` on systems that support many threads (assume 36 as the minimum) can improve performance [by up to 23%](https://github.com/ggml-org/llama.cpp/pull/7606).
7272

7373
The pre-built binaries are compiled without OpenMP since OpenMP isn't always available on all systems, and has to be installed separately.
7474

src/config.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ export const localXpacksStoreDirectory = path.join(xpackDirectory, "store");
3636
export const localXpacksCacheDirectory = path.join(xpackDirectory, "cache");
3737
export const buildMetadataFileName = "_nlcBuildMetadata.json";
3838
export const xpmVersion = "^0.16.3";
39-
export const builtinLlamaCppGitHubRepo = "ggerganov/llama.cpp";
39+
export const builtinLlamaCppGitHubRepo = "ggml-org/llama.cpp";
4040
export const builtinLlamaCppRelease = await getBinariesGithubRelease();
4141

4242
export const isCI = env.get("CI")

src/evaluator/LlamaGrammar.ts

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -39,8 +39,8 @@ export class LlamaGrammar {
3939
/**
4040
* > GBNF files are supported.
4141
* > More info here: [
42-
* github:ggerganov/llama.cpp:grammars/README.md
43-
* ](https://github.com/ggerganov/llama.cpp/blob/f5fe98d11bdf9e7797bcfb05c0c3601ffc4b9d26/grammars/README.md)
42+
* github:ggml-org/llama.cpp:grammars/README.md
43+
* ](https://github.com/ggml-org/llama.cpp/blob/f5fe98d11bdf9e7797bcfb05c0c3601ffc4b9d26/grammars/README.md)
4444
*
4545
* Prefer to create a new instance of this class by using `llama.createGrammar(...)`.
4646
* @deprecated Use `llama.createGrammar(...)` instead.

0 commit comments

Comments
 (0)