Skip to content

Commit e92149a

Browse files
authored
llama-cli -> llama-server. (#1477)
1 parent a37b99b commit e92149a

File tree

2 files changed

+8
-14
lines changed

2 files changed

+8
-14
lines changed

packages/tasks/src/local-apps.spec.ts

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ describe("local-apps", () => {
1313
const snippet = snippetFunc(model);
1414

1515
expect(snippet[0].content).toEqual(`# Load and run the model:
16-
llama-cli -hf bartowski/Llama-3.2-3B-Instruct-GGUF:{{QUANT_TAG}}`);
16+
llama-server -hf bartowski/Llama-3.2-3B-Instruct-GGUF:{{QUANT_TAG}}`);
1717
});
1818

1919
it("llama.cpp non-conversational", async () => {
@@ -26,8 +26,7 @@ llama-cli -hf bartowski/Llama-3.2-3B-Instruct-GGUF:{{QUANT_TAG}}`);
2626
const snippet = snippetFunc(model);
2727

2828
expect(snippet[0].content).toEqual(`# Load and run the model:
29-
llama-cli -hf mlabonne/gemma-2b-GGUF:{{QUANT_TAG}} \\
30-
-p "Once upon a time,"`);
29+
llama-server -hf mlabonne/gemma-2b-GGUF:{{QUANT_TAG}}`);
3130
});
3231

3332
it("vLLM conversational llm", async () => {

packages/tasks/src/local-apps.ts

Lines changed: 6 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -108,23 +108,18 @@ function getQuantTag(filepath?: string): string {
108108
const snippetLlamacpp = (model: ModelData, filepath?: string): LocalAppSnippet[] => {
109109
const command = (binary: string) => {
110110
const snippet = ["# Load and run the model:", `${binary} -hf ${model.id}${getQuantTag(filepath)}`];
111-
if (!model.tags.includes("conversational")) {
112-
// for non-conversational models, add a prompt
113-
snippet[snippet.length - 1] += " \\";
114-
snippet.push(' -p "Once upon a time,"');
115-
}
116111
return snippet.join("\n");
117112
};
118113
return [
119114
{
120115
title: "Install from brew",
121116
setup: "brew install llama.cpp",
122-
content: command("llama-cli"),
117+
content: command("llama-server"),
123118
},
124119
{
125120
title: "Install from WinGet (Windows)",
126121
setup: "winget install llama.cpp",
127-
content: command("llama-cli"),
122+
content: command("llama-server"),
128123
},
129124
{
130125
title: "Use pre-built binary",
@@ -133,17 +128,17 @@ const snippetLlamacpp = (model: ModelData, filepath?: string): LocalAppSnippet[]
133128
"# Download pre-built binary from:",
134129
"# https://github.com/ggerganov/llama.cpp/releases",
135130
].join("\n"),
136-
content: command("./llama-cli"),
131+
content: command("./llama-server"),
137132
},
138133
{
139134
title: "Build from source code",
140135
setup: [
141136
"git clone https://github.com/ggerganov/llama.cpp.git",
142137
"cd llama.cpp",
143-
"cmake -B build -DLLAMA_CURL=ON",
144-
"cmake --build build -j --target llama-cli",
138+
"cmake -B build",
139+
"cmake --build build -j --target llama-server",
145140
].join("\n"),
146-
content: command("./build/bin/llama-cli"),
141+
content: command("./build/bin/llama-server"),
147142
},
148143
];
149144
};

0 commit comments

Comments
 (0)