diff --git a/packages/tasks/src/model-libraries-snippets.spec.ts b/packages/tasks/src/model-libraries-snippets.spec.ts index fa87d82423..81d28f5540 100644 --- a/packages/tasks/src/model-libraries-snippets.spec.ts +++ b/packages/tasks/src/model-libraries-snippets.spec.ts @@ -12,7 +12,9 @@ describe("model-libraries-snippets", () => { }; const snippet = llama_cpp_python(model); - expect(snippet.join("\n")).toEqual(`from llama_cpp import Llama + expect(snippet.join("\n")).toEqual(`# !pip install llama-cpp-python + +from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="bartowski/Llama-3.2-3B-Instruct-GGUF", @@ -37,7 +39,9 @@ llm.create_chat_completion( }; const snippet = llama_cpp_python(model); - expect(snippet.join("\n")).toEqual(`from llama_cpp import Llama + expect(snippet.join("\n")).toEqual(`# !pip install llama-cpp-python + +from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="mlabonne/gemma-2b-GGUF", diff --git a/packages/tasks/src/model-libraries-snippets.ts b/packages/tasks/src/model-libraries-snippets.ts index 798fb8358b..d180658810 100644 --- a/packages/tasks/src/model-libraries-snippets.ts +++ b/packages/tasks/src/model-libraries-snippets.ts @@ -678,7 +678,9 @@ model.score("query", ["doc1", "doc2", "doc3"])`, export const llama_cpp_python = (model: ModelData): string[] => { const snippets = [ - `from llama_cpp import Llama + `# !pip install llama-cpp-python + +from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="${model.id}",