Skip to content

Commit 3334a0b

Browse files
committed
Use bartowski/Qwen2.5-7B-Instruct-GGUF/Qwen2.5-7B-Instruct-Q8_0.gguf as default model
1 parent 1b2dd46 commit 3334a0b

File tree

6 files changed

+7
-7
lines changed

6 files changed

+7
-7
lines changed

demo/app.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@
2323
@st.cache_resource
2424
def load_text_to_text_model():
2525
return load_llama_cpp_model(
26-
model_id="bartowski/Qwen2.5-3B-Instruct-GGUF/Qwen2.5-3B-Instruct-f16.gguf"
26+
model_id="bartowski/Qwen2.5-7B-Instruct-GGUF/Qwen2.5-7B-Instruct-Q8_0.gguf"
2727
)
2828

2929

demo/notebook.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -205,7 +205,7 @@
205205
")\n",
206206
"\n",
207207
"text_model = load_llama_cpp_model(\n",
208-
" \"bartowski/Qwen2.5-3B-Instruct-GGUF/Qwen2.5-3B-Instruct-f16.gguf\"\n",
208+
" \"bartowski/Qwen2.5-7B-Instruct-GGUF/Qwen2.5-7B-Instruct-Q8_0.gguf\"\n",
209209
")\n",
210210
"speech_model = load_tts_model(\"hexgrad/kLegacy/v0.19/kokoro-v0_19.pth\")"
211211
]

docs/step-by-step-guide.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -106,7 +106,7 @@ from document_to_podcast.inference.text_to_text import text_to_text, text_to_tex
106106

107107
# Load the model
108108
model = load_llama_cpp_model(
109-
"bartowski/Qwen2.5-3B-Instruct-GGUF/Qwen2.5-3B-Instruct-f16.gguf"
109+
"bartowski/Qwen2.5-7B-Instruct-GGUF/Qwen2.5-7B-Instruct-Q8_0.gguf"
110110
)
111111

112112
# Define your input and system prompt

example_data/config.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
input_file: "example_data/introducing-mozilla-ai-investing-in-trustworthy-ai.html"
22
output_folder: "example_data/"
3-
text_to_text_model: "bartowski/Qwen2.5-3B-Instruct-GGUF/Qwen2.5-3B-Instruct-f16.gguf"
3+
text_to_text_model: "bartowski/Qwen2.5-7B-Instruct-GGUF/Qwen2.5-7B-Instruct-Q8_0.gguf"
44
text_to_speech_model: "OuteAI/OuteTTS-0.1-350M-GGUF/OuteTTS-0.1-350M-FP16.gguf"
55
text_to_text_prompt: |
66
You are a podcast scriptwriter generating engaging and natural-sounding conversations in JSON format.

src/document_to_podcast/cli.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@
2626
def document_to_podcast(
2727
input_file: str | None = None,
2828
output_folder: str | None = None,
29-
text_to_text_model: str = "bartowski/Qwen2.5-3B-Instruct-GGUF/Qwen2.5-3B-Instruct-f16.gguf",
29+
text_to_text_model: str = "bartowski/Qwen2.5-7B-Instruct-GGUF/Qwen2.5-7B-Instruct-Q8_0.gguf",
3030
text_to_text_prompt: str = DEFAULT_PROMPT,
3131
text_to_speech_model: str = "OuteAI/OuteTTS-0.2-500M-GGUF/OuteTTS-0.2-500M-FP16.gguf",
3232
speakers: list[Speaker] | None = None,
@@ -58,7 +58,7 @@ def document_to_podcast(
5858
5959
Need to be a gguf file.
6060
61-
Defaults to `bartowski/Qwen2.5-3B-Instruct-GGUF/Qwen2.5-3B-Instruct-f16.gguf`.
61+
Defaults to `bartowski/Qwen2.5-7B-Instruct-GGUF/Qwen2.5-7B-Instruct-Q8_0.gguf`.
6262
6363
text_to_text_prompt (str, optional): The prompt for the text-to-text model.
6464
Defaults to DEFAULT_PROMPT.

src/document_to_podcast/inference/model_loaders.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ def load_llama_cpp_model(model_id: str) -> Llama:
1010
Loads the given model_id using Llama.from_pretrained.
1111
1212
Examples:
13-
>>> model = load_llama_cpp_model("bartowski/Qwen2.5-3B-Instruct-GGUF/Qwen2.5-3B-Instruct-f16.gguf")
13+
>>> model = load_llama_cpp_model("bartowski/Qwen2.5-7B-Instruct-GGUF/Qwen2.5-7B-Instruct-Q8_0.gguf")
1414
1515
Args:
1616
model_id (str): The model id to load.

0 commit comments

Comments
 (0)