Skip to content

Commit 8c3a6b6

Browse files
committed
fix: harmonizing of model names and selectors
1 parent 64b0dd8 commit 8c3a6b6

File tree

2 files changed

+40
-50
lines changed

2 files changed

+40
-50
lines changed

agentic_rag/gradio_app.py

Lines changed: 33 additions & 33 deletions
Original file line numberDiff line numberDiff line change
@@ -298,46 +298,46 @@ def create_interface():
298298
# HF models first if token is available
299299
if hf_token:
300300
model_choices.extend([
301-
"Local (Mistral)",
302-
"Local (Mistral) - 4-bit Quantized",
303-
"Local (Mistral) - 8-bit Quantized",
301+
"mistral",
302+
"mistral-4bit",
303+
"mistral-8bit",
304304
])
305305
# Then Ollama models (don't require HF token)
306306
model_choices.extend([
307-
"Ollama - llama3",
308-
"Ollama - phi-3",
309-
"Ollama - qwen2",
307+
"llama3",
308+
"phi-3",
309+
"qwen2",
310310
# New Ollama models
311-
"Ollama - gemma3:1b",
312-
"Ollama - gemma3",
313-
"Ollama - gemma3:12b",
314-
"Ollama - gemma3:27b",
315-
"Ollama - qwq",
316-
"Ollama - deepseek-r1",
317-
"Ollama - deepseek-r1:671b",
318-
"Ollama - llama3.3",
319-
"Ollama - llama3.2",
320-
"Ollama - llama3.2:1b",
321-
"Ollama - llama3.2-vision",
322-
"Ollama - llama3.2-vision:90b",
323-
"Ollama - llama3.1",
324-
"Ollama - llama3.1:405b",
325-
"Ollama - phi4",
326-
"Ollama - phi4-mini",
327-
"Ollama - mistral",
328-
"Ollama - moondream",
329-
"Ollama - neural-chat",
330-
"Ollama - starling-lm",
331-
"Ollama - codellama",
332-
"Ollama - llama2-uncensored",
333-
"Ollama - llava",
334-
"Ollama - granite3.2"
311+
"gemma3:1b",
312+
"gemma3",
313+
"gemma3:12b",
314+
"gemma3:27b",
315+
"qwq",
316+
"deepseek-r1",
317+
"deepseek-r1:671b",
318+
"llama3.3",
319+
"llama3.2",
320+
"llama3.2:1b",
321+
"llama3.2-vision",
322+
"llama3.2-vision:90b",
323+
"llama3.1",
324+
"llama3.1:405b",
325+
"phi4",
326+
"phi4-mini",
327+
"mistral",
328+
"moondream",
329+
"neural-chat",
330+
"starling-lm",
331+
"codellama",
332+
"llama2-uncensored",
333+
"llava",
334+
"granite3.2"
335335
])
336336
if openai_key:
337-
model_choices.append("OpenAI")
337+
model_choices.append("openai")
338338

339-
# Set default model to Ollama - qwen2
340-
default_model = "Ollama - qwen2"
339+
# Set default model to qwen2
340+
default_model = "qwen2"
341341

342342
# Model Management Tab (First Tab)
343343
with gr.Tab("Model Management"):

agentic_rag/local_rag_agent.py

Lines changed: 7 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -170,35 +170,25 @@ def __init__(self, vector_store: VectorStore = None, model_name: str = None,
170170
self.model_name = model_name
171171
# skip_analysis parameter kept for backward compatibility but no longer used
172172

173-
# Check if this is an Ollama model
174-
self.is_ollama = model_name and (model_name.startswith("ollama:") or "Ollama - " in model_name)
173+
# Check if this is an Ollama model (anything not Mistral is considered Ollama)
174+
self.is_ollama = not (model_name and "mistral" in model_name.lower())
175175

176176
if self.is_ollama:
177-
# Extract the actual model name from the prefix
178-
# If model_name contains 'ollama:' prefix, remove it
179-
# If model_name is from gradio interface (e.g., "Ollama - llama3"), extract just the model name
180-
if model_name.startswith("ollama:"):
181-
ollama_model_name = model_name.replace("ollama:", "")
182-
elif "Ollama - " in model_name:
183-
ollama_model_name = model_name.replace("Ollama - ", "").strip()
184-
else:
185-
ollama_model_name = model_name
186-
187177
# Add :latest suffix if not present
188-
if not ollama_model_name.endswith(":latest"):
189-
ollama_model_name = f"{ollama_model_name}:latest"
178+
if not model_name.endswith(":latest"):
179+
model_name = f"{model_name}:latest"
190180

191181
# Load Ollama model
192182
print("\nLoading Ollama model...")
193-
print(f"Model: {ollama_model_name}")
183+
print(f"Model: {model_name}")
194184
print("Note: Make sure Ollama is running on your system.")
195185

196186
# Initialize Ollama model handler
197-
self.ollama_handler = OllamaModelHandler(ollama_model_name)
187+
self.ollama_handler = OllamaModelHandler(model_name)
198188

199189
# Create pipeline-like interface
200190
self.pipeline = self.ollama_handler
201-
print(f"Using Ollama model: {ollama_model_name}")
191+
print(f"Using Ollama model: {model_name}")
202192
else:
203193
# Only initialize Mistral if no model is specified
204194
if not model_name:

0 commit comments

Comments
 (0)