Skip to content

JulianDataScienceExplorerV2/Local-LLM-Chatbot-GUI-Ollama

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Local LLM Chatbot GUI (Ollama + Python)

A beautiful, local-first graphical user interface (GUI) written in Python to chat with state-of-the-art open source Large Language Models (LLMs) running completely offline via Ollama.

Una hermosa interfaz gráfica local (GUI) escrita en Python para chatear con LLMs open-source de última generación corriendo 100% offline a través de Ollama.


Captura de pantalla 2026-02-25 224441

Features / Caracteristicas

  • 100% Private: Runs entirely on your local hardware. No API keys, no internet connection required.
  • Graphical Interface: Custom desktop UI built using Python GUI libraries.
  • Model Switching: Select between LLaMA 3, Mistral, Gemma, or any model installed locally.
  • Streaming Responses: Real-time token generation for instant feedback.

Tech Stack / Lenguajes

  • Language: Python 3.10+
  • Backend API: requests (connecting to localhost Ollama listener)
  • GUI Framework: Tkinter (Customized Desktop Window Engine)
  • LLM Engine: Ollama inference wrapper

Requirements / Requisitos

  1. You must have Ollama installed and running locally.
  2. Pull at least one model:
    ollama pull llama3

How to Run / Como Ejecutar

# Clone the client
git clone https://github.com/JulianDataScienceExplorerV2/Local-LLM-Chatbot-GUI-Ollama.git
cd Local-LLM-Chatbot-GUI-Ollama

# Run the Chat interface
python app.py

Julian David Urrego Lancheros
Generative AI & Python Developer

About

Local LLM Chatbot GUI. Interacts with open-source models offline via Ollama. Built in Python with a sleek interface. / Interfaz grafica local para chatear offline con modelos open-source usando Ollama.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages