This is an Advanced Retrieval-Augmented Generation (RAG) system built locally to chat with private documents. This version is optimized to run 100% locally on your machine using Ollama, saving costs and ensuring data privacy.
- Framework: LangChain π¦π
- LLM: Llama 3.2:1b (Local via Ollama) π¦
- Embeddings: mxbai-embed-large (1024-dim) π’
- Vector Store: FAISS (Facebook AI Similarity Search) β‘
- Database: Pickle (Metadata storage) πΎ
- Environment: Python 3.10+ π
- Clone the Repository π
git clone https://github.com/Shahryar-Sohail/local-rag/
cd local-rag- Create & Activate Virtual Environment π¦
python -m venv .venv.venv\Scripts\activate- Install Dependencies π¦
pip install -r requirements.txt- Setup Local Models (Ollama) π₯
ollama pull llama3.2:1b
ollama pull mxbai-embed-large5.π Running the Project To test the backend pipeline and see the AI in action:
python app.py