A basic implementation guide for building a Retrieval-Augmented Generation (RAG) system using LlamaIndex.
- Python 3.10+
- Ollama
- Install Ollama:
macOS
curl -fsSL https://ollama.com/install.sh | shLinux
curl -fsSL https://ollama.com/install.sh | sh- Pull the Llama 2 model:
ollama pull llama3.2This project demonstrates how to:
- Set up a basic RAG system using LlamaIndex
- Integrate with Ollama for local LLM inference
- Process and index documents for retrieval
- Generate contextual responses using the indexed knowledge
- Clone this repository
- Follow the installation steps above
- Run the Jupyter notebook
workflow.ipynbto see the RAG system in action
Make sure Ollama is running in the background before executing the notebook:
ollama serve