The Multi-Agent Museum Assistant project aims to provide an interactive and intelligent assistant for museum visitors. The system leverages multiple agents to enhance the visitor experience by providing detailed information and guidance throughout the museum.
- Interactive Assistance: Provides real-time information and guidance to museum visitors.
- Multi-Agent System: Utilizes multiple agents to handle different tasks and improve efficiency.
- Customized Museum Experience: Uses LLMs and RAG to provide customized answers to the visitors
The following steps require CUDA installed on your system
-
Install Ollama:
curl -fsSL https://ollama.com/install.sh | sh -
Download Ollama Models:
ollama serve # Start Ollama ollama pull deepseek-r1 # LLM model (see src/config.py) ollama pull bge-m3 # embedding model (see src/config.py)
-
Install Docker (Optional):
3.1. Add Docker's official GPG key:sudo apt-get update sudo apt-get install ca-certificates curl sudo install -m 0755 -d /etc/apt/keyrings sudo curl -fsSL https://download.docker.com/linux/ubuntu/gpg -o /etc/apt/keyrings/docker.asc sudo chmod a+r /etc/apt/keyrings/docker.asc
3.2. Install Docker:
sudo apt-get install docker-ce docker-ce-cli containerd.io docker-buildx-plugin docker-compose-plugin sudo apt-get install -y docker-ce-rootless-extras
3.3. Install GPU support for containers:
curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg \ && curl -s -L https://nvidia.github.io/libnvidia-container/stable/deb/nvidia-container-toolkit.list | \ sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \ sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
sudo apt-get update sudo apt-get install -y nvidia-container-toolkit
3.4. Verify that the installation is successful by running the hello-world image:
sudo docker run hello-world
-
Download Local Translations Server (Optional):
You can either:
Download the repo in a separate project (follow the provided installation and execution steps)git clone https://github.com/atzoriandrea/Local-Translation-Server.git
or Download and run this docker image (link) with the following command
docker load --input translation-server.tar docker run --gpus all -d -p 2048:2048 translation-server # It will listen on port 2048
A docker image of this system will be provided soon. In the meanwhile...
To set up the project, follow these steps:
-
Clone the repository:
git clone https://github.com/yourusername/multi-agent-museum-assistant.git cd multi-agent-museum-assistant -
Set up the Python environment:
python3 -m venv .venv source .venv/bin/activate -
Install the dependencies:
pip install --no-cache-dir -r requirements.txt
-
Get your DeepL API key:
Create and download an API key from this link (You MUST be registered to DeepL)Copy your key into the
api_keyfield insidesrc/api.py
To run the application, use the following commands:
cd src
streamlit run orchestrator.py