Priority: HIGH Estimated Time: 30-45 minutes Risk Level: LOW (drop-in replacement, old setup stays as backup)
Deploy Razer AIKit as a Docker sidecar to Visionary Tool Server, adding 22 new MCP tools for local LLM inference with 280K+ HuggingFace models.
This is a DROP-IN REPLACEMENT:
- Same port 8082
- Same ngrok domain
- All existing client configs work unchanged
Location: /home/claude/razer-aikit-deployment/ (or provided by user)
- docker-compose.aikit.yml →
D:\DEV_PROJECTS\GitHub\Claude_Opus_ChatGPT_App_Project\ - Dockerfile.aikit →
D:\DEV_PROJECTS\GitHub\Claude_Opus_ChatGPT_App_Project\ - app/tools/razer_aikit.py →
D:\DEV_PROJECTS\GitHub\Claude_Opus_ChatGPT_App_Project\app\tools\ - RAZER_AIKIT_DEPLOYMENT.md →
D:\DEV_PROJECTS\GitHub\Claude_Opus_ChatGPT_App_Project\
New-Item -ItemType Directory -Force -Path "D:\Visionary_Models\aikit"
New-Item -ItemType Directory -Force -Path "D:\Visionary_Models\aikit-cache"Copy all 4 files to their respective locations (listed above).
Edit D:\DEV_PROJECTS\GitHub\Claude_Opus_ChatGPT_App_Project\app\server.py:
Add import:
from app.tools import razer_aikitCheck .env.master for HUGGINGFACE_API_KEY. If missing:
HUGGINGFACE_API_KEY=hf_xxxxxxxxxxxx
(User must provide their token from https://huggingface.co/settings/tokens)
cd D:\DEV_PROJECTS\GitHub\Claude_Opus_ChatGPT_App_Project
docker compose -f docker-compose.aikit.yml build --no-cachedocker compose -f docker-compose.aikit.yml up -d# Wait 60 seconds for AIKit to initialize
Start-Sleep -Seconds 60
# Check health endpoints
curl http://localhost:8000/health
curl http://localhost:8082/healthcurl -X POST http://localhost:8000/v1/chat/completions `
-H "Content-Type: application/json" `
-d '{"model": "deepseek-ai/deepseek-coder-1.3b-instruct", "messages": [{"role": "user", "content": "Hello"}], "max_tokens": 50}'-
docker compose psshows 2-3 healthy containers - AIKit health returns
{"status":"ok"} - Tool Server health shows
tools: 334(312 + 22) - Test chat completion returns valid response
- GPU visible inside container (
docker exec razer-aikit nvidia-smi)
Fix: Restart Docker Desktop, ensure WSL2 integration enabled
Fix: Stop existing services on 8000/8082 or change ports in docker-compose
Fix: Check HuggingFace token, verify internet connectivity
Fix: Use smaller model (1.3B instead of 7B)
If anything fails:
docker compose -f docker-compose.aikit.yml down
# Original setup on port 8082 unaffectedAfter execution, provide:
- Output of
docker compose ps - Output of health checks
- Output of test chat completion
- Any errors encountered
- Tool count from Tool Server health
- All existing clients auto-connect (same port 8082, same ngrok domain)
- Test
aikit_chattool via MCP - Try larger models:
Qwen/Qwen2.5-7B-Instruct - Enable ngrok tunnel if mobile access needed
- Old setup stays on ice as backup (don't delete yet)