Replies: 1 comment
-
✅ Implemented in v10.2.0!Great news! External embedding API support (including TEI) is now available in v10.2.0! 🎉 What's NewYou can now use external OpenAI-compatible embedding APIs instead of local models, including:
Configuration# TEI example (as requested in this discussion)
export MCP_EXTERNAL_EMBEDDING_URL=http://localhost:8080/v1/embeddings
export MCP_EXTERNAL_EMBEDDING_MODEL=nomic-ai/nomic-embed-text-v1.5
# vLLM example
export MCP_EXTERNAL_EMBEDDING_URL=http://localhost:8890/v1/embeddings
export MCP_EXTERNAL_EMBEDDING_MODEL=nomic-ai/nomic-embed-text-v1.5
# Ollama example
export MCP_EXTERNAL_EMBEDDING_URL=http://localhost:11434/v1/embeddings
export MCP_EXTERNAL_EMBEDDING_MODEL=nomic-embed-text
# Optional: API key for authenticated endpoints
export MCP_EXTERNAL_EMBEDDING_API_KEY=sk-xxxKey Features
Multi-User Deployment BenefitsThis addresses the multi-user deployment concerns mentioned in this discussion:
Important Note
DocumentationComplete setup guide with examples for all supported backends: Get Started# Install or upgrade
pip install --upgrade mcp-memory-service
# Or specify exact version
pip install mcp-memory-service==10.2.0CreditsSpecial thanks to @isiahw1 for implementing this feature! Release: https://github.com/doobidoo/mcp-memory-service/releases/tag/v10.2.0 |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
🚀 Proposal: Optional TEI Backend for Multi-User Deployments
Problem Statement
MCP Memory Service currently uses ONNX Runtime for embedding generation, which works excellently for single-user deployments:
However, this architecture may have limitations for high-concurrency server deployments:
Proposed Solution
Integrate HuggingFace Text Embeddings Inference (TEI) as an optional embedding backend for users who need higher throughput:
Key Benefits
Architecture
Backward Compatible Integration:
Deployment Flow:
graph LR A[User Request] --> B{Backend Type?} B -->|onnx| C[ONNX In-Process] B -->|tei| D[TEI HTTP Server] C --> E[Embeddings] D --> ETEI Advantages
Implementation Plan
Phase 1: Validation (2 weeks)
all-MiniLM-L6-v2,paraphrase-multilingual-mpnet-base-v2)Phase 2: Optional Backend (4 weeks)
TEIEmbeddingClientclassMCP_EMBEDDING_BACKENDconfig flagPhase 3: Production Hardening (2 weeks)
Total Effort: 8 weeks (1 developer, part-time)
Open Questions for Community
Related Context
MCP_EMBEDDING_MODELenv variable (PR fix: Use EMBEDDING_MODEL_NAME from config instead of hardcoded value #276, issue Bug: MCP_EMBEDDING_MODEL environment variable ignored in server.py eager/lazy init #275)Decision Criteria
Recommended to proceed if:
Not recommended if:
Next Steps
Call to Action:
References:
Beta Was this translation helpful? Give feedback.
All reactions