This is a Redis LangCache semantic caching demo for Python using:
- make
- python>=3.10
- uv
- docker
- Optional outside tests
Copy and edit the .env file:
cp .env.example .envYour .env file needs three things:
- A Redis Cloud connection string (
REDIS_URL) - LangCache service credentials (
LANGCACHE_API_URL,LANGCACHE_CACHE_ID,LANGCACHE_API_KEY)
See Connecting to Redis Cloud and Setting up LangCache for details.
Available settings:
APP_ENV=development|test|productionLOG_LEVEL=DEBUG|INFO|WARNING|ERROR|CRITICALLOG_STREAM_KEY=logsPORT=8080REDIS_URL=redis://...LANGCACHE_API_URL=https://...LANGCACHE_CACHE_ID=your-cache-idLANGCACHE_API_KEY=your-api-keyLANGCACHE_CACHE_THRESHOLD=0.65LANGCACHE_KNOWLEDGE_THRESHOLD=0.35
For docker, .env.docker should use container-internal addresses. Example:
APP_ENV=production
LOG_LEVEL=INFO
LOG_STREAM_KEY=logs
REDIS_URL="redis://redis:6379"
LANGCACHE_API_URL="https://your-langcache-host"
LANGCACHE_CACHE_ID="your-cache-id"
LANGCACHE_API_KEY="your-api-key"Next, spin up docker containers:
make dockerYou should have a server running on http://localhost:<port> where the port is set in your .env file (default is 8080). You can test the following routes:
POST /api/langcache/ask- Ask the semantic cache a support question with{ "question": "How do I reset my password?" }GET /api/langcache/stats- Read request, hit, miss, and hit-rate metrics
The answer payload includes:
cacheHit- whether LangCache returned a cached responsematchedPrompt- the prompt that matched the querysimilarity- how similar the query was to the matched promptentryId- the LangCache entry IDsource-"cache"for hits,"knowledge_base"for misses
- A user sends a question to the
/api/langcache/askendpoint - The app searches LangCache for a semantically similar cached response
- If LangCache finds a match (cache hit), the cached response is returned immediately
- If no match is found (cache miss), the app generates an answer from the knowledge base, stores it in LangCache for future queries, and returns it
LangCache handles embedding generation and vector similarity search. The knowledge base serves as a simulated LLM for this demo.
Requests and component logs are written to stdout. They are also shipped to Redis as stream entries via XADD on the key configured by LOG_STREAM_KEY (default logs).
The test suite lives in __test__ and can be run with:
make testIf REDIS_URL points at the default local Redis and nothing is listening on it, the tests will start the redis service with docker compose automatically and stop it afterward.
Install dependencies and run the dev server:
make install
make devFor a production-style server:
make serveRun make to see the list of available commands.
Useful targets:
make formatmake lintmake updatemake lock
If you don't yet have a database setup in Redis Cloud get started here for free.
To connect to a Redis Cloud database, log into the console and find the following:
- The
public endpoint - Your
username - Your
password
Combine them into a connection string and put it in .env and .env.docker. Example:
REDIS_URL="redis://default:<password>@redis-#####.c###.us-west-2-#.ec2.redns.redis-cloud.com:#####"Run make test to verify the connection.
LangCache is available on Redis Cloud. After creating a LangCache service, grab these values from the Configuration page:
- The LangCache API base URL
- The Cache ID
- The API key (only shown at creation time)
Add them to your .env file:
LANGCACHE_API_URL="https://your-langcache-host"
LANGCACHE_CACHE_ID="your-cache-id"
LANGCACHE_API_KEY="your-api-key"