You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -55,7 +56,6 @@ Need specific sample code to help get started with Redis? Start here.
55
56
|[/redis-intro/redis_intro.ipynb](python-recipes/redis-intro/redis_intro.ipynb)| The place to start if brand new to Redis |
56
57
|[/vector-search/00_redispy.ipynb](python-recipes/vector-search/00_redispy.ipynb)| Vector search with Redis python client |
57
58
|[/vector-search/01_redisvl.ipynb](python-recipes/vector-search/01_redisvl.ipynb)| Vector search with Redis Vector Library |
58
-
59
59
## Getting started with RAG
60
60
61
61
**Retrieval Augmented Generation** (aka RAG) is a technique to enhance the ability of an LLM to respond to user queries. The **retrieval** part of RAG is supported by a vector database, which can return semantically relevant results to a user’s query, serving as contextual information to **augment** the **generative** capabilities of an LLM.
@@ -71,6 +71,12 @@ To get started with RAG, either from scratch or using a popular framework like L
71
71
|[/RAG/05_nvidia_ai_rag_redis.ipynb](python-recipes/RAG/05_nvidia_ai_rag_redis.ipynb)| RAG using Redis and Nvidia |
72
72
|[/RAG/06_ragas_evaluation.ipynb](python-recipes/RAG/06_ragas_evaluation.ipynb)| Utilize RAGAS framework to evaluate RAG performance |
73
73
74
+
## LLM Session Management
75
+
LLMs are stateless. To maintain context within a conversation chat sessions must be stored and resent to the LLM. Redis manages the storage and retrieval of chat sessions to maintain context and conversational relevance.
76
+
| Recipe | Description |
77
+
| --- | --- |
78
+
|[/llm-session-manager/00_session_manager.ipynb](python-recipes/llm-session-manager/00_llm_session_manager.ipynb)| LLM session manager with semantic similarity |
79
+
|[/llm-session-manager/01_multiple_sessions.ipynb](python-recipes/llm-session-manager/01_multiple_sessions.ipynb)| Handle multiple simultaneous chats with one instance |
74
80
75
81
## Semantic Cache
76
82
An estimated 31% of LLM queries are potentially redundant ([source](https://arxiv.org/pdf/2403.02694)). Redis enables semantic caching to help cut down on LLM costs quickly.
0 commit comments