|
10 | 10 | "\n",
|
11 | 11 | "# Build a RAG chatbot\n",
|
12 | 12 | "\n",
|
13 |
| - "This notebook shows you how to build a simple RAG chatbot in Python using Pinecone as the vector database, [OpenAI](https://docs.pinecone.io/integrations/openai) for the embedding model and LLM, and [LangChain](https://docs.pinecone.io/integrations/langchain) for the RAG workflow.\n", |
| 13 | + "This page shows you how to build a simple RAG chatbot in Python using Pinecone for the vector database and embedding model, [OpenAI](https://docs.pinecone.io/integrations/openai) for the LLM, and [LangChain](https://docs.pinecone.io/integrations/langchain) for the RAG workflow.\n", |
14 | 14 | "\n",
|
15 | 15 | "## How it works\n",
|
16 | 16 | "\n",
|
|
129 | 129 | "source": [
|
130 | 130 | "## Store knowledge in Pinecone\n",
|
131 | 131 | "\n",
|
132 |
| - "For this guide, you'll use a document about a fictional product called the WonderVector5000 that LLMs do not have any information about. You'll use LangChain to chunk the document into smaller segments, convert each segment into vectors using OpenAI, and then upsert your vectors into your Pinecone index.\n" |
| 132 | + "For this guide, you'll use a document about a fictional product called the WonderVector5000 that LLMs do not have any information about. You'll use LangChain to chunk the document into smaller segments, convert each segment into vectors using, and then upsert your vectors into your Pinecone index.\n" |
133 | 133 | ]
|
134 | 134 | },
|
135 | 135 | {
|
|
277 | 277 | "id": "1GKq1M_5tiHX"
|
278 | 278 | },
|
279 | 279 | "source": [
|
280 |
| - "Initialize a LangChain embedding object. Note that this step uses the OpenAI API key you set as an environment variable earlier. OpenAI is a paid service, so running the remainder of this guide may incur some small cost." |
| 280 | + "Initialize a LangChain embedding object. Note that this step uses a Pinecone API key you set as an environment variable earlier." |
281 | 281 | ]
|
282 | 282 | },
|
283 | 283 | {
|
|
370 | 370 | "id": "fWsPiLBZ67wG"
|
371 | 371 | },
|
372 | 372 | "source": [
|
373 |
| - "Now create a serverless index in Pinecone for storing the embeddings of your document, setting the index dimensions and distance metric to match those of the OpenAI `text-embedding-3-small` model you'll use to create the embeddings:\n", |
| 373 | + "Now create a serverless index in Pinecone for storing the embeddings of your document, setting the index dimensions and distance metric to match those of the Pinecone `multilingual-e5-large` model you'll use to create the embeddings:\n", |
374 | 374 | "\n"
|
375 | 375 | ]
|
376 | 376 | },
|
|
2035 | 2035 | "\n",
|
2036 | 2036 | "Now that your document is stored as embeddings in Pinecone, when you send questions to the LLM, you can add relevant knowledge from your Pinecone index to ensure that the LLM returns an accurate response.\n",
|
2037 | 2037 | "\n",
|
| 2038 | + "OpenAI is a paid service, so running the remainder of this guide may incur some small cost.\n", |
| 2039 | + "\n", |
2038 | 2040 | "Initialize a LangChain object for interacting with the `gpt-3.5-turbo` LLM:"
|
2039 | 2041 | ]
|
2040 | 2042 | },
|
|
0 commit comments