Skip to content

Commit f2d1556

Browse files
Merge pull request #270718 from wmwxwa/patch-8
Move up related concepts pt 3.md
2 parents ea680c0 + cbd33bf commit f2d1556

File tree

1 file changed

+16
-16
lines changed

1 file changed

+16
-16
lines changed

articles/cosmos-db/vector-database.md

Lines changed: 16 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -78,7 +78,22 @@ The process of creating good prompts for a scenario is called prompt engineering
7878

7979
Tokens are small chunks of text generated by splitting the input text into smaller segments. These segments can either be words or groups of characters, varying in length from a single character to an entire word. For instance, the word hamburger would be divided into tokens such as ham, bur, and ger while a short and common word like pear would be considered a single token. LLMs like ChatGPT, GPT-3.5, or GPT-4 break words into tokens for processing.
8080

81-
Here are multiple ways to implement RAG on your data by using our vector database functionalities:
81+
### Retrieval-augmented generation
82+
83+
Retrieval-augmentated generation (RAG) is an architecture that augments the capabilities of LLMs like ChatGPT, GPT-3.5, or GPT-4 by adding an information retrieval system like vector search that provides grounding data, such as those stored in a vector database. This approach allows your LLM to generate contextually relevant and accurate responses based on your custom data sourced from vectorized documents, images, audio, video, etc.
84+
85+
A simple RAG pattern using Azure Cosmos DB for NoSQL could be:
86+
87+
1. Insert data into an Azure Cosmos DB for NoSQL database and collection
88+
2. Create embeddings from a data property using Azure OpenAI Embeddings
89+
3. Link the Azure Cosmos DB for NoSQL to Azure Cognitive Search (for vector indexing/search)
90+
4. Create a vector index over the embeddings properties
91+
5. Create a function to perform vector similarity search based on a user prompt
92+
6. Perform question answering over the data using an Azure OpenAI Completions model
93+
94+
The RAG pattern, with prompt engineering, serves the purpose of enhancing response quality by offering more contextual information to the model. RAG enables the model to apply a broader knowledge base by incorporating relevant external sources into the generation process, resulting in more comprehensive and informed responses. For more information on "grounding" LLMs, see [grounding LLMs](https://techcommunity.microsoft.com/t5/fasttrack-for-azure/grounding-llms/ba-p/3843857).
95+
96+
Here are multiple ways to implement RAG on your data by using our integrated vector database functionalities:
8297

8398
## How to implement integrated vector database functionalities
8499

@@ -125,21 +140,6 @@ The natively integrated vector database in our NoSQL API will become available i
125140
> [!div class="nextstepaction"]
126141
> [Use the Azure Cosmos DB lifetime free tier](free-tier.md)
127142
128-
### Retrieval-augmented generation
129-
130-
Retrieval-augmentated generation (RAG) is an architecture that augments the capabilities of LLMs like ChatGPT, GPT-3.5, or GPT-4 by adding an information retrieval system like vector search that provides grounding data, such as those stored in a vector database. This approach allows your LLM to generate contextually relevant and accurate responses based on your custom data sourced from vectorized documents, images, audio, video, etc.
131-
132-
A simple RAG pattern using Azure Cosmos DB for NoSQL could be:
133-
134-
1. Insert data into an Azure Cosmos DB for NoSQL database and collection
135-
2. Create embeddings from a data property using Azure OpenAI Embeddings
136-
3. Link the Azure Cosmos DB for NoSQL to Azure Cognitive Search (for vector indexing/search)
137-
4. Create a vector index over the embeddings properties
138-
5. Create a function to perform vector similarity search based on a user prompt
139-
6. Perform question answering over the data using an Azure OpenAI Completions model
140-
141-
The RAG pattern, with prompt engineering, serves the purpose of enhancing response quality by offering more contextual information to the model. RAG enables the model to apply a broader knowledge base by incorporating relevant external sources into the generation process, resulting in more comprehensive and informed responses. For more information on "grounding" LLMs, see [grounding LLMs](https://techcommunity.microsoft.com/t5/fasttrack-for-azure/grounding-llms/ba-p/3843857).
142-
143143
## Related content
144144

145145
- [Azure Cosmos DB for MongoDB vCore Integrated Vector Database](mongodb/vcore/vector-search.md)

0 commit comments

Comments
 (0)