Skip to content

Commit c39675a

Browse files
authored
Merge pull request #248848 from sandeepsnairms/DMA-ADS
MongoDB vCore Vector Search -add sample git links
2 parents d964f82 + 2a1ef5b commit c39675a

File tree

2 files changed

+50
-0
lines changed

2 files changed

+50
-0
lines changed

articles/cosmos-db/mongodb/vcore/TOC.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,8 @@
1616
items:
1717
- name: Node.js
1818
href: tutorial-nodejs-web-app.md
19+
- name: Build AI Apps with Vector Search
20+
href: tutorial-vector-search-in-ai-apps.md
1921
- name: Concepts
2022
items:
2123
- name: Vector search
Lines changed: 48 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,48 @@
1+
---
2+
title: Build AI Apps with Azure Cosmos DB for MongoDB vCore Vector Search
3+
titleSuffix: Build AI Apps with Azure Cosmos DB for MongoDB vCore Vector Search
4+
description: Enhance AI-powered Applications with Retrieval Augmented Generation (RAG) using Azure Cosmos DB for MongoDB vCore Vector Search.
5+
ms.service: cosmos-db
6+
ms.subservice: mongodb-vcore
7+
ms.topic: tutorial
8+
author: sandnair
9+
ms.author: sandnair
10+
ms.reviewer: sandnair
11+
ms.date: 08/22/2023
12+
---
13+
14+
# AI Apps with Azure Cosmos DB for MongoDB vCore Vector Search
15+
16+
[!INCLUDE[MongoDB vCore](../../includes/appliesto-mongodb-vcore.md)]
17+
18+
## Introduction
19+
20+
Large Language Models (LLMs) available in Azure OpenAI are potent tools that can elevate the capabilities of your AI-driven applications. To fully unleash the potential of LLMs, giving them access to timely and relevant data from your application's data store is crucial. This process, known as Retrieval Augmented Generation (RAG), can be seamlessly accomplished using Azure Cosmos DB. In this tutorial, we delve into the core concepts of RAG and provide links to tutorials and sample code that exemplify powerful RAG strategies using Azure Cosmos DB for MongoDB vCore vector search.
21+
22+
Retrieval Augmented Generation (RAG) elevates AI-powered applications by incorporating external knowledge and data into model inputs. With Azure Cosmos DB for MongoDB vCore's vector search, this process becomes seamless, ensuring that the most pertinent information is effortlessly integrated into your AI models. By applying the power of [embeddings](../../../ai-services/openai/tutorials/embeddings.md) and vector search, you can provide your AI applications with the context they need to excel. Through the provided tutorials and code samples, you can become proficient in harnessing RAG to create smarter and more context-aware AI solutions.
23+
24+
## Understanding Retrieval Augmented Generation (RAG)
25+
26+
Retrieval Augmented Generation harnesses external knowledge and models to efficiently manage custom data or domain-specific expertise. This involves extracting pertinent information from an external data source and seamlessly integrating it into the model's input through prompt engineering. A robust approach is essential to identify the most pertinent data from the external source within the [token limitations of a request](../../../ai-services/openai/quotas-limits.md). This limitation is elegantly addressed by using embeddings, which convert data into vectors, capturing the semantic essence of the text and enabling context comprehension beyond simple keywords.
27+
28+
## What is vector search?
29+
30+
[Vector search](./vector-search.md) is an approach that enables the discovery of analogous items based on shared data characteristics, deviating from the necessity for precise matches within a property field. This method proves invaluable in various applications like text similarity searches, image association, recommendation systems, and anomaly detection. Its functionality revolves around the utilization of vector representations (sequences of numerical values) generated from your data via machine learning models or embeddings APIs. Examples of such APIs encompass [Azure OpenAI Embeddings](/azure/ai-services/openai/how-to/embeddings) or [Hugging Face on Azure](https://azure.microsoft.com/solutions/hugging-face-on-azure/). The technique gauges the disparity between your query vector and the data vectors. The data vectors that exhibit the closest proximity to your query vector are identified as semantically akin.
31+
32+
33+
## Utilizing Vector Search with Azure Cosmos DB for MongoDB vCore
34+
35+
RAG's power is truly harnessed through the native vector search capability within Azure Cosmos DB for MongoDB vCore. This enables a seamless fusion of AI-focused applications with stored data in Azure Cosmos DB. Vector search optimally stores, indexes, and searches high-dimensional vector data directly within Azure Cosmos DB for MongoDB vCore alongside other application data. This eliminates the need to migrate data to costlier alternatives for vector search functionality.
36+
37+
## Code samples and tutorials
38+
39+
- [**.NET Retail Chatbot Demo**](https://github.com/AzureCosmosDB/VectorSearchAiAssistant/tree/mongovcorev2): Learn how to build a chatbot using .NET that demonstrates RAG's potential in a retail context.
40+
- [**.NET Tutorial - Recipe Chatbot**](https://github.com/microsoft/AzureDataRetrievalAugmentedGenerationSamples/tree/main/C%23/CosmosDB-MongoDBvCore): Walk through creating a recipe chatbot using .NET, showcasing RAG's application in a culinary scenario.
41+
- [**Python Notebook Tutorial**](https://github.com/microsoft/AzureDataRetrievalAugmentedGenerationSamples/tree/main/Python/CosmosDB-MongoDB-vCore) - Azure Product Chatbot: Explore a Python notebook tutorial that guides you through constructing an Azure product chatbot, highlighting RAG's benefits.
42+
43+
44+
## Next steps
45+
46+
> [!div class="nextstepaction"]
47+
> [Introduction to Azure Cosmos DB for MongoDB vCore](introduction.md)
48+

0 commit comments

Comments
 (0)