Skip to content

Commit 39932fb

Browse files
committed
Updated RAG quickstarts and TOC
1 parent 24fb555 commit 39932fb

File tree

3 files changed

+29
-32
lines changed

3 files changed

+29
-32
lines changed

articles/search/search-get-started-portal-image-search.md

Lines changed: 21 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ ms.custom:
1515

1616
In this quickstart, you use the **Import and vectorize data** wizard in the Azure portal to get started with [multimodal search](multimodal-search-overview.md). The wizard simplifies the process of extracting, chunking, vectorizing, and loading both text and images into a searchable index.
1717

18-
Unlike [Quickstart: Vector search in the Azure portal](search-get-started-portal-import-vectors.md), which processes simple images, this quickstart supports advanced image processing for multimodal RAG scenarios.
18+
Unlike [Quickstart: Vector search in the Azure portal](search-get-started-portal-import-vectors.md), which processes simple text-containing images, this quickstart supports advanced image processing for multimodal RAG scenarios.
1919

2020
This quickstart uses a multimodal PDF from the [azure-search-sample-data](https://github.com/Azure-Samples/azure-search-sample-data/tree/main/sustainable-ai-pdf) repo. However, you can use different files and still complete this quickstart.
2121

@@ -55,7 +55,7 @@ For content embedding, you can choose either image verbalization (followed by te
5555

5656
<sup>1</sup> The endpoint of your Azure OpenAI resource must have a [custom subdomain](/azure/ai-services/cognitive-services-custom-subdomains), such as `https://my-unique-name.openai.azure.com`. If you created your resource in the [Azure portal](https://portal.azure.com/), this subdomain was automatically generated during resource setup.
5757

58-
<sup>2</sup> Azure OpenAI resources (with access to embedding models) that were created in the [Azure AI Foundry portal](https://ai.azure.com/) aren't supported. You must create an Azure OpenAI resource in the Azure portal.
58+
<sup>2</sup> Azure OpenAI resources (with access to embedding models) that were created in the [Azure AI Foundry portal](https://ai.azure.com/?cid=learnDocs) aren't supported. You must create an Azure OpenAI resource in the Azure portal.
5959

6060
<sup>3</sup> For billing purposes, you must [attach your multi-service resource](cognitive-search-attach-cognitive-services.md) to the skillset in your Azure AI Search service. Unless you use a [keyless connection](cognitive-search-attach-cognitive-services.md#bill-through-a-keyless-connection) to create the skillset, both resources must be in the same region.
6161

@@ -101,7 +101,7 @@ On your Azure AI Search service:
101101

102102
### [**Azure Storage**](#tab/storage-perms)
103103

104-
Azure Storage is both the data source for your documents and the destination for extracted images. Your search service requires access to these storage containers, which you create in the next section of this quickstart.
104+
Azure Storage is both the data source for your documents and the destination for extracted images. Your search service requires access to these storage containers, which you create in the next section.
105105

106106
On your Azure Storage account:
107107

@@ -111,7 +111,7 @@ On your Azure Storage account:
111111

112112
### Conditional roles
113113

114-
The following tabs cover all wizard-compatible resources for multimodal search. Only select the tabs that apply to your chosen [extraction method](#supported-extraction-methods) and [embedding method](#supported-embedding-methods).
114+
The following tabs cover all wizard-compatible resources for multimodal search. Select only the tabs that apply to your chosen [extraction method](#supported-extraction-methods) and [embedding method](#supported-embedding-methods).
115115

116116
### [**Azure OpenAI**](#tab/openai-perms)
117117

@@ -123,7 +123,7 @@ On your Azure OpenAI resource:
123123

124124
### [**Azure AI Foundry**](#tab/ai-foundry-perms)
125125

126-
The Azure AI Foundry model catalog provides large language models (LLMs) for image verbalization and embedding models for text and image vectorization. Your search service requires access to call the [GenAI Prompt skill](cognitive-search-skill-genai-prompt.md) and [AML skill](cognitive-search-aml-skill.md).
126+
The Azure AI Foundry model catalog provides LLMs for image verbalization and embedding models for text and image vectorization. Your search service requires access to call the [GenAI Prompt skill](cognitive-search-skill-genai-prompt.md) and [AML skill](cognitive-search-aml-skill.md).
127127

128128
On your Azure AI Foundry project:
129129

@@ -162,7 +162,7 @@ The wizard offers several options for content embedding. Image verbalization req
162162
163163
To deploy the models for this quickstart:
164164

165-
1. Sign in to the [Azure AI Foundry portal](https://ai.azure.com).
165+
1. Sign in to the [Azure AI Foundry portal](https://ai.azure.com/?cid=learnDocs).
166166

167167
1. Select your Azure OpenAI resource or Azure AI Foundry project.
168168

@@ -210,9 +210,9 @@ Depending on your chosen [extraction method](#supported-extraction-methods), the
210210

211211
### [**Default extraction**](#tab/document-extraction)
212212

213-
The default method calls the [Document Extraction skill](cognitive-search-skill-document-extraction.md) to extract content and metadata from documents, which includes generating normalized images. The [Text Split skill](cognitive-search-skill-textsplit.md) is then used to split the extracted content into pages.
213+
The default method calls the [Document Extraction skill](cognitive-search-skill-document-extraction.md) to extract text content and generate normalized images from your documents. The [Text Split skill](cognitive-search-skill-textsplit.md) is then called to split the extracted text content into pages.
214214

215-
To use Document Extraction skill:
215+
To use the Document Extraction skill:
216216

217217
1. On the **Content extraction** page, select **Default**.
218218

@@ -222,13 +222,13 @@ To use Document Extraction skill:
222222

223223
### [**Enhanced extraction**](#tab/document-intelligence)
224224

225-
Your Azure AI multi-service resource provides access to the [Document Layout skill](cognitive-search-skill-document-intelligence-layout.md), which extracts page numbers, bounding polygons, and other location metadata from both text and images. The Document Layout skill also breaks documents into smaller, more manageable chunks.
225+
Your Azure AI multi-service resource provides access to [Azure AI Document Intelligence](/azure/ai-services/document-intelligence/overview), which calls the [Document Layout skill](cognitive-search-skill-document-intelligence-layout.md) to recognize document structure and extract text and images relationally. It does so by attaching location metadata, such as page numbers and bounding polygons, to each image. The Document Layout skill also breaks text content into smaller, more manageable chunks.
226226

227227
To use the Document Layout skill:
228228

229229
1. On the **Content extraction** page, select **AI Document Intelligence**.
230230

231-
1. Specify your Azure subscription and Azure AI multi-service resource.
231+
1. Specify your Azure subscription and multi-service resource.
232232

233233
1. For the authentication type, select **System assigned identity**.
234234

@@ -248,7 +248,7 @@ During this step, the wizard uses your chosen [embedding method](#supported-embe
248248

249249
The wizard calls one skill to create descriptive text for images (image verbalization) and another skill to create vector embeddings for both text and images.
250250

251-
For image verbalization, the [GenAI Prompt skill](cognitive-search-skill-genai-prompt.md) uses the LLM you deployed to analyze each extracted image and produce a natural-language description.
251+
For image verbalization, the [GenAI Prompt skill](cognitive-search-skill-genai-prompt.md) uses your deployed LLM to analyze each extracted image and produce a natural-language description.
252252

253253
For embeddings, the [Azure OpenAI Embedding skill](cognitive-search-skill-azure-openai-embedding.md), [AML skill](cognitive-search-aml-skill.md), or [Azure AI Vision multimodal embeddings skill](cognitive-search-skill-vision-vectorize.md) uses your deployed embedding model to convert text chunks and verbalized descriptions into high-dimensional vectors. These vectors enable similarity and hybrid retrieval.
254254

@@ -332,14 +332,14 @@ On the **Advanced settings** page, you can optionally add fields to the index sc
332332

333333
| Field | Applies to | Description | Attributes |
334334
|--|--|--|--|
335-
| content_id | Text and image vectors | String field. Document key for the index. | Searchable, retrievable, sortable, filterable, and facetable. |
336-
| document_title | Text and image vectors | String field. Human-readable document title, page title, or page number. | Searchable, retrievable, sortable, filterable, and facetable. |
335+
| content_id | Text and image vectors | String field. Document key for the index. | Retrievable, sortable, and searchable. |
336+
| document_title | Text and image vectors | String field. Human-readable document title. | Retrievable and searchable. |
337337
| text_document_id | Text vectors | String field. Identifies the parent document from which the text chunk originates. | Retrievable and filterable. |
338-
| image_document_id | Image vectors | String field. Identifies the parent document from which the image originates. | Searchable, retrievable, sortable, filterable, and facetable. |
339-
| content_text | Text vectors | String field. Human-readable version of the text chunk. | Searchable, retrievable, sortable, filterable, and facetable. |
340-
| content_embedding | Image vectors | Collection(Edm.Single). Vector representation of the image verbalization. | Searchable and retrievable. |
341-
| content_path | Text and image vectors | String field. Path to the content in the storage container. | Retrievable, sortable, filterable, and facetable. |
342-
| locationMetadata | Text and image vectors | Edm.ComplexType. Contains metadata about the content's location. | Varies by field. |
338+
| image_document_id | Image vectors | String field. Identifies the parent document from which the image originates. | Retrievable and filterable. |
339+
| content_text | Text vectors | String field. Human-readable version of the text chunk. | Retrievable and searchable. |
340+
| content_embedding | Text and image vectors | Collection(Edm.Single). Vector representation of text and images. | Retrievable and searchable. |
341+
| content_path | Text and image vectors | String field. Path to the content in the storage container. | Retrievable and searchable. |
342+
| locationMetadata | Image vectors | Edm.ComplexType. Contains metadata about the image's location in the documents. | Varies by field. |
343343

344344
You can't modify the generated fields or their attributes, but you can add fields if your data source provides them. For example, Azure Blob Storage provides a collection of metadata fields.
345345

@@ -351,9 +351,6 @@ To add fields to the index schema:
351351

352352
1. Select a source field from the available fields, enter a field name for the index, and accept (or override) the default data type.
353353

354-
> [!NOTE]
355-
> Metadata fields are searchable but not retrievable, filterable, facetable, or sortable.
356-
357354
1. If you want to restore the schema to its original version, select **Reset**.
358355

359356
## Schedule indexing
@@ -392,9 +389,9 @@ When the wizard completes the configuration, it creates the following objects:
392389

393390
+ The [Document Extraction skill](cognitive-search-skill-document-extraction.md) or [Document Layout skill](cognitive-search-skill-document-intelligence-layout.md) extracts text and images from source documents. The [Text Split skill](cognitive-search-skill-textsplit.md) accompanies the Document Extraction skill for data chunking, while the Document Layout skill has built-in chunking.
394391

395-
+ The [GenAI Prompt skill](cognitive-search-skill-genai-prompt.md) verbalizes images. If you're using direct multimodal embeddings, this skill is absent.
392+
+ The [GenAI Prompt skill](cognitive-search-skill-genai-prompt.md) verbalizes images in natural language. If you're using direct multimodal embeddings, this skill is absent.
396393

397-
+ The [Azure OpenAI Embedding skill](cognitive-search-skill-azure-openai-embedding.md), [AML skill](cognitive-search-aml-skill.md), or [Azure AI Vision multimodal embeddings skill](cognitive-search-skill-vision-vectorize.md) is called twice to vectorize text and image content.
394+
+ The [Azure OpenAI Embedding skill](cognitive-search-skill-azure-openai-embedding.md), [AML skill](cognitive-search-aml-skill.md), or [Azure AI Vision multimodal embeddings skill](cognitive-search-skill-vision-vectorize.md) is called once for text vectorization and once for image vectorization.
398395

399396
+ The [Shaper skill](cognitive-search-skill-shaper.md) enriches the output with metadata and creates new images with contextual information.
400397

@@ -403,7 +400,7 @@ When the wizard completes the configuration, it creates the following objects:
403400
404401
## Check results
405402

406-
This quickstart creates a multimodal index that supports [hybrid search](hybrid-search-overview.md) over both text and verbalized images. However, it doesn't support images as query inputs, which requires integrated vectorization using an embedding skill and an equivalent vectorizer. For more information, see [Query with Search explorer](search-explorer.md).
403+
This quickstart creates a multimodal index that supports [hybrid search](hybrid-search-overview.md) over both text and images. Unless you use direct multimodal embeddings, the index doesn't accept images as query inputs, which requires the [AML skill](cognitive-search-aml-skill.md) or [Azure AI Vision multimodal embeddings skill](cognitive-search-skill-vision-vectorize.md) with an equivalent vectorizer. For more information, see [Query with Search explorer](search-explorer.md).
407404

408405
Hybrid search is a combination of full-text queries and vector queries. When you issue a hybrid query, the search engine computes the semantic similarity between your query and the indexed vectors and ranks the results accordingly. For the index created in this quickstart, the results surface content from the `content_text` field that closely aligns with your query.
409406

articles/search/search-get-started-portal-import-vectors.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ For integrated vectorization, you must use one of the following embedding models
5252

5353
<sup>1</sup> The endpoint of your Azure OpenAI resource must have a [custom subdomain](/azure/ai-services/cognitive-services-custom-subdomains), such as `https://my-unique-name.openai.azure.com`. If you created your resource in the [Azure portal](https://portal.azure.com/), this subdomain was automatically generated during resource setup.
5454

55-
<sup>2</sup> Azure OpenAI resources (with access to embedding models) that were created in the [Azure AI Foundry portal](https://ai.azure.com/) aren't supported. Only Azure OpenAI resources created in the Azure portal are compatible with the [Azure OpenAI Embedding skill](cognitive-search-skill-azure-openai-embedding.md).
55+
<sup>2</sup> Azure OpenAI resources (with access to embedding models) that were created in the [Azure AI Foundry portal](https://ai.azure.com/?cid=learnDocs) aren't supported. Only Azure OpenAI resources created in the Azure portal are compatible with the [Azure OpenAI Embedding skill](cognitive-search-skill-azure-openai-embedding.md).
5656

5757
<sup>3</sup> For billing purposes, you must [attach your multi-service resource](cognitive-search-attach-cognitive-services.md) to the skillset in your Azure AI Search service. Unless you use a [keyless connection](cognitive-search-attach-cognitive-services.md#bill-through-a-keyless-connection) to create the skillset, both resources must be in the same region.
5858

@@ -182,7 +182,7 @@ This section points you to the content that works for this quickstart. Before yo
182182

183183
## Prepare embedding model
184184

185-
The wizard can use embedding models deployed from Azure OpenAI, Azure AI Vision, or from the model catalog in the [Azure AI Foundry portal](https://ai.azure.com/). Before you proceed, make sure you completed the prerequisites for [role-based access](#role-based-access).
185+
The wizard can use embedding models deployed from Azure OpenAI, Azure AI Vision, or from the model catalog in the [Azure AI Foundry portal](https://ai.azure.com/?cid=learnDocs). Before you proceed, make sure you completed the prerequisites for [role-based access](#role-based-access).
186186

187187
### [Azure OpenAI](#tab/model-aoai)
188188

@@ -204,7 +204,7 @@ The wizard supports text-embedding-ada-002, text-embedding-3-large, and text-emb
204204

205205
1. To deploy an embedding model:
206206

207-
1. Sign in to the [Azure AI Foundry portal](https://ai.azure.com/) and select your Azure OpenAI resource.
207+
1. Sign in to the [Azure AI Foundry portal](https://ai.azure.com/?cid=learnDocs) and select your Azure OpenAI resource.
208208

209209
1. From the left pane, select **Model catalog**.
210210

@@ -255,7 +255,7 @@ For the model catalog, you should have an [Azure AI Foundry project](/azure/ai-f
255255

256256
1. To deploy an embedding model:
257257

258-
1. Sign in to the [Azure AI Foundry portal](https://ai.azure.com/) and select your project.
258+
1. Sign in to the [Azure AI Foundry portal](https://ai.azure.com/?cid=learnDocs) and select your project.
259259

260260
1. From the left pane, select **Model catalog**.
261261

@@ -409,7 +409,7 @@ However, if you work with content that includes useful images, you can apply AI
409409

410410
Azure AI Search and your Azure AI resource must be in the same region or configured for [keyless billing connections](cognitive-search-attach-cognitive-services.md).
411411

412-
1. On the **Vectorize your images** page, specify the kind of connection the wizard should make. For image vectorization, the wizard can connect to embedding models in the [Azure AI Foundry portal](https://ai.azure.com/) or Azure AI Vision.
412+
1. On the **Vectorize your images** page, specify the kind of connection the wizard should make. For image vectorization, the wizard can connect to embedding models in the [Azure AI Foundry portal](https://ai.azure.com/?cid=learnDocs) or Azure AI Vision.
413413

414414
1. Specify the subscription.
415415

articles/search/toc.yml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -32,11 +32,11 @@ items:
3232
href: search-get-started-rbac.md
3333
- name: Azure portal
3434
items:
35-
- name: Keyword search wizard
35+
- name: Create a search index
3636
href: search-get-started-portal.md
37-
- name: RAG wizard
37+
- name: Create a vector index
3838
href: search-get-started-portal-import-vectors.md
39-
- name: Multimodal RAG wizard
39+
- name: Create a multimodal index
4040
href: search-get-started-portal-image-search.md
4141
- name: Create a demo app
4242
href: search-create-app-portal.md

0 commit comments

Comments
 (0)