You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/search/cognitive-search-defining-skillset.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ ms.service: azure-ai-search
8
8
ms.custom:
9
9
- ignite-2023
10
10
ms.topic: conceptual
11
-
ms.date: 12/06/2024
11
+
ms.date: 04/18/2025
12
12
---
13
13
14
14
# Create a skillset in Azure AI Search
@@ -192,7 +192,7 @@ Skills read from and write to an enriched document. Skill inputs specify the ori
192
192
|`source`: `/document/some-named-field`| For text-based skills, such as entity recognition or key phrase extraction, the origin should be a field that contains sufficient text to be analyzed, such as a *description* or *summary*. |
193
193
|`source`: `/document/normalized_images/*`| For image content, the source is image that's been normalized during document cracking. |
194
194
195
-
If the skill iterates over an array, both context and input source should include `/*` in the correct positions.
195
+
If the skill iterates over an array, both context and input source should include `/*` in the correct positions. For more information about the complete syntax, see [Skill context and input annotation language](cognitive-search-skill-annotation-language.md).
Copy file name to clipboardExpand all lines: articles/search/cognitive-search-skill-vision-vectorize.md
+38-24Lines changed: 38 additions & 24 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -9,44 +9,55 @@ ms.custom:
9
9
- build-2024
10
10
- references_regions
11
11
ms.topic: reference
12
-
ms.date: 08/05/2024
12
+
ms.date: 04/18/2025
13
13
---
14
14
15
15
# Azure AI Vision multimodal embeddings skill
16
16
17
17
> [!IMPORTANT]
18
-
> This skill is in public preview under [Supplemental Terms of Use](https://azure.microsoft.com/support/legal/preview-supplemental-terms/). The [2024-05-01-Preview REST API](/rest/api/searchservice/skillsets/create-or-update?view=rest-searchservice-2024-05-01-Preview&preserve-view=true)supports this feature.
18
+
> This skill is in public preview under [Supplemental Terms of Use](https://azure.microsoft.com/support/legal/preview-supplemental-terms/). The [2024-05-01-Preview REST API](/rest/api/searchservice/skillsets/create-or-update?view=rest-searchservice-2024-05-01-Preview&preserve-view=true)and newer preview APIs support this feature.
19
19
20
20
The **Azure AI Vision multimodal embeddings** skill uses Azure AI Vision's [multimodal embeddings API](/azure/ai-services/computer-vision/concept-image-retrieval) to generate embeddings for image or text input.
21
21
22
-
The skill is only supported in search services located in a region that supports the [Azure AI Vision Multimodal embeddings API](/azure/ai-services/computer-vision/how-to/image-retrieval). Review [region availability for multimodal embeddings](/azure/ai-services/computer-vision/overview-image-analysis?tabs=4-0#region-availability). Your data is processed in the [Geo](https://azure.microsoft.com/explore/global-infrastructure/data-residency/) where your model is deployed.
22
+
This skill must be [attached to a billable Azure AI multi-service resource](cognitive-search-attach-cognitive-services.md) for transactions that exceed 20 documents per indexer per day. Execution of built-in skills is charged at the existing [Azure AI services pay-as-you go price](https://azure.microsoft.com/pricing/details/cognitive-services/).
23
23
24
-
> [!NOTE]
25
-
> This skill is bound to Azure AI services and requires [a billable resource](cognitive-search-attach-cognitive-services.md) for transactions that exceed 20 documents per indexer per day. Execution of built-in skills is charged at the existing [Azure AI services pay-as-you go price](https://azure.microsoft.com/pricing/details/cognitive-services/).
26
-
>
27
-
> In addition, image extraction is [billable by Azure AI Search](https://azure.microsoft.com/pricing/details/search/).
28
-
>
24
+
In addition, image extraction is [billable by Azure AI Search](https://azure.microsoft.com/pricing/details/search/).
25
+
26
+
Location of resources is an important consideration. Because you're using a preview API version to create a skillset that contains preview skills, you have the option of a [keyless connection](cognitive-search-attach-cognitive-services.md#bill-through-a-keyless-connection), which relaxes the region requirement. However, if you're connecting with an API key, then Azure AI Search and Azure AI multi-service must be in the same region.
27
+
28
+
+ First, find a [supported region for multimodal embeddings](/azure/ai-services/computer-vision/overview-image-analysis?tabs=4-0#region-availability).
29
+
30
+
+ Second, verify the [region provides AI enrichment](search-region-support.md).
31
+
32
+
The Azure AI multi-service resource is used for billing purposes only. Content processing occurs on separate resources managed and maintained by Azure AI Search within the same geo. Your data is processed in the [Geo](https://azure.microsoft.com/explore/global-infrastructure/data-residency/) where your resource is deployed.
29
33
30
34
## @odata.type
31
35
32
36
Microsoft.Skills.Vision.VectorizeSkill
33
37
34
38
## Data limits
35
39
36
-
The input limits for the skill can be found in [the Azure AI Vision documentation](/azure/ai-services/computer-vision/concept-image-retrieval#input-requirements) for images and text respectively. Consider using the [Text Split skill](cognitive-search-skill-textsplit.md) if you need data chunking for text inputs.
40
+
The input limits for the skill can be found in the [Azure AI Vision documentation](/azure/ai-services/computer-vision/concept-image-retrieval#input-requirements) for images and text respectively. Consider using the [Text Split skill](cognitive-search-skill-textsplit.md) if you need data chunking for text inputs.
41
+
42
+
Applicable inputs include:
43
+
44
+
+ Image input file size must be less than 20 megabytes (MB). Image size must be greater than 10 x 10 pixels and less than 16,000 x 16,000 pixels.
45
+
+ Text input string must be between (inclusive) one word and 70 words.
37
46
38
47
## Skill parameters
39
48
40
49
Parameters are case-sensitive.
41
50
42
51
| Inputs | Description |
43
52
|---------------------|-------------|
44
-
|`modelVersion`| (Required) The model version to be passed to the Azure AI Vision multimodal embeddings API for generating embeddings. It's important that all embeddings stored in a given index field are generated using the same `modelVersion`. For information about version support for this model, refer to [multimodal embeddings](/azure/ai-services/computer-vision/concept-image-retrieval#what-are-vector-embeddings).|
53
+
|`modelVersion`| (Required) The model version (`2023-04-15`) to be passed to the Azure AI Vision multimodal embeddings API for generating embeddings. Vector embeddings can only be compared and matched if they're from the same model type. Images vectorized by one model won't be searchable through a different model. The latest Image Analysis API offers two models, version `2023-04-15` which supports text search in many languages, and the legacy `2022-04-11` model which supports only English. Azure AI Search uses the newer version. |
45
54
46
55
## Skill inputs
47
56
57
+
Skill definition inputs include name, source, and inputs. The following table provides valid values for name of the input. You can also specify recursive inputs. For more information, see the [REST API reference](/rest/api/searchservice/skillsets/create?view=rest-searchservice-2025-03-01-preview#inputfieldmappingentry&preserve-view=true) and [Create a skillset](cognitive-search-defining-skillset.md).
58
+
48
59
| Input | Description |
49
-
|--------------------|-------------|
60
+
|--------|-------------|
50
61
|`text`| The input text to be vectorized. If you're using data chunking, the source might be `/document/pages/*`. |
51
62
|`image`| Complex Type. Currently only works with "/document/normalized_images" field, produced by the Azure blob indexer when ```imageAction``` is set to a value other than ```none```. |
52
63
|`url`| The URL to download the image to be vectorized. |
@@ -62,15 +73,15 @@ Only one of `text`, `image` or `url`/`queryString` can be configured for a singl
62
73
63
74
## Sample definition
64
75
65
-
For text input, consider a record that has the following fields:
76
+
For text input, consider a blob that has the following content:
66
77
67
78
```json
68
79
{
69
-
"content": "Microsoft released Windows 10."
80
+
"content": "Forests, grasslands, deserts, and mountains are all part of the Patagonian landscape that spans more than a million square kilometers of South America."
70
81
}
71
82
```
72
83
73
-
Then your skill definition might look like this:
84
+
For text inputs, your skill definition might look like this:
74
85
75
86
```json
76
87
{
@@ -85,14 +96,15 @@ Then your skill definition might look like this:
85
96
],
86
97
"outputs": [
87
98
{
88
-
"name": "vector"
99
+
"name": "vector",
100
+
"targetName": "text_vector"
89
101
}
90
102
]
91
103
}
92
104
93
105
```
94
106
95
-
For image input, your skill definition might look like this:
107
+
For image input, a second skill definition in the same skillset might look like this:
96
108
97
109
```json
98
110
{
@@ -107,13 +119,14 @@ For image input, your skill definition might look like this:
107
119
],
108
120
"outputs": [
109
121
{
110
-
"name": "vector"
122
+
"name": "vector",
123
+
"targetName": "image_vector"
111
124
}
112
125
]
113
126
}
114
127
```
115
128
116
-
If you want to vectorize images directly from your blob storage datasource, your skill definition might look like this:
129
+
If you want to vectorize images directly from your blob storage data source rather than extract images during indexing, your skill definition should specify a URL, and perhaps a SAS token depending on storage security. For this scenario, your skill definition might look like this:
117
130
118
131
```json
119
132
{
@@ -132,19 +145,20 @@ If you want to vectorize images directly from your blob storage datasource, your
132
145
],
133
146
"outputs": [
134
147
{
135
-
"name": "vector"
148
+
"name": "vector",
149
+
"targetName": "image_vector"
136
150
}
137
151
]
138
152
}
139
153
```
140
154
141
155
## Sample output
142
156
143
-
For the given input text, a vectorized embedding output is produced.
157
+
For the given input, a vectorized embedding output is produced. Output is 1,024 dimensions, which is the number of dimensions supported by the Azure AI Vision multimodal API.
144
158
145
159
```json
146
160
{
147
-
"vector": [
161
+
"text_vector": [
148
162
0.018990106880664825,
149
163
-0.0073809814639389515,
150
164
....
@@ -153,7 +167,7 @@ For the given input text, a vectorized embedding output is produced.
153
167
}
154
168
```
155
169
156
-
The output resides in memory. To send this output to a field in the search index, you must define an [outputFieldMapping](cognitive-search-output-field-mapping.md) that maps the vectorized embedding output (which is an array) to a [vector field](vector-search-how-to-create-index.md). Assuming the skill output resides in the document's **vector** node, and **content_vector** is the field in the search index, the outputFieldMapping in indexer should look like:
170
+
The output resides in memory. To send this output to a field in the search index, you must define an [outputFieldMapping](cognitive-search-output-field-mapping.md) that maps the vectorized embedding output (which is an array) to a [vector field](vector-search-how-to-create-index.md). Assuming the skill output resides in the document's **vector** node, and **content_vector** is the field in the search index, the outputFieldMapping in the indexer should look like:
157
171
158
172
```json
159
173
"outputFieldMappings": [
@@ -164,7 +178,7 @@ The output resides in memory. To send this output to a field in the search index
164
178
]
165
179
```
166
180
167
-
For mapping image embeddings to the index, you'll need to use the [Index Projections](index-projections-concept-intro.md) feature. The payload for `indexProjections` might look something like this:
181
+
For mapping image embeddings to the index, youuse [index projections](index-projections-concept-intro.md). The payload for `indexProjections` might look something like the following example. image_content_vector is a field in the index, and it's populated with the content found in the **vector** of the **normalized_images** array.
168
182
169
183
```json
170
184
"indexProjections": {
@@ -175,7 +189,7 @@ For mapping image embeddings to the index, you'll need to use the [Index Project
Copy file name to clipboardExpand all lines: articles/search/search-region-support.md
+9-9Lines changed: 9 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -27,7 +27,7 @@ Some features take a dependency on other Azure services or infrastructure that a
27
27
|[Availability zones](search-reliability.md#availability-zone-support)| Divides a region's data centers into distinct physical location groups, providing high availability within the same geo. | Regional support is noted in this article. |
28
28
|[Semantic ranker](semantic-search-overview.md)| Takes a dependency on Microsoft-hosted models in specific regions. | Regional support is noted in this article. |
29
29
|[Query rewrite](semantic-how-to-query-rewrite.md)| Takes a dependency on Microsoft-hosted models in specific regions. | Regional support is noted in this article. |
30
-
|[AI service integration](cognitive-search-concept-intro.md)| Refers to [built-in skills](cognitive-search-predefined-skills.md) that make internal calls to Azure AI for enrichment and transformation during indexing. Integration requires that Azure AI Search coexists with an [Azure AI services multi-service account](/azure/ai-services/multi-service-resource#azure-ai-services-resource-for-azure-ai-search-skills) in the same physical region. You can bypass region requirements if you use [identity-based connections](cognitive-search-attach-cognitive-services.md#bill-through-a-keyless-connection), currently in public preview. | Regional support is noted in this article. |
30
+
|[AI enrichment](cognitive-search-concept-intro.md)| Refers to [built-in skills](cognitive-search-predefined-skills.md) that make internal calls to Azure AI for enrichment and transformation during indexing. Integration requires that Azure AI Search coexists with an [Azure AI services multi-service account](/azure/ai-services/multi-service-resource#azure-ai-services-resource-for-azure-ai-search-skills) in the same physical region. You can bypass region requirements if you use [identity-based connections](cognitive-search-attach-cognitive-services.md#bill-through-a-keyless-connection), currently in public preview. | Regional support is noted in this article. |
31
31
|[Azure OpenAI integration](vector-search-integrated-vectorization.md)| Refers to the AzureOpenAIEmbedding skill and vectorizer that make internal calls to deployed embedding models on Azure OpenAI. | Check [Azure OpenAI model region availability](/azure/ai-services/openai/concepts/models#model-summary-table-and-region-availability) for the most current list of regions for each embedding and chat model. Specific Azure OpenAI models are in fewer regions, so check for model availability first, and then verify Azure AI Search is available in the same region.|
32
32
|[Azure AI Foundry integration](vector-search-integrated-vectorization-ai-studio.md)| Refers to skills and vectorizers that make internal calls to the models hosted in the model catalog. | Check [Azure AI Foundry region availability](/azure/ai-foundry/reference/region-support) for the most current list of regions. |
33
33
|[Azure AI Vision 4.0 multimodal APIs](search-get-started-portal-image-search.md)| Refers to the Azure AI Vision multimodal embeddings skill and vectorizer that call the multimodal embedding API. | Check the [Azure AI Vision region list](/azure/ai-services/computer-vision/overview-image-analysis#region-availability) first, and then verify Azure AI Search is available in the same region.|
@@ -36,11 +36,11 @@ Some features take a dependency on other Azure services or infrastructure that a
36
36
37
37
You can create an Azure AI Search resource in any of the following Azure public regions. Almost all of these regions support [higher capacity tiers](search-limits-quotas-capacity.md#service-limits). Exceptions are noted where they apply.
38
38
39
-
AI service integration refers to internal connections to an Azure AI services multi-service account and doesn't include Azure OpenAI integration.
39
+
AI enrichment refers to internal connections to an Azure AI services multi-service account and doesn't include Azure OpenAI integration.
40
40
41
41
### Americas
42
42
43
-
| Region | AI service integration| Availability zones | Semantic ranker | Query rewrite |
43
+
| Region | AI enrichment| Availability zones | Semantic ranker | Query rewrite |
44
44
|--|--|--|--|--|
45
45
| Brazil South | ✅ || ✅ ||
46
46
| Canada Central | ✅ | ✅ | ✅ ||
@@ -58,7 +58,7 @@ AI service integration refers to internal connections to an Azure AI services mu
58
58
59
59
### Europe
60
60
61
-
| Region | AI service integration| Availability zones | Semantic ranker | Query rewrite |
61
+
| Region | AI enrichment| Availability zones | Semantic ranker | Query rewrite |
62
62
|--|--|--|--|--|
63
63
| North Europe <sup>1</sup> | ✅ | ✅ | ✅ | ✅ |
64
64
| West Europe | ✅ | ✅ | ✅ ||
@@ -80,7 +80,7 @@ AI service integration refers to internal connections to an Azure AI services mu
80
80
81
81
### Middle East
82
82
83
-
| Region | AI service integration| Availability zones | Semantic ranker | Query rewrite |
83
+
| Region | AI enrichment| Availability zones | Semantic ranker | Query rewrite |
84
84
|--|--|--|--|--|
85
85
| Israel Central <sup>1</sup> || ✅ |||
86
86
| Qatar Central <sup>1</sup> || ✅ |||
@@ -90,13 +90,13 @@ AI service integration refers to internal connections to an Azure AI services mu
90
90
91
91
### Africa
92
92
93
-
| Region | AI service integration| Availability zones | Semantic ranker | Query rewrite |
93
+
| Region | AI enrichment| Availability zones | Semantic ranker | Query rewrite |
94
94
|--|--|--|--|--|
95
95
| South Africa North | ✅ | ✅ |||
96
96
97
97
### Asia Pacific
98
98
99
-
| Region | AI service integration| Availability zones | Semantic ranker | Query rewrite |
99
+
| Region | AI enrichment| Availability zones | Semantic ranker | Query rewrite |
100
100
|--|--|--|--|--|
101
101
| Australia East | ✅ | ✅ | ✅ ||
102
102
| Australia Southeast ||| ✅ ||
@@ -115,15 +115,15 @@ AI service integration refers to internal connections to an Azure AI services mu
115
115
116
116
## Azure Government regions
117
117
118
-
| Region | AI service integration| Availability zones | Semantic ranker | Query rewrite |
118
+
| Region | AI enrichment| Availability zones | Semantic ranker | Query rewrite |
119
119
|--|--|--|--|--|
120
120
| Arizona | ✅ || ✅ ||
121
121
| Texas |||||
122
122
| Virginia | ✅ | ✅ | ✅ ||
123
123
124
124
## Azure operated by 21Vianet
125
125
126
-
| Region | AI service integration| Availability zones | Semantic ranker | Query rewrite |
126
+
| Region | AI enrichment| Availability zones | Semantic ranker | Query rewrite |
0 commit comments