Skip to content

Commit 1423f01

Browse files
authored
Merge pull request #7282 from HeidiSteen/heidist-0901
[release branch] GA updates for 2025-09-01
2 parents 0ea9cf6 + ea4e949 commit 1423f01

22 files changed

+167
-599
lines changed

articles/search/cognitive-search-skill-document-intelligence-layout.md

Lines changed: 15 additions & 22 deletions
Original file line numberDiff line numberDiff line change
@@ -9,14 +9,12 @@ ms.custom:
99
- references_regions
1010
- ignite-2024
1111
ms.topic: reference
12-
ms.date: 09/19/2025
12+
ms.date: 09/28/2025
1313
ms.update-cycle: 365-days
1414
---
1515

1616
# Document Layout skill
1717

18-
[!INCLUDE [Feature preview](./includes/previews/preview-generic.md)]
19-
2018
The **Document Layout** skill analyzes a document to detect structure and characteristics, and produces a syntactical representation of the document in Markdown or Text format. You can use it to extract text and images, where image extraction includes location metadata that preserves image position within the document. Image proximity to related content is beneficial in Retrieval Augmented Generation (RAG) workloads and [multimodal search](multimodal-search-overview.md) scenarios.
2119

2220
This article is the reference documentation for the Document Layout skill. For usage information, see [How to chunk and vectorize by document layout](search-how-to-semantic-chunking.md).
@@ -37,11 +35,12 @@ This skill is bound to a [billable Azure AI multi-service resource](cognitive-se
3735
This skill has the following limitations:
3836

3937
+ The skill isn't suitable for large documents requiring more than 5 minutes of processing in the AI Document Intelligence layout model. The skill times out, but charges still apply to the AI Services multi-services resource if it attaches to the skillset for billing purposes. Ensure documents are optimized to stay within processing limits to avoid unnecessary costs.
38+
4039
+ Since this skill calls the Azure AI Document Intelligence layout model, all documented [service behaviors for different document types](/azure/ai-services/document-intelligence/prebuilt/layout#pages) for different file types apply to its output. For example, Word (DOCX) and PDF files may produce different results due to differences in how images are handled. If consistent image behavior across DOCX and PDF is required, consider converting documents to PDF or reviewing the [multimodal search documentation](multimodal-search-overview.md) for alternative approaches.
4140

4241
## Supported regions
4342

44-
The Document Layout skill calls the [Document Intelligence Public preview version 2024-07-31-preview](/rest/api/aiservices/operation-groups?view=rest-aiservices-v4.0%20(2024-07-31-preview)&preserve-view=true).
43+
The Document Layout skill calls the [Document Intelligence 2024-11-30 API](/rest/api/aiservices/operation-groups).
4544

4645
Supported regions vary by modality and how the skill connects to the Document Intelligence layout model.
4746

@@ -70,12 +69,6 @@ This skill recognizes the following file formats.
7069

7170
Refer to [Azure AI Document Intelligence layout model supported languages](/azure/ai-services/document-intelligence/language-support/ocr?view=doc-intel-3.1.0&tabs=read-print%2Clayout-print%2Cgeneral#layout&preserve-view=true) for printed text.
7271

73-
## Supported parameters
74-
75-
Several parameters are version-specific. The skills parameter table notes the API version in which a parameter was introduced so that you know how to configure the skill. To use version-specific features such as image and location metadata extraction in [2025-05-01-preview REST API](/rest/api/searchservice/skillsets/create?view=rest-searchservice-2025-05-01-preview&preserve-view=true), you can use the Azure portal, or target 2025-05-01-preview, or check an Azure SDK change log to see if it supports the new parameters.
76-
77-
The Azure portal supports most preview features and can be used to create or update a skillset. For updates to the Document Layout skill, edit the skillset JSON definition to add new preview parameters.
78-
7972
## @odata.type
8073

8174
Microsoft.Skills.Util.DocumentIntelligenceLayoutSkill
@@ -89,21 +82,21 @@ Microsoft.Skills.Util.DocumentIntelligenceLayoutSkill
8982

9083
## Skill parameters
9184

92-
Parameters are case-sensitive.
85+
Parameters are case-sensitive. Several parameters were introduced in specific preview versions of the REST API. We recommend using the generally available version (2025-09-01) or the latest preview (2025-08-01-preview) for full access to all parameters.
9386

94-
| Parameter name | Version | Allowed Values | Description |
95-
|--------------------|-------------|-------------|-------------|
96-
| `outputMode` | [2024-11-01-preview](/rest/api/searchservice/skillsets/create-or-update?view=rest-searchservice-2024-11-01-preview&preserve-view=true) |`oneToMany` | Controls the cardinality of the output produced by the skill. |
97-
| `markdownHeaderDepth` | [2024-11-01-preview](/rest/api/searchservice/skillsets/create-or-update?view=rest-searchservice-2024-11-01-preview&preserve-view=true) |`h1`, `h2`, `h3`, `h4`, `h5`, `h6(default)` | Only applies if `outputFormat` is set to `markdown`. This parameter describes the deepest nesting level that should be considered. For instance, if the markdownHeaderDepth is `h3`, any sections that are deeper such as `h4`, are rolled into `h3`. |
98-
| `outputFormat` | [2025-05-01-preview](/rest/api/searchservice/skillsets/create-or-update?view=rest-searchservice-2025-05-01-preview&preserve-view=true) |`markdown(default)`, `text` | **New**. Controls the format of the output generated by the skill. |
99-
| `extractionOptions` | [2025-05-01-preview](/rest/api/searchservice/skillsets/create-or-update?view=rest-searchservice-2025-05-01-preview&preserve-view=true) |`["images"]`, `["images", "locationMetadata"]`, `["locationMetadata"]` | **New**. Identify any extra content extracted from the document. Define an array of enums that correspond to the content to be included in the output. For instance, if the `extractionOptions` is `["images", "locationMetadata"]`, the output includes images and location metadata which provides page location information related to where the content was extracted, such as a page number or section. This parameter applies to both output formats. |
100-
| `chunkingProperties` | [2025-05-01-preview](/rest/api/searchservice/skillsets/create-or-update?view=rest-searchservice-2025-05-01-preview&preserve-view=true) | See below. | **New**. Only applies if `outputFormat` is set to `text`. Options that encapsulate how to chunk text content while recomputing other metadata. |
87+
| Parameter name | Allowed Values | Description |
88+
|--------------------|----------------|-------------|
89+
| `outputMode` |`oneToMany` | Controls the cardinality of the output produced by the skill. |
90+
| `markdownHeaderDepth` |`h1`, `h2`, `h3`, `h4`, `h5`, `h6(default)` | Only applies if `outputFormat` is set to `markdown`. This parameter describes the deepest nesting level that should be considered. For instance, if the markdownHeaderDepth is `h3`, any sections that are deeper such as `h4`, are rolled into `h3`. |
91+
| `outputFormat` |`markdown(default)`, `text` | **New**. Controls the format of the output generated by the skill. |
92+
| `extractionOptions` |`["images"]`, `["images", "locationMetadata"]`, `["locationMetadata"]` | **New**. Identify any extra content extracted from the document. Define an array of enums that correspond to the content to be included in the output. For instance, if the `extractionOptions` is `["images", "locationMetadata"]`, the output includes images and location metadata which provides page location information related to where the content was extracted, such as a page number or section. This parameter applies to both output formats. |
93+
| `chunkingProperties` | See below. | **New**. Only applies if `outputFormat` is set to `text`. Options that encapsulate how to chunk text content while recomputing other metadata. |
10194

10295
| ChunkingProperties Parameter | Version | Allowed Values | Description |
10396
|--------------------|-------------|-------------|-------------|
104-
| `unit` | [2025-05-01-preview](/rest/api/searchservice/skillsets/create-or-update?view=rest-searchservice-2025-05-01-preview&preserve-view=true) | `Characters`. currently the only allowed value. Chunk length is measured in characters, as opposed to words or tokens | **New**. Controls the cardinality of the chunk unit. |
105-
| `maximumLength` | [2025-05-01-preview](/rest/api/searchservice/skillsets/create-or-update?view=rest-searchservice-2025-05-01-preview&preserve-view=true) | Any integer between 300-50000 | **New**. The maximum chunk length in characters as measured by String.Length. |
106-
| `overlapLength` | [2025-05-01-preview](/rest/api/searchservice/skillsets/create-or-update?view=rest-searchservice-2025-05-01-preview&preserve-view=true) | Integer. The value needs to be less than the half of the `maximumLength` | **New**. The length of overlap provided between two text chunks. |
97+
| `unit` | `Characters`. currently the only allowed value. Chunk length is measured in characters, as opposed to words or tokens | **New**. Controls the cardinality of the chunk unit. |
98+
| `maximumLength` | Any integer between 300-50000 | **New**. The maximum chunk length in characters as measured by String.Length. |
99+
| `overlapLength` | Integer. The value needs to be less than the half of the `maximumLength` | **New**. The length of overlap provided between two text chunks. |
107100

108101
## Skill inputs
109102

@@ -203,7 +196,7 @@ The value of the `markdownHeaderDepth` controls the number of keys in the "secti
203196

204197
## Example for text output mode and image and metadata extraction
205198

206-
This example demonstrates how to use the new parameters introduced in the **2025-05-01-preview** to output text content in fixed-sized chunks and extract images along with location metadata from the document.
199+
This example demonstrates how to output text content in fixed-sized chunks and extract images along with location metadata from the document.
207200

208201
### Sample definition for text output mode and image and metadata extraction
209202

articles/search/cognitive-search-skill-genai-prompt.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ The GenAI Prompt skill is available in the [latest preview REST API](/rest/api/s
3939

4040
- For image verbalization, the model you use to analyze the image determines what image formats are supported.
4141

42-
- For GPT-5 model, the `temperature` parameter is not supported in the same way as previous models. If defined, it must be set to `1.0`, as other values will result in errors.
42+
- For GPT-5 models, the `temperature` parameter is not supported in the same way as previous models. If defined, it must be set to `1.0`, as other values will result in errors.
4343

4444
- Billing is based on the pricing of the model you use.
4545

articles/search/hybrid-search-ranking.md

Lines changed: 8 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ ms.service: azure-ai-search
88
ms.custom:
99
- ignite-2023
1010
ms.topic: conceptual
11-
ms.date: 08/27/2025
11+
ms.date: 09/28/2025
1212
---
1313

1414
# Relevance scoring in hybrid search using Reciprocal Rank Fusion (RRF)
@@ -17,9 +17,6 @@ Reciprocal Rank Fusion (RRF) is an algorithm that evaluates the search scores fr
1717

1818
RRF is based on the concept of *reciprocal rank*, which is the inverse of the rank of the first relevant document in a list of search results. The goal of the technique is to take into account the position of the items in the original rankings, and give higher importance to items that are ranked higher in multiple lists. This can help improve the overall quality and reliability of the final ranking, making it more useful for the task of fusing multiple ordered search results.
1919

20-
> [!NOTE]
21-
> The [latest preview REST API](/rest/api/searchservice/documents/search-post?view=rest-searchservice-2025-08-01-preview&preserve-view=true) can deconstruct an RRF-ranked search score into its component subscores. This gives you transparency into all-up score composition. For more information, see [Unpack search scores (preview)](#unpack-a-search-score-into-subscores-preview) in this article.
22-
2320
## How RRF ranking works
2421

2522
RRF works by taking the search results from multiple methods, assigning a reciprocal rank score to each document in the results, and then combining the scores to create a new ranking. The concept is that documents appearing in the top positions across multiple search methods are likely to be more relevant and should be ranked higher in the combined result.
@@ -59,22 +56,20 @@ The following chart identifies the scoring property returned on each match, algo
5956

6057
Semantic ranking occurs after RRF merging of results. Its score (`@search.rerankerScore`) is always reported separately in the query response. Semantic ranker can rerank full text and hybrid search results, assuming those results include fields having semantically rich content. It can rerank pure vector queries if the search documents include text fields that contain semantically relevant content.
6158

62-
## Unpack a search score into subscores (preview)
63-
64-
Using the [latest preview REST API](/rest/api/searchservice/documents/search-post?view=rest-searchservice-2025-08-01-preview&preserve-view=true), you can deconstruct a search score to view its subscores.
59+
## Unpack a search score into subscores
6560

66-
For vector queries, this information can help you determine an appropriate value for [vector weighting](vector-search-how-to-query.md#vector-weighting) or [setting minimum thresholds](vector-search-how-to-query.md#set-thresholds-to-exclude-low-scoring-results-preview).
61+
You can deconstruct a search score to view its subscores. For vector queries, this information can help you determine an appropriate value for [vector weighting](vector-search-how-to-query.md#vector-weighting) or [setting minimum thresholds](vector-search-how-to-query.md#set-thresholds-to-exclude-low-scoring-results-preview).
6762

6863
To get subscores:
6964

70-
+ Use the [latest preview Search Documents REST API](/rest/api/searchservice/documents/search-post?view=rest-searchservice-2025-08-01-preview&preserve-view=true#request-body) or an Azure SDK beta package that provides the feature.
65+
+ Use the [Search Documents REST API](/rest/api/searchservice/documents/search-post#request-body) or an Azure SDK package that provides the feature.
7166

7267
+ Modify a query request, adding a new `debug` parameter set to either `vector`, `semantic` if using semantic ranker, or `all`.
7368

7469
Here's an example of hybrid query that returns subscores in debug mode:
7570

7671
```http
77-
POST https://{{search-service-name}}.search.windows.net/indexes/{{index-name}}/docs/search?api-version=2025-08-01-preview
72+
POST https://{{search-service-name}}.search.windows.net/indexes/{{index-name}}/docs/search?api-version=2025-09-01
7873
7974
{
8075
"vectorQueries": [
@@ -114,7 +109,7 @@ POST https://{{search-service-name}}.search.windows.net/indexes/{{index-name}}/d
114109

115110
## Weighted scores
116111

117-
Using the [stable REST API version](/rest/api/searchservice/documents/search-post) and newer preview API versions, you can [weight vector queries](vector-search-how-to-query.md#vector-weighting) to increase or decrease their importance in a hybrid query.
112+
You can also [weight vector queries](vector-search-how-to-query.md#vector-weighting) to increase or decrease their importance in a hybrid query.
118113

119114
Recall that when computing RRF for a certain document, the search engine looks at the rank of that document for each result set where it shows up. Assume a document shows up in three separate search results, where the results are from two vector queries and one text BM25-ranked query. The position of the document varies in each result.
120115

@@ -142,5 +137,5 @@ For more information, see [How to work with search results](search-pagination-pa
142137

143138
## See also
144139

145-
+ [Learn more about hybrid search](hybrid-search-overview.md)
146-
+ [Learn more about vector search](vector-search-overview.md)
140+
+ [Hybrid search](hybrid-search-overview.md)
141+
+ [Vector search](vector-search-overview.md)

articles/search/search-agentic-retrieval-how-to-create.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,8 @@ Use Azure OpenAI or an equivalent open source model:
6565
+ `gpt-4.1-nano`
6666
+ `gpt-4.1-mini`
6767
+ `gpt-5`
68+
+ `gpt-5-nano`
69+
+ `gpt-5-mini`
6870

6971
## Configure access
7072

articles/search/search-agentic-retrieval-how-to-index.md

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -47,6 +47,7 @@ Here's an example index that works for agentic retrieval. It meets the criteria
4747
```json
4848
{
4949
"name": "earth_at_night",
50+
"description": "Contains images an descriptions of our planet in darkness as captured from space by Earth-observing satellites and astronauts on the International Space Station over the past 25 years.",
5051
"fields": [
5152
{
5253
"name": "id", "type": "Edm.String",
@@ -166,6 +167,15 @@ All `searchable` fields are included in query execution. There's no support for
166167
> + Fields selected in the response string are semantic fields (title, terms, content)
167168
> + Fields in reference source data are all `retrievable` fields, assuming reference source data is true -->
168169

170+
## Add a description
171+
172+
An index `description` field is exposed programmatically, which means you can pass this description to LLMs and Model Context Protocol (MCP) servers as an input when deciding to use a specific index for a query. This human-readable text is invaluable when a system must access several indexes and make a decision based on the description.
173+
174+
An index description is a schema update, and you can add it without having to rebuild the entire index.
175+
176+
+ String length is 4,000 characters maximum.
177+
+ Content must be human-readable, in Unicode. Your use case should determine which language to use (for example, English or another language).
178+
169179
## Add a semantic configuration
170180

171181
The index must have at least one semantic configuration. The semantic configuration must have:

articles/search/search-agentic-retrieval-how-to-pipeline.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -48,6 +48,8 @@ Use one of the following chat completion models with your AI agent:
4848
+ `gpt-4.1-nano`
4949
+ `gpt-4.1-mini`
5050
+ `gpt-5`
51+
+ `gpt-5-nano`
52+
+ `gpt-5-mini`
5153

5254
### Package version requirements
5355

articles/search/search-api-migration.md

Lines changed: 8 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ ms.custom:
1111
- build-2024
1212
- ignite-2024
1313
ms.topic: conceptual
14-
ms.date: 08/27/2025
14+
ms.date: 09/27/2025
1515
---
1616

1717
# Upgrade to the latest REST API in Azure AI Search
@@ -22,7 +22,7 @@ Here are the most recent versions of the REST APIs:
2222

2323
| Targeted operations | REST API | Status |
2424
|---------------------|----------|--------|
25-
| Data plane | [`2024-07-01`](/rest/api/searchservice/search-service-api-versions#2024-07-01) | Stable |
25+
| Data plane | [`2025-09-01`](/rest/api/searchservice/search-service-api-versions#2025-09-01) | Stable |
2626
| Data plane | [`2025-08-01-preview`](/rest/api/searchservice/search-service-api-versions#2025-08-01-preview&preserve-view=true) | Preview |
2727
| Control plane | [`2025-05-01`](/rest/api/searchmanagement/operation-groups?view=rest-searchmanagement-2025-05-01&preserve-view=true) | Stable |
2828
| Control plane | [`2025-02-01-preview`](/rest/api/searchmanagement/operation-groups?view=rest-searchmanagement-2025-02-01-preview&preserve-view=true) | Preview |
@@ -90,6 +90,12 @@ See [Migrate from preview version](semantic-code-migration.md) to transition you
9090

9191
Upgrade guidance assumes upgrade from the most recent previous version. If your code is based on an old API version, we recommend upgrading through each successive version to get to the newest version.
9292

93+
### Upgrade to 2025-09-01
94+
95+
[`2025-09-01`](/rest/api/searchservice/search-service-api-versions#2025-09-01) is the latest stable REST API version and it adds general availability for the OneLake indexer, Document Layout skill, and other APIs.
96+
97+
There are no breaking changes if you're upgrading from `2024-07-01` and not using any preview features. To use the new stable release, change the API version and test your code.
98+
9399
### Upgrade to 2025-08-01-preview
94100

95101
[`2025-08-01-preview`](/rest/api/searchservice/search-service-api-versions#2025-08-01-preview) introduces the following breaking changes to knowledge agents created using `2025-05-01-preview`:

0 commit comments

Comments
 (0)