You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/search/search-limits-quotas-capacity.md
+15-12Lines changed: 15 additions & 12 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ author: HeidiSteen
8
8
ms.author: heidist
9
9
ms.service: cognitive-search
10
10
ms.topic: conceptual
11
-
ms.date: 11/16/2023
11
+
ms.date: 01/02/2024
12
12
ms.custom:
13
13
- references_regions
14
14
- ignite-2023
@@ -18,7 +18,7 @@ ms.custom:
18
18
19
19
Maximum limits on storage, workloads, and quantities of indexes and other objects depend on whether you [provision Azure AI Search](search-create-service-portal.md) at **Free**, **Basic**, **Standard**, or **Storage Optimized** pricing tiers.
20
20
21
-
+**Free** is a multi-tenant shared service that comes with your Azure subscription.
21
+
+**Free** is a multitenant shared service that comes with your Azure subscription.
22
22
23
23
+**Basic** provides dedicated computing resources for production workloads at a smaller scale, but shares some networking infrastructure with other tenants.
24
24
@@ -29,7 +29,7 @@ Maximum limits on storage, workloads, and quantities of indexes and other object
@@ -59,7 +59,7 @@ You might find some variation in maximum limits if your service happens to be pr
59
59
60
60
## Document limits
61
61
62
-
There are no longer any document limits per service in Azure AI Search, however, there's a limit of approximately 24 billion documents per index on Basic, S1, S2, S3, L1, and L2 search services. For S3 HD, the limit is 2 billion documents per index. Each element of a complex collection counts as a separate document in terms of these limits.
62
+
You can have approximately 24 billion documents per index on Basic, S1, S2, S3, L1, and L2 search services. For S3 HD, the limit is 2 billion documents per index. Each instance of a complex collection counts as a separate document in terms of these limits.
63
63
64
64
### Document size limits per API call
65
65
@@ -73,11 +73,11 @@ When estimating document size, remember to consider only those fields that can b
73
73
74
74
When you index documents with vector fields, Azure AI Search constructs internal vector indexes using the algorithm parameters you provide. The size of these vector indexes is restricted by the memory reserved for vector search for your service's tier (or SKU).
75
75
76
-
The service enforces a vector index size quota **for every partition** in your search service. Each extra partition increases the available vector index size quota. This quota is a hard limit to ensure your service remains healthy, which means that further indexing attempts once the limit is exceeded results in failure. You may resume indexing once you free up available quota by either deleting some vector documents or by scaling up in partitions.
76
+
The service enforces a vector index size quota **for every partition** in your search service. Each extra partition increases the available vector index size quota. This quota is a hard limit to ensure your service remains healthy, which means that further indexing attempts once the limit is exceeded results in failure. You can resume indexing once you free up available quota by either deleting some vector documents or by scaling up in partitions.
77
77
78
78
The table describes the vector index size quota per partition across the service tiers (or SKU). For context, it includes:
79
79
80
-
+[Storage limits](#storage-limits) for each tier, repeated here for context.
80
+
+[Partition storage limits](#service-limits) for each tier, repeated here for context.
81
81
+ Amount of each partition (in GB) available for vector indexes (created when you add vector fields to an index).
82
82
+ Approximate number of embeddings (floating point values) per partition.
83
83
@@ -125,8 +125,8 @@ Maximum running times exist to provide balance and stability to the service as a
125
125
| Maximum skillsets <sup>4</sup> |3 |5 or 15 |50 |200 |200 |N/A |10 |10 |
126
126
| Maximum indexing load per invocation |10,000 documents |Limited only by maximum documents |Limited only by maximum documents |Limited only by maximum documents |Limited only by maximum documents |N/A |No limit |No limit |
| Blob indexer: maximum characters of content extracted from a blob |32,000 |64,000 |4 million |8 million |16 million |N/A |4 million |4 million |
132
132
@@ -138,9 +138,9 @@ Maximum running times exist to provide balance and stability to the service as a
138
138
139
139
<sup>4</sup> Maximum of 30 skills per skillset.
140
140
141
-
<sup>5</sup> AI enrichment and image analysis are computationally intensive and consume disproportionate amounts of available processing power. Running time for these workloads has been shortened to give other jobs in the queue more opportunity to run.
141
+
<sup>5</sup> Regarding the 2 or 24 hour maximum duration for indexers: a 2-hour maximum is the most common and it's what you should plan for. The 24-hour limit is from an older indexer implementation. If you have unscheduled indexers that run continuously for 24 hours, it's because those indexers couldn't be migrated to the newer runtime behavior. For extra large data sets, indexers can be made to run longer than maximum limits if you put them on a [2-hour run time schedule](search-howto-schedule-indexers.md). When the first 2-hour interval is complete, the indexer picks up where it left off to start the next 2-hour interval.
142
142
143
-
<sup>6</sup> Indexer execution and combined indexer-skillset execution is subject to a 2-hour maximum duration. Currently, some indexers have a longer 24-hour maximum execution window, but that behavior isn't the norm. The longer window only applies if a service or its indexers can't be internally migrated to the newer runtime behavior. If more than 2 hours are needed to complete an indexer or indexer-skillset process, [schedule the indexer](search-howto-schedule-indexers.md) to run at 2-hour intervals.
143
+
<sup>6</sup> Skillset execution, and image analysis in particular, are computationally intensive and consume disproportionate amounts of available processing power. Running time for these workloads has been shortened to give other jobs in the queue more opportunity to run.
144
144
145
145
> [!NOTE]
146
146
> As stated in the [Index limits](#index-limits), indexers will also enforce the upper limit of 3000 elements across all complex collections per document starting with the latest GA API version that supports complex types (`2019-05-06`) onwards. This means that if you've created your indexer with a prior API version, you will not be subject to this limit. To preserve maximum compatibility, an indexer that was created with a prior API version and then updated with an API version `2019-05-06` or later, will still be **excluded** from the limits. Customers should be aware of the adverse impact of having very large complex collections (as stated previously) and we highly recommend creating any new indexers with the latest GA API version.
@@ -156,7 +156,7 @@ Indexers can access other Azure resources [over private endpoints](search-indexe
<sup>1</sup> AI enrichment and image analysis are computationally intensive and consume disproportionate amounts of available processing power. For this reason, private connections are disabled on lower tiers to avoid an adverse impact on the performance and stability of the search service itself.
159
+
<sup>1</sup> AI enrichment and image analysis are computationally intensive and consume disproportionate amounts of available processing power. For this reason, private connections are disabled on lower tiers to ensure the performance and stability of the search service itself.
160
160
161
161
<sup>2</sup> The number of distinct resource types are computed as the number of unique `groupId` values used across all shared private link resources for a given search service, irrespective of the status of the resource.
162
162
@@ -179,7 +179,7 @@ Maximum number of [index aliases](search-how-to-alias.md) varies by tier. In all
179
179
180
180
## Data limits (AI enrichment)
181
181
182
-
An [AI enrichment pipeline](cognitive-search-concept-intro.md) that makes calls to Azure AI Language resource for [entity recognition](cognitive-search-skill-entity-recognition-v3.md), [entity linking](cognitive-search-skill-entity-linking-v3.md), [key phrase extraction](cognitive-search-skill-keyphrases.md), [sentiment analysis](cognitive-search-skill-sentiment-v3.md), [language detection](cognitive-search-skill-language-detection.md), and [personal-information detection](cognitive-search-skill-pii-detection.md) is subject to data limits. The maximum size of a record should be 50,000 characters as measured by [`String.Length`](/dotnet/api/system.string.length). If you need to break up your data before sending it to the sentiment analyzer, use the [Text Split skill](cognitive-search-skill-textsplit.md).
182
+
An [AI enrichment pipeline](cognitive-search-concept-intro.md) that makes calls to an Azure AI Language resource for [entity recognition](cognitive-search-skill-entity-recognition-v3.md), [entity linking](cognitive-search-skill-entity-linking-v3.md), [key phrase extraction](cognitive-search-skill-keyphrases.md), [sentiment analysis](cognitive-search-skill-sentiment-v3.md), [language detection](cognitive-search-skill-language-detection.md), and [personal-information detection](cognitive-search-skill-pii-detection.md) is subject to data limits. The maximum size of a record should be 50,000 characters as measured by [`String.Length`](/dotnet/api/system.string.length). If you need to break up your data before sending it to the sentiment analyzer, use the [Text Split skill](cognitive-search-skill-textsplit.md).
183
183
184
184
## Throttling limits
185
185
@@ -198,6 +198,7 @@ Static rate request limits for operations related to a service:
198
198
+ Service Statistics (GET /servicestats): 4 per second per search unit
199
199
200
200
## API request limits
201
+
201
202
* Maximum of 16 MB per request <sup>1</sup>
202
203
* Maximum 8 KB URL length
203
204
* Maximum 1000 documents per batch of index uploads, merges, or deletes
@@ -213,10 +214,12 @@ Static rate request limits for operations related to a service:
213
214
Limits on query size and composition exist because unbounded queries can destabilize your search service. Typically, such queries are created programmatically. If your application generates search queries programmatically, we recommend designing it in such a way that it doesn't generate queries of unbounded size.
214
215
215
216
## API response limits
217
+
216
218
* Maximum 1000 documents returned per page of search results
217
219
* Maximum 100 suggestions returned per Suggest API request
218
220
219
221
## API key limits
222
+
220
223
API keys are used for service authentication. There are two types. Admin keys are specified in the request header and grant full read-write access to the service. Query keys are read-only, specified on the URL, and typically distributed to client applications.
Copy file name to clipboardExpand all lines: articles/search/vector-search-index-size.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -114,7 +114,7 @@ The storage size of one vector is determined by its dimensionality. Multiply the
114
114
115
115
For `Edm.Single`, the size of the data type is 4 bytes.
116
116
117
-
### Memory Overhead from the Selected Algorithm
117
+
### Memory overhead from the selected algorithm
118
118
119
119
Every approximate nearest neighbor (ANN) algorithm generates extra data structures in memory to enable efficient searching. These structures consume extra space within memory.
120
120
@@ -161,4 +161,4 @@ Disk storage overhead of vector data is roughly three times the size of vector i
161
161
162
162
### Storage vs. vector index size quotas
163
163
164
-
The storage and vector index size quotas aren't separate quotas. Vector indexes contribute to the [storage quota for the search service](search-limits-quotas-capacity.md#storage-limits) as a whole. For example, if your storage quota is exhausted but there's remaining vector quota, you can't index any more documents, regardless if they're vector documents, until you scale up in partitions to increase storage quota or delete documents (either text or vector) to reduce storage usage. Similarly, if vector quota is exhausted but there's remaining storage quota, further indexing attempts fail until vector quota is freed, either by deleting some vector documents or by scaling up in partitions.
164
+
The storage and vector index size quotas aren't separate quotas. Vector indexes contribute to the [storage quota for the search service](search-limits-quotas-capacity.md#service-limits) as a whole. For example, if your storage quota is exhausted but there's remaining vector quota, you can't index any more documents, regardless if they're vector documents, until you scale up in partitions to increase storage quota or delete documents (either text or vector) to reduce storage usage. Similarly, if vector quota is exhausted but there's remaining storage quota, further indexing attempts fail until vector quota is freed, either by deleting some vector documents or by scaling up in partitions.
Copy file name to clipboardExpand all lines: includes/azure-search-limits-per-service.md
+11-10Lines changed: 11 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,21 +5,22 @@
5
5
author: HeidiSteen
6
6
ms.service: cognitive-search
7
7
ms.topic: include
8
-
ms.date: 05/11/2020
8
+
ms.date: 01/02/2024
9
9
ms.author: heidist
10
10
ms.custom: include file
11
11
---
12
12
13
-
A search service is constrained by disk space or by a hard limit on the maximum number of indexes or indexers, whichever comes first. The following table documents storage limits. For maximum object limits, see [Limits by resource](../articles/search/search-limits-quotas-capacity.md#index-limits).
13
+
A search service is subject to a maximum storage limit (partition size multiplied by the number of partitions) or by a hard limit on the [maximum number of indexes](../articles/search/search-limits-quotas-capacity.md#index-limits) or [indexers](../articles/search/search-limits-quotas-capacity.md#indexer-limits), whichever comes first.
<sup>1</sup> Basic has one fixed partition. Additional search units can be used to add replicas for larger query volumes.
22
+
<sup>1</sup> Basic has one fixed partition. You can specify up to 3 more search units to add replicas for larger query volumes and high availability.
24
23
25
-
<sup>2</sup> Service level agreements are in effect for billable services on dedicated resources. Free services and preview features have no SLA. For billable services, SLAs take effect when you provision sufficient redundancy for your service. Two or more replicas are required for query (read) SLAs. Three or more replicas are required for query and indexing (read-write) SLAs. The number of partitions isn't an SLA consideration.
24
+
<sup>2</sup> Service level agreements apply to billable services having dedicated resources. Free services and preview features have no SLA. For billable services, SLAs take effect when you provision sufficient redundancy for your service. Two or more replicas are required for query (read) SLAs. Three or more replicas are required for query and indexing (read-write) SLAs. The number of partitions isn't an SLA consideration. See [Reliability in Azure AI Search](/azure/search/search-reliability#high-availability) to learn more about replicas and high availability.
25
+
26
+
<sup>3</sup> Free services don't have a dedicated partition. The 50-MB storage limit refers to the maximum space allocated to a free search service on infrastructure shared with other customers.
Copy file name to clipboardExpand all lines: includes/azure-search-limits-per-subscription.md
+7-7Lines changed: 7 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,22 +4,22 @@
4
4
author: HeidiSteen
5
5
ms.service: cognitive-search
6
6
ms.topic: include
7
-
ms.date: 07/17/2023
7
+
ms.date: 01/02/2024
8
8
ms.author: heidist
9
9
ms.custom:
10
10
- include file
11
11
- ignite-2023
12
12
---
13
13
14
-
You can create multiple *billable* search services (Basic and higher), limited only by the number of services allowed at each tier. For example, you could create up to 16 services at the Basic tier and another 16 services at the S1 tier within the same subscription. For more information about tiers, see [Choose an SKU or tier for Azure AI Search](../articles/search/search-sku-tier.md).
14
+
You can create multiple *billable* search services (Basic and higher), up to the maximum number of services allowed at each tier. For example, you could create up to 16 services at the Basic tier and another 16 services at the S1 tier within the same subscription. For more information about tiers, see [Choose a tier (or SKU) for Azure AI Search](../articles/search/search-sku-tier.md).
15
15
16
16
Maximum service limits can be raised upon request. If you need more services within the same subscription, [file a support request](../articles/search/search-create-service-portal.md#add-more-services-to-a-subscription).
| Maximum scale in search units (SU)<sup>2</sup> |N/A |3 SU |36 SU |36 SU |36 SU |36 SU |36 SU |36 SU |
21
+
| Maximum search units (SU)<sup>2</sup> |N/A |3 SU |36 SU |36 SU |36 SU |36 SU |36 SU |36 SU |
22
22
23
-
<sup>1</sup> You can have one free search service per Azure subscription. The free tier is based on infrastructure that's shared with other customers. Because the hardware isn't dedicated, scale-up isn't supported, and storage is limited to 50 MB.
23
+
<sup>1</sup> You can have one free search service per Azure subscription. The free tier is based on infrastructure shared with other customers. Because the hardware isn't dedicated, scale-up isn't supported, and storage is limited to 50 MB.
24
24
25
-
<sup>2</sup> Search units are billing units, allocated as either a *replica* or a *partition*. You need both resources for storage, indexing, and query operations. To learn more about SU computations, see [Scale resource levels for query and index workloads](../articles/search/search-capacity-planning.md).
25
+
<sup>2</sup> Search units (SU) are billing units, allocated as either a *replica* or a *partition*. You need both. To learn more about SU combinations, see [Estimate and manage capacity of a search service](../articles/search/search-capacity-planning.md).
0 commit comments