Skip to content

Commit 38d55ba

Browse files
Merge pull request #270470 from gmndrg/main
Best practices for integration with Azure OpenAI embedding models
2 parents c6ae0f6 + ccadbed commit 38d55ba

File tree

2 files changed

+15
-2
lines changed

2 files changed

+15
-2
lines changed

articles/search/cognitive-search-skill-azure-openai-embedding.md

Lines changed: 10 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ ms.service: cognitive-search
88
ms.custom:
99
- ignite-2023
1010
ms.topic: reference
11-
ms.date: 02/21/2024
11+
ms.date: 03/28/2024
1212
---
1313

1414
# Azure OpenAI Embedding skill
@@ -113,6 +113,15 @@ The output resides in memory. To send this output to a field in the search index
113113
]
114114
```
115115

116+
## Best practices
117+
118+
The following are some best practices you need to consider when utilizing this skill:
119+
- If you are hitting your Azure OpenAI TPM (Tokens per minute) limit, consider the [quota limits advisory](../ai-services/openai/quotas-limits.md) so you can address accordingly. Refer to the [Azure OpenAI monitoring](../ai-services/openai/how-to/monitoring.md) documentation for more information about your Azure OpenAI instance performance.
120+
- The Azure OpenAI embeddings model deployment you use for this skill should be ideally separate from the deployment used for other use cases, including the [query vectorizer](vector-search-how-to-configure-vectorizer.md). This helps each deployment to be tailored to its specific use case, leading to optimized performance and identifying traffic from the indexer and the index embedding calls easily.
121+
- Your Azure OpenAI instance should be in the same region or at least geographically close to the region where your AI Search service is hosted. This reduces latency and improves the speed of data transfer between the services.
122+
- If you have a larger than default Azure OpenAI TPM (Tokens per minute) limit as published in [quotas and limits](../ai-services/openai/quotas-limits.md) documentation, open a [support case](../azure-portal/supportability/how-to-create-azure-support-request.md) with the Azure AI Search team, so this can be adjusted accordingly. This helps your indexing process not being unnecessarily slowed down by the documented default TPM limit, if you have higher limits.
123+
124+
116125
## Errors and warnings
117126

118127
| Condition | Result |

articles/search/vector-search-how-to-configure-vectorizer.md

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ ms.service: cognitive-search
99
ms.custom:
1010
- ignite-2023
1111
ms.topic: how-to
12-
ms.date: 03/27/2024
12+
ms.date: 03/28/2024
1313
---
1414

1515
# Configure a vectorizer in a search index
@@ -222,6 +222,10 @@ OperationEvent
222222
| where Name == "Query.Search" and AdditionalInfo["QueryMetadata"]["Vectors"] has "TextLength"
223223
```
224224

225+
## Best practices
226+
227+
If you are setting up an Azure OpenAI vectorizer, consider the same [best practices](cognitive-search-skill-azure-openai-embedding.md#best-practices) that we recommend for the Azure OpenAI embedding skill.
228+
225229
## See also
226230

227231
+ [Integrated vectorization (preview)](vector-search-integrated-vectorization.md)

0 commit comments

Comments
 (0)