Skip to content

Commit 24f9c9b

Browse files
Merge pull request #6225 from MicrosoftDocs/main
Auto Publish – main to live - 2025-07-25 17:05 UTC
2 parents ca98a5c + 63a9b26 commit 24f9c9b

File tree

7 files changed

+22
-13
lines changed

7 files changed

+22
-13
lines changed

articles/ai-foundry/agents/concepts/threads-runs-messages.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -34,11 +34,11 @@ A custom AI that uses AI models in conjunction with tools.
3434

3535
## Threads
3636

37-
Threads are conversation sessions between an agent and a user. They store messages and automatically handle truncation to fit content into a model’s context. When you create a thread, you can append new messages to it as users respond.
37+
Threads are conversation sessions between an agent and a user. They store messages and automatically handle truncation to fit content into a model’s context. When you create a thread, you can append new messages (maximum of 100,000 per thread) to it as users respond.
3838

3939
## Messages
4040

41-
Messages are the individual pieces of communication within a thread. They can be created by either the agent or the user and can include text, or other files. Messages are stored as a list within the thread, allowing for a structured and organized conversation flow.
41+
Messages are the individual pieces of communication within a thread. They can be created by either the agent or the user and can include text, or other files. Messages are stored as a list within the thread, allowing for a structured and organized conversation flow. You can attach up to 100,000 messages to a single thread.
4242

4343
## Runs
4444

articles/ai-foundry/agents/quotas-limits.md

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -21,10 +21,13 @@ The following sections provide you with a guide to the default quotas and limits
2121

2222
| Limit Name | Limit Value |
2323
|--|--|
24-
| Max files per agent/thread | 10,000 |
25-
| Max file size for agents & fine-tuning | 512 MB |
26-
| Max size for all uploaded files for agents |200 GB |
27-
| agents token limit | 2,000,000 token limit |
24+
| Maximum number of files per agent/thread | 10,000 |
25+
| Maximum file size for agents & fine-tuning | 512 MB |
26+
| Maximum size for all uploaded files for agents | 300 GB |
27+
| Maximum file size in tokens for attaching to a vector store | 2,000,000 tokens |
28+
| Maximum number of messages per thread | 100,000 |
29+
| Maximum size of `text` content per message | 1,500,000 characters |
30+
| Maximum number of tools registered per agent | 128 |
2831

2932
## Quotas and limits for Azure OpenAI models
3033

articles/ai-foundry/openai/how-to/fine-tuning-deploy.md

Lines changed: 8 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ manager: nitinme
66
ms.service: azure-ai-openai
77
ms.custom: build-2023, build-2023-dataai, devx-track-python, references_regions
88
ms.topic: how-to
9-
ms.date: 07/02/2025
9+
ms.date: 07/25/2025
1010
author: mrbullwinkle
1111
ms.author: mbullwin
1212
---
@@ -272,7 +272,7 @@ curl -X PUT "https://management.azure.com/subscriptions/<SUBSCRIPTION>/resourceG
272272
}'
273273
```
274274

275-
### Deploy a model with Azure CLI
275+
## [CLI](#tab/cli)
276276

277277
The following example shows how to use the Azure CLI to deploy your customized model. With the Azure CLI, you must specify a name for the deployment of your customized model. For more information about how to use the Azure CLI to deploy customized models, see [`az cognitiveservices account deployment`](/cli/azure/cognitiveservices/account/deployment).
278278

@@ -297,6 +297,7 @@ az cognitiveservices account deployment create
297297
--sku-capacity "1"
298298
--sku-name "Standard"
299299
```
300+
300301
---
301302

302303
[!INCLUDE [Fine-tuning deletion](../includes/fine-tune.md)]
@@ -343,6 +344,11 @@ curl $AZURE_OPENAI_ENDPOINT/openai/deployments/<deployment_name>/chat/completion
343344
-H "api-key: $AZURE_OPENAI_API_KEY" \
344345
-d '{"messages":[{"role": "system", "content": "You are a helpful assistant."},{"role": "user", "content": "Does Azure OpenAI support customer managed keys?"},{"role": "assistant", "content": "Yes, customer managed keys are supported by Azure OpenAI."},{"role": "user", "content": "Do other Azure services support this too?"}]}'
345346
```
347+
348+
## [CLI](#tab/cli)
349+
350+
Azure CLI is only for control plane operations such as resource creation and [model deployment](/cli/azure/cognitiveservices/account/deployment). For inference operations, use the [REST API](/azure/ai-foundry/openai/reference-preview-latest), or the [language based SDKs](../supported-languages.md).
351+
346352
---
347353

348354
### Prompt caching

articles/ai-services/speech-service/quickstarts/setup-platform.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ author: eric-urban
66
manager: nitinme
77
ms.service: azure-ai-speech
88
ms.topic: quickstart
9-
ms.date: 7/12/2025
9+
ms.date: 7/25/2025
1010
ms.author: eur
1111
ms.custom: devx-track-python, devx-track-js, devx-track-csharp, mode-other, devx-track-dotnet, devx-track-extended-java, devx-track-go, ignite-2023, linux-related-content
1212
zone_pivot_groups: programming-languages-ai-services
@@ -49,7 +49,7 @@ zone_pivot_groups: programming-languages-ai-services
4949

5050
## Code samples
5151

52-
In depth samples are available in the [Azure-Samples/cognitive-services-speech-sdk](https://aka.ms/csspeech/samples) repository on GitHub. There are samples for C# (including UWP and Unity), C++, Java, JavaScript (including Browser and Node.js), Objective-C, Python, and Swift. Code samples for Go are available in the [Microsoft/cognitive-services-speech-sdk-go](https://github.com/Microsoft/cognitive-services-speech-sdk-go) repository on GitHub.
52+
Code samples are available in the [Azure-Samples/cognitive-services-speech-sdk](https://aka.ms/csspeech/samples) repository on GitHub. There are samples for C# (including Universal Windows Platform (UWP) and Unity), C++, Java, JavaScript (including Browser and Node.js), Objective-C, Python, and Swift. Code samples for Go are available in the [Microsoft/cognitive-services-speech-sdk-go](https://github.com/Microsoft/cognitive-services-speech-sdk-go) repository on GitHub.
5353

5454
## Related content
5555

articles/machine-learning/data-science-virtual-machine/dsvm-tutorial-resource-manager.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ If your environment meets the prerequisites and you know how to use ARM template
2828

2929
## Prerequisites
3030

31-
* An Azure subscription. If you don't have an Azure subscription, create a [free account](https://azure.microsoft.com/free/services/machine-learning/) before you begin.
31+
* An Azure subscription. If you don't have an Azure subscription, create a [free account](https://azure.microsoft.com/products/machine-learning/) before you begin.
3232

3333
* You need the [Azure CLI](/cli/azure/install-azure-cli) to use the CLI commands in this document from your **local environment**.
3434

articles/machine-learning/data-science-virtual-machine/reference-ubuntu-vm.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ available in the `py38_pytorch` environment.
2828

2929
H2O is a fast, in-memory, distributed machine learning and predictive analytics platform. A Python package is installed in both the root and py35 Anaconda environments. An R package is also installed.
3030

31-
To open H2O from the command line, run `java -jar /dsvm/tools/h2o/current/h2o.jar`. You can configure various available[command-line options](http://docs.h2o.ai/h2o/latest-stable/h2o-docs/starting-h2o.html#from-the-command-line). Browse to the Flow web UI to `http://localhost:54321` to get started. JupyterHub offers sample notebooks.
31+
To open H2O from the command line, run `java -jar /dsvm/tools/h2o/current/h2o.jar`. You can configure various available command-line options. Browse to the Flow web UI to `http://localhost:54321` to get started. JupyterHub offers sample notebooks.
3232

3333
### TensorFlow
3434

articles/machine-learning/prompt-flow/troubleshoot-guidance.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -146,7 +146,7 @@ You may encounter 409 error from Azure OpenAI, it means you have reached the rat
146146

147147
In this case, if you find the message `request canceled` in the logs, it might be because the OpenAI API call is taking too long and exceeding the timeout limit.
148148

149-
An OpenAI API timeout could be caused by a network issue or a complex request that requires more processing time. For more information, see [OpenAI API timeout](https://help.openai.com/en/articles/6897186-timeout).
149+
An OpenAI API timeout could be caused by a network issue or a complex request that requires more processing time. For more information, see [OpenAI API timeout](https://platform.openai.com/docs/actions/production#timeouts).
150150

151151
Wait a few seconds and retry your request. This action usually resolves any network issues.
152152

0 commit comments

Comments
 (0)