Skip to content

Commit 3f133d9

Browse files
authored
Merge pull request #5392 from MicrosoftDocs/main
6/5/2025 AM Publish
2 parents 08bcab8 + 3acd8c7 commit 3f133d9

File tree

8 files changed

+81
-43
lines changed

8 files changed

+81
-43
lines changed

articles/ai-services/agents/how-to/tools/bing-grounding.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -37,6 +37,13 @@ Grounding with Bing returns relevant search results to the customer's model depl
3737
3838
The authorization will happen between Grounding with Bing Search service and Azure AI Foundry Agent Service. Any Bing search query that is generated and sent to Bing for the purposes of grounding is transferred, along with the resource key, outside of the Azure compliance boundary to the Grounding with Bing Search service. Grounding with Bing Search is subject to Bing's terms and do not have the same compliance standards and certifications as the Azure AI Foundry Agent Service, as described in the [Grounding with Bing Search Terms of Use](https://www.microsoft.com/bing/apis/grounding-legal). It is your responsibility to assess whether the use of Grounding with Bing Search in your agent meets your needs and requirements.
3939

40+
## Supported capabilities and known issues
41+
- Grounding with Bing Search tool is designed to retrieve real-time information from web, NOT specific web domains.
42+
- NOT Recommended to **summarize** an entire web page.
43+
- Within one run, the AI model will evaluate the tool outputs and may decide to invoke the tool again for more information and context. AI model may also decide which piece(s) of tool outputs are used to generate the response.
44+
- Azure AI Agent service will return **AI model generated response** as output so end-to-end latency will be impacted pre-/post-processing of LLMs.
45+
- Grounding with Bing Search tool does NOT return the tool output to developers and end users.
46+
4047
## Usage support
4148

4249
|Azure AI foundry support | Python SDK | C# SDK | JavaScript SDK | REST API |Basic agent setup | Standard agent setup |

articles/ai-services/agents/how-to/virtual-networks.md

Lines changed: 31 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -54,6 +54,7 @@ For customers without an existing virtual network, the Standard Setup with Priva
5454
* `Microsoft.Search`
5555
* `Microsoft.Network`
5656
* `Microsoft.App`
57+
* `Microsoft.ContainerService`
5758
* To use Bing Search tool: `Microsoft.Bing`
5859

5960
```console
@@ -104,6 +105,36 @@ For customers without an existing virtual network, the Standard Setup with Priva
104105
```console
105106
az deployment group create --resource-group {my_resource_group} --template-file main-create.bicep
106107
```
108+
109+
1. Run the CheckCapabilityHostReadiness.ps1 and edit the command to add your subscription ID, resource group name, and your newly deployed AI Services account resource name.
110+
111+
```
112+
.\CheckCapabilityHostReadiness.ps1 -subscriptionId "<your-sub-id>" -resourcegroup "<new-rg-name>" -accountname "<your-aiservices-name>"
113+
```
114+
115+
If you don't want to run the PowerShell script, you can run a bash script instead, from the file CheckCapabilityHostReadiness.sh. Run the following two commands:
116+
117+
```
118+
chmod +x CheckCapabilityHostReadiness.sh
119+
./CheckCapabilityHostReadiness.sh "<your-sub-id>" "<new-rg-name>" "<your-aiservices-name>"
120+
```
121+
122+
1. Deploy the main-project-caphost-create.bicep
123+
124+
```
125+
az deployment group create --resource-group <new-rg-name> --template-file main-project-caphost-create.bicep
126+
```
127+
128+
After running this script, you're required to input the following values:
129+
130+
```
131+
Please provide string value for 'accountName' (? for help): <your-account-name>
132+
Please provide string value for 'projectName' (? for help): <your-project-name>
133+
Please provide string value for 'aiSearchName' (? for help): <your-search-name>
134+
Please provide string value for 'azureStorageName' (? for help): <your-storage-name>
135+
Please provide string value for 'cosmosDBName' (? for help): <your-cosmosdb-name>
136+
```
137+
107138
For more details, see the [README](https://github.com/azure-ai-foundry/foundry-samples/tree/main/samples/microsoft/infrastructure-setup/15-private-network-standard-agent-setup).
108139
109140
## Deep Dive Standard Setup with Private Networking Template

articles/ai-services/agents/includes/quickstart-python.md

Lines changed: 28 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -79,32 +79,32 @@ with project_client:
7979
)
8080
print(f"Created agent, ID: {agent.id}")
8181

82-
# Create a thread for communication
83-
thread = project_client.agents.threads.create()
84-
print(f"Created thread, ID: {thread.id}")
85-
86-
# Add a message to the thread
87-
message = project_client.agents.messages.create(
88-
thread_id=thread.id,
89-
role="user", # Role of the message sender
90-
content="What is the weather in Seattle today?", # Message content
91-
)
92-
print(f"Created message, ID: {message['id']}")
93-
94-
# Create and process an agent run
95-
run = project_client.agents.runs.create_and_process(thread_id=thread.id, agent_id=agent.id)
96-
print(f"Run finished with status: {run.status}")
97-
98-
# Check if the run failed
99-
if run.status == "failed":
100-
print(f"Run failed: {run.last_error}")
101-
102-
# Fetch and log all messages
103-
messages = project_client.agents.messages.list(thread_id=thread.id)
104-
for message in messages.data:
105-
print(f"Role: {message.role}, Content: {message.content}")
106-
107-
# Delete the agent when done
108-
project_client.agents.delete_agent(agent.id)
109-
print("Deleted agent")
82+
# Create a thread for communication
83+
thread = project_client.agents.threads.create()
84+
print(f"Created thread, ID: {thread.id}")
85+
86+
# Add a message to the thread
87+
message = project_client.agents.messages.create(
88+
thread_id=thread.id,
89+
role="user", # Role of the message sender
90+
content="What is the weather in Seattle today?", # Message content
91+
)
92+
print(f"Created message, ID: {message['id']}")
93+
94+
# Create and process an agent run
95+
run = project_client.agents.runs.create_and_process(thread_id=thread.id, agent_id=agent.id)
96+
print(f"Run finished with status: {run.status}")
97+
98+
# Check if the run failed
99+
if run.status == "failed":
100+
print(f"Run failed: {run.last_error}")
101+
102+
# Fetch and log all messages
103+
messages = project_client.agents.messages.list(thread_id=thread.id)
104+
for message in messages:
105+
print(f"Role: {message.role}, Content: {message.content}")
106+
107+
# Delete the agent when done
108+
project_client.agents.delete_agent(agent.id)
109+
print("Deleted agent")
110110
```

articles/ai-services/openai/how-to/provisioned-throughput-onboarding.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -77,14 +77,14 @@ The amount of throughput (measured in tokens per minute or TPM) a deployment get
7777

7878
For example, for `gpt-4.1:2025-04-14`, 1 output token counts as 4 input tokens towards your utilization limit which matches the [pricing](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/). Older models use a different ratio and for a deeper understanding on how different ratios of input and output tokens impact the throughput your workload needs, see the [Azure AI Foundry PTU quota calculator](https://ai.azure.com/resource/calculator).
7979

80-
|Topic| **o4-mini** | **gpt-4.1** | **gpt-4.1-mini** | **gpt-4.1-nano** | **o3** | **o3-mini** | **o1** | **gpt-4o** | **gpt-4o-mini** | **DeepSeek-R1** | **DeepSeek-V3-0324** | **MAI-DS-R1** |
81-
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
82-
|Global & data zone provisioned minimum deployment| 15 | 15|15| 15 | 15 |15|15|15|15| 100|100|100|
83-
|Global & data zone provisioned scale increment| 5 | 5|5| 5 | 5 |5|5|5|5| 100|100|100|
84-
|Regional provisioned minimum deployment|25| 50|25| 25 |50 | 25|25|50|25| NA|NA|NA|
85-
|Regional provisioned scale increment|25| 50|25| 25 | 50 | 25|50|50|25|NA|NA|NA|
86-
|Input TPM per PTU|5,400 | 3,000|14,900| 59,400 | 600 | 2,500|230|2,500|37,000|4,000|4,000|4,000|
87-
|Latency Target Value| 66 Tokens Per Second | 40 Tokens Per Second|50 Tokens Per Second| 60 Tokens Per Second | 40 Tokens Per Second | 66 Tokens Per Second |25 Tokens Per Second|25 Tokens Per Second|33 Tokens Per Second|50 Tokens Per Second|50 Tokens Per Second|50 Tokens Per Second|
80+
|Topic| **o4-mini** | **gpt-4.1** | **gpt-4.1-mini** | **gpt-4.1-nano** | **o3** | **o3-mini** | **o1** | **gpt-4o** | **gpt-4o-mini** | **DeepSeek-R1** | **DeepSeek-V3-0324** |
81+
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
82+
|Global & data zone provisioned minimum deployment| 15 | 15|15| 15 | 15 |15|15|15|15| 100|100|
83+
|Global & data zone provisioned scale increment| 5 | 5|5| 5 | 5 |5|5|5|5| 100|100|
84+
|Regional provisioned minimum deployment|25| 50|25| 25 |50 | 25|25|50|25| NA|NA|
85+
|Regional provisioned scale increment|25| 50|25| 25 | 50 | 25|50|50|25|NA|NA|
86+
|Input TPM per PTU|5,400 | 3,000|14,900| 59,400 | 600 | 2,500|230|2,500|37,000|4,000|4,000|
87+
|Latency Target Value| 66 Tokens Per Second | 40 Tokens Per Second|50 Tokens Per Second| 60 Tokens Per Second | 40 Tokens Per Second | 66 Tokens Per Second |25 Tokens Per Second|25 Tokens Per Second|33 Tokens Per Second|50 Tokens Per Second|50 Tokens Per Second|
8888

8989

9090
For a full list, see the [Azure AI Foundry calculator](https://ai.azure.com/resource/calculator).

articles/machine-learning/feature-retrieval-concepts.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -96,7 +96,7 @@ serialization_version: 2
9696
The feature store point-in-time join can create training data in two ways:
9797
9898
- The `get_offline_features()` API function in the feature store SDK in a Spark session/job
99-
- The Azure Machine Learning build-in feature retrieval (pipeline) component
99+
- The Azure Machine Learning built-in feature retrieval (pipeline) component
100100

101101
In the first option, the feature retrieval specification itself is optional because the user can provide the list of features on that API. However, if a feature retrieval specification is provided, the `resolve_feature_retrieval_spec()` function in the feature store SDK can load the list of features that the specification defined. That function then passes that list to the `get_offline_features()` API function.
102102

articles/machine-learning/troubleshooting-managed-feature-store.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -279,7 +279,7 @@ Once the `begin_create_or_update()` call returns successfully, the next `feature
279279
- [Observation Data isn't Joined with any feature values](#observation-data-isnt-joined-with-any-feature-values)
280280
- [User or Managed Identity doesn't have proper RBAC permission on the feature store](#user-or-managed-identity-doesnt-have-proper-rbac-permission-on-the-feature-store)
281281
- [User or Managed Identity doesn't have proper RBAC permission to Read from the Source Storage or Offline store](#user-or-managed-identity-doesnt-have-proper-rbac-permission-to-read-from-the-source-storage-or-offline-store)
282-
- [Training job fails to read data generated by the build-in Feature Retrieval Component](#training-job-fails-to-read-data-generated-by-the-build-in-feature-retrieval-component)
282+
- [Training job fails to read data generated by the built-in Feature Retrieval Component](#training-job-fails-to-read-data-generated-by-the-built-in-feature-retrieval-component)
283283
- [`generate_feature_retrieval_spec()` fails due to use of local feature set specification](#generate_feature_retrieval_spec-fails-due-to-use-of-local-feature-set-specification)
284284
- [The `get_offline_features()` query takes a long time](#the-get_offline_features-query-takes-a-long-time)
285285

@@ -450,7 +450,7 @@ An error occurred while calling o1025.parquet.
450450

451451
`Storage Blob Data Reader` is the minimum recommended access requirement. Users can also assign roles - for example, `Storage Blob Data Contributor` or `Storage Blob Data Owner` - with more privileges.
452452

453-
### Training job fails to read data generated by the build-in Feature Retrieval Component
453+
### Training job fails to read data generated by the built-in Feature Retrieval Component
454454

455455
#### Symptom
456456

articles/machine-learning/v1/how-to-debug-visual-studio-code.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -369,7 +369,7 @@ Local web service deployments require a working Docker installation on your loca
369369
370370
1. To configure VS Code to communicate with the Docker image, create a new debug configuration:
371371
372-
1. From VS Code, select the __Debug__ menu in the __Run__ extention and then select __Open configurations__. A file named __launch.json__ opens.
372+
1. From VS Code, select the __Debug__ menu in the __Run__ extension and then select __Open configurations__. A file named __launch.json__ opens.
373373
374374
1. In the __launch.json__ file, find the __"configurations"__ item (the line that contains `"configurations": [`), and insert the following text after it.
375375
@@ -509,7 +509,7 @@ Local web service deployments require a working Docker installation on your loca
509509
510510
This command attaches your `score.py` locally to the one in the container. Therefore, any changes made in the editor are automatically reflected in the container
511511
512-
2. For a better experience, you can go into the container with a new VS Code interface. Select the `Docker` extention from the VS Code side bar, find your local container created, in this documentation its `debug:1`. Right-click this container and select `"Attach Visual Studio Code"`, then a new VS Code interface will be opened automatically, and this interface shows the inside of your created container.
512+
2. For a better experience, you can go into the container with a new VS Code interface. Select the `Docker` extension from the VS Code side bar, find your local container created, in this documentation its `debug:1`. Right-click this container and select `"Attach Visual Studio Code"`, then a new VS Code interface will be opened automatically, and this interface shows the inside of your created container.
513513
514514
![The container VS Code interface](../media/how-to-troubleshoot-deployment/container-interface.png)
515515
@@ -522,7 +522,7 @@ Local web service deployments require a working Docker installation on your loca
522522
523523
![The container run console output](../media/how-to-troubleshoot-deployment/container-run.png)
524524
525-
4. To attach VS Code to debugpy inside the container, open VS Code, and use the F5 key or select __Debug__. When prompted, select the __Azure Machine Learning Deployment: Docker Debug__ configuration. You can also select the __Run__ extention icon from the side bar, the __Azure Machine Learning Deployment: Docker Debug__ entry from the Debug dropdown menu, and then use the green arrow to attach the debugger.
525+
4. To attach VS Code to debugpy inside the container, open VS Code, and use the F5 key or select __Debug__. When prompted, select the __Azure Machine Learning Deployment: Docker Debug__ configuration. You can also select the __Run__ extension icon from the side bar, the __Azure Machine Learning Deployment: Docker Debug__ entry from the Debug dropdown menu, and then use the green arrow to attach the debugger.
526526
527527
![The debug icon, start debugging button, and configuration selector](../media/how-to-troubleshoot-deployment/start-debugging.png)
528528

articles/machine-learning/v1/how-to-monitor-datasets.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -194,7 +194,7 @@ As described later, a dataset monitor runs at a set frequency (daily, weekly, mo
194194
The **backfill** function runs a backfill job, for a specified start and end date range. A backfill job fills in expected missing data points in a data set, as a way to ensure data accuracy and completeness.
195195

196196
> [!NOTE]
197-
> Azure Machine Learning model monitoring doesn't support manual **backfill** function, if you want to redo the model monitor for a specif time range, you can create another model monitor for that specific time range.
197+
> Azure Machine Learning model monitoring doesn't support manual **backfill** function, if you want to redo the model monitor for a specific time range, you can create another model monitor for that specific time range.
198198
199199
# [Python SDK](#tab/python)
200200
<a name="sdk-monitor"></a>

0 commit comments

Comments
 (0)