Skip to content

Commit 64d7b9e

Browse files
authored
Merge pull request #2141 from JKirsch1/update-article-about-monitoring-online-endpoints
Freshness - Machine Learning HowTo 180 days
2 parents 09c026e + 4d3309a commit 64d7b9e

23 files changed

+198
-156
lines changed

articles/machine-learning/how-to-authenticate-online-endpoint.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -728,7 +728,7 @@ The __Test__ tab of the deployment's detail page supports scoring for endpoints
728728

729729
## Log and monitor traffic
730730

731-
To enable traffic logging in the diagnostics settings for the endpoint, follow the steps in [How to enable/disable logs](how-to-monitor-online-endpoints.md#how-to-enabledisable-logs).
731+
To enable traffic logging in the diagnostics settings for the endpoint, follow the steps in [Turn on logs](how-to-monitor-online-endpoints.md#turn-on-logs).
732732

733733
If the diagnostic setting is enabled, you can check the `AmlOnlineEndpointTrafficLogs` table to see the auth mode and user identity.
734734

articles/machine-learning/how-to-autoscale-endpoints.md

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -455,7 +455,10 @@ The following steps continue the rule configuration on the __Custom autoscale__
455455

456456
## Find IDs for supported metrics
457457

458-
If you want to use other metrics in code to set up autoscale rules by using the Azure CLI or the SDK, see the table in [Available metrics](how-to-monitor-online-endpoints.md#available-metrics).
458+
You can use other metrics when you use the Azure CLI or the SDK to set up autoscale rules.
459+
460+
- For the names of endpoint metrics to use in code, see the values in the __Name in REST API__ column in the table in [Supported metrics for Microsoft.MachineLearningServices/workspaces/onlineEndpoints](monitor-azure-machine-learning-reference.md#supported-metrics-for-microsoftmachinelearningservicesworkspacesonlineendpoints).
461+
- For the names of deployment metrics to use in code, see the values in the __Name in REST API__ column in the tables in [Supported metrics for Microsoft.MachineLearningServices/workspaces/onlineEndpoints/deployments](monitor-azure-machine-learning-reference.md#supported-metrics-for-microsoftmachinelearningservicesworkspacesonlineendpointsdeployments).
459462

460463
## Create scale rule based on schedule
461464

articles/machine-learning/how-to-deploy-online-endpoints.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1230,7 +1230,7 @@ To view metrics and set alerts based on your SLA, complete the steps that are de
12301230

12311231
### (Optional) Integrate with Log Analytics
12321232

1233-
The `get-logs` command for CLI or the `get_logs` method for SDK provides only the last few hundred lines of logs from an automatically selected instance. However, Log Analytics provides a way to durably store and analyze logs. For more information on using logging, see [Monitor online endpoints](how-to-monitor-online-endpoints.md#logs).
1233+
The `get-logs` command for CLI or the `get_logs` method for SDK provides only the last few hundred lines of logs from an automatically selected instance. However, Log Analytics provides a way to durably store and analyze logs. For more information on using logging, see [Monitor online endpoints](how-to-monitor-online-endpoints.md#use-logs).
12341234

12351235
<!-- [!INCLUDE [Email Notification Include](includes/machine-learning-email-notifications.md)] -->
12361236

articles/machine-learning/how-to-manage-quotas.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -123,7 +123,7 @@ Azure Machine Learning online endpoints and batch endpoints have resource limits
123123
> [!IMPORTANT]
124124
> These limits are *regional*, meaning that you can use up to these limits per each region you're using. For example, if your current limit for number of endpoints per subscription is 100, you can create 100 endpoints in the East US region, 100 endpoints in the West US region, and 100 endpoints in each of the other supported regions in a single subscription. Same principle applies to all the other limits.
125125
126-
To determine the current usage for an endpoint, [view the metrics](how-to-monitor-online-endpoints.md#metrics).
126+
To determine the current usage for an endpoint, [view the metrics](how-to-monitor-online-endpoints.md#use-metrics).
127127

128128
To request an exception from the Azure Machine Learning product team, use the steps in the [Endpoint limit increases](#endpoint-limit-increases).
129129

articles/machine-learning/how-to-monitor-online-endpoints.md

Lines changed: 149 additions & 110 deletions
Large diffs are not rendered by default.

articles/machine-learning/how-to-safely-rollout-online-endpoints.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -577,7 +577,7 @@ Mirroring has the following limitations:
577577
* Mirroring is supported for the CLI (v2) (version 2.4.0 or above) and Python SDK (v2) (version 1.0.0 or above). If you use an older version of CLI/SDK to update an endpoint, you'll lose the mirror traffic setting.
578578
* Mirroring isn't currently supported for Kubernetes online endpoints.
579579
* You can mirror traffic to only one deployment in an endpoint.
580-
* The maximum percentage of traffic you can mirror is 50%. This limit is to reduce the effect on your [endpoint bandwidth quota](how-to-manage-quotas.md#azure-machine-learning-online-endpoints-and-batch-endpoints) (default 5 MBPS)—your endpoint bandwidth is throttled if you exceed the allocated quota. For information on monitoring bandwidth throttling, see [Monitor managed online endpoints](how-to-monitor-online-endpoints.md#metrics-at-endpoint-scope).
580+
* The maximum percentage of traffic you can mirror is 50%. This limit is to reduce the effect on your [endpoint bandwidth quota](how-to-manage-quotas.md#azure-machine-learning-online-endpoints-and-batch-endpoints) (default 5 MBPS)—your endpoint bandwidth is throttled if you exceed the allocated quota. For information on monitoring bandwidth throttling, see [Monitor managed online endpoints](how-to-monitor-online-endpoints.md#metrics-at-the-endpoint-scope).
581581

582582
Also note the following behaviors:
583583

articles/machine-learning/includes/endpoint-monitor-console-reference.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -2,15 +2,15 @@
22
author: Blackmist
33
ms.service: azure-machine-learning
44
ms.topic: include
5-
ms.date: 04/22/2023
5+
ms.date: 01/02/2025
66
ms.author: larryfr
77
---
88

99
| Property | Description |
1010
|:--- |:--- |
11-
| TimeGenerated | The timestamp (UTC) of when the log was generated.
12-
| OperationName | The operation associated with log record.
13-
| InstanceId | The ID of the instance that generated this log record.
14-
| DeploymentName | The name of the deployment associated with the log record.
15-
| ContainerName | The name of the container where the log was generated.
16-
| Message | The content of the log.
11+
| TimeGenerated | The UTC time stamp of the time the log is generated |
12+
| OperationName | The operation associated with the log record |
13+
| InstanceId | The ID of the instance that generates the log record |
14+
| DeploymentName | The name of the deployment associated with the log record |
15+
| ContainerName | The name of the container where the log is generated |
16+
| Message | The content of the log |

articles/machine-learning/includes/endpoint-monitor-event-reference.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -2,15 +2,15 @@
22
author: Blackmist
33
ms.service: azure-machine-learning
44
ms.topic: include
5-
ms.date: 04/22/2023
5+
ms.date: 01/02/2025
66
ms.author: larryfr
77
---
88

99
| Property | Description |
1010
|:--- |:--- |
11-
| TimeGenerated | The timestamp (UTC) of when the log was generated.
12-
| OperationName | The operation associated with log record.
13-
| InstanceId | The ID of the instance that generated this log record.
14-
| DeploymentName | The name of the deployment associated with the log record.
15-
| Name | The name of the event.
16-
| Message | The content of the event.
11+
| TimeGenerated | The UTC time stamp of the time the log is generated |
12+
| OperationName | The operation associated with the log record |
13+
| InstanceId | The ID of the instance that generates the log record |
14+
| DeploymentName | The name of the deployment associated with the log record |
15+
| Name | The name of the event |
16+
| Message | The content of the event |

articles/machine-learning/includes/endpoint-monitor-traffic-reference.md

Lines changed: 23 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -2,31 +2,31 @@
22
author: Blackmist
33
ms.service: azure-machine-learning
44
ms.topic: include
5-
ms.date: 04/22/2023
5+
ms.date: 01/02/2025
66
ms.author: larryfr
77
---
88

99
| Property | Description |
1010
|:--- |:--- |
11-
| Method | The requested method from client.
12-
| Path | The requested path from client.
13-
| SubscriptionId | The machine learning subscription ID of the online endpoint.
14-
| AzureMLWorkspaceId | The machine learning workspace ID of the online endpoint.
15-
| AzureMLWorkspaceName | The machine learning workspace name of the online endpoint.
16-
| EndpointName | The name of the online endpoint.
17-
| DeploymentName | The name of the online deployment.
18-
| Protocol | The protocol of the request.
19-
| ResponseCode | The final response code returned to the client.
20-
| ResponseCodeReason | The final response code reason returned to the client.
21-
| ModelStatusCode | The response status code from model.
22-
| ModelStatusReason | The response status reason from model.
23-
| RequestPayloadSize | The total bytes received from the client.
24-
| ResponsePayloadSize | The total bytes sent back to the client.
25-
| UserAgent | The user-agent header of the request, including comments but truncated to a max of 70 characters.
26-
| XRequestId | The request ID generated by Azure Machine Learning for internal tracing.
27-
| XMSClientRequestId | The tracking ID generated by the client.
28-
| TotalDurationMs | Duration in milliseconds from the request start time to the last response byte sent back to the client. If the client disconnected, it measures from the start time to client disconnect time.
29-
| RequestDurationMs | Duration in milliseconds from the request start time to the last byte of the request received from the client.
30-
| ResponseDurationMs | Duration in milliseconds from the request start time to the first response byte read from the model.
31-
| RequestThrottlingDelayMs | Delay in milliseconds in request data transfer due to network throttling.
32-
| ResponseThrottlingDelayMs | Delay in milliseconds in response data transfer due to network throttling.
11+
| Method | The method that the client requests. |
12+
| Path | The path that the client requests. |
13+
| SubscriptionId | The machine learning subscription ID of the online endpoint. |
14+
| AzureMLWorkspaceId | The machine learning workspace ID of the online endpoint. |
15+
| AzureMLWorkspaceName | The machine learning workspace name of the online endpoint. |
16+
| EndpointName | The name of the online endpoint. |
17+
| DeploymentName | The name of the online deployment. |
18+
| Protocol | The protocol of the request. |
19+
| ResponseCode | The final response code that's returned to the client. |
20+
| ResponseCodeReason | The final response code reason that's returned to the client. |
21+
| ModelStatusCode | The response status code from the model. |
22+
| ModelStatusReason | The response status reason from the model. |
23+
| RequestPayloadSize | The total bytes received from the client. |
24+
| ResponsePayloadSize | The total bytes sent back to the client. |
25+
| UserAgent | The user-agent header of the request, including comments but truncated to a maximum of 70 characters. |
26+
| XRequestId | The request ID that Azure Machine Learning generates for internal tracing. |
27+
| XMSClientRequestId | The tracking ID that the client generates. |
28+
| TotalDurationMs | The duration in milliseconds from the request start time to the time the last response byte is sent back to the client. If the client disconnects, the duration is taken from the start time to the client disconnect time. |
29+
| RequestDurationMs | The duration in milliseconds from the request start time to the time the last byte of the request is received from the client. |
30+
| ResponseDurationMs | The duration in milliseconds from the request start time to the time the first response byte is read from the model. |
31+
| RequestThrottlingDelayMs | The delay in milliseconds in the request data transfer due to network throttling. |
32+
| ResponseThrottlingDelayMs | The delay in milliseconds in the response data transfer due to network throttling. |
57.9 KB
Loading

0 commit comments

Comments
 (0)