Skip to content

Commit 2a8727c

Browse files
authored
Merge pull request #3527 from MicrosoftDocs/main
3/13/2025 AM Publish
2 parents 838ba6c + e2cec84 commit 2a8727c

File tree

5 files changed

+83
-26
lines changed

5 files changed

+83
-26
lines changed

articles/ai-foundry/model-inference/how-to/manage-costs.md

Lines changed: 12 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.date: 1/21/2025
1212

1313
# Plan to manage costs for model inference in Azure AI Services
1414

15-
This article describes how you can plan for and manage costs for model inference in Azure AI Services. After you start using model inference in Azure AI Services resources, use **Cost Management features** to set budgets and monitor costs.
15+
This article describes how you can view, plan for, and manage costs for model inference in Azure AI Services.
1616

1717
Although this article is about planning for and managing costs for model inference in Azure AI Services, you're billed for all Azure services and resources used in your Azure subscription.
1818

@@ -24,11 +24,9 @@ Although this article is about planning for and managing costs for model inferen
2424

2525
## Understand model inference billing model
2626

27-
Models deployed in Azure AI Services are charged per 1,000 tokens. Language models understand and process text by breaking it down into tokens. For reference, each token is roughly four characters for typical English text. Costs per token vary depending on which model series you choose. Models that can process images break down images in tokens too. The number of tokens per image depends on the model and the resolution of the input image.
27+
Language models understand and process inputs by breaking them down into tokens. For reference, each token is roughly four characters for typical English text. Models that can process images or audio break down them into tokens too for billing purposes. The number of tokens per image or audio content depends on the model and the resolution of the input.
2828

29-
Token costs are for both input and output. For example, suppose you have a 1,000 token JavaScript code sample that you ask a model to convert to Python. You would be charged approximately 1,000 tokens for the initial input request sent, and 1,000 more tokens for the output that is received in response for a total of 2,000 tokens.
30-
31-
In practice, for this type of completion call, the token input/output wouldn't be perfectly 1:1. A conversion from one programming language to another could result in a longer or shorter output depending on many factors. One such factor is the value assigned to the `max_tokens` parameter.
29+
Costs per token vary depending on which model series you choose but in all cases models deployed in Azure AI Services are charged per 1,000 tokens. Token costs are for both input and output. For example, suppose you have a 1,000 token JavaScript code sample that you ask a model to convert to Python. You would be charged approximately 1,000 tokens for the initial input request sent, and 1,000 more tokens for the output that is received in response for a total of 2,000 tokens.
3230

3331
### Cost breakdown
3432

@@ -57,16 +55,19 @@ The following sections explain the entries in details.
5755

5856
### Azure OpenAI and Microsoft models
5957

60-
Azure OpenAI and Microsoft's family of models (like Phi) are charged directly and they show up as billing meters under each Azure AI services resource. This billing happens directly through Microsoft. When you inspect your bill, you notice billing meters accounting for inputs and outputs for each consumed model.
58+
Azure OpenAI models and models offered as first-party consumption services from Microsoft (including DeepSeek family and Phi family of models) are charged directly and they show up as billing meters under each Azure AI services resource. This billing happens directly through Microsoft. When you inspect your bill, you notice billing meters accounting for inputs and outputs for each consumed model.
6159

6260
:::image type="content" source="../media/manage-cost/cost-by-meter-1p.png" alt-text="Screenshot of cost analysis dashboard scoped to the resource group where the Azure AI Services resource is deployed, highlighting the meters for Azure OpenAI and Microsoft's models. Cost is group by meter." lightbox="../media/manage-cost/cost-by-meter-1p.png":::
6361

6462
### Provider models
6563

66-
Models provided by another provider, like Mistral AI, Cohere, Meta AI, or AI21 Labs, are billed using Azure Marketplace. As opposite to Microsoft billing meters, those entries are associated with the resource group where your Azure AI services is deployed instead of to the Azure AI Services resource itself. You see entries under the **Service Name** *SaaS* accounting for inputs and outputs for each consumed model.
64+
Models provided by another provider, like Mistral AI, Cohere, Meta AI, or AI21 Labs, are billed using Azure Marketplace. As opposite to Microsoft billing meters, those entries are associated with the resource group where your Azure AI services is deployed instead of to the Azure AI Services resource itself. Given model providers charge you directly, you see entries under the category **Marketplace** and **Service Name** *SaaS* accounting for inputs and outputs for each consumed model.
6765

6866
:::image type="content" source="../media/manage-cost/cost-by-meter-saas.png" alt-text="Screenshot of cost analysis dashboard scoped to the resource group where the Azure AI Services resource is deployed, highlighting the meters for models billed throughout Azure Marketplace. Cost is group by meter." lightbox="../media/manage-cost/cost-by-meter-saas.png":::
6967

68+
> [!IMPORTANT]
69+
> This distinction between Azure OpenAI, Microsoft-offered models, and provider models only affects how the model is made available to you and how you are charged. In all cases, models are hosted within Azure cloud and there is no interaction with external services or providers.
70+
7071
### Using Azure Prepayment
7172

7273
You can pay for Azure OpenAI and Microsoft's models charges with your Azure Prepayment credit. However, you can't use Azure Prepayment credit to pay for charges for other provider models given they're billed through Azure Marketplace.
@@ -78,10 +79,6 @@ For example, a 400 error due to a content filter or input limit, or a 408 error
7879

7980
If the service doesn't perform processing, you aren't charged. For example, a 401 error due to authentication or a 429 error due to exceeding the Rate Limit.
8081

81-
## Other costs
82-
83-
Enabling capabilities such as sending data to Azure Monitor Logs and alerting incurs extra costs for those services. These costs are visible under those other services and at the subscription level, but aren't visible when scoped just to your Azure AI services resource.
84-
8582
## Monitor costs
8683

8784
Azure resource usage unit costs vary by time intervals, such as seconds, minutes, hours, and days, or by unit usage, such as bytes and megabytes. As soon as Azure AI services use starts, costs can be incurred and you can see the costs in the [cost analysis](/azure/cost-management/quick-acm-cost-analysis?WT.mc_id=costmanagementcontent_docsacmhorizontal_-inproduct-learn).
@@ -127,6 +124,10 @@ You can create budgets with filters for specific resources or services in Azure
127124

128125
You can also [export your cost data](/azure/cost-management-billing/costs/tutorial-export-acm-data?WT.mc_id=costmanagementcontent_docsacmhorizontal_-inproduct-learn) to a storage account, which is helpful when you need others to do extra data analysis for costs. For example, a finance team can analyze the data using Excel or Power BI. You can export your costs on a daily, weekly, or monthly schedule and set a custom date range. We recommend exporting cost data as the way to retrieve cost datasets.
129126

127+
## Other costs
128+
129+
Enabling capabilities such as sending data to Azure Monitor Logs and alerting incurs extra costs for those services. These costs are visible under those other services and at the subscription level, but aren't visible when scoped just to your Azure AI services resource.
130+
130131
## Next steps
131132

132133
- Learn [how to optimize your cloud investment with cost management](/azure/cost-management-billing/costs/cost-mgt-best-practices?WT.mc_id=costmanagementcontent_docsacmhorizontal_-inproduct-learn).

articles/machine-learning/azure-machine-learning-ci-image-release-notes.md

Lines changed: 30 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -20,6 +20,36 @@ Azure Machine Learning checks and validates any machine learning packages that m
2020

2121
Main updates provided with each image version are described in the below sections.
2222

23+
## February 11, 2025
24+
25+
Image Version: `25.01.31`
26+
27+
Release Notes:
28+
29+
SDK Version: `1.59.0`
30+
31+
## January 15, 2025
32+
33+
Image Version: `24.12.31`
34+
35+
Release Notes:
36+
37+
SDK Version: `1.57.0`
38+
39+
Jupyter-core: `5.7.2`
40+
41+
nvdia_docker2: installed
42+
43+
gnomeshell: removed
44+
45+
ml: '2.32.4'
46+
47+
Nvidia Driver: `535.216.03`
48+
49+
`CUDA`: `12.2`
50+
51+
'nginx': Server status was Failed. nginx issue fixed and the status is Running.
52+
2353
## December 18, 2024
2454

2555
Image Version: `24.12.09`

articles/machine-learning/data-science-virtual-machine/release-notes.md

Lines changed: 35 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -21,6 +21,41 @@ Azure portal users can find the latest image available for provisioning the Data
2121

2222
Visit the [list of known issues](reference-known-issues.md) to learn about known bugs and workarounds.
2323

24+
## February 18, 2025
25+
26+
[Data Science VM – Ubuntu 20.04](https://azuremarketplace.microsoft.com/marketplace/apps/microsoft-dsvm.ubuntu-2004?tab=Overview)
27+
28+
Version `25.02.13`
29+
30+
- SDK `1.59.0`
31+
- NVDIA `535.183.01`
32+
- Cuda `cuda_12.2.r12`
33+
- Python `3.10.8`
34+
35+
## February 7, 2025
36+
37+
[Data Science Virtual Machine - Windows 2022](https://azuremarketplace.microsoft.com/marketplace/apps/microsoft-dsvm.dsvm-win-2022?tab=Overview)
38+
39+
Version `25.02.03`
40+
41+
- SDK `1.59.0`
42+
43+
## February 4, 2025
44+
45+
[Data Science Virtual Machine - Windows 2019](https://azuremarketplace.microsoft.com/marketplace/apps/microsoft-dsvm.dsvm-win-2019?tab=Overview)
46+
47+
Version `25.01.31`
48+
49+
- SDK `1.59.0`
50+
51+
## January 22, 2025
52+
53+
[Data Science VM – Ubuntu 22.04](https://azuremarketplace.microsoft.com/marketplace/apps/microsoft-dsvm.ubuntu-2004?tab=Overview)
54+
55+
Version `25.01.20`
56+
57+
- SDK `1.59.0`
58+
2459
## October 21, 2024
2560

2661
[Data Science Virtual Machine – Ubuntu 22.04](https://azuremarketplace.microsoft.com/marketplace/apps/microsoft-dsvm.ubuntu-2204?tab=Overview)

articles/machine-learning/how-to-migrate-from-v1.md

Lines changed: 0 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -40,15 +40,6 @@ You should use v2 if you're starting a new machine learning project or workflow.
4040

4141
A new v2 project can reuse existing v1 resources like workspaces and compute and existing assets like models and environments created using v1.
4242

43-
Some feature gaps in v2 include:
44-
45-
- Spark support in jobs - currently in preview in v2.
46-
- Publishing jobs (pipelines in v1) as endpoints. You can however, schedule pipelines without publishing.
47-
- Support for SQL/database datastores.
48-
- Ability to use classic prebuilt components in the designer with v2.
49-
50-
You should then ensure the features you need in v2 meet your organization's requirements, such as being generally available.
51-
5243
> [!IMPORTANT]
5344
> New features in Azure Machine Learning will only be launched in v2.
5445

articles/machine-learning/how-to-monitor-kubernetes-online-enpoint-inference-server-log.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -6,19 +6,19 @@ services: machine-learning
66
ms.service: azure-machine-learning
77
ms.subservice: mlops
88
ms.topic: conceptual
9-
author: zetiaatgithub
10-
ms.author: zetia
9+
author: Blackmist
10+
ms.author: larryfr
11+
ms.reviewer: zetia
1112
ms.custom: devplatv2
12-
ms.date: 09/26/2023
13+
ms.date: 03/12/2025
1314
---
1415

1516
# Monitor Kubernetes Online Endpoint inference server logs
1617

1718
[!INCLUDE [dev v2](includes/machine-learning-dev-v2.md)]
1819

1920

20-
To diagnose online issues and monitor Azure Machine Learning model inference server metrics, we usually need to collect model inference server logs.
21-
21+
To diagnose online issues and monitor Azure Machine Learning model inference server metrics, you usually need to collect model inference server logs. In this article, learn how to collect inference server logs from Azure Kubernetes Service (AKS) and Azure Arc enabled Kubernetes clusters. The logs are collected in **Log Analytics** workspace, which is a part of **Azure Monitor**.
2222

2323
## AKS cluster
2424

@@ -28,7 +28,7 @@ In AKS cluster, you can use the built-in ability to collect container logs. Foll
2828

2929
:::image type="content" source="./media/how-to-attach-kubernetes-to-workspace/aks-portal-monitor-logs.png" alt-text="Diagram illustrating how to configure Azure monitor in AKS." lightbox="./media/how-to-attach-kubernetes-to-workspace/aks-portal-monitor-logs.png":::
3030

31-
1. Click **Configure Monitoring** to enable Azure Monitor for your AKS. In the **Advanced Settings** section, you can specify an existing **Log Analytics** or create a new one for collecting logs.
31+
1. Select **Configure Monitoring** to enable Azure Monitor for your AKS. In the **Advanced Settings** section, you can specify an existing **Log Analytics** or create a new one for collecting logs.
3232

3333
:::image type="content" source="./media/how-to-attach-kubernetes-to-workspace/aks-portal-config-az-monitor.png" alt-text="Diagram illustrating how to configure container insight in AKS monitor." lightbox="./media/how-to-attach-kubernetes-to-workspace/aks-portal-config-az-monitor.png":::
3434

0 commit comments

Comments
 (0)