You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/foundry-models/includes/use-chat-reasoning/about-reasoning.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -9,7 +9,7 @@ author: santiagxf
9
9
10
10
## Reasoning models
11
11
12
-
Reasoning models can reach higher levels of performance in domains like math, coding, science, strategy, and logistics. The way these models produce outputs is by explicitly using chain of thought to explore all possible paths before generating an answer. They verify their answers as they produce them, which helps to arrive at better, more accurate conclusions. As a result, reasoning models might require less context in prompting in order to produce effective results.
12
+
Reasoning models can reach higher levels of performance in domains like math, coding, science, strategy, and logistics. The way these models produce outputs is by explicitly using chain of thought to explore all possible paths before generating an answer. They verify their answers as they produce them, which helps to arrive at more accurate conclusions. As a result, reasoning models might require less context in prompting in order to produce effective results.
13
13
14
14
This way of scaling a model's performance is referred to as *inference compute time* as it trades performance against higher latency and cost. In contrast, other approaches might scale through *training compute time*.
15
15
@@ -18,4 +18,4 @@ Reasoning models produce two types of content as outputs:
18
18
* Reasoning completions
19
19
* Output completions
20
20
21
-
Both of these completions count towards content generated from the model. Therefore, they contribute to the token limits and costs associated with the model. Some models, like `DeepSeek-R1`, might respond with the reasoning content. Others, like `o1`, respond only with the output completions.
21
+
Both of these completions count towards content generated from the model. Therefore, they contribute to the token limits and costs associated with the model. Some models, like `DeepSeek-R1`, might respond with the reasoning content. Others, like `o1`, only output the completions.
Copy file name to clipboardExpand all lines: articles/ai-foundry/foundry-models/tutorials/get-started-deepseek-r1.md
+10-9Lines changed: 10 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -38,6 +38,8 @@ Foundry Models is a capability in Azure AI Foundry resources in Azure. You can c
38
38
To create an Azure AI project that supports deployment for DeepSeek-R1, follow these steps. You can also create the resources, using [Azure CLI](../how-to/quickstart-create-resources.md?pivots=programming-language-cli) or [infrastructure as code, with Bicep](../how-to/quickstart-create-resources.md?pivots=programming-language-bicep).
1. Sign in to [Azure AI Foundry portal](https://ai.azure.com/?cid=learnDocs).
42
44
43
45
1. Go to the preview features icon on the header of the landing page and make sure that the **Deploy models to Azure AI Foundry resources** feature is turned on.
@@ -49,7 +51,7 @@ To create an Azure AI project that supports deployment for DeepSeek-R1, follow t
49
51
50
52
1. Search for the **DeepSeek-R1** model and open its model card.
51
53
52
-
1. Select **Use this model**. This opens up a wizard to create an Azure AI Foundry project and resources that you'll work in. You can keep the default name for the project or change it.
54
+
1. Select **Use this model**. This action opens up a wizard to create an Azure AI Foundry project and resources that you'll work in. You can keep the default name for the project or change it.
53
55
54
56
> [!TIP]
55
57
> **Are you using Azure OpenAI in Azure AI Foundry Models?** When you're connected to Azure AI Foundry portal using an Azure OpenAI resource, only Azure OpenAI models show up in the catalog. To view the full list of models, including DeepSeek-R1, use the top **Announcements** section and locate the card with the option **Explore more models**.
@@ -64,24 +66,23 @@ To create an Azure AI project that supports deployment for DeepSeek-R1, follow t
64
66
| -------------- | ----------- |
65
67
| Resource group | The main container for all the resources in Azure. This helps get resources that work together organized. It also helps to have a scope for the costs associated with the entire project. |
66
68
| Region | The region of the resources that you're creating. |
67
-
| AI Foundry resource | The resource enabling access to the flagship models in Azure AI model catalog. In this tutorial, a new account is created, but Azure AI Foundry resources (formerly known as Azure AI Services) can be shared across multiple hubs and projects. Hubs use a connection to the resource to have access to the model deployments available there. To learn how you can create connections to Azure AI Foundry resources to consume models you can read[Connect your AI project](../../model-inference/how-to/configure-project-connection.md). |
69
+
| AI Foundry resource | The resource enabling access to the flagship models in Azure AI model catalog. In this tutorial, a new account is created, but Azure AI Foundry resources (formerly known as Azure AI Services) can be shared across multiple hubs and projects. Hubs use a connection to the resource to have access to the model deployments available there. To learn how you can create connections to Azure AI Foundry resources to consume models, see[Connect your AI project](../../model-inference/how-to/configure-project-connection.md). |
68
70
69
71
1. Select **Create** to create the Foundry project alongside the other defaults. Wait until the project creation is complete. This process takes a few minutes.
70
72
71
73
## Deploy the model
72
74
73
75
1. Once the project and resources are created, a deployment wizard appears. DeepSeek-R1 is offered as a Microsoft first party consumption service. You can review our privacy and security commitments under [Data, privacy, and Security](../../../ai-studio/how-to/concept-data-privacy.md).
74
76
75
-
> [!TIP]
76
-
> Review the pricing details for the model by selecting the [Pricing and terms tab](https://aka.ms/DeepSeekPricing).
77
+
1. Review the pricing details for the model by selecting the [Pricing and terms tab](https://aka.ms/DeepSeekPricing).
77
78
78
79
1. Select **Agree and Proceed** to continue with the deployment.
79
80
80
81
1. You can configure the deployment settings at this time. By default, the deployment receives the name of the model you're deploying. The deployment name is used in the `model` parameter for requests to route to this particular model deployment. This allows you to also configure specific names for your models when you attach specific configurations.
81
82
82
-
1. Azure AI Foundry automatically selects the Foundry resource created earlier with your project. Use the **Customize** option to change the connection based on your needs. DeepSeek-R1 is currently offered under the **Global Standard** deployment type which offers higher throughput and performance.
83
+
1. Azure AI Foundry automatically selects the Foundry resource created earlier with your project. Use the **Customize** option to change the connection based on your needs. DeepSeek-R1 is currently offered under the **Global Standard** deployment type, which offers higher throughput and performance.
83
84
84
-
:::image type="content" source="../media/quickstart-get-started-deepseek-r1/model-deploy.png" alt-text="Screenshot showing how to deploy the model." lightbox="../media/quickstart-get-started-deepseek-r1/model-deploy.png":::
85
+
:::image type="content" source="../media/quickstart-get-started-deepseek-r1/deployment-wizard.png" alt-text="Screenshot showing how to deploy the model." lightbox="../media/quickstart-get-started-deepseek-r1/deployment-wizard.png":::
85
86
86
87
1. Select **Deploy**.
87
88
@@ -96,7 +97,7 @@ You can get started by using the model in the playground to have an idea of the
96
97
97
98
2. In the **Deployment** drop down, the deployment you created is already automatically selected.
98
99
99
-
3. Configure the system prompt as needed. In general, reasoning models don't use system messages in the same way that other types of models.
100
+
3. Configure the system prompt as needed. In general, reasoning models don't use system messages in the same way as other types of models.
100
101
101
102
:::image type="content" source="../media/quickstart-get-started-deepseek-r1/playground-chat-models.png" alt-text="Screenshot showing how to select a model deployment to use in playground, configure the system message, and test it out." lightbox="../media/quickstart-get-started-deepseek-r1/playground-chat-models.png":::
102
103
@@ -119,11 +120,11 @@ You can use the Azure AI Model Inference package to consume the model in code:
Reasoning may generate longer responses and consume a larger number of tokens. You can see the [rate limits](../../model-inference/quotas-limits.md) that apply to DeepSeek-R1 models. Consider having a retry strategy to handle rate limits being applied. You can also [request increases to the default limits](../quotas-limits.md#request-increases-to-the-default-limits).
123
+
Reasoning might generate longer responses and consume a larger number of tokens. You can see the [rate limits](../../model-inference/quotas-limits.md) that apply to DeepSeek-R1 models. Consider having a retry strategy to handle rate limits being applied. You can also [request increases to the default limits](../quotas-limits.md#request-increases-to-the-default-limits).
123
124
124
125
### Reasoning content
125
126
126
-
Some reasoning models, like DeepSeek-R1, generate completions and include the reasoning behind it. The reasoning associated with the completion is included in the response's content within the tags `<think>` and `</think>`. The model may select on which scenarios to generate reasoning content. You following example shows how to do it in Python:
127
+
Some reasoning models, like DeepSeek-R1, generate completions and include the reasoning behind it. The reasoning associated with the completion is included in the response's content within the tags `<think>` and `</think>`. The model might select which scenarios for which to generate reasoning content. The following example shows how to generate the reasoning content, using Python:
0 commit comments