You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-studio/how-to/deploy-models-deepseek.md
+16-17Lines changed: 16 additions & 17 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
1
---
2
-
title: How to use DeepSeek-R1 with Azure AI Foundry
2
+
title: How to use DeepSeek-R1 reasoning model with Azure AI Foundry
3
3
titleSuffix: Azure AI Foundry
4
-
description: Learn how to use DeepSeek-R1 with Azure AI Foundry.
4
+
description: Learn how to use DeepSeek-R1 reasoning model with Azure AI Foundry.
5
5
manager: scottpolly
6
6
author: msakande
7
7
reviewer: santiagxf
@@ -25,7 +25,7 @@ DeepSeek-R1 excels at reasoning tasks using a step-by-step training process, suc
25
25
26
26
::: zone pivot="programming-language-python"
27
27
28
-
## DeepSeek-R1 (preview)
28
+
## DeepSeek-R1
29
29
30
30
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
31
31
@@ -86,7 +86,6 @@ from azure.core.credentials import AzureKeyCredential
You can extract the reasoning content from the response to understand the model's thought process as follows:
170
169
171
170
172
171
```python
@@ -277,7 +276,7 @@ except HttpResponseError as ex:
277
276
278
277
::: zone pivot="programming-language-javascript"
279
278
280
-
## DeepSeek-R1 (preview)
279
+
## DeepSeek-R1
281
280
282
281
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
@@ -425,7 +424,7 @@ var response = await client.path("/chat/completions").post({
425
424
});
426
425
```
427
426
428
-
The response is as follows:
427
+
You can extract the reasoning content from the response to understand the model's thought process as follows:
429
428
430
429
431
430
```javascript
@@ -550,7 +549,7 @@ catch (error) {
550
549
551
550
::: zone pivot="programming-language-csharp"
552
551
553
-
## DeepSeek-R1 (preview)
552
+
## DeepSeek-R1
554
553
555
554
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
849
848
@@ -904,7 +903,7 @@ The response is as follows:
904
903
{
905
904
"model_name": "DeepSeek-R1",
906
905
"model_type": "chat-completions",
907
-
"model_provider_name": "DeepSeek"
906
+
"model_provider_name": "Deepseek"
908
907
}
909
908
```
910
909
@@ -980,7 +979,7 @@ Some reasoning models, like DeepSeek-R1, generate completions and include the re
980
979
}
981
980
```
982
981
983
-
The response is as follows:
982
+
You can extract the reasoning content from the response to understand the model's thought process as follows:
984
983
985
984
986
985
```json
@@ -1128,7 +1127,7 @@ The following example shows how to handle events when the model detects harmful
1128
1127
1129
1128
## More inference examples
1130
1129
1131
-
For more examples of how to use DeepSeek models, see the following examples and tutorials:
1130
+
For more examples of how to use Deepseek models, see the following examples and tutorials:
## Cost and quota considerations forDeepSeek models deployed as serverless API endpoints
1139
+
## Cost and quota considerations forDeepseek models deployed as serverless API endpoints
1141
1140
1142
1141
Quota is managed per deployment. Each deployment has a rate limit of200,000 tokens per minute and 1,000API requests per minute. However, we currently limit one deployment per model per project. Contact Microsoft Azure Support if the current rate limits aren't sufficient for your scenarios.
0 commit comments