Skip to content

Commit 84e3ddb

Browse files
committed
fix
1 parent 6ffd176 commit 84e3ddb

File tree

1 file changed

+16
-17
lines changed

1 file changed

+16
-17
lines changed

articles/ai-studio/how-to/deploy-models-deepseek.md

Lines changed: 16 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
2-
title: How to use DeepSeek-R1 with Azure AI Foundry
2+
title: How to use DeepSeek-R1 reasoning model with Azure AI Foundry
33
titleSuffix: Azure AI Foundry
4-
description: Learn how to use DeepSeek-R1 with Azure AI Foundry.
4+
description: Learn how to use DeepSeek-R1 reasoning model with Azure AI Foundry.
55
manager: scottpolly
66
author: msakande
77
reviewer: santiagxf
@@ -25,7 +25,7 @@ DeepSeek-R1 excels at reasoning tasks using a step-by-step training process, suc
2525

2626
::: zone pivot="programming-language-python"
2727

28-
## DeepSeek-R1 (preview)
28+
## DeepSeek-R1
2929

3030
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
3131

@@ -86,7 +86,6 @@ from azure.core.credentials import AzureKeyCredential
8686
client = ChatCompletionsClient(
8787
endpoint=os.environ["AZURE_INFERENCE_ENDPOINT"],
8888
credential=AzureKeyCredential(os.environ["AZURE_INFERENCE_CREDENTIAL"]),
89-
model="DeepSeek-R1"
9089
)
9190
```
9291

@@ -111,7 +110,7 @@ print("Model provider name:", model_info.model_provider_name)
111110
```console
112111
Model name: DeepSeek-R1
113112
Model type: chat-completions
114-
Model provider name: DeepSeek
113+
Model provider name: Deepseek
115114
```
116115

117116
### Create a chat completion request
@@ -166,7 +165,7 @@ response = client.complete(
166165
)
167166
```
168167

169-
The response is as follows:
168+
You can extract the reasoning content from the response to understand the model's thought process as follows:
170169

171170

172171
```python
@@ -277,7 +276,7 @@ except HttpResponseError as ex:
277276

278277
::: zone pivot="programming-language-javascript"
279278

280-
## DeepSeek-R1 (preview)
279+
## DeepSeek-R1
281280

282281
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
283282

@@ -360,7 +359,7 @@ console.log("Model provider name: ", model_info.body.model_provider_name)
360359
```console
361360
Model name: DeepSeek-R1
362361
Model type: chat-completions
363-
Model provider name: DeepSeek
362+
Model provider name: Deepseek
364363
```
365364

366365
### Create a chat completion request
@@ -425,7 +424,7 @@ var response = await client.path("/chat/completions").post({
425424
});
426425
```
427426

428-
The response is as follows:
427+
You can extract the reasoning content from the response to understand the model's thought process as follows:
429428

430429

431430
```javascript
@@ -550,7 +549,7 @@ catch (error) {
550549
551550
::: zone pivot="programming-language-csharp"
552551
553-
## DeepSeek-R1 (preview)
552+
## DeepSeek-R1
554553
555554
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
556555
@@ -653,7 +652,7 @@ Console.WriteLine($"Model provider name: {modelInfo.Value.ModelProviderName}");
653652
```console
654653
Model name: DeepSeek-R1
655654
Model type: chat-completions
656-
Model provider name: DeepSeek
655+
Model provider name: Deepseek
657656
```
658657

659658
### Create a chat completion request
@@ -712,7 +711,7 @@ ChatCompletionsOptions requestOptions = new ChatCompletionsOptions()
712711
Response<ChatCompletions> response = client.Complete(requestOptions);
713712
```
714713

715-
The response is as follows:
714+
You can extract the reasoning content from the response to understand the model's thought process as follows:
716715
717716
718717
```csharp
@@ -843,7 +842,7 @@ catch (RequestFailedException ex)
843842
844843
::: zone pivot="programming-language-rest"
845844
846-
## DeepSeek-R1 (preview)
845+
## DeepSeek-R1
847846
848847
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
849848
@@ -904,7 +903,7 @@ The response is as follows:
904903
{
905904
"model_name": "DeepSeek-R1",
906905
"model_type": "chat-completions",
907-
"model_provider_name": "DeepSeek"
906+
"model_provider_name": "Deepseek"
908907
}
909908
```
910909
@@ -980,7 +979,7 @@ Some reasoning models, like DeepSeek-R1, generate completions and include the re
980979
}
981980
```
982981
983-
The response is as follows:
982+
You can extract the reasoning content from the response to understand the model's thought process as follows:
984983

985984

986985
```json
@@ -1128,7 +1127,7 @@ The following example shows how to handle events when the model detects harmful
11281127

11291128
## More inference examples
11301129

1131-
For more examples of how to use DeepSeek models, see the following examples and tutorials:
1130+
For more examples of how to use Deepseek models, see the following examples and tutorials:
11321131

11331132
| Description | Language | Sample |
11341133
|-------------------------------------------|-------------------|-----------------------------------------------------------------|
@@ -1137,7 +1136,7 @@ For more examples of how to use DeepSeek models, see the following examples and
11371136
| Azure AI Inference package for C# | C# | [Link](https://aka.ms/azsdk/azure-ai-inference/csharp/samples) |
11381137
| Azure AI Inference package for Java | Java | [Link](https://github.com/Azure/azure-sdk-for-java/tree/main/sdk/ai/azure-ai-inference/src/samples) |
11391138

1140-
## Cost and quota considerations for DeepSeek models deployed as serverless API endpoints
1139+
## Cost and quota considerations for Deepseek models deployed as serverless API endpoints
11411140

11421141
Quota is managed per deployment. Each deployment has a rate limit of 200,000 tokens per minute and 1,000 API requests per minute. However, we currently limit one deployment per model per project. Contact Microsoft Azure Support if the current rate limits aren't sufficient for your scenarios.
11431142

0 commit comments

Comments
 (0)