Skip to content

Commit 6ffd176

Browse files
committed
fix
1 parent 9ba7056 commit 6ffd176

File tree

1 file changed

+5
-5
lines changed

1 file changed

+5
-5
lines changed

articles/ai-studio/how-to/deploy-models-deepseek.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ ms.custom: references_regions, generated
1414
zone_pivot_groups: azure-ai-model-catalog-samples-chat
1515
---
1616

17-
# How to use DeepSeek-R1
17+
# How to use DeepSeek-R1 reasoning model
1818

1919
[!INCLUDE [Feature preview](~/reusable-content/ce-skilling/azure/includes/ai-studio/includes/feature-preview.md)]
2020

@@ -25,7 +25,7 @@ DeepSeek-R1 excels at reasoning tasks using a step-by-step training process, suc
2525

2626
::: zone pivot="programming-language-python"
2727

28-
## DeepSeek-R1
28+
## DeepSeek-R1 (preview)
2929

3030
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
3131

@@ -277,7 +277,7 @@ except HttpResponseError as ex:
277277

278278
::: zone pivot="programming-language-javascript"
279279

280-
## DeepSeek-R1
280+
## DeepSeek-R1 (preview)
281281

282282
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
283283

@@ -550,7 +550,7 @@ catch (error) {
550550
551551
::: zone pivot="programming-language-csharp"
552552
553-
## DeepSeek-R1
553+
## DeepSeek-R1 (preview)
554554
555555
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
556556
@@ -843,7 +843,7 @@ catch (RequestFailedException ex)
843843

844844
::: zone pivot="programming-language-rest"
845845

846-
## DeepSeek-R1
846+
## DeepSeek-R1 (preview)
847847

848848
DeepSeek-R1 builds on the progress of earlier reasoning-focused models that improved performance by extending Chain-of-Thought (CoT) reasoning. DeepSeek-R1 takes things further by combining reinforcement learning (RL) with fine-tuning on carefully chosen datasets. It evolved from an earlier version, DeepSeek-R1-Zero, which relied solely on RL and showed strong reasoning skills but had issues like hard-to-read outputs and language inconsistencies. To address these limitations, DeepSeek-R1 incorporates a small amount of cold-start data and follows a refined training pipeline that blends reasoning-oriented RL with supervised fine-tuning on curated datasets, resulting in a model that achieves state-of-the-art performance on reasoning benchmarks.
849849

0 commit comments

Comments
 (0)