Skip to content

Commit 4eb1e44

Browse files
committed
update
1 parent 6596e1e commit 4eb1e44

File tree

2 files changed

+1
-3
lines changed

2 files changed

+1
-3
lines changed

articles/ai-services/openai/how-to/structured-outputs.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ services: cognitive-services
66
manager: nitinme
77
ms.service: azure-ai-openai
88
ms.topic: how-to
9-
ms.date: 12/17/2024
9+
ms.date: 12/18/2024
1010
author: mrbullwinkle
1111
ms.author: mbullwin
1212
recommendations: false

articles/ai-services/openai/whats-new.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -37,8 +37,6 @@ To learn more about the advanced `o1` series models see, [getting started with o
3737
| `o1-preview` | See the [models table](#global-standard-model-availability). |
3838
| `o1-mini` | See the [models table](#global-provisioned-managed-model-availability). |
3939

40-
41-
4240
### Preference fine-tuning (preview)
4341

4442
[Direct preference optimization (DPO)](./how-to/fine-tuning.md#direct-preference-optimization-dpo-preview) is a new alignment technique for large language models, designed to adjust model weights based on human preferences. Unlike reinforcement learning from human feedback (RLHF), DPO does not require fitting a reward model and uses simpler data (binary preferences) for training. This method is computationally lighter and faster, making it equally effective at alignment while being more efficient. DPO is especially useful in scenarios where subjective elements like tone, style, or specific content preferences are important. We’re excited to announce the public preview of DPO in Azure OpenAI Service, starting with the `gpt-4o-2024-08-06` model.

0 commit comments

Comments
 (0)