You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/openai/concepts/customizing-llms.md
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,7 +3,7 @@ title: Azure OpenAI Service getting started with customizing a large language mo
3
3
titleSuffix: Azure OpenAI Service
4
4
description: Learn more about the concepts behind customizing an LLM with Azure OpenAI.
5
5
ms.topic: conceptual
6
-
ms.date: 03/26/2024
6
+
ms.date: 09/20/2024
7
7
ms.service: azure-ai-openai
8
8
manager: nitinme
9
9
author: mrbullwinkle
@@ -76,9 +76,9 @@ Fine-tuning requires the use of high-quality training data, in a [special exampl
76
76
77
77
### Illustrative use case
78
78
79
-
An IT department has been using GPT-4 to convert natural language queries to SQL, but they have found that the responses are not always reliably grounded in their schema, and the cost is prohibitively high.
79
+
An IT department has been using GPT-4o to convert natural language queries to SQL, but they have found that the responses are not always reliably grounded in their schema, and the cost is prohibitively high.
80
80
81
-
They fine-tune GPT-3.5-Turbo with hundreds of requests and correct responses and produce a model that performs better than the base model with lower costs and latency.
81
+
They fine-tune GPT-4o mini with hundreds of requests and correct responses and produce a model that performs better than the base model with lower costs and latency.
82
82
83
83
### Things to consider
84
84
@@ -90,13 +90,13 @@ They fine-tune GPT-3.5-Turbo with hundreds of requests and correct responses and
90
90
91
91
- Fine-tuning costs:
92
92
93
-
- Fine-tuning can reduce costs across two dimensions: (1) by using fewer tokens depending on the task (2) by using a smaller model (for example GPT 3.5 Turbo can potentially be fine-tuned to achieve the same quality of GPT-4 on a particular task).
93
+
- Fine-tuning can reduce costs across two dimensions: (1) by using fewer tokens depending on the task (2) by using a smaller model (for example GPT-4o mini can potentially be fine-tuned to achieve the same quality of GPT-4o on a particular task).
94
94
95
95
- Fine-tuning has upfront costs for training the model. And additional hourly costs for hosting the custom model once it's deployed.
96
96
97
97
### Getting started
98
98
99
99
-[When to use Azure OpenAI fine-tuning](./fine-tuning-considerations.md)
100
100
-[Customize a model with fine-tuning](../how-to/fine-tuning.md)
Copy file name to clipboardExpand all lines: articles/ai-services/openai/concepts/model-versions.md
+2-6Lines changed: 2 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI
4
4
description: Learn about model versions in Azure OpenAI.
5
5
ms.service: azure-ai-openai
6
6
ms.topic: conceptual
7
-
ms.date: 10/30/2023
7
+
ms.date: 09/20/2024
8
8
manager: nitinme
9
9
author: mrbullwinkle #ChrisHMSFT
10
10
ms.author: mbullwin #chrhoder
@@ -15,15 +15,11 @@ recommendations: false
15
15
16
16
Azure OpenAI Service is committed to providing the best generative AI models for customers. As part of this commitment, Azure OpenAI Service regularly releases new model versions to incorporate the latest features and improvements from OpenAI.
17
17
18
-
In particular, the GPT-3.5 Turbo and GPT-4 models see regular updates with new features. For example, versions 0613 of GPT-3.5 Turbo and GPT-4 introduced function calling. Function calling is a popular feature that allows the model to create structured outputs that can be used to call external tools.
19
-
20
18
## How model versions work
21
19
22
20
We want to make it easy for customers to stay up to date as models improve. Customers can choose to start with a particular version and to automatically update as new versions are released.
23
21
24
-
When a customer deploys GPT-3.5-Turbo and GPT-4 on Azure OpenAI Service, the standard behavior is to deploy the current default version – for example, GPT-4 version 0314. When the default version changes to say GPT-4 version 0613, the deployment is automatically updated to version 0613 so that customer deployments feature the latest capabilities of the model.
25
-
26
-
Customers can also deploy a specific version like GPT-4 0613 and choose an update policy, which can include the following options:
22
+
When you deploy a model you can choose an update policy, which can include the following options:
27
23
28
24
* Deployments set to **Auto-update to default** automatically update to use the new default version.
29
25
* Deployments set to **Upgrade when expired** automatically update when its current version is retired.
Copy file name to clipboardExpand all lines: articles/ai-services/openai/concepts/red-teaming.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI Service
4
4
description: Learn about how red teaming and adversarial testing are an essential practice in the responsible development of systems and features using large language models (LLMs)
Copy file name to clipboardExpand all lines: articles/ai-services/openai/concepts/system-message.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI Service
4
4
description: Learn about how to construct system messages also know as metaprompts to guide an AI system's behavior.
5
5
ms.service: azure-ai-openai
6
6
ms.topic: conceptual
7
-
ms.date: 03/26/2024
7
+
ms.date: 09/20/2024
8
8
ms.custom:
9
9
- ignite-2023
10
10
manager: nitinme
@@ -69,7 +69,7 @@ Here are some examples of lines you can include:
69
69
70
70
## Provide examples to demonstrate the intended behavior of the model
71
71
72
-
When using the system message to demonstrate the intended behavior of the model in your scenario, it is helpful to provide specific examples. When providing examples, consider the following:
72
+
When using the system message to demonstrate the intended behavior of the model in your scenario, it's helpful to provide specific examples. When providing examples, consider the following:
73
73
74
74
-**Describe difficult use cases** where the prompt is ambiguous or complicated, to give the model more visibility into how to approach such cases.
75
75
@@ -166,7 +166,7 @@ Here are some examples of lines you can include to potentially mitigate differen
166
166
167
167
Indirect attacks, also referred to as Indirect Prompt Attacks, or Cross Domain Prompt Injection Attacks, are a type of prompt injection technique where malicious instructions are hidden in the ancillary documents that are fed into Generative AI Models. We’ve found system messages to be an effective mitigation for these attacks, by way of spotlighting.
168
168
169
-
**Spotlighting** is a family of techniques that helps large language models (LLMs) distinguish between valid system instructions and potentially untrustworthy external inputs. It is based on the idea of transforming the input text in a way that makes it more salient to the model, while preserving its semantic content and task performance.
169
+
**Spotlighting** is a family of techniques that helps large language models (LLMs) distinguish between valid system instructions and potentially untrustworthy external inputs. It's based on the idea of transforming the input text in a way that makes it more salient to the model, while preserving its semantic content and task performance.
170
170
171
171
-**Delimiters** are a natural starting point to help mitigate indirect attacks. Including delimiters in your system message helps to explicitly demarcate the location of the input text in the system message. You can choose one or more special tokens to prepend and append the input text, and the model will be made aware of this boundary. By using delimiters, the model will only handle documents if they contain the appropriate delimiters, which reduces the success rate of indirect attacks. However, since delimiters can be subverted by clever adversaries, we recommend you continue on to the other spotlighting approaches.
172
172
@@ -182,7 +182,7 @@ Below is an example of a potential system message, for a retail company deployin
182
182
183
183
:::image type="content" source="../media/concepts/system-message/template.png" alt-text="Screenshot of metaprompts influencing a chatbot conversation." lightbox="../media/concepts/system-message/template.png":::
184
184
185
-
Finally, remember that system messages, or metaprompts, are not "one size fits all." Use of these type of examples has varying degrees of success in different applications. It is important to try different wording, ordering, and structure of system message text to reduce identified harms, and to test the variations to see what works best for a given scenario.
185
+
Finally, remember that system messages, or metaprompts, are not "one size fits all." Use of these type of examples has varying degrees of success in different applications. It's important to try different wording, ordering, and structure of system message text to reduce identified harms, and to test the variations to see what works best for a given scenario.
Copy file name to clipboardExpand all lines: articles/ai-services/openai/how-to/latency.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,7 +5,7 @@ description: Learn about performance and latency with Azure OpenAI
5
5
manager: nitinme
6
6
ms.service: azure-ai-openai
7
7
ms.topic: how-to
8
-
ms.date: 02/07/2024
8
+
ms.date: 09/20/2024
9
9
author: mrbullwinkle
10
10
ms.author: mbullwin
11
11
recommendations: false
@@ -52,7 +52,7 @@ There are several factors that you can control to improve per-call latency of yo
52
52
53
53
### Model selection
54
54
55
-
Latency varies based on what model you're using. For an identical request, expect that different models have different latencies for the chat completions call. If your use case requires the lowest latency models with the fastest response times, we recommend the latest models in the [GPT-3.5 Turbo model series](../concepts/models.md#gpt-35-models).
55
+
Latency varies based on what model you're using. For an identical request, expect that different models have different latencies for the chat completions call. If your use case requires the lowest latency models with the fastest response times, we recommend the latest [GPT-4o mini model](../concepts/models.md).
56
56
57
57
### Generation size and Max tokens
58
58
@@ -128,7 +128,7 @@ Time from the first token to the last token, divided by the number of generated
128
128
129
129
## Summary
130
130
131
-
***Model latency**: If model latency is important to you, we recommend trying out our latest models in the [GPT-3.5 Turbo model series](../concepts/models.md).
131
+
***Model latency**: If model latency is important to you, we recommend trying out the [GPT-4o mini model](../concepts/models.md).
132
132
133
133
***Lower max tokens**: OpenAI has found that even in cases where the total number of tokens generated is similar the request with the higher value set for the max token parameter will have more latency.
Copy file name to clipboardExpand all lines: articles/ai-services/openai/how-to/migration.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,13 +7,13 @@ ms.author: mbullwin
7
7
ms.service: azure-ai-openai
8
8
ms.custom: devx-track-python
9
9
ms.topic: how-to
10
-
ms.date: 02/26/2024
10
+
ms.date: 09/26/2024
11
11
manager: nitinme
12
12
---
13
13
14
14
# Migrating to the OpenAI Python API library 1.x
15
15
16
-
OpenAI has just released a new version of the [OpenAI Python API library](https://github.com/openai/openai-python/). This guide is supplemental to [OpenAI's migration guide](https://github.com/openai/openai-python/discussions/742) and will help bring you up to speed on the changes specific to Azure OpenAI.
16
+
OpenAI released a new version of the [OpenAI Python API library](https://github.com/openai/openai-python/). This guide is supplemental to [OpenAI's migration guide](https://github.com/openai/openai-python/discussions/742) and will help bring you up to speed on the changes specific to Azure OpenAI.
0 commit comments