You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/openai/concepts/models.md
+7-5Lines changed: 7 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -102,7 +102,7 @@ GPT-4 version 0125-preview is an updated version of the GPT-4 Turbo preview prev
102
102
103
103
> [!IMPORTANT]
104
104
>
105
-
> -`gpt-4` version 0125-preview replaces version 1106-preview. Deployments of `gpt-4` version 1106-preview set to "Auto-update to default" and "Upgrade when expired" will start to be upgraded on March 8th, 2024 and will complete upgrades within 2 weeks. Deployments of `gpt-4` version 1106-preview set to "No autoupgrade" will stop working starting February 20, 2024. If you have a deployment of `gpt-4` version 1106-preview, you can test version `0125-preview` in the available regions below.
105
+
> -`gpt-4` version 0125-preview replaces version 1106-preview. Deployments of `gpt-4` version 1106-preview set to "Auto-update to default" and "Upgrade when expired" will start to be upgraded on March 8th, 2024 and will complete upgrades within 2 weeks. Deployments of `gpt-4` version 1106-preview set to "No autoupgrade" will stop working starting March 8th, 2024. If you have a deployment of `gpt-4` version 1106-preview, you can test version `0125-preview` in the available regions below.
106
106
107
107
| Model ID | Max Request (tokens) | Training Data (up to) |
108
108
| --- | :--- | :---: |
@@ -111,7 +111,7 @@ GPT-4 version 0125-preview is an updated version of the GPT-4 Turbo preview prev
**<sup>1</sup>** GPT-4 Turbo Preview = `gpt-4` (0125-preview). To deploy this model, under **Deployments** select model **gpt-4**. For **Model version** select **0125-preview**.
@@ -147,6 +147,9 @@ The following GPT-4 models are available with [Azure Government](/azure/azure-go
147
147
148
148
### GPT-3.5 models
149
149
150
+
> [!IMPORTANT]
151
+
> The NEW `gpt-35-turbo (0125)` model has various improvements, including higher accuracy at responding in requested formats and a fix for a bug which caused a text encoding issue for non-English language function calls.
152
+
150
153
GPT-3.5 Turbo is used with the Chat Completion API. GPT-3.5 Turbo version 0301 can also be used with the Completions API. GPT-3.5 Turbo versions 0613 and 1106 only support the Chat Completions API.
151
154
152
155
GPT-3.5 Turbo version 0301 is the first version of the model released. Version 0613 is the second version of the model and adds function calling support.
@@ -158,6 +161,7 @@ See [model versions](../concepts/model-versions.md) to learn about how Azure Ope
158
161
159
162
### GPT-3.5-Turbo model availability
160
163
164
+
161
165
#### Public cloud regions
162
166
163
167
| Model ID | Model Availability | Max Request (tokens) | Training Data (up to) |
@@ -167,12 +171,10 @@ See [model versions](../concepts/model-versions.md) to learn about how Azure Ope
167
171
|`gpt-35-turbo-16k` (0613) | Australia East <br> Canada East <br> East US <br> East US 2 <br> France Central <br> Japan East <br> North Central US <br> Sweden Central <br> Switzerland North<br> UK South | 16,384 | Sep 2021 |
168
172
|`gpt-35-turbo-instruct` (0914) | East US <br> Sweden Central | 4,097 |Sep 2021 |
169
173
|`gpt-35-turbo` (1106) | Australia East <br> Canada East <br> France Central <br> South India <br> Sweden Central<br> UK South <br> West US | Input: 16,385<br> Output: 4,096 | Sep 2021|
170
-
|`gpt-35-turbo`**<sup>2</sup>** (0125) | Canada East <br> North Central US <br> South Central US | 16,385 | Sep 2021 |
174
+
|`gpt-35-turbo` (0125)**NEW**| Canada East <br> North Central US <br> South Central US | 16,385 | Sep 2021 |
171
175
172
176
**<sup>1</sup>** This model will accept requests > 4,096 tokens. It is not recommended to exceed the 4,096 input token limit as the newer version of the model are capped at 4,096 tokens. If you encounter issues when exceeding 4,096 input tokens with this model this configuration is not officially supported.
173
177
174
-
**<sup>2</sup>** This model has various improvements, including higher accuracy at responding in requested formats and a fix for a bug which caused a text encoding issue for non-English language function calls.
175
-
176
178
### Embeddings models
177
179
178
180
These models can only be used with Embedding API requests.
0 commit comments