You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/openai/includes/api-versions/latest-inference-preview.md
+8-8Lines changed: 8 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -298,7 +298,7 @@ Creates a completion for the chat message
298
298
| prediction |[PredictionContent](#predictioncontent)| Configuration for a Predicted Output, which can greatly improve response times when large parts of the model response are known ahead of time. This is most common when you are regenerating a file with only minor changes to most of the content. | No ||
299
299
| presence_penalty | number | Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.<br> | No | 0 |
300
300
| reasoning_effort | enum |**o1 models only** <br><br> Constrains effort on reasoning for <br>reasoning models.<br><br>Currently supported values are `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.<br>Possible values: `low`, `medium`, `high`| No ||
301
-
| response_format | [ResponseFormatText](#responseformattext) or [ResponseFormatJsonObject](#responseformatjsonobject) or [ResponseFormatJsonSchema](#responseformatjsonschema) | An object specifying the format that the model must output. Compatible with [GPT-4o](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models), [GPT-4o mini](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models), [GPT-4 Turbo](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models) and all [GPT-3.5](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-35) Turbo models newer than `gpt-3.5-turbo-1106`.<br><br>Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which guarantees the model will match your supplied JSON schema.<br><br>Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the message the model generates is valid JSON.<br><br>**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.<br> | No | |
301
+
| response_format | [ResponseFormatText](#responseformattext) or [ResponseFormatJsonObject](#responseformatjsonobject) or [ResponseFormatJsonSchema](#responseformatjsonschema) | An object specifying the format that the model must output. Compatible with [GPT-4o](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models), [GPT-4o mini](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models), [GPT-4 Turbo](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models) and all [GPT-3.5](../../../../ai-services/openai/concepts/models.md#gpt-35) Turbo models newer than `gpt-3.5-turbo-1106`.<br><br>Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which guarantees the model will match your supplied JSON schema.<br><br>Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the message the model generates is valid JSON.<br><br>**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.<br> | No | |
302
302
| seed | integer | This feature is in Beta.<br>If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result.<br>Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend.<br> | No ||
303
303
| stop | string or array | Up to 4 sequences where the API will stop generating further tokens.<br> | No ||
304
304
| store | boolean | Whether or not to store the output of this chat completion request for use in our model distillation or evaluation products. | No ||
@@ -4938,7 +4938,7 @@ Represents a completion response from the API. Note: both the streamed and non-s
4938
4938
| prediction |[PredictionContent](#predictioncontent)| Configuration for a Predicted Output, which can greatly improve response times when large parts of the model response are known ahead of time. This is most common when you are regenerating a file with only minor changes to most of the content. | No ||
4939
4939
| presence_penalty | number | Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.<br> | No | 0 |
4940
4940
| reasoning_effort | enum |**o1 models only** <br><br> Constrains effort on reasoning for <br>reasoning models.<br><br>Currently supported values are `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.<br>Possible values: `low`, `medium`, `high`| No ||
4941
-
| response_format | [ResponseFormatText](#responseformattext) or [ResponseFormatJsonObject](#responseformatjsonobject) or [ResponseFormatJsonSchema](#responseformatjsonschema) | An object specifying the format that the model must output. Compatible with [GPT-4o](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models), [GPT-4o mini](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models), [GPT-4 Turbo](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models) and all [GPT-3.5](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-35) Turbo models newer than `gpt-3.5-turbo-1106`.<br><br>Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which guarantees the model will match your supplied JSON schema.<br><br>Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the message the model generates is valid JSON.<br><br>**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.<br> | No | |
4941
+
| response_format | [ResponseFormatText](#responseformattext) or [ResponseFormatJsonObject](#responseformatjsonobject) or [ResponseFormatJsonSchema](#responseformatjsonschema) | An object specifying the format that the model must output. Compatible with [GPT-4o](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models), [GPT-4o mini](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models), [GPT-4 Turbo](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models) and all [GPT-3.5](../../../../ai-services/openai/concepts/models.md#gpt-35) Turbo models newer than `gpt-3.5-turbo-1106`.<br><br>Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which guarantees the model will match your supplied JSON schema.<br><br>Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the message the model generates is valid JSON.<br><br>**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.<br> | No | |
4942
4942
| seed | integer | This feature is in Beta.<br>If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result.<br>Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend.<br> | No ||
4943
4943
| stop | string or array | Up to 4 sequences where the API will stop generating further tokens.<br> | No ||
4944
4944
| store | boolean | Whether or not to store the output of this chat completion request for use in our model distillation or evaluation products. | No ||
@@ -4969,7 +4969,7 @@ User security context contains several parameters that describe the AI applicati
| description | string | A description of what the function does, used by the model to choose when and how to call the function. | No ||
4971
4971
| name | string | The name of the function to be called. Must be a-z, A-Z, 0-9, or contain underscores and dashes, with a maximum length of 64. | Yes ||
4972
-
| parameters |[FunctionParameters](#functionparameters)| The parameters the functions accepts, described as a JSON Schema object. See the guide](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. <br><br>Omitting `parameters` defines a function with an empty parameter list. | No ||
4972
+
| parameters |[FunctionParameters](#functionparameters)| The parameters the functions accepts, described as a JSON Schema object. See [the guide](../../../../ai-services/openai/how-to/function-calling.md) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. <br><br>Omitting `parameters` defines a function with an empty parameter list. | No ||
4973
4973
4974
4974
### chatCompletionFunctionCallOption
4975
4975
@@ -4982,7 +4982,7 @@ Specifying a particular function via `{"name": "my_function"}` forces the model
4982
4982
4983
4983
### chatCompletionFunctionParameters
4984
4984
4985
-
The parameters the functions accepts, described as a JSON Schema object. See the [guide/](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format.
4985
+
The parameters the functions accepts, described as a JSON Schema object. See the [guide](../../../../ai-services/openai/how-to/function-calling.md) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format.
4986
4986
4987
4987
No properties defined for this component.
4988
4988
@@ -5108,7 +5108,7 @@ This component can be one of the following:
5108
5108
| Name | Type | Description | Required | Default |
| └─ detail | enum | Specifies the detail level of the image. Learn more in the [Vision guide](https://learn.microsoft.com/azure/ai-services/openai/how-to/gpt-with-vision?tabs=rest%2Csystem-assigned%2Cresource#detail-parameter-settings-in-image-processing-low-high-auto).<br>Possible values: `auto`, `low`, `high`| No ||
5111
+
| └─ detail | enum | Specifies the detail level of the image. Learn more in the [Vision guide](../../../../ai-services/openai/how-to/gpt-with-vision.md?tabs=rest%2Csystem-assigned%2Cresource#detail-parameter-settings).<br>Possible values: `auto`, `low`, `high`| No ||
5112
5112
| └─ url | string | Either a URL of the image or the base64 encoded image data. | No ||
5113
5113
| type | enum | The type of the content part.<br>Possible values: `image_url`| Yes ||
5114
5114
@@ -5866,7 +5866,7 @@ Usage statistics for the completion request.
5866
5866
5867
5867
### FunctionParameters
5868
5868
5869
-
The parameters the functions accepts, described as a JSON Schema object. See the guide](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format.
5869
+
The parameters the functions accepts, described as a JSON Schema object. See the [guide](../../../../ai-services/openai/how-to/function-calling.md) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format.
5870
5870
5871
5871
Omitting `parameters` defines a function with an empty parameter list.
5872
5872
@@ -5879,7 +5879,7 @@ No properties defined for this component.
| description | string | A description of what the function does, used by the model to choose when and how to call the function. | No ||
5881
5881
| name | string | The name of the function to be called. Must be a-z, A-Z, 0-9, or contain underscores and dashes, with a maximum length of 64. | Yes ||
5882
-
| parameters |[FunctionParameters](#functionparameters)| The parameters the functions accepts, described as a JSON Schema object. See the guide](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. <br><br>Omitting `parameters` defines a function with an empty parameter list. | No ||
5882
+
| parameters |[FunctionParameters](#functionparameters)| The parameters the functions accepts, described as a JSON Schema object. See the [guide](../../../../ai-services/openai/how-to/function-calling.md) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. <br><br>Omitting `parameters` defines a function with an empty parameter list. | No ||
5883
5883
| strict | boolean | Whether to enable strict schema adherence when generating the function call. If set to true, the model will follow the exact schema defined in the `parameters` field. Only a subset of JSON Schema is supported when `strict` is `true`. | No | False |
5884
5884
5885
5885
### ResponseFormatText
@@ -6309,7 +6309,7 @@ Represents an `assistant` that can call the model and use tools.
6309
6309
| function | object | The function definition. | Yes ||
6310
6310
| └─ description | string | A description of what the function does, used by the model to choose when and how to call the function. | No ||
6311
6311
| └─ name | string | The name of the function to be called. Must be a-z, A-Z, 0-9, or contain underscores and dashes, with a maximum length of 64. | No ||
6312
-
| └─ parameters |[chatCompletionFunctionParameters](#chatcompletionfunctionparameters)| The parameters the functions accepts, described as a JSON Schema object. See the [guide/](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. | No ||
6312
+
| └─ parameters |[chatCompletionFunctionParameters](#chatcompletionfunctionparameters)| The parameters the functions accepts, described as a JSON Schema object. See the [guide](../../../../ai-services/openai/how-to/function-calling.md) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. | No ||
6313
6313
| type | string | The type of tool being defined: `function`| Yes ||
0 commit comments