Skip to content

Commit be115e9

Browse files
Merge pull request #4149 from dksimpson/us356678-dks-35
Bulk - Fix validation issues
2 parents 20af0cd + dd7e4cd commit be115e9

File tree

1 file changed

+8
-8
lines changed

1 file changed

+8
-8
lines changed

articles/ai-services/openai/includes/api-versions/latest-inference-preview.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -298,7 +298,7 @@ Creates a completion for the chat message
298298
| prediction | [PredictionContent](#predictioncontent) | Configuration for a Predicted Output, which can greatly improve response times when large parts of the model response are known ahead of time. This is most common when you are regenerating a file with only minor changes to most of the content. | No | |
299299
| presence_penalty | number | Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.<br> | No | 0 |
300300
| reasoning_effort | enum | **o1 models only** <br><br> Constrains effort on reasoning for <br>reasoning models.<br><br>Currently supported values are `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.<br>Possible values: `low`, `medium`, `high` | No | |
301-
| response_format | [ResponseFormatText](#responseformattext) or [ResponseFormatJsonObject](#responseformatjsonobject) or [ResponseFormatJsonSchema](#responseformatjsonschema) | An object specifying the format that the model must output. Compatible with [GPT-4o](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models), [GPT-4o mini](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models), [GPT-4 Turbo](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models) and all [GPT-3.5](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-35) Turbo models newer than `gpt-3.5-turbo-1106`.<br><br>Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which guarantees the model will match your supplied JSON schema.<br><br>Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the message the model generates is valid JSON.<br><br>**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.<br> | No | |
301+
| response_format | [ResponseFormatText](#responseformattext) or [ResponseFormatJsonObject](#responseformatjsonobject) or [ResponseFormatJsonSchema](#responseformatjsonschema) | An object specifying the format that the model must output. Compatible with [GPT-4o](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models), [GPT-4o mini](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models), [GPT-4 Turbo](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models) and all [GPT-3.5](../../../../ai-services/openai/concepts/models.md#gpt-35) Turbo models newer than `gpt-3.5-turbo-1106`.<br><br>Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which guarantees the model will match your supplied JSON schema.<br><br>Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the message the model generates is valid JSON.<br><br>**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.<br> | No | |
302302
| seed | integer | This feature is in Beta.<br>If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result.<br>Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend.<br> | No | |
303303
| stop | string or array | Up to 4 sequences where the API will stop generating further tokens.<br> | No | |
304304
| store | boolean | Whether or not to store the output of this chat completion request for use in our model distillation or evaluation products. | No | |
@@ -4938,7 +4938,7 @@ Represents a completion response from the API. Note: both the streamed and non-s
49384938
| prediction | [PredictionContent](#predictioncontent) | Configuration for a Predicted Output, which can greatly improve response times when large parts of the model response are known ahead of time. This is most common when you are regenerating a file with only minor changes to most of the content. | No | |
49394939
| presence_penalty | number | Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.<br> | No | 0 |
49404940
| reasoning_effort | enum | **o1 models only** <br><br> Constrains effort on reasoning for <br>reasoning models.<br><br>Currently supported values are `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.<br>Possible values: `low`, `medium`, `high` | No | |
4941-
| response_format | [ResponseFormatText](#responseformattext) or [ResponseFormatJsonObject](#responseformatjsonobject) or [ResponseFormatJsonSchema](#responseformatjsonschema) | An object specifying the format that the model must output. Compatible with [GPT-4o](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models), [GPT-4o mini](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models), [GPT-4 Turbo](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-4-and-gpt-4-turbo-models) and all [GPT-3.5](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#gpt-35) Turbo models newer than `gpt-3.5-turbo-1106`.<br><br>Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which guarantees the model will match your supplied JSON schema.<br><br>Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the message the model generates is valid JSON.<br><br>**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.<br> | No | |
4941+
| response_format | [ResponseFormatText](#responseformattext) or [ResponseFormatJsonObject](#responseformatjsonobject) or [ResponseFormatJsonSchema](#responseformatjsonschema) | An object specifying the format that the model must output. Compatible with [GPT-4o](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models), [GPT-4o mini](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models), [GPT-4 Turbo](../../../../ai-services/openai/concepts/models.md#gpt-4-and-gpt-4-turbo-models) and all [GPT-3.5](../../../../ai-services/openai/concepts/models.md#gpt-35) Turbo models newer than `gpt-3.5-turbo-1106`.<br><br>Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which guarantees the model will match your supplied JSON schema.<br><br>Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the message the model generates is valid JSON.<br><br>**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.<br> | No | |
49424942
| seed | integer | This feature is in Beta.<br>If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result.<br>Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend.<br> | No | |
49434943
| stop | string or array | Up to 4 sequences where the API will stop generating further tokens.<br> | No | |
49444944
| store | boolean | Whether or not to store the output of this chat completion request for use in our model distillation or evaluation products. | No | |
@@ -4969,7 +4969,7 @@ User security context contains several parameters that describe the AI applicati
49694969
|------|------|-------------|----------|---------|
49704970
| description | string | A description of what the function does, used by the model to choose when and how to call the function. | No | |
49714971
| name | string | The name of the function to be called. Must be a-z, A-Z, 0-9, or contain underscores and dashes, with a maximum length of 64. | Yes | |
4972-
| parameters | [FunctionParameters](#functionparameters) | The parameters the functions accepts, described as a JSON Schema object. See the guide](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. <br><br>Omitting `parameters` defines a function with an empty parameter list. | No | |
4972+
| parameters | [FunctionParameters](#functionparameters) | The parameters the functions accepts, described as a JSON Schema object. See [the guide](../../../../ai-services/openai/how-to/function-calling.md) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. <br><br>Omitting `parameters` defines a function with an empty parameter list. | No | |
49734973

49744974
### chatCompletionFunctionCallOption
49754975

@@ -4982,7 +4982,7 @@ Specifying a particular function via `{"name": "my_function"}` forces the model
49824982

49834983
### chatCompletionFunctionParameters
49844984

4985-
The parameters the functions accepts, described as a JSON Schema object. See the [guide/](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format.
4985+
The parameters the functions accepts, described as a JSON Schema object. See the [guide](../../../../ai-services/openai/how-to/function-calling.md) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format.
49864986

49874987
No properties defined for this component.
49884988

@@ -5108,7 +5108,7 @@ This component can be one of the following:
51085108
| Name | Type | Description | Required | Default |
51095109
|------|------|-------------|----------|---------|
51105110
| image_url | object | | Yes | |
5111-
| └─ detail | enum | Specifies the detail level of the image. Learn more in the [Vision guide](https://learn.microsoft.com/azure/ai-services/openai/how-to/gpt-with-vision?tabs=rest%2Csystem-assigned%2Cresource#detail-parameter-settings-in-image-processing-low-high-auto).<br>Possible values: `auto`, `low`, `high` | No | |
5111+
| └─ detail | enum | Specifies the detail level of the image. Learn more in the [Vision guide](../../../../ai-services/openai/how-to/gpt-with-vision.md?tabs=rest%2Csystem-assigned%2Cresource#detail-parameter-settings).<br>Possible values: `auto`, `low`, `high` | No | |
51125112
| └─ url | string | Either a URL of the image or the base64 encoded image data. | No | |
51135113
| type | enum | The type of the content part.<br>Possible values: `image_url` | Yes | |
51145114

@@ -5866,7 +5866,7 @@ Usage statistics for the completion request.
58665866

58675867
### FunctionParameters
58685868

5869-
The parameters the functions accepts, described as a JSON Schema object. See the guide](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format.
5869+
The parameters the functions accepts, described as a JSON Schema object. See the [guide](../../../../ai-services/openai/how-to/function-calling.md) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format.
58705870

58715871
Omitting `parameters` defines a function with an empty parameter list.
58725872

@@ -5879,7 +5879,7 @@ No properties defined for this component.
58795879
|------|------|-------------|----------|---------|
58805880
| description | string | A description of what the function does, used by the model to choose when and how to call the function. | No | |
58815881
| name | string | The name of the function to be called. Must be a-z, A-Z, 0-9, or contain underscores and dashes, with a maximum length of 64. | Yes | |
5882-
| parameters | [FunctionParameters](#functionparameters) | The parameters the functions accepts, described as a JSON Schema object. See the guide](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. <br><br>Omitting `parameters` defines a function with an empty parameter list. | No | |
5882+
| parameters | [FunctionParameters](#functionparameters) | The parameters the functions accepts, described as a JSON Schema object. See the [guide](../../../../ai-services/openai/how-to/function-calling.md) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. <br><br>Omitting `parameters` defines a function with an empty parameter list. | No | |
58835883
| strict | boolean | Whether to enable strict schema adherence when generating the function call. If set to true, the model will follow the exact schema defined in the `parameters` field. Only a subset of JSON Schema is supported when `strict` is `true`. | No | False |
58845884

58855885
### ResponseFormatText
@@ -6309,7 +6309,7 @@ Represents an `assistant` that can call the model and use tools.
63096309
| function | object | The function definition. | Yes | |
63106310
| └─ description | string | A description of what the function does, used by the model to choose when and how to call the function. | No | |
63116311
| └─ name | string | The name of the function to be called. Must be a-z, A-Z, 0-9, or contain underscores and dashes, with a maximum length of 64. | No | |
6312-
| └─ parameters | [chatCompletionFunctionParameters](#chatcompletionfunctionparameters) | The parameters the functions accepts, described as a JSON Schema object. See the [guide/](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. | No | |
6312+
| └─ parameters | [chatCompletionFunctionParameters](#chatcompletionfunctionparameters) | The parameters the functions accepts, described as a JSON Schema object. See the [guide](../../../../ai-services/openai/how-to/function-calling.md) for examples, and the [JSON Schema reference](https://json-schema.org/understanding-json-schema/) for documentation about the format. | No | |
63136313
| type | string | The type of tool being defined: `function` | Yes | |
63146314

63156315

0 commit comments

Comments
 (0)