You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"instructions": "You are a helpful assistant designed to make me better at coding!",
2889
2889
"tools": [],
2890
2890
"file_ids": [],
2891
-
"metadata": {}
2891
+
"metadata": {},
2892
+
"top_p": 1.0,
2893
+
"temperature": 1.0,
2894
+
"response_format": "auto"
2892
2895
},
2893
2896
{
2894
2897
"id": "asst_abc456",
@@ -2900,7 +2903,10 @@ paths:
2900
2903
"instructions": "You are a helpful assistant designed to make me better at coding!",
2901
2904
"tools": [],
2902
2905
"file_ids": [],
2903
-
"metadata": {}
2906
+
"metadata": {},
2907
+
"top_p": 1.0,
2908
+
"temperature": 1.0,
2909
+
"response_format": "auto"
2904
2910
},
2905
2911
{
2906
2912
"id": "asst_abc789",
@@ -2912,7 +2918,10 @@ paths:
2912
2918
"instructions": null,
2913
2919
"tools": [],
2914
2920
"file_ids": [],
2915
-
"metadata": {}
2921
+
"metadata": {},
2922
+
"top_p": 1.0,
2923
+
"temperature": 1.0,
2924
+
"response_format": "auto"
2916
2925
}
2917
2926
],
2918
2927
"first_id": "asst_abc123",
@@ -3001,7 +3010,10 @@ paths:
3001
3010
}
3002
3011
],
3003
3012
"file_ids": [],
3004
-
"metadata": {}
3013
+
"metadata": {},
3014
+
"top_p": 1.0,
3015
+
"temperature": 1.0,
3016
+
"response_format": "auto"
3005
3017
}
3006
3018
- title: Files
3007
3019
request:
@@ -3064,7 +3076,10 @@ paths:
3064
3076
"file_ids": [
3065
3077
"file-abc123"
3066
3078
],
3067
-
"metadata": {}
3079
+
"metadata": {},
3080
+
"top_p": 1.0,
3081
+
"temperature": 1.0,
3082
+
"response_format": "auto"
3068
3083
}
3069
3084
3070
3085
/assistants/{assistant_id}:
@@ -3238,7 +3253,10 @@ paths:
3238
3253
"file-abc123",
3239
3254
"file-abc456"
3240
3255
],
3241
-
"metadata": {}
3256
+
"metadata": {},
3257
+
"top_p": 1.0,
3258
+
"temperature": 1.0,
3259
+
"response_format": "auto"
3242
3260
}
3243
3261
delete:
3244
3262
operationId: deleteAssistant
@@ -4501,7 +4519,8 @@ paths:
4501
4519
"completion_tokens": 456,
4502
4520
"total_tokens": 579
4503
4521
},
4504
-
"temperature": 1,
4522
+
"temperature": 1.0,
4523
+
"top_p": 1.0,
4505
4524
"max_prompt_tokens": 1000,
4506
4525
"max_completion_tokens": 1000,
4507
4526
"truncation_strategy": {
@@ -4542,7 +4561,8 @@ paths:
4542
4561
"completion_tokens": 456,
4543
4562
"total_tokens": 579
4544
4563
},
4545
-
"temperature": 1,
4564
+
"temperature": 1.0,
4565
+
"top_p": 1.0,
4546
4566
"max_prompt_tokens": 1000,
4547
4567
"max_completion_tokens": 1000,
4548
4568
"truncation_strategy": {
@@ -4651,7 +4671,8 @@ paths:
4651
4671
],
4652
4672
"metadata": {},
4653
4673
"usage": null,
4654
-
"temperature": 1,
4674
+
"temperature": 1.0,
4675
+
"top_p": 1.0,
4655
4676
"max_prompt_tokens": 1000,
4656
4677
"max_completion_tokens": 1000,
4657
4678
"truncation_strategy": {
@@ -4996,7 +5017,8 @@ paths:
4996
5017
"completion_tokens": 456,
4997
5018
"total_tokens": 579
4998
5019
},
4999
-
"temperature": 1,
5020
+
"temperature": 1.0,
5021
+
"top_p": 1.0,
5000
5022
"max_prompt_tokens": 1000,
5001
5023
"max_completion_tokens": 1000,
5002
5024
"truncation_strategy": {
@@ -5119,7 +5141,8 @@ paths:
5119
5141
"completion_tokens": 456,
5120
5142
"total_tokens": 579
5121
5143
},
5122
-
"temperature": 1,
5144
+
"temperature": 1.0,
5145
+
"top_p": 1.0,
5123
5146
"max_prompt_tokens": 1000,
5124
5147
"max_completion_tokens": 1000,
5125
5148
"truncation_strategy": {
@@ -5266,7 +5289,8 @@ paths:
5266
5289
"file_ids": [],
5267
5290
"metadata": {},
5268
5291
"usage": null,
5269
-
"temperature": 1,
5292
+
"temperature": 1.0,
5293
+
"top_p": 1.0,
5270
5294
"max_prompt_tokens": 1000,
5271
5295
"max_completion_tokens": 1000,
5272
5296
"truncation_strategy": {
@@ -5474,7 +5498,8 @@ paths:
5474
5498
"file_ids": [],
5475
5499
"metadata": {},
5476
5500
"usage": null,
5477
-
"temperature": 1
5501
+
"temperature": 1.0,
5502
+
"top_p": 1.0,
5478
5503
}
5479
5504
5480
5505
/threads/{thread_id}/runs/{run_id}/steps:
@@ -7775,7 +7800,7 @@ components:
7775
7800
description: |
7776
7801
The ID of an uploaded file that contains training data.
7777
7802
7778
-
See [upload file](/docs/api-reference/files/upload) for how to upload a file.
7803
+
See [upload file](/docs/api-reference/files/create) for how to upload a file.
7779
7804
7780
7805
Your dataset must be formatted as a JSONL file. Additionally, you must upload your file with the purpose `fine-tune`.
7781
7806
@@ -8770,6 +8795,33 @@ components:
8770
8795
- total_tokens
8771
8796
nullable: true
8772
8797
8798
+
AssistantsApiResponseFormatOption:
8799
+
description: |
8800
+
Specifies the format that the model must output. Compatible with [GPT-4 Turbo](/docs/models/gpt-4-and-gpt-4-turbo) and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
8801
+
8802
+
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the message the model generates is valid JSON.
8803
+
8804
+
**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.
An object describing the expected output of the model. If `json_object` only `function` type `tools` are allowed to be passed to the Run. If `text` the model can return text or any value needed.
8817
+
properties:
8818
+
type:
8819
+
type: string
8820
+
enum: ["text", "json_object"]
8821
+
example: "json_object"
8822
+
default: "text"
8823
+
description: Must be one of `text` or `json_object`.
8824
+
8773
8825
AssistantObject:
8774
8826
type: object
8775
8827
title: Assistant
@@ -8918,6 +8970,29 @@ components:
8918
8970
type: object
8919
8971
x-oaiTypeLabel: map
8920
8972
nullable: true
8973
+
temperature:
8974
+
description: &run_temperature_description |
8975
+
What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.
8976
+
type: number
8977
+
minimum: 0
8978
+
maximum: 2
8979
+
default: 1
8980
+
example: 1
8981
+
nullable: true
8982
+
top_p:
8983
+
type: number
8984
+
minimum: 0
8985
+
maximum: 1
8986
+
default: 1
8987
+
example: 1
8988
+
nullable: true
8989
+
description: &run_top_p_description |
8990
+
An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered.
8991
+
8992
+
We generally recommend altering this or temperature but not both.
What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.
9049
+
type: number
9050
+
minimum: 0
9051
+
maximum: 2
9052
+
default: 1
9053
+
example: 1
9054
+
nullable: true
9055
+
top_p:
9056
+
type: number
9057
+
minimum: 0
9058
+
maximum: 1
9059
+
default: 1
9060
+
example: 1
9061
+
nullable: true
9062
+
description: &run_top_p_description |
9063
+
An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered.
9064
+
9065
+
We generally recommend altering this or temperature but not both.
Specifies the format that the model must output. Compatible with [GPT-4 Turbo](/docs/models/gpt-4-and-gpt-4-turbo) and all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
9107
-
9108
-
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the message the model generates is valid JSON.
9109
-
9110
-
**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.
An object describing the expected output of the model. If `json_object` only `function` type `tools` are allowed to be passed to the Run. If `text` the model can return text or any value needed.
9123
-
properties:
9124
-
type:
9125
-
type: string
9126
-
enum: ["text", "json_object"]
9127
-
example: "json_object"
9128
-
default: "text"
9129
-
description: Must be one of `text` or `json_object`.
9130
-
9131
9202
RunObject:
9132
9203
type: object
9133
9204
title: A run on a thread
@@ -9264,6 +9335,10 @@ components:
9264
9335
description: The sampling temperature used for this run. If not set, defaults to 1.
9265
9336
type: number
9266
9337
nullable: true
9338
+
top_p:
9339
+
description: The nucleus sampling value used for this run. If not set, defaults to 1.
9340
+
type: number
9341
+
nullable: true
9267
9342
max_prompt_tokens:
9268
9343
type: integer
9269
9344
nullable: true
@@ -9339,7 +9414,8 @@ components:
9339
9414
"completion_tokens": 456,
9340
9415
"total_tokens": 579
9341
9416
},
9342
-
"temperature": 1,
9417
+
"temperature": 1.0,
9418
+
"top_p": 1.0,
9343
9419
"max_prompt_tokens": 1000,
9344
9420
"max_completion_tokens": 1000,
9345
9421
"truncation_strategy": {
@@ -9422,8 +9498,18 @@ components:
9422
9498
default: 1
9423
9499
example: 1
9424
9500
nullable: true
9425
-
description: &run_temperature_description |
9426
-
What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.
9501
+
description: *run_temperature_description
9502
+
top_p:
9503
+
type: number
9504
+
minimum: 0
9505
+
maximum: 1
9506
+
default: 1
9507
+
example: 1
9508
+
nullable: true
9509
+
description: &run_top_p_description |
9510
+
An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered.
9511
+
9512
+
We generally recommend altering this or temperature but not both.
0 commit comments