@@ -19,6 +19,10 @@ export const openAiNativeModels = {
1919 description : "GPT-5: The best model for coding and agentic tasks across domains" ,
2020 // supportsVerbosity is a new capability; ensure ModelInfo includes it
2121 supportsVerbosity : true ,
22+ usesResponsesApi : true ,
23+ // Q: Why do we not send the temperature for GPT-5?
24+ // A: Because OpenAI does not support temperature over the API for GPT-5.
25+ supportsTemperature : false ,
2226 } ,
2327 "gpt-5-mini-2025-08-07" : {
2428 maxTokens : 128000 ,
@@ -32,6 +36,8 @@ export const openAiNativeModels = {
3236 cacheReadsPrice : 0.03 ,
3337 description : "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks" ,
3438 supportsVerbosity : true ,
39+ usesResponsesApi : true ,
40+ supportsTemperature : false ,
3541 } ,
3642 "gpt-5-nano-2025-08-07" : {
3743 maxTokens : 128000 ,
@@ -45,6 +51,8 @@ export const openAiNativeModels = {
4551 cacheReadsPrice : 0.01 ,
4652 description : "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5" ,
4753 supportsVerbosity : true ,
54+ usesResponsesApi : true ,
55+ supportsTemperature : false ,
4856 } ,
4957 "gpt-4.1" : {
5058 maxTokens : 32_768 ,
@@ -83,6 +91,8 @@ export const openAiNativeModels = {
8391 cacheReadsPrice : 0.5 ,
8492 supportsReasoningEffort : true ,
8593 reasoningEffort : "medium" ,
94+ usesResponsesApi : true ,
95+ supportsTemperature : false ,
8696 } ,
8797 "o3-high" : {
8898 maxTokens : 100_000 ,
@@ -93,6 +103,8 @@ export const openAiNativeModels = {
93103 outputPrice : 8.0 ,
94104 cacheReadsPrice : 0.5 ,
95105 reasoningEffort : "high" ,
106+ usesResponsesApi : true ,
107+ supportsTemperature : false ,
96108 } ,
97109 "o3-low" : {
98110 maxTokens : 100_000 ,
@@ -103,6 +115,8 @@ export const openAiNativeModels = {
103115 outputPrice : 8.0 ,
104116 cacheReadsPrice : 0.5 ,
105117 reasoningEffort : "low" ,
118+ usesResponsesApi : true ,
119+ supportsTemperature : false ,
106120 } ,
107121 "o4-mini" : {
108122 maxTokens : 100_000 ,
@@ -114,6 +128,8 @@ export const openAiNativeModels = {
114128 cacheReadsPrice : 0.275 ,
115129 supportsReasoningEffort : true ,
116130 reasoningEffort : "medium" ,
131+ usesResponsesApi : true ,
132+ supportsTemperature : false ,
117133 } ,
118134 "o4-mini-high" : {
119135 maxTokens : 100_000 ,
@@ -124,6 +140,8 @@ export const openAiNativeModels = {
124140 outputPrice : 4.4 ,
125141 cacheReadsPrice : 0.275 ,
126142 reasoningEffort : "high" ,
143+ usesResponsesApi : true ,
144+ supportsTemperature : false ,
127145 } ,
128146 "o4-mini-low" : {
129147 maxTokens : 100_000 ,
@@ -134,6 +152,8 @@ export const openAiNativeModels = {
134152 outputPrice : 4.4 ,
135153 cacheReadsPrice : 0.275 ,
136154 reasoningEffort : "low" ,
155+ usesResponsesApi : true ,
156+ supportsTemperature : false ,
137157 } ,
138158 "o3-mini" : {
139159 maxTokens : 100_000 ,
@@ -145,6 +165,8 @@ export const openAiNativeModels = {
145165 cacheReadsPrice : 0.55 ,
146166 supportsReasoningEffort : true ,
147167 reasoningEffort : "medium" ,
168+ usesResponsesApi : true ,
169+ supportsTemperature : false ,
148170 } ,
149171 "o3-mini-high" : {
150172 maxTokens : 100_000 ,
@@ -155,6 +177,8 @@ export const openAiNativeModels = {
155177 outputPrice : 4.4 ,
156178 cacheReadsPrice : 0.55 ,
157179 reasoningEffort : "high" ,
180+ usesResponsesApi : true ,
181+ supportsTemperature : false ,
158182 } ,
159183 "o3-mini-low" : {
160184 maxTokens : 100_000 ,
@@ -165,6 +189,8 @@ export const openAiNativeModels = {
165189 outputPrice : 4.4 ,
166190 cacheReadsPrice : 0.55 ,
167191 reasoningEffort : "low" ,
192+ usesResponsesApi : true ,
193+ supportsTemperature : false ,
168194 } ,
169195 o1 : {
170196 maxTokens : 100_000 ,
@@ -174,6 +200,8 @@ export const openAiNativeModels = {
174200 inputPrice : 15 ,
175201 outputPrice : 60 ,
176202 cacheReadsPrice : 7.5 ,
203+ usesResponsesApi : true ,
204+ supportsTemperature : false ,
177205 } ,
178206 "o1-preview" : {
179207 maxTokens : 32_768 ,
@@ -183,6 +211,8 @@ export const openAiNativeModels = {
183211 inputPrice : 15 ,
184212 outputPrice : 60 ,
185213 cacheReadsPrice : 7.5 ,
214+ usesResponsesApi : true ,
215+ supportsTemperature : false ,
186216 } ,
187217 "o1-mini" : {
188218 maxTokens : 65_536 ,
@@ -192,6 +222,8 @@ export const openAiNativeModels = {
192222 inputPrice : 1.1 ,
193223 outputPrice : 4.4 ,
194224 cacheReadsPrice : 0.55 ,
225+ usesResponsesApi : true ,
226+ supportsTemperature : false ,
195227 } ,
196228 "gpt-4.5-preview" : {
197229 maxTokens : 16_384 ,
@@ -228,6 +260,7 @@ export const openAiNativeModels = {
228260 inputPrice : 1.5 ,
229261 outputPrice : 6 ,
230262 cacheReadsPrice : 0 ,
263+ usesResponsesApi : true ,
231264 description :
232265 "Codex Mini: Cloud-based software engineering agent powered by codex-1, a version of o3 optimized for coding tasks. Trained with reinforcement learning to generate human-style code, adhere to instructions, and iteratively run tests." ,
233266 } ,
@@ -247,6 +280,5 @@ export const openAiModelInfoSaneDefaults: ModelInfo = {
247280export const azureOpenAiDefaultApiVersion = "2024-08-01-preview"
248281
249282export const OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0
250- export const GPT5_DEFAULT_TEMPERATURE = 1.0
251283
252284export const OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions"
0 commit comments