@@ -24,7 +24,7 @@ class CompletionCreateParams < OpenAI::Internal::Type::BaseModel
24
24
-> { OpenAI ::Internal ::Type ::ArrayOf [ union : OpenAI ::Models ::Chat ::ChatCompletionMessageParam ] }
25
25
26
26
# @!attribute model
27
- # Model ID used to generate the response, like `gpt-4o` or `o1 `. OpenAI offers a
27
+ # Model ID used to generate the response, like `gpt-4o` or `o3 `. OpenAI offers a
28
28
# wide range of models with different capabilities, performance characteristics,
29
29
# and price points. Refer to the
30
30
# [model guide](https://platform.openai.com/docs/models) to browse and compare
@@ -121,7 +121,7 @@ class CompletionCreateParams < OpenAI::Internal::Type::BaseModel
121
121
#
122
122
# This value is now deprecated in favor of `max_completion_tokens`, and is not
123
123
# compatible with
124
- # [o1 series models](https://platform.openai.com/docs/guides/reasoning).
124
+ # [o- series models](https://platform.openai.com/docs/guides/reasoning).
125
125
#
126
126
# @return [Integer, nil]
127
127
optional :max_tokens , Integer , nil? : true
@@ -240,6 +240,9 @@ class CompletionCreateParams < OpenAI::Internal::Type::BaseModel
240
240
# latency guarentee.
241
241
# - If set to 'default', the request will be processed using the default service
242
242
# tier with a lower uptime SLA and no latency guarentee.
243
+ # - If set to 'flex', the request will be processed with the Flex Processing
244
+ # service tier.
245
+ # [Learn more](https://platform.openai.com/docs/guides/flex-processing).
243
246
# - When not set, the default behavior is 'auto'.
244
247
#
245
248
# When this parameter is set, the response body will include the `service_tier`
@@ -249,6 +252,8 @@ class CompletionCreateParams < OpenAI::Internal::Type::BaseModel
249
252
optional :service_tier , enum : -> { OpenAI ::Models ::Chat ::CompletionCreateParams ::ServiceTier } , nil? : true
250
253
251
254
# @!attribute stop
255
+ # Not supported with latest reasoning models `o3` and `o4-mini`.
256
+ #
252
257
# Up to 4 sequences where the API will stop generating further tokens. The
253
258
# returned text will not contain the stop sequence.
254
259
#
@@ -422,7 +427,7 @@ class CompletionCreateParams < OpenAI::Internal::Type::BaseModel
422
427
423
428
# def initialize: (Hash | OpenAI::Internal::Type::BaseModel) -> void
424
429
425
- # Model ID used to generate the response, like `gpt-4o` or `o1 `. OpenAI offers a
430
+ # Model ID used to generate the response, like `gpt-4o` or `o3 `. OpenAI offers a
426
431
# wide range of models with different capabilities, performance characteristics,
427
432
# and price points. Refer to the
428
433
# [model guide](https://platform.openai.com/docs/models) to browse and compare
@@ -432,7 +437,7 @@ module Model
432
437
433
438
variant String
434
439
435
- # Model ID used to generate the response, like `gpt-4o` or `o1 `. OpenAI
440
+ # Model ID used to generate the response, like `gpt-4o` or `o3 `. OpenAI
436
441
# offers a wide range of models with different capabilities, performance
437
442
# characteristics, and price points. Refer to the [model guide](https://platform.openai.com/docs/models)
438
443
# to browse and compare available models.
@@ -589,6 +594,9 @@ module ResponseFormat
589
594
# latency guarentee.
590
595
# - If set to 'default', the request will be processed using the default service
591
596
# tier with a lower uptime SLA and no latency guarentee.
597
+ # - If set to 'flex', the request will be processed with the Flex Processing
598
+ # service tier.
599
+ # [Learn more](https://platform.openai.com/docs/guides/flex-processing).
592
600
# - When not set, the default behavior is 'auto'.
593
601
#
594
602
# When this parameter is set, the response body will include the `service_tier`
@@ -598,6 +606,7 @@ module ServiceTier
598
606
599
607
AUTO = :auto
600
608
DEFAULT = :default
609
+ FLEX = :flex
601
610
602
611
finalize!
603
612
@@ -606,6 +615,8 @@ module ServiceTier
606
615
# def self.values; end
607
616
end
608
617
618
+ # Not supported with latest reasoning models `o3` and `o4-mini`.
619
+ #
609
620
# Up to 4 sequences where the API will stop generating further tokens. The
610
621
# returned text will not contain the stop sequence.
611
622
module Stop
0 commit comments