From 90c44400f8713b7d2d0b51142f4ed5509dbca713 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 8 Oct 2025 14:55:34 +0000 Subject: [PATCH 1/4] chore: ignore linter error for tests having large collections --- .rubocop.yml | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/.rubocop.yml b/.rubocop.yml index decbf212..1097e9aa 100644 --- a/.rubocop.yml +++ b/.rubocop.yml @@ -121,6 +121,10 @@ Metrics/BlockLength: Metrics/ClassLength: Enabled: false +Metrics/CollectionLiteralLength: + Exclude: + - "test/**/*" + Metrics/CyclomaticComplexity: Enabled: false From d18c5af9d05ae63616f2c83fb228c15f37cdddb0 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Thu, 9 Oct 2025 16:26:25 +0000 Subject: [PATCH 2/4] chore: simplify model references --- rbi/openai/models/eval_create_response.rbi | 8 ++++---- rbi/openai/models/eval_list_response.rbi | 8 ++++---- rbi/openai/models/eval_retrieve_response.rbi | 8 ++++---- rbi/openai/models/eval_update_response.rbi | 8 ++++---- sig/openai/models/eval_create_response.rbs | 4 ++-- sig/openai/models/eval_list_response.rbs | 4 ++-- sig/openai/models/eval_retrieve_response.rbs | 4 ++-- sig/openai/models/eval_update_response.rbs | 4 ++-- 8 files changed, 24 insertions(+), 24 deletions(-) diff --git a/rbi/openai/models/eval_create_response.rbi b/rbi/openai/models/eval_create_response.rbi index 6513b6f7..891ce3fe 100644 --- a/rbi/openai/models/eval_create_response.rbi +++ b/rbi/openai/models/eval_create_response.rbi @@ -70,8 +70,8 @@ module OpenAI testing_criteria: T::Array[ T.any( - OpenAI::Models::Graders::LabelModelGrader::OrHash, - OpenAI::Models::Graders::StringCheckGrader::OrHash, + OpenAI::Graders::LabelModelGrader::OrHash, + OpenAI::Graders::StringCheckGrader::OrHash, OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash, OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderPython::OrHash, OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderScoreModel::OrHash @@ -223,8 +223,8 @@ module OpenAI Variants = T.type_alias do T.any( - OpenAI::Models::Graders::LabelModelGrader, - OpenAI::Models::Graders::StringCheckGrader, + OpenAI::Graders::LabelModelGrader, + OpenAI::Graders::StringCheckGrader, OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderTextSimilarity, OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderPython, OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderScoreModel diff --git a/rbi/openai/models/eval_list_response.rbi b/rbi/openai/models/eval_list_response.rbi index 1158ef42..c1d1d4d7 100644 --- a/rbi/openai/models/eval_list_response.rbi +++ b/rbi/openai/models/eval_list_response.rbi @@ -68,8 +68,8 @@ module OpenAI testing_criteria: T::Array[ T.any( - OpenAI::Models::Graders::LabelModelGrader::OrHash, - OpenAI::Models::Graders::StringCheckGrader::OrHash, + OpenAI::Graders::LabelModelGrader::OrHash, + OpenAI::Graders::StringCheckGrader::OrHash, OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash, OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderPython::OrHash, OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderScoreModel::OrHash @@ -221,8 +221,8 @@ module OpenAI Variants = T.type_alias do T.any( - OpenAI::Models::Graders::LabelModelGrader, - OpenAI::Models::Graders::StringCheckGrader, + OpenAI::Graders::LabelModelGrader, + OpenAI::Graders::StringCheckGrader, OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderTextSimilarity, OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderPython, OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderScoreModel diff --git a/rbi/openai/models/eval_retrieve_response.rbi b/rbi/openai/models/eval_retrieve_response.rbi index 66b5570d..5cdaafb0 100644 --- a/rbi/openai/models/eval_retrieve_response.rbi +++ b/rbi/openai/models/eval_retrieve_response.rbi @@ -72,8 +72,8 @@ module OpenAI testing_criteria: T::Array[ T.any( - OpenAI::Models::Graders::LabelModelGrader::OrHash, - OpenAI::Models::Graders::StringCheckGrader::OrHash, + OpenAI::Graders::LabelModelGrader::OrHash, + OpenAI::Graders::StringCheckGrader::OrHash, OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash, OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderPython::OrHash, OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderScoreModel::OrHash @@ -225,8 +225,8 @@ module OpenAI Variants = T.type_alias do T.any( - OpenAI::Models::Graders::LabelModelGrader, - OpenAI::Models::Graders::StringCheckGrader, + OpenAI::Graders::LabelModelGrader, + OpenAI::Graders::StringCheckGrader, OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderTextSimilarity, OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderPython, OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderScoreModel diff --git a/rbi/openai/models/eval_update_response.rbi b/rbi/openai/models/eval_update_response.rbi index 07d46351..780e123c 100644 --- a/rbi/openai/models/eval_update_response.rbi +++ b/rbi/openai/models/eval_update_response.rbi @@ -70,8 +70,8 @@ module OpenAI testing_criteria: T::Array[ T.any( - OpenAI::Models::Graders::LabelModelGrader::OrHash, - OpenAI::Models::Graders::StringCheckGrader::OrHash, + OpenAI::Graders::LabelModelGrader::OrHash, + OpenAI::Graders::StringCheckGrader::OrHash, OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash, OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderPython::OrHash, OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderScoreModel::OrHash @@ -223,8 +223,8 @@ module OpenAI Variants = T.type_alias do T.any( - OpenAI::Models::Graders::LabelModelGrader, - OpenAI::Models::Graders::StringCheckGrader, + OpenAI::Graders::LabelModelGrader, + OpenAI::Graders::StringCheckGrader, OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderTextSimilarity, OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderPython, OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderScoreModel diff --git a/sig/openai/models/eval_create_response.rbs b/sig/openai/models/eval_create_response.rbs index 0d4799af..032726a5 100644 --- a/sig/openai/models/eval_create_response.rbs +++ b/sig/openai/models/eval_create_response.rbs @@ -85,8 +85,8 @@ module OpenAI end type testing_criterion = - OpenAI::Models::Graders::LabelModelGrader - | OpenAI::Models::Graders::StringCheckGrader + OpenAI::Graders::LabelModelGrader + | OpenAI::Graders::StringCheckGrader | OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderTextSimilarity | OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderPython | OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderScoreModel diff --git a/sig/openai/models/eval_list_response.rbs b/sig/openai/models/eval_list_response.rbs index ef01d7c9..7f139f33 100644 --- a/sig/openai/models/eval_list_response.rbs +++ b/sig/openai/models/eval_list_response.rbs @@ -85,8 +85,8 @@ module OpenAI end type testing_criterion = - OpenAI::Models::Graders::LabelModelGrader - | OpenAI::Models::Graders::StringCheckGrader + OpenAI::Graders::LabelModelGrader + | OpenAI::Graders::StringCheckGrader | OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderTextSimilarity | OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderPython | OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderScoreModel diff --git a/sig/openai/models/eval_retrieve_response.rbs b/sig/openai/models/eval_retrieve_response.rbs index 10f46696..3f4b7fcd 100644 --- a/sig/openai/models/eval_retrieve_response.rbs +++ b/sig/openai/models/eval_retrieve_response.rbs @@ -85,8 +85,8 @@ module OpenAI end type testing_criterion = - OpenAI::Models::Graders::LabelModelGrader - | OpenAI::Models::Graders::StringCheckGrader + OpenAI::Graders::LabelModelGrader + | OpenAI::Graders::StringCheckGrader | OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderTextSimilarity | OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderPython | OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderScoreModel diff --git a/sig/openai/models/eval_update_response.rbs b/sig/openai/models/eval_update_response.rbs index 532a5435..5ec19559 100644 --- a/sig/openai/models/eval_update_response.rbs +++ b/sig/openai/models/eval_update_response.rbs @@ -85,8 +85,8 @@ module OpenAI end type testing_criterion = - OpenAI::Models::Graders::LabelModelGrader - | OpenAI::Models::Graders::StringCheckGrader + OpenAI::Graders::LabelModelGrader + | OpenAI::Graders::StringCheckGrader | OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderTextSimilarity | OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderPython | OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderScoreModel From ac3e58bbee0c919ac84c4b3ac8b67955bca7ba88 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 10 Oct 2025 01:08:58 +0000 Subject: [PATCH 3/4] feat(api): comparison filter in/not in --- .stats.yml | 6 +-- .../models/beta/assistant_create_params.rb | 3 ++ .../models/beta/assistant_update_params.rb | 3 ++ .../models/beta/threads/run_create_params.rb | 3 ++ .../models/chat/completion_create_params.rb | 3 ++ lib/openai/models/comparison_filter.rb | 35 +++++++++++--- ...create_eval_completions_run_data_source.rb | 3 ++ .../models/evals/run_cancel_response.rb | 6 +++ lib/openai/models/evals/run_create_params.rb | 6 +++ .../models/evals/run_create_response.rb | 6 +++ lib/openai/models/evals/run_list_response.rb | 6 +++ .../models/evals/run_retrieve_response.rb | 6 +++ .../models/graders/score_model_grader.rb | 3 ++ lib/openai/models/reasoning.rb | 3 ++ lib/openai/models/reasoning_effort.rb | 3 ++ .../models/vector_stores/vector_store_file.rb | 6 +-- lib/openai/resources/files.rb | 2 +- .../models/beta/assistant_create_params.rbi | 6 +++ .../models/beta/assistant_update_params.rbi | 6 +++ .../models/beta/threads/run_create_params.rbi | 6 +++ .../models/chat/completion_create_params.rbi | 6 +++ rbi/openai/models/comparison_filter.rbi | 47 +++++++++++++++++-- ...reate_eval_completions_run_data_source.rbi | 6 +++ .../models/evals/run_cancel_response.rbi | 12 +++++ rbi/openai/models/evals/run_create_params.rbi | 12 +++++ .../models/evals/run_create_response.rbi | 12 +++++ rbi/openai/models/evals/run_list_response.rbi | 12 +++++ .../models/evals/run_retrieve_response.rbi | 12 +++++ .../models/graders/score_model_grader.rbi | 6 +++ rbi/openai/models/reasoning.rbi | 6 +++ rbi/openai/models/reasoning_effort.rbi | 3 ++ .../vector_stores/vector_store_file.rbi | 6 +-- rbi/openai/resources/beta/assistants.rbi | 6 +++ rbi/openai/resources/beta/threads/runs.rbi | 6 +++ rbi/openai/resources/chat/completions.rbi | 6 +++ rbi/openai/resources/files.rbi | 2 +- sig/openai/models/comparison_filter.rbs | 16 ++++++- 37 files changed, 275 insertions(+), 22 deletions(-) diff --git a/.stats.yml b/.stats.yml index f7aa916f..e68631a0 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 135 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-d64cf80d2ebddf175c5578f68226a3d5bbd3f7fd8d62ccac2205f3fc05a355ee.yml -openapi_spec_hash: d51e0d60d0c536f210b597a211bc5af0 -config_hash: e7c42016df9c6bd7bd6ff15101b9bc9b +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-e66e85fb7f72477256dca1acb6b23396989d381c5c1b318de564195436bcb93f.yml +openapi_spec_hash: 0a4bbb5aa0ae532a072bd6b3854e70b1 +config_hash: 89bf7bb3a1f9439ffc6ea0e7dc57ba9b diff --git a/lib/openai/models/beta/assistant_create_params.rb b/lib/openai/models/beta/assistant_create_params.rb index 3fe326c7..d1903311 100644 --- a/lib/openai/models/beta/assistant_create_params.rb +++ b/lib/openai/models/beta/assistant_create_params.rb @@ -55,6 +55,9 @@ class AssistantCreateParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/beta/assistant_update_params.rb b/lib/openai/models/beta/assistant_update_params.rb index 75ec6a82..3adb9322 100644 --- a/lib/openai/models/beta/assistant_update_params.rb +++ b/lib/openai/models/beta/assistant_update_params.rb @@ -55,6 +55,9 @@ class AssistantUpdateParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/beta/threads/run_create_params.rb b/lib/openai/models/beta/threads/run_create_params.rb index 2d44abe8..57269c0b 100644 --- a/lib/openai/models/beta/threads/run_create_params.rb +++ b/lib/openai/models/beta/threads/run_create_params.rb @@ -113,6 +113,9 @@ class RunCreateParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/chat/completion_create_params.rb b/lib/openai/models/chat/completion_create_params.rb index b9e16488..72303164 100644 --- a/lib/openai/models/chat/completion_create_params.rb +++ b/lib/openai/models/chat/completion_create_params.rb @@ -197,6 +197,9 @@ class CompletionCreateParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/comparison_filter.rb b/lib/openai/models/comparison_filter.rb index a8a32298..96acf57f 100644 --- a/lib/openai/models/comparison_filter.rb +++ b/lib/openai/models/comparison_filter.rb @@ -10,7 +10,8 @@ class ComparisonFilter < OpenAI::Internal::Type::BaseModel required :key, String # @!attribute type - # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`. + # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`, + # `nin`. # # - `eq`: equals # - `ne`: not equal @@ -18,6 +19,8 @@ class ComparisonFilter < OpenAI::Internal::Type::BaseModel # - `gte`: greater than or equal # - `lt`: less than # - `lte`: less than or equal + # - `in`: in + # - `nin`: not in # # @return [Symbol, OpenAI::Models::ComparisonFilter::Type] required :type, enum: -> { OpenAI::ComparisonFilter::Type } @@ -26,7 +29,7 @@ class ComparisonFilter < OpenAI::Internal::Type::BaseModel # The value to compare against the attribute key; supports string, number, or # boolean types. # - # @return [String, Float, Boolean] + # @return [String, Float, Boolean, Array] required :value, union: -> { OpenAI::ComparisonFilter::Value } # @!method initialize(key:, type:, value:) @@ -38,11 +41,12 @@ class ComparisonFilter < OpenAI::Internal::Type::BaseModel # # @param key [String] The key to compare against the value. # - # @param type [Symbol, OpenAI::Models::ComparisonFilter::Type] Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`. + # @param type [Symbol, OpenAI::Models::ComparisonFilter::Type] Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`, ` # - # @param value [String, Float, Boolean] The value to compare against the attribute key; supports string, number, or bool + # @param value [String, Float, Boolean, Array] The value to compare against the attribute key; supports string, number, or bool - # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`. + # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`, + # `nin`. # # - `eq`: equals # - `ne`: not equal @@ -50,6 +54,8 @@ class ComparisonFilter < OpenAI::Internal::Type::BaseModel # - `gte`: greater than or equal # - `lt`: less than # - `lte`: less than or equal + # - `in`: in + # - `nin`: not in # # @see OpenAI::Models::ComparisonFilter#type module Type @@ -79,8 +85,25 @@ module Value variant OpenAI::Internal::Type::Boolean + variant -> { OpenAI::Models::ComparisonFilter::Value::UnionMember3Array } + + module UnionMember3 + extend OpenAI::Internal::Type::Union + + variant String + + variant Float + + # @!method self.variants + # @return [Array(String, Float)] + end + # @!method self.variants - # @return [Array(String, Float, Boolean)] + # @return [Array(String, Float, Boolean, Array)] + + # @type [OpenAI::Internal::Type::Converter] + UnionMember3Array = + OpenAI::Internal::Type::ArrayOf[union: -> { OpenAI::ComparisonFilter::Value::UnionMember3 }] end end end diff --git a/lib/openai/models/evals/create_eval_completions_run_data_source.rb b/lib/openai/models/evals/create_eval_completions_run_data_source.rb index 24e5106e..920f3fd0 100644 --- a/lib/openai/models/evals/create_eval_completions_run_data_source.rb +++ b/lib/openai/models/evals/create_eval_completions_run_data_source.rb @@ -466,6 +466,9 @@ class SamplingParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/evals/run_cancel_response.rb b/lib/openai/models/evals/run_cancel_response.rb index f3886959..0b0f2b47 100644 --- a/lib/openai/models/evals/run_cancel_response.rb +++ b/lib/openai/models/evals/run_cancel_response.rb @@ -320,6 +320,9 @@ class Responses < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true @@ -661,6 +664,9 @@ class SamplingParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/evals/run_create_params.rb b/lib/openai/models/evals/run_create_params.rb index 739dfe6f..f4ec6c70 100644 --- a/lib/openai/models/evals/run_create_params.rb +++ b/lib/openai/models/evals/run_create_params.rb @@ -232,6 +232,9 @@ class Responses < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true @@ -589,6 +592,9 @@ class SamplingParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/evals/run_create_response.rb b/lib/openai/models/evals/run_create_response.rb index 0dd7985c..32d9244d 100644 --- a/lib/openai/models/evals/run_create_response.rb +++ b/lib/openai/models/evals/run_create_response.rb @@ -320,6 +320,9 @@ class Responses < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true @@ -661,6 +664,9 @@ class SamplingParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/evals/run_list_response.rb b/lib/openai/models/evals/run_list_response.rb index 0789f15a..b1947e4d 100644 --- a/lib/openai/models/evals/run_list_response.rb +++ b/lib/openai/models/evals/run_list_response.rb @@ -320,6 +320,9 @@ class Responses < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true @@ -661,6 +664,9 @@ class SamplingParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/evals/run_retrieve_response.rb b/lib/openai/models/evals/run_retrieve_response.rb index f531c975..9a2d8896 100644 --- a/lib/openai/models/evals/run_retrieve_response.rb +++ b/lib/openai/models/evals/run_retrieve_response.rb @@ -320,6 +320,9 @@ class Responses < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true @@ -665,6 +668,9 @@ class SamplingParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/graders/score_model_grader.rb b/lib/openai/models/graders/score_model_grader.rb index eeb6d628..8a7eb20e 100644 --- a/lib/openai/models/graders/score_model_grader.rb +++ b/lib/openai/models/graders/score_model_grader.rb @@ -226,6 +226,9 @@ class SamplingParams < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :reasoning_effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/reasoning.rb b/lib/openai/models/reasoning.rb index f2718df3..6a7b66c9 100644 --- a/lib/openai/models/reasoning.rb +++ b/lib/openai/models/reasoning.rb @@ -10,6 +10,9 @@ class Reasoning < OpenAI::Internal::Type::BaseModel # effort can result in faster responses and fewer tokens used on reasoning in a # response. # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. + # # @return [Symbol, OpenAI::Models::ReasoningEffort, nil] optional :effort, enum: -> { OpenAI::ReasoningEffort }, nil?: true diff --git a/lib/openai/models/reasoning_effort.rb b/lib/openai/models/reasoning_effort.rb index 486b6d31..02a47244 100644 --- a/lib/openai/models/reasoning_effort.rb +++ b/lib/openai/models/reasoning_effort.rb @@ -7,6 +7,9 @@ module Models # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. module ReasoningEffort extend OpenAI::Internal::Type::Enum diff --git a/lib/openai/models/vector_stores/vector_store_file.rb b/lib/openai/models/vector_stores/vector_store_file.rb index c6e737e3..0b330a50 100644 --- a/lib/openai/models/vector_stores/vector_store_file.rb +++ b/lib/openai/models/vector_stores/vector_store_file.rb @@ -101,7 +101,7 @@ class VectorStoreFile < OpenAI::Internal::Type::BaseModel # @see OpenAI::Models::VectorStores::VectorStoreFile#last_error class LastError < OpenAI::Internal::Type::BaseModel # @!attribute code - # One of `server_error` or `rate_limit_exceeded`. + # One of `server_error`, `unsupported_file`, or `invalid_file`. # # @return [Symbol, OpenAI::Models::VectorStores::VectorStoreFile::LastError::Code] required :code, enum: -> { OpenAI::VectorStores::VectorStoreFile::LastError::Code } @@ -116,11 +116,11 @@ class LastError < OpenAI::Internal::Type::BaseModel # The last error associated with this vector store file. Will be `null` if there # are no errors. # - # @param code [Symbol, OpenAI::Models::VectorStores::VectorStoreFile::LastError::Code] One of `server_error` or `rate_limit_exceeded`. + # @param code [Symbol, OpenAI::Models::VectorStores::VectorStoreFile::LastError::Code] One of `server_error`, `unsupported_file`, or `invalid_file`. # # @param message [String] A human-readable description of the error. - # One of `server_error` or `rate_limit_exceeded`. + # One of `server_error`, `unsupported_file`, or `invalid_file`. # # @see OpenAI::Models::VectorStores::VectorStoreFile::LastError#code module Code diff --git a/lib/openai/resources/files.rb b/lib/openai/resources/files.rb index 52d9a808..39a6207c 100644 --- a/lib/openai/resources/files.rb +++ b/lib/openai/resources/files.rb @@ -105,7 +105,7 @@ def list(params = {}) ) end - # Delete a file. + # Delete a file and remove it from all vector stores. # # @overload delete(file_id, request_options: {}) # diff --git a/rbi/openai/models/beta/assistant_create_params.rbi b/rbi/openai/models/beta/assistant_create_params.rbi index f4783b7e..034e3437 100644 --- a/rbi/openai/models/beta/assistant_create_params.rbi +++ b/rbi/openai/models/beta/assistant_create_params.rbi @@ -50,6 +50,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) } attr_accessor :reasoning_effort @@ -212,6 +215,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Specifies the format that the model must output. Compatible with # [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), diff --git a/rbi/openai/models/beta/assistant_update_params.rbi b/rbi/openai/models/beta/assistant_update_params.rbi index 217e526e..f07f224c 100644 --- a/rbi/openai/models/beta/assistant_update_params.rbi +++ b/rbi/openai/models/beta/assistant_update_params.rbi @@ -70,6 +70,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) } attr_accessor :reasoning_effort @@ -236,6 +239,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Specifies the format that the model must output. Compatible with # [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), diff --git a/rbi/openai/models/beta/threads/run_create_params.rbi b/rbi/openai/models/beta/threads/run_create_params.rbi index ccb3b696..ddcb19d4 100644 --- a/rbi/openai/models/beta/threads/run_create_params.rbi +++ b/rbi/openai/models/beta/threads/run_create_params.rbi @@ -116,6 +116,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) } attr_accessor :reasoning_effort @@ -334,6 +337,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Specifies the format that the model must output. Compatible with # [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), diff --git a/rbi/openai/models/chat/completion_create_params.rbi b/rbi/openai/models/chat/completion_create_params.rbi index 14fa8ae1..e934b1e2 100644 --- a/rbi/openai/models/chat/completion_create_params.rbi +++ b/rbi/openai/models/chat/completion_create_params.rbi @@ -230,6 +230,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) } attr_accessor :reasoning_effort @@ -667,6 +670,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # An object specifying the format that the model must output. # diff --git a/rbi/openai/models/comparison_filter.rbi b/rbi/openai/models/comparison_filter.rbi index 9ae08eed..c2849c0a 100644 --- a/rbi/openai/models/comparison_filter.rbi +++ b/rbi/openai/models/comparison_filter.rbi @@ -12,7 +12,8 @@ module OpenAI sig { returns(String) } attr_accessor :key - # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`. + # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`, + # `nin`. # # - `eq`: equals # - `ne`: not equal @@ -20,6 +21,8 @@ module OpenAI # - `gte`: greater than or equal # - `lt`: less than # - `lte`: less than or equal + # - `in`: in + # - `nin`: not in sig { returns(OpenAI::ComparisonFilter::Type::OrSymbol) } attr_accessor :type @@ -40,7 +43,8 @@ module OpenAI def self.new( # The key to compare against the value. key:, - # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`. + # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`, + # `nin`. # # - `eq`: equals # - `ne`: not equal @@ -48,6 +52,8 @@ module OpenAI # - `gte`: greater than or equal # - `lt`: less than # - `lte`: less than or equal + # - `in`: in + # - `nin`: not in type:, # The value to compare against the attribute key; supports string, number, or # boolean types. @@ -67,7 +73,8 @@ module OpenAI def to_hash end - # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`. + # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`, + # `nin`. # # - `eq`: equals # - `ne`: not equal @@ -75,6 +82,8 @@ module OpenAI # - `gte`: greater than or equal # - `lt`: less than # - `lte`: less than or equal + # - `in`: in + # - `nin`: not in module Type extend OpenAI::Internal::Type::Enum @@ -103,13 +112,43 @@ module OpenAI module Value extend OpenAI::Internal::Type::Union - Variants = T.type_alias { T.any(String, Float, T::Boolean) } + Variants = + T.type_alias do + T.any( + String, + Float, + T::Boolean, + T::Array[OpenAI::ComparisonFilter::Value::UnionMember3::Variants] + ) + end + + module UnionMember3 + extend OpenAI::Internal::Type::Union + + Variants = T.type_alias { T.any(String, Float) } + + sig do + override.returns( + T::Array[OpenAI::ComparisonFilter::Value::UnionMember3::Variants] + ) + end + def self.variants + end + end sig do override.returns(T::Array[OpenAI::ComparisonFilter::Value::Variants]) end def self.variants end + + UnionMember3Array = + T.let( + OpenAI::Internal::Type::ArrayOf[ + union: OpenAI::ComparisonFilter::Value::UnionMember3 + ], + OpenAI::Internal::Type::Converter + ) end end end diff --git a/rbi/openai/models/evals/create_eval_completions_run_data_source.rbi b/rbi/openai/models/evals/create_eval_completions_run_data_source.rbi index c04a1c2f..7515578d 100644 --- a/rbi/openai/models/evals/create_eval_completions_run_data_source.rbi +++ b/rbi/openai/models/evals/create_eval_completions_run_data_source.rbi @@ -892,6 +892,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) } attr_accessor :reasoning_effort @@ -992,6 +995,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # An object specifying the format that the model must output. # diff --git a/rbi/openai/models/evals/run_cancel_response.rbi b/rbi/openai/models/evals/run_cancel_response.rbi index 6502568e..9f4e862f 100644 --- a/rbi/openai/models/evals/run_cancel_response.rbi +++ b/rbi/openai/models/evals/run_cancel_response.rbi @@ -515,6 +515,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig do returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) end @@ -574,6 +577,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Sampling temperature. This is a query parameter used to select responses. temperature: nil, @@ -1120,6 +1126,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) } attr_accessor :reasoning_effort @@ -1241,6 +1250,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # A seed value to initialize the randomness, during sampling. seed: nil, diff --git a/rbi/openai/models/evals/run_create_params.rbi b/rbi/openai/models/evals/run_create_params.rbi index f772910e..9ba78fc7 100644 --- a/rbi/openai/models/evals/run_create_params.rbi +++ b/rbi/openai/models/evals/run_create_params.rbi @@ -425,6 +425,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) } attr_accessor :reasoning_effort @@ -482,6 +485,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Sampling temperature. This is a query parameter used to select responses. temperature: nil, @@ -1078,6 +1084,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) } attr_accessor :reasoning_effort @@ -1216,6 +1225,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # A seed value to initialize the randomness, during sampling. seed: nil, diff --git a/rbi/openai/models/evals/run_create_response.rbi b/rbi/openai/models/evals/run_create_response.rbi index f127a170..e60c4240 100644 --- a/rbi/openai/models/evals/run_create_response.rbi +++ b/rbi/openai/models/evals/run_create_response.rbi @@ -515,6 +515,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig do returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) end @@ -574,6 +577,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Sampling temperature. This is a query parameter used to select responses. temperature: nil, @@ -1120,6 +1126,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) } attr_accessor :reasoning_effort @@ -1241,6 +1250,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # A seed value to initialize the randomness, during sampling. seed: nil, diff --git a/rbi/openai/models/evals/run_list_response.rbi b/rbi/openai/models/evals/run_list_response.rbi index 2fba6c40..0ec5ecc2 100644 --- a/rbi/openai/models/evals/run_list_response.rbi +++ b/rbi/openai/models/evals/run_list_response.rbi @@ -511,6 +511,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig do returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) end @@ -570,6 +573,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Sampling temperature. This is a query parameter used to select responses. temperature: nil, @@ -1116,6 +1122,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) } attr_accessor :reasoning_effort @@ -1237,6 +1246,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # A seed value to initialize the randomness, during sampling. seed: nil, diff --git a/rbi/openai/models/evals/run_retrieve_response.rbi b/rbi/openai/models/evals/run_retrieve_response.rbi index bf9b002e..5bd74162 100644 --- a/rbi/openai/models/evals/run_retrieve_response.rbi +++ b/rbi/openai/models/evals/run_retrieve_response.rbi @@ -517,6 +517,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig do returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) end @@ -576,6 +579,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Sampling temperature. This is a query parameter used to select responses. temperature: nil, @@ -1122,6 +1128,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) } attr_accessor :reasoning_effort @@ -1243,6 +1252,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # A seed value to initialize the randomness, during sampling. seed: nil, diff --git a/rbi/openai/models/graders/score_model_grader.rbi b/rbi/openai/models/graders/score_model_grader.rbi index 8491c99e..6c8ddd69 100644 --- a/rbi/openai/models/graders/score_model_grader.rbi +++ b/rbi/openai/models/graders/score_model_grader.rbi @@ -399,6 +399,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) } attr_accessor :reasoning_effort @@ -432,6 +435,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # A seed value to initialize the randomness, during sampling. seed: nil, diff --git a/rbi/openai/models/reasoning.rbi b/rbi/openai/models/reasoning.rbi index f6364147..530a4b32 100644 --- a/rbi/openai/models/reasoning.rbi +++ b/rbi/openai/models/reasoning.rbi @@ -11,6 +11,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) } attr_accessor :effort @@ -46,6 +49,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. effort: nil, # **Deprecated:** use `summary` instead. # diff --git a/rbi/openai/models/reasoning_effort.rbi b/rbi/openai/models/reasoning_effort.rbi index fb0629b1..83ef54a6 100644 --- a/rbi/openai/models/reasoning_effort.rbi +++ b/rbi/openai/models/reasoning_effort.rbi @@ -7,6 +7,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. module ReasoningEffort extend OpenAI::Internal::Type::Enum diff --git a/rbi/openai/models/vector_stores/vector_store_file.rbi b/rbi/openai/models/vector_stores/vector_store_file.rbi index 9207da17..3d403625 100644 --- a/rbi/openai/models/vector_stores/vector_store_file.rbi +++ b/rbi/openai/models/vector_stores/vector_store_file.rbi @@ -190,7 +190,7 @@ module OpenAI ) end - # One of `server_error` or `rate_limit_exceeded`. + # One of `server_error`, `unsupported_file`, or `invalid_file`. sig do returns( OpenAI::VectorStores::VectorStoreFile::LastError::Code::TaggedSymbol @@ -212,7 +212,7 @@ module OpenAI ).returns(T.attached_class) end def self.new( - # One of `server_error` or `rate_limit_exceeded`. + # One of `server_error`, `unsupported_file`, or `invalid_file`. code:, # A human-readable description of the error. message: @@ -231,7 +231,7 @@ module OpenAI def to_hash end - # One of `server_error` or `rate_limit_exceeded`. + # One of `server_error`, `unsupported_file`, or `invalid_file`. module Code extend OpenAI::Internal::Type::Enum diff --git a/rbi/openai/resources/beta/assistants.rbi b/rbi/openai/resources/beta/assistants.rbi index 6489d48e..fb286a85 100644 --- a/rbi/openai/resources/beta/assistants.rbi +++ b/rbi/openai/resources/beta/assistants.rbi @@ -65,6 +65,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Specifies the format that the model must output. Compatible with # [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), @@ -192,6 +195,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Specifies the format that the model must output. Compatible with # [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), diff --git a/rbi/openai/resources/beta/threads/runs.rbi b/rbi/openai/resources/beta/threads/runs.rbi index d6e83222..06008148 100644 --- a/rbi/openai/resources/beta/threads/runs.rbi +++ b/rbi/openai/resources/beta/threads/runs.rbi @@ -128,6 +128,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Body param: Specifies the format that the model must output. Compatible with # [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), @@ -307,6 +310,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # Body param: Specifies the format that the model must output. Compatible with # [GPT-4o](https://platform.openai.com/docs/models#gpt-4o), diff --git a/rbi/openai/resources/chat/completions.rbi b/rbi/openai/resources/chat/completions.rbi index 43ec1239..54da2ae3 100644 --- a/rbi/openai/resources/chat/completions.rbi +++ b/rbi/openai/resources/chat/completions.rbi @@ -222,6 +222,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # An object specifying the format that the model must output. # @@ -539,6 +542,9 @@ module OpenAI # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning # effort can result in faster responses and fewer tokens used on reasoning in a # response. + # + # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning + # effort. reasoning_effort: nil, # An object specifying the format that the model must output. # diff --git a/rbi/openai/resources/files.rbi b/rbi/openai/resources/files.rbi index 0b28857e..dd7f80af 100644 --- a/rbi/openai/resources/files.rbi +++ b/rbi/openai/resources/files.rbi @@ -89,7 +89,7 @@ module OpenAI ) end - # Delete a file. + # Delete a file and remove it from all vector stores. sig do params( file_id: String, diff --git a/sig/openai/models/comparison_filter.rbs b/sig/openai/models/comparison_filter.rbs index f5f686ba..c27c9860 100644 --- a/sig/openai/models/comparison_filter.rbs +++ b/sig/openai/models/comparison_filter.rbs @@ -41,12 +41,26 @@ module OpenAI def self?.values: -> ::Array[OpenAI::Models::ComparisonFilter::type_] end - type value = String | Float | bool + type value = + String + | Float + | bool + | ::Array[OpenAI::Models::ComparisonFilter::Value::union_member3] module Value extend OpenAI::Internal::Type::Union + type union_member3 = String | Float + + module UnionMember3 + extend OpenAI::Internal::Type::Union + + def self?.variants: -> ::Array[OpenAI::Models::ComparisonFilter::Value::union_member3] + end + def self?.variants: -> ::Array[OpenAI::Models::ComparisonFilter::value] + + UnionMember3Array: OpenAI::Internal::Type::Converter end end end From 89f1f947eba9c609308eea1f16b2d2f8eb431957 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 10 Oct 2025 01:09:36 +0000 Subject: [PATCH 4/4] release: 0.31.0 --- .release-please-manifest.json | 2 +- CHANGELOG.md | 14 ++++++++++++++ Gemfile.lock | 2 +- README.md | 2 +- lib/openai/version.rb | 2 +- 5 files changed, 18 insertions(+), 4 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 554e34bb..f81bf992 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "0.30.0" + ".": "0.31.0" } \ No newline at end of file diff --git a/CHANGELOG.md b/CHANGELOG.md index ad2d911a..6af26dd4 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,19 @@ # Changelog +## 0.31.0 (2025-10-10) + +Full Changelog: [v0.30.0...v0.31.0](https://github.com/openai/openai-ruby/compare/v0.30.0...v0.31.0) + +### Features + +* **api:** comparison filter in/not in ([ac3e58b](https://github.com/openai/openai-ruby/commit/ac3e58bbee0c919ac84c4b3ac8b67955bca7ba88)) + + +### Chores + +* ignore linter error for tests having large collections ([90c4440](https://github.com/openai/openai-ruby/commit/90c44400f8713b7d2d0b51142f4ed5509dbca713)) +* simplify model references ([d18c5af](https://github.com/openai/openai-ruby/commit/d18c5af9d05ae63616f2c83fb228c15f37cdddb0)) + ## 0.30.0 (2025-10-06) Full Changelog: [v0.29.0...v0.30.0](https://github.com/openai/openai-ruby/compare/v0.29.0...v0.30.0) diff --git a/Gemfile.lock b/Gemfile.lock index 9d3559ae..5ee1897c 100644 --- a/Gemfile.lock +++ b/Gemfile.lock @@ -11,7 +11,7 @@ GIT PATH remote: . specs: - openai (0.30.0) + openai (0.31.0) connection_pool GEM diff --git a/README.md b/README.md index 6a5eacc9..cc9ae07b 100644 --- a/README.md +++ b/README.md @@ -15,7 +15,7 @@ To use this gem, install via Bundler by adding the following to your application ```ruby -gem "openai", "~> 0.30.0" +gem "openai", "~> 0.31.0" ``` diff --git a/lib/openai/version.rb b/lib/openai/version.rb index 417d3f9a..4b592eec 100644 --- a/lib/openai/version.rb +++ b/lib/openai/version.rb @@ -1,5 +1,5 @@ # frozen_string_literal: true module OpenAI - VERSION = "0.30.0" + VERSION = "0.31.0" end