You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Constrains effort on reasoning for reasoning models. Currently supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
586
+
587
+
Note: The `gpt-5-pro` model defaults to (and only supports) high reasoning effort.
Specifies the processing type used for serving the request.
609
+
610
+
If set to `auto`, then the request will be processed with the service tier configured in the Project settings. If set to `default`, then the request will be processed with the standard pricing and performance for the selected model. If set to `flex` or `priority`, then the request will be processed with the corresponding service tier. When not set, the default behavior is `auto`.
611
+
612
+
When the service tier parameter is set, the response body will include the `service_tier` value based on the processing mode actually used to serve the request. This response value may be different from the value set in the parameter.
Constrains effort on reasoning for reasoning models. Currently supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
1517
+
1518
+
Note: The `gpt-5-pro` model defaults to (and only supports) high reasoning effort.
Specifies the processing type used for serving the request.
1540
+
1541
+
If set to `auto`, then the request will be processed with the service tier configured in the Project settings. If set to `default`, then the request will be processed with the standard pricing and performance for the selected model. If set to `flex` or `priority`, then the request will be processed with the corresponding service tier. When not set, the default behavior is `auto`.
1542
+
1543
+
When the service tier parameter is set, the response body will include the `service_tier` value based on the processing mode actually used to serve the request. This response value may be different from the value set in the parameter.
Constrains effort on reasoning for reasoning models. Currently supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
586
+
587
+
Note: The `gpt-5-pro` model defaults to (and only supports) high reasoning effort.
Specifies the processing type used for serving the request.
609
+
610
+
If set to `auto`, then the request will be processed with the service tier configured in the Project settings. If set to `default`, then the request will be processed with the standard pricing and performance for the selected model. If set to `flex` or `priority`, then the request will be processed with the corresponding service tier. When not set, the default behavior is `auto`.
611
+
612
+
When the service tier parameter is set, the response body will include the `service_tier` value based on the processing mode actually used to serve the request. This response value may be different from the value set in the parameter.
Constrains effort on reasoning for reasoning models. Currently supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
1517
+
1518
+
Note: The `gpt-5-pro` model defaults to (and only supports) high reasoning effort.
Specifies the processing type used for serving the request.
1540
+
1541
+
If set to `auto`, then the request will be processed with the service tier configured in the Project settings. If set to `default`, then the request will be processed with the standard pricing and performance for the selected model. If set to `flex` or `priority`, then the request will be processed with the corresponding service tier. When not set, the default behavior is `auto`.
1542
+
1543
+
When the service tier parameter is set, the response body will include the `service_tier` value based on the processing mode actually used to serve the request. This response value may be different from the value set in the parameter.
Copy file name to clipboardExpand all lines: model-providers/openai/openai-vanilla/runtime/src/main/java/io/quarkiverse/langchain4j/openai/runtime/OpenAiRecorder.java
Copy file name to clipboardExpand all lines: model-providers/openai/openai-vanilla/runtime/src/main/java/io/quarkiverse/langchain4j/openai/runtime/config/ChatModelConfig.java
+25Lines changed: 25 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -96,4 +96,29 @@ public interface ChatModelConfig {
96
96
* @return
97
97
*/
98
98
Optional<List<String>> stop();
99
+
100
+
/**
101
+
* Constrains effort on reasoning for reasoning models.
102
+
* Currently supported values are {@code minimal}, {@code low}, {@code medium}, and {@code high}.
103
+
* Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
104
+
* <p>
105
+
* Note: The {@code gpt-5-pro} model defaults to (and only supports) high reasoning effort.
106
+
*/
107
+
Optional<String> reasoningEffort();
108
+
109
+
/**
110
+
* Specifies the processing type used for serving the request.
111
+
* <p>
112
+
* If set to {@code auto}, then the request will be processed with the service tier configured in the Project settings.
113
+
* If set to {@code default}, then the request will be processed with the standard pricing and performance for the selected
114
+
* model.
115
+
* If set to {@code flex} or {@code priority}, then the request will be processed with the corresponding service tier.
116
+
* When not set, the default behavior is {@code auto}.
117
+
* <p>
118
+
* When the service tier parameter is set, the response body will include the {@code service_tier} value based on the
119
+
* processing mode actually used to serve the request.
120
+
* This response value may be different from the value set in the parameter.
0 commit comments