@@ -281,9 +281,8 @@ export interface ChatCompletion {
281
281
* - If set to 'default', then the request will be processed with the standard
282
282
* pricing and performance for the selected model.
283
283
* - If set to '[flex](https://platform.openai.com/docs/guides/flex-processing)' or
284
- * 'priority', then the request will be processed with the corresponding service
285
- * tier. [Contact sales](https://openai.com/contact-sales) to learn more about
286
- * Priority processing.
284
+ * '[priority](https://openai.com/api-priority-processing/)', then the request
285
+ * will be processed with the corresponding service tier.
287
286
* - When not set, the default behavior is 'auto'.
288
287
*
289
288
* When the `service_tier` parameter is set, the response body will include the
@@ -294,7 +293,8 @@ export interface ChatCompletion {
294
293
service_tier ?: 'auto' | 'default' | 'flex' | 'scale' | 'priority' | null ;
295
294
296
295
/**
297
- * This fingerprint represents the backend configuration that the model runs with.
296
+ * @deprecated This fingerprint represents the backend configuration that the model
297
+ * runs with.
298
298
*
299
299
* Can be used in conjunction with the `seed` request parameter to understand when
300
300
* backend changes have been made that might impact determinism.
@@ -535,9 +535,8 @@ export interface ChatCompletionChunk {
535
535
* - If set to 'default', then the request will be processed with the standard
536
536
* pricing and performance for the selected model.
537
537
* - If set to '[flex](https://platform.openai.com/docs/guides/flex-processing)' or
538
- * 'priority', then the request will be processed with the corresponding service
539
- * tier. [Contact sales](https://openai.com/contact-sales) to learn more about
540
- * Priority processing.
538
+ * '[priority](https://openai.com/api-priority-processing/)', then the request
539
+ * will be processed with the corresponding service tier.
541
540
* - When not set, the default behavior is 'auto'.
542
541
*
543
542
* When the `service_tier` parameter is set, the response body will include the
@@ -548,9 +547,9 @@ export interface ChatCompletionChunk {
548
547
service_tier ?: 'auto' | 'default' | 'flex' | 'scale' | 'priority' | null ;
549
548
550
549
/**
551
- * This fingerprint represents the backend configuration that the model runs with.
552
- * Can be used in conjunction with the `seed` request parameter to understand when
553
- * backend changes have been made that might impact determinism.
550
+ * @deprecated This fingerprint represents the backend configuration that the model
551
+ * runs with. Can be used in conjunction with the `seed` request parameter to
552
+ * understand when backend changes have been made that might impact determinism.
554
553
*/
555
554
system_fingerprint ?: string ;
556
555
@@ -1664,11 +1663,11 @@ export interface ChatCompletionCreateParamsBase {
1664
1663
safety_identifier ?: string ;
1665
1664
1666
1665
/**
1667
- * This feature is in Beta. If specified, our system will make a best effort to
1668
- * sample deterministically, such that repeated requests with the same `seed` and
1669
- * parameters should return the same result. Determinism is not guaranteed, and you
1670
- * should refer to the `system_fingerprint` response parameter to monitor changes
1671
- * in the backend.
1666
+ * @deprecated This feature is in Beta. If specified, our system will make a best
1667
+ * effort to sample deterministically, such that repeated requests with the same
1668
+ * `seed` and parameters should return the same result. Determinism is not
1669
+ * guaranteed, and you should refer to the `system_fingerprint` response parameter
1670
+ * to monitor changes in the backend.
1672
1671
*/
1673
1672
seed ?: number | null ;
1674
1673
@@ -1681,9 +1680,8 @@ export interface ChatCompletionCreateParamsBase {
1681
1680
* - If set to 'default', then the request will be processed with the standard
1682
1681
* pricing and performance for the selected model.
1683
1682
* - If set to '[flex](https://platform.openai.com/docs/guides/flex-processing)' or
1684
- * 'priority', then the request will be processed with the corresponding service
1685
- * tier. [Contact sales](https://openai.com/contact-sales) to learn more about
1686
- * Priority processing.
1683
+ * '[priority](https://openai.com/api-priority-processing/)', then the request
1684
+ * will be processed with the corresponding service tier.
1687
1685
* - When not set, the default behavior is 'auto'.
1688
1686
*
1689
1687
* When the `service_tier` parameter is set, the response body will include the
@@ -1735,6 +1733,8 @@ export interface ChatCompletionCreateParamsBase {
1735
1733
*/
1736
1734
temperature ?: number | null ;
1737
1735
1736
+ text ?: ChatCompletionCreateParams . Text ;
1737
+
1738
1738
/**
1739
1739
* Controls which (if any) tool is called by the model. `none` means the model will
1740
1740
* not call any tool and instead generates a message. `auto` means the model can
@@ -1825,6 +1825,15 @@ export namespace ChatCompletionCreateParams {
1825
1825
parameters ?: Shared . FunctionParameters ;
1826
1826
}
1827
1827
1828
+ export interface Text {
1829
+ /**
1830
+ * Constrains the verbosity of the model's response. Lower values will result in
1831
+ * more concise responses, while higher values will result in more verbose
1832
+ * responses. Currently supported values are `low`, `medium`, and `high`.
1833
+ */
1834
+ verbosity ?: 'low' | 'medium' | 'high' | null ;
1835
+ }
1836
+
1828
1837
/**
1829
1838
* This tool searches the web for relevant results to use in a response. Learn more
1830
1839
* about the
0 commit comments