Skip to content

Commit 0859662

Browse files
HavenDVgithub-actions[bot]
andauthored
feat: Updated OpenAPI spec (#237)
Co-authored-by: github-actions[bot] <[email protected]>
1 parent 321feb9 commit 0859662

File tree

7 files changed

+66
-0
lines changed

7 files changed

+66
-0
lines changed

src/libs/Cohere/Generated/Cohere.CohereClient.Chat.g.cs

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -740,6 +740,11 @@ partial void ProcessChatResponseContent(
740740
/// - AUTO: Cohere Platform Only<br/>
741741
/// - AUTO_PRESERVE_ORDER: Azure, AWS Sagemaker/Bedrock, Private Deployments
742742
/// </param>
743+
/// <param name="rawPrompting">
744+
/// When enabled, the user's prompt will be sent to the model without<br/>
745+
/// any pre-processing.<br/>
746+
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
747+
/// </param>
743748
/// <param name="responseFormat">
744749
/// Configuration for forcing the model output to adhere to the specified format. Supported on [Command R 03-2024](https://docs.cohere.com/docs/command-r), [Command R+ 04-2024](https://docs.cohere.com/docs/command-r-plus) and newer models.<br/>
745750
/// The model can be forced into outputting JSON objects (with up to 5 levels of nesting) by setting `{ "type": "json_object" }`.<br/>
@@ -834,6 +839,7 @@ partial void ProcessChatResponseContent(
834839
string? model = default,
835840
string? preamble = default,
836841
global::Cohere.ChatRequestPromptTruncation? promptTruncation = default,
842+
bool? rawPrompting = default,
837843
global::Cohere.ResponseFormat? responseFormat = default,
838844
global::Cohere.ChatRequestSafetyMode? safetyMode = default,
839845
bool? searchQueriesOnly = default,
@@ -862,6 +868,7 @@ partial void ProcessChatResponseContent(
862868
Preamble = preamble,
863869
PresencePenalty = presencePenalty,
864870
PromptTruncation = promptTruncation,
871+
RawPrompting = rawPrompting,
865872
ResponseFormat = responseFormat,
866873
SafetyMode = safetyMode,
867874
SearchQueriesOnly = searchQueriesOnly,

src/libs/Cohere/Generated/Cohere.CohereClient.Chatv2.g.cs

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -667,6 +667,11 @@ partial void ProcessChatv2ResponseContent(
667667
/// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.<br/>
668668
/// Used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
669669
/// </param>
670+
/// <param name="rawPrompting">
671+
/// When enabled, the user's prompt will be sent to the model without<br/>
672+
/// any pre-processing.<br/>
673+
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
674+
/// </param>
670675
/// <param name="reasoningEffort">
671676
/// The reasoning effort level of the model. This affects the model's performance and the time it takes to generate a response.
672677
/// </param>
@@ -733,6 +738,7 @@ partial void ProcessChatv2ResponseContent(
733738
int? maxTokens = default,
734739
float? p = default,
735740
float? presencePenalty = default,
741+
bool? rawPrompting = default,
736742
global::Cohere.ReasoningEffort? reasoningEffort = default,
737743
global::Cohere.ResponseFormatV2? responseFormat = default,
738744
global::Cohere.Chatv2RequestSafetyMode? safetyMode = default,
@@ -757,6 +763,7 @@ partial void ProcessChatv2ResponseContent(
757763
Model = model,
758764
P = p,
759765
PresencePenalty = presencePenalty,
766+
RawPrompting = rawPrompting,
760767
ReasoningEffort = reasoningEffort,
761768
ResponseFormat = responseFormat,
762769
SafetyMode = safetyMode,

src/libs/Cohere/Generated/Cohere.ICohereClient.Chat.g.cs

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -127,6 +127,11 @@ public partial interface ICohereClient
127127
/// - AUTO: Cohere Platform Only<br/>
128128
/// - AUTO_PRESERVE_ORDER: Azure, AWS Sagemaker/Bedrock, Private Deployments
129129
/// </param>
130+
/// <param name="rawPrompting">
131+
/// When enabled, the user's prompt will be sent to the model without<br/>
132+
/// any pre-processing.<br/>
133+
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
134+
/// </param>
130135
/// <param name="responseFormat">
131136
/// Configuration for forcing the model output to adhere to the specified format. Supported on [Command R 03-2024](https://docs.cohere.com/docs/command-r), [Command R+ 04-2024](https://docs.cohere.com/docs/command-r-plus) and newer models.<br/>
132137
/// The model can be forced into outputting JSON objects (with up to 5 levels of nesting) by setting `{ "type": "json_object" }`.<br/>
@@ -221,6 +226,7 @@ public partial interface ICohereClient
221226
string? model = default,
222227
string? preamble = default,
223228
global::Cohere.ChatRequestPromptTruncation? promptTruncation = default,
229+
bool? rawPrompting = default,
224230
global::Cohere.ResponseFormat? responseFormat = default,
225231
global::Cohere.ChatRequestSafetyMode? safetyMode = default,
226232
bool? searchQueriesOnly = default,

src/libs/Cohere/Generated/Cohere.ICohereClient.Chatv2.g.cs

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -62,6 +62,11 @@ public partial interface ICohereClient
6262
/// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.<br/>
6363
/// Used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
6464
/// </param>
65+
/// <param name="rawPrompting">
66+
/// When enabled, the user's prompt will be sent to the model without<br/>
67+
/// any pre-processing.<br/>
68+
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
69+
/// </param>
6570
/// <param name="reasoningEffort">
6671
/// The reasoning effort level of the model. This affects the model's performance and the time it takes to generate a response.
6772
/// </param>
@@ -128,6 +133,7 @@ public partial interface ICohereClient
128133
int? maxTokens = default,
129134
float? p = default,
130135
float? presencePenalty = default,
136+
bool? rawPrompting = default,
131137
global::Cohere.ReasoningEffort? reasoningEffort = default,
132138
global::Cohere.ResponseFormatV2? responseFormat = default,
133139
global::Cohere.Chatv2RequestSafetyMode? safetyMode = default,

src/libs/Cohere/Generated/Cohere.Models.ChatRequest.g.cs

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -159,6 +159,14 @@ public sealed partial class ChatRequest
159159
[global::System.Text.Json.Serialization.JsonConverter(typeof(global::Cohere.JsonConverters.ChatRequestPromptTruncationJsonConverter))]
160160
public global::Cohere.ChatRequestPromptTruncation? PromptTruncation { get; set; }
161161

162+
/// <summary>
163+
/// When enabled, the user's prompt will be sent to the model without<br/>
164+
/// any pre-processing.<br/>
165+
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
166+
/// </summary>
167+
[global::System.Text.Json.Serialization.JsonPropertyName("raw_prompting")]
168+
public bool? RawPrompting { get; set; }
169+
162170
/// <summary>
163171
/// Configuration for forcing the model output to adhere to the specified format. Supported on [Command R 03-2024](https://docs.cohere.com/docs/command-r), [Command R+ 04-2024](https://docs.cohere.com/docs/command-r-plus) and newer models.<br/>
164172
/// The model can be forced into outputting JSON objects (with up to 5 levels of nesting) by setting `{ "type": "json_object" }`.<br/>
@@ -369,6 +377,11 @@ public sealed partial class ChatRequest
369377
/// - AUTO: Cohere Platform Only<br/>
370378
/// - AUTO_PRESERVE_ORDER: Azure, AWS Sagemaker/Bedrock, Private Deployments
371379
/// </param>
380+
/// <param name="rawPrompting">
381+
/// When enabled, the user's prompt will be sent to the model without<br/>
382+
/// any pre-processing.<br/>
383+
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
384+
/// </param>
372385
/// <param name="responseFormat">
373386
/// Configuration for forcing the model output to adhere to the specified format. Supported on [Command R 03-2024](https://docs.cohere.com/docs/command-r), [Command R+ 04-2024](https://docs.cohere.com/docs/command-r-plus) and newer models.<br/>
374387
/// The model can be forced into outputting JSON objects (with up to 5 levels of nesting) by setting `{ "type": "json_object" }`.<br/>
@@ -461,6 +474,7 @@ public ChatRequest(
461474
string? preamble,
462475
double? presencePenalty,
463476
global::Cohere.ChatRequestPromptTruncation? promptTruncation,
477+
bool? rawPrompting,
464478
global::Cohere.ResponseFormat? responseFormat,
465479
global::Cohere.ChatRequestSafetyMode? safetyMode,
466480
bool? searchQueriesOnly,
@@ -487,6 +501,7 @@ public ChatRequest(
487501
this.Preamble = preamble;
488502
this.PresencePenalty = presencePenalty;
489503
this.PromptTruncation = promptTruncation;
504+
this.RawPrompting = rawPrompting;
490505
this.ResponseFormat = responseFormat;
491506
this.SafetyMode = safetyMode;
492507
this.SearchQueriesOnly = searchQueriesOnly;

src/libs/Cohere/Generated/Cohere.Models.Chatv2Request.g.cs

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -80,6 +80,14 @@ public sealed partial class Chatv2Request
8080
[global::System.Text.Json.Serialization.JsonPropertyName("presence_penalty")]
8181
public float? PresencePenalty { get; set; }
8282

83+
/// <summary>
84+
/// When enabled, the user's prompt will be sent to the model without<br/>
85+
/// any pre-processing.<br/>
86+
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
87+
/// </summary>
88+
[global::System.Text.Json.Serialization.JsonPropertyName("raw_prompting")]
89+
public bool? RawPrompting { get; set; }
90+
8391
/// <summary>
8492
/// The reasoning effort level of the model. This affects the model's performance and the time it takes to generate a response.
8593
/// </summary>
@@ -213,6 +221,11 @@ public sealed partial class Chatv2Request
213221
/// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.<br/>
214222
/// Used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
215223
/// </param>
224+
/// <param name="rawPrompting">
225+
/// When enabled, the user's prompt will be sent to the model without<br/>
226+
/// any pre-processing.<br/>
227+
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
228+
/// </param>
216229
/// <param name="reasoningEffort">
217230
/// The reasoning effort level of the model. This affects the model's performance and the time it takes to generate a response.
218231
/// </param>
@@ -279,6 +292,7 @@ public Chatv2Request(
279292
int? maxTokens,
280293
float? p,
281294
float? presencePenalty,
295+
bool? rawPrompting,
282296
global::Cohere.ReasoningEffort? reasoningEffort,
283297
global::Cohere.ResponseFormatV2? responseFormat,
284298
global::Cohere.Chatv2RequestSafetyMode? safetyMode,
@@ -300,6 +314,7 @@ public Chatv2Request(
300314
this.MaxTokens = maxTokens;
301315
this.P = p;
302316
this.PresencePenalty = presencePenalty;
317+
this.RawPrompting = rawPrompting;
303318
this.ReasoningEffort = reasoningEffort;
304319
this.ResponseFormat = responseFormat;
305320
this.SafetyMode = safetyMode;

src/libs/Cohere/openapi.yaml

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -148,6 +148,11 @@ paths:
148148
description: "Defaults to `AUTO` when `connectors` are specified and `OFF` in all other cases.\n\nDictates how the prompt will be constructed.\n\nWith `prompt_truncation` set to \"AUTO\", some elements from `chat_history` and `documents` will be dropped in an attempt to construct a prompt that fits within the model's context length limit. During this process the order of the documents and chat history will be changed and ranked by relevance.\n\nWith `prompt_truncation` set to \"AUTO_PRESERVE_ORDER\", some elements from `chat_history` and `documents` will be dropped in an attempt to construct a prompt that fits within the model's context length limit. During this process the order of the documents and chat history will be preserved as they are inputted into the API.\n\nWith `prompt_truncation` set to \"OFF\", no elements will be dropped. If the sum of the inputs exceeds the model's context length limit, a `TooManyTokens` error will be returned.\n\nCompatible Deployments:\n - AUTO: Cohere Platform Only\n - AUTO_PRESERVE_ORDER: Azure, AWS Sagemaker/Bedrock, Private Deployments\n"
149149
x-fern-audiences:
150150
- public
151+
raw_prompting:
152+
type: boolean
153+
description: "When enabled, the user's prompt will be sent to the model without\nany pre-processing.\n\nCompatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments\n"
154+
x-fern-audiences:
155+
- sdk-only
151156
response_format:
152157
$ref: '#/components/schemas/ResponseFormat'
153158
safety_mode:
@@ -7440,6 +7445,11 @@ paths:
74407445
format: float
74417446
x-fern-audiences:
74427447
- public
7448+
raw_prompting:
7449+
type: boolean
7450+
description: "When enabled, the user's prompt will be sent to the model without\nany pre-processing.\n\nCompatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments\n"
7451+
x-fern-audiences:
7452+
- sdk-only
74437453
reasoning_effort:
74447454
$ref: '#/components/schemas/ReasoningEffort'
74457455
response_format:

0 commit comments

Comments
 (0)