Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -89,8 +89,6 @@ namespace Together
typeof(global::Together.JsonConverters.ChatCompletionToolMessageParamRoleNullableJsonConverter),
typeof(global::Together.JsonConverters.ChatCompletionFunctionMessageParamRoleJsonConverter),
typeof(global::Together.JsonConverters.ChatCompletionFunctionMessageParamRoleNullableJsonConverter),
typeof(global::Together.JsonConverters.ChatCompletionRequestModelJsonConverter),
typeof(global::Together.JsonConverters.ChatCompletionRequestModelNullableJsonConverter),
typeof(global::Together.JsonConverters.ChatCompletionRequestContextLengthExceededBehaviorJsonConverter),
typeof(global::Together.JsonConverters.ChatCompletionRequestContextLengthExceededBehaviorNullableJsonConverter),
typeof(global::Together.JsonConverters.ChatCompletionRequestFunctionCallEnumJsonConverter),
Expand Down Expand Up @@ -300,7 +298,6 @@ namespace Together
typeof(global::Together.JsonConverters.AllOfJsonConverter<global::Together.FinishReason?, object>),
typeof(global::Together.JsonConverters.AllOfJsonConverter<object, global::Together.LogprobsPart>),
typeof(global::Together.JsonConverters.OneOfJsonConverter<global::Together.ChatCompletionUserMessageContentMultimodalItemVariant1, global::Together.ChatCompletionUserMessageContentMultimodalItemVariant2, global::Together.ChatCompletionUserMessageContentMultimodalItemVideo, global::Together.ChatCompletionUserMessageContentMultimodalItemAudio, global::Together.ChatCompletionUserMessageContentMultimodalItemInputAudio>),
typeof(global::Together.JsonConverters.AnyOfJsonConverter<global::Together.ChatCompletionRequestModel?, string>),
typeof(global::Together.JsonConverters.OneOfJsonConverter<global::Together.ChatCompletionRequestFunctionCallEnum?, global::Together.ChatCompletionRequestFunctionCallEnum2>),
typeof(global::Together.JsonConverters.OneOfJsonConverter<string, global::Together.ToolChoice2>),
typeof(global::Together.JsonConverters.AllOfJsonConverter<global::Together.UsageData, object>),
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -428,8 +428,7 @@ partial void ProcessChatCompletionsResponseContent(
/// A list of messages comprising the conversation so far.
/// </param>
/// <param name="model">
/// The name of the model to query. [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)<br/>
/// Example: Qwen/Qwen3.5-9B
/// The name of the model to query. [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)
/// </param>
/// <param name="maxTokens">
/// The maximum number of tokens to generate.
Expand Down Expand Up @@ -515,7 +514,7 @@ partial void ProcessChatCompletionsResponseContent(
/// <exception cref="global::System.InvalidOperationException"></exception>
public async global::System.Threading.Tasks.Task<global::Together.ChatCompletionResponse> ChatCompletionsAsync(
global::System.Collections.Generic.IList<global::Together.ChatCompletionMessageParam> messages,
global::Together.AnyOf<global::Together.ChatCompletionRequestModel?, string> model,
string model,
int? maxTokens = default,
global::System.Collections.Generic.IList<string>? stop = default,
float? temperature = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -169,8 +169,7 @@ partial void ProcessChatCompletionsAsStreamResponse(
/// A list of messages comprising the conversation so far.
/// </param>
/// <param name="model">
/// The name of the model to query. [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)<br/>
/// Example: Qwen/Qwen3.5-9B
/// The name of the model to query. [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)
/// </param>
/// <param name="maxTokens">
/// The maximum number of tokens to generate.
Expand Down Expand Up @@ -256,7 +255,7 @@ partial void ProcessChatCompletionsAsStreamResponse(
/// <exception cref="global::System.InvalidOperationException"></exception>
public async global::System.Collections.Generic.IAsyncEnumerable<global::Together.ChatCompletionStream> ChatCompletionsAsStreamAsync(
global::System.Collections.Generic.IList<global::Together.ChatCompletionMessageParam> messages,
global::Together.AnyOf<global::Together.ChatCompletionRequestModel?, string> model,
string model,
int? maxTokens = default,
global::System.Collections.Generic.IList<string>? stop = default,
float? temperature = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -246,13 +246,16 @@ partial void ProcessCreateEndpointResponseContent(
/// Creates a new dedicated endpoint for serving models. The endpoint will automatically start after creation. You can deploy any supported model on hardware configurations that meet the model's requirements.
/// </summary>
/// <param name="displayName">
/// A human-readable name for the endpoint
/// A human-readable name for the endpoint<br/>
/// Example: My Llama3 70b endpoint
/// </param>
/// <param name="model">
/// The model to deploy on this endpoint
/// The model to deploy on this endpoint<br/>
/// Example: deepseek-ai/DeepSeek-R1
/// </param>
/// <param name="hardware">
/// The hardware configuration to use for this endpoint
/// The hardware configuration to use for this endpoint<br/>
/// Example: 1x_nvidia_a100_80gb_sxm
/// </param>
/// <param name="autoscaling">
/// Configuration for automatic scaling of the endpoint
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,8 @@ partial void ProcessListHardwareResponseContent(
/// <param name="model">
/// Filter hardware configurations by model compatibility. When provided,<br/>
/// the response includes availability status for each compatible configuration.<br/>
/// Example: meta-llama/Llama-3-70b-chat-hf
/// [See all of Together AI's dedicated models](https://docs.together.ai/docs/dedicated-models)<br/>
/// Example: deepseek-ai/DeepSeek-R1
/// </param>
/// <param name="cancellationToken">The token to cancel the operation with</param>
/// <exception cref="global::Together.ApiException"></exception>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,8 +24,7 @@ public partial interface IChatClient
/// A list of messages comprising the conversation so far.
/// </param>
/// <param name="model">
/// The name of the model to query. [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)<br/>
/// Example: Qwen/Qwen3.5-9B
/// The name of the model to query. [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)
/// </param>
/// <param name="maxTokens">
/// The maximum number of tokens to generate.
Expand Down Expand Up @@ -111,7 +110,7 @@ public partial interface IChatClient
/// <exception cref="global::System.InvalidOperationException"></exception>
global::System.Threading.Tasks.Task<global::Together.ChatCompletionResponse> ChatCompletionsAsync(
global::System.Collections.Generic.IList<global::Together.ChatCompletionMessageParam> messages,
global::Together.AnyOf<global::Together.ChatCompletionRequestModel?, string> model,
string model,
int? maxTokens = default,
global::System.Collections.Generic.IList<string>? stop = default,
float? temperature = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,8 +24,7 @@ public partial interface IChatClient
/// A list of messages comprising the conversation so far.
/// </param>
/// <param name="model">
/// The name of the model to query. [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)<br/>
/// Example: Qwen/Qwen3.5-9B
/// The name of the model to query. [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)
/// </param>
/// <param name="maxTokens">
/// The maximum number of tokens to generate.
Expand Down Expand Up @@ -111,7 +110,7 @@ public partial interface IChatClient
/// <exception cref="global::System.InvalidOperationException"></exception>
global::System.Collections.Generic.IAsyncEnumerable<global::Together.ChatCompletionStream> ChatCompletionsAsStreamAsync(
global::System.Collections.Generic.IList<global::Together.ChatCompletionMessageParam> messages,
global::Together.AnyOf<global::Together.ChatCompletionRequestModel?, string> model,
string model,
int? maxTokens = default,
global::System.Collections.Generic.IList<string>? stop = default,
float? temperature = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,13 +21,16 @@ public partial interface IEndpointsClient
/// Creates a new dedicated endpoint for serving models. The endpoint will automatically start after creation. You can deploy any supported model on hardware configurations that meet the model's requirements.
/// </summary>
/// <param name="displayName">
/// A human-readable name for the endpoint
/// A human-readable name for the endpoint<br/>
/// Example: My Llama3 70b endpoint
/// </param>
/// <param name="model">
/// The model to deploy on this endpoint
/// The model to deploy on this endpoint<br/>
/// Example: deepseek-ai/DeepSeek-R1
/// </param>
/// <param name="hardware">
/// The hardware configuration to use for this endpoint
/// The hardware configuration to use for this endpoint<br/>
/// Example: 1x_nvidia_a100_80gb_sxm
/// </param>
/// <param name="autoscaling">
/// Configuration for automatic scaling of the endpoint
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,8 @@ public partial interface IHardwareClient
/// <param name="model">
/// Filter hardware configurations by model compatibility. When provided,<br/>
/// the response includes availability status for each compatible configuration.<br/>
/// Example: meta-llama/Llama-3-70b-chat-hf
/// [See all of Together AI's dedicated models](https://docs.together.ai/docs/dedicated-models)<br/>
/// Example: deepseek-ai/DeepSeek-R1
/// </param>
/// <param name="cancellationToken">The token to cancel the operation with</param>
/// <exception cref="global::Together.ApiException"></exception>
Expand Down

This file was deleted.

This file was deleted.

Loading
Loading