diff --git a/src/collections/config/types/generative.ts b/src/collections/config/types/generative.ts index ab25fc56..667bc347 100644 --- a/src/collections/config/types/generative.ts +++ b/src/collections/config/types/generative.ts @@ -58,13 +58,6 @@ export type GenerativeMistralConfig = { temperature?: number; }; -export type GenerativeOctoAIConfig = { - baseURL?: string; - maxTokens?: number; - model?: string; - temperature?: number; -}; - export type GenerativeOllamaConfig = { apiEndpoint?: string; model?: string; @@ -102,7 +95,6 @@ export type GenerativeConfig = | GenerativeGoogleConfig | GenerativeFriendliAIConfig | GenerativeMistralConfig - | GenerativeOctoAIConfig | GenerativeOllamaConfig | GenerativeOpenAIConfig | GenerativePaLMConfig @@ -127,8 +119,6 @@ export type GenerativeConfigType = G extends 'generative-anthropic' ? GenerativeFriendliAIConfig : G extends 'generative-mistral' ? GenerativeMistralConfig - : G extends 'generative-octoai' - ? GenerativeOctoAIConfig : G extends 'generative-ollama' ? GenerativeOllamaConfig : G extends 'generative-openai' @@ -152,7 +142,6 @@ export type GenerativeSearch = | 'generative-google' | 'generative-friendliai' | 'generative-mistral' - | 'generative-octoai' | 'generative-ollama' | 'generative-openai' | GenerativePalm diff --git a/src/collections/config/types/vectorizer.ts b/src/collections/config/types/vectorizer.ts index c2257e5a..a0a2872e 100644 --- a/src/collections/config/types/vectorizer.ts +++ b/src/collections/config/types/vectorizer.ts @@ -33,7 +33,6 @@ export type Vectorizer = | 'text2vec-huggingface' | 'text2vec-jina' | 'text2vec-mistral' - | 'text2vec-octoai' | 'text2vec-ollama' | 'text2vec-openai' | Text2VecPalmVectorizer @@ -290,20 +289,6 @@ export type Text2VecMistralConfig = { vectorizeCollectionName?: boolean; }; -/** - * The configuration for text vectorization using the OctoAI module. - * - * See the [documentation](https://weaviate.io/developers/weaviate/model-providers/octoai/embeddings) for detailed usage. - */ -export type Text2VecOctoAIConfig = { - /** The base URL to use where API requests should go. */ - baseURL?: string; - /** The model to use. */ - model?: string; - /** Whether to vectorize the collection name. */ - vectorizeCollectionName?: boolean; -}; - /** * The configuration for text vectorization using the Ollama module. * @@ -447,8 +432,6 @@ export type VectorizerConfigType = V extends 'img2vec-neural' ? Text2VecJinaConfig | undefined : V extends 'text2vec-mistral' ? Text2VecMistralConfig | undefined - : V extends 'text2vec-octoai' - ? Text2VecOctoAIConfig | undefined : V extends 'text2vec-ollama' ? Text2VecOllamaConfig | undefined : V extends 'text2vec-openai' diff --git a/src/collections/configure/generative.ts b/src/collections/configure/generative.ts index e8d66ded..730f2bcb 100644 --- a/src/collections/configure/generative.ts +++ b/src/collections/configure/generative.ts @@ -8,7 +8,6 @@ import { GenerativeFriendliAIConfig, GenerativeGoogleConfig, GenerativeMistralConfig, - GenerativeOctoAIConfig, GenerativeOllamaConfig, GenerativeOpenAIConfig, GenerativePaLMConfig, @@ -23,7 +22,6 @@ import { GenerativeDatabricksConfigCreate, GenerativeFriendliAIConfigCreate, GenerativeMistralConfigCreate, - GenerativeOctoAIConfigCreate, GenerativeOllamaConfigCreate, GenerativeOpenAIConfigCreate, GenerativePaLMConfigCreate, @@ -171,22 +169,6 @@ export default { config, }; }, - /** - * Create a `ModuleConfig<'generative-octoai', GenerativeOpenAIConfig | undefined>` object for use when performing AI generation using the `generative-octoai` module. - * - * See the [documentation](https://weaviate.io/developers/weaviate/model-providers/octoai/generative) for detailed usage. - * - * @param {GenerativeOctoAIConfigCreate} [config] The configuration for the `generative-octoai` module. - * @returns {ModuleConfig<'generative-octoai', GenerativeOctoAIConfig | undefined>} The configuration object. - */ - octoai( - config?: GenerativeOctoAIConfigCreate - ): ModuleConfig<'generative-octoai', GenerativeOctoAIConfig | undefined> { - return { - name: 'generative-octoai', - config, - }; - }, /** * Create a `ModuleConfig<'generative-ollama', GenerativeOllamaConfig | undefined>` object for use when performing AI generation using the `generative-ollama` module. * diff --git a/src/collections/configure/types/generative.ts b/src/collections/configure/types/generative.ts index 59df7f88..2b1a18cf 100644 --- a/src/collections/configure/types/generative.ts +++ b/src/collections/configure/types/generative.ts @@ -5,7 +5,6 @@ import { GenerativeDatabricksConfig, GenerativeFriendliAIConfig, GenerativeMistralConfig, - GenerativeOctoAIConfig, GenerativeOllamaConfig, GenerativePaLMConfig, } from '../../index.js'; @@ -45,8 +44,6 @@ export type GenerativeFriendliAIConfigCreate = GenerativeFriendliAIConfig; export type GenerativeMistralConfigCreate = GenerativeMistralConfig; -export type GenerativeOctoAIConfigCreate = GenerativeOctoAIConfig; - export type GenerativeOllamaConfigCreate = GenerativeOllamaConfig; export type GenerativeOpenAIConfigCreate = GenerativeOpenAIConfigBaseCreate & { @@ -64,7 +61,6 @@ export type GenerativeConfigCreate = | GenerativeDatabricksConfigCreate | GenerativeFriendliAIConfigCreate | GenerativeMistralConfigCreate - | GenerativeOctoAIConfigCreate | GenerativeOllamaConfigCreate | GenerativeOpenAIConfigCreate | GenerativePaLMConfigCreate @@ -85,8 +81,6 @@ export type GenerativeConfigCreateType = G extends 'generative-anthropic' ? GenerativeFriendliAIConfigCreate : G extends 'generative-mistral' ? GenerativeMistralConfigCreate - : G extends 'generative-octoai' - ? GenerativeOctoAIConfigCreate : G extends 'generative-ollama' ? GenerativeOllamaConfigCreate : G extends 'generative-openai' diff --git a/src/collections/configure/types/vectorizer.ts b/src/collections/configure/types/vectorizer.ts index e30e5e07..bba7877e 100644 --- a/src/collections/configure/types/vectorizer.ts +++ b/src/collections/configure/types/vectorizer.ts @@ -13,7 +13,6 @@ import { Text2VecHuggingFaceConfig, Text2VecJinaConfig, Text2VecMistralConfig, - Text2VecOctoAIConfig, Text2VecOllamaConfig, Text2VecOpenAIConfig, Text2VecTransformersConfig, @@ -154,8 +153,6 @@ export type Text2VecJinaConfigCreate = Text2VecJinaConfig; export type Text2VecMistralConfigCreate = Text2VecMistralConfig; -export type Text2VecOctoAIConfigCreate = Text2VecOctoAIConfig; - export type Text2VecOllamaConfigCreate = Text2VecOllamaConfig; export type Text2VecOpenAIConfigCreate = Text2VecOpenAIConfig; @@ -197,8 +194,6 @@ export type VectorizerConfigCreateType = V extends 'img2vec-neural' ? Text2VecJinaConfigCreate | undefined : V extends 'text2vec-mistral' ? Text2VecMistralConfigCreate | undefined - : V extends 'text2vec-octoai' - ? Text2VecOctoAIConfigCreate | undefined : V extends 'text2vec-ollama' ? Text2VecOllamaConfigCreate | undefined : V extends 'text2vec-openai' diff --git a/src/collections/configure/unit.test.ts b/src/collections/configure/unit.test.ts index 5291dc09..6ff632ad 100644 --- a/src/collections/configure/unit.test.ts +++ b/src/collections/configure/unit.test.ts @@ -8,7 +8,6 @@ import { GenerativeFriendliAIConfig, GenerativeGoogleConfig, GenerativeMistralConfig, - GenerativeOctoAIConfig, GenerativeOllamaConfig, GenerativeOpenAIConfig, ModuleConfig, @@ -1025,45 +1024,6 @@ describe('Unit testing of the vectorizer factory class', () => { }); }); - it('should create the correct Text2VecOctoAIConfig type with defaults', () => { - const config = configure.vectorizer.text2VecOctoAI(); - expect(config).toEqual>({ - name: undefined, - vectorIndex: { - name: 'hnsw', - config: undefined, - }, - vectorizer: { - name: 'text2vec-octoai', - config: undefined, - }, - }); - }); - - it('should create the correct Text2VecOctoAIConfig type with all values', () => { - const config = configure.vectorizer.text2VecOctoAI({ - name: 'test', - baseURL: 'base-url', - model: 'model', - vectorizeCollectionName: true, - }); - expect(config).toEqual>({ - name: 'test', - vectorIndex: { - name: 'hnsw', - config: undefined, - }, - vectorizer: { - name: 'text2vec-octoai', - config: { - baseURL: 'base-url', - model: 'model', - vectorizeCollectionName: true, - }, - }, - }); - }); - it('should create the correct Text2VecOllamaConfig type with defaults', () => { const config = configure.vectorizer.text2VecOllama(); expect(config).toEqual>({ @@ -1546,32 +1506,6 @@ describe('Unit testing of the generative factory class', () => { }); }); - it('should create the correct GenerativeOctoAIConfig type with required & default values', () => { - const config = configure.generative.octoai(); - expect(config).toEqual>({ - name: 'generative-octoai', - config: undefined, - }); - }); - - it('should create the correct GenerativeOctoAIConfig type with all values', () => { - const config = configure.generative.octoai({ - baseURL: 'base-url', - maxTokens: 100, - model: 'model', - temperature: 0.5, - }); - expect(config).toEqual>({ - name: 'generative-octoai', - config: { - baseURL: 'base-url', - maxTokens: 100, - model: 'model', - temperature: 0.5, - }, - }); - }); - it('should create the correct GenerativeOllamaConfig type with required & default values', () => { const config = configure.generative.ollama(); expect(config).toEqual>({ diff --git a/src/collections/configure/vectorizer.ts b/src/collections/configure/vectorizer.ts index f3f53367..3ad042df 100644 --- a/src/collections/configure/vectorizer.ts +++ b/src/collections/configure/vectorizer.ts @@ -439,27 +439,6 @@ export const vectorizer = { }, }); }, - /** - * Create a `VectorConfigCreate` object with the vectorizer set to `'text2vec-octoai'`. - * - * See the [documentation](https://weaviate.io/developers/weaviate/model-providers/octoai/embeddings) for detailed usage. - * - * @param {ConfigureTextVectorizerOptions} [opts] The configuration for the `text2vec-octoai` vectorizer. - * @returns {VectorConfigCreate, N, I, 'text2vec-octoai'>} The configuration object. - */ - text2VecOctoAI: ( - opts?: ConfigureTextVectorizerOptions - ): VectorConfigCreate, N, I, 'text2vec-octoai'> => { - const { name, sourceProperties, vectorIndexConfig, ...config } = opts || {}; - return makeVectorizer(name, { - sourceProperties, - vectorIndexConfig, - vectorizerConfig: { - name: 'text2vec-octoai', - config: Object.keys(config).length === 0 ? undefined : config, - }, - }); - }, /** * Create a `VectorConfigCreate` object with the vectorizer set to `'text2vec-openai'`. * diff --git a/src/proto/v1/generative.ts b/src/proto/v1/generative.ts index 3d61af30..12b1619f 100644 --- a/src/proto/v1/generative.ts +++ b/src/proto/v1/generative.ts @@ -31,7 +31,11 @@ export interface GenerativeSearch_Single { export interface GenerativeSearch_Grouped { task: string; - properties?: TextArray | undefined; + properties?: + | TextArray + | undefined; + /** only allow one at the beginning, but multiple in the future */ + queries: GenerativeProvider[]; } export interface GenerativeProvider { @@ -42,10 +46,11 @@ export interface GenerativeProvider { cohere?: GenerativeCohere | undefined; dummy?: GenerativeDummy | undefined; mistral?: GenerativeMistral | undefined; - octoai?: GenerativeOctoAI | undefined; ollama?: GenerativeOllama | undefined; openai?: GenerativeOpenAI | undefined; google?: GenerativeGoogle | undefined; + databricks?: GenerativeDatabricks | undefined; + friendliai?: GenerativeFriendliAI | undefined; } export interface GenerativeAnthropic { @@ -67,6 +72,11 @@ export interface GenerativeAnyscale { export interface GenerativeAWS { model?: string | undefined; temperature?: number | undefined; + service?: string | undefined; + region?: string | undefined; + endpoint?: string | undefined; + targetModel?: string | undefined; + targetVariant?: string | undefined; } export interface GenerativeCohere { @@ -92,15 +102,6 @@ export interface GenerativeMistral { topP?: number | undefined; } -export interface GenerativeOctoAI { - baseUrl?: string | undefined; - maxTokens?: number | undefined; - model?: string | undefined; - n?: number | undefined; - temperature?: number | undefined; - topP?: number | undefined; -} - export interface GenerativeOllama { apiEndpoint?: string | undefined; model?: string | undefined; @@ -109,7 +110,6 @@ export interface GenerativeOllama { export interface GenerativeOpenAI { frequencyPenalty?: number | undefined; - logProbs?: boolean | undefined; maxTokens?: number | undefined; model: string; n?: number | undefined; @@ -117,7 +117,11 @@ export interface GenerativeOpenAI { stop?: TextArray | undefined; temperature?: number | undefined; topP?: number | undefined; - topLogProbs?: number | undefined; + baseUrl?: string | undefined; + apiVersion?: string | undefined; + resourceName?: string | undefined; + deploymentId?: string | undefined; + isAzure?: boolean | undefined; } export interface GenerativeGoogle { @@ -129,6 +133,33 @@ export interface GenerativeGoogle { topK?: number | undefined; topP?: number | undefined; stopSequences?: TextArray | undefined; + apiEndpoint?: string | undefined; + projectId?: string | undefined; + endpointId?: string | undefined; + region?: string | undefined; +} + +export interface GenerativeDatabricks { + endpoint?: string | undefined; + model?: string | undefined; + frequencyPenalty?: number | undefined; + logProbs?: boolean | undefined; + topLogProbs?: number | undefined; + maxTokens?: number | undefined; + n?: number | undefined; + presencePenalty?: number | undefined; + stop?: TextArray | undefined; + temperature?: number | undefined; + topP?: number | undefined; +} + +export interface GenerativeFriendliAI { + baseUrl?: string | undefined; + model?: string | undefined; + maxTokens?: number | undefined; + temperature?: number | undefined; + n?: number | undefined; + topP?: number | undefined; } export interface GenerativeAnthropicMetadata { @@ -184,16 +215,6 @@ export interface GenerativeMistralMetadata_Usage { totalTokens?: number | undefined; } -export interface GenerativeOctoAIMetadata { - usage?: GenerativeOctoAIMetadata_Usage | undefined; -} - -export interface GenerativeOctoAIMetadata_Usage { - promptTokens?: number | undefined; - completionTokens?: number | undefined; - totalTokens?: number | undefined; -} - export interface GenerativeOllamaMetadata { } @@ -232,6 +253,26 @@ export interface GenerativeGoogleMetadata_UsageMetadata { totalTokenCount?: number | undefined; } +export interface GenerativeDatabricksMetadata { + usage?: GenerativeDatabricksMetadata_Usage | undefined; +} + +export interface GenerativeDatabricksMetadata_Usage { + promptTokens?: number | undefined; + completionTokens?: number | undefined; + totalTokens?: number | undefined; +} + +export interface GenerativeFriendliAIMetadata { + usage?: GenerativeFriendliAIMetadata_Usage | undefined; +} + +export interface GenerativeFriendliAIMetadata_Usage { + promptTokens?: number | undefined; + completionTokens?: number | undefined; + totalTokens?: number | undefined; +} + export interface GenerativeMetadata { anthropic?: GenerativeAnthropicMetadata | undefined; anyscale?: GenerativeAnyscaleMetadata | undefined; @@ -239,10 +280,11 @@ export interface GenerativeMetadata { cohere?: GenerativeCohereMetadata | undefined; dummy?: GenerativeDummyMetadata | undefined; mistral?: GenerativeMistralMetadata | undefined; - octoai?: GenerativeOctoAIMetadata | undefined; ollama?: GenerativeOllamaMetadata | undefined; openai?: GenerativeOpenAIMetadata | undefined; google?: GenerativeGoogleMetadata | undefined; + databricks?: GenerativeDatabricksMetadata | undefined; + friendliai?: GenerativeFriendliAIMetadata | undefined; } export interface GenerativeReply { @@ -482,7 +524,7 @@ export const GenerativeSearch_Single = { }; function createBaseGenerativeSearch_Grouped(): GenerativeSearch_Grouped { - return { task: "", properties: undefined }; + return { task: "", properties: undefined, queries: [] }; } export const GenerativeSearch_Grouped = { @@ -493,6 +535,9 @@ export const GenerativeSearch_Grouped = { if (message.properties !== undefined) { TextArray.encode(message.properties, writer.uint32(18).fork()).ldelim(); } + for (const v of message.queries) { + GenerativeProvider.encode(v!, writer.uint32(26).fork()).ldelim(); + } return writer; }, @@ -517,6 +562,13 @@ export const GenerativeSearch_Grouped = { message.properties = TextArray.decode(reader, reader.uint32()); continue; + case 3: + if (tag !== 26) { + break; + } + + message.queries.push(GenerativeProvider.decode(reader, reader.uint32())); + continue; } if ((tag & 7) === 4 || tag === 0) { break; @@ -530,6 +582,9 @@ export const GenerativeSearch_Grouped = { return { task: isSet(object.task) ? globalThis.String(object.task) : "", properties: isSet(object.properties) ? TextArray.fromJSON(object.properties) : undefined, + queries: globalThis.Array.isArray(object?.queries) + ? object.queries.map((e: any) => GenerativeProvider.fromJSON(e)) + : [], }; }, @@ -541,6 +596,9 @@ export const GenerativeSearch_Grouped = { if (message.properties !== undefined) { obj.properties = TextArray.toJSON(message.properties); } + if (message.queries?.length) { + obj.queries = message.queries.map((e) => GenerativeProvider.toJSON(e)); + } return obj; }, @@ -553,6 +611,7 @@ export const GenerativeSearch_Grouped = { message.properties = (object.properties !== undefined && object.properties !== null) ? TextArray.fromPartial(object.properties) : undefined; + message.queries = object.queries?.map((e) => GenerativeProvider.fromPartial(e)) || []; return message; }, }; @@ -566,10 +625,11 @@ function createBaseGenerativeProvider(): GenerativeProvider { cohere: undefined, dummy: undefined, mistral: undefined, - octoai: undefined, ollama: undefined, openai: undefined, google: undefined, + databricks: undefined, + friendliai: undefined, }; } @@ -596,17 +656,20 @@ export const GenerativeProvider = { if (message.mistral !== undefined) { GenerativeMistral.encode(message.mistral, writer.uint32(58).fork()).ldelim(); } - if (message.octoai !== undefined) { - GenerativeOctoAI.encode(message.octoai, writer.uint32(66).fork()).ldelim(); - } if (message.ollama !== undefined) { - GenerativeOllama.encode(message.ollama, writer.uint32(74).fork()).ldelim(); + GenerativeOllama.encode(message.ollama, writer.uint32(66).fork()).ldelim(); } if (message.openai !== undefined) { - GenerativeOpenAI.encode(message.openai, writer.uint32(82).fork()).ldelim(); + GenerativeOpenAI.encode(message.openai, writer.uint32(74).fork()).ldelim(); } if (message.google !== undefined) { - GenerativeGoogle.encode(message.google, writer.uint32(90).fork()).ldelim(); + GenerativeGoogle.encode(message.google, writer.uint32(82).fork()).ldelim(); + } + if (message.databricks !== undefined) { + GenerativeDatabricks.encode(message.databricks, writer.uint32(90).fork()).ldelim(); + } + if (message.friendliai !== undefined) { + GenerativeFriendliAI.encode(message.friendliai, writer.uint32(98).fork()).ldelim(); } return writer; }, @@ -672,28 +735,35 @@ export const GenerativeProvider = { break; } - message.octoai = GenerativeOctoAI.decode(reader, reader.uint32()); + message.ollama = GenerativeOllama.decode(reader, reader.uint32()); continue; case 9: if (tag !== 74) { break; } - message.ollama = GenerativeOllama.decode(reader, reader.uint32()); + message.openai = GenerativeOpenAI.decode(reader, reader.uint32()); continue; case 10: if (tag !== 82) { break; } - message.openai = GenerativeOpenAI.decode(reader, reader.uint32()); + message.google = GenerativeGoogle.decode(reader, reader.uint32()); continue; case 11: if (tag !== 90) { break; } - message.google = GenerativeGoogle.decode(reader, reader.uint32()); + message.databricks = GenerativeDatabricks.decode(reader, reader.uint32()); + continue; + case 12: + if (tag !== 98) { + break; + } + + message.friendliai = GenerativeFriendliAI.decode(reader, reader.uint32()); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -713,10 +783,11 @@ export const GenerativeProvider = { cohere: isSet(object.cohere) ? GenerativeCohere.fromJSON(object.cohere) : undefined, dummy: isSet(object.dummy) ? GenerativeDummy.fromJSON(object.dummy) : undefined, mistral: isSet(object.mistral) ? GenerativeMistral.fromJSON(object.mistral) : undefined, - octoai: isSet(object.octoai) ? GenerativeOctoAI.fromJSON(object.octoai) : undefined, ollama: isSet(object.ollama) ? GenerativeOllama.fromJSON(object.ollama) : undefined, openai: isSet(object.openai) ? GenerativeOpenAI.fromJSON(object.openai) : undefined, google: isSet(object.google) ? GenerativeGoogle.fromJSON(object.google) : undefined, + databricks: isSet(object.databricks) ? GenerativeDatabricks.fromJSON(object.databricks) : undefined, + friendliai: isSet(object.friendliai) ? GenerativeFriendliAI.fromJSON(object.friendliai) : undefined, }; }, @@ -743,9 +814,6 @@ export const GenerativeProvider = { if (message.mistral !== undefined) { obj.mistral = GenerativeMistral.toJSON(message.mistral); } - if (message.octoai !== undefined) { - obj.octoai = GenerativeOctoAI.toJSON(message.octoai); - } if (message.ollama !== undefined) { obj.ollama = GenerativeOllama.toJSON(message.ollama); } @@ -755,6 +823,12 @@ export const GenerativeProvider = { if (message.google !== undefined) { obj.google = GenerativeGoogle.toJSON(message.google); } + if (message.databricks !== undefined) { + obj.databricks = GenerativeDatabricks.toJSON(message.databricks); + } + if (message.friendliai !== undefined) { + obj.friendliai = GenerativeFriendliAI.toJSON(message.friendliai); + } return obj; }, @@ -780,9 +854,6 @@ export const GenerativeProvider = { message.mistral = (object.mistral !== undefined && object.mistral !== null) ? GenerativeMistral.fromPartial(object.mistral) : undefined; - message.octoai = (object.octoai !== undefined && object.octoai !== null) - ? GenerativeOctoAI.fromPartial(object.octoai) - : undefined; message.ollama = (object.ollama !== undefined && object.ollama !== null) ? GenerativeOllama.fromPartial(object.ollama) : undefined; @@ -792,6 +863,12 @@ export const GenerativeProvider = { message.google = (object.google !== undefined && object.google !== null) ? GenerativeGoogle.fromPartial(object.google) : undefined; + message.databricks = (object.databricks !== undefined && object.databricks !== null) + ? GenerativeDatabricks.fromPartial(object.databricks) + : undefined; + message.friendliai = (object.friendliai !== undefined && object.friendliai !== null) + ? GenerativeFriendliAI.fromPartial(object.friendliai) + : undefined; return message; }, }; @@ -1045,7 +1122,15 @@ export const GenerativeAnyscale = { }; function createBaseGenerativeAWS(): GenerativeAWS { - return { model: undefined, temperature: undefined }; + return { + model: undefined, + temperature: undefined, + service: undefined, + region: undefined, + endpoint: undefined, + targetModel: undefined, + targetVariant: undefined, + }; } export const GenerativeAWS = { @@ -1056,6 +1141,21 @@ export const GenerativeAWS = { if (message.temperature !== undefined) { writer.uint32(65).double(message.temperature); } + if (message.service !== undefined) { + writer.uint32(74).string(message.service); + } + if (message.region !== undefined) { + writer.uint32(82).string(message.region); + } + if (message.endpoint !== undefined) { + writer.uint32(90).string(message.endpoint); + } + if (message.targetModel !== undefined) { + writer.uint32(98).string(message.targetModel); + } + if (message.targetVariant !== undefined) { + writer.uint32(106).string(message.targetVariant); + } return writer; }, @@ -1080,6 +1180,41 @@ export const GenerativeAWS = { message.temperature = reader.double(); continue; + case 9: + if (tag !== 74) { + break; + } + + message.service = reader.string(); + continue; + case 10: + if (tag !== 82) { + break; + } + + message.region = reader.string(); + continue; + case 11: + if (tag !== 90) { + break; + } + + message.endpoint = reader.string(); + continue; + case 12: + if (tag !== 98) { + break; + } + + message.targetModel = reader.string(); + continue; + case 13: + if (tag !== 106) { + break; + } + + message.targetVariant = reader.string(); + continue; } if ((tag & 7) === 4 || tag === 0) { break; @@ -1093,6 +1228,11 @@ export const GenerativeAWS = { return { model: isSet(object.model) ? globalThis.String(object.model) : undefined, temperature: isSet(object.temperature) ? globalThis.Number(object.temperature) : undefined, + service: isSet(object.service) ? globalThis.String(object.service) : undefined, + region: isSet(object.region) ? globalThis.String(object.region) : undefined, + endpoint: isSet(object.endpoint) ? globalThis.String(object.endpoint) : undefined, + targetModel: isSet(object.targetModel) ? globalThis.String(object.targetModel) : undefined, + targetVariant: isSet(object.targetVariant) ? globalThis.String(object.targetVariant) : undefined, }; }, @@ -1104,6 +1244,21 @@ export const GenerativeAWS = { if (message.temperature !== undefined) { obj.temperature = message.temperature; } + if (message.service !== undefined) { + obj.service = message.service; + } + if (message.region !== undefined) { + obj.region = message.region; + } + if (message.endpoint !== undefined) { + obj.endpoint = message.endpoint; + } + if (message.targetModel !== undefined) { + obj.targetModel = message.targetModel; + } + if (message.targetVariant !== undefined) { + obj.targetVariant = message.targetVariant; + } return obj; }, @@ -1114,6 +1269,11 @@ export const GenerativeAWS = { const message = createBaseGenerativeAWS(); message.model = object.model ?? undefined; message.temperature = object.temperature ?? undefined; + message.service = object.service ?? undefined; + message.region = object.region ?? undefined; + message.endpoint = object.endpoint ?? undefined; + message.targetModel = object.targetModel ?? undefined; + message.targetVariant = object.targetVariant ?? undefined; return message; }, }; @@ -1471,44 +1631,28 @@ export const GenerativeMistral = { }, }; -function createBaseGenerativeOctoAI(): GenerativeOctoAI { - return { - baseUrl: undefined, - maxTokens: undefined, - model: undefined, - n: undefined, - temperature: undefined, - topP: undefined, - }; +function createBaseGenerativeOllama(): GenerativeOllama { + return { apiEndpoint: undefined, model: undefined, temperature: undefined }; } -export const GenerativeOctoAI = { - encode(message: GenerativeOctoAI, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.baseUrl !== undefined) { - writer.uint32(10).string(message.baseUrl); - } - if (message.maxTokens !== undefined) { - writer.uint32(16).int64(message.maxTokens); +export const GenerativeOllama = { + encode(message: GenerativeOllama, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.apiEndpoint !== undefined) { + writer.uint32(10).string(message.apiEndpoint); } if (message.model !== undefined) { - writer.uint32(26).string(message.model); - } - if (message.n !== undefined) { - writer.uint32(32).int64(message.n); + writer.uint32(18).string(message.model); } if (message.temperature !== undefined) { - writer.uint32(41).double(message.temperature); - } - if (message.topP !== undefined) { - writer.uint32(49).double(message.topP); + writer.uint32(25).double(message.temperature); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOctoAI { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOllama { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeOctoAI(); + const message = createBaseGenerativeOllama(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { @@ -1517,43 +1661,22 @@ export const GenerativeOctoAI = { break; } - message.baseUrl = reader.string(); + message.apiEndpoint = reader.string(); continue; case 2: - if (tag !== 16) { - break; - } - - message.maxTokens = longToNumber(reader.int64() as Long); - continue; - case 3: - if (tag !== 26) { + if (tag !== 18) { break; } message.model = reader.string(); continue; - case 4: - if (tag !== 32) { - break; - } - - message.n = longToNumber(reader.int64() as Long); - continue; - case 5: - if (tag !== 41) { + case 3: + if (tag !== 25) { break; } message.temperature = reader.double(); continue; - case 6: - if (tag !== 49) { - break; - } - - message.topP = reader.double(); - continue; } if ((tag & 7) === 4 || tag === 0) { break; @@ -1563,190 +1686,98 @@ export const GenerativeOctoAI = { return message; }, - fromJSON(object: any): GenerativeOctoAI { + fromJSON(object: any): GenerativeOllama { return { - baseUrl: isSet(object.baseUrl) ? globalThis.String(object.baseUrl) : undefined, - maxTokens: isSet(object.maxTokens) ? globalThis.Number(object.maxTokens) : undefined, + apiEndpoint: isSet(object.apiEndpoint) ? globalThis.String(object.apiEndpoint) : undefined, model: isSet(object.model) ? globalThis.String(object.model) : undefined, - n: isSet(object.n) ? globalThis.Number(object.n) : undefined, temperature: isSet(object.temperature) ? globalThis.Number(object.temperature) : undefined, - topP: isSet(object.topP) ? globalThis.Number(object.topP) : undefined, }; }, - toJSON(message: GenerativeOctoAI): unknown { + toJSON(message: GenerativeOllama): unknown { const obj: any = {}; - if (message.baseUrl !== undefined) { - obj.baseUrl = message.baseUrl; - } - if (message.maxTokens !== undefined) { - obj.maxTokens = Math.round(message.maxTokens); + if (message.apiEndpoint !== undefined) { + obj.apiEndpoint = message.apiEndpoint; } if (message.model !== undefined) { obj.model = message.model; } - if (message.n !== undefined) { - obj.n = Math.round(message.n); - } if (message.temperature !== undefined) { obj.temperature = message.temperature; } - if (message.topP !== undefined) { - obj.topP = message.topP; - } return obj; }, - create(base?: DeepPartial): GenerativeOctoAI { - return GenerativeOctoAI.fromPartial(base ?? {}); + create(base?: DeepPartial): GenerativeOllama { + return GenerativeOllama.fromPartial(base ?? {}); }, - fromPartial(object: DeepPartial): GenerativeOctoAI { - const message = createBaseGenerativeOctoAI(); - message.baseUrl = object.baseUrl ?? undefined; - message.maxTokens = object.maxTokens ?? undefined; + fromPartial(object: DeepPartial): GenerativeOllama { + const message = createBaseGenerativeOllama(); + message.apiEndpoint = object.apiEndpoint ?? undefined; message.model = object.model ?? undefined; - message.n = object.n ?? undefined; message.temperature = object.temperature ?? undefined; - message.topP = object.topP ?? undefined; return message; }, }; -function createBaseGenerativeOllama(): GenerativeOllama { - return { apiEndpoint: undefined, model: undefined, temperature: undefined }; +function createBaseGenerativeOpenAI(): GenerativeOpenAI { + return { + frequencyPenalty: undefined, + maxTokens: undefined, + model: "", + n: undefined, + presencePenalty: undefined, + stop: undefined, + temperature: undefined, + topP: undefined, + baseUrl: undefined, + apiVersion: undefined, + resourceName: undefined, + deploymentId: undefined, + isAzure: undefined, + }; } -export const GenerativeOllama = { - encode(message: GenerativeOllama, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.apiEndpoint !== undefined) { - writer.uint32(10).string(message.apiEndpoint); +export const GenerativeOpenAI = { + encode(message: GenerativeOpenAI, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.frequencyPenalty !== undefined) { + writer.uint32(9).double(message.frequencyPenalty); } - if (message.model !== undefined) { - writer.uint32(18).string(message.model); + if (message.maxTokens !== undefined) { + writer.uint32(16).int64(message.maxTokens); } - if (message.temperature !== undefined) { - writer.uint32(25).double(message.temperature); - } - return writer; - }, - - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOllama { - const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); - let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeOllama(); - while (reader.pos < end) { - const tag = reader.uint32(); - switch (tag >>> 3) { - case 1: - if (tag !== 10) { - break; - } - - message.apiEndpoint = reader.string(); - continue; - case 2: - if (tag !== 18) { - break; - } - - message.model = reader.string(); - continue; - case 3: - if (tag !== 25) { - break; - } - - message.temperature = reader.double(); - continue; - } - if ((tag & 7) === 4 || tag === 0) { - break; - } - reader.skipType(tag & 7); - } - return message; - }, - - fromJSON(object: any): GenerativeOllama { - return { - apiEndpoint: isSet(object.apiEndpoint) ? globalThis.String(object.apiEndpoint) : undefined, - model: isSet(object.model) ? globalThis.String(object.model) : undefined, - temperature: isSet(object.temperature) ? globalThis.Number(object.temperature) : undefined, - }; - }, - - toJSON(message: GenerativeOllama): unknown { - const obj: any = {}; - if (message.apiEndpoint !== undefined) { - obj.apiEndpoint = message.apiEndpoint; - } - if (message.model !== undefined) { - obj.model = message.model; - } - if (message.temperature !== undefined) { - obj.temperature = message.temperature; - } - return obj; - }, - - create(base?: DeepPartial): GenerativeOllama { - return GenerativeOllama.fromPartial(base ?? {}); - }, - fromPartial(object: DeepPartial): GenerativeOllama { - const message = createBaseGenerativeOllama(); - message.apiEndpoint = object.apiEndpoint ?? undefined; - message.model = object.model ?? undefined; - message.temperature = object.temperature ?? undefined; - return message; - }, -}; - -function createBaseGenerativeOpenAI(): GenerativeOpenAI { - return { - frequencyPenalty: undefined, - logProbs: undefined, - maxTokens: undefined, - model: "", - n: undefined, - presencePenalty: undefined, - stop: undefined, - temperature: undefined, - topP: undefined, - topLogProbs: undefined, - }; -} - -export const GenerativeOpenAI = { - encode(message: GenerativeOpenAI, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.frequencyPenalty !== undefined) { - writer.uint32(9).double(message.frequencyPenalty); - } - if (message.logProbs !== undefined) { - writer.uint32(16).bool(message.logProbs); - } - if (message.maxTokens !== undefined) { - writer.uint32(24).int64(message.maxTokens); - } - if (message.model !== "") { - writer.uint32(34).string(message.model); + if (message.model !== "") { + writer.uint32(26).string(message.model); } if (message.n !== undefined) { - writer.uint32(40).int64(message.n); + writer.uint32(32).int64(message.n); } if (message.presencePenalty !== undefined) { - writer.uint32(49).double(message.presencePenalty); + writer.uint32(41).double(message.presencePenalty); } if (message.stop !== undefined) { - TextArray.encode(message.stop, writer.uint32(58).fork()).ldelim(); + TextArray.encode(message.stop, writer.uint32(50).fork()).ldelim(); } if (message.temperature !== undefined) { - writer.uint32(65).double(message.temperature); + writer.uint32(57).double(message.temperature); } if (message.topP !== undefined) { - writer.uint32(73).double(message.topP); + writer.uint32(65).double(message.topP); } - if (message.topLogProbs !== undefined) { - writer.uint32(80).int64(message.topLogProbs); + if (message.baseUrl !== undefined) { + writer.uint32(74).string(message.baseUrl); + } + if (message.apiVersion !== undefined) { + writer.uint32(82).string(message.apiVersion); + } + if (message.resourceName !== undefined) { + writer.uint32(90).string(message.resourceName); + } + if (message.deploymentId !== undefined) { + writer.uint32(98).string(message.deploymentId); + } + if (message.isAzure !== undefined) { + writer.uint32(104).bool(message.isAzure); } return writer; }, @@ -1770,63 +1801,84 @@ export const GenerativeOpenAI = { break; } - message.logProbs = reader.bool(); + message.maxTokens = longToNumber(reader.int64() as Long); continue; case 3: - if (tag !== 24) { + if (tag !== 26) { break; } - message.maxTokens = longToNumber(reader.int64() as Long); + message.model = reader.string(); continue; case 4: - if (tag !== 34) { + if (tag !== 32) { break; } - message.model = reader.string(); + message.n = longToNumber(reader.int64() as Long); continue; case 5: - if (tag !== 40) { + if (tag !== 41) { break; } - message.n = longToNumber(reader.int64() as Long); + message.presencePenalty = reader.double(); continue; case 6: - if (tag !== 49) { + if (tag !== 50) { break; } - message.presencePenalty = reader.double(); + message.stop = TextArray.decode(reader, reader.uint32()); continue; case 7: - if (tag !== 58) { + if (tag !== 57) { break; } - message.stop = TextArray.decode(reader, reader.uint32()); + message.temperature = reader.double(); continue; case 8: if (tag !== 65) { break; } - message.temperature = reader.double(); + message.topP = reader.double(); continue; case 9: - if (tag !== 73) { + if (tag !== 74) { break; } - message.topP = reader.double(); + message.baseUrl = reader.string(); continue; case 10: - if (tag !== 80) { + if (tag !== 82) { break; } - message.topLogProbs = longToNumber(reader.int64() as Long); + message.apiVersion = reader.string(); + continue; + case 11: + if (tag !== 90) { + break; + } + + message.resourceName = reader.string(); + continue; + case 12: + if (tag !== 98) { + break; + } + + message.deploymentId = reader.string(); + continue; + case 13: + if (tag !== 104) { + break; + } + + message.isAzure = reader.bool(); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -1840,7 +1892,6 @@ export const GenerativeOpenAI = { fromJSON(object: any): GenerativeOpenAI { return { frequencyPenalty: isSet(object.frequencyPenalty) ? globalThis.Number(object.frequencyPenalty) : undefined, - logProbs: isSet(object.logProbs) ? globalThis.Boolean(object.logProbs) : undefined, maxTokens: isSet(object.maxTokens) ? globalThis.Number(object.maxTokens) : undefined, model: isSet(object.model) ? globalThis.String(object.model) : "", n: isSet(object.n) ? globalThis.Number(object.n) : undefined, @@ -1848,7 +1899,11 @@ export const GenerativeOpenAI = { stop: isSet(object.stop) ? TextArray.fromJSON(object.stop) : undefined, temperature: isSet(object.temperature) ? globalThis.Number(object.temperature) : undefined, topP: isSet(object.topP) ? globalThis.Number(object.topP) : undefined, - topLogProbs: isSet(object.topLogProbs) ? globalThis.Number(object.topLogProbs) : undefined, + baseUrl: isSet(object.baseUrl) ? globalThis.String(object.baseUrl) : undefined, + apiVersion: isSet(object.apiVersion) ? globalThis.String(object.apiVersion) : undefined, + resourceName: isSet(object.resourceName) ? globalThis.String(object.resourceName) : undefined, + deploymentId: isSet(object.deploymentId) ? globalThis.String(object.deploymentId) : undefined, + isAzure: isSet(object.isAzure) ? globalThis.Boolean(object.isAzure) : undefined, }; }, @@ -1857,9 +1912,6 @@ export const GenerativeOpenAI = { if (message.frequencyPenalty !== undefined) { obj.frequencyPenalty = message.frequencyPenalty; } - if (message.logProbs !== undefined) { - obj.logProbs = message.logProbs; - } if (message.maxTokens !== undefined) { obj.maxTokens = Math.round(message.maxTokens); } @@ -1881,8 +1933,20 @@ export const GenerativeOpenAI = { if (message.topP !== undefined) { obj.topP = message.topP; } - if (message.topLogProbs !== undefined) { - obj.topLogProbs = Math.round(message.topLogProbs); + if (message.baseUrl !== undefined) { + obj.baseUrl = message.baseUrl; + } + if (message.apiVersion !== undefined) { + obj.apiVersion = message.apiVersion; + } + if (message.resourceName !== undefined) { + obj.resourceName = message.resourceName; + } + if (message.deploymentId !== undefined) { + obj.deploymentId = message.deploymentId; + } + if (message.isAzure !== undefined) { + obj.isAzure = message.isAzure; } return obj; }, @@ -1893,7 +1957,6 @@ export const GenerativeOpenAI = { fromPartial(object: DeepPartial): GenerativeOpenAI { const message = createBaseGenerativeOpenAI(); message.frequencyPenalty = object.frequencyPenalty ?? undefined; - message.logProbs = object.logProbs ?? undefined; message.maxTokens = object.maxTokens ?? undefined; message.model = object.model ?? ""; message.n = object.n ?? undefined; @@ -1901,7 +1964,11 @@ export const GenerativeOpenAI = { message.stop = (object.stop !== undefined && object.stop !== null) ? TextArray.fromPartial(object.stop) : undefined; message.temperature = object.temperature ?? undefined; message.topP = object.topP ?? undefined; - message.topLogProbs = object.topLogProbs ?? undefined; + message.baseUrl = object.baseUrl ?? undefined; + message.apiVersion = object.apiVersion ?? undefined; + message.resourceName = object.resourceName ?? undefined; + message.deploymentId = object.deploymentId ?? undefined; + message.isAzure = object.isAzure ?? undefined; return message; }, }; @@ -1916,6 +1983,10 @@ function createBaseGenerativeGoogle(): GenerativeGoogle { topK: undefined, topP: undefined, stopSequences: undefined, + apiEndpoint: undefined, + projectId: undefined, + endpointId: undefined, + region: undefined, }; } @@ -1945,6 +2016,18 @@ export const GenerativeGoogle = { if (message.stopSequences !== undefined) { TextArray.encode(message.stopSequences, writer.uint32(66).fork()).ldelim(); } + if (message.apiEndpoint !== undefined) { + writer.uint32(74).string(message.apiEndpoint); + } + if (message.projectId !== undefined) { + writer.uint32(82).string(message.projectId); + } + if (message.endpointId !== undefined) { + writer.uint32(90).string(message.endpointId); + } + if (message.region !== undefined) { + writer.uint32(98).string(message.region); + } return writer; }, @@ -2011,6 +2094,34 @@ export const GenerativeGoogle = { message.stopSequences = TextArray.decode(reader, reader.uint32()); continue; + case 9: + if (tag !== 74) { + break; + } + + message.apiEndpoint = reader.string(); + continue; + case 10: + if (tag !== 82) { + break; + } + + message.projectId = reader.string(); + continue; + case 11: + if (tag !== 90) { + break; + } + + message.endpointId = reader.string(); + continue; + case 12: + if (tag !== 98) { + break; + } + + message.region = reader.string(); + continue; } if ((tag & 7) === 4 || tag === 0) { break; @@ -2030,6 +2141,10 @@ export const GenerativeGoogle = { topK: isSet(object.topK) ? globalThis.Number(object.topK) : undefined, topP: isSet(object.topP) ? globalThis.Number(object.topP) : undefined, stopSequences: isSet(object.stopSequences) ? TextArray.fromJSON(object.stopSequences) : undefined, + apiEndpoint: isSet(object.apiEndpoint) ? globalThis.String(object.apiEndpoint) : undefined, + projectId: isSet(object.projectId) ? globalThis.String(object.projectId) : undefined, + endpointId: isSet(object.endpointId) ? globalThis.String(object.endpointId) : undefined, + region: isSet(object.region) ? globalThis.String(object.region) : undefined, }; }, @@ -2059,6 +2174,18 @@ export const GenerativeGoogle = { if (message.stopSequences !== undefined) { obj.stopSequences = TextArray.toJSON(message.stopSequences); } + if (message.apiEndpoint !== undefined) { + obj.apiEndpoint = message.apiEndpoint; + } + if (message.projectId !== undefined) { + obj.projectId = message.projectId; + } + if (message.endpointId !== undefined) { + obj.endpointId = message.endpointId; + } + if (message.region !== undefined) { + obj.region = message.region; + } return obj; }, @@ -2077,26 +2204,72 @@ export const GenerativeGoogle = { message.stopSequences = (object.stopSequences !== undefined && object.stopSequences !== null) ? TextArray.fromPartial(object.stopSequences) : undefined; + message.apiEndpoint = object.apiEndpoint ?? undefined; + message.projectId = object.projectId ?? undefined; + message.endpointId = object.endpointId ?? undefined; + message.region = object.region ?? undefined; return message; }, }; -function createBaseGenerativeAnthropicMetadata(): GenerativeAnthropicMetadata { - return { usage: undefined }; +function createBaseGenerativeDatabricks(): GenerativeDatabricks { + return { + endpoint: undefined, + model: undefined, + frequencyPenalty: undefined, + logProbs: undefined, + topLogProbs: undefined, + maxTokens: undefined, + n: undefined, + presencePenalty: undefined, + stop: undefined, + temperature: undefined, + topP: undefined, + }; } -export const GenerativeAnthropicMetadata = { - encode(message: GenerativeAnthropicMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.usage !== undefined) { - GenerativeAnthropicMetadata_Usage.encode(message.usage, writer.uint32(10).fork()).ldelim(); +export const GenerativeDatabricks = { + encode(message: GenerativeDatabricks, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.endpoint !== undefined) { + writer.uint32(10).string(message.endpoint); + } + if (message.model !== undefined) { + writer.uint32(18).string(message.model); + } + if (message.frequencyPenalty !== undefined) { + writer.uint32(25).double(message.frequencyPenalty); + } + if (message.logProbs !== undefined) { + writer.uint32(32).bool(message.logProbs); + } + if (message.topLogProbs !== undefined) { + writer.uint32(40).int64(message.topLogProbs); + } + if (message.maxTokens !== undefined) { + writer.uint32(48).int64(message.maxTokens); + } + if (message.n !== undefined) { + writer.uint32(56).int64(message.n); + } + if (message.presencePenalty !== undefined) { + writer.uint32(65).double(message.presencePenalty); + } + if (message.stop !== undefined) { + TextArray.encode(message.stop, writer.uint32(74).fork()).ldelim(); + } + if (message.temperature !== undefined) { + writer.uint32(81).double(message.temperature); + } + if (message.topP !== undefined) { + writer.uint32(89).double(message.topP); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeAnthropicMetadata { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeDatabricks { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeAnthropicMetadata(); + const message = createBaseGenerativeDatabricks(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { @@ -2105,18 +2278,338 @@ export const GenerativeAnthropicMetadata = { break; } - message.usage = GenerativeAnthropicMetadata_Usage.decode(reader, reader.uint32()); + message.endpoint = reader.string(); continue; - } - if ((tag & 7) === 4 || tag === 0) { - break; - } - reader.skipType(tag & 7); - } - return message; - }, - - fromJSON(object: any): GenerativeAnthropicMetadata { + case 2: + if (tag !== 18) { + break; + } + + message.model = reader.string(); + continue; + case 3: + if (tag !== 25) { + break; + } + + message.frequencyPenalty = reader.double(); + continue; + case 4: + if (tag !== 32) { + break; + } + + message.logProbs = reader.bool(); + continue; + case 5: + if (tag !== 40) { + break; + } + + message.topLogProbs = longToNumber(reader.int64() as Long); + continue; + case 6: + if (tag !== 48) { + break; + } + + message.maxTokens = longToNumber(reader.int64() as Long); + continue; + case 7: + if (tag !== 56) { + break; + } + + message.n = longToNumber(reader.int64() as Long); + continue; + case 8: + if (tag !== 65) { + break; + } + + message.presencePenalty = reader.double(); + continue; + case 9: + if (tag !== 74) { + break; + } + + message.stop = TextArray.decode(reader, reader.uint32()); + continue; + case 10: + if (tag !== 81) { + break; + } + + message.temperature = reader.double(); + continue; + case 11: + if (tag !== 89) { + break; + } + + message.topP = reader.double(); + continue; + } + if ((tag & 7) === 4 || tag === 0) { + break; + } + reader.skipType(tag & 7); + } + return message; + }, + + fromJSON(object: any): GenerativeDatabricks { + return { + endpoint: isSet(object.endpoint) ? globalThis.String(object.endpoint) : undefined, + model: isSet(object.model) ? globalThis.String(object.model) : undefined, + frequencyPenalty: isSet(object.frequencyPenalty) ? globalThis.Number(object.frequencyPenalty) : undefined, + logProbs: isSet(object.logProbs) ? globalThis.Boolean(object.logProbs) : undefined, + topLogProbs: isSet(object.topLogProbs) ? globalThis.Number(object.topLogProbs) : undefined, + maxTokens: isSet(object.maxTokens) ? globalThis.Number(object.maxTokens) : undefined, + n: isSet(object.n) ? globalThis.Number(object.n) : undefined, + presencePenalty: isSet(object.presencePenalty) ? globalThis.Number(object.presencePenalty) : undefined, + stop: isSet(object.stop) ? TextArray.fromJSON(object.stop) : undefined, + temperature: isSet(object.temperature) ? globalThis.Number(object.temperature) : undefined, + topP: isSet(object.topP) ? globalThis.Number(object.topP) : undefined, + }; + }, + + toJSON(message: GenerativeDatabricks): unknown { + const obj: any = {}; + if (message.endpoint !== undefined) { + obj.endpoint = message.endpoint; + } + if (message.model !== undefined) { + obj.model = message.model; + } + if (message.frequencyPenalty !== undefined) { + obj.frequencyPenalty = message.frequencyPenalty; + } + if (message.logProbs !== undefined) { + obj.logProbs = message.logProbs; + } + if (message.topLogProbs !== undefined) { + obj.topLogProbs = Math.round(message.topLogProbs); + } + if (message.maxTokens !== undefined) { + obj.maxTokens = Math.round(message.maxTokens); + } + if (message.n !== undefined) { + obj.n = Math.round(message.n); + } + if (message.presencePenalty !== undefined) { + obj.presencePenalty = message.presencePenalty; + } + if (message.stop !== undefined) { + obj.stop = TextArray.toJSON(message.stop); + } + if (message.temperature !== undefined) { + obj.temperature = message.temperature; + } + if (message.topP !== undefined) { + obj.topP = message.topP; + } + return obj; + }, + + create(base?: DeepPartial): GenerativeDatabricks { + return GenerativeDatabricks.fromPartial(base ?? {}); + }, + fromPartial(object: DeepPartial): GenerativeDatabricks { + const message = createBaseGenerativeDatabricks(); + message.endpoint = object.endpoint ?? undefined; + message.model = object.model ?? undefined; + message.frequencyPenalty = object.frequencyPenalty ?? undefined; + message.logProbs = object.logProbs ?? undefined; + message.topLogProbs = object.topLogProbs ?? undefined; + message.maxTokens = object.maxTokens ?? undefined; + message.n = object.n ?? undefined; + message.presencePenalty = object.presencePenalty ?? undefined; + message.stop = (object.stop !== undefined && object.stop !== null) ? TextArray.fromPartial(object.stop) : undefined; + message.temperature = object.temperature ?? undefined; + message.topP = object.topP ?? undefined; + return message; + }, +}; + +function createBaseGenerativeFriendliAI(): GenerativeFriendliAI { + return { + baseUrl: undefined, + model: undefined, + maxTokens: undefined, + temperature: undefined, + n: undefined, + topP: undefined, + }; +} + +export const GenerativeFriendliAI = { + encode(message: GenerativeFriendliAI, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.baseUrl !== undefined) { + writer.uint32(10).string(message.baseUrl); + } + if (message.model !== undefined) { + writer.uint32(18).string(message.model); + } + if (message.maxTokens !== undefined) { + writer.uint32(24).int64(message.maxTokens); + } + if (message.temperature !== undefined) { + writer.uint32(33).double(message.temperature); + } + if (message.n !== undefined) { + writer.uint32(40).int64(message.n); + } + if (message.topP !== undefined) { + writer.uint32(49).double(message.topP); + } + return writer; + }, + + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeFriendliAI { + const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); + let end = length === undefined ? reader.len : reader.pos + length; + const message = createBaseGenerativeFriendliAI(); + while (reader.pos < end) { + const tag = reader.uint32(); + switch (tag >>> 3) { + case 1: + if (tag !== 10) { + break; + } + + message.baseUrl = reader.string(); + continue; + case 2: + if (tag !== 18) { + break; + } + + message.model = reader.string(); + continue; + case 3: + if (tag !== 24) { + break; + } + + message.maxTokens = longToNumber(reader.int64() as Long); + continue; + case 4: + if (tag !== 33) { + break; + } + + message.temperature = reader.double(); + continue; + case 5: + if (tag !== 40) { + break; + } + + message.n = longToNumber(reader.int64() as Long); + continue; + case 6: + if (tag !== 49) { + break; + } + + message.topP = reader.double(); + continue; + } + if ((tag & 7) === 4 || tag === 0) { + break; + } + reader.skipType(tag & 7); + } + return message; + }, + + fromJSON(object: any): GenerativeFriendliAI { + return { + baseUrl: isSet(object.baseUrl) ? globalThis.String(object.baseUrl) : undefined, + model: isSet(object.model) ? globalThis.String(object.model) : undefined, + maxTokens: isSet(object.maxTokens) ? globalThis.Number(object.maxTokens) : undefined, + temperature: isSet(object.temperature) ? globalThis.Number(object.temperature) : undefined, + n: isSet(object.n) ? globalThis.Number(object.n) : undefined, + topP: isSet(object.topP) ? globalThis.Number(object.topP) : undefined, + }; + }, + + toJSON(message: GenerativeFriendliAI): unknown { + const obj: any = {}; + if (message.baseUrl !== undefined) { + obj.baseUrl = message.baseUrl; + } + if (message.model !== undefined) { + obj.model = message.model; + } + if (message.maxTokens !== undefined) { + obj.maxTokens = Math.round(message.maxTokens); + } + if (message.temperature !== undefined) { + obj.temperature = message.temperature; + } + if (message.n !== undefined) { + obj.n = Math.round(message.n); + } + if (message.topP !== undefined) { + obj.topP = message.topP; + } + return obj; + }, + + create(base?: DeepPartial): GenerativeFriendliAI { + return GenerativeFriendliAI.fromPartial(base ?? {}); + }, + fromPartial(object: DeepPartial): GenerativeFriendliAI { + const message = createBaseGenerativeFriendliAI(); + message.baseUrl = object.baseUrl ?? undefined; + message.model = object.model ?? undefined; + message.maxTokens = object.maxTokens ?? undefined; + message.temperature = object.temperature ?? undefined; + message.n = object.n ?? undefined; + message.topP = object.topP ?? undefined; + return message; + }, +}; + +function createBaseGenerativeAnthropicMetadata(): GenerativeAnthropicMetadata { + return { usage: undefined }; +} + +export const GenerativeAnthropicMetadata = { + encode(message: GenerativeAnthropicMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.usage !== undefined) { + GenerativeAnthropicMetadata_Usage.encode(message.usage, writer.uint32(10).fork()).ldelim(); + } + return writer; + }, + + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeAnthropicMetadata { + const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); + let end = length === undefined ? reader.len : reader.pos + length; + const message = createBaseGenerativeAnthropicMetadata(); + while (reader.pos < end) { + const tag = reader.uint32(); + switch (tag >>> 3) { + case 1: + if (tag !== 10) { + break; + } + + message.usage = GenerativeAnthropicMetadata_Usage.decode(reader, reader.uint32()); + continue; + } + if ((tag & 7) === 4 || tag === 0) { + break; + } + reader.skipType(tag & 7); + } + return message; + }, + + fromJSON(object: any): GenerativeAnthropicMetadata { return { usage: isSet(object.usage) ? GenerativeAnthropicMetadata_Usage.fromJSON(object.usage) : undefined }; }, @@ -2874,22 +3367,65 @@ export const GenerativeMistralMetadata_Usage = { }, }; -function createBaseGenerativeOctoAIMetadata(): GenerativeOctoAIMetadata { +function createBaseGenerativeOllamaMetadata(): GenerativeOllamaMetadata { + return {}; +} + +export const GenerativeOllamaMetadata = { + encode(_: GenerativeOllamaMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + return writer; + }, + + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOllamaMetadata { + const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); + let end = length === undefined ? reader.len : reader.pos + length; + const message = createBaseGenerativeOllamaMetadata(); + while (reader.pos < end) { + const tag = reader.uint32(); + switch (tag >>> 3) { + } + if ((tag & 7) === 4 || tag === 0) { + break; + } + reader.skipType(tag & 7); + } + return message; + }, + + fromJSON(_: any): GenerativeOllamaMetadata { + return {}; + }, + + toJSON(_: GenerativeOllamaMetadata): unknown { + const obj: any = {}; + return obj; + }, + + create(base?: DeepPartial): GenerativeOllamaMetadata { + return GenerativeOllamaMetadata.fromPartial(base ?? {}); + }, + fromPartial(_: DeepPartial): GenerativeOllamaMetadata { + const message = createBaseGenerativeOllamaMetadata(); + return message; + }, +}; + +function createBaseGenerativeOpenAIMetadata(): GenerativeOpenAIMetadata { return { usage: undefined }; } -export const GenerativeOctoAIMetadata = { - encode(message: GenerativeOctoAIMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { +export const GenerativeOpenAIMetadata = { + encode(message: GenerativeOpenAIMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { if (message.usage !== undefined) { - GenerativeOctoAIMetadata_Usage.encode(message.usage, writer.uint32(10).fork()).ldelim(); + GenerativeOpenAIMetadata_Usage.encode(message.usage, writer.uint32(10).fork()).ldelim(); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOctoAIMetadata { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOpenAIMetadata { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeOctoAIMetadata(); + const message = createBaseGenerativeOpenAIMetadata(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { @@ -2898,7 +3434,7 @@ export const GenerativeOctoAIMetadata = { break; } - message.usage = GenerativeOctoAIMetadata_Usage.decode(reader, reader.uint32()); + message.usage = GenerativeOpenAIMetadata_Usage.decode(reader, reader.uint32()); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -2909,36 +3445,36 @@ export const GenerativeOctoAIMetadata = { return message; }, - fromJSON(object: any): GenerativeOctoAIMetadata { - return { usage: isSet(object.usage) ? GenerativeOctoAIMetadata_Usage.fromJSON(object.usage) : undefined }; + fromJSON(object: any): GenerativeOpenAIMetadata { + return { usage: isSet(object.usage) ? GenerativeOpenAIMetadata_Usage.fromJSON(object.usage) : undefined }; }, - toJSON(message: GenerativeOctoAIMetadata): unknown { + toJSON(message: GenerativeOpenAIMetadata): unknown { const obj: any = {}; if (message.usage !== undefined) { - obj.usage = GenerativeOctoAIMetadata_Usage.toJSON(message.usage); + obj.usage = GenerativeOpenAIMetadata_Usage.toJSON(message.usage); } return obj; }, - create(base?: DeepPartial): GenerativeOctoAIMetadata { - return GenerativeOctoAIMetadata.fromPartial(base ?? {}); + create(base?: DeepPartial): GenerativeOpenAIMetadata { + return GenerativeOpenAIMetadata.fromPartial(base ?? {}); }, - fromPartial(object: DeepPartial): GenerativeOctoAIMetadata { - const message = createBaseGenerativeOctoAIMetadata(); + fromPartial(object: DeepPartial): GenerativeOpenAIMetadata { + const message = createBaseGenerativeOpenAIMetadata(); message.usage = (object.usage !== undefined && object.usage !== null) - ? GenerativeOctoAIMetadata_Usage.fromPartial(object.usage) + ? GenerativeOpenAIMetadata_Usage.fromPartial(object.usage) : undefined; return message; }, }; -function createBaseGenerativeOctoAIMetadata_Usage(): GenerativeOctoAIMetadata_Usage { +function createBaseGenerativeOpenAIMetadata_Usage(): GenerativeOpenAIMetadata_Usage { return { promptTokens: undefined, completionTokens: undefined, totalTokens: undefined }; } -export const GenerativeOctoAIMetadata_Usage = { - encode(message: GenerativeOctoAIMetadata_Usage, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { +export const GenerativeOpenAIMetadata_Usage = { + encode(message: GenerativeOpenAIMetadata_Usage, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { if (message.promptTokens !== undefined) { writer.uint32(8).int64(message.promptTokens); } @@ -2946,15 +3482,181 @@ export const GenerativeOctoAIMetadata_Usage = { writer.uint32(16).int64(message.completionTokens); } if (message.totalTokens !== undefined) { - writer.uint32(24).int64(message.totalTokens); + writer.uint32(24).int64(message.totalTokens); + } + return writer; + }, + + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOpenAIMetadata_Usage { + const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); + let end = length === undefined ? reader.len : reader.pos + length; + const message = createBaseGenerativeOpenAIMetadata_Usage(); + while (reader.pos < end) { + const tag = reader.uint32(); + switch (tag >>> 3) { + case 1: + if (tag !== 8) { + break; + } + + message.promptTokens = longToNumber(reader.int64() as Long); + continue; + case 2: + if (tag !== 16) { + break; + } + + message.completionTokens = longToNumber(reader.int64() as Long); + continue; + case 3: + if (tag !== 24) { + break; + } + + message.totalTokens = longToNumber(reader.int64() as Long); + continue; + } + if ((tag & 7) === 4 || tag === 0) { + break; + } + reader.skipType(tag & 7); + } + return message; + }, + + fromJSON(object: any): GenerativeOpenAIMetadata_Usage { + return { + promptTokens: isSet(object.promptTokens) ? globalThis.Number(object.promptTokens) : undefined, + completionTokens: isSet(object.completionTokens) ? globalThis.Number(object.completionTokens) : undefined, + totalTokens: isSet(object.totalTokens) ? globalThis.Number(object.totalTokens) : undefined, + }; + }, + + toJSON(message: GenerativeOpenAIMetadata_Usage): unknown { + const obj: any = {}; + if (message.promptTokens !== undefined) { + obj.promptTokens = Math.round(message.promptTokens); + } + if (message.completionTokens !== undefined) { + obj.completionTokens = Math.round(message.completionTokens); + } + if (message.totalTokens !== undefined) { + obj.totalTokens = Math.round(message.totalTokens); + } + return obj; + }, + + create(base?: DeepPartial): GenerativeOpenAIMetadata_Usage { + return GenerativeOpenAIMetadata_Usage.fromPartial(base ?? {}); + }, + fromPartial(object: DeepPartial): GenerativeOpenAIMetadata_Usage { + const message = createBaseGenerativeOpenAIMetadata_Usage(); + message.promptTokens = object.promptTokens ?? undefined; + message.completionTokens = object.completionTokens ?? undefined; + message.totalTokens = object.totalTokens ?? undefined; + return message; + }, +}; + +function createBaseGenerativeGoogleMetadata(): GenerativeGoogleMetadata { + return { metadata: undefined, usageMetadata: undefined }; +} + +export const GenerativeGoogleMetadata = { + encode(message: GenerativeGoogleMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.metadata !== undefined) { + GenerativeGoogleMetadata_Metadata.encode(message.metadata, writer.uint32(10).fork()).ldelim(); + } + if (message.usageMetadata !== undefined) { + GenerativeGoogleMetadata_UsageMetadata.encode(message.usageMetadata, writer.uint32(18).fork()).ldelim(); + } + return writer; + }, + + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeGoogleMetadata { + const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); + let end = length === undefined ? reader.len : reader.pos + length; + const message = createBaseGenerativeGoogleMetadata(); + while (reader.pos < end) { + const tag = reader.uint32(); + switch (tag >>> 3) { + case 1: + if (tag !== 10) { + break; + } + + message.metadata = GenerativeGoogleMetadata_Metadata.decode(reader, reader.uint32()); + continue; + case 2: + if (tag !== 18) { + break; + } + + message.usageMetadata = GenerativeGoogleMetadata_UsageMetadata.decode(reader, reader.uint32()); + continue; + } + if ((tag & 7) === 4 || tag === 0) { + break; + } + reader.skipType(tag & 7); + } + return message; + }, + + fromJSON(object: any): GenerativeGoogleMetadata { + return { + metadata: isSet(object.metadata) ? GenerativeGoogleMetadata_Metadata.fromJSON(object.metadata) : undefined, + usageMetadata: isSet(object.usageMetadata) + ? GenerativeGoogleMetadata_UsageMetadata.fromJSON(object.usageMetadata) + : undefined, + }; + }, + + toJSON(message: GenerativeGoogleMetadata): unknown { + const obj: any = {}; + if (message.metadata !== undefined) { + obj.metadata = GenerativeGoogleMetadata_Metadata.toJSON(message.metadata); + } + if (message.usageMetadata !== undefined) { + obj.usageMetadata = GenerativeGoogleMetadata_UsageMetadata.toJSON(message.usageMetadata); + } + return obj; + }, + + create(base?: DeepPartial): GenerativeGoogleMetadata { + return GenerativeGoogleMetadata.fromPartial(base ?? {}); + }, + fromPartial(object: DeepPartial): GenerativeGoogleMetadata { + const message = createBaseGenerativeGoogleMetadata(); + message.metadata = (object.metadata !== undefined && object.metadata !== null) + ? GenerativeGoogleMetadata_Metadata.fromPartial(object.metadata) + : undefined; + message.usageMetadata = (object.usageMetadata !== undefined && object.usageMetadata !== null) + ? GenerativeGoogleMetadata_UsageMetadata.fromPartial(object.usageMetadata) + : undefined; + return message; + }, +}; + +function createBaseGenerativeGoogleMetadata_TokenCount(): GenerativeGoogleMetadata_TokenCount { + return { totalBillableCharacters: undefined, totalTokens: undefined }; +} + +export const GenerativeGoogleMetadata_TokenCount = { + encode(message: GenerativeGoogleMetadata_TokenCount, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.totalBillableCharacters !== undefined) { + writer.uint32(8).int64(message.totalBillableCharacters); + } + if (message.totalTokens !== undefined) { + writer.uint32(16).int64(message.totalTokens); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOctoAIMetadata_Usage { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeGoogleMetadata_TokenCount { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeOctoAIMetadata_Usage(); + const message = createBaseGenerativeGoogleMetadata_TokenCount(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { @@ -2963,20 +3665,13 @@ export const GenerativeOctoAIMetadata_Usage = { break; } - message.promptTokens = longToNumber(reader.int64() as Long); + message.totalBillableCharacters = longToNumber(reader.int64() as Long); continue; case 2: if (tag !== 16) { break; } - message.completionTokens = longToNumber(reader.int64() as Long); - continue; - case 3: - if (tag !== 24) { - break; - } - message.totalTokens = longToNumber(reader.int64() as Long); continue; } @@ -2988,21 +3683,19 @@ export const GenerativeOctoAIMetadata_Usage = { return message; }, - fromJSON(object: any): GenerativeOctoAIMetadata_Usage { + fromJSON(object: any): GenerativeGoogleMetadata_TokenCount { return { - promptTokens: isSet(object.promptTokens) ? globalThis.Number(object.promptTokens) : undefined, - completionTokens: isSet(object.completionTokens) ? globalThis.Number(object.completionTokens) : undefined, + totalBillableCharacters: isSet(object.totalBillableCharacters) + ? globalThis.Number(object.totalBillableCharacters) + : undefined, totalTokens: isSet(object.totalTokens) ? globalThis.Number(object.totalTokens) : undefined, }; }, - toJSON(message: GenerativeOctoAIMetadata_Usage): unknown { + toJSON(message: GenerativeGoogleMetadata_TokenCount): unknown { const obj: any = {}; - if (message.promptTokens !== undefined) { - obj.promptTokens = Math.round(message.promptTokens); - } - if (message.completionTokens !== undefined) { - obj.completionTokens = Math.round(message.completionTokens); + if (message.totalBillableCharacters !== undefined) { + obj.totalBillableCharacters = Math.round(message.totalBillableCharacters); } if (message.totalTokens !== undefined) { obj.totalTokens = Math.round(message.totalTokens); @@ -3010,77 +3703,36 @@ export const GenerativeOctoAIMetadata_Usage = { return obj; }, - create(base?: DeepPartial): GenerativeOctoAIMetadata_Usage { - return GenerativeOctoAIMetadata_Usage.fromPartial(base ?? {}); + create(base?: DeepPartial): GenerativeGoogleMetadata_TokenCount { + return GenerativeGoogleMetadata_TokenCount.fromPartial(base ?? {}); }, - fromPartial(object: DeepPartial): GenerativeOctoAIMetadata_Usage { - const message = createBaseGenerativeOctoAIMetadata_Usage(); - message.promptTokens = object.promptTokens ?? undefined; - message.completionTokens = object.completionTokens ?? undefined; + fromPartial(object: DeepPartial): GenerativeGoogleMetadata_TokenCount { + const message = createBaseGenerativeGoogleMetadata_TokenCount(); + message.totalBillableCharacters = object.totalBillableCharacters ?? undefined; message.totalTokens = object.totalTokens ?? undefined; return message; }, }; -function createBaseGenerativeOllamaMetadata(): GenerativeOllamaMetadata { - return {}; +function createBaseGenerativeGoogleMetadata_TokenMetadata(): GenerativeGoogleMetadata_TokenMetadata { + return { inputTokenCount: undefined, outputTokenCount: undefined }; } -export const GenerativeOllamaMetadata = { - encode(_: GenerativeOllamaMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - return writer; - }, - - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOllamaMetadata { - const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); - let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeOllamaMetadata(); - while (reader.pos < end) { - const tag = reader.uint32(); - switch (tag >>> 3) { - } - if ((tag & 7) === 4 || tag === 0) { - break; - } - reader.skipType(tag & 7); +export const GenerativeGoogleMetadata_TokenMetadata = { + encode(message: GenerativeGoogleMetadata_TokenMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.inputTokenCount !== undefined) { + GenerativeGoogleMetadata_TokenCount.encode(message.inputTokenCount, writer.uint32(10).fork()).ldelim(); } - return message; - }, - - fromJSON(_: any): GenerativeOllamaMetadata { - return {}; - }, - - toJSON(_: GenerativeOllamaMetadata): unknown { - const obj: any = {}; - return obj; - }, - - create(base?: DeepPartial): GenerativeOllamaMetadata { - return GenerativeOllamaMetadata.fromPartial(base ?? {}); - }, - fromPartial(_: DeepPartial): GenerativeOllamaMetadata { - const message = createBaseGenerativeOllamaMetadata(); - return message; - }, -}; - -function createBaseGenerativeOpenAIMetadata(): GenerativeOpenAIMetadata { - return { usage: undefined }; -} - -export const GenerativeOpenAIMetadata = { - encode(message: GenerativeOpenAIMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.usage !== undefined) { - GenerativeOpenAIMetadata_Usage.encode(message.usage, writer.uint32(10).fork()).ldelim(); + if (message.outputTokenCount !== undefined) { + GenerativeGoogleMetadata_TokenCount.encode(message.outputTokenCount, writer.uint32(18).fork()).ldelim(); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOpenAIMetadata { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeGoogleMetadata_TokenMetadata { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeOpenAIMetadata(); + const message = createBaseGenerativeGoogleMetadata_TokenMetadata(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { @@ -3089,7 +3741,14 @@ export const GenerativeOpenAIMetadata = { break; } - message.usage = GenerativeOpenAIMetadata_Usage.decode(reader, reader.uint32()); + message.inputTokenCount = GenerativeGoogleMetadata_TokenCount.decode(reader, reader.uint32()); + continue; + case 2: + if (tag !== 18) { + break; + } + + message.outputTokenCount = GenerativeGoogleMetadata_TokenCount.decode(reader, reader.uint32()); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -3100,75 +3759,68 @@ export const GenerativeOpenAIMetadata = { return message; }, - fromJSON(object: any): GenerativeOpenAIMetadata { - return { usage: isSet(object.usage) ? GenerativeOpenAIMetadata_Usage.fromJSON(object.usage) : undefined }; + fromJSON(object: any): GenerativeGoogleMetadata_TokenMetadata { + return { + inputTokenCount: isSet(object.inputTokenCount) + ? GenerativeGoogleMetadata_TokenCount.fromJSON(object.inputTokenCount) + : undefined, + outputTokenCount: isSet(object.outputTokenCount) + ? GenerativeGoogleMetadata_TokenCount.fromJSON(object.outputTokenCount) + : undefined, + }; }, - toJSON(message: GenerativeOpenAIMetadata): unknown { + toJSON(message: GenerativeGoogleMetadata_TokenMetadata): unknown { const obj: any = {}; - if (message.usage !== undefined) { - obj.usage = GenerativeOpenAIMetadata_Usage.toJSON(message.usage); + if (message.inputTokenCount !== undefined) { + obj.inputTokenCount = GenerativeGoogleMetadata_TokenCount.toJSON(message.inputTokenCount); + } + if (message.outputTokenCount !== undefined) { + obj.outputTokenCount = GenerativeGoogleMetadata_TokenCount.toJSON(message.outputTokenCount); } return obj; }, - create(base?: DeepPartial): GenerativeOpenAIMetadata { - return GenerativeOpenAIMetadata.fromPartial(base ?? {}); + create(base?: DeepPartial): GenerativeGoogleMetadata_TokenMetadata { + return GenerativeGoogleMetadata_TokenMetadata.fromPartial(base ?? {}); }, - fromPartial(object: DeepPartial): GenerativeOpenAIMetadata { - const message = createBaseGenerativeOpenAIMetadata(); - message.usage = (object.usage !== undefined && object.usage !== null) - ? GenerativeOpenAIMetadata_Usage.fromPartial(object.usage) + fromPartial(object: DeepPartial): GenerativeGoogleMetadata_TokenMetadata { + const message = createBaseGenerativeGoogleMetadata_TokenMetadata(); + message.inputTokenCount = (object.inputTokenCount !== undefined && object.inputTokenCount !== null) + ? GenerativeGoogleMetadata_TokenCount.fromPartial(object.inputTokenCount) + : undefined; + message.outputTokenCount = (object.outputTokenCount !== undefined && object.outputTokenCount !== null) + ? GenerativeGoogleMetadata_TokenCount.fromPartial(object.outputTokenCount) : undefined; return message; }, }; -function createBaseGenerativeOpenAIMetadata_Usage(): GenerativeOpenAIMetadata_Usage { - return { promptTokens: undefined, completionTokens: undefined, totalTokens: undefined }; +function createBaseGenerativeGoogleMetadata_Metadata(): GenerativeGoogleMetadata_Metadata { + return { tokenMetadata: undefined }; } -export const GenerativeOpenAIMetadata_Usage = { - encode(message: GenerativeOpenAIMetadata_Usage, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.promptTokens !== undefined) { - writer.uint32(8).int64(message.promptTokens); - } - if (message.completionTokens !== undefined) { - writer.uint32(16).int64(message.completionTokens); - } - if (message.totalTokens !== undefined) { - writer.uint32(24).int64(message.totalTokens); +export const GenerativeGoogleMetadata_Metadata = { + encode(message: GenerativeGoogleMetadata_Metadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.tokenMetadata !== undefined) { + GenerativeGoogleMetadata_TokenMetadata.encode(message.tokenMetadata, writer.uint32(10).fork()).ldelim(); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeOpenAIMetadata_Usage { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeGoogleMetadata_Metadata { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeOpenAIMetadata_Usage(); + const message = createBaseGenerativeGoogleMetadata_Metadata(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { case 1: - if (tag !== 8) { - break; - } - - message.promptTokens = longToNumber(reader.int64() as Long); - continue; - case 2: - if (tag !== 16) { - break; - } - - message.completionTokens = longToNumber(reader.int64() as Long); - continue; - case 3: - if (tag !== 24) { + if (tag !== 10) { break; } - message.totalTokens = longToNumber(reader.int64() as Long); + message.tokenMetadata = GenerativeGoogleMetadata_TokenMetadata.decode(reader, reader.uint32()); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -3179,75 +3831,79 @@ export const GenerativeOpenAIMetadata_Usage = { return message; }, - fromJSON(object: any): GenerativeOpenAIMetadata_Usage { + fromJSON(object: any): GenerativeGoogleMetadata_Metadata { return { - promptTokens: isSet(object.promptTokens) ? globalThis.Number(object.promptTokens) : undefined, - completionTokens: isSet(object.completionTokens) ? globalThis.Number(object.completionTokens) : undefined, - totalTokens: isSet(object.totalTokens) ? globalThis.Number(object.totalTokens) : undefined, + tokenMetadata: isSet(object.tokenMetadata) + ? GenerativeGoogleMetadata_TokenMetadata.fromJSON(object.tokenMetadata) + : undefined, }; }, - toJSON(message: GenerativeOpenAIMetadata_Usage): unknown { + toJSON(message: GenerativeGoogleMetadata_Metadata): unknown { const obj: any = {}; - if (message.promptTokens !== undefined) { - obj.promptTokens = Math.round(message.promptTokens); - } - if (message.completionTokens !== undefined) { - obj.completionTokens = Math.round(message.completionTokens); - } - if (message.totalTokens !== undefined) { - obj.totalTokens = Math.round(message.totalTokens); + if (message.tokenMetadata !== undefined) { + obj.tokenMetadata = GenerativeGoogleMetadata_TokenMetadata.toJSON(message.tokenMetadata); } return obj; }, - create(base?: DeepPartial): GenerativeOpenAIMetadata_Usage { - return GenerativeOpenAIMetadata_Usage.fromPartial(base ?? {}); + create(base?: DeepPartial): GenerativeGoogleMetadata_Metadata { + return GenerativeGoogleMetadata_Metadata.fromPartial(base ?? {}); }, - fromPartial(object: DeepPartial): GenerativeOpenAIMetadata_Usage { - const message = createBaseGenerativeOpenAIMetadata_Usage(); - message.promptTokens = object.promptTokens ?? undefined; - message.completionTokens = object.completionTokens ?? undefined; - message.totalTokens = object.totalTokens ?? undefined; + fromPartial(object: DeepPartial): GenerativeGoogleMetadata_Metadata { + const message = createBaseGenerativeGoogleMetadata_Metadata(); + message.tokenMetadata = (object.tokenMetadata !== undefined && object.tokenMetadata !== null) + ? GenerativeGoogleMetadata_TokenMetadata.fromPartial(object.tokenMetadata) + : undefined; return message; }, }; -function createBaseGenerativeGoogleMetadata(): GenerativeGoogleMetadata { - return { metadata: undefined, usageMetadata: undefined }; +function createBaseGenerativeGoogleMetadata_UsageMetadata(): GenerativeGoogleMetadata_UsageMetadata { + return { promptTokenCount: undefined, candidatesTokenCount: undefined, totalTokenCount: undefined }; } -export const GenerativeGoogleMetadata = { - encode(message: GenerativeGoogleMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.metadata !== undefined) { - GenerativeGoogleMetadata_Metadata.encode(message.metadata, writer.uint32(10).fork()).ldelim(); +export const GenerativeGoogleMetadata_UsageMetadata = { + encode(message: GenerativeGoogleMetadata_UsageMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.promptTokenCount !== undefined) { + writer.uint32(8).int64(message.promptTokenCount); } - if (message.usageMetadata !== undefined) { - GenerativeGoogleMetadata_UsageMetadata.encode(message.usageMetadata, writer.uint32(18).fork()).ldelim(); + if (message.candidatesTokenCount !== undefined) { + writer.uint32(16).int64(message.candidatesTokenCount); + } + if (message.totalTokenCount !== undefined) { + writer.uint32(24).int64(message.totalTokenCount); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeGoogleMetadata { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeGoogleMetadata_UsageMetadata { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeGoogleMetadata(); + const message = createBaseGenerativeGoogleMetadata_UsageMetadata(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { case 1: - if (tag !== 10) { + if (tag !== 8) { break; } - message.metadata = GenerativeGoogleMetadata_Metadata.decode(reader, reader.uint32()); + message.promptTokenCount = longToNumber(reader.int64() as Long); continue; case 2: - if (tag !== 18) { + if (tag !== 16) { break; } - message.usageMetadata = GenerativeGoogleMetadata_UsageMetadata.decode(reader, reader.uint32()); + message.candidatesTokenCount = longToNumber(reader.int64() as Long); + continue; + case 3: + if (tag !== 24) { + break; + } + + message.totalTokenCount = longToNumber(reader.int64() as Long); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -3258,76 +3914,67 @@ export const GenerativeGoogleMetadata = { return message; }, - fromJSON(object: any): GenerativeGoogleMetadata { + fromJSON(object: any): GenerativeGoogleMetadata_UsageMetadata { return { - metadata: isSet(object.metadata) ? GenerativeGoogleMetadata_Metadata.fromJSON(object.metadata) : undefined, - usageMetadata: isSet(object.usageMetadata) - ? GenerativeGoogleMetadata_UsageMetadata.fromJSON(object.usageMetadata) + promptTokenCount: isSet(object.promptTokenCount) ? globalThis.Number(object.promptTokenCount) : undefined, + candidatesTokenCount: isSet(object.candidatesTokenCount) + ? globalThis.Number(object.candidatesTokenCount) : undefined, + totalTokenCount: isSet(object.totalTokenCount) ? globalThis.Number(object.totalTokenCount) : undefined, }; }, - toJSON(message: GenerativeGoogleMetadata): unknown { + toJSON(message: GenerativeGoogleMetadata_UsageMetadata): unknown { const obj: any = {}; - if (message.metadata !== undefined) { - obj.metadata = GenerativeGoogleMetadata_Metadata.toJSON(message.metadata); + if (message.promptTokenCount !== undefined) { + obj.promptTokenCount = Math.round(message.promptTokenCount); } - if (message.usageMetadata !== undefined) { - obj.usageMetadata = GenerativeGoogleMetadata_UsageMetadata.toJSON(message.usageMetadata); + if (message.candidatesTokenCount !== undefined) { + obj.candidatesTokenCount = Math.round(message.candidatesTokenCount); + } + if (message.totalTokenCount !== undefined) { + obj.totalTokenCount = Math.round(message.totalTokenCount); } return obj; }, - create(base?: DeepPartial): GenerativeGoogleMetadata { - return GenerativeGoogleMetadata.fromPartial(base ?? {}); + create(base?: DeepPartial): GenerativeGoogleMetadata_UsageMetadata { + return GenerativeGoogleMetadata_UsageMetadata.fromPartial(base ?? {}); }, - fromPartial(object: DeepPartial): GenerativeGoogleMetadata { - const message = createBaseGenerativeGoogleMetadata(); - message.metadata = (object.metadata !== undefined && object.metadata !== null) - ? GenerativeGoogleMetadata_Metadata.fromPartial(object.metadata) - : undefined; - message.usageMetadata = (object.usageMetadata !== undefined && object.usageMetadata !== null) - ? GenerativeGoogleMetadata_UsageMetadata.fromPartial(object.usageMetadata) - : undefined; + fromPartial(object: DeepPartial): GenerativeGoogleMetadata_UsageMetadata { + const message = createBaseGenerativeGoogleMetadata_UsageMetadata(); + message.promptTokenCount = object.promptTokenCount ?? undefined; + message.candidatesTokenCount = object.candidatesTokenCount ?? undefined; + message.totalTokenCount = object.totalTokenCount ?? undefined; return message; }, }; -function createBaseGenerativeGoogleMetadata_TokenCount(): GenerativeGoogleMetadata_TokenCount { - return { totalBillableCharacters: undefined, totalTokens: undefined }; +function createBaseGenerativeDatabricksMetadata(): GenerativeDatabricksMetadata { + return { usage: undefined }; } -export const GenerativeGoogleMetadata_TokenCount = { - encode(message: GenerativeGoogleMetadata_TokenCount, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.totalBillableCharacters !== undefined) { - writer.uint32(8).int64(message.totalBillableCharacters); - } - if (message.totalTokens !== undefined) { - writer.uint32(16).int64(message.totalTokens); +export const GenerativeDatabricksMetadata = { + encode(message: GenerativeDatabricksMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.usage !== undefined) { + GenerativeDatabricksMetadata_Usage.encode(message.usage, writer.uint32(10).fork()).ldelim(); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeGoogleMetadata_TokenCount { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeDatabricksMetadata { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeGoogleMetadata_TokenCount(); + const message = createBaseGenerativeDatabricksMetadata(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { case 1: - if (tag !== 8) { - break; - } - - message.totalBillableCharacters = longToNumber(reader.int64() as Long); - continue; - case 2: - if (tag !== 16) { + if (tag !== 10) { break; } - message.totalTokens = longToNumber(reader.int64() as Long); + message.usage = GenerativeDatabricksMetadata_Usage.decode(reader, reader.uint32()); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -3338,72 +3985,75 @@ export const GenerativeGoogleMetadata_TokenCount = { return message; }, - fromJSON(object: any): GenerativeGoogleMetadata_TokenCount { - return { - totalBillableCharacters: isSet(object.totalBillableCharacters) - ? globalThis.Number(object.totalBillableCharacters) - : undefined, - totalTokens: isSet(object.totalTokens) ? globalThis.Number(object.totalTokens) : undefined, - }; + fromJSON(object: any): GenerativeDatabricksMetadata { + return { usage: isSet(object.usage) ? GenerativeDatabricksMetadata_Usage.fromJSON(object.usage) : undefined }; }, - toJSON(message: GenerativeGoogleMetadata_TokenCount): unknown { + toJSON(message: GenerativeDatabricksMetadata): unknown { const obj: any = {}; - if (message.totalBillableCharacters !== undefined) { - obj.totalBillableCharacters = Math.round(message.totalBillableCharacters); - } - if (message.totalTokens !== undefined) { - obj.totalTokens = Math.round(message.totalTokens); + if (message.usage !== undefined) { + obj.usage = GenerativeDatabricksMetadata_Usage.toJSON(message.usage); } return obj; }, - create(base?: DeepPartial): GenerativeGoogleMetadata_TokenCount { - return GenerativeGoogleMetadata_TokenCount.fromPartial(base ?? {}); + create(base?: DeepPartial): GenerativeDatabricksMetadata { + return GenerativeDatabricksMetadata.fromPartial(base ?? {}); }, - fromPartial(object: DeepPartial): GenerativeGoogleMetadata_TokenCount { - const message = createBaseGenerativeGoogleMetadata_TokenCount(); - message.totalBillableCharacters = object.totalBillableCharacters ?? undefined; - message.totalTokens = object.totalTokens ?? undefined; + fromPartial(object: DeepPartial): GenerativeDatabricksMetadata { + const message = createBaseGenerativeDatabricksMetadata(); + message.usage = (object.usage !== undefined && object.usage !== null) + ? GenerativeDatabricksMetadata_Usage.fromPartial(object.usage) + : undefined; return message; }, }; -function createBaseGenerativeGoogleMetadata_TokenMetadata(): GenerativeGoogleMetadata_TokenMetadata { - return { inputTokenCount: undefined, outputTokenCount: undefined }; +function createBaseGenerativeDatabricksMetadata_Usage(): GenerativeDatabricksMetadata_Usage { + return { promptTokens: undefined, completionTokens: undefined, totalTokens: undefined }; } -export const GenerativeGoogleMetadata_TokenMetadata = { - encode(message: GenerativeGoogleMetadata_TokenMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.inputTokenCount !== undefined) { - GenerativeGoogleMetadata_TokenCount.encode(message.inputTokenCount, writer.uint32(10).fork()).ldelim(); +export const GenerativeDatabricksMetadata_Usage = { + encode(message: GenerativeDatabricksMetadata_Usage, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.promptTokens !== undefined) { + writer.uint32(8).int64(message.promptTokens); } - if (message.outputTokenCount !== undefined) { - GenerativeGoogleMetadata_TokenCount.encode(message.outputTokenCount, writer.uint32(18).fork()).ldelim(); + if (message.completionTokens !== undefined) { + writer.uint32(16).int64(message.completionTokens); + } + if (message.totalTokens !== undefined) { + writer.uint32(24).int64(message.totalTokens); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeGoogleMetadata_TokenMetadata { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeDatabricksMetadata_Usage { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeGoogleMetadata_TokenMetadata(); + const message = createBaseGenerativeDatabricksMetadata_Usage(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { case 1: - if (tag !== 10) { + if (tag !== 8) { break; } - message.inputTokenCount = GenerativeGoogleMetadata_TokenCount.decode(reader, reader.uint32()); + message.promptTokens = longToNumber(reader.int64() as Long); continue; case 2: - if (tag !== 18) { + if (tag !== 16) { break; } - message.outputTokenCount = GenerativeGoogleMetadata_TokenCount.decode(reader, reader.uint32()); + message.completionTokens = longToNumber(reader.int64() as Long); + continue; + case 3: + if (tag !== 24) { + break; + } + + message.totalTokens = longToNumber(reader.int64() as Long); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -3414,59 +4064,56 @@ export const GenerativeGoogleMetadata_TokenMetadata = { return message; }, - fromJSON(object: any): GenerativeGoogleMetadata_TokenMetadata { + fromJSON(object: any): GenerativeDatabricksMetadata_Usage { return { - inputTokenCount: isSet(object.inputTokenCount) - ? GenerativeGoogleMetadata_TokenCount.fromJSON(object.inputTokenCount) - : undefined, - outputTokenCount: isSet(object.outputTokenCount) - ? GenerativeGoogleMetadata_TokenCount.fromJSON(object.outputTokenCount) - : undefined, + promptTokens: isSet(object.promptTokens) ? globalThis.Number(object.promptTokens) : undefined, + completionTokens: isSet(object.completionTokens) ? globalThis.Number(object.completionTokens) : undefined, + totalTokens: isSet(object.totalTokens) ? globalThis.Number(object.totalTokens) : undefined, }; }, - toJSON(message: GenerativeGoogleMetadata_TokenMetadata): unknown { + toJSON(message: GenerativeDatabricksMetadata_Usage): unknown { const obj: any = {}; - if (message.inputTokenCount !== undefined) { - obj.inputTokenCount = GenerativeGoogleMetadata_TokenCount.toJSON(message.inputTokenCount); + if (message.promptTokens !== undefined) { + obj.promptTokens = Math.round(message.promptTokens); } - if (message.outputTokenCount !== undefined) { - obj.outputTokenCount = GenerativeGoogleMetadata_TokenCount.toJSON(message.outputTokenCount); + if (message.completionTokens !== undefined) { + obj.completionTokens = Math.round(message.completionTokens); + } + if (message.totalTokens !== undefined) { + obj.totalTokens = Math.round(message.totalTokens); } return obj; }, - create(base?: DeepPartial): GenerativeGoogleMetadata_TokenMetadata { - return GenerativeGoogleMetadata_TokenMetadata.fromPartial(base ?? {}); + create(base?: DeepPartial): GenerativeDatabricksMetadata_Usage { + return GenerativeDatabricksMetadata_Usage.fromPartial(base ?? {}); }, - fromPartial(object: DeepPartial): GenerativeGoogleMetadata_TokenMetadata { - const message = createBaseGenerativeGoogleMetadata_TokenMetadata(); - message.inputTokenCount = (object.inputTokenCount !== undefined && object.inputTokenCount !== null) - ? GenerativeGoogleMetadata_TokenCount.fromPartial(object.inputTokenCount) - : undefined; - message.outputTokenCount = (object.outputTokenCount !== undefined && object.outputTokenCount !== null) - ? GenerativeGoogleMetadata_TokenCount.fromPartial(object.outputTokenCount) - : undefined; + fromPartial(object: DeepPartial): GenerativeDatabricksMetadata_Usage { + const message = createBaseGenerativeDatabricksMetadata_Usage(); + message.promptTokens = object.promptTokens ?? undefined; + message.completionTokens = object.completionTokens ?? undefined; + message.totalTokens = object.totalTokens ?? undefined; return message; }, }; -function createBaseGenerativeGoogleMetadata_Metadata(): GenerativeGoogleMetadata_Metadata { - return { tokenMetadata: undefined }; +function createBaseGenerativeFriendliAIMetadata(): GenerativeFriendliAIMetadata { + return { usage: undefined }; } -export const GenerativeGoogleMetadata_Metadata = { - encode(message: GenerativeGoogleMetadata_Metadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.tokenMetadata !== undefined) { - GenerativeGoogleMetadata_TokenMetadata.encode(message.tokenMetadata, writer.uint32(10).fork()).ldelim(); +export const GenerativeFriendliAIMetadata = { + encode(message: GenerativeFriendliAIMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.usage !== undefined) { + GenerativeFriendliAIMetadata_Usage.encode(message.usage, writer.uint32(10).fork()).ldelim(); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeGoogleMetadata_Metadata { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeFriendliAIMetadata { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeGoogleMetadata_Metadata(); + const message = createBaseGenerativeFriendliAIMetadata(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { @@ -3475,7 +4122,7 @@ export const GenerativeGoogleMetadata_Metadata = { break; } - message.tokenMetadata = GenerativeGoogleMetadata_TokenMetadata.decode(reader, reader.uint32()); + message.usage = GenerativeFriendliAIMetadata_Usage.decode(reader, reader.uint32()); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -3486,56 +4133,52 @@ export const GenerativeGoogleMetadata_Metadata = { return message; }, - fromJSON(object: any): GenerativeGoogleMetadata_Metadata { - return { - tokenMetadata: isSet(object.tokenMetadata) - ? GenerativeGoogleMetadata_TokenMetadata.fromJSON(object.tokenMetadata) - : undefined, - }; + fromJSON(object: any): GenerativeFriendliAIMetadata { + return { usage: isSet(object.usage) ? GenerativeFriendliAIMetadata_Usage.fromJSON(object.usage) : undefined }; }, - toJSON(message: GenerativeGoogleMetadata_Metadata): unknown { + toJSON(message: GenerativeFriendliAIMetadata): unknown { const obj: any = {}; - if (message.tokenMetadata !== undefined) { - obj.tokenMetadata = GenerativeGoogleMetadata_TokenMetadata.toJSON(message.tokenMetadata); + if (message.usage !== undefined) { + obj.usage = GenerativeFriendliAIMetadata_Usage.toJSON(message.usage); } return obj; }, - create(base?: DeepPartial): GenerativeGoogleMetadata_Metadata { - return GenerativeGoogleMetadata_Metadata.fromPartial(base ?? {}); + create(base?: DeepPartial): GenerativeFriendliAIMetadata { + return GenerativeFriendliAIMetadata.fromPartial(base ?? {}); }, - fromPartial(object: DeepPartial): GenerativeGoogleMetadata_Metadata { - const message = createBaseGenerativeGoogleMetadata_Metadata(); - message.tokenMetadata = (object.tokenMetadata !== undefined && object.tokenMetadata !== null) - ? GenerativeGoogleMetadata_TokenMetadata.fromPartial(object.tokenMetadata) + fromPartial(object: DeepPartial): GenerativeFriendliAIMetadata { + const message = createBaseGenerativeFriendliAIMetadata(); + message.usage = (object.usage !== undefined && object.usage !== null) + ? GenerativeFriendliAIMetadata_Usage.fromPartial(object.usage) : undefined; return message; }, }; -function createBaseGenerativeGoogleMetadata_UsageMetadata(): GenerativeGoogleMetadata_UsageMetadata { - return { promptTokenCount: undefined, candidatesTokenCount: undefined, totalTokenCount: undefined }; +function createBaseGenerativeFriendliAIMetadata_Usage(): GenerativeFriendliAIMetadata_Usage { + return { promptTokens: undefined, completionTokens: undefined, totalTokens: undefined }; } -export const GenerativeGoogleMetadata_UsageMetadata = { - encode(message: GenerativeGoogleMetadata_UsageMetadata, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { - if (message.promptTokenCount !== undefined) { - writer.uint32(8).int64(message.promptTokenCount); +export const GenerativeFriendliAIMetadata_Usage = { + encode(message: GenerativeFriendliAIMetadata_Usage, writer: _m0.Writer = _m0.Writer.create()): _m0.Writer { + if (message.promptTokens !== undefined) { + writer.uint32(8).int64(message.promptTokens); } - if (message.candidatesTokenCount !== undefined) { - writer.uint32(16).int64(message.candidatesTokenCount); + if (message.completionTokens !== undefined) { + writer.uint32(16).int64(message.completionTokens); } - if (message.totalTokenCount !== undefined) { - writer.uint32(24).int64(message.totalTokenCount); + if (message.totalTokens !== undefined) { + writer.uint32(24).int64(message.totalTokens); } return writer; }, - decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeGoogleMetadata_UsageMetadata { + decode(input: _m0.Reader | Uint8Array, length?: number): GenerativeFriendliAIMetadata_Usage { const reader = input instanceof _m0.Reader ? input : _m0.Reader.create(input); let end = length === undefined ? reader.len : reader.pos + length; - const message = createBaseGenerativeGoogleMetadata_UsageMetadata(); + const message = createBaseGenerativeFriendliAIMetadata_Usage(); while (reader.pos < end) { const tag = reader.uint32(); switch (tag >>> 3) { @@ -3544,21 +4187,21 @@ export const GenerativeGoogleMetadata_UsageMetadata = { break; } - message.promptTokenCount = longToNumber(reader.int64() as Long); + message.promptTokens = longToNumber(reader.int64() as Long); continue; case 2: if (tag !== 16) { break; } - message.candidatesTokenCount = longToNumber(reader.int64() as Long); + message.completionTokens = longToNumber(reader.int64() as Long); continue; case 3: if (tag !== 24) { break; } - message.totalTokenCount = longToNumber(reader.int64() as Long); + message.totalTokens = longToNumber(reader.int64() as Long); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -3569,38 +4212,36 @@ export const GenerativeGoogleMetadata_UsageMetadata = { return message; }, - fromJSON(object: any): GenerativeGoogleMetadata_UsageMetadata { + fromJSON(object: any): GenerativeFriendliAIMetadata_Usage { return { - promptTokenCount: isSet(object.promptTokenCount) ? globalThis.Number(object.promptTokenCount) : undefined, - candidatesTokenCount: isSet(object.candidatesTokenCount) - ? globalThis.Number(object.candidatesTokenCount) - : undefined, - totalTokenCount: isSet(object.totalTokenCount) ? globalThis.Number(object.totalTokenCount) : undefined, + promptTokens: isSet(object.promptTokens) ? globalThis.Number(object.promptTokens) : undefined, + completionTokens: isSet(object.completionTokens) ? globalThis.Number(object.completionTokens) : undefined, + totalTokens: isSet(object.totalTokens) ? globalThis.Number(object.totalTokens) : undefined, }; }, - toJSON(message: GenerativeGoogleMetadata_UsageMetadata): unknown { + toJSON(message: GenerativeFriendliAIMetadata_Usage): unknown { const obj: any = {}; - if (message.promptTokenCount !== undefined) { - obj.promptTokenCount = Math.round(message.promptTokenCount); + if (message.promptTokens !== undefined) { + obj.promptTokens = Math.round(message.promptTokens); } - if (message.candidatesTokenCount !== undefined) { - obj.candidatesTokenCount = Math.round(message.candidatesTokenCount); + if (message.completionTokens !== undefined) { + obj.completionTokens = Math.round(message.completionTokens); } - if (message.totalTokenCount !== undefined) { - obj.totalTokenCount = Math.round(message.totalTokenCount); + if (message.totalTokens !== undefined) { + obj.totalTokens = Math.round(message.totalTokens); } return obj; }, - create(base?: DeepPartial): GenerativeGoogleMetadata_UsageMetadata { - return GenerativeGoogleMetadata_UsageMetadata.fromPartial(base ?? {}); + create(base?: DeepPartial): GenerativeFriendliAIMetadata_Usage { + return GenerativeFriendliAIMetadata_Usage.fromPartial(base ?? {}); }, - fromPartial(object: DeepPartial): GenerativeGoogleMetadata_UsageMetadata { - const message = createBaseGenerativeGoogleMetadata_UsageMetadata(); - message.promptTokenCount = object.promptTokenCount ?? undefined; - message.candidatesTokenCount = object.candidatesTokenCount ?? undefined; - message.totalTokenCount = object.totalTokenCount ?? undefined; + fromPartial(object: DeepPartial): GenerativeFriendliAIMetadata_Usage { + const message = createBaseGenerativeFriendliAIMetadata_Usage(); + message.promptTokens = object.promptTokens ?? undefined; + message.completionTokens = object.completionTokens ?? undefined; + message.totalTokens = object.totalTokens ?? undefined; return message; }, }; @@ -3613,10 +4254,11 @@ function createBaseGenerativeMetadata(): GenerativeMetadata { cohere: undefined, dummy: undefined, mistral: undefined, - octoai: undefined, ollama: undefined, openai: undefined, google: undefined, + databricks: undefined, + friendliai: undefined, }; } @@ -3640,17 +4282,20 @@ export const GenerativeMetadata = { if (message.mistral !== undefined) { GenerativeMistralMetadata.encode(message.mistral, writer.uint32(50).fork()).ldelim(); } - if (message.octoai !== undefined) { - GenerativeOctoAIMetadata.encode(message.octoai, writer.uint32(58).fork()).ldelim(); - } if (message.ollama !== undefined) { - GenerativeOllamaMetadata.encode(message.ollama, writer.uint32(66).fork()).ldelim(); + GenerativeOllamaMetadata.encode(message.ollama, writer.uint32(58).fork()).ldelim(); } if (message.openai !== undefined) { - GenerativeOpenAIMetadata.encode(message.openai, writer.uint32(74).fork()).ldelim(); + GenerativeOpenAIMetadata.encode(message.openai, writer.uint32(66).fork()).ldelim(); } if (message.google !== undefined) { - GenerativeGoogleMetadata.encode(message.google, writer.uint32(82).fork()).ldelim(); + GenerativeGoogleMetadata.encode(message.google, writer.uint32(74).fork()).ldelim(); + } + if (message.databricks !== undefined) { + GenerativeDatabricksMetadata.encode(message.databricks, writer.uint32(82).fork()).ldelim(); + } + if (message.friendliai !== undefined) { + GenerativeFriendliAIMetadata.encode(message.friendliai, writer.uint32(90).fork()).ldelim(); } return writer; }, @@ -3709,28 +4354,35 @@ export const GenerativeMetadata = { break; } - message.octoai = GenerativeOctoAIMetadata.decode(reader, reader.uint32()); + message.ollama = GenerativeOllamaMetadata.decode(reader, reader.uint32()); continue; case 8: if (tag !== 66) { break; } - message.ollama = GenerativeOllamaMetadata.decode(reader, reader.uint32()); + message.openai = GenerativeOpenAIMetadata.decode(reader, reader.uint32()); continue; case 9: if (tag !== 74) { break; } - message.openai = GenerativeOpenAIMetadata.decode(reader, reader.uint32()); + message.google = GenerativeGoogleMetadata.decode(reader, reader.uint32()); continue; case 10: if (tag !== 82) { break; } - message.google = GenerativeGoogleMetadata.decode(reader, reader.uint32()); + message.databricks = GenerativeDatabricksMetadata.decode(reader, reader.uint32()); + continue; + case 11: + if (tag !== 90) { + break; + } + + message.friendliai = GenerativeFriendliAIMetadata.decode(reader, reader.uint32()); continue; } if ((tag & 7) === 4 || tag === 0) { @@ -3749,10 +4401,11 @@ export const GenerativeMetadata = { cohere: isSet(object.cohere) ? GenerativeCohereMetadata.fromJSON(object.cohere) : undefined, dummy: isSet(object.dummy) ? GenerativeDummyMetadata.fromJSON(object.dummy) : undefined, mistral: isSet(object.mistral) ? GenerativeMistralMetadata.fromJSON(object.mistral) : undefined, - octoai: isSet(object.octoai) ? GenerativeOctoAIMetadata.fromJSON(object.octoai) : undefined, ollama: isSet(object.ollama) ? GenerativeOllamaMetadata.fromJSON(object.ollama) : undefined, openai: isSet(object.openai) ? GenerativeOpenAIMetadata.fromJSON(object.openai) : undefined, google: isSet(object.google) ? GenerativeGoogleMetadata.fromJSON(object.google) : undefined, + databricks: isSet(object.databricks) ? GenerativeDatabricksMetadata.fromJSON(object.databricks) : undefined, + friendliai: isSet(object.friendliai) ? GenerativeFriendliAIMetadata.fromJSON(object.friendliai) : undefined, }; }, @@ -3776,9 +4429,6 @@ export const GenerativeMetadata = { if (message.mistral !== undefined) { obj.mistral = GenerativeMistralMetadata.toJSON(message.mistral); } - if (message.octoai !== undefined) { - obj.octoai = GenerativeOctoAIMetadata.toJSON(message.octoai); - } if (message.ollama !== undefined) { obj.ollama = GenerativeOllamaMetadata.toJSON(message.ollama); } @@ -3788,6 +4438,12 @@ export const GenerativeMetadata = { if (message.google !== undefined) { obj.google = GenerativeGoogleMetadata.toJSON(message.google); } + if (message.databricks !== undefined) { + obj.databricks = GenerativeDatabricksMetadata.toJSON(message.databricks); + } + if (message.friendliai !== undefined) { + obj.friendliai = GenerativeFriendliAIMetadata.toJSON(message.friendliai); + } return obj; }, @@ -3814,9 +4470,6 @@ export const GenerativeMetadata = { message.mistral = (object.mistral !== undefined && object.mistral !== null) ? GenerativeMistralMetadata.fromPartial(object.mistral) : undefined; - message.octoai = (object.octoai !== undefined && object.octoai !== null) - ? GenerativeOctoAIMetadata.fromPartial(object.octoai) - : undefined; message.ollama = (object.ollama !== undefined && object.ollama !== null) ? GenerativeOllamaMetadata.fromPartial(object.ollama) : undefined; @@ -3826,6 +4479,12 @@ export const GenerativeMetadata = { message.google = (object.google !== undefined && object.google !== null) ? GenerativeGoogleMetadata.fromPartial(object.google) : undefined; + message.databricks = (object.databricks !== undefined && object.databricks !== null) + ? GenerativeDatabricksMetadata.fromPartial(object.databricks) + : undefined; + message.friendliai = (object.friendliai !== undefined && object.friendliai !== null) + ? GenerativeFriendliAIMetadata.fromPartial(object.friendliai) + : undefined; return message; }, };