1616 */
1717
1818import { BaseConversationalTask , BaseTextGenerationTask } from "./providerHelper" ;
19- import type {
20- ChatCompletionOutput ,
21- TextGenerationOutput ,
22- TextGenerationOutputFinishReason ,
23- } from "@huggingface/tasks" ;
19+ import type { ChatCompletionOutput , TextGenerationOutput , TextGenerationOutputFinishReason } from "@huggingface/tasks" ;
2420import { InferenceOutputError } from "../lib/InferenceOutputError" ;
2521import type { BodyParams } from "../types" ;
2622import { omit } from "../utils/omit" ;
@@ -54,9 +50,9 @@ export class OvhCloudTextGenerationTask extends BaseTextGenerationTask {
5450 ...omit ( params . args , [ "inputs" , "parameters" ] ) ,
5551 ...( params . args . parameters
5652 ? {
57- max_tokens : ( params . args . parameters as Record < string , unknown > ) . max_new_tokens ,
58- ...omit ( params . args . parameters as Record < string , unknown > , "max_new_tokens" ) ,
59- }
53+ max_tokens : ( params . args . parameters as Record < string , unknown > ) . max_new_tokens ,
54+ ...omit ( params . args . parameters as Record < string , unknown > , "max_new_tokens" ) ,
55+ }
6056 : undefined ) ,
6157 prompt : params . args . inputs ,
6258 } ;
@@ -76,5 +72,4 @@ export class OvhCloudTextGenerationTask extends BaseTextGenerationTask {
7672 }
7773 throw new InferenceOutputError ( "Expected OVHcloud text generation response format" ) ;
7874 }
79-
80- }
75+ }
0 commit comments