@@ -1231,7 +1231,7 @@ def create(
12311231 length_penalty : float = 1.0 ,
12321232 no_repeat_ngram_size : int = 0 ,
12331233 early_stopping : bool = False ,
1234- aggregate_paths : bool = False ,
1234+ aggregate_paths : bool = True ,
12351235 # Entropy specific params
12361236 top_k : int = 27 ,
12371237 min_p : float = 0.03 ,
@@ -1263,10 +1263,10 @@ def create(
12631263 cot_params = {
12641264 "k" : k ,
12651265 "num_beams" : num_beams ,
1266- "max_new_tokens" : max_tokens if max_tokens is not None else 4096 ,
1266+ "max_new_tokens" : max_tokens if max_tokens is not None else 512 ,
12671267 "temperature" : temperature ,
12681268 "top_p" : top_p ,
1269- "repetition_penalty" : 1.0 + frequency_penalty ,
1269+ "repetition_penalty" : 1.0 ,
12701270 "length_penalty" : length_penalty ,
12711271 "no_repeat_ngram_size" : no_repeat_ngram_size ,
12721272 "early_stopping" : early_stopping ,
@@ -1293,7 +1293,7 @@ def create(
12931293 # Use directly available parameters for entropy decoding
12941294
12951295 entropy_params = {
1296- "max_new_tokens" : max_tokens if max_tokens is not None else 4096 ,
1296+ "max_new_tokens" : max_tokens if max_tokens is not None else 512 ,
12971297 "temperature" : 0.666 ,
12981298 "top_p" : 0.90 ,
12991299 "top_k" : top_k ,
0 commit comments