@@ -27,7 +27,7 @@ def __init__(
2727 y_column : Optional [str ] = None ,
2828 task_description : Optional [str ] = None ,
2929 n_subsamples : int = 30 ,
30- eval_strategy : EvalStrategy = "full" ,
30+ eval_strategy : " EvalStrategy" = "full" ,
3131 seed : int = 42 ,
3232 config : Optional ["ExperimentConfig" ] = None ,
3333 ) -> None :
@@ -70,7 +70,7 @@ def __init__(
7070 self .eval_cache : Dict [Tuple [str , str , str ], float ] = {} # (prompt, x, y): scores per datapoint
7171 self .seq_cache : Dict [Tuple [str , str , str ], str ] = {} # (prompt, x, y): generating sequence per datapoint
7272
73- def subsample (self , eval_strategy : EvalStrategy = None ) -> Tuple [List [str ], List [str ]]:
73+ def subsample (self , eval_strategy : " EvalStrategy" = None ) -> Tuple [List [str ], List [str ]]:
7474 """Subsample the dataset based on the specified parameters.
7575
7676 Args:
@@ -170,7 +170,7 @@ def evaluate(
170170 system_prompts : Optional [Union [str , List [str ]]] = None ,
171171 return_agg_scores : Literal [True ] = True ,
172172 return_seq : Literal [False ] = False ,
173- eval_strategy : Optional [EvalStrategy ] = None ,
173+ eval_strategy : Optional [" EvalStrategy" ] = None ,
174174 ) -> List [float ]:
175175 ...
176176
@@ -182,7 +182,7 @@ def evaluate(
182182 system_prompts : Optional [Union [str , List [str ]]] = None ,
183183 return_agg_scores : Literal [False ] = False ,
184184 return_seq : Literal [False ] = False ,
185- eval_strategy : Optional [EvalStrategy ] = None ,
185+ eval_strategy : Optional [" EvalStrategy" ] = None ,
186186 ) -> List [List [float ]]:
187187 ...
188188
@@ -194,7 +194,7 @@ def evaluate(
194194 system_prompts : Optional [Union [str , List [str ]]] = None ,
195195 return_agg_scores : Literal [False ] = False ,
196196 return_seq : Literal [True ] = True ,
197- eval_strategy : Optional [EvalStrategy ] = None ,
197+ eval_strategy : Optional [" EvalStrategy" ] = None ,
198198 ) -> Tuple [List [List [float ]], List [List [str ]]]:
199199 ...
200200
@@ -206,7 +206,7 @@ def evaluate(
206206 system_prompts : Optional [Union [str , List [str ]]] = None ,
207207 return_agg_scores : Literal [True ] = True ,
208208 return_seq : Literal [False ] = False ,
209- eval_strategy : Optional [EvalStrategy ] = None ,
209+ eval_strategy : Optional [" EvalStrategy" ] = None ,
210210 ) -> List [float ]:
211211 ...
212212
@@ -218,7 +218,7 @@ def evaluate(
218218 system_prompts : Optional [Union [str , List [str ]]] = None ,
219219 return_agg_scores : Literal [False ] = False ,
220220 return_seq : Literal [False ] = False ,
221- eval_strategy : Optional [EvalStrategy ] = None ,
221+ eval_strategy : Optional [" EvalStrategy" ] = None ,
222222 ) -> List [List [float ]]:
223223 ...
224224
@@ -230,7 +230,7 @@ def evaluate(
230230 system_prompts : Optional [Union [str , List [str ]]] = None ,
231231 return_agg_scores : Literal [False ] = False ,
232232 return_seq : Literal [True ] = True ,
233- eval_strategy : Optional [EvalStrategy ] = None ,
233+ eval_strategy : Optional [" EvalStrategy" ] = None ,
234234 ) -> Tuple [List [List [float ]], List [List [str ]]]:
235235 ...
236236
@@ -241,7 +241,7 @@ def evaluate(
241241 system_prompts : Optional [Union [str , List [str ]]] = None ,
242242 return_agg_scores : bool = True ,
243243 return_seq : bool = False ,
244- eval_strategy : Optional [EvalStrategy ] = None ,
244+ eval_strategy : Optional [" EvalStrategy" ] = None ,
245245 ) -> Union [List [float ], List [List [float ]], Tuple [List [List [float ]], List [List [str ]]]]:
246246 """Evaluate a set of prompts using a given predictor.
247247
0 commit comments