@@ -171,7 +171,12 @@ def placeholder_regex(
171171 "multi_line" : multi_line ,
172172 }
173173
174- return EvaluatorDetails (slug = "placeholder-regex" , version = None , config = config , required_input_fields = ["text" , "placeholder_value" ])
174+ return EvaluatorDetails (
175+ slug = "placeholder-regex" ,
176+ version = None ,
177+ config = config ,
178+ required_input_fields = ["text" , "placeholder_value" ],
179+ )
175180
176181 @staticmethod
177182 def char_count (
@@ -204,7 +209,12 @@ def char_count_ratio(
204209 """
205210 config : Dict [str , Any ] = {}
206211
207- return EvaluatorDetails (slug = "char-count-ratio" , version = None , config = config , required_input_fields = ["numerator_text" , "denominator_text" ])
212+ return EvaluatorDetails (
213+ slug = "char-count-ratio" ,
214+ version = None ,
215+ config = config ,
216+ required_input_fields = ["numerator_text" , "denominator_text" ],
217+ )
208218
209219 @staticmethod
210220 def word_count () -> EvaluatorDetails :
@@ -236,7 +246,12 @@ def word_count_ratio(
236246 """
237247 config : Dict [str , Any ] = {}
238248
239- return EvaluatorDetails (slug = "word-count-ratio" , version = None , config = config , required_input_fields = ["numerator_text" , "denominator_text" ])
249+ return EvaluatorDetails (
250+ slug = "word-count-ratio" ,
251+ version = None ,
252+ config = config ,
253+ required_input_fields = ["numerator_text" , "denominator_text" ],
254+ )
240255
241256 @staticmethod
242257 def answer_relevancy (
@@ -253,7 +268,12 @@ def answer_relevancy(
253268 """
254269 config : Dict [str , Any ] = {}
255270
256- return EvaluatorDetails (slug = "answer-relevancy" , version = None , config = config , required_input_fields = ["question" , "answer" ])
271+ return EvaluatorDetails (
272+ slug = "answer-relevancy" ,
273+ version = None ,
274+ config = config ,
275+ required_input_fields = ["question" , "answer" ],
276+ )
257277
258278 @staticmethod
259279 def faithfulness (
@@ -271,7 +291,12 @@ def faithfulness(
271291 """
272292 config : Dict [str , Any ] = {}
273293
274- return EvaluatorDetails (slug = "faithfulness" , version = None , config = config , required_input_fields = ["question" , "completion" , "context" ])
294+ return EvaluatorDetails (
295+ slug = "faithfulness" ,
296+ version = None ,
297+ config = config ,
298+ required_input_fields = ["question" , "completion" , "context" ],
299+ )
275300
276301 @staticmethod
277302 def profanity_detector () -> EvaluatorDetails :
@@ -354,7 +379,12 @@ def semantic_similarity(
354379 """
355380 config : Dict [str , Any ] = {}
356381
357- return EvaluatorDetails (slug = "semantic-similarity" , version = None , config = config , required_input_fields = ["completion" , "reference" ])
382+ return EvaluatorDetails (
383+ slug = "semantic-similarity" ,
384+ version = None ,
385+ config = config ,
386+ required_input_fields = ["completion" , "reference" ],
387+ )
358388
359389 @staticmethod
360390 def agent_goal_accuracy (
@@ -365,13 +395,19 @@ def agent_goal_accuracy(
365395 Required task output fields:
366396 - question: The input question or goal
367397 - completion: The agent's completion
398+ - reference: The reference answer or goal
368399
369400 Returns:
370401 EvaluatorDetails configured for agent goal accuracy evaluation
371402 """
372403 config : Dict [str , Any ] = {}
373404
374- return EvaluatorDetails (slug = "agent-goal-accuracy" , version = None , config = config , required_input_fields = ["question" , "completion" , "reference" ])
405+ return EvaluatorDetails (
406+ slug = "agent-goal-accuracy" ,
407+ version = None ,
408+ config = config ,
409+ required_input_fields = ["question" , "completion" , "reference" ],
410+ )
375411
376412 @staticmethod
377413 def topic_adherence (
@@ -380,6 +416,7 @@ def topic_adherence(
380416 Topic adherence evaluator - validates topic adherence.
381417
382418 Required task output fields:
419+ - question: The input question or goal
383420 - completion: The completion text to evaluate
384421 - reference_topics: The expected topic or topics
385422
@@ -388,7 +425,12 @@ def topic_adherence(
388425 """
389426 config : Dict [str , Any ] = {}
390427
391- return EvaluatorDetails (slug = "topic-adherence" , version = None , config = config , required_input_fields = ["question" , "completion" , "reference_topics" ])
428+ return EvaluatorDetails (
429+ slug = "topic-adherence" ,
430+ version = None ,
431+ config = config ,
432+ required_input_fields = ["question" , "completion" , "reference_topics" ],
433+ )
392434
393435 @staticmethod
394436 def perplexity (
@@ -403,5 +445,10 @@ def perplexity(
403445 EvaluatorDetails configured for perplexity measurement
404446 """
405447 config : Dict [str , Any ] = {}
406-
407- return EvaluatorDetails (slug = "perplexity" , version = None , config = config , required_input_fields = ["prompt" ])
448+
449+ return EvaluatorDetails (
450+ slug = "perplexity" ,
451+ version = None ,
452+ config = config ,
453+ required_input_fields = ["prompt" ],
454+ )
0 commit comments