Skip to content

Commit f64177c

Browse files
committed
lint
1 parent 7f1729d commit f64177c

File tree

3 files changed

+60
-11
lines changed

3 files changed

+60
-11
lines changed

packages/traceloop-sdk/traceloop/sdk/evaluator/config.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,8 @@ class EvaluatorDetails(BaseModel):
1010
slug: The evaluator slug/identifier
1111
version: Optional version of the evaluator
1212
config: Optional configuration dictionary for the evaluator
13-
required_input_fields: Optional list of required fields to the evaluator input. These fields must be present in the task output.
13+
required_input_fields: Optional list of required fields to the evaluator
14+
input. These fields must be present in the task output.
1415
1516
Example:
1617
>>> EvaluatorDetails(slug="pii-detector", config={"probability_threshold": 0.8}, required_input_fields=["text"])

packages/traceloop-sdk/traceloop/sdk/evaluator/evaluators_made_by_traceloop.py

Lines changed: 57 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -171,7 +171,12 @@ def placeholder_regex(
171171
"multi_line": multi_line,
172172
}
173173

174-
return EvaluatorDetails(slug="placeholder-regex", version=None, config=config, required_input_fields=["text", "placeholder_value"])
174+
return EvaluatorDetails(
175+
slug="placeholder-regex",
176+
version=None,
177+
config=config,
178+
required_input_fields=["text", "placeholder_value"],
179+
)
175180

176181
@staticmethod
177182
def char_count(
@@ -204,7 +209,12 @@ def char_count_ratio(
204209
"""
205210
config: Dict[str, Any] = {}
206211

207-
return EvaluatorDetails(slug="char-count-ratio", version=None, config=config, required_input_fields=["numerator_text", "denominator_text"])
212+
return EvaluatorDetails(
213+
slug="char-count-ratio",
214+
version=None,
215+
config=config,
216+
required_input_fields=["numerator_text", "denominator_text"],
217+
)
208218

209219
@staticmethod
210220
def word_count() -> EvaluatorDetails:
@@ -236,7 +246,12 @@ def word_count_ratio(
236246
"""
237247
config: Dict[str, Any] = {}
238248

239-
return EvaluatorDetails(slug="word-count-ratio", version=None, config=config, required_input_fields=["numerator_text", "denominator_text"])
249+
return EvaluatorDetails(
250+
slug="word-count-ratio",
251+
version=None,
252+
config=config,
253+
required_input_fields=["numerator_text", "denominator_text"],
254+
)
240255

241256
@staticmethod
242257
def answer_relevancy(
@@ -253,7 +268,12 @@ def answer_relevancy(
253268
"""
254269
config: Dict[str, Any] = {}
255270

256-
return EvaluatorDetails(slug="answer-relevancy", version=None, config=config, required_input_fields=["question", "answer"])
271+
return EvaluatorDetails(
272+
slug="answer-relevancy",
273+
version=None,
274+
config=config,
275+
required_input_fields=["question", "answer"],
276+
)
257277

258278
@staticmethod
259279
def faithfulness(
@@ -271,7 +291,12 @@ def faithfulness(
271291
"""
272292
config: Dict[str, Any] = {}
273293

274-
return EvaluatorDetails(slug="faithfulness", version=None, config=config, required_input_fields=["question", "completion", "context"])
294+
return EvaluatorDetails(
295+
slug="faithfulness",
296+
version=None,
297+
config=config,
298+
required_input_fields=["question", "completion", "context"],
299+
)
275300

276301
@staticmethod
277302
def profanity_detector() -> EvaluatorDetails:
@@ -354,7 +379,12 @@ def semantic_similarity(
354379
"""
355380
config: Dict[str, Any] = {}
356381

357-
return EvaluatorDetails(slug="semantic-similarity", version=None, config=config, required_input_fields=["completion", "reference"])
382+
return EvaluatorDetails(
383+
slug="semantic-similarity",
384+
version=None,
385+
config=config,
386+
required_input_fields=["completion", "reference"],
387+
)
358388

359389
@staticmethod
360390
def agent_goal_accuracy(
@@ -365,13 +395,19 @@ def agent_goal_accuracy(
365395
Required task output fields:
366396
- question: The input question or goal
367397
- completion: The agent's completion
398+
- reference: The reference answer or goal
368399
369400
Returns:
370401
EvaluatorDetails configured for agent goal accuracy evaluation
371402
"""
372403
config: Dict[str, Any] = {}
373404

374-
return EvaluatorDetails(slug="agent-goal-accuracy", version=None, config=config, required_input_fields=["question", "completion", "reference"])
405+
return EvaluatorDetails(
406+
slug="agent-goal-accuracy",
407+
version=None,
408+
config=config,
409+
required_input_fields=["question", "completion", "reference"],
410+
)
375411

376412
@staticmethod
377413
def topic_adherence(
@@ -380,6 +416,7 @@ def topic_adherence(
380416
Topic adherence evaluator - validates topic adherence.
381417
382418
Required task output fields:
419+
- question: The input question or goal
383420
- completion: The completion text to evaluate
384421
- reference_topics: The expected topic or topics
385422
@@ -388,7 +425,12 @@ def topic_adherence(
388425
"""
389426
config: Dict[str, Any] = {}
390427

391-
return EvaluatorDetails(slug="topic-adherence", version=None, config=config, required_input_fields=["question", "completion", "reference_topics"])
428+
return EvaluatorDetails(
429+
slug="topic-adherence",
430+
version=None,
431+
config=config,
432+
required_input_fields=["question", "completion", "reference_topics"],
433+
)
392434

393435
@staticmethod
394436
def perplexity(
@@ -403,5 +445,10 @@ def perplexity(
403445
EvaluatorDetails configured for perplexity measurement
404446
"""
405447
config: Dict[str, Any] = {}
406-
407-
return EvaluatorDetails(slug="perplexity", version=None, config=config, required_input_fields=["prompt"])
448+
449+
return EvaluatorDetails(
450+
slug="perplexity",
451+
version=None,
452+
config=config,
453+
required_input_fields=["prompt"],
454+
)

packages/traceloop-sdk/traceloop/sdk/experiment/experiment.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -161,6 +161,7 @@ async def _run_locally(
161161
errors: List[str] = []
162162

163163
evaluators_to_validate = [evaluator for evaluator in evaluators if isinstance(evaluator, EvaluatorDetails)]
164+
164165
async def run_single_row(row: Optional[Dict[str, Any]]) -> TaskResponse:
165166
try:
166167
task_result = await task(row)

0 commit comments

Comments
 (0)