Skip to content

Commit f31ea42

Browse files
authored
examples: fix api changes (#2129)
1 parent 19ecd56 commit f31ea42

File tree

4 files changed

+15
-15
lines changed

4 files changed

+15
-15
lines changed

experimental/ragas_examples/agent_evals/evals.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -5,13 +5,13 @@
55

66
math_agent = get_default_agent()
77

8-
@numeric_metric(name="correctness")
8+
@numeric_metric(name="correctness", allowed_values=(0.0, 1.0))
99
def correctness_metric(prediction: float, actual: float):
1010
"""Calculate correctness of the prediction."""
1111
if isinstance(prediction, str) and "ERROR" in prediction:
1212
return 0.0
1313
result = 1.0 if abs(prediction - actual) < 1e-5 else 0.0
14-
return MetricResult(result=result, reason=f"Prediction: {prediction}, Actual: {actual}")
14+
return MetricResult(value=result, reason=f"Prediction: {prediction}, Actual: {actual}")
1515

1616

1717
def load_dataset():
@@ -56,13 +56,13 @@ async def run_experiment(row):
5656
"expected_answer": expected_answer,
5757
"prediction": prediction.get("result"),
5858
"log_file": prediction.get("log_file"),
59-
"correctness": correctness.result
59+
"correctness": correctness.value
6060
}
6161

6262

6363
async def main():
6464
dataset = load_dataset()
65-
_ = await run_experiment.run_async(dataset)
65+
_ = await run_experiment.arun(dataset)
6666

6767

6868
if __name__ == "__main__":

experimental/ragas_examples/prompt_evals/evals.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -4,10 +4,10 @@
44

55
from .prompt import run_prompt
66

7-
@discrete_metric(name="accuracy", values=["pass", "fail"])
7+
@discrete_metric(name="accuracy", allowed_values=["pass", "fail"])
88
def my_metric(prediction: str, actual: str):
99
"""Calculate accuracy of the prediction."""
10-
return MetricResult(result="pass", reason="") if prediction == actual else MetricResult(result="fail", reason="")
10+
return MetricResult(value="pass", reason="") if prediction == actual else MetricResult(value="fail", reason="")
1111

1212

1313
@experiment()
@@ -22,7 +22,7 @@ async def run_experiment(row):
2222
experiment_view = {
2323
**row,
2424
"response":response,
25-
"score":score.result,
25+
"score":score.value,
2626
}
2727
return experiment_view
2828

@@ -56,7 +56,7 @@ def load_dataset():
5656

5757
async def main():
5858
dataset = load_dataset()
59-
experiment_results = await run_experiment.run_async(dataset)
59+
experiment_results = await run_experiment.arun(dataset)
6060
print("Experiment completed successfully!")
6161
print("Experiment results:", experiment_results)
6262

experimental/ragas_examples/rag_eval/evals.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -33,8 +33,8 @@ def load_dataset():
3333

3434
my_metric = DiscreteMetric(
3535
name="correctness",
36-
prompt = "Check if the response contains points mentioned from the grading notes and return 'pass' or 'fail'.\nResponse: {response} Grading Notes: {grading_notes}",
37-
values=["pass", "fail"],
36+
prompt="Check if the response contains points mentioned from the grading notes and return 'pass' or 'fail'.\nResponse: {response} Grading Notes: {grading_notes}",
37+
allowed_values=["pass", "fail"],
3838
)
3939

4040
@experiment()
@@ -50,7 +50,7 @@ async def run_experiment(row):
5050
experiment_view = {
5151
**row,
5252
"response": response,
53-
"score": score.result,
53+
"score": score.value,
5454
"log_file": response.get("logs", " "),
5555
}
5656
return experiment_view
@@ -59,7 +59,7 @@ async def run_experiment(row):
5959
async def main():
6060
dataset = load_dataset()
6161
print("dataset loaded successfully", dataset)
62-
await run_experiment.run_async(dataset)
62+
await run_experiment.arun(dataset)
6363

6464
if __name__ == "__main__":
6565
import asyncio

experimental/ragas_examples/workflow_eval/evals.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -80,7 +80,7 @@ def load_dataset():
8080
my_metric = DiscreteMetric(
8181
name="response_quality",
8282
prompt="Evaluate the response based on the pass criteria: {pass_criteria}. Does the response meet the criteria? Return 'pass' or 'fail'.\nResponse: {response}",
83-
values=["pass", "fail"],
83+
allowed_values=["pass", "fail"],
8484
)
8585

8686

@@ -99,15 +99,15 @@ async def run_experiment(row):
9999
experiment_view = {
100100
**row,
101101
"response": response.get("response_template", " "),
102-
"score": score.result,
102+
"score": score.value,
103103
"score_reason": score.reason,
104104
}
105105
return experiment_view
106106

107107

108108
async def main():
109109
dataset = load_dataset()
110-
_ = await run_experiment.run_async(dataset)
110+
_ = await run_experiment.arun(dataset)
111111

112112
if __name__ == "__main__":
113113
import asyncio

0 commit comments

Comments
 (0)