@@ -58,6 +58,7 @@ def evaluate(
5858 metrics : t .Optional [t .Sequence [Metric ]] = None ,
5959 llm : t .Optional [BaseRagasLLM | LangchainLLM ] = None ,
6060 embeddings : t .Optional [BaseRagasEmbeddings | LangchainEmbeddings ] = None ,
61+ experiment_name : t .Optional [str ] = None ,
6162 callbacks : Callbacks = None ,
6263 run_config : t .Optional [RunConfig ] = None ,
6364 token_usage_parser : t .Optional [TokenUsageParser ] = None ,
@@ -87,6 +88,8 @@ def evaluate(
8788 The embeddings to use for the metrics. If not provided then ragas will use
8889 the default embeddings for metrics which require embeddings. This can we overridden by the embeddings specified in
8990 the metric level with `metric.embeddings`.
91+ experiment_name: str, optional
92+ The name of the experiment to track. This is used to track the evaluation in the tracing tools.
9093 callbacks: Callbacks, optional
9194 Lifecycle Langchain Callbacks to run during evaluation. Check the
9295 [langchain documentation](https://python.langchain.com/docs/modules/callbacks/)
@@ -246,7 +249,7 @@ def evaluate(
246249 # new evaluation chain
247250 row_run_managers = []
248251 evaluation_rm , evaluation_group_cm = new_group (
249- name = RAGAS_EVALUATION_CHAIN_NAME ,
252+ name = experiment_name or RAGAS_EVALUATION_CHAIN_NAME ,
250253 inputs = {},
251254 callbacks = callbacks ,
252255 metadata = {"type" : ChainType .EVALUATION },
0 commit comments