@@ -1530,7 +1530,7 @@ async def mock_async_task(inputs: TaskInput) -> TaskOutput:
15301530 return TaskOutput (answer = 'Paris' )
15311531 return TaskOutput (answer = 'Unknown' ) # pragma: no cover
15321532
1533- await example_dataset .evaluate (mock_async_task )
1533+ await example_dataset .evaluate (mock_async_task , metadata = { 'key' : 'value' } )
15341534
15351535 spans = capfire .exporter .exported_spans_as_dict (parse_json_attributes = True )
15361536 spans .sort (key = lambda s : s ['start_time' ])
@@ -1556,6 +1556,7 @@ async def mock_async_task(inputs: TaskInput) -> TaskOutput:
15561556 'gen_ai.operation.name' : {},
15571557 'n_cases' : {},
15581558 'name' : {},
1559+ 'metadata' : {'type' : 'object' },
15591560 'logfire.experiment.metadata' : {
15601561 'type' : 'object' ,
15611562 'properties' : {
@@ -1571,11 +1572,13 @@ async def mock_async_task(inputs: TaskInput) -> TaskOutput:
15711572 'type' : 'object' ,
15721573 },
15731574 'logfire.msg' : 'evaluate mock_async_task' ,
1575+ 'metadata' : {'key' : 'value' },
15741576 'logfire.msg_template' : 'evaluate {name}' ,
15751577 'logfire.span_type' : 'span' ,
15761578 'n_cases' : 2 ,
15771579 'logfire.experiment.metadata' : {
15781580 'n_cases' : 2 ,
1581+ 'metadata' : {'key' : 'value' },
15791582 'averages' : {
15801583 'name' : 'Averages' ,
15811584 'scores' : {'confidence' : 1.0 },
0 commit comments