Skip to content

Commit 4ba0ad0

Browse files
committed
fixes
1 parent fa707fc commit 4ba0ad0

File tree

1 file changed

+7
-8
lines changed

1 file changed

+7
-8
lines changed

articles/ai-studio/concepts/evaluation-approach-gen-ai.md

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -72,23 +72,22 @@ After deployment, the AI application enters the post-production evaluation phase
7272

7373
By [continuously monitoring the AI application’s behavior in production](https://aka.ms/AzureAIMonitoring), you can maintain high-quality user experiences and swiftly address any issues that surface.
7474

75-
76-
## Conclusion
75+
## Conclusion
7776

7877
GenAIOps is all about establishing a reliable and repeatable process for managing generative AI applications across their lifecycle. Evaluation plays a vital role at each stage, from base model selection, through pre-production testing, to ongoing post-production monitoring. By systematically measuring and addressing risks and refining AI systems at every step, teams can build generative AI solutions that are not only powerful but also trustworthy and safe for real-world use.
7978

80-
Cheat sheet:
79+
Cheat sheet:
8180

82-
| Purpose | Process | Parameters |
83-
| -----| -----| ----|
81+
| Purpose | Process | Parameters |
82+
| -----| -----| ----|
8483
| What are you evaluating for? | Identify or build relevant evaluators | - [Quality and performance](./evaluation-metrics-built-in.md?tabs=warning#generation-quality-metrics) ( [Quality and performance sample notebook](https://github.com/Azure-Samples/rag-data-openai-python-promptflow/blob/main/src/evaluation/evaluate.py)) </br> - [Safety and Security](./evaluation-metrics-built-in.md?tabs=warning#risk-and-safety-metrics)) ([Safety and Security sample notebook]((https://github.com/Azure-Samples/rag-data-openai-python-promptflow/blob/main/src/evaluation/evaluatesafetyrisks.py))) </br> [Custom](../how-to/develop/evaluate-sdk.md#custom-evaluators) ([Custom sample notebook](https://github.com/Azure-Samples/rag-data-openai-python-promptflow/blob/main/src/evaluation/evaluate.py))] |
8584
| What data should you use? | Upload or generate relevant dataset | [Generic simulator for measuring Quality and Performance](./concept-synthetic-data.md) ( Generic simulator sample notebook|(https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/system/finetune/Llama-notebooks/datagen/synthetic-data-generation.ipynb)] </br> - Adversarial simulator for measuring Safety and Security [[Adversarial simulator Docs](../how-to/develop/simulator-interaction-data.md), [Adversarial simulator sample notebook](https://github.com/Azure-Samples/rag-data-openai-python-promptflow/blob/main/src/evaluation/simulate_and_evaluate_online_endpoint.ipynb)] |
8685
| What resources should conduct the evaluation? | Run evaluation | - Local run <br> </br> - Remote cloud run |
87-
| How did my model/app perform? | Analyze results | [View aggregate scores, view details, score details, compare eval runs](..//how-to/evaluate-results.md)]
86+
| How did my model/app perform? | Analyze results | [View aggregate scores, view details, score details, compare eval runs](..//how-to/evaluate-results.md)] |
8887

89-
## Next steps
88+
## Related content
9089

9190
- [Evaluate your generative AI apps via the playground](../how-to/evaluate-prompts-playground.md)
9291
- [Evaluate your generative AI apps with the Azure AI Foundry SDK or portal](../how-to/evaluate-generative-ai-app.md)
93-
- [Evaluation and monitoring metrics for generative AI](..concepts/evaluation-metrics-built-in.md)
92+
- [Evaluation and monitoring metrics for generative AI](evaluation-metrics-built-in.md)
9493
- [Transparency Note for Azure AI Foundry safety evaluations](safety-evaluations-transparency-note.md)

0 commit comments

Comments
 (0)