You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: sdk/ai/azure-ai-projects/README.md
+116Lines changed: 116 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -718,6 +718,119 @@ print("Deleted agent")
718
718
719
719
<!-- END SNIPPET -->
720
720
721
+
### Evaluation
722
+
723
+
Evaluation in Azure AI Project client library is designed to assess the performance of generative AI applications in the cloud. The output of Generative AI application is quantitively measured with mathematical based metrics, AI-assisted quality and safety metrics. Metrics are defined as evaluators. Built-in or custom evaluators can provide comprehensive insights into the application's capabilities and limitations.
724
+
725
+
#### Evaluator
726
+
727
+
Evaluators are custom or prebuilt classes or functions that are designed to measure the quality of the outputs from language models or generative AI applications.
728
+
729
+
Evaluators are made available via [azure-ai-evaluation][azure_ai_evaluation] SDK for local experience and also in [Evaluator Library][evaluator_library] in Azure AI Studio for using them in the cloud.
730
+
731
+
More details on built-in and custom evaluators can be found [here][evaluators].
732
+
733
+
#### Run Evaluation in cloud:
734
+
735
+
To run evaluation in cloud the following are needed:
736
+
737
+
- Evaluators
738
+
- Data to be evaluated
739
+
-[Optional] Azure Open AI model.
740
+
741
+
##### Evaluators
742
+
743
+
For running evaluator in cloud, evaluator `ID` is needed. To get it via code you use [azure-ai-evaluation][azure_ai_evaluation]
744
+
745
+
```
746
+
# pip install azure-ai-evaluation
747
+
748
+
from azure.ai.evaluation import RelevanceEvaluator
749
+
750
+
evaluator_id = RelevanceEvaluator.id
751
+
```
752
+
753
+
##### Data to be evaluated
754
+
755
+
Evaluation in the cloud supports data in form of `jsonl` file. Data can be uploaded via the helper method `upload_file` on the project client.
Azure AI Studio project comes with a default Azure Open AI endpoint which can be easily accessed using following code. This gives you the endpoint details for you Azure OpenAI endpoint. Some of the evaluators need model that supports chat completion.
NOTE: For running evaluators locally refer to [Evaluate with the Azure AI Evaluation SDK][evaluators].
833
+
721
834
#### Tracing
722
835
723
836
You can add an Application Insights Azure resource to your Azure AI Studio project. See the Tracing tab in your studio. If one was enabled, you can get the Application Insights connection string, configure your Agents, and observe the full execution path through Azure Monitor. Typically, you might want to start tracing before you create an Agent.
@@ -873,3 +986,6 @@ additional questions or comments.
0 commit comments