Skip to content

Commit 7fd561f

Browse files
committed
fixes
1 parent cc313b1 commit 7fd561f

File tree

1 file changed

+4
-3
lines changed
  • articles/ai-studio/includes/evaluations/from-data

1 file changed

+4
-3
lines changed

articles/ai-studio/includes/evaluations/from-data/python.md

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -141,7 +141,7 @@ from azure.ai.resources.client import AIClient
141141
client = AIClient.from_config(DefaultAzureCredential())
142142
```
143143
> [!NOTE]
144-
> If only risk and safety metrics are passed into `metrics_list` then the `model_config` parameter in the interface below is optional. The Azure AI Studio safety evaluations back-end service provisions a GPT-4 model that can generate content risk severity scores and reasoning to enable you to evaluate your application for content harms.
144+
> If only risk and safety metrics are passed into `metrics_list` then the `model_config` parameter in the following interface is optional. The Azure AI Studio safety evaluations back-end service provisions a GPT-4 model that can generate content risk severity scores and reasoning to enable you to evaluate your application for content harms.
145145
146146
### Evaluate question answering: `qa`
147147

@@ -243,7 +243,7 @@ The contents of `eval_results.jsonl` looks like this:
243243

244244
The outputs of your risk and safety metrics will provide the following:
245245

246-
- `{metric_name}_defect_rate` which measures % of instances which surpassed the severity threshold (set to default 4) and is the aggregate metric over the whole dataset.
246+
- `{metric_name}_defect_rate`, which measures % of instances that surpassed the severity threshold (set to default 4) and is the aggregate metric over the whole dataset.
247247
- `{metric_name}_score` with a range between 0 and 7 severity for each data point. You can read more about the descriptions of each [content risk and severity scale](../../../concepts/evaluation-metrics-built-in.md).
248248
- `{metric_name}_reasoning` with a text reasoning for why a certain severity score was given for each data point.
249249

@@ -258,7 +258,8 @@ result.download_evaluation_artifacts("./myevalresults")
258258
259259
### Evaluate Conversation: `chat`
260260

261-
The same interface can be used with `evaluate()` for the conversation scenario but with data mapping required only for model output `y_pred` and `task_type="chat"` shown below
261+
The same interface can be used with `evaluate()` for the conversation scenario but with data mapping required only for model output `y_pred` and `task_type="chat"`.
262+
262263
```python
263264
task_type="chat",
264265
data_mapping={

0 commit comments

Comments
 (0)