Skip to content

Commit 43b496d

Browse files
committed
formatting
1 parent f5ba655 commit 43b496d

File tree

1 file changed

+4
-2
lines changed

1 file changed

+4
-2
lines changed

articles/ai-studio/concepts/model-benchmarks.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,13 +36,15 @@ Model benchmarks assess LLMs and SLMs across the following categories: quality,
3636

3737
Azure AI assesses the quality of LLMs and SLMs across various metrics that are grouped into two main categories: accuracy, and prompt assisted metrics:
3838

39-
**Accuracy**
39+
40+
- Accuracy
4041

4142
| Metric | Description |
4243
|--------|-------------|
4344
| Accuracy | Accuracy scores are available at the dataset and the model levels. At the dataset level, the score is the average value of an accuracy metric computed over all examples in the dataset. The accuracy metric used is `exact-match` in all cases, except for the _HumanEval_ dataset that uses a `pass@1` metric. Exact match compares model generated text with the correct answer according to the dataset, reporting one if the generated text matches the answer exactly and zero otherwise. The `pass@1` metric measures the proportion of model solutions that pass a set of unit tests in a code generation task. At the model level, the accuracy score is the average of the dataset-level accuracies for each model. |
4445

45-
**Prompt assisted metrics**
46+
47+
- Prompt assisted metrics
4648

4749
| Metric | Description |
4850
|--------|-------------|

0 commit comments

Comments
 (0)