Skip to content

Commit b8fe059

Browse files
authored
Update README (#27)
* update readme * update metrics info * fix type checks * add citations * update readme * add emojis * update readme * update readme
1 parent ddc5d76 commit b8fe059

File tree

1 file changed

+8
-10
lines changed

1 file changed

+8
-10
lines changed

README.md

Lines changed: 8 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -26,9 +26,9 @@
2626

2727
<h4 align="center">
2828
<p>
29-
<a href="#Installation">Installation</a> |
30-
<a href="#quickstart">Quick Example</a> |
31-
<a href="#metrics">Metrics List</a> |
29+
<a href="#beers-installation">Installation</a> |
30+
<a href="#beers-quickstart">Quickstart</a> |
31+
<a href="#beers-metrics">Metrics</a> |
3232
<a href="https://huggingface.co/explodinggradients">Hugging Face</a>
3333
<p>
3434
</h4>
@@ -37,7 +37,7 @@ ragas is a framework that helps you evaluate your Retrieval Augmented Generation
3737

3838
ragas provides you with the tools based on the latest research for evaluating LLM generated text to give you insights about your RAG pipeline. ragas can be integrated with your CI/CD to provide continuous check to ensure performance.
3939

40-
## 🛡 Installation
40+
## :beers: Installation
4141

4242
```bash
4343
pip install ragas
@@ -48,7 +48,7 @@ git clone https://github.com/explodinggradients/ragas && cd ragas
4848
pip install -e .
4949
```
5050

51-
## 🔥 Quickstart
51+
## :beers: Quickstart
5252

5353
This is a small example program you can run to see ragas in action!
5454
```python
@@ -75,13 +75,11 @@ e = Evaluation(
7575
results = e.eval(ds["ground_truth"], ds["generated_text"])
7676
print(results)
7777
```
78-
If you want a more in-depth explanation of core components, check out our [quick-start notebook](./examples/quickstart.ipynb)
79-
## 🧰 Metrics
78+
If you want a more in-depth explanation of core components, check out our quick-start notebook
79+
## :beers: Metrics
8080

8181
### ✏️ Character based
8282

83-
Character based metrics focus on analyzing text at the character level.
84-
8583
- **Levenshtein distance** the number of single character edits (additional, insertion, deletion) required to change your generated text to ground truth text.
8684
- **Levenshtein** **ratio** is obtained by dividing the Levenshtein distance by sum of number of characters in generated text and ground truth. This type of metrics is suitable where one works with short and precise texts.
8785

@@ -113,7 +111,7 @@ Model based methods uses language models combined with NLP techniques to compare
113111

114112
- **$Q^2$**
115113

116-
Best used to measure factual consistencies between ground truth and generated text. Scores can range from 0 to 1. Higher score indicates better factual consistency between ground truth and generated answer. Employs QA-QG paradigm followed by NLI to compare ground truth and generated answer. Q2Score score is highly correlated with human judgement.
114+
Best used to measure factual consistencies between ground truth and generated text. Scores can range from 0 to 1. Higher score indicates better factual consistency between ground truth and generated answer. Employs QA-QG paradigm followed by NLI to compare ground truth and generated answer. $Q^2$ score is highly correlated with human judgement.
117115

118116
📜 Checkout [citations](./citations.md) for related publications.
119117

0 commit comments

Comments
 (0)