You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* added info about data
* fix some errors
* fix emojis
* fix blue
* char
* quickstart
* download spacy if not found
* finish quickstart
* fix linting issues
* update badges
---------
Co-authored-by: Jithin James <[email protected]>
ragas is a framework that helps you evaluate your Retrieval Augmented Generation (RAG) pipelines. RAG denotes a class of LLM applications that use external data to augment the LLM’s context. There are existing tools and frameworks that help you build these pipelines but evaluating it and quantifying your pipeline performance can be hard.. This is were ragas (RAG Assessment) comes in
36
37
37
38
ragas provides you with the tools based on the latest research for evaluating LLM generated text to give you insights about your RAG pipeline. ragas can be integrated with your CI/CD to provide continuous check to ensure performance.
38
39
39
-
## Installation 🛡
40
+
## 🛡 Installation
40
41
41
42
```bash
42
43
pip install ragas
@@ -47,7 +48,7 @@ git clone https://github.com/explodinggradients/ragas && cd ragas
47
48
pip install -e .
48
49
```
49
50
50
-
## Quickstart 🔥
51
+
## 🔥 Quickstart
51
52
52
53
This is a small example program you can run to see ragas in action!
If you want a more in-depth explanation of core components, check out our quick-start notebook
78
+
If you want a more in-depth explanation of core components, check out our [quick-start notebook](./examples/quickstart.ipynb)
78
79
## 🧰 Metrics
79
80
80
81
### ✏️ Character based
81
82
83
+
Character based metrics focus on analyzing text at the character level.
84
+
82
85
-**Levenshtein distance** the number of single character edits (additional, insertion, deletion) required to change your generated text to ground truth text.
83
86
-**Levenshtein****ratio** is obtained by dividing the Levenshtein distance by sum of number of characters in generated text and ground truth. This type of metrics is suitable where one works with short and precise texts.
84
87
@@ -92,7 +95,7 @@ N-gram based metrics as name indicates uses n-grams for comparing generated answ
92
95
93
96
-**BLEU** (BiLingual Evaluation Understudy)
94
97
95
-
It measures precision by comparing clipped n-grams in generated text to ground truth text. These matches do not consider the ordering of words.
98
+
It measures precision by comparing clipped n-grams in generated text to ground truth text. These matches do not consider the ordering of words.
0 commit comments