Skip to content

Commit e498420

Browse files
authored
Fix typos (#180)
1 parent cd9a3ce commit e498420

File tree

2 files changed

+14
-14
lines changed

2 files changed

+14
-14
lines changed

docs/getstarted/evaluation.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ execution:
99
(heading-target)=
1010
# Evaluation
1111

12-
welcome to the ragas quickstart. We're going to get you up and running with ragas as qickly as you can so that you can go back to improving your Retrieval Augmented Generation pipelines while this library makes sure your changes are improving your entire pipeline.
12+
welcome to the ragas quickstart. We're going to get you up and running with ragas as quickly as you can so that you can go back to improving your Retrieval Augmented Generation pipelines while this library makes sure your changes are improving your entire pipeline.
1313

1414
to kick things of lets start with the data
1515

@@ -32,7 +32,7 @@ os.environ["OPENAI_API_KEY"] = "your-openai-key"
3232
Ragas performs a `ground_truth` free evaluation of your RAG pipelines. This is because for most people building a gold labeled dataset which represents in the distribution they get in production is a very expensive process.
3333

3434
```{note}
35-
While originially ragas was aimed at `ground_truth` free evalutions there is some aspects of the RAG pipeline that need `ground_truth` in order to measure. We're in the process of building a testset generation features that will make it easier. Checkout [issue#136](https://github.com/explodinggradients/ragas/issues/136) for more details.
35+
While originally ragas was aimed at `ground_truth` free evaluations there is some aspects of the RAG pipeline that need `ground_truth` in order to measure. We're in the process of building a testset generation features that will make it easier. Checkout [issue#136](https://github.com/explodinggradients/ragas/issues/136) for more details.
3636
```
3737

3838
Hence to work with ragas all you need are the following data
@@ -41,7 +41,7 @@ Hence to work with ragas all you need are the following data
4141
- contexts: `list[list[str]]` - The contexts which where passed into the LLM to answer the question.
4242
- ground_truths: `list[list[str]]` - The ground truth answer to the questions. (only required if you are using context_recall)
4343

44-
Ideally your list of questions should reflect the questions your users give, including those that you have been problamatic in the past.
44+
Ideally your list of questions should reflect the questions your users give, including those that you have been problematic in the past.
4545

4646
Here we're using an example dataset from on of the baselines we created for the [Financial Opinion Mining and Question Answering (fiqa) Dataset](https://sites.google.com/view/fiqa/) we created.
4747

@@ -80,9 +80,9 @@ from ragas.metrics import (
8080
```
8181
here you can see that we are using 4 metrics, but what do the represent?
8282

83-
1. faithfulness - the factual consistancy of the answer to the context base on the question.
83+
1. faithfulness - the factual consistency of the answer to the context base on the question.
8484
2. context_precision - a measure of how relevant the retrieved context is to the question. Conveys quality of the retrieval pipeline.
85-
3. answer_relevancy - a measure of how relevent the answer is to the question
85+
3. answer_relevancy - a measure of how relevant the answer is to the question
8686
4. context_recall: measures the ability of the retriever to retrieve all the necessary information needed to answer the question.
8787

8888

@@ -92,7 +92,7 @@ by default these metrics are using OpenAI's API to compute the score. If you usi
9292

9393
## Evaluation
9494

95-
Running the evalutation is as simple as calling evaluate on the `Dataset` with the metrics of your choice.
95+
Running the evaluation is as simple as calling evaluate on the `Dataset` with the metrics of your choice.
9696

9797
```{code-block} python
9898
:caption: evaluate using sample dataset

docs/getstarted/quickstart.ipynb

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
" <h1> Quickstart </h1>\n",
1616
"</p>\n",
1717
"\n",
18-
"welcome to the ragas quickstart. We're going to get you up and running with ragas as qickly as you can so that you can go back to improving your Retrieval Augmented Generation pipelines while this library makes sure your changes are improving your entire pipeline.\n",
18+
"Welcome to the ragas quickstart. We're going to get you up and running with ragas as quickly as you can so that you can go back to improving your Retrieval Augmented Generation pipelines while this library makes sure your changes are improving your entire pipeline.\n",
1919
"\n",
2020
"to kick things of lets start with the data\n",
2121
"\n",
@@ -62,17 +62,17 @@
6262
"\n",
6363
"Ragas performs a `ground_truth` free evaluation of your RAG pipelines. This is because for most people building a gold labeled dataset which represents in the distribution they get in production is a very expensive process.\n",
6464
"\n",
65-
"**Note:** *While originially ragas was aimed at `ground_truth` free evalutions there is some aspects of the RAG pipeline that need `ground_truth` in order to measure. We're in the process of building a testset generation features that will make it easier. Checkout [issue#136](https://github.com/explodinggradients/ragas/issues/136) for more details.*\n",
65+
"**Note:** *While originally ragas was aimed at `ground_truth` free evaluations there is some aspects of the RAG pipeline that need `ground_truth` in order to measure. We're in the process of building a testset generation features that will make it easier. Checkout [issue#136](https://github.com/explodinggradients/ragas/issues/136) for more details.*\n",
6666
"\n",
6767
"Hence to work with ragas all you need are the following data\n",
6868
"- question: `list[str]` - These are the questions you RAG pipeline will be evaluated on. \n",
6969
"- answer: `list[str]` - The answer generated from the RAG pipeline and give to the user.\n",
7070
"- contexts: `list[list[str]]` - The contexts which where passed into the LLM to answer the question.\n",
7171
"- ground_truths: `list[list[str]]` - The ground truth answer to the questions. (only required if you are using context_recall)\n",
7272
"\n",
73-
"Ideally your list of questions should reflect the questions your users give, including those that you have been problamatic in the past.\n",
73+
"Ideally your list of questions should reflect the questions your users give, including those that you have been problematic in the past.\n",
7474
"\n",
75-
"Here we're using an example dataset from on of the baselines we created for the [Financial Opinion Mining and Question Answering (fiqa) Dataset](https://sites.google.com/view/fiqa/) we created. If you want to want to know more about the baseline, feel free to check the `experiements/baseline` section"
75+
"Here we're using an example dataset from on of the baselines we created for the [Financial Opinion Mining and Question Answering (fiqa) Dataset](https://sites.google.com/view/fiqa/) we created. If you want to want to know more about the baseline, feel free to check the `experiments/baseline` section"
7676
]
7777
},
7878
{
@@ -167,10 +167,10 @@
167167
"here you can see that we are using 4 metrics, but what do the represent?\n",
168168
"\n",
169169
"1. context_precision - a measure of how relevant the retrieved context is to the question. Conveys quality of the retrieval pipeline.\n",
170-
"2. answer_relevancy - a measure of how relevent the answer is to the question\n",
171-
"3. faithfulness - the factual consistancy of the answer to the context base on the question.\n",
170+
"2. answer_relevancy - a measure of how relevant the answer is to the question\n",
171+
"3. faithfulness - the factual consistency of the answer to the context base on the question.\n",
172172
"4. context_recall: measures the ability of the retriever to retrieve all the necessary information needed to answer the question. \n",
173-
"5. harmfulness (AspectCritique) - in general, `AspectCritique` is a metric that can be used to quantify various aspects of the answer. Aspects like harmfulness, maliciousness, coherence, correctness, concisenes are available by default but you can easily define your own. Check the [docs](./metrics.md) for more info.\n",
173+
"5. harmfulness (AspectCritique) - in general, `AspectCritique` is a metric that can be used to quantify various aspects of the answer. Aspects like harmfulness, maliciousness, coherence, correctness, conciseness are available by default but you can easily define your own. Check the [docs](./metrics.md) for more info.\n",
174174
"\n",
175175
"**Note:** *by default these metrics are using OpenAI's API to compute the score. If you using this metric make sure you set the environment key `OPENAI_API_KEY` with your API key. You can also try other LLMs for evaluation, check the [llm guide](./guides/llms.ipynb) to learn more*\n",
176176
"\n",
@@ -184,7 +184,7 @@
184184
"source": [
185185
"## Evaluation\n",
186186
"\n",
187-
"Running the evalutation is as simple as calling evaluate on the `Dataset` with the metrics of your choice."
187+
"Running the evaluation is as simple as calling evaluate on the `Dataset` with the metrics of your choice."
188188
]
189189
},
190190
{

0 commit comments

Comments
 (0)