Skip to content

Commit c9889e6

Browse files
authored
Fix formatting errors (#3979)
https://docs.zenml.io/user-guides/llmops-guide/finetuning-embeddings you can see that formatting is off.
1 parent cb1fb7e commit c9889e6

File tree

1 file changed

+5
-31
lines changed
  • docs/book/user-guide/llmops-guide/finetuning-embeddings

1 file changed

+5
-31
lines changed

docs/book/user-guide/llmops-guide/finetuning-embeddings/README.md

Lines changed: 5 additions & 31 deletions
Original file line numberDiff line numberDiff line change
@@ -2,42 +2,16 @@
22
description: Finetune embeddings on custom synthetic data to improve retrieval performance.
33
---
44

5-
We previously learned [how to use RAG with ZenML](../rag-with-zenml/README.md) to
6-
build a production-ready RAG pipeline. In this section, we will explore how to
7-
optimize and maintain your embedding models through synthetic data generation and
8-
human feedback. So far, we've been using off-the-shelf embeddings, which provide
9-
a good baseline and decent performance on standard tasks. However, you can often
10-
significantly improve performance by finetuning embeddings on your own domain-specific data.
5+
We previously learned [how to use RAG with ZenML](../rag-with-zenml/README.md) to build a production-ready RAG pipeline. In this section, we will explore how to optimize and maintain your embedding models through synthetic data generation and human feedback. So far, we've been using off-the-shelf embeddings, which provide a good baseline and decent performance on standard tasks. However, you can often significantly improve performance by finetuning embeddings on your own domain-specific data.
116

12-
Our RAG pipeline uses a retrieval-based approach, where it first retrieves the
13-
most relevant documents from our vector database, and then uses a language model
14-
to generate a response based on those documents. By finetuning our embeddings on
15-
a dataset of technical documentation similar to our target domain, we can improve
16-
the retrieval step and overall performance of the RAG pipeline.
7+
Our RAG pipeline uses a retrieval-based approach, where it first retrieves the most relevant documents from our vector database, and then uses a language model to generate a response based on those documents. By finetuning our embeddings on a dataset of technical documentation similar to our target domain, we can improve the retrieval step and overall performance of the RAG pipeline.
178

18-
The work of finetuning embeddings based on synthetic data and human feedback is
19-
a multi-step process. We'll go through the following steps:
9+
The work of finetuning embeddings based on synthetic data and human feedback is a multi-step process. We'll go through the following steps:
2010

2111
- [generating synthetic data with `distilabel`](synthetic-data-generation.md)
2212
- [finetuning embeddings with Sentence Transformers](finetuning-embeddings-with-sentence-transformers.md)
2313
- [evaluating finetuned embeddings and using ZenML's model control plane to get a systematic overview](evaluating-finetuned-embeddings.md)
2414

25-
Besides ZenML, we will do this by using two open source libraries:
26-
[`argilla`](https://github.com/argilla-io/argilla/) and
27-
[`distilabel`](https://github.com/argilla-io/distilabel). Both of these
28-
libraries focus optimizing model outputs through improving data quality,
29-
however, each one of them takes a different approach to tackle the same problem.
30-
`distilabel` provides a scalable and reliable approach to distilling knowledge
31-
from LLMs by generating synthetic data or providing AI feedback with LLMs as
32-
judges. `argilla` enables AI engineers and domain experts to collaborate on data
33-
projects by allowing them to organize and explore data through within an
34-
interactive and engaging UI. Both libraries can be used individually but they
35-
work better together. We'll showcase their use via ZenML pipelines.
15+
Besides ZenML, we will do this by using two open source libraries: [`argilla`](https://github.com/argilla-io/argilla/) and [`distilabel`](https://github.com/argilla-io/distilabel). Both of these libraries focus optimizing model outputs through improving data quality, however, each one of them takes a different approach to tackle the same problem. `distilabel` provides a scalable and reliable approach to distilling knowledge from LLMs by generating synthetic data or providing AI feedback with LLMs as judges. `argilla` enables AI engineers and domain experts to collaborate on data projects by allowing them to organize and explore data through within an interactive and engaging UI. Both libraries can be used individually but they work better together. We'll showcase their use via ZenML pipelines.
3616

37-
To follow along with the example explained in this guide, please follow the
38-
instructions in [the `llm-complete-guide` repository](https://github.com/zenml-io/zenml-projects/tree/main/llm-complete-guide) where the full code is also
39-
available. This specific section on embeddings finetuning can be run locally or
40-
using cloud compute as you prefer.
41-
42-
<!-- For scarf -->
43-
<figure><img alt="ZenML Scarf" referrerpolicy="no-referrer-when-downgrade" src="https://static.scarf.sh/a.png?x-pxid=f0b4f458-0a54-4fcd-aa95-d5ee424815bc" /></figure>
17+
To follow along with the example explained in this guide, please follow the instructions in [the `llm-complete-guide` repository](https://github.com/zenml-io/zenml-projects/tree/main/llm-complete-guide) where the full code is also available. This specific section on embeddings finetuning can be run locally or using cloud compute as you prefer.

0 commit comments

Comments
 (0)