Skip to content

Commit 0f6046b

Browse files
committed
Update Blog “llm-agentic-tool-mesh-empowering-gen-ai-with-retrieval-augmented-generation-rag”
1 parent de424b8 commit 0f6046b

File tree

1 file changed

+13
-13
lines changed

1 file changed

+13
-13
lines changed

content/blog/llm-agentic-tool-mesh-empowering-gen-ai-with-retrieval-augmented-generation-rag.md

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -29,22 +29,22 @@ RAG is a technique that enhances the capabilities of language models by providin
2929

3030
This approach improves the accuracy and relevance of generated responses, especially in domains requiring up-to-date or specialized information. The key benefits of RAG are:
3131

32-
* **Enhanced accuracy**: By accessing external data, models can provide more precise and factual responses.
33-
* **Domain specialization**: Enables models to handle specialized topics by leveraging domain-specific documents.
34-
* **Reduced hallucinations**: Minimizes the generation of incorrect or nonsensical information by grounding responses in real data.
32+
* **Enhanced accuracy**: Provides more precise and factual responses by accessing more data available externally
33+
* **Domain specialization**: Enables models to handle specialized topics by leveraging domain-specific documents
34+
* **Reduced hallucinations**: Minimizes the generation of incorrect or nonsensical information by grounding responses in real data
3535

3636
# RAG in LLM Agentic Tool Mesh
3737

38-
In the LLM Agentic Tool Mesh platform, RAG is a crucial process that enhances language models' capabilities by integrating external knowledge. LLM Agentic Tool implements RAG through two main stages:
38+
In the LLM Agentic Tool Mesh platform, RAG is a **crucial tool** that **can enhance** **a language model's** capabilities by integrating external knowledge. LLM Agentic Tool implements RAG through two main stages:
3939

40-
* **Injection**.
41-
* **Retrieval**.
40+
* **Injection**
41+
* **Retrieval**
4242

4343
Each stage is designed to standardize and optimize data use, ensuring generated content is both relevant and accurate.
4444

4545
## The injection process
4646

47-
The injection process involves preparing and integrating data into a storage system where it can be efficiently retrieved during the generation process.
47+
The injection process involves preparing and integrating data into a storage system where it can be efficiently retrieved when content is being generated.
4848

4949
This process is abstracted into three key steps:
5050

@@ -144,7 +144,7 @@ else:
144144

145145
### Loading
146146

147-
As part of the loading phase, the process includes injecting the transformed data into the chosen storage solution, such as a vector database, and further adapting the data as needed, such as by chunking it into smaller pieces for efficient retrieval.
147+
As part of the loading phase, the process includes injecting the transformed data into the chosen storage solution, such as a vector database, and further adapting the data as needed, such as chunking it into smaller pieces for efficient retrieval.
148148

149149
Example usage
150150

@@ -182,9 +182,9 @@ else:
182182

183183
Once the data has been injected and is ready for use, the retrieval process focuses on fetching the most relevant information based on a given input query. This ensures that the language model has access to the right data to generate accurate and contextually relevant outputs.
184184

185-
1. **Data retrieval**: Use various methods, such as dense or sparse retrieval, to fetch the most relevant data from storage.
186-
2. **Metadata filtering**: Apply metadata filters to narrow down search results, ensuring the retrieved data matches the specific needs of the query.
187-
3. **Chunk Expansion**: Expand the retrieved data chunks to provide comprehensive information for the language model.
185+
1. **Data retrieval**: Uses various methods, such as dense or sparse retrieval, to fetch the most relevant data from storage
186+
2. **Metadata filtering**: Applies metadata filters to narrow down search results, ensuring the retrieved data matches the specific needs of the query
187+
3. **Chunk Expansion**: Expands the retrieved data chunks to provide comprehensive information for the language model
188188

189189
![](/img/retrieve.png)
190190

@@ -224,7 +224,7 @@ else:
224224

225225
# LLM Agentic Tool Mesh in action: Agentic tool using RAG
226226

227-
In the [LLM Agentic Tool Mesh GitHub](https://github.com/HewlettPackard/llmesh), there is an example of a RAG-based tool that provides quick and accurate access to 5G specifications: the **telco expert**.
227+
In the [LLM Agentic Tool Mesh GitHub](https://github.com/HewlettPackard/llmesh), there is an example of a RAG-based tool that provides quick and accurate access to 5G specifications: the **telco expert (inside folder `examples/tool_rag`)**.
228228

229229
This agentic tool leverages the RAG services in LLM Agentic Tool Mesh to read telco standards, build or use a vector store from them, and then uses a query engine to find and return relevant information based on user queries.
230230

@@ -251,7 +251,7 @@ def telco_expert(query: str) -> str:
251251
return summary_answer + "\n\n" + chunk_answer
252252
```
253253

254-
The functionalities showed are:
254+
The functionalities shown are:
255255

256256
* **Data injection**: Loads telco standards into a vector store.
257257
* **Query augmentation**: Enhances the user's query for better retrieval.

0 commit comments

Comments
 (0)