Skip to content

Commit 12d1360

Browse files
Update RAG documentation: refine titles and section headings for clarity and consistency across setup, pipeline, and memory observation guides.
1 parent e7f1752 commit 12d1360

File tree

5 files changed

+16
-16
lines changed

5 files changed

+16
-16
lines changed

content/learning-paths/laptops-and-desktops/dgx_spark_rag/1_rag.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
11
---
2-
title: Understanding RAG on Grace–Blackwell (GB10)
2+
title: Build a RAG pipeline on Arm-based Grace–Blackwell (GB10) systems
33
weight: 2
44

55
### FIXED, DO NOT MODIFY
66
layout: learningpathall
77
---
88

9-
## Before you start
9+
## Get started
1010

1111
Before starting this Learning Path, you should complete [Unlock quantized LLM performance on Arm-based NVIDIA DGX Spark](/learning-paths/laptops-and-desktops/dgx_spark_llamacpp/) to learn about the CPU and GPU builds of llama.cpp. This background is recommended for building the RAG solution on llama.cpp.
1212

content/learning-paths/laptops-and-desktops/dgx_spark_rag/2_rag_setup.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
---
2-
title: Configure your development environment and prepare models
2+
title: Configure the RAG development environment and models
33
weight: 3
44
layout: "learningpathall"
55
---

content/learning-paths/laptops-and-desktops/dgx_spark_rag/2b_rag_setup.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,16 +1,16 @@
11
---
2-
title: Add documents to the vector database
2+
title: Add documents to the RAG vector database
33
weight: 4
44
layout: "learningpathall"
55
---
66

7-
## Prepare a sample document corpus
7+
## Prepare a sample document corpus for RAG
88

99
You are now ready to add your documents to the RAG database that will be used for retrieval and reasoning.
1010

1111
This converts your raw knowledge documents into clean, chunked text segments that can later be vectorized and indexed by FAISS.
1212

13-
## Understanding FAISS for vector search
13+
## Use FAISS for efficient vector search on Arm
1414

1515
FAISS (Facebook AI Similarity Search) is an open-source library developed by Meta AI for efficient similarity search and clustering of dense vectors. It's particularly well-suited for RAG applications because it can quickly find the most relevant document chunks from large collections.
1616

@@ -21,7 +21,7 @@ Key advantages of FAISS for this application:
2121
- Speed: Uses advanced indexing algorithms to perform nearest-neighbor searches in milliseconds
2222
- Flexibility: Supports multiple distance metrics (L2, cosine similarity) and index types
2323

24-
### Create a workspace and data folder
24+
## Set up youe RAG workspace and data folder
2525

2626
Create a directory structure for your data:
2727

@@ -57,7 +57,7 @@ Use `wget` to batch download all the PDFs into `~/rag/pdf`.
5757
wget -P ~/rag/pdf -i datasheet.txt
5858
```
5959

60-
### Convert PDF into txt file
60+
## Convert PDF documents to text files
6161

6262
Then, create a Python file named `pdf2text.py` with the code below:
6363

@@ -109,7 +109,7 @@ At the end of the output you see:
109109
Total converted PDFs: 12
110110
```
111111

112-
### Verify your corpus
112+
## Verify your document corpus
113113

114114
You should now see a number of files in your folder. Run the command below to inspect the results:
115115

@@ -119,7 +119,7 @@ find ~/rag/text/ -type f -name "*.txt" -exec cat {} + | wc -l
119119

120120
It will show how many lines are in total. The number is around 100,000.
121121

122-
## Build an Embedding and Search Index
122+
## Build an embedding and search index with FAISS
123123

124124
Convert your prepared text corpus into vector embeddings and store them in a FAISS index for efficient semantic search.
125125

@@ -133,7 +133,7 @@ This stage enables your RAG pipeline to retrieve the most relevant text chunks w
133133

134134
Use e5-base-v2 to encode the documents and create a FAISS vector index.
135135

136-
### Create the FAISS builder script
136+
### Create and run the FAISS builder script
137137

138138

139139
```bash

content/learning-paths/laptops-and-desktops/dgx_spark_rag/3_rag_pipeline.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
---
2-
title: Implementing the RAG pipeline
2+
title: Build and run the RAG pipeline on Arm-based Grace–Blackwell (GB10) systems
33
weight: 5
44
layout: "learningpathall"
55
---
66

7-
## Integrating retrieval and generation
7+
## Integrate retrieval and generation on Arm
88

99
In the previous sections, you prepared the environment, validated the e5-base-v2 embedding model, and verified that the Llama 3.1 8B Instruct model runs successfully on the Grace–Blackwell (GB10) platform.
1010

content/learning-paths/laptops-and-desktops/dgx_spark_rag/4_rag_memory_observation.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
---
2-
title: Observe unified memory performance
2+
title: Monitor unified memory performance on Arm-based Grace–Blackwell systems
33
weight: 6
44
layout: "learningpathall"
55
---
@@ -25,7 +25,7 @@ Open two terminals on your GB10 system and use them as listed in the table below
2525
You should also have your original terminals open that you used to run the `llama-server` and the RAG queries in the previous section. You will run these again and use the two new terminals for observation.
2626

2727

28-
### Prepare for Unified Memory Observation Experiments
28+
### Prepare for unified memory observation
2929

3030
Ensure the RAG pipeline is stopped before starting the observation.
3131

@@ -200,7 +200,7 @@ The GPU executes compute kernels with GPU utilization at 96%, without reading fr
200200

201201
The `utilization.memory=0` and `memory.used=[N/A]` metrics are clear signs that data sharing, not data copying, is happening.
202202

203-
### Observe and interpret unified memory behavior
203+
### Interpret unified memory behavior
204204

205205
This experiment confirms the Grace–Blackwell Unified Memory architecture in action:
206206
- The CPU and GPU share the same address space.

0 commit comments

Comments
 (0)