You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: content/learning-paths/laptops-and-desktops/dgx_spark_rag/1_rag.md
+7-8Lines changed: 7 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,5 @@
1
1
---
2
-
title: Build a RAG pipeline on Arm-based Grace–Blackwell systems
2
+
title: Explore building a RAG pipeline on Arm-based Grace–Blackwell systems
3
3
weight: 2
4
4
5
5
### FIXED, DO NOT MODIFY
@@ -8,14 +8,13 @@ layout: learningpathall
8
8
9
9
## Get started
10
10
11
-
Before starting this Learning Path, you should complete [Unlock quantized LLM performance on Arm-based NVIDIA DGX Spark](/learning-paths/laptops-and-desktops/dgx_spark_llamacpp/) to learn about the CPU and GPU builds of llama.cpp. This background is recommended for building the RAG solution on llama.cpp.
11
+
Before getting started, you should complete the Learning Path[Unlock quantized LLM performance on Arm-based NVIDIA DGX Spark](/learning-paths/laptops-and-desktops/dgx_spark_llamacpp/) to learn about the CPU and GPU builds of llama.cpp. This background is recommended for building the RAG solution on llama.cpp.
12
12
13
13
The NVIDIA DGX Spark is also referred to as the Grace-Blackwell platform or GB10, the name of the NVIDIA Grace-Blackwell Superchip.
14
14
15
15
## What is RAG?
16
16
17
-
Retrieval-Augmented Generation (RAG) combines information retrieval with language-model generation.
18
-
Instead of relying solely on pre-trained weights, a RAG system retrieves relevant text from a document corpus and passes it to a language model to create factual, context-aware responses.
17
+
Retrieval-Augmented Generation (RAG) combines information retrieval with language-model generation. Instead of relying solely on pre-trained weights, a RAG system retrieves relevant text from a document corpus and passes it to a language model to create factual, context-aware responses.
19
18
20
19
Here is a typical pipeline:
21
20
@@ -35,9 +34,9 @@ Its unique CPU–GPU design and unified memory enable seamless data exchange, ma
35
34
36
35
The GB10 platform includes:
37
36
38
-
- Grace CPU (Armv9.2 architecture) – 20 cores including 10 Cortex-X925 cores and 10 Cortex-A725 cores
39
-
- Blackwell GPU – CUDA 13.0 Tensor Core architecture
40
-
- Unified Memory (128 GB NVLink-C2C) – Shared address space between CPU and GPU which allows both processors to access the same 128 GB unified memory region without copy operations.
37
+
- Grace CPU (Armv9.2 architecture) - 20 cores including 10 Cortex-X925 cores and 10 Cortex-A725 cores
38
+
- Blackwell GPU - CUDA 13.0 Tensor Core architecture
39
+
- Unified Memory (128 GB NVLink-C2C) - Shared address space between CPU and GPU which allows both processors to access the same 128 GB unified memory region without copy operations.
41
40
42
41
The GB10 provides the following benefits for RAG applications:
43
42
@@ -102,7 +101,7 @@ The technology stack you will use is listed below:
102
101
| Unified Memory Architecture | Unified LPDDR5X shared memory | Grace CPU and Blackwell GPU | Enables zero-copy data sharing between CPU and GPU for improved latency and efficiency. |
103
102
104
103
105
-
## Prerequisites Check
104
+
## Check your setup
106
105
107
106
Before starting, run the following commands to confirm your hardware is ready:
Copy file name to clipboardExpand all lines: content/learning-paths/laptops-and-desktops/dgx_spark_rag/_index.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,7 +2,7 @@
2
2
title: Build a RAG pipeline on Arm-based NVIDIA DGX Spark
3
3
minutes_to_complete: 60
4
4
5
-
who_is_this_for: This is an advanced topic for developers who want to understand and implement a Retrieval-Augmented Generation (RAG) pipeline on the NVIDIA DGX Spark platform. It is ideal for those interested in exploring how Arm-based Grace CPUs manage local document retrieval and orchestration, while Blackwell GPUs accelerate large language model inference through the open-source llama.cpp REST server.
5
+
who_is_this_for: This is an advanced topic for developers who want to build a Retrieval-Augmented Generation (RAG) pipeline on the NVIDIA DGX Spark platform. You'll learn how Arm-based Grace CPUs handle document retrieval and orchestration, while Blackwell GPUs speed up large language model inference using the open-source llama.cpp REST server. This is a great fit if you're interested in combining Arm CPU management with GPU-accelerated AI workloads.
6
6
7
7
learning_objectives:
8
8
- Describe how a RAG system combines document retrieval and language model generation
@@ -11,7 +11,7 @@ learning_objectives:
11
11
- Build a reproducible RAG application that demonstrates efficient hybrid computing
12
12
13
13
prerequisites:
14
-
- An NVIDIA DGX Spark system with at least 15 GB of available disk space.
14
+
- An NVIDIA DGX Spark system with at least 15 GB of available disk space
0 commit comments