Skip to content

Commit ce3bbcb

Browse files
Starting content review.
1 parent 4b5469e commit ce3bbcb

File tree

2 files changed

+10
-4
lines changed

2 files changed

+10
-4
lines changed

content/learning-paths/servers-and-cloud-computing/deepseek-cpu/_index.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,12 +10,12 @@ minutes_to_complete: 30
1010
who_is_this_for: This is an introductory topic for developers interested in running DeepSeek-R1 on Arm-based servers.
1111

1212
learning_objectives:
13-
- Download and build llama.cpp on your Arm server.
13+
- Download and build llama.cpp on your Arm-based server.
1414
- Download a pre-quantized DeepSeek-R1 model from Hugging Face.
1515
- Run the pre-quantized model on your Arm CPU and measure the performance.
1616

1717
prerequisites:
18-
- An [Arm based instance](/learning-paths/servers-and-cloud-computing/csp/) from a cloud service provider or an on-premise Arm server. This Learning Path was tested on an AWS Graviton4 r8g.24xlarge instance.
18+
- An [Arm-based instance](/learning-paths/servers-and-cloud-computing/csp/) from a cloud service provider or an on-premise Arm server. This Learning Path was tested on an AWS Graviton4 r8g.24xlarge instance.
1919

2020
author:
2121
- Tianyu Li

content/learning-paths/servers-and-cloud-computing/deepseek-cpu/deepseek-chatbot.md

Lines changed: 8 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,9 +10,15 @@ layout: learningpathall
1010
The instructions in this Learning Path are for any Arm server running Ubuntu 24.04 LTS. You need an Arm server instance with at least 64 cores and 512GB of RAM to run this example. Configure disk storage up to at least 400 GB. The instructions have been tested on an AWS Graviton4 r8g.24xlarge instance.
1111

1212

13-
## Overview
13+
## Background and what you'll build
14+
15+
Arm CPUs are widely used in ML and AI use cases. In this Learning Path, you will learn how to run a generative AI inference-based use case of a LLM chatbot on Arm-based CPUs by deploying the [DeepSeek-R1 671B LLM](https://huggingface.co/bartowski/DeepSeek-R1-GGUF) on your Arm-based CPU using `llama.cpp`, optimized for Arm hardware. You'll:
16+
17+
- Build and run `llama.cpp` with Arm-specific performance improvements.
18+
- Download a quantized GGUF model from Hugging Face.
19+
- Run and measure performance on a large Arm instance (e.g., AWS Graviton4).
20+
1421

15-
Arm CPUs are widely used in traditional ML and AI use cases. In this Learning Path, you learn how to run generative AI inference-based use cases like a LLM chatbot on Arm-based CPUs. You do this by deploying the [DeepSeek-R1 GGUF models](https://huggingface.co/bartowski/DeepSeek-R1-GGUF) on your Arm-based CPU using `llama.cpp`.
1622

1723
[llama.cpp](https://github.com/ggerganov/llama.cpp) is an open source C/C++ project developed by Georgi Gerganov that enables efficient LLM inference on a variety of hardware - both locally, and in the cloud.
1824

0 commit comments

Comments
 (0)