Skip to content

Commit 5d86186

Browse files
committed
minor changes
1 parent 0e17e45 commit 5d86186

File tree

2 files changed

+14
-1
lines changed

2 files changed

+14
-1
lines changed

content/learning-paths/servers-and-cloud-computing/llama_cpp_streamline/Analyzing_token_generation_at_Prefill_and_Decode_stage.md

Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -54,6 +54,19 @@ To add Annotation Markers to llama-cli, change the llama-cli code *llama.cpp/too
5454
```
5555
and the Annotation Marker code in the 'main' function,
5656

57+
Firstly, add the Streamline Annotation setup code after *common_init*,
58+
```c
59+
common_init();
60+
61+
//Add the Annotation setup code
62+
ANNOTATE_SETUP;
63+
64+
```
65+
66+
67+
then add the Annotation Marker generation code here,
68+
69+
5770
```c
5871
for (int i = 0; i < (int) embd.size(); i += params.n_batch) {
5972
int n_eval = (int) embd.size() - i;

content/learning-paths/servers-and-cloud-computing/llama_cpp_streamline/_index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
---
2-
title: Use Streamline to analyze LLM running on CPU with llama.cpp
2+
title: Use Streamline to analyze LLM running on CPU with llama.cpp and KleidiAI
33

44
minutes_to_complete: 50
55

0 commit comments

Comments
 (0)