@@ -246,21 +246,21 @@ Instructions for adding support for new models: [HOWTO-add-model.md](docs/develo
246246| [ OpenCL] ( docs/backend/OPENCL.md ) | Adreno GPU |
247247
248248## Software architecture
249-
250249``` mermaid
251250block-beta
252251columns 1
253252
254253block:llamacpp
255- llamacpp["llama_cpp"]
256- style llamacpp fill:#3c3,color:#000,stroke:#000
254+ llamacpp["llama_cpp"]
255+ style llamacpp fill:#3c3,color:#000,stroke:#000
257256end
258257
259- block:ggml
260- ggml ["GGML"]
261- style ggml fill:#3c3,color:#000,stroke:#000
258+ block:ggml_backend
259+ ggml_backend ["GGML backend subsystem "]
260+ style ggml_backend fill:#3c3,color:#000,stroke:#000
262261
263- ggml_cpu["ggml-cpu"]
262+ block:ggmlbackends
263+ ggml_cpu["ggml-cpu"]
264264 ggml_metal["ggml-metal"]
265265 ggml_sycl["ggml-sycl"]
266266 ggml_cuda["ggml-cuda"]
@@ -272,25 +272,43 @@ block:ggml
272272 ggml_nnpa["ggml-nnpa"]
273273 ggml_ane["ggml-ane"]
274274
275- style ggml_cpu fill:#888,color:#000,stroke:#000
276- style ggml_metal fill:#888,color:#000,stroke:#000
277- style ggml_sycl fill:#888,color:#000,stroke:#000
278- style ggml_cuda fill:#888,color:#000,stroke:#000
279- style ggml_hip fill:#888,color:#000,stroke:#000
280- style ggml_vulkan fill:#888,color:#000,stroke:#000
281- style ggml_cann fill:#888,color:#000,stroke:#000
275+ style ggml_cpu fill:#888,color:#000,stroke:#000
276+ style ggml_metal fill:#888,color:#000,stroke:#000
277+ style ggml_sycl fill:#888,color:#000,stroke:#000
278+ style ggml_cuda fill:#888,color:#000,stroke:#000
279+ style ggml_hip fill:#888,color:#000,stroke:#000
280+ style ggml_vulkan fill:#888,color:#000,stroke:#000
281+ style ggml_cann fill:#888,color:#000,stroke:#000
282282
283283 style ggml_opencl fill:#cc3,color:#000,stroke:#000
284284 style ggml_qnn fill:#cc3,color:#000,stroke:#000
285285 style ggml_ane fill:#fff,color:#000,stroke:#f00,stroke-width:2,stroke-dasharray:5
286286 style ggml_nnpa fill:#cc3,color:#000,stroke:#000
287+ end
287288end
288289
289290block:ggml_pal
290- ggml_pal["GGML Platform Abstraction Layer"]
291- style ggml_pal fill:#c33,color:#000,stroke:#000
291+ ggml_pal["Platform Abstraction Layer"]
292+ style ggml_pal fill:#c33,color:#000,stroke:#000
292293end
293294
295+ block:ggml_backendsubsystem
296+ ggml_backendsubsystem["GGML backend subsystem"]
297+ style ggml_backendsubsystem fill:#3c3,color:#000,stroke:#000
298+ end
299+
300+ block:group1:2
301+ columns 2
302+ block:ggml_tensor
303+ ggml_tensor["GGML tensor"]
304+ style ggml_tensor fill:#3c3,color:#000,stroke:#000
305+ end
306+
307+ block:ggml_cgraph
308+ ggml_cgraph["GGML cgraph"]
309+ style ggml_cgraph fill:#3c3,color:#000,stroke:#000
310+ end
311+ end
294312
295313block:OS
296314 Windows
@@ -345,11 +363,12 @@ flowchart LR
345363 classDef TODO fill:#c33,color:#000,stroke:#000
346364 classDef NEW fill:#fff,color:#000,stroke:#f00,stroke-width:2,stroke-dasharray:5
347365 subgraph Legend
348- direction LR
366+ direction LR
349367 EXIST:::EXIST ~~~ TODO:::TODO ~~~ WIP:::WIP ~~~ DONE:::DONE ~~~ NEW:::NEW
350368 end
351369```
352370
371+
353372## Building the project
354373
355374The main product of this project is the ` llama ` library. Its C-style interface can be found in [ include/llama.h] ( include/llama.h ) .
0 commit comments