You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Now, just use `node-llama-cpp` as you normally would.
88
+
89
+
## Intel AMX {#intel-amx}
90
+
> Intel AMX (Advanced Matrix Extensions) is a dedicated hardware block found on Intel Xeon processors
91
+
> that helps optimize and accelerate matrix multiplication operations.
92
+
>
93
+
> It's available on the 4th Gen and newer Intel Xeon processors.
94
+
95
+
Intel AMX can improve CPU inference performance [by 2x and up to even 14x](https://github.com/ggerganov/llama.cpp/pull/7707) faster inference times on supported CPUs (on specific conditions).
96
+
97
+
If you're using a 4th Gen or newer Intel Xeon processor,
98
+
you might want to [build `llama.cpp` from source](./building-from-source.md) to utilize these hardware-specific optimizations available on your hardware.
99
+
100
+
To do this, run this command inside your project on the machine you run your project on:
101
+
```shell
102
+
npx --no node-llama-cpp source download
103
+
```
104
+
105
+
Alternatively, you can force `node-llama-cpp` to not use its prebuilt binaries
106
+
and instead build from source when calling [`getLlama`](../api/functions/getLlama.md) for the first time on a Xeon CPU:
0 commit comments