|
| 1 | +package llamacpp |
| 2 | + |
| 3 | +/////////////////////////////////////////////////////////////////////////////// |
| 4 | +// CGO |
| 5 | + |
| 6 | +/* |
| 7 | +#cgo pkg-config: llamacpp |
| 8 | +#include <llama.h> |
| 9 | +#include <stdlib.h> |
| 10 | +*/ |
| 11 | +import "C" |
| 12 | +import "unsafe" |
| 13 | + |
| 14 | +/////////////////////////////////////////////////////////////////////////////// |
| 15 | +// TYPES |
| 16 | + |
| 17 | +type ( |
| 18 | + Ggml_numa_strategy C.enum_ggml_numa_strategy |
| 19 | + Llama_model C.struct_llama_model |
| 20 | + Llama_model_params C.struct_llama_model_params |
| 21 | +) |
| 22 | + |
| 23 | +/////////////////////////////////////////////////////////////////////////////// |
| 24 | +// GLOBALS |
| 25 | + |
| 26 | +const ( |
| 27 | + GGML_NUMA_STRATEGY_DISABLED = C.GGML_NUMA_STRATEGY_DISABLED |
| 28 | + GGML_NUMA_STRATEGY_DISTRIBUTE = C.GGML_NUMA_STRATEGY_DISTRIBUTE |
| 29 | + GGML_NUMA_STRATEGY_ISOLATE = C.GGML_NUMA_STRATEGY_ISOLATE |
| 30 | + GGML_NUMA_STRATEGY_NUMACTL = C.GGML_NUMA_STRATEGY_NUMACTL |
| 31 | + GGML_NUMA_STRATEGY_MIRROR = C.GGML_NUMA_STRATEGY_MIRROR |
| 32 | +) |
| 33 | + |
| 34 | +/////////////////////////////////////////////////////////////////////////////// |
| 35 | +// LIFECYCLE |
| 36 | + |
| 37 | +// Initialize the llama + ggml backend. If numa is true, use NUMA optimizations |
| 38 | +// Call once at the start of the program |
| 39 | +func Llama_backend_init() { |
| 40 | + C.llama_backend_init() |
| 41 | +} |
| 42 | + |
| 43 | +func Llama_numa_init(numa Ggml_numa_strategy) { |
| 44 | + C.llama_numa_init(C.enum_ggml_numa_strategy(numa)) |
| 45 | +} |
| 46 | + |
| 47 | +// Call once at the end of the program - currently only used for MPI |
| 48 | +func Llama_backend_free() { |
| 49 | + C.llama_backend_free() |
| 50 | +} |
| 51 | + |
| 52 | +// Load a model from a file |
| 53 | +func Llama_load_model_from_file(path string, params Llama_model_params) *Llama_model { |
| 54 | + cPath := C.CString(path) |
| 55 | + defer C.free(unsafe.Pointer(cPath)) |
| 56 | + return (*Llama_model)(C.llama_load_model_from_file(cPath, C.struct_llama_model_params(params))) |
| 57 | +} |
| 58 | + |
| 59 | +// Free a model |
| 60 | +func Llama_free_model(model *Llama_model) { |
| 61 | + C.llama_free_model((*C.struct_llama_model)(model)) |
| 62 | +} |
0 commit comments