@@ -129,14 +129,14 @@ pub fn build(b: *std.build.Builder) !void {
129129 const clip = make .obj ("clip" , "examples/llava/clip.cpp" );
130130 const llava = make .obj ("llava" , "examples/llava/llava.cpp" );
131131
132- _ = make .exe ("main" , "examples/main/main.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo , sampling , console , grammar_parser });
133- _ = make .exe ("quantize" , "examples/quantize/quantize.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo });
134- _ = make .exe ("perplexity" , "examples/perplexity/perplexity.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo });
135- _ = make .exe ("embedding" , "examples/embedding/embedding.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo });
136- _ = make .exe ("finetune" , "examples/finetune/finetune.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo , train });
132+ _ = make .exe ("main" , "examples/main/main.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo , console , grammar_parser });
133+ _ = make .exe ("quantize" , "examples/quantize/quantize.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo });
134+ _ = make .exe ("perplexity" , "examples/perplexity/perplexity.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo });
135+ _ = make .exe ("embedding" , "examples/embedding/embedding.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo });
136+ _ = make .exe ("finetune" , "examples/finetune/finetune.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo , train });
137137 _ = make .exe ("train-text-from-scratch" , "examples/train-text-from-scratch/train-text-from-scratch.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo , train });
138138
139- const server = make .exe ("server" , "examples/server/server.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo , sampling , grammar_parser , clip , llava });
139+ const server = make .exe ("server" , "examples/server/server.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo , grammar_parser , clip , llava });
140140 if (server .target .isWindows ()) {
141141 server .linkSystemLibrary ("ws2_32" );
142142 }
0 commit comments