@@ -108,23 +108,18 @@ function getQuantTag(filepath?: string): string {
108108const snippetLlamacpp = ( model : ModelData , filepath ?: string ) : LocalAppSnippet [ ] => {
109109 const command = ( binary : string ) => {
110110 const snippet = [ "# Load and run the model:" , `${ binary } -hf ${ model . id } ${ getQuantTag ( filepath ) } ` ] ;
111- if ( ! model . tags . includes ( "conversational" ) ) {
112- // for non-conversational models, add a prompt
113- snippet [ snippet . length - 1 ] += " \\" ;
114- snippet . push ( ' -p "Once upon a time,"' ) ;
115- }
116111 return snippet . join ( "\n" ) ;
117112 } ;
118113 return [
119114 {
120115 title : "Install from brew" ,
121116 setup : "brew install llama.cpp" ,
122- content : command ( "llama-cli " ) ,
117+ content : command ( "llama-server " ) ,
123118 } ,
124119 {
125120 title : "Install from WinGet (Windows)" ,
126121 setup : "winget install llama.cpp" ,
127- content : command ( "llama-cli " ) ,
122+ content : command ( "llama-server " ) ,
128123 } ,
129124 {
130125 title : "Use pre-built binary" ,
@@ -133,17 +128,17 @@ const snippetLlamacpp = (model: ModelData, filepath?: string): LocalAppSnippet[]
133128 "# Download pre-built binary from:" ,
134129 "# https://github.com/ggerganov/llama.cpp/releases" ,
135130 ] . join ( "\n" ) ,
136- content : command ( "./llama-cli " ) ,
131+ content : command ( "./llama-server " ) ,
137132 } ,
138133 {
139134 title : "Build from source code" ,
140135 setup : [
141136 "git clone https://github.com/ggerganov/llama.cpp.git" ,
142137 "cd llama.cpp" ,
143- "cmake -B build -DLLAMA_CURL=ON " ,
144- "cmake --build build -j --target llama-cli " ,
138+ "cmake -B build" ,
139+ "cmake --build build -j --target llama-server " ,
145140 ] . join ( "\n" ) ,
146- content : command ( "./build/bin/llama-cli " ) ,
141+ content : command ( "./build/bin/llama-server " ) ,
147142 } ,
148143 ] ;
149144} ;
0 commit comments