|
2 | 2 | "name": "llama-vscode", |
3 | 3 | "displayName": "llama-vscode", |
4 | 4 | "description": "Local LLM-assisted text completion using llama.cpp", |
5 | | - "version": "0.0.9-beta-3", |
| 5 | + "version": "0.0.10-beta-1", |
6 | 6 | "publisher": "ggml-org", |
7 | 7 | "repository": "https://github.com/ggml-org/llama.vscode", |
8 | 8 | "engines": { |
|
31 | 31 | "commands": [ |
32 | 32 | { |
33 | 33 | "command": "extension.triggerInlineCompletion", |
34 | | - "title": "Trigger Inline Completion" |
| 34 | + "title": "llama-vscode: Trigger Inline Completion" |
35 | 35 | }, |
36 | 36 | { |
37 | 37 | "command": "extension.triggerNoCacheCompletion", |
38 | | - "title": "Trigger No Cache Completion" |
| 38 | + "title": "llama-vscode: Trigger No Cache Completion" |
39 | 39 | }, |
40 | 40 | { |
41 | 41 | "command": "extension.copyIntercept", |
42 | | - "title": "Copy Intercept" |
| 42 | + "title": "llama-vscode: Copy Intercept" |
43 | 43 | }, |
44 | 44 | { |
45 | 45 | "command": "extension.cutIntercept", |
46 | | - "title": "Cut Intercept" |
| 46 | + "title": "llama-vscode: Cut Intercept" |
47 | 47 | }, |
48 | 48 | { |
49 | 49 | "command": "extension.acceptFirstLine", |
50 | | - "title": "Accept First Line" |
| 50 | + "title": "llama-vscode: Accept First Line" |
51 | 51 | }, |
52 | 52 | { |
53 | 53 | "command": "extension.acceptFirstWord", |
54 | | - "title": "Accept First Word" |
| 54 | + "title": "llama-vscode: Accept First Word" |
55 | 55 | }, |
56 | 56 | { |
57 | 57 | "command": "extension.copyChunks", |
58 | | - "title": "Copy Chunks" |
| 58 | + "title": "llama-vscode: Copy Chunks" |
59 | 59 | }, |
60 | 60 | { |
61 | 61 | "command": "extension.showMenu", |
62 | | - "title": "Show Menu" |
| 62 | + "title": "llama-vscode: Show Menu" |
63 | 63 | }, |
64 | 64 | { |
65 | 65 | "command": "extension.askAi", |
66 | | - "title": "Ask AI" |
| 66 | + "title": "llama-vscode: Ask AI" |
67 | 67 | }, |
68 | 68 | { |
69 | 69 | "command": "extension.askAiWithContext", |
70 | | - "title": "Ask AI With Context" |
| 70 | + "title": "llama-vscode: Ask AI With Context" |
71 | 71 | }, |
72 | 72 | { |
73 | 73 | "command": "extension.editSelectedText", |
74 | | - "title": "Edit Selected Text with AI" |
| 74 | + "title": "llama-vscode: Edit Selected Text with AI" |
75 | 75 | }, |
76 | 76 | { |
77 | 77 | "command": "extension.acceptTextEdit", |
78 | | - "title": "Accept Text Edit Suggestion" |
| 78 | + "title": "llama-vscode: Accept Text Edit Suggestion" |
79 | 79 | }, |
80 | 80 | { |
81 | 81 | "command": "extension.rejectTextEdit", |
82 | | - "title": "Reject Text Edit Suggestion" |
| 82 | + "title": "llama-vscode: Reject Text Edit Suggestion" |
83 | 83 | } |
84 | 84 | ], |
85 | 85 | "keybindings": [ |
|
122 | 122 | "command": "extension.acceptFirstWord", |
123 | 123 | "key": "ctrl+right", |
124 | 124 | "when": "editorTextFocus && inlineSuggestionVisible" |
125 | | - } |
126 | | - , |
| 125 | + }, |
127 | 126 | { |
128 | 127 | "command": "extension.showMenu", |
129 | 128 | "key": "ctrl+shift+m", |
|
169 | 168 | "default": "cd c:/ai ; ./llama-server.exe -m qwen2.5-coder-3b-instruct-q6_k.gguf -ngl 99 --port 8011 --path C:/llama.cpp/llama.cpp/examples/server/webui/dist", |
170 | 169 | "description": "Shell command for starting chat llama.cpp server, executed from the menu" |
171 | 170 | }, |
| 171 | + "llama-vscode.launch_embeddings": { |
| 172 | + "type": "string", |
| 173 | + "default": "cd c:/ai ; ./llama-server.exe -m all-MiniLM-L6-v2-Q8_0.gguf --port 8010", |
| 174 | + "description": "Shell command for starting chat llama.cpp server, executed from the menu" |
| 175 | + }, |
172 | 176 | "llama-vscode.launch_training_completion": { |
173 | 177 | "type": "string", |
174 | 178 | "default": "", |
|
199 | 203 | "default": "http://127.0.0.1:8011", |
200 | 204 | "description": "The URL to be used by the extension for chat with ai." |
201 | 205 | }, |
| 206 | + "llama-vscode.endpoint_embeddings": { |
| 207 | + "type": "string", |
| 208 | + "default": "http://127.0.0.1:8010", |
| 209 | + "description": "The URL to be used by the extension for creating embeddings." |
| 210 | + }, |
202 | 211 | "llama-vscode.auto": { |
203 | 212 | "type": "boolean", |
204 | 213 | "default": true, |
|
274 | 283 | "default": 1000, |
275 | 284 | "description": "how often to process queued chunks in normal mode" |
276 | 285 | }, |
| 286 | + "llama-vscode.rag_chunk_max_chars": { |
| 287 | + "type": "number", |
| 288 | + "default": 2000, |
| 289 | + "description": "Max number of chars per RAG chunk" |
| 290 | + }, |
| 291 | + "llama-vscode.rag_max_lines_per_chunk": { |
| 292 | + "type": "number", |
| 293 | + "default": 60, |
| 294 | + "description": "Max number of lines per RAG chunk" |
| 295 | + }, |
| 296 | + "llama-vscode.rag_max_chars_per_chunk_line": { |
| 297 | + "type": "number", |
| 298 | + "default": 300, |
| 299 | + "description": "max chars for a chunk line, the rest of the line is cut" |
| 300 | + }, |
| 301 | + "llama-vscode.rag_max_chunks": { |
| 302 | + "type": "number", |
| 303 | + "default": 30000, |
| 304 | + "description": "max cunks for the RAG search" |
| 305 | + }, |
| 306 | + "llama-vscode.rag_max_bm25_filter_chunks": { |
| 307 | + "type": "number", |
| 308 | + "default": 47, |
| 309 | + "description": "max RAG chunks to filter with BM25 algorithm" |
| 310 | + }, |
| 311 | + "llama-vscode.rag_max_embedding_filter_chunks": { |
| 312 | + "type": "number", |
| 313 | + "default": 5, |
| 314 | + "description": "max RAG chunks to provide as context to the LLM" |
| 315 | + }, |
| 316 | + "llama-vscode.rag_max_context_files": { |
| 317 | + "type": "number", |
| 318 | + "default": 3, |
| 319 | + "description": "max number of complete files to send as context to the LLM" |
| 320 | + }, |
| 321 | + "llama-vscode.rag_max_context_file_chars": { |
| 322 | + "type": "number", |
| 323 | + "default": 5000, |
| 324 | + "description": "max chars for a context file. If the file is bigger it will be cut to avoid too big context." |
| 325 | + }, |
277 | 326 | "llama-vscode.language": { |
278 | 327 | "type": "string", |
279 | 328 | "default": "en", |
|
285 | 334 | "description": "Enable/disable completions" |
286 | 335 | }, |
287 | 336 | "llama-vscode.languageSettings": { |
288 | | - "type": "object", |
289 | | - "default": { |
290 | | - "*": true |
291 | | - }, |
292 | | - "additionalProperties": { |
293 | | - "type": "boolean" |
294 | | - }, |
295 | | - "description": "Enable/disable suggestions for specific languages" |
| 337 | + "type": "object", |
| 338 | + "default": { |
| 339 | + "*": true |
| 340 | + }, |
| 341 | + "additionalProperties": { |
| 342 | + "type": "boolean" |
| 343 | + }, |
| 344 | + "description": "Enable/disable suggestions for specific languages" |
296 | 345 | }, |
297 | 346 | "llama-vscode.use_openai_endpoint": { |
298 | 347 | "type": "boolean", |
|
326 | 375 | }, |
327 | 376 | "dependencies": { |
328 | 377 | "axios": "^1.1.2", |
| 378 | + "ignore": "^7.0.4", |
329 | 379 | "openai": "^4.80.1" |
330 | 380 | }, |
331 | 381 | "devDependencies": { |
|
0 commit comments