We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent f27d39d commit a50be90Copy full SHA for a50be90
interpreter/terminal_interface/profiles/defaults/cortex-llama32.py
@@ -0,0 +1,21 @@
1
+"""
2
+This is an Open Interpreter profile for using Llama 3.2:3b served locally by Cortex.
3
+
4
+This profile configures Open Interpreter to use a locally hosted Llama 3.2 model through Cortex.
5
6
7
8
+from interpreter import interpreter
9
10
11
+# Update the model to match t
12
+interpreter.llm.model = "llama3.2:3b-gguf-q8-0"
13
+interpreter.llm.context_window = 8192
14
+interpreter.llm.max_tokens = 4096
15
+interpreter.llm.api_base = "http://127.0.0.1:39281/v1"
16
+interpreter.llm.supports_functions = False
17
+interpreter.llm.supports_vision = False
18
19
+interpreter.offline = True
20
+interpreter.loop = True
21
+interpreter.auto_run = False
0 commit comments