@@ -129,7 +129,7 @@ async def test_fastapi_server():
129
129
# Sending POST request
130
130
post_url = "http://localhost:8000/settings"
131
131
settings = {
132
- "model" : "gpt-3.5 -turbo" ,
132
+ "model" : "gpt-4 -turbo" ,
133
133
"messages" : [
134
134
{
135
135
"role" : "user" ,
@@ -183,7 +183,7 @@ async def test_fastapi_server():
183
183
# Send another POST request
184
184
post_url = "http://localhost:8000/settings"
185
185
settings = {
186
- "model" : "gpt-3.5 -turbo" ,
186
+ "model" : "gpt-4 -turbo" ,
187
187
"messages" : [
188
188
{
189
189
"role" : "user" ,
@@ -267,7 +267,7 @@ def test_m_vision():
267
267
]
268
268
269
269
interpreter .llm .supports_vision = False
270
- interpreter .llm .model = "gpt-4-turbo "
270
+ interpreter .llm .model = "gpt-4o "
271
271
interpreter .llm .supports_functions = True
272
272
interpreter .llm .context_window = 110000
273
273
interpreter .llm .max_tokens = 4096
@@ -304,7 +304,7 @@ def test_skills():
304
304
305
305
import json
306
306
307
- interpreter .llm .model = "gpt-4-turbo "
307
+ interpreter .llm .model = "gpt-4o "
308
308
309
309
messages = ["USER: Hey can you search the web for me?\n AI: Sure!" ]
310
310
@@ -558,7 +558,7 @@ def setup_function():
558
558
interpreter .reset ()
559
559
interpreter .llm .temperature = 0
560
560
interpreter .auto_run = True
561
- interpreter .llm .model = "gpt-4-turbo "
561
+ interpreter .llm .model = "gpt-4o "
562
562
interpreter .llm .context_window = 123000
563
563
interpreter .llm .max_tokens = 4096
564
564
interpreter .llm .supports_functions = True
0 commit comments