@@ -218,7 +218,11 @@ def start_terminal_interface(interpreter):
218
218
"nargs" : "?" , # This means you can pass in nothing if you want
219
219
},
220
220
{"name" : "profiles" , "help_text" : "opens profiles directory" , "type" : bool },
221
- {"name" : "local_models" , "help_text" : "opens local models directory" , "type" : bool },
221
+ {
222
+ "name" : "local_models" ,
223
+ "help_text" : "opens local models directory" ,
224
+ "type" : bool ,
225
+ },
222
226
{
223
227
"name" : "conversations" ,
224
228
"help_text" : "list conversations to resume" ,
@@ -304,7 +308,7 @@ def start_terminal_interface(interpreter):
304
308
if args .profiles :
305
309
open_storage_dir ("profiles" )
306
310
return
307
-
311
+
308
312
if args .local_models :
309
313
open_storage_dir ("models" )
310
314
return
@@ -381,8 +385,7 @@ def start_terminal_interface(interpreter):
381
385
except :
382
386
# Doesn't matter
383
387
pass
384
-
385
-
388
+
386
389
if interpreter .llm .api_base :
387
390
if (
388
391
not interpreter .llm .model .lower ().startswith ("openai/" )
@@ -391,12 +394,11 @@ def start_terminal_interface(interpreter):
391
394
and not interpreter .llm .model .lower ().startswith ("jan" )
392
395
and not interpreter .llm .model .lower ().startswith ("local" )
393
396
):
394
- interpreter .llm .model = "openai/" + interpreter .llm .model
397
+ interpreter .llm .model = "openai/" + interpreter .llm .model
395
398
elif interpreter .llm .model .lower ().startswith ("jan/" ):
396
399
# Strip jan/ from the model name
397
400
interpreter .llm .model = interpreter .llm .model [4 :]
398
401
399
-
400
402
# If --conversations is used, run conversation_navigator
401
403
if args .conversations :
402
404
conversation_navigator (interpreter )
0 commit comments