@@ -260,48 +260,30 @@ def start_terminal_interface(interpreter):
260
260
nickname = arg .get ("nickname" )
261
261
default = arg .get ("default" )
262
262
263
+ # Construct argument name flags
264
+ flags = (
265
+ [f"-{ nickname } " , f'--{ arg ["name" ]} ' ] if nickname else [f'--{ arg ["name" ]} ' ]
266
+ )
267
+
263
268
if arg ["type" ] == bool :
264
- if nickname :
265
- parser .add_argument (
266
- f"-{ nickname } " ,
267
- f'--{ arg ["name" ]} ' ,
268
- dest = arg ["name" ],
269
- help = arg ["help_text" ],
270
- action = action ,
271
- default = default ,
272
- )
273
- else :
274
- parser .add_argument (
275
- f'--{ arg ["name" ]} ' ,
276
- dest = arg ["name" ],
277
- help = arg ["help_text" ],
278
- action = action ,
279
- default = default ,
280
- )
269
+ parser .add_argument (
270
+ * flags ,
271
+ dest = arg ["name" ],
272
+ help = arg ["help_text" ],
273
+ action = action ,
274
+ default = default ,
275
+ )
281
276
else :
282
277
choices = arg .get ("choices" )
283
-
284
- if nickname :
285
- parser .add_argument (
286
- f"-{ nickname } " ,
287
- f'--{ arg ["name" ]} ' ,
288
- dest = arg ["name" ],
289
- help = arg ["help_text" ],
290
- type = arg ["type" ],
291
- choices = choices ,
292
- default = default ,
293
- nargs = arg .get ("nargs" ),
294
- )
295
- else :
296
- parser .add_argument (
297
- f'--{ arg ["name" ]} ' ,
298
- dest = arg ["name" ],
299
- help = arg ["help_text" ],
300
- type = arg ["type" ],
301
- choices = choices ,
302
- default = default ,
303
- nargs = arg .get ("nargs" ),
304
- )
278
+ parser .add_argument (
279
+ * flags ,
280
+ dest = arg ["name" ],
281
+ help = arg ["help_text" ],
282
+ type = arg ["type" ],
283
+ choices = choices ,
284
+ default = default ,
285
+ nargs = arg .get ("nargs" ),
286
+ )
305
287
306
288
args = parser .parse_args ()
307
289
@@ -357,15 +339,21 @@ def start_terminal_interface(interpreter):
357
339
358
340
### Set some helpful settings we know are likely to be true
359
341
360
- if interpreter .llm .model .startswith ("gpt-4" ) or interpreter .llm .model .startswith ("openai/gpt-4" ):
342
+ if interpreter .llm .model .startswith ("gpt-4" ) or interpreter .llm .model .startswith (
343
+ "openai/gpt-4"
344
+ ):
361
345
if interpreter .llm .context_window is None :
362
346
interpreter .llm .context_window = 128000
363
347
if interpreter .llm .max_tokens is None :
364
348
interpreter .llm .max_tokens = 4096
365
349
if interpreter .llm .supports_functions is None :
366
- interpreter .llm .supports_functions = False if "vision" in interpreter .llm .model else True
350
+ interpreter .llm .supports_functions = (
351
+ False if "vision" in interpreter .llm .model else True
352
+ )
367
353
368
- if interpreter .llm .model .startswith ("gpt-3.5-turbo" ) or interpreter .llm .model .startswith ("openai/gpt-3.5-turbo" ):
354
+ if interpreter .llm .model .startswith (
355
+ "gpt-3.5-turbo"
356
+ ) or interpreter .llm .model .startswith ("openai/gpt-3.5-turbo" ):
369
357
if interpreter .llm .context_window is None :
370
358
interpreter .llm .context_window = 16000
371
359
if interpreter .llm .max_tokens is None :
0 commit comments