@@ -136,6 +136,9 @@ def download_model(models_dir, models, interpreter):
136
136
)
137
137
]
138
138
answers = inquirer .prompt (questions )
139
+
140
+ if answers == None :
141
+ exit ()
139
142
140
143
# Get the selected model
141
144
selected_model = next (
@@ -201,6 +204,8 @@ def download_model(models_dir, models, interpreter):
201
204
]
202
205
answers = inquirer .prompt (questions )
203
206
207
+ if answers == None :
208
+ exit ()
204
209
205
210
selected_model = answers ["model" ]
206
211
@@ -266,7 +271,11 @@ def list_ollama_models():
266
271
),
267
272
]
268
273
name_answer = inquirer .prompt (name_question )
269
- selected_name = name_answer ["name" ] if name_answer else None
274
+
275
+ if name_answer == None :
276
+ exit ()
277
+
278
+ selected_name = name_answer ["name" ]
270
279
271
280
if selected_name == "llama3" :
272
281
# If the user selects llama3, we need to check if it's installed, and if not, install it
@@ -316,7 +325,11 @@ def list_ollama_models():
316
325
),
317
326
]
318
327
model_name_answer = inquirer .prompt (model_name_question )
319
- jan_model_name = model_name_answer ["jan_model_name" ] if model_name_answer else None
328
+
329
+ if model_name_answer == None :
330
+ exit ()
331
+
332
+ jan_model_name = model_name_answer ["jan_model_name" ]
320
333
interpreter .llm .model = f"jan/{ jan_model_name } "
321
334
interpreter .display_message (f"\n Using Jan model: `{ jan_model_name } ` \n " )
322
335
time .sleep (1 )
@@ -360,6 +373,9 @@ def list_ollama_models():
360
373
)
361
374
]
362
375
answers = inquirer .prompt (questions )
376
+
377
+ if answers == None :
378
+ exit ()
363
379
364
380
if answers ["model" ] == " ↓ Download new model" :
365
381
model_path = download_model (models_dir , models , interpreter )
0 commit comments