@@ -375,11 +375,8 @@ def load_api_key(self, token_path):
375375 ApiModel = data [in_use ]['Model' ] if in_use in data and 'Model' in data [in_use ] else None
376376 if in_use == 'OpenAI' :
377377 ApiKey = data [in_use ]['OpenaiApiKey' ]
378- print ('OpenaiApiKey : ' , ApiKey )
379378 HttpProxyHost = data [in_use ]['HttpProxyHost' ]
380- print ('HttpProxyHost : ' , HttpProxyHost )
381379 HttpProxyPort = data [in_use ]['HttpProxyPort' ]
382- print ('HttpProxyPort : ' , HttpProxyPort )
383380 openaiApiHost = data [in_use ]['ApiHost' ]
384381 if openaiApiHost is not None and len (str (openaiApiHost )) > 0 :
385382 ApiHost = openaiApiHost
@@ -391,7 +388,6 @@ def load_api_key(self, token_path):
391388 if ApiModel is None or "" == ApiModel .strip ():
392389 ApiModel = "gpt-4o"
393390 # Other default models are configured through the client
394- print ('DeepBIApiKey : ' , ApiKey )
395391 ApiHost = CONFIG .ApiHost
396392 elif "ZhiPuAI" == in_use :
397393 ApiKey = data [in_use ]['ApiKey' ]
@@ -410,11 +406,8 @@ def load_api_key(self, token_path):
410406 else :
411407 # 这里是默认的 openai 所有配置都有问题
412408 ApiKey = data ['OpenaiApiKey' ]
413- print ('OpenaiApiKey : ' , ApiKey )
414409 HttpProxyHost = data ['HttpProxyHost' ]
415- print ('HttpProxyHost : ' , HttpProxyHost )
416410 HttpProxyPort = data ['HttpProxyPort' ]
417- print ('HttpProxyPort : ' , HttpProxyPort )
418411 ApiType = "openai"
419412 ApiModel = data [in_use ]['Model' ] if in_use in data and 'Model' in data [in_use ] else None
420413 # all llm config
0 commit comments