44from PySide6 .QtCore import Signal
55from PySide6 .QtWidgets import QStackedWidget , QWidget , QSizePolicy , QHBoxLayout , QVBoxLayout , QMessageBox
66
7- from pyqt_openai .config_loader import CONFIG_MANAGER
8- from pyqt_openai .globals import LLAMAINDEX_WRAPPER , DB , get_openai_chat_model , get_argument , ChatThread
97from pyqt_openai .chat_widget .center .chatBrowser import ChatBrowser
108from pyqt_openai .chat_widget .center .chatHome import ChatHome
119from pyqt_openai .chat_widget .center .menuWidget import MenuWidget
1210from pyqt_openai .chat_widget .center .prompt import Prompt
13- from pyqt_openai .chat_widget .chatThread import LlamaOpenAIThread
11+ from pyqt_openai .chat_widget .llamaOpenAIThread import LlamaOpenAIThread
12+ from pyqt_openai .config_loader import CONFIG_MANAGER
13+ from pyqt_openai .globals import LLAMAINDEX_WRAPPER , DB
14+ from pyqt_openai .util .script import get_argument , ChatThread
1415from pyqt_openai .lang .translations import LangClass
1516from pyqt_openai .models import ChatMessageContainer
1617from pyqt_openai .widgets .notifier import NotifierWidget
@@ -21,7 +22,7 @@ class ChatWidget(QWidget):
2122 onMenuCloseClicked = Signal ()
2223
2324 def __init__ (self , parent = None ):
24- super (ChatWidget , self ).__init__ (parent )
25+ super ().__init__ (parent )
2526 self .__initVal ()
2627 self .__initUi ()
2728
@@ -115,7 +116,7 @@ def __chat(self):
115116 try :
116117 # Get necessary parameters
117118 stream = CONFIG_MANAGER .get_general_property ('stream' )
118- model = CONFIG_MANAGER .get_general_property ('model' )
119+ model = CONFIG_MANAGER .get_general_property ('g4f_model' ) if self . __is_g4f else CONFIG_MANAGER . get_general_property ( ' model' )
119120 system = CONFIG_MANAGER .get_general_property ('system' )
120121 temperature = CONFIG_MANAGER .get_general_property ('temperature' )
121122 max_tokens = CONFIG_MANAGER .get_general_property ('max_tokens' )
@@ -125,11 +126,16 @@ def __chat(self):
125126 presence_penalty = CONFIG_MANAGER .get_general_property ('presence_penalty' )
126127 use_llama_index = CONFIG_MANAGER .get_general_property ('use_llama_index' )
127128 use_max_tokens = CONFIG_MANAGER .get_general_property ('use_max_tokens' )
129+ provider = CONFIG_MANAGER .get_general_property ('provider' )
130+ g4f_use_chat_history = CONFIG_MANAGER .get_general_property ('g4f_use_chat_history' )
128131
129132 # Get image files
130133 images = self .__prompt .getImageBuffers ()
131134
132- messages = self .__browser .getMessages (CONFIG_MANAGER .get_general_property ('maximum_messages_in_parameter' ))
135+ maximum_messages_in_parameter = CONFIG_MANAGER .get_general_property ('maximum_messages_in_parameter' )
136+ messages = self .__browser .getMessages (maximum_messages_in_parameter )
137+ if self .__is_g4f and not g4f_use_chat_history :
138+ messages = []
133139
134140 cur_text = self .__prompt .getContent ()
135141
@@ -197,7 +203,7 @@ def __chat(self):
197203 # Run a different thread based on whether the llama-index is enabled or not.
198204 self .__t = LlamaOpenAIThread (param , container , LLAMAINDEX_WRAPPER , query_text )
199205 else :
200- self .__t = ChatThread (param , info = container , is_g4f = self .__is_g4f )
206+ self .__t = ChatThread (param , info = container , is_g4f = self .__is_g4f , provider = provider )
201207
202208 self .__t .started .connect (self .__beforeGenerated )
203209 self .__t .replyGenerated .connect (self .__browser .showLabel )
0 commit comments