@@ -48,12 +48,7 @@ def _process_input(self, chats, token: Optional[str] = None):
4848 elif m .role == tt .Message .HUMAN :
4949 final_messages .append ({"role" : "user" , "content" : m .value })
5050 elif m .role == tt .Message .GPT :
51- final_messages .append (
52- {
53- "role" : "assistant" ,
54- "content" : m .value .strip (),
55- }
56- )
51+ final_messages .append ({"role" : "assistant" , "content" : m .value })
5752 elif m .role == tt .Message .FUNCTION_CALL :
5853 _m = tu .from_json (m .value ) if isinstance (m .value , str ) else m .value
5954 final_messages .append (
@@ -97,7 +92,7 @@ def chat(
9792 self ,
9893 chats : tt .Thread | str ,
9994 model : Optional [str ] = None ,
100- max_tokens : int = 1024 ,
95+ max_tokens : int = None ,
10196 temperature : float = 1 ,
10297 parallel_tool_calls : bool = False ,
10398 token : Optional [str ] = None ,
@@ -126,7 +121,7 @@ def stream_chat(
126121 self ,
127122 chats : tt .Thread | str ,
128123 model : Optional [str ] = None ,
129- max_tokens : int = 1024 ,
124+ max_tokens : int = None ,
130125 temperature : float = 1 ,
131126 parallel_tool_calls : bool = False ,
132127 token : Optional [str ] = None ,
@@ -144,13 +139,14 @@ def stream_chat(
144139 "messages" : messages ,
145140 "model" : model or self .model_id ,
146141 "stream" : True ,
147- "max_tokens" : max_tokens ,
148- "parallel_tool_calls" : parallel_tool_calls ,
149142 }
143+ if max_tokens :
144+ data ["max_tokens" ] = max_tokens
150145 if isinstance (chats , tt .Thread ) and len (chats .tools ):
151146 data ["tools" ] = [
152147 {"type" : "function" , "function" : x .to_dict ()} for x in chats .tools
153148 ]
149+ data ["parallel_tool_calls" ] = parallel_tool_calls
154150 if debug :
155151 fp = "sample_oai.json"
156152 print ("Saving at path " + fp )
0 commit comments