@@ -60,7 +60,7 @@ def generate_text(
6060 n : int = 1 ,
6161 temperature : float = 1e-8 ,
6262 stop : t .Optional [t .List [str ]] = None ,
63- callbacks : Callbacks = [] ,
63+ callbacks : Callbacks = None ,
6464 ) -> LLMResult :
6565 ...
6666
@@ -71,7 +71,7 @@ async def agenerate_text(
7171 n : int = 1 ,
7272 temperature : float = 1e-8 ,
7373 stop : t .Optional [t .List [str ]] = None ,
74- callbacks : Callbacks = [] ,
74+ callbacks : Callbacks = None ,
7575 ) -> LLMResult :
7676 ...
7777
@@ -81,7 +81,7 @@ async def generate(
8181 n : int = 1 ,
8282 temperature : float = 1e-8 ,
8383 stop : t .Optional [t .List [str ]] = None ,
84- callbacks : Callbacks = [] ,
84+ callbacks : Callbacks = None ,
8585 is_async : bool = True ,
8686 ) -> LLMResult :
8787 """Generate text using the given event loop."""
@@ -119,9 +119,7 @@ class LangchainLLMWrapper(BaseRagasLLM):
119119 """
120120
121121 def __init__ (
122- self ,
123- langchain_llm : BaseLanguageModel ,
124- run_config : t .Optional [RunConfig ] = None
122+ self , langchain_llm : BaseLanguageModel , run_config : t .Optional [RunConfig ] = None
125123 ):
126124 self .langchain_llm = langchain_llm
127125 if run_config is None :
@@ -134,7 +132,7 @@ def generate_text(
134132 n : int = 1 ,
135133 temperature : float = 1e-8 ,
136134 stop : t .Optional [t .List [str ]] = None ,
137- callbacks : t . Optional [ Callbacks ] = None ,
135+ callbacks : Callbacks = None ,
138136 ) -> LLMResult :
139137 temperature = self .get_temperature (n = n )
140138 if is_multiple_completion_supported (self .langchain_llm ):
@@ -164,7 +162,7 @@ async def agenerate_text(
164162 n : int = 1 ,
165163 temperature : float = 1e-8 ,
166164 stop : t .Optional [t .List [str ]] = None ,
167- callbacks : t . Optional [ Callbacks ] = None ,
165+ callbacks : Callbacks = None ,
168166 ) -> LLMResult :
169167 temperature = self .get_temperature (n = n )
170168 if is_multiple_completion_supported (self .langchain_llm ):
@@ -206,8 +204,7 @@ def set_run_config(self, run_config: RunConfig):
206204
207205
208206def llm_factory (
209- model : str = "gpt-3.5-turbo-16k" ,
210- run_config : t .Optional [RunConfig ] = None
207+ model : str = "gpt-3.5-turbo-16k" , run_config : t .Optional [RunConfig ] = None
211208) -> BaseRagasLLM :
212209 timeout = None
213210 if run_config is not None :
0 commit comments