@@ -21,7 +21,8 @@ class ScriptCreatorGraph(AbstractGraph):
2121 source (str): The source of the graph.
2222 config (dict): Configuration parameters for the graph.
2323 llm_model: An instance of a language model client, configured for generating answers.
24- embedder_model: An instance of an embedding model client, configured for generating embeddings.
24+ embedder_model: An instance of an embedding model client,
25+ configured for generating embeddings.
2526 verbose (bool): A flag indicating whether to show print statements during execution.
2627 headless (bool): A flag indicating whether to run the graph in headless mode.
2728 model_token (int): The token limit for the language model.
@@ -44,7 +45,7 @@ class ScriptCreatorGraph(AbstractGraph):
4445 def __init__ (self , prompt : str , source : str , config : dict ):
4546
4647 self .library = config ['library' ]
47-
48+
4849 super ().__init__ (prompt , config , source )
4950
5051 self .input_key = "url" if source .startswith ("http" ) else "local_dir"
@@ -61,25 +62,29 @@ def _create_graph(self) -> BaseGraph:
6162 input = "url | local_dir" ,
6263 output = ["doc" ],
6364 node_config = {
64- "headless" : True if self .config is None else self .config .get ("headless" , True )}
65+ "headless" : True if self .config is None else self .config .get ("headless" , True ),
66+ "verbose" : self .verbose }
6567 )
6668 parse_node = ParseNode (
6769 input = "doc" ,
6870 output = ["parsed_doc" ],
69- node_config = {"chunk_size" : self .model_token }
71+ node_config = {"chunk_size" : self .model_token ,
72+ "verbose" : self .verbose }
7073 )
7174 rag_node = RAGNode (
7275 input = "user_prompt & (parsed_doc | doc)" ,
7376 output = ["relevant_chunks" ],
7477 node_config = {
7578 "llm" : self .llm_model ,
76- "embedder_model" : self .embedder_model
79+ "embedder_model" : self .embedder_model ,
80+ "verbose" : self .verbose
7781 }
7882 )
7983 generate_scraper_node = GenerateScraperNode (
8084 input = "user_prompt & (relevant_chunks | parsed_doc | doc)" ,
8185 output = ["answer" ],
82- node_config = {"llm" : self .llm_model },
86+ node_config = {"llm" : self .llm_model ,
87+ "verbose" : self .verbose },
8388 library = self .library ,
8489 website = self .source
8590 )
@@ -106,7 +111,7 @@ def run(self) -> str:
106111 Returns:
107112 str: The answer to the prompt.
108113 """
109-
114+
110115 inputs = {"user_prompt" : self .prompt , self .input_key : self .source }
111116 self .final_state , self .execution_info = self .graph .execute (inputs )
112117
0 commit comments