@@ -44,8 +44,8 @@ def export_conversation(checkpointer):
4444 return add_signature (checkpointer .serde .dumps (checkpointer .storage ).decode ('utf-8' ), key )
4545
4646
47- def react (task , nc : Nextcloud ):
48- safe_tools , dangerous_tools = get_tools (nc )
47+ async def react (task , nc : Nextcloud ):
48+ safe_tools , dangerous_tools = await get_tools (nc )
4949
5050 model .bind_nextcloud (nc )
5151
@@ -54,7 +54,7 @@ def react(task, nc: Nextcloud):
5454 + safe_tools
5555 )
5656
57- def call_model (
57+ async def call_model (
5858 state : AgentState ,
5959 config : RunnableConfig ,
6060 ):
@@ -77,12 +77,12 @@ def call_model(
7777""" .replace ("{CURRENT_DATE}" , current_date )
7878 )
7979
80- response = bound_model .invoke ([system_prompt ] + state ["messages" ], config )
80+ response = await bound_model .ainvoke ([system_prompt ] + state ["messages" ], config )
8181 # We return a list, because this will get added to the existing list
8282 return {"messages" : [response ]}
8383
8484 checkpointer = MemorySaver ()
85- graph = get_graph (call_model , safe_tools , dangerous_tools , checkpointer )
85+ graph = await get_graph (call_model , safe_tools , dangerous_tools , checkpointer )
8686
8787 load_conversation (checkpointer , task ['input' ]['conversation_token' ])
8888
@@ -105,7 +105,7 @@ def call_model(
105105 else :
106106 new_input = {"messages" : [("user" , task ['input' ]['input' ])]}
107107
108- for event in graph .stream (new_input , thread , stream_mode = "values" ):
108+ async for event in graph .astream (new_input , thread , stream_mode = "values" ):
109109 last_message = event ['messages' ][- 1 ]
110110 for message in event ['messages' ]:
111111 if isinstance (message , HumanMessage ):
0 commit comments