@@ -1112,9 +1112,9 @@ def generate_client_response_streaming(
11121112 model_input : Messages ,
11131113) -> Generator [YieldMessage , Any , tuple [Message , Any ]]:
11141114 msg_stream : Generator [Message , Any , Any ]
1115- model_input_str = messages_to_str (block .model , model_input )
11161115 match block :
11171116 case BamModelBlock ():
1117+ model_input_str = messages_to_str (block .model , model_input )
11181118 msg_stream = BamModel .generate_text_stream (
11191119 model_id = block .model ,
11201120 prompt_id = block .prompt_id ,
@@ -1172,9 +1172,9 @@ def generate_client_response_single(
11721172 model_input : Messages ,
11731173) -> Generator [YieldMessage , Any , tuple [Message , Any ]]:
11741174 msg : Message
1175- model_input_str = messages_to_str (block .model , model_input )
11761175 match block :
11771176 case BamModelBlock ():
1177+ model_input_str = messages_to_str (block .model , model_input )
11781178 msg , raw_result = BamModel .generate_text (
11791179 model_id = block .model ,
11801180 prompt_id = block .prompt_id ,
@@ -1203,9 +1203,9 @@ def generate_client_response_batching( # pylint: disable=too-many-arguments
12031203 # model: str,
12041204 model_input : Messages ,
12051205) -> Generator [YieldMessage , Any , Message ]:
1206- model_input_str = messages_to_str (block .model , model_input )
12071206 match block :
12081207 case BamModelBlock ():
1208+ model_input_str = messages_to_str (block .model , model_input )
12091209 msg = yield ModelCallMessage (
12101210 model_id = block .model ,
12111211 prompt_id = block .prompt_id ,
0 commit comments