@@ -1051,7 +1051,9 @@ def get_transformed_inputs(kwargs):
10511051 if "input" in litellm_params :
10521052 append_log (state , "Model Input" , litellm_params ["input" ])
10531053 else :
1054- append_log (state , "Model Input" , model_input )
1054+ append_log (
1055+ state , "Model Input" , messages_to_str (concrete_block .model , model_input )
1056+ )
10551057 background : Messages = [msg ]
10561058 result = msg ["content" ]
10571059 append_log (state , "Model Output" , result )
@@ -1093,7 +1095,7 @@ def generate_client_response_streaming(
10931095 model_input : Messages ,
10941096) -> Generator [YieldMessage , Any , Message ]:
10951097 msg_stream : Generator [Message , Any , None ]
1096- model_input_str = messages_to_str (model_input )
1098+ model_input_str = messages_to_str (block . model , model_input )
10971099 match block :
10981100 case BamModelBlock ():
10991101 msg_stream = BamModel .generate_text_stream (
@@ -1148,7 +1150,7 @@ def generate_client_response_single(
11481150 model_input : Messages ,
11491151) -> Generator [YieldMessage , Any , Message ]:
11501152 msg : Message
1151- model_input_str = messages_to_str (model_input )
1153+ model_input_str = messages_to_str (block . model , model_input )
11521154 match block :
11531155 case BamModelBlock ():
11541156 msg = BamModel .generate_text (
@@ -1178,7 +1180,7 @@ def generate_client_response_batching( # pylint: disable=too-many-arguments
11781180 # model: str,
11791181 model_input : Messages ,
11801182) -> Generator [YieldMessage , Any , Message ]:
1181- model_input_str = messages_to_str (model_input )
1183+ model_input_str = messages_to_str (block . model , model_input )
11821184 match block :
11831185 case BamModelBlock ():
11841186 msg = yield ModelCallMessage (
0 commit comments