1212from memos .llms .utils import remove_thinking_tags
1313from memos .log import get_logger
1414from memos .types import MessageList
15- from memos .utils import timed
15+ from memos .utils import timed_with_status
1616
1717
1818logger = get_logger (__name__ )
@@ -28,7 +28,7 @@ def __init__(self, config: OpenAILLMConfig):
2828 )
2929 logger .info ("OpenAI LLM instance initialized" )
3030
31- @timed ( log = True , log_prefix = "OpenAI LLM" , log_args = ["model_name_or_path" ])
31+ @timed_with_status ( log_prefix = "OpenAI LLM" , log_args = ["model_name_or_path" ])
3232 def generate (self , messages : MessageList , ** kwargs ) -> str :
3333 """Generate a response from OpenAI LLM, optionally overriding generation params."""
3434 response = self .client .chat .completions .create (
@@ -55,7 +55,7 @@ def generate(self, messages: MessageList, **kwargs) -> str:
5555 return reasoning_content + response_content
5656 return response_content
5757
58- @timed ( log = True , log_prefix = "OpenAI LLM" , log_args = ["model_name_or_path" ])
58+ @timed_with_status ( log_prefix = "OpenAI LLM" , log_args = ["model_name_or_path" ])
5959 def generate_stream (self , messages : MessageList , ** kwargs ) -> Generator [str , None , None ]:
6060 """Stream response from OpenAI LLM with optional reasoning support."""
6161 if kwargs .get ("tools" ):
0 commit comments