@@ -722,7 +722,9 @@ async def _handle_guess_result(
722722 self , result : dict , user_id : str , skill , text : str ,
723723 client , today , thread_history , channel_id : str , guess_text : str ,
724724 ) -> str :
725- """Process the result of a locked-in guess."""
725+ """Process the result of a locked-in guess. NOTE: Currently disabled."""
726+ return f"<@{ user_id } > The diagnosis game is currently disabled. Stay tuned for updates!"
727+ # --- DISABLED: original implementation below ---
726728 from ..slack_client import post_message
727729
728730 if result ["correct" ]:
@@ -804,6 +806,11 @@ async def _medhack_llm_response(
804806 extra_instruction : str = "" ,
805807 ) -> str :
806808 """Generate an in-character clinical response for the diagnosis game.
809+ NOTE: Currently disabled.
810+ """
811+ return "The patient simulator is currently disabled. Stay tuned for updates!"
812+ # --- DISABLED: original implementation below ---
813+ """
807814
808815 Thread history is included so the LLM can follow the conversation flow
809816 within a thread. Each patient case lives in its own Slack thread, so
@@ -861,6 +868,8 @@ async def _medhack_llm_response(
861868 {"role" : "user" , "content" : prompt }
862869 ], model = "gpt-5" , max_tokens = 4096 , reasoning_effort = "high" )
863870 return response .content
871+ """
872+ # --- END DISABLED ---
864873
865874 async def _execute_with_llm(
866875 self,
0 commit comments