Skip to content

Commit 1124c88

Browse files
authored
Merge pull request #93 from MLAI-AUS-Inc/disable-medhack-patient-sim-and-announcements
Disable MedHack patient simulator and winner announcement
2 parents 29a7bee + 2427ff6 commit 1124c88

File tree

1 file changed

+10
-1
lines changed

1 file changed

+10
-1
lines changed

roo-standalone/roo/skills/executor.py

Lines changed: 10 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -722,7 +722,9 @@ async def _handle_guess_result(
722722
self, result: dict, user_id: str, skill, text: str,
723723
client, today, thread_history, channel_id: str, guess_text: str,
724724
) -> str:
725-
"""Process the result of a locked-in guess."""
725+
"""Process the result of a locked-in guess. NOTE: Currently disabled."""
726+
return f"<@{user_id}> The diagnosis game is currently disabled. Stay tuned for updates!"
727+
# --- DISABLED: original implementation below ---
726728
from ..slack_client import post_message
727729

728730
if result["correct"]:
@@ -804,6 +806,11 @@ async def _medhack_llm_response(
804806
extra_instruction: str = "",
805807
) -> str:
806808
"""Generate an in-character clinical response for the diagnosis game.
809+
NOTE: Currently disabled.
810+
"""
811+
return "The patient simulator is currently disabled. Stay tuned for updates!"
812+
# --- DISABLED: original implementation below ---
813+
"""
807814
808815
Thread history is included so the LLM can follow the conversation flow
809816
within a thread. Each patient case lives in its own Slack thread, so
@@ -861,6 +868,8 @@ async def _medhack_llm_response(
861868
{"role": "user", "content": prompt}
862869
], model="gpt-5", max_tokens=4096, reasoning_effort="high")
863870
return response.content
871+
"""
872+
# --- END DISABLED ---
864873
865874
async def _execute_with_llm(
866875
self,

0 commit comments

Comments
 (0)