-
Notifications
You must be signed in to change notification settings - Fork 0
Enhancement: update MCP example #75
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Open
ryanignatius
wants to merge
3
commits into
main
Choose a base branch
from
e/update-example
base: main
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Open
Changes from all commits
Commits
Show all changes
3 commits
Select commit
Hold shift + click to select a range
File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
93 changes: 93 additions & 0 deletions
93
examples/custom-pipeline/mcp_pipeline/components/chat_history_manager.py
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,93 @@ | ||
| """Chat history manager component for MCP Pipeline. | ||
|
|
||
| Authors: | ||
| Samuel Lusandi (samuel.lusandi@gdplabs.id) | ||
| """ | ||
|
|
||
| import datetime | ||
| import uuid | ||
|
|
||
| from typing import Any | ||
|
|
||
| from gllm_core.event import EventEmitter as EventEmitter | ||
| from gllm_core.schema import Component | ||
| from gllm_datastore.sql_data_store import SQLAlchemySQLDataStore | ||
| from gllm_misc.chat_history_manager.schema import Message | ||
|
|
||
|
|
||
| class McpChatHistoryManager(Component): | ||
| """Chat history manager component.""" | ||
|
|
||
| def __init__( | ||
| self, | ||
| data_store: SQLAlchemySQLDataStore, | ||
| ): | ||
| """Initialize the chat history manager.""" | ||
| self.data_store = data_store | ||
|
|
||
| async def _run(self, **kwargs: str) -> Any: | ||
| """Run the chat history manager component. | ||
|
|
||
| Args: | ||
| kwargs (Any): The keyword arguments, which may contain the operation. | ||
|
|
||
| Returns: | ||
| Any: The result of the operation. | ||
| """ | ||
| if not self.data_store: | ||
| return None | ||
|
|
||
| operation = kwargs.get("operation") | ||
|
|
||
| if operation == "read": | ||
| return await self.read(kwargs) | ||
|
|
||
| if operation == "write": | ||
| return await self.write(kwargs) | ||
|
|
||
| async def read(self, kwargs: dict[str, Any]) -> Any: | ||
| """Read the chat history. | ||
|
|
||
| Args: | ||
| kwargs (dict[str, Any]): The keyword arguments. | ||
|
|
||
| Returns: | ||
| Any: The chat history. | ||
| """ | ||
| return None | ||
|
|
||
| async def write(self, kwargs: dict[str, Any]) -> Any: | ||
| """Write the chat history. | ||
|
|
||
| Args: | ||
| kwargs (dict[str, Any]): The keyword arguments. | ||
|
|
||
| Returns: | ||
| Any: The chat history. | ||
| """ | ||
| user_message = Message( | ||
| id=kwargs.get("user_message_id") or str(uuid.uuid4()), | ||
| conversation_id=kwargs.get("conversation_id"), | ||
| role="user", | ||
| content=kwargs.get("query"), | ||
| parent_id=kwargs.get("parent_id") or kwargs.get("conversation_id"), | ||
| created_time=datetime.datetime.now(datetime.UTC), | ||
| is_active=kwargs.get("is_active", True), | ||
| feedback=None, | ||
| source="user", | ||
| metadata_="{}", | ||
| ) | ||
| assistant_message = Message( | ||
| id=kwargs.get("assistant_message_id") or str(uuid.uuid4()), | ||
| conversation_id=kwargs.get("conversation_id"), | ||
| role="assistant", | ||
| content=kwargs.get("response"), | ||
| parent_id=user_message.id, | ||
| created_time=datetime.datetime.now(datetime.UTC), | ||
| is_active=kwargs.get("is_active", True), | ||
| feedback=None, | ||
| source=kwargs.get("source"), | ||
| metadata_="{}", | ||
| ) | ||
| self.data_store.create([user_message, assistant_message]) | ||
| return "" | ||
212 changes: 212 additions & 0 deletions
212
examples/custom-pipeline/mcp_pipeline/components/response_synthesizer.py
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,212 @@ | ||
| """Response Synthesizer for MCP Pipeline. | ||
|
|
||
| Authors: | ||
| Samuel Lusandi (samuel.lusandi@gdplabs.id) | ||
| """ | ||
|
|
||
| import json | ||
| import uuid | ||
| import time | ||
|
|
||
| from typing import Any | ||
|
|
||
| from gllm_core.event import EventEmitter | ||
| from gllm_core.constants import EventLevel, EventType | ||
| from gllm_inference.schema import Attachment, MultimodalContent, PromptRole | ||
| from gllm_generation.response_synthesizer.response_synthesizer import BaseResponseSynthesizer | ||
|
|
||
| from langchain_mcp_adapters.client import MultiServerMCPClient | ||
| from langchain_openai import ChatOpenAI | ||
| from langchain_core.language_models import BaseLanguageModel | ||
| from langgraph.prebuilt import create_react_agent | ||
|
|
||
| from mcp_pipeline.mcp_config import get_mcp_servers | ||
|
|
||
|
|
||
| def get_language_model(model: str, key: str) -> BaseLanguageModel: | ||
| provider = model.split("/")[0] | ||
| model_name = model.split("/")[1] | ||
| if provider == "openai": | ||
| return ChatOpenAI(model=model_name, api_key=key) | ||
| else: | ||
| raise ValueError(f"Unsupported model: {model}") | ||
|
|
||
|
|
||
| class MaximumToolCallsException(Exception): | ||
| """Exception raised when the maximum number of tool calls is reached.""" | ||
| pass | ||
|
|
||
|
|
||
| class McpResponseSynthesizer(BaseResponseSynthesizer): | ||
|
|
||
| MAX_TOOL_CALLS = 10 | ||
|
|
||
| def __init__(self, model: str, key: str, mcp_server_url: str): | ||
| super().__init__() | ||
| self.model = model | ||
| self.key = key | ||
| self.mcp_server_url = mcp_server_url | ||
|
|
||
| async def synthesize_response( | ||
| self, | ||
| query: str | None = None, | ||
| state_variables: dict[str, Any] | None = None, | ||
| history: list[tuple[PromptRole, str | list[Any]]] | None = None, | ||
| extra_contents: list[MultimodalContent] | None = None, | ||
| attachments: list[Attachment] | None = None, | ||
| hyperparameters: dict[str, Any] | None = None, | ||
| event_emitter: EventEmitter | None = None, | ||
| system_multimodal_contents: list[Any] | None = None, | ||
| user_multimodal_contents: list[Any] | None = None, | ||
| ) -> str: | ||
| """Synthesizes a response based on the provided query. | ||
|
|
||
| This abstract method must be implemented by subclasses to define the logic for generating a response. It | ||
| may optionally take an input `query`, some other input variables passed through `state_variables`, and an | ||
| `event_emitter`. It returns the synthesized response as a string. | ||
|
|
||
| Args: | ||
| query (str | None, optional): The input query used to synthesize the response. Defaults to None. | ||
| state_variables (dict[str, Any] | None, optional): Additional state variables to assist in generating the | ||
| response. Defaults to None. | ||
| history (list[tuple[PromptRole, str | list[Any]]] | None, optional): The chat history of the conversation | ||
| to be considered in generating the response. Defaults to None. | ||
| extra_contents (list[MultimodalContent] | None, optional): The additional multimodal contents to be considered | ||
| in generating the response. Defaults to None. | ||
| attachments (list[Attachment] | None, optional): The attachments to be considered in generating the response. | ||
| Defaults to None. | ||
| hyperparameters (dict[str, Any] | None, optional): The hyperparameters to be considered in generating the response. | ||
| Defaults to None. | ||
| event_emitter (EventEmitter | None, optional): The event emitter for handling events during response | ||
| synthesis. Defaults to None. | ||
| system_multimodal_contents (list[Any] | None, optional): The system multimodal contents to be considered | ||
| in generating the response. Defaults to None. | ||
| user_multimodal_contents (list[Any] | None, optional): The user multimodal contents to be considered in | ||
| generating the response. Defaults to None. | ||
|
|
||
| Returns: | ||
| str: The synthesized response. | ||
|
|
||
| Raises: | ||
| NotImplementedError: If the method is not implemented in a subclass. | ||
| """ | ||
| start_time = time.time() | ||
| client = MultiServerMCPClient(get_mcp_servers(self.mcp_server_url)) | ||
| tools = await client.get_tools() | ||
| for tool in tools: | ||
| tool.name = tool.name.replace("::", "__") | ||
|
|
||
| model = self.model | ||
| key = self.key | ||
|
|
||
| agent = create_react_agent( | ||
| name="HelloAgent", | ||
| prompt=f"""You are a helpful assistant that can utilize all tools given to you to solve the user's input. | ||
|
|
||
| When there is anything related to relative time, you *must* call the get_current_time tool. Otherwise you will not be able to | ||
| provide an accurate response. The timezone *must* be UTC+7 Asia/Jakarta. | ||
|
|
||
| Here are a few things you need to do for specific tasks: | ||
|
|
||
| Regarding Slack: | ||
| - Ping specific user on slack: You must first find the user's slack ID and then ping them using the format <@user_id> | ||
| - Reminder the following when you use IDs: | ||
| * IDs prefixed with `T` indicate Teams/Organizations (T*******) | ||
| * IDs prefixed with `U` indicate Users (U*******) | ||
| * IDs prefixed with `C` indicate Channels (C*******) | ||
| * IDs prefixed with `D` indicate Direct Messages (D*******) | ||
| - Regarding sending a message to slack channel: If a user specifies a channel, and IF the channel cannot be found, do not give up yet; | ||
| it is likely a private channel. | ||
| * Try sending it directly, the slack_post_message tool will throw an error if the channel truly cannot be found. | ||
| * Remember to just use the name of the channel as ID directly! | ||
|
|
||
| Regarding Github: | ||
| - If the user does not specify a repository owner, you *must* assume it is `GDP-ADMIN`. (i.e., `bosa` becomes `GDP-ADMIN/bosa`) | ||
| - If the user does not specify a repository owner and a repository name, you do not have to proceed; simply tell the user that you need more information. | ||
| - Regarding issues and pull requests in github: | ||
| * If the result is under 1000 items, you *must* use the search endpoints (i.e., search_repositories, search_issues, etc.) because | ||
| your query can become a lot more versatile, and the results will be much more accurate. You can filter by date, filter by user, | ||
| etc. | ||
| * If the result is over 1000 items, you *must* use the list endpoints (i.e., list_repositories, list_issues, etc.) because | ||
| the search endpoints will not return all the results. However, do know that this endpoint doesn't have as much query capabilities, | ||
| i.e., there's no query, no filter by date, etc. | ||
| """, | ||
| model=get_language_model(model, key), | ||
| tools=tools, | ||
| ) | ||
|
|
||
| final_response = "" | ||
| processed_tool_ids = set() | ||
|
|
||
| tool_call_count = 0 | ||
| message_prefix = "" | ||
| try: | ||
| async for chunk in agent.astream({"messages": query}, stream_mode="values"): | ||
| if isinstance(chunk, dict) and 'messages' in chunk: | ||
| messages = chunk['messages'] | ||
|
|
||
| for message in messages: | ||
| if hasattr(message, 'additional_kwargs') and 'tool_calls' in message.additional_kwargs: | ||
| tool_calls = message.additional_kwargs['tool_calls'] | ||
| for tool_call in tool_calls: | ||
| if tool_call['id'] not in processed_tool_ids: | ||
| processed_tool_ids.add(tool_call['id']) | ||
| tool_call_count += 1 | ||
|
|
||
| tool_info = f"Called tool `{tool_call['function']['name']}`" | ||
|
|
||
| step_id = str(uuid.uuid4()) | ||
| tool_running_data = { | ||
| "data_type": "process", | ||
| "data_value": { | ||
| "id": step_id, | ||
| "message": tool_info, | ||
| "status": "running", | ||
| "time": round(time.time() - start_time, 2) | ||
| }, | ||
| } | ||
| tool_finished_data = { | ||
| "data_type": "process", | ||
| "data_value": { | ||
| "id": step_id, | ||
| "message": tool_info, | ||
| "status": "finished", | ||
| "time": round(time.time() - start_time, 2) | ||
| }, | ||
| } | ||
|
|
||
| if event_emitter: | ||
| await event_emitter.emit( | ||
| json.dumps(tool_running_data), | ||
| event_level=EventLevel.INFO, | ||
| event_type=EventType.DATA | ||
| ) | ||
| await event_emitter.emit( | ||
| json.dumps(tool_finished_data), | ||
| event_level=EventLevel.INFO, | ||
| event_type=EventType.DATA | ||
| ) | ||
|
|
||
| if tool_call_count >= self.MAX_TOOL_CALLS: | ||
| raise MaximumToolCallsException("Maximum tool calls reached") | ||
| except MaximumToolCallsException as e: | ||
| print(f"Error during tool calls: {e}") | ||
| message_prefix = "We've reached the maximum number of tool calls. This is what we have so far:\n\n" | ||
| except Exception as e: | ||
| print(f"Error during tool calls: {e}") | ||
| message_prefix = "An error occurred while processing your request. This is what we have so far:\n\n" | ||
|
|
||
| if 'messages' in chunk: | ||
| messages = chunk['messages'] | ||
| last_ai_message = next((msg for msg in reversed(messages) | ||
| if hasattr(msg, 'content') and msg.content), None) | ||
| final_response = message_prefix + (last_ai_message.content if last_ai_message else "No response generated") | ||
|
|
||
| if event_emitter and final_response: | ||
| await event_emitter.emit( | ||
| final_response, | ||
| event_level=EventLevel.INFO, | ||
| event_type=EventType.RESPONSE | ||
| ) | ||
|
|
||
| return final_response |
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Using
datetime.UTCwill raise an AttributeError; consider usingdatetime.timezone.utcor a valid timezone object.