Discord AI Agent
Discord bot for supporting AI/LLM chat applications powered by the Model Context Protocol (MCP), allowing for numerous integrations
For general MCP resources, see Arkestra:cookbook/mcp/README.md
uv pip install -U .The Discord AI Agent supports multiple LLM providers through an extensible architecture:
- Generic/OpenAI-compatible (default): Works with LM Studio, Ollama, and other OpenAI-compatible endpoints
- OpenAI Cloud: Direct integration with OpenAI's API
- Anthropic Claude: Direct integration with Anthropic's Claude API
Install with specific provider support:
# For OpenAI cloud support
uv pip install -U ".[openai]"
# For Claude support
uv pip install -U ".[claude]"
# For both OpenAI and Claude
uv pip install -U ".[openai,claude]"
# For RSS functionality
uv pip install -U ".[rss]"
# For all features
uv pip install -U ".[openai,claude,rss]"API keys should be provided via environment variables for security:
# For OpenAI/generic providers
export OPENAI_API_KEY="your-openai-api-key"
# For Claude provider
export ANTHROPIC_API_KEY="your-anthropic-api-key"cp config/example.main.toml config/main.tomlThen edit main.toml as needed. It specifies your LLM endpoint and model context resources such as MCP servers.
For LM Studio, Ollama, or other OpenAI-compatible endpoints:
[llm_endpoint]
api_type = "generic"
base_url = "http://localhost:1234/v1"
model = "qwen3-1.7b-mlx"For direct OpenAI API access:
[llm_endpoint]
api_type = "openai"
model = "gpt-4o-mini"For direct Claude API access:
[llm_endpoint]
api_type = "claude"
model = "claude-3-5-sonnet-20241022"Copy the appropriate example configuration:
# For OpenAI cloud
cp config/openai.example.main.toml config/main.toml
# For Claude
cp config/claude.example.main.toml config/main.toml
# For local LM Studio (default)
cp config/example.main.toml config/main.tomlFollow the same pattern with config/*news*.example.b4a.toml depending on what context tools you plan to use, e.g.
cp config/toys.example.b4a.toml config/toys.b4a.tomlIf you included toys.b4a.toml in your main.toml you'll need to have that MCP server running. In a separate terminal:
cd demo_server
uv pip install -Ur requirements.txt --constraint=constraints.txt
uvicorn toy_mcp_server:create_app --factory --host 127.0.0.1 --port 8902To use a different port, make sure you also update the b4a.toml file
Make sure you set up any other MCP or other resources you've specified in your B4A. Now you can run the bot.
# Assumes you've exported DISCORD_TOKEN="YOUR_TOKEN"
python mcp_discord_bot.py --discord-token $DISCORD_TOKEN --config-path configStructlog/rich tracebacks can be elaborate, so there is a --classic-tracebacks option to tame them
For very copious logging you can add --loglevel DEBUG
Note: you can use the environment rather than --discord-token & --config-path
export AIBOT_DISCORD_TOKEN="YOUR_TOKEN"
export AIBOT_DISCORD_CONFIG_PATH="./config"
python mcp_discord_bot.py # Reads from env varsLaunch servers & bot, then DM the bot
-
Use command
/context_info -
Use command
/context_details -
Try out some sample queries listed below
-
Check some of your registered tools (just raw tool calls without LLM intervention)
/invoke_tool tool_name:add tool_input:{"a": 1234, "b": 5678}
/invoke_tool tool_name:magic_8_ball tool_input:{}
- Check some of your registered RSS feeds, including some different
/invoke_tool tool_name:query_rss_feed tool_input:{"feed_name": "Reddit r/Boulder"}
/invoke_tool tool_name:query_rss_feed tool_input:{"feed_name": "Reddit r/LocalLLaMA", "query": "new AI model"}
/invoke_tool tool_name:query_rss_feed tool_input:{"feed_name": "Reddit r/LocalLLaMA", "limit": 3}
- Set up a standing prompt using
/set_standing_prompt
Pick schedule:"Hourly", then write a prompt, for example, if you do have the RSS query tool set up to include Reddit's LocalLLaMa community, you could try:
Summarize any discussion of new AI models from LocalLlama
/set_standing_prompt schedule:Hourly prompt:Summarize any discussion of new AI models from LocalLlama
- Ask the magic 8-ball if I should deploy to production on Friday
- Consult the magic 8-ball about my chances of winning the lottery
- Magic 8-ball, tell me: will it rain tomorrow?
- Any recent observations about Qwen 3 among Local LLM enthusiasts today?
- /set_standing_prompt schedule:Hourly prompt:Summarize the latest news about AI development from LocalLlama. If there's any error retrieving news, or there's nothing new, don't assume anything or make anything up. Just give the plain facts as they are.
To enable persistent chat history storage using a PostgreSQL database with the PGVector extension, Ensure you have a running PostgreSQL database server with the pgvector extension available, then set up the environment variables accordingly. See below for notes on DB setup.
uv pip install ogbujipt pgvector asyncpg sentence-transformersSet enabled = true in the [pgvector_history] section of the config file, e.g. main.toml. You must then set some required environment variables, either based on a PG connection string or on elaborated credentials.
Make sure it's in the AIBOT_PG_USER_CONNECT_STRING environment variable. If this variable exists, it will supersede any other PG env vars.
Other environment variables:
- Required:
AIBOT_PG_DB_NAME: Name of the database to useAIBOT_PG_USER: Role name to use. It's a single-role authentication systemAIBOT_PG_PASSWORD: Role's password
- Optional (Defaults Shown):
AIBOT_PG_HOST: Database host (default:localhost)AIBOT_PG_PORT: Database port (default:5432)AIBOT_PG_TABLE_NAME: Table name for storing history (default:discord_chat_history)AIBOT_PG_EMBEDDING_MODEL: Sentence Transformer model to use for embeddings (default:all-MiniLM-L6-v2)
- Database Setup: [cite: 2]. You can use Docker for a quick setup (similar to the demo notebook [cite: 3]):
# Example using default user/pass/db - change values as needed!
docker run --name pg_chat_history -d -p 5432:5432 \
-e POSTGRES_DB=YOUR_PGVECTOR_DB_NAME \
-e POSTGRES_USER=YOUR_PGVECTOR_USER \
-e POSTGRES_PASSWORD=YOUR_PGVECTOR_PASSWORD \
pgvector/pgvectorReplace YOUR_PGVECTOR_DB_NAME, YOUR_PGVECTOR_USER, and YOUR_PGVECTOR_PASSWORD with the values you set in the environment variables.
The bot will automatically connect to the database, create the table (if it doesn't exist), and start storing/retrieving chat history from PGVector. If any required variables are missing or the connection fails, it will fall back to the default in-memory history.
- discord.py
- mcp-sse-client MCP client library
uv pip install mlx-omni-server
mlx-omni-server --port 1234
# uv pip install mlx mlx_lm
# mlx_lm.server --model mlx-community/Llama-3.2-3B-Instruct-4bit --port 1234Note: with mlx-omni-server ran into RuntimeError: Failed to generate completion: generate_step() got an unexpected keyword argument 'user'
Fixed with this patch:
diff --git a/chat/mlx/mlx_model.py b/chat/mlx/mlx_model.py
index da7aef5..094ae9c 100644
--- a/chat/mlx/mlx_model.py
+++ b/chat/mlx/mlx_model.py
@@ -45,6 +45,9 @@ class MLXModel(BaseTextModel):
def _get_generation_params(self, request: ChatCompletionRequest) -> Dict[str, Any]:
params = request.get_extra_params()
+ # Exclude user. See #37
+ if "user" in params:
+ del params["user"]
known_params = {
"top_k",
"min_tokens_to_keep",There are many local MLX models from which you can pick
For SSE servers, you can check with curl, e.g.
curl -N http://localhost:8901/sseYou can try a simple connection as follows, to make sure there is no exception:
import os
import asyncio
from sentence_transformers import SentenceTransformer
from ogbujipt.embedding.pgvector import MessageDB
emodel_name = os.environ.get('AIBOT_PG_EMBEDDING_MODEL', 'all-MiniLM-L6-v2')
emodel = SentenceTransformer(emodel_name)
su_conn_str = os.environ.get('AIBOT_PG_SUPERUSER_CONNECT_STRING')
tname = os.environ.get('AIBOT_PG_TABLE_NAME', 'discord_chat_history')
db = asyncio.run(await MessageDB.from_conn_string(su_conn_str, emodel, tname))You can use the "Connect" icon at the top to get connection string info
Unless you buy an IPV4 add-on you need to use the session pooler version of the connection string,
or you'll get nodename nor servname provided, or not known. Ref: https://github.com/orgs/supabase/discussions/33534
If so, don't forget to include the tenant ID (e.g. [USER].hjdsfghjfbdhsk; the part after the dot) or you'll get InternalServerError: Tenant or user not found
It's probably a good idea to have an app-level user, in order to assert least privilege.
You can just use util/supabase_setup.py, which you should run only once.
op run --env-file .env -- python util/supabase_setup.pyTo make sure asyncpg doesn't cause probs with this
Disable automatic use of prepared statements by passing
statement_cache_size=0toasyncpg.connect()andasyncpg.create_pool()(and, obviously, avoid the use ofConnection.prepare()
Tables are in the public schema in Supabase, are automatically exposed via Supabase's auto-generated REST API. Without RLS enabled, any table in the public schema is accessible through the API to anyone who has this anon key, which is exposed in client-side code.
If you do have your Supabase table in the public schema, we recommend you enable RLS on it, without creating any policies. That will blocks all public API access to the table, while the configured service role can bypass RLS. For example:
ALTER TABLE public.discord_chat_history ENABLE ROW LEVEL SECURITY;