CLI for Graphsignal: login (store API key) and fetch debug context from api.graphsignal.com. You can also install a skill so your AI coding agent (Cursor, Claude Code, Codex, etc.) can run the CLI and use the returned context to help debug inference, profiles, or errors.
pip install graphsignal-debugStore your Graphsignal API key in ~/.graphsignal/config.yml:
graphsignal-debug loginYou will be prompted for your API key.
Fetch debug context for a time range. Requires being logged in.
graphsignal-debug fetch --start 2026-03-10T00:00:00Z --end 2026-03-12T00:00:00ZOptional --tags filter (semicolon-separated key:value pairs). Tags must match exactly the tags sent to Graphsignal when the app was instrumented:
graphsignal-debug fetch --start 2026-03-10T00:00:00Z --end 2026-03-12T00:00:00Z --tags "env:prod"The command calls GET /api/v1/debug_context/ on https://api.graphsignal.com with start_time_ns, end_time_ns, and optional tags, and prints the response context.
Install the Graphsignal skill so your AI coding agent can run graphsignal-debug fetch for a time range and use the returned context (profiles, errors, traces) to help you debug.
Claude Code — Clone the repo into Claude's personal skills directory:
git clone https://github.com/graphsignal/graphsignal-debug ~/.claude/skills/graphsignal-debugOther agents (Cursor, Codex, Gemini) — Use the skills.sh registry:
npx skills add graphsignal/graphsignal-debugInstall the CLI first: pip install graphsignal-debug, then graphsignal-debug login with your API key.
- Cursor — Use the skill when working in Cursor with agent/composer.
- Claude Code — Use with Claude Code (e.g. via Claude CLI or supported IDEs).
- Codex — Use with Codex agent workflows.
- Gemini CLI — Use with Gemini from the command line.
Once the skill is installed, you can ask the agent to:
- Find the root cause of a latency spike — e.g. "Fetch Graphsignal data for the last 2 hours and find the root cause of the latency spike" or "What's causing the slowdown? Use Graphsignal debug context from 10am to noon today."
- Explain errors or failures — e.g. "Get debug context for the last 24 hours and summarize any errors or failures" or "Why did inference fail around 3pm? Pull Graphsignal data for that window."
- Inspect profiles and bottlenecks — e.g. "Fetch Graphsignal context for yesterday and identify the main performance bottlenecks" or "Which operations are taking the most time? Use Graphsignal data from the last 6 hours."
The agent will call graphsignal-debug fetch --start <ISO> --end <ISO> (and optional --tags when you specify deployment or service tags), then analyze the returned context to answer your question.