Record your screen. Say what's wrong. Your AI agent fixes it.
Quick Start · Context-Aware Capture · Why markupR · MCP Server · CLI · Integrations · Contributing
Desktop app workflow is the default: record + narrate + stop, then ship context-rich markdown (frames + cursor/window/focus hints when available) directly to your agent.
AI coding agents can't see your screen. When you find a bug, you context-switch into writing mode -- describing the layout issue in text, manually screenshotting, cropping, and dragging images into the right spot. You speak at 150 words per minute but type at 60. The context is lost in translation.
markupR is a desktop capture app first. You hit a hotkey, narrate what you see, and stop. Then it runs a post-session pipeline that aligns transcript timestamps with the recording, extracts the right frames, and outputs structured Markdown your agent can execute against immediately.
- Record -- press a hotkey, talk through what you see
- Process -- Whisper transcribes, ffmpeg extracts frames at the exact moments you described
- Enrich -- capture context is attached to shot markers (cursor position, active window/app, focused element hints when available)
- Output -- structured Markdown with screenshots and context your agent can trust
Cmd+Shift+F --> talk --> Cmd+Shift+F --> Cmd+V into your agent
Download from markupr.com or GitHub Releases.
macOS install note: Apple notarization is currently rolling out. If macOS warns on first launch, use Right-click -> Open once to trust the app. If needed, run:
xattr -dr com.apple.quarantine /Applications/markupR.app
- Press
Cmd+Shift+F(macOS) orCtrl+Shift+F(Windows) to start - Narrate what you see and mark shots when needed
- Press the hotkey again to stop
- Paste the generated report path into Claude Code, Cursor, Windsurf, or any coding agent
npx --package markupr markupr-mcpnpx markupr analyze ./recording.movUse this when you already have a video file. The desktop app remains the primary capture workflow.
Every important frame can carry extra machine-usable context, not just pixels.
- Cursor coordinates at capture time
- Active app + window title (best-effort from OS context)
- Focused element hints (role/text/title hints when available)
- Trigger metadata (
manual,pause, orvoice-command)
This makes the report a high-signal liaison between you and your agent: what you said, what you saw, and where your attention was.
Local-first. Whisper runs on your device. Your recordings, transcripts, and screenshots never leave your machine. No cloud dependency, no account required.
AI-native output. The Markdown output is structured for LLM consumption -- headings, categories, severity levels, inline screenshots, and capture-context hints. Not a raw transcript with random images.
Works everywhere. Desktop app for daily flow. CLI for scripts and CI/CD. MCP server for agent integration. GitHub Action for PR feedback. Same pipeline, four interfaces.
Open source. MIT licensed. No telemetry, no tracking, no analytics. Read the source, fork it, ship it.
# Feedback Session -- Feb 5, 2026
## FB-001: Button sizing issue
The submit button is way too small on mobile. I'm trying to tap it
and keep hitting the cancel link underneath. Needs more vertical
padding, maybe 12px minimum tap target.

## FB-002: Loading state feels janky
After the spinner disappears, the content pops in with no transition.
There's a visible layout shift -- the sidebar jumps left by about
20 pixels.
Each screenshot is extracted from the exact video frame matching your narration timestamp, with context hints attached when available. See full examples in examples/.
Give your AI coding agent eyes and ears. Add markupR as an MCP server and it can capture screenshots, record your screen with voice, and receive structured reports -- all mid-conversation.
Claude Code (~/.claude/settings.json):
{
"mcpServers": {
"markupR": {
"command": "npx",
"args": ["--yes", "--package", "markupr", "markupr-mcp"]
}
}
}Cursor / Windsurf -- same config in your MCP settings.
| Tool | Description |
|---|---|
capture_screenshot |
Grab the current screen and attach context metadata (cursor + active app/window + focus hints when available). |
capture_with_voice |
Record screen + mic for a set duration. Returns a structured report. |
analyze_video |
Process any existing .mov or .mp4 into Markdown with extracted frames (fallback path for externally captured recordings). |
analyze_screenshot |
Run a screenshot through the AI analysis pipeline. |
start_recording |
Begin an interactive recording session. |
stop_recording |
End the session and run the full pipeline. |
You: "The sidebar is overlapping the main content on mobile. Can you see it?"
Agent: [calls capture_screenshot]
"I can see the issue -- the sidebar has position: fixed but no z-index,
and it's 280px wide with no responsive breakpoint. Let me fix the CSS..."
[fixes the code]
No copy-pasting screenshots. No rewriting what you already know. The agent gets structured report context and acts.
Full MCP documentation: README-MCP.md
# Run without installing
npx markupr analyze ./recording.mov
# Or install globally
npm install -g markuprmarkupr analyze <video> -- Process an existing screen recording into structured Markdown.
markupr analyze ./bug-demo.mov
markupr analyze ./recording.mov --output ./reports
markupr analyze ./recording.mov --template github-issue
markupr analyze ./recording.mov --no-frames # transcript onlymarkupr watch [directory] -- Watch for new recordings and auto-process them.
markupr watch ~/Desktop --output ./reportsmarkupr push github <report> -- Create GitHub issues from a feedback report.
markupr push github ./report.md --repo myorg/myapp
markupr push github ./report.md --repo myorg/myapp --dry-runmarkupr push linear <report> -- Create Linear issues from a feedback report.
markupr push linear ./report.md --team ENGmarkdown (default) | json | github-issue | linear | jira | html
- Node.js 18+
- ffmpeg on your PATH (
brew install ffmpeg/apt install ffmpeg/choco install ffmpeg)
Run markupR in CI to get visual feedback on pull requests:
- uses: eddiesanjuan/markupr-action@v1
with:
video-path: ./recordings/
github-token: ${{ secrets.GITHUB_TOKEN }}- Press
Cmd+Shift+F(macOS) orCtrl+Shift+F(Windows) - Narrate what you see and mark shots as needed
- Press the hotkey again to stop
- Paste the file path from your clipboard into Claude Code, Cursor, or any AI agent
+-----------+
Screen + Voice -> | Whisper | -> Timestamped transcript
+-----------+
|
+-----------+
| Analyzer | -> Key moments identified
+-----------+
|
+-----------+
| ffmpeg | -> Frames extracted at exact timestamps
+-----------+
|
+-----------+
| Generator | -> Structured Markdown with inline screenshots
+-----------+
The pipeline degrades gracefully. No ffmpeg? Transcript-only output. No Whisper model? Timer-based screenshots. No API keys? Everything runs locally.
Desktop app capture remains the default path. CLI/MCP analyze_video remains available when you need to process an existing recording.
For architecture details, see CLAUDE.md.
git clone https://github.com/eddiesanjuan/markupr.git
cd markupR
npm install
npm run dev| Command | Description |
|---|---|
npm run dev |
Development mode with hot reload |
npm run build |
Build everything (desktop + CLI + MCP) |
npm test |
Run all tests |
npm run lint |
Lint |
npm run typecheck |
Type check |
- Fork the repository
- Create a feature branch:
git checkout -b feature/your-feature - Run tests:
npm test && npm run lint && npm run typecheck - Open a Pull Request
See CONTRIBUTING.md for full guidelines.
MIT -- see LICENSE.
Built by Eddie San Juan
markupr.com
