Skip to content

Latest commit

 

History

History
335 lines (257 loc) · 15.2 KB

File metadata and controls

335 lines (257 loc) · 15.2 KB

Backlog — AI Chief Data Brain

Vision: From querying data → to understanding business → to autonomously improving it

Sprint methodology: Pick top 10 → design architecture + UX → implement → test → lint → commit → push → verify CI → mark done → next task. After sprint completion, wait for approval, re-prioritize, start next sprint.


Sprint 1 — Complete

# Task Status Priority Dependencies Est. Complexity
1 Foundation: Data Graph + Memory + Trust Layers done P0 None (prerequisite for all) High
2 Autonomous Insight Feed done P0 Task 1 High
3 Anomaly Intelligence (upgrade) done P0 Tasks 1, 2 Medium
4 Opportunity Detector done P0 Tasks 1, 2 Medium
5 Loss Detector done P0 Tasks 1, 2 Medium
6 Insight → Action Engine done P0 Tasks 3, 4, 5 Medium
7 Cross-Source Reconciliation Engine done P1 Task 1 High
8 Semantic Layer Auto-Build done P1 Task 1 High
9 Query-less Exploration done P1 Tasks 1, 2 Medium
10 Temporal Intelligence Engine done P1 Task 1 Medium

Sprint 1 — Task Details

Task 1: Foundation — Data Graph + Memory + Trust Layers

Goal: Build the three core modules that every subsequent feature depends on.

Data Graph Module

  • Unified registry of metrics, sources, relationships, and dependencies
  • Auto-populated from DB index pipeline (extend probe_service)
  • Models: MetricDefinition, MetricRelationship, SourceNode
  • Service: DataGraphService — build, query, update, visualize graph
  • API: /api/data-graph/{project_id} (GET, POST refresh)

Memory Layer

  • Persistent store for findings: what was discovered, confirmed/rejected, confidence, decay
  • Extends existing AgentLearning pattern but for insights
  • Models: InsightRecord (finding, type, confidence, status, source_metrics, timestamps)
  • Service: InsightMemoryService — store, query, confirm, reject, decay
  • Orchestrator integration: check memory before generating new insights
  • API: /api/insights/{project_id} (GET, PATCH confirm/reject)

Trust Layer

  • Confidence scoring, source validation, traceability for every insight
  • Model: TrustScore (insight_id, confidence, sources, validation_method)
  • Service: TrustService — score calculation, source verification
  • Wrapper: TrustedInsight dataclass for all agent outputs
  • Frontend: confidence badges on insights

Edge cases: Empty DB index, single source only, no codebase connected, stale metrics, conflicting relationships, LLM unavailable during graph build.


Task 2: Autonomous Insight Feed

Goal: System proactively analyzes data and surfaces insights without user questions.

Backend:

  • New agent: InsightFeedAgent — runs on schedule or on-demand
  • Scheduler integration: background task that periodically scans connected sources
  • Analysis pipeline: observe metrics → compare to history → detect changes → generate insights
  • Store insights in Memory Layer with confidence scores
  • Prioritization: critical problems > growth opportunities > informational

Frontend:

  • New InsightFeedPanel component — dedicated feed in sidebar or main area
  • Cards: "3 things that changed in 24 hours", "1 critical problem", "2 growth opportunities"
  • Real-time updates via SSE
  • Dismiss, confirm, drill-down actions

API: /api/feed/{project_id} (GET latest, POST trigger scan)

Edge cases: No data changes in 24h, too many insights (need ranking/dedup), LLM quota limits, user hasn't connected any sources yet, connection down during scan.


Task 3: Anomaly Intelligence (upgrade)

Goal: Upgrade DataSanityChecker from "something changed" to "why / where / how critical."

Backend:

  • Extend DataSanityCheckerAnomalyIntelligenceEngine
  • Root cause analysis: when anomaly detected, run follow-up queries to find "why"
  • Severity scoring: business impact estimation (revenue, users affected)
  • Context enrichment: pull related metrics, similar past events from Memory Layer
  • LLM-powered explanation generation

Frontend:

  • Upgraded anomaly cards with severity indicators, root cause, and suggested actions
  • Timeline view of anomalies
  • Link anomalies to related insights in the feed

Edge cases: Anomaly is just noise (high false positive rate), cascading anomalies from single root cause, insufficient data for root cause analysis.


Task 4: Opportunity Detector

Goal: Find segments with high LTV, undermonetized users, channels with potential.

Backend:

  • New agent: OpportunityAgent
  • Segment analysis: automatic cohort comparison on key metrics
  • Pattern detection: "Users from X convert N% better"
  • Gap analysis: where traffic/conversion potential exists
  • Store opportunities in Memory Layer

Frontend:

  • Opportunity cards with impact estimates, evidence, and suggested actions
  • Integration with Insight Feed

Edge cases: Not enough data for segmentation, single-product businesses, no revenue data available, privacy considerations for user segmentation.


Task 5: Loss Detector

Goal: Find revenue leaks, inefficient spend, conversion drops.

Backend:

  • New agent: LossDetectorAgent
  • Funnel analysis: identify where users/revenue drop off
  • Spend analysis: flag inefficient channels/campaigns
  • Regression detection: compare current vs historical conversion rates
  • Quantify losses: "$X/month lost due to Y"

Frontend:

  • Loss cards with monetary impact, trend visualization, and fix suggestions
  • Integration with Insight Feed

Edge cases: No funnel data, no spend data, seasonal drops misidentified as losses, data latency causing false alarms.


Task 6: Insight → Action Engine ✅ done

Goal: Every insight gets a concrete recommended action with expected impact.

Backend:

  • New service: ActionEngine
  • Takes any insight (anomaly, opportunity, loss) and generates:
    • What to do (specific, actionable)
    • Expected impact (quantified)
    • Confidence level
    • Priority
  • LLM-powered action generation with data context
  • Track action outcomes when user reports back

Frontend:

  • Action cards attached to every insight
  • "Expected: +X% if you do Y" format
  • Action tracking: mark as done, report outcome

Edge cases: Insight too vague for action, action requires external systems not connected, multiple conflicting actions, user lacks permissions to act.


Task 7: Cross-Source Reconciliation Engine ✅ done

Goal: Compare data across sources (DB vs Stripe, Ads vs CRM) and find discrepancies.

Backend:

  • New agent: ReconciliationAgent
  • Compare matching metrics across connections
  • Detect: missing records, value mismatches, timing differences
  • Report discrepancies with severity and likely cause
  • Leverages Data Graph to find comparable metrics

Frontend:

  • Reconciliation dashboard: side-by-side comparison
  • Discrepancy list with severity
  • Drill-down into specific mismatches

Edge cases: Different schemas across sources, timezone mismatches, currency differences, eventual consistency delays, no overlapping metrics between sources.


Task 8: Semantic Layer Auto-Build ✅ done

Goal: Auto-discover metrics, normalize definitions, replace tribal knowledge.

Backend:

  • Extend DB index pipeline to extract metric definitions
  • LLM-powered: analyze column names, types, relationships → generate metric catalog
  • Normalize: unify definitions across connections (e.g., "revenue" means the same everywhere)
  • Service: SemanticLayerService — build, query, update catalog
  • Store in Data Graph

Frontend:

  • Metric catalog browser
  • Edit/confirm metric definitions
  • Link metrics to business terms

Edge cases: Ambiguous column names, different definitions across DBs, custom aggregation formulas, calculated fields.


Task 9: Query-less Exploration ✅ done

Goal: User says "What's wrong?" and agent autonomously investigates.

Backend:

  • New exploration mode in Orchestrator
  • Autonomous investigation pipeline:
    1. Scan recent insights and anomalies
    2. Run diagnostic queries across sources
    3. Build hypothesis → test → confirm/reject
    4. Compile findings into structured report
  • Leverage Memory Layer for context

Frontend:

  • "Explore" button or natural language trigger
  • Progressive disclosure: show investigation steps as they happen
  • Final report with findings, severity, actions

Edge cases: No clear issues found (positive report), too many issues (need prioritization), investigation takes too long (timeout/streaming), user interrupts mid-investigation.


Task 10: Temporal Intelligence Engine ✅ done

Goal: Understand trends, seasonality, lags — not just snapshots.

Backend:

  • New service: TemporalIntelligenceService
  • Time series analysis: decompose into trend + seasonality + residual
  • Lag detection: find delayed effects between metrics
  • Anomaly detection in temporal context (adjust for seasonality)
  • Pure Python (statsmodels-lite or custom) — no heavy ML deps

Frontend:

  • Time series charts with trend overlay
  • Seasonality indicators
  • "This is normal for this time" context on insights

Edge cases: Insufficient history for seasonality detection, irregular time intervals, missing data points, multiple seasonality patterns.


Backlog Queue — Ideas

These are not prioritized within the queue. They will be prioritized when Sprint 1 is complete.

# Idea Category Notes
11 Cross-Source Causal Graph Engine Core Build cause-effect relationships between metrics across sources
12 Data Hypothesis Generator Multiplier Auto-generate growth and problem hypotheses
13 Auto Cohort Discovery Multiplier Find unexpected segments and hidden patterns automatically
14 Behavioral Pattern Mining Multiplier Find action chains and leading indicators ("users who do X → convert 3x")
15 KPI Dependency Mapping Multiplier Map what influences what and how strongly
16 Smart Alerts with Context Memory Multiplier Alerts that know past events, similar cases, resolved problems
17 Data Confidence Engine Multiplier Per-insight confidence with sources and validations
18 Auto Benchmarking Engine Multiplier Compare with history, industry, internal segments
19 Predictive Scenario Engine Advanced "What if you increase price by 10%?" simulations
20 Cross-Company Pattern Learning Advanced/Moat Learn from anonymized patterns across companies
21 Autonomous Optimization Loops Advanced/Moat Find → propose → test → optimize automatically
22 Multi-database JOIN queries Platform Query across multiple connections in a single question
23 Natural language data transformations Platform "Show me revenue per user cohort by signup month"
24 AI-powered data documentation generation Platform Auto-generate docs for every table, column, relationship
25 Real-time collaborative analysis Collaboration Multiple users exploring data together
26 Shared insight feeds with public links Collaboration Share feed with stakeholders
27 Slack/Discord integration Collaboration Push insights to team channels
28 PDF/email report generation Output Scheduled report delivery
29 Custom visualization builder Platform User-defined chart types
30 Plugin/extension system Platform Third-party integrations
31 SSO (SAML, OIDC) Enterprise Enterprise authentication
32 Granular RBAC Enterprise Fine-grained permissions per source/metric
33 Audit log export Enterprise Compliance and security
34 BigQuery connector Connectors Google BigQuery support
35 Snowflake connector Connectors Snowflake support
36 DuckDB connector Connectors DuckDB/local analytics
37 SQLite connector Connectors SQLite support
38 Redshift connector Connectors AWS Redshift support
39 Query result caching with invalidation Performance Cache repeated queries
40 Dark/light theme toggle UX Theme switching
41 Keyboard-driven navigation UX Power user shortcuts
42 Natural language query templates UX Reusable question patterns
43 Mobile-responsive improvements UX Better mobile experience
44 Data lineage visualization Data Quality See where data flows
45 Schema change detection Data Quality Alert when schema changes
46 Data freshness monitoring Data Quality Track when sources last updated
47 Metric alerting with thresholds Monitoring Set alerts on any metric
48 Goal tracking Business Track progress toward business goals
49 Competitive intelligence integration Advanced Pull and compare industry data
50 Natural language API (headless mode) Platform Use as API without UI
51 Webhook notifications Integration Push events to external systems
52 Data catalog search UX Search across all metrics and tables
53 Conversation branching UX Fork a chat to explore alternatives
54 Insight sharing with annotations Collaboration Share specific insights with team notes
55 Cost attribution analysis Business Attribute costs to features/teams
56 Revenue attribution modeling Business Multi-touch attribution
57 Churn prediction Advanced Predict which users will churn
58 LTV prediction Advanced Predict user lifetime value
59 A/B test analysis agent Advanced Automated experiment analysis
60 Data quality score per source Data Quality Overall health score per connection

Completed

# Task Completed Date Notes
1 Foundation: Data Graph + Memory + Trust Layers 2026-03-22 Models, services, API, frontend panel
2 Autonomous Insight Feed 2026-03-22 InsightFeedAgent, /api/feed/, InsightFeedPanel
3 Anomaly Intelligence (upgrade) 2026-03-22 AnomalyIntelligenceEngine, AnomalyReportCard
4 Opportunity Detector 2026-03-22 OpportunityDetector, OpportunityCard
5 Loss Detector 2026-03-22 LossDetector, LossReportCard
6 Insight → Action Engine 2026-03-22 ActionEngine, ActionCard
7 Cross-Source Reconciliation Engine 2026-03-22 ReconciliationEngine, ReconciliationCard
8 Semantic Layer Auto-Build 2026-03-22 SemanticLayerService, MetricCatalogPanel
9 Query-less Exploration 2026-03-22 ExplorationEngine, ExplorationReport
10 Temporal Intelligence Engine 2026-03-22 TemporalIntelligenceService, TemporalReport

Process

  1. Take next pending task from Sprint
  2. Design architecture (backend modules, models, services, API)
  3. Design UX/UI (frontend components, interactions, responsive)
  4. Analyze edge cases
  5. Create 30+ step implementation plan
  6. Implement
  7. Test (unit + integration + frontend)
  8. Lint check (ruff, mypy, eslint, tsc)
  9. Commit + push
  10. Verify GitHub Actions pass
  11. If CI fails → fix → re-push
  12. Mark task as done in this file
  13. Proceed to next task
  14. After Sprint complete → wait for approval → re-prioritize → next Sprint