Skip to content

Deploy live SPARQL playground with datasets for projects & credit classes #42

@clevinson

Description

@clevinson

MVP: Initial hosted RDF Triple Store / Knowledge Graph Integration

This issue tracks initial work for getting our graph infrastructure up and running. The goal is to automate schema publishing to a queryable RDF graph (hosted in Neptune) and start loading real project data. Mostly backend and infra-focused, but with an eye toward future UX improvements.

✅ MVP Tasks

  • Get Apache Neptune instance running, and explore how to integrate it @clevinson
    • Spin up the graph database and start poking around—make sure it’ll work for loading and querying RDF generated from our LinkML YAML schemas.
  • Set up GitHub Action to convert YAML in schema/data/playground/* into RDF and post to Neptune @paul121
    • Automate the pipeline: convert LinkML → RDF → load to Neptune on push to main.
  • Refactor schema/data folder to follow new pattern (schema/data/playground) and convert existing credit class JSON-LD to YAML @clevinson
    • Move everything into a consistent structure and rewrite the JSON-LD examples as YAML so they’re easier to maintain.
  • Generate project datasets for C02, C06, BT01 @S4mmyb
    • Draft project instances for a few real projects to test loading and queries.

🌀 Possible Additional MVP Tasks

  • Add root identifiers to credit class & project schemas (@id in JSON-LD / identifier: true in LinkML)
  • Update slot_uri for projectStartDate and projectEndDate
  • Add hosted SPARQL UI (e.g. http://yasgui.triply.cc) — ideally with LLM integration to help generate queries

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions