Skip to content

nulib/meadow

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5,258 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Meadow

Build Coverage Status

Prerequisites

Initial startup:

  • From the meadow project root, cd app.
  • Install Elixir dependencies with mix deps.get
  • Run mix meadow.setup. This creates the Sequins pipeline, S3 buckets, and database.
  • Install Node.js dependencies with mix assets.install
    • assets.install looks for all package-lock.json files project-wide and runs npm install in each directory found, so you don't need to run npm install in individual directories.
  • run sgport open all 3001
  • Start the Phoenix server with mix phx.server (or iex -S mix phx.server if you want to an interactive shell).

Now you can visit https://[YOURENV].dev.rdc.library.northwestern.edu:3001/ from your browser.

Stopping the application

You can stop the Phoenix server with Ctrl + C twice

Clearing and resetting data

If you need to clear your data and reset the entire development environment, from meadow/app run:

mix ecto.reset
mix meadow.search.clear
mix meadow.pipeline.purge
clean-s3 dev -y

...then

mix deps.get
mix meadow.setup
mix phx.server

Dependencies

You may need to run mix deps.get again if new dependencies have been added

You may need to run mix assets.install if new node packages have been added

Database

If you need to reset the database you can run mix ecto.reset which will drop + create + migrate the database

If you just want to run the migrations but leave the data intact, you can just do mix ecto.migrate

If you would like to interact directly with the database

Run the Elixir test suite

Start/Provision Test Environment

In one terminal:

make localstack-provision

Run Tests

cd app
export AWS_LOCALSTACK=true
mix test [test args...]

Note: Do not try to run Meadow with export AWS_LOCALSTACK=true set.

Stop/Deprovision Test Environment

Back in the meadow root directory:

make localstack-stop

GraphQL API

You can visit the GraphiQL interface at: https://[YOURENV].dev.rdc.library.northwestern.edu:3001/api/graphiql

Livebook Integration

To start meadow with superuser Livebook integration, run: MEADOW_ROOT/bin/meadow-livebook [iex arguments]

For example, from Meadow's root directory: ./bin/meadow-livebook phx.server

Opensearch Dashboard

  • To start: es-proxy start
  • To stop: es-proxy stop
  • See the console output for the url to the dashboard

Digital Collections API

In order to see data and thumbnails from your current environment, you'll need to run the DC API alongside Meadow. Follow the instructions for Running the API locally and [Running the API locally via our AWS dev domain](Running the API locally via our AWS dev domain) to get it running.

Reindexing data

To force an Elasticsearch re-index, and not wait for the 2-minute cycle to kick in when updating a Meadow item:

Run the interactive shell in a terminal tab

iex -S mix

And force a re-index:

Meadow.Data.Indexer.reindex_all()

AI Agent Plans

Meadow supports AI agent-generated plans for batch modifications to works. The system uses a two-table structure that allows agents to propose work-specific changes based on high-level prompts.

Data Model

Plans - High-level task definitions

  • prompt: Natural language instruction (e.g., "Add a date_created EDTF string for the work based on the work's existing description, creator, and temporal subjects")
  • query: OpenSearch query string identifying target works
    • Collection query: "collection.id:abc-123"
    • Specific works: "id:(work-id-1 OR work-id-2 OR work-id-3)"
  • status: :pending, :proposed, :approved, :rejected, :completed, or :error`

PlanChanges - Work-specific modifications

  • plan_id: Foreign key to parent plan
  • work_id: Specific work being modified
  • add: Map of values to append to existing work data
  • delete: Map of values to remove from existing work data
  • replace: Map of values to fully replace in work data
  • status: Individual approval/rejection tracking

Each PlanChange must specify at least one operation (add, delete, or replace).

PlanChange payloads

  • add merges values into existing metadata. For lists (like subjects or notes) the values are appended when they are not already present. Scalar fields (e.g., title) are merged according to the context (:append for add, :replace for replace).
  • delete removes the provided values verbatim. For controlled vocabularies this means the JSON structure must match what is stored in the database (role/term maps). The planner normalizes structs and string-keyed maps automatically when applying changes.
  • replace overwrites existing values for the provided keys. Use this when the existing content should be replaced entirely instead of appended or removed.

Controlled metadata entries (subjects, creators, contributors, etc.) follow the shape below. For subjects you must supply both the role (with at least id/scheme) and the term.id; extra fields such as label or variants are ignored when applying but can be included when working with structs in IEx:

%{
  descriptive_metadata: %{
    subject: [
      %{
        role: %{id: "TOPICAL", scheme: "subject_role"},
        term: %{
          id: "http://id.loc.gov/authorities/subjects/sh85141086",
          label: "Universities and colleges",
          variants: ["Colleges", "Higher education institutions"]
        }
      }
    ]
  }
}

When constructing PlanChanges you can mix-and-match operations as needed. For example, to remove an outdated subject and add a new one in a single change:

delete: %{
  descriptive_metadata: %{
    subject: [
      %{role: %{id: "TOPICAL", scheme: "subject_role"}, term: %{id: "mock1:result2"}}
    ]
  }
},
add: %{
  descriptive_metadata: %{
    subject: [
      %{role: %{id: "TOPICAL", scheme: "subject_role"}, term: %{id: "mock1:result5"}}
    ]
  }
}

Example Workflows

Adding new metadata:

# 1. Create a plan with a query - PlanChanges are auto-generated for matching works
{:ok, plan} = Meadow.Data.Planner.create_plan(%{
  prompt: "Add a date_created EDTF string for the work based on the work's existing description, creator, and temporal subjects",
  query: "collection.id:abc-123"
})

# 2. Agent updates each auto-generated PlanChange with work-specific values
changes = Meadow.Data.Planner.list_plan_changes(plan.id)

change_a = Enum.at(changes, 0)
{:ok, updated_change_a} = Meadow.Data.Planner.update_plan_change(change_a, %{
  add: %{descriptive_metadata: %{date_created: ["1896-11-10"]}}
})

change_b = Enum.at(changes, 1)
{:ok, updated_change_b} = Meadow.Data.Planner.update_plan_change(change_b, %{
  add: %{descriptive_metadata: %{date_created: ["1923-05"]}}
})

Removing unwanted values:

# Remove extraneous subject headings
{:ok, change} = Meadow.Data.Planner.create_plan_change(%{
  plan_id: plan.id,
  work_id: "work-id",
  delete: %{
    descriptive_metadata: %{
      subject: [
        %{role: %{id: "TOPICAL", scheme: "subject_role"}, term: %{id: "http://example.org/photograph"}},
        %{role: %{id: "TOPICAL", scheme: "subject_role"}, term: %{id: "http://example.org/image"}}
      ]
    }
  }
})

Replacing existing values:

# Replace the title
{:ok, change} = Meadow.Data.Planner.create_plan_change(%{
  plan_id: plan.id,
  work_id: "work-id",
  replace: %{descriptive_metadata: %{title: "New Title"}}
})

Reviewing and applying:

# 3. User reviews and approves
{:ok, _} = Meadow.Data.Planner.approve_plan(plan, "user@example.com")
{:ok, _} = Meadow.Data.Planner.approve_plan_change(change_a, "user@example.com")
{:ok, _} = Meadow.Data.Planner.approve_plan_change(change_b, "user@example.com")

# 4. Apply approved changes
{:ok, completed_plan} = Meadow.Data.Planner.apply_plan(plan)

Doing development on the Meadow Pipeline lambdas

In the AWS developer environment, the lambdas associated with the pipeline are shared amongst developers. In order to do development and see whether it's working you can override the configuration to use the SAM pipeline the deployed lambdas.

In one terminal:

make pipeline-start

In another terminal:

cd app
USE_SAM_LAMBDAS=true iex -S mix phx.server

Deploying lambdas with SAM

The pipeline infrastructure is defined in infrastructure/pipeline/template.yaml and can be deployed with the AWS SAM CLI. There are make tasks to assist. Make sure AWS_PROFILE is set to the correct admin profile and logged in, and then:

make pipeline-deploy ENV=staging

TypeScript/GraphQL Types

Meadow now supports TypeScript and GraphQL types in the React app. To generate types, run the following commands:

# Generate a local JSON version of GraphQL schema
mix graphql.schema.export -o priv/graphql/schema.json

# Generate TypeScript types for the UI
cd assets
npm run generate-types

Types will be generated in meadow/app/assets/js/__generated__. You can import them into React components like so:

import type { FileSet, Work as WorkType } from "@js/__generated__/graphql";

const SomeComponent = ({ work }: { work: WorkType }) => {
  // ...
};

Terraform

Meadow's Terraform code is stored in this repo. To run Terraform commands, you'll need to do the configuration setup

UI Customization

Meadow runs in Development, Staging and Production environments. To help distinguish environments (and avoid potential errors), Staging and Development environments support alternate, background colors.

Production

  • A wrapper CSS class of is-production-environment wraps the main HTML element (in case anyone wants to target a selector for any reason).

Staging

  • Supports a toggle background color switch in the site header
  • Customize your own dev background color by updating the hex value for localStorage property devBg
  • A wrapper CSS class of is-staging-environment wraps the main HTML element.

Development

  • Supports a toggle background color switch in the site header
  • Customize your own dev background color by updating the hex value for localStorage property devBg
  • A wrapper CSS class of is-development-environment wraps the main HTML element.

About

A Northwestern Libraries repository application

Topics

Resources

Stars

Watchers

Forks

Packages

No packages published

Contributors 12