-
Notifications
You must be signed in to change notification settings - Fork 0
Description
Right now in
Right now in effective altruism, there seem to be a whole bunch of different Airtable tables and other sorts of data tables. These are typically spread out through a bunch of different websites. It's become a pain.
In particular, this is a bottleneck to forecasting on structured data. If there were better table infrastructure, then that could be used as the foundation for scalable forecasting. We could both have forecasts on all items of existing data tables and on items of future data tables.
If we could import/synchronize these into one system, that could help with:
- Discoverability
- Metadata additions, like comments, and ratings of accuracy.
- A very fast interface for going through them.
- An API to access each part of it (very useful for some forecasting challenges)
Relevant data includes:
- Much of QURI's stuff, like our list of projects (future and previous)
- Our World In Data data tables
- Much of Vipul Naik's websites
- There are a bunch of one-off effective altruism tables out there.
- Michael Aird has a lot of databases/collections. Here's one.
I'm not sure how to best do this. Some options include:
- Host everything using custom instances of Datasette
- Use some enterprise solution that ingests data, like Snowflake.
- Make a more hacky solution with PostGres, directly.
It's possible there's some neat solution to this, like posting all of the data to Github repos, or to some other "Github but for data" alternative. dolthub is one option. This Hacker News Thread discusses other options, with one relevant table being linked to here.
Datahub is also sort of neat: https://datahub.io/
https://www.datopian.com/ does consulting. They might be a good group to reach out to.