You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
fix(docs): update links after veda docs site re-org (#313)
* update links after veda docs site re-org
* Update README.md
Co-authored-by: Jennifer Tran <12633533+botanical@users.noreply.github.com>
* Update README.md
Co-authored-by: Jennifer Tran <12633533+botanical@users.noreply.github.com>
* Update README.md
Co-authored-by: Jennifer Tran <12633533+botanical@users.noreply.github.com>
* Update README.md
Co-authored-by: Jennifer Tran <12633533+botanical@users.noreply.github.com>
* formatting and ingest ui link
* upgrade black to patch version
---------
Co-authored-by: Jennifer Tran <12633533+botanical@users.noreply.github.com>
Copy file name to clipboardExpand all lines: README.md
+12-10Lines changed: 12 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -20,25 +20,25 @@ To add data to VEDA you will:
20
20
21
21
### Step 1: Stage your files
22
22
23
-
Upload files to the staging bucket `s3://veda-data-store-staging` (which you can do with a VEDA JupyterHub account--request access [here](https://docs.openveda.cloud/nasa-veda-platform/scientific-computing/#veda-sponsored-jupyterhub-service)) or a self-hosted bucket in s3 has shared read access to VEDA service. [See docs.openveda.cloud for additional details on preparing files.](https://docs.openveda.cloud/instance-management/adding-content/dataset-ingestion/file-preparation.html)
23
+
Upload files to the staging bucket `s3://veda-data-store-staging` (which you can do with a VEDA JupyterHub account--request access [here](https://docs.openveda.cloud/user-guide/scientific-computing/getting-access.html)) or a self-hosted bucket in s3 has shared read access to VEDA service. [See docs.openveda.cloud for additional details on preparing files.](https://docs.openveda.cloud/user-guide/content-curation/dataset-ingestion/file-preparation.html)
24
24
25
25
### Step 2: Generate STAC metadata in the staging catalog
26
26
27
-
Metadata must first be added to the Staging Catalog [staging.openveda.cloud/api/stac](https://staging.openveda.cloud/api/stac). You will need to create a dataset config file using the veda-ingest-ui and submit it to the `/workflows/dataset/publish` endpoint to generate STAC Collection metadata and generate Item records for the files you have uploaded in Step 1.
27
+
Metadata must first be added to the Staging Catalog [staging.openveda.cloud/api/stac](https://staging.openveda.cloud/api/stac). You will need to create a dataset config file with the [ingest.openveda.cloud](https://ingest.openveda.cloud) form to generate STAC Collection metadata and generate Item records for the files you have staged in Step 1.
28
28
29
-
* Use the veda-ingest-ui form to generate a dataset config and open a veda-data PR
29
+
* Use the [ingest.openveda.cloud](https://ingest.openveda.cloud) form to generate a dataset config and open a veda-data PR
30
30
31
31
* OR manually generate a dataset-config JSON and open a veda-data PR
32
32
33
33
* When a veda-data PR is opened, a github action will automatically (1) POST the config to airflow and stage the collection and items in the staging catalog instance and (2) open a veda-config dashboard preview for the dataset.
34
34
35
-
> See detailed steps for the [dataset submission process](https://docs.openveda.cloud/instance-management/adding-content/dataset-ingestion/) in the contribuing section of [veda-docs](https://nasa-impact.github.io/veda-docs) where you can also find this full ingestion workflow example [geoglam ingest notebook](https://docs.openveda.cloud/instance-management/adding-content/dataset-ingestion/example-template/example-geoglam-ingest.html)
35
+
> See detailed steps for the [dataset submission process](https://docs.openveda.cloud/user-guide/content-curation/dataset-ingestion/) in the Content Curation section of the User Guide in [veda-docs](https://nasa-impact.github.io/veda-docs) where you can also find this full ingestion workflow example [geoglam ingest notebook](https://docs.openveda.cloud/user-guide/content-curation/dataset-ingestion/example-template/example-geoglam-ingest.html)
36
36
37
37
### Step 3: Acceptance testing
38
38
39
39
Perform acceptance testing appropriate for your data. This should include reviewing the [staging.openveda.cloud STAC browser](https://staging.openveda.cloud) and reviewing the corresponding veda-config PR dashboard preview.
40
40
41
-
> See [veda-docs/instance-management/adding-content/dashboard-configuration](https://docs.openveda.cloud/instance-management/adding-content/dashboard-configuration/dataset-configuration.html)for more information about configuring a dashboard preview).
41
+
> See the [Content Curation/Dashboard Configuration](https://docs.openveda.cloud/user-guide/content-curation/dashboard-configuration/) section of the User Guide in docs.openveda.cloud for more information about configuring a dashboard preview.
42
42
43
43
### Step 4: Promote to production
44
44
@@ -48,7 +48,9 @@ After acceptance testing, request approval--when your PR is merged, the dataset
48
48
49
49
### Step 5 [Optional]: Share your data
50
50
51
-
Share your data in the [VEDA Dashboard](https://www.earthdata.nasa.gov/dashboard/) by submitting a PR to [veda-config](https://github.com/NASA-IMPACT/veda-config) ([see veda-docs/contributing/dashboard-configuration](https://nasa-impact.github.io/veda-docs/contributing/dashboard-configuration/dataset-configuration.html)) and add jupyterhub hosted usage examples to [veda-docs/contributing/docs-and-notebooks](https://nasa-impact.github.io/veda-docs/contributing/docs-and-notebooks.html)
51
+
* Share your data in the VEDA Dashboard [User Guide > Content Curation > Dashboard Configuration](https://docs.openveda.cloud/user-guide/content-curation/dashboard-configuration/)
52
+
53
+
* Add JupyterHub hosted usage examples to docs.openveda.cloud [User Guide > Content Curation > Usage Example Notebook Submission](https://docs.openveda.cloud/user-guide/content-curation/docs-and-notebooks.html)
52
54
53
55
## Project ingestion data structure
54
56
@@ -59,7 +61,7 @@ Various ingestion workflows are supported and documented below but only the conf
59
61
60
62
### `<stage>/collections/`
61
63
62
-
The `ingestion-data/collections/` directory holds json files representing the data for VEDA collection metadata (STAC). STAC Collection metadata can be generated from an id, title, description using Pystac. See this [veda-docs/contributing notebook example](https://nasa-impact.github.io/veda-docs/notebooks/veda-operations/stac-collection-creation.html) to get started.
64
+
The `ingestion-data/collections/` directory holds JSON files representing the data for VEDA collection metadata (STAC). STAC Collection metadata can be generated from an `id`, `title`, and `description` using PySTAC. See [User Guide > Content Curation > Dataset Ingestion > STAC Collection Creation](https://docs.openveda.cloud/user-guide/notebooks/veda-operations/stac-collection-creation.html) to get started.
63
65
64
66
Should follow the following format:
65
67
@@ -147,7 +149,7 @@ Should follow the following format:
147
149
148
150
### `<stage>/discovery-items/`
149
151
150
-
The `ingestion-data/discovery-items/` directory holds json files representing the inputs for initiating the discovery, ingest and publication workflows.
152
+
The `ingestion-data/discovery-items/` directory holds JSON files representing the inputs for initiating the discovery, ingest and publication workflows.
151
153
Can either be a single input event or a list of input events.
152
154
153
155
Should follow the following format:
@@ -174,7 +176,7 @@ Should follow the following format:
174
176
175
177
### `<stage>/dataset-config/`
176
178
177
-
The `ingestion-data/dataset-config/` directory holds json files that can be used with the `dataset/publish` workflows endpoint, combining both collection metadata and discovery items. For an example of this ingestion workflow, see this [geoglam ingest notebook in nasa-impact.github.io/veda-docs/contributing/dataset-ingeston](https://nasa-impact.github.io/veda-docs/contributing/dataset-ingestion/example-template/example-geoglam-ingest.html).
179
+
The `ingestion-data/dataset-config/` directory holds JSON files that can be used with the `dataset/publish` workflows endpoint, combining both collection metadata and discovery items. For an example of this ingestion workflow, see this [geoglam ingest notebook in the Content Curation section of the VEDA User Guide](https://docs.openveda.cloud/user-guide/content-curation/dataset-ingestion/example-template/example-geoglam-ingest.html).
@@ -242,7 +244,7 @@ Should follow the following format:
242
244
243
245
### `production/promotion-config`
244
246
245
-
This directory contains the configuration needed to execute a stand-alone airflow DAG that transfers assets to production and generates production metadata. The promotion-config json uses the same schema and values from the [staging/dataset-config JSON](#stagedataset-config) with an additional `transfer` field which should be set to true when S3 objects need to be transferred from a staging location to the production data store. The veda data promotion pipeline copies data from a specified staging bucket and prefix to a permanent location in `s3://veda-data-store` using the collection_id as a prefix and publishes STAC metadata to the produciton catalog.
247
+
This directory contains the configuration needed to execute a stand-alone airflow DAG that transfers assets to production and generates production metadata. The promotion-config JSON uses the same schema and values from the [staging/dataset-config JSON](#stagedataset-config) with an additional `transfer` field which should be set to true when S3 objects need to be transferred from a staging location to the production data store. The veda data promotion pipeline copies data from a specified staging bucket and prefix to a permanent location in `s3://veda-data-store` using the collection_id as a prefix and publishes STAC metadata to the produciton catalog.
0 commit comments