Skip to content

Commit 863a56f

Browse files
anayeayebotanical
andauthored
fix(docs): update links after veda docs site re-org (#313)
* update links after veda docs site re-org * Update README.md Co-authored-by: Jennifer Tran <12633533+botanical@users.noreply.github.com> * Update README.md Co-authored-by: Jennifer Tran <12633533+botanical@users.noreply.github.com> * Update README.md Co-authored-by: Jennifer Tran <12633533+botanical@users.noreply.github.com> * Update README.md Co-authored-by: Jennifer Tran <12633533+botanical@users.noreply.github.com> * formatting and ingest ui link * upgrade black to patch version --------- Co-authored-by: Jennifer Tran <12633533+botanical@users.noreply.github.com>
1 parent 228ce94 commit 863a56f

File tree

3 files changed

+16
-14
lines changed

3 files changed

+16
-14
lines changed

.pre-commit-config.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ repos:
1212
hooks:
1313
- id: markdownlint-cli2
1414
- repo: https://github.com/psf/black-pre-commit-mirror
15-
rev: 23.9.1
15+
rev: 24.3.0
1616
hooks:
1717
- id: black
1818
- id: black-jupyter

README.md

Lines changed: 12 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -20,25 +20,25 @@ To add data to VEDA you will:
2020

2121
### Step 1: Stage your files
2222

23-
Upload files to the staging bucket `s3://veda-data-store-staging` (which you can do with a VEDA JupyterHub account--request access [here](https://docs.openveda.cloud/nasa-veda-platform/scientific-computing/#veda-sponsored-jupyterhub-service)) or a self-hosted bucket in s3 has shared read access to VEDA service. [See docs.openveda.cloud for additional details on preparing files.](https://docs.openveda.cloud/instance-management/adding-content/dataset-ingestion/file-preparation.html)
23+
Upload files to the staging bucket `s3://veda-data-store-staging` (which you can do with a VEDA JupyterHub account--request access [here](https://docs.openveda.cloud/user-guide/scientific-computing/getting-access.html)) or a self-hosted bucket in s3 has shared read access to VEDA service. [See docs.openveda.cloud for additional details on preparing files.](https://docs.openveda.cloud/user-guide/content-curation/dataset-ingestion/file-preparation.html)
2424

2525
### Step 2: Generate STAC metadata in the staging catalog
2626

27-
Metadata must first be added to the Staging Catalog [staging.openveda.cloud/api/stac](https://staging.openveda.cloud/api/stac). You will need to create a dataset config file using the veda-ingest-ui and submit it to the `/workflows/dataset/publish` endpoint to generate STAC Collection metadata and generate Item records for the files you have uploaded in Step 1.
27+
Metadata must first be added to the Staging Catalog [staging.openveda.cloud/api/stac](https://staging.openveda.cloud/api/stac). You will need to create a dataset config file with the [ingest.openveda.cloud](https://ingest.openveda.cloud) form to generate STAC Collection metadata and generate Item records for the files you have staged in Step 1.
2828

29-
* Use the veda-ingest-ui form to generate a dataset config and open a veda-data PR
29+
* Use the [ingest.openveda.cloud](https://ingest.openveda.cloud) form to generate a dataset config and open a veda-data PR
3030

3131
* OR manually generate a dataset-config JSON and open a veda-data PR
3232

3333
* When a veda-data PR is opened, a github action will automatically (1) POST the config to airflow and stage the collection and items in the staging catalog instance and (2) open a veda-config dashboard preview for the dataset.
3434

35-
> See detailed steps for the [dataset submission process](https://docs.openveda.cloud/instance-management/adding-content/dataset-ingestion/) in the contribuing section of [veda-docs](https://nasa-impact.github.io/veda-docs) where you can also find this full ingestion workflow example [geoglam ingest notebook](https://docs.openveda.cloud/instance-management/adding-content/dataset-ingestion/example-template/example-geoglam-ingest.html)
35+
> See detailed steps for the [dataset submission process](https://docs.openveda.cloud/user-guide/content-curation/dataset-ingestion/) in the Content Curation section of the User Guide in [veda-docs](https://nasa-impact.github.io/veda-docs) where you can also find this full ingestion workflow example [geoglam ingest notebook](https://docs.openveda.cloud/user-guide/content-curation/dataset-ingestion/example-template/example-geoglam-ingest.html)
3636
3737
### Step 3: Acceptance testing
3838

3939
Perform acceptance testing appropriate for your data. This should include reviewing the [staging.openveda.cloud STAC browser](https://staging.openveda.cloud) and reviewing the corresponding veda-config PR dashboard preview.
4040

41-
> See [veda-docs/instance-management/adding-content/dashboard-configuration](https://docs.openveda.cloud/instance-management/adding-content/dashboard-configuration/dataset-configuration.html) for more information about configuring a dashboard preview).
41+
> See the [Content Curation/Dashboard Configuration](https://docs.openveda.cloud/user-guide/content-curation/dashboard-configuration/) section of the User Guide in docs.openveda.cloud for more information about configuring a dashboard preview.
4242
4343
### Step 4: Promote to production
4444

@@ -48,7 +48,9 @@ After acceptance testing, request approval--when your PR is merged, the dataset
4848
4949
### Step 5 [Optional]: Share your data
5050

51-
Share your data in the [VEDA Dashboard](https://www.earthdata.nasa.gov/dashboard/) by submitting a PR to [veda-config](https://github.com/NASA-IMPACT/veda-config) ([see veda-docs/contributing/dashboard-configuration](https://nasa-impact.github.io/veda-docs/contributing/dashboard-configuration/dataset-configuration.html)) and add jupyterhub hosted usage examples to [veda-docs/contributing/docs-and-notebooks](https://nasa-impact.github.io/veda-docs/contributing/docs-and-notebooks.html)
51+
* Share your data in the VEDA Dashboard [User Guide > Content Curation > Dashboard Configuration](https://docs.openveda.cloud/user-guide/content-curation/dashboard-configuration/)
52+
53+
* Add JupyterHub hosted usage examples to docs.openveda.cloud [User Guide > Content Curation > Usage Example Notebook Submission](https://docs.openveda.cloud/user-guide/content-curation/docs-and-notebooks.html)
5254

5355
## Project ingestion data structure
5456

@@ -59,7 +61,7 @@ Various ingestion workflows are supported and documented below but only the conf
5961

6062
### `<stage>/collections/`
6163

62-
The `ingestion-data/collections/` directory holds json files representing the data for VEDA collection metadata (STAC). STAC Collection metadata can be generated from an id, title, description using Pystac. See this [veda-docs/contributing notebook example](https://nasa-impact.github.io/veda-docs/notebooks/veda-operations/stac-collection-creation.html) to get started.
64+
The `ingestion-data/collections/` directory holds JSON files representing the data for VEDA collection metadata (STAC). STAC Collection metadata can be generated from an `id`, `title`, and `description` using PySTAC. See [User Guide > Content Curation > Dataset Ingestion > STAC Collection Creation](https://docs.openveda.cloud/user-guide/notebooks/veda-operations/stac-collection-creation.html) to get started.
6365

6466
Should follow the following format:
6567

@@ -147,7 +149,7 @@ Should follow the following format:
147149

148150
### `<stage>/discovery-items/`
149151

150-
The `ingestion-data/discovery-items/` directory holds json files representing the inputs for initiating the discovery, ingest and publication workflows.
152+
The `ingestion-data/discovery-items/` directory holds JSON files representing the inputs for initiating the discovery, ingest and publication workflows.
151153
Can either be a single input event or a list of input events.
152154

153155
Should follow the following format:
@@ -174,7 +176,7 @@ Should follow the following format:
174176

175177
### `<stage>/dataset-config/`
176178

177-
The `ingestion-data/dataset-config/` directory holds json files that can be used with the `dataset/publish` workflows endpoint, combining both collection metadata and discovery items. For an example of this ingestion workflow, see this [geoglam ingest notebook in nasa-impact.github.io/veda-docs/contributing/dataset-ingeston](https://nasa-impact.github.io/veda-docs/contributing/dataset-ingestion/example-template/example-geoglam-ingest.html).
179+
The `ingestion-data/dataset-config/` directory holds JSON files that can be used with the `dataset/publish` workflows endpoint, combining both collection metadata and discovery items. For an example of this ingestion workflow, see this [geoglam ingest notebook in the Content Curation section of the VEDA User Guide](https://docs.openveda.cloud/user-guide/content-curation/dataset-ingestion/example-template/example-geoglam-ingest.html).
178180

179181
<details>
180182
<summary><b>/dataset-config/collection_id.json</b></summary>
@@ -242,7 +244,7 @@ Should follow the following format:
242244

243245
### `production/promotion-config`
244246

245-
This directory contains the configuration needed to execute a stand-alone airflow DAG that transfers assets to production and generates production metadata. The promotion-config json uses the same schema and values from the [staging/dataset-config JSON](#stagedataset-config) with an additional `transfer` field which should be set to true when S3 objects need to be transferred from a staging location to the production data store. The veda data promotion pipeline copies data from a specified staging bucket and prefix to a permanent location in `s3://veda-data-store` using the collection_id as a prefix and publishes STAC metadata to the produciton catalog.
247+
This directory contains the configuration needed to execute a stand-alone airflow DAG that transfers assets to production and generates production metadata. The promotion-config JSON uses the same schema and values from the [staging/dataset-config JSON](#stagedataset-config) with an additional `transfer` field which should be set to true when S3 objects need to be transferred from a staging location to the production data store. The veda data promotion pipeline copies data from a specified staging bucket and prefix to a permanent location in `s3://veda-data-store` using the collection_id as a prefix and publishes STAC metadata to the produciton catalog.
246248

247249
<details>
248250
<summary><b>production/promotion-config/collection_id.json</b></summary>

transformation-scripts/collection-and-item-ingest.ipynb

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -130,9 +130,9 @@
130130
"client = CognitoClient(\n",
131131
" client_id=test_client_id if testing_mode else mcp_prod_client_id,\n",
132132
" user_pool_id=test_user_pool_id if testing_mode else mcp_prod_user_pool_id,\n",
133-
" identity_pool_id=test_identity_pool_id\n",
134-
" if testing_mode\n",
135-
" else mcp_prod_identity_pool_id,\n",
133+
" identity_pool_id=(\n",
134+
" test_identity_pool_id if testing_mode else mcp_prod_identity_pool_id\n",
135+
" ),\n",
136136
")\n",
137137
"_ = client.login()"
138138
]

0 commit comments

Comments
 (0)