Skip to content

Commit 070425c

Browse files
Add extra guidance for sharing large datasets on the Hub (#1288)
* docs: add extra guidance for sharing large datasets on the Hub * wording improvements * Update docs/hub/repositories-recommendations.md Co-authored-by: Omar Sanseviero <[email protected]> * Update docs/hub/repositories-recommendations.md Co-authored-by: Omar Sanseviero <[email protected]> * make all requirements mandatory * formatting --------- Co-authored-by: Omar Sanseviero <[email protected]>
1 parent efb313a commit 070425c

File tree

1 file changed

+19
-0
lines changed

1 file changed

+19
-0
lines changed

docs/hub/repositories-recommendations.md

Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -57,3 +57,22 @@ happen (in rare cases) that even if the timeout is raised client-side, the proce
5757
completed server-side. This can be checked manually by browsing the repo on the Hub. To prevent this timeout, we recommend
5858
adding around 50-100 files per commit.
5959

60+
## Sharing large datasets on the Hub
61+
62+
One key way in which Hugging Face supports the machine learning ecosystem is by hosting datasets on the Hub, including very large datasets. To ensure we can effectively support the open-source ecosystem, we ask that you let us know in advance via datasets at huggingface.co or on [our Discord](http://hf.co/join/discord) if you are uploading datasets above a couple of hundred GBs or TBs of data.
63+
64+
When you get in touch with us, please let us know:
65+
66+
- What is the dataset, and who/what is it likely to be useful for?
67+
- The size of the dataset.
68+
- The format you plan to use for sharing your dataset.
69+
70+
For hosting large datasets on the Hub, we require the following for your dataset:
71+
72+
- A dataset card: we want to ensure that your dataset can be used effectively by the community and one of the key ways of enabling this is via a dataset card. This [guidance](docs/hub/datasets-cards.md) provides an overview of how to write a dataset card.
73+
- You are sharing the dataset to enable community reuse. If you plan to upload a dataset you anticipate won't have any further reuse, other platforms are likely more suitable.
74+
- You must follow the repository limitations outlined above.
75+
- Using file formats that are well integrated with the Hugging Face ecosystem. We have good support for [Parquet](https://huggingface.co/docs/datasets/v2.19.0/en/loading#parquet) and [WebDataset](https://huggingface.co/docs/datasets/v2.19.0/en/loading#webdataset) formats, which are often good options for sharing large datasets efficiently. This will also ensure the dataset viewer works for your dataset.
76+
- Avoid the use of custom loading scripts when using datasets. In our experience, datasets that require custom code to use often end up with limited reuse.
77+
78+
Please get in touch with us if any of these requirements are difficult for you to meet because of the type of data or domain you are working in.

0 commit comments

Comments
 (0)