Skip to content

Commit 48ea61b

Browse files
lappemicPolina Kazakovalhoestq
authored
add config_name and data_files to datasetcard.md specs (#1292)
* add config_name and data_files to datasetcard.md specs * Apply suggesions Co-authored-by: Polina Kazakova <[email protected]> * Apply suggestion Co-authored-by: Quentin Lhoest <[email protected]> * Remove #TODO * Update datasetcard.md Co-authored-by: Polina Kazakova <[email protected]> --------- Co-authored-by: Polina Kazakova <[email protected]> Co-authored-by: Quentin Lhoest <[email protected]>
1 parent 52a3ab2 commit 48ea61b

File tree

1 file changed

+11
-3
lines changed

1 file changed

+11
-3
lines changed

datasetcard.md

Lines changed: 11 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -33,9 +33,17 @@ task_ids:
3333
- {subtask_0} # Example: extractive-qa
3434
- {subtask_1} # Example: multi-class-image-classification
3535
paperswithcode_id: {paperswithcode_id} # Dataset id on PapersWithCode (from the URL). Example for SQuAD: squad
36-
configs: # Optional for datasets with multiple configurations like glue.
37-
- {config_0} # Example for glue: sst2
38-
- {config_1} # Example for glue: cola
36+
configs: # Optional. This can be used to pass additional parameters to the dataset loader, such as `data_files`, `data_dir`, and any builder-specific parameters
37+
- config_name: {config_name_0} # Example: default
38+
data_files:
39+
- split: {split_name_0} # Example: train
40+
path: {file_path_0} # Example: data.csv
41+
- split: {split_name_1} # Example: test
42+
path: {file_path_1} # Example: holdout.csv
43+
- config_name: {config_name_1} # Example: processed
44+
data_files:
45+
- split: {split_name_3} # Example: train
46+
path: {file_path_3} # Example: data_processed.csv
3947

4048
# Optional. This part can be used to store the feature types and size of the dataset to be used in python. This can be automatically generated using the datasets-cli.
4149
dataset_info:

0 commit comments

Comments
 (0)