Skip to content

Commit 85b8cac

Browse files
author
sanchit-gandhi
committed
more renaming
1 parent cd26d4f commit 85b8cac

File tree

4 files changed

+8
-8
lines changed

4 files changed

+8
-8
lines changed

README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -27,21 +27,21 @@ In the proceeding example, we follow Stability's approach by taking audio embedd
2727
model, and training the linear classifier on a combination of three open-source datasets:
2828
1. The English Accented (`en_accented`) subset of [Voxpopuli](https://huggingface.co/datasets/facebook/voxpopuli)
2929
2. The train split of [VCTK](https://huggingface.co/datasets/vctk)
30-
3. The dev split of [EdAcc](https://huggingface.co/datasets/sanchit-gandhi/edacc)
30+
3. The dev split of [EdAcc](https://huggingface.co/datasets/edinburghcstr/edacc)
3131

32-
The model is subsequently evaluated on the test split of [EdAcc](https://huggingface.co/datasets/sanchit-gandhi/edacc)
32+
The model is subsequently evaluated on the test split of [EdAcc](https://huggingface.co/datasets/edinburghcstr/edacc)
3333
to give the final classification accuracy.
3434

3535
```bash
3636
#!/usr/bin/env bash
3737

3838
python run_audio_classification.py \
3939
--model_name_or_path "facebook/mms-lid-126" \
40-
--train_dataset_name "vctk+facebook/voxpopuli+sanchit-gandhi/edacc" \
40+
--train_dataset_name "vctk+facebook/voxpopuli+edinburghcstr/edacc" \
4141
--train_dataset_config_name "main+en_accented+default" \
4242
--train_split_name "train+test+validation" \
4343
--train_label_column_name "accent+accent+accent" \
44-
--eval_dataset_name "sanchit-gandhi/edacc" \
44+
--eval_dataset_name "edinburghcstr/edacc" \
4545
--eval_dataset_config_name "default" \
4646
--eval_split_name "test" \
4747
--eval_label_column_name "accent" \

dataset_concatenation_scripts/run_dataset_concatenation.sh

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
#!/usr/bin/env bash
22

33
python run_dataset_concatenation.py \
4-
--dataset_name "sanchit-gandhi/vctk+facebook/voxpopuli+sanchit-gandhi/edacc-normalized" \
4+
--dataset_name "sanchit-gandhi/vctk+facebook/voxpopuli+edinburghcstr/edacc-normalized" \
55
--dataset_config_name "default+en_accented+default" \
66
--dataset_split_name "train+test+validation" \
77
--label_column_name "accent+accent+accent" \
@@ -11,7 +11,7 @@ python run_dataset_concatenation.py \
1111
--output_dir "./concatenated-dataset"
1212

1313
python run_dataset_concatenation.py \
14-
--dataset_name "sanchit-gandhi/edacc-normalized" \
14+
--dataset_name "edinburghcstr/edacc-normalized" \
1515
--dataset_config_name "default" \
1616
--dataset_split_name "test" \
1717
--label_column_name "accent" \

edacc/prepare_edacc.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ def main():
7373
"How would you describe your accent in English? (e.g. Italian, Glaswegian)"
7474
]
7575

76-
accent_dataset = load_dataset("sanchit-gandhi/edacc_accents", split="train")
76+
accent_dataset = load_dataset("edinburghcstr/edacc_accents", split="train")
7777

7878
def format_dataset(batch):
7979
batch["speaker_id"] = (

edacc/run_edacc.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,5 +3,5 @@
33
python prepare_edacc.py \
44
--dataset_dir "/fsx/sanchit/edacc/edacc_v1.0" \
55
--output_dir "/fsx/sanchit/edacc_processed" \
6-
--hub_dataset_id "sanchit-gandhi/edacc-normalized" \
6+
--hub_dataset_id "edinburghcstr/edacc-normalized" \
77
--push_to_hub

0 commit comments

Comments
 (0)