Skip to content

Commit a87db05

Browse files
committed
Add quick inde
1 parent 0e5f273 commit a87db05

File tree

2 files changed

+33
-27
lines changed

2 files changed

+33
-27
lines changed

README.md

Lines changed: 32 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,5 @@
11
# Parler-TTS
22

3-
[[Paper we reproduce]](https://arxiv.org/abs/2402.01912)
4-
[[Models]](https://huggingface.co/parler-tts)
5-
[[Training Code]](training)
6-
[[Interactive Demo]](https://huggingface.co/spaces/parler-tts/parler_tts_mini)
7-
83
> [!IMPORTANT]
94
> We're proud to release Parler-TTS v0.1, our first 300M parameter model, trained on 10.5K hours of audio data.
105
> In the coming weeks, we'll be working on scaling up to 50k hours of data, in preparation for the v1 model.
@@ -15,6 +10,15 @@ Contrarily to other TTS models, Parler-TTS is a **fully open-source** release. A
1510

1611
This repository contains the inference and training code for Parler-TTS. It is designed to accompany the [Data-Speech](https://github.com/huggingface/dataspeech) repository for dataset annotation.
1712

13+
14+
## 📖 Quick Index
15+
* [Installation](#installation)
16+
* [Usage](#usage)
17+
* [Training](#training)
18+
* [Demo](https://huggingface.co/spaces/parler-tts/parler_tts_mini)
19+
* [Model weights and datasets](https://huggingface.co/parler-tts)
20+
21+
1822
## Usage
1923

2024
> [!TIP]
@@ -44,7 +48,7 @@ audio_arr = generation.cpu().numpy().squeeze()
4448
sf.write("parler_tts_out.wav", audio_arr, model.config.sampling_rate)
4549
```
4650

47-
## Installation steps
51+
## Installation
4852

4953
Parler-TTS has light-weight dependencies and can be installed in one line:
5054

@@ -66,26 +70,6 @@ Special thanks to:
6670
- Descript for the [DAC codec model](https://github.com/descriptinc/descript-audio-codec)
6771
- Hugging Face 🤗 for providing compute resources and time to explore!
6872

69-
## Contribution
70-
71-
Contributions are welcome, as the project offers many possibilities for improvement and exploration.
72-
73-
Namely, we're looking at ways to improve both quality and speed:
74-
- Datasets:
75-
- Train on more data
76-
- Add more features such as accents
77-
- Training:
78-
- Add PEFT compatibility to do Lora fine-tuning.
79-
- Add possibility to train without description column.
80-
- Add notebook training.
81-
- Explore multilingual training.
82-
- Explore mono-speaker finetuning.
83-
- Explore more architectures.
84-
- Optimization:
85-
- Compilation and static cache
86-
- Support to FA2 and SDPA
87-
- Evaluation:
88-
- Add more evaluation metrics
8973

9074
## Citation
9175

@@ -112,3 +96,25 @@ If you found this repository useful, please consider citing this work and also t
11296
primaryClass={cs.SD}
11397
}
11498
```
99+
100+
## Contribution
101+
102+
Contributions are welcome, as the project offers many possibilities for improvement and exploration.
103+
104+
Namely, we're looking at ways to improve both quality and speed:
105+
- Datasets:
106+
- Train on more data
107+
- Add more features such as accents
108+
- Training:
109+
- Add PEFT compatibility to do Lora fine-tuning.
110+
- Add possibility to train without description column.
111+
- Add notebook training.
112+
- Explore multilingual training.
113+
- Explore mono-speaker finetuning.
114+
- Explore more architectures.
115+
- Optimization:
116+
- Compilation and static cache
117+
- Support to FA2 and SDPA
118+
- Evaluation:
119+
- Add more evaluation metrics
120+

training/TRAINING.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -207,5 +207,5 @@ Thus, the script generalises to any number of training datasets.
207207

208208

209209
> [!IMPORTANT]
210-
> Starting training a new model from scratch can easily be overwhelming, here how the training of v0.01 looked like: [logs](https://api.wandb.ai/links/ylacombe/ea449l81)
210+
> Starting training a new model from scratch can easily be overwhelming,so here's what training looked like for v0.1: [logs](https://api.wandb.ai/links/ylacombe/ea449l81)
211211

0 commit comments

Comments
 (0)