-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathtrain.py
More file actions
75 lines (62 loc) · 2.69 KB
/
train.py
File metadata and controls
75 lines (62 loc) · 2.69 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
import torch, torch.nn.functional as F, sys, time
from torch.utils.data import DataLoader, RandomSampler
from matcha_tts import MatchaTTS
from matcha_tts.hparams import Hyperparameters
from dataset import ProcessedDataset, collate_samples, TOKENS
DEVICE = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
ENABLE_AMP = True
LOG_INTERVAL = 100
CHECKPOINT_INTERVAL = 5000
if __name__ == "__main__":
dataset = ProcessedDataset("datasets/ljspeech-processed/")
dataloader = DataLoader(
dataset,
batch_size=32,
sampler=RandomSampler(dataset, replacement=True, num_samples=1_000_000_000),
pin_memory=True,
num_workers=1,
collate_fn=collate_samples,
)
hparams = Hyperparameters(num_symbols=len(TOKENS))
model = MatchaTTS(hparams).to(DEVICE)
optimizer = torch.optim.AdamW(model.parameters(), 1e-4)
scaler = torch.GradScaler(enabled=ENABLE_AMP)
batches = 0
if len(sys.argv) == 2:
checkpoint = torch.load(sys.argv[1])
model.load_state_dict(checkpoint["model"])
optimizer.load_state_dict(checkpoint["optimizer"])
scaler.load_state_dict(checkpoint["scaler"])
batches = checkpoint["batches"]
model.train()
running_start = time.time()
running_loss = 0.0
for text, text_lengths, mels, mels_lengths in dataloader:
text = text.to(DEVICE, non_blocking=True)
text_lengths = text_lengths.to(DEVICE, non_blocking=True)
mels = mels.to(DEVICE, non_blocking=True)
mels_lengths = mels_lengths.to(DEVICE, non_blocking=True)
optimizer.zero_grad()
with torch.autocast(DEVICE.type, enabled=ENABLE_AMP):
main_loss, prior_loss, duration_loss = model(text, text_lengths, mels, mels_lengths)
loss = main_loss + prior_loss + duration_loss
scaler.scale(loss).backward()
scaler.unscale_(optimizer)
torch.nn.utils.clip_grad_value_(model.parameters(), 5.0)
scaler.step(optimizer)
scaler.update()
running_loss += loss.item()
batches += 1
if batches % LOG_INTERVAL == 0:
avg_running_loss = running_loss / LOG_INTERVAL
batches_per_sec = LOG_INTERVAL / (time.time() - running_start)
print(f"[{batches}] loss: {avg_running_loss}, {batches_per_sec:.2f} batches/sec", flush=True)
running_start = time.time()
running_loss = 0.0
if batches % CHECKPOINT_INTERVAL == 0:
torch.save({
"model": model.state_dict(),
"optimizer": optimizer.state_dict(),
"scaler": scaler.state_dict(),
"batches": batches,
}, f"checkpoints/{batches:06}-model.pth")