-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathtrain.py
More file actions
36 lines (30 loc) · 1.1 KB
/
train.py
File metadata and controls
36 lines (30 loc) · 1.1 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
import torch
import torch.nn as nn
import torch.optim as optim
from data import vocab_to_int, dataloader
from model import TransformerSentiment
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
print(f"Using device: {device}")
vocab_size = len(vocab_to_int)
model = TransformerSentiment(vocab_size, max_len=512)
model.to(device)
model_save_path = 'model.pth'
optimizer_save_path = 'optimizer.pth'
criterion = nn.BCELoss()
optimizer = optim.Adam(model.parameters(), lr=0.001)
epochs = 50
for epoch in range(epochs):
model.train()
total_loss = 0
for reviews, labels in dataloader:
optimizer.zero_grad()
src_mask = (reviews == 0)
outputs = model(reviews, src_mask).squeeze()
loss = criterion(outputs, labels)
loss.backward()
optimizer.step()
total_loss += loss.item()
print(f"Epoch {epoch+1}, Loss: {total_loss / len(dataloader)}")
torch.save(model.state_dict(), model_save_path)
torch.save(optimizer.state_dict(), optimizer_save_path)
print(f"Checkpoint saved at step {iter}.")