forked from DavidBert/ModIA_TP1
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathtrain_mnist.py
More file actions
99 lines (83 loc) · 3.46 KB
/
train_mnist.py
File metadata and controls
99 lines (83 loc) · 3.46 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
import argparse
from statistics import mean
import torch
import torchvision
import torchvision.transforms as transforms
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
from torch.utils.tensorboard import SummaryWriter
from tqdm import tqdm
from mnist_net import MNISTNet
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
def train(net, optimizer, loader, writer, epochs=10):
criterion = nn.CrossEntropyLoss()
for epoch in range(epochs):
running_loss = []
t = tqdm(loader)
for x, y in t:
x, y = x.to(device), y.to(device)
outputs = net(x)
loss = criterion(outputs, y)
running_loss.append(loss.item())
optimizer.zero_grad()
loss.backward()
optimizer.step()
t.set_description(f'training loss: {mean(running_loss)}')
writer.add_scalar('training loss', mean(running_loss), epoch)
def test(model, dataloader):
test_corrects = 0
total = 0
with torch.no_grad():
for x, y in dataloader:
x = x.to(device)
y = y.to(device)
y_hat = model(x).argmax(1)
test_corrects += y_hat.eq(y).sum().item()
total += y.size(0)
return test_corrects / total
if __name__=='__main__':
writer = SummaryWriter(f'runs/MNIST')
parser = argparse.ArgumentParser()
parser.add_argument('--exp_name', type=str, default = 'MNIST', help='experiment name')
parser.add_argument('--batch_size', type=int, default=1, help='batch size')
parser.add_argument('--lr', type=float, default=0.001, help='learning rate')
parser.add_argument('--epochs', type=int, default=1, help='training epochs')
args = parser.parse_args()
exp_name = args.exp_name
epochs = args.epochs
batch_size = args.batch_size
lr = args.lr
# transforms
transform = transforms.Compose(
[transforms.ToTensor(),
transforms.Normalize((0.5,), (0.5,))])
# datasets
trainset = torchvision.datasets.MNIST('./data', download=True, train=True, transform=transform)
testset = torchvision.datasets.MNIST('./data', download=True, train=False, transform=transform)
# dataloaders
trainloader = torch.utils.data.DataLoader(trainset, batch_size=batch_size, shuffle=True, num_workers=2)
testloader = torch.utils.data.DataLoader(testset, batch_size=batch_size, shuffle=False, num_workers=2)
net = MNISTNet()
# setting net on device(GPU if available, else CPU)
net = net.to(device)
#optimizer = optim.SGD(net.parameters(), lr=lr, momentum=0.9)
optimizer = optim.Adam(net.parameters(), lr=lr)
train(net, optimizer, trainloader, writer, epochs)
test_acc = test(net, testloader)
print(f'Test accuracy:{test_acc}')
torch.save(net.state_dict(), "mnist_net.pth")
# add embeddings to tensorboard
perm = torch.randperm(len(trainset.data))
images, labels = trainset.data[perm][:256], trainset.targets[perm][:256]
images = images.unsqueeze(1).float().to(device)
with torch.no_grad():
embeddings = net.get_features(images)
writer.add_embedding(embeddings,
metadata=labels,
label_img=images, global_step=1)
# save networks computational graph in tensorboard
writer.add_graph(net, images)
# save a dataset sample in tensorboard
img_grid = torchvision.utils.make_grid(images[:64])
writer.add_image('mnist_images', img_grid)