constellationnet/train.py

73 lines
2.0 KiB
Python
Raw Normal View History

2019-12-13 20:17:57 +00:00
import constellation
from constellation import util
2019-12-13 17:11:09 +00:00
import torch
# Number of symbols to learn
order = 4
# Number of training examples in an epoch
epoch_size = 10000
# Number of epochs
2019-12-13 20:17:57 +00:00
num_epochs = 20000
2019-12-13 17:11:09 +00:00
# Number of epochs to skip between every loss report
2019-12-13 20:17:57 +00:00
loss_report_epoch_skip = 500
2019-12-13 17:11:09 +00:00
2019-12-13 20:17:57 +00:00
# File in which the trained model is saved
2019-12-13 22:10:40 +00:00
output_file = 'output/constellation-order-{}.tc'.format(order)
2019-12-13 20:17:57 +00:00
2019-12-13 22:10:40 +00:00
model = constellation.ConstellationNet(order=order)
# Train the model with random data
model.train()
2019-12-13 20:17:57 +00:00
print('Starting training with {} epochs\n'.format(num_epochs))
2019-12-13 17:11:09 +00:00
criterion = torch.nn.CrossEntropyLoss()
optimizer = torch.optim.Adam(model.parameters())
running_loss = 0
for epoch in range(num_epochs):
2019-12-13 20:17:57 +00:00
classes_dataset = util.get_random_messages(epoch_size, order)
onehot_dataset = util.messages_to_onehot(classes_dataset, order)
2019-12-13 17:11:09 +00:00
optimizer.zero_grad()
predictions = model(onehot_dataset)
loss = criterion(predictions, classes_dataset)
loss.backward()
optimizer.step()
# Report loss
running_loss += loss.item()
if epoch % loss_report_epoch_skip == loss_report_epoch_skip - 1:
print('Epoch {}/{}'.format(epoch + 1, num_epochs))
print('Loss is {}'.format(running_loss))
running_loss = 0
2019-12-13 20:17:57 +00:00
print('\nFinished training\n')
# Print some examples of reconstruction
2019-12-13 22:10:40 +00:00
model.eval()
print('Reconstruction examples:')
print('Input vector\t\t\tOutput vector after softmax')
2019-12-13 20:17:57 +00:00
2019-12-13 22:10:40 +00:00
with torch.no_grad():
onehot_example = util.messages_to_onehot(torch.arange(0, order), order)
2019-12-13 20:17:57 +00:00
raw_output = model(onehot_example)
raw_output.required_grad = False
reconstructed_example = torch.nn.functional.softmax(raw_output, dim=1)
2019-12-13 22:10:40 +00:00
for index in range(order):
2019-12-13 20:17:57 +00:00
print('{}\t\t{}'.format(
2019-12-13 22:10:40 +00:00
onehot_example[index].tolist(),
2019-12-13 20:17:57 +00:00
'[{}]'.format(', '.join(
'{:.5f}'.format(x)
2019-12-13 22:10:40 +00:00
for x in reconstructed_example[index].tolist()
2019-12-13 20:17:57 +00:00
))
))
2019-12-13 17:11:09 +00:00
2019-12-13 20:17:57 +00:00
print('\nSaving model as {}'.format(output_file))
torch.save(model.state_dict(), output_file)