forked from wiseodd/controlled-text-generation
-
Notifications
You must be signed in to change notification settings - Fork 0
/
train_vae.py
109 lines (79 loc) · 2.63 KB
/
train_vae.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
import math
import os
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.autograd as autograd
import torch.optim as optim
import numpy as np
from torch.autograd import Variable
from ctextgen.dataset import *
from ctextgen.model import RNN_VAE
import argparse
parser = argparse.ArgumentParser(
description='Conditional Text Generation: Train VAE as in Bowman, 2016, with c ~ p(c)'
)
parser.add_argument('--gpu', default=False, action='store_true',
help='whether to run in the GPU')
parser.add_argument('--save', default=False, action='store_true',
help='whether to save model or not')
args = parser.parse_args()
mb_size = 32
z_dim = 20
h_dim = 64
lr = 1e-3
lr_decay_every = 1000000
n_iter = 20000
log_interval = 1000
z_dim = h_dim
c_dim = 2
dataset = SST_Dataset()
model = RNN_VAE(
dataset.n_vocab, h_dim, z_dim, c_dim, p_word_dropout=0.3,
pretrained_embeddings=dataset.get_vocab_vectors(), freeze_embeddings=False,
gpu=args.gpu
)
def main():
# Annealing for KL term
kld_start_inc = 3000
kld_weight = 0.01
kld_max = 0.15
kld_inc = (kld_max - kld_weight) / (n_iter - kld_start_inc)
trainer = optim.Adam(model.vae_params, lr=lr)
for it in range(n_iter):
inputs, labels = dataset.next_batch(args.gpu)
recon_loss, kl_loss = model.forward(inputs)
loss = recon_loss + kld_weight * kl_loss
# Anneal kl_weight
if it > kld_start_inc and kld_weight < kld_max:
kld_weight += kld_inc
loss.backward()
grad_norm = torch.nn.utils.clip_grad_norm(model.vae_params, 5)
trainer.step()
trainer.zero_grad()
if it % log_interval == 0:
z = model.sample_z_prior(1)
c = model.sample_c_prior(1)
sample_idxs = model.sample_sentence(z, c)
sample_sent = dataset.idxs2sentence(sample_idxs)
print('Iter-{}; Loss: {:.4f}; Recon: {:.4f}; KL: {:.4f}; Grad_norm: {:.4f};'
.format(it, loss.data[0], recon_loss.data[0], kl_loss.data[0], grad_norm))
print('Sample: "{}"'.format(sample_sent))
print()
# Anneal learning rate
new_lr = lr * (0.5 ** (it // lr_decay_every))
for param_group in trainer.param_groups:
param_group['lr'] = new_lr
def save_model():
if not os.path.exists('models/'):
os.makedirs('models/')
torch.save(model.state_dict(), 'models/vae.bin')
if __name__ == '__main__':
try:
main()
except KeyboardInterrupt:
if args.save:
save_model()
exit(0)
if args.save:
save_model()