-
Notifications
You must be signed in to change notification settings - Fork 0
/
train_rnn.py
228 lines (193 loc) · 7.31 KB
/
train_rnn.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
#!/usr/bin/env python
from keras.models import load_model
from keras.callbacks import ModelCheckpoint, EarlyStopping, History
from callback import *
from sklearn.model_selection import train_test_split
from rnn import preprocess_text, build_model, vectorize_data
from frequent_flush import Frequent_flush
from model_file import ModelFile
import numpy as np
import sys
import os
import getopt
import time
TRAIN_FILE = 'quotes/author-quote.txt'
MODEL_DIR = ''
NUM_EPOCHS = 30
QUICK_MODE = False
LEARNING_RATE = .001
DECAY = .5
LAYER_SIZE = 256
NUM_LAYERS = 2
DROPOUT = 0.2
SEQ_LEN = 90
BUFFER_OUTPUT = False
# Dealing with runtime options
def main(argv):
training_file, new_model, model_dir, num_epochs, quick_mode, \
learning_rate, decay, layer_size, num_layers, dropout, seq_len, \
buffer_output = parse_options(argv)
if(not buffer_output):
Frequent_flush(1).start()
print("Using Input File: ", training_file)
print("Preprocessing...")
char_to_idx, idx_to_char, vocab, text, sentences, next_chars = \
preprocess_text(filename=training_file, SEQ_LEN=SEQ_LEN)
X, y = vectorize_data(SEQ_LEN, vocab, sentences, char_to_idx, next_chars)
if(quick_mode):
X = X[:1000]
y = y[:1000]
print("Using SEQ_LEN of", SEQ_LEN)
print("Vocab size:", vocab)
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2)
# Building a simple LSTM
if new_model:
print("Building model")
print("Size of Layers: ", layer_size)
print("Depth of Network: ", num_layers)
print("Using dropout: ", dropout)
model = build_model(num_layers=num_layers, seq_len=seq_len, vocab=vocab,
layer_size=layer_size, dropout=dropout, lr=LEARNING_RATE,
decay=DECAY)
model.summary()
model_dir = create_model_dir(model)
print("Saved model data to:", os.path.abspath(model_dir))
history = {'acc':[], 'val_acc':[], 'loss':[], 'val_loss':[]}
initial_epoch = 0
else:
model, history, initial_epoch = load_model_dir(model_dir)
model.summary()
print("Creating callbacks")
callbacks_list = create_callbacks(model_dir, writer=True,
idx_to_char=idx_to_char, char_to_idx=char_to_idx, text=text,
seq_len=SEQ_LEN, vocab=vocab)
print("\nTraining model\n")
history = train_model(model, X_train, y_train, X_test, y_test,
callbacks_list, num_epochs, history, initial_epoch)
print("Saving training history")
history_filename = os.path.join(model_dir, 'model_history.npy')
np.save(history_filename, history)
def create_model_dir(model):
if not os.path.exists('models'):
os.makedirs('models')
timestr = time.strftime("%m%d-%H%M")
model_name = "CharRNN_" + timestr
model_dir = os.path.join('models', model_name)
if not os.path.exists(model_dir):
os.makedirs(model_dir)
model.save(os.path.join(model_dir, "model.h5"))
return model_dir
def load_model_dir(model_dir):
model_file = ModelFile(model_directory=model_dir)
if(model_file.check_exists()):
print("Loading model")
else:
print("Incomplete model. Make sure that the folder exists and" +
" contains weights, history, and model.h5")
sys.exit(0)
model_dir = model_file.model_directory
model_filename = model_file.model
weights_file = model_file.weights
history_file = model_file.history
model = load_model(model_filename)
print("Using weights file ", weights_file)
model.load_weights(weights_file)
print("Using history file ", history_file)
history = np.load(history_file).item()
initial_epoch = len(history['acc'])
print("Starting from epoch ", initial_epoch)
return model,history,initial_epoch
def create_callbacks(model_dir, history=True, checkp=True, earlyStop=False,
writer=False, idx_to_char = {}, char_to_idx = {}, text = "",
seq_len = 300 , vocab = 0):
callbacks = []
weights_filepath = os.path.join(model_dir,
"weights-improvement-{epoch:03d}-{loss:.4f}.hdf5")
composition_dir = os.path.join(model_dir, 'compositions')
composition_filepath = os.path.join(composition_dir,
"epoch_{epoch:03d}_composition_reduced.txt")
checkpoint = ModelCheckpoint(weights_filepath, save_best_only=True,
verbose=1)
esCallback = EarlyStopping(min_delta=0, patience=10, verbose=1)
hisCallback = History()
writerCallback = Write_Text(idx_to_char, char_to_idx, text, seq_len, vocab,
filepath=composition_filepath)
if history:
callbacks.append(hisCallback)
if checkp:
callbacks.append(checkpoint)
if earlyStop:
callbacks.append(esCallback)
if writer:
if not os.path.exists(composition_dir):
os.makedirs(composition_dir)
callbacks.append(writerCallback)
return callbacks
def train_model(model, X_train, y_train, X_test, y_test, callbacks_list,
num_epochs, history, initial_epoch=0):
for e in range(num_epochs):
epochs = e + initial_epoch
try:
print("\nEPOCH {}\n".format(epochs))
hist = model.fit(X_train, y_train, validation_data=(X_test,y_test),
batch_size=128, epochs=epochs+1, callbacks=callbacks_list,
initial_epoch=epochs)
for k, v in hist.history.items():
history[k] = history[k] + v
except KeyboardInterrupt:
print("Exiting training loop")
break
return history
def parse_options(argv):
'''
Takes a list of strings, and returns a list of hyper parameters in the order
given below
'''
global TRAIN_FILE
global MODEL_DIR
global NUM_EPOCHS
global QUICK_MODE
global DECAY
global LEARNING_RATE
global LAYER_SIZE
global NUM_LAYERS
global DROPOUT
global SEQ_LEN
global BUFFER_OUTPUT
new_model = True
try:
opts, args = getopt.getopt(argv, 'i:e:qm:w:h:b', ['inputFile=','epochs=',
'quickmode' , 'learningRate=','decay=', 'numLayers=', 'layerSize=',
'sequenceLength=', 'historyFile=', 'bufferOutput', 'dropout=',
'modelDirectory'])
for opt, arg in opts:
if opt in ('-i', '--inputFile'):
TRAIN_FILE = arg
elif opt in ('-m', '--modelDirectory'):
MODEL_DIR = arg
new_model = False
elif opt in ('-e', '--epochs'):
NUM_EPOCHS = int(arg)
elif opt in ('-q', '--quickmode'):
QUICK_MODE = True
elif opt == '--learningRate':
LEARNING_RATE = float(arg)
elif opt == '--decay':
DECAY = float(arg)
elif opt == '--layerSize':
LAYER_SIZE = int(arg)
elif opt == '--numLayers':
NUM_LAYERS = int(arg)
elif opt == '--dropout':
DROPOUT = float(arg)
elif opt == '--sequenceLength':
SEQ_LEN = int(arg)
elif opt in ('-b', '--bufferOutput'):
BUFFER_OUTPUT = True
except getopt.GetoptError as e:
print(e)
return (TRAIN_FILE, new_model, MODEL_DIR, NUM_EPOCHS, QUICK_MODE,
LEARNING_RATE, DECAY, LAYER_SIZE, NUM_LAYERS, DROPOUT,
SEQ_LEN, BUFFER_OUTPUT)
if __name__ == "__main__":
main(sys.argv[1:])