-
Notifications
You must be signed in to change notification settings - Fork 12
/
behavior_model_cnn_attention.py
340 lines (298 loc) · 12.6 KB
/
behavior_model_cnn_attention.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
"""
Created on Wed Mar 15 09:12:22 2017
@author: aitor
"""
import json
import sys
from gensim.models import Word2Vec
import h5py
from keras.callbacks import ModelCheckpoint
from keras.layers import Dot, Bidirectional, Concatenate, Convolution2D, Dense, Dropout, Embedding, Flatten, GRU, Input, Lambda, MaxPooling2D, Multiply, Reshape
from keras.models import load_model, Model
from keras.preprocessing.text import Tokenizer
import matplotlib
matplotlib.use('Agg')
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
# Kasteren dataset
DIR = './sensor2vec/kasteren_dataset/'
# Dataset with vectors but without the action timestamps
DATASET_CSV = DIR + 'base_kasteren_reduced.csv'
DATASET_NO_TIME = DIR + 'dataset_no_time.json'
# dataset with actions transformed with time periods
DATASET_ACTION_PERIODS = DIR + 'kasteren_action_periods.csv'
# List of unique activities in the dataset
UNIQUE_ACTIVITIES = DIR + 'unique_activities.json'
# List of unique actions in the dataset
UNIQUE_ACTIONS = DIR + 'unique_actions.json'
# List of unique actions in the dataset taking into account time periods
UNIQUE_TIME_ACTIONS = DIR + 'unique_time_actions.json'
# Action vectors
#ACTION_VECTORS = DIR + 'actions_vectors.json'
# Word2Vec model
WORD2VEC_MODEL = DIR + 'actions.model'
# Word2Vec model taking into account time periods
WORD2VEC_TIME_MODEL = DIR + 'actions_time.model'
#number of input actions for the model
INPUT_ACTIONS = 5
#Number of elements in the action's embbeding vector
ACTION_EMBEDDING_LENGTH = 50
#best model in the training
BEST_MODEL = 'best_model.hdf5'
# if time is being taken into account
TIME = False
"""
Load the best model saved in the checkpoint callback
"""
def select_best_model():
model = load_model(BEST_MODEL)
return model
"""
Function used to visualize the training history
metrics: Visualized metrics,
save: if the png are saved to disk
history: training history to be visualized
"""
def plot_training_info(metrics, save, history):
# summarize history for accuracy
if 'accuracy' in metrics:
plt.plot(history['acc'])
plt.plot(history['val_acc'])
plt.title('model accuracy')
plt.ylabel('accuracy')
plt.xlabel('epoch')
plt.legend(['train', 'test'], loc='upper left')
if save == True:
plt.savefig('accuracy.png')
plt.gcf().clear()
else:
plt.show()
# summarize history for loss
if 'loss' in metrics:
plt.plot(history['loss'])
plt.plot(history['val_loss'])
plt.title('model loss')
plt.ylabel('loss')
plt.xlabel('epoch')
#plt.ylim(1e-3, 1e-2)
plt.yscale("log")
plt.legend(['train', 'test'], loc='upper left')
if save == True:
plt.savefig('loss.png')
plt.gcf().clear()
else:
plt.show()
"""
Prepares the training examples of secuences based on the total actions, using
embeddings to represent them.
Input
df:Pandas DataFrame with timestamp, sensor, action, event and activity
unique_actions: list of actions
Output:
X: array with action index sequences
y: array with action index for next action
tokenizer: instance of Tokenizer class used for action/index convertion
"""
def prepare_x_y(df, unique_actions):
#recover all the actions in order.
actions = df['action'].values
# print actions.tolist()
# print actions.tolist().index('HallBedroomDoor_1')
# Use tokenizer to generate indices for every action
# Very important to put lower=False, since the Word2Vec model
# has the action names with some capital letters
tokenizer = Tokenizer(lower=False)
tokenizer.fit_on_texts(actions.tolist())
action_index = tokenizer.word_index
# print action_index
#translate actions to indexes
actions_by_index = []
print len(actions)
for action in actions:
# print action
actions_by_index.append(action_index[action])
#Create the trainning sets of sequences with a lenght of INPUT_ACTIONS
last_action = len(actions) - 1
X = []
y = []
for i in range(last_action-INPUT_ACTIONS):
X.append(actions_by_index[i:i+INPUT_ACTIONS])
#represent the target action as a onehot for the softmax
target_action = ''.join(i for i in actions[i+INPUT_ACTIONS] if not i.isdigit()) # remove the period if it exists
target_action_onehot = np.zeros(len(unique_actions))
target_action_onehot[unique_actions.index(target_action)] = 1.0
y.append(target_action_onehot)
return X, y, tokenizer
"""
Prepares the training examples of secuences based on the total actions, using
one hot vectors to represent them
Input
df:Pandas DataFrame with timestamp, sensor, action, event and activity
unique_actions: list of actions
Output:
X: array with action index sequences
y: array with action index for next action
"""
def prepare_x_y_onehot(df, unique_actions):
#recover all the actions in order.
actions = df['action'].values
#translate actions to onehots
actions_by_onehot = []
for action in actions:
onehot = [0] * len(unique_actions)
action_index = unique_actions.index(action)
onehot[action_index] = 1
actions_by_onehot.append(onehot)
#Create the trainning sets of sequences with a lenght of INPUT_ACTIONS
last_action = len(actions) - 1
X = []
y = []
for i in range(last_action-INPUT_ACTIONS):
X.append(actions_by_onehot[i:i+INPUT_ACTIONS])
#represent the target action as a onehot for the softmax
target_action = actions_by_onehot[i+INPUT_ACTIONS]
y.append(target_action)
return X, y
"""
Function to create the embedding matrix, which will be used to initialize
the embedding layer of the network
Input:
tokenizer: instance of Tokenizer class used for action/index convertion
Output:
embedding_matrix: matrix with the embedding vectors for each action
"""
def create_embedding_matrix(tokenizer):
if TIME:
model = Word2Vec.load(WORD2VEC_TIME_MODEL)
else:
model = Word2Vec.load(WORD2VEC_MODEL)
action_index = tokenizer.word_index
embedding_matrix = np.zeros((len(action_index) + 1, ACTION_EMBEDDING_LENGTH))
unknown_words = {}
for action, i in action_index.items():
try:
embedding_vector = model[action]
embedding_matrix[i] = embedding_vector
except:
if action in unknown_words:
unknown_words[action] += 1
else:
unknown_words[action] = 1
print "Number of unknown tokens: " + str(len(unknown_words))
print unknown_words
return embedding_matrix
def main(argv):
print '*' * 20
print 'Loading dataset...'
sys.stdout.flush()
#dataset of activities
if TIME:
DATASET = DATASET_ACTION_PERIODS
else:
DATASET = DATASET_CSV
df_dataset = pd.read_csv(DATASET, parse_dates=[[0, 1]], header=None, index_col=0, sep=' ')
df_dataset.columns = ['sensor', 'action', 'event', 'activity']
df_dataset.index.names = ["timestamp"]
# we only need the actions without the period to calculate the onehot vector for y, because we are only predicting the actions
unique_actions = json.load(open(UNIQUE_ACTIONS, 'r'))
total_actions = len(unique_actions)
print '*' * 20
print 'Preparing dataset...'
sys.stdout.flush()
# Prepare sequences using action indices
# Each action will be an index which will point to an action vector
# in the weights matrix of the Embedding layer of the network input
X, y, tokenizer = prepare_x_y(df_dataset, unique_actions)
# Create the embedding matrix for the embedding layer initialization
embedding_matrix = create_embedding_matrix(tokenizer)
#divide the examples in training and validation
total_examples = len(X)
test_per = 0.2
limit = int(test_per * total_examples)
X_train = X[limit:]
X_test = X[:limit]
y_train = y[limit:]
y_test = y[:limit]
print 'Different actions:', total_actions
print 'Total examples:', total_examples
print 'Train examples:', len(X_train), len(y_train)
print 'Test examples:', len(X_test), len(y_test)
sys.stdout.flush()
X_train = np.array(X_train)
y_train = np.array(y_train)
X_test = np.array(X_test)
y_test = np.array(y_test)
print 'Shape (X,y):'
print X_train.shape
print y_train.shape
print '*' * 20
print 'Building model...'
sys.stdout.flush()
#input pipeline
input_actions = Input(shape=(INPUT_ACTIONS,), dtype='int32', name='input_actions')
embedding_actions = Embedding(input_dim=embedding_matrix.shape[0], output_dim=embedding_matrix.shape[1], weights=[embedding_matrix], input_length=INPUT_ACTIONS, trainable=True, name='embedding_actions')(input_actions)
#attention mechanism
bidirectional_gru = Bidirectional(GRU(512, input_shape=(INPUT_ACTIONS, ACTION_EMBEDDING_LENGTH), name='bidirectional_gru'))(embedding_actions)
# TODO: check time distributed, and return sequences
dense_att_1 = Dense(512, activation = 'tanh',name = 'dense_att_1')(bidirectional_gru)
dense_att_2 = Dense(INPUT_ACTIONS, activation = 'softmax',name = 'dense_att_2')(dense_att_1)
reshape_att = Reshape((INPUT_ACTIONS, 1), name = 'reshape_att')(dense_att_2) #so we can multiply it with embeddings
#apply the attention
apply_att = Multiply()([embedding_actions, reshape_att])
#convolutions
reshape = Reshape((INPUT_ACTIONS, ACTION_EMBEDDING_LENGTH, 1), name = 'reshape')(apply_att) #add channel dimension for the CNNs
#branching convolutions
ngram_2 = Convolution2D(200, 2, ACTION_EMBEDDING_LENGTH, border_mode='valid',activation='relu', name = 'conv_2')(reshape)
maxpool_2 = MaxPooling2D(pool_size=(INPUT_ACTIONS-2+1,1), name = 'pooling_2')(ngram_2)
ngram_3 = Convolution2D(200, 3, ACTION_EMBEDDING_LENGTH, border_mode='valid',activation='relu', name = 'conv_3')(reshape)
maxpool_3 = MaxPooling2D(pool_size=(INPUT_ACTIONS-3+1,1), name = 'pooling_3')(ngram_3)
ngram_4 = Convolution2D(200, 4, ACTION_EMBEDDING_LENGTH, border_mode='valid',activation='relu', name = 'conv_4')(reshape)
maxpool_4 = MaxPooling2D(pool_size=(INPUT_ACTIONS-4+1,1), name = 'pooling_4')(ngram_4)
ngram_5 = Convolution2D(200, 5, ACTION_EMBEDDING_LENGTH, border_mode='valid',activation='relu', name = 'conv_5')(reshape)
maxpool_5 = MaxPooling2D(pool_size=(INPUT_ACTIONS-5+1,1), name = 'pooling_5')(ngram_5)
#1 branch again
merged = Concatenate(axis=2)([maxpool_2, maxpool_3, maxpool_4, maxpool_5])
flatten = Flatten(name = 'flatten')(merged)
dense_1 = Dense(256, activation = 'relu',name = 'dense_1')(flatten)
drop_1 = Dropout(0.8, name = 'drop_1')(dense_1)
#action prediction
output_actions = Dense(total_actions, activation='softmax', name='main_output')(drop_1)
model = Model(input=[input_actions], output=[output_actions])
model.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy', 'mse', 'mae'])
print(model.summary())
sys.stdout.flush()
print '*' * 20
print 'Training model...'
sys.stdout.flush()
BATCH_SIZE = 128
checkpoint = ModelCheckpoint(BEST_MODEL, monitor='val_acc', verbose=0, save_best_only=True, save_weights_only=False, mode='auto')
history = model.fit(X_train, y_train, batch_size=BATCH_SIZE, nb_epoch=1000, validation_data=(X_test, y_test), shuffle=True, callbacks=[checkpoint])
print '*' * 20
print 'Plotting history...'
sys.stdout.flush()
plot_training_info(['accuracy', 'loss'], True, history.history)
print '*' * 20
print 'Evaluating best model...'
sys.stdout.flush()
model = load_model(BEST_MODEL)
metrics = model.evaluate(X_test, y_test, batch_size=BATCH_SIZE)
print metrics
predictions = model.predict(X_test, BATCH_SIZE)
correct = [0] * 5
prediction_range = 5
for i, prediction in enumerate(predictions):
correct_answer = y_test[i].tolist().index(1)
best_n = np.sort(prediction)[::-1][:prediction_range]
for j in range(prediction_range):
if prediction.tolist().index(best_n[j]) == correct_answer:
for k in range(j,prediction_range):
correct[k] += 1
accuracies = []
for i in range(prediction_range):
print '%s prediction accuracy: %s' % (i+1, (correct[i] * 1.0) / len(y_test))
accuracies.append((correct[i] * 1.0) / len(y_test))
print accuracies
print '************ FIN ************\n' * 3
if __name__ == "__main__":
main(sys.argv)