forked from happynoom/DeepTrade_keras
-
Notifications
You must be signed in to change notification settings - Fork 0
/
windpuller.py
99 lines (86 loc) · 4.54 KB
/
windpuller.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
# -*- coding: utf-8 -*-
# Copyright 2017 The Xiaoyu Fang. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ==============================================================================
from keras.layers import Dense, LSTM, Activation, BatchNormalization, Dropout, LocallyConnected1D, \
GaussianNoise
from keras import initializers
from activations import ReLU
from activations import BiReLU
from renormalization import BatchRenormalization
from keras.models import Sequential
from keras.optimizers import SGD, RMSprop
from keras.models import load_model
from keras.initializers import Constant
from keras import regularizers
import keras.backend as K
def risk_estimation(y_true, y_pred):
return -100. * K.mean(y_true * y_pred)
def pairwise_logit(y_true, y_pred):
loss_mat = K.log(1 + K.exp(K.sign(K.transpose(y_true) - y_true) * (y_pred - K.transpose(y_pred))))
return K.mean(K.mean(loss_mat))
class WindPuller(object):
def __init__(self, input_shape = None, lr=0.01, n_layers=2, n_hidden=8, rate_dropout=0.2, loss=risk_estimation):
self.input_shape = input_shape
self.lr = lr
self.n_layers = n_layers
self.n_hidden = n_hidden
self.rate_dropout = rate_dropout
self.loss = loss
self.model = None
def build_model(self):
print("initializing..., learing rate %s, n_layers %s, n_hidden %s, dropout rate %s." % (
self.lr, self.n_layers, self.n_hidden, self.rate_dropout))
self.model = Sequential()
self.model.add(GaussianNoise(stddev=0.01, input_shape=self.input_shape))
# self.model.add(LocallyConnected1D(self.input_shape[1] * 2, 3))
# self.model.add(Dropout(rate=0.5))
for i in range(0, self.n_layers - 1):
self.model.add(LSTM(self.n_hidden * 4, return_sequences=True, activation='tanh',
recurrent_activation='hard_sigmoid', kernel_initializer='glorot_uniform',
recurrent_initializer='orthogonal', bias_initializer='zeros',
dropout=self.rate_dropout, recurrent_dropout=self.rate_dropout))
self.model.add(LSTM(self.n_hidden, return_sequences=False, activation='tanh',
recurrent_activation='hard_sigmoid', kernel_initializer='glorot_uniform',
recurrent_initializer='orthogonal', bias_initializer='zeros',
dropout=self.rate_dropout, recurrent_dropout=self.rate_dropout))
self.model.add(Dense(1, kernel_initializer=initializers.glorot_uniform()))
# self.model.add(BatchNormalization(axis=-1, moving_mean_initializer=Constant(value=0.5),
# moving_variance_initializer=Constant(value=0.25)))
if self.loss == risk_estimation:
self.model.add(BatchNormalization(axis=-1, beta_initializer='ones'))
self.model.add(ReLU(alpha=0.0, max_value=1.0))
opt = RMSprop(lr=self.lr)
self.model.compile(loss=self.loss,
optimizer=opt,
metrics=['accuracy'])
def fit(self, x, y, batch_size=32, nb_epoch=100, verbose=1, callbacks=None,
validation_split=0., validation_data=None, shuffle=True,
class_weight=None, sample_weight=None, initial_epoch=0, **kwargs):
self.model.fit(x, y, batch_size, nb_epoch, verbose, callbacks,
validation_split, validation_data, shuffle, class_weight, sample_weight,
initial_epoch, **kwargs)
def save(self, path):
self.model.save(path)
@staticmethod
def load_model(path):
wp = WindPuller()
wp.model = load_model(path, custom_objects={'risk_estimation': risk_estimation})
return wp
def evaluate(self, x, y, batch_size=32, verbose=1,
sample_weight=None, **kwargs):
return self.model.evaluate(x, y, batch_size, verbose,
sample_weight)
def predict(self, x, batch_size=32, verbose=0):
return self.model.predict(x, batch_size, verbose)