Using Native tensorflow RNNLayer with dropout within keras model
I have a model implemented in Keras, but I need to implement the same model in tensorflow. So, I am looking to implement only the RNN layer of the model and keep the rest the same, that is, the prediction method, fitting the model... are all implemented in keras. Therefore, here is the code:
Keras model:
def emotion_model(max_seq_len, num_features, learning_rate, num_units_1, num_units_2, bidirectional, dropout, num_targets):
# Input layer
inputs = Input(shape=(max_seq_len, num_features))
# 1st layer
net = LSTM(num_units_1, return_sequences=True, dropout=dropout, recurrent_dropout=dropout)(net)
# 2nd layer
net = LSTM(num_units_2, return_sequences=True, dropout=dropout, recurrent_dropout=dropout)(net)
# Output layer
outputs =
out1 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out1)
if num_targets >= 2:
out2 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out2)
if num_targets == 3:
out3 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out3)
# Create and compile model
rmsprop = RMSprop(lr=learning_rate)
model = Model(inputs=inputs, outputs=outputs)
model.compile(optimizer=rmsprop, loss=ccc_loss) # CCC-based loss function
return model
Now, I would like to replace the LSTM layers above with the equivalent code in tensorflow. Therefore, in a different Module I have implemented the following:
def baseline_model(inputs, cell_Size1, cell_Size2, dropout):
with tf.variable_scope('model', reuse=tf.AUTO_REUSE):
cell1 = tf.nn.rnn_cell.LSTMCell(cell_Size1)
cell1 = tf.nn.rnn_cell.DropoutWrapper(cell1, input_keep_prob=1.0 - dropout, state_keep_prob=1.0 - dropout)
cell2 = tf.nn.rnn_cell.LSTMCell(cell_Size2)
cell2 = tf.nn.rnn_cell.DropoutWrapper(cell2, input_keep_prob=1.0 - dropout, state_keep_prob=1.0 - dropout)
cell = tf.nn.rnn_cell.MultiRNNCell([cell1, cell2], state_is_tuple=True)
# output1: shape=[1, time_steps, 32]
output, new_state = tf.nn.dynamic_rnn(cell, inputs, dtype=tf.float32)
return output
I have tried net = Lambda(partial(baseline_model, dropout))(net)
where I removed the cell_size1
and cell_size2
from the method "baseline_model" arguments, yet didn't work
Second, I have tried dumping directly the LSTM layers implemented in tensorflow instead of the LSTM
layers in keras above, and this doesn't solve my problem.
Any help is much appreciated!!
python tensorflow keras rnn
add a comment |
I have a model implemented in Keras, but I need to implement the same model in tensorflow. So, I am looking to implement only the RNN layer of the model and keep the rest the same, that is, the prediction method, fitting the model... are all implemented in keras. Therefore, here is the code:
Keras model:
def emotion_model(max_seq_len, num_features, learning_rate, num_units_1, num_units_2, bidirectional, dropout, num_targets):
# Input layer
inputs = Input(shape=(max_seq_len, num_features))
# 1st layer
net = LSTM(num_units_1, return_sequences=True, dropout=dropout, recurrent_dropout=dropout)(net)
# 2nd layer
net = LSTM(num_units_2, return_sequences=True, dropout=dropout, recurrent_dropout=dropout)(net)
# Output layer
outputs =
out1 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out1)
if num_targets >= 2:
out2 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out2)
if num_targets == 3:
out3 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out3)
# Create and compile model
rmsprop = RMSprop(lr=learning_rate)
model = Model(inputs=inputs, outputs=outputs)
model.compile(optimizer=rmsprop, loss=ccc_loss) # CCC-based loss function
return model
Now, I would like to replace the LSTM layers above with the equivalent code in tensorflow. Therefore, in a different Module I have implemented the following:
def baseline_model(inputs, cell_Size1, cell_Size2, dropout):
with tf.variable_scope('model', reuse=tf.AUTO_REUSE):
cell1 = tf.nn.rnn_cell.LSTMCell(cell_Size1)
cell1 = tf.nn.rnn_cell.DropoutWrapper(cell1, input_keep_prob=1.0 - dropout, state_keep_prob=1.0 - dropout)
cell2 = tf.nn.rnn_cell.LSTMCell(cell_Size2)
cell2 = tf.nn.rnn_cell.DropoutWrapper(cell2, input_keep_prob=1.0 - dropout, state_keep_prob=1.0 - dropout)
cell = tf.nn.rnn_cell.MultiRNNCell([cell1, cell2], state_is_tuple=True)
# output1: shape=[1, time_steps, 32]
output, new_state = tf.nn.dynamic_rnn(cell, inputs, dtype=tf.float32)
return output
I have tried net = Lambda(partial(baseline_model, dropout))(net)
where I removed the cell_size1
and cell_size2
from the method "baseline_model" arguments, yet didn't work
Second, I have tried dumping directly the LSTM layers implemented in tensorflow instead of the LSTM
layers in keras above, and this doesn't solve my problem.
Any help is much appreciated!!
python tensorflow keras rnn
add a comment |
I have a model implemented in Keras, but I need to implement the same model in tensorflow. So, I am looking to implement only the RNN layer of the model and keep the rest the same, that is, the prediction method, fitting the model... are all implemented in keras. Therefore, here is the code:
Keras model:
def emotion_model(max_seq_len, num_features, learning_rate, num_units_1, num_units_2, bidirectional, dropout, num_targets):
# Input layer
inputs = Input(shape=(max_seq_len, num_features))
# 1st layer
net = LSTM(num_units_1, return_sequences=True, dropout=dropout, recurrent_dropout=dropout)(net)
# 2nd layer
net = LSTM(num_units_2, return_sequences=True, dropout=dropout, recurrent_dropout=dropout)(net)
# Output layer
outputs =
out1 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out1)
if num_targets >= 2:
out2 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out2)
if num_targets == 3:
out3 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out3)
# Create and compile model
rmsprop = RMSprop(lr=learning_rate)
model = Model(inputs=inputs, outputs=outputs)
model.compile(optimizer=rmsprop, loss=ccc_loss) # CCC-based loss function
return model
Now, I would like to replace the LSTM layers above with the equivalent code in tensorflow. Therefore, in a different Module I have implemented the following:
def baseline_model(inputs, cell_Size1, cell_Size2, dropout):
with tf.variable_scope('model', reuse=tf.AUTO_REUSE):
cell1 = tf.nn.rnn_cell.LSTMCell(cell_Size1)
cell1 = tf.nn.rnn_cell.DropoutWrapper(cell1, input_keep_prob=1.0 - dropout, state_keep_prob=1.0 - dropout)
cell2 = tf.nn.rnn_cell.LSTMCell(cell_Size2)
cell2 = tf.nn.rnn_cell.DropoutWrapper(cell2, input_keep_prob=1.0 - dropout, state_keep_prob=1.0 - dropout)
cell = tf.nn.rnn_cell.MultiRNNCell([cell1, cell2], state_is_tuple=True)
# output1: shape=[1, time_steps, 32]
output, new_state = tf.nn.dynamic_rnn(cell, inputs, dtype=tf.float32)
return output
I have tried net = Lambda(partial(baseline_model, dropout))(net)
where I removed the cell_size1
and cell_size2
from the method "baseline_model" arguments, yet didn't work
Second, I have tried dumping directly the LSTM layers implemented in tensorflow instead of the LSTM
layers in keras above, and this doesn't solve my problem.
Any help is much appreciated!!
python tensorflow keras rnn
I have a model implemented in Keras, but I need to implement the same model in tensorflow. So, I am looking to implement only the RNN layer of the model and keep the rest the same, that is, the prediction method, fitting the model... are all implemented in keras. Therefore, here is the code:
Keras model:
def emotion_model(max_seq_len, num_features, learning_rate, num_units_1, num_units_2, bidirectional, dropout, num_targets):
# Input layer
inputs = Input(shape=(max_seq_len, num_features))
# 1st layer
net = LSTM(num_units_1, return_sequences=True, dropout=dropout, recurrent_dropout=dropout)(net)
# 2nd layer
net = LSTM(num_units_2, return_sequences=True, dropout=dropout, recurrent_dropout=dropout)(net)
# Output layer
outputs =
out1 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out1)
if num_targets >= 2:
out2 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out2)
if num_targets == 3:
out3 = TimeDistributed(Dense(1))(net) # linear activation
outputs.append(out3)
# Create and compile model
rmsprop = RMSprop(lr=learning_rate)
model = Model(inputs=inputs, outputs=outputs)
model.compile(optimizer=rmsprop, loss=ccc_loss) # CCC-based loss function
return model
Now, I would like to replace the LSTM layers above with the equivalent code in tensorflow. Therefore, in a different Module I have implemented the following:
def baseline_model(inputs, cell_Size1, cell_Size2, dropout):
with tf.variable_scope('model', reuse=tf.AUTO_REUSE):
cell1 = tf.nn.rnn_cell.LSTMCell(cell_Size1)
cell1 = tf.nn.rnn_cell.DropoutWrapper(cell1, input_keep_prob=1.0 - dropout, state_keep_prob=1.0 - dropout)
cell2 = tf.nn.rnn_cell.LSTMCell(cell_Size2)
cell2 = tf.nn.rnn_cell.DropoutWrapper(cell2, input_keep_prob=1.0 - dropout, state_keep_prob=1.0 - dropout)
cell = tf.nn.rnn_cell.MultiRNNCell([cell1, cell2], state_is_tuple=True)
# output1: shape=[1, time_steps, 32]
output, new_state = tf.nn.dynamic_rnn(cell, inputs, dtype=tf.float32)
return output
I have tried net = Lambda(partial(baseline_model, dropout))(net)
where I removed the cell_size1
and cell_size2
from the method "baseline_model" arguments, yet didn't work
Second, I have tried dumping directly the LSTM layers implemented in tensorflow instead of the LSTM
layers in keras above, and this doesn't solve my problem.
Any help is much appreciated!!
python tensorflow keras rnn
python tensorflow keras rnn
asked Nov 11 '18 at 17:35
I. AI. A
714430
714430
add a comment |
add a comment |
0
active
oldest
votes
Your Answer
StackExchange.ifUsing("editor", function ()
StackExchange.using("externalEditor", function ()
StackExchange.using("snippets", function ()
StackExchange.snippets.init();
);
);
, "code-snippets");
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "1"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53251390%2fusing-native-tensorflow-rnnlayer-with-dropout-within-keras-model%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
0
active
oldest
votes
0
active
oldest
votes
active
oldest
votes
active
oldest
votes
Thanks for contributing an answer to Stack Overflow!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53251390%2fusing-native-tensorflow-rnnlayer-with-dropout-within-keras-model%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown