amaye15 commited on
Commit
aeb0a86
·
1 Parent(s): fe12381

Update modeling_autoencoder.py

Browse files
Files changed (1) hide show
  1. modeling_autoencoder.py +5 -3
modeling_autoencoder.py CHANGED
@@ -202,9 +202,11 @@ def create_layers(
202
  layers.append(nn.Linear(input_dim, output_dim))
203
  elif layer_types in ['lstm', 'rnn', 'gru']:
204
  rnn_layer = getattr(nn, layer_types.upper()) # Dynamically get the RNN layer class.
205
- half_output_dim = output_dim // (1 if bidirectional else 1)
206
- layers.append(rnn_layer(input_dim, half_output_dim, batch_first=True, bidirectional=bidirectional))
207
-
 
 
208
  # Add dropout layer between layers, except for the last layer.
209
  if (idx != num_layers - 1) and (dropout_rate is not None):
210
  layers.append(nn.Dropout(dropout_rate))
 
202
  layers.append(nn.Linear(input_dim, output_dim))
203
  elif layer_types in ['lstm', 'rnn', 'gru']:
204
  rnn_layer = getattr(nn, layer_types.upper()) # Dynamically get the RNN layer class.
205
+ half_output_dim = output_dim // (2 if bidirectional else 1)
206
+ if idx == 0:
207
+ layers.append(rnn_layer(input_dim, half_output_dim, batch_first=True, bidirectional=bidirectional))
208
+ else:
209
+ layers.append(rnn_layer(input_dim*2, half_output_dim, batch_first=True, bidirectional=bidirectional))
210
  # Add dropout layer between layers, except for the last layer.
211
  if (idx != num_layers - 1) and (dropout_rate is not None):
212
  layers.append(nn.Dropout(dropout_rate))