Spaces:
Sleeping
Sleeping
Upload 13 files
Browse files- aglib/meliad/transformer/configs/recurrent/bias_dual.gin +15 -0
- aglib/meliad/transformer/configs/recurrent/bias_single.gin +15 -0
- aglib/meliad/transformer/configs/recurrent/bias_skip.gin +15 -0
- aglib/meliad/transformer/configs/recurrent/block_feedback_512.gin +9 -0
- aglib/meliad/transformer/configs/recurrent/lstm_dual.gin +15 -0
- aglib/meliad/transformer/configs/recurrent/lstm_single.gin +15 -0
- aglib/meliad/transformer/configs/recurrent/lstm_skip.gin +15 -0
- aglib/meliad/transformer/configs/recurrent/states_1024.gin +5 -0
- aglib/meliad/transformer/configs/recurrent/states_128.gin +5 -0
- aglib/meliad/transformer/configs/recurrent/states_2048.gin +5 -0
- aglib/meliad/transformer/configs/recurrent/states_256.gin +5 -0
- aglib/meliad/transformer/configs/recurrent/states_512.gin +5 -0
- aglib/meliad/transformer/configs/recurrent/tall_24L.gin +8 -0
aglib/meliad/transformer/configs/recurrent/bias_dual.gin
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
decoder_stack.DecoderStack:
|
3 |
+
dstack_window_length = 0
|
4 |
+
recurrent_layer_indices = (-3,)
|
5 |
+
feedback_recurrence = False
|
6 |
+
|
7 |
+
transformer_layer.TransformerLayer:
|
8 |
+
window_length = 512
|
9 |
+
recurrent_num_states = 512
|
10 |
+
recurrent_gate_type = "bias"
|
11 |
+
recurrent_single_gate = False
|
12 |
+
recurrent_skip_ffn = False
|
13 |
+
|
14 |
+
training_loop.Trainer:
|
15 |
+
generate_every_steps = 0 # Not yet supported with recurrence.
|
aglib/meliad/transformer/configs/recurrent/bias_single.gin
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
decoder_stack.DecoderStack:
|
3 |
+
dstack_window_length = 0
|
4 |
+
recurrent_layer_indices = (-3,)
|
5 |
+
feedback_recurrence = False
|
6 |
+
|
7 |
+
transformer_layer.TransformerLayer:
|
8 |
+
window_length = 512
|
9 |
+
recurrent_num_states = 512
|
10 |
+
recurrent_gate_type = "bias"
|
11 |
+
recurrent_single_gate = True
|
12 |
+
recurrent_skip_ffn = False
|
13 |
+
|
14 |
+
training_loop.Trainer:
|
15 |
+
generate_every_steps = 0 # Not yet supported with recurrence.
|
aglib/meliad/transformer/configs/recurrent/bias_skip.gin
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
decoder_stack.DecoderStack:
|
3 |
+
dstack_window_length = 0
|
4 |
+
recurrent_layer_indices = (-3,)
|
5 |
+
feedback_recurrence = False
|
6 |
+
|
7 |
+
transformer_layer.TransformerLayer:
|
8 |
+
window_length = 512
|
9 |
+
recurrent_num_states = 512
|
10 |
+
recurrent_gate_type = "bias"
|
11 |
+
recurrent_single_gate = False
|
12 |
+
recurrent_skip_ffn = True
|
13 |
+
|
14 |
+
training_loop.Trainer:
|
15 |
+
generate_every_steps = 0 # Not yet supported with recurrence.
|
aglib/meliad/transformer/configs/recurrent/block_feedback_512.gin
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
# Apply after the gate config.
|
3 |
+
|
4 |
+
transformer_layer.TransformerLayer:
|
5 |
+
window_length = 512
|
6 |
+
|
7 |
+
decoder_stack.DecoderStack:
|
8 |
+
dstack_window_length = 512
|
9 |
+
feedback_recurrence = True
|
aglib/meliad/transformer/configs/recurrent/lstm_dual.gin
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
decoder_stack.DecoderStack:
|
3 |
+
dstack_window_length = 0
|
4 |
+
recurrent_layer_indices = (-3,)
|
5 |
+
feedback_recurrence = False
|
6 |
+
|
7 |
+
transformer_layer.TransformerLayer:
|
8 |
+
window_length = 512
|
9 |
+
recurrent_num_states = 512
|
10 |
+
recurrent_gate_type = "lstm"
|
11 |
+
recurrent_single_gate = False
|
12 |
+
recurrent_skip_ffn = False
|
13 |
+
|
14 |
+
training_loop.Trainer:
|
15 |
+
generate_every_steps = 0 # Not yet supported with recurrence.
|
aglib/meliad/transformer/configs/recurrent/lstm_single.gin
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
decoder_stack.DecoderStack:
|
3 |
+
dstack_window_length = 0
|
4 |
+
recurrent_layer_indices = (-3,)
|
5 |
+
feedback_recurrence = False
|
6 |
+
|
7 |
+
transformer_layer.TransformerLayer:
|
8 |
+
window_length = 512
|
9 |
+
recurrent_num_states = 512
|
10 |
+
recurrent_gate_type = "lstm"
|
11 |
+
recurrent_single_gate = True
|
12 |
+
recurrent_skip_ffn = False
|
13 |
+
|
14 |
+
training_loop.Trainer:
|
15 |
+
generate_every_steps = 0 # Not yet supported with recurrence.
|
aglib/meliad/transformer/configs/recurrent/lstm_skip.gin
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
decoder_stack.DecoderStack:
|
3 |
+
dstack_window_length = 0
|
4 |
+
recurrent_layer_indices = (-3,)
|
5 |
+
feedback_recurrence = False
|
6 |
+
|
7 |
+
transformer_layer.TransformerLayer:
|
8 |
+
window_length = 512
|
9 |
+
recurrent_num_states = 512
|
10 |
+
recurrent_gate_type = "lstm"
|
11 |
+
recurrent_single_gate = False
|
12 |
+
recurrent_skip_ffn = True
|
13 |
+
|
14 |
+
training_loop.Trainer:
|
15 |
+
generate_every_steps = 0 # Not yet supported with recurrence.
|
aglib/meliad/transformer/configs/recurrent/states_1024.gin
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
# Apply after the gate config.
|
3 |
+
|
4 |
+
transformer_layer.TransformerLayer:
|
5 |
+
recurrent_num_states = 1024
|
aglib/meliad/transformer/configs/recurrent/states_128.gin
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
# Apply after the gate config.
|
3 |
+
|
4 |
+
transformer_layer.TransformerLayer:
|
5 |
+
recurrent_num_states = 128
|
aglib/meliad/transformer/configs/recurrent/states_2048.gin
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
# Apply after the gate config.
|
3 |
+
|
4 |
+
transformer_layer.TransformerLayer:
|
5 |
+
recurrent_num_states = 2048
|
aglib/meliad/transformer/configs/recurrent/states_256.gin
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
# Apply after the gate config.
|
3 |
+
|
4 |
+
transformer_layer.TransformerLayer:
|
5 |
+
recurrent_num_states = 256
|
aglib/meliad/transformer/configs/recurrent/states_512.gin
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
# Apply after the gate config.
|
3 |
+
|
4 |
+
transformer_layer.TransformerLayer:
|
5 |
+
recurrent_num_states = 512
|
aglib/meliad/transformer/configs/recurrent/tall_24L.gin
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
# Use a "tall" 24L configuration, with two recurrent layers.
|
3 |
+
# Apply after the gate config.
|
4 |
+
|
5 |
+
NUM_LAYERS = 24
|
6 |
+
|
7 |
+
decoder_stack.DecoderStack:
|
8 |
+
recurrent_layer_indices = (9,19,)
|