HugoVoxx commited on
Commit
b5b13c7
·
verified ·
1 Parent(s): 10846e9

Upload 13 files

Browse files
aglib/meliad/transformer/configs/recurrent/bias_dual.gin ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ decoder_stack.DecoderStack:
3
+ dstack_window_length = 0
4
+ recurrent_layer_indices = (-3,)
5
+ feedback_recurrence = False
6
+
7
+ transformer_layer.TransformerLayer:
8
+ window_length = 512
9
+ recurrent_num_states = 512
10
+ recurrent_gate_type = "bias"
11
+ recurrent_single_gate = False
12
+ recurrent_skip_ffn = False
13
+
14
+ training_loop.Trainer:
15
+ generate_every_steps = 0 # Not yet supported with recurrence.
aglib/meliad/transformer/configs/recurrent/bias_single.gin ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ decoder_stack.DecoderStack:
3
+ dstack_window_length = 0
4
+ recurrent_layer_indices = (-3,)
5
+ feedback_recurrence = False
6
+
7
+ transformer_layer.TransformerLayer:
8
+ window_length = 512
9
+ recurrent_num_states = 512
10
+ recurrent_gate_type = "bias"
11
+ recurrent_single_gate = True
12
+ recurrent_skip_ffn = False
13
+
14
+ training_loop.Trainer:
15
+ generate_every_steps = 0 # Not yet supported with recurrence.
aglib/meliad/transformer/configs/recurrent/bias_skip.gin ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ decoder_stack.DecoderStack:
3
+ dstack_window_length = 0
4
+ recurrent_layer_indices = (-3,)
5
+ feedback_recurrence = False
6
+
7
+ transformer_layer.TransformerLayer:
8
+ window_length = 512
9
+ recurrent_num_states = 512
10
+ recurrent_gate_type = "bias"
11
+ recurrent_single_gate = False
12
+ recurrent_skip_ffn = True
13
+
14
+ training_loop.Trainer:
15
+ generate_every_steps = 0 # Not yet supported with recurrence.
aglib/meliad/transformer/configs/recurrent/block_feedback_512.gin ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # Apply after the gate config.
3
+
4
+ transformer_layer.TransformerLayer:
5
+ window_length = 512
6
+
7
+ decoder_stack.DecoderStack:
8
+ dstack_window_length = 512
9
+ feedback_recurrence = True
aglib/meliad/transformer/configs/recurrent/lstm_dual.gin ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ decoder_stack.DecoderStack:
3
+ dstack_window_length = 0
4
+ recurrent_layer_indices = (-3,)
5
+ feedback_recurrence = False
6
+
7
+ transformer_layer.TransformerLayer:
8
+ window_length = 512
9
+ recurrent_num_states = 512
10
+ recurrent_gate_type = "lstm"
11
+ recurrent_single_gate = False
12
+ recurrent_skip_ffn = False
13
+
14
+ training_loop.Trainer:
15
+ generate_every_steps = 0 # Not yet supported with recurrence.
aglib/meliad/transformer/configs/recurrent/lstm_single.gin ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ decoder_stack.DecoderStack:
3
+ dstack_window_length = 0
4
+ recurrent_layer_indices = (-3,)
5
+ feedback_recurrence = False
6
+
7
+ transformer_layer.TransformerLayer:
8
+ window_length = 512
9
+ recurrent_num_states = 512
10
+ recurrent_gate_type = "lstm"
11
+ recurrent_single_gate = True
12
+ recurrent_skip_ffn = False
13
+
14
+ training_loop.Trainer:
15
+ generate_every_steps = 0 # Not yet supported with recurrence.
aglib/meliad/transformer/configs/recurrent/lstm_skip.gin ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ decoder_stack.DecoderStack:
3
+ dstack_window_length = 0
4
+ recurrent_layer_indices = (-3,)
5
+ feedback_recurrence = False
6
+
7
+ transformer_layer.TransformerLayer:
8
+ window_length = 512
9
+ recurrent_num_states = 512
10
+ recurrent_gate_type = "lstm"
11
+ recurrent_single_gate = False
12
+ recurrent_skip_ffn = True
13
+
14
+ training_loop.Trainer:
15
+ generate_every_steps = 0 # Not yet supported with recurrence.
aglib/meliad/transformer/configs/recurrent/states_1024.gin ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+
2
+ # Apply after the gate config.
3
+
4
+ transformer_layer.TransformerLayer:
5
+ recurrent_num_states = 1024
aglib/meliad/transformer/configs/recurrent/states_128.gin ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+
2
+ # Apply after the gate config.
3
+
4
+ transformer_layer.TransformerLayer:
5
+ recurrent_num_states = 128
aglib/meliad/transformer/configs/recurrent/states_2048.gin ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+
2
+ # Apply after the gate config.
3
+
4
+ transformer_layer.TransformerLayer:
5
+ recurrent_num_states = 2048
aglib/meliad/transformer/configs/recurrent/states_256.gin ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+
2
+ # Apply after the gate config.
3
+
4
+ transformer_layer.TransformerLayer:
5
+ recurrent_num_states = 256
aglib/meliad/transformer/configs/recurrent/states_512.gin ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+
2
+ # Apply after the gate config.
3
+
4
+ transformer_layer.TransformerLayer:
5
+ recurrent_num_states = 512
aglib/meliad/transformer/configs/recurrent/tall_24L.gin ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+
2
+ # Use a "tall" 24L configuration, with two recurrent layers.
3
+ # Apply after the gate config.
4
+
5
+ NUM_LAYERS = 24
6
+
7
+ decoder_stack.DecoderStack:
8
+ recurrent_layer_indices = (9,19,)