File size: 1,093 Bytes
158b61b |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 |
## Where the vocab(s) will be written
save_data: cnndm/run/example
# Prevent overwriting existing files in the folder
overwrite: False
# filter long examples
src_seq_length: 10000
tgt_seq_length: 10000
src_seq_length_trunc: 400
tgt_seq_length_trunc: 100
# common vocabulary for source and target
share_vocab: True
# Corpus opts:
data:
cnndm:
path_src: cnndm/train.txt.src
path_tgt: cnndm/train.txt.tgt.tagged
transforms: []
weight: 1
valid:
path_src: cnndm/val.txt.src
path_tgt: cnndm/val.txt.tgt.tagged
transforms: []
src_vocab_size: 50000
tgt_vocab_size: 50000
src_vocab: cnndm/run/example.vocab.src
tgt_vocab: cnndm/run/example.vocab.tgt
save_model: cnndm/run/model
copy_attn: true
global_attention: mlp
word_vec_size: 128
rnn_size: 512
layers: 1
encoder_type: brnn
train_steps: 200000
max_grad_norm: 2
dropout: 0
batch_size: 16
valid_batch_size: 16
optim: adagrad
learning_rate: 0.15
adagrad_accumulator_init: 0.1
reuse_copy_attn: true
copy_loss_by_seqlength: true
bridge: true
seed: 777
world_size: 2
gpu_ranks: [0, 1] |