| save_data: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/models |
| |
| src_vocab: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/models/run/bpe.vocab.src |
| tgt_vocab: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/models/run/bpe.vocab.tgt |
| overwrite: True |
|
|
| |
|
|
| data: |
| |
| |
| |
| |
| |
|
|
| |
| |
| |
| |
|
|
| pt-gl: |
| path_src: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/train.pt35k.txt |
| path_tgt: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/train.gl35k.txt |
| transforms: [bpe, filtertoolong] |
| |
| |
| |
| |
|
|
| |
| |
| |
|
|
|
|
| valid: |
| path_src: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/valid.pt35k.txt |
| path_tgt: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/valid.gl35k.txt |
| transforms: [bpe, filtertoolong] |
| |
| |
|
|
| |
| src_subword_model: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/pt_35k.code |
| tgt_subword_model: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/gl_35k.code |
| |
| |
| |
| |
| src_subword_type: bpe |
| tgt_subord_type: bpe |
| src_subword_nbest: 1 |
| src_subword_alpha: 0.0 |
| tgt_subword_nbest: 1 |
| tgt_subword_alpha: 0.0 |
|
|
| |
| src_seq_length: 150 |
| tgt_seq_length: 150 |
|
|
| |
| skip_empty_level: silent |
|
|
| |
| |
| |
| src_embeddings: /mnt/netapp1/Proxecto_NOS/mt/treino_data/embeddings/pt.emb.txt |
| tgt_embeddings: /mnt/netapp1/Proxecto_NOS/mt/treino_data/embeddings/gl.emb.txt |
|
|
| |
| embeddings_type: "word2vec" |
|
|
| |
| |
|
|
| |
| save_model: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/models/ |
| keep_checkpoint: 50 |
| save_checkpoint_steps: 10000 |
| average_decay: 0.0005 |
| seed: 1234 |
| report_every: 1000 |
| train_steps: 400000 |
| valid_steps: 10000 |
|
|
| |
| queue_size: 10000 |
| bucket_size: 32768 |
| world_size: 1 |
| gpu_ranks: [0] |
| batch_type: "tokens" |
| batch_size: 4096 |
| valid_batch_size: 64 |
| batch_size_multiple: 1 |
| max_generator_batches: 2 |
| accum_count: [4] |
| accum_steps: [0] |
|
|
| |
| model_dtype: "fp16" |
| optim: "adam" |
| learning_rate: 2 |
| warmup_steps: 8000 |
| decay_method: "noam" |
| adam_beta2: 0.998 |
| max_grad_norm: 0 |
| label_smoothing: 0.1 |
| param_init: 0 |
| param_init_glorot: true |
| normalization: "tokens" |
|
|
| |
| encoder_type: transformer |
| decoder_type: transformer |
| position_encoding: true |
| max_len: 6000 |
| |
| enc_layers: 12 |
| dec_layers: 12 |
| heads: 16 |
| |
| hidden_size: 512 |
| word_vec_size: 512 |
| transformer_ff: 2048 |
| dropout_steps: [0] |
| dropout: [0.1] |
| attention_dropout: [0.1] |
| share_decoder_embeddings: true |
| share_embeddings: false |
|
|