jramompichel commited on
Commit
ca6abb5
verified
1 Parent(s): 269e64c

Upload 2 files

Browse files
Files changed (2) hide show
  1. nos-coda_iacobus-pt-gl.pt +3 -0
  2. pt-gl.yaml +128 -0
nos-coda_iacobus-pt-gl.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5539a720d32eb58a40d895095c511d8f840ac899eb81565404251e252297d39f
3
+ size 1436008923
pt-gl.yaml ADDED
@@ -0,0 +1,128 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ save_data: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/models
2
+ ## Where the vocab(s) will be written
3
+ src_vocab: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/models/run/bpe.vocab.src
4
+ tgt_vocab: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/models/run/bpe.vocab.tgt
5
+ overwrite: True
6
+
7
+ # Corpus opts:
8
+
9
+ data:
10
+ # en-es:
11
+ # path_src: /mnt/netapp1/Proxecto_NOS/mt/corpus/nmt-pld/en-es/train.en10k.txt
12
+ # path_tgt: /mnt/netapp1/Proxecto_NOS/mt/corpus/nmt-pld/en-es/train.es10k.txt
13
+ #transforms: [bpe, filtertoolong]
14
+ #weight: 100
15
+
16
+ # en-pt:
17
+ # path_src: /mnt/netapp1/Proxecto_NOS/mt/corpus/nmt-pld/en-pt/train.en10k.txt
18
+ # path_tgt: /mnt/netapp1/Proxecto_NOS/mt/corpus/nmt-pld/en-pt/train.pt10k.txt
19
+ # weight: 5
20
+
21
+ pt-gl:
22
+ path_src: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/train.pt35k.txt
23
+ path_tgt: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/train.gl35k.txt
24
+ transforms: [bpe, filtertoolong]
25
+
26
+ # en-it:
27
+ # path_src: /mnt/netapp1/Proxecto_NOS/mt/corpus/nmt-pld/en-it/train.en10k.txt
28
+ # path_tgt: /mnt/netapp1/Proxecto_NOS/mt/corpus/nmt-pld/en-it/train.it10k.txt
29
+
30
+ # en-ro:
31
+ # path_src: /mnt/netapp1/Proxecto_NOS/mt/corpus/nmt-pld/en-ro/train.en10k.txt
32
+ # path_src: /mnt/netapp1/Proxecto_NOS/mt/corpus/nmt-pld/en-ro/train.ro10k.txt
33
+
34
+
35
+ valid:
36
+ path_src: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/valid.pt35k.txt
37
+ path_tgt: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/valid.gl35k.txt
38
+ transforms: [bpe, filtertoolong]
39
+
40
+ ### Transform related opts:
41
+
42
+ #### Subword
43
+ src_subword_model: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/pt_35k.code
44
+ tgt_subword_model: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/gl_35k.code
45
+ #src_subword_vocab: /home/compartido/paulo/modelos/run/bpe.vocab.src
46
+ #tgt_subword_vocab: /home/compartido/paulo/modelos/run/bpe.vocab.tgt
47
+ #src_subword_model: ../sentencepiece/en-gl/en.sp.model
48
+ #tgt_subword_model: ../sentencepiece/en-gl/gl.sp.model
49
+ src_subword_type: bpe
50
+ tgt_subord_type: bpe
51
+ src_subword_nbest: 1
52
+ src_subword_alpha: 0.0
53
+ tgt_subword_nbest: 1
54
+ tgt_subword_alpha: 0.0
55
+
56
+ #### Filter
57
+ src_seq_length: 150
58
+ tgt_seq_length: 150
59
+
60
+ # silently ignore empty lines in the data
61
+ skip_empty_level: silent
62
+
63
+ ##embeddings
64
+ #src_embeddings: /mnt/lustre/scratch/nlsas//home/usc/ci/pgo/modelos/embeddings/en.emb.txt
65
+ #tgt_embeddings: /mnt/lustre/scratch/nlsas//home/usc/ci/pgo/modelos/embeddings/gl.emb.txt
66
+ src_embeddings: /mnt/netapp1/Proxecto_NOS/mt/treino_data/embeddings/pt.emb.txt
67
+ tgt_embeddings: /mnt/netapp1/Proxecto_NOS/mt/treino_data/embeddings/gl.emb.txt
68
+
69
+ ## supported types: GloVe, word2vec
70
+ embeddings_type: "word2vec"
71
+
72
+ # word_vec_size need to match with the pretrained embeddings dimensions
73
+ #word_vec_size: 300
74
+
75
+ # General opts
76
+ save_model: /mnt/netapp1/Proxecto_NOS/mt/corpus/iacobus/pt-gl/aut/models/
77
+ keep_checkpoint: 50
78
+ save_checkpoint_steps: 10000
79
+ average_decay: 0.0005
80
+ seed: 1234
81
+ report_every: 1000
82
+ train_steps: 400000
83
+ valid_steps: 10000
84
+
85
+ # Batching
86
+ queue_size: 10000
87
+ bucket_size: 32768
88
+ world_size: 1
89
+ gpu_ranks: [0]
90
+ batch_type: "tokens"
91
+ batch_size: 4096
92
+ valid_batch_size: 64
93
+ batch_size_multiple: 1
94
+ max_generator_batches: 2
95
+ accum_count: [4]
96
+ accum_steps: [0]
97
+
98
+ # Optimization
99
+ model_dtype: "fp16"
100
+ optim: "adam"
101
+ learning_rate: 2
102
+ warmup_steps: 8000
103
+ decay_method: "noam"
104
+ adam_beta2: 0.998
105
+ max_grad_norm: 0
106
+ label_smoothing: 0.1
107
+ param_init: 0
108
+ param_init_glorot: true
109
+ normalization: "tokens"
110
+
111
+ # Model
112
+ encoder_type: transformer
113
+ decoder_type: transformer
114
+ position_encoding: true
115
+ max_len: 6000
116
+ #max_relative_positions: 20
117
+ enc_layers: 12
118
+ dec_layers: 12
119
+ heads: 16
120
+ #rnn_size: 512
121
+ hidden_size: 512
122
+ word_vec_size: 512
123
+ transformer_ff: 2048
124
+ dropout_steps: [0]
125
+ dropout: [0.1]
126
+ attention_dropout: [0.1]
127
+ share_decoder_embeddings: true
128
+ share_embeddings: false