yonathanstwn commited on
Commit
bea3f57
1 Parent(s): d9bcaae

End of training

Browse files
.gitattributes CHANGED
@@ -32,5 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
- last-checkpoint/tokenizer.json filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
35
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
last-checkpoint/generation_config.json → generation_config.json RENAMED
File without changes
last-checkpoint/config.json DELETED
@@ -1,35 +0,0 @@
1
- {
2
- "_name_or_path": "./models/nllb-en-id",
3
- "activation_dropout": 0.0,
4
- "activation_function": "relu",
5
- "architectures": [
6
- "M2M100ForConditionalGeneration"
7
- ],
8
- "attention_dropout": 0.1,
9
- "bos_token_id": 0,
10
- "d_model": 1024,
11
- "decoder_attention_heads": 16,
12
- "decoder_ffn_dim": 4096,
13
- "decoder_layerdrop": 0,
14
- "decoder_layers": 12,
15
- "decoder_start_token_id": 2,
16
- "dropout": 0.1,
17
- "encoder_attention_heads": 16,
18
- "encoder_ffn_dim": 4096,
19
- "encoder_layerdrop": 0,
20
- "encoder_layers": 12,
21
- "eos_token_id": 2,
22
- "init_std": 0.02,
23
- "is_encoder_decoder": true,
24
- "max_length": 200,
25
- "max_position_embeddings": 1024,
26
- "model_type": "m2m_100",
27
- "num_hidden_layers": 12,
28
- "pad_token_id": 1,
29
- "scale_embedding": true,
30
- "tokenizer_class": "NllbTokenizer",
31
- "torch_dtype": "float32",
32
- "transformers_version": "4.26.1",
33
- "use_cache": true,
34
- "vocab_size": 256206
35
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cde4ef0d03921cf301b7e4277f3075c28c45fe455ef24b0a627014198d3c94ff
3
- size 4920898947
 
 
 
 
last-checkpoint/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:399e4ee176be40862a03b661a1ead63796f58728991959abc071fd0eb140a6c9
3
- size 2468874377
 
 
 
 
last-checkpoint/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:00942863925632570829ac94b5d27075d207d665dd053a5a65d5cd92501a9c4d
3
- size 14575
 
 
 
 
last-checkpoint/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b853d56fd50fc738f70b55e5a38b11254561d0eeeea3283e462a1b133bae651e
3
- size 627
 
 
 
 
last-checkpoint/sentencepiece.bpe.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:14bb8dfb35c0ffdea7bc01e56cea38b9e3d5efcdcb9c251d6b40538e1aab555a
3
- size 4852054
 
 
 
 
last-checkpoint/special_tokens_map.json DELETED
@@ -1,219 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "ace_Arab",
4
- "ace_Latn",
5
- "acm_Arab",
6
- "acq_Arab",
7
- "aeb_Arab",
8
- "afr_Latn",
9
- "ajp_Arab",
10
- "aka_Latn",
11
- "amh_Ethi",
12
- "apc_Arab",
13
- "arb_Arab",
14
- "ars_Arab",
15
- "ary_Arab",
16
- "arz_Arab",
17
- "asm_Beng",
18
- "ast_Latn",
19
- "awa_Deva",
20
- "ayr_Latn",
21
- "azb_Arab",
22
- "azj_Latn",
23
- "bak_Cyrl",
24
- "bam_Latn",
25
- "ban_Latn",
26
- "bel_Cyrl",
27
- "bem_Latn",
28
- "ben_Beng",
29
- "bho_Deva",
30
- "bjn_Arab",
31
- "bjn_Latn",
32
- "bod_Tibt",
33
- "bos_Latn",
34
- "bug_Latn",
35
- "bul_Cyrl",
36
- "cat_Latn",
37
- "ceb_Latn",
38
- "ces_Latn",
39
- "cjk_Latn",
40
- "ckb_Arab",
41
- "crh_Latn",
42
- "cym_Latn",
43
- "dan_Latn",
44
- "deu_Latn",
45
- "dik_Latn",
46
- "dyu_Latn",
47
- "dzo_Tibt",
48
- "ell_Grek",
49
- "eng_Latn",
50
- "epo_Latn",
51
- "est_Latn",
52
- "eus_Latn",
53
- "ewe_Latn",
54
- "fao_Latn",
55
- "pes_Arab",
56
- "fij_Latn",
57
- "fin_Latn",
58
- "fon_Latn",
59
- "fra_Latn",
60
- "fur_Latn",
61
- "fuv_Latn",
62
- "gla_Latn",
63
- "gle_Latn",
64
- "glg_Latn",
65
- "grn_Latn",
66
- "guj_Gujr",
67
- "hat_Latn",
68
- "hau_Latn",
69
- "heb_Hebr",
70
- "hin_Deva",
71
- "hne_Deva",
72
- "hrv_Latn",
73
- "hun_Latn",
74
- "hye_Armn",
75
- "ibo_Latn",
76
- "ilo_Latn",
77
- "ind_Latn",
78
- "isl_Latn",
79
- "ita_Latn",
80
- "jav_Latn",
81
- "jpn_Jpan",
82
- "kab_Latn",
83
- "kac_Latn",
84
- "kam_Latn",
85
- "kan_Knda",
86
- "kas_Arab",
87
- "kas_Deva",
88
- "kat_Geor",
89
- "knc_Arab",
90
- "knc_Latn",
91
- "kaz_Cyrl",
92
- "kbp_Latn",
93
- "kea_Latn",
94
- "khm_Khmr",
95
- "kik_Latn",
96
- "kin_Latn",
97
- "kir_Cyrl",
98
- "kmb_Latn",
99
- "kon_Latn",
100
- "kor_Hang",
101
- "kmr_Latn",
102
- "lao_Laoo",
103
- "lvs_Latn",
104
- "lij_Latn",
105
- "lim_Latn",
106
- "lin_Latn",
107
- "lit_Latn",
108
- "lmo_Latn",
109
- "ltg_Latn",
110
- "ltz_Latn",
111
- "lua_Latn",
112
- "lug_Latn",
113
- "luo_Latn",
114
- "lus_Latn",
115
- "mag_Deva",
116
- "mai_Deva",
117
- "mal_Mlym",
118
- "mar_Deva",
119
- "min_Latn",
120
- "mkd_Cyrl",
121
- "plt_Latn",
122
- "mlt_Latn",
123
- "mni_Beng",
124
- "khk_Cyrl",
125
- "mos_Latn",
126
- "mri_Latn",
127
- "zsm_Latn",
128
- "mya_Mymr",
129
- "nld_Latn",
130
- "nno_Latn",
131
- "nob_Latn",
132
- "npi_Deva",
133
- "nso_Latn",
134
- "nus_Latn",
135
- "nya_Latn",
136
- "oci_Latn",
137
- "gaz_Latn",
138
- "ory_Orya",
139
- "pag_Latn",
140
- "pan_Guru",
141
- "pap_Latn",
142
- "pol_Latn",
143
- "por_Latn",
144
- "prs_Arab",
145
- "pbt_Arab",
146
- "quy_Latn",
147
- "ron_Latn",
148
- "run_Latn",
149
- "rus_Cyrl",
150
- "sag_Latn",
151
- "san_Deva",
152
- "sat_Beng",
153
- "scn_Latn",
154
- "shn_Mymr",
155
- "sin_Sinh",
156
- "slk_Latn",
157
- "slv_Latn",
158
- "smo_Latn",
159
- "sna_Latn",
160
- "snd_Arab",
161
- "som_Latn",
162
- "sot_Latn",
163
- "spa_Latn",
164
- "als_Latn",
165
- "srd_Latn",
166
- "srp_Cyrl",
167
- "ssw_Latn",
168
- "sun_Latn",
169
- "swe_Latn",
170
- "swh_Latn",
171
- "szl_Latn",
172
- "tam_Taml",
173
- "tat_Cyrl",
174
- "tel_Telu",
175
- "tgk_Cyrl",
176
- "tgl_Latn",
177
- "tha_Thai",
178
- "tir_Ethi",
179
- "taq_Latn",
180
- "taq_Tfng",
181
- "tpi_Latn",
182
- "tsn_Latn",
183
- "tso_Latn",
184
- "tuk_Latn",
185
- "tum_Latn",
186
- "tur_Latn",
187
- "twi_Latn",
188
- "tzm_Tfng",
189
- "uig_Arab",
190
- "ukr_Cyrl",
191
- "umb_Latn",
192
- "urd_Arab",
193
- "uzn_Latn",
194
- "vec_Latn",
195
- "vie_Latn",
196
- "war_Latn",
197
- "wol_Latn",
198
- "xho_Latn",
199
- "ydd_Hebr",
200
- "yor_Latn",
201
- "yue_Hant",
202
- "zho_Hans",
203
- "zho_Hant",
204
- "zul_Latn"
205
- ],
206
- "bos_token": "<s>",
207
- "cls_token": "<s>",
208
- "eos_token": "</s>",
209
- "mask_token": {
210
- "content": "<mask>",
211
- "lstrip": true,
212
- "normalized": true,
213
- "rstrip": false,
214
- "single_word": false
215
- },
216
- "pad_token": "<pad>",
217
- "sep_token": "</s>",
218
- "unk_token": "<unk>"
219
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8dec87343dd7b147b84072891b984b69471e9b7539eee4e544a0450ffc88e0e8
3
- size 17331294
 
 
 
 
last-checkpoint/tokenizer_config.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "additional_special_tokens": null,
3
- "bos_token": "<s>",
4
- "cls_token": "<s>",
5
- "eos_token": "</s>",
6
- "mask_token": {
7
- "__type": "AddedToken",
8
- "content": "<mask>",
9
- "lstrip": true,
10
- "normalized": true,
11
- "rstrip": false,
12
- "single_word": false
13
- },
14
- "model_max_length": 1024,
15
- "name_or_path": "./models/nllb-en-id",
16
- "pad_token": "<pad>",
17
- "sep_token": "</s>",
18
- "sp_model_kwargs": {},
19
- "special_tokens_map_file": null,
20
- "src_lang": "eng_Latn",
21
- "tgt_lang": "ind_Latn",
22
- "tokenizer_class": "NllbTokenizer",
23
- "unk_token": "<unk>"
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/trainer_state.json DELETED
@@ -1,166 +0,0 @@
1
- {
2
- "best_metric": 0.8704027533531189,
3
- "best_model_checkpoint": "models/nllb-ecolindo/checkpoint-210190",
4
- "epoch": 10.0,
5
- "global_step": 210190,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 1.0,
12
- "learning_rate": 9.174596246180708e-06,
13
- "loss": 1.3017,
14
- "step": 21019
15
- },
16
- {
17
- "epoch": 1.0,
18
- "eval_bleu": 35.0676,
19
- "eval_loss": 0.9831313490867615,
20
- "eval_runtime": 94.4525,
21
- "eval_samples_per_second": 21.175,
22
- "eval_steps_per_second": 0.667,
23
- "step": 21019
24
- },
25
- {
26
- "epoch": 2.0,
27
- "learning_rate": 8.15519666327174e-06,
28
- "loss": 1.0088,
29
- "step": 42038
30
- },
31
- {
32
- "epoch": 2.0,
33
- "eval_bleu": 36.3191,
34
- "eval_loss": 0.9318408966064453,
35
- "eval_runtime": 72.147,
36
- "eval_samples_per_second": 27.721,
37
- "eval_steps_per_second": 0.873,
38
- "step": 42038
39
- },
40
- {
41
- "epoch": 3.0,
42
- "learning_rate": 7.135797080362773e-06,
43
- "loss": 0.9472,
44
- "step": 63057
45
- },
46
- {
47
- "epoch": 3.0,
48
- "eval_bleu": 36.5221,
49
- "eval_loss": 0.9090332388877869,
50
- "eval_runtime": 43.552,
51
- "eval_samples_per_second": 45.922,
52
- "eval_steps_per_second": 1.447,
53
- "step": 63057
54
- },
55
- {
56
- "epoch": 4.0,
57
- "learning_rate": 6.116397497453806e-06,
58
- "loss": 0.9078,
59
- "step": 84076
60
- },
61
- {
62
- "epoch": 4.0,
63
- "eval_bleu": 36.3949,
64
- "eval_loss": 0.8919472694396973,
65
- "eval_runtime": 57.3849,
66
- "eval_samples_per_second": 34.852,
67
- "eval_steps_per_second": 1.098,
68
- "step": 84076
69
- },
70
- {
71
- "epoch": 5.0,
72
- "learning_rate": 5.096997914544838e-06,
73
- "loss": 0.8789,
74
- "step": 105095
75
- },
76
- {
77
- "epoch": 5.0,
78
- "eval_bleu": 37.3689,
79
- "eval_loss": 0.8832682967185974,
80
- "eval_runtime": 56.9673,
81
- "eval_samples_per_second": 35.108,
82
- "eval_steps_per_second": 1.106,
83
- "step": 105095
84
- },
85
- {
86
- "epoch": 6.0,
87
- "learning_rate": 4.07759833163587e-06,
88
- "loss": 0.8576,
89
- "step": 126114
90
- },
91
- {
92
- "epoch": 6.0,
93
- "eval_bleu": 37.2262,
94
- "eval_loss": 0.8784551620483398,
95
- "eval_runtime": 55.8811,
96
- "eval_samples_per_second": 35.79,
97
- "eval_steps_per_second": 1.127,
98
- "step": 126114
99
- },
100
- {
101
- "epoch": 7.0,
102
- "learning_rate": 3.058198748726903e-06,
103
- "loss": 0.8403,
104
- "step": 147133
105
- },
106
- {
107
- "epoch": 7.0,
108
- "eval_bleu": 37.2933,
109
- "eval_loss": 0.8747518062591553,
110
- "eval_runtime": 44.6649,
111
- "eval_samples_per_second": 44.778,
112
- "eval_steps_per_second": 1.411,
113
- "step": 147133
114
- },
115
- {
116
- "epoch": 8.0,
117
- "learning_rate": 2.038799165817935e-06,
118
- "loss": 0.8281,
119
- "step": 168152
120
- },
121
- {
122
- "epoch": 8.0,
123
- "eval_bleu": 37.2952,
124
- "eval_loss": 0.8716793656349182,
125
- "eval_runtime": 43.4982,
126
- "eval_samples_per_second": 45.979,
127
- "eval_steps_per_second": 1.448,
128
- "step": 168152
129
- },
130
- {
131
- "epoch": 9.0,
132
- "learning_rate": 1.0193995829089675e-06,
133
- "loss": 0.8186,
134
- "step": 189171
135
- },
136
- {
137
- "epoch": 9.0,
138
- "eval_bleu": 37.2801,
139
- "eval_loss": 0.8708682656288147,
140
- "eval_runtime": 41.2854,
141
- "eval_samples_per_second": 48.443,
142
- "eval_steps_per_second": 1.526,
143
- "step": 189171
144
- },
145
- {
146
- "epoch": 10.0,
147
- "learning_rate": 0.0,
148
- "loss": 0.813,
149
- "step": 210190
150
- },
151
- {
152
- "epoch": 10.0,
153
- "eval_bleu": 37.2396,
154
- "eval_loss": 0.8704027533531189,
155
- "eval_runtime": 44.7957,
156
- "eval_samples_per_second": 44.647,
157
- "eval_steps_per_second": 1.406,
158
- "step": 210190
159
- }
160
- ],
161
- "max_steps": 210190,
162
- "num_train_epochs": 10,
163
- "total_flos": 3.338536761820938e+17,
164
- "trial_name": null,
165
- "trial_params": null
166
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d9db2c86ed401d06f5076171976f58cd756593b824f876e75dd4564d3c0960e
3
- size 3643