yonathanstwn commited on
Commit
132bd91
1 Parent(s): 6889e06

End of training

Browse files
.gitattributes CHANGED
@@ -32,7 +32,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
- last-checkpoint/source.spm filter=lfs diff=lfs merge=lfs -text
36
- last-checkpoint/target.spm filter=lfs diff=lfs merge=lfs -text
37
  source.spm filter=lfs diff=lfs merge=lfs -text
38
  target.spm filter=lfs diff=lfs merge=lfs -text
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
35
  source.spm filter=lfs diff=lfs merge=lfs -text
36
  target.spm filter=lfs diff=lfs merge=lfs -text
last-checkpoint/generation_config.json → generation_config.json RENAMED
File without changes
last-checkpoint/config.json DELETED
@@ -1,61 +0,0 @@
1
- {
2
- "_name_or_path": "./models/opus-mt-id-en",
3
- "_num_labels": 3,
4
- "activation_dropout": 0.0,
5
- "activation_function": "swish",
6
- "add_bias_logits": false,
7
- "add_final_layer_norm": false,
8
- "architectures": [
9
- "MarianMTModel"
10
- ],
11
- "attention_dropout": 0.0,
12
- "bad_words_ids": [
13
- [
14
- 54795
15
- ]
16
- ],
17
- "bos_token_id": 0,
18
- "classif_dropout": 0.0,
19
- "classifier_dropout": 0.0,
20
- "d_model": 512,
21
- "decoder_attention_heads": 8,
22
- "decoder_ffn_dim": 2048,
23
- "decoder_layerdrop": 0.0,
24
- "decoder_layers": 6,
25
- "decoder_start_token_id": 54795,
26
- "decoder_vocab_size": 54796,
27
- "dropout": 0.1,
28
- "encoder_attention_heads": 8,
29
- "encoder_ffn_dim": 2048,
30
- "encoder_layerdrop": 0.0,
31
- "encoder_layers": 6,
32
- "eos_token_id": 0,
33
- "forced_eos_token_id": 0,
34
- "id2label": {
35
- "0": "LABEL_0",
36
- "1": "LABEL_1",
37
- "2": "LABEL_2"
38
- },
39
- "init_std": 0.02,
40
- "is_encoder_decoder": true,
41
- "label2id": {
42
- "LABEL_0": 0,
43
- "LABEL_1": 1,
44
- "LABEL_2": 2
45
- },
46
- "max_length": 512,
47
- "max_position_embeddings": 512,
48
- "model_type": "marian",
49
- "normalize_before": false,
50
- "normalize_embedding": false,
51
- "num_beams": 6,
52
- "num_hidden_layers": 6,
53
- "pad_token_id": 54795,
54
- "scale_embedding": true,
55
- "share_encoder_decoder_embeddings": true,
56
- "static_position_embeddings": true,
57
- "torch_dtype": "float32",
58
- "transformers_version": "4.26.1",
59
- "use_cache": true,
60
- "vocab_size": 54796
61
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:be02ba762351448fdb43c9de8fc7524a5673db9940f359e3f2c141dae0eafcf9
3
- size 577701381
 
 
 
 
last-checkpoint/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7fb8a6cc6995e380012e737b779e9da8e70c04b5cc384c0cf6ac8ee12c9d9bcf
3
- size 289081861
 
 
 
 
last-checkpoint/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a21f7066a08653d668e2c1fd19c96892fdabde30fb66ef567cd6df78f62941c
3
- size 14575
 
 
 
 
last-checkpoint/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:71b837d714ec73dc76f669f3b9cdbd9ff3164ece76a650cb1f470331c7d8a12d
3
- size 627
 
 
 
 
last-checkpoint/source.spm DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a8fefe71c7f26cb0c6aa1b9f0cc0f8d18006b20fe41c547af7f25b9c8333465
3
- size 800687
 
 
 
 
last-checkpoint/special_tokens_map.json DELETED
@@ -1,5 +0,0 @@
1
- {
2
- "eos_token": "</s>",
3
- "pad_token": "<pad>",
4
- "unk_token": "<unk>"
5
- }
 
 
 
 
 
 
last-checkpoint/target.spm DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e88300911c2c573ec5526777a1e84bae698d20925b82dcef9c7248bb0e537ed0
3
- size 795925
 
 
 
 
last-checkpoint/tokenizer_config.json DELETED
@@ -1,13 +0,0 @@
1
- {
2
- "eos_token": "</s>",
3
- "model_max_length": 512,
4
- "name_or_path": "./models/opus-mt-id-en",
5
- "pad_token": "<pad>",
6
- "separate_vocabs": false,
7
- "source_lang": "id",
8
- "sp_model_kwargs": {},
9
- "special_tokens_map_file": null,
10
- "target_lang": "en",
11
- "tokenizer_class": "MarianTokenizer",
12
- "unk_token": "<unk>"
13
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/trainer_state.json DELETED
@@ -1,91 +0,0 @@
1
- {
2
- "best_metric": 0.7666811347007751,
3
- "best_model_checkpoint": "models/opus-mt-id-en-ccmatrix-v2/checkpoint-742190",
4
- "epoch": 5.0,
5
- "global_step": 742190,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 1.0,
12
- "learning_rate": 8.043349273222341e-06,
13
- "loss": 0.803,
14
- "step": 148438
15
- },
16
- {
17
- "epoch": 1.0,
18
- "eval_bleu": 50.3216,
19
- "eval_loss": 0.8185181617736816,
20
- "eval_runtime": 6046.3001,
21
- "eval_samples_per_second": 41.348,
22
- "eval_steps_per_second": 1.292,
23
- "step": 148438
24
- },
25
- {
26
- "epoch": 2.0,
27
- "learning_rate": 6.032511954916757e-06,
28
- "loss": 0.7212,
29
- "step": 296876
30
- },
31
- {
32
- "epoch": 2.0,
33
- "eval_bleu": 51.282,
34
- "eval_loss": 0.7903820872306824,
35
- "eval_runtime": 6090.6592,
36
- "eval_samples_per_second": 41.046,
37
- "eval_steps_per_second": 1.283,
38
- "step": 296876
39
- },
40
- {
41
- "epoch": 3.0,
42
- "learning_rate": 4.0216746366111706e-06,
43
- "loss": 0.6913,
44
- "step": 445314
45
- },
46
- {
47
- "epoch": 3.0,
48
- "eval_bleu": 51.7806,
49
- "eval_loss": 0.7790645956993103,
50
- "eval_runtime": 5967.3621,
51
- "eval_samples_per_second": 41.895,
52
- "eval_steps_per_second": 1.309,
53
- "step": 445314
54
- },
55
- {
56
- "epoch": 4.0,
57
- "learning_rate": 2.0108373183055853e-06,
58
- "loss": 0.6727,
59
- "step": 593752
60
- },
61
- {
62
- "epoch": 4.0,
63
- "eval_bleu": 52.0263,
64
- "eval_loss": 0.7690777778625488,
65
- "eval_runtime": 5991.9694,
66
- "eval_samples_per_second": 41.723,
67
- "eval_steps_per_second": 1.304,
68
- "step": 593752
69
- },
70
- {
71
- "epoch": 5.0,
72
- "learning_rate": 0.0,
73
- "loss": 0.6609,
74
- "step": 742190
75
- },
76
- {
77
- "epoch": 5.0,
78
- "eval_bleu": 52.173,
79
- "eval_loss": 0.7666811347007751,
80
- "eval_runtime": 5978.9596,
81
- "eval_samples_per_second": 41.813,
82
- "eval_steps_per_second": 1.307,
83
- "step": 742190
84
- }
85
- ],
86
- "max_steps": 742190,
87
- "num_train_epochs": 5,
88
- "total_flos": 2.7115179224334336e+17,
89
- "trial_name": null,
90
- "trial_params": null
91
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e0ec848403d2a6fefc7f1bfa6e79cfc2188115a47663d8ad03f4bd0c89ec7b0
3
- size 3643
 
 
 
 
last-checkpoint/vocab.json DELETED
The diff for this file is too large to render. See raw diff