eglkan1 commited on
Commit
5be88f8
1 Parent(s): 70a0580

End of training

Browse files
README.md CHANGED
@@ -5,6 +5,7 @@ tags:
5
  - generated_from_trainer
6
  metrics:
7
  - rouge
 
8
  model-index:
9
  - name: mT5-TextSimp-LT-BatchSize8-lr5e-5
10
  results: []
@@ -17,10 +18,11 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/mt5-base](https://huggingface.co/google/mt5-base) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.0969
21
- - Rouge1: 0.6185
22
- - Rouge2: 0.4427
23
- - Rougel: 0.6087
 
24
  - Gen Len: 38.0501
25
 
26
  ## Model description
@@ -51,21 +53,21 @@ The following hyperparameters were used during training:
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Gen Len |
55
- |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:-------:|
56
- | 32.3498 | 0.96 | 200 | 26.0719 | 0.0008 | 0.0 | 0.0008 | 512.0 |
57
- | 5.8297 | 1.91 | 400 | 4.1306 | 0.0059 | 0.0 | 0.0058 | 45.0573 |
58
- | 0.7087 | 2.87 | 600 | 0.6039 | 0.003 | 0.0 | 0.0029 | 39.0501 |
59
- | 0.4166 | 3.83 | 800 | 0.1958 | 0.3954 | 0.2416 | 0.3823 | 39.0501 |
60
- | 0.2193 | 4.78 | 1000 | 0.1172 | 0.5244 | 0.3536 | 0.514 | 38.0501 |
61
- | 0.1371 | 5.74 | 1200 | 0.1029 | 0.5936 | 0.4122 | 0.5839 | 38.0501 |
62
- | 0.1971 | 6.7 | 1400 | 0.0974 | 0.6077 | 0.4302 | 0.5984 | 38.0501 |
63
- | 0.1653 | 7.66 | 1600 | 0.0969 | 0.6185 | 0.4427 | 0.6087 | 38.0501 |
64
 
65
 
66
  ### Framework versions
67
 
68
- - Transformers 4.36.2
69
- - Pytorch 2.1.1
70
- - Datasets 2.16.1
71
- - Tokenizers 0.15.0
 
5
  - generated_from_trainer
6
  metrics:
7
  - rouge
8
+ - sacrebleu
9
  model-index:
10
  - name: mT5-TextSimp-LT-BatchSize8-lr5e-5
11
  results: []
 
18
 
19
  This model is a fine-tuned version of [google/mt5-base](https://huggingface.co/google/mt5-base) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.0983
22
+ - Rouge1: 0.6245
23
+ - Rouge2: 0.4439
24
+ - Rougel: 0.6142
25
+ - Sacrebleu: 35.7192
26
  - Gen Len: 38.0501
27
 
28
  ## Model description
 
53
 
54
  ### Training results
55
 
56
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Sacrebleu | Gen Len |
57
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:--------:|
58
+ | 35.3898 | 0.96 | 200 | 27.6372 | 0.0019 | 0.0 | 0.0018 | 0.0003 | 512.0 |
59
+ | 3.5712 | 1.91 | 400 | 1.9615 | 0.0171 | 0.0 | 0.0167 | 0.0225 | 39.0501 |
60
+ | 0.6489 | 2.87 | 600 | 0.5638 | 0.0052 | 0.0 | 0.0051 | 0.0256 | 39.0501 |
61
+ | 0.6017 | 3.83 | 800 | 3.2823 | 0.2419 | 0.1287 | 0.2318 | 0.6457 | 130.3556 |
62
+ | 0.3784 | 4.78 | 1000 | 0.1340 | 0.5092 | 0.3277 | 0.4978 | 26.7005 | 38.0549 |
63
+ | 0.1521 | 5.74 | 1200 | 0.1092 | 0.5782 | 0.3973 | 0.5672 | 33.2443 | 38.0501 |
64
+ | 0.2096 | 6.7 | 1400 | 0.1001 | 0.6149 | 0.4342 | 0.6046 | 34.6518 | 38.0501 |
65
+ | 0.1719 | 7.66 | 1600 | 0.0983 | 0.6245 | 0.4439 | 0.6142 | 35.7192 | 38.0501 |
66
 
67
 
68
  ### Framework versions
69
 
70
+ - Transformers 4.33.0
71
+ - Pytorch 2.1.2+cu121
72
+ - Datasets 2.14.4
73
+ - Tokenizers 0.13.3
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "decoder_start_token_id": 0,
4
  "eos_token_id": 1,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.36.2"
7
  }
 
3
  "decoder_start_token_id": 0,
4
  "eos_token_id": 1,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.33.0"
7
  }
special_tokens_map.json CHANGED
@@ -1,23 +1,5 @@
1
  {
2
- "eos_token": {
3
- "content": "</s>",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "pad_token": {
10
- "content": "<pad>",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "unk_token": {
17
- "content": "<unk>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- }
23
  }
 
1
  {
2
+ "eos_token": "</s>",
3
+ "pad_token": "<pad>",
4
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  }
tokenizer_config.json CHANGED
@@ -1,31 +1,5 @@
1
  {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "<pad>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "1": {
12
- "content": "</s>",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "2": {
20
- "content": "<unk>",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- }
27
- },
28
- "additional_special_tokens": [],
29
  "clean_up_tokenization_spaces": true,
30
  "eos_token": "</s>",
31
  "extra_ids": 0,
@@ -34,5 +8,6 @@
34
  "pad_token": "<pad>",
35
  "sp_model_kwargs": {},
36
  "tokenizer_class": "T5Tokenizer",
 
37
  "unk_token": "<unk>"
38
  }
 
1
  {
2
+ "additional_special_tokens": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "clean_up_tokenization_spaces": true,
4
  "eos_token": "</s>",
5
  "extra_ids": 0,
 
8
  "pad_token": "<pad>",
9
  "sp_model_kwargs": {},
10
  "tokenizer_class": "T5Tokenizer",
11
+ "tokenizer_file": null,
12
  "unk_token": "<unk>"
13
  }