andreaparker commited on
Commit
5e86ffa
1 Parent(s): a689e31
Files changed (3) hide show
  1. README.md +16 -10
  2. generation_config.json +1 -1
  3. pytorch_model.bin +1 -1
README.md CHANGED
@@ -1,4 +1,6 @@
1
  ---
 
 
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
@@ -21,7 +23,7 @@ model-index:
21
  metrics:
22
  - name: Rouge1
23
  type: rouge
24
- value: 46.5595
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -31,12 +33,12 @@ should probably proofread and complete it, then remove this comment. -->
31
 
32
  This model is a fine-tuned version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) on the samsum dataset.
33
  It achieves the following results on the evaluation set:
34
- - Loss: 1.3822
35
- - Rouge1: 46.5595
36
- - Rouge2: 22.9243
37
- - Rougel: 39.0902
38
- - Rougelsum: 42.934
39
- - Gen Len: 17.2308
40
 
41
  ## Model description
42
 
@@ -61,18 +63,22 @@ The following hyperparameters were used during training:
61
  - seed: 42
62
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
  - lr_scheduler_type: linear
64
- - num_epochs: 1
65
 
66
  ### Training results
67
 
68
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
69
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
70
- | 1.5156 | 1.0 | 1842 | 1.3822 | 46.5595 | 22.9243 | 39.0902 | 42.934 | 17.2308 |
 
 
 
 
71
 
72
 
73
  ### Framework versions
74
 
75
- - Transformers 4.27.2
76
  - Pytorch 1.13.1+cu116
77
  - Datasets 2.10.1
78
  - Tokenizers 0.13.2
 
1
  ---
2
+ language:
3
+ - en
4
  license: apache-2.0
5
  tags:
6
  - generated_from_trainer
 
23
  metrics:
24
  - name: Rouge1
25
  type: rouge
26
+ value: 47.4145
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
33
 
34
  This model is a fine-tuned version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) on the samsum dataset.
35
  It achieves the following results on the evaluation set:
36
+ - Loss: 1.3772
37
+ - Rouge1: 47.4145
38
+ - Rouge2: 23.9579
39
+ - Rougel: 40.0508
40
+ - Rougelsum: 43.7144
41
+ - Gen Len: 17.3162
42
 
43
  ## Model description
44
 
 
63
  - seed: 42
64
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
  - lr_scheduler_type: linear
66
+ - num_epochs: 5
67
 
68
  ### Training results
69
 
70
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
71
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
72
+ | 1.4264 | 1.0 | 1842 | 1.3829 | 46.4916 | 23.1227 | 39.444 | 42.9025 | 17.0977 |
73
+ | 1.3527 | 2.0 | 3684 | 1.3732 | 47.0694 | 23.4769 | 39.5942 | 43.2226 | 17.4554 |
74
+ | 1.2554 | 3.0 | 5526 | 1.3709 | 46.8801 | 23.3161 | 39.5423 | 43.1581 | 17.2027 |
75
+ | 1.2503 | 4.0 | 7368 | 1.3736 | 47.4138 | 23.7437 | 40.0016 | 43.6108 | 17.2198 |
76
+ | 1.1675 | 5.0 | 9210 | 1.3772 | 47.4145 | 23.9579 | 40.0508 | 43.7144 | 17.3162 |
77
 
78
 
79
  ### Framework versions
80
 
81
+ - Transformers 4.27.3
82
  - Pytorch 1.13.1+cu116
83
  - Datasets 2.10.1
84
  - Tokenizers 0.13.2
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
- "transformers_version": "4.27.2"
6
  }
 
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
+ "transformers_version": "4.27.3"
6
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6967554cd20ce84ece990620be96d288d88ac16d7952849c2e8c873f5a279769
3
  size 990408885
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a70e77be9b6270e9ec28accaee7fb761fe6627fbc979d38c205591fd4bb33bfa
3
  size 990408885