dantedgp commited on
Commit
c80d60c
·
verified ·
1 Parent(s): 9a62bfa

End of training

Browse files
Files changed (1) hide show
  1. README.md +71 -74
README.md CHANGED
@@ -1,74 +1,71 @@
1
- ---
2
- license: apache-2.0
3
- base_model: google/flan-t5-small
4
- tags:
5
- - generated_from_trainer
6
- metrics:
7
- - rouge
8
- model-index:
9
- - name: flan-t5-small-finetuned-question-generation
10
- results: []
11
- widget:
12
- - text: "ask: Mitochondria context: The mitochondria, commonly known as the powerhouse of the cell, oxidizes pyruvate and harnesses the energy of electrons comming from glucose by making them cascade down the electron chain."
13
- pipeline: "text2text-generation"
14
- ---
15
-
16
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
- should probably proofread and complete it, then remove this comment. -->
18
-
19
- # flan-t5-small-finetuned-question-generation
20
-
21
- This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on an unknown dataset.
22
- It achieves the following results on the evaluation set:
23
- - Loss: 1.5888
24
- - Rouge1: 49.9368
25
- - Rouge2: 27.2185
26
- - Rougel: 46.1729
27
- - Rougelsum: 46.1900
28
- - Gen Len: 13.7801
29
-
30
- ## Model description
31
-
32
- More information needed
33
-
34
- ## Intended uses & limitations
35
-
36
- More information needed
37
-
38
- ## Training and evaluation data
39
-
40
- More information needed
41
-
42
- ## Training procedure
43
-
44
- ### Training hyperparameters
45
-
46
- The following hyperparameters were used during training:
47
- - learning_rate: 5.6e-05
48
- - train_batch_size: 8
49
- - eval_batch_size: 8
50
- - seed: 42
51
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
- - lr_scheduler_type: linear
53
- - num_epochs: 8
54
-
55
- ### Training results
56
-
57
- | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
58
- |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
59
- | 1.8453 | 1.0 | 10917 | 1.6093 | 47.9986 | 25.3028 | 44.3072 | 44.3345 | 13.6285 |
60
- | 1.6998 | 2.0 | 21834 | 1.5955 | 49.1185 | 26.2998 | 45.4201 | 45.4194 | 13.4627 |
61
- | 1.6257 | 3.0 | 32751 | 1.5933 | 49.4151 | 26.7655 | 45.6769 | 45.6809 | 13.7603 |
62
- | 1.5722 | 4.0 | 43668 | 1.5822 | 49.4727 | 26.8015 | 45.8010 | 45.8247 | 13.6301 |
63
- | 1.5337 | 5.0 | 54585 | 1.5881 | 49.5572 | 26.8461 | 45.8811 | 45.8919 | 13.7029 |
64
- | 1.5045 | 6.0 | 65502 | 1.5852 | 49.7863 | 27.0921 | 46.0625 | 46.0677 | 13.8099 |
65
- | 1.4822 | 7.0 | 76419 | 1.5867 | 49.7769 | 27.0550 | 46.0091 | 46.0228 | 13.8027 |
66
- | 1.4674 | 8.0 | 87336 | 1.5888 | 49.9368 | 27.2185 | 46.1729 | 46.1900 | 13.7801 |
67
-
68
-
69
- ### Framework versions
70
-
71
- - Transformers 4.42.4
72
- - Pytorch 2.3.1
73
- - Datasets 2.20.0
74
- - Tokenizers 0.19.1
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: google/flan-t5-small
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - rouge
8
+ model-index:
9
+ - name: flan-t5-small-finetuned-question-generation
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # flan-t5-small-finetuned-question-generation
17
+
18
+ This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on an unknown dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 1.5998
21
+ - Rouge1: 50.1718
22
+ - Rouge2: 27.5603
23
+ - Rougel: 46.3981
24
+ - Rougelsum: 46.3975
25
+ - Gen Len: 13.7948
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 0.0001
45
+ - train_batch_size: 8
46
+ - eval_batch_size: 8
47
+ - seed: 42
48
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
+ - lr_scheduler_type: linear
50
+ - num_epochs: 8
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
55
+ |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
56
+ | 1.819 | 1.0 | 10913 | 1.6159 | 48.8496 | 26.1270 | 45.1331 | 45.1442 | 13.8064 |
57
+ | 1.6487 | 2.0 | 21826 | 1.5947 | 48.8142 | 26.2209 | 45.1475 | 45.1482 | 13.8229 |
58
+ | 1.5546 | 3.0 | 32739 | 1.5910 | 49.6261 | 27.1655 | 45.9472 | 45.9535 | 13.9086 |
59
+ | 1.4862 | 4.0 | 43652 | 1.5887 | 49.9953 | 27.4630 | 46.2824 | 46.2841 | 13.7223 |
60
+ | 1.4327 | 5.0 | 54565 | 1.5950 | 50.1663 | 27.6038 | 46.4602 | 46.4721 | 13.7106 |
61
+ | 1.3907 | 6.0 | 65478 | 1.5910 | 49.9510 | 27.4795 | 46.2230 | 46.2218 | 13.8172 |
62
+ | 1.3598 | 7.0 | 76391 | 1.5973 | 50.1049 | 27.4804 | 46.3268 | 46.3300 | 13.7966 |
63
+ | 1.3388 | 8.0 | 87304 | 1.5998 | 50.1718 | 27.5603 | 46.3981 | 46.3975 | 13.7948 |
64
+
65
+
66
+ ### Framework versions
67
+
68
+ - Transformers 4.42.4
69
+ - Pytorch 2.3.1
70
+ - Datasets 2.20.0
71
+ - Tokenizers 0.19.1