Model save
Browse files- README.md +41 -21
- model.safetensors +1 -1
README.md
CHANGED
@@ -3,23 +3,11 @@ license: apache-2.0
|
|
3 |
base_model: google-t5/t5-base
|
4 |
tags:
|
5 |
- generated_from_trainer
|
6 |
-
datasets:
|
7 |
-
- fedora-copr/pep-sum
|
8 |
metrics:
|
9 |
- rouge
|
10 |
model-index:
|
11 |
- name: pep_summarization
|
12 |
-
results:
|
13 |
-
- task:
|
14 |
-
name: Summarization
|
15 |
-
type: summarization
|
16 |
-
dataset:
|
17 |
-
name: fedora-copr/pep-sum
|
18 |
-
type: fedora-copr/pep-sum
|
19 |
-
metrics:
|
20 |
-
- name: Rouge1
|
21 |
-
type: rouge
|
22 |
-
value: 87.9903
|
23 |
---
|
24 |
|
25 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -27,14 +15,14 @@ should probably proofread and complete it, then remove this comment. -->
|
|
27 |
|
28 |
# pep_summarization
|
29 |
|
30 |
-
This model is a fine-tuned version of [google-t5/t5-base](https://huggingface.co/google-t5/t5-base) on
|
31 |
It achieves the following results on the evaluation set:
|
32 |
-
- Loss: 0.
|
33 |
-
- Rouge1:
|
34 |
-
- Rouge2:
|
35 |
-
- Rougel:
|
36 |
-
- Rougelsum:
|
37 |
-
- Gen Len:
|
38 |
|
39 |
## Model description
|
40 |
|
@@ -59,10 +47,42 @@ The following hyperparameters were used during training:
|
|
59 |
- seed: 42
|
60 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
61 |
- lr_scheduler_type: linear
|
62 |
-
- num_epochs:
|
63 |
|
64 |
### Training results
|
65 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
66 |
|
67 |
|
68 |
### Framework versions
|
|
|
3 |
base_model: google-t5/t5-base
|
4 |
tags:
|
5 |
- generated_from_trainer
|
|
|
|
|
6 |
metrics:
|
7 |
- rouge
|
8 |
model-index:
|
9 |
- name: pep_summarization
|
10 |
+
results: []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
---
|
12 |
|
13 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
15 |
|
16 |
# pep_summarization
|
17 |
|
18 |
+
This model is a fine-tuned version of [google-t5/t5-base](https://huggingface.co/google-t5/t5-base) on an unknown dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
+
- Loss: 0.0564
|
21 |
+
- Rouge1: 89.1468
|
22 |
+
- Rouge2: 88.6354
|
23 |
+
- Rougel: 89.0016
|
24 |
+
- Rougelsum: 89.0138
|
25 |
+
- Gen Len: 63.7246
|
26 |
|
27 |
## Model description
|
28 |
|
|
|
47 |
- seed: 42
|
48 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
49 |
- lr_scheduler_type: linear
|
50 |
+
- num_epochs: 30.0
|
51 |
|
52 |
### Training results
|
53 |
|
54 |
+
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
|
55 |
+
|:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
|
56 |
+
| No log | 1.0 | 69 | 0.0463 | 84.7175 | 84.1187 | 84.7778 | 84.4607 | 74.1884 |
|
57 |
+
| No log | 2.0 | 138 | 0.0312 | 87.2197 | 86.9176 | 87.1927 | 87.1243 | 70.0 |
|
58 |
+
| No log | 3.0 | 207 | 0.0357 | 87.3839 | 87.2143 | 87.4316 | 87.3834 | 68.0580 |
|
59 |
+
| No log | 4.0 | 276 | 0.0334 | 87.8426 | 87.5124 | 87.8504 | 87.7767 | 68.0580 |
|
60 |
+
| No log | 5.0 | 345 | 0.0330 | 89.2541 | 88.8329 | 89.2476 | 89.1951 | 65.8551 |
|
61 |
+
| No log | 6.0 | 414 | 0.0352 | 89.8437 | 89.6094 | 90.0088 | 89.8354 | 67.9565 |
|
62 |
+
| No log | 7.0 | 483 | 0.0351 | 87.6113 | 87.1275 | 87.5987 | 87.4656 | 68.8841 |
|
63 |
+
| 0.0508 | 8.0 | 552 | 0.0346 | 90.0332 | 89.523 | 89.93 | 89.9648 | 64.9275 |
|
64 |
+
| 0.0508 | 9.0 | 621 | 0.0341 | 90.2056 | 89.7318 | 90.0764 | 90.1856 | 60.2174 |
|
65 |
+
| 0.0508 | 10.0 | 690 | 0.0405 | 90.2441 | 89.7403 | 90.1241 | 90.1975 | 62.4928 |
|
66 |
+
| 0.0508 | 11.0 | 759 | 0.0422 | 89.9563 | 89.3932 | 89.8517 | 89.919 | 62.6232 |
|
67 |
+
| 0.0508 | 12.0 | 828 | 0.0462 | 88.9553 | 88.5149 | 88.8596 | 88.8863 | 64.5507 |
|
68 |
+
| 0.0508 | 13.0 | 897 | 0.0462 | 88.3505 | 87.8014 | 88.2999 | 88.1348 | 68.6087 |
|
69 |
+
| 0.0508 | 14.0 | 966 | 0.0453 | 89.2841 | 88.7915 | 89.0835 | 89.1838 | 63.7971 |
|
70 |
+
| 0.0047 | 15.0 | 1035 | 0.0475 | 89.207 | 88.8346 | 89.1459 | 89.1182 | 65.4348 |
|
71 |
+
| 0.0047 | 16.0 | 1104 | 0.0526 | 89.7978 | 89.3703 | 89.7601 | 89.7866 | 65.9275 |
|
72 |
+
| 0.0047 | 17.0 | 1173 | 0.0517 | 88.0891 | 87.7321 | 88.1064 | 88.0137 | 66.4058 |
|
73 |
+
| 0.0047 | 18.0 | 1242 | 0.0503 | 90.3002 | 89.7609 | 90.1585 | 90.218 | 62.1014 |
|
74 |
+
| 0.0047 | 19.0 | 1311 | 0.0545 | 88.9807 | 88.5391 | 88.8142 | 88.8417 | 65.6957 |
|
75 |
+
| 0.0047 | 20.0 | 1380 | 0.0547 | 89.2547 | 88.8381 | 89.1517 | 89.158 | 65.1739 |
|
76 |
+
| 0.0047 | 21.0 | 1449 | 0.0560 | 88.2792 | 87.9155 | 88.2849 | 88.1559 | 66.0870 |
|
77 |
+
| 0.0019 | 22.0 | 1518 | 0.0575 | 88.0891 | 87.7321 | 88.1064 | 88.0137 | 66.4058 |
|
78 |
+
| 0.0019 | 23.0 | 1587 | 0.0576 | 87.7192 | 87.309 | 87.7299 | 87.5507 | 66.0435 |
|
79 |
+
| 0.0019 | 24.0 | 1656 | 0.0558 | 89.0175 | 88.5301 | 88.8811 | 88.906 | 64.1594 |
|
80 |
+
| 0.0019 | 25.0 | 1725 | 0.0561 | 89.0175 | 88.5301 | 88.8811 | 88.906 | 64.1594 |
|
81 |
+
| 0.0019 | 26.0 | 1794 | 0.0559 | 90.1169 | 89.6101 | 89.9618 | 90.0139 | 62.4203 |
|
82 |
+
| 0.0019 | 27.0 | 1863 | 0.0569 | 89.1468 | 88.6354 | 89.0016 | 89.0138 | 63.7246 |
|
83 |
+
| 0.0019 | 28.0 | 1932 | 0.0562 | 89.1468 | 88.6354 | 89.0016 | 89.0138 | 63.7246 |
|
84 |
+
| 0.0013 | 29.0 | 2001 | 0.0563 | 89.1468 | 88.6354 | 89.0016 | 89.0138 | 63.7246 |
|
85 |
+
| 0.0013 | 30.0 | 2070 | 0.0564 | 89.1468 | 88.6354 | 89.0016 | 89.0138 | 63.7246 |
|
86 |
|
87 |
|
88 |
### Framework versions
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 891644712
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5145bac43f06ff212b5fecf4a2ed05638e855a0b5551c0b9af1a062efeba7de3
|
3 |
size 891644712
|