dtruong46me commited on
Commit
21db756
1 Parent(s): cbdf9f8

End of training

Browse files
Files changed (2) hide show
  1. README.md +34 -34
  2. training_args.bin +1 -1
README.md CHANGED
@@ -18,10 +18,10 @@ should probably proofread and complete it, then remove this comment. -->
18
  This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
  - Loss: 0.2540
21
- - Rouge1: 39.369
22
- - Rouge2: 17.7119
23
- - Rougel: 34.2704
24
- - Rougelsum: 36.3125
25
  - Gen Len: 19.97
26
 
27
  ## Model description
@@ -55,36 +55,36 @@ The following hyperparameters were used during training:
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
57
  |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
58
- | 0.3676 | 1.0 | 1557 | 0.2753 | 36.7091 | 14.1132 | 31.2961 | 33.1854 | 19.942 |
59
- | 0.3135 | 2.0 | 3115 | 0.2658 | 37.8359 | 15.2418 | 32.1763 | 34.1253 | 19.97 |
60
- | 0.2992 | 3.0 | 4672 | 0.2596 | 38.3886 | 15.7182 | 32.5452 | 34.5302 | 19.942 |
61
- | 0.2888 | 4.0 | 6230 | 0.2559 | 37.6916 | 15.086 | 32.2052 | 34.2042 | 19.94 |
62
- | 0.281 | 5.0 | 7787 | 0.2549 | 38.4108 | 15.8645 | 32.802 | 34.8545 | 19.952 |
63
- | 0.2734 | 6.0 | 9345 | 0.2533 | 38.7542 | 15.9833 | 33.2205 | 35.2353 | 19.954 |
64
- | 0.2679 | 7.0 | 10902 | 0.2529 | 38.7307 | 16.2092 | 33.1806 | 35.2169 | 19.96 |
65
- | 0.2619 | 8.0 | 12460 | 0.2528 | 39.0498 | 16.462 | 33.784 | 35.7682 | 19.968 |
66
- | 0.2576 | 9.0 | 14017 | 0.2528 | 38.744 | 16.5548 | 33.3367 | 35.3423 | 19.948 |
67
- | 0.253 | 10.0 | 15575 | 0.2523 | 38.5709 | 16.3816 | 33.2589 | 35.1747 | 19.956 |
68
- | 0.2494 | 11.0 | 17132 | 0.2516 | 38.7621 | 16.5616 | 33.4274 | 35.5053 | 19.958 |
69
- | 0.2456 | 12.0 | 18690 | 0.2514 | 38.356 | 16.2738 | 33.1215 | 34.8369 | 19.964 |
70
- | 0.2419 | 13.0 | 20247 | 0.2520 | 38.482 | 16.339 | 32.958 | 35.0048 | 19.972 |
71
- | 0.2388 | 14.0 | 21805 | 0.2514 | 38.9139 | 17.2759 | 33.6644 | 35.5293 | 19.97 |
72
- | 0.2363 | 15.0 | 23362 | 0.2530 | 38.9259 | 16.8451 | 33.5228 | 35.3245 | 19.976 |
73
- | 0.2336 | 16.0 | 24920 | 0.2519 | 38.904 | 17.0261 | 33.8007 | 35.5953 | 19.958 |
74
- | 0.2313 | 17.0 | 26477 | 0.2518 | 38.9099 | 17.1375 | 33.7251 | 35.6641 | 19.97 |
75
- | 0.229 | 18.0 | 28035 | 0.2518 | 38.7184 | 16.751 | 33.4921 | 35.4046 | 19.964 |
76
- | 0.2272 | 19.0 | 29592 | 0.2522 | 39.0706 | 17.0511 | 33.8572 | 35.8149 | 19.964 |
77
- | 0.2252 | 20.0 | 31150 | 0.2527 | 38.7257 | 17.0736 | 33.7989 | 35.5765 | 19.968 |
78
- | 0.2234 | 21.0 | 32707 | 0.2527 | 38.8724 | 17.2135 | 33.8956 | 35.8487 | 19.97 |
79
- | 0.2217 | 22.0 | 34265 | 0.2532 | 38.9055 | 17.0727 | 33.8365 | 35.6688 | 19.968 |
80
- | 0.2206 | 23.0 | 35822 | 0.2521 | 39.4982 | 17.6918 | 34.2308 | 36.2109 | 19.962 |
81
- | 0.2198 | 24.0 | 37380 | 0.2532 | 39.56 | 17.888 | 34.3548 | 36.3312 | 19.964 |
82
- | 0.2184 | 25.0 | 38937 | 0.2533 | 39.2472 | 17.3858 | 33.9809 | 36.0246 | 19.972 |
83
- | 0.2173 | 26.0 | 40495 | 0.2536 | 38.9687 | 17.3762 | 34.0394 | 35.9552 | 19.972 |
84
- | 0.2166 | 27.0 | 42052 | 0.2532 | 39.2298 | 17.7185 | 34.207 | 36.1568 | 19.974 |
85
- | 0.2159 | 28.0 | 43610 | 0.2539 | 39.3157 | 17.7234 | 34.296 | 36.2733 | 19.972 |
86
- | 0.2154 | 29.0 | 45167 | 0.2543 | 39.327 | 17.6213 | 34.2962 | 36.2503 | 19.974 |
87
- | 0.2152 | 29.99 | 46710 | 0.2540 | 39.369 | 17.7119 | 34.2704 | 36.3125 | 19.97 |
88
 
89
 
90
  ### Framework versions
 
18
  This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
  - Loss: 0.2540
21
+ - Rouge1: 39.4088
22
+ - Rouge2: 17.6509
23
+ - Rougel: 34.241
24
+ - Rougelsum: 36.3257
25
  - Gen Len: 19.97
26
 
27
  ## Model description
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
57
  |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
58
+ | 0.3676 | 1.0 | 1557 | 0.2753 | 36.6587 | 14.0807 | 31.2838 | 33.1779 | 19.942 |
59
+ | 0.3135 | 2.0 | 3115 | 0.2658 | 37.8343 | 15.2461 | 32.2261 | 34.1553 | 19.97 |
60
+ | 0.2992 | 3.0 | 4672 | 0.2596 | 38.3851 | 15.6982 | 32.5124 | 34.5772 | 19.942 |
61
+ | 0.2888 | 4.0 | 6230 | 0.2559 | 37.6648 | 15.1146 | 32.1953 | 34.2139 | 19.94 |
62
+ | 0.281 | 5.0 | 7787 | 0.2549 | 38.3654 | 15.8444 | 32.775 | 34.9156 | 19.952 |
63
+ | 0.2734 | 6.0 | 9345 | 0.2533 | 38.7474 | 16.0237 | 33.155 | 35.3048 | 19.954 |
64
+ | 0.2679 | 7.0 | 10902 | 0.2529 | 38.7094 | 16.1904 | 33.2149 | 35.2449 | 19.96 |
65
+ | 0.2619 | 8.0 | 12460 | 0.2528 | 39.034 | 16.4682 | 33.7757 | 35.82 | 19.968 |
66
+ | 0.2576 | 9.0 | 14017 | 0.2528 | 38.769 | 16.5015 | 33.3685 | 35.4211 | 19.948 |
67
+ | 0.253 | 10.0 | 15575 | 0.2523 | 38.5811 | 16.3423 | 33.2559 | 35.2143 | 19.956 |
68
+ | 0.2494 | 11.0 | 17132 | 0.2516 | 38.7084 | 16.5171 | 33.4486 | 35.5503 | 19.958 |
69
+ | 0.2456 | 12.0 | 18690 | 0.2514 | 38.3763 | 16.2338 | 33.1431 | 34.8647 | 19.964 |
70
+ | 0.2419 | 13.0 | 20247 | 0.2520 | 38.455 | 16.2491 | 32.9546 | 35.0263 | 19.972 |
71
+ | 0.2388 | 14.0 | 21805 | 0.2514 | 38.9372 | 17.1821 | 33.6449 | 35.5621 | 19.97 |
72
+ | 0.2363 | 15.0 | 23362 | 0.2530 | 38.9104 | 16.742 | 33.5194 | 35.3391 | 19.976 |
73
+ | 0.2336 | 16.0 | 24920 | 0.2519 | 38.8698 | 16.9396 | 33.7987 | 35.6173 | 19.958 |
74
+ | 0.2313 | 17.0 | 26477 | 0.2518 | 38.8774 | 17.0545 | 33.7151 | 35.6844 | 19.97 |
75
+ | 0.229 | 18.0 | 28035 | 0.2518 | 38.7073 | 16.7039 | 33.4976 | 35.4177 | 19.964 |
76
+ | 0.2272 | 19.0 | 29592 | 0.2522 | 39.0868 | 16.948 | 33.8953 | 35.8788 | 19.964 |
77
+ | 0.2252 | 20.0 | 31150 | 0.2527 | 38.7854 | 16.9882 | 33.8017 | 35.6314 | 19.968 |
78
+ | 0.2234 | 21.0 | 32707 | 0.2527 | 38.9196 | 17.1419 | 33.9139 | 35.8599 | 19.97 |
79
+ | 0.2217 | 22.0 | 34265 | 0.2532 | 38.9227 | 17.0561 | 33.8032 | 35.6876 | 19.968 |
80
+ | 0.2206 | 23.0 | 35822 | 0.2521 | 39.5234 | 17.6253 | 34.2157 | 36.2645 | 19.962 |
81
+ | 0.2198 | 24.0 | 37380 | 0.2532 | 39.6108 | 17.8336 | 34.3222 | 36.3369 | 19.964 |
82
+ | 0.2184 | 25.0 | 38937 | 0.2533 | 39.3052 | 17.2967 | 33.9684 | 36.0207 | 19.972 |
83
+ | 0.2173 | 26.0 | 40495 | 0.2536 | 39.019 | 17.3083 | 34.0561 | 35.9826 | 19.972 |
84
+ | 0.2166 | 27.0 | 42052 | 0.2532 | 39.2553 | 17.6306 | 34.1763 | 36.1479 | 19.974 |
85
+ | 0.2159 | 28.0 | 43610 | 0.2539 | 39.3659 | 17.6526 | 34.276 | 36.2856 | 19.972 |
86
+ | 0.2154 | 29.0 | 45167 | 0.2543 | 39.3868 | 17.5653 | 34.2637 | 36.2704 | 19.974 |
87
+ | 0.2152 | 29.99 | 46710 | 0.2540 | 39.4088 | 17.6509 | 34.241 | 36.3257 | 19.97 |
88
 
89
 
90
  ### Framework versions
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4e100ded8a472b40d70c7af33dfd7f15aa5e718e79382e1e697826db15a0fef
3
  size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba90fb9fcd8a5878da26e645d8d8fa4c6f1fe40f026c7a76a5a5698e831d8aa1
3
  size 4856