ubaada commited on
Commit
5c5032c
1 Parent(s): 48d5c9c

ubaada/pegasus-x-large-booksum-16k

Browse files
Files changed (3) hide show
  1. README.md +12 -6
  2. model.safetensors +1 -1
  3. training_args.bin +1 -1
README.md CHANGED
@@ -17,10 +17,10 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 1.9565
21
- - Rouge1: 0.1423
22
- - Rouge2: 0.0178
23
- - Rougel: 0.0928
24
 
25
  ## Model description
26
 
@@ -39,7 +39,7 @@ More information needed
39
  ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
- - learning_rate: 8e-05
43
  - train_batch_size: 8
44
  - eval_batch_size: 1
45
  - seed: 42
@@ -50,7 +50,7 @@ The following hyperparameters were used during training:
50
  - total_eval_batch_size: 2
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
- - num_epochs: 3
54
 
55
  ### Training results
56
 
@@ -59,6 +59,12 @@ The following hyperparameters were used during training:
59
  | 1.5731 | 0.9996 | 600 | 1.9730 | 0.1342 | 0.0151 | 0.0912 |
60
  | 1.3694 | 1.9996 | 1200 | 1.9623 | 0.1371 | 0.0175 | 0.0909 |
61
  | 1.9561 | 2.9992 | 1800 | 1.9565 | 0.1423 | 0.0178 | 0.0928 |
 
 
 
 
 
 
62
 
63
 
64
  ### Framework versions
 
17
 
18
  This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 1.9401
21
+ - Rouge1: 0.1934
22
+ - Rouge2: 0.0269
23
+ - Rougel: 0.1151
24
 
25
  ## Model description
26
 
 
39
  ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
+ - learning_rate: 4e-05
43
  - train_batch_size: 8
44
  - eval_batch_size: 1
45
  - seed: 42
 
50
  - total_eval_batch_size: 2
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
+ - num_epochs: 9
54
 
55
  ### Training results
56
 
 
59
  | 1.5731 | 0.9996 | 600 | 1.9730 | 0.1342 | 0.0151 | 0.0912 |
60
  | 1.3694 | 1.9996 | 1200 | 1.9623 | 0.1371 | 0.0175 | 0.0909 |
61
  | 1.9561 | 2.9992 | 1800 | 1.9565 | 0.1423 | 0.0178 | 0.0928 |
62
+ | 1.0882 | 3.9996 | 2400 | 1.9548 | 0.1417 | 0.0186 | 0.0900 |
63
+ | 1.4872 | 4.9992 | 3000 | 1.9412 | 0.1581 | 0.0212 | 0.1006 |
64
+ | 1.4126 | 5.9988 | 3600 | 1.9486 | 0.1589 | 0.0188 | 0.0986 |
65
+ | 1.1634 | 7.0 | 4201 | 1.9464 | 0.1756 | 0.0229 | 0.1046 |
66
+ | 0.9541 | 7.9996 | 4801 | 1.9401 | 0.1791 | 0.0243 | 0.1078 |
67
+ | 0.9153 | 8.9975 | 5400 | 1.9401 | 0.1934 | 0.0269 | 0.1151 |
68
 
69
 
70
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84f21eff3ce6deda24ca6e879a245eb12d871429288d8890407f15c1c1bb1a82
3
  size 990386200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e40fbb8361d246d53f4edae35d42da61366d34201eec62f22ac6d4b04a2e99b2
3
  size 990386200
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:296154d90ec6ea346feed2cef6a2e1a56591d88c5f82648a51e3b732668f6e6b
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c611d9f9ff8ec098da6eb03453782a6458e06a90779c1927d1d621248f690dfb
3
  size 6776