GlycerinLOL commited on
Commit
4115343
1 Parent(s): 420d47f

Model save

Browse files
README.md ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: google/pegasus-large
3
+ tags:
4
+ - generated_from_trainer
5
+ metrics:
6
+ - rouge
7
+ - precision
8
+ - recall
9
+ - f1
10
+ model-index:
11
+ - name: LLM_Teached_Pegasus_100k_FS
12
+ results: []
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+ # LLM_Teached_Pegasus_100k_FS
19
+
20
+ This model is a fine-tuned version of [google/pegasus-large](https://huggingface.co/google/pegasus-large) on an unknown dataset.
21
+ It achieves the following results on the evaluation set:
22
+ - Loss: 1.4469
23
+ - Rouge1: 0.4939
24
+ - Rouge2: 0.2453
25
+ - Rougel: 0.4133
26
+ - Rougelsum: 0.4134
27
+ - Gen Len: 25.9629
28
+ - Precision: 0.9133
29
+ - Recall: 0.9138
30
+ - F1: 0.9134
31
+
32
+ ## Model description
33
+
34
+ More information needed
35
+
36
+ ## Intended uses & limitations
37
+
38
+ More information needed
39
+
40
+ ## Training and evaluation data
41
+
42
+ More information needed
43
+
44
+ ## Training procedure
45
+
46
+ ### Training hyperparameters
47
+
48
+ The following hyperparameters were used during training:
49
+ - learning_rate: 2e-05
50
+ - train_batch_size: 16
51
+ - eval_batch_size: 16
52
+ - seed: 42
53
+ - gradient_accumulation_steps: 6
54
+ - total_train_batch_size: 96
55
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
56
+ - lr_scheduler_type: linear
57
+ - num_epochs: 16
58
+ - mixed_precision_training: Native AMP
59
+
60
+ ### Training results
61
+
62
+ | Training Loss | Epoch | Step | F1 | Gen Len | Validation Loss | Precision | Recall | Rouge1 | Rouge2 | Rougel | Rougelsum |
63
+ |:-------------:|:-----:|:-----:|:------:|:-------:|:---------------:|:---------:|:------:|:------:|:------:|:------:|:---------:|
64
+ | 1.781 | 2.0 | 1388 | 0.9088 | 26.8891 | 1.5797 | 0.908 | 0.91 | 0.4708 | 0.2219 | 0.3892 | 0.389 |
65
+ | 1.6618 | 3.0 | 2083 | 0.91 | 26.7282 | 1.5411 | 0.9094 | 0.9111 | 0.4776 | 0.2303 | 0.3977 | 0.3973 |
66
+ | 1.626 | 4.0 | 2776 | 0.911 | 26.7596 | 1.5171 | 0.9102 | 0.9121 | 0.4834 | 0.2345 | 0.402 | 0.402 |
67
+ | 1.5918 | 5.0 | 3471 | 0.9112 | 26.6476 | 1.5001 | 0.9106 | 0.9122 | 0.4853 | 0.2365 | 0.4045 | 0.4045 |
68
+ | 1.5586 | 6.0 | 4164 | 0.9116 | 26.7778 | 1.4880 | 0.9108 | 0.9127 | 0.4875 | 0.2373 | 0.4063 | 0.4063 |
69
+ | 1.5375 | 7.0 | 4858 | 1.4768 | 0.4898 | 0.24 | 0.4083 | 0.4083 | 26.3991| 0.9116 | 0.9128 | 0.912 |
70
+ | 1.5146 | 8.0 | 5553 | 1.4686 | 0.4907 | 0.241 | 0.4088 | 0.4089 | 26.156 | 0.9123 | 0.9133 | 0.9126 |
71
+ | 1.5006 | 9.0 | 6247 | 1.4636 | 0.4914 | 0.2419 | 0.4097 | 0.4099 | 26.2629| 0.9122 | 0.9135 | 0.9127 |
72
+ | 1.49 | 10.0 | 6942 | 1.4580 | 0.4911 | 0.2429 | 0.4109 | 0.411 | 26.0273| 0.9125 | 0.9133 | 0.9127 |
73
+ | 1.4749 | 11.0 | 7636 | 1.4546 | 0.4932 | 0.244 | 0.4121 | 0.4123 | 26.2304| 0.9127 | 0.9138 | 0.9131 |
74
+ | 1.4661 | 12.0 | 8331 | 1.4514 | 0.4937 | 0.2448 | 0.4126 | 0.4127 | 25.8778| 0.9133 | 0.9136 | 0.9132 |
75
+ | 1.4575 | 13.0 | 9025 | 1.4499 | 0.4947 | 0.2453 | 0.4139 | 0.414 | 26.1151| 0.913 | 0.914 | 0.9133 |
76
+ | 1.4511 | 14.0 | 9720 | 1.4478 | 0.4939 | 0.2451 | 0.4133 | 0.4134 | 26.0287| 0.9131 | 0.9138 | 0.9133 |
77
+ | 1.4519 | 15.0 | 10414 | 1.4471 | 0.4938 | 0.2451 | 0.4134 | 0.4134 | 25.9078| 0.9132 | 0.9137 | 0.9133 |
78
+ | 1.4439 | 15.99 | 11104 | 1.4469 | 0.4939 | 0.2453 | 0.4133 | 0.4134 | 25.9629| 0.9133 | 0.9138 | 0.9134 |
79
+
80
+
81
+ ### Framework versions
82
+
83
+ - Transformers 4.36.0
84
+ - Pytorch 2.0.1+cu117
85
+ - Datasets 2.14.5
86
+ - Tokenizers 0.15.0
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 0,
3
+ "decoder_start_token_id": 0,
4
+ "eos_token_id": 1,
5
+ "forced_eos_token_id": 1,
6
+ "length_penalty": 0.8,
7
+ "max_length": 256,
8
+ "num_beams": 8,
9
+ "pad_token_id": 0,
10
+ "transformers_version": "4.36.0"
11
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32ca42f41fd84eea8c581c1717ff06971857f13cbe656545240c49e28574b00f
3
  size 2283652852
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4953494d63f199ef2201eef22bddd467dee8ee50d5b6c0052014933986a15967
3
  size 2283652852
runs/Mar15_08-53-06_d9n20yctr1710463501031-gcv7j/events.out.tfevents.1710464000.d9n20yctr1710463501031-gcv7j.2441.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e95949f3ae4dbf2d47ade3c3698040a9d4e5fe55c1eb2f6dee0da0f80c3af22b
3
- size 15089
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cc1c21d66f59dea7abc59ab3af696149293da7c920b5e8885ae627a2709d514
3
+ size 16117