GlycerinLOL commited on
Commit
c33e842
1 Parent(s): 4f34701

Model save

Browse files
README.md ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: google/pegasus-xsum
3
+ tags:
4
+ - generated_from_trainer
5
+ datasets:
6
+ - samsum
7
+ metrics:
8
+ - rouge
9
+ - precision
10
+ - recall
11
+ - f1
12
+ model-index:
13
+ - name: Pegasus_xsum_samsum
14
+ results:
15
+ - task:
16
+ name: Sequence-to-sequence Language Modeling
17
+ type: text2text-generation
18
+ dataset:
19
+ name: samsum
20
+ type: samsum
21
+ config: samsum
22
+ split: validation
23
+ args: samsum
24
+ metrics:
25
+ - name: Rouge1
26
+ type: rouge
27
+ value: 0.5072
28
+ - name: Precision
29
+ type: precision
30
+ value: 0.9247
31
+ - name: Recall
32
+ type: recall
33
+ value: 0.9099
34
+ - name: F1
35
+ type: f1
36
+ value: 0.917
37
+ ---
38
+
39
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
40
+ should probably proofread and complete it, then remove this comment. -->
41
+
42
+ # Pegasus_xsum_samsum
43
+
44
+ This model is a fine-tuned version of [google/pegasus-xsum](https://huggingface.co/google/pegasus-xsum) on the samsum dataset.
45
+ It achieves the following results on the evaluation set:
46
+ - Loss: 1.4709
47
+ - Rouge1: 0.5072
48
+ - Rouge2: 0.2631
49
+ - Rougel: 0.4243
50
+ - Rougelsum: 0.4244
51
+ - Gen Len: 19.1479
52
+ - Precision: 0.9247
53
+ - Recall: 0.9099
54
+ - F1: 0.917
55
+
56
+ ## Model description
57
+
58
+ More information needed
59
+
60
+ ## Intended uses & limitations
61
+
62
+ More information needed
63
+
64
+ ## Training and evaluation data
65
+
66
+ More information needed
67
+
68
+ ## Training procedure
69
+
70
+ ### Training hyperparameters
71
+
72
+ The following hyperparameters were used during training:
73
+ - learning_rate: 2e-05
74
+ - train_batch_size: 4
75
+ - eval_batch_size: 4
76
+ - seed: 42
77
+ - gradient_accumulation_steps: 4
78
+ - total_train_batch_size: 16
79
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
80
+ - lr_scheduler_type: linear
81
+ - num_epochs: 4
82
+ - mixed_precision_training: Native AMP
83
+
84
+ ### Training results
85
+
86
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Precision | Recall | F1 |
87
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:---------:|:------:|:------:|
88
+ | 1.9542 | 1.0 | 920 | 1.5350 | 0.4928 | 0.2436 | 0.4085 | 0.4086 | 18.5672 | 0.9229 | 0.9074 | 0.9149 |
89
+ | 1.6331 | 2.0 | 1841 | 1.4914 | 0.5037 | 0.257 | 0.4202 | 0.4206 | 18.8154 | 0.9246 | 0.9092 | 0.9166 |
90
+ | 1.5694 | 3.0 | 2762 | 1.4761 | 0.5071 | 0.259 | 0.4212 | 0.4214 | 19.4487 | 0.9241 | 0.9103 | 0.917 |
91
+ | 1.5374 | 4.0 | 3680 | 1.4709 | 0.5072 | 0.2631 | 0.4243 | 0.4244 | 19.1479 | 0.9247 | 0.9099 | 0.917 |
92
+
93
+
94
+ ### Framework versions
95
+
96
+ - Transformers 4.36.0
97
+ - Pytorch 2.0.1+cu117
98
+ - Datasets 2.14.5
99
+ - Tokenizers 0.15.0
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 0,
3
+ "decoder_start_token_id": 0,
4
+ "eos_token_id": 1,
5
+ "forced_eos_token_id": 1,
6
+ "length_penalty": 0.6,
7
+ "max_length": 64,
8
+ "num_beams": 8,
9
+ "pad_token_id": 0,
10
+ "transformers_version": "4.36.0"
11
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42c28efe74cc339edb90ec0d9e446c806d8af9d331f4041d8bfe61b85ffe7c9c
3
  size 2279458540
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14147f92302869d444c720356b6c1263f091e779662754de20a5bce0b2eb0ee8
3
  size 2279458540
runs/Feb02_13-43-49_o8amirctr1706761897220-scx9w/events.out.tfevents.1706852634.o8amirctr1706761897220-scx9w.28324.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0c7160e7db54d16ce8267382c3deb946ba292f772319c82fd397ceff82b83c5
3
- size 8393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2caadc6e993654c43fad2a2f1a52febf281d1097b5ad573259145f6110e1c4b9
3
+ size 9421