tanatapanun commited on
Commit
a466619
1 Parent(s): 5608404

Model save

Browse files
README.md ADDED
@@ -0,0 +1,84 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: GanjinZero/biobart-base
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - rouge
8
+ model-index:
9
+ - name: fine-tuned-BioBART-20-epochs-1024-input-128-output
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # fine-tuned-BioBART-20-epochs-1024-input-128-output
17
+
18
+ This model is a fine-tuned version of [GanjinZero/biobart-base](https://huggingface.co/GanjinZero/biobart-base) on the None dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 1.6050
21
+ - Rouge1: 0.1704
22
+ - Rouge2: 0.0496
23
+ - Rougel: 0.138
24
+ - Rougelsum: 0.1356
25
+ - Gen Len: 34.1
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 0.0001
45
+ - train_batch_size: 8
46
+ - eval_batch_size: 8
47
+ - seed: 42
48
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
+ - lr_scheduler_type: linear
50
+ - lr_scheduler_warmup_ratio: 0.1
51
+ - num_epochs: 20
52
+
53
+ ### Training results
54
+
55
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
56
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
57
+ | No log | 1.0 | 151 | 6.5303 | 0.0 | 0.0 | 0.0 | 0.0 | 12.58 |
58
+ | No log | 2.0 | 302 | 1.9967 | 0.1046 | 0.0318 | 0.0908 | 0.0904 | 26.22 |
59
+ | No log | 3.0 | 453 | 1.6736 | 0.0447 | 0.0076 | 0.036 | 0.0353 | 15.66 |
60
+ | 4.5402 | 4.0 | 604 | 1.5728 | 0.1397 | 0.0344 | 0.1068 | 0.1079 | 34.51 |
61
+ | 4.5402 | 5.0 | 755 | 1.5231 | 0.1675 | 0.0345 | 0.1325 | 0.1328 | 34.4 |
62
+ | 4.5402 | 6.0 | 906 | 1.4986 | 0.1195 | 0.0287 | 0.0863 | 0.0873 | 38.66 |
63
+ | 1.1958 | 7.0 | 1057 | 1.4791 | 0.1478 | 0.0379 | 0.1172 | 0.1176 | 35.41 |
64
+ | 1.1958 | 8.0 | 1208 | 1.4802 | 0.1459 | 0.0368 | 0.1066 | 0.108 | 32.5 |
65
+ | 1.1958 | 9.0 | 1359 | 1.4841 | 0.1687 | 0.0289 | 0.1342 | 0.1345 | 30.89 |
66
+ | 0.7933 | 10.0 | 1510 | 1.5005 | 0.1457 | 0.035 | 0.1125 | 0.1103 | 34.3 |
67
+ | 0.7933 | 11.0 | 1661 | 1.5101 | 0.1808 | 0.0364 | 0.1498 | 0.1505 | 31.33 |
68
+ | 0.7933 | 12.0 | 1812 | 1.5262 | 0.1882 | 0.0419 | 0.1553 | 0.1549 | 31.65 |
69
+ | 0.7933 | 13.0 | 1963 | 1.5481 | 0.167 | 0.032 | 0.1381 | 0.139 | 31.04 |
70
+ | 0.5232 | 14.0 | 2114 | 1.5494 | 0.1723 | 0.0442 | 0.1407 | 0.138 | 34.88 |
71
+ | 0.5232 | 15.0 | 2265 | 1.5590 | 0.1801 | 0.0318 | 0.142 | 0.1413 | 37.99 |
72
+ | 0.5232 | 16.0 | 2416 | 1.5829 | 0.1608 | 0.0353 | 0.1249 | 0.1249 | 33.97 |
73
+ | 0.3565 | 17.0 | 2567 | 1.5837 | 0.1535 | 0.0354 | 0.1159 | 0.115 | 35.96 |
74
+ | 0.3565 | 18.0 | 2718 | 1.5977 | 0.1565 | 0.0349 | 0.1244 | 0.1227 | 34.29 |
75
+ | 0.3565 | 19.0 | 2869 | 1.6002 | 0.169 | 0.0428 | 0.1358 | 0.1331 | 34.84 |
76
+ | 0.2734 | 20.0 | 3020 | 1.6050 | 0.1704 | 0.0496 | 0.138 | 0.1356 | 34.1 |
77
+
78
+
79
+ ### Framework versions
80
+
81
+ - Transformers 4.36.2
82
+ - Pytorch 1.12.1+cu113
83
+ - Datasets 2.16.1
84
+ - Tokenizers 0.15.0
generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 0,
3
+ "decoder_start_token_id": 2,
4
+ "early_stopping": true,
5
+ "eos_token_id": 2,
6
+ "forced_eos_token_id": 2,
7
+ "max_length": 128,
8
+ "no_repeat_ngram_size": 3,
9
+ "num_beams": 4,
10
+ "pad_token_id": 1,
11
+ "transformers_version": "4.36.2"
12
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4bbe800ddd5fa60f2c0a9f742ca7349d4b1ce17715a98e5a02d819f6b21b7acc
3
  size 557912620
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecc3b313cb09cada4fd677204d1d24370970a3c7f5ce7a0f574412dee7cb779b
3
  size 557912620
runs/Jan25_12-38-46_william-gpu-3090-10-8vlnc/events.out.tfevents.1706186328.william-gpu-3090-10-8vlnc.7963.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8334250208ae2b1846b675862c00cc890391989ecb7c7ad4933dea388c476258
3
- size 16642
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65ebf0edc732f75d5a70a2c0dbc8b061ebb1a4f4baa9ee7b3b7ccdb2244f4e8d
3
+ size 17521