DiwasDiwas commited on
Commit
8a9b9b0
1 Parent(s): e31a5ac

DiwasDiwas/t5-small-MedicoSummarizer

Browse files
README.md CHANGED
@@ -1,19 +1,28 @@
1
  ---
 
 
2
  tags:
3
- - generated_from_keras_callback
 
 
4
  model-index:
5
  - name: t5-small-MedicoSummarizer
6
  results: []
7
  ---
8
 
9
- <!-- This model card has been generated automatically according to the information Keras had access to. You should
10
- probably proofread and complete it, then remove this comment. -->
11
 
12
  # t5-small-MedicoSummarizer
13
 
14
- This model was trained from scratch on an unknown dataset.
15
  It achieves the following results on the evaluation set:
16
-
 
 
 
 
 
17
 
18
  ## Model description
19
 
@@ -32,16 +41,34 @@ More information needed
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
- - optimizer: None
36
- - training_precision: float32
 
 
 
 
 
 
37
 
38
  ### Training results
39
 
 
 
 
 
 
 
 
 
 
 
 
 
40
 
41
 
42
  ### Framework versions
43
 
44
  - Transformers 4.35.2
45
- - TensorFlow 2.14.0
46
  - Datasets 2.15.0
47
  - Tokenizers 0.15.0
 
1
  ---
2
+ license: apache-2.0
3
+ base_model: t5-small
4
  tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - rouge
8
  model-index:
9
  - name: t5-small-MedicoSummarizer
10
  results: []
11
  ---
12
 
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # t5-small-MedicoSummarizer
17
 
18
+ This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 2.9721
21
+ - Rouge1: 0.3222
22
+ - Rouge2: 0.0728
23
+ - Rougel: 0.1933
24
+ - Rougelsum: 0.1932
25
+ - Gen Len: 122.633
26
 
27
  ## Model description
28
 
 
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
+ - learning_rate: 2e-05
45
+ - train_batch_size: 16
46
+ - eval_batch_size: 16
47
+ - seed: 42
48
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
+ - lr_scheduler_type: linear
50
+ - num_epochs: 10
51
+ - mixed_precision_training: Native AMP
52
 
53
  ### Training results
54
 
55
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
56
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
57
+ | 3.4732 | 1.0 | 625 | 3.1206 | 0.295 | 0.0629 | 0.1743 | 0.1743 | 117.058 |
58
+ | 3.2979 | 2.0 | 1250 | 3.0677 | 0.3031 | 0.0677 | 0.1821 | 0.1821 | 116.531 |
59
+ | 3.2329 | 3.0 | 1875 | 3.0342 | 0.31 | 0.0696 | 0.1862 | 0.1861 | 119.239 |
60
+ | 3.1848 | 4.0 | 2500 | 3.0100 | 0.3122 | 0.0699 | 0.186 | 0.1858 | 119.044 |
61
+ | 3.1689 | 5.0 | 3125 | 2.9970 | 0.3188 | 0.0713 | 0.1908 | 0.1907 | 122.084 |
62
+ | 3.153 | 6.0 | 3750 | 2.9855 | 0.3187 | 0.0703 | 0.1927 | 0.1926 | 122.33 |
63
+ | 3.1564 | 7.0 | 4375 | 2.9801 | 0.32 | 0.0706 | 0.1929 | 0.1928 | 122.479 |
64
+ | 3.136 | 8.0 | 5000 | 2.9747 | 0.3228 | 0.0727 | 0.1936 | 0.1935 | 122.804 |
65
+ | 3.1248 | 9.0 | 5625 | 2.9726 | 0.321 | 0.0721 | 0.1926 | 0.1925 | 122.685 |
66
+ | 3.1309 | 10.0 | 6250 | 2.9721 | 0.3222 | 0.0728 | 0.1933 | 0.1932 | 122.633 |
67
 
68
 
69
  ### Framework versions
70
 
71
  - Transformers 4.35.2
72
+ - Pytorch 2.1.0+cu118
73
  - Datasets 2.15.0
74
  - Tokenizers 0.15.0
all_results.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
  "epoch": 10.0,
3
- "eval_gen_len": 122.957,
4
- "eval_loss": 2.9784293174743652,
5
- "eval_rouge1": 0.326,
6
- "eval_rouge2": 0.0772,
7
- "eval_rougeL": 0.1955,
8
- "eval_rougeLsum": 0.1956,
9
- "eval_runtime": 166.3431,
10
- "eval_samples_per_second": 6.012,
11
- "eval_steps_per_second": 0.379
12
  }
 
1
  {
2
  "epoch": 10.0,
3
+ "eval_gen_len": 122.633,
4
+ "eval_loss": 2.972090721130371,
5
+ "eval_rouge1": 0.3222,
6
+ "eval_rouge2": 0.0728,
7
+ "eval_rougeL": 0.1933,
8
+ "eval_rougeLsum": 0.1932,
9
+ "eval_runtime": 169.7847,
10
+ "eval_samples_per_second": 5.89,
11
+ "eval_steps_per_second": 0.371
12
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e186e4cef0e63e5fe7ed9db4d33c4087561a9b79f66d9e481b9fa55cd65fad6a
3
- size 242041904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49a8c9343f09310cec45c4402da19adfe439523c11c48479060997b3c3fd3671
3
+ size 242041896
runs/Dec11_23-40-51_b90ba2bbe6f9/events.out.tfevents.1702338060.b90ba2bbe6f9.525.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41e7b2b658597a0f3073d63a2ba65874dc26a802ca17e482ee55e1cb341f8c1d
3
+ size 12707
runs/Dec11_23-40-51_b90ba2bbe6f9/events.out.tfevents.1702345311.b90ba2bbe6f9.525.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d321670daf5a94dab32ddac03f025e02ad1a6e9ca1f897878daa8f7dc08674fd
3
+ size 613
special_tokens_map.json CHANGED
@@ -101,25 +101,7 @@
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
- "eos_token": {
105
- "content": "</s>",
106
- "lstrip": false,
107
- "normalized": false,
108
- "rstrip": false,
109
- "single_word": false
110
- },
111
- "pad_token": {
112
- "content": "<pad>",
113
- "lstrip": false,
114
- "normalized": false,
115
- "rstrip": false,
116
- "single_word": false
117
- },
118
- "unk_token": {
119
- "content": "<unk>",
120
- "lstrip": false,
121
- "normalized": false,
122
- "rstrip": false,
123
- "single_word": false
124
- }
125
  }
 
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
+ "eos_token": "</s>",
105
+ "pad_token": "<pad>",
106
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
107
  }
tokenizer_config.json CHANGED
@@ -931,12 +931,8 @@
931
  "clean_up_tokenization_spaces": true,
932
  "eos_token": "</s>",
933
  "extra_ids": 100,
934
- "max_length": 128,
935
  "model_max_length": 512,
936
  "pad_token": "<pad>",
937
- "stride": 0,
938
  "tokenizer_class": "T5Tokenizer",
939
- "truncation_side": "right",
940
- "truncation_strategy": "longest_first",
941
  "unk_token": "<unk>"
942
  }
 
931
  "clean_up_tokenization_spaces": true,
932
  "eos_token": "</s>",
933
  "extra_ids": 100,
 
934
  "model_max_length": 512,
935
  "pad_token": "<pad>",
 
936
  "tokenizer_class": "T5Tokenizer",
 
 
937
  "unk_token": "<unk>"
938
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0427ff56d63b412706fb908700359a5bcb36d2f85207da5887b17e20650b185
3
- size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1a30541da037788c1b3cb624e1eccd2d8c54b283a414f9fc1258837dec33647
3
+ size 4856