jpodivin commited on
Commit
9746291
1 Parent(s): 6838f76

End of training

Browse files
Files changed (5) hide show
  1. README.md +14 -2
  2. all_results.json +13 -13
  3. eval_results.json +9 -9
  4. train_results.json +4 -4
  5. trainer_state.json +49 -49
README.md CHANGED
@@ -3,11 +3,23 @@ license: apache-2.0
3
  base_model: google-t5/t5-base
4
  tags:
5
  - generated_from_trainer
 
 
6
  metrics:
7
  - rouge
8
  model-index:
9
  - name: pep_summarization
10
- results: []
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -15,7 +27,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # pep_summarization
17
 
18
- This model is a fine-tuned version of [google-t5/t5-base](https://huggingface.co/google-t5/t5-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
  - Loss: 0.9288
21
  - Rouge1: 39.7591
 
3
  base_model: google-t5/t5-base
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - fedora-copr/pep-sum
8
  metrics:
9
  - rouge
10
  model-index:
11
  - name: pep_summarization
12
+ results:
13
+ - task:
14
+ name: Summarization
15
+ type: summarization
16
+ dataset:
17
+ name: fedora-copr/pep-sum
18
+ type: fedora-copr/pep-sum
19
+ metrics:
20
+ - name: Rouge1
21
+ type: rouge
22
+ value: 39.7591
23
  ---
24
 
25
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
27
 
28
  # pep_summarization
29
 
30
+ This model is a fine-tuned version of [google-t5/t5-base](https://huggingface.co/google-t5/t5-base) on the fedora-copr/pep-sum dataset.
31
  It achieves the following results on the evaluation set:
32
  - Loss: 0.9288
33
  - Rouge1: 39.7591
all_results.json CHANGED
@@ -1,18 +1,18 @@
1
  {
2
  "epoch": 5.0,
3
- "eval_gen_len": 62.05797101449275,
4
- "eval_loss": 0.0359114333987236,
5
- "eval_rouge1": 84.8601,
6
- "eval_rouge2": 84.2185,
7
- "eval_rougeL": 84.7842,
8
- "eval_rougeLsum": 84.8657,
9
- "eval_runtime": 9.4102,
10
  "eval_samples": 69,
11
- "eval_samples_per_second": 7.332,
12
- "eval_steps_per_second": 1.913,
13
- "train_loss": 0.16054312664529552,
14
- "train_runtime": 120.7301,
15
  "train_samples": 276,
16
- "train_samples_per_second": 11.43,
17
- "train_steps_per_second": 2.858
18
  }
 
1
  {
2
  "epoch": 5.0,
3
+ "eval_gen_len": 68.04347826086956,
4
+ "eval_loss": 0.9288389682769775,
5
+ "eval_rouge1": 39.7591,
6
+ "eval_rouge2": 24.1983,
7
+ "eval_rougeL": 32.2869,
8
+ "eval_rougeLsum": 36.1493,
9
+ "eval_runtime": 9.8095,
10
  "eval_samples": 69,
11
+ "eval_samples_per_second": 7.034,
12
+ "eval_steps_per_second": 1.835,
13
+ "train_loss": 0.8950336843297102,
14
+ "train_runtime": 114.301,
15
  "train_samples": 276,
16
+ "train_samples_per_second": 12.073,
17
+ "train_steps_per_second": 3.018
18
  }
eval_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 5.0,
3
- "eval_gen_len": 62.05797101449275,
4
- "eval_loss": 0.0359114333987236,
5
- "eval_rouge1": 84.8601,
6
- "eval_rouge2": 84.2185,
7
- "eval_rougeL": 84.7842,
8
- "eval_rougeLsum": 84.8657,
9
- "eval_runtime": 9.4102,
10
  "eval_samples": 69,
11
- "eval_samples_per_second": 7.332,
12
- "eval_steps_per_second": 1.913
13
  }
 
1
  {
2
  "epoch": 5.0,
3
+ "eval_gen_len": 68.04347826086956,
4
+ "eval_loss": 0.9288389682769775,
5
+ "eval_rouge1": 39.7591,
6
+ "eval_rouge2": 24.1983,
7
+ "eval_rougeL": 32.2869,
8
+ "eval_rougeLsum": 36.1493,
9
+ "eval_runtime": 9.8095,
10
  "eval_samples": 69,
11
+ "eval_samples_per_second": 7.034,
12
+ "eval_steps_per_second": 1.835
13
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 5.0,
3
- "train_loss": 0.16054312664529552,
4
- "train_runtime": 120.7301,
5
  "train_samples": 276,
6
- "train_samples_per_second": 11.43,
7
- "train_steps_per_second": 2.858
8
  }
 
1
  {
2
  "epoch": 5.0,
3
+ "train_loss": 0.8950336843297102,
4
+ "train_runtime": 114.301,
5
  "train_samples": 276,
6
+ "train_samples_per_second": 12.073,
7
+ "train_steps_per_second": 3.018
8
  }
trainer_state.json CHANGED
@@ -10,77 +10,77 @@
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
- "eval_gen_len": 49.08695652173913,
14
- "eval_loss": 0.2411632537841797,
15
- "eval_rouge1": 74.9533,
16
- "eval_rouge2": 71.1831,
17
- "eval_rougeL": 74.2512,
18
- "eval_rougeLsum": 74.5914,
19
- "eval_runtime": 9.2355,
20
- "eval_samples_per_second": 7.471,
21
- "eval_steps_per_second": 1.949,
22
  "step": 69
23
  },
24
  {
25
  "epoch": 2.0,
26
- "eval_gen_len": 88.17391304347827,
27
- "eval_loss": 0.08010436594486237,
28
- "eval_rouge1": 78.7144,
29
- "eval_rouge2": 77.7186,
30
- "eval_rougeL": 78.6962,
31
- "eval_rougeLsum": 78.6452,
32
- "eval_runtime": 11.7911,
33
- "eval_samples_per_second": 5.852,
34
- "eval_steps_per_second": 1.527,
35
  "step": 138
36
  },
37
  {
38
  "epoch": 3.0,
39
- "eval_gen_len": 73.0,
40
- "eval_loss": 0.08717019110918045,
41
- "eval_rouge1": 83.0112,
42
- "eval_rouge2": 82.1689,
43
- "eval_rougeL": 82.8847,
44
- "eval_rougeLsum": 82.9534,
45
- "eval_runtime": 10.9663,
46
- "eval_samples_per_second": 6.292,
47
- "eval_steps_per_second": 1.641,
48
  "step": 207
49
  },
50
  {
51
  "epoch": 4.0,
52
- "eval_gen_len": 60.88405797101449,
53
- "eval_loss": 0.06106267869472504,
54
- "eval_rouge1": 86.2191,
55
- "eval_rouge2": 85.8311,
56
- "eval_rougeL": 86.308,
57
- "eval_rougeLsum": 86.4542,
58
- "eval_runtime": 9.6061,
59
- "eval_samples_per_second": 7.183,
60
- "eval_steps_per_second": 1.874,
61
  "step": 276
62
  },
63
  {
64
  "epoch": 5.0,
65
- "eval_gen_len": 62.05797101449275,
66
- "eval_loss": 0.0359114333987236,
67
- "eval_rouge1": 84.8601,
68
- "eval_rouge2": 84.2185,
69
- "eval_rougeL": 84.7842,
70
- "eval_rougeLsum": 84.8657,
71
- "eval_runtime": 9.6516,
72
- "eval_samples_per_second": 7.149,
73
- "eval_steps_per_second": 1.865,
74
  "step": 345
75
  },
76
  {
77
  "epoch": 5.0,
78
  "step": 345,
79
  "total_flos": 1680723777945600.0,
80
- "train_loss": 0.16054312664529552,
81
- "train_runtime": 120.7301,
82
- "train_samples_per_second": 11.43,
83
- "train_steps_per_second": 2.858
84
  }
85
  ],
86
  "logging_steps": 500,
 
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
+ "eval_gen_len": 68.04347826086956,
14
+ "eval_loss": 0.9288747310638428,
15
+ "eval_rouge1": 39.7591,
16
+ "eval_rouge2": 24.1983,
17
+ "eval_rougeL": 32.2869,
18
+ "eval_rougeLsum": 36.1493,
19
+ "eval_runtime": 9.3009,
20
+ "eval_samples_per_second": 7.419,
21
+ "eval_steps_per_second": 1.935,
22
  "step": 69
23
  },
24
  {
25
  "epoch": 2.0,
26
+ "eval_gen_len": 68.04347826086956,
27
+ "eval_loss": 0.9288538694381714,
28
+ "eval_rouge1": 39.7591,
29
+ "eval_rouge2": 24.1983,
30
+ "eval_rougeL": 32.2869,
31
+ "eval_rougeLsum": 36.1493,
32
+ "eval_runtime": 9.2625,
33
+ "eval_samples_per_second": 7.449,
34
+ "eval_steps_per_second": 1.943,
35
  "step": 138
36
  },
37
  {
38
  "epoch": 3.0,
39
+ "eval_gen_len": 68.04347826086956,
40
+ "eval_loss": 0.9288433790206909,
41
+ "eval_rouge1": 39.7591,
42
+ "eval_rouge2": 24.1983,
43
+ "eval_rougeL": 32.2869,
44
+ "eval_rougeLsum": 36.1493,
45
+ "eval_runtime": 9.3044,
46
+ "eval_samples_per_second": 7.416,
47
+ "eval_steps_per_second": 1.935,
48
  "step": 207
49
  },
50
  {
51
  "epoch": 4.0,
52
+ "eval_gen_len": 68.04347826086956,
53
+ "eval_loss": 0.9288394451141357,
54
+ "eval_rouge1": 39.7591,
55
+ "eval_rouge2": 24.1983,
56
+ "eval_rougeL": 32.2869,
57
+ "eval_rougeLsum": 36.1493,
58
+ "eval_runtime": 9.3367,
59
+ "eval_samples_per_second": 7.39,
60
+ "eval_steps_per_second": 1.928,
61
  "step": 276
62
  },
63
  {
64
  "epoch": 5.0,
65
+ "eval_gen_len": 68.04347826086956,
66
+ "eval_loss": 0.9288389682769775,
67
+ "eval_rouge1": 39.7591,
68
+ "eval_rouge2": 24.1983,
69
+ "eval_rougeL": 32.2869,
70
+ "eval_rougeLsum": 36.1493,
71
+ "eval_runtime": 9.2668,
72
+ "eval_samples_per_second": 7.446,
73
+ "eval_steps_per_second": 1.942,
74
  "step": 345
75
  },
76
  {
77
  "epoch": 5.0,
78
  "step": 345,
79
  "total_flos": 1680723777945600.0,
80
+ "train_loss": 0.8950336843297102,
81
+ "train_runtime": 114.301,
82
+ "train_samples_per_second": 12.073,
83
+ "train_steps_per_second": 3.018
84
  }
85
  ],
86
  "logging_steps": 500,