spacemanidol commited on
Commit
563d215
1 Parent(s): 3863e23

Upload 12 files

Browse files
README.md CHANGED
@@ -6,7 +6,7 @@ datasets:
6
  metrics:
7
  - rouge
8
  model-index:
9
- - name: base-2-2-t
10
  results:
11
  - task:
12
  name: Summarization
@@ -20,22 +20,22 @@ model-index:
20
  metrics:
21
  - name: Rouge1
22
  type: rouge
23
- value: 5.7782
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
27
  should probably proofread and complete it, then remove this comment. -->
28
 
29
- # base-2-2-t
30
 
31
- This model is a fine-tuned version of [x/base-2-2](https://huggingface.co/x/base-2-2) on the xsum dataset.
32
  It achieves the following results on the evaluation set:
33
- - Loss: nan
34
- - Rouge1: 5.7782
35
- - Rouge2: 0.064
36
- - Rougel: 5.1661
37
- - Rougelsum: 5.1836
38
- - Gen Len: 122.6594
39
 
40
  ## Model description
41
 
@@ -55,15 +55,14 @@ More information needed
55
 
56
  The following hyperparameters were used during training:
57
  - learning_rate: 0.0001
58
- - train_batch_size: 2
59
- - eval_batch_size: 4
60
  - seed: 42
61
- - gradient_accumulation_steps: 32
62
  - total_train_batch_size: 64
63
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
64
  - lr_scheduler_type: constant
65
  - num_epochs: 3.0
66
- - mixed_precision_training: Native AMP
67
 
68
  ### Training results
69
 
@@ -72,6 +71,6 @@ The following hyperparameters were used during training:
72
  ### Framework versions
73
 
74
  - Transformers 4.27.0.dev0
75
- - Pytorch 1.13.0+cu117
76
- - Datasets 2.7.1
77
- - Tokenizers 0.12.1
 
6
  metrics:
7
  - rouge
8
  model-index:
9
+ - name: base-2-2
10
  results:
11
  - task:
12
  name: Summarization
 
20
  metrics:
21
  - name: Rouge1
22
  type: rouge
23
+ value: 32.4824
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
27
  should probably proofread and complete it, then remove this comment. -->
28
 
29
+ # base-2-2
30
 
31
+ This model is a fine-tuned version of [x/base-2-2/](https://huggingface.co/x/base-2-2/) on the xsum dataset.
32
  It achieves the following results on the evaluation set:
33
+ - Loss: 2.2905
34
+ - Rouge1: 32.4824
35
+ - Rouge2: 10.6748
36
+ - Rougel: 25.5004
37
+ - Rougelsum: 25.4918
38
+ - Gen Len: 27.9846
39
 
40
  ## Model description
41
 
 
55
 
56
  The following hyperparameters were used during training:
57
  - learning_rate: 0.0001
58
+ - train_batch_size: 16
59
+ - eval_batch_size: 16
60
  - seed: 42
61
+ - gradient_accumulation_steps: 4
62
  - total_train_batch_size: 64
63
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
64
  - lr_scheduler_type: constant
65
  - num_epochs: 3.0
 
66
 
67
  ### Training results
68
 
 
71
  ### Framework versions
72
 
73
  - Transformers 4.27.0.dev0
74
+ - Pytorch 1.12.1+cu113
75
+ - Datasets 2.10.0
76
+ - Tokenizers 0.13.2
all_results.json CHANGED
@@ -1,18 +1,18 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_gen_len": 122.65939789882582,
4
- "eval_loss": NaN,
5
- "eval_rouge1": 5.7782,
6
- "eval_rouge2": 0.064,
7
- "eval_rougeL": 5.1661,
8
- "eval_rougeLsum": 5.1836,
9
- "eval_runtime": 1957.8518,
10
  "eval_samples": 11327,
11
- "eval_samples_per_second": 5.785,
12
- "eval_steps_per_second": 1.446,
13
- "train_loss": NaN,
14
- "train_runtime": 19100.0105,
15
  "train_samples": 204017,
16
- "train_samples_per_second": 32.045,
17
- "train_steps_per_second": 0.501
18
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_gen_len": 27.9846384744416,
4
+ "eval_loss": 2.290543556213379,
5
+ "eval_rouge1": 32.4824,
6
+ "eval_rouge2": 10.6748,
7
+ "eval_rougeL": 25.5004,
8
+ "eval_rougeLsum": 25.4918,
9
+ "eval_runtime": 600.3206,
10
  "eval_samples": 11327,
11
+ "eval_samples_per_second": 18.868,
12
+ "eval_steps_per_second": 1.179,
13
+ "train_loss": 2.8207567790121972,
14
+ "train_runtime": 17071.3808,
15
  "train_samples": 204017,
16
+ "train_samples_per_second": 35.852,
17
+ "train_steps_per_second": 0.56
18
  }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "x/base-2-2",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
 
1
  {
2
+ "_name_or_path": "x/base-2-2/",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
eval_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_gen_len": 122.65939789882582,
4
- "eval_loss": NaN,
5
- "eval_rouge1": 5.7782,
6
- "eval_rouge2": 0.064,
7
- "eval_rougeL": 5.1661,
8
- "eval_rougeLsum": 5.1836,
9
- "eval_runtime": 1957.8518,
10
  "eval_samples": 11327,
11
- "eval_samples_per_second": 5.785,
12
- "eval_steps_per_second": 1.446
13
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_gen_len": 27.9846384744416,
4
+ "eval_loss": 2.290543556213379,
5
+ "eval_rouge1": 32.4824,
6
+ "eval_rouge2": 10.6748,
7
+ "eval_rougeL": 25.5004,
8
+ "eval_rougeLsum": 25.4918,
9
+ "eval_runtime": 600.3206,
10
  "eval_samples": 11327,
11
+ "eval_samples_per_second": 18.868,
12
+ "eval_steps_per_second": 1.179
13
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:14afc6b7bbd6d6ee78962270c0c675e901abfd6a7a600c62a50044c8af40931a
3
- size 461739927
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33cf4dc6be11a9d601974d98197f4e1a50a57091cbc4489490286491604d6e2d
3
+ size 461738751
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 3.0,
3
- "train_loss": NaN,
4
- "train_runtime": 19100.0105,
5
  "train_samples": 204017,
6
- "train_samples_per_second": 32.045,
7
- "train_steps_per_second": 0.501
8
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "train_loss": 2.8207567790121972,
4
+ "train_runtime": 17071.3808,
5
  "train_samples": 204017,
6
+ "train_samples_per_second": 35.852,
7
+ "train_steps_per_second": 0.56
8
  }
trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.999754923585174,
5
- "global_step": 9561,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -10,130 +10,130 @@
10
  {
11
  "epoch": 0.16,
12
  "learning_rate": 0.0001,
13
- "loss": Infinity,
14
  "step": 500
15
  },
16
  {
17
  "epoch": 0.31,
18
  "learning_rate": 0.0001,
19
- "loss": NaN,
20
  "step": 1000
21
  },
22
  {
23
  "epoch": 0.47,
24
  "learning_rate": 0.0001,
25
- "loss": NaN,
26
  "step": 1500
27
  },
28
  {
29
  "epoch": 0.63,
30
  "learning_rate": 0.0001,
31
- "loss": NaN,
32
  "step": 2000
33
  },
34
  {
35
  "epoch": 0.78,
36
  "learning_rate": 0.0001,
37
- "loss": NaN,
38
  "step": 2500
39
  },
40
  {
41
  "epoch": 0.94,
42
  "learning_rate": 0.0001,
43
- "loss": NaN,
44
  "step": 3000
45
  },
46
  {
47
  "epoch": 1.1,
48
  "learning_rate": 0.0001,
49
- "loss": NaN,
50
  "step": 3500
51
  },
52
  {
53
- "epoch": 1.26,
54
  "learning_rate": 0.0001,
55
- "loss": NaN,
56
  "step": 4000
57
  },
58
  {
59
  "epoch": 1.41,
60
  "learning_rate": 0.0001,
61
- "loss": NaN,
62
  "step": 4500
63
  },
64
  {
65
  "epoch": 1.57,
66
  "learning_rate": 0.0001,
67
- "loss": NaN,
68
  "step": 5000
69
  },
70
  {
71
  "epoch": 1.73,
72
  "learning_rate": 0.0001,
73
- "loss": NaN,
74
  "step": 5500
75
  },
76
  {
77
  "epoch": 1.88,
78
  "learning_rate": 0.0001,
79
- "loss": NaN,
80
  "step": 6000
81
  },
82
  {
83
  "epoch": 2.04,
84
  "learning_rate": 0.0001,
85
- "loss": NaN,
86
  "step": 6500
87
  },
88
  {
89
  "epoch": 2.2,
90
  "learning_rate": 0.0001,
91
- "loss": NaN,
92
  "step": 7000
93
  },
94
  {
95
  "epoch": 2.35,
96
  "learning_rate": 0.0001,
97
- "loss": NaN,
98
  "step": 7500
99
  },
100
  {
101
  "epoch": 2.51,
102
  "learning_rate": 0.0001,
103
- "loss": NaN,
104
  "step": 8000
105
  },
106
  {
107
  "epoch": 2.67,
108
  "learning_rate": 0.0001,
109
- "loss": NaN,
110
  "step": 8500
111
  },
112
  {
113
  "epoch": 2.82,
114
  "learning_rate": 0.0001,
115
- "loss": NaN,
116
  "step": 9000
117
  },
118
  {
119
  "epoch": 2.98,
120
  "learning_rate": 0.0001,
121
- "loss": NaN,
122
  "step": 9500
123
  },
124
  {
125
  "epoch": 3.0,
126
- "step": 9561,
127
- "total_flos": 2.1720696365776896e+17,
128
- "train_loss": NaN,
129
- "train_runtime": 19100.0105,
130
- "train_samples_per_second": 32.045,
131
- "train_steps_per_second": 0.501
132
  }
133
  ],
134
- "max_steps": 9561,
135
  "num_train_epochs": 3,
136
- "total_flos": 2.1720696365776896e+17,
137
  "trial_name": null,
138
  "trial_params": null
139
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "global_step": 9564,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
10
  {
11
  "epoch": 0.16,
12
  "learning_rate": 0.0001,
13
+ "loss": 3.8829,
14
  "step": 500
15
  },
16
  {
17
  "epoch": 0.31,
18
  "learning_rate": 0.0001,
19
+ "loss": 3.2846,
20
  "step": 1000
21
  },
22
  {
23
  "epoch": 0.47,
24
  "learning_rate": 0.0001,
25
+ "loss": 3.1209,
26
  "step": 1500
27
  },
28
  {
29
  "epoch": 0.63,
30
  "learning_rate": 0.0001,
31
+ "loss": 3.0109,
32
  "step": 2000
33
  },
34
  {
35
  "epoch": 0.78,
36
  "learning_rate": 0.0001,
37
+ "loss": 2.9449,
38
  "step": 2500
39
  },
40
  {
41
  "epoch": 0.94,
42
  "learning_rate": 0.0001,
43
+ "loss": 2.8897,
44
  "step": 3000
45
  },
46
  {
47
  "epoch": 1.1,
48
  "learning_rate": 0.0001,
49
+ "loss": 2.8286,
50
  "step": 3500
51
  },
52
  {
53
+ "epoch": 1.25,
54
  "learning_rate": 0.0001,
55
+ "loss": 2.7752,
56
  "step": 4000
57
  },
58
  {
59
  "epoch": 1.41,
60
  "learning_rate": 0.0001,
61
+ "loss": 2.7459,
62
  "step": 4500
63
  },
64
  {
65
  "epoch": 1.57,
66
  "learning_rate": 0.0001,
67
+ "loss": 2.7212,
68
  "step": 5000
69
  },
70
  {
71
  "epoch": 1.73,
72
  "learning_rate": 0.0001,
73
+ "loss": 2.7011,
74
  "step": 5500
75
  },
76
  {
77
  "epoch": 1.88,
78
  "learning_rate": 0.0001,
79
+ "loss": 2.6711,
80
  "step": 6000
81
  },
82
  {
83
  "epoch": 2.04,
84
  "learning_rate": 0.0001,
85
+ "loss": 2.6403,
86
  "step": 6500
87
  },
88
  {
89
  "epoch": 2.2,
90
  "learning_rate": 0.0001,
91
+ "loss": 2.6013,
92
  "step": 7000
93
  },
94
  {
95
  "epoch": 2.35,
96
  "learning_rate": 0.0001,
97
+ "loss": 2.5927,
98
  "step": 7500
99
  },
100
  {
101
  "epoch": 2.51,
102
  "learning_rate": 0.0001,
103
+ "loss": 2.5639,
104
  "step": 8000
105
  },
106
  {
107
  "epoch": 2.67,
108
  "learning_rate": 0.0001,
109
+ "loss": 2.5589,
110
  "step": 8500
111
  },
112
  {
113
  "epoch": 2.82,
114
  "learning_rate": 0.0001,
115
+ "loss": 2.5567,
116
  "step": 9000
117
  },
118
  {
119
  "epoch": 2.98,
120
  "learning_rate": 0.0001,
121
+ "loss": 2.5386,
122
  "step": 9500
123
  },
124
  {
125
  "epoch": 3.0,
126
+ "step": 9564,
127
+ "total_flos": 3.339310425474171e+17,
128
+ "train_loss": 2.8207567790121972,
129
+ "train_runtime": 17071.3808,
130
+ "train_samples_per_second": 35.852,
131
+ "train_steps_per_second": 0.56
132
  }
133
  ],
134
+ "max_steps": 9564,
135
  "num_train_epochs": 3,
136
+ "total_flos": 3.339310425474171e+17,
137
  "trial_name": null,
138
  "trial_params": null
139
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d23a48d8b342476368a58b9a74cdd2e6c3978449f98fe1c928c28b153e4b1b66
3
- size 3643
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:328b5771a3827918d71ac94e88bd16e5563d0aa456542e0fc338d0c4ca9a3d92
3
+ size 3695