spacemanidol commited on
Commit
efa074c
1 Parent(s): f7373a6

Upload 13 files

Browse files
README.md CHANGED
@@ -6,7 +6,7 @@ datasets:
6
  metrics:
7
  - rouge
8
  model-index:
9
- - name: small-2-2-t
10
  results:
11
  - task:
12
  name: Summarization
@@ -20,22 +20,22 @@ model-index:
20
  metrics:
21
  - name: Rouge1
22
  type: rouge
23
- value: 26.601
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
27
  should probably proofread and complete it, then remove this comment. -->
28
 
29
- # small-2-2-t
30
 
31
- This model is a fine-tuned version of [models/small-2-2](https://huggingface.co/models/small-2-2) on the xsum dataset.
32
  It achieves the following results on the evaluation set:
33
- - Loss: 2.9093
34
- - Rouge1: 26.601
35
- - Rouge2: 7.1924
36
- - Rougel: 21.2586
37
- - Rougelsum: 21.2573
38
- - Gen Len: 26.3125
39
 
40
  ## Model description
41
 
@@ -55,10 +55,10 @@ More information needed
55
 
56
  The following hyperparameters were used during training:
57
  - learning_rate: 0.0001
58
- - train_batch_size: 32
59
- - eval_batch_size: 16
60
  - seed: 42
61
- - gradient_accumulation_steps: 2
62
  - total_train_batch_size: 64
63
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
64
  - lr_scheduler_type: constant
@@ -71,6 +71,6 @@ The following hyperparameters were used during training:
71
  ### Framework versions
72
 
73
  - Transformers 4.27.0.dev0
74
- - Pytorch 1.12.1+cu113
75
- - Datasets 2.10.0
76
- - Tokenizers 0.13.2
 
6
  metrics:
7
  - rouge
8
  model-index:
9
+ - name: small-2-2
10
  results:
11
  - task:
12
  name: Summarization
 
20
  metrics:
21
  - name: Rouge1
22
  type: rouge
23
+ value: 26.7124
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
27
  should probably proofread and complete it, then remove this comment. -->
28
 
29
+ # small-2-2
30
 
31
+ This model is a fine-tuned version of [x/small-2-2/](https://huggingface.co/x/small-2-2/) on the xsum dataset.
32
  It achieves the following results on the evaluation set:
33
+ - Loss: 2.9112
34
+ - Rouge1: 26.7124
35
+ - Rouge2: 7.3066
36
+ - Rougel: 21.3783
37
+ - Rougelsum: 21.3771
38
+ - Gen Len: 26.3514
39
 
40
  ## Model description
41
 
 
55
 
56
  The following hyperparameters were used during training:
57
  - learning_rate: 0.0001
58
+ - train_batch_size: 4
59
+ - eval_batch_size: 4
60
  - seed: 42
61
+ - gradient_accumulation_steps: 16
62
  - total_train_batch_size: 64
63
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
64
  - lr_scheduler_type: constant
 
71
  ### Framework versions
72
 
73
  - Transformers 4.27.0.dev0
74
+ - Pytorch 1.13.0+cu117
75
+ - Datasets 2.7.1
76
+ - Tokenizers 0.12.1
all_results.json CHANGED
@@ -1,18 +1,18 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_gen_len": 26.312527588946764,
4
- "eval_loss": 2.9093124866485596,
5
- "eval_rouge1": 26.601,
6
- "eval_rouge2": 7.1924,
7
- "eval_rougeL": 21.2586,
8
- "eval_rougeLsum": 21.2573,
9
- "eval_runtime": 219.9179,
10
  "eval_samples": 11327,
11
- "eval_samples_per_second": 51.506,
12
- "eval_steps_per_second": 3.219,
13
- "train_loss": 3.5344422363128087,
14
- "train_runtime": 4207.4456,
15
  "train_samples": 204017,
16
- "train_samples_per_second": 145.469,
17
- "train_steps_per_second": 2.273
18
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_gen_len": 26.351372825990996,
4
+ "eval_loss": 2.9112093448638916,
5
+ "eval_rouge1": 26.7124,
6
+ "eval_rouge2": 7.3066,
7
+ "eval_rougeL": 21.3783,
8
+ "eval_rougeLsum": 21.3771,
9
+ "eval_runtime": 383.7315,
10
  "eval_samples": 11327,
11
+ "eval_samples_per_second": 29.518,
12
+ "eval_steps_per_second": 7.38,
13
+ "train_loss": 3.5328583012638624,
14
+ "train_runtime": 7182.8846,
15
  "train_samples": 204017,
16
+ "train_samples_per_second": 85.21,
17
+ "train_steps_per_second": 1.331
18
  }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "models/small-2-2",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
 
1
  {
2
+ "_name_or_path": "x/small-2-2/",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
eval_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_gen_len": 26.312527588946764,
4
- "eval_loss": 2.9093124866485596,
5
- "eval_rouge1": 26.601,
6
- "eval_rouge2": 7.1924,
7
- "eval_rougeL": 21.2586,
8
- "eval_rougeLsum": 21.2573,
9
- "eval_runtime": 219.9179,
10
  "eval_samples": 11327,
11
- "eval_samples_per_second": 51.506,
12
- "eval_steps_per_second": 3.219
13
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_gen_len": 26.351372825990996,
4
+ "eval_loss": 2.9112093448638916,
5
+ "eval_rouge1": 26.7124,
6
+ "eval_rouge2": 7.3066,
7
+ "eval_rougeL": 21.3783,
8
+ "eval_rougeLsum": 21.3771,
9
+ "eval_runtime": 383.7315,
10
  "eval_samples": 11327,
11
+ "eval_samples_per_second": 29.518,
12
+ "eval_steps_per_second": 7.38
13
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38bf9caea956ad1781e4c9e2ad1af5ba1f8c629bc1388f9c6c4c36361fd50829
3
- size 175679497
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b144b5388dedfbfa6856b537a14d03e4a277eb349bc3107dc28b545282cb8a99
3
+ size 175680397
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86
3
+ size 791656
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 3.0,
3
- "train_loss": 3.5344422363128087,
4
- "train_runtime": 4207.4456,
5
  "train_samples": 204017,
6
- "train_samples_per_second": 145.469,
7
- "train_steps_per_second": 2.273
8
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "train_loss": 3.5328583012638624,
4
+ "train_runtime": 7182.8846,
5
  "train_samples": 204017,
6
+ "train_samples_per_second": 85.21,
7
+ "train_steps_per_second": 1.331
8
  }
trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.0,
5
- "global_step": 9564,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -10,19 +10,19 @@
10
  {
11
  "epoch": 0.16,
12
  "learning_rate": 0.0001,
13
- "loss": 4.7835,
14
  "step": 500
15
  },
16
  {
17
  "epoch": 0.31,
18
  "learning_rate": 0.0001,
19
- "loss": 4.0029,
20
  "step": 1000
21
  },
22
  {
23
  "epoch": 0.47,
24
  "learning_rate": 0.0001,
25
- "loss": 3.8273,
26
  "step": 1500
27
  },
28
  {
@@ -34,106 +34,106 @@
34
  {
35
  "epoch": 0.78,
36
  "learning_rate": 0.0001,
37
- "loss": 3.6487,
38
  "step": 2500
39
  },
40
  {
41
  "epoch": 0.94,
42
  "learning_rate": 0.0001,
43
- "loss": 3.5852,
44
  "step": 3000
45
  },
46
  {
47
  "epoch": 1.1,
48
  "learning_rate": 0.0001,
49
- "loss": 3.5361,
50
  "step": 3500
51
  },
52
  {
53
- "epoch": 1.25,
54
  "learning_rate": 0.0001,
55
- "loss": 3.4853,
56
  "step": 4000
57
  },
58
  {
59
  "epoch": 1.41,
60
  "learning_rate": 0.0001,
61
- "loss": 3.4501,
62
  "step": 4500
63
  },
64
  {
65
  "epoch": 1.57,
66
  "learning_rate": 0.0001,
67
- "loss": 3.4219,
68
  "step": 5000
69
  },
70
  {
71
  "epoch": 1.73,
72
  "learning_rate": 0.0001,
73
- "loss": 3.3996,
74
  "step": 5500
75
  },
76
  {
77
  "epoch": 1.88,
78
  "learning_rate": 0.0001,
79
- "loss": 3.3681,
80
  "step": 6000
81
  },
82
  {
83
  "epoch": 2.04,
84
  "learning_rate": 0.0001,
85
- "loss": 3.3408,
86
  "step": 6500
87
  },
88
  {
89
  "epoch": 2.2,
90
  "learning_rate": 0.0001,
91
- "loss": 3.3105,
92
  "step": 7000
93
  },
94
  {
95
  "epoch": 2.35,
96
  "learning_rate": 0.0001,
97
- "loss": 3.3033,
98
  "step": 7500
99
  },
100
  {
101
  "epoch": 2.51,
102
  "learning_rate": 0.0001,
103
- "loss": 3.2689,
104
  "step": 8000
105
  },
106
  {
107
  "epoch": 2.67,
108
  "learning_rate": 0.0001,
109
- "loss": 3.2592,
110
  "step": 8500
111
  },
112
  {
113
  "epoch": 2.82,
114
  "learning_rate": 0.0001,
115
- "loss": 3.2535,
116
  "step": 9000
117
  },
118
  {
119
  "epoch": 2.98,
120
  "learning_rate": 0.0001,
121
- "loss": 3.2336,
122
  "step": 9500
123
  },
124
  {
125
  "epoch": 3.0,
126
- "step": 9564,
127
- "total_flos": 1.0309255643489894e+17,
128
- "train_loss": 3.5344422363128087,
129
- "train_runtime": 4207.4456,
130
- "train_samples_per_second": 145.469,
131
- "train_steps_per_second": 2.273
132
  }
133
  ],
134
- "max_steps": 9564,
135
  "num_train_epochs": 3,
136
- "total_flos": 1.0309255643489894e+17,
137
  "trial_name": null,
138
  "trial_params": null
139
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.999745123027154,
5
+ "global_step": 9561,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
10
  {
11
  "epoch": 0.16,
12
  "learning_rate": 0.0001,
13
+ "loss": 4.7889,
14
  "step": 500
15
  },
16
  {
17
  "epoch": 0.31,
18
  "learning_rate": 0.0001,
19
+ "loss": 4.007,
20
  "step": 1000
21
  },
22
  {
23
  "epoch": 0.47,
24
  "learning_rate": 0.0001,
25
+ "loss": 3.8226,
26
  "step": 1500
27
  },
28
  {
 
34
  {
35
  "epoch": 0.78,
36
  "learning_rate": 0.0001,
37
+ "loss": 3.643,
38
  "step": 2500
39
  },
40
  {
41
  "epoch": 0.94,
42
  "learning_rate": 0.0001,
43
+ "loss": 3.5814,
44
  "step": 3000
45
  },
46
  {
47
  "epoch": 1.1,
48
  "learning_rate": 0.0001,
49
+ "loss": 3.5365,
50
  "step": 3500
51
  },
52
  {
53
+ "epoch": 1.26,
54
  "learning_rate": 0.0001,
55
+ "loss": 3.4833,
56
  "step": 4000
57
  },
58
  {
59
  "epoch": 1.41,
60
  "learning_rate": 0.0001,
61
+ "loss": 3.4472,
62
  "step": 4500
63
  },
64
  {
65
  "epoch": 1.57,
66
  "learning_rate": 0.0001,
67
+ "loss": 3.4199,
68
  "step": 5000
69
  },
70
  {
71
  "epoch": 1.73,
72
  "learning_rate": 0.0001,
73
+ "loss": 3.398,
74
  "step": 5500
75
  },
76
  {
77
  "epoch": 1.88,
78
  "learning_rate": 0.0001,
79
+ "loss": 3.3639,
80
  "step": 6000
81
  },
82
  {
83
  "epoch": 2.04,
84
  "learning_rate": 0.0001,
85
+ "loss": 3.3425,
86
  "step": 6500
87
  },
88
  {
89
  "epoch": 2.2,
90
  "learning_rate": 0.0001,
91
+ "loss": 3.3083,
92
  "step": 7000
93
  },
94
  {
95
  "epoch": 2.35,
96
  "learning_rate": 0.0001,
97
+ "loss": 3.2997,
98
  "step": 7500
99
  },
100
  {
101
  "epoch": 2.51,
102
  "learning_rate": 0.0001,
103
+ "loss": 3.2649,
104
  "step": 8000
105
  },
106
  {
107
  "epoch": 2.67,
108
  "learning_rate": 0.0001,
109
+ "loss": 3.258,
110
  "step": 8500
111
  },
112
  {
113
  "epoch": 2.82,
114
  "learning_rate": 0.0001,
115
+ "loss": 3.2514,
116
  "step": 9000
117
  },
118
  {
119
  "epoch": 2.98,
120
  "learning_rate": 0.0001,
121
+ "loss": 3.2301,
122
  "step": 9500
123
  },
124
  {
125
  "epoch": 3.0,
126
+ "step": 9561,
127
+ "total_flos": 8.150411322232013e+16,
128
+ "train_loss": 3.5328583012638624,
129
+ "train_runtime": 7182.8846,
130
+ "train_samples_per_second": 85.21,
131
+ "train_steps_per_second": 1.331
132
  }
133
  ],
134
+ "max_steps": 9561,
135
  "num_train_epochs": 3,
136
+ "total_flos": 8.150411322232013e+16,
137
  "trial_name": null,
138
  "trial_params": null
139
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f4452815b843d7fbd4d6ccc6494c80d29fe0d2e902d2471d312ebbebceb16fb
3
- size 3631
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c4c81cb7753ab52e2ab7a0063a8c2ebcdffab508d0c985f8c7eb5a96bdafd2b
3
+ size 3643