spacemanidol commited on
Commit
fad998a
1 Parent(s): f26eda9

Upload 13 files

Browse files
README.md CHANGED
@@ -6,7 +6,7 @@ datasets:
6
  metrics:
7
  - rouge
8
  model-index:
9
- - name: small-3-6-t
10
  results:
11
  - task:
12
  name: Summarization
@@ -20,22 +20,22 @@ model-index:
20
  metrics:
21
  - name: Rouge1
22
  type: rouge
23
- value: 31.7432
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
27
  should probably proofread and complete it, then remove this comment. -->
28
 
29
- # small-3-6-t
30
 
31
- This model is a fine-tuned version of [models/small-3-6](https://huggingface.co/models/small-3-6) on the xsum dataset.
32
  It achieves the following results on the evaluation set:
33
- - Loss: 2.1322
34
- - Rouge1: 31.7432
35
- - Rouge2: 10.077
36
- - Rougel: 24.8565
37
- - Rougelsum: 24.8444
38
- - Gen Len: 28.8587
39
 
40
  ## Model description
41
 
@@ -55,10 +55,10 @@ More information needed
55
 
56
  The following hyperparameters were used during training:
57
  - learning_rate: 0.0001
58
- - train_batch_size: 32
59
- - eval_batch_size: 16
60
  - seed: 42
61
- - gradient_accumulation_steps: 2
62
  - total_train_batch_size: 64
63
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
64
  - lr_scheduler_type: constant
@@ -71,6 +71,6 @@ The following hyperparameters were used during training:
71
  ### Framework versions
72
 
73
  - Transformers 4.27.0.dev0
74
- - Pytorch 1.12.1+cu113
75
- - Datasets 2.10.0
76
- - Tokenizers 0.13.2
 
6
  metrics:
7
  - rouge
8
  model-index:
9
+ - name: small-3-6
10
  results:
11
  - task:
12
  name: Summarization
 
20
  metrics:
21
  - name: Rouge1
22
  type: rouge
23
+ value: 31.8175
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
27
  should probably proofread and complete it, then remove this comment. -->
28
 
29
+ # small-3-6
30
 
31
+ This model is a fine-tuned version of [x/small-3-6/](https://huggingface.co/x/small-3-6/) on the xsum dataset.
32
  It achieves the following results on the evaluation set:
33
+ - Loss: 2.1335
34
+ - Rouge1: 31.8175
35
+ - Rouge2: 10.1063
36
+ - Rougel: 24.9237
37
+ - Rougelsum: 24.9165
38
+ - Gen Len: 28.4315
39
 
40
  ## Model description
41
 
 
55
 
56
  The following hyperparameters were used during training:
57
  - learning_rate: 0.0001
58
+ - train_batch_size: 4
59
+ - eval_batch_size: 4
60
  - seed: 42
61
+ - gradient_accumulation_steps: 16
62
  - total_train_batch_size: 64
63
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
64
  - lr_scheduler_type: constant
 
71
  ### Framework versions
72
 
73
  - Transformers 4.27.0.dev0
74
+ - Pytorch 1.13.0+cu117
75
+ - Datasets 2.7.1
76
+ - Tokenizers 0.12.1
all_results.json CHANGED
@@ -1,18 +1,18 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_gen_len": 28.858656307936787,
4
- "eval_loss": 2.132248878479004,
5
- "eval_rouge1": 31.7432,
6
- "eval_rouge2": 10.077,
7
- "eval_rougeL": 24.8565,
8
- "eval_rougeLsum": 24.8444,
9
- "eval_runtime": 553.66,
10
  "eval_samples": 11327,
11
- "eval_samples_per_second": 20.458,
12
- "eval_steps_per_second": 1.279,
13
- "train_loss": 2.37733912099011,
14
- "train_runtime": 8397.5704,
15
  "train_samples": 204017,
16
- "train_samples_per_second": 72.884,
17
- "train_steps_per_second": 1.139
18
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_gen_len": 28.431535269709542,
4
+ "eval_loss": 2.1334991455078125,
5
+ "eval_rouge1": 31.8175,
6
+ "eval_rouge2": 10.1063,
7
+ "eval_rougeL": 24.9237,
8
+ "eval_rougeLsum": 24.9165,
9
+ "eval_runtime": 1009.3627,
10
  "eval_samples": 11327,
11
+ "eval_samples_per_second": 11.222,
12
+ "eval_steps_per_second": 2.806,
13
+ "train_loss": 2.374575984090233,
14
+ "train_runtime": 14184.3644,
15
  "train_samples": 204017,
16
+ "train_samples_per_second": 43.15,
17
+ "train_steps_per_second": 0.674
18
  }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "models/small-3-6",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
 
1
  {
2
+ "_name_or_path": "x/small-3-6/",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
eval_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_gen_len": 28.858656307936787,
4
- "eval_loss": 2.132248878479004,
5
- "eval_rouge1": 31.7432,
6
- "eval_rouge2": 10.077,
7
- "eval_rougeL": 24.8565,
8
- "eval_rougeLsum": 24.8444,
9
- "eval_runtime": 553.66,
10
  "eval_samples": 11327,
11
- "eval_samples_per_second": 20.458,
12
- "eval_steps_per_second": 1.279
13
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_gen_len": 28.431535269709542,
4
+ "eval_loss": 2.1334991455078125,
5
+ "eval_rouge1": 31.8175,
6
+ "eval_rouge2": 10.1063,
7
+ "eval_rougeL": 24.9237,
8
+ "eval_rougeLsum": 24.9165,
9
+ "eval_runtime": 1009.3627,
10
  "eval_samples": 11327,
11
+ "eval_samples_per_second": 11.222,
12
+ "eval_steps_per_second": 2.806
13
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:682d6e315377557db86cd39f41c08b2f2a3761cecb5e3b5441b3f7afa4bae9ea
3
- size 270131021
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca5d0d9b6f61d273138c4dee5e168e949a587e4d43422dd8f236e7d4a3bd181b
3
+ size 270132533
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86
3
+ size 791656
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 3.0,
3
- "train_loss": 2.37733912099011,
4
- "train_runtime": 8397.5704,
5
  "train_samples": 204017,
6
- "train_samples_per_second": 72.884,
7
- "train_steps_per_second": 1.139
8
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "train_loss": 2.374575984090233,
4
+ "train_runtime": 14184.3644,
5
  "train_samples": 204017,
6
+ "train_samples_per_second": 43.15,
7
+ "train_steps_per_second": 0.674
8
  }
trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.0,
5
- "global_step": 9564,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -10,130 +10,130 @@
10
  {
11
  "epoch": 0.16,
12
  "learning_rate": 0.0001,
13
- "loss": 2.5455,
14
  "step": 500
15
  },
16
  {
17
  "epoch": 0.31,
18
  "learning_rate": 0.0001,
19
- "loss": 2.4773,
20
  "step": 1000
21
  },
22
  {
23
  "epoch": 0.47,
24
  "learning_rate": 0.0001,
25
- "loss": 2.4581,
26
  "step": 1500
27
  },
28
  {
29
  "epoch": 0.63,
30
  "learning_rate": 0.0001,
31
- "loss": 2.4289,
32
  "step": 2000
33
  },
34
  {
35
  "epoch": 0.78,
36
  "learning_rate": 0.0001,
37
- "loss": 2.4264,
38
  "step": 2500
39
  },
40
  {
41
  "epoch": 0.94,
42
  "learning_rate": 0.0001,
43
- "loss": 2.4111,
44
  "step": 3000
45
  },
46
  {
47
  "epoch": 1.1,
48
  "learning_rate": 0.0001,
49
- "loss": 2.3918,
50
  "step": 3500
51
  },
52
  {
53
- "epoch": 1.25,
54
  "learning_rate": 0.0001,
55
- "loss": 2.373,
56
  "step": 4000
57
  },
58
  {
59
  "epoch": 1.41,
60
  "learning_rate": 0.0001,
61
- "loss": 2.3684,
62
  "step": 4500
63
  },
64
  {
65
  "epoch": 1.57,
66
  "learning_rate": 0.0001,
67
- "loss": 2.364,
68
  "step": 5000
69
  },
70
  {
71
  "epoch": 1.73,
72
  "learning_rate": 0.0001,
73
- "loss": 2.3606,
74
  "step": 5500
75
  },
76
  {
77
  "epoch": 1.88,
78
  "learning_rate": 0.0001,
79
- "loss": 2.3512,
80
  "step": 6000
81
  },
82
  {
83
  "epoch": 2.04,
84
  "learning_rate": 0.0001,
85
- "loss": 2.3383,
86
  "step": 6500
87
  },
88
  {
89
  "epoch": 2.2,
90
  "learning_rate": 0.0001,
91
- "loss": 2.3184,
92
  "step": 7000
93
  },
94
  {
95
  "epoch": 2.35,
96
  "learning_rate": 0.0001,
97
- "loss": 2.3238,
98
  "step": 7500
99
  },
100
  {
101
  "epoch": 2.51,
102
  "learning_rate": 0.0001,
103
- "loss": 2.3079,
104
  "step": 8000
105
  },
106
  {
107
  "epoch": 2.67,
108
  "learning_rate": 0.0001,
109
- "loss": 2.3105,
110
  "step": 8500
111
  },
112
  {
113
  "epoch": 2.82,
114
  "learning_rate": 0.0001,
115
- "loss": 2.3148,
116
  "step": 9000
117
  },
118
  {
119
  "epoch": 2.98,
120
  "learning_rate": 0.0001,
121
- "loss": 2.3072,
122
  "step": 9500
123
  },
124
  {
125
  "epoch": 3.0,
126
- "step": 9564,
127
- "total_flos": 1.916909651434537e+17,
128
- "train_loss": 2.37733912099011,
129
- "train_runtime": 8397.5704,
130
- "train_samples_per_second": 72.884,
131
- "train_steps_per_second": 1.139
132
  }
133
  ],
134
- "max_steps": 9564,
135
  "num_train_epochs": 3,
136
- "total_flos": 1.916909651434537e+17,
137
  "trial_name": null,
138
  "trial_params": null
139
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.999745123027154,
5
+ "global_step": 9561,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
10
  {
11
  "epoch": 0.16,
12
  "learning_rate": 0.0001,
13
+ "loss": 2.5416,
14
  "step": 500
15
  },
16
  {
17
  "epoch": 0.31,
18
  "learning_rate": 0.0001,
19
+ "loss": 2.4758,
20
  "step": 1000
21
  },
22
  {
23
  "epoch": 0.47,
24
  "learning_rate": 0.0001,
25
+ "loss": 2.4544,
26
  "step": 1500
27
  },
28
  {
29
  "epoch": 0.63,
30
  "learning_rate": 0.0001,
31
+ "loss": 2.4235,
32
  "step": 2000
33
  },
34
  {
35
  "epoch": 0.78,
36
  "learning_rate": 0.0001,
37
+ "loss": 2.4206,
38
  "step": 2500
39
  },
40
  {
41
  "epoch": 0.94,
42
  "learning_rate": 0.0001,
43
+ "loss": 2.4084,
44
  "step": 3000
45
  },
46
  {
47
  "epoch": 1.1,
48
  "learning_rate": 0.0001,
49
+ "loss": 2.3928,
50
  "step": 3500
51
  },
52
  {
53
+ "epoch": 1.26,
54
  "learning_rate": 0.0001,
55
+ "loss": 2.3703,
56
  "step": 4000
57
  },
58
  {
59
  "epoch": 1.41,
60
  "learning_rate": 0.0001,
61
+ "loss": 2.3678,
62
  "step": 4500
63
  },
64
  {
65
  "epoch": 1.57,
66
  "learning_rate": 0.0001,
67
+ "loss": 2.3605,
68
  "step": 5000
69
  },
70
  {
71
  "epoch": 1.73,
72
  "learning_rate": 0.0001,
73
+ "loss": 2.3572,
74
  "step": 5500
75
  },
76
  {
77
  "epoch": 1.88,
78
  "learning_rate": 0.0001,
79
+ "loss": 2.3473,
80
  "step": 6000
81
  },
82
  {
83
  "epoch": 2.04,
84
  "learning_rate": 0.0001,
85
+ "loss": 2.3415,
86
  "step": 6500
87
  },
88
  {
89
  "epoch": 2.2,
90
  "learning_rate": 0.0001,
91
+ "loss": 2.3158,
92
  "step": 7000
93
  },
94
  {
95
  "epoch": 2.35,
96
  "learning_rate": 0.0001,
97
+ "loss": 2.3191,
98
  "step": 7500
99
  },
100
  {
101
  "epoch": 2.51,
102
  "learning_rate": 0.0001,
103
+ "loss": 2.3028,
104
  "step": 8000
105
  },
106
  {
107
  "epoch": 2.67,
108
  "learning_rate": 0.0001,
109
+ "loss": 2.3086,
110
  "step": 8500
111
  },
112
  {
113
  "epoch": 2.82,
114
  "learning_rate": 0.0001,
115
+ "loss": 2.3124,
116
  "step": 9000
117
  },
118
  {
119
  "epoch": 2.98,
120
  "learning_rate": 0.0001,
121
+ "loss": 2.3034,
122
  "step": 9500
123
  },
124
  {
125
  "epoch": 3.0,
126
+ "step": 9561,
127
+ "total_flos": 1.5154927442907955e+17,
128
+ "train_loss": 2.374575984090233,
129
+ "train_runtime": 14184.3644,
130
+ "train_samples_per_second": 43.15,
131
+ "train_steps_per_second": 0.674
132
  }
133
  ],
134
+ "max_steps": 9561,
135
  "num_train_epochs": 3,
136
+ "total_flos": 1.5154927442907955e+17,
137
  "trial_name": null,
138
  "trial_params": null
139
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b06acb0c42c995c39614ca9c1505acc13befd60ef91232c34d33675a0e2843b
3
- size 3631
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8115bce279b297a884a2fc911eea452fbe3fd1f655c5e468581ac9fe6fa7120
3
+ size 3643