chansung commited on
Commit
69e6926
·
verified ·
1 Parent(s): 8f05ef0

Model save

Browse files
README.md CHANGED
@@ -2,13 +2,12 @@
2
  license: gemma
3
  library_name: peft
4
  tags:
5
- - alignment-handbook
6
  - trl
7
  - sft
8
  - generated_from_trainer
9
  base_model: google/gemma-7b
10
  datasets:
11
- - chansung/no_robots_only_coding
12
  model-index:
13
  - name: gemma-7b-sft-qlora-no-robots15
14
  results: []
@@ -19,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  # gemma-7b-sft-qlora-no-robots15
21
 
22
- This model is a fine-tuned version of [google/gemma-7b](https://huggingface.co/google/gemma-7b) on the chansung/no_robots_only_coding dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 1.2830
25
 
26
  ## Model description
27
 
@@ -58,20 +57,20 @@ The following hyperparameters were used during training:
58
 
59
  | Training Loss | Epoch | Step | Validation Loss |
60
  |:-------------:|:-----:|:----:|:---------------:|
61
- | 21.906 | 0.91 | 5 | 7.6533 |
62
- | 13.5603 | 2.0 | 11 | 6.6442 |
63
- | 10.2605 | 2.91 | 16 | 6.0815 |
64
- | 9.9129 | 4.0 | 22 | 3.1148 |
65
- | 4.5895 | 4.91 | 27 | 1.6583 |
66
- | 1.6316 | 6.0 | 33 | 1.4155 |
67
- | 1.4115 | 6.91 | 38 | 1.3543 |
68
- | 1.2971 | 8.0 | 44 | 1.3133 |
69
- | 1.1321 | 8.91 | 49 | 1.2903 |
70
- | 0.9739 | 10.0 | 55 | 1.2820 |
71
- | 0.917 | 10.91 | 60 | 1.2888 |
72
- | 0.8541 | 12.0 | 66 | 1.2781 |
73
- | 0.8659 | 12.91 | 71 | 1.2892 |
74
- | 0.8354 | 13.64 | 75 | 1.2830 |
75
 
76
 
77
  ### Framework versions
 
2
  license: gemma
3
  library_name: peft
4
  tags:
 
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
  base_model: google/gemma-7b
9
  datasets:
10
+ - generator
11
  model-index:
12
  - name: gemma-7b-sft-qlora-no-robots15
13
  results: []
 
18
 
19
  # gemma-7b-sft-qlora-no-robots15
20
 
21
+ This model is a fine-tuned version of [google/gemma-7b](https://huggingface.co/google/gemma-7b) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 1.2808
24
 
25
  ## Model description
26
 
 
57
 
58
  | Training Loss | Epoch | Step | Validation Loss |
59
  |:-------------:|:-----:|:----:|:---------------:|
60
+ | 21.9058 | 0.91 | 5 | 7.6562 |
61
+ | 13.5645 | 2.0 | 11 | 6.6359 |
62
+ | 10.2613 | 2.91 | 16 | 6.0754 |
63
+ | 9.903 | 4.0 | 22 | 3.1116 |
64
+ | 4.594 | 4.91 | 27 | 1.6371 |
65
+ | 1.6122 | 6.0 | 33 | 1.4160 |
66
+ | 1.3971 | 6.91 | 38 | 1.3411 |
67
+ | 1.2757 | 8.0 | 44 | 1.3074 |
68
+ | 1.1233 | 8.91 | 49 | 1.2756 |
69
+ | 0.9741 | 10.0 | 55 | 1.2736 |
70
+ | 0.9266 | 10.91 | 60 | 1.2791 |
71
+ | 0.8584 | 12.0 | 66 | 1.2753 |
72
+ | 0.8714 | 12.91 | 71 | 1.2842 |
73
+ | 0.8421 | 13.64 | 75 | 1.2808 |
74
 
75
 
76
  ### Framework versions
adapter_config.json CHANGED
@@ -20,12 +20,12 @@
20
  "revision": null,
21
  "target_modules": [
22
  "q_proj",
23
- "v_proj",
24
  "o_proj",
25
  "k_proj",
26
- "up_proj",
27
  "gate_proj",
28
- "down_proj"
 
29
  ],
30
  "task_type": "CAUSAL_LM"
31
  }
 
20
  "revision": null,
21
  "target_modules": [
22
  "q_proj",
 
23
  "o_proj",
24
  "k_proj",
25
+ "down_proj",
26
  "gate_proj",
27
+ "up_proj",
28
+ "v_proj"
29
  ],
30
  "task_type": "CAUSAL_LM"
31
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be6e49bde386e61c34c9dd54410a8a0f5f837bcefb36a8042ba70849c4ef6788
3
  size 100060536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f315f40214c028f258b809817d77512f3182e49cd92012ca7e6622d69636efef
3
  size 100060536
all_results.json CHANGED
@@ -1,13 +1,8 @@
1
  {
2
  "epoch": 13.64,
3
- "eval_loss": 1.2830009460449219,
4
- "eval_runtime": 0.6019,
5
- "eval_samples": 16,
6
- "eval_samples_per_second": 3.323,
7
- "eval_steps_per_second": 1.661,
8
- "train_loss": 4.788167775472005,
9
- "train_runtime": 283.7979,
10
  "train_samples": 926,
11
- "train_samples_per_second": 4.651,
12
  "train_steps_per_second": 0.264
13
  }
 
1
  {
2
  "epoch": 13.64,
3
+ "train_loss": 4.785474745432536,
4
+ "train_runtime": 283.6494,
 
 
 
 
 
5
  "train_samples": 926,
6
+ "train_samples_per_second": 4.654,
7
  "train_steps_per_second": 0.264
8
  }
runs/Apr11_21-36-25_deep-diver-main-swift-parrot-1-0-0/events.out.tfevents.1712885927.deep-diver-main-swift-parrot-1-0-0.521.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30297c206f33c1c3586f8b8dfd4ca982a24997cec1fa34745472cb89e5edd976
3
+ size 12773
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 13.64,
3
- "train_loss": 4.788167775472005,
4
- "train_runtime": 283.7979,
5
  "train_samples": 926,
6
- "train_samples_per_second": 4.651,
7
  "train_steps_per_second": 0.264
8
  }
 
1
  {
2
  "epoch": 13.64,
3
+ "train_loss": 4.785474745432536,
4
+ "train_runtime": 283.6494,
5
  "train_samples": 926,
6
+ "train_samples_per_second": 4.654,
7
  "train_steps_per_second": 0.264
8
  }
trainer_state.json CHANGED
@@ -19,87 +19,87 @@
19
  "epoch": 0.91,
20
  "grad_norm": 35.25,
21
  "learning_rate": 0.000125,
22
- "loss": 21.906,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.91,
27
- "eval_loss": 7.653346538543701,
28
- "eval_runtime": 0.5608,
29
- "eval_samples_per_second": 3.567,
30
- "eval_steps_per_second": 1.783,
31
  "step": 5
32
  },
33
  {
34
  "epoch": 1.82,
35
- "grad_norm": 9.3125,
36
  "learning_rate": 0.00019956059820218982,
37
- "loss": 13.5603,
38
  "step": 10
39
  },
40
  {
41
  "epoch": 2.0,
42
- "eval_loss": 6.644189357757568,
43
- "eval_runtime": 0.5604,
44
- "eval_samples_per_second": 3.569,
45
- "eval_steps_per_second": 1.784,
46
  "step": 11
47
  },
48
  {
49
  "epoch": 2.73,
50
- "grad_norm": 4.625,
51
  "learning_rate": 0.00019466156752904343,
52
- "loss": 10.2605,
53
  "step": 15
54
  },
55
  {
56
  "epoch": 2.91,
57
- "eval_loss": 6.081549644470215,
58
- "eval_runtime": 0.5806,
59
- "eval_samples_per_second": 3.445,
60
- "eval_steps_per_second": 1.722,
61
  "step": 16
62
  },
63
  {
64
  "epoch": 3.64,
65
- "grad_norm": 15.125,
66
  "learning_rate": 0.00018458320592590975,
67
- "loss": 9.9129,
68
  "step": 20
69
  },
70
  {
71
  "epoch": 4.0,
72
- "eval_loss": 3.114753484725952,
73
- "eval_runtime": 0.5656,
74
- "eval_samples_per_second": 3.536,
75
- "eval_steps_per_second": 1.768,
76
  "step": 22
77
  },
78
  {
79
  "epoch": 4.55,
80
- "grad_norm": 2.96875,
81
  "learning_rate": 0.00016987694277788417,
82
- "loss": 4.5895,
83
  "step": 25
84
  },
85
  {
86
  "epoch": 4.91,
87
- "eval_loss": 1.6582958698272705,
88
- "eval_runtime": 0.5842,
89
- "eval_samples_per_second": 3.424,
90
- "eval_steps_per_second": 1.712,
91
  "step": 27
92
  },
93
  {
94
  "epoch": 5.45,
95
- "grad_norm": 3.171875,
96
  "learning_rate": 0.0001513474193514842,
97
- "loss": 1.6316,
98
  "step": 30
99
  },
100
  {
101
  "epoch": 6.0,
102
- "eval_loss": 1.415539264678955,
103
  "eval_runtime": 0.5662,
104
  "eval_samples_per_second": 3.532,
105
  "eval_steps_per_second": 1.766,
@@ -107,138 +107,138 @@
107
  },
108
  {
109
  "epoch": 6.36,
110
- "grad_norm": 7.1875,
111
  "learning_rate": 0.0001300084635000341,
112
- "loss": 1.4115,
113
  "step": 35
114
  },
115
  {
116
  "epoch": 6.91,
117
- "eval_loss": 1.3542958498001099,
118
- "eval_runtime": 0.58,
119
- "eval_samples_per_second": 3.448,
120
- "eval_steps_per_second": 1.724,
121
  "step": 38
122
  },
123
  {
124
  "epoch": 7.27,
125
- "grad_norm": 1.296875,
126
  "learning_rate": 0.0001070276188945293,
127
- "loss": 1.2971,
128
  "step": 40
129
  },
130
  {
131
  "epoch": 8.0,
132
- "eval_loss": 1.313336968421936,
133
- "eval_runtime": 0.5751,
134
- "eval_samples_per_second": 3.478,
135
- "eval_steps_per_second": 1.739,
136
  "step": 44
137
  },
138
  {
139
  "epoch": 8.18,
140
- "grad_norm": 1.6171875,
141
  "learning_rate": 8.366226381814697e-05,
142
- "loss": 1.1321,
143
  "step": 45
144
  },
145
  {
146
  "epoch": 8.91,
147
- "eval_loss": 1.2903474569320679,
148
- "eval_runtime": 0.5848,
149
- "eval_samples_per_second": 3.42,
150
- "eval_steps_per_second": 1.71,
151
  "step": 49
152
  },
153
  {
154
  "epoch": 9.09,
155
- "grad_norm": 0.9375,
156
  "learning_rate": 6.119081473277501e-05,
157
- "loss": 1.062,
158
  "step": 50
159
  },
160
  {
161
  "epoch": 10.0,
162
  "grad_norm": 0.90234375,
163
  "learning_rate": 4.084277875864776e-05,
164
- "loss": 0.9739,
165
  "step": 55
166
  },
167
  {
168
  "epoch": 10.0,
169
- "eval_loss": 1.2820332050323486,
170
- "eval_runtime": 0.5646,
171
- "eval_samples_per_second": 3.542,
172
- "eval_steps_per_second": 1.771,
173
  "step": 55
174
  },
175
  {
176
  "epoch": 10.91,
177
- "grad_norm": 1.2109375,
178
  "learning_rate": 2.3731482188961818e-05,
179
- "loss": 0.917,
180
  "step": 60
181
  },
182
  {
183
  "epoch": 10.91,
184
- "eval_loss": 1.2888375520706177,
185
- "eval_runtime": 0.5836,
186
- "eval_samples_per_second": 3.427,
187
- "eval_steps_per_second": 1.714,
188
  "step": 60
189
  },
190
  {
191
  "epoch": 11.82,
192
- "grad_norm": 0.609375,
193
  "learning_rate": 1.0793155744261351e-05,
194
- "loss": 0.8541,
195
  "step": 65
196
  },
197
  {
198
  "epoch": 12.0,
199
- "eval_loss": 1.27806556224823,
200
- "eval_runtime": 0.5683,
201
- "eval_samples_per_second": 3.519,
202
- "eval_steps_per_second": 1.76,
203
  "step": 66
204
  },
205
  {
206
  "epoch": 12.73,
207
- "grad_norm": 0.6796875,
208
  "learning_rate": 2.735709467518699e-06,
209
- "loss": 0.8659,
210
  "step": 70
211
  },
212
  {
213
  "epoch": 12.91,
214
- "eval_loss": 1.2891546487808228,
215
- "eval_runtime": 0.7414,
216
- "eval_samples_per_second": 2.698,
217
- "eval_steps_per_second": 1.349,
218
  "step": 71
219
  },
220
  {
221
  "epoch": 13.64,
222
- "grad_norm": 0.515625,
223
  "learning_rate": 0.0,
224
- "loss": 0.8354,
225
  "step": 75
226
  },
227
  {
228
  "epoch": 13.64,
229
- "eval_loss": 1.2830009460449219,
230
- "eval_runtime": 0.5659,
231
- "eval_samples_per_second": 3.534,
232
- "eval_steps_per_second": 1.767,
233
  "step": 75
234
  },
235
  {
236
  "epoch": 13.64,
237
  "step": 75,
238
  "total_flos": 1.1503415671442637e+17,
239
- "train_loss": 4.788167775472005,
240
- "train_runtime": 283.7979,
241
- "train_samples_per_second": 4.651,
242
  "train_steps_per_second": 0.264
243
  }
244
  ],
 
19
  "epoch": 0.91,
20
  "grad_norm": 35.25,
21
  "learning_rate": 0.000125,
22
+ "loss": 21.9058,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.91,
27
+ "eval_loss": 7.6562018394470215,
28
+ "eval_runtime": 0.5686,
29
+ "eval_samples_per_second": 3.517,
30
+ "eval_steps_per_second": 1.759,
31
  "step": 5
32
  },
33
  {
34
  "epoch": 1.82,
35
+ "grad_norm": 9.0,
36
  "learning_rate": 0.00019956059820218982,
37
+ "loss": 13.5645,
38
  "step": 10
39
  },
40
  {
41
  "epoch": 2.0,
42
+ "eval_loss": 6.635939121246338,
43
+ "eval_runtime": 0.5597,
44
+ "eval_samples_per_second": 3.573,
45
+ "eval_steps_per_second": 1.787,
46
  "step": 11
47
  },
48
  {
49
  "epoch": 2.73,
50
+ "grad_norm": 4.5625,
51
  "learning_rate": 0.00019466156752904343,
52
+ "loss": 10.2613,
53
  "step": 15
54
  },
55
  {
56
  "epoch": 2.91,
57
+ "eval_loss": 6.0754241943359375,
58
+ "eval_runtime": 0.5741,
59
+ "eval_samples_per_second": 3.484,
60
+ "eval_steps_per_second": 1.742,
61
  "step": 16
62
  },
63
  {
64
  "epoch": 3.64,
65
+ "grad_norm": 14.6875,
66
  "learning_rate": 0.00018458320592590975,
67
+ "loss": 9.903,
68
  "step": 20
69
  },
70
  {
71
  "epoch": 4.0,
72
+ "eval_loss": 3.1116435527801514,
73
+ "eval_runtime": 0.5647,
74
+ "eval_samples_per_second": 3.542,
75
+ "eval_steps_per_second": 1.771,
76
  "step": 22
77
  },
78
  {
79
  "epoch": 4.55,
80
+ "grad_norm": 32.75,
81
  "learning_rate": 0.00016987694277788417,
82
+ "loss": 4.594,
83
  "step": 25
84
  },
85
  {
86
  "epoch": 4.91,
87
+ "eval_loss": 1.6371122598648071,
88
+ "eval_runtime": 0.5752,
89
+ "eval_samples_per_second": 3.477,
90
+ "eval_steps_per_second": 1.739,
91
  "step": 27
92
  },
93
  {
94
  "epoch": 5.45,
95
+ "grad_norm": 3.125,
96
  "learning_rate": 0.0001513474193514842,
97
+ "loss": 1.6122,
98
  "step": 30
99
  },
100
  {
101
  "epoch": 6.0,
102
+ "eval_loss": 1.4159561395645142,
103
  "eval_runtime": 0.5662,
104
  "eval_samples_per_second": 3.532,
105
  "eval_steps_per_second": 1.766,
 
107
  },
108
  {
109
  "epoch": 6.36,
110
+ "grad_norm": 1.03125,
111
  "learning_rate": 0.0001300084635000341,
112
+ "loss": 1.3971,
113
  "step": 35
114
  },
115
  {
116
  "epoch": 6.91,
117
+ "eval_loss": 1.3411411046981812,
118
+ "eval_runtime": 0.5782,
119
+ "eval_samples_per_second": 3.459,
120
+ "eval_steps_per_second": 1.73,
121
  "step": 38
122
  },
123
  {
124
  "epoch": 7.27,
125
+ "grad_norm": 1.65625,
126
  "learning_rate": 0.0001070276188945293,
127
+ "loss": 1.2757,
128
  "step": 40
129
  },
130
  {
131
  "epoch": 8.0,
132
+ "eval_loss": 1.307417631149292,
133
+ "eval_runtime": 0.5668,
134
+ "eval_samples_per_second": 3.529,
135
+ "eval_steps_per_second": 1.764,
136
  "step": 44
137
  },
138
  {
139
  "epoch": 8.18,
140
+ "grad_norm": 1.796875,
141
  "learning_rate": 8.366226381814697e-05,
142
+ "loss": 1.1233,
143
  "step": 45
144
  },
145
  {
146
  "epoch": 8.91,
147
+ "eval_loss": 1.275590419769287,
148
+ "eval_runtime": 0.5782,
149
+ "eval_samples_per_second": 3.459,
150
+ "eval_steps_per_second": 1.729,
151
  "step": 49
152
  },
153
  {
154
  "epoch": 9.09,
155
+ "grad_norm": 0.99609375,
156
  "learning_rate": 6.119081473277501e-05,
157
+ "loss": 1.0599,
158
  "step": 50
159
  },
160
  {
161
  "epoch": 10.0,
162
  "grad_norm": 0.90234375,
163
  "learning_rate": 4.084277875864776e-05,
164
+ "loss": 0.9741,
165
  "step": 55
166
  },
167
  {
168
  "epoch": 10.0,
169
+ "eval_loss": 1.273629903793335,
170
+ "eval_runtime": 0.5636,
171
+ "eval_samples_per_second": 3.549,
172
+ "eval_steps_per_second": 1.774,
173
  "step": 55
174
  },
175
  {
176
  "epoch": 10.91,
177
+ "grad_norm": 0.62890625,
178
  "learning_rate": 2.3731482188961818e-05,
179
+ "loss": 0.9266,
180
  "step": 60
181
  },
182
  {
183
  "epoch": 10.91,
184
+ "eval_loss": 1.279077410697937,
185
+ "eval_runtime": 0.5768,
186
+ "eval_samples_per_second": 3.467,
187
+ "eval_steps_per_second": 1.734,
188
  "step": 60
189
  },
190
  {
191
  "epoch": 11.82,
192
+ "grad_norm": 0.5234375,
193
  "learning_rate": 1.0793155744261351e-05,
194
+ "loss": 0.8584,
195
  "step": 65
196
  },
197
  {
198
  "epoch": 12.0,
199
+ "eval_loss": 1.2753326892852783,
200
+ "eval_runtime": 0.5671,
201
+ "eval_samples_per_second": 3.527,
202
+ "eval_steps_per_second": 1.763,
203
  "step": 66
204
  },
205
  {
206
  "epoch": 12.73,
207
+ "grad_norm": 0.515625,
208
  "learning_rate": 2.735709467518699e-06,
209
+ "loss": 0.8714,
210
  "step": 70
211
  },
212
  {
213
  "epoch": 12.91,
214
+ "eval_loss": 1.2842026948928833,
215
+ "eval_runtime": 0.7036,
216
+ "eval_samples_per_second": 2.842,
217
+ "eval_steps_per_second": 1.421,
218
  "step": 71
219
  },
220
  {
221
  "epoch": 13.64,
222
+ "grad_norm": 0.56640625,
223
  "learning_rate": 0.0,
224
+ "loss": 0.8421,
225
  "step": 75
226
  },
227
  {
228
  "epoch": 13.64,
229
+ "eval_loss": 1.2808171510696411,
230
+ "eval_runtime": 0.5646,
231
+ "eval_samples_per_second": 3.543,
232
+ "eval_steps_per_second": 1.771,
233
  "step": 75
234
  },
235
  {
236
  "epoch": 13.64,
237
  "step": 75,
238
  "total_flos": 1.1503415671442637e+17,
239
+ "train_loss": 4.785474745432536,
240
+ "train_runtime": 283.6494,
241
+ "train_samples_per_second": 4.654,
242
  "train_steps_per_second": 0.264
243
  }
244
  ],
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ffbfc1f7707f8c234a55a9fc0b314774d4ec0938f941afe549a7d87e7a4373e1
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8767d3875a56a1eab9d37ca54199f7864f60adc0471cee54828002448777d1fe
3
  size 5112