jan-hq commited on
Commit
8f66fca
1 Parent(s): c1ec575

Model save

Browse files
README.md ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama2
3
+ library_name: peft
4
+ tags:
5
+ - trl
6
+ - sft
7
+ - generated_from_trainer
8
+ datasets:
9
+ - generator
10
+ base_model: Yellow-AI-NLP/komodo-7b-base
11
+ model-index:
12
+ - name: komodo-7b-chat
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # komodo-7b-chat
20
+
21
+ This model is a fine-tuned version of [Yellow-AI-NLP/komodo-7b-base](https://huggingface.co/Yellow-AI-NLP/komodo-7b-base) on the generator dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 1.2932
24
+
25
+ ## Model description
26
+
27
+ More information needed
28
+
29
+ ## Intended uses & limitations
30
+
31
+ More information needed
32
+
33
+ ## Training and evaluation data
34
+
35
+ More information needed
36
+
37
+ ## Training procedure
38
+
39
+ ### Training hyperparameters
40
+
41
+ The following hyperparameters were used during training:
42
+ - learning_rate: 2e-05
43
+ - train_batch_size: 2
44
+ - eval_batch_size: 2
45
+ - seed: 42
46
+ - distributed_type: multi-GPU
47
+ - num_devices: 2
48
+ - gradient_accumulation_steps: 16
49
+ - total_train_batch_size: 64
50
+ - total_eval_batch_size: 4
51
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
+ - lr_scheduler_type: cosine
53
+ - lr_scheduler_warmup_ratio: 0.1
54
+ - num_epochs: 3
55
+
56
+ ### Training results
57
+
58
+ | Training Loss | Epoch | Step | Validation Loss |
59
+ |:-------------:|:-----:|:----:|:---------------:|
60
+ | 1.3204 | 0.99 | 67 | 1.3371 |
61
+ | 1.3083 | 1.99 | 134 | 1.2950 |
62
+ | 1.2921 | 2.98 | 201 | 1.2932 |
63
+
64
+
65
+ ### Framework versions
66
+
67
+ - PEFT 0.8.2
68
+ - Transformers 4.37.2
69
+ - Pytorch 2.1.2+cu121
70
+ - Datasets 2.14.6
71
+ - Tokenizers 0.15.0
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:781f96eb8287538ab87fce3d85f56e873ed68fda57f262a5836535529a3bbeb6
3
  size 80014016
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c5c6093e962de9f9a72f05ea64bb4c56cbd7ee06cb069ffd7c2ff6f03e27c9b
3
  size 80014016
all_results.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.98,
3
+ "eval_loss": 1.2931873798370361,
4
+ "eval_runtime": 60.4437,
5
+ "eval_samples": 1384,
6
+ "eval_samples_per_second": 8.223,
7
+ "eval_steps_per_second": 2.068,
8
+ "train_loss": 1.3530203068434303,
9
+ "train_runtime": 5442.6995,
10
+ "train_samples": 12450,
11
+ "train_samples_per_second": 2.376,
12
+ "train_steps_per_second": 0.037
13
+ }
eval_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.98,
3
+ "eval_loss": 1.2931873798370361,
4
+ "eval_runtime": 60.4437,
5
+ "eval_samples": 1384,
6
+ "eval_samples_per_second": 8.223,
7
+ "eval_steps_per_second": 2.068
8
+ }
runs/Mar18_05-31-15_333df911e7ea/events.out.tfevents.1710741086.333df911e7ea.87998.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:edd1402a9a6fac3b79b8843b23a10dd0292b4ca98d83796ceb92afd397e525d8
3
- size 7990
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19cbe8ecf692d8067ea7f067ff20e6066349d9b3bc1d6c0d8be7b1a63b13a0cc
3
+ size 12011
runs/Mar18_05-31-15_333df911e7ea/events.out.tfevents.1710746588.333df911e7ea.87998.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f85df354edd5d0b045918d9c2a2713fd2850f61e853b157a473bb07e9d9c935
3
+ size 359
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.98,
3
+ "train_loss": 1.3530203068434303,
4
+ "train_runtime": 5442.6995,
5
+ "train_samples": 12450,
6
+ "train_samples_per_second": 2.376,
7
+ "train_steps_per_second": 0.037
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,300 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.9833024118738405,
5
+ "eval_steps": 500,
6
+ "global_step": 201,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.01,
13
+ "learning_rate": 9.523809523809525e-07,
14
+ "loss": 1.5933,
15
+ "step": 1
16
+ },
17
+ {
18
+ "epoch": 0.07,
19
+ "learning_rate": 4.761904761904762e-06,
20
+ "loss": 1.4782,
21
+ "step": 5
22
+ },
23
+ {
24
+ "epoch": 0.15,
25
+ "learning_rate": 9.523809523809525e-06,
26
+ "loss": 1.5256,
27
+ "step": 10
28
+ },
29
+ {
30
+ "epoch": 0.22,
31
+ "learning_rate": 1.4285714285714287e-05,
32
+ "loss": 1.4709,
33
+ "step": 15
34
+ },
35
+ {
36
+ "epoch": 0.3,
37
+ "learning_rate": 1.904761904761905e-05,
38
+ "loss": 1.5224,
39
+ "step": 20
40
+ },
41
+ {
42
+ "epoch": 0.37,
43
+ "learning_rate": 1.9975640502598243e-05,
44
+ "loss": 1.5039,
45
+ "step": 25
46
+ },
47
+ {
48
+ "epoch": 0.45,
49
+ "learning_rate": 1.9876883405951378e-05,
50
+ "loss": 1.4679,
51
+ "step": 30
52
+ },
53
+ {
54
+ "epoch": 0.52,
55
+ "learning_rate": 1.9702957262759964e-05,
56
+ "loss": 1.4341,
57
+ "step": 35
58
+ },
59
+ {
60
+ "epoch": 0.59,
61
+ "learning_rate": 1.945518575599317e-05,
62
+ "loss": 1.4291,
63
+ "step": 40
64
+ },
65
+ {
66
+ "epoch": 0.67,
67
+ "learning_rate": 1.913545457642601e-05,
68
+ "loss": 1.4145,
69
+ "step": 45
70
+ },
71
+ {
72
+ "epoch": 0.74,
73
+ "learning_rate": 1.874619707139396e-05,
74
+ "loss": 1.4082,
75
+ "step": 50
76
+ },
77
+ {
78
+ "epoch": 0.82,
79
+ "learning_rate": 1.8290375725550417e-05,
80
+ "loss": 1.3682,
81
+ "step": 55
82
+ },
83
+ {
84
+ "epoch": 0.89,
85
+ "learning_rate": 1.777145961456971e-05,
86
+ "loss": 1.3838,
87
+ "step": 60
88
+ },
89
+ {
90
+ "epoch": 0.96,
91
+ "learning_rate": 1.7193398003386514e-05,
92
+ "loss": 1.3204,
93
+ "step": 65
94
+ },
95
+ {
96
+ "epoch": 0.99,
97
+ "eval_loss": 1.3371310234069824,
98
+ "eval_runtime": 60.5562,
99
+ "eval_samples_per_second": 8.207,
100
+ "eval_steps_per_second": 2.064,
101
+ "step": 67
102
+ },
103
+ {
104
+ "epoch": 1.04,
105
+ "learning_rate": 1.6560590289905074e-05,
106
+ "loss": 1.3366,
107
+ "step": 70
108
+ },
109
+ {
110
+ "epoch": 1.11,
111
+ "learning_rate": 1.5877852522924733e-05,
112
+ "loss": 1.357,
113
+ "step": 75
114
+ },
115
+ {
116
+ "epoch": 1.19,
117
+ "learning_rate": 1.5150380749100545e-05,
118
+ "loss": 1.3083,
119
+ "step": 80
120
+ },
121
+ {
122
+ "epoch": 1.26,
123
+ "learning_rate": 1.4383711467890776e-05,
124
+ "loss": 1.3313,
125
+ "step": 85
126
+ },
127
+ {
128
+ "epoch": 1.34,
129
+ "learning_rate": 1.3583679495453e-05,
130
+ "loss": 1.3263,
131
+ "step": 90
132
+ },
133
+ {
134
+ "epoch": 1.41,
135
+ "learning_rate": 1.2756373558169992e-05,
136
+ "loss": 1.343,
137
+ "step": 95
138
+ },
139
+ {
140
+ "epoch": 1.48,
141
+ "learning_rate": 1.190808995376545e-05,
142
+ "loss": 1.3439,
143
+ "step": 100
144
+ },
145
+ {
146
+ "epoch": 1.56,
147
+ "learning_rate": 1.1045284632676535e-05,
148
+ "loss": 1.2764,
149
+ "step": 105
150
+ },
151
+ {
152
+ "epoch": 1.63,
153
+ "learning_rate": 1.0174524064372837e-05,
154
+ "loss": 1.3138,
155
+ "step": 110
156
+ },
157
+ {
158
+ "epoch": 1.71,
159
+ "learning_rate": 9.302435262558748e-06,
160
+ "loss": 1.3108,
161
+ "step": 115
162
+ },
163
+ {
164
+ "epoch": 1.78,
165
+ "learning_rate": 8.43565534959769e-06,
166
+ "loss": 1.2986,
167
+ "step": 120
168
+ },
169
+ {
170
+ "epoch": 1.86,
171
+ "learning_rate": 7.580781044003324e-06,
172
+ "loss": 1.2803,
173
+ "step": 125
174
+ },
175
+ {
176
+ "epoch": 1.93,
177
+ "learning_rate": 6.744318455428436e-06,
178
+ "loss": 1.3083,
179
+ "step": 130
180
+ },
181
+ {
182
+ "epoch": 1.99,
183
+ "eval_loss": 1.2949973344802856,
184
+ "eval_runtime": 60.6302,
185
+ "eval_samples_per_second": 8.197,
186
+ "eval_steps_per_second": 2.062,
187
+ "step": 134
188
+ },
189
+ {
190
+ "epoch": 2.0,
191
+ "learning_rate": 5.932633569242e-06,
192
+ "loss": 1.3041,
193
+ "step": 135
194
+ },
195
+ {
196
+ "epoch": 2.08,
197
+ "learning_rate": 5.151903797536631e-06,
198
+ "loss": 1.2992,
199
+ "step": 140
200
+ },
201
+ {
202
+ "epoch": 2.15,
203
+ "learning_rate": 4.408070965292534e-06,
204
+ "loss": 1.2984,
205
+ "step": 145
206
+ },
207
+ {
208
+ "epoch": 2.23,
209
+ "learning_rate": 3.7067960895016277e-06,
210
+ "loss": 1.284,
211
+ "step": 150
212
+ },
213
+ {
214
+ "epoch": 2.3,
215
+ "learning_rate": 3.0534162954100264e-06,
216
+ "loss": 1.3143,
217
+ "step": 155
218
+ },
219
+ {
220
+ "epoch": 2.37,
221
+ "learning_rate": 2.45290419777228e-06,
222
+ "loss": 1.3138,
223
+ "step": 160
224
+ },
225
+ {
226
+ "epoch": 2.45,
227
+ "learning_rate": 1.9098300562505266e-06,
228
+ "loss": 1.3192,
229
+ "step": 165
230
+ },
231
+ {
232
+ "epoch": 2.52,
233
+ "learning_rate": 1.4283269929788779e-06,
234
+ "loss": 1.3079,
235
+ "step": 170
236
+ },
237
+ {
238
+ "epoch": 2.6,
239
+ "learning_rate": 1.012059537008332e-06,
240
+ "loss": 1.3073,
241
+ "step": 175
242
+ },
243
+ {
244
+ "epoch": 2.67,
245
+ "learning_rate": 6.641957350279838e-07,
246
+ "loss": 1.3209,
247
+ "step": 180
248
+ },
249
+ {
250
+ "epoch": 2.75,
251
+ "learning_rate": 3.8738304061681107e-07,
252
+ "loss": 1.3137,
253
+ "step": 185
254
+ },
255
+ {
256
+ "epoch": 2.82,
257
+ "learning_rate": 1.8372816552336025e-07,
258
+ "loss": 1.2787,
259
+ "step": 190
260
+ },
261
+ {
262
+ "epoch": 2.89,
263
+ "learning_rate": 5.4781046317267103e-08,
264
+ "loss": 1.2959,
265
+ "step": 195
266
+ },
267
+ {
268
+ "epoch": 2.97,
269
+ "learning_rate": 1.5230484360873043e-09,
270
+ "loss": 1.2921,
271
+ "step": 200
272
+ },
273
+ {
274
+ "epoch": 2.98,
275
+ "eval_loss": 1.2931873798370361,
276
+ "eval_runtime": 60.4523,
277
+ "eval_samples_per_second": 8.221,
278
+ "eval_steps_per_second": 2.068,
279
+ "step": 201
280
+ },
281
+ {
282
+ "epoch": 2.98,
283
+ "step": 201,
284
+ "total_flos": 1.0527084092262973e+18,
285
+ "train_loss": 1.3530203068434303,
286
+ "train_runtime": 5442.6995,
287
+ "train_samples_per_second": 2.376,
288
+ "train_steps_per_second": 0.037
289
+ }
290
+ ],
291
+ "logging_steps": 5,
292
+ "max_steps": 201,
293
+ "num_input_tokens_seen": 0,
294
+ "num_train_epochs": 3,
295
+ "save_steps": 100,
296
+ "total_flos": 1.0527084092262973e+18,
297
+ "train_batch_size": 2,
298
+ "trial_name": null,
299
+ "trial_params": null
300
+ }