TinyPixel commited on
Commit
0041342
1 Parent(s): 14854af

Upload folder using huggingface_hub

Browse files
adapter_config.json CHANGED
@@ -21,10 +21,10 @@
21
  "target_modules": [
22
  "q_proj",
23
  "o_proj",
24
- "gate_proj",
25
  "k_proj",
26
  "up_proj",
27
  "v_proj",
 
28
  "down_proj"
29
  ],
30
  "task_type": "CAUSAL_LM"
 
21
  "target_modules": [
22
  "q_proj",
23
  "o_proj",
 
24
  "k_proj",
25
  "up_proj",
26
  "v_proj",
27
+ "gate_proj",
28
  "down_proj"
29
  ],
30
  "task_type": "CAUSAL_LM"
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa405f6df81cd6621d9fecf2363a7b3292f01b7790557a885c26421a53061412
3
  size 100198584
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cce465d74780ff1c8fa14ad786f4777949911336c60860b4e56c9b9678538a0
3
  size 100198584
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5739b541d9500aa245f1e5d17fa010f4ce2fe6299f41ccd02e7a615ae4291b4f
3
  size 200545402
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6977072d659bb8859c1b457d18cbf53ab50a4eb6878e72de42443069ddf94d2b
3
  size 200545402
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:96134f70cea7e3c39093ded4a11b22c9c544ebe4cafa4eef41fc7db0ed207135
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96b5c3a92d13d9774559b1e8b732f80e8225d87c7f39ade0eb68f762f1a0d92b
3
  size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4372b8bc129cb832d61f89de539110be4b375eeb5fe693144c4dd880623a9aaf
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6468d7475bc9eb4c0b562b23ba181567a539eb1d912340d31b8d2ddab88bd51c
3
  size 1064
special_tokens_map.json CHANGED
@@ -1,4 +1,20 @@
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "bos_token": {
3
  "content": "<|endoftext|>",
4
  "lstrip": false,
@@ -14,7 +30,7 @@
14
  "single_word": false
15
  },
16
  "pad_token": {
17
- "content": "[PAD]",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
 
1
  {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<|prompter|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<|assistant|>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ }
17
+ ],
18
  "bos_token": {
19
  "content": "<|endoftext|>",
20
  "lstrip": false,
 
30
  "single_word": false
31
  },
32
  "pad_token": {
33
+ "content": "<|padding|>",
34
  "lstrip": false,
35
  "normalized": false,
36
  "rstrip": false,
tokenizer.json CHANGED
@@ -230,7 +230,16 @@
230
  },
231
  {
232
  "id": 50277,
233
- "content": "[PAD]",
 
 
 
 
 
 
 
 
 
234
  "single_word": false,
235
  "lstrip": false,
236
  "rstrip": false,
 
230
  },
231
  {
232
  "id": 50277,
233
+ "content": "<|prompter|>",
234
+ "single_word": false,
235
+ "lstrip": false,
236
+ "rstrip": false,
237
+ "normalized": false,
238
+ "special": true
239
+ },
240
+ {
241
+ "id": 50278,
242
+ "content": "<|assistant|>",
243
  "single_word": false,
244
  "lstrip": false,
245
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -202,7 +202,15 @@
202
  "special": false
203
  },
204
  "50277": {
205
- "content": "[PAD]",
 
 
 
 
 
 
 
 
206
  "lstrip": false,
207
  "normalized": false,
208
  "rstrip": false,
@@ -210,11 +218,15 @@
210
  "special": true
211
  }
212
  },
 
 
 
 
213
  "bos_token": "<|endoftext|>",
214
  "clean_up_tokenization_spaces": true,
215
  "eos_token": "<|endoftext|>",
216
  "model_max_length": 1000000000000000019884624838656,
217
- "pad_token": "[PAD]",
218
  "tokenizer_class": "GPTNeoXTokenizer",
219
  "unk_token": "<|endoftext|>"
220
  }
 
202
  "special": false
203
  },
204
  "50277": {
205
+ "content": "<|prompter|>",
206
+ "lstrip": false,
207
+ "normalized": false,
208
+ "rstrip": false,
209
+ "single_word": false,
210
+ "special": true
211
+ },
212
+ "50278": {
213
+ "content": "<|assistant|>",
214
  "lstrip": false,
215
  "normalized": false,
216
  "rstrip": false,
 
218
  "special": true
219
  }
220
  },
221
+ "additional_special_tokens": [
222
+ "<|prompter|>",
223
+ "<|assistant|>"
224
+ ],
225
  "bos_token": "<|endoftext|>",
226
  "clean_up_tokenization_spaces": true,
227
  "eos_token": "<|endoftext|>",
228
  "model_max_length": 1000000000000000019884624838656,
229
+ "pad_token": "<|padding|>",
230
  "tokenizer_class": "GPTNeoXTokenizer",
231
  "unk_token": "<|endoftext|>"
232
  }
trainer_state.json CHANGED
@@ -11,145 +11,145 @@
11
  {
12
  "epoch": 0.04,
13
  "learning_rate": 2e-05,
14
- "loss": 1.8529,
15
  "step": 2
16
  },
17
  {
18
  "epoch": 0.08,
19
  "learning_rate": 2e-05,
20
- "loss": 2.219,
21
  "step": 4
22
  },
23
  {
24
  "epoch": 0.12,
25
  "learning_rate": 2e-05,
26
- "loss": 2.2383,
27
  "step": 6
28
  },
29
  {
30
  "epoch": 0.16,
31
  "learning_rate": 2e-05,
32
- "loss": 2.3014,
33
  "step": 8
34
  },
35
  {
36
  "epoch": 0.21,
37
  "learning_rate": 2e-05,
38
- "loss": 2.4572,
39
  "step": 10
40
  },
41
  {
42
  "epoch": 0.25,
43
  "learning_rate": 2e-05,
44
- "loss": 2.7778,
45
  "step": 12
46
  },
47
  {
48
  "epoch": 0.29,
49
  "learning_rate": 2e-05,
50
- "loss": 1.9011,
51
  "step": 14
52
  },
53
  {
54
  "epoch": 0.33,
55
  "learning_rate": 2e-05,
56
- "loss": 1.8648,
57
  "step": 16
58
  },
59
  {
60
  "epoch": 0.37,
61
  "learning_rate": 2e-05,
62
- "loss": 2.1674,
63
  "step": 18
64
  },
65
  {
66
  "epoch": 0.41,
67
  "learning_rate": 2e-05,
68
- "loss": 2.0898,
69
  "step": 20
70
  },
71
  {
72
  "epoch": 0.45,
73
  "learning_rate": 2e-05,
74
- "loss": 2.326,
75
  "step": 22
76
  },
77
  {
78
  "epoch": 0.49,
79
  "learning_rate": 2e-05,
80
- "loss": 2.5082,
81
  "step": 24
82
  },
83
  {
84
  "epoch": 0.53,
85
  "learning_rate": 2e-05,
86
- "loss": 1.8068,
87
  "step": 26
88
  },
89
  {
90
  "epoch": 0.57,
91
  "learning_rate": 2e-05,
92
- "loss": 2.0524,
93
  "step": 28
94
  },
95
  {
96
  "epoch": 0.62,
97
  "learning_rate": 2e-05,
98
- "loss": 2.2911,
99
  "step": 30
100
  },
101
  {
102
  "epoch": 0.66,
103
  "learning_rate": 2e-05,
104
- "loss": 2.0351,
105
  "step": 32
106
  },
107
  {
108
  "epoch": 0.7,
109
  "learning_rate": 2e-05,
110
- "loss": 2.0601,
111
  "step": 34
112
  },
113
  {
114
  "epoch": 0.74,
115
  "learning_rate": 2e-05,
116
- "loss": 2.3926,
117
  "step": 36
118
  },
119
  {
120
  "epoch": 0.78,
121
  "learning_rate": 2e-05,
122
- "loss": 1.8453,
123
  "step": 38
124
  },
125
  {
126
  "epoch": 0.82,
127
  "learning_rate": 2e-05,
128
- "loss": 2.1355,
129
  "step": 40
130
  },
131
  {
132
  "epoch": 0.86,
133
  "learning_rate": 2e-05,
134
- "loss": 2.2368,
135
  "step": 42
136
  },
137
  {
138
  "epoch": 0.9,
139
  "learning_rate": 2e-05,
140
- "loss": 2.2696,
141
  "step": 44
142
  },
143
  {
144
  "epoch": 0.94,
145
  "learning_rate": 2e-05,
146
- "loss": 2.0637,
147
  "step": 46
148
  },
149
  {
150
  "epoch": 0.98,
151
  "learning_rate": 2e-05,
152
- "loss": 2.3634,
153
  "step": 48
154
  }
155
  ],
@@ -158,7 +158,7 @@
158
  "num_input_tokens_seen": 0,
159
  "num_train_epochs": 1,
160
  "save_steps": 500,
161
- "total_flos": 6747601288636416.0,
162
  "train_batch_size": 1,
163
  "trial_name": null,
164
  "trial_params": null
 
11
  {
12
  "epoch": 0.04,
13
  "learning_rate": 2e-05,
14
+ "loss": 1.8569,
15
  "step": 2
16
  },
17
  {
18
  "epoch": 0.08,
19
  "learning_rate": 2e-05,
20
+ "loss": 2.2234,
21
  "step": 4
22
  },
23
  {
24
  "epoch": 0.12,
25
  "learning_rate": 2e-05,
26
+ "loss": 2.2528,
27
  "step": 6
28
  },
29
  {
30
  "epoch": 0.16,
31
  "learning_rate": 2e-05,
32
+ "loss": 2.3016,
33
  "step": 8
34
  },
35
  {
36
  "epoch": 0.21,
37
  "learning_rate": 2e-05,
38
+ "loss": 2.5323,
39
  "step": 10
40
  },
41
  {
42
  "epoch": 0.25,
43
  "learning_rate": 2e-05,
44
+ "loss": 3.003,
45
  "step": 12
46
  },
47
  {
48
  "epoch": 0.29,
49
  "learning_rate": 2e-05,
50
+ "loss": 1.9097,
51
  "step": 14
52
  },
53
  {
54
  "epoch": 0.33,
55
  "learning_rate": 2e-05,
56
+ "loss": 1.8799,
57
  "step": 16
58
  },
59
  {
60
  "epoch": 0.37,
61
  "learning_rate": 2e-05,
62
+ "loss": 2.1973,
63
  "step": 18
64
  },
65
  {
66
  "epoch": 0.41,
67
  "learning_rate": 2e-05,
68
+ "loss": 2.1347,
69
  "step": 20
70
  },
71
  {
72
  "epoch": 0.45,
73
  "learning_rate": 2e-05,
74
+ "loss": 2.4056,
75
  "step": 22
76
  },
77
  {
78
  "epoch": 0.49,
79
  "learning_rate": 2e-05,
80
+ "loss": 2.7531,
81
  "step": 24
82
  },
83
  {
84
  "epoch": 0.53,
85
  "learning_rate": 2e-05,
86
+ "loss": 1.819,
87
  "step": 26
88
  },
89
  {
90
  "epoch": 0.57,
91
  "learning_rate": 2e-05,
92
+ "loss": 2.0862,
93
  "step": 28
94
  },
95
  {
96
  "epoch": 0.62,
97
  "learning_rate": 2e-05,
98
+ "loss": 2.3401,
99
  "step": 30
100
  },
101
  {
102
  "epoch": 0.66,
103
  "learning_rate": 2e-05,
104
+ "loss": 2.1033,
105
  "step": 32
106
  },
107
  {
108
  "epoch": 0.7,
109
  "learning_rate": 2e-05,
110
+ "loss": 2.1674,
111
  "step": 34
112
  },
113
  {
114
  "epoch": 0.74,
115
  "learning_rate": 2e-05,
116
+ "loss": 2.7192,
117
  "step": 36
118
  },
119
  {
120
  "epoch": 0.78,
121
  "learning_rate": 2e-05,
122
+ "loss": 1.8597,
123
  "step": 38
124
  },
125
  {
126
  "epoch": 0.82,
127
  "learning_rate": 2e-05,
128
+ "loss": 2.1725,
129
  "step": 40
130
  },
131
  {
132
  "epoch": 0.86,
133
  "learning_rate": 2e-05,
134
+ "loss": 2.2864,
135
  "step": 42
136
  },
137
  {
138
  "epoch": 0.9,
139
  "learning_rate": 2e-05,
140
+ "loss": 2.3444,
141
  "step": 44
142
  },
143
  {
144
  "epoch": 0.94,
145
  "learning_rate": 2e-05,
146
+ "loss": 2.1558,
147
  "step": 46
148
  },
149
  {
150
  "epoch": 0.98,
151
  "learning_rate": 2e-05,
152
+ "loss": 2.661,
153
  "step": 48
154
  }
155
  ],
 
158
  "num_input_tokens_seen": 0,
159
  "num_train_epochs": 1,
160
  "save_steps": 500,
161
+ "total_flos": 6711085690343424.0,
162
  "train_batch_size": 1,
163
  "trial_name": null,
164
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5bad23ca216ebbd9b73663d7744425ce42a5f56e20a4740af2ac918b77e5f4e
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8384811e1aeffd902b702fa834e38d6e3247b65bb333cb619770845d7087ddab
3
  size 4728