lomahony commited on
Commit
3e5dcd8
1 Parent(s): e9160ae

Upload 8 files

Browse files
base-2.8b-eval-files/EleutherAI-pythia-2.8b-0shot-shelloutput.txt ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bootstrapping for stddev: perplexity
2
+ hf (pretrained=EleutherAI/pythia-2.8b), limit: None, num_fewshot: 0, batch_size: 16
3
+ | Task |Version|Filter| Metric | Value | |Stderr|
4
+ |--------------|-------|------|---------------|------:|---|-----:|
5
+ |arc_challenge |Yaml |none |acc | 0.2952|± |0.0133|
6
+ | | |none |acc_norm | 0.3311|± |0.0138|
7
+ |arc_easy |Yaml |none |acc | 0.6448|± |0.0098|
8
+ | | |none |acc_norm | 0.5871|± |0.0101|
9
+ |boolq |Yaml |none |acc | 0.6474|± |0.0084|
10
+ |hellaswag |Yaml |none |acc | 0.4532|± |0.0050|
11
+ | | |none |acc_norm | 0.5926|± |0.0049|
12
+ |lambada_openai|Yaml |none |perplexity | 5.0370|± |0.1191|
13
+ | | |none |acc | 0.6472|± |0.0067|
14
+ |openbookqa |Yaml |none |acc | 0.2400|± |0.0191|
15
+ | | |none |acc_norm | 0.3580|± |0.0215|
16
+ |piqa |Yaml |none |acc | 0.7394|± |0.0102|
17
+ | | |none |acc_norm | 0.7356|± |0.0103|
18
+ |sciq |Yaml |none |acc | 0.8870|± |0.0100|
19
+ | | |none |acc_norm | 0.8350|± |0.0117|
20
+ |wikitext |Yaml |none |word_perplexity|20.0571| | |
21
+ | | |none |byte_perplexity| 1.6394| | |
22
+ | | |none |bits_per_byte | 0.7132| | |
23
+ |winogrande |Yaml |none |acc | 0.5975|± |0.0138|
24
+
base-2.8b-eval-files/EleutherAI-pythia-2.8b-0shot/results.json ADDED
@@ -0,0 +1,404 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.295221843003413,
5
+ "acc_stderr,none": 0.013329750293382316,
6
+ "acc_norm,none": 0.3310580204778157,
7
+ "acc_norm_stderr,none": 0.013752062419817836
8
+ },
9
+ "arc_easy": {
10
+ "acc,none": 0.6447811447811448,
11
+ "acc_stderr,none": 0.009820245899287126,
12
+ "acc_norm,none": 0.5871212121212122,
13
+ "acc_norm_stderr,none": 0.010102837421104667
14
+ },
15
+ "boolq": {
16
+ "acc,none": 0.6474006116207951,
17
+ "acc_stderr,none": 0.008356412493562119
18
+ },
19
+ "hellaswag": {
20
+ "acc,none": 0.4531965743875722,
21
+ "acc_stderr,none": 0.004967872475383267,
22
+ "acc_norm,none": 0.5926110336586338,
23
+ "acc_norm_stderr,none": 0.00490344168000382
24
+ },
25
+ "lambada_openai": {
26
+ "perplexity,none": 5.0369907596068435,
27
+ "perplexity_stderr,none": 0.11909165322070424,
28
+ "acc,none": 0.6471958082670289,
29
+ "acc_stderr,none": 0.006657279471298494
30
+ },
31
+ "openbookqa": {
32
+ "acc,none": 0.24,
33
+ "acc_stderr,none": 0.019118866653759753,
34
+ "acc_norm,none": 0.358,
35
+ "acc_norm_stderr,none": 0.02146143486285912
36
+ },
37
+ "piqa": {
38
+ "acc,none": 0.7393906420021763,
39
+ "acc_stderr,none": 0.010241826155811627,
40
+ "acc_norm,none": 0.735582154515778,
41
+ "acc_norm_stderr,none": 0.01028978724476717
42
+ },
43
+ "sciq": {
44
+ "acc,none": 0.887,
45
+ "acc_stderr,none": 0.010016552866696846,
46
+ "acc_norm,none": 0.835,
47
+ "acc_norm_stderr,none": 0.011743632866916171
48
+ },
49
+ "wikitext": {
50
+ "word_perplexity,none": 20.05713713364702,
51
+ "byte_perplexity,none": 1.6394022363791052,
52
+ "bits_per_byte,none": 0.7131698710530711
53
+ },
54
+ "winogrande": {
55
+ "acc,none": 0.5974743488555643,
56
+ "acc_stderr,none": 0.013782866831703043
57
+ }
58
+ },
59
+ "configs": {
60
+ "arc_challenge": {
61
+ "task": "arc_challenge",
62
+ "group": [
63
+ "ai2_arc",
64
+ "multiple_choice"
65
+ ],
66
+ "dataset_path": "ai2_arc",
67
+ "dataset_name": "ARC-Challenge",
68
+ "training_split": "train",
69
+ "validation_split": "validation",
70
+ "test_split": "test",
71
+ "doc_to_text": "Question: {{question}}\nAnswer:",
72
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
73
+ "doc_to_choice": "{{choices.text}}",
74
+ "description": "",
75
+ "target_delimiter": " ",
76
+ "fewshot_delimiter": "\n\n",
77
+ "num_fewshot": 0,
78
+ "metric_list": [
79
+ {
80
+ "metric": "acc",
81
+ "aggregation": "mean",
82
+ "higher_is_better": true
83
+ },
84
+ {
85
+ "metric": "acc_norm",
86
+ "aggregation": "mean",
87
+ "higher_is_better": true
88
+ }
89
+ ],
90
+ "output_type": "multiple_choice",
91
+ "repeats": 1,
92
+ "should_decontaminate": true,
93
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
94
+ },
95
+ "arc_easy": {
96
+ "task": "arc_easy",
97
+ "group": [
98
+ "ai2_arc",
99
+ "multiple_choice"
100
+ ],
101
+ "dataset_path": "ai2_arc",
102
+ "dataset_name": "ARC-Easy",
103
+ "training_split": "train",
104
+ "validation_split": "validation",
105
+ "test_split": "test",
106
+ "doc_to_text": "Question: {{question}}\nAnswer:",
107
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
108
+ "doc_to_choice": "{{choices.text}}",
109
+ "description": "",
110
+ "target_delimiter": " ",
111
+ "fewshot_delimiter": "\n\n",
112
+ "num_fewshot": 0,
113
+ "metric_list": [
114
+ {
115
+ "metric": "acc",
116
+ "aggregation": "mean",
117
+ "higher_is_better": true
118
+ },
119
+ {
120
+ "metric": "acc_norm",
121
+ "aggregation": "mean",
122
+ "higher_is_better": true
123
+ }
124
+ ],
125
+ "output_type": "multiple_choice",
126
+ "repeats": 1,
127
+ "should_decontaminate": true,
128
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
129
+ },
130
+ "boolq": {
131
+ "task": "boolq",
132
+ "group": [
133
+ "super-glue-lm-eval-v1"
134
+ ],
135
+ "dataset_path": "super_glue",
136
+ "dataset_name": "boolq",
137
+ "training_split": "train",
138
+ "validation_split": "validation",
139
+ "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:",
140
+ "doc_to_target": "label",
141
+ "doc_to_choice": [
142
+ "no",
143
+ "yes"
144
+ ],
145
+ "description": "",
146
+ "target_delimiter": " ",
147
+ "fewshot_delimiter": "\n\n",
148
+ "num_fewshot": 0,
149
+ "metric_list": [
150
+ {
151
+ "metric": "acc"
152
+ }
153
+ ],
154
+ "output_type": "multiple_choice",
155
+ "repeats": 1,
156
+ "should_decontaminate": true,
157
+ "doc_to_decontamination_query": "passage"
158
+ },
159
+ "hellaswag": {
160
+ "task": "hellaswag",
161
+ "group": [
162
+ "multiple_choice"
163
+ ],
164
+ "dataset_path": "hellaswag",
165
+ "training_split": "train",
166
+ "validation_split": "validation",
167
+ "doc_to_text": "{% set text = activity_label ~ ': ' ~ ctx_a ~ ' ' ~ ctx_b.capitalize() %}{{text|trim|replace(' [title]', '. ')|regex_replace('\\[.*?\\]', '')|replace(' ', ' ')}}",
168
+ "doc_to_target": "{{label}}",
169
+ "doc_to_choice": "{{endings|map('trim')|map('replace', ' [title]', '. ')|map('regex_replace', '\\[.*?\\]', '')|map('replace', ' ', ' ')|list}}",
170
+ "description": "",
171
+ "target_delimiter": " ",
172
+ "fewshot_delimiter": "\n\n",
173
+ "num_fewshot": 0,
174
+ "metric_list": [
175
+ {
176
+ "metric": "acc",
177
+ "aggregation": "mean",
178
+ "higher_is_better": true
179
+ },
180
+ {
181
+ "metric": "acc_norm",
182
+ "aggregation": "mean",
183
+ "higher_is_better": true
184
+ }
185
+ ],
186
+ "output_type": "multiple_choice",
187
+ "repeats": 1,
188
+ "should_decontaminate": false
189
+ },
190
+ "lambada_openai": {
191
+ "task": "lambada_openai",
192
+ "group": [
193
+ "lambada",
194
+ "loglikelihood",
195
+ "perplexity"
196
+ ],
197
+ "dataset_path": "EleutherAI/lambada_openai",
198
+ "dataset_name": "default",
199
+ "test_split": "test",
200
+ "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
201
+ "doc_to_target": "{{' '+text.split(' ')[-1]}}",
202
+ "description": "",
203
+ "target_delimiter": " ",
204
+ "fewshot_delimiter": "\n\n",
205
+ "num_fewshot": 0,
206
+ "metric_list": [
207
+ {
208
+ "metric": "perplexity",
209
+ "aggregation": "perplexity",
210
+ "higher_is_better": false
211
+ },
212
+ {
213
+ "metric": "acc",
214
+ "aggregation": "mean",
215
+ "higher_is_better": true
216
+ }
217
+ ],
218
+ "output_type": "loglikelihood",
219
+ "repeats": 1,
220
+ "should_decontaminate": true,
221
+ "doc_to_decontamination_query": "{{text}}"
222
+ },
223
+ "openbookqa": {
224
+ "task": "openbookqa",
225
+ "group": [
226
+ "multiple_choice"
227
+ ],
228
+ "dataset_path": "openbookqa",
229
+ "dataset_name": "main",
230
+ "training_split": "train",
231
+ "validation_split": "validation",
232
+ "test_split": "test",
233
+ "doc_to_text": "question_stem",
234
+ "doc_to_target": "{{choices.label.index(answerKey.lstrip())}}",
235
+ "doc_to_choice": "{{choices.text}}",
236
+ "description": "",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "num_fewshot": 0,
240
+ "metric_list": [
241
+ {
242
+ "metric": "acc",
243
+ "aggregation": "mean",
244
+ "higher_is_better": true
245
+ },
246
+ {
247
+ "metric": "acc_norm",
248
+ "aggregation": "mean",
249
+ "higher_is_better": true
250
+ }
251
+ ],
252
+ "output_type": "multiple_choice",
253
+ "repeats": 1,
254
+ "should_decontaminate": true,
255
+ "doc_to_decontamination_query": "question_stem"
256
+ },
257
+ "piqa": {
258
+ "task": "piqa",
259
+ "group": [
260
+ "multiple_choice"
261
+ ],
262
+ "dataset_path": "piqa",
263
+ "training_split": "train",
264
+ "validation_split": "validation",
265
+ "doc_to_text": "Question: {{goal}}\nAnswer:",
266
+ "doc_to_target": "label",
267
+ "doc_to_choice": "{{[sol1, sol2]}}",
268
+ "description": "",
269
+ "target_delimiter": " ",
270
+ "fewshot_delimiter": "\n\n",
271
+ "num_fewshot": 0,
272
+ "metric_list": [
273
+ {
274
+ "metric": "acc",
275
+ "aggregation": "mean",
276
+ "higher_is_better": true
277
+ },
278
+ {
279
+ "metric": "acc_norm",
280
+ "aggregation": "mean",
281
+ "higher_is_better": true
282
+ }
283
+ ],
284
+ "output_type": "multiple_choice",
285
+ "repeats": 1,
286
+ "should_decontaminate": true,
287
+ "doc_to_decontamination_query": "goal"
288
+ },
289
+ "sciq": {
290
+ "task": "sciq",
291
+ "group": [
292
+ "multiple_choice"
293
+ ],
294
+ "dataset_path": "sciq",
295
+ "training_split": "train",
296
+ "validation_split": "validation",
297
+ "test_split": "test",
298
+ "doc_to_text": "{{support.lstrip()}}\nQuestion: {{question}}\nAnswer:",
299
+ "doc_to_target": 3,
300
+ "doc_to_choice": "{{[distractor1, distractor2, distractor3, correct_answer]}}",
301
+ "description": "",
302
+ "target_delimiter": " ",
303
+ "fewshot_delimiter": "\n\n",
304
+ "num_fewshot": 0,
305
+ "metric_list": [
306
+ {
307
+ "metric": "acc",
308
+ "aggregation": "mean",
309
+ "higher_is_better": true
310
+ },
311
+ {
312
+ "metric": "acc_norm",
313
+ "aggregation": "mean",
314
+ "higher_is_better": true
315
+ }
316
+ ],
317
+ "output_type": "multiple_choice",
318
+ "repeats": 1,
319
+ "should_decontaminate": true,
320
+ "doc_to_decontamination_query": "{{support}} {{question}}"
321
+ },
322
+ "wikitext": {
323
+ "task": "wikitext",
324
+ "group": [
325
+ "perplexity",
326
+ "loglikelihood_rolling"
327
+ ],
328
+ "dataset_path": "EleutherAI/wikitext_document_level",
329
+ "dataset_name": "wikitext-2-raw-v1",
330
+ "training_split": "train",
331
+ "validation_split": "validation",
332
+ "test_split": "test",
333
+ "doc_to_text": "",
334
+ "doc_to_target": "<function wikitext_detokenizer at 0x7fbdfeb25120>",
335
+ "description": "",
336
+ "target_delimiter": " ",
337
+ "fewshot_delimiter": "\n\n",
338
+ "num_fewshot": 0,
339
+ "metric_list": [
340
+ {
341
+ "metric": "word_perplexity"
342
+ },
343
+ {
344
+ "metric": "byte_perplexity"
345
+ },
346
+ {
347
+ "metric": "bits_per_byte"
348
+ }
349
+ ],
350
+ "output_type": "loglikelihood_rolling",
351
+ "repeats": 1,
352
+ "should_decontaminate": true,
353
+ "doc_to_decontamination_query": "{{page}}"
354
+ },
355
+ "winogrande": {
356
+ "task": "winogrande",
357
+ "dataset_path": "winogrande",
358
+ "dataset_name": "winogrande_xl",
359
+ "training_split": "train",
360
+ "validation_split": "validation",
361
+ "doc_to_text": "<function doc_to_text at 0x7fbdfeb25360>",
362
+ "doc_to_target": "<function doc_to_target at 0x7fbdfeb256c0>",
363
+ "doc_to_choice": "<function doc_to_choice at 0x7fbdfeb25a20>",
364
+ "description": "",
365
+ "target_delimiter": " ",
366
+ "fewshot_delimiter": "\n\n",
367
+ "num_fewshot": 0,
368
+ "metric_list": [
369
+ {
370
+ "metric": "acc",
371
+ "aggregation": "mean",
372
+ "higher_is_better": true
373
+ }
374
+ ],
375
+ "output_type": "multiple_choice",
376
+ "repeats": 1,
377
+ "should_decontaminate": true,
378
+ "doc_to_decontamination_query": "sentence"
379
+ }
380
+ },
381
+ "versions": {
382
+ "arc_challenge": "Yaml",
383
+ "arc_easy": "Yaml",
384
+ "boolq": "Yaml",
385
+ "hellaswag": "Yaml",
386
+ "lambada_openai": "Yaml",
387
+ "openbookqa": "Yaml",
388
+ "piqa": "Yaml",
389
+ "sciq": "Yaml",
390
+ "wikitext": "Yaml",
391
+ "winogrande": "Yaml"
392
+ },
393
+ "config": {
394
+ "model": "hf",
395
+ "model_args": "pretrained=EleutherAI/pythia-2.8b",
396
+ "batch_size": "16",
397
+ "batch_sizes": [],
398
+ "device": null,
399
+ "use_cache": null,
400
+ "limit": null,
401
+ "bootstrap_iters": 100000
402
+ },
403
+ "git_hash": "d1a44c8"
404
+ }
base-2.8b-eval-files/EleutherAI-pythia-2.8b-5shot-shelloutput.txt ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bootstrapping for stddev: perplexity
2
+ hf (pretrained=EleutherAI/pythia-2.8b), limit: None, num_fewshot: 5, batch_size: 8
3
+ | Task |Version|Filter| Metric | Value | |Stderr|
4
+ |--------------|-------|------|---------------|------:|---|-----:|
5
+ |arc_challenge |Yaml |none |acc | 0.3123|± |0.0135|
6
+ | | |none |acc_norm | 0.3549|± |0.0140|
7
+ |arc_easy |Yaml |none |acc | 0.6700|± |0.0096|
8
+ | | |none |acc_norm | 0.6759|± |0.0096|
9
+ |boolq |Yaml |none |acc | 0.6664|± |0.0082|
10
+ |hellaswag |Yaml |none |acc | 0.4504|± |0.0050|
11
+ | | |none |acc_norm | 0.6031|± |0.0049|
12
+ |lambada_openai|Yaml |none |perplexity | 6.4479|± |0.1602|
13
+ | | |none |acc | 0.5948|± |0.0068|
14
+ |openbookqa |Yaml |none |acc | 0.2480|± |0.0193|
15
+ | | |none |acc_norm | 0.3720|± |0.0216|
16
+ |piqa |Yaml |none |acc | 0.7427|± |0.0102|
17
+ | | |none |acc_norm | 0.7443|± |0.0102|
18
+ |sciq |Yaml |none |acc | 0.9440|± |0.0073|
19
+ | | |none |acc_norm | 0.9400|± |0.0075|
20
+ |wikitext |Yaml |none |word_perplexity|20.0571| | |
21
+ | | |none |byte_perplexity| 1.6394| | |
22
+ | | |none |bits_per_byte | 0.7132| | |
23
+ |winogrande |Yaml |none |acc | 0.5943|± |0.0138|
24
+
base-2.8b-eval-files/EleutherAI-pythia-2.8b-5shot/results.json ADDED
@@ -0,0 +1,404 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.3122866894197952,
5
+ "acc_stderr,none": 0.013542598541688065,
6
+ "acc_norm,none": 0.35494880546075086,
7
+ "acc_norm_stderr,none": 0.01398303690409409
8
+ },
9
+ "arc_easy": {
10
+ "acc,none": 0.67003367003367,
11
+ "acc_stderr,none": 0.009648311574241036,
12
+ "acc_norm,none": 0.6759259259259259,
13
+ "acc_norm_stderr,none": 0.009603728850095387
14
+ },
15
+ "boolq": {
16
+ "acc,none": 0.6663608562691131,
17
+ "acc_stderr,none": 0.008246805985556868
18
+ },
19
+ "hellaswag": {
20
+ "acc,none": 0.450408285202151,
21
+ "acc_stderr,none": 0.0049651776330499236,
22
+ "acc_norm,none": 0.6030671181039634,
23
+ "acc_norm_stderr,none": 0.0048826194841666
24
+ },
25
+ "lambada_openai": {
26
+ "perplexity,none": 6.447861085824152,
27
+ "perplexity_stderr,none": 0.16023935097204245,
28
+ "acc,none": 0.5947991461284688,
29
+ "acc_stderr,none": 0.0068396269826581525
30
+ },
31
+ "openbookqa": {
32
+ "acc,none": 0.248,
33
+ "acc_stderr,none": 0.019332342821239103,
34
+ "acc_norm,none": 0.372,
35
+ "acc_norm_stderr,none": 0.0216371979857224
36
+ },
37
+ "piqa": {
38
+ "acc,none": 0.7426550598476604,
39
+ "acc_stderr,none": 0.01019992106479251,
40
+ "acc_norm,none": 0.7442872687704026,
41
+ "acc_norm_stderr,none": 0.010178690109459857
42
+ },
43
+ "sciq": {
44
+ "acc,none": 0.944,
45
+ "acc_stderr,none": 0.007274401481697066,
46
+ "acc_norm,none": 0.94,
47
+ "acc_norm_stderr,none": 0.007513751157474921
48
+ },
49
+ "wikitext": {
50
+ "word_perplexity,none": 20.05713713364702,
51
+ "byte_perplexity,none": 1.6394022363791052,
52
+ "bits_per_byte,none": 0.7131698710530711
53
+ },
54
+ "winogrande": {
55
+ "acc,none": 0.5943172849250198,
56
+ "acc_stderr,none": 0.013800206336014201
57
+ }
58
+ },
59
+ "configs": {
60
+ "arc_challenge": {
61
+ "task": "arc_challenge",
62
+ "group": [
63
+ "ai2_arc",
64
+ "multiple_choice"
65
+ ],
66
+ "dataset_path": "ai2_arc",
67
+ "dataset_name": "ARC-Challenge",
68
+ "training_split": "train",
69
+ "validation_split": "validation",
70
+ "test_split": "test",
71
+ "doc_to_text": "Question: {{question}}\nAnswer:",
72
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
73
+ "doc_to_choice": "{{choices.text}}",
74
+ "description": "",
75
+ "target_delimiter": " ",
76
+ "fewshot_delimiter": "\n\n",
77
+ "num_fewshot": 5,
78
+ "metric_list": [
79
+ {
80
+ "metric": "acc",
81
+ "aggregation": "mean",
82
+ "higher_is_better": true
83
+ },
84
+ {
85
+ "metric": "acc_norm",
86
+ "aggregation": "mean",
87
+ "higher_is_better": true
88
+ }
89
+ ],
90
+ "output_type": "multiple_choice",
91
+ "repeats": 1,
92
+ "should_decontaminate": true,
93
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
94
+ },
95
+ "arc_easy": {
96
+ "task": "arc_easy",
97
+ "group": [
98
+ "ai2_arc",
99
+ "multiple_choice"
100
+ ],
101
+ "dataset_path": "ai2_arc",
102
+ "dataset_name": "ARC-Easy",
103
+ "training_split": "train",
104
+ "validation_split": "validation",
105
+ "test_split": "test",
106
+ "doc_to_text": "Question: {{question}}\nAnswer:",
107
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
108
+ "doc_to_choice": "{{choices.text}}",
109
+ "description": "",
110
+ "target_delimiter": " ",
111
+ "fewshot_delimiter": "\n\n",
112
+ "num_fewshot": 5,
113
+ "metric_list": [
114
+ {
115
+ "metric": "acc",
116
+ "aggregation": "mean",
117
+ "higher_is_better": true
118
+ },
119
+ {
120
+ "metric": "acc_norm",
121
+ "aggregation": "mean",
122
+ "higher_is_better": true
123
+ }
124
+ ],
125
+ "output_type": "multiple_choice",
126
+ "repeats": 1,
127
+ "should_decontaminate": true,
128
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
129
+ },
130
+ "boolq": {
131
+ "task": "boolq",
132
+ "group": [
133
+ "super-glue-lm-eval-v1"
134
+ ],
135
+ "dataset_path": "super_glue",
136
+ "dataset_name": "boolq",
137
+ "training_split": "train",
138
+ "validation_split": "validation",
139
+ "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:",
140
+ "doc_to_target": "label",
141
+ "doc_to_choice": [
142
+ "no",
143
+ "yes"
144
+ ],
145
+ "description": "",
146
+ "target_delimiter": " ",
147
+ "fewshot_delimiter": "\n\n",
148
+ "num_fewshot": 5,
149
+ "metric_list": [
150
+ {
151
+ "metric": "acc"
152
+ }
153
+ ],
154
+ "output_type": "multiple_choice",
155
+ "repeats": 1,
156
+ "should_decontaminate": true,
157
+ "doc_to_decontamination_query": "passage"
158
+ },
159
+ "hellaswag": {
160
+ "task": "hellaswag",
161
+ "group": [
162
+ "multiple_choice"
163
+ ],
164
+ "dataset_path": "hellaswag",
165
+ "training_split": "train",
166
+ "validation_split": "validation",
167
+ "doc_to_text": "{% set text = activity_label ~ ': ' ~ ctx_a ~ ' ' ~ ctx_b.capitalize() %}{{text|trim|replace(' [title]', '. ')|regex_replace('\\[.*?\\]', '')|replace(' ', ' ')}}",
168
+ "doc_to_target": "{{label}}",
169
+ "doc_to_choice": "{{endings|map('trim')|map('replace', ' [title]', '. ')|map('regex_replace', '\\[.*?\\]', '')|map('replace', ' ', ' ')|list}}",
170
+ "description": "",
171
+ "target_delimiter": " ",
172
+ "fewshot_delimiter": "\n\n",
173
+ "num_fewshot": 5,
174
+ "metric_list": [
175
+ {
176
+ "metric": "acc",
177
+ "aggregation": "mean",
178
+ "higher_is_better": true
179
+ },
180
+ {
181
+ "metric": "acc_norm",
182
+ "aggregation": "mean",
183
+ "higher_is_better": true
184
+ }
185
+ ],
186
+ "output_type": "multiple_choice",
187
+ "repeats": 1,
188
+ "should_decontaminate": false
189
+ },
190
+ "lambada_openai": {
191
+ "task": "lambada_openai",
192
+ "group": [
193
+ "lambada",
194
+ "loglikelihood",
195
+ "perplexity"
196
+ ],
197
+ "dataset_path": "EleutherAI/lambada_openai",
198
+ "dataset_name": "default",
199
+ "test_split": "test",
200
+ "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
201
+ "doc_to_target": "{{' '+text.split(' ')[-1]}}",
202
+ "description": "",
203
+ "target_delimiter": " ",
204
+ "fewshot_delimiter": "\n\n",
205
+ "num_fewshot": 5,
206
+ "metric_list": [
207
+ {
208
+ "metric": "perplexity",
209
+ "aggregation": "perplexity",
210
+ "higher_is_better": false
211
+ },
212
+ {
213
+ "metric": "acc",
214
+ "aggregation": "mean",
215
+ "higher_is_better": true
216
+ }
217
+ ],
218
+ "output_type": "loglikelihood",
219
+ "repeats": 1,
220
+ "should_decontaminate": true,
221
+ "doc_to_decontamination_query": "{{text}}"
222
+ },
223
+ "openbookqa": {
224
+ "task": "openbookqa",
225
+ "group": [
226
+ "multiple_choice"
227
+ ],
228
+ "dataset_path": "openbookqa",
229
+ "dataset_name": "main",
230
+ "training_split": "train",
231
+ "validation_split": "validation",
232
+ "test_split": "test",
233
+ "doc_to_text": "question_stem",
234
+ "doc_to_target": "{{choices.label.index(answerKey.lstrip())}}",
235
+ "doc_to_choice": "{{choices.text}}",
236
+ "description": "",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "num_fewshot": 5,
240
+ "metric_list": [
241
+ {
242
+ "metric": "acc",
243
+ "aggregation": "mean",
244
+ "higher_is_better": true
245
+ },
246
+ {
247
+ "metric": "acc_norm",
248
+ "aggregation": "mean",
249
+ "higher_is_better": true
250
+ }
251
+ ],
252
+ "output_type": "multiple_choice",
253
+ "repeats": 1,
254
+ "should_decontaminate": true,
255
+ "doc_to_decontamination_query": "question_stem"
256
+ },
257
+ "piqa": {
258
+ "task": "piqa",
259
+ "group": [
260
+ "multiple_choice"
261
+ ],
262
+ "dataset_path": "piqa",
263
+ "training_split": "train",
264
+ "validation_split": "validation",
265
+ "doc_to_text": "Question: {{goal}}\nAnswer:",
266
+ "doc_to_target": "label",
267
+ "doc_to_choice": "{{[sol1, sol2]}}",
268
+ "description": "",
269
+ "target_delimiter": " ",
270
+ "fewshot_delimiter": "\n\n",
271
+ "num_fewshot": 5,
272
+ "metric_list": [
273
+ {
274
+ "metric": "acc",
275
+ "aggregation": "mean",
276
+ "higher_is_better": true
277
+ },
278
+ {
279
+ "metric": "acc_norm",
280
+ "aggregation": "mean",
281
+ "higher_is_better": true
282
+ }
283
+ ],
284
+ "output_type": "multiple_choice",
285
+ "repeats": 1,
286
+ "should_decontaminate": true,
287
+ "doc_to_decontamination_query": "goal"
288
+ },
289
+ "sciq": {
290
+ "task": "sciq",
291
+ "group": [
292
+ "multiple_choice"
293
+ ],
294
+ "dataset_path": "sciq",
295
+ "training_split": "train",
296
+ "validation_split": "validation",
297
+ "test_split": "test",
298
+ "doc_to_text": "{{support.lstrip()}}\nQuestion: {{question}}\nAnswer:",
299
+ "doc_to_target": 3,
300
+ "doc_to_choice": "{{[distractor1, distractor2, distractor3, correct_answer]}}",
301
+ "description": "",
302
+ "target_delimiter": " ",
303
+ "fewshot_delimiter": "\n\n",
304
+ "num_fewshot": 5,
305
+ "metric_list": [
306
+ {
307
+ "metric": "acc",
308
+ "aggregation": "mean",
309
+ "higher_is_better": true
310
+ },
311
+ {
312
+ "metric": "acc_norm",
313
+ "aggregation": "mean",
314
+ "higher_is_better": true
315
+ }
316
+ ],
317
+ "output_type": "multiple_choice",
318
+ "repeats": 1,
319
+ "should_decontaminate": true,
320
+ "doc_to_decontamination_query": "{{support}} {{question}}"
321
+ },
322
+ "wikitext": {
323
+ "task": "wikitext",
324
+ "group": [
325
+ "perplexity",
326
+ "loglikelihood_rolling"
327
+ ],
328
+ "dataset_path": "EleutherAI/wikitext_document_level",
329
+ "dataset_name": "wikitext-2-raw-v1",
330
+ "training_split": "train",
331
+ "validation_split": "validation",
332
+ "test_split": "test",
333
+ "doc_to_text": "",
334
+ "doc_to_target": "<function wikitext_detokenizer at 0x7f80942fd120>",
335
+ "description": "",
336
+ "target_delimiter": " ",
337
+ "fewshot_delimiter": "\n\n",
338
+ "num_fewshot": 5,
339
+ "metric_list": [
340
+ {
341
+ "metric": "word_perplexity"
342
+ },
343
+ {
344
+ "metric": "byte_perplexity"
345
+ },
346
+ {
347
+ "metric": "bits_per_byte"
348
+ }
349
+ ],
350
+ "output_type": "loglikelihood_rolling",
351
+ "repeats": 1,
352
+ "should_decontaminate": true,
353
+ "doc_to_decontamination_query": "{{page}}"
354
+ },
355
+ "winogrande": {
356
+ "task": "winogrande",
357
+ "dataset_path": "winogrande",
358
+ "dataset_name": "winogrande_xl",
359
+ "training_split": "train",
360
+ "validation_split": "validation",
361
+ "doc_to_text": "<function doc_to_text at 0x7f80942fd360>",
362
+ "doc_to_target": "<function doc_to_target at 0x7f80942fd6c0>",
363
+ "doc_to_choice": "<function doc_to_choice at 0x7f80942fda20>",
364
+ "description": "",
365
+ "target_delimiter": " ",
366
+ "fewshot_delimiter": "\n\n",
367
+ "num_fewshot": 5,
368
+ "metric_list": [
369
+ {
370
+ "metric": "acc",
371
+ "aggregation": "mean",
372
+ "higher_is_better": true
373
+ }
374
+ ],
375
+ "output_type": "multiple_choice",
376
+ "repeats": 1,
377
+ "should_decontaminate": true,
378
+ "doc_to_decontamination_query": "sentence"
379
+ }
380
+ },
381
+ "versions": {
382
+ "arc_challenge": "Yaml",
383
+ "arc_easy": "Yaml",
384
+ "boolq": "Yaml",
385
+ "hellaswag": "Yaml",
386
+ "lambada_openai": "Yaml",
387
+ "openbookqa": "Yaml",
388
+ "piqa": "Yaml",
389
+ "sciq": "Yaml",
390
+ "wikitext": "Yaml",
391
+ "winogrande": "Yaml"
392
+ },
393
+ "config": {
394
+ "model": "hf",
395
+ "model_args": "pretrained=EleutherAI/pythia-2.8b",
396
+ "batch_size": "8",
397
+ "batch_sizes": [],
398
+ "device": null,
399
+ "use_cache": null,
400
+ "limit": null,
401
+ "bootstrap_iters": 100000
402
+ },
403
+ "git_hash": "d1a44c8"
404
+ }
dpo-2.8b-eval-files/dpo-pythia-2.8b-0shot-shelloutput.txt ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bootstrapping for stddev: perplexity
2
+ hf (pretrained=lomahony/eleuther-pythia2.8b-hh-dpo), limit: None, num_fewshot: 0, batch_size: 16
3
+ | Task |Version|Filter| Metric | Value | |Stderr|
4
+ |--------------|-------|------|---------------|------:|---|-----:|
5
+ |arc_challenge |Yaml |none |acc | 0.3302|± |0.0137|
6
+ | | |none |acc_norm | 0.3490|± |0.0139|
7
+ |arc_easy |Yaml |none |acc | 0.6625|± |0.0097|
8
+ | | |none |acc_norm | 0.5918|± |0.0101|
9
+ |boolq |Yaml |none |acc | 0.6248|± |0.0085|
10
+ |hellaswag |Yaml |none |acc | 0.4677|± |0.0050|
11
+ | | |none |acc_norm | 0.6072|± |0.0049|
12
+ |lambada_openai|Yaml |none |perplexity | 4.4821|± |0.1220|
13
+ | | |none |acc | 0.6350|± |0.0067|
14
+ |openbookqa |Yaml |none |acc | 0.2640|± |0.0197|
15
+ | | |none |acc_norm | 0.3960|± |0.0219|
16
+ |piqa |Yaml |none |acc | 0.7535|± |0.0101|
17
+ | | |none |acc_norm | 0.7454|± |0.0102|
18
+ |sciq |Yaml |none |acc | 0.8630|± |0.0109|
19
+ | | |none |acc_norm | 0.8030|± |0.0126|
20
+ |wikitext |Yaml |none |word_perplexity|21.9279| | |
21
+ | | |none |byte_perplexity| 1.6637| | |
22
+ | | |none |bits_per_byte | 0.7344| | |
23
+ |winogrande |Yaml |none |acc | 0.5967|± |0.0138|
24
+
dpo-2.8b-eval-files/dpo-pythia-2.8b-0shot/results.json ADDED
@@ -0,0 +1,404 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.3302047781569966,
5
+ "acc_stderr,none": 0.013743085603760427,
6
+ "acc_norm,none": 0.34897610921501704,
7
+ "acc_norm_stderr,none": 0.013928933461382497
8
+ },
9
+ "arc_easy": {
10
+ "acc,none": 0.6624579124579124,
11
+ "acc_stderr,none": 0.009703117820790301,
12
+ "acc_norm,none": 0.5917508417508418,
13
+ "acc_norm_stderr,none": 0.010085566195791252
14
+ },
15
+ "boolq": {
16
+ "acc,none": 0.6247706422018349,
17
+ "acc_stderr,none": 0.008468397820914273
18
+ },
19
+ "hellaswag": {
20
+ "acc,none": 0.46773551085441145,
21
+ "acc_stderr,none": 0.00497938187671261,
22
+ "acc_norm,none": 0.6072495518820952,
23
+ "acc_norm_stderr,none": 0.004873640184773437
24
+ },
25
+ "lambada_openai": {
26
+ "perplexity,none": 4.4821107933393245,
27
+ "perplexity_stderr,none": 0.12203127098816188,
28
+ "acc,none": 0.6349699204346982,
29
+ "acc_stderr,none": 0.006707380989588293
30
+ },
31
+ "openbookqa": {
32
+ "acc,none": 0.264,
33
+ "acc_stderr,none": 0.019732885585922087,
34
+ "acc_norm,none": 0.396,
35
+ "acc_norm_stderr,none": 0.02189352994166582
36
+ },
37
+ "piqa": {
38
+ "acc,none": 0.7535364526659413,
39
+ "acc_stderr,none": 0.01005481078967182,
40
+ "acc_norm,none": 0.7453754080522307,
41
+ "acc_norm_stderr,none": 0.01016443223706048
42
+ },
43
+ "sciq": {
44
+ "acc,none": 0.863,
45
+ "acc_stderr,none": 0.010878848714333304,
46
+ "acc_norm,none": 0.803,
47
+ "acc_norm_stderr,none": 0.012583693787968144
48
+ },
49
+ "wikitext": {
50
+ "word_perplexity,none": 21.92791944465979,
51
+ "byte_perplexity,none": 1.6636811907626483,
52
+ "bits_per_byte,none": 0.734378998038975
53
+ },
54
+ "winogrande": {
55
+ "acc,none": 0.5966850828729282,
56
+ "acc_stderr,none": 0.013787257285896238
57
+ }
58
+ },
59
+ "configs": {
60
+ "arc_challenge": {
61
+ "task": "arc_challenge",
62
+ "group": [
63
+ "ai2_arc",
64
+ "multiple_choice"
65
+ ],
66
+ "dataset_path": "ai2_arc",
67
+ "dataset_name": "ARC-Challenge",
68
+ "training_split": "train",
69
+ "validation_split": "validation",
70
+ "test_split": "test",
71
+ "doc_to_text": "Question: {{question}}\nAnswer:",
72
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
73
+ "doc_to_choice": "{{choices.text}}",
74
+ "description": "",
75
+ "target_delimiter": " ",
76
+ "fewshot_delimiter": "\n\n",
77
+ "num_fewshot": 0,
78
+ "metric_list": [
79
+ {
80
+ "metric": "acc",
81
+ "aggregation": "mean",
82
+ "higher_is_better": true
83
+ },
84
+ {
85
+ "metric": "acc_norm",
86
+ "aggregation": "mean",
87
+ "higher_is_better": true
88
+ }
89
+ ],
90
+ "output_type": "multiple_choice",
91
+ "repeats": 1,
92
+ "should_decontaminate": true,
93
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
94
+ },
95
+ "arc_easy": {
96
+ "task": "arc_easy",
97
+ "group": [
98
+ "ai2_arc",
99
+ "multiple_choice"
100
+ ],
101
+ "dataset_path": "ai2_arc",
102
+ "dataset_name": "ARC-Easy",
103
+ "training_split": "train",
104
+ "validation_split": "validation",
105
+ "test_split": "test",
106
+ "doc_to_text": "Question: {{question}}\nAnswer:",
107
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
108
+ "doc_to_choice": "{{choices.text}}",
109
+ "description": "",
110
+ "target_delimiter": " ",
111
+ "fewshot_delimiter": "\n\n",
112
+ "num_fewshot": 0,
113
+ "metric_list": [
114
+ {
115
+ "metric": "acc",
116
+ "aggregation": "mean",
117
+ "higher_is_better": true
118
+ },
119
+ {
120
+ "metric": "acc_norm",
121
+ "aggregation": "mean",
122
+ "higher_is_better": true
123
+ }
124
+ ],
125
+ "output_type": "multiple_choice",
126
+ "repeats": 1,
127
+ "should_decontaminate": true,
128
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
129
+ },
130
+ "boolq": {
131
+ "task": "boolq",
132
+ "group": [
133
+ "super-glue-lm-eval-v1"
134
+ ],
135
+ "dataset_path": "super_glue",
136
+ "dataset_name": "boolq",
137
+ "training_split": "train",
138
+ "validation_split": "validation",
139
+ "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:",
140
+ "doc_to_target": "label",
141
+ "doc_to_choice": [
142
+ "no",
143
+ "yes"
144
+ ],
145
+ "description": "",
146
+ "target_delimiter": " ",
147
+ "fewshot_delimiter": "\n\n",
148
+ "num_fewshot": 0,
149
+ "metric_list": [
150
+ {
151
+ "metric": "acc"
152
+ }
153
+ ],
154
+ "output_type": "multiple_choice",
155
+ "repeats": 1,
156
+ "should_decontaminate": true,
157
+ "doc_to_decontamination_query": "passage"
158
+ },
159
+ "hellaswag": {
160
+ "task": "hellaswag",
161
+ "group": [
162
+ "multiple_choice"
163
+ ],
164
+ "dataset_path": "hellaswag",
165
+ "training_split": "train",
166
+ "validation_split": "validation",
167
+ "doc_to_text": "{% set text = activity_label ~ ': ' ~ ctx_a ~ ' ' ~ ctx_b.capitalize() %}{{text|trim|replace(' [title]', '. ')|regex_replace('\\[.*?\\]', '')|replace(' ', ' ')}}",
168
+ "doc_to_target": "{{label}}",
169
+ "doc_to_choice": "{{endings|map('trim')|map('replace', ' [title]', '. ')|map('regex_replace', '\\[.*?\\]', '')|map('replace', ' ', ' ')|list}}",
170
+ "description": "",
171
+ "target_delimiter": " ",
172
+ "fewshot_delimiter": "\n\n",
173
+ "num_fewshot": 0,
174
+ "metric_list": [
175
+ {
176
+ "metric": "acc",
177
+ "aggregation": "mean",
178
+ "higher_is_better": true
179
+ },
180
+ {
181
+ "metric": "acc_norm",
182
+ "aggregation": "mean",
183
+ "higher_is_better": true
184
+ }
185
+ ],
186
+ "output_type": "multiple_choice",
187
+ "repeats": 1,
188
+ "should_decontaminate": false
189
+ },
190
+ "lambada_openai": {
191
+ "task": "lambada_openai",
192
+ "group": [
193
+ "lambada",
194
+ "loglikelihood",
195
+ "perplexity"
196
+ ],
197
+ "dataset_path": "EleutherAI/lambada_openai",
198
+ "dataset_name": "default",
199
+ "test_split": "test",
200
+ "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
201
+ "doc_to_target": "{{' '+text.split(' ')[-1]}}",
202
+ "description": "",
203
+ "target_delimiter": " ",
204
+ "fewshot_delimiter": "\n\n",
205
+ "num_fewshot": 0,
206
+ "metric_list": [
207
+ {
208
+ "metric": "perplexity",
209
+ "aggregation": "perplexity",
210
+ "higher_is_better": false
211
+ },
212
+ {
213
+ "metric": "acc",
214
+ "aggregation": "mean",
215
+ "higher_is_better": true
216
+ }
217
+ ],
218
+ "output_type": "loglikelihood",
219
+ "repeats": 1,
220
+ "should_decontaminate": true,
221
+ "doc_to_decontamination_query": "{{text}}"
222
+ },
223
+ "openbookqa": {
224
+ "task": "openbookqa",
225
+ "group": [
226
+ "multiple_choice"
227
+ ],
228
+ "dataset_path": "openbookqa",
229
+ "dataset_name": "main",
230
+ "training_split": "train",
231
+ "validation_split": "validation",
232
+ "test_split": "test",
233
+ "doc_to_text": "question_stem",
234
+ "doc_to_target": "{{choices.label.index(answerKey.lstrip())}}",
235
+ "doc_to_choice": "{{choices.text}}",
236
+ "description": "",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "num_fewshot": 0,
240
+ "metric_list": [
241
+ {
242
+ "metric": "acc",
243
+ "aggregation": "mean",
244
+ "higher_is_better": true
245
+ },
246
+ {
247
+ "metric": "acc_norm",
248
+ "aggregation": "mean",
249
+ "higher_is_better": true
250
+ }
251
+ ],
252
+ "output_type": "multiple_choice",
253
+ "repeats": 1,
254
+ "should_decontaminate": true,
255
+ "doc_to_decontamination_query": "question_stem"
256
+ },
257
+ "piqa": {
258
+ "task": "piqa",
259
+ "group": [
260
+ "multiple_choice"
261
+ ],
262
+ "dataset_path": "piqa",
263
+ "training_split": "train",
264
+ "validation_split": "validation",
265
+ "doc_to_text": "Question: {{goal}}\nAnswer:",
266
+ "doc_to_target": "label",
267
+ "doc_to_choice": "{{[sol1, sol2]}}",
268
+ "description": "",
269
+ "target_delimiter": " ",
270
+ "fewshot_delimiter": "\n\n",
271
+ "num_fewshot": 0,
272
+ "metric_list": [
273
+ {
274
+ "metric": "acc",
275
+ "aggregation": "mean",
276
+ "higher_is_better": true
277
+ },
278
+ {
279
+ "metric": "acc_norm",
280
+ "aggregation": "mean",
281
+ "higher_is_better": true
282
+ }
283
+ ],
284
+ "output_type": "multiple_choice",
285
+ "repeats": 1,
286
+ "should_decontaminate": true,
287
+ "doc_to_decontamination_query": "goal"
288
+ },
289
+ "sciq": {
290
+ "task": "sciq",
291
+ "group": [
292
+ "multiple_choice"
293
+ ],
294
+ "dataset_path": "sciq",
295
+ "training_split": "train",
296
+ "validation_split": "validation",
297
+ "test_split": "test",
298
+ "doc_to_text": "{{support.lstrip()}}\nQuestion: {{question}}\nAnswer:",
299
+ "doc_to_target": 3,
300
+ "doc_to_choice": "{{[distractor1, distractor2, distractor3, correct_answer]}}",
301
+ "description": "",
302
+ "target_delimiter": " ",
303
+ "fewshot_delimiter": "\n\n",
304
+ "num_fewshot": 0,
305
+ "metric_list": [
306
+ {
307
+ "metric": "acc",
308
+ "aggregation": "mean",
309
+ "higher_is_better": true
310
+ },
311
+ {
312
+ "metric": "acc_norm",
313
+ "aggregation": "mean",
314
+ "higher_is_better": true
315
+ }
316
+ ],
317
+ "output_type": "multiple_choice",
318
+ "repeats": 1,
319
+ "should_decontaminate": true,
320
+ "doc_to_decontamination_query": "{{support}} {{question}}"
321
+ },
322
+ "wikitext": {
323
+ "task": "wikitext",
324
+ "group": [
325
+ "perplexity",
326
+ "loglikelihood_rolling"
327
+ ],
328
+ "dataset_path": "EleutherAI/wikitext_document_level",
329
+ "dataset_name": "wikitext-2-raw-v1",
330
+ "training_split": "train",
331
+ "validation_split": "validation",
332
+ "test_split": "test",
333
+ "doc_to_text": "",
334
+ "doc_to_target": "<function wikitext_detokenizer at 0x7feb0f9c1120>",
335
+ "description": "",
336
+ "target_delimiter": " ",
337
+ "fewshot_delimiter": "\n\n",
338
+ "num_fewshot": 0,
339
+ "metric_list": [
340
+ {
341
+ "metric": "word_perplexity"
342
+ },
343
+ {
344
+ "metric": "byte_perplexity"
345
+ },
346
+ {
347
+ "metric": "bits_per_byte"
348
+ }
349
+ ],
350
+ "output_type": "loglikelihood_rolling",
351
+ "repeats": 1,
352
+ "should_decontaminate": true,
353
+ "doc_to_decontamination_query": "{{page}}"
354
+ },
355
+ "winogrande": {
356
+ "task": "winogrande",
357
+ "dataset_path": "winogrande",
358
+ "dataset_name": "winogrande_xl",
359
+ "training_split": "train",
360
+ "validation_split": "validation",
361
+ "doc_to_text": "<function doc_to_text at 0x7feb0f9c1360>",
362
+ "doc_to_target": "<function doc_to_target at 0x7feb0f9c16c0>",
363
+ "doc_to_choice": "<function doc_to_choice at 0x7feb0f9c1a20>",
364
+ "description": "",
365
+ "target_delimiter": " ",
366
+ "fewshot_delimiter": "\n\n",
367
+ "num_fewshot": 0,
368
+ "metric_list": [
369
+ {
370
+ "metric": "acc",
371
+ "aggregation": "mean",
372
+ "higher_is_better": true
373
+ }
374
+ ],
375
+ "output_type": "multiple_choice",
376
+ "repeats": 1,
377
+ "should_decontaminate": true,
378
+ "doc_to_decontamination_query": "sentence"
379
+ }
380
+ },
381
+ "versions": {
382
+ "arc_challenge": "Yaml",
383
+ "arc_easy": "Yaml",
384
+ "boolq": "Yaml",
385
+ "hellaswag": "Yaml",
386
+ "lambada_openai": "Yaml",
387
+ "openbookqa": "Yaml",
388
+ "piqa": "Yaml",
389
+ "sciq": "Yaml",
390
+ "wikitext": "Yaml",
391
+ "winogrande": "Yaml"
392
+ },
393
+ "config": {
394
+ "model": "hf",
395
+ "model_args": "pretrained=lomahony/eleuther-pythia2.8b-hh-dpo",
396
+ "batch_size": "16",
397
+ "batch_sizes": [],
398
+ "device": null,
399
+ "use_cache": null,
400
+ "limit": null,
401
+ "bootstrap_iters": 100000
402
+ },
403
+ "git_hash": "d1a44c8"
404
+ }
dpo-2.8b-eval-files/dpo-pythia-2.8b-5shot-shelloutput.txt ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bootstrapping for stddev: perplexity
2
+ hf (pretrained=lomahony/eleuther-pythia2.8b-hh-dpo), limit: None, num_fewshot: 5, batch_size: 8
3
+ | Task |Version|Filter| Metric | Value | |Stderr|
4
+ |--------------|-------|------|---------------|------:|---|-----:|
5
+ |arc_challenge |Yaml |none |acc | 0.3353|± |0.0138|
6
+ | | |none |acc_norm | 0.3788|± |0.0142|
7
+ |arc_easy |Yaml |none |acc | 0.6890|± |0.0095|
8
+ | | |none |acc_norm | 0.6936|± |0.0095|
9
+ |boolq |Yaml |none |acc | 0.6495|± |0.0083|
10
+ |hellaswag |Yaml |none |acc | 0.4680|± |0.0050|
11
+ | | |none |acc_norm | 0.6124|± |0.0049|
12
+ |lambada_openai|Yaml |none |perplexity | 5.9966|± |0.1685|
13
+ | | |none |acc | 0.5789|± |0.0069|
14
+ |openbookqa |Yaml |none |acc | 0.2720|± |0.0199|
15
+ | | |none |acc_norm | 0.3740|± |0.0217|
16
+ |piqa |Yaml |none |acc | 0.7535|± |0.0101|
17
+ | | |none |acc_norm | 0.7573|± |0.0100|
18
+ |sciq |Yaml |none |acc | 0.9340|± |0.0079|
19
+ | | |none |acc_norm | 0.9190|± |0.0086|
20
+ |wikitext |Yaml |none |word_perplexity|21.9279| | |
21
+ | | |none |byte_perplexity| 1.6637| | |
22
+ | | |none |bits_per_byte | 0.7344| | |
23
+ |winogrande |Yaml |none |acc | 0.5967|± |0.0138|
24
+
dpo-2.8b-eval-files/dpo-pythia-2.8b-5shot/results.json ADDED
@@ -0,0 +1,404 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.33532423208191126,
5
+ "acc_stderr,none": 0.01379618294778556,
6
+ "acc_norm,none": 0.378839590443686,
7
+ "acc_norm_stderr,none": 0.014175915490000324
8
+ },
9
+ "arc_easy": {
10
+ "acc,none": 0.688973063973064,
11
+ "acc_stderr,none": 0.009498790639757615,
12
+ "acc_norm,none": 0.6936026936026936,
13
+ "acc_norm_stderr,none": 0.00945945357339833
14
+ },
15
+ "boolq": {
16
+ "acc,none": 0.6495412844036698,
17
+ "acc_stderr,none": 0.00834476963472486
18
+ },
19
+ "hellaswag": {
20
+ "acc,none": 0.468034256124278,
21
+ "acc_stderr,none": 0.004979573765575858,
22
+ "acc_norm,none": 0.612427803226449,
23
+ "acc_norm_stderr,none": 0.004862003566798546
24
+ },
25
+ "lambada_openai": {
26
+ "perplexity,none": 5.99663804430596,
27
+ "perplexity_stderr,none": 0.1684912908221153,
28
+ "acc,none": 0.5788860857752766,
29
+ "acc_stderr,none": 0.006878732547908385
30
+ },
31
+ "openbookqa": {
32
+ "acc,none": 0.272,
33
+ "acc_stderr,none": 0.019920483209566065,
34
+ "acc_norm,none": 0.374,
35
+ "acc_norm_stderr,none": 0.021660710347204484
36
+ },
37
+ "piqa": {
38
+ "acc,none": 0.7535364526659413,
39
+ "acc_stderr,none": 0.010054810789671818,
40
+ "acc_norm,none": 0.7573449401523396,
41
+ "acc_norm_stderr,none": 0.010002002569708696
42
+ },
43
+ "sciq": {
44
+ "acc,none": 0.934,
45
+ "acc_stderr,none": 0.007855297938697587,
46
+ "acc_norm,none": 0.919,
47
+ "acc_norm_stderr,none": 0.00863212103213995
48
+ },
49
+ "wikitext": {
50
+ "word_perplexity,none": 21.92791944465979,
51
+ "byte_perplexity,none": 1.6636811907626483,
52
+ "bits_per_byte,none": 0.734378998038975
53
+ },
54
+ "winogrande": {
55
+ "acc,none": 0.5966850828729282,
56
+ "acc_stderr,none": 0.013787257285896238
57
+ }
58
+ },
59
+ "configs": {
60
+ "arc_challenge": {
61
+ "task": "arc_challenge",
62
+ "group": [
63
+ "ai2_arc",
64
+ "multiple_choice"
65
+ ],
66
+ "dataset_path": "ai2_arc",
67
+ "dataset_name": "ARC-Challenge",
68
+ "training_split": "train",
69
+ "validation_split": "validation",
70
+ "test_split": "test",
71
+ "doc_to_text": "Question: {{question}}\nAnswer:",
72
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
73
+ "doc_to_choice": "{{choices.text}}",
74
+ "description": "",
75
+ "target_delimiter": " ",
76
+ "fewshot_delimiter": "\n\n",
77
+ "num_fewshot": 5,
78
+ "metric_list": [
79
+ {
80
+ "metric": "acc",
81
+ "aggregation": "mean",
82
+ "higher_is_better": true
83
+ },
84
+ {
85
+ "metric": "acc_norm",
86
+ "aggregation": "mean",
87
+ "higher_is_better": true
88
+ }
89
+ ],
90
+ "output_type": "multiple_choice",
91
+ "repeats": 1,
92
+ "should_decontaminate": true,
93
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
94
+ },
95
+ "arc_easy": {
96
+ "task": "arc_easy",
97
+ "group": [
98
+ "ai2_arc",
99
+ "multiple_choice"
100
+ ],
101
+ "dataset_path": "ai2_arc",
102
+ "dataset_name": "ARC-Easy",
103
+ "training_split": "train",
104
+ "validation_split": "validation",
105
+ "test_split": "test",
106
+ "doc_to_text": "Question: {{question}}\nAnswer:",
107
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
108
+ "doc_to_choice": "{{choices.text}}",
109
+ "description": "",
110
+ "target_delimiter": " ",
111
+ "fewshot_delimiter": "\n\n",
112
+ "num_fewshot": 5,
113
+ "metric_list": [
114
+ {
115
+ "metric": "acc",
116
+ "aggregation": "mean",
117
+ "higher_is_better": true
118
+ },
119
+ {
120
+ "metric": "acc_norm",
121
+ "aggregation": "mean",
122
+ "higher_is_better": true
123
+ }
124
+ ],
125
+ "output_type": "multiple_choice",
126
+ "repeats": 1,
127
+ "should_decontaminate": true,
128
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
129
+ },
130
+ "boolq": {
131
+ "task": "boolq",
132
+ "group": [
133
+ "super-glue-lm-eval-v1"
134
+ ],
135
+ "dataset_path": "super_glue",
136
+ "dataset_name": "boolq",
137
+ "training_split": "train",
138
+ "validation_split": "validation",
139
+ "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:",
140
+ "doc_to_target": "label",
141
+ "doc_to_choice": [
142
+ "no",
143
+ "yes"
144
+ ],
145
+ "description": "",
146
+ "target_delimiter": " ",
147
+ "fewshot_delimiter": "\n\n",
148
+ "num_fewshot": 5,
149
+ "metric_list": [
150
+ {
151
+ "metric": "acc"
152
+ }
153
+ ],
154
+ "output_type": "multiple_choice",
155
+ "repeats": 1,
156
+ "should_decontaminate": true,
157
+ "doc_to_decontamination_query": "passage"
158
+ },
159
+ "hellaswag": {
160
+ "task": "hellaswag",
161
+ "group": [
162
+ "multiple_choice"
163
+ ],
164
+ "dataset_path": "hellaswag",
165
+ "training_split": "train",
166
+ "validation_split": "validation",
167
+ "doc_to_text": "{% set text = activity_label ~ ': ' ~ ctx_a ~ ' ' ~ ctx_b.capitalize() %}{{text|trim|replace(' [title]', '. ')|regex_replace('\\[.*?\\]', '')|replace(' ', ' ')}}",
168
+ "doc_to_target": "{{label}}",
169
+ "doc_to_choice": "{{endings|map('trim')|map('replace', ' [title]', '. ')|map('regex_replace', '\\[.*?\\]', '')|map('replace', ' ', ' ')|list}}",
170
+ "description": "",
171
+ "target_delimiter": " ",
172
+ "fewshot_delimiter": "\n\n",
173
+ "num_fewshot": 5,
174
+ "metric_list": [
175
+ {
176
+ "metric": "acc",
177
+ "aggregation": "mean",
178
+ "higher_is_better": true
179
+ },
180
+ {
181
+ "metric": "acc_norm",
182
+ "aggregation": "mean",
183
+ "higher_is_better": true
184
+ }
185
+ ],
186
+ "output_type": "multiple_choice",
187
+ "repeats": 1,
188
+ "should_decontaminate": false
189
+ },
190
+ "lambada_openai": {
191
+ "task": "lambada_openai",
192
+ "group": [
193
+ "lambada",
194
+ "loglikelihood",
195
+ "perplexity"
196
+ ],
197
+ "dataset_path": "EleutherAI/lambada_openai",
198
+ "dataset_name": "default",
199
+ "test_split": "test",
200
+ "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
201
+ "doc_to_target": "{{' '+text.split(' ')[-1]}}",
202
+ "description": "",
203
+ "target_delimiter": " ",
204
+ "fewshot_delimiter": "\n\n",
205
+ "num_fewshot": 5,
206
+ "metric_list": [
207
+ {
208
+ "metric": "perplexity",
209
+ "aggregation": "perplexity",
210
+ "higher_is_better": false
211
+ },
212
+ {
213
+ "metric": "acc",
214
+ "aggregation": "mean",
215
+ "higher_is_better": true
216
+ }
217
+ ],
218
+ "output_type": "loglikelihood",
219
+ "repeats": 1,
220
+ "should_decontaminate": true,
221
+ "doc_to_decontamination_query": "{{text}}"
222
+ },
223
+ "openbookqa": {
224
+ "task": "openbookqa",
225
+ "group": [
226
+ "multiple_choice"
227
+ ],
228
+ "dataset_path": "openbookqa",
229
+ "dataset_name": "main",
230
+ "training_split": "train",
231
+ "validation_split": "validation",
232
+ "test_split": "test",
233
+ "doc_to_text": "question_stem",
234
+ "doc_to_target": "{{choices.label.index(answerKey.lstrip())}}",
235
+ "doc_to_choice": "{{choices.text}}",
236
+ "description": "",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "num_fewshot": 5,
240
+ "metric_list": [
241
+ {
242
+ "metric": "acc",
243
+ "aggregation": "mean",
244
+ "higher_is_better": true
245
+ },
246
+ {
247
+ "metric": "acc_norm",
248
+ "aggregation": "mean",
249
+ "higher_is_better": true
250
+ }
251
+ ],
252
+ "output_type": "multiple_choice",
253
+ "repeats": 1,
254
+ "should_decontaminate": true,
255
+ "doc_to_decontamination_query": "question_stem"
256
+ },
257
+ "piqa": {
258
+ "task": "piqa",
259
+ "group": [
260
+ "multiple_choice"
261
+ ],
262
+ "dataset_path": "piqa",
263
+ "training_split": "train",
264
+ "validation_split": "validation",
265
+ "doc_to_text": "Question: {{goal}}\nAnswer:",
266
+ "doc_to_target": "label",
267
+ "doc_to_choice": "{{[sol1, sol2]}}",
268
+ "description": "",
269
+ "target_delimiter": " ",
270
+ "fewshot_delimiter": "\n\n",
271
+ "num_fewshot": 5,
272
+ "metric_list": [
273
+ {
274
+ "metric": "acc",
275
+ "aggregation": "mean",
276
+ "higher_is_better": true
277
+ },
278
+ {
279
+ "metric": "acc_norm",
280
+ "aggregation": "mean",
281
+ "higher_is_better": true
282
+ }
283
+ ],
284
+ "output_type": "multiple_choice",
285
+ "repeats": 1,
286
+ "should_decontaminate": true,
287
+ "doc_to_decontamination_query": "goal"
288
+ },
289
+ "sciq": {
290
+ "task": "sciq",
291
+ "group": [
292
+ "multiple_choice"
293
+ ],
294
+ "dataset_path": "sciq",
295
+ "training_split": "train",
296
+ "validation_split": "validation",
297
+ "test_split": "test",
298
+ "doc_to_text": "{{support.lstrip()}}\nQuestion: {{question}}\nAnswer:",
299
+ "doc_to_target": 3,
300
+ "doc_to_choice": "{{[distractor1, distractor2, distractor3, correct_answer]}}",
301
+ "description": "",
302
+ "target_delimiter": " ",
303
+ "fewshot_delimiter": "\n\n",
304
+ "num_fewshot": 5,
305
+ "metric_list": [
306
+ {
307
+ "metric": "acc",
308
+ "aggregation": "mean",
309
+ "higher_is_better": true
310
+ },
311
+ {
312
+ "metric": "acc_norm",
313
+ "aggregation": "mean",
314
+ "higher_is_better": true
315
+ }
316
+ ],
317
+ "output_type": "multiple_choice",
318
+ "repeats": 1,
319
+ "should_decontaminate": true,
320
+ "doc_to_decontamination_query": "{{support}} {{question}}"
321
+ },
322
+ "wikitext": {
323
+ "task": "wikitext",
324
+ "group": [
325
+ "perplexity",
326
+ "loglikelihood_rolling"
327
+ ],
328
+ "dataset_path": "EleutherAI/wikitext_document_level",
329
+ "dataset_name": "wikitext-2-raw-v1",
330
+ "training_split": "train",
331
+ "validation_split": "validation",
332
+ "test_split": "test",
333
+ "doc_to_text": "",
334
+ "doc_to_target": "<function wikitext_detokenizer at 0x7f452e1a9120>",
335
+ "description": "",
336
+ "target_delimiter": " ",
337
+ "fewshot_delimiter": "\n\n",
338
+ "num_fewshot": 5,
339
+ "metric_list": [
340
+ {
341
+ "metric": "word_perplexity"
342
+ },
343
+ {
344
+ "metric": "byte_perplexity"
345
+ },
346
+ {
347
+ "metric": "bits_per_byte"
348
+ }
349
+ ],
350
+ "output_type": "loglikelihood_rolling",
351
+ "repeats": 1,
352
+ "should_decontaminate": true,
353
+ "doc_to_decontamination_query": "{{page}}"
354
+ },
355
+ "winogrande": {
356
+ "task": "winogrande",
357
+ "dataset_path": "winogrande",
358
+ "dataset_name": "winogrande_xl",
359
+ "training_split": "train",
360
+ "validation_split": "validation",
361
+ "doc_to_text": "<function doc_to_text at 0x7f452e1a9360>",
362
+ "doc_to_target": "<function doc_to_target at 0x7f452e1a96c0>",
363
+ "doc_to_choice": "<function doc_to_choice at 0x7f452e1a9a20>",
364
+ "description": "",
365
+ "target_delimiter": " ",
366
+ "fewshot_delimiter": "\n\n",
367
+ "num_fewshot": 5,
368
+ "metric_list": [
369
+ {
370
+ "metric": "acc",
371
+ "aggregation": "mean",
372
+ "higher_is_better": true
373
+ }
374
+ ],
375
+ "output_type": "multiple_choice",
376
+ "repeats": 1,
377
+ "should_decontaminate": true,
378
+ "doc_to_decontamination_query": "sentence"
379
+ }
380
+ },
381
+ "versions": {
382
+ "arc_challenge": "Yaml",
383
+ "arc_easy": "Yaml",
384
+ "boolq": "Yaml",
385
+ "hellaswag": "Yaml",
386
+ "lambada_openai": "Yaml",
387
+ "openbookqa": "Yaml",
388
+ "piqa": "Yaml",
389
+ "sciq": "Yaml",
390
+ "wikitext": "Yaml",
391
+ "winogrande": "Yaml"
392
+ },
393
+ "config": {
394
+ "model": "hf",
395
+ "model_args": "pretrained=lomahony/eleuther-pythia2.8b-hh-dpo",
396
+ "batch_size": "8",
397
+ "batch_sizes": [],
398
+ "device": null,
399
+ "use_cache": null,
400
+ "limit": null,
401
+ "bootstrap_iters": 100000
402
+ },
403
+ "git_hash": "d1a44c8"
404
+ }