lomahony commited on
Commit
11e25ea
1 Parent(s): 9c699a1

Upload 8 files

Browse files
base-12b-eval-files/EleutherAI-pythia-12b-0shot-shelloutput.txt ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bootstrapping for stddev: perplexity
2
+ hf (pretrained=EleutherAI/pythia-12b,parallelize=True), limit: None, num_fewshot: 0, batch_size: 4
3
+ | Task |Version|Filter| Metric | Value | |Stderr|
4
+ |--------------|-------|------|---------------|------:|---|-----:|
5
+ |arc_challenge |Yaml |none |acc | 0.3157|± |0.0136|
6
+ | | |none |acc_norm | 0.3515|± |0.0140|
7
+ |arc_easy |Yaml |none |acc | 0.7033|± |0.0094|
8
+ | | |none |acc_norm | 0.6372|± |0.0099|
9
+ |boolq |Yaml |none |acc | 0.6722|± |0.0082|
10
+ |hellaswag |Yaml |none |acc | 0.5038|± |0.0050|
11
+ | | |none |acc_norm | 0.6728|± |0.0047|
12
+ |lambada_openai|Yaml |none |perplexity | 3.9283|± |0.0838|
13
+ | | |none |acc | 0.7056|± |0.0063|
14
+ |openbookqa |Yaml |none |acc | 0.2640|± |0.0197|
15
+ | | |none |acc_norm | 0.3800|± |0.0217|
16
+ |piqa |Yaml |none |acc | 0.7612|± |0.0099|
17
+ | | |none |acc_norm | 0.7699|± |0.0098|
18
+ |sciq |Yaml |none |acc | 0.9040|± |0.0093|
19
+ | | |none |acc_norm | 0.8500|± |0.0113|
20
+ |wikitext |Yaml |none |word_perplexity|16.1038| | |
21
+ | | |none |byte_perplexity| 1.5811| | |
22
+ | | |none |bits_per_byte | 0.6610| | |
23
+ |winogrande |Yaml |none |acc | 0.6354|± |0.0135|
24
+
base-12b-eval-files/EleutherAI-pythia-12b-0shot/results.json ADDED
@@ -0,0 +1,404 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.31569965870307165,
5
+ "acc_stderr,none": 0.013582571095815295,
6
+ "acc_norm,none": 0.3515358361774744,
7
+ "acc_norm_stderr,none": 0.013952413699600931
8
+ },
9
+ "arc_easy": {
10
+ "acc,none": 0.7032828282828283,
11
+ "acc_stderr,none": 0.009373559492986851,
12
+ "acc_norm,none": 0.6372053872053872,
13
+ "acc_norm_stderr,none": 0.009865936757013938
14
+ },
15
+ "boolq": {
16
+ "acc,none": 0.67217125382263,
17
+ "acc_stderr,none": 0.008210243237673385
18
+ },
19
+ "hellaswag": {
20
+ "acc,none": 0.5037841067516431,
21
+ "acc_stderr,none": 0.004989638507409938,
22
+ "acc_norm,none": 0.6727743477394941,
23
+ "acc_norm_stderr,none": 0.004682414968323615
24
+ },
25
+ "lambada_openai": {
26
+ "perplexity,none": 3.928303151284302,
27
+ "perplexity_stderr,none": 0.08380809965264784,
28
+ "acc,none": 0.7056083834659421,
29
+ "acc_stderr,none": 0.006349750457397497
30
+ },
31
+ "openbookqa": {
32
+ "acc,none": 0.264,
33
+ "acc_stderr,none": 0.019732885585922098,
34
+ "acc_norm,none": 0.38,
35
+ "acc_norm_stderr,none": 0.02172888143870172
36
+ },
37
+ "piqa": {
38
+ "acc,none": 0.7611534276387377,
39
+ "acc_stderr,none": 0.009948120385337496,
40
+ "acc_norm,none": 0.7698585418933623,
41
+ "acc_norm_stderr,none": 0.009820832826839813
42
+ },
43
+ "sciq": {
44
+ "acc,none": 0.904,
45
+ "acc_stderr,none": 0.009320454434783203,
46
+ "acc_norm,none": 0.85,
47
+ "acc_norm_stderr,none": 0.01129723982340931
48
+ },
49
+ "wikitext": {
50
+ "word_perplexity,none": 16.10381134238307,
51
+ "byte_perplexity,none": 1.5811322783219102,
52
+ "bits_per_byte,none": 0.6609580693383097
53
+ },
54
+ "winogrande": {
55
+ "acc,none": 0.6353591160220995,
56
+ "acc_stderr,none": 0.013527746622429842
57
+ }
58
+ },
59
+ "configs": {
60
+ "arc_challenge": {
61
+ "task": "arc_challenge",
62
+ "group": [
63
+ "ai2_arc",
64
+ "multiple_choice"
65
+ ],
66
+ "dataset_path": "ai2_arc",
67
+ "dataset_name": "ARC-Challenge",
68
+ "training_split": "train",
69
+ "validation_split": "validation",
70
+ "test_split": "test",
71
+ "doc_to_text": "Question: {{question}}\nAnswer:",
72
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
73
+ "doc_to_choice": "{{choices.text}}",
74
+ "description": "",
75
+ "target_delimiter": " ",
76
+ "fewshot_delimiter": "\n\n",
77
+ "num_fewshot": 0,
78
+ "metric_list": [
79
+ {
80
+ "metric": "acc",
81
+ "aggregation": "mean",
82
+ "higher_is_better": true
83
+ },
84
+ {
85
+ "metric": "acc_norm",
86
+ "aggregation": "mean",
87
+ "higher_is_better": true
88
+ }
89
+ ],
90
+ "output_type": "multiple_choice",
91
+ "repeats": 1,
92
+ "should_decontaminate": true,
93
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
94
+ },
95
+ "arc_easy": {
96
+ "task": "arc_easy",
97
+ "group": [
98
+ "ai2_arc",
99
+ "multiple_choice"
100
+ ],
101
+ "dataset_path": "ai2_arc",
102
+ "dataset_name": "ARC-Easy",
103
+ "training_split": "train",
104
+ "validation_split": "validation",
105
+ "test_split": "test",
106
+ "doc_to_text": "Question: {{question}}\nAnswer:",
107
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
108
+ "doc_to_choice": "{{choices.text}}",
109
+ "description": "",
110
+ "target_delimiter": " ",
111
+ "fewshot_delimiter": "\n\n",
112
+ "num_fewshot": 0,
113
+ "metric_list": [
114
+ {
115
+ "metric": "acc",
116
+ "aggregation": "mean",
117
+ "higher_is_better": true
118
+ },
119
+ {
120
+ "metric": "acc_norm",
121
+ "aggregation": "mean",
122
+ "higher_is_better": true
123
+ }
124
+ ],
125
+ "output_type": "multiple_choice",
126
+ "repeats": 1,
127
+ "should_decontaminate": true,
128
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
129
+ },
130
+ "boolq": {
131
+ "task": "boolq",
132
+ "group": [
133
+ "super-glue-lm-eval-v1"
134
+ ],
135
+ "dataset_path": "super_glue",
136
+ "dataset_name": "boolq",
137
+ "training_split": "train",
138
+ "validation_split": "validation",
139
+ "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:",
140
+ "doc_to_target": "label",
141
+ "doc_to_choice": [
142
+ "no",
143
+ "yes"
144
+ ],
145
+ "description": "",
146
+ "target_delimiter": " ",
147
+ "fewshot_delimiter": "\n\n",
148
+ "num_fewshot": 0,
149
+ "metric_list": [
150
+ {
151
+ "metric": "acc"
152
+ }
153
+ ],
154
+ "output_type": "multiple_choice",
155
+ "repeats": 1,
156
+ "should_decontaminate": true,
157
+ "doc_to_decontamination_query": "passage"
158
+ },
159
+ "hellaswag": {
160
+ "task": "hellaswag",
161
+ "group": [
162
+ "multiple_choice"
163
+ ],
164
+ "dataset_path": "hellaswag",
165
+ "training_split": "train",
166
+ "validation_split": "validation",
167
+ "doc_to_text": "{% set text = activity_label ~ ': ' ~ ctx_a ~ ' ' ~ ctx_b.capitalize() %}{{text|trim|replace(' [title]', '. ')|regex_replace('\\[.*?\\]', '')|replace(' ', ' ')}}",
168
+ "doc_to_target": "{{label}}",
169
+ "doc_to_choice": "{{endings|map('trim')|map('replace', ' [title]', '. ')|map('regex_replace', '\\[.*?\\]', '')|map('replace', ' ', ' ')|list}}",
170
+ "description": "",
171
+ "target_delimiter": " ",
172
+ "fewshot_delimiter": "\n\n",
173
+ "num_fewshot": 0,
174
+ "metric_list": [
175
+ {
176
+ "metric": "acc",
177
+ "aggregation": "mean",
178
+ "higher_is_better": true
179
+ },
180
+ {
181
+ "metric": "acc_norm",
182
+ "aggregation": "mean",
183
+ "higher_is_better": true
184
+ }
185
+ ],
186
+ "output_type": "multiple_choice",
187
+ "repeats": 1,
188
+ "should_decontaminate": false
189
+ },
190
+ "lambada_openai": {
191
+ "task": "lambada_openai",
192
+ "group": [
193
+ "lambada",
194
+ "loglikelihood",
195
+ "perplexity"
196
+ ],
197
+ "dataset_path": "EleutherAI/lambada_openai",
198
+ "dataset_name": "default",
199
+ "test_split": "test",
200
+ "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
201
+ "doc_to_target": "{{' '+text.split(' ')[-1]}}",
202
+ "description": "",
203
+ "target_delimiter": " ",
204
+ "fewshot_delimiter": "\n\n",
205
+ "num_fewshot": 0,
206
+ "metric_list": [
207
+ {
208
+ "metric": "perplexity",
209
+ "aggregation": "perplexity",
210
+ "higher_is_better": false
211
+ },
212
+ {
213
+ "metric": "acc",
214
+ "aggregation": "mean",
215
+ "higher_is_better": true
216
+ }
217
+ ],
218
+ "output_type": "loglikelihood",
219
+ "repeats": 1,
220
+ "should_decontaminate": true,
221
+ "doc_to_decontamination_query": "{{text}}"
222
+ },
223
+ "openbookqa": {
224
+ "task": "openbookqa",
225
+ "group": [
226
+ "multiple_choice"
227
+ ],
228
+ "dataset_path": "openbookqa",
229
+ "dataset_name": "main",
230
+ "training_split": "train",
231
+ "validation_split": "validation",
232
+ "test_split": "test",
233
+ "doc_to_text": "question_stem",
234
+ "doc_to_target": "{{choices.label.index(answerKey.lstrip())}}",
235
+ "doc_to_choice": "{{choices.text}}",
236
+ "description": "",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "num_fewshot": 0,
240
+ "metric_list": [
241
+ {
242
+ "metric": "acc",
243
+ "aggregation": "mean",
244
+ "higher_is_better": true
245
+ },
246
+ {
247
+ "metric": "acc_norm",
248
+ "aggregation": "mean",
249
+ "higher_is_better": true
250
+ }
251
+ ],
252
+ "output_type": "multiple_choice",
253
+ "repeats": 1,
254
+ "should_decontaminate": true,
255
+ "doc_to_decontamination_query": "question_stem"
256
+ },
257
+ "piqa": {
258
+ "task": "piqa",
259
+ "group": [
260
+ "multiple_choice"
261
+ ],
262
+ "dataset_path": "piqa",
263
+ "training_split": "train",
264
+ "validation_split": "validation",
265
+ "doc_to_text": "Question: {{goal}}\nAnswer:",
266
+ "doc_to_target": "label",
267
+ "doc_to_choice": "{{[sol1, sol2]}}",
268
+ "description": "",
269
+ "target_delimiter": " ",
270
+ "fewshot_delimiter": "\n\n",
271
+ "num_fewshot": 0,
272
+ "metric_list": [
273
+ {
274
+ "metric": "acc",
275
+ "aggregation": "mean",
276
+ "higher_is_better": true
277
+ },
278
+ {
279
+ "metric": "acc_norm",
280
+ "aggregation": "mean",
281
+ "higher_is_better": true
282
+ }
283
+ ],
284
+ "output_type": "multiple_choice",
285
+ "repeats": 1,
286
+ "should_decontaminate": true,
287
+ "doc_to_decontamination_query": "goal"
288
+ },
289
+ "sciq": {
290
+ "task": "sciq",
291
+ "group": [
292
+ "multiple_choice"
293
+ ],
294
+ "dataset_path": "sciq",
295
+ "training_split": "train",
296
+ "validation_split": "validation",
297
+ "test_split": "test",
298
+ "doc_to_text": "{{support.lstrip()}}\nQuestion: {{question}}\nAnswer:",
299
+ "doc_to_target": 3,
300
+ "doc_to_choice": "{{[distractor1, distractor2, distractor3, correct_answer]}}",
301
+ "description": "",
302
+ "target_delimiter": " ",
303
+ "fewshot_delimiter": "\n\n",
304
+ "num_fewshot": 0,
305
+ "metric_list": [
306
+ {
307
+ "metric": "acc",
308
+ "aggregation": "mean",
309
+ "higher_is_better": true
310
+ },
311
+ {
312
+ "metric": "acc_norm",
313
+ "aggregation": "mean",
314
+ "higher_is_better": true
315
+ }
316
+ ],
317
+ "output_type": "multiple_choice",
318
+ "repeats": 1,
319
+ "should_decontaminate": true,
320
+ "doc_to_decontamination_query": "{{support}} {{question}}"
321
+ },
322
+ "wikitext": {
323
+ "task": "wikitext",
324
+ "group": [
325
+ "perplexity",
326
+ "loglikelihood_rolling"
327
+ ],
328
+ "dataset_path": "EleutherAI/wikitext_document_level",
329
+ "dataset_name": "wikitext-2-raw-v1",
330
+ "training_split": "train",
331
+ "validation_split": "validation",
332
+ "test_split": "test",
333
+ "doc_to_text": "",
334
+ "doc_to_target": "<function wikitext_detokenizer at 0x7f1b6d8b9120>",
335
+ "description": "",
336
+ "target_delimiter": " ",
337
+ "fewshot_delimiter": "\n\n",
338
+ "num_fewshot": 0,
339
+ "metric_list": [
340
+ {
341
+ "metric": "word_perplexity"
342
+ },
343
+ {
344
+ "metric": "byte_perplexity"
345
+ },
346
+ {
347
+ "metric": "bits_per_byte"
348
+ }
349
+ ],
350
+ "output_type": "loglikelihood_rolling",
351
+ "repeats": 1,
352
+ "should_decontaminate": true,
353
+ "doc_to_decontamination_query": "{{page}}"
354
+ },
355
+ "winogrande": {
356
+ "task": "winogrande",
357
+ "dataset_path": "winogrande",
358
+ "dataset_name": "winogrande_xl",
359
+ "training_split": "train",
360
+ "validation_split": "validation",
361
+ "doc_to_text": "<function doc_to_text at 0x7f1b6d8b9360>",
362
+ "doc_to_target": "<function doc_to_target at 0x7f1b6d8b96c0>",
363
+ "doc_to_choice": "<function doc_to_choice at 0x7f1b6d8b9a20>",
364
+ "description": "",
365
+ "target_delimiter": " ",
366
+ "fewshot_delimiter": "\n\n",
367
+ "num_fewshot": 0,
368
+ "metric_list": [
369
+ {
370
+ "metric": "acc",
371
+ "aggregation": "mean",
372
+ "higher_is_better": true
373
+ }
374
+ ],
375
+ "output_type": "multiple_choice",
376
+ "repeats": 1,
377
+ "should_decontaminate": true,
378
+ "doc_to_decontamination_query": "sentence"
379
+ }
380
+ },
381
+ "versions": {
382
+ "arc_challenge": "Yaml",
383
+ "arc_easy": "Yaml",
384
+ "boolq": "Yaml",
385
+ "hellaswag": "Yaml",
386
+ "lambada_openai": "Yaml",
387
+ "openbookqa": "Yaml",
388
+ "piqa": "Yaml",
389
+ "sciq": "Yaml",
390
+ "wikitext": "Yaml",
391
+ "winogrande": "Yaml"
392
+ },
393
+ "config": {
394
+ "model": "hf",
395
+ "model_args": "pretrained=EleutherAI/pythia-12b,parallelize=True",
396
+ "batch_size": "4",
397
+ "batch_sizes": [],
398
+ "device": null,
399
+ "use_cache": null,
400
+ "limit": null,
401
+ "bootstrap_iters": 100000
402
+ },
403
+ "git_hash": "d1a44c8"
404
+ }
base-12b-eval-files/EleutherAI-pythia-12b-5shot-shelloutput.txt ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bootstrapping for stddev: perplexity
2
+ hf (pretrained=EleutherAI/pythia-12b,parallelize=True), limit: None, num_fewshot: 5, batch_size: 4
3
+ | Task |Version|Filter| Metric | Value | |Stderr|
4
+ |--------------|-------|------|---------------|------:|---|-----:|
5
+ |arc_challenge |Yaml |none |acc | 0.3686|± |0.0141|
6
+ | | |none |acc_norm | 0.4010|± |0.0143|
7
+ |arc_easy |Yaml |none |acc | 0.7088|± |0.0093|
8
+ | | |none |acc_norm | 0.7146|± |0.0093|
9
+ |boolq |Yaml |none |acc | 0.6795|± |0.0082|
10
+ |hellaswag |Yaml |none |acc | 0.5043|± |0.0050|
11
+ | | |none |acc_norm | 0.6830|± |0.0046|
12
+ |lambada_openai|Yaml |none |perplexity | 4.7835|± |0.1073|
13
+ | | |none |acc | 0.6625|± |0.0066|
14
+ |openbookqa |Yaml |none |acc | 0.2940|± |0.0204|
15
+ | | |none |acc_norm | 0.3980|± |0.0219|
16
+ |piqa |Yaml |none |acc | 0.7699|± |0.0098|
17
+ | | |none |acc_norm | 0.7731|± |0.0098|
18
+ |sciq |Yaml |none |acc | 0.9480|± |0.0070|
19
+ | | |none |acc_norm | 0.9510|± |0.0068|
20
+ |wikitext |Yaml |none |word_perplexity|16.1038| | |
21
+ | | |none |byte_perplexity| 1.5811| | |
22
+ | | |none |bits_per_byte | 0.6610| | |
23
+ |winogrande |Yaml |none |acc | 0.6354|± |0.0135|
24
+
base-12b-eval-files/EleutherAI-pythia-12b-5shot/results.json ADDED
@@ -0,0 +1,404 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.36860068259385664,
5
+ "acc_stderr,none": 0.014097810678042194,
6
+ "acc_norm,none": 0.40102389078498296,
7
+ "acc_norm_stderr,none": 0.014322255790719867
8
+ },
9
+ "arc_easy": {
10
+ "acc,none": 0.7087542087542088,
11
+ "acc_stderr,none": 0.009322788837938856,
12
+ "acc_norm,none": 0.7146464646464646,
13
+ "acc_norm_stderr,none": 0.00926628058499775
14
+ },
15
+ "boolq": {
16
+ "acc,none": 0.6795107033639144,
17
+ "acc_stderr,none": 0.008162016261049398
18
+ },
19
+ "hellaswag": {
20
+ "acc,none": 0.504282015534754,
21
+ "acc_stderr,none": 0.0049895984262495335,
22
+ "acc_norm,none": 0.6830312686715794,
23
+ "acc_norm_stderr,none": 0.004643441945489853
24
+ },
25
+ "lambada_openai": {
26
+ "perplexity,none": 4.783522630921447,
27
+ "perplexity_stderr,none": 0.10727737476272488,
28
+ "acc,none": 0.6625266834853484,
29
+ "acc_stderr,none": 0.006587694938528712
30
+ },
31
+ "openbookqa": {
32
+ "acc,none": 0.294,
33
+ "acc_stderr,none": 0.020395095484936614,
34
+ "acc_norm,none": 0.398,
35
+ "acc_norm_stderr,none": 0.02191237788577998
36
+ },
37
+ "piqa": {
38
+ "acc,none": 0.7698585418933623,
39
+ "acc_stderr,none": 0.009820832826839817,
40
+ "acc_norm,none": 0.7731229597388466,
41
+ "acc_norm_stderr,none": 0.00977158425921514
42
+ },
43
+ "sciq": {
44
+ "acc,none": 0.948,
45
+ "acc_stderr,none": 0.0070246242138171325,
46
+ "acc_norm,none": 0.951,
47
+ "acc_norm_stderr,none": 0.006829761756140924
48
+ },
49
+ "wikitext": {
50
+ "word_perplexity,none": 16.10381134238307,
51
+ "byte_perplexity,none": 1.5811322783219102,
52
+ "bits_per_byte,none": 0.6609580693383097
53
+ },
54
+ "winogrande": {
55
+ "acc,none": 0.6353591160220995,
56
+ "acc_stderr,none": 0.013527746622429842
57
+ }
58
+ },
59
+ "configs": {
60
+ "arc_challenge": {
61
+ "task": "arc_challenge",
62
+ "group": [
63
+ "ai2_arc",
64
+ "multiple_choice"
65
+ ],
66
+ "dataset_path": "ai2_arc",
67
+ "dataset_name": "ARC-Challenge",
68
+ "training_split": "train",
69
+ "validation_split": "validation",
70
+ "test_split": "test",
71
+ "doc_to_text": "Question: {{question}}\nAnswer:",
72
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
73
+ "doc_to_choice": "{{choices.text}}",
74
+ "description": "",
75
+ "target_delimiter": " ",
76
+ "fewshot_delimiter": "\n\n",
77
+ "num_fewshot": 5,
78
+ "metric_list": [
79
+ {
80
+ "metric": "acc",
81
+ "aggregation": "mean",
82
+ "higher_is_better": true
83
+ },
84
+ {
85
+ "metric": "acc_norm",
86
+ "aggregation": "mean",
87
+ "higher_is_better": true
88
+ }
89
+ ],
90
+ "output_type": "multiple_choice",
91
+ "repeats": 1,
92
+ "should_decontaminate": true,
93
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
94
+ },
95
+ "arc_easy": {
96
+ "task": "arc_easy",
97
+ "group": [
98
+ "ai2_arc",
99
+ "multiple_choice"
100
+ ],
101
+ "dataset_path": "ai2_arc",
102
+ "dataset_name": "ARC-Easy",
103
+ "training_split": "train",
104
+ "validation_split": "validation",
105
+ "test_split": "test",
106
+ "doc_to_text": "Question: {{question}}\nAnswer:",
107
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
108
+ "doc_to_choice": "{{choices.text}}",
109
+ "description": "",
110
+ "target_delimiter": " ",
111
+ "fewshot_delimiter": "\n\n",
112
+ "num_fewshot": 5,
113
+ "metric_list": [
114
+ {
115
+ "metric": "acc",
116
+ "aggregation": "mean",
117
+ "higher_is_better": true
118
+ },
119
+ {
120
+ "metric": "acc_norm",
121
+ "aggregation": "mean",
122
+ "higher_is_better": true
123
+ }
124
+ ],
125
+ "output_type": "multiple_choice",
126
+ "repeats": 1,
127
+ "should_decontaminate": true,
128
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
129
+ },
130
+ "boolq": {
131
+ "task": "boolq",
132
+ "group": [
133
+ "super-glue-lm-eval-v1"
134
+ ],
135
+ "dataset_path": "super_glue",
136
+ "dataset_name": "boolq",
137
+ "training_split": "train",
138
+ "validation_split": "validation",
139
+ "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:",
140
+ "doc_to_target": "label",
141
+ "doc_to_choice": [
142
+ "no",
143
+ "yes"
144
+ ],
145
+ "description": "",
146
+ "target_delimiter": " ",
147
+ "fewshot_delimiter": "\n\n",
148
+ "num_fewshot": 5,
149
+ "metric_list": [
150
+ {
151
+ "metric": "acc"
152
+ }
153
+ ],
154
+ "output_type": "multiple_choice",
155
+ "repeats": 1,
156
+ "should_decontaminate": true,
157
+ "doc_to_decontamination_query": "passage"
158
+ },
159
+ "hellaswag": {
160
+ "task": "hellaswag",
161
+ "group": [
162
+ "multiple_choice"
163
+ ],
164
+ "dataset_path": "hellaswag",
165
+ "training_split": "train",
166
+ "validation_split": "validation",
167
+ "doc_to_text": "{% set text = activity_label ~ ': ' ~ ctx_a ~ ' ' ~ ctx_b.capitalize() %}{{text|trim|replace(' [title]', '. ')|regex_replace('\\[.*?\\]', '')|replace(' ', ' ')}}",
168
+ "doc_to_target": "{{label}}",
169
+ "doc_to_choice": "{{endings|map('trim')|map('replace', ' [title]', '. ')|map('regex_replace', '\\[.*?\\]', '')|map('replace', ' ', ' ')|list}}",
170
+ "description": "",
171
+ "target_delimiter": " ",
172
+ "fewshot_delimiter": "\n\n",
173
+ "num_fewshot": 5,
174
+ "metric_list": [
175
+ {
176
+ "metric": "acc",
177
+ "aggregation": "mean",
178
+ "higher_is_better": true
179
+ },
180
+ {
181
+ "metric": "acc_norm",
182
+ "aggregation": "mean",
183
+ "higher_is_better": true
184
+ }
185
+ ],
186
+ "output_type": "multiple_choice",
187
+ "repeats": 1,
188
+ "should_decontaminate": false
189
+ },
190
+ "lambada_openai": {
191
+ "task": "lambada_openai",
192
+ "group": [
193
+ "lambada",
194
+ "loglikelihood",
195
+ "perplexity"
196
+ ],
197
+ "dataset_path": "EleutherAI/lambada_openai",
198
+ "dataset_name": "default",
199
+ "test_split": "test",
200
+ "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
201
+ "doc_to_target": "{{' '+text.split(' ')[-1]}}",
202
+ "description": "",
203
+ "target_delimiter": " ",
204
+ "fewshot_delimiter": "\n\n",
205
+ "num_fewshot": 5,
206
+ "metric_list": [
207
+ {
208
+ "metric": "perplexity",
209
+ "aggregation": "perplexity",
210
+ "higher_is_better": false
211
+ },
212
+ {
213
+ "metric": "acc",
214
+ "aggregation": "mean",
215
+ "higher_is_better": true
216
+ }
217
+ ],
218
+ "output_type": "loglikelihood",
219
+ "repeats": 1,
220
+ "should_decontaminate": true,
221
+ "doc_to_decontamination_query": "{{text}}"
222
+ },
223
+ "openbookqa": {
224
+ "task": "openbookqa",
225
+ "group": [
226
+ "multiple_choice"
227
+ ],
228
+ "dataset_path": "openbookqa",
229
+ "dataset_name": "main",
230
+ "training_split": "train",
231
+ "validation_split": "validation",
232
+ "test_split": "test",
233
+ "doc_to_text": "question_stem",
234
+ "doc_to_target": "{{choices.label.index(answerKey.lstrip())}}",
235
+ "doc_to_choice": "{{choices.text}}",
236
+ "description": "",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "num_fewshot": 5,
240
+ "metric_list": [
241
+ {
242
+ "metric": "acc",
243
+ "aggregation": "mean",
244
+ "higher_is_better": true
245
+ },
246
+ {
247
+ "metric": "acc_norm",
248
+ "aggregation": "mean",
249
+ "higher_is_better": true
250
+ }
251
+ ],
252
+ "output_type": "multiple_choice",
253
+ "repeats": 1,
254
+ "should_decontaminate": true,
255
+ "doc_to_decontamination_query": "question_stem"
256
+ },
257
+ "piqa": {
258
+ "task": "piqa",
259
+ "group": [
260
+ "multiple_choice"
261
+ ],
262
+ "dataset_path": "piqa",
263
+ "training_split": "train",
264
+ "validation_split": "validation",
265
+ "doc_to_text": "Question: {{goal}}\nAnswer:",
266
+ "doc_to_target": "label",
267
+ "doc_to_choice": "{{[sol1, sol2]}}",
268
+ "description": "",
269
+ "target_delimiter": " ",
270
+ "fewshot_delimiter": "\n\n",
271
+ "num_fewshot": 5,
272
+ "metric_list": [
273
+ {
274
+ "metric": "acc",
275
+ "aggregation": "mean",
276
+ "higher_is_better": true
277
+ },
278
+ {
279
+ "metric": "acc_norm",
280
+ "aggregation": "mean",
281
+ "higher_is_better": true
282
+ }
283
+ ],
284
+ "output_type": "multiple_choice",
285
+ "repeats": 1,
286
+ "should_decontaminate": true,
287
+ "doc_to_decontamination_query": "goal"
288
+ },
289
+ "sciq": {
290
+ "task": "sciq",
291
+ "group": [
292
+ "multiple_choice"
293
+ ],
294
+ "dataset_path": "sciq",
295
+ "training_split": "train",
296
+ "validation_split": "validation",
297
+ "test_split": "test",
298
+ "doc_to_text": "{{support.lstrip()}}\nQuestion: {{question}}\nAnswer:",
299
+ "doc_to_target": 3,
300
+ "doc_to_choice": "{{[distractor1, distractor2, distractor3, correct_answer]}}",
301
+ "description": "",
302
+ "target_delimiter": " ",
303
+ "fewshot_delimiter": "\n\n",
304
+ "num_fewshot": 5,
305
+ "metric_list": [
306
+ {
307
+ "metric": "acc",
308
+ "aggregation": "mean",
309
+ "higher_is_better": true
310
+ },
311
+ {
312
+ "metric": "acc_norm",
313
+ "aggregation": "mean",
314
+ "higher_is_better": true
315
+ }
316
+ ],
317
+ "output_type": "multiple_choice",
318
+ "repeats": 1,
319
+ "should_decontaminate": true,
320
+ "doc_to_decontamination_query": "{{support}} {{question}}"
321
+ },
322
+ "wikitext": {
323
+ "task": "wikitext",
324
+ "group": [
325
+ "perplexity",
326
+ "loglikelihood_rolling"
327
+ ],
328
+ "dataset_path": "EleutherAI/wikitext_document_level",
329
+ "dataset_name": "wikitext-2-raw-v1",
330
+ "training_split": "train",
331
+ "validation_split": "validation",
332
+ "test_split": "test",
333
+ "doc_to_text": "",
334
+ "doc_to_target": "<function wikitext_detokenizer at 0x7f9e93ad1120>",
335
+ "description": "",
336
+ "target_delimiter": " ",
337
+ "fewshot_delimiter": "\n\n",
338
+ "num_fewshot": 5,
339
+ "metric_list": [
340
+ {
341
+ "metric": "word_perplexity"
342
+ },
343
+ {
344
+ "metric": "byte_perplexity"
345
+ },
346
+ {
347
+ "metric": "bits_per_byte"
348
+ }
349
+ ],
350
+ "output_type": "loglikelihood_rolling",
351
+ "repeats": 1,
352
+ "should_decontaminate": true,
353
+ "doc_to_decontamination_query": "{{page}}"
354
+ },
355
+ "winogrande": {
356
+ "task": "winogrande",
357
+ "dataset_path": "winogrande",
358
+ "dataset_name": "winogrande_xl",
359
+ "training_split": "train",
360
+ "validation_split": "validation",
361
+ "doc_to_text": "<function doc_to_text at 0x7f9e93ad1360>",
362
+ "doc_to_target": "<function doc_to_target at 0x7f9e93ad16c0>",
363
+ "doc_to_choice": "<function doc_to_choice at 0x7f9e93ad1a20>",
364
+ "description": "",
365
+ "target_delimiter": " ",
366
+ "fewshot_delimiter": "\n\n",
367
+ "num_fewshot": 5,
368
+ "metric_list": [
369
+ {
370
+ "metric": "acc",
371
+ "aggregation": "mean",
372
+ "higher_is_better": true
373
+ }
374
+ ],
375
+ "output_type": "multiple_choice",
376
+ "repeats": 1,
377
+ "should_decontaminate": true,
378
+ "doc_to_decontamination_query": "sentence"
379
+ }
380
+ },
381
+ "versions": {
382
+ "arc_challenge": "Yaml",
383
+ "arc_easy": "Yaml",
384
+ "boolq": "Yaml",
385
+ "hellaswag": "Yaml",
386
+ "lambada_openai": "Yaml",
387
+ "openbookqa": "Yaml",
388
+ "piqa": "Yaml",
389
+ "sciq": "Yaml",
390
+ "wikitext": "Yaml",
391
+ "winogrande": "Yaml"
392
+ },
393
+ "config": {
394
+ "model": "hf",
395
+ "model_args": "pretrained=EleutherAI/pythia-12b,parallelize=True",
396
+ "batch_size": "4",
397
+ "batch_sizes": [],
398
+ "device": null,
399
+ "use_cache": null,
400
+ "limit": null,
401
+ "bootstrap_iters": 100000
402
+ },
403
+ "git_hash": "d1a44c8"
404
+ }
sft-12b-eval-files/sft-pythia-12b-0shot-shelloutput.txt ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bootstrapping for stddev: perplexity
2
+ hf (pretrained=lomahony/eleuther-pythia12b-hh-sft,parallelize=True), limit: None, num_fewshot: 0, batch_size: 4
3
+ | Task |Version|Filter| Metric | Value | |Stderr|
4
+ |--------------|-------|------|---------------|------:|---|-----:|
5
+ |arc_challenge |Yaml |none |acc | 0.3131|± |0.0136|
6
+ | | |none |acc_norm | 0.3447|± |0.0139|
7
+ |arc_easy |Yaml |none |acc | 0.7041|± |0.0094|
8
+ | | |none |acc_norm | 0.6178|± |0.0100|
9
+ |boolq |Yaml |none |acc | 0.6945|± |0.0081|
10
+ |hellaswag |Yaml |none |acc | 0.5056|± |0.0050|
11
+ | | |none |acc_norm | 0.6685|± |0.0047|
12
+ |lambada_openai|Yaml |none |perplexity | 3.5396|± |0.0776|
13
+ | | |none |acc | 0.7093|± |0.0063|
14
+ |openbookqa |Yaml |none |acc | 0.2760|± |0.0200|
15
+ | | |none |acc_norm | 0.3820|± |0.0218|
16
+ |piqa |Yaml |none |acc | 0.7655|± |0.0099|
17
+ | | |none |acc_norm | 0.7655|± |0.0099|
18
+ |sciq |Yaml |none |acc | 0.9060|± |0.0092|
19
+ | | |none |acc_norm | 0.8410|± |0.0116|
20
+ |wikitext |Yaml |none |word_perplexity|16.4972| | |
21
+ | | |none |byte_perplexity| 1.5874| | |
22
+ | | |none |bits_per_byte | 0.6667| | |
23
+ |winogrande |Yaml |none |acc | 0.6354|± |0.0135|
24
+
sft-12b-eval-files/sft-pythia-12b-0shot/results.json ADDED
@@ -0,0 +1,404 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.31313993174061433,
5
+ "acc_stderr,none": 0.013552671543623489,
6
+ "acc_norm,none": 0.3447098976109215,
7
+ "acc_norm_stderr,none": 0.013888816286782114
8
+ },
9
+ "arc_easy": {
10
+ "acc,none": 0.7041245791245792,
11
+ "acc_stderr,none": 0.009365854134140067,
12
+ "acc_norm,none": 0.6178451178451179,
13
+ "acc_norm_stderr,none": 0.00997074728129243
14
+ },
15
+ "boolq": {
16
+ "acc,none": 0.6944954128440367,
17
+ "acc_stderr,none": 0.00805630868516481
18
+ },
19
+ "hellaswag": {
20
+ "acc,none": 0.5055765783708425,
21
+ "acc_stderr,none": 0.0049894710550909506,
22
+ "acc_norm,none": 0.6684923322047401,
23
+ "acc_norm_stderr,none": 0.004697929774670318
24
+ },
25
+ "lambada_openai": {
26
+ "perplexity,none": 3.5396245492779395,
27
+ "perplexity_stderr,none": 0.07756934951997978,
28
+ "acc,none": 0.7092955559868038,
29
+ "acc_stderr,none": 0.006326325417865827
30
+ },
31
+ "openbookqa": {
32
+ "acc,none": 0.276,
33
+ "acc_stderr,none": 0.02001121929807353,
34
+ "acc_norm,none": 0.382,
35
+ "acc_norm_stderr,none": 0.021750820591250827
36
+ },
37
+ "piqa": {
38
+ "acc,none": 0.76550598476605,
39
+ "acc_stderr,none": 0.009885203143240548,
40
+ "acc_norm,none": 0.76550598476605,
41
+ "acc_norm_stderr,none": 0.009885203143240548
42
+ },
43
+ "sciq": {
44
+ "acc,none": 0.906,
45
+ "acc_stderr,none": 0.00923305200078773,
46
+ "acc_norm,none": 0.841,
47
+ "acc_norm_stderr,none": 0.01156947936827129
48
+ },
49
+ "wikitext": {
50
+ "word_perplexity,none": 16.49715176032047,
51
+ "byte_perplexity,none": 1.5874349349461754,
52
+ "bits_per_byte,none": 0.6666974605750762
53
+ },
54
+ "winogrande": {
55
+ "acc,none": 0.6353591160220995,
56
+ "acc_stderr,none": 0.01352774662242984
57
+ }
58
+ },
59
+ "configs": {
60
+ "arc_challenge": {
61
+ "task": "arc_challenge",
62
+ "group": [
63
+ "ai2_arc",
64
+ "multiple_choice"
65
+ ],
66
+ "dataset_path": "ai2_arc",
67
+ "dataset_name": "ARC-Challenge",
68
+ "training_split": "train",
69
+ "validation_split": "validation",
70
+ "test_split": "test",
71
+ "doc_to_text": "Question: {{question}}\nAnswer:",
72
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
73
+ "doc_to_choice": "{{choices.text}}",
74
+ "description": "",
75
+ "target_delimiter": " ",
76
+ "fewshot_delimiter": "\n\n",
77
+ "num_fewshot": 0,
78
+ "metric_list": [
79
+ {
80
+ "metric": "acc",
81
+ "aggregation": "mean",
82
+ "higher_is_better": true
83
+ },
84
+ {
85
+ "metric": "acc_norm",
86
+ "aggregation": "mean",
87
+ "higher_is_better": true
88
+ }
89
+ ],
90
+ "output_type": "multiple_choice",
91
+ "repeats": 1,
92
+ "should_decontaminate": true,
93
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
94
+ },
95
+ "arc_easy": {
96
+ "task": "arc_easy",
97
+ "group": [
98
+ "ai2_arc",
99
+ "multiple_choice"
100
+ ],
101
+ "dataset_path": "ai2_arc",
102
+ "dataset_name": "ARC-Easy",
103
+ "training_split": "train",
104
+ "validation_split": "validation",
105
+ "test_split": "test",
106
+ "doc_to_text": "Question: {{question}}\nAnswer:",
107
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
108
+ "doc_to_choice": "{{choices.text}}",
109
+ "description": "",
110
+ "target_delimiter": " ",
111
+ "fewshot_delimiter": "\n\n",
112
+ "num_fewshot": 0,
113
+ "metric_list": [
114
+ {
115
+ "metric": "acc",
116
+ "aggregation": "mean",
117
+ "higher_is_better": true
118
+ },
119
+ {
120
+ "metric": "acc_norm",
121
+ "aggregation": "mean",
122
+ "higher_is_better": true
123
+ }
124
+ ],
125
+ "output_type": "multiple_choice",
126
+ "repeats": 1,
127
+ "should_decontaminate": true,
128
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
129
+ },
130
+ "boolq": {
131
+ "task": "boolq",
132
+ "group": [
133
+ "super-glue-lm-eval-v1"
134
+ ],
135
+ "dataset_path": "super_glue",
136
+ "dataset_name": "boolq",
137
+ "training_split": "train",
138
+ "validation_split": "validation",
139
+ "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:",
140
+ "doc_to_target": "label",
141
+ "doc_to_choice": [
142
+ "no",
143
+ "yes"
144
+ ],
145
+ "description": "",
146
+ "target_delimiter": " ",
147
+ "fewshot_delimiter": "\n\n",
148
+ "num_fewshot": 0,
149
+ "metric_list": [
150
+ {
151
+ "metric": "acc"
152
+ }
153
+ ],
154
+ "output_type": "multiple_choice",
155
+ "repeats": 1,
156
+ "should_decontaminate": true,
157
+ "doc_to_decontamination_query": "passage"
158
+ },
159
+ "hellaswag": {
160
+ "task": "hellaswag",
161
+ "group": [
162
+ "multiple_choice"
163
+ ],
164
+ "dataset_path": "hellaswag",
165
+ "training_split": "train",
166
+ "validation_split": "validation",
167
+ "doc_to_text": "{% set text = activity_label ~ ': ' ~ ctx_a ~ ' ' ~ ctx_b.capitalize() %}{{text|trim|replace(' [title]', '. ')|regex_replace('\\[.*?\\]', '')|replace(' ', ' ')}}",
168
+ "doc_to_target": "{{label}}",
169
+ "doc_to_choice": "{{endings|map('trim')|map('replace', ' [title]', '. ')|map('regex_replace', '\\[.*?\\]', '')|map('replace', ' ', ' ')|list}}",
170
+ "description": "",
171
+ "target_delimiter": " ",
172
+ "fewshot_delimiter": "\n\n",
173
+ "num_fewshot": 0,
174
+ "metric_list": [
175
+ {
176
+ "metric": "acc",
177
+ "aggregation": "mean",
178
+ "higher_is_better": true
179
+ },
180
+ {
181
+ "metric": "acc_norm",
182
+ "aggregation": "mean",
183
+ "higher_is_better": true
184
+ }
185
+ ],
186
+ "output_type": "multiple_choice",
187
+ "repeats": 1,
188
+ "should_decontaminate": false
189
+ },
190
+ "lambada_openai": {
191
+ "task": "lambada_openai",
192
+ "group": [
193
+ "lambada",
194
+ "loglikelihood",
195
+ "perplexity"
196
+ ],
197
+ "dataset_path": "EleutherAI/lambada_openai",
198
+ "dataset_name": "default",
199
+ "test_split": "test",
200
+ "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
201
+ "doc_to_target": "{{' '+text.split(' ')[-1]}}",
202
+ "description": "",
203
+ "target_delimiter": " ",
204
+ "fewshot_delimiter": "\n\n",
205
+ "num_fewshot": 0,
206
+ "metric_list": [
207
+ {
208
+ "metric": "perplexity",
209
+ "aggregation": "perplexity",
210
+ "higher_is_better": false
211
+ },
212
+ {
213
+ "metric": "acc",
214
+ "aggregation": "mean",
215
+ "higher_is_better": true
216
+ }
217
+ ],
218
+ "output_type": "loglikelihood",
219
+ "repeats": 1,
220
+ "should_decontaminate": true,
221
+ "doc_to_decontamination_query": "{{text}}"
222
+ },
223
+ "openbookqa": {
224
+ "task": "openbookqa",
225
+ "group": [
226
+ "multiple_choice"
227
+ ],
228
+ "dataset_path": "openbookqa",
229
+ "dataset_name": "main",
230
+ "training_split": "train",
231
+ "validation_split": "validation",
232
+ "test_split": "test",
233
+ "doc_to_text": "question_stem",
234
+ "doc_to_target": "{{choices.label.index(answerKey.lstrip())}}",
235
+ "doc_to_choice": "{{choices.text}}",
236
+ "description": "",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "num_fewshot": 0,
240
+ "metric_list": [
241
+ {
242
+ "metric": "acc",
243
+ "aggregation": "mean",
244
+ "higher_is_better": true
245
+ },
246
+ {
247
+ "metric": "acc_norm",
248
+ "aggregation": "mean",
249
+ "higher_is_better": true
250
+ }
251
+ ],
252
+ "output_type": "multiple_choice",
253
+ "repeats": 1,
254
+ "should_decontaminate": true,
255
+ "doc_to_decontamination_query": "question_stem"
256
+ },
257
+ "piqa": {
258
+ "task": "piqa",
259
+ "group": [
260
+ "multiple_choice"
261
+ ],
262
+ "dataset_path": "piqa",
263
+ "training_split": "train",
264
+ "validation_split": "validation",
265
+ "doc_to_text": "Question: {{goal}}\nAnswer:",
266
+ "doc_to_target": "label",
267
+ "doc_to_choice": "{{[sol1, sol2]}}",
268
+ "description": "",
269
+ "target_delimiter": " ",
270
+ "fewshot_delimiter": "\n\n",
271
+ "num_fewshot": 0,
272
+ "metric_list": [
273
+ {
274
+ "metric": "acc",
275
+ "aggregation": "mean",
276
+ "higher_is_better": true
277
+ },
278
+ {
279
+ "metric": "acc_norm",
280
+ "aggregation": "mean",
281
+ "higher_is_better": true
282
+ }
283
+ ],
284
+ "output_type": "multiple_choice",
285
+ "repeats": 1,
286
+ "should_decontaminate": true,
287
+ "doc_to_decontamination_query": "goal"
288
+ },
289
+ "sciq": {
290
+ "task": "sciq",
291
+ "group": [
292
+ "multiple_choice"
293
+ ],
294
+ "dataset_path": "sciq",
295
+ "training_split": "train",
296
+ "validation_split": "validation",
297
+ "test_split": "test",
298
+ "doc_to_text": "{{support.lstrip()}}\nQuestion: {{question}}\nAnswer:",
299
+ "doc_to_target": 3,
300
+ "doc_to_choice": "{{[distractor1, distractor2, distractor3, correct_answer]}}",
301
+ "description": "",
302
+ "target_delimiter": " ",
303
+ "fewshot_delimiter": "\n\n",
304
+ "num_fewshot": 0,
305
+ "metric_list": [
306
+ {
307
+ "metric": "acc",
308
+ "aggregation": "mean",
309
+ "higher_is_better": true
310
+ },
311
+ {
312
+ "metric": "acc_norm",
313
+ "aggregation": "mean",
314
+ "higher_is_better": true
315
+ }
316
+ ],
317
+ "output_type": "multiple_choice",
318
+ "repeats": 1,
319
+ "should_decontaminate": true,
320
+ "doc_to_decontamination_query": "{{support}} {{question}}"
321
+ },
322
+ "wikitext": {
323
+ "task": "wikitext",
324
+ "group": [
325
+ "perplexity",
326
+ "loglikelihood_rolling"
327
+ ],
328
+ "dataset_path": "EleutherAI/wikitext_document_level",
329
+ "dataset_name": "wikitext-2-raw-v1",
330
+ "training_split": "train",
331
+ "validation_split": "validation",
332
+ "test_split": "test",
333
+ "doc_to_text": "",
334
+ "doc_to_target": "<function wikitext_detokenizer at 0x7f91f8185120>",
335
+ "description": "",
336
+ "target_delimiter": " ",
337
+ "fewshot_delimiter": "\n\n",
338
+ "num_fewshot": 0,
339
+ "metric_list": [
340
+ {
341
+ "metric": "word_perplexity"
342
+ },
343
+ {
344
+ "metric": "byte_perplexity"
345
+ },
346
+ {
347
+ "metric": "bits_per_byte"
348
+ }
349
+ ],
350
+ "output_type": "loglikelihood_rolling",
351
+ "repeats": 1,
352
+ "should_decontaminate": true,
353
+ "doc_to_decontamination_query": "{{page}}"
354
+ },
355
+ "winogrande": {
356
+ "task": "winogrande",
357
+ "dataset_path": "winogrande",
358
+ "dataset_name": "winogrande_xl",
359
+ "training_split": "train",
360
+ "validation_split": "validation",
361
+ "doc_to_text": "<function doc_to_text at 0x7f91f8185360>",
362
+ "doc_to_target": "<function doc_to_target at 0x7f91f81856c0>",
363
+ "doc_to_choice": "<function doc_to_choice at 0x7f91f8185a20>",
364
+ "description": "",
365
+ "target_delimiter": " ",
366
+ "fewshot_delimiter": "\n\n",
367
+ "num_fewshot": 0,
368
+ "metric_list": [
369
+ {
370
+ "metric": "acc",
371
+ "aggregation": "mean",
372
+ "higher_is_better": true
373
+ }
374
+ ],
375
+ "output_type": "multiple_choice",
376
+ "repeats": 1,
377
+ "should_decontaminate": true,
378
+ "doc_to_decontamination_query": "sentence"
379
+ }
380
+ },
381
+ "versions": {
382
+ "arc_challenge": "Yaml",
383
+ "arc_easy": "Yaml",
384
+ "boolq": "Yaml",
385
+ "hellaswag": "Yaml",
386
+ "lambada_openai": "Yaml",
387
+ "openbookqa": "Yaml",
388
+ "piqa": "Yaml",
389
+ "sciq": "Yaml",
390
+ "wikitext": "Yaml",
391
+ "winogrande": "Yaml"
392
+ },
393
+ "config": {
394
+ "model": "hf",
395
+ "model_args": "pretrained=lomahony/eleuther-pythia12b-hh-sft,parallelize=True",
396
+ "batch_size": "4",
397
+ "batch_sizes": [],
398
+ "device": null,
399
+ "use_cache": null,
400
+ "limit": null,
401
+ "bootstrap_iters": 100000
402
+ },
403
+ "git_hash": "d1a44c8"
404
+ }
sft-12b-eval-files/sft-pythia-12b-5shot-shelloutput.txt ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bootstrapping for stddev: perplexity
2
+ hf (pretrained=lomahony/eleuther-pythia12b-hh-sft,parallelize=True), limit: None, num_fewshot: 5, batch_size: 4
3
+ | Task |Version|Filter| Metric | Value | |Stderr|
4
+ |--------------|-------|------|---------------|------:|---|-----:|
5
+ |arc_challenge |Yaml |none |acc | 0.3712|± |0.0141|
6
+ | | |none |acc_norm | 0.4010|± |0.0143|
7
+ |arc_easy |Yaml |none |acc | 0.7189|± |0.0092|
8
+ | | |none |acc_norm | 0.7285|± |0.0091|
9
+ |boolq |Yaml |none |acc | 0.7174|± |0.0079|
10
+ |hellaswag |Yaml |none |acc | 0.5076|± |0.0050|
11
+ | | |none |acc_norm | 0.6764|± |0.0047|
12
+ |lambada_openai|Yaml |none |perplexity | 4.3625|± |0.1008|
13
+ | | |none |acc | 0.6703|± |0.0065|
14
+ |openbookqa |Yaml |none |acc | 0.3120|± |0.0207|
15
+ | | |none |acc_norm | 0.4000|± |0.0219|
16
+ |piqa |Yaml |none |acc | 0.7688|± |0.0098|
17
+ | | |none |acc_norm | 0.7704|± |0.0098|
18
+ |sciq |Yaml |none |acc | 0.9470|± |0.0071|
19
+ | | |none |acc_norm | 0.9500|± |0.0069|
20
+ |wikitext |Yaml |none |word_perplexity|16.4972| | |
21
+ | | |none |byte_perplexity| 1.5874| | |
22
+ | | |none |bits_per_byte | 0.6667| | |
23
+ |winogrande |Yaml |none |acc | 0.6354|± |0.0135|
24
+
sft-12b-eval-files/sft-pythia-12b-5shot/results.json ADDED
@@ -0,0 +1,404 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.371160409556314,
5
+ "acc_stderr,none": 0.014117971901142818,
6
+ "acc_norm,none": 0.40102389078498296,
7
+ "acc_norm_stderr,none": 0.014322255790719867
8
+ },
9
+ "arc_easy": {
10
+ "acc,none": 0.7188552188552189,
11
+ "acc_stderr,none": 0.009224735470286998,
12
+ "acc_norm,none": 0.7285353535353535,
13
+ "acc_norm_stderr,none": 0.009125362970360623
14
+ },
15
+ "boolq": {
16
+ "acc,none": 0.7174311926605504,
17
+ "acc_stderr,none": 0.007874895085575213
18
+ },
19
+ "hellaswag": {
20
+ "acc,none": 0.5075682135032862,
21
+ "acc_stderr,none": 0.004989209770743232,
22
+ "acc_norm,none": 0.6763592909778928,
23
+ "acc_norm_stderr,none": 0.004669085411342217
24
+ },
25
+ "lambada_openai": {
26
+ "perplexity,none": 4.362525202130014,
27
+ "perplexity_stderr,none": 0.10078560700527144,
28
+ "acc,none": 0.6702891519503202,
29
+ "acc_stderr,none": 0.006549524731584282
30
+ },
31
+ "openbookqa": {
32
+ "acc,none": 0.312,
33
+ "acc_stderr,none": 0.020740596536488073,
34
+ "acc_norm,none": 0.4,
35
+ "acc_norm_stderr,none": 0.021930844120728505
36
+ },
37
+ "piqa": {
38
+ "acc,none": 0.7687704026115343,
39
+ "acc_stderr,none": 0.009837063180625326,
40
+ "acc_norm,none": 0.7704026115342764,
41
+ "acc_norm_stderr,none": 0.009812682950815206
42
+ },
43
+ "sciq": {
44
+ "acc,none": 0.947,
45
+ "acc_stderr,none": 0.007088105617246442,
46
+ "acc_norm,none": 0.95,
47
+ "acc_norm_stderr,none": 0.006895472974897893
48
+ },
49
+ "wikitext": {
50
+ "word_perplexity,none": 16.49715176032047,
51
+ "byte_perplexity,none": 1.5874349349461754,
52
+ "bits_per_byte,none": 0.6666974605750762
53
+ },
54
+ "winogrande": {
55
+ "acc,none": 0.6353591160220995,
56
+ "acc_stderr,none": 0.01352774662242984
57
+ }
58
+ },
59
+ "configs": {
60
+ "arc_challenge": {
61
+ "task": "arc_challenge",
62
+ "group": [
63
+ "ai2_arc",
64
+ "multiple_choice"
65
+ ],
66
+ "dataset_path": "ai2_arc",
67
+ "dataset_name": "ARC-Challenge",
68
+ "training_split": "train",
69
+ "validation_split": "validation",
70
+ "test_split": "test",
71
+ "doc_to_text": "Question: {{question}}\nAnswer:",
72
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
73
+ "doc_to_choice": "{{choices.text}}",
74
+ "description": "",
75
+ "target_delimiter": " ",
76
+ "fewshot_delimiter": "\n\n",
77
+ "num_fewshot": 5,
78
+ "metric_list": [
79
+ {
80
+ "metric": "acc",
81
+ "aggregation": "mean",
82
+ "higher_is_better": true
83
+ },
84
+ {
85
+ "metric": "acc_norm",
86
+ "aggregation": "mean",
87
+ "higher_is_better": true
88
+ }
89
+ ],
90
+ "output_type": "multiple_choice",
91
+ "repeats": 1,
92
+ "should_decontaminate": true,
93
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
94
+ },
95
+ "arc_easy": {
96
+ "task": "arc_easy",
97
+ "group": [
98
+ "ai2_arc",
99
+ "multiple_choice"
100
+ ],
101
+ "dataset_path": "ai2_arc",
102
+ "dataset_name": "ARC-Easy",
103
+ "training_split": "train",
104
+ "validation_split": "validation",
105
+ "test_split": "test",
106
+ "doc_to_text": "Question: {{question}}\nAnswer:",
107
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
108
+ "doc_to_choice": "{{choices.text}}",
109
+ "description": "",
110
+ "target_delimiter": " ",
111
+ "fewshot_delimiter": "\n\n",
112
+ "num_fewshot": 5,
113
+ "metric_list": [
114
+ {
115
+ "metric": "acc",
116
+ "aggregation": "mean",
117
+ "higher_is_better": true
118
+ },
119
+ {
120
+ "metric": "acc_norm",
121
+ "aggregation": "mean",
122
+ "higher_is_better": true
123
+ }
124
+ ],
125
+ "output_type": "multiple_choice",
126
+ "repeats": 1,
127
+ "should_decontaminate": true,
128
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:"
129
+ },
130
+ "boolq": {
131
+ "task": "boolq",
132
+ "group": [
133
+ "super-glue-lm-eval-v1"
134
+ ],
135
+ "dataset_path": "super_glue",
136
+ "dataset_name": "boolq",
137
+ "training_split": "train",
138
+ "validation_split": "validation",
139
+ "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:",
140
+ "doc_to_target": "label",
141
+ "doc_to_choice": [
142
+ "no",
143
+ "yes"
144
+ ],
145
+ "description": "",
146
+ "target_delimiter": " ",
147
+ "fewshot_delimiter": "\n\n",
148
+ "num_fewshot": 5,
149
+ "metric_list": [
150
+ {
151
+ "metric": "acc"
152
+ }
153
+ ],
154
+ "output_type": "multiple_choice",
155
+ "repeats": 1,
156
+ "should_decontaminate": true,
157
+ "doc_to_decontamination_query": "passage"
158
+ },
159
+ "hellaswag": {
160
+ "task": "hellaswag",
161
+ "group": [
162
+ "multiple_choice"
163
+ ],
164
+ "dataset_path": "hellaswag",
165
+ "training_split": "train",
166
+ "validation_split": "validation",
167
+ "doc_to_text": "{% set text = activity_label ~ ': ' ~ ctx_a ~ ' ' ~ ctx_b.capitalize() %}{{text|trim|replace(' [title]', '. ')|regex_replace('\\[.*?\\]', '')|replace(' ', ' ')}}",
168
+ "doc_to_target": "{{label}}",
169
+ "doc_to_choice": "{{endings|map('trim')|map('replace', ' [title]', '. ')|map('regex_replace', '\\[.*?\\]', '')|map('replace', ' ', ' ')|list}}",
170
+ "description": "",
171
+ "target_delimiter": " ",
172
+ "fewshot_delimiter": "\n\n",
173
+ "num_fewshot": 5,
174
+ "metric_list": [
175
+ {
176
+ "metric": "acc",
177
+ "aggregation": "mean",
178
+ "higher_is_better": true
179
+ },
180
+ {
181
+ "metric": "acc_norm",
182
+ "aggregation": "mean",
183
+ "higher_is_better": true
184
+ }
185
+ ],
186
+ "output_type": "multiple_choice",
187
+ "repeats": 1,
188
+ "should_decontaminate": false
189
+ },
190
+ "lambada_openai": {
191
+ "task": "lambada_openai",
192
+ "group": [
193
+ "lambada",
194
+ "loglikelihood",
195
+ "perplexity"
196
+ ],
197
+ "dataset_path": "EleutherAI/lambada_openai",
198
+ "dataset_name": "default",
199
+ "test_split": "test",
200
+ "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
201
+ "doc_to_target": "{{' '+text.split(' ')[-1]}}",
202
+ "description": "",
203
+ "target_delimiter": " ",
204
+ "fewshot_delimiter": "\n\n",
205
+ "num_fewshot": 5,
206
+ "metric_list": [
207
+ {
208
+ "metric": "perplexity",
209
+ "aggregation": "perplexity",
210
+ "higher_is_better": false
211
+ },
212
+ {
213
+ "metric": "acc",
214
+ "aggregation": "mean",
215
+ "higher_is_better": true
216
+ }
217
+ ],
218
+ "output_type": "loglikelihood",
219
+ "repeats": 1,
220
+ "should_decontaminate": true,
221
+ "doc_to_decontamination_query": "{{text}}"
222
+ },
223
+ "openbookqa": {
224
+ "task": "openbookqa",
225
+ "group": [
226
+ "multiple_choice"
227
+ ],
228
+ "dataset_path": "openbookqa",
229
+ "dataset_name": "main",
230
+ "training_split": "train",
231
+ "validation_split": "validation",
232
+ "test_split": "test",
233
+ "doc_to_text": "question_stem",
234
+ "doc_to_target": "{{choices.label.index(answerKey.lstrip())}}",
235
+ "doc_to_choice": "{{choices.text}}",
236
+ "description": "",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "num_fewshot": 5,
240
+ "metric_list": [
241
+ {
242
+ "metric": "acc",
243
+ "aggregation": "mean",
244
+ "higher_is_better": true
245
+ },
246
+ {
247
+ "metric": "acc_norm",
248
+ "aggregation": "mean",
249
+ "higher_is_better": true
250
+ }
251
+ ],
252
+ "output_type": "multiple_choice",
253
+ "repeats": 1,
254
+ "should_decontaminate": true,
255
+ "doc_to_decontamination_query": "question_stem"
256
+ },
257
+ "piqa": {
258
+ "task": "piqa",
259
+ "group": [
260
+ "multiple_choice"
261
+ ],
262
+ "dataset_path": "piqa",
263
+ "training_split": "train",
264
+ "validation_split": "validation",
265
+ "doc_to_text": "Question: {{goal}}\nAnswer:",
266
+ "doc_to_target": "label",
267
+ "doc_to_choice": "{{[sol1, sol2]}}",
268
+ "description": "",
269
+ "target_delimiter": " ",
270
+ "fewshot_delimiter": "\n\n",
271
+ "num_fewshot": 5,
272
+ "metric_list": [
273
+ {
274
+ "metric": "acc",
275
+ "aggregation": "mean",
276
+ "higher_is_better": true
277
+ },
278
+ {
279
+ "metric": "acc_norm",
280
+ "aggregation": "mean",
281
+ "higher_is_better": true
282
+ }
283
+ ],
284
+ "output_type": "multiple_choice",
285
+ "repeats": 1,
286
+ "should_decontaminate": true,
287
+ "doc_to_decontamination_query": "goal"
288
+ },
289
+ "sciq": {
290
+ "task": "sciq",
291
+ "group": [
292
+ "multiple_choice"
293
+ ],
294
+ "dataset_path": "sciq",
295
+ "training_split": "train",
296
+ "validation_split": "validation",
297
+ "test_split": "test",
298
+ "doc_to_text": "{{support.lstrip()}}\nQuestion: {{question}}\nAnswer:",
299
+ "doc_to_target": 3,
300
+ "doc_to_choice": "{{[distractor1, distractor2, distractor3, correct_answer]}}",
301
+ "description": "",
302
+ "target_delimiter": " ",
303
+ "fewshot_delimiter": "\n\n",
304
+ "num_fewshot": 5,
305
+ "metric_list": [
306
+ {
307
+ "metric": "acc",
308
+ "aggregation": "mean",
309
+ "higher_is_better": true
310
+ },
311
+ {
312
+ "metric": "acc_norm",
313
+ "aggregation": "mean",
314
+ "higher_is_better": true
315
+ }
316
+ ],
317
+ "output_type": "multiple_choice",
318
+ "repeats": 1,
319
+ "should_decontaminate": true,
320
+ "doc_to_decontamination_query": "{{support}} {{question}}"
321
+ },
322
+ "wikitext": {
323
+ "task": "wikitext",
324
+ "group": [
325
+ "perplexity",
326
+ "loglikelihood_rolling"
327
+ ],
328
+ "dataset_path": "EleutherAI/wikitext_document_level",
329
+ "dataset_name": "wikitext-2-raw-v1",
330
+ "training_split": "train",
331
+ "validation_split": "validation",
332
+ "test_split": "test",
333
+ "doc_to_text": "",
334
+ "doc_to_target": "<function wikitext_detokenizer at 0x7f05d3475120>",
335
+ "description": "",
336
+ "target_delimiter": " ",
337
+ "fewshot_delimiter": "\n\n",
338
+ "num_fewshot": 5,
339
+ "metric_list": [
340
+ {
341
+ "metric": "word_perplexity"
342
+ },
343
+ {
344
+ "metric": "byte_perplexity"
345
+ },
346
+ {
347
+ "metric": "bits_per_byte"
348
+ }
349
+ ],
350
+ "output_type": "loglikelihood_rolling",
351
+ "repeats": 1,
352
+ "should_decontaminate": true,
353
+ "doc_to_decontamination_query": "{{page}}"
354
+ },
355
+ "winogrande": {
356
+ "task": "winogrande",
357
+ "dataset_path": "winogrande",
358
+ "dataset_name": "winogrande_xl",
359
+ "training_split": "train",
360
+ "validation_split": "validation",
361
+ "doc_to_text": "<function doc_to_text at 0x7f05d3475360>",
362
+ "doc_to_target": "<function doc_to_target at 0x7f05d34756c0>",
363
+ "doc_to_choice": "<function doc_to_choice at 0x7f05d3475a20>",
364
+ "description": "",
365
+ "target_delimiter": " ",
366
+ "fewshot_delimiter": "\n\n",
367
+ "num_fewshot": 5,
368
+ "metric_list": [
369
+ {
370
+ "metric": "acc",
371
+ "aggregation": "mean",
372
+ "higher_is_better": true
373
+ }
374
+ ],
375
+ "output_type": "multiple_choice",
376
+ "repeats": 1,
377
+ "should_decontaminate": true,
378
+ "doc_to_decontamination_query": "sentence"
379
+ }
380
+ },
381
+ "versions": {
382
+ "arc_challenge": "Yaml",
383
+ "arc_easy": "Yaml",
384
+ "boolq": "Yaml",
385
+ "hellaswag": "Yaml",
386
+ "lambada_openai": "Yaml",
387
+ "openbookqa": "Yaml",
388
+ "piqa": "Yaml",
389
+ "sciq": "Yaml",
390
+ "wikitext": "Yaml",
391
+ "winogrande": "Yaml"
392
+ },
393
+ "config": {
394
+ "model": "hf",
395
+ "model_args": "pretrained=lomahony/eleuther-pythia12b-hh-sft,parallelize=True",
396
+ "batch_size": "4",
397
+ "batch_sizes": [],
398
+ "device": null,
399
+ "use_cache": null,
400
+ "limit": null,
401
+ "bootstrap_iters": 100000
402
+ },
403
+ "git_hash": "d1a44c8"
404
+ }