evaluation / bloom_176B_0shot /bloom_176B_bigbench_sentence_ambiguity.json
Muennighoff's picture
Add eval
1e119e1
raw
history blame
548 Bytes
{
"results": {
"bigbench_sentence_ambiguity": {
"multiple_choice_grade": 0.5666666666666667,
"multiple_choice_grade_stderr": 0.0645132433593152
}
},
"versions": {
"bigbench_sentence_ambiguity": 0
},
"config": {
"model": "hf-causal-experimental",
"model_args": "pretrained=/pfs/lustrep2/scratch/project_462000185/muennighoff/bloom",
"num_fewshot": 0,
"batch_size": null,
"device": "cuda:0",
"no_cache": true,
"limit": null,
"bootstrap_iters": 100000,
"description_dict": {}
}
}