Muennighoff's picture
Add MTEB evaluation
5d4e702
raw
history blame
293 Bytes
{
"dataset_version": null,
"mteb_version": "0.0.2",
"test": {
"accuracy": 0.46390000000000003,
"accuracy_stderr": 0.018785366645343925,
"evaluation_time": 31.09,
"f1": 0.42264248858567033,
"f1_stderr": 0.011565342464001555,
"main_score": 0.46390000000000003
}
}