Muennighoff's picture
Add MTEB evaluation
039f451
raw
history blame contribute delete
400 Bytes
{
"test": {
"accuracy": 0.7126109999999999,
"accuracy_stderr": 0.0371266464483395,
"ap": 0.6616336378255403,
"ap_stderr": 0.038282983973039475,
"evaluation_time": 330535.51,
"f1": 0.7089719145825304,
"f1_stderr": 0.03949505400938336,
"main_score": 0.7126109999999999
},
"dataset_version": null,
"mteb_version": "0.0.2"
}