Muennighoff's picture
Add MTEB evaluation
039f451
raw
history blame contribute delete
368 Bytes
{
"test": {
"en": {
"accuracy": 0.7242134062927497,
"accuracy_stderr": 0.012571448106282981,
"f1": 0.5503624810959269,
"f1_stderr": 0.011737816802681789,
"main_score": 0.7242134062927497
},
"evaluation_time": 612.9
},
"dataset_version": null,
"mteb_version": "0.0.2"
}