Sentence Similarity
sentence-transformers
PyTorch
English
bert
feature-extraction
mteb
custom_code
Eval Results
6 papers
lodestone-base-4096-v1 / mteb_results /CQADupstackAndroidRetrieval.json
dylanAtHum's picture
Add CQADupstack Benchmarks
c25c4c3 unverified
{
"dataset_revision": null,
"mteb_dataset_name": "CQADupstackAndroidRetrieval",
"mteb_version": "1.1.0",
"test": {
"evaluation_time": 66.88,
"map_at_1": 0.26595,
"map_at_10": 0.36556,
"map_at_100": 0.37984,
"map_at_1000": 0.38134,
"map_at_3": 0.33417,
"map_at_5": 0.3516,
"mrr_at_1": 0.32761,
"mrr_at_10": 0.41799,
"mrr_at_100": 0.42526,
"mrr_at_1000": 0.42582,
"mrr_at_3": 0.3939,
"mrr_at_5": 0.40727,
"ndcg_at_1": 0.32761,
"ndcg_at_10": 0.42549,
"ndcg_at_100": 0.47915,
"ndcg_at_1000": 0.50475,
"ndcg_at_3": 0.3793,
"ndcg_at_5": 0.39939,
"precision_at_1": 0.32761,
"precision_at_10": 0.08312,
"precision_at_100": 0.01403,
"precision_at_1000": 0.00197,
"precision_at_3": 0.18741,
"precision_at_5": 0.13448,
"recall_at_1": 0.26595,
"recall_at_10": 0.54332,
"recall_at_100": 0.76936,
"recall_at_1000": 0.93914,
"recall_at_3": 0.40666,
"recall_at_5": 0.46513
}
}