lodestone-base-4096-v1 / mteb_results /CQADupstackWordpressRetrieval.json
dylanAtHum's picture
Add CQADupstack Benchmarks
c25c4c3 unverified
raw
history blame
997 Bytes
{
"dataset_revision": null,
"mteb_dataset_name": "CQADupstackWordpressRetrieval",
"mteb_version": "1.1.0",
"test": {
"evaluation_time": 299.49,
"map_at_1": 0.13943,
"map_at_10": 0.2042,
"map_at_100": 0.21335,
"map_at_1000": 0.2144,
"map_at_3": 0.17865,
"map_at_5": 0.1936,
"mrr_at_1": 0.15712,
"mrr_at_10": 0.22345,
"mrr_at_100": 0.23228,
"mrr_at_1000": 0.23304,
"mrr_at_3": 0.19901,
"mrr_at_5": 0.21325,
"ndcg_at_1": 0.15712,
"ndcg_at_10": 0.24801,
"ndcg_at_100": 0.29799,
"ndcg_at_1000": 0.32514,
"ndcg_at_3": 0.19751,
"ndcg_at_5": 0.22252,
"precision_at_1": 0.15712,
"precision_at_10": 0.04177,
"precision_at_100": 0.00738,
"precision_at_1000": 0.00106,
"precision_at_3": 0.08688,
"precision_at_5": 0.06617,
"recall_at_1": 0.13943,
"recall_at_10": 0.36913,
"recall_at_100": 0.60519,
"recall_at_1000": 0.81206,
"recall_at_3": 0.23007,
"recall_at_5": 0.29082
}
}