Sentence Similarity
sentence-transformers
PyTorch
English
bert
feature-extraction
mteb
custom_code
Eval Results
text-embeddings-inference
6 papers
lodestone-base-4096-v1 / mteb_results /CQADupstackWebmastersRetrieval.json
dylanAtHum's picture
Add CQADupstack Benchmarks
c25c4c3 unverified
{
"dataset_revision": null,
"mteb_dataset_name": "CQADupstackWebmastersRetrieval",
"mteb_version": "1.1.0",
"test": {
"evaluation_time": 63.51,
"map_at_1": 0.22807,
"map_at_10": 0.30014,
"map_at_100": 0.31422,
"map_at_1000": 0.31652,
"map_at_3": 0.27447,
"map_at_5": 0.28711,
"mrr_at_1": 0.27668,
"mrr_at_10": 0.34489,
"mrr_at_100": 0.35453,
"mrr_at_1000": 0.35526,
"mrr_at_3": 0.32477,
"mrr_at_5": 0.33603,
"ndcg_at_1": 0.27668,
"ndcg_at_10": 0.34983,
"ndcg_at_100": 0.40535,
"ndcg_at_1000": 0.43747,
"ndcg_at_3": 0.31027,
"ndcg_at_5": 0.32608,
"precision_at_1": 0.27668,
"precision_at_10": 0.06838,
"precision_at_100": 0.01411,
"precision_at_1000": 0.00236,
"precision_at_3": 0.14295,
"precision_at_5": 0.10435,
"recall_at_1": 0.22807,
"recall_at_10": 0.43545,
"recall_at_100": 0.69398,
"recall_at_1000": 0.90706,
"recall_at_3": 0.32183,
"recall_at_5": 0.36563
}
}