|
{ |
|
"dataset_revision": "d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46", |
|
"mteb_dataset_name": "SprintDuplicateQuestions", |
|
"mteb_version": "1.1.0", |
|
"test": { |
|
"cos_sim": { |
|
"accuracy": 0.997039603960396, |
|
"accuracy_threshold": 0.8038696646690369, |
|
"ap": 0.9040809844250262, |
|
"f1": 0.8453181583031557, |
|
"f1_threshold": 0.8038696646690369, |
|
"precision": 0.8756698821007503, |
|
"recall": 0.817 |
|
}, |
|
"dot": { |
|
"accuracy": 0.997039603960396, |
|
"accuracy_threshold": 0.8038696050643921, |
|
"ap": 0.9040809844250263, |
|
"f1": 0.8453181583031557, |
|
"f1_threshold": 0.8038696050643921, |
|
"precision": 0.8756698821007503, |
|
"recall": 0.817 |
|
}, |
|
"euclidean": { |
|
"accuracy": 0.997039603960396, |
|
"accuracy_threshold": 0.6263071894645691, |
|
"ap": 0.904080982863383, |
|
"f1": 0.8453181583031557, |
|
"f1_threshold": 0.6263071894645691, |
|
"precision": 0.8756698821007503, |
|
"recall": 0.817 |
|
}, |
|
"evaluation_time": 6.81, |
|
"manhattan": { |
|
"accuracy": 0.997, |
|
"accuracy_threshold": 13.793682098388672, |
|
"ap": 0.9039771161966652, |
|
"f1": 0.8432989690721648, |
|
"f1_threshold": 13.853282928466797, |
|
"precision": 0.8702127659574468, |
|
"recall": 0.818 |
|
}, |
|
"max": { |
|
"accuracy": 0.997039603960396, |
|
"ap": 0.9040809844250263, |
|
"f1": 0.8453181583031557 |
|
} |
|
} |
|
} |