Results for st-polish-paraphrase-from-distilroberta

#11
results/st-polish-paraphrase-from-distilroberta/8TagsClustering.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "8TagsClustering",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "evaluation_time": 132.63,
7
+ "v_measure": 0.3110829518397503,
8
+ "v_measure_std": 0.045013682647523846
9
+ }
10
+ }
results/st-polish-paraphrase-from-distilroberta/AbusiveClauses.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "AbusiveClauses",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "accuracy": 0.6461627570228786,
7
+ "accuracy_stderr": 0.04703192340643617,
8
+ "ap": 0.7525956230518026,
9
+ "ap_stderr": 0.02719372192727684,
10
+ "evaluation_time": 202.16,
11
+ "f1": 0.6206314459866696,
12
+ "f1_stderr": 0.05182465207095916,
13
+ "main_score": 0.6461627570228786
14
+ }
15
+ }
results/st-polish-paraphrase-from-distilroberta/AllegroReviews.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "AllegroReviews",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "accuracy": 0.3450298210735586,
7
+ "accuracy_stderr": 0.04178368855061296,
8
+ "evaluation_time": 239.68,
9
+ "f1": 0.3118628854698987,
10
+ "f1_stderr": 0.02978356589058331,
11
+ "main_score": 0.3450298210735586
12
+ }
13
+ }
results/st-polish-paraphrase-from-distilroberta/CBD.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "CBD",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "accuracy": 0.7027,
7
+ "accuracy_stderr": 0.06656583207622362,
8
+ "ap": 0.2309733548781348,
9
+ "ap_stderr": 0.04309169873643487,
10
+ "evaluation_time": 51.47,
11
+ "f1": 0.5912637043467891,
12
+ "f1_stderr": 0.05928864394548397,
13
+ "main_score": 0.7027
14
+ }
15
+ }
results/st-polish-paraphrase-from-distilroberta/CDSC-E.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "CDSC-E",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "accuracy": 0.89,
8
+ "accuracy_threshold": 0.9488884210586548,
9
+ "ap": 0.7600097929689587,
10
+ "f1": 0.6686746987951808,
11
+ "f1_threshold": 0.9480955600738525,
12
+ "precision": 0.7816901408450704,
13
+ "recall": 0.5842105263157895
14
+ },
15
+ "dot": {
16
+ "accuracy": 0.857,
17
+ "accuracy_threshold": 166.82081604003906,
18
+ "ap": 0.5871703435202477,
19
+ "f1": 0.5686653771760155,
20
+ "f1_threshold": 129.34066772460938,
21
+ "precision": 0.44954128440366975,
22
+ "recall": 0.7736842105263158
23
+ },
24
+ "euclidean": {
25
+ "accuracy": 0.885,
26
+ "accuracy_threshold": 2.933295726776123,
27
+ "ap": 0.7515648731041904,
28
+ "f1": 0.6631578947368421,
29
+ "f1_threshold": 5.008637428283691,
30
+ "precision": 0.6631578947368421,
31
+ "recall": 0.6631578947368421
32
+ },
33
+ "evaluation_time": 18.61,
34
+ "manhattan": {
35
+ "accuracy": 0.887,
36
+ "accuracy_threshold": 81.82913208007812,
37
+ "ap": 0.7518401101929159,
38
+ "f1": 0.663013698630137,
39
+ "f1_threshold": 102.7909927368164,
40
+ "precision": 0.6914285714285714,
41
+ "recall": 0.6368421052631579
42
+ },
43
+ "max": {
44
+ "accuracy": 0.89,
45
+ "ap": 0.7600097929689587,
46
+ "f1": 0.6686746987951808
47
+ }
48
+ }
49
+ }
results/st-polish-paraphrase-from-distilroberta/CDSC-R.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "CDSC-R",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.8898345846242938,
8
+ "spearman": 0.8961978942744416
9
+ },
10
+ "euclidean": {
11
+ "pearson": 0.9039045713550165,
12
+ "spearman": 0.8983285476565798
13
+ },
14
+ "evaluation_time": 43.85,
15
+ "manhattan": {
16
+ "pearson": 0.9028956457451006,
17
+ "spearman": 0.8972744340455947
18
+ }
19
+ }
20
+ }
results/st-polish-paraphrase-from-distilroberta/MassiveIntentClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
3
+ "mteb_dataset_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "evaluation_time": 132.87,
7
+ "pl": {
8
+ "accuracy": 0.6480497646267654,
9
+ "accuracy_stderr": 0.012043153010920776,
10
+ "f1": 0.6303438023996079,
11
+ "f1_stderr": 0.010787617435549525,
12
+ "main_score": 0.6480497646267654
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 117.44,
17
+ "pl": {
18
+ "accuracy": 0.6539596655189375,
19
+ "accuracy_stderr": 0.014145832273062134,
20
+ "f1": 0.6311905309567906,
21
+ "f1_stderr": 0.011607856216746485,
22
+ "main_score": 0.6539596655189375
23
+ }
24
+ }
25
+ }
results/st-polish-paraphrase-from-distilroberta/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
+ "mteb_dataset_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "evaluation_time": 68.73,
7
+ "pl": {
8
+ "accuracy": 0.700168123739072,
9
+ "accuracy_stderr": 0.01735183600567955,
10
+ "f1": 0.6973781126384635,
11
+ "f1_stderr": 0.011896976130469463,
12
+ "main_score": 0.700168123739072
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 55.43,
17
+ "pl": {
18
+ "accuracy": 0.7025577963600591,
19
+ "accuracy_stderr": 0.02185321275120188,
20
+ "f1": 0.6980674385249522,
21
+ "f1_stderr": 0.016427028807319447,
22
+ "main_score": 0.7025577963600591
23
+ }
24
+ }
25
+ }
results/st-polish-paraphrase-from-distilroberta/PPC.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "PPC",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "accuracy": 0.883,
8
+ "accuracy_threshold": 0.8584332466125488,
9
+ "ap": 0.9329073067322579,
10
+ "f1": 0.9075369075369075,
11
+ "f1_threshold": 0.825851559638977,
12
+ "precision": 0.8550512445095169,
13
+ "recall": 0.9668874172185431
14
+ },
15
+ "dot": {
16
+ "accuracy": 0.752,
17
+ "accuracy_threshold": 128.82369995117188,
18
+ "ap": 0.86640535018982,
19
+ "f1": 0.8066465256797583,
20
+ "f1_threshold": 110.18665313720703,
21
+ "precision": 0.7416666666666667,
22
+ "recall": 0.8841059602649006
23
+ },
24
+ "euclidean": {
25
+ "accuracy": 0.887,
26
+ "accuracy_threshold": 6.4312896728515625,
27
+ "ap": 0.9327063012204004,
28
+ "f1": 0.9093774625689519,
29
+ "f1_threshold": 7.100665092468262,
30
+ "precision": 0.8676691729323308,
31
+ "recall": 0.9552980132450332
32
+ },
33
+ "evaluation_time": 30.77,
34
+ "manhattan": {
35
+ "accuracy": 0.888,
36
+ "accuracy_threshold": 152.89788818359375,
37
+ "ap": 0.9331080134023166,
38
+ "f1": 0.9119496855345912,
39
+ "f1_threshold": 156.10391235351562,
40
+ "precision": 0.8682634730538922,
41
+ "recall": 0.9602649006622517
42
+ },
43
+ "max": {
44
+ "accuracy": 0.888,
45
+ "ap": 0.9331080134023166,
46
+ "f1": 0.9119496855345912
47
+ }
48
+ }
49
+ }
results/st-polish-paraphrase-from-distilroberta/PSC.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "PSC",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "accuracy": 0.9703153988868275,
8
+ "accuracy_threshold": 0.6410723924636841,
9
+ "ap": 0.9910327047985201,
10
+ "f1": 0.9525222551928784,
11
+ "f1_threshold": 0.6144498586654663,
12
+ "precision": 0.9277456647398844,
13
+ "recall": 0.9786585365853658
14
+ },
15
+ "dot": {
16
+ "accuracy": 0.9554730983302412,
17
+ "accuracy_threshold": 48.173728942871094,
18
+ "ap": 0.9606948765620352,
19
+ "f1": 0.9286754002911208,
20
+ "f1_threshold": 44.893150329589844,
21
+ "precision": 0.8885793871866295,
22
+ "recall": 0.9725609756097561
23
+ },
24
+ "euclidean": {
25
+ "accuracy": 0.9517625231910947,
26
+ "accuracy_threshold": 7.655665397644043,
27
+ "ap": 0.9724073320491486,
28
+ "f1": 0.9226190476190477,
29
+ "f1_threshold": 7.677097797393799,
30
+ "precision": 0.9011627906976745,
31
+ "recall": 0.9451219512195121
32
+ },
33
+ "evaluation_time": 121.52,
34
+ "manhattan": {
35
+ "accuracy": 0.9499072356215214,
36
+ "accuracy_threshold": 161.79623413085938,
37
+ "ap": 0.9723500795454882,
38
+ "f1": 0.9191616766467067,
39
+ "f1_threshold": 164.94464111328125,
40
+ "precision": 0.9029411764705882,
41
+ "recall": 0.9359756097560976
42
+ },
43
+ "max": {
44
+ "accuracy": 0.9703153988868275,
45
+ "ap": 0.9910327047985201,
46
+ "f1": 0.9525222551928784
47
+ }
48
+ }
49
+ }
results/st-polish-paraphrase-from-distilroberta/PolEmo2.0-IN.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "PolEmo2.0-IN",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "accuracy": 0.6707756232686981,
7
+ "accuracy_stderr": 0.03598470161067884,
8
+ "evaluation_time": 251.61,
9
+ "f1": 0.6651927606236673,
10
+ "f1_stderr": 0.0326670738484192,
11
+ "main_score": 0.6707756232686981
12
+ }
13
+ }
results/st-polish-paraphrase-from-distilroberta/PolEmo2.0-OUT.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "PolEmo2.0-OUT",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "accuracy": 0.3856275303643725,
7
+ "accuracy_stderr": 0.05773490862681293,
8
+ "evaluation_time": 195.95,
9
+ "f1": 0.3327499365030043,
10
+ "f1_stderr": 0.04245736209871192,
11
+ "main_score": 0.3856275303643725
12
+ }
13
+ }
results/st-polish-paraphrase-from-distilroberta/SICK-E-PL.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "SICK-E-PL",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "accuracy": 0.8414186710150836,
8
+ "accuracy_threshold": 0.9119361639022827,
9
+ "ap": 0.7963023934906186,
10
+ "f1": 0.7201365187713311,
11
+ "f1_threshold": 0.8689566850662231,
12
+ "precision": 0.6913499344692006,
13
+ "recall": 0.7514245014245015
14
+ },
15
+ "dot": {
16
+ "accuracy": 0.795760293518141,
17
+ "accuracy_threshold": 155.5910186767578,
18
+ "ap": 0.6841926705773022,
19
+ "f1": 0.6415425065731815,
20
+ "f1_threshold": 131.79505920410156,
21
+ "precision": 0.5438335809806835,
22
+ "recall": 0.782051282051282
23
+ },
24
+ "euclidean": {
25
+ "accuracy": 0.8383611903791276,
26
+ "accuracy_threshold": 5.6178741455078125,
27
+ "ap": 0.7875560712800018,
28
+ "f1": 0.7141891891891893,
29
+ "f1_threshold": 6.6637773513793945,
30
+ "precision": 0.679305912596401,
31
+ "recall": 0.7528490028490028
32
+ },
33
+ "evaluation_time": 70.54,
34
+ "manhattan": {
35
+ "accuracy": 0.8379535262943335,
36
+ "accuracy_threshold": 115.29086303710938,
37
+ "ap": 0.7886951592975753,
38
+ "f1": 0.7144265527440025,
39
+ "f1_threshold": 149.68344116210938,
40
+ "precision": 0.6632092739475289,
41
+ "recall": 0.7742165242165242
42
+ },
43
+ "max": {
44
+ "accuracy": 0.8414186710150836,
45
+ "ap": 0.7963023934906186,
46
+ "f1": 0.7201365187713311
47
+ }
48
+ }
49
+ }
results/st-polish-paraphrase-from-distilroberta/SICK-R-PL.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": null,
3
+ "mteb_dataset_name": "SICK-R-PL",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.8271547647112909,
8
+ "spearman": 0.7637008919163761
9
+ },
10
+ "euclidean": {
11
+ "pearson": 0.8035940908992067,
12
+ "spearman": 0.7567176506649443
13
+ },
14
+ "evaluation_time": 131.41,
15
+ "manhattan": {
16
+ "pearson": 0.8024392672884499,
17
+ "spearman": 0.7561711016653669
18
+ }
19
+ }
20
+ }
results/st-polish-paraphrase-from-distilroberta/STS22.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "6d1ba47164174a496b7fa5d3569dae26a6813b80",
3
+ "mteb_dataset_name": "STS22",
4
+ "mteb_version": "1.1.1.dev0",
5
+ "test": {
6
+ "evaluation_time": 101.93,
7
+ "pl": {
8
+ "cos_sim": {
9
+ "pearson": 0.396312225002525,
10
+ "spearman": 0.40346668940521924
11
+ },
12
+ "euclidean": {
13
+ "pearson": 0.3027778152050383,
14
+ "spearman": 0.4024228369767347
15
+ },
16
+ "manhattan": {
17
+ "pearson": 0.30317179990717374,
18
+ "spearman": 0.4001470417293502
19
+ }
20
+ }
21
+ }
22
+ }