Muennighoff commited on
Commit
bda9a2d
1 Parent(s): 529b27a

Add nordic results

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. paths.json +0 -0
  2. results.py +12 -0
  3. results/DanskBERT/AngryTweetsClassification.json +13 -0
  4. results/DanskBERT/BornholmBitextMining.json +13 -0
  5. results/DanskBERT/DKHateClassification.json +15 -0
  6. results/DanskBERT/DalajClassification.json +15 -0
  7. results/DanskBERT/DanishPoliticalCommentsClassification.json +13 -0
  8. results/DanskBERT/LccSentimentClassification.json +13 -0
  9. results/DanskBERT/MassiveIntentClassification.json +53 -0
  10. results/DanskBERT/MassiveScenarioClassification.json +53 -0
  11. results/DanskBERT/NoRecClassification.json +13 -0
  12. results/DanskBERT/NordicLangClassification.json +13 -0
  13. results/DanskBERT/NorwegianParliament.json +25 -0
  14. results/DanskBERT/ScalaDaClassification.json +15 -0
  15. results/DanskBERT/ScalaNbClassification.json +15 -0
  16. results/DanskBERT/ScalaSvClassification.json +15 -0
  17. results/DanskBERT/SwerecClassification.json +13 -0
  18. results/all-MiniLM-L6-v2/AngryTweetsClassification.json +13 -0
  19. results/all-MiniLM-L6-v2/BornholmBitextMining.json +13 -0
  20. results/all-MiniLM-L6-v2/DKHateClassification.json +15 -0
  21. results/all-MiniLM-L6-v2/DalajClassification.json +15 -0
  22. results/all-MiniLM-L6-v2/DanishPoliticalCommentsClassification.json +13 -0
  23. results/all-MiniLM-L6-v2/LccSentimentClassification.json +13 -0
  24. results/all-MiniLM-L6-v2/MassiveIntentClassification.json +47 -2
  25. results/all-MiniLM-L6-v2/MassiveScenarioClassification.json +46 -1
  26. results/all-MiniLM-L6-v2/NoRecClassification.json +13 -0
  27. results/all-MiniLM-L6-v2/NordicLangClassification.json +13 -0
  28. results/all-MiniLM-L6-v2/NorwegianParliament.json +25 -0
  29. results/all-MiniLM-L6-v2/ScalaDaClassification.json +15 -0
  30. results/all-MiniLM-L6-v2/ScalaNbClassification.json +15 -0
  31. results/all-MiniLM-L6-v2/ScalaSvClassification.json +15 -0
  32. results/all-MiniLM-L6-v2/SwerecClassification.json +13 -0
  33. results/bert-base-swedish-cased/AngryTweetsClassification.json +13 -0
  34. results/bert-base-swedish-cased/BornholmBitextMining.json +13 -0
  35. results/bert-base-swedish-cased/DKHateClassification.json +15 -0
  36. results/bert-base-swedish-cased/DalajClassification.json +15 -0
  37. results/bert-base-swedish-cased/DanishPoliticalCommentsClassification.json +13 -0
  38. results/bert-base-swedish-cased/LccSentimentClassification.json +13 -0
  39. results/bert-base-swedish-cased/MassiveIntentClassification.json +53 -0
  40. results/bert-base-swedish-cased/MassiveScenarioClassification.json +53 -0
  41. results/bert-base-swedish-cased/NoRecClassification.json +13 -0
  42. results/bert-base-swedish-cased/NordicLangClassification.json +13 -0
  43. results/bert-base-swedish-cased/NorwegianParliament.json +25 -0
  44. results/bert-base-swedish-cased/ScalaDaClassification.json +15 -0
  45. results/bert-base-swedish-cased/ScalaNbClassification.json +15 -0
  46. results/bert-base-swedish-cased/ScalaSvClassification.json +15 -0
  47. results/bert-base-swedish-cased/SweFAQRetrieval.json +38 -0
  48. results/bert-base-swedish-cased/SwerecClassification.json +13 -0
  49. results/dfm-encoder-large-v1/AngryTweetsClassification.json +13 -0
  50. results/dfm-encoder-large-v1/BornholmBitextMining.json +13 -0
paths.json CHANGED
The diff for this file is too large to render. See raw diff
 
results.py CHANGED
@@ -33,10 +33,16 @@ MODELS = [
33
  "all-MiniLM-L6-v2",
34
  "all-mpnet-base-v2",
35
  "allenai-specter",
 
36
  "bert-base-uncased",
37
  "contriever-base-msmarco",
38
  "cross-en-de-roberta-sentence-transformer",
 
 
39
  "distiluse-base-multilingual-cased-v2",
 
 
 
40
  "gbert-base",
41
  "gbert-large",
42
  "gelectra-base",
@@ -49,8 +55,13 @@ MODELS = [
49
  "gtr-t5-xxl",
50
  "komninos",
51
  "msmarco-bert-co-condensor",
 
 
 
 
52
  "paraphrase-multilingual-MiniLM-L12-v2",
53
  "paraphrase-multilingual-mpnet-base-v2",
 
54
  "sentence-t5-base",
55
  "sentence-t5-large",
56
  "sentence-t5-xl",
@@ -70,6 +81,7 @@ MODELS = [
70
  "text-search-davinci-001",
71
  "unsup-simcse-bert-base-uncased",
72
  "use-cmlm-multilingual",
 
73
  "xlm-roberta-large",
74
  ]
75
 
 
33
  "all-MiniLM-L6-v2",
34
  "all-mpnet-base-v2",
35
  "allenai-specter",
36
+ "bert-base-swedish-cased",
37
  "bert-base-uncased",
38
  "contriever-base-msmarco",
39
  "cross-en-de-roberta-sentence-transformer",
40
+ "dfm-encoder-large-v1",
41
+ "dfm-sentence-encoder-large-1",
42
  "distiluse-base-multilingual-cased-v2",
43
+ "DanskBERT",
44
+ "electra-small-nordic",
45
+ "electra-small-swedish-cased-discriminator",
46
  "gbert-base",
47
  "gbert-large",
48
  "gelectra-base",
 
55
  "gtr-t5-xxl",
56
  "komninos",
57
  "msmarco-bert-co-condensor",
58
+ "nb-bert-base",
59
+ "nb-bert-large",
60
+ "norbert3-base",
61
+ "norbert3-large",
62
  "paraphrase-multilingual-MiniLM-L12-v2",
63
  "paraphrase-multilingual-mpnet-base-v2",
64
+ "sentence-bert-swedish-cased",
65
  "sentence-t5-base",
66
  "sentence-t5-large",
67
  "sentence-t5-xl",
 
81
  "text-search-davinci-001",
82
  "unsup-simcse-bert-base-uncased",
83
  "use-cmlm-multilingual",
84
+ "xlm-roberta-base",
85
  "xlm-roberta-large",
86
  ]
87
 
results/DanskBERT/AngryTweetsClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "20b0e6081892e78179356fada741b7afa381443d",
3
+ "mteb_dataset_name": "AngryTweetsClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.5427889207258835,
7
+ "accuracy_stderr": 0.01993435732781835,
8
+ "evaluation_time": 32.51,
9
+ "f1": 0.5339328890266668,
10
+ "f1_stderr": 0.017878917919632555,
11
+ "main_score": 0.5427889207258835
12
+ }
13
+ }
results/DanskBERT/BornholmBitextMining.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3bc5cfb4ec514264fe2db5615fac9016f7251552",
3
+ "mteb_dataset_name": "BornholmBitextMining",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.09,
7
+ "evaluation_time": 13.65,
8
+ "f1": 0.06338898363570435,
9
+ "main_score": 0.06338898363570435,
10
+ "precision": 0.059330365960386246,
11
+ "recall": 0.09
12
+ }
13
+ }
results/DanskBERT/DKHateClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "59d12749a3c91a186063c7d729ec392fda94681c",
3
+ "mteb_dataset_name": "DKHateClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.5930091185410334,
7
+ "accuracy_stderr": 0.09222758041789648,
8
+ "ap": 0.8946529290076611,
9
+ "ap_stderr": 0.011813322596830038,
10
+ "evaluation_time": 34.66,
11
+ "f1": 0.48968337481112795,
12
+ "f1_stderr": 0.06219457537723159,
13
+ "main_score": 0.5930091185410334
14
+ }
15
+ }
results/DanskBERT/DalajClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7ebf0b4caa7b2ae39698a889de782c09e6f5ee56",
3
+ "mteb_dataset_name": "DalajClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.5031531531531531,
7
+ "accuracy_stderr": 0.005868835234592509,
8
+ "ap": 0.5016233522210309,
9
+ "ap_stderr": 0.002961416663478842,
10
+ "evaluation_time": 27.67,
11
+ "f1": 0.5006833169570167,
12
+ "f1_stderr": 0.0058928047682858435,
13
+ "main_score": 0.5031531531531531
14
+ }
15
+ }
results/DanskBERT/DanishPoliticalCommentsClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "edbb03726c04a0efab14fc8c3b8b79e4d420e5a1",
3
+ "mteb_dataset_name": "DanishPoliticalCommentsClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "train": {
6
+ "accuracy": 0.3981265611990008,
7
+ "accuracy_stderr": 0.02927618530034054,
8
+ "evaluation_time": 81.08,
9
+ "f1": 0.3652691159014313,
10
+ "f1_stderr": 0.02154097328963034,
11
+ "main_score": 0.3981265611990008
12
+ }
13
+ }
results/DanskBERT/LccSentimentClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "de7ba3406ee55ea2cc52a0a41408fa6aede6d3c6",
3
+ "mteb_dataset_name": "LccSentimentClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.58,
7
+ "accuracy_stderr": 0.040879225911349036,
8
+ "evaluation_time": 10.96,
9
+ "f1": 0.5646156675447656,
10
+ "f1_stderr": 0.03460749732787966,
11
+ "main_score": 0.58
12
+ }
13
+ }
results/DanskBERT/MassiveIntentClassification.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
3
+ "mteb_dataset_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "da": {
7
+ "accuracy": 0.5468392737054472,
8
+ "accuracy_stderr": 0.011536817088604236,
9
+ "f1": 0.5039000994078626,
10
+ "f1_stderr": 0.006389390451655079,
11
+ "main_score": 0.5468392737054472
12
+ },
13
+ "evaluation_time": 207.42,
14
+ "nb": {
15
+ "accuracy": 0.45379959650302626,
16
+ "accuracy_stderr": 0.012549015365232539,
17
+ "f1": 0.4271217484445252,
18
+ "f1_stderr": 0.011229470448672847,
19
+ "main_score": 0.45379959650302626
20
+ },
21
+ "sv": {
22
+ "accuracy": 0.40817081371889713,
23
+ "accuracy_stderr": 0.006134927155520256,
24
+ "f1": 0.3841954212578219,
25
+ "f1_stderr": 0.008971093419384164,
26
+ "main_score": 0.40817081371889713
27
+ }
28
+ },
29
+ "validation": {
30
+ "da": {
31
+ "accuracy": 0.5582390555828824,
32
+ "accuracy_stderr": 0.011006445314683956,
33
+ "f1": 0.5151428356070841,
34
+ "f1_stderr": 0.008861586598311164,
35
+ "main_score": 0.5582390555828824
36
+ },
37
+ "evaluation_time": 189.86,
38
+ "nb": {
39
+ "accuracy": 0.46242006886374815,
40
+ "accuracy_stderr": 0.013734839963464364,
41
+ "f1": 0.43005970598528265,
42
+ "f1_stderr": 0.013601150407790828,
43
+ "main_score": 0.46242006886374815
44
+ },
45
+ "sv": {
46
+ "accuracy": 0.4048696507624201,
47
+ "accuracy_stderr": 0.009431994322367262,
48
+ "f1": 0.38552685366886996,
49
+ "f1_stderr": 0.007369118783426967,
50
+ "main_score": 0.4048696507624201
51
+ }
52
+ }
53
+ }
results/DanskBERT/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
+ "mteb_dataset_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "da": {
7
+ "accuracy": 0.5955615332885004,
8
+ "accuracy_stderr": 0.01577699539582055,
9
+ "f1": 0.5794034655925369,
10
+ "f1_stderr": 0.013384379870283205,
11
+ "main_score": 0.5955615332885004
12
+ },
13
+ "evaluation_time": 95.61,
14
+ "nb": {
15
+ "accuracy": 0.47552118359112316,
16
+ "accuracy_stderr": 0.018872921265933037,
17
+ "f1": 0.4640926498110063,
18
+ "f1_stderr": 0.01476824680252287,
19
+ "main_score": 0.47552118359112316
20
+ },
21
+ "sv": {
22
+ "accuracy": 0.40141223940820436,
23
+ "accuracy_stderr": 0.01808423841391776,
24
+ "f1": 0.3822692785842296,
25
+ "f1_stderr": 0.013452966848637938,
26
+ "main_score": 0.40141223940820436
27
+ }
28
+ },
29
+ "validation": {
30
+ "da": {
31
+ "accuracy": 0.594392523364486,
32
+ "accuracy_stderr": 0.01934448803580611,
33
+ "f1": 0.5796578284338747,
34
+ "f1_stderr": 0.01787448188066123,
35
+ "main_score": 0.594392523364486
36
+ },
37
+ "evaluation_time": 75.72,
38
+ "nb": {
39
+ "accuracy": 0.4664043285784555,
40
+ "accuracy_stderr": 0.018904987337221752,
41
+ "f1": 0.4617714246937276,
42
+ "f1_stderr": 0.013988693425569182,
43
+ "main_score": 0.4664043285784555
44
+ },
45
+ "sv": {
46
+ "accuracy": 0.39596655189375307,
47
+ "accuracy_stderr": 0.015554735170528184,
48
+ "f1": 0.379545671770527,
49
+ "f1_stderr": 0.013408542819112435,
50
+ "main_score": 0.39596655189375307
51
+ }
52
+ }
53
+ }
results/DanskBERT/NoRecClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "07b99ab3363c2e7f8f87015b01c21f4d9b917ce3",
3
+ "mteb_dataset_name": "NoRecClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.460595703125,
7
+ "accuracy_stderr": 0.02193578721014052,
8
+ "evaluation_time": 36.11,
9
+ "f1": 0.43854924377503063,
10
+ "f1_stderr": 0.016858535482613627,
11
+ "main_score": 0.460595703125
12
+ }
13
+ }
results/DanskBERT/NordicLangClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e254179d18ab0165fdb6dbef91178266222bee2a",
3
+ "mteb_dataset_name": "NordicLangClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.7424666666666666,
7
+ "accuracy_stderr": 0.012513281832605807,
8
+ "evaluation_time": 82.57,
9
+ "f1": 0.7379219364810606,
10
+ "f1_stderr": 0.012082505150569824,
11
+ "main_score": 0.7424666666666666
12
+ }
13
+ }
results/DanskBERT/NorwegianParliament.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f7393532774c66312378d30b197610b43d751972",
3
+ "mteb_dataset_name": "NorwegianParliament",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.5679166666666666,
7
+ "accuracy_stderr": 0.029674741410461815,
8
+ "ap": 0.539603032753131,
9
+ "ap_stderr": 0.01869489598457436,
10
+ "evaluation_time": 230.39,
11
+ "f1": 0.5666235749808163,
12
+ "f1_stderr": 0.0296227793144478,
13
+ "main_score": 0.5679166666666666
14
+ },
15
+ "validation": {
16
+ "accuracy": 0.5728333333333333,
17
+ "accuracy_stderr": 0.02726719640887195,
18
+ "ap": 0.5427835002745661,
19
+ "ap_stderr": 0.01766074252735676,
20
+ "evaluation_time": 247.74,
21
+ "f1": 0.5714697907489126,
22
+ "f1_stderr": 0.026460942004883435,
23
+ "main_score": 0.5728333333333333
24
+ }
25
+ }
results/DanskBERT/ScalaDaClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1de08520a7b361e92ffa2a2201ebd41942c54675",
3
+ "mteb_dataset_name": "ScalaDaClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.66591796875,
7
+ "accuracy_stderr": 0.019314964556917027,
8
+ "ap": 0.6179104217471408,
9
+ "ap_stderr": 0.021368618593014095,
10
+ "evaluation_time": 32.34,
11
+ "f1": 0.6616198709006833,
12
+ "f1_stderr": 0.017917908374838355,
13
+ "main_score": 0.66591796875
14
+ }
15
+ }
results/DanskBERT/ScalaNbClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "237111a078ad5a834a55c57803d40bbe410ed03b",
3
+ "mteb_dataset_name": "ScalaNbClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.59990234375,
7
+ "accuracy_stderr": 0.02391900414221448,
8
+ "ap": 0.5608231549243188,
9
+ "ap_stderr": 0.016255913124253032,
10
+ "evaluation_time": 32.38,
11
+ "f1": 0.5990010371256689,
12
+ "f1_stderr": 0.024566392985160047,
13
+ "main_score": 0.59990234375
14
+ }
15
+ }
results/DanskBERT/ScalaSvClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1b48e3dcb02872335ff985ff938a054a4ed99008",
3
+ "mteb_dataset_name": "ScalaSvClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.537939453125,
7
+ "accuracy_stderr": 0.03139997732805539,
8
+ "ap": 0.5213109065272618,
9
+ "ap_stderr": 0.018150283855429068,
10
+ "evaluation_time": 39.93,
11
+ "f1": 0.5357569438251415,
12
+ "f1_stderr": 0.03126635766907699,
13
+ "main_score": 0.537939453125
14
+ }
15
+ }
results/DanskBERT/SwerecClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3c62f26bafdc4c4e1c16401ad4b32f0a94b46612",
3
+ "mteb_dataset_name": "SwerecClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.619482421875,
7
+ "accuracy_stderr": 0.020710321291315156,
8
+ "evaluation_time": 143.39,
9
+ "f1": 0.5558447995104121,
10
+ "f1_stderr": 0.021773450519163632,
11
+ "main_score": 0.619482421875
12
+ }
13
+ }
results/all-MiniLM-L6-v2/AngryTweetsClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "20b0e6081892e78179356fada741b7afa381443d",
3
+ "mteb_dataset_name": "AngryTweetsClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.42492836676217766,
7
+ "accuracy_stderr": 0.025562172565760207,
8
+ "evaluation_time": 11.79,
9
+ "f1": 0.4146959958482899,
10
+ "f1_stderr": 0.018608050832297794,
11
+ "main_score": 0.42492836676217766
12
+ }
13
+ }
results/all-MiniLM-L6-v2/BornholmBitextMining.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3bc5cfb4ec514264fe2db5615fac9016f7251552",
3
+ "mteb_dataset_name": "BornholmBitextMining",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.36,
7
+ "evaluation_time": 3.48,
8
+ "f1": 0.2968132161955691,
9
+ "main_score": 0.2968132161955691,
10
+ "precision": 0.27690919913419915,
11
+ "recall": 0.36
12
+ }
13
+ }
results/all-MiniLM-L6-v2/DKHateClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "59d12749a3c91a186063c7d729ec392fda94681c",
3
+ "mteb_dataset_name": "DKHateClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.5504559270516718,
7
+ "accuracy_stderr": 0.08179003177509295,
8
+ "ap": 0.1339388090920717,
9
+ "ap_stderr": 0.011425193008238908,
10
+ "evaluation_time": 8.1,
11
+ "f1": 0.4487544754943351,
12
+ "f1_stderr": 0.04439449341359171,
13
+ "main_score": 0.5504559270516718
14
+ }
15
+ }
results/all-MiniLM-L6-v2/DalajClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7ebf0b4caa7b2ae39698a889de782c09e6f5ee56",
3
+ "mteb_dataset_name": "DalajClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.49898648648648647,
7
+ "accuracy_stderr": 0.009521594515282868,
8
+ "ap": 0.4995893380614416,
9
+ "ap_stderr": 0.004686992032945654,
10
+ "evaluation_time": 8.53,
11
+ "f1": 0.49645260794205,
12
+ "f1_stderr": 0.010845998499157633,
13
+ "main_score": 0.49898648648648647
14
+ }
15
+ }
results/all-MiniLM-L6-v2/DanishPoliticalCommentsClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "edbb03726c04a0efab14fc8c3b8b79e4d420e5a1",
3
+ "mteb_dataset_name": "DanishPoliticalCommentsClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "train": {
6
+ "accuracy": 0.2696269982238011,
7
+ "accuracy_stderr": 0.02164496501945602,
8
+ "evaluation_time": 33.56,
9
+ "f1": 0.2394547042365672,
10
+ "f1_stderr": 0.014135416015409444,
11
+ "main_score": 0.2696269982238011
12
+ }
13
+ }
results/all-MiniLM-L6-v2/LccSentimentClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "de7ba3406ee55ea2cc52a0a41408fa6aede6d3c6",
3
+ "mteb_dataset_name": "LccSentimentClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.3846666666666666,
7
+ "accuracy_stderr": 0.03664241622309678,
8
+ "evaluation_time": 3.74,
9
+ "f1": 0.3650136884557438,
10
+ "f1_stderr": 0.03540233062350939,
11
+ "main_score": 0.3846666666666666
12
+ }
13
+ }
results/all-MiniLM-L6-v2/MassiveIntentClassification.json CHANGED
@@ -7,9 +7,54 @@
7
  "f1_stderr": 0.01101873248863987,
8
  "main_score": 0.6740416946872899
9
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  "evaluation_time": 20.8
11
  },
12
- "mteb_version": "0.0.2",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  "mteb_dataset_name": "MassiveIntentClassification",
14
- "dataset_revision": "072a486a144adf7f4479a4a0dddb2152e161e1ea"
15
  }
 
7
  "f1_stderr": 0.01101873248863987,
8
  "main_score": 0.6740416946872899
9
  },
10
+ "da": {
11
+ "accuracy": 0.4098520511096167,
12
+ "accuracy_stderr": 0.013546600921886436,
13
+ "f1": 0.39031952711786977,
14
+ "f1_stderr": 0.01465861207914092,
15
+ "main_score": 0.4098520511096167
16
+ },
17
+ "nb": {
18
+ "accuracy": 0.3934431741761937,
19
+ "accuracy_stderr": 0.010632494048802496,
20
+ "f1": 0.3725650363380114,
21
+ "f1_stderr": 0.0047186150764588165,
22
+ "main_score": 0.3934431741761937
23
+ },
24
+ "sv": {
25
+ "accuracy": 0.38103564223268327,
26
+ "accuracy_stderr": 0.015896221776346243,
27
+ "f1": 0.36129998730373103,
28
+ "f1_stderr": 0.009758357654155333,
29
+ "main_score": 0.38103564223268327
30
+ },
31
  "evaluation_time": 20.8
32
  },
33
+ "validation": {
34
+ "da": {
35
+ "accuracy": 0.396212493851451,
36
+ "accuracy_stderr": 0.016305413904394876,
37
+ "f1": 0.37526615524002804,
38
+ "f1_stderr": 0.016303136707962292,
39
+ "main_score": 0.396212493851451
40
+ },
41
+ "evaluation_time": 77.55,
42
+ "nb": {
43
+ "accuracy": 0.38145597638957207,
44
+ "accuracy_stderr": 0.008588270625147559,
45
+ "f1": 0.3557170505087034,
46
+ "f1_stderr": 0.008932333735728392,
47
+ "main_score": 0.38145597638957207
48
+ },
49
+ "sv": {
50
+ "accuracy": 0.37968519429414654,
51
+ "accuracy_stderr": 0.01211163638368146,
52
+ "f1": 0.3555505582605072,
53
+ "f1_stderr": 0.00793761524455261,
54
+ "main_score": 0.37968519429414654
55
+ }
56
+ },
57
+ "mteb_version": "1.0.3.dev0",
58
  "mteb_dataset_name": "MassiveIntentClassification",
59
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7"
60
  }
results/all-MiniLM-L6-v2/MassiveScenarioClassification.json CHANGED
@@ -7,9 +7,54 @@
7
  "f1_stderr": 0.0116798506229795,
8
  "main_score": 0.7575655682582381
9
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  "evaluation_time": 17.89
11
  },
12
- "mteb_version": "0.0.2",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  "mteb_dataset_name": "MassiveScenarioClassification",
14
  "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634"
15
  }
 
7
  "f1_stderr": 0.0116798506229795,
8
  "main_score": 0.7575655682582381
9
  },
10
+ "da": {
11
+ "accuracy": 0.47007397444519167,
12
+ "accuracy_stderr": 0.016525460362257268,
13
+ "f1": 0.44591837506202986,
14
+ "f1_stderr": 0.011715003849608418,
15
+ "main_score": 0.47007397444519167
16
+ },
17
+ "nb": {
18
+ "accuracy": 0.44670477471418957,
19
+ "accuracy_stderr": 0.01725133080004754,
20
+ "f1": 0.42175654729028544,
21
+ "f1_stderr": 0.015659543471027913,
22
+ "main_score": 0.44670477471418957
23
+ },
24
+ "sv": {
25
+ "accuracy": 0.4293207800941493,
26
+ "accuracy_stderr": 0.018548430093723933,
27
+ "f1": 0.3990548911517426,
28
+ "f1_stderr": 0.015416056576756836,
29
+ "main_score": 0.4293207800941493
30
+ },
31
  "evaluation_time": 17.89
32
  },
33
+ "validation": {
34
+ "da": {
35
+ "accuracy": 0.44692572552877524,
36
+ "accuracy_stderr": 0.012853520608224854,
37
+ "f1": 0.43967186299229216,
38
+ "f1_stderr": 0.00882990108429513,
39
+ "main_score": 0.44692572552877524
40
+ },
41
+ "evaluation_time": 27.56,
42
+ "nb": {
43
+ "accuracy": 0.4338908017707821,
44
+ "accuracy_stderr": 0.012673911928725553,
45
+ "f1": 0.4207858934675214,
46
+ "f1_stderr": 0.008274838306434095,
47
+ "main_score": 0.4338908017707821
48
+ },
49
+ "sv": {
50
+ "accuracy": 0.429021151008362,
51
+ "accuracy_stderr": 0.021686224117471994,
52
+ "f1": 0.4083151156205155,
53
+ "f1_stderr": 0.01881794915834628,
54
+ "main_score": 0.429021151008362
55
+ }
56
+ },
57
+ "mteb_version": "1.0.3.dev0",
58
  "mteb_dataset_name": "MassiveScenarioClassification",
59
  "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634"
60
  }
results/all-MiniLM-L6-v2/NoRecClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "07b99ab3363c2e7f8f87015b01c21f4d9b917ce3",
3
+ "mteb_dataset_name": "NoRecClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.4001953125,
7
+ "accuracy_stderr": 0.02129286985736529,
8
+ "evaluation_time": 11.55,
9
+ "f1": 0.3800296115219944,
10
+ "f1_stderr": 0.016433278119896217,
11
+ "main_score": 0.4001953125
12
+ }
13
+ }
results/all-MiniLM-L6-v2/NordicLangClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e254179d18ab0165fdb6dbef91178266222bee2a",
3
+ "mteb_dataset_name": "NordicLangClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.5471,
7
+ "accuracy_stderr": 0.019781051539288805,
8
+ "evaluation_time": 20.1,
9
+ "f1": 0.5337340133025357,
10
+ "f1_stderr": 0.021836739801343485,
11
+ "main_score": 0.5471
12
+ }
13
+ }
results/all-MiniLM-L6-v2/NorwegianParliament.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f7393532774c66312378d30b197610b43d751972",
3
+ "mteb_dataset_name": "NorwegianParliament",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.548,
7
+ "accuracy_stderr": 0.024218909049656952,
8
+ "ap": 0.5271332062355649,
9
+ "ap_stderr": 0.014648312153288252,
10
+ "evaluation_time": 35.46,
11
+ "f1": 0.5465938348735246,
12
+ "f1_stderr": 0.022992567784370865,
13
+ "main_score": 0.548
14
+ },
15
+ "validation": {
16
+ "accuracy": 0.5610833333333333,
17
+ "accuracy_stderr": 0.03167773929223275,
18
+ "ap": 0.5358471090493995,
19
+ "ap_stderr": 0.02036818966895372,
20
+ "evaluation_time": 35.8,
21
+ "f1": 0.5596622220610591,
22
+ "f1_stderr": 0.03076837885753835,
23
+ "main_score": 0.5610833333333333
24
+ }
25
+ }
results/all-MiniLM-L6-v2/ScalaDaClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1de08520a7b361e92ffa2a2201ebd41942c54675",
3
+ "mteb_dataset_name": "ScalaDaClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.500341796875,
7
+ "accuracy_stderr": 0.00619193763738963,
8
+ "ap": 0.5002121838569542,
9
+ "ap_stderr": 0.0031046569373535055,
10
+ "evaluation_time": 12.3,
11
+ "f1": 0.49707836995043886,
12
+ "f1_stderr": 0.0056844458165261,
13
+ "main_score": 0.500341796875
14
+ }
15
+ }
results/all-MiniLM-L6-v2/ScalaNbClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "237111a078ad5a834a55c57803d40bbe410ed03b",
3
+ "mteb_dataset_name": "ScalaNbClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.501708984375,
7
+ "accuracy_stderr": 0.005376639537486631,
8
+ "ap": 0.5008894270087827,
9
+ "ap_stderr": 0.0027141576703846324,
10
+ "evaluation_time": 11.42,
11
+ "f1": 0.4975764910012903,
12
+ "f1_stderr": 0.0057738834379346465,
13
+ "main_score": 0.501708984375
14
+ }
15
+ }
results/all-MiniLM-L6-v2/ScalaSvClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1b48e3dcb02872335ff985ff938a054a4ed99008",
3
+ "mteb_dataset_name": "ScalaSvClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.4994140625,
7
+ "accuracy_stderr": 0.004620905532937882,
8
+ "ap": 0.49973047923031955,
9
+ "ap_stderr": 0.002309036685793766,
10
+ "evaluation_time": 11.09,
11
+ "f1": 0.49683291347295633,
12
+ "f1_stderr": 0.004218877367923389,
13
+ "main_score": 0.4994140625
14
+ }
15
+ }
results/all-MiniLM-L6-v2/SwerecClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3c62f26bafdc4c4e1c16401ad4b32f0a94b46612",
3
+ "mteb_dataset_name": "SwerecClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.6083984375,
7
+ "accuracy_stderr": 0.0200147669635808,
8
+ "evaluation_time": 32.27,
9
+ "f1": 0.5383846800461662,
10
+ "f1_stderr": 0.01947712920967888,
11
+ "main_score": 0.6083984375
12
+ }
13
+ }
results/bert-base-swedish-cased/AngryTweetsClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "20b0e6081892e78179356fada741b7afa381443d",
3
+ "mteb_dataset_name": "AngryTweetsClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.4458452722063037,
7
+ "accuracy_stderr": 0.028880980134025704,
8
+ "evaluation_time": 36.3,
9
+ "f1": 0.4353629739721945,
10
+ "f1_stderr": 0.029111483185951324,
11
+ "main_score": 0.4458452722063037
12
+ }
13
+ }
results/bert-base-swedish-cased/BornholmBitextMining.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3bc5cfb4ec514264fe2db5615fac9016f7251552",
3
+ "mteb_dataset_name": "BornholmBitextMining",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.092,
7
+ "evaluation_time": 9.85,
8
+ "f1": 0.06602699130934425,
9
+ "main_score": 0.06602699130934425,
10
+ "precision": 0.0603645764405845,
11
+ "recall": 0.092
12
+ }
13
+ }
results/bert-base-swedish-cased/DKHateClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "59d12749a3c91a186063c7d729ec392fda94681c",
3
+ "mteb_dataset_name": "DKHateClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.5553191489361702,
7
+ "accuracy_stderr": 0.08659200521947144,
8
+ "ap": 0.14138288673519145,
9
+ "ap_stderr": 0.020328032456458166,
10
+ "evaluation_time": 46.95,
11
+ "f1": 0.45947587346589847,
12
+ "f1_stderr": 0.058762937812424136,
13
+ "main_score": 0.5553191489361702
14
+ }
15
+ }
results/bert-base-swedish-cased/DalajClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7ebf0b4caa7b2ae39698a889de782c09e6f5ee56",
3
+ "mteb_dataset_name": "DalajClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.5176801801801801,
7
+ "accuracy_stderr": 0.028431283803400806,
8
+ "ap": 0.5101123206773084,
9
+ "ap_stderr": 0.015255890577832387,
10
+ "evaluation_time": 17.51,
11
+ "f1": 0.5152889708732695,
12
+ "f1_stderr": 0.029514988893198382,
13
+ "main_score": 0.5176801801801801
14
+ }
15
+ }
results/bert-base-swedish-cased/DanishPoliticalCommentsClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "edbb03726c04a0efab14fc8c3b8b79e4d420e5a1",
3
+ "mteb_dataset_name": "DanishPoliticalCommentsClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "train": {
6
+ "accuracy": 0.28972024866785084,
7
+ "accuracy_stderr": 0.019830922542745974,
8
+ "evaluation_time": 113.09,
9
+ "f1": 0.2568525534958728,
10
+ "f1_stderr": 0.009279289648007597,
11
+ "main_score": 0.28972024866785084
12
+ }
13
+ }
results/bert-base-swedish-cased/LccSentimentClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "de7ba3406ee55ea2cc52a0a41408fa6aede6d3c6",
3
+ "mteb_dataset_name": "LccSentimentClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.41200000000000003,
7
+ "accuracy_stderr": 0.06469071717711042,
8
+ "evaluation_time": 13.74,
9
+ "f1": 0.3942888799816736,
10
+ "f1_stderr": 0.05673115864109014,
11
+ "main_score": 0.41200000000000003
12
+ }
13
+ }
results/bert-base-swedish-cased/MassiveIntentClassification.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
3
+ "mteb_dataset_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "da": {
7
+ "accuracy": 0.37982515131136513,
8
+ "accuracy_stderr": 0.011440520880727738,
9
+ "f1": 0.3448174943380284,
10
+ "f1_stderr": 0.009748846860402905,
11
+ "main_score": 0.37982515131136513
12
+ },
13
+ "evaluation_time": 211.15,
14
+ "nb": {
15
+ "accuracy": 0.3574983187626093,
16
+ "accuracy_stderr": 0.01486943985314982,
17
+ "f1": 0.3338735327824476,
18
+ "f1_stderr": 0.011441579699662988,
19
+ "main_score": 0.3574983187626093
20
+ },
21
+ "sv": {
22
+ "accuracy": 0.5275386684599865,
23
+ "accuracy_stderr": 0.00907799693547891,
24
+ "f1": 0.4917374109552649,
25
+ "f1_stderr": 0.007623106603827818,
26
+ "main_score": 0.5275386684599865
27
+ }
28
+ },
29
+ "validation": {
30
+ "da": {
31
+ "accuracy": 0.3873585833743237,
32
+ "accuracy_stderr": 0.017159982148168464,
33
+ "f1": 0.35577459105364684,
34
+ "f1_stderr": 0.013577678220635193,
35
+ "main_score": 0.3873585833743237
36
+ },
37
+ "evaluation_time": 189.08,
38
+ "nb": {
39
+ "accuracy": 0.366699458927693,
40
+ "accuracy_stderr": 0.01506522475665767,
41
+ "f1": 0.33492665872289507,
42
+ "f1_stderr": 0.01340872506343075,
43
+ "main_score": 0.366699458927693
44
+ },
45
+ "sv": {
46
+ "accuracy": 0.5293654697491392,
47
+ "accuracy_stderr": 0.008084871937552125,
48
+ "f1": 0.49166671254083794,
49
+ "f1_stderr": 0.01180695722613713,
50
+ "main_score": 0.5293654697491392
51
+ }
52
+ }
53
+ }
results/bert-base-swedish-cased/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
+ "mteb_dataset_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "da": {
7
+ "accuracy": 0.40437121721587077,
8
+ "accuracy_stderr": 0.0217415653668484,
9
+ "f1": 0.3723288747252761,
10
+ "f1_stderr": 0.019283422341780296,
11
+ "main_score": 0.40437121721587077
12
+ },
13
+ "evaluation_time": 95.92,
14
+ "nb": {
15
+ "accuracy": 0.35763281775386685,
16
+ "accuracy_stderr": 0.02580146772548029,
17
+ "f1": 0.3412599346756071,
18
+ "f1_stderr": 0.0237566412933957,
19
+ "main_score": 0.35763281775386685
20
+ },
21
+ "sv": {
22
+ "accuracy": 0.5609280430396772,
23
+ "accuracy_stderr": 0.024760603110657826,
24
+ "f1": 0.547979935150986,
25
+ "f1_stderr": 0.02207147319053608,
26
+ "main_score": 0.5609280430396772
27
+ }
28
+ },
29
+ "validation": {
30
+ "da": {
31
+ "accuracy": 0.39454008853910477,
32
+ "accuracy_stderr": 0.021005746966988717,
33
+ "f1": 0.3718680770887862,
34
+ "f1_stderr": 0.019314996889392557,
35
+ "main_score": 0.39454008853910477
36
+ },
37
+ "evaluation_time": 79.78,
38
+ "nb": {
39
+ "accuracy": 0.35125430398425966,
40
+ "accuracy_stderr": 0.023884953034132034,
41
+ "f1": 0.33753794511461643,
42
+ "f1_stderr": 0.021553023585863994,
43
+ "main_score": 0.35125430398425966
44
+ },
45
+ "sv": {
46
+ "accuracy": 0.5571077225774717,
47
+ "accuracy_stderr": 0.02686172612607831,
48
+ "f1": 0.5479549119696884,
49
+ "f1_stderr": 0.02461664558075998,
50
+ "main_score": 0.5571077225774717
51
+ }
52
+ }
53
+ }
results/bert-base-swedish-cased/NoRecClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "07b99ab3363c2e7f8f87015b01c21f4d9b917ce3",
3
+ "mteb_dataset_name": "NoRecClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.4390625,
7
+ "accuracy_stderr": 0.028534013227877218,
8
+ "evaluation_time": 37.11,
9
+ "f1": 0.4202061121201135,
10
+ "f1_stderr": 0.024742648756657316,
11
+ "main_score": 0.4390625
12
+ }
13
+ }
results/bert-base-swedish-cased/NordicLangClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e254179d18ab0165fdb6dbef91178266222bee2a",
3
+ "mteb_dataset_name": "NordicLangClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.6245,
7
+ "accuracy_stderr": 0.01292306293243035,
8
+ "evaluation_time": 77.67,
9
+ "f1": 0.621903954791346,
10
+ "f1_stderr": 0.012823268129236837,
11
+ "main_score": 0.6245
12
+ }
13
+ }
results/bert-base-swedish-cased/NorwegianParliament.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f7393532774c66312378d30b197610b43d751972",
3
+ "mteb_dataset_name": "NorwegianParliament",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.5755833333333334,
7
+ "accuracy_stderr": 0.0293731973442154,
8
+ "ap": 0.5448983994542387,
9
+ "ap_stderr": 0.019931830281335862,
10
+ "evaluation_time": 242.33,
11
+ "f1": 0.5723464665881223,
12
+ "f1_stderr": 0.029062139074027413,
13
+ "main_score": 0.5755833333333334
14
+ },
15
+ "validation": {
16
+ "accuracy": 0.57925,
17
+ "accuracy_stderr": 0.01670183788955243,
18
+ "ap": 0.5470070655435579,
19
+ "ap_stderr": 0.012722250147546812,
20
+ "evaluation_time": 246.19,
21
+ "f1": 0.5759161391268088,
22
+ "f1_stderr": 0.01648544668375403,
23
+ "main_score": 0.57925
24
+ }
25
+ }
results/bert-base-swedish-cased/ScalaDaClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1de08520a7b361e92ffa2a2201ebd41942c54675",
3
+ "mteb_dataset_name": "ScalaDaClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.535302734375,
7
+ "accuracy_stderr": 0.014917851250069142,
8
+ "ap": 0.5190339482077201,
9
+ "ap_stderr": 0.008496708329137853,
10
+ "evaluation_time": 40.14,
11
+ "f1": 0.5317564506131404,
12
+ "f1_stderr": 0.015658989073144235,
13
+ "main_score": 0.535302734375
14
+ }
15
+ }
results/bert-base-swedish-cased/ScalaNbClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "237111a078ad5a834a55c57803d40bbe410ed03b",
3
+ "mteb_dataset_name": "ScalaNbClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.536279296875,
7
+ "accuracy_stderr": 0.01425421682507119,
8
+ "ap": 0.5197845375798126,
9
+ "ap_stderr": 0.008407439404250613,
10
+ "evaluation_time": 35.26,
11
+ "f1": 0.5337152569584676,
12
+ "f1_stderr": 0.01295236142071566,
13
+ "main_score": 0.536279296875
14
+ }
15
+ }
results/bert-base-swedish-cased/ScalaSvClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1b48e3dcb02872335ff985ff938a054a4ed99008",
3
+ "mteb_dataset_name": "ScalaSvClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.606884765625,
7
+ "accuracy_stderr": 0.03894051387383141,
8
+ "ap": 0.5669525516433389,
9
+ "ap_stderr": 0.02720295446228148,
10
+ "evaluation_time": 27.84,
11
+ "f1": 0.6040820549557981,
12
+ "f1_stderr": 0.042546675006439924,
13
+ "main_score": 0.606884765625
14
+ }
15
+ }
results/bert-base-swedish-cased/SweFAQRetrieval.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7ebf0b4caa7b2ae39698a889de782c09e6f5ee56",
3
+ "mteb_dataset_name": "SweFAQRetrieval",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "evaluation_time": 27.16,
7
+ "map_at_1": 0.0,
8
+ "map_at_10": 0.0,
9
+ "map_at_100": 0.0,
10
+ "map_at_1000": 0.0,
11
+ "map_at_3": 0.0,
12
+ "map_at_5": 0.0,
13
+ "mrr_at_1": 0.0,
14
+ "mrr_at_10": 0.0,
15
+ "mrr_at_100": 0.0,
16
+ "mrr_at_1000": 0.0,
17
+ "mrr_at_3": 0.0,
18
+ "mrr_at_5": 0.0,
19
+ "ndcg_at_1": 0.0,
20
+ "ndcg_at_10": 0.0,
21
+ "ndcg_at_100": 0.0,
22
+ "ndcg_at_1000": 0.0,
23
+ "ndcg_at_3": 0.0,
24
+ "ndcg_at_5": 0.0,
25
+ "precision_at_1": 0.0,
26
+ "precision_at_10": 0.0,
27
+ "precision_at_100": 0.0,
28
+ "precision_at_1000": 0.0,
29
+ "precision_at_3": 0.0,
30
+ "precision_at_5": 0.0,
31
+ "recall_at_1": 0.0,
32
+ "recall_at_10": 0.0,
33
+ "recall_at_100": 0.0,
34
+ "recall_at_1000": 0.0,
35
+ "recall_at_3": 0.0,
36
+ "recall_at_5": 0.0
37
+ }
38
+ }
results/bert-base-swedish-cased/SwerecClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3c62f26bafdc4c4e1c16401ad4b32f0a94b46612",
3
+ "mteb_dataset_name": "SwerecClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.6984375,
7
+ "accuracy_stderr": 0.033370156611782956,
8
+ "evaluation_time": 115.72,
9
+ "f1": 0.6401058854724648,
10
+ "f1_stderr": 0.031149126296143328,
11
+ "main_score": 0.6984375
12
+ }
13
+ }
results/dfm-encoder-large-v1/AngryTweetsClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "20b0e6081892e78179356fada741b7afa381443d",
3
+ "mteb_dataset_name": "AngryTweetsClassification",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.5380133715377269,
7
+ "accuracy_stderr": 0.028188040634955128,
8
+ "evaluation_time": 98.1,
9
+ "f1": 0.5278713940222668,
10
+ "f1_stderr": 0.025453796294134995,
11
+ "main_score": 0.5380133715377269
12
+ }
13
+ }
results/dfm-encoder-large-v1/BornholmBitextMining.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3bc5cfb4ec514264fe2db5615fac9016f7251552",
3
+ "mteb_dataset_name": "BornholmBitextMining",
4
+ "mteb_version": "1.0.3.dev0",
5
+ "test": {
6
+ "accuracy": 0.166,
7
+ "evaluation_time": 32.52,
8
+ "f1": 0.11648071969709954,
9
+ "main_score": 0.11648071969709954,
10
+ "precision": 0.1041448634080987,
11
+ "recall": 0.166
12
+ }
13
+ }