File size: 1,548 Bytes
bda9a2d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
{
  "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
  "mteb_dataset_name": "MassiveIntentClassification",
  "mteb_version": "1.0.3.dev0",
  "test": {
    "da": {
      "accuracy": 0.37982515131136513,
      "accuracy_stderr": 0.011440520880727738,
      "f1": 0.3448174943380284,
      "f1_stderr": 0.009748846860402905,
      "main_score": 0.37982515131136513
    },
    "evaluation_time": 211.15,
    "nb": {
      "accuracy": 0.3574983187626093,
      "accuracy_stderr": 0.01486943985314982,
      "f1": 0.3338735327824476,
      "f1_stderr": 0.011441579699662988,
      "main_score": 0.3574983187626093
    },
    "sv": {
      "accuracy": 0.5275386684599865,
      "accuracy_stderr": 0.00907799693547891,
      "f1": 0.4917374109552649,
      "f1_stderr": 0.007623106603827818,
      "main_score": 0.5275386684599865
    }
  },
  "validation": {
    "da": {
      "accuracy": 0.3873585833743237,
      "accuracy_stderr": 0.017159982148168464,
      "f1": 0.35577459105364684,
      "f1_stderr": 0.013577678220635193,
      "main_score": 0.3873585833743237
    },
    "evaluation_time": 189.08,
    "nb": {
      "accuracy": 0.366699458927693,
      "accuracy_stderr": 0.01506522475665767,
      "f1": 0.33492665872289507,
      "f1_stderr": 0.01340872506343075,
      "main_score": 0.366699458927693
    },
    "sv": {
      "accuracy": 0.5293654697491392,
      "accuracy_stderr": 0.008084871937552125,
      "f1": 0.49166671254083794,
      "f1_stderr": 0.01180695722613713,
      "main_score": 0.5293654697491392
    }
  }
}