Muennighoff commited on
Commit
17564ed
1 Parent(s): c28ee3a

Fix SummEval scores

Browse files
results/SGPT-1.3B-weightedmean-msmarco-specb-bitfit/SummEval.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "mteb_version": "0.0.2",
3
- "test": {
4
- "cos_sim": {
5
- "pearson": 0.2557602918901377,
6
- "spearman": 0.25440272876996695
7
- },
8
- "dot": {
9
- "pearson": 0.24909680980895063,
10
- "spearman": 0.24032627570006823
11
- },
12
- "evaluation_time": 507.55
13
- },
14
  "mteb_dataset_name": "SummEval",
15
- "dataset_revision": "8753c2788d36c01fc6f05d03fe3f7268d63f9122"
 
 
 
 
 
 
 
 
 
 
 
16
  }
 
1
  {
2
+ "dataset_revision": "cda12ad7615edc362dbf25a00fdd61d3b1eaf93c",
 
 
 
 
 
 
 
 
 
 
 
3
  "mteb_dataset_name": "SummEval",
4
+ "mteb_version": "1.0.2.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.3117573968015504,
8
+ "spearman": 0.3043371643155132
9
+ },
10
+ "dot": {
11
+ "pearson": 0.30164319483092744,
12
+ "spearman": 0.29207082242868754
13
+ },
14
+ "evaluation_time": 36.76
15
+ }
16
  }
results/SGPT-125M-weightedmean-msmarco-specb-bitfit-que/SummEval.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "test": {
3
- "cos_sim": {
4
- "pearson": 0.295961822471627,
5
- "spearman": 0.2890145030911965
6
- },
7
- "dot": {
8
- "pearson": 0.2917474339962901,
9
- "spearman": 0.2736297597081396
10
- },
11
- "evaluation_time": 39.55
12
- },
13
- "mteb_version": "0.0.2",
14
  "mteb_dataset_name": "SummEval",
15
- "dataset_revision": "8753c2788d36c01fc6f05d03fe3f7268d63f9122"
 
 
 
 
 
 
 
 
 
 
 
16
  }
 
1
  {
2
+ "dataset_revision": "cda12ad7615edc362dbf25a00fdd61d3b1eaf93c",
 
 
 
 
 
 
 
 
 
 
 
3
  "mteb_dataset_name": "SummEval",
4
+ "mteb_version": "1.0.2.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.29603788751645216,
8
+ "spearman": 0.29705103354786033
9
+ },
10
+ "dot": {
11
+ "pearson": 0.2807425338095399,
12
+ "spearman": 0.26841406359135367
13
+ },
14
+ "evaluation_time": 5.16
15
+ }
16
  }
results/SGPT-125M-weightedmean-msmarco-specb-bitfit/SummEval.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "test": {
3
- "cos_sim": {
4
- "pearson": 0.295961822471627,
5
- "spearman": 0.2890145030911965
6
- },
7
- "dot": {
8
- "pearson": 0.2917474339962901,
9
- "spearman": 0.2736297597081396
10
- },
11
- "evaluation_time": 39.55
12
- },
13
- "mteb_version": "0.0.2",
14
  "mteb_dataset_name": "SummEval",
15
- "dataset_revision": "8753c2788d36c01fc6f05d03fe3f7268d63f9122"
 
 
 
 
 
 
 
 
 
 
 
16
  }
 
1
  {
2
+ "dataset_revision": "cda12ad7615edc362dbf25a00fdd61d3b1eaf93c",
 
 
 
 
 
 
 
 
 
 
 
3
  "mteb_dataset_name": "SummEval",
4
+ "mteb_version": "1.0.2.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.29603788751645216,
8
+ "spearman": 0.29705103354786033
9
+ },
10
+ "dot": {
11
+ "pearson": 0.2807425338095399,
12
+ "spearman": 0.26841406359135367
13
+ },
14
+ "evaluation_time": 5.16
15
+ }
16
  }
results/SGPT-2.7B-weightedmean-msmarco-specb-bitfit/SummEval.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "mteb_version": "0.0.2",
3
- "test": {
4
- "cos_sim": {
5
- "pearson": 0.26047224542079067,
6
- "spearman": 0.27870478281195465
7
- },
8
- "dot": {
9
- "pearson": 0.2518242068570122,
10
- "spearman": 0.25116243491984985
11
- },
12
- "evaluation_time": 819.63
13
- },
14
  "mteb_dataset_name": "SummEval",
15
- "dataset_revision": "8753c2788d36c01fc6f05d03fe3f7268d63f9122"
 
 
 
 
 
 
 
 
 
 
 
16
  }
 
1
  {
2
+ "dataset_revision": "cda12ad7615edc362dbf25a00fdd61d3b1eaf93c",
 
 
 
 
 
 
 
 
 
 
 
3
  "mteb_dataset_name": "SummEval",
4
+ "mteb_version": "1.0.2.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.30027210161713946,
8
+ "spearman": 0.31030178065751735
9
+ },
10
+ "dot": {
11
+ "pearson": 0.3009179785685587,
12
+ "spearman": 0.30408303252207813
13
+ },
14
+ "evaluation_time": 73.08
15
+ }
16
  }
results/SGPT-5.8B-weightedmean-msmarco-specb-bitfit/SummEval.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "test": {
3
- "cos_sim": {
4
- "pearson": 0.2457438758817976,
5
- "spearman": 0.24747448399760644
6
- },
7
- "dot": {
8
- "pearson": 0.2658901758418499,
9
- "spearman": 0.25653620812462785
10
- },
11
- "evaluation_time": 1684.65
12
  },
13
- "mteb_version": "0.0.2",
14
- "mteb_dataset_name": "SummEval",
15
- "dataset_revision": "8753c2788d36c01fc6f05d03fe3f7268d63f9122"
 
 
 
16
  }
 
1
  {
2
+ "dataset_revision": "cda12ad7615edc362dbf25a00fdd61d3b1eaf93c",
3
+ "mteb_dataset_name": "SummEval",
4
+ "mteb_version": "1.0.2.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.30737881131277356,
8
+ "spearman": 0.3145979323917254
 
 
 
9
  },
10
+ "dot": {
11
+ "pearson": 0.2624686017962023,
12
+ "spearman": 0.25006732878791743
13
+ },
14
+ "evaluation_time": 157.09
15
+ }
16
  }
results/sgpt-bloom-7b1-msmarco/SummEval.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "mteb_version": "0.0.2",
3
- "test": {
4
- "cos_sim": {
5
- "pearson": 0.24276810356195672,
6
- "spearman": 0.24989532429036973
7
- },
8
- "dot": {
9
- "pearson": 0.26468920496607706,
10
- "spearman": 0.27254486462692024
11
- },
12
- "evaluation_time": 1594.54
13
- },
14
  "mteb_dataset_name": "SummEval",
15
- "dataset_revision": "8753c2788d36c01fc6f05d03fe3f7268d63f9122"
 
 
 
 
 
 
 
 
 
 
 
16
  }
 
1
  {
2
+ "dataset_revision": "cda12ad7615edc362dbf25a00fdd61d3b1eaf93c",
 
 
 
 
 
 
 
 
 
 
 
3
  "mteb_dataset_name": "SummEval",
4
+ "mteb_version": "1.0.2.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.31963331462886957,
8
+ "spearman": 0.3359510652629926
9
+ },
10
+ "dot": {
11
+ "pearson": 0.29033733540882123,
12
+ "spearman": 0.31550290638315504
13
+ },
14
+ "evaluation_time": 157.22
15
+ }
16
  }
results/sup-simcse-bert-base-uncased/SummEval.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "mteb_version": "0.0.2",
3
- "test": {
4
- "cos_sim": {
5
- "pearson": 0.24526207933011906,
6
- "spearman": 0.23308772096174835
7
- },
8
- "dot": {
9
- "pearson": 0.26133788672820163,
10
- "spearman": 0.24994794298618017
11
- },
12
- "evaluation_time": 338.87
13
- },
14
  "mteb_dataset_name": "SummEval",
15
- "dataset_revision": "8753c2788d36c01fc6f05d03fe3f7268d63f9122"
 
 
 
 
 
 
 
 
 
 
 
16
  }
 
1
  {
2
+ "dataset_revision": "cda12ad7615edc362dbf25a00fdd61d3b1eaf93c",
 
 
 
 
 
 
 
 
 
 
 
3
  "mteb_dataset_name": "SummEval",
4
+ "mteb_version": "1.0.2.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.3123244669972182,
8
+ "spearman": 0.31167684471295304
9
+ },
10
+ "dot": {
11
+ "pearson": 0.26725173854002066,
12
+ "spearman": 0.27639669292477526
13
+ },
14
+ "evaluation_time": 3.52
15
+ }
16
  }
results/unsup-simcse-bert-base-uncased/SummEval.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "mteb_version": "0.0.2",
3
- "test": {
4
- "cos_sim": {
5
- "pearson": 0.31363741791653166,
6
- "spearman": 0.3114893410144058
7
- },
8
- "dot": {
9
- "pearson": 0.29171225057718525,
10
- "spearman": 0.2883749298393864
11
- },
12
- "evaluation_time": 20.89
13
- },
14
  "mteb_dataset_name": "SummEval",
15
- "dataset_revision": "8753c2788d36c01fc6f05d03fe3f7268d63f9122"
 
 
 
 
 
 
 
 
 
 
 
16
  }
 
1
  {
2
+ "dataset_revision": "cda12ad7615edc362dbf25a00fdd61d3b1eaf93c",
 
 
 
 
 
 
 
 
 
 
 
3
  "mteb_dataset_name": "SummEval",
4
+ "mteb_version": "1.0.2.dev0",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.3136374010149441,
8
+ "spearman": 0.31150793544756983
9
+ },
10
+ "dot": {
11
+ "pearson": 0.29171228392763654,
12
+ "spearman": 0.28837352172904324
13
+ },
14
+ "evaluation_time": 3.5
15
+ }
16
  }