nouamanetazi HF staff commited on
Commit
b475468
1 Parent(s): 3a2f43c

add model column

Browse files
LASER2_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.7891044776119402
3
- AmazonPolarityClassification,accuracy,0.6100945
4
- AmazonReviewsClassification,accuracy,0.30578
5
- Banking77Classification,accuracy,0.5775974025974026
6
- EmotionClassification,accuracy,0.24830000000000002
7
- ImdbClassification,accuracy,0.57584
8
- MassiveIntentClassification,accuracy,0.4790181573638197
9
- MassiveScenarioClassification,accuracy,0.558238063214526
10
- MTOPDomainClassification,accuracy,0.7610579115367078
11
- MTOPIntentClassification,accuracy,0.5047423620611035
12
- ToxicConversationsClassification,accuracy,0.5404779999999999
13
- TweetSentimentExtractionClassification,accuracy,0.4872665534804754
14
- ArxivClusteringP2P,v_measure,0.1776823856238192
15
- ArxivClusteringS2S,v_measure,0.1239260518556585
16
- BiorxivClusteringP2P,v_measure,0.12399936477309108
17
- BiorxivClusteringS2S,v_measure,0.08827421024926384
18
- MedrxivClusteringP2P,v_measure,0.17908142247465778
19
- MedrxivClusteringS2S,v_measure,0.1662952889451872
20
- RedditClustering,v_measure,0.0996202584007419
21
- RedditClusteringP2P,v_measure,0.264185580282609
22
- StackExchangeClustering,v_measure,0.15794084761797694
23
- StackExchangeClusteringP2P,v_measure,0.18627980064437472
24
- TwentyNewsgroupsClustering,v_measure,0.11378130183913912
25
- SprintDuplicateQuestions,ap,0.6554452704075004
26
- TwitterSemEval2015,ap,0.5956789548901481
27
- TwitterURLCorpus,ap,0.8147393118163185
28
- AskUbuntuDupQuestions,map,0.4898595820868522
29
- MindSmallReranking,map,0.24788913161151171
30
- SciDocsRR,map,0.5498656851897086
31
- StackOverflowDupQuestions,map,0.36983704984940896
32
- ArguAna,ndcg_at_10,0.12856
33
- ClimateFEVER,ndcg_at_10,0.0036
34
- CQADupstackRetrieval,ndcg_at_10,0.04119
35
- DBPedia,ndcg_at_10,0.01526
36
- FEVER,ndcg_at_10,0.00767
37
- FiQA2018,ndcg_at_10,0.01725
38
- HotpotQA,ndcg_at_10,0.05504
39
- MSMARCO,ndcg_at_10,0.03593
40
- NFCorpus,ndcg_at_10,0.02439
41
- NQ,ndcg_at_10,0.00642
42
- QuoraRetrieval,ndcg_at_10,0.71145
43
- SCIDOCS,ndcg_at_10,0.00777
44
- SciFact,ndcg_at_10,0.04038
45
- Touche2020,ndcg_at_10,0.0106
46
- TRECCOVID,ndcg_at_10,0.10972
47
- BIOSSES,cosine_spearman,0.6200864182463187
48
- SICK-R,cosine_spearman,0.6285761430269268
49
- STS12,cosine_spearman,0.6260375600516247
50
- STS13,cosine_spearman,0.5961753521490076
51
- STS14,cosine_spearman,0.5702962555736466
52
- STS15,cosine_spearman,0.7156560343816677
53
- STS16,cosine_spearman,0.7075338261767046
54
- STS17,cosine_spearman,0.7672812705124583
55
- STS22,cosine_spearman,0.3974427076488482
56
- STSBenchmark,cosine_spearman,0.6977241730309153
57
- SummEval,cosine_spearman,0.268033428993562
 
1
+ model,dataset,metric,value
2
+ LASER2,AmazonCounterfactualClassification,accuracy,0.7891044776119402
3
+ LASER2,AmazonPolarityClassification,accuracy,0.6100945
4
+ LASER2,AmazonReviewsClassification,accuracy,0.30578
5
+ LASER2,Banking77Classification,accuracy,0.5775974025974026
6
+ LASER2,EmotionClassification,accuracy,0.24830000000000002
7
+ LASER2,ImdbClassification,accuracy,0.57584
8
+ LASER2,MassiveIntentClassification,accuracy,0.4790181573638197
9
+ LASER2,MassiveScenarioClassification,accuracy,0.558238063214526
10
+ LASER2,MTOPDomainClassification,accuracy,0.7610579115367078
11
+ LASER2,MTOPIntentClassification,accuracy,0.5047423620611035
12
+ LASER2,ToxicConversationsClassification,accuracy,0.5404779999999999
13
+ LASER2,TweetSentimentExtractionClassification,accuracy,0.4872665534804754
14
+ LASER2,ArxivClusteringP2P,v_measure,0.1776823856238192
15
+ LASER2,ArxivClusteringS2S,v_measure,0.1239260518556585
16
+ LASER2,BiorxivClusteringP2P,v_measure,0.12399936477309108
17
+ LASER2,BiorxivClusteringS2S,v_measure,0.08827421024926384
18
+ LASER2,MedrxivClusteringP2P,v_measure,0.17908142247465778
19
+ LASER2,MedrxivClusteringS2S,v_measure,0.1662952889451872
20
+ LASER2,RedditClustering,v_measure,0.0996202584007419
21
+ LASER2,RedditClusteringP2P,v_measure,0.264185580282609
22
+ LASER2,StackExchangeClustering,v_measure,0.15794084761797694
23
+ LASER2,StackExchangeClusteringP2P,v_measure,0.18627980064437472
24
+ LASER2,TwentyNewsgroupsClustering,v_measure,0.11378130183913912
25
+ LASER2,SprintDuplicateQuestions,ap,0.6554452704075004
26
+ LASER2,TwitterSemEval2015,ap,0.5956789548901481
27
+ LASER2,TwitterURLCorpus,ap,0.8147393118163185
28
+ LASER2,AskUbuntuDupQuestions,map,0.4898595820868522
29
+ LASER2,MindSmallReranking,map,0.24788913161151171
30
+ LASER2,SciDocsRR,map,0.5498656851897086
31
+ LASER2,StackOverflowDupQuestions,map,0.36983704984940896
32
+ LASER2,ArguAna,ndcg_at_10,0.12856
33
+ LASER2,ClimateFEVER,ndcg_at_10,0.0036
34
+ LASER2,CQADupstackRetrieval,ndcg_at_10,0.04119
35
+ LASER2,DBPedia,ndcg_at_10,0.01526
36
+ LASER2,FEVER,ndcg_at_10,0.00767
37
+ LASER2,FiQA2018,ndcg_at_10,0.01725
38
+ LASER2,HotpotQA,ndcg_at_10,0.05504
39
+ LASER2,MSMARCO,ndcg_at_10,0.03593
40
+ LASER2,NFCorpus,ndcg_at_10,0.02439
41
+ LASER2,NQ,ndcg_at_10,0.00642
42
+ LASER2,QuoraRetrieval,ndcg_at_10,0.71145
43
+ LASER2,SCIDOCS,ndcg_at_10,0.00777
44
+ LASER2,SciFact,ndcg_at_10,0.04038
45
+ LASER2,Touche2020,ndcg_at_10,0.0106
46
+ LASER2,TRECCOVID,ndcg_at_10,0.10972
47
+ LASER2,BIOSSES,cosine_spearman,0.6200864182463187
48
+ LASER2,SICK-R,cosine_spearman,0.6285761430269268
49
+ LASER2,STS12,cosine_spearman,0.6260375600516247
50
+ LASER2,STS13,cosine_spearman,0.5961753521490076
51
+ LASER2,STS14,cosine_spearman,0.5702962555736466
52
+ LASER2,STS15,cosine_spearman,0.7156560343816677
53
+ LASER2,STS16,cosine_spearman,0.7075338261767046
54
+ LASER2,STS17,cosine_spearman,0.7672812705124583
55
+ LASER2,STS22,cosine_spearman,0.3974427076488482
56
+ LASER2,STSBenchmark,cosine_spearman,0.6977241730309153
57
+ LASER2,SummEval,cosine_spearman,0.268033428993562
SGPT-125M-weightedmean-msmarco-specb-bitfit_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.6123880597014926
3
- AmazonPolarityClassification,accuracy,0.65401225
4
- AmazonReviewsClassification,accuracy,0.31165999999999994
5
- Banking77Classification,accuracy,0.7770454545454545
6
- EmotionClassification,accuracy,0.39075000000000004
7
- ImdbClassification,accuracy,0.586696
8
- MassiveIntentClassification,accuracy,0.6140887693342301
9
- MassiveScenarioClassification,accuracy,0.6973772696704774
10
- MTOPDomainClassification,accuracy,0.8695622435020519
11
- MTOPIntentClassification,accuracy,0.6225262197902417
12
- ToxicConversationsClassification,accuracy,0.6265799999999999
13
- TweetSentimentExtractionClassification,accuracy,0.5240803621958121
14
- ArxivClusteringP2P,v_measure,0.3970858340673288
15
- ArxivClusteringS2S,v_measure,0.2824284771372105
16
- BiorxivClusteringP2P,v_measure,0.33632603955439844
17
- BiorxivClusteringS2S,v_measure,0.27038042665369927
18
- MedrxivClusteringP2P,v_measure,0.31374938993074253
19
- MedrxivClusteringS2S,v_measure,0.26871455379644094
20
- RedditClustering,v_measure,0.4023390747226228
21
- RedditClusteringP2P,v_measure,0.49090518272935624
22
- StackExchangeClustering,v_measure,0.5274481093815175
23
- StackExchangeClusteringP2P,v_measure,0.3265999453562101
24
- TwentyNewsgroupsClustering,v_measure,0.3212697126747911
25
- SprintDuplicateQuestions,ap,0.8988577913120002
26
- TwitterSemEval2015,ap,0.5474680676121269
27
- TwitterURLCorpus,ap,0.8105760818661524
28
- AskUbuntuDupQuestions,map,0.5583700395192394
29
- MindSmallReranking,map,0.30402396942935334
30
- SciDocsRR,map,0.7133941904192648
31
- StackOverflowDupQuestions,map,0.44744984645554653
32
- ArguAna,ndcg_at_10,0.47481
33
- ClimateFEVER,ndcg_at_10,0.21345
34
- CQADupstackRetrieval,ndcg_at_10,0.2724866666666666
35
- DBPedia,ndcg_at_10,0.22365
36
- FEVER,ndcg_at_10,
37
- FiQA2018,ndcg_at_10,0.21472
38
- HotpotQA,ndcg_at_10,
39
- MSMARCO,ndcg_at_10,0.54238
40
- NFCorpus,ndcg_at_10,0.2303
41
- NQ,ndcg_at_10,0.27858
42
- QuoraRetrieval,ndcg_at_10,0.82688
43
- SCIDOCS,ndcg_at_10,0.12677
44
- SciFact,ndcg_at_10,0.57881
45
- Touche2020,ndcg_at_10,0.21525
46
- TRECCOVID,ndcg_at_10,0.57921
47
- BIOSSES,cosine_spearman,0.7520954502580506
48
- SICK-R,cosine_spearman,0.7520954502580506
49
- STS12,cosine_spearman,0.6652980061546658
50
- STS13,cosine_spearman,0.7616628863742361
51
- STS14,cosine_spearman,0.6904572664009687
52
- STS15,cosine_spearman,0.7923677712825851
53
- STS16,cosine_spearman,0.7606792422441928
54
- STS17,cosine_spearman,0.8495398260629698
55
- STS22,cosine_spearman,0.6566183708171826
56
- STSBenchmark,cosine_spearman,0.7533716094627373
57
- SummEval,cosine_spearman,0.2890145030911965
 
1
+ model,dataset,metric,value
2
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,AmazonCounterfactualClassification,accuracy,0.6123880597014926
3
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,AmazonPolarityClassification,accuracy,0.65401225
4
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,AmazonReviewsClassification,accuracy,0.31165999999999994
5
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,Banking77Classification,accuracy,0.7770454545454545
6
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,EmotionClassification,accuracy,0.39075000000000004
7
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,ImdbClassification,accuracy,0.586696
8
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,MassiveIntentClassification,accuracy,0.6140887693342301
9
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,MassiveScenarioClassification,accuracy,0.6973772696704774
10
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,MTOPDomainClassification,accuracy,0.8695622435020519
11
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,MTOPIntentClassification,accuracy,0.6225262197902417
12
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,ToxicConversationsClassification,accuracy,0.6265799999999999
13
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,TweetSentimentExtractionClassification,accuracy,0.5240803621958121
14
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,ArxivClusteringP2P,v_measure,0.3970858340673288
15
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,ArxivClusteringS2S,v_measure,0.2824284771372105
16
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,BiorxivClusteringP2P,v_measure,0.33632603955439844
17
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,BiorxivClusteringS2S,v_measure,0.27038042665369927
18
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,MedrxivClusteringP2P,v_measure,0.31374938993074253
19
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,MedrxivClusteringS2S,v_measure,0.26871455379644094
20
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,RedditClustering,v_measure,0.4023390747226228
21
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,RedditClusteringP2P,v_measure,0.49090518272935624
22
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,StackExchangeClustering,v_measure,0.5274481093815175
23
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,StackExchangeClusteringP2P,v_measure,0.3265999453562101
24
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,TwentyNewsgroupsClustering,v_measure,0.3212697126747911
25
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,SprintDuplicateQuestions,ap,0.8988577913120002
26
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,TwitterSemEval2015,ap,0.5474680676121269
27
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,TwitterURLCorpus,ap,0.8105760818661524
28
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,AskUbuntuDupQuestions,map,0.5583700395192394
29
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,MindSmallReranking,map,0.30402396942935334
30
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,SciDocsRR,map,0.7133941904192648
31
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,StackOverflowDupQuestions,map,0.44744984645554653
32
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,ArguAna,ndcg_at_10,0.47481
33
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,ClimateFEVER,ndcg_at_10,0.21345
34
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,CQADupstackRetrieval,ndcg_at_10,0.2724866666666666
35
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,DBPedia,ndcg_at_10,0.22365
36
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,FEVER,ndcg_at_10,0.55509
37
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,FiQA2018,ndcg_at_10,0.21472
38
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,HotpotQA,ndcg_at_10,0.39335
39
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,MSMARCO,ndcg_at_10,0.54238
40
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,NFCorpus,ndcg_at_10,0.2303
41
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,NQ,ndcg_at_10,0.27858
42
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,QuoraRetrieval,ndcg_at_10,0.82688
43
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,SCIDOCS,ndcg_at_10,0.12677
44
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,SciFact,ndcg_at_10,0.57881
45
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,Touche2020,ndcg_at_10,0.21525
46
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,TRECCOVID,ndcg_at_10,0.57921
47
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,BIOSSES,cosine_spearman,0.7520954502580506
48
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,SICK-R,cosine_spearman,0.7520954502580506
49
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,STS12,cosine_spearman,0.6652980061546658
50
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,STS13,cosine_spearman,0.7616628863742361
51
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,STS14,cosine_spearman,0.6904572664009687
52
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,STS15,cosine_spearman,0.7923677712825851
53
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,STS16,cosine_spearman,0.7606792422441928
54
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,STS17,cosine_spearman,0.8495398260629698
55
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,STS22,cosine_spearman,0.6566183708171826
56
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,STSBenchmark,cosine_spearman,0.7533716094627373
57
+ SGPT-125M-weightedmean-msmarco-specb-bitfit,SummEval,cosine_spearman,0.2890145030911965
SGPT-125M-weightedmean-nli-bitfit_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.6588059701492537
3
- AmazonPolarityClassification,accuracy,0.74938225
4
- AmazonReviewsClassification,accuracy,0.35098
5
- Banking77Classification,accuracy,0.7467857142857142
6
- EmotionClassification,accuracy,0.42225
7
- ImdbClassification,accuracy,0.62896
8
- MassiveIntentClassification,accuracy,0.5808002689979825
9
- MassiveScenarioClassification,accuracy,0.6633826496301277
10
- MTOPDomainClassification,accuracy,0.8151846785225718
11
- MTOPIntentClassification,accuracy,0.582421340629275
12
- ToxicConversationsClassification,accuracy,0.627862
13
- TweetSentimentExtractionClassification,accuracy,0.5482173174872665
14
- ArxivClusteringP2P,v_measure,0.3474248247787077
15
- ArxivClusteringS2S,v_measure,0.2467870651472156
16
- BiorxivClusteringP2P,v_measure,0.2893427045246491
17
- BiorxivClusteringS2S,v_measure,0.23080939123955474
18
- MedrxivClusteringP2P,v_measure,0.28301902023313874
19
- MedrxivClusteringS2S,v_measure,0.24932123582259286
20
- RedditClustering,v_measure,0.3375741018380938
21
- RedditClusteringP2P,v_measure,0.7765601981302583
22
- StackExchangeClustering,v_measure,0.4459127540530939
23
- StackExchangeClusteringP2P,v_measure,0.28230204578753637
24
- TwentyNewsgroupsClustering,v_measure,0.2824295128553035
25
- SprintDuplicateQuestions,ap,0.7773085502917281
26
- TwitterSemEval2015,ap,0.5709409536692154
27
- TwitterURLCorpus,ap,0.805106377126291
28
- AskUbuntuDupQuestions,map,0.5263439984994702
29
- MindSmallReranking,map,0.2926934104146833
30
- SciDocsRR,map,0.6835710819755543
31
- StackOverflowDupQuestions,map,0.3996520488022785
32
- ArguAna,ndcg_at_10,0.3104
33
- ClimateFEVER,ndcg_at_10,0.11007
34
- CQADupstackRetrieval,ndcg_at_10,0.20290166666666667
35
- DBPedia,ndcg_at_10,0.10868
36
- FEVER,ndcg_at_10,0.18399
37
- FiQA2018,ndcg_at_10,0.0894
38
- HotpotQA,ndcg_at_10,0.17727
39
- MSMARCO,ndcg_at_10,0.18159
40
- NFCorpus,ndcg_at_10,0.11805
41
- NQ,ndcg_at_10,0.07626
42
- QuoraRetrieval,ndcg_at_10,0.7896
43
- SCIDOCS,ndcg_at_10,0.07133
44
- SciFact,ndcg_at_10,0.31787
45
- Touche2020,ndcg_at_10,0.1227
46
- TRECCOVID,ndcg_at_10,0.3931
47
- BIOSSES,cosine_spearman,0.7092812216947605
48
- SICK-R,cosine_spearman,0.7092812216947605
49
- STS12,cosine_spearman,0.6916914010333395
50
- STS13,cosine_spearman,0.7723446219694267
51
- STS14,cosine_spearman,0.7098892671568665
52
- STS15,cosine_spearman,0.7973794782598049
53
- STS16,cosine_spearman,0.7793230986462234
54
- STS17,cosine_spearman,0.8732720847297224
55
- STS22,cosine_spearman,0.5963676007113087
56
- STSBenchmark,cosine_spearman,0.7954269211027273
57
- SummEval,cosine_spearman,0.30256847004390486
 
1
+ model,dataset,metric,value
2
+ SGPT-125M-weightedmean-nli-bitfit,AmazonCounterfactualClassification,accuracy,0.6588059701492537
3
+ SGPT-125M-weightedmean-nli-bitfit,AmazonPolarityClassification,accuracy,0.74938225
4
+ SGPT-125M-weightedmean-nli-bitfit,AmazonReviewsClassification,accuracy,0.35098
5
+ SGPT-125M-weightedmean-nli-bitfit,Banking77Classification,accuracy,0.7467857142857142
6
+ SGPT-125M-weightedmean-nli-bitfit,EmotionClassification,accuracy,0.42225
7
+ SGPT-125M-weightedmean-nli-bitfit,ImdbClassification,accuracy,0.62896
8
+ SGPT-125M-weightedmean-nli-bitfit,MassiveIntentClassification,accuracy,0.5808002689979825
9
+ SGPT-125M-weightedmean-nli-bitfit,MassiveScenarioClassification,accuracy,0.6633826496301277
10
+ SGPT-125M-weightedmean-nli-bitfit,MTOPDomainClassification,accuracy,0.8151846785225718
11
+ SGPT-125M-weightedmean-nli-bitfit,MTOPIntentClassification,accuracy,0.582421340629275
12
+ SGPT-125M-weightedmean-nli-bitfit,ToxicConversationsClassification,accuracy,0.627862
13
+ SGPT-125M-weightedmean-nli-bitfit,TweetSentimentExtractionClassification,accuracy,0.5482173174872665
14
+ SGPT-125M-weightedmean-nli-bitfit,ArxivClusteringP2P,v_measure,0.3474248247787077
15
+ SGPT-125M-weightedmean-nli-bitfit,ArxivClusteringS2S,v_measure,0.2467870651472156
16
+ SGPT-125M-weightedmean-nli-bitfit,BiorxivClusteringP2P,v_measure,0.2893427045246491
17
+ SGPT-125M-weightedmean-nli-bitfit,BiorxivClusteringS2S,v_measure,0.23080939123955474
18
+ SGPT-125M-weightedmean-nli-bitfit,MedrxivClusteringP2P,v_measure,0.28301902023313874
19
+ SGPT-125M-weightedmean-nli-bitfit,MedrxivClusteringS2S,v_measure,0.24932123582259286
20
+ SGPT-125M-weightedmean-nli-bitfit,RedditClustering,v_measure,0.3375741018380938
21
+ SGPT-125M-weightedmean-nli-bitfit,RedditClusteringP2P,v_measure,0.7765601981302583
22
+ SGPT-125M-weightedmean-nli-bitfit,StackExchangeClustering,v_measure,0.4459127540530939
23
+ SGPT-125M-weightedmean-nli-bitfit,StackExchangeClusteringP2P,v_measure,0.28230204578753637
24
+ SGPT-125M-weightedmean-nli-bitfit,TwentyNewsgroupsClustering,v_measure,0.2824295128553035
25
+ SGPT-125M-weightedmean-nli-bitfit,SprintDuplicateQuestions,ap,0.7773085502917281
26
+ SGPT-125M-weightedmean-nli-bitfit,TwitterSemEval2015,ap,0.5709409536692154
27
+ SGPT-125M-weightedmean-nli-bitfit,TwitterURLCorpus,ap,0.805106377126291
28
+ SGPT-125M-weightedmean-nli-bitfit,AskUbuntuDupQuestions,map,0.5263439984994702
29
+ SGPT-125M-weightedmean-nli-bitfit,MindSmallReranking,map,0.2926934104146833
30
+ SGPT-125M-weightedmean-nli-bitfit,SciDocsRR,map,0.6835710819755543
31
+ SGPT-125M-weightedmean-nli-bitfit,StackOverflowDupQuestions,map,0.3996520488022785
32
+ SGPT-125M-weightedmean-nli-bitfit,ArguAna,ndcg_at_10,0.3104
33
+ SGPT-125M-weightedmean-nli-bitfit,ClimateFEVER,ndcg_at_10,0.11007
34
+ SGPT-125M-weightedmean-nli-bitfit,CQADupstackRetrieval,ndcg_at_10,0.20290166666666667
35
+ SGPT-125M-weightedmean-nli-bitfit,DBPedia,ndcg_at_10,0.10868
36
+ SGPT-125M-weightedmean-nli-bitfit,FEVER,ndcg_at_10,0.18399
37
+ SGPT-125M-weightedmean-nli-bitfit,FiQA2018,ndcg_at_10,0.0894
38
+ SGPT-125M-weightedmean-nli-bitfit,HotpotQA,ndcg_at_10,0.17727
39
+ SGPT-125M-weightedmean-nli-bitfit,MSMARCO,ndcg_at_10,0.18159
40
+ SGPT-125M-weightedmean-nli-bitfit,NFCorpus,ndcg_at_10,0.11805
41
+ SGPT-125M-weightedmean-nli-bitfit,NQ,ndcg_at_10,0.07626
42
+ SGPT-125M-weightedmean-nli-bitfit,QuoraRetrieval,ndcg_at_10,0.7896
43
+ SGPT-125M-weightedmean-nli-bitfit,SCIDOCS,ndcg_at_10,0.07133
44
+ SGPT-125M-weightedmean-nli-bitfit,SciFact,ndcg_at_10,0.31787
45
+ SGPT-125M-weightedmean-nli-bitfit,Touche2020,ndcg_at_10,0.1227
46
+ SGPT-125M-weightedmean-nli-bitfit,TRECCOVID,ndcg_at_10,0.3931
47
+ SGPT-125M-weightedmean-nli-bitfit,BIOSSES,cosine_spearman,0.7092812216947605
48
+ SGPT-125M-weightedmean-nli-bitfit,SICK-R,cosine_spearman,0.7092812216947605
49
+ SGPT-125M-weightedmean-nli-bitfit,STS12,cosine_spearman,0.6916914010333395
50
+ SGPT-125M-weightedmean-nli-bitfit,STS13,cosine_spearman,0.7723446219694267
51
+ SGPT-125M-weightedmean-nli-bitfit,STS14,cosine_spearman,0.7098892671568665
52
+ SGPT-125M-weightedmean-nli-bitfit,STS15,cosine_spearman,0.7973794782598049
53
+ SGPT-125M-weightedmean-nli-bitfit,STS16,cosine_spearman,0.7793230986462234
54
+ SGPT-125M-weightedmean-nli-bitfit,STS17,cosine_spearman,0.8732720847297224
55
+ SGPT-125M-weightedmean-nli-bitfit,STS22,cosine_spearman,0.5963676007113087
56
+ SGPT-125M-weightedmean-nli-bitfit,STSBenchmark,cosine_spearman,0.7954269211027273
57
+ SGPT-125M-weightedmean-nli-bitfit,SummEval,cosine_spearman,0.30256847004390486
SGPT-5.8B-weightedmean-msmarco-specb-bitfit_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.6922388059701492
3
- AmazonPolarityClassification,accuracy,0.7126109999999999
4
- AmazonReviewsClassification,accuracy,0.39191999999999994
5
- Banking77Classification,accuracy,0.8449350649350649
6
- EmotionClassification,accuracy,0.49655000000000005
7
- ImdbClassification,accuracy,0.666376
8
- MassiveIntentClassification,accuracy,0.703866845998655
9
- MassiveScenarioClassification,accuracy,0.7627774041694687
10
- MTOPDomainClassification,accuracy,0.9346557227542178
11
- MTOPIntentClassification,accuracy,0.7242134062927497
12
- ToxicConversationsClassification,accuracy,0.67709
13
- TweetSentimentExtractionClassification,accuracy,0.5685059422750425
14
- ArxivClusteringP2P,v_measure,0.4559037428592033
15
- ArxivClusteringS2S,v_measure,0.3886371701986363
16
- BiorxivClusteringP2P,v_measure,0.3655145972298938
17
- BiorxivClusteringS2S,v_measure,0.3369901851846774
18
- MedrxivClusteringP2P,v_measure,0.3151174592577334
19
- MedrxivClusteringS2S,v_measure,0.28764235987575365
20
- RedditClustering,v_measure,0.4045148482612238
21
- RedditClusteringP2P,v_measure,0.5574949067303913
22
- StackExchangeClustering,v_measure,0.5920812266121527
23
- StackExchangeClusteringP2P,v_measure,0.33954248554638056
24
- TwentyNewsgroupsClustering,v_measure,0.3946372298609047
25
- SprintDuplicateQuestions,ap,0.9384024096781063
26
- TwitterSemEval2015,ap,0.6686677647503386
27
- TwitterURLCorpus,ap,0.8528960532524223
28
- AskUbuntuDupQuestions,map,0.6162556869142777
29
- MindSmallReranking,map,
30
- SciDocsRR,map,0.8078825425914723
31
- StackOverflowDupQuestions,map,0.5152800990025549
32
- ArguAna,ndcg_at_10,0.52203
33
- ClimateFEVER,ndcg_at_10,
34
- CQADupstackRetrieval,ndcg_at_10,0.3940316666666666
35
- DBPedia,ndcg_at_10,
36
- FEVER,ndcg_at_10,
37
- FiQA2018,ndcg_at_10,0.37463
38
- HotpotQA,ndcg_at_10,
39
- MSMARCO,ndcg_at_10,
40
- NFCorpus,ndcg_at_10,0.36161
41
- NQ,ndcg_at_10,
42
- QuoraRetrieval,ndcg_at_10,0.87899
43
- SCIDOCS,ndcg_at_10,0.19899
44
- SciFact,ndcg_at_10,0.74485
45
- Touche2020,ndcg_at_10,
46
- TRECCOVID,ndcg_at_10,0.77592
47
- BIOSSES,cosine_spearman,0.862501580394454
48
- SICK-R,cosine_spearman,0.862501580394454
49
- STS12,cosine_spearman,0.675020551515597
50
- STS13,cosine_spearman,0.7916333501772059
51
- STS14,cosine_spearman,0.7446048326701329
52
- STS15,cosine_spearman,0.8447194637929274
53
- STS16,cosine_spearman,0.8096089203722137
54
- STS17,cosine_spearman,0.8778202647220289
55
- STS22,cosine_spearman,0.6934889515492327
56
- STSBenchmark,cosine_spearman,0.8138977341532744
57
- SummEval,cosine_spearman,0.24747448399760644
 
1
+ model,dataset,metric,value
2
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,AmazonCounterfactualClassification,accuracy,0.6922388059701492
3
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,AmazonPolarityClassification,accuracy,0.7126109999999999
4
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,AmazonReviewsClassification,accuracy,0.39191999999999994
5
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,Banking77Classification,accuracy,0.8449350649350649
6
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,EmotionClassification,accuracy,0.49655000000000005
7
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ImdbClassification,accuracy,0.666376
8
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MassiveIntentClassification,accuracy,0.703866845998655
9
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MassiveScenarioClassification,accuracy,0.7627774041694687
10
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MTOPDomainClassification,accuracy,0.9346557227542178
11
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MTOPIntentClassification,accuracy,0.7242134062927497
12
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ToxicConversationsClassification,accuracy,0.67709
13
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,TweetSentimentExtractionClassification,accuracy,0.5685059422750425
14
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ArxivClusteringP2P,v_measure,0.4559037428592033
15
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ArxivClusteringS2S,v_measure,0.3886371701986363
16
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,BiorxivClusteringP2P,v_measure,0.3655145972298938
17
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,BiorxivClusteringS2S,v_measure,0.3369901851846774
18
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MedrxivClusteringP2P,v_measure,0.3151174592577334
19
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MedrxivClusteringS2S,v_measure,0.28764235987575365
20
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,RedditClustering,v_measure,0.4045148482612238
21
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,RedditClusteringP2P,v_measure,0.5574949067303913
22
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,StackExchangeClustering,v_measure,0.5920812266121527
23
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,StackExchangeClusteringP2P,v_measure,0.33954248554638056
24
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,TwentyNewsgroupsClustering,v_measure,0.3946372298609047
25
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SprintDuplicateQuestions,ap,0.9384024096781063
26
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,TwitterSemEval2015,ap,0.6686677647503386
27
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,TwitterURLCorpus,ap,0.8528960532524223
28
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,AskUbuntuDupQuestions,map,0.6162556869142777
29
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MindSmallReranking,map,
30
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SciDocsRR,map,0.8078825425914723
31
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,StackOverflowDupQuestions,map,0.5152800990025549
32
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ArguAna,ndcg_at_10,0.52203
33
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ClimateFEVER,ndcg_at_10,
34
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,CQADupstackRetrieval,ndcg_at_10,0.3940316666666666
35
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,DBPedia,ndcg_at_10,
36
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,FEVER,ndcg_at_10,
37
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,FiQA2018,ndcg_at_10,0.37463
38
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,HotpotQA,ndcg_at_10,
39
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MSMARCO,ndcg_at_10,
40
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,NFCorpus,ndcg_at_10,0.36161
41
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,NQ,ndcg_at_10,
42
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,QuoraRetrieval,ndcg_at_10,0.87899
43
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SCIDOCS,ndcg_at_10,0.19899
44
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SciFact,ndcg_at_10,0.74485
45
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,Touche2020,ndcg_at_10,
46
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,TRECCOVID,ndcg_at_10,0.77592
47
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,BIOSSES,cosine_spearman,0.862501580394454
48
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SICK-R,cosine_spearman,0.862501580394454
49
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS12,cosine_spearman,0.675020551515597
50
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS13,cosine_spearman,0.7916333501772059
51
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS14,cosine_spearman,0.7446048326701329
52
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS15,cosine_spearman,0.8447194637929274
53
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS16,cosine_spearman,0.8096089203722137
54
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS17,cosine_spearman,0.8778202647220289
55
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS22,cosine_spearman,0.6934889515492327
56
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STSBenchmark,cosine_spearman,0.8138977341532744
57
+ SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SummEval,cosine_spearman,0.24747448399760644
SGPT-5.8B-weightedmean-nli-bitfit_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.7407462686567164
3
- AmazonPolarityClassification,accuracy,0.8230920000000002
4
- AmazonReviewsClassification,accuracy,0.41584000000000004
5
- Banking77Classification,accuracy,0.8174025974025974
6
- EmotionClassification,accuracy,0.4991999999999999
7
- ImdbClassification,accuracy,0.7432920000000001
8
- MassiveIntentClassification,accuracy,0.7000336247478144
9
- MassiveScenarioClassification,accuracy,0.7502689979825151
10
- MTOPDomainClassification,accuracy,0.896374829001368
11
- MTOPIntentClassification,accuracy,0.7068171454628362
12
- ToxicConversationsClassification,accuracy,0.6993339999999999
13
- TweetSentimentExtractionClassification,accuracy,0.6243916242218449
14
- ArxivClusteringP2P,v_measure,0.40553923271901693
15
- ArxivClusteringS2S,v_measure,0.3249323183712211
16
- BiorxivClusteringP2P,v_measure,0.3359451202614059
17
- BiorxivClusteringS2S,v_measure,0.29128241446157166
18
- MedrxivClusteringP2P,v_measure,0.30327566856300814
19
- MedrxivClusteringS2S,v_measure,0.2801650210863619
20
- RedditClustering,v_measure,0.42171313610410677
21
- RedditClusteringP2P,v_measure,0.4801815621479994
22
- StackExchangeClustering,v_measure,0.5413314692311623
23
- StackExchangeClusteringP2P,v_measure,0.31115181648287144
24
- TwentyNewsgroupsClustering,v_measure,0.37202082549859794
25
- SprintDuplicateQuestions,ap,0.8053523657342799
26
- TwitterSemEval2015,ap,0.6599787692764192
27
- TwitterURLCorpus,ap,0.8454288910204419
28
- AskUbuntuDupQuestions,map,0.5589811361443445
29
- MindSmallReranking,map,
30
- SciDocsRR,map,0.7753549990038017
31
- StackOverflowDupQuestions,map,0.44771112666694335
32
- ArguAna,ndcg_at_10,0.3507
33
- ClimateFEVER,ndcg_at_10,0.17574
34
- CQADupstackRetrieval,ndcg_at_10,0.29978333333333335
35
- DBPedia,ndcg_at_10,0.26098
36
- FEVER,ndcg_at_10,0.38644
37
- FiQA2018,ndcg_at_10,0.18591
38
- HotpotQA,ndcg_at_10,0.33995
39
- MSMARCO,ndcg_at_10,
40
- NFCorpus,ndcg_at_10,0.28256
41
- NQ,ndcg_at_10,0.24628
42
- QuoraRetrieval,ndcg_at_10,0.8468
43
- SCIDOCS,ndcg_at_10,0.13554
44
- SciFact,ndcg_at_10,0.46662
45
- Touche2020,ndcg_at_10,0.16178
46
- TRECCOVID,ndcg_at_10,0.55353
47
- BIOSSES,cosine_spearman,0.7950000423261177
48
- SICK-R,cosine_spearman,0.7958562467776268
49
- STS12,cosine_spearman,0.7428502198729446
50
- STS13,cosine_spearman,0.8535148434923192
51
- STS14,cosine_spearman,0.7920612995350826
52
- STS15,cosine_spearman,0.8552171407629341
53
- STS16,cosine_spearman,0.8253853238521991
54
- STS17,cosine_spearman,0.904389369935707
55
- STS22,cosine_spearman,0.6319719780439462
56
- STSBenchmark,cosine_spearman,0.8566816193002651
57
- SummEval,cosine_spearman,0.30384175038360195
 
1
+ model,dataset,metric,value
2
+ SGPT-5.8B-weightedmean-nli-bitfit,AmazonCounterfactualClassification,accuracy,0.7407462686567164
3
+ SGPT-5.8B-weightedmean-nli-bitfit,AmazonPolarityClassification,accuracy,0.8230920000000002
4
+ SGPT-5.8B-weightedmean-nli-bitfit,AmazonReviewsClassification,accuracy,0.41584000000000004
5
+ SGPT-5.8B-weightedmean-nli-bitfit,Banking77Classification,accuracy,0.8174025974025974
6
+ SGPT-5.8B-weightedmean-nli-bitfit,EmotionClassification,accuracy,0.4991999999999999
7
+ SGPT-5.8B-weightedmean-nli-bitfit,ImdbClassification,accuracy,0.7432920000000001
8
+ SGPT-5.8B-weightedmean-nli-bitfit,MassiveIntentClassification,accuracy,0.7000336247478144
9
+ SGPT-5.8B-weightedmean-nli-bitfit,MassiveScenarioClassification,accuracy,0.7502689979825151
10
+ SGPT-5.8B-weightedmean-nli-bitfit,MTOPDomainClassification,accuracy,0.896374829001368
11
+ SGPT-5.8B-weightedmean-nli-bitfit,MTOPIntentClassification,accuracy,0.7068171454628362
12
+ SGPT-5.8B-weightedmean-nli-bitfit,ToxicConversationsClassification,accuracy,0.6993339999999999
13
+ SGPT-5.8B-weightedmean-nli-bitfit,TweetSentimentExtractionClassification,accuracy,0.6243916242218449
14
+ SGPT-5.8B-weightedmean-nli-bitfit,ArxivClusteringP2P,v_measure,0.40553923271901693
15
+ SGPT-5.8B-weightedmean-nli-bitfit,ArxivClusteringS2S,v_measure,0.3249323183712211
16
+ SGPT-5.8B-weightedmean-nli-bitfit,BiorxivClusteringP2P,v_measure,0.3359451202614059
17
+ SGPT-5.8B-weightedmean-nli-bitfit,BiorxivClusteringS2S,v_measure,0.29128241446157166
18
+ SGPT-5.8B-weightedmean-nli-bitfit,MedrxivClusteringP2P,v_measure,0.30327566856300814
19
+ SGPT-5.8B-weightedmean-nli-bitfit,MedrxivClusteringS2S,v_measure,0.2801650210863619
20
+ SGPT-5.8B-weightedmean-nli-bitfit,RedditClustering,v_measure,0.42171313610410677
21
+ SGPT-5.8B-weightedmean-nli-bitfit,RedditClusteringP2P,v_measure,0.4801815621479994
22
+ SGPT-5.8B-weightedmean-nli-bitfit,StackExchangeClustering,v_measure,0.5413314692311623
23
+ SGPT-5.8B-weightedmean-nli-bitfit,StackExchangeClusteringP2P,v_measure,0.31115181648287144
24
+ SGPT-5.8B-weightedmean-nli-bitfit,TwentyNewsgroupsClustering,v_measure,0.37202082549859794
25
+ SGPT-5.8B-weightedmean-nli-bitfit,SprintDuplicateQuestions,ap,0.8053523657342799
26
+ SGPT-5.8B-weightedmean-nli-bitfit,TwitterSemEval2015,ap,0.6599787692764192
27
+ SGPT-5.8B-weightedmean-nli-bitfit,TwitterURLCorpus,ap,0.8454288910204419
28
+ SGPT-5.8B-weightedmean-nli-bitfit,AskUbuntuDupQuestions,map,0.5589811361443445
29
+ SGPT-5.8B-weightedmean-nli-bitfit,MindSmallReranking,map,
30
+ SGPT-5.8B-weightedmean-nli-bitfit,SciDocsRR,map,0.7753549990038017
31
+ SGPT-5.8B-weightedmean-nli-bitfit,StackOverflowDupQuestions,map,0.44771112666694335
32
+ SGPT-5.8B-weightedmean-nli-bitfit,ArguAna,ndcg_at_10,0.3507
33
+ SGPT-5.8B-weightedmean-nli-bitfit,ClimateFEVER,ndcg_at_10,0.17574
34
+ SGPT-5.8B-weightedmean-nli-bitfit,CQADupstackRetrieval,ndcg_at_10,0.29978333333333335
35
+ SGPT-5.8B-weightedmean-nli-bitfit,DBPedia,ndcg_at_10,0.26098
36
+ SGPT-5.8B-weightedmean-nli-bitfit,FEVER,ndcg_at_10,0.38644
37
+ SGPT-5.8B-weightedmean-nli-bitfit,FiQA2018,ndcg_at_10,0.18591
38
+ SGPT-5.8B-weightedmean-nli-bitfit,HotpotQA,ndcg_at_10,0.33995
39
+ SGPT-5.8B-weightedmean-nli-bitfit,MSMARCO,ndcg_at_10,
40
+ SGPT-5.8B-weightedmean-nli-bitfit,NFCorpus,ndcg_at_10,0.28256
41
+ SGPT-5.8B-weightedmean-nli-bitfit,NQ,ndcg_at_10,0.24628
42
+ SGPT-5.8B-weightedmean-nli-bitfit,QuoraRetrieval,ndcg_at_10,0.8468
43
+ SGPT-5.8B-weightedmean-nli-bitfit,SCIDOCS,ndcg_at_10,0.13554
44
+ SGPT-5.8B-weightedmean-nli-bitfit,SciFact,ndcg_at_10,0.46662
45
+ SGPT-5.8B-weightedmean-nli-bitfit,Touche2020,ndcg_at_10,0.16178
46
+ SGPT-5.8B-weightedmean-nli-bitfit,TRECCOVID,ndcg_at_10,0.55353
47
+ SGPT-5.8B-weightedmean-nli-bitfit,BIOSSES,cosine_spearman,0.7950000423261177
48
+ SGPT-5.8B-weightedmean-nli-bitfit,SICK-R,cosine_spearman,0.7958562467776268
49
+ SGPT-5.8B-weightedmean-nli-bitfit,STS12,cosine_spearman,0.7428502198729446
50
+ SGPT-5.8B-weightedmean-nli-bitfit,STS13,cosine_spearman,0.8535148434923192
51
+ SGPT-5.8B-weightedmean-nli-bitfit,STS14,cosine_spearman,0.7920612995350826
52
+ SGPT-5.8B-weightedmean-nli-bitfit,STS15,cosine_spearman,0.8552171407629341
53
+ SGPT-5.8B-weightedmean-nli-bitfit,STS16,cosine_spearman,0.8253853238521991
54
+ SGPT-5.8B-weightedmean-nli-bitfit,STS17,cosine_spearman,0.904389369935707
55
+ SGPT-5.8B-weightedmean-nli-bitfit,STS22,cosine_spearman,0.6319719780439462
56
+ SGPT-5.8B-weightedmean-nli-bitfit,STSBenchmark,cosine_spearman,0.8566816193002651
57
+ SGPT-5.8B-weightedmean-nli-bitfit,SummEval,cosine_spearman,0.30384175038360195
all-MiniLM-L6-v2_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.641492537313433
3
- AmazonPolarityClassification,accuracy,0.62582975
4
- AmazonReviewsClassification,accuracy,0.31786
5
- Banking77Classification,accuracy,0.7975000000000001
6
- EmotionClassification,accuracy,0.3843
7
- ImdbClassification,accuracy,0.606612
8
- MassiveIntentClassification,accuracy,0.6740416946872899
9
- MassiveScenarioClassification,accuracy,0.7575655682582381
10
- MTOPDomainClassification,accuracy,0.9156178750569997
11
- MTOPIntentClassification,accuracy,0.6218194254445966
12
- ToxicConversationsClassification,accuracy,0.669918
13
- TweetSentimentExtractionClassification,accuracy,0.554102999434069
14
- ArxivClusteringP2P,v_measure,0.46545950790501556
15
- ArxivClusteringS2S,v_measure,0.3785709823840442
16
- BiorxivClusteringP2P,v_measure,0.38483019141351227
17
- BiorxivClusteringS2S,v_measure,0.33170209943399803
18
- MedrxivClusteringP2P,v_measure,0.3440873490143895
19
- MedrxivClusteringS2S,v_measure,0.3229220750053091
20
- RedditClustering,v_measure,0.5066969274980475
21
- RedditClusteringP2P,v_measure,0.8150529020581544
22
- StackExchangeClustering,v_measure,0.5336142166203601
23
- StackExchangeClusteringP2P,v_measure,0.3800182562780098
24
- TwentyNewsgroupsClustering,v_measure,0.4686027142764777
25
- SprintDuplicateQuestions,ap,0.9455063045792446
26
- TwitterSemEval2015,ap,0.6785802440228593
27
- TwitterURLCorpus,ap,0.847016595145111
28
- AskUbuntuDupQuestions,map,0.634768168123733
29
- MindSmallReranking,map,0.30798042020200267
30
- SciDocsRR,map,0.8711941318470207
31
- StackOverflowDupQuestions,map,0.5076213438431608
32
- ArguAna,ndcg_at_10,0.50167
33
- ClimateFEVER,ndcg_at_10,0.20272
34
- CQADupstackRetrieval,ndcg_at_10,0.4132475
35
- DBPedia,ndcg_at_10,0.3233
36
- FEVER,ndcg_at_10,0.51934
37
- FiQA2018,ndcg_at_10,0.36867
38
- HotpotQA,ndcg_at_10,0.46513
39
- MSMARCO,ndcg_at_10,0.63685
40
- NFCorpus,ndcg_at_10,0.31594
41
- NQ,ndcg_at_10,0.43869
42
- QuoraRetrieval,ndcg_at_10,0.87556
43
- SCIDOCS,ndcg_at_10,0.21641
44
- SciFact,ndcg_at_10,0.64508
45
- Touche2020,ndcg_at_10,0.16904
46
- TRECCOVID,ndcg_at_10,0.47246
47
- BIOSSES,cosine_spearman,0.8164041444909368
48
- SICK-R,cosine_spearman,0.7758245130495687
49
- STS12,cosine_spearman,0.7236900735029991
50
- STS13,cosine_spearman,0.8060316722220763
51
- STS14,cosine_spearman,0.7558912800301661
52
- STS15,cosine_spearman,0.8538966051883823
53
- STS16,cosine_spearman,0.7898945343973262
54
- STS17,cosine_spearman,0.8758779089494525
55
- STS22,cosine_spearman,0.6721465212910986
56
- STSBenchmark,cosine_spearman,0.8203246731235654
57
- SummEval,cosine_spearman,0.3080875767124449
 
1
+ model,dataset,metric,value
2
+ all-MiniLM-L6-v2,AmazonCounterfactualClassification,accuracy,0.641492537313433
3
+ all-MiniLM-L6-v2,AmazonPolarityClassification,accuracy,0.62582975
4
+ all-MiniLM-L6-v2,AmazonReviewsClassification,accuracy,0.31786
5
+ all-MiniLM-L6-v2,Banking77Classification,accuracy,0.7975000000000001
6
+ all-MiniLM-L6-v2,EmotionClassification,accuracy,0.3843
7
+ all-MiniLM-L6-v2,ImdbClassification,accuracy,0.606612
8
+ all-MiniLM-L6-v2,MassiveIntentClassification,accuracy,0.6740416946872899
9
+ all-MiniLM-L6-v2,MassiveScenarioClassification,accuracy,0.7575655682582381
10
+ all-MiniLM-L6-v2,MTOPDomainClassification,accuracy,0.9156178750569997
11
+ all-MiniLM-L6-v2,MTOPIntentClassification,accuracy,0.6218194254445966
12
+ all-MiniLM-L6-v2,ToxicConversationsClassification,accuracy,0.669918
13
+ all-MiniLM-L6-v2,TweetSentimentExtractionClassification,accuracy,0.554102999434069
14
+ all-MiniLM-L6-v2,ArxivClusteringP2P,v_measure,0.46545950790501556
15
+ all-MiniLM-L6-v2,ArxivClusteringS2S,v_measure,0.3785709823840442
16
+ all-MiniLM-L6-v2,BiorxivClusteringP2P,v_measure,0.38483019141351227
17
+ all-MiniLM-L6-v2,BiorxivClusteringS2S,v_measure,0.33170209943399803
18
+ all-MiniLM-L6-v2,MedrxivClusteringP2P,v_measure,0.3440873490143895
19
+ all-MiniLM-L6-v2,MedrxivClusteringS2S,v_measure,0.3229220750053091
20
+ all-MiniLM-L6-v2,RedditClustering,v_measure,0.5066969274980475
21
+ all-MiniLM-L6-v2,RedditClusteringP2P,v_measure,0.8150529020581544
22
+ all-MiniLM-L6-v2,StackExchangeClustering,v_measure,0.5336142166203601
23
+ all-MiniLM-L6-v2,StackExchangeClusteringP2P,v_measure,0.3800182562780098
24
+ all-MiniLM-L6-v2,TwentyNewsgroupsClustering,v_measure,0.4686027142764777
25
+ all-MiniLM-L6-v2,SprintDuplicateQuestions,ap,0.9455063045792446
26
+ all-MiniLM-L6-v2,TwitterSemEval2015,ap,0.6785802440228593
27
+ all-MiniLM-L6-v2,TwitterURLCorpus,ap,0.847016595145111
28
+ all-MiniLM-L6-v2,AskUbuntuDupQuestions,map,0.634768168123733
29
+ all-MiniLM-L6-v2,MindSmallReranking,map,0.30798042020200267
30
+ all-MiniLM-L6-v2,SciDocsRR,map,0.8711941318470207
31
+ all-MiniLM-L6-v2,StackOverflowDupQuestions,map,0.5076213438431608
32
+ all-MiniLM-L6-v2,ArguAna,ndcg_at_10,0.50167
33
+ all-MiniLM-L6-v2,ClimateFEVER,ndcg_at_10,0.20272
34
+ all-MiniLM-L6-v2,CQADupstackRetrieval,ndcg_at_10,0.4132475
35
+ all-MiniLM-L6-v2,DBPedia,ndcg_at_10,0.3233
36
+ all-MiniLM-L6-v2,FEVER,ndcg_at_10,0.51934
37
+ all-MiniLM-L6-v2,FiQA2018,ndcg_at_10,0.36867
38
+ all-MiniLM-L6-v2,HotpotQA,ndcg_at_10,0.46513
39
+ all-MiniLM-L6-v2,MSMARCO,ndcg_at_10,0.63685
40
+ all-MiniLM-L6-v2,NFCorpus,ndcg_at_10,0.31594
41
+ all-MiniLM-L6-v2,NQ,ndcg_at_10,0.43869
42
+ all-MiniLM-L6-v2,QuoraRetrieval,ndcg_at_10,0.87556
43
+ all-MiniLM-L6-v2,SCIDOCS,ndcg_at_10,0.21641
44
+ all-MiniLM-L6-v2,SciFact,ndcg_at_10,0.64508
45
+ all-MiniLM-L6-v2,Touche2020,ndcg_at_10,0.16904
46
+ all-MiniLM-L6-v2,TRECCOVID,ndcg_at_10,0.47246
47
+ all-MiniLM-L6-v2,BIOSSES,cosine_spearman,0.8164041444909368
48
+ all-MiniLM-L6-v2,SICK-R,cosine_spearman,0.7758245130495687
49
+ all-MiniLM-L6-v2,STS12,cosine_spearman,0.7236900735029991
50
+ all-MiniLM-L6-v2,STS13,cosine_spearman,0.8060316722220763
51
+ all-MiniLM-L6-v2,STS14,cosine_spearman,0.7558912800301661
52
+ all-MiniLM-L6-v2,STS15,cosine_spearman,0.8538966051883823
53
+ all-MiniLM-L6-v2,STS16,cosine_spearman,0.7898945343973262
54
+ all-MiniLM-L6-v2,STS17,cosine_spearman,0.8758779089494525
55
+ all-MiniLM-L6-v2,STS22,cosine_spearman,0.6721465212910986
56
+ all-MiniLM-L6-v2,STSBenchmark,cosine_spearman,0.8203246731235654
57
+ all-MiniLM-L6-v2,SummEval,cosine_spearman,0.3080875767124449
all-mpnet-base-v2_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.652686567164179
3
- AmazonPolarityClassification,accuracy,0.6713145
4
- AmazonReviewsClassification,accuracy,0.31920000000000004
5
- Banking77Classification,accuracy,0.8186038961038961
6
- EmotionClassification,accuracy,0.39725
7
- ImdbClassification,accuracy,0.707156
8
- MassiveIntentClassification,accuracy,0.6956624075319435
9
- MassiveScenarioClassification,accuracy,0.7601210490921319
10
- MTOPDomainClassification,accuracy,0.9207934336525309
11
- MTOPIntentClassification,accuracy,0.7020975832193344
12
- ToxicConversationsClassification,accuracy,0.608612
13
- TweetSentimentExtractionClassification,accuracy,0.5546406338426711
14
- ArxivClusteringP2P,v_measure,0.48378863409223744
15
- ArxivClusteringS2S,v_measure,0.3972488615315985
16
- BiorxivClusteringP2P,v_measure,0.3961660513362518
17
- BiorxivClusteringS2S,v_measure,0.3502442407186902
18
- MedrxivClusteringP2P,v_measure,0.3558002813186373
19
- MedrxivClusteringS2S,v_measure,0.32872725562410443
20
- RedditClustering,v_measure,0.5482425669843732
21
- RedditClusteringP2P,v_measure,0.5676897267804937
22
- StackExchangeClustering,v_measure,0.5380048409076928
23
- StackExchangeClusteringP2P,v_measure,0.34280269334397545
24
- TwentyNewsgroupsClustering,v_measure,0.49744553480838094
25
- SprintDuplicateQuestions,ap,0.9014551416831161
26
- TwitterSemEval2015,ap,0.7385398650568216
27
- TwitterURLCorpus,ap,0.8510831084479727
28
- AskUbuntuDupQuestions,map,0.6585199009344481
29
- MindSmallReranking,map,0.3096534360486133
30
- SciDocsRR,map,0.8865396986895777
31
- StackOverflowDupQuestions,map,0.5197907654945493
32
- ArguAna,ndcg_at_10,0.46521
33
- ClimateFEVER,ndcg_at_10,0.21971
34
- CQADupstackRetrieval,ndcg_at_10,0.44959166666666667
35
- DBPedia,ndcg_at_10,0.32088
36
- FEVER,ndcg_at_10,0.50863
37
- FiQA2018,ndcg_at_10,0.49963
38
- HotpotQA,ndcg_at_10,0.39294
39
- MSMARCO,ndcg_at_10,0.66682
40
- NFCorpus,ndcg_at_10,0.33289
41
- NQ,ndcg_at_10,0.5045
42
- QuoraRetrieval,ndcg_at_10,0.87462
43
- SCIDOCS,ndcg_at_10,0.23765
44
- SciFact,ndcg_at_10,0.6557
45
- Touche2020,ndcg_at_10,0.19934
46
- TRECCOVID,ndcg_at_10,0.51326
47
- BIOSSES,cosine_spearman,0.8043195317854409
48
- SICK-R,cosine_spearman,0.8058541928440089
49
- STS12,cosine_spearman,0.726342785382245
50
- STS13,cosine_spearman,0.8348499016384896
51
- STS14,cosine_spearman,0.7800013713421681
52
- STS15,cosine_spearman,0.856627539164948
53
- STS16,cosine_spearman,0.8003011315645663
54
- STS17,cosine_spearman,0.9059523263123733
55
- STS22,cosine_spearman,0.6794534221542502
56
- STSBenchmark,cosine_spearman,0.83421898505618
57
- SummEval,cosine_spearman,0.27485869639926136
 
1
+ model,dataset,metric,value
2
+ all-mpnet-base-v2,AmazonCounterfactualClassification,accuracy,0.652686567164179
3
+ all-mpnet-base-v2,AmazonPolarityClassification,accuracy,0.6713145
4
+ all-mpnet-base-v2,AmazonReviewsClassification,accuracy,0.31920000000000004
5
+ all-mpnet-base-v2,Banking77Classification,accuracy,0.8186038961038961
6
+ all-mpnet-base-v2,EmotionClassification,accuracy,0.39725
7
+ all-mpnet-base-v2,ImdbClassification,accuracy,0.707156
8
+ all-mpnet-base-v2,MassiveIntentClassification,accuracy,0.6956624075319435
9
+ all-mpnet-base-v2,MassiveScenarioClassification,accuracy,0.7601210490921319
10
+ all-mpnet-base-v2,MTOPDomainClassification,accuracy,0.9207934336525309
11
+ all-mpnet-base-v2,MTOPIntentClassification,accuracy,0.7020975832193344
12
+ all-mpnet-base-v2,ToxicConversationsClassification,accuracy,0.608612
13
+ all-mpnet-base-v2,TweetSentimentExtractionClassification,accuracy,0.5546406338426711
14
+ all-mpnet-base-v2,ArxivClusteringP2P,v_measure,0.48378863409223744
15
+ all-mpnet-base-v2,ArxivClusteringS2S,v_measure,0.3972488615315985
16
+ all-mpnet-base-v2,BiorxivClusteringP2P,v_measure,0.3961660513362518
17
+ all-mpnet-base-v2,BiorxivClusteringS2S,v_measure,0.3502442407186902
18
+ all-mpnet-base-v2,MedrxivClusteringP2P,v_measure,0.3558002813186373
19
+ all-mpnet-base-v2,MedrxivClusteringS2S,v_measure,0.32872725562410443
20
+ all-mpnet-base-v2,RedditClustering,v_measure,0.5482425669843732
21
+ all-mpnet-base-v2,RedditClusteringP2P,v_measure,0.5676897267804937
22
+ all-mpnet-base-v2,StackExchangeClustering,v_measure,0.5380048409076928
23
+ all-mpnet-base-v2,StackExchangeClusteringP2P,v_measure,0.34280269334397545
24
+ all-mpnet-base-v2,TwentyNewsgroupsClustering,v_measure,0.49744553480838094
25
+ all-mpnet-base-v2,SprintDuplicateQuestions,ap,0.9014551416831161
26
+ all-mpnet-base-v2,TwitterSemEval2015,ap,0.7385398650568216
27
+ all-mpnet-base-v2,TwitterURLCorpus,ap,0.8510831084479727
28
+ all-mpnet-base-v2,AskUbuntuDupQuestions,map,0.6585199009344481
29
+ all-mpnet-base-v2,MindSmallReranking,map,0.3096534360486133
30
+ all-mpnet-base-v2,SciDocsRR,map,0.8865396986895777
31
+ all-mpnet-base-v2,StackOverflowDupQuestions,map,0.5197907654945493
32
+ all-mpnet-base-v2,ArguAna,ndcg_at_10,0.46521
33
+ all-mpnet-base-v2,ClimateFEVER,ndcg_at_10,0.21971
34
+ all-mpnet-base-v2,CQADupstackRetrieval,ndcg_at_10,0.44959166666666667
35
+ all-mpnet-base-v2,DBPedia,ndcg_at_10,0.32088
36
+ all-mpnet-base-v2,FEVER,ndcg_at_10,0.50863
37
+ all-mpnet-base-v2,FiQA2018,ndcg_at_10,0.49963
38
+ all-mpnet-base-v2,HotpotQA,ndcg_at_10,0.39294
39
+ all-mpnet-base-v2,MSMARCO,ndcg_at_10,0.66682
40
+ all-mpnet-base-v2,NFCorpus,ndcg_at_10,0.33289
41
+ all-mpnet-base-v2,NQ,ndcg_at_10,0.5045
42
+ all-mpnet-base-v2,QuoraRetrieval,ndcg_at_10,0.87462
43
+ all-mpnet-base-v2,SCIDOCS,ndcg_at_10,0.23765
44
+ all-mpnet-base-v2,SciFact,ndcg_at_10,0.6557
45
+ all-mpnet-base-v2,Touche2020,ndcg_at_10,0.19934
46
+ all-mpnet-base-v2,TRECCOVID,ndcg_at_10,0.51326
47
+ all-mpnet-base-v2,BIOSSES,cosine_spearman,0.8043195317854409
48
+ all-mpnet-base-v2,SICK-R,cosine_spearman,0.8058541928440089
49
+ all-mpnet-base-v2,STS12,cosine_spearman,0.726342785382245
50
+ all-mpnet-base-v2,STS13,cosine_spearman,0.8348499016384896
51
+ all-mpnet-base-v2,STS14,cosine_spearman,0.7800013713421681
52
+ all-mpnet-base-v2,STS15,cosine_spearman,0.856627539164948
53
+ all-mpnet-base-v2,STS16,cosine_spearman,0.8003011315645663
54
+ all-mpnet-base-v2,STS17,cosine_spearman,0.9059523263123733
55
+ all-mpnet-base-v2,STS22,cosine_spearman,0.6794534221542502
56
+ all-mpnet-base-v2,STSBenchmark,cosine_spearman,0.83421898505618
57
+ all-mpnet-base-v2,SummEval,cosine_spearman,0.27485869639926136
bert-base-uncased_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.7425373134328358
3
- AmazonPolarityClassification,accuracy,0.7132945
4
- AmazonReviewsClassification,accuracy,0.33564
5
- Banking77Classification,accuracy,0.6340584415584415
6
- EmotionClassification,accuracy,0.3528
7
- ImdbClassification,accuracy,0.653456
8
- MassiveIntentClassification,accuracy,0.5988231338264962
9
- MassiveScenarioClassification,accuracy,0.6427706792199059
10
- MTOPDomainClassification,accuracy,0.8262653898768809
11
- MTOPIntentClassification,accuracy,0.6813725490196079
12
- ToxicConversationsClassification,accuracy,0.6999679999999999
13
- TweetSentimentExtractionClassification,accuracy,0.5180814940577249
14
- ArxivClusteringP2P,v_measure,0.3518932830729758
15
- ArxivClusteringS2S,v_measure,0.2750822951507033
16
- BiorxivClusteringP2P,v_measure,0.3012278640553615
17
- BiorxivClusteringS2S,v_measure,0.2476601053047804
18
- MedrxivClusteringP2P,v_measure,0.26087688306606044
19
- MedrxivClusteringS2S,v_measure,0.23604914608225602
20
- RedditClustering,v_measure,0.27241718557705
21
- RedditClusteringP2P,v_measure,0.43323854460464056
22
- StackExchangeClustering,v_measure,0.4358261870547655
23
- StackExchangeClusteringP2P,v_measure,0.2654851675795123
24
- TwentyNewsgroupsClustering,v_measure,0.23354321403189832
25
- SprintDuplicateQuestions,ap,0.36808703728970593
26
- TwitterSemEval2015,ap,0.5589788752976392
27
- TwitterURLCorpus,ap,0.7628732783437441
28
- AskUbuntuDupQuestions,map,0.4584088706528255
29
- MindSmallReranking,map,0.28366637355845425
30
- SciDocsRR,map,0.6493728987830145
31
- StackOverflowDupQuestions,map,0.34615478798860166
32
- ArguAna,ndcg_at_10,0.28294
33
- ClimateFEVER,ndcg_at_10,0.0541
34
- CQADupstackRetrieval,ndcg_at_10,0.05506583333333333
35
- DBPedia,ndcg_at_10,0.04132
36
- FEVER,ndcg_at_10,0.033
37
- FiQA2018,ndcg_at_10,0.02191
38
- HotpotQA,ndcg_at_10,0.0826
39
- MSMARCO,ndcg_at_10,0.06176
40
- NFCorpus,ndcg_at_10,0.04304
41
- NQ,ndcg_at_10,0.02615
42
- QuoraRetrieval,ndcg_at_10,0.61029
43
- SCIDOCS,ndcg_at_10,0.02815
44
- SciFact,ndcg_at_10,0.13339
45
- Touche2020,ndcg_at_10,0.00967
46
- TRECCOVID,ndcg_at_10,0.14745
47
- BIOSSES,cosine_spearman,0.5469823428818151
48
- SICK-R,cosine_spearman,0.5864506948179484
49
- STS12,cosine_spearman,0.30871788407575457
50
- STS13,cosine_spearman,0.5989485045425808
51
- STS14,cosine_spearman,0.4772791352844941
52
- STS15,cosine_spearman,0.6028567383446486
53
- STS16,cosine_spearman,0.6373272693604403
54
- STS17,cosine_spearman,0.6410023907260639
55
- STS22,cosine_spearman,0.563667996535454
56
- STSBenchmark,cosine_spearman,0.4729108172174081
57
- SummEval,cosine_spearman,0.2981716355664815
 
1
+ model,dataset,metric,value
2
+ bert-base-uncased,AmazonCounterfactualClassification,accuracy,0.7425373134328358
3
+ bert-base-uncased,AmazonPolarityClassification,accuracy,0.7132945
4
+ bert-base-uncased,AmazonReviewsClassification,accuracy,0.33564
5
+ bert-base-uncased,Banking77Classification,accuracy,0.6340584415584415
6
+ bert-base-uncased,EmotionClassification,accuracy,0.3528
7
+ bert-base-uncased,ImdbClassification,accuracy,0.653456
8
+ bert-base-uncased,MassiveIntentClassification,accuracy,0.5988231338264962
9
+ bert-base-uncased,MassiveScenarioClassification,accuracy,0.6427706792199059
10
+ bert-base-uncased,MTOPDomainClassification,accuracy,0.8262653898768809
11
+ bert-base-uncased,MTOPIntentClassification,accuracy,0.6813725490196079
12
+ bert-base-uncased,ToxicConversationsClassification,accuracy,0.6999679999999999
13
+ bert-base-uncased,TweetSentimentExtractionClassification,accuracy,0.5180814940577249
14
+ bert-base-uncased,ArxivClusteringP2P,v_measure,0.3518932830729758
15
+ bert-base-uncased,ArxivClusteringS2S,v_measure,0.2750822951507033
16
+ bert-base-uncased,BiorxivClusteringP2P,v_measure,0.3012278640553615
17
+ bert-base-uncased,BiorxivClusteringS2S,v_measure,0.2476601053047804
18
+ bert-base-uncased,MedrxivClusteringP2P,v_measure,0.26087688306606044
19
+ bert-base-uncased,MedrxivClusteringS2S,v_measure,0.23604914608225602
20
+ bert-base-uncased,RedditClustering,v_measure,0.27241718557705
21
+ bert-base-uncased,RedditClusteringP2P,v_measure,0.43323854460464056
22
+ bert-base-uncased,StackExchangeClustering,v_measure,0.4358261870547655
23
+ bert-base-uncased,StackExchangeClusteringP2P,v_measure,0.2654851675795123
24
+ bert-base-uncased,TwentyNewsgroupsClustering,v_measure,0.23354321403189832
25
+ bert-base-uncased,SprintDuplicateQuestions,ap,0.36808703728970593
26
+ bert-base-uncased,TwitterSemEval2015,ap,0.5589788752976392
27
+ bert-base-uncased,TwitterURLCorpus,ap,0.7628732783437441
28
+ bert-base-uncased,AskUbuntuDupQuestions,map,0.4584088706528255
29
+ bert-base-uncased,MindSmallReranking,map,0.28366637355845425
30
+ bert-base-uncased,SciDocsRR,map,0.6493728987830145
31
+ bert-base-uncased,StackOverflowDupQuestions,map,0.34615478798860166
32
+ bert-base-uncased,ArguAna,ndcg_at_10,0.28294
33
+ bert-base-uncased,ClimateFEVER,ndcg_at_10,0.0541
34
+ bert-base-uncased,CQADupstackRetrieval,ndcg_at_10,0.05506583333333333
35
+ bert-base-uncased,DBPedia,ndcg_at_10,0.04132
36
+ bert-base-uncased,FEVER,ndcg_at_10,0.033
37
+ bert-base-uncased,FiQA2018,ndcg_at_10,0.02191
38
+ bert-base-uncased,HotpotQA,ndcg_at_10,0.0826
39
+ bert-base-uncased,MSMARCO,ndcg_at_10,0.06176
40
+ bert-base-uncased,NFCorpus,ndcg_at_10,0.04304
41
+ bert-base-uncased,NQ,ndcg_at_10,0.02615
42
+ bert-base-uncased,QuoraRetrieval,ndcg_at_10,0.61029
43
+ bert-base-uncased,SCIDOCS,ndcg_at_10,0.02815
44
+ bert-base-uncased,SciFact,ndcg_at_10,0.13339
45
+ bert-base-uncased,Touche2020,ndcg_at_10,0.00967
46
+ bert-base-uncased,TRECCOVID,ndcg_at_10,0.14745
47
+ bert-base-uncased,BIOSSES,cosine_spearman,0.5469823428818151
48
+ bert-base-uncased,SICK-R,cosine_spearman,0.5864506948179484
49
+ bert-base-uncased,STS12,cosine_spearman,0.30871788407575457
50
+ bert-base-uncased,STS13,cosine_spearman,0.5989485045425808
51
+ bert-base-uncased,STS14,cosine_spearman,0.4772791352844941
52
+ bert-base-uncased,STS15,cosine_spearman,0.6028567383446486
53
+ bert-base-uncased,STS16,cosine_spearman,0.6373272693604403
54
+ bert-base-uncased,STS17,cosine_spearman,0.6410023907260639
55
+ bert-base-uncased,STS22,cosine_spearman,0.563667996535454
56
+ bert-base-uncased,STSBenchmark,cosine_spearman,0.4729108172174081
57
+ bert-base-uncased,SummEval,cosine_spearman,0.2981716355664815
contriever-base-msmarco_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.7219402985074627
3
- AmazonPolarityClassification,accuracy,0.6863262499999999
4
- AmazonReviewsClassification,accuracy,0.37422
5
- Banking77Classification,accuracy,0.8001623376623377
6
- EmotionClassification,accuracy,0.44770000000000004
7
- ImdbClassification,accuracy,0.670388
8
- MassiveIntentClassification,accuracy,0.6777740416946872
9
- MassiveScenarioClassification,accuracy,0.7599865501008743
10
- MTOPDomainClassification,accuracy,0.9317829457364342
11
- MTOPIntentClassification,accuracy,0.6931372549019608
12
- ToxicConversationsClassification,accuracy,0.6776540000000001
13
- TweetSentimentExtractionClassification,accuracy,0.5610356536502547
14
- ArxivClusteringP2P,v_measure,0.42609650289899825
15
- ArxivClusteringS2S,v_measure,0.32317533856057756
16
- BiorxivClusteringP2P,v_measure,0.34974621074365847
17
- BiorxivClusteringS2S,v_measure,0.29083876572182377
18
- MedrxivClusteringP2P,v_measure,0.3118675108904548
19
- MedrxivClusteringS2S,v_measure,0.2727180592101203
20
- RedditClustering,v_measure,0.5489383973987073
21
- RedditClusteringP2P,v_measure,0.5757965400787036
22
- StackExchangeClustering,v_measure,0.6314901527677189
23
- StackExchangeClusteringP2P,v_measure,0.3225478381926452
24
- TwentyNewsgroupsClustering,v_measure,0.46822120585974714
25
- SprintDuplicateQuestions,ap,0.9554507015913917
26
- TwitterSemEval2015,ap,0.6684972822513366
27
- TwitterURLCorpus,ap,0.8520869965638479
28
- AskUbuntuDupQuestions,map,0.5668809109301288
29
- MindSmallReranking,map,0.3157849515126634
30
- SciDocsRR,map,0.7650730255518788
31
- StackOverflowDupQuestions,map,0.4778138013068855
32
- ArguAna,ndcg_at_10,0.4832
33
- ClimateFEVER,ndcg_at_10,0.24788
34
- CQADupstackRetrieval,ndcg_at_10,0.3366891666666667
35
- DBPedia,ndcg_at_10,0.38103
36
- FEVER,ndcg_at_10,0.59294
37
- FiQA2018,ndcg_at_10,0.2742
38
- HotpotQA,ndcg_at_10,0.56808
39
- MSMARCO,ndcg_at_10,0.62494
40
- NFCorpus,ndcg_at_10,0.31315
41
- NQ,ndcg_at_10,0.41833
42
- QuoraRetrieval,ndcg_at_10,0.86719
43
- SCIDOCS,ndcg_at_10,0.17125
44
- SciFact,ndcg_at_10,0.6551
45
- Touche2020,ndcg_at_10,0.15789
46
- TRECCOVID,ndcg_at_10,0.44769
47
- BIOSSES,cosine_spearman,0.8331816352111794
48
- SICK-R,cosine_spearman,0.702003740199912
49
- STS12,cosine_spearman,0.643399816144592
50
- STS13,cosine_spearman,0.8003351945178154
51
- STS14,cosine_spearman,0.7450891052510561
52
- STS15,cosine_spearman,0.8329896673390621
53
- STS16,cosine_spearman,0.7967145670824619
54
- STS17,cosine_spearman,0.863170912160602
55
- STS22,cosine_spearman,0.6463838183557634
56
- STSBenchmark,cosine_spearman,0.7880678352766406
57
- SummEval,cosine_spearman,0.30361250428863945
 
1
+ model,dataset,metric,value
2
+ contriever-base-msmarco,AmazonCounterfactualClassification,accuracy,0.7219402985074627
3
+ contriever-base-msmarco,AmazonPolarityClassification,accuracy,0.6863262499999999
4
+ contriever-base-msmarco,AmazonReviewsClassification,accuracy,0.37422
5
+ contriever-base-msmarco,Banking77Classification,accuracy,0.8001623376623377
6
+ contriever-base-msmarco,EmotionClassification,accuracy,0.44770000000000004
7
+ contriever-base-msmarco,ImdbClassification,accuracy,0.670388
8
+ contriever-base-msmarco,MassiveIntentClassification,accuracy,0.6777740416946872
9
+ contriever-base-msmarco,MassiveScenarioClassification,accuracy,0.7599865501008743
10
+ contriever-base-msmarco,MTOPDomainClassification,accuracy,0.9317829457364342
11
+ contriever-base-msmarco,MTOPIntentClassification,accuracy,0.6931372549019608
12
+ contriever-base-msmarco,ToxicConversationsClassification,accuracy,0.6776540000000001
13
+ contriever-base-msmarco,TweetSentimentExtractionClassification,accuracy,0.5610356536502547
14
+ contriever-base-msmarco,ArxivClusteringP2P,v_measure,0.42609650289899825
15
+ contriever-base-msmarco,ArxivClusteringS2S,v_measure,0.32317533856057756
16
+ contriever-base-msmarco,BiorxivClusteringP2P,v_measure,0.34974621074365847
17
+ contriever-base-msmarco,BiorxivClusteringS2S,v_measure,0.29083876572182377
18
+ contriever-base-msmarco,MedrxivClusteringP2P,v_measure,0.3118675108904548
19
+ contriever-base-msmarco,MedrxivClusteringS2S,v_measure,0.2727180592101203
20
+ contriever-base-msmarco,RedditClustering,v_measure,0.5489383973987073
21
+ contriever-base-msmarco,RedditClusteringP2P,v_measure,0.5757965400787036
22
+ contriever-base-msmarco,StackExchangeClustering,v_measure,0.6314901527677189
23
+ contriever-base-msmarco,StackExchangeClusteringP2P,v_measure,0.3225478381926452
24
+ contriever-base-msmarco,TwentyNewsgroupsClustering,v_measure,0.46822120585974714
25
+ contriever-base-msmarco,SprintDuplicateQuestions,ap,0.9554507015913917
26
+ contriever-base-msmarco,TwitterSemEval2015,ap,0.6684972822513366
27
+ contriever-base-msmarco,TwitterURLCorpus,ap,0.8520869965638479
28
+ contriever-base-msmarco,AskUbuntuDupQuestions,map,0.5668809109301288
29
+ contriever-base-msmarco,MindSmallReranking,map,0.3157849515126634
30
+ contriever-base-msmarco,SciDocsRR,map,0.7650730255518788
31
+ contriever-base-msmarco,StackOverflowDupQuestions,map,0.4778138013068855
32
+ contriever-base-msmarco,ArguAna,ndcg_at_10,0.4832
33
+ contriever-base-msmarco,ClimateFEVER,ndcg_at_10,0.24788
34
+ contriever-base-msmarco,CQADupstackRetrieval,ndcg_at_10,0.3366891666666667
35
+ contriever-base-msmarco,DBPedia,ndcg_at_10,0.38103
36
+ contriever-base-msmarco,FEVER,ndcg_at_10,0.59294
37
+ contriever-base-msmarco,FiQA2018,ndcg_at_10,0.2742
38
+ contriever-base-msmarco,HotpotQA,ndcg_at_10,0.56808
39
+ contriever-base-msmarco,MSMARCO,ndcg_at_10,0.62494
40
+ contriever-base-msmarco,NFCorpus,ndcg_at_10,0.31315
41
+ contriever-base-msmarco,NQ,ndcg_at_10,0.41833
42
+ contriever-base-msmarco,QuoraRetrieval,ndcg_at_10,0.86719
43
+ contriever-base-msmarco,SCIDOCS,ndcg_at_10,0.17125
44
+ contriever-base-msmarco,SciFact,ndcg_at_10,0.6551
45
+ contriever-base-msmarco,Touche2020,ndcg_at_10,0.15789
46
+ contriever-base-msmarco,TRECCOVID,ndcg_at_10,0.44769
47
+ contriever-base-msmarco,BIOSSES,cosine_spearman,0.8331816352111794
48
+ contriever-base-msmarco,SICK-R,cosine_spearman,0.702003740199912
49
+ contriever-base-msmarco,STS12,cosine_spearman,0.643399816144592
50
+ contriever-base-msmarco,STS13,cosine_spearman,0.8003351945178154
51
+ contriever-base-msmarco,STS14,cosine_spearman,0.7450891052510561
52
+ contriever-base-msmarco,STS15,cosine_spearman,0.8329896673390621
53
+ contriever-base-msmarco,STS16,cosine_spearman,0.7967145670824619
54
+ contriever-base-msmarco,STS17,cosine_spearman,0.863170912160602
55
+ contriever-base-msmarco,STS22,cosine_spearman,0.6463838183557634
56
+ contriever-base-msmarco,STSBenchmark,cosine_spearman,0.7880678352766406
57
+ contriever-base-msmarco,SummEval,cosine_spearman,0.30361250428863945
glove.6B.300d_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.5691044776119403
3
- AmazonPolarityClassification,accuracy,0.6032235
4
- AmazonReviewsClassification,accuracy,0.2967000000000001
5
- Banking77Classification,accuracy,0.6768831168831169
6
- EmotionClassification,accuracy,0.36929999999999996
7
- ImdbClassification,accuracy,0.625748
8
- MassiveIntentClassification,accuracy,0.5618695359784802
9
- MassiveScenarioClassification,accuracy,0.6602555480833894
10
- MTOPDomainClassification,accuracy,0.791062471500228
11
- MTOPIntentClassification,accuracy,0.5585499316005472
12
- ToxicConversationsClassification,accuracy,0.654038
13
- TweetSentimentExtractionClassification,accuracy,0.5079513299377476
14
- ArxivClusteringP2P,v_measure,0.3255904912728055
15
- ArxivClusteringS2S,v_measure,0.23143703465994314
16
- BiorxivClusteringP2P,v_measure,0.2926940233440429
17
- BiorxivClusteringS2S,v_measure,0.19184208889576307
18
- MedrxivClusteringP2P,v_measure,0.2612166047215606
19
- MedrxivClusteringS2S,v_measure,0.20375267261606717
20
- RedditClustering,v_measure,0.2845571223825329
21
- RedditClusteringP2P,v_measure,0.7826896969479853
22
- StackExchangeClustering,v_measure,0.35797310020384626
23
- StackExchangeClusteringP2P,v_measure,0.28510988741557386
24
- TwentyNewsgroupsClustering,v_measure,0.25831825117696844
25
- SprintDuplicateQuestions,ap,0.8696278773063988
26
- TwitterSemEval2015,ap,0.48454642053467223
27
- TwitterURLCorpus,ap,0.7734926843948666
28
- AskUbuntuDupQuestions,map,0.4957211733829243
29
- MindSmallReranking,map,0.27006016291330964
30
- SciDocsRR,map,0.6255895801677928
31
- StackOverflowDupQuestions,map,0.3403090098227517
32
- ArguAna,ndcg_at_10,0.363
33
- ClimateFEVER,ndcg_at_10,0.14441
34
- CQADupstackRetrieval,ndcg_at_10,0.154745
35
- DBPedia,ndcg_at_10,0.18285
36
- FEVER,ndcg_at_10,0.1499
37
- FiQA2018,ndcg_at_10,0.10087
38
- HotpotQA,ndcg_at_10,0.1918
39
- MSMARCO,ndcg_at_10,0.27094
40
- NFCorpus,ndcg_at_10,0.13871
41
- NQ,ndcg_at_10,0.1287
42
- QuoraRetrieval,ndcg_at_10,0.71318
43
- SCIDOCS,ndcg_at_10,0.08041
44
- SciFact,ndcg_at_10,0.29575
45
- Touche2020,ndcg_at_10,0.13987
46
- TRECCOVID,ndcg_at_10,0.36219
47
- BIOSSES,cosine_spearman,0.44927529728339055
48
- SICK-R,cosine_spearman,0.5542938456737263
49
- STS12,cosine_spearman,0.5463983809943868
50
- STS13,cosine_spearman,0.691628709341984
51
- STS14,cosine_spearman,0.6081333831404667
52
- STS15,cosine_spearman,0.723122589323431
53
- STS16,cosine_spearman,0.6533668138063549
54
- STS17,cosine_spearman,0.7794897670754327
55
- STS22,cosine_spearman,0.5635046917979596
56
- STSBenchmark,cosine_spearman,0.6153556159874409
57
- SummEval,cosine_spearman,0.2886781899399968
 
1
+ model,dataset,metric,value
2
+ glove.6B.300d,AmazonCounterfactualClassification,accuracy,0.5691044776119403
3
+ glove.6B.300d,AmazonPolarityClassification,accuracy,0.6032235
4
+ glove.6B.300d,AmazonReviewsClassification,accuracy,0.2967000000000001
5
+ glove.6B.300d,Banking77Classification,accuracy,0.6768831168831169
6
+ glove.6B.300d,EmotionClassification,accuracy,0.36929999999999996
7
+ glove.6B.300d,ImdbClassification,accuracy,0.625748
8
+ glove.6B.300d,MassiveIntentClassification,accuracy,0.5618695359784802
9
+ glove.6B.300d,MassiveScenarioClassification,accuracy,0.6602555480833894
10
+ glove.6B.300d,MTOPDomainClassification,accuracy,0.791062471500228
11
+ glove.6B.300d,MTOPIntentClassification,accuracy,0.5585499316005472
12
+ glove.6B.300d,ToxicConversationsClassification,accuracy,0.654038
13
+ glove.6B.300d,TweetSentimentExtractionClassification,accuracy,0.5079513299377476
14
+ glove.6B.300d,ArxivClusteringP2P,v_measure,0.3255904912728055
15
+ glove.6B.300d,ArxivClusteringS2S,v_measure,0.23143703465994314
16
+ glove.6B.300d,BiorxivClusteringP2P,v_measure,0.2926940233440429
17
+ glove.6B.300d,BiorxivClusteringS2S,v_measure,0.19184208889576307
18
+ glove.6B.300d,MedrxivClusteringP2P,v_measure,0.2612166047215606
19
+ glove.6B.300d,MedrxivClusteringS2S,v_measure,0.20375267261606717
20
+ glove.6B.300d,RedditClustering,v_measure,0.2845571223825329
21
+ glove.6B.300d,RedditClusteringP2P,v_measure,0.7826896969479853
22
+ glove.6B.300d,StackExchangeClustering,v_measure,0.35797310020384626
23
+ glove.6B.300d,StackExchangeClusteringP2P,v_measure,0.28510988741557386
24
+ glove.6B.300d,TwentyNewsgroupsClustering,v_measure,0.25831825117696844
25
+ glove.6B.300d,SprintDuplicateQuestions,ap,0.8696278773063988
26
+ glove.6B.300d,TwitterSemEval2015,ap,0.48454642053467223
27
+ glove.6B.300d,TwitterURLCorpus,ap,0.7734926843948666
28
+ glove.6B.300d,AskUbuntuDupQuestions,map,0.4957211733829243
29
+ glove.6B.300d,MindSmallReranking,map,0.27006016291330964
30
+ glove.6B.300d,SciDocsRR,map,0.6255895801677928
31
+ glove.6B.300d,StackOverflowDupQuestions,map,0.3403090098227517
32
+ glove.6B.300d,ArguAna,ndcg_at_10,0.363
33
+ glove.6B.300d,ClimateFEVER,ndcg_at_10,0.14441
34
+ glove.6B.300d,CQADupstackRetrieval,ndcg_at_10,0.154745
35
+ glove.6B.300d,DBPedia,ndcg_at_10,0.18285
36
+ glove.6B.300d,FEVER,ndcg_at_10,0.1499
37
+ glove.6B.300d,FiQA2018,ndcg_at_10,0.10087
38
+ glove.6B.300d,HotpotQA,ndcg_at_10,0.1918
39
+ glove.6B.300d,MSMARCO,ndcg_at_10,0.27094
40
+ glove.6B.300d,NFCorpus,ndcg_at_10,0.13871
41
+ glove.6B.300d,NQ,ndcg_at_10,0.1287
42
+ glove.6B.300d,QuoraRetrieval,ndcg_at_10,0.71318
43
+ glove.6B.300d,SCIDOCS,ndcg_at_10,0.08041
44
+ glove.6B.300d,SciFact,ndcg_at_10,0.29575
45
+ glove.6B.300d,Touche2020,ndcg_at_10,0.13987
46
+ glove.6B.300d,TRECCOVID,ndcg_at_10,0.36219
47
+ glove.6B.300d,BIOSSES,cosine_spearman,0.44927529728339055
48
+ glove.6B.300d,SICK-R,cosine_spearman,0.5542938456737263
49
+ glove.6B.300d,STS12,cosine_spearman,0.5463983809943868
50
+ glove.6B.300d,STS13,cosine_spearman,0.691628709341984
51
+ glove.6B.300d,STS14,cosine_spearman,0.6081333831404667
52
+ glove.6B.300d,STS15,cosine_spearman,0.723122589323431
53
+ glove.6B.300d,STS16,cosine_spearman,0.6533668138063549
54
+ glove.6B.300d,STS17,cosine_spearman,0.7794897670754327
55
+ glove.6B.300d,STS22,cosine_spearman,0.5635046917979596
56
+ glove.6B.300d,STSBenchmark,cosine_spearman,0.6153556159874409
57
+ glove.6B.300d,SummEval,cosine_spearman,0.2886781899399968
gtr-t5-base_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.6932835820895522
3
- AmazonPolarityClassification,accuracy,0.67818775
4
- AmazonReviewsClassification,accuracy,0.38482
5
- Banking77Classification,accuracy,0.7925649350649351
6
- EmotionClassification,accuracy,0.422
7
- ImdbClassification,accuracy,0.659944
8
- MassiveIntentClassification,accuracy,0.6705110961667786
9
- MassiveScenarioClassification,accuracy,0.7539677202420982
10
- MTOPDomainClassification,accuracy,0.9242134062927496
11
- MTOPIntentClassification,accuracy,0.6244186046511628
12
- ToxicConversationsClassification,accuracy,0.6660360000000001
13
- TweetSentimentExtractionClassification,accuracy,0.5602150537634408
14
- ArxivClusteringP2P,v_measure,0.3548695236674728
15
- ArxivClusteringS2S,v_measure,0.27180108855677926
16
- BiorxivClusteringP2P,v_measure,0.2765822038298308
17
- BiorxivClusteringS2S,v_measure,0.23251707197080881
18
- MedrxivClusteringP2P,v_measure,0.2756514819237031
19
- MedrxivClusteringS2S,v_measure,0.2513156119229876
20
- RedditClustering,v_measure,0.5613269289874916
21
- RedditClusteringP2P,v_measure,0.585344949362798
22
- StackExchangeClustering,v_measure,0.6421338133178497
23
- StackExchangeClusteringP2P,v_measure,0.3301453796958954
24
- TwentyNewsgroupsClustering,v_measure,0.46718954435107624
25
- SprintDuplicateQuestions,ap,0.9454933848900439
26
- TwitterSemEval2015,ap,0.7222675077030647
27
- TwitterURLCorpus,ap,0.8477048332956029
28
- AskUbuntuDupQuestions,map,0.6086144157716742
29
- MindSmallReranking,map,0.313303339508014
30
- SciDocsRR,map,0.737089144684744
31
- StackOverflowDupQuestions,map,0.510120398337891
32
- ArguAna,ndcg_at_10,0.50828
33
- ClimateFEVER,ndcg_at_10,0.24884
34
- CQADupstackRetrieval,ndcg_at_10,0.34548749999999995
35
- DBPedia,ndcg_at_10,0.35244
36
- FEVER,ndcg_at_10,0.68932
37
- FiQA2018,ndcg_at_10,0.35147
38
- HotpotQA,ndcg_at_10,0.54926
39
- MSMARCO,ndcg_at_10,0.68745
40
- NFCorpus,ndcg_at_10,0.30223
41
- NQ,ndcg_at_10,0.50469
42
- QuoraRetrieval,ndcg_at_10,0.87977
43
- SCIDOCS,ndcg_at_10,0.14
44
- SciFact,ndcg_at_10,0.59738
45
- Touche2020,ndcg_at_10,0.25891
46
- TRECCOVID,ndcg_at_10,0.56047
47
- BIOSSES,cosine_spearman,0.7900420980306923
48
- SICK-R,cosine_spearman,0.7900420980306923
49
- STS12,cosine_spearman,0.6859358272758398
50
- STS13,cosine_spearman,0.7909019447848391
51
- STS14,cosine_spearman,0.7464141703710193
52
- STS15,cosine_spearman,0.8484757279716068
53
- STS16,cosine_spearman,0.8156700264541153
54
- STS17,cosine_spearman,0.8579615701004483
55
- STS22,cosine_spearman,0.6616950180914625
56
- STSBenchmark,cosine_spearman,0.7957572102026641
57
- SummEval,cosine_spearman,0.29669058153671907
 
1
+ model,dataset,metric,value
2
+ gtr-t5-base,AmazonCounterfactualClassification,accuracy,0.6932835820895522
3
+ gtr-t5-base,AmazonPolarityClassification,accuracy,0.67818775
4
+ gtr-t5-base,AmazonReviewsClassification,accuracy,0.38482
5
+ gtr-t5-base,Banking77Classification,accuracy,0.7925649350649351
6
+ gtr-t5-base,EmotionClassification,accuracy,0.422
7
+ gtr-t5-base,ImdbClassification,accuracy,0.659944
8
+ gtr-t5-base,MassiveIntentClassification,accuracy,0.6705110961667786
9
+ gtr-t5-base,MassiveScenarioClassification,accuracy,0.7539677202420982
10
+ gtr-t5-base,MTOPDomainClassification,accuracy,0.9242134062927496
11
+ gtr-t5-base,MTOPIntentClassification,accuracy,0.6244186046511628
12
+ gtr-t5-base,ToxicConversationsClassification,accuracy,0.6660360000000001
13
+ gtr-t5-base,TweetSentimentExtractionClassification,accuracy,0.5602150537634408
14
+ gtr-t5-base,ArxivClusteringP2P,v_measure,0.3548695236674728
15
+ gtr-t5-base,ArxivClusteringS2S,v_measure,0.27180108855677926
16
+ gtr-t5-base,BiorxivClusteringP2P,v_measure,0.2765822038298308
17
+ gtr-t5-base,BiorxivClusteringS2S,v_measure,0.23251707197080881
18
+ gtr-t5-base,MedrxivClusteringP2P,v_measure,0.2756514819237031
19
+ gtr-t5-base,MedrxivClusteringS2S,v_measure,0.2513156119229876
20
+ gtr-t5-base,RedditClustering,v_measure,0.5613269289874916
21
+ gtr-t5-base,RedditClusteringP2P,v_measure,0.585344949362798
22
+ gtr-t5-base,StackExchangeClustering,v_measure,0.6421338133178497
23
+ gtr-t5-base,StackExchangeClusteringP2P,v_measure,0.3301453796958954
24
+ gtr-t5-base,TwentyNewsgroupsClustering,v_measure,0.46718954435107624
25
+ gtr-t5-base,SprintDuplicateQuestions,ap,0.9454933848900439
26
+ gtr-t5-base,TwitterSemEval2015,ap,0.7222675077030647
27
+ gtr-t5-base,TwitterURLCorpus,ap,0.8477048332956029
28
+ gtr-t5-base,AskUbuntuDupQuestions,map,0.6086144157716742
29
+ gtr-t5-base,MindSmallReranking,map,0.313303339508014
30
+ gtr-t5-base,SciDocsRR,map,0.737089144684744
31
+ gtr-t5-base,StackOverflowDupQuestions,map,0.510120398337891
32
+ gtr-t5-base,ArguAna,ndcg_at_10,0.50828
33
+ gtr-t5-base,ClimateFEVER,ndcg_at_10,0.24884
34
+ gtr-t5-base,CQADupstackRetrieval,ndcg_at_10,0.34548749999999995
35
+ gtr-t5-base,DBPedia,ndcg_at_10,0.35244
36
+ gtr-t5-base,FEVER,ndcg_at_10,0.68932
37
+ gtr-t5-base,FiQA2018,ndcg_at_10,0.35147
38
+ gtr-t5-base,HotpotQA,ndcg_at_10,0.54926
39
+ gtr-t5-base,MSMARCO,ndcg_at_10,0.68745
40
+ gtr-t5-base,NFCorpus,ndcg_at_10,0.30223
41
+ gtr-t5-base,NQ,ndcg_at_10,0.50469
42
+ gtr-t5-base,QuoraRetrieval,ndcg_at_10,0.87977
43
+ gtr-t5-base,SCIDOCS,ndcg_at_10,0.14
44
+ gtr-t5-base,SciFact,ndcg_at_10,0.59738
45
+ gtr-t5-base,Touche2020,ndcg_at_10,0.25891
46
+ gtr-t5-base,TRECCOVID,ndcg_at_10,0.56047
47
+ gtr-t5-base,BIOSSES,cosine_spearman,0.7900420980306923
48
+ gtr-t5-base,SICK-R,cosine_spearman,0.7900420980306923
49
+ gtr-t5-base,STS12,cosine_spearman,0.6859358272758398
50
+ gtr-t5-base,STS13,cosine_spearman,0.7909019447848391
51
+ gtr-t5-base,STS14,cosine_spearman,0.7464141703710193
52
+ gtr-t5-base,STS15,cosine_spearman,0.8484757279716068
53
+ gtr-t5-base,STS16,cosine_spearman,0.8156700264541153
54
+ gtr-t5-base,STS17,cosine_spearman,0.8579615701004483
55
+ gtr-t5-base,STS22,cosine_spearman,0.6616950180914625
56
+ gtr-t5-base,STSBenchmark,cosine_spearman,0.7957572102026641
57
+ gtr-t5-base,SummEval,cosine_spearman,0.29669058153671907
gtr-t5-xxl_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.6729850746268656
3
- AmazonPolarityClassification,accuracy,0.7505425
4
- AmazonReviewsClassification,accuracy,0.37298
5
- Banking77Classification,accuracy,0.8232142857142858
6
- EmotionClassification,accuracy,0.43190000000000006
7
- ImdbClassification,accuracy,0.708028
8
- MassiveIntentClassification,accuracy,0.7060860793544049
9
- MassiveScenarioClassification,accuracy,0.7777404169468729
10
- MTOPDomainClassification,accuracy,0.9384404924760602
11
- MTOPIntentClassification,accuracy,0.6771317829457364
12
- ToxicConversationsClassification,accuracy,0.684844
13
- TweetSentimentExtractionClassification,accuracy,0.5453593661573288
14
- ArxivClusteringP2P,v_measure,0.37901257967755886
15
- ArxivClusteringS2S,v_measure,0.32386812476732035
16
- BiorxivClusteringP2P,v_measure,0.30479899128697724
17
- BiorxivClusteringS2S,v_measure,0.274998862465458
18
- MedrxivClusteringP2P,v_measure,0.29122279823498076
19
- MedrxivClusteringS2S,v_measure,0.27559097352717216
20
- RedditClustering,v_measure,0.6412600439035501
21
- RedditClusteringP2P,v_measure,0.6284408311967189
22
- StackExchangeClustering,v_measure,0.7143121835042072
23
- StackExchangeClusteringP2P,v_measure,0.32846363140242907
24
- TwentyNewsgroupsClustering,v_measure,0.5043598055321865
25
- SprintDuplicateQuestions,ap,0.9568031991336098
26
- TwitterSemEval2015,ap,0.7754306712482096
27
- TwitterURLCorpus,ap,0.851341871186482
28
- AskUbuntuDupQuestions,map,0.632297938852618
29
- MindSmallReranking,map,0.31927428563936344
30
- SciDocsRR,map,0.7796354659897269
31
- StackOverflowDupQuestions,map,0.5349865450479917
32
- ArguAna,ndcg_at_10,0.53769
33
- ClimateFEVER,ndcg_at_10,
34
- CQADupstackRetrieval,ndcg_at_10,0.3855908333333333
35
- DBPedia,ndcg_at_10,
36
- FEVER,ndcg_at_10,
37
- FiQA2018,ndcg_at_10,0.46776
38
- HotpotQA,ndcg_at_10,
39
- MSMARCO,ndcg_at_10,
40
- NFCorpus,ndcg_at_10,0.34179
41
- NQ,ndcg_at_10,
42
- QuoraRetrieval,ndcg_at_10,0.89094
43
- SCIDOCS,ndcg_at_10,0.15881
44
- SciFact,ndcg_at_10,0.66772
45
- Touche2020,ndcg_at_10,0.2676
46
- TRECCOVID,ndcg_at_10,0.51903
47
- BIOSSES,cosine_spearman,0.819072879998681
48
- SICK-R,cosine_spearman,0.7429383974498337
49
- STS12,cosine_spearman,0.701248199396564
50
- STS13,cosine_spearman,0.8271549276218907
51
- STS14,cosine_spearman,0.7824131910280874
52
- STS15,cosine_spearman,0.8625849084541269
53
- STS16,cosine_spearman,0.8161499741247533
54
- STS17,cosine_spearman,0.8518454365763111
55
- STS22,cosine_spearman,0.6576131885658799
56
- STSBenchmark,cosine_spearman,0.7772574493807944
57
- SummEval,cosine_spearman,0.3063683987345714
 
1
+ model,dataset,metric,value
2
+ gtr-t5-xxl,AmazonCounterfactualClassification,accuracy,0.6729850746268656
3
+ gtr-t5-xxl,AmazonPolarityClassification,accuracy,0.7505425
4
+ gtr-t5-xxl,AmazonReviewsClassification,accuracy,0.37298
5
+ gtr-t5-xxl,Banking77Classification,accuracy,0.8232142857142858
6
+ gtr-t5-xxl,EmotionClassification,accuracy,0.43190000000000006
7
+ gtr-t5-xxl,ImdbClassification,accuracy,0.708028
8
+ gtr-t5-xxl,MassiveIntentClassification,accuracy,0.7060860793544049
9
+ gtr-t5-xxl,MassiveScenarioClassification,accuracy,0.7777404169468729
10
+ gtr-t5-xxl,MTOPDomainClassification,accuracy,0.9384404924760602
11
+ gtr-t5-xxl,MTOPIntentClassification,accuracy,0.6771317829457364
12
+ gtr-t5-xxl,ToxicConversationsClassification,accuracy,0.684844
13
+ gtr-t5-xxl,TweetSentimentExtractionClassification,accuracy,0.5453593661573288
14
+ gtr-t5-xxl,ArxivClusteringP2P,v_measure,0.37901257967755886
15
+ gtr-t5-xxl,ArxivClusteringS2S,v_measure,0.32386812476732035
16
+ gtr-t5-xxl,BiorxivClusteringP2P,v_measure,0.30479899128697724
17
+ gtr-t5-xxl,BiorxivClusteringS2S,v_measure,0.274998862465458
18
+ gtr-t5-xxl,MedrxivClusteringP2P,v_measure,0.29122279823498076
19
+ gtr-t5-xxl,MedrxivClusteringS2S,v_measure,0.27559097352717216
20
+ gtr-t5-xxl,RedditClustering,v_measure,0.6412600439035501
21
+ gtr-t5-xxl,RedditClusteringP2P,v_measure,0.6284408311967189
22
+ gtr-t5-xxl,StackExchangeClustering,v_measure,0.7143121835042072
23
+ gtr-t5-xxl,StackExchangeClusteringP2P,v_measure,0.32846363140242907
24
+ gtr-t5-xxl,TwentyNewsgroupsClustering,v_measure,0.5043598055321865
25
+ gtr-t5-xxl,SprintDuplicateQuestions,ap,0.9568031991336098
26
+ gtr-t5-xxl,TwitterSemEval2015,ap,0.7754306712482096
27
+ gtr-t5-xxl,TwitterURLCorpus,ap,0.851341871186482
28
+ gtr-t5-xxl,AskUbuntuDupQuestions,map,0.632297938852618
29
+ gtr-t5-xxl,MindSmallReranking,map,0.31927428563936344
30
+ gtr-t5-xxl,SciDocsRR,map,0.7796354659897269
31
+ gtr-t5-xxl,StackOverflowDupQuestions,map,0.5349865450479917
32
+ gtr-t5-xxl,ArguAna,ndcg_at_10,0.53769
33
+ gtr-t5-xxl,ClimateFEVER,ndcg_at_10,
34
+ gtr-t5-xxl,CQADupstackRetrieval,ndcg_at_10,0.3855908333333333
35
+ gtr-t5-xxl,DBPedia,ndcg_at_10,
36
+ gtr-t5-xxl,FEVER,ndcg_at_10,
37
+ gtr-t5-xxl,FiQA2018,ndcg_at_10,0.46776
38
+ gtr-t5-xxl,HotpotQA,ndcg_at_10,
39
+ gtr-t5-xxl,MSMARCO,ndcg_at_10,
40
+ gtr-t5-xxl,NFCorpus,ndcg_at_10,0.34179
41
+ gtr-t5-xxl,NQ,ndcg_at_10,
42
+ gtr-t5-xxl,QuoraRetrieval,ndcg_at_10,0.89094
43
+ gtr-t5-xxl,SCIDOCS,ndcg_at_10,0.15881
44
+ gtr-t5-xxl,SciFact,ndcg_at_10,0.66772
45
+ gtr-t5-xxl,Touche2020,ndcg_at_10,0.2676
46
+ gtr-t5-xxl,TRECCOVID,ndcg_at_10,0.51903
47
+ gtr-t5-xxl,BIOSSES,cosine_spearman,0.819072879998681
48
+ gtr-t5-xxl,SICK-R,cosine_spearman,0.7429383974498337
49
+ gtr-t5-xxl,STS12,cosine_spearman,0.701248199396564
50
+ gtr-t5-xxl,STS13,cosine_spearman,0.8271549276218907
51
+ gtr-t5-xxl,STS14,cosine_spearman,0.7824131910280874
52
+ gtr-t5-xxl,STS15,cosine_spearman,0.8625849084541269
53
+ gtr-t5-xxl,STS16,cosine_spearman,0.8161499741247533
54
+ gtr-t5-xxl,STS17,cosine_spearman,0.8518454365763111
55
+ gtr-t5-xxl,STS22,cosine_spearman,0.6576131885658799
56
+ gtr-t5-xxl,STSBenchmark,cosine_spearman,0.7772574493807944
57
+ gtr-t5-xxl,SummEval,cosine_spearman,0.3063683987345714
komninos_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.6053731343283582
3
- AmazonPolarityClassification,accuracy,0.5958574999999999
4
- AmazonReviewsClassification,accuracy,0.31013999999999997
5
- Banking77Classification,accuracy,0.6704870129870131
6
- EmotionClassification,accuracy,0.3318
7
- ImdbClassification,accuracy,0.639756
8
- MassiveIntentClassification,accuracy,0.5721250840618696
9
- MassiveScenarioClassification,accuracy,0.6611297915265636
10
- MTOPDomainClassification,accuracy,0.7857045143638851
11
- MTOPIntentClassification,accuracy,0.5707250341997263
12
- ToxicConversationsClassification,accuracy,0.677564
13
- TweetSentimentExtractionClassification,accuracy,0.4968307866440294
14
- ArxivClusteringP2P,v_measure,0.3472624365591666
15
- ArxivClusteringS2S,v_measure,0.26008268382733774
16
- BiorxivClusteringP2P,v_measure,0.2975913127921992
17
- BiorxivClusteringS2S,v_measure,0.20711732977389513
18
- MedrxivClusteringP2P,v_measure,0.2665246941866784
19
- MedrxivClusteringS2S,v_measure,0.21503100153759527
20
- RedditClustering,v_measure,0.288407473136122
21
- RedditClusteringP2P,v_measure,0.0737029180225715
22
- StackExchangeClustering,v_measure,0.39038573398030996
23
- StackExchangeClusteringP2P,v_measure,0.3022655763114615
24
- TwentyNewsgroupsClustering,v_measure,0.27420294690851976
25
- SprintDuplicateQuestions,ap,0.8555090397855645
26
- TwitterSemEval2015,ap,0.5385004776679795
27
- TwitterURLCorpus,ap,0.7941032575876537
28
- AskUbuntuDupQuestions,map,0.5087892970617054
29
- MindSmallReranking,map,0.28921918247348344
30
- SciDocsRR,map,0.6355311507513624
31
- StackOverflowDupQuestions,map,0.3564753567606952
32
- ArguAna,ndcg_at_10,0.30959
33
- ClimateFEVER,ndcg_at_10,0.14867
34
- CQADupstackRetrieval,ndcg_at_10,0.16788249999999996
35
- DBPedia,ndcg_at_10,0.15877
36
- FEVER,ndcg_at_10,0.15558
37
- FiQA2018,ndcg_at_10,0.10488
38
- HotpotQA,ndcg_at_10,0.20768
39
- MSMARCO,ndcg_at_10,0.28197
40
- NFCorpus,ndcg_at_10,0.11787
41
- NQ,ndcg_at_10,0.12751
42
- QuoraRetrieval,ndcg_at_10,0.71575
43
- SCIDOCS,ndcg_at_10,0.0847
44
- SciFact,ndcg_at_10,0.29526
45
- Touche2020,ndcg_at_10,0.13173
46
- TRECCOVID,ndcg_at_10,0.35923
47
- BIOSSES,cosine_spearman,0.502481328896023
48
- SICK-R,cosine_spearman,0.5548989810259632
49
- STS12,cosine_spearman,0.5351287401411937
50
- STS13,cosine_spearman,0.7080105270649192
51
- STS14,cosine_spearman,0.6356156596006354
52
- STS15,cosine_spearman,0.740812262403237
53
- STS16,cosine_spearman,0.646011290069885
54
- STS17,cosine_spearman,0.7691005426258415
55
- STS22,cosine_spearman,0.5389239838658062
56
- STSBenchmark,cosine_spearman,0.6155281748776228
57
- SummEval,cosine_spearman,0.3049084135949722
 
1
+ model,dataset,metric,value
2
+ komninos,AmazonCounterfactualClassification,accuracy,0.6053731343283582
3
+ komninos,AmazonPolarityClassification,accuracy,0.5958574999999999
4
+ komninos,AmazonReviewsClassification,accuracy,0.31013999999999997
5
+ komninos,Banking77Classification,accuracy,0.6704870129870131
6
+ komninos,EmotionClassification,accuracy,0.3318
7
+ komninos,ImdbClassification,accuracy,0.639756
8
+ komninos,MassiveIntentClassification,accuracy,0.5721250840618696
9
+ komninos,MassiveScenarioClassification,accuracy,0.6611297915265636
10
+ komninos,MTOPDomainClassification,accuracy,0.7857045143638851
11
+ komninos,MTOPIntentClassification,accuracy,0.5707250341997263
12
+ komninos,ToxicConversationsClassification,accuracy,0.677564
13
+ komninos,TweetSentimentExtractionClassification,accuracy,0.4968307866440294
14
+ komninos,ArxivClusteringP2P,v_measure,0.3472624365591666
15
+ komninos,ArxivClusteringS2S,v_measure,0.26008268382733774
16
+ komninos,BiorxivClusteringP2P,v_measure,0.2975913127921992
17
+ komninos,BiorxivClusteringS2S,v_measure,0.20711732977389513
18
+ komninos,MedrxivClusteringP2P,v_measure,0.2665246941866784
19
+ komninos,MedrxivClusteringS2S,v_measure,0.21503100153759527
20
+ komninos,RedditClustering,v_measure,0.288407473136122
21
+ komninos,RedditClusteringP2P,v_measure,0.0737029180225715
22
+ komninos,StackExchangeClustering,v_measure,0.39038573398030996
23
+ komninos,StackExchangeClusteringP2P,v_measure,0.3022655763114615
24
+ komninos,TwentyNewsgroupsClustering,v_measure,0.27420294690851976
25
+ komninos,SprintDuplicateQuestions,ap,0.8555090397855645
26
+ komninos,TwitterSemEval2015,ap,0.5385004776679795
27
+ komninos,TwitterURLCorpus,ap,0.7941032575876537
28
+ komninos,AskUbuntuDupQuestions,map,0.5087892970617054
29
+ komninos,MindSmallReranking,map,0.28921918247348344
30
+ komninos,SciDocsRR,map,0.6355311507513624
31
+ komninos,StackOverflowDupQuestions,map,0.3564753567606952
32
+ komninos,ArguAna,ndcg_at_10,0.30959
33
+ komninos,ClimateFEVER,ndcg_at_10,0.14867
34
+ komninos,CQADupstackRetrieval,ndcg_at_10,0.16788249999999996
35
+ komninos,DBPedia,ndcg_at_10,0.15877
36
+ komninos,FEVER,ndcg_at_10,0.15558
37
+ komninos,FiQA2018,ndcg_at_10,0.10488
38
+ komninos,HotpotQA,ndcg_at_10,0.20768
39
+ komninos,MSMARCO,ndcg_at_10,0.28197
40
+ komninos,NFCorpus,ndcg_at_10,0.11787
41
+ komninos,NQ,ndcg_at_10,0.12751
42
+ komninos,QuoraRetrieval,ndcg_at_10,0.71575
43
+ komninos,SCIDOCS,ndcg_at_10,0.0847
44
+ komninos,SciFact,ndcg_at_10,0.29526
45
+ komninos,Touche2020,ndcg_at_10,0.13173
46
+ komninos,TRECCOVID,ndcg_at_10,0.35923
47
+ komninos,BIOSSES,cosine_spearman,0.502481328896023
48
+ komninos,SICK-R,cosine_spearman,0.5548989810259632
49
+ komninos,STS12,cosine_spearman,0.5351287401411937
50
+ komninos,STS13,cosine_spearman,0.7080105270649192
51
+ komninos,STS14,cosine_spearman,0.6356156596006354
52
+ komninos,STS15,cosine_spearman,0.740812262403237
53
+ komninos,STS16,cosine_spearman,0.646011290069885
54
+ komninos,STS17,cosine_spearman,0.7691005426258415
55
+ komninos,STS22,cosine_spearman,0.5389239838658062
56
+ komninos,STSBenchmark,cosine_spearman,0.6155281748776228
57
+ komninos,SummEval,cosine_spearman,0.3049084135949722
msmarco-bert-co-condensor_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.6405970149253731
3
- AmazonPolarityClassification,accuracy,0.66883225
4
- AmazonReviewsClassification,accuracy,0.34852
5
- Banking77Classification,accuracy,0.8235389610389611
6
- EmotionClassification,accuracy,0.4191
7
- ImdbClassification,accuracy,0.6016519999999999
8
- MassiveIntentClassification,accuracy,0.7040013449899125
9
- MassiveScenarioClassification,accuracy,0.73728984532616
10
- MTOPDomainClassification,accuracy,0.9133834929320566
11
- MTOPIntentClassification,accuracy,0.7106931144550843
12
- ToxicConversationsClassification,accuracy,0.6400720000000001
13
- TweetSentimentExtractionClassification,accuracy,0.5573571024335031
14
- ArxivClusteringP2P,v_measure,0.369429585124196
15
- ArxivClusteringS2S,v_measure,0.2903421184003919
16
- BiorxivClusteringP2P,v_measure,0.3234945427009579
17
- BiorxivClusteringS2S,v_measure,0.28155732412289625
18
- MedrxivClusteringP2P,v_measure,0.3023226235842047
19
- MedrxivClusteringS2S,v_measure,0.2700892040743902
20
- RedditClustering,v_measure,0.48044277851861944
21
- RedditClusteringP2P,v_measure,0.5352963896499805
22
- StackExchangeClustering,v_measure,0.5953788370911354
23
- StackExchangeClusteringP2P,v_measure,0.30481638824199886
24
- TwentyNewsgroupsClustering,v_measure,0.3867893425770875
25
- SprintDuplicateQuestions,ap,0.9609255002552
26
- TwitterSemEval2015,ap,0.6594999377083578
27
- TwitterURLCorpus,ap,0.8317318134999252
28
- AskUbuntuDupQuestions,map,0.5898600276337803
29
- MindSmallReranking,map,0.2713298598749459
30
- SciDocsRR,map,0.7277578387510362
31
- StackOverflowDupQuestions,map,0.4848067606975988
32
- ArguAna,ndcg_at_10,0.45154
33
- ClimateFEVER,ndcg_at_10,0.16957
34
- CQADupstackRetrieval,ndcg_at_10,0.2771733333333333
35
- DBPedia,ndcg_at_10,0.27859
36
- FEVER,ndcg_at_10,0.45684
37
- FiQA2018,ndcg_at_10,0.15619
38
- HotpotQA,ndcg_at_10,0.35608
39
- MSMARCO,ndcg_at_10,0.56235
40
- NFCorpus,ndcg_at_10,0.22291
41
- NQ,ndcg_at_10,0.29855
42
- QuoraRetrieval,ndcg_at_10,0.86506
43
- SCIDOCS,ndcg_at_10,0.10131
44
- SciFact,ndcg_at_10,0.52311
45
- Touche2020,ndcg_at_10,0.08573
46
- TRECCOVID,ndcg_at_10,0.40536
47
- BIOSSES,cosine_spearman,0.7731501468710659
48
- SICK-R,cosine_spearman,0.719994667180189
49
- STS12,cosine_spearman,0.6819498646316081
50
- STS13,cosine_spearman,0.8040441089650371
51
- STS14,cosine_spearman,0.7401841404523465
52
- STS15,cosine_spearman,0.8257380490897778
53
- STS16,cosine_spearman,0.7977999310698289
54
- STS17,cosine_spearman,0.8593569295808465
55
- STS22,cosine_spearman,0.675433916658919
56
- STSBenchmark,cosine_spearman,0.7696651521767748
57
- SummEval,cosine_spearman,0.29504172709559223
 
1
+ model,dataset,metric,value
2
+ msmarco-bert-co-condensor,AmazonCounterfactualClassification,accuracy,0.6405970149253731
3
+ msmarco-bert-co-condensor,AmazonPolarityClassification,accuracy,0.66883225
4
+ msmarco-bert-co-condensor,AmazonReviewsClassification,accuracy,0.34852
5
+ msmarco-bert-co-condensor,Banking77Classification,accuracy,0.8235389610389611
6
+ msmarco-bert-co-condensor,EmotionClassification,accuracy,0.4191
7
+ msmarco-bert-co-condensor,ImdbClassification,accuracy,0.6016519999999999
8
+ msmarco-bert-co-condensor,MassiveIntentClassification,accuracy,0.7040013449899125
9
+ msmarco-bert-co-condensor,MassiveScenarioClassification,accuracy,0.73728984532616
10
+ msmarco-bert-co-condensor,MTOPDomainClassification,accuracy,0.9133834929320566
11
+ msmarco-bert-co-condensor,MTOPIntentClassification,accuracy,0.7106931144550843
12
+ msmarco-bert-co-condensor,ToxicConversationsClassification,accuracy,0.6400720000000001
13
+ msmarco-bert-co-condensor,TweetSentimentExtractionClassification,accuracy,0.5573571024335031
14
+ msmarco-bert-co-condensor,ArxivClusteringP2P,v_measure,0.369429585124196
15
+ msmarco-bert-co-condensor,ArxivClusteringS2S,v_measure,0.2903421184003919
16
+ msmarco-bert-co-condensor,BiorxivClusteringP2P,v_measure,0.3234945427009579
17
+ msmarco-bert-co-condensor,BiorxivClusteringS2S,v_measure,0.28155732412289625
18
+ msmarco-bert-co-condensor,MedrxivClusteringP2P,v_measure,0.3023226235842047
19
+ msmarco-bert-co-condensor,MedrxivClusteringS2S,v_measure,0.2700892040743902
20
+ msmarco-bert-co-condensor,RedditClustering,v_measure,0.48044277851861944
21
+ msmarco-bert-co-condensor,RedditClusteringP2P,v_measure,0.5352963896499805
22
+ msmarco-bert-co-condensor,StackExchangeClustering,v_measure,0.5953788370911354
23
+ msmarco-bert-co-condensor,StackExchangeClusteringP2P,v_measure,0.30481638824199886
24
+ msmarco-bert-co-condensor,TwentyNewsgroupsClustering,v_measure,0.3867893425770875
25
+ msmarco-bert-co-condensor,SprintDuplicateQuestions,ap,0.9609255002552
26
+ msmarco-bert-co-condensor,TwitterSemEval2015,ap,0.6594999377083578
27
+ msmarco-bert-co-condensor,TwitterURLCorpus,ap,0.8317318134999252
28
+ msmarco-bert-co-condensor,AskUbuntuDupQuestions,map,0.5898600276337803
29
+ msmarco-bert-co-condensor,MindSmallReranking,map,0.2713298598749459
30
+ msmarco-bert-co-condensor,SciDocsRR,map,0.7277578387510362
31
+ msmarco-bert-co-condensor,StackOverflowDupQuestions,map,0.4848067606975988
32
+ msmarco-bert-co-condensor,ArguAna,ndcg_at_10,0.45154
33
+ msmarco-bert-co-condensor,ClimateFEVER,ndcg_at_10,0.16957
34
+ msmarco-bert-co-condensor,CQADupstackRetrieval,ndcg_at_10,0.2771733333333333
35
+ msmarco-bert-co-condensor,DBPedia,ndcg_at_10,0.27859
36
+ msmarco-bert-co-condensor,FEVER,ndcg_at_10,0.45684
37
+ msmarco-bert-co-condensor,FiQA2018,ndcg_at_10,0.15619
38
+ msmarco-bert-co-condensor,HotpotQA,ndcg_at_10,0.35608
39
+ msmarco-bert-co-condensor,MSMARCO,ndcg_at_10,0.56235
40
+ msmarco-bert-co-condensor,NFCorpus,ndcg_at_10,0.22291
41
+ msmarco-bert-co-condensor,NQ,ndcg_at_10,0.29855
42
+ msmarco-bert-co-condensor,QuoraRetrieval,ndcg_at_10,0.86506
43
+ msmarco-bert-co-condensor,SCIDOCS,ndcg_at_10,0.10131
44
+ msmarco-bert-co-condensor,SciFact,ndcg_at_10,0.52311
45
+ msmarco-bert-co-condensor,Touche2020,ndcg_at_10,0.08573
46
+ msmarco-bert-co-condensor,TRECCOVID,ndcg_at_10,0.40536
47
+ msmarco-bert-co-condensor,BIOSSES,cosine_spearman,0.7731501468710659
48
+ msmarco-bert-co-condensor,SICK-R,cosine_spearman,0.719994667180189
49
+ msmarco-bert-co-condensor,STS12,cosine_spearman,0.6819498646316081
50
+ msmarco-bert-co-condensor,STS13,cosine_spearman,0.8040441089650371
51
+ msmarco-bert-co-condensor,STS14,cosine_spearman,0.7401841404523465
52
+ msmarco-bert-co-condensor,STS15,cosine_spearman,0.8257380490897778
53
+ msmarco-bert-co-condensor,STS16,cosine_spearman,0.7977999310698289
54
+ msmarco-bert-co-condensor,STS17,cosine_spearman,0.8593569295808465
55
+ msmarco-bert-co-condensor,STS22,cosine_spearman,0.675433916658919
56
+ msmarco-bert-co-condensor,STSBenchmark,cosine_spearman,0.7696651521767748
57
+ msmarco-bert-co-condensor,SummEval,cosine_spearman,0.29504172709559223
sentence-t5-base_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.7582089552238807
3
- AmazonPolarityClassification,accuracy,0.8511737500000001
4
- AmazonReviewsClassification,accuracy,0.44943999999999995
5
- Banking77Classification,accuracy,0.7647727272727273
6
- EmotionClassification,accuracy,0.5135500000000001
7
- ImdbClassification,accuracy,0.773372
8
- MassiveIntentClassification,accuracy,0.6974445191661063
9
- MassiveScenarioClassification,accuracy,0.7232347007397445
10
- MTOPDomainClassification,accuracy,0.903374373005016
11
- MTOPIntentClassification,accuracy,0.6332421340629275
12
- ToxicConversationsClassification,accuracy,0.6819919999999999
13
- TweetSentimentExtractionClassification,accuracy,0.6271080928126768
14
- ArxivClusteringP2P,v_measure,0.39275291662236395
15
- ArxivClusteringS2S,v_measure,0.27261128959373326
16
- BiorxivClusteringP2P,v_measure,0.33985330956534987
17
- BiorxivClusteringS2S,v_measure,0.22921149832439514
18
- MedrxivClusteringP2P,v_measure,0.33201641285078753
19
- MedrxivClusteringS2S,v_measure,0.2613067187536383
20
- RedditClustering,v_measure,0.5293053483970847
21
- RedditClusteringP2P,v_measure,0.596686566444821
22
- StackExchangeClustering,v_measure,0.6313072579524015
23
- StackExchangeClusteringP2P,v_measure,0.3568114137558022
24
- TwentyNewsgroupsClustering,v_measure,0.48099101871064487
25
- SprintDuplicateQuestions,ap,0.9122814257221482
26
- TwitterSemEval2015,ap,0.7825096573546108
27
- TwitterURLCorpus,ap,0.8604575990028458
28
- AskUbuntuDupQuestions,map,0.5972875661091822
29
- MindSmallReranking,map,0.30196798710053224
30
- SciDocsRR,map,0.7396228332723271
31
- StackOverflowDupQuestions,map,0.4846269194141537
32
- ArguAna,ndcg_at_10,0.44846
33
- ClimateFEVER,ndcg_at_10,0.10367
34
- CQADupstackRetrieval,ndcg_at_10,0.3523175
35
- DBPedia,ndcg_at_10,0.27766
36
- FEVER,ndcg_at_10,0.26165
37
- FiQA2018,ndcg_at_10,0.34832
38
- HotpotQA,ndcg_at_10,0.33198
39
- MSMARCO,ndcg_at_10,0.4111
40
- NFCorpus,ndcg_at_10,0.28645
41
- NQ,ndcg_at_10,0.36324
42
- QuoraRetrieval,ndcg_at_10,0.85491
43
- SCIDOCS,ndcg_at_10,0.14155
44
- SciFact,ndcg_at_10,0.4576
45
- Touche2020,ndcg_at_10,0.20296
46
- TRECCOVID,ndcg_at_10,0.40699
47
- BIOSSES,cosine_spearman,0.7589088585182279
48
- SICK-R,cosine_spearman,0.7589088585182279
49
- STS12,cosine_spearman,0.780511871449349
50
- STS13,cosine_spearman,0.8584897342040492
51
- STS14,cosine_spearman,0.8218926664662587
52
- STS15,cosine_spearman,0.8745860981918768
53
- STS16,cosine_spearman,0.8403233567112526
54
- STS17,cosine_spearman,0.8956885540021488
55
- STS22,cosine_spearman,0.6265994888539158
56
- STSBenchmark,cosine_spearman,0.8552030817522575
57
- SummEval,cosine_spearman,0.313940211538976
 
1
+ model,dataset,metric,value
2
+ sentence-t5-base,AmazonCounterfactualClassification,accuracy,0.7582089552238807
3
+ sentence-t5-base,AmazonPolarityClassification,accuracy,0.8511737500000001
4
+ sentence-t5-base,AmazonReviewsClassification,accuracy,0.44943999999999995
5
+ sentence-t5-base,Banking77Classification,accuracy,0.7647727272727273
6
+ sentence-t5-base,EmotionClassification,accuracy,0.5135500000000001
7
+ sentence-t5-base,ImdbClassification,accuracy,0.773372
8
+ sentence-t5-base,MassiveIntentClassification,accuracy,0.6974445191661063
9
+ sentence-t5-base,MassiveScenarioClassification,accuracy,0.7232347007397445
10
+ sentence-t5-base,MTOPDomainClassification,accuracy,0.903374373005016
11
+ sentence-t5-base,MTOPIntentClassification,accuracy,0.6332421340629275
12
+ sentence-t5-base,ToxicConversationsClassification,accuracy,0.6819919999999999
13
+ sentence-t5-base,TweetSentimentExtractionClassification,accuracy,0.6271080928126768
14
+ sentence-t5-base,ArxivClusteringP2P,v_measure,0.39275291662236395
15
+ sentence-t5-base,ArxivClusteringS2S,v_measure,0.27261128959373326
16
+ sentence-t5-base,BiorxivClusteringP2P,v_measure,0.33985330956534987
17
+ sentence-t5-base,BiorxivClusteringS2S,v_measure,0.22921149832439514
18
+ sentence-t5-base,MedrxivClusteringP2P,v_measure,0.33201641285078753
19
+ sentence-t5-base,MedrxivClusteringS2S,v_measure,0.2613067187536383
20
+ sentence-t5-base,RedditClustering,v_measure,0.5293053483970847
21
+ sentence-t5-base,RedditClusteringP2P,v_measure,0.596686566444821
22
+ sentence-t5-base,StackExchangeClustering,v_measure,0.6313072579524015
23
+ sentence-t5-base,StackExchangeClusteringP2P,v_measure,0.3568114137558022
24
+ sentence-t5-base,TwentyNewsgroupsClustering,v_measure,0.48099101871064487
25
+ sentence-t5-base,SprintDuplicateQuestions,ap,0.9122814257221482
26
+ sentence-t5-base,TwitterSemEval2015,ap,0.7825096573546108
27
+ sentence-t5-base,TwitterURLCorpus,ap,0.8604575990028458
28
+ sentence-t5-base,AskUbuntuDupQuestions,map,0.5972875661091822
29
+ sentence-t5-base,MindSmallReranking,map,0.30196798710053224
30
+ sentence-t5-base,SciDocsRR,map,0.7396228332723271
31
+ sentence-t5-base,StackOverflowDupQuestions,map,0.4846269194141537
32
+ sentence-t5-base,ArguAna,ndcg_at_10,0.44846
33
+ sentence-t5-base,ClimateFEVER,ndcg_at_10,0.10367
34
+ sentence-t5-base,CQADupstackRetrieval,ndcg_at_10,0.3523175
35
+ sentence-t5-base,DBPedia,ndcg_at_10,0.27766
36
+ sentence-t5-base,FEVER,ndcg_at_10,0.26165
37
+ sentence-t5-base,FiQA2018,ndcg_at_10,0.34832
38
+ sentence-t5-base,HotpotQA,ndcg_at_10,0.33198
39
+ sentence-t5-base,MSMARCO,ndcg_at_10,0.4111
40
+ sentence-t5-base,NFCorpus,ndcg_at_10,0.28645
41
+ sentence-t5-base,NQ,ndcg_at_10,0.36324
42
+ sentence-t5-base,QuoraRetrieval,ndcg_at_10,0.85491
43
+ sentence-t5-base,SCIDOCS,ndcg_at_10,0.14155
44
+ sentence-t5-base,SciFact,ndcg_at_10,0.4576
45
+ sentence-t5-base,Touche2020,ndcg_at_10,0.20296
46
+ sentence-t5-base,TRECCOVID,ndcg_at_10,0.40699
47
+ sentence-t5-base,BIOSSES,cosine_spearman,0.7589088585182279
48
+ sentence-t5-base,SICK-R,cosine_spearman,0.7589088585182279
49
+ sentence-t5-base,STS12,cosine_spearman,0.780511871449349
50
+ sentence-t5-base,STS13,cosine_spearman,0.8584897342040492
51
+ sentence-t5-base,STS14,cosine_spearman,0.8218926664662587
52
+ sentence-t5-base,STS15,cosine_spearman,0.8745860981918768
53
+ sentence-t5-base,STS16,cosine_spearman,0.8403233567112526
54
+ sentence-t5-base,STS17,cosine_spearman,0.8956885540021488
55
+ sentence-t5-base,STS22,cosine_spearman,0.6265994888539158
56
+ sentence-t5-base,STSBenchmark,cosine_spearman,0.8552030817522575
57
+ sentence-t5-base,SummEval,cosine_spearman,0.313940211538976
sentence-t5-xxl_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.7707462686567165
3
- AmazonPolarityClassification,accuracy,0.9278587500000001
4
- AmazonReviewsClassification,accuracy,0.48926
5
- Banking77Classification,accuracy,0.8230844155844155
6
- EmotionClassification,accuracy,0.4857
7
- ImdbClassification,accuracy,0.902268
8
- MassiveIntentClassification,accuracy,0.7344317417619368
9
- MassiveScenarioClassification,accuracy,0.7481842636180229
10
- MTOPDomainClassification,accuracy,0.9249430004559963
11
- MTOPIntentClassification,accuracy,0.683264933880529
12
- ToxicConversationsClassification,accuracy,0.700366
13
- TweetSentimentExtractionClassification,accuracy,0.6200905489530277
14
- ArxivClusteringP2P,v_measure,0.428912091513744
15
- ArxivClusteringS2S,v_measure,0.334692173332607
16
- BiorxivClusteringP2P,v_measure,0.3652762269967934
17
- BiorxivClusteringS2S,v_measure,0.2866305685584199
18
- MedrxivClusteringP2P,v_measure,0.3208579812374771
19
- MedrxivClusteringS2S,v_measure,0.2681601616447503
20
- RedditClustering,v_measure,0.5898542282844221
21
- RedditClusteringP2P,v_measure,0.6445543643659679
22
- StackExchangeClustering,v_measure,0.7077797692376002
23
- StackExchangeClusteringP2P,v_measure,0.3525246972831145
24
- TwentyNewsgroupsClustering,v_measure,0.5093024400468489
25
- SprintDuplicateQuestions,ap,0.8888675641034331
26
- TwitterSemEval2015,ap,0.802843387858549
27
- TwitterURLCorpus,ap,0.8601198098811131
28
- AskUbuntuDupQuestions,map,0.6615532012360636
29
- MindSmallReranking,map,0.3059647884122016
30
- SciDocsRR,map,0.7609412848204858
31
- StackOverflowDupQuestions,map,0.5285484177221984
32
- ArguAna,ndcg_at_10,0.39847
33
- ClimateFEVER,ndcg_at_10,
34
- CQADupstackRetrieval,ndcg_at_10,0.44654833333333344
35
- DBPedia,ndcg_at_10,
36
- FEVER,ndcg_at_10,
37
- FiQA2018,ndcg_at_10,0.46677
38
- HotpotQA,ndcg_at_10,
39
- MSMARCO,ndcg_at_10,
40
- NFCorpus,ndcg_at_10,0.35077
41
- NQ,ndcg_at_10,0.5287
42
- QuoraRetrieval,ndcg_at_10,0.85959
43
- SCIDOCS,ndcg_at_10,0.17173
44
- SciFact,ndcg_at_10,0.5538
45
- Touche2020,ndcg_at_10,0.21647
46
- TRECCOVID,ndcg_at_10,0.59481
47
- BIOSSES,cosine_spearman,0.8042954348508337
48
- SICK-R,cosine_spearman,0.8042954348508337
49
- STS12,cosine_spearman,0.7884633095838921
50
- STS13,cosine_spearman,0.8893834486158684
51
- STS14,cosine_spearman,0.8485684801116044
52
- STS15,cosine_spearman,0.8931982869751905
53
- STS16,cosine_spearman,0.846657396318755
54
- STS17,cosine_spearman,0.8946272351246337
55
- STS22,cosine_spearman,0.6532938965959216
56
- STSBenchmark,cosine_spearman,0.8401139141474803
57
- SummEval,cosine_spearman,0.30077056036328337
 
1
+ model,dataset,metric,value
2
+ sentence-t5-xxl,AmazonCounterfactualClassification,accuracy,0.7707462686567165
3
+ sentence-t5-xxl,AmazonPolarityClassification,accuracy,0.9278587500000001
4
+ sentence-t5-xxl,AmazonReviewsClassification,accuracy,0.48926
5
+ sentence-t5-xxl,Banking77Classification,accuracy,0.8230844155844155
6
+ sentence-t5-xxl,EmotionClassification,accuracy,0.4857
7
+ sentence-t5-xxl,ImdbClassification,accuracy,0.902268
8
+ sentence-t5-xxl,MassiveIntentClassification,accuracy,0.7344317417619368
9
+ sentence-t5-xxl,MassiveScenarioClassification,accuracy,0.7481842636180229
10
+ sentence-t5-xxl,MTOPDomainClassification,accuracy,0.9249430004559963
11
+ sentence-t5-xxl,MTOPIntentClassification,accuracy,0.683264933880529
12
+ sentence-t5-xxl,ToxicConversationsClassification,accuracy,0.700366
13
+ sentence-t5-xxl,TweetSentimentExtractionClassification,accuracy,0.6200905489530277
14
+ sentence-t5-xxl,ArxivClusteringP2P,v_measure,0.428912091513744
15
+ sentence-t5-xxl,ArxivClusteringS2S,v_measure,0.334692173332607
16
+ sentence-t5-xxl,BiorxivClusteringP2P,v_measure,0.3652762269967934
17
+ sentence-t5-xxl,BiorxivClusteringS2S,v_measure,0.2866305685584199
18
+ sentence-t5-xxl,MedrxivClusteringP2P,v_measure,0.3208579812374771
19
+ sentence-t5-xxl,MedrxivClusteringS2S,v_measure,0.2681601616447503
20
+ sentence-t5-xxl,RedditClustering,v_measure,0.5898542282844221
21
+ sentence-t5-xxl,RedditClusteringP2P,v_measure,0.6445543643659679
22
+ sentence-t5-xxl,StackExchangeClustering,v_measure,0.7077797692376002
23
+ sentence-t5-xxl,StackExchangeClusteringP2P,v_measure,0.3525246972831145
24
+ sentence-t5-xxl,TwentyNewsgroupsClustering,v_measure,0.5093024400468489
25
+ sentence-t5-xxl,SprintDuplicateQuestions,ap,0.8888675641034331
26
+ sentence-t5-xxl,TwitterSemEval2015,ap,0.802843387858549
27
+ sentence-t5-xxl,TwitterURLCorpus,ap,0.8601198098811131
28
+ sentence-t5-xxl,AskUbuntuDupQuestions,map,0.6615532012360636
29
+ sentence-t5-xxl,MindSmallReranking,map,0.3059647884122016
30
+ sentence-t5-xxl,SciDocsRR,map,0.7609412848204858
31
+ sentence-t5-xxl,StackOverflowDupQuestions,map,0.5285484177221984
32
+ sentence-t5-xxl,ArguAna,ndcg_at_10,0.39847
33
+ sentence-t5-xxl,ClimateFEVER,ndcg_at_10,
34
+ sentence-t5-xxl,CQADupstackRetrieval,ndcg_at_10,0.44654833333333344
35
+ sentence-t5-xxl,DBPedia,ndcg_at_10,
36
+ sentence-t5-xxl,FEVER,ndcg_at_10,
37
+ sentence-t5-xxl,FiQA2018,ndcg_at_10,0.46677
38
+ sentence-t5-xxl,HotpotQA,ndcg_at_10,
39
+ sentence-t5-xxl,MSMARCO,ndcg_at_10,
40
+ sentence-t5-xxl,NFCorpus,ndcg_at_10,0.35077
41
+ sentence-t5-xxl,NQ,ndcg_at_10,0.5287
42
+ sentence-t5-xxl,QuoraRetrieval,ndcg_at_10,0.85959
43
+ sentence-t5-xxl,SCIDOCS,ndcg_at_10,0.17173
44
+ sentence-t5-xxl,SciFact,ndcg_at_10,0.5538
45
+ sentence-t5-xxl,Touche2020,ndcg_at_10,0.21647
46
+ sentence-t5-xxl,TRECCOVID,ndcg_at_10,0.59481
47
+ sentence-t5-xxl,BIOSSES,cosine_spearman,0.8042954348508337
48
+ sentence-t5-xxl,SICK-R,cosine_spearman,0.8042954348508337
49
+ sentence-t5-xxl,STS12,cosine_spearman,0.7884633095838921
50
+ sentence-t5-xxl,STS13,cosine_spearman,0.8893834486158684
51
+ sentence-t5-xxl,STS14,cosine_spearman,0.8485684801116044
52
+ sentence-t5-xxl,STS15,cosine_spearman,0.8931982869751905
53
+ sentence-t5-xxl,STS16,cosine_spearman,0.846657396318755
54
+ sentence-t5-xxl,STS17,cosine_spearman,0.8946272351246337
55
+ sentence-t5-xxl,STS22,cosine_spearman,0.6532938965959216
56
+ sentence-t5-xxl,STSBenchmark,cosine_spearman,0.8401139141474803
57
+ sentence-t5-xxl,SummEval,cosine_spearman,0.30077056036328337
sgpt-bloom-7b1-msmarco_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,
3
- AmazonPolarityClassification,accuracy,
4
- AmazonReviewsClassification,accuracy,
5
- Banking77Classification,accuracy,
6
- EmotionClassification,accuracy,
7
- ImdbClassification,accuracy,
8
- MassiveIntentClassification,accuracy,
9
- MassiveScenarioClassification,accuracy,
10
- MTOPDomainClassification,accuracy,
11
- MTOPIntentClassification,accuracy,
12
- ToxicConversationsClassification,accuracy,
13
- TweetSentimentExtractionClassification,accuracy,
14
- ArxivClusteringP2P,v_measure,
15
- ArxivClusteringS2S,v_measure,
16
- BiorxivClusteringP2P,v_measure,
17
- BiorxivClusteringS2S,v_measure,
18
- MedrxivClusteringP2P,v_measure,
19
- MedrxivClusteringS2S,v_measure,
20
- RedditClustering,v_measure,
21
- RedditClusteringP2P,v_measure,
22
- StackExchangeClustering,v_measure,
23
- StackExchangeClusteringP2P,v_measure,
24
- TwentyNewsgroupsClustering,v_measure,
25
- SprintDuplicateQuestions,ap,
26
- TwitterSemEval2015,ap,
27
- TwitterURLCorpus,ap,
28
- AskUbuntuDupQuestions,map,
29
- MindSmallReranking,map,
30
- SciDocsRR,map,
31
- StackOverflowDupQuestions,map,
32
- ArguAna,ndcg_at_10,
33
- ClimateFEVER,ndcg_at_10,
34
- CQADupstackRetrieval,ndcg_at_10,
35
- DBPedia,ndcg_at_10,
36
- FEVER,ndcg_at_10,
37
- FiQA2018,ndcg_at_10,
38
- HotpotQA,ndcg_at_10,
39
- MSMARCO,ndcg_at_10,
40
- NFCorpus,ndcg_at_10,
41
- NQ,ndcg_at_10,
42
- QuoraRetrieval,ndcg_at_10,
43
- SCIDOCS,ndcg_at_10,
44
- SciFact,ndcg_at_10,
45
- Touche2020,ndcg_at_10,
46
- TRECCOVID,ndcg_at_10,
47
- BIOSSES,cosine_spearman,
48
- SICK-R,cosine_spearman,
49
- STS12,cosine_spearman,
50
- STS13,cosine_spearman,
51
- STS14,cosine_spearman,
52
- STS15,cosine_spearman,
53
- STS16,cosine_spearman,
54
- STS17,cosine_spearman,
55
- STS22,cosine_spearman,
56
- STSBenchmark,cosine_spearman,
57
- SummEval,cosine_spearman,
 
1
+ model,dataset,metric,value
2
+ sgpt-bloom-7b1-msmarco,AmazonCounterfactualClassification,accuracy,
3
+ sgpt-bloom-7b1-msmarco,AmazonPolarityClassification,accuracy,
4
+ sgpt-bloom-7b1-msmarco,AmazonReviewsClassification,accuracy,
5
+ sgpt-bloom-7b1-msmarco,Banking77Classification,accuracy,
6
+ sgpt-bloom-7b1-msmarco,EmotionClassification,accuracy,
7
+ sgpt-bloom-7b1-msmarco,ImdbClassification,accuracy,
8
+ sgpt-bloom-7b1-msmarco,MassiveIntentClassification,accuracy,
9
+ sgpt-bloom-7b1-msmarco,MassiveScenarioClassification,accuracy,
10
+ sgpt-bloom-7b1-msmarco,MTOPDomainClassification,accuracy,
11
+ sgpt-bloom-7b1-msmarco,MTOPIntentClassification,accuracy,
12
+ sgpt-bloom-7b1-msmarco,ToxicConversationsClassification,accuracy,
13
+ sgpt-bloom-7b1-msmarco,TweetSentimentExtractionClassification,accuracy,
14
+ sgpt-bloom-7b1-msmarco,ArxivClusteringP2P,v_measure,
15
+ sgpt-bloom-7b1-msmarco,ArxivClusteringS2S,v_measure,
16
+ sgpt-bloom-7b1-msmarco,BiorxivClusteringP2P,v_measure,
17
+ sgpt-bloom-7b1-msmarco,BiorxivClusteringS2S,v_measure,
18
+ sgpt-bloom-7b1-msmarco,MedrxivClusteringP2P,v_measure,
19
+ sgpt-bloom-7b1-msmarco,MedrxivClusteringS2S,v_measure,
20
+ sgpt-bloom-7b1-msmarco,RedditClustering,v_measure,
21
+ sgpt-bloom-7b1-msmarco,RedditClusteringP2P,v_measure,
22
+ sgpt-bloom-7b1-msmarco,StackExchangeClustering,v_measure,
23
+ sgpt-bloom-7b1-msmarco,StackExchangeClusteringP2P,v_measure,
24
+ sgpt-bloom-7b1-msmarco,TwentyNewsgroupsClustering,v_measure,
25
+ sgpt-bloom-7b1-msmarco,SprintDuplicateQuestions,ap,
26
+ sgpt-bloom-7b1-msmarco,TwitterSemEval2015,ap,
27
+ sgpt-bloom-7b1-msmarco,TwitterURLCorpus,ap,
28
+ sgpt-bloom-7b1-msmarco,AskUbuntuDupQuestions,map,
29
+ sgpt-bloom-7b1-msmarco,MindSmallReranking,map,
30
+ sgpt-bloom-7b1-msmarco,SciDocsRR,map,
31
+ sgpt-bloom-7b1-msmarco,StackOverflowDupQuestions,map,
32
+ sgpt-bloom-7b1-msmarco,ArguAna,ndcg_at_10,
33
+ sgpt-bloom-7b1-msmarco,ClimateFEVER,ndcg_at_10,
34
+ sgpt-bloom-7b1-msmarco,CQADupstackRetrieval,ndcg_at_10,
35
+ sgpt-bloom-7b1-msmarco,DBPedia,ndcg_at_10,
36
+ sgpt-bloom-7b1-msmarco,FEVER,ndcg_at_10,
37
+ sgpt-bloom-7b1-msmarco,FiQA2018,ndcg_at_10,
38
+ sgpt-bloom-7b1-msmarco,HotpotQA,ndcg_at_10,
39
+ sgpt-bloom-7b1-msmarco,MSMARCO,ndcg_at_10,
40
+ sgpt-bloom-7b1-msmarco,NFCorpus,ndcg_at_10,
41
+ sgpt-bloom-7b1-msmarco,NQ,ndcg_at_10,
42
+ sgpt-bloom-7b1-msmarco,QuoraRetrieval,ndcg_at_10,
43
+ sgpt-bloom-7b1-msmarco,SCIDOCS,ndcg_at_10,
44
+ sgpt-bloom-7b1-msmarco,SciFact,ndcg_at_10,
45
+ sgpt-bloom-7b1-msmarco,Touche2020,ndcg_at_10,
46
+ sgpt-bloom-7b1-msmarco,TRECCOVID,ndcg_at_10,
47
+ sgpt-bloom-7b1-msmarco,BIOSSES,cosine_spearman,
48
+ sgpt-bloom-7b1-msmarco,SICK-R,cosine_spearman,
49
+ sgpt-bloom-7b1-msmarco,STS12,cosine_spearman,
50
+ sgpt-bloom-7b1-msmarco,STS13,cosine_spearman,
51
+ sgpt-bloom-7b1-msmarco,STS14,cosine_spearman,
52
+ sgpt-bloom-7b1-msmarco,STS15,cosine_spearman,
53
+ sgpt-bloom-7b1-msmarco,STS16,cosine_spearman,
54
+ sgpt-bloom-7b1-msmarco,STS17,cosine_spearman,
55
+ sgpt-bloom-7b1-msmarco,STS22,cosine_spearman,
56
+ sgpt-bloom-7b1-msmarco,STSBenchmark,cosine_spearman,
57
+ sgpt-bloom-7b1-msmarco,SummEval,cosine_spearman,
sup-simcse-bert-base-uncased_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.7574626865671641
3
- AmazonPolarityClassification,accuracy,0.8247332500000001
4
- AmazonReviewsClassification,accuracy,0.39598000000000005
5
- Banking77Classification,accuracy,0.7575974025974027
6
- EmotionClassification,accuracy,0.4481
7
- ImdbClassification,accuracy,0.7353120000000001
8
- MassiveIntentClassification,accuracy,0.6595158036314727
9
- MassiveScenarioClassification,accuracy,0.7078345662407532
10
- MTOPDomainClassification,accuracy,0.8428636570907433
11
- MTOPIntentClassification,accuracy,0.6314409484724123
12
- ToxicConversationsClassification,accuracy,0.720444
13
- TweetSentimentExtractionClassification,accuracy,0.5973401245048104
14
- ArxivClusteringP2P,v_measure,0.35178847292844123
15
- ArxivClusteringS2S,v_measure,0.2753955423760812
16
- BiorxivClusteringP2P,v_measure,0.3015030662227709
17
- BiorxivClusteringS2S,v_measure,0.24667587465067067
18
- MedrxivClusteringP2P,v_measure,0.2624813884919727
19
- MedrxivClusteringS2S,v_measure,0.24119581645899563
20
- RedditClustering,v_measure,0.4022796798210619
21
- RedditClusteringP2P,v_measure,0.4773761279854133
22
- StackExchangeClustering,v_measure,0.4755177976262692
23
- StackExchangeClusteringP2P,v_measure,0.29453502158512795
24
- TwentyNewsgroupsClustering,v_measure,0.3486285218255636
25
- SprintDuplicateQuestions,ap,0.6938793582706408
26
- TwitterSemEval2015,ap,0.6775361224485238
27
- TwitterURLCorpus,ap,0.8388869258620997
28
- AskUbuntuDupQuestions,map,0.5180158192605682
29
- MindSmallReranking,map,0.2929834424526365
30
- SciDocsRR,map,0.7013643643529786
31
- StackOverflowDupQuestions,map,0.3890480943394549
32
- ArguAna,ndcg_at_10,0.38331
33
- ClimateFEVER,ndcg_at_10,0.11985
34
- CQADupstackRetrieval,ndcg_at_10,0.1449675
35
- DBPedia,ndcg_at_10,0.19729
36
- FEVER,ndcg_at_10,0.20411
37
- FiQA2018,ndcg_at_10,0.1041
38
- HotpotQA,ndcg_at_10,0.22895
39
- MSMARCO,ndcg_at_10,0.25193
40
- NFCorpus,ndcg_at_10,0.12418
41
- NQ,ndcg_at_10,0.1608
42
- QuoraRetrieval,ndcg_at_10,0.79621
43
- SCIDOCS,ndcg_at_10,0.07534
44
- SciFact,ndcg_at_10,0.29593
45
- Touche2020,ndcg_at_10,0.09895
46
- TRECCOVID,ndcg_at_10,0.22929
47
- BIOSSES,cosine_spearman,0.6837806406467221
48
- SICK-R,cosine_spearman,0.8077099458895838
49
- STS12,cosine_spearman,0.7529685124437535
50
- STS13,cosine_spearman,0.8466699046070492
51
- STS14,cosine_spearman,0.8018946888447316
52
- STS15,cosine_spearman,0.8539881217395949
53
- STS16,cosine_spearman,0.8081920987955081
54
- STS17,cosine_spearman,0.8943673830561076
55
- STS22,cosine_spearman,0.6195505621412989
56
- STSBenchmark,cosine_spearman,0.8424697229482505
57
- SummEval,cosine_spearman,0.23308772096174835
 
1
+ model,dataset,metric,value
2
+ sup-simcse-bert-base-uncased,AmazonCounterfactualClassification,accuracy,0.7574626865671641
3
+ sup-simcse-bert-base-uncased,AmazonPolarityClassification,accuracy,0.8247332500000001
4
+ sup-simcse-bert-base-uncased,AmazonReviewsClassification,accuracy,0.39598000000000005
5
+ sup-simcse-bert-base-uncased,Banking77Classification,accuracy,0.7575974025974027
6
+ sup-simcse-bert-base-uncased,EmotionClassification,accuracy,0.4481
7
+ sup-simcse-bert-base-uncased,ImdbClassification,accuracy,0.7353120000000001
8
+ sup-simcse-bert-base-uncased,MassiveIntentClassification,accuracy,0.6595158036314727
9
+ sup-simcse-bert-base-uncased,MassiveScenarioClassification,accuracy,0.7078345662407532
10
+ sup-simcse-bert-base-uncased,MTOPDomainClassification,accuracy,0.8428636570907433
11
+ sup-simcse-bert-base-uncased,MTOPIntentClassification,accuracy,0.6314409484724123
12
+ sup-simcse-bert-base-uncased,ToxicConversationsClassification,accuracy,0.720444
13
+ sup-simcse-bert-base-uncased,TweetSentimentExtractionClassification,accuracy,0.5973401245048104
14
+ sup-simcse-bert-base-uncased,ArxivClusteringP2P,v_measure,0.35178847292844123
15
+ sup-simcse-bert-base-uncased,ArxivClusteringS2S,v_measure,0.2753955423760812
16
+ sup-simcse-bert-base-uncased,BiorxivClusteringP2P,v_measure,0.3015030662227709
17
+ sup-simcse-bert-base-uncased,BiorxivClusteringS2S,v_measure,0.24667587465067067
18
+ sup-simcse-bert-base-uncased,MedrxivClusteringP2P,v_measure,0.2624813884919727
19
+ sup-simcse-bert-base-uncased,MedrxivClusteringS2S,v_measure,0.24119581645899563
20
+ sup-simcse-bert-base-uncased,RedditClustering,v_measure,0.4022796798210619
21
+ sup-simcse-bert-base-uncased,RedditClusteringP2P,v_measure,0.4773761279854133
22
+ sup-simcse-bert-base-uncased,StackExchangeClustering,v_measure,0.4755177976262692
23
+ sup-simcse-bert-base-uncased,StackExchangeClusteringP2P,v_measure,0.29453502158512795
24
+ sup-simcse-bert-base-uncased,TwentyNewsgroupsClustering,v_measure,0.3486285218255636
25
+ sup-simcse-bert-base-uncased,SprintDuplicateQuestions,ap,0.6938793582706408
26
+ sup-simcse-bert-base-uncased,TwitterSemEval2015,ap,0.6775361224485238
27
+ sup-simcse-bert-base-uncased,TwitterURLCorpus,ap,0.8388869258620997
28
+ sup-simcse-bert-base-uncased,AskUbuntuDupQuestions,map,0.5180158192605682
29
+ sup-simcse-bert-base-uncased,MindSmallReranking,map,0.2929834424526365
30
+ sup-simcse-bert-base-uncased,SciDocsRR,map,0.7013643643529786
31
+ sup-simcse-bert-base-uncased,StackOverflowDupQuestions,map,0.3890480943394549
32
+ sup-simcse-bert-base-uncased,ArguAna,ndcg_at_10,0.38331
33
+ sup-simcse-bert-base-uncased,ClimateFEVER,ndcg_at_10,0.11985
34
+ sup-simcse-bert-base-uncased,CQADupstackRetrieval,ndcg_at_10,0.1449675
35
+ sup-simcse-bert-base-uncased,DBPedia,ndcg_at_10,0.19729
36
+ sup-simcse-bert-base-uncased,FEVER,ndcg_at_10,0.20411
37
+ sup-simcse-bert-base-uncased,FiQA2018,ndcg_at_10,0.1041
38
+ sup-simcse-bert-base-uncased,HotpotQA,ndcg_at_10,0.22895
39
+ sup-simcse-bert-base-uncased,MSMARCO,ndcg_at_10,0.25193
40
+ sup-simcse-bert-base-uncased,NFCorpus,ndcg_at_10,0.12418
41
+ sup-simcse-bert-base-uncased,NQ,ndcg_at_10,0.1608
42
+ sup-simcse-bert-base-uncased,QuoraRetrieval,ndcg_at_10,0.79621
43
+ sup-simcse-bert-base-uncased,SCIDOCS,ndcg_at_10,0.07534
44
+ sup-simcse-bert-base-uncased,SciFact,ndcg_at_10,0.29593
45
+ sup-simcse-bert-base-uncased,Touche2020,ndcg_at_10,0.09895
46
+ sup-simcse-bert-base-uncased,TRECCOVID,ndcg_at_10,0.22929
47
+ sup-simcse-bert-base-uncased,BIOSSES,cosine_spearman,0.6837806406467221
48
+ sup-simcse-bert-base-uncased,SICK-R,cosine_spearman,0.8077099458895838
49
+ sup-simcse-bert-base-uncased,STS12,cosine_spearman,0.7529685124437535
50
+ sup-simcse-bert-base-uncased,STS13,cosine_spearman,0.8466699046070492
51
+ sup-simcse-bert-base-uncased,STS14,cosine_spearman,0.8018946888447316
52
+ sup-simcse-bert-base-uncased,STS15,cosine_spearman,0.8539881217395949
53
+ sup-simcse-bert-base-uncased,STS16,cosine_spearman,0.8081920987955081
54
+ sup-simcse-bert-base-uncased,STS17,cosine_spearman,0.8943673830561076
55
+ sup-simcse-bert-base-uncased,STS22,cosine_spearman,0.6195505621412989
56
+ sup-simcse-bert-base-uncased,STSBenchmark,cosine_spearman,0.8424697229482505
57
+ sup-simcse-bert-base-uncased,SummEval,cosine_spearman,0.23308772096174835
unsup-simcse-bert-base-uncased_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.6708955223880597
3
- AmazonPolarityClassification,accuracy,0.7447925
4
- AmazonReviewsClassification,accuracy,0.33846000000000004
5
- Banking77Classification,accuracy,0.735487012987013
6
- EmotionClassification,accuracy,0.42219999999999996
7
- ImdbClassification,accuracy,0.696284
8
- MassiveIntentClassification,accuracy,0.5983860121049093
9
- MassiveScenarioClassification,accuracy,0.662542030934768
10
- MTOPDomainClassification,accuracy,0.817122663018696
11
- MTOPIntentClassification,accuracy,0.5922708618331054
12
- ToxicConversationsClassification,accuracy,0.688172
13
- TweetSentimentExtractionClassification,accuracy,0.5335880022637238
14
- ArxivClusteringP2P,v_measure,0.3260699748580497
15
- ArxivClusteringS2S,v_measure,0.2468226709235907
16
- BiorxivClusteringP2P,v_measure,0.24902697594424988
17
- BiorxivClusteringS2S,v_measure,0.1955213639220657
18
- MedrxivClusteringP2P,v_measure,0.23602716067058646
19
- MedrxivClusteringS2S,v_measure,0.21973507645898568
20
- RedditClustering,v_measure,0.32177188380770716
21
- RedditClusteringP2P,v_measure,0.4513625539595167
22
- StackExchangeClustering,v_measure,0.43070159841748956
23
- StackExchangeClusteringP2P,v_measure,0.2850446706405736
24
- TwentyNewsgroupsClustering,v_measure,0.23205752453237888
25
- SprintDuplicateQuestions,ap,0.6940657828591321
26
- TwitterSemEval2015,ap,0.6020770691133284
27
- TwitterURLCorpus,ap,0.8136867122131795
28
- AskUbuntuDupQuestions,map,0.5156997143788069
29
- MindSmallReranking,map,0.2861705451549499
30
- SciDocsRR,map,0.6633446765426328
31
- StackOverflowDupQuestions,map,0.39352125808339417
32
- ArguAna,ndcg_at_10,0.38343
33
- ClimateFEVER,ndcg_at_10,0.11799
34
- CQADupstackRetrieval,ndcg_at_10,0.13222750000000003
35
- DBPedia,ndcg_at_10,0.15041
36
- FEVER,ndcg_at_10,0.21055
37
- FiQA2018,ndcg_at_10,0.09836
38
- HotpotQA,ndcg_at_10,0.19752
39
- MSMARCO,ndcg_at_10,0.22566
40
- NFCorpus,ndcg_at_10,0.09879
41
- NQ,ndcg_at_10,0.11692
42
- QuoraRetrieval,ndcg_at_10,0.78031
43
- SCIDOCS,ndcg_at_10,0.05496
44
- SciFact,ndcg_at_10,0.25716
45
- Touche2020,ndcg_at_10,0.08897
46
- TRECCOVID,ndcg_at_10,0.26196
47
- BIOSSES,cosine_spearman,0.7231128621592763
48
- SICK-R,cosine_spearman,0.7224371728496006
49
- STS12,cosine_spearman,0.6604669141177121
50
- STS13,cosine_spearman,0.8148757487870733
51
- STS14,cosine_spearman,0.7360735790527075
52
- STS15,cosine_spearman,0.7972498368805553
53
- STS16,cosine_spearman,0.7812450491845039
54
- STS17,cosine_spearman,0.8357872706551984
55
- STS22,cosine_spearman,0.5965048713602169
56
- STSBenchmark,cosine_spearman,0.7651512563863317
57
- SummEval,cosine_spearman,0.3114893410144058
 
1
+ model,dataset,metric,value
2
+ unsup-simcse-bert-base-uncased,AmazonCounterfactualClassification,accuracy,0.6708955223880597
3
+ unsup-simcse-bert-base-uncased,AmazonPolarityClassification,accuracy,0.7447925
4
+ unsup-simcse-bert-base-uncased,AmazonReviewsClassification,accuracy,0.33846000000000004
5
+ unsup-simcse-bert-base-uncased,Banking77Classification,accuracy,0.735487012987013
6
+ unsup-simcse-bert-base-uncased,EmotionClassification,accuracy,0.42219999999999996
7
+ unsup-simcse-bert-base-uncased,ImdbClassification,accuracy,0.696284
8
+ unsup-simcse-bert-base-uncased,MassiveIntentClassification,accuracy,0.5983860121049093
9
+ unsup-simcse-bert-base-uncased,MassiveScenarioClassification,accuracy,0.662542030934768
10
+ unsup-simcse-bert-base-uncased,MTOPDomainClassification,accuracy,0.817122663018696
11
+ unsup-simcse-bert-base-uncased,MTOPIntentClassification,accuracy,0.5922708618331054
12
+ unsup-simcse-bert-base-uncased,ToxicConversationsClassification,accuracy,0.688172
13
+ unsup-simcse-bert-base-uncased,TweetSentimentExtractionClassification,accuracy,0.5335880022637238
14
+ unsup-simcse-bert-base-uncased,ArxivClusteringP2P,v_measure,0.3260699748580497
15
+ unsup-simcse-bert-base-uncased,ArxivClusteringS2S,v_measure,0.2468226709235907
16
+ unsup-simcse-bert-base-uncased,BiorxivClusteringP2P,v_measure,0.24902697594424988
17
+ unsup-simcse-bert-base-uncased,BiorxivClusteringS2S,v_measure,0.1955213639220657
18
+ unsup-simcse-bert-base-uncased,MedrxivClusteringP2P,v_measure,0.23602716067058646
19
+ unsup-simcse-bert-base-uncased,MedrxivClusteringS2S,v_measure,0.21973507645898568
20
+ unsup-simcse-bert-base-uncased,RedditClustering,v_measure,0.32177188380770716
21
+ unsup-simcse-bert-base-uncased,RedditClusteringP2P,v_measure,0.4513625539595167
22
+ unsup-simcse-bert-base-uncased,StackExchangeClustering,v_measure,0.43070159841748956
23
+ unsup-simcse-bert-base-uncased,StackExchangeClusteringP2P,v_measure,0.2850446706405736
24
+ unsup-simcse-bert-base-uncased,TwentyNewsgroupsClustering,v_measure,0.23205752453237888
25
+ unsup-simcse-bert-base-uncased,SprintDuplicateQuestions,ap,0.6940657828591321
26
+ unsup-simcse-bert-base-uncased,TwitterSemEval2015,ap,0.6020770691133284
27
+ unsup-simcse-bert-base-uncased,TwitterURLCorpus,ap,0.8136867122131795
28
+ unsup-simcse-bert-base-uncased,AskUbuntuDupQuestions,map,0.5156997143788069
29
+ unsup-simcse-bert-base-uncased,MindSmallReranking,map,0.2861705451549499
30
+ unsup-simcse-bert-base-uncased,SciDocsRR,map,0.6633446765426328
31
+ unsup-simcse-bert-base-uncased,StackOverflowDupQuestions,map,0.39352125808339417
32
+ unsup-simcse-bert-base-uncased,ArguAna,ndcg_at_10,0.38343
33
+ unsup-simcse-bert-base-uncased,ClimateFEVER,ndcg_at_10,0.11799
34
+ unsup-simcse-bert-base-uncased,CQADupstackRetrieval,ndcg_at_10,0.13222750000000003
35
+ unsup-simcse-bert-base-uncased,DBPedia,ndcg_at_10,0.15041
36
+ unsup-simcse-bert-base-uncased,FEVER,ndcg_at_10,0.21055
37
+ unsup-simcse-bert-base-uncased,FiQA2018,ndcg_at_10,0.09836
38
+ unsup-simcse-bert-base-uncased,HotpotQA,ndcg_at_10,0.19752
39
+ unsup-simcse-bert-base-uncased,MSMARCO,ndcg_at_10,0.22566
40
+ unsup-simcse-bert-base-uncased,NFCorpus,ndcg_at_10,0.09879
41
+ unsup-simcse-bert-base-uncased,NQ,ndcg_at_10,0.11692
42
+ unsup-simcse-bert-base-uncased,QuoraRetrieval,ndcg_at_10,0.78031
43
+ unsup-simcse-bert-base-uncased,SCIDOCS,ndcg_at_10,0.05496
44
+ unsup-simcse-bert-base-uncased,SciFact,ndcg_at_10,0.25716
45
+ unsup-simcse-bert-base-uncased,Touche2020,ndcg_at_10,0.08897
46
+ unsup-simcse-bert-base-uncased,TRECCOVID,ndcg_at_10,0.26196
47
+ unsup-simcse-bert-base-uncased,BIOSSES,cosine_spearman,0.7231128621592763
48
+ unsup-simcse-bert-base-uncased,SICK-R,cosine_spearman,0.7224371728496006
49
+ unsup-simcse-bert-base-uncased,STS12,cosine_spearman,0.6604669141177121
50
+ unsup-simcse-bert-base-uncased,STS13,cosine_spearman,0.8148757487870733
51
+ unsup-simcse-bert-base-uncased,STS14,cosine_spearman,0.7360735790527075
52
+ unsup-simcse-bert-base-uncased,STS15,cosine_spearman,0.7972498368805553
53
+ unsup-simcse-bert-base-uncased,STS16,cosine_spearman,0.7812450491845039
54
+ unsup-simcse-bert-base-uncased,STS17,cosine_spearman,0.8357872706551984
55
+ unsup-simcse-bert-base-uncased,STS22,cosine_spearman,0.5965048713602169
56
+ unsup-simcse-bert-base-uncased,STSBenchmark,cosine_spearman,0.7651512563863317
57
+ unsup-simcse-bert-base-uncased,SummEval,cosine_spearman,0.3114893410144058